2025-11-03T16:56:52.1064368Z Current runner version: '2.329.0' 2025-11-03T16:56:52.1070207Z Runner name: 'i-05b4d4af410353347' 2025-11-03T16:56:52.1071190Z Runner group name: 'default' 2025-11-03T16:56:52.1072015Z Machine name: 'ip-10-0-8-115' 2025-11-03T16:56:52.1074661Z ##[group]GITHUB_TOKEN Permissions 2025-11-03T16:56:52.1076799Z Contents: read 2025-11-03T16:56:52.1077355Z Metadata: read 2025-11-03T16:56:52.1078055Z ##[endgroup] 2025-11-03T16:56:52.1079922Z Secret source: Actions 2025-11-03T16:56:52.1080880Z Prepare workflow directory 2025-11-03T16:56:52.1552479Z Prepare all required actions 2025-11-03T16:56:52.1586830Z Getting action download info 2025-11-03T16:56:52.5361681Z Download action repository 'pytorch/test-infra@main' (SHA:62d0084d5e2a6926f8884d6714e7fe289eb96af5) 2025-11-03T16:56:55.0356526Z Download action repository 'pytorch/pytorch@main' (SHA:104b86861862ccc47f6df6528347a017bef813ad) 2025-11-03T16:57:11.8407943Z Download action repository 'actions/setup-python@a26af69be951a213d495a4c3e4e4022e16d87065' (SHA:a26af69be951a213d495a4c3e4e4022e16d87065) 2025-11-03T16:57:12.2274195Z Download action repository 'aws-actions/configure-aws-credentials@ececac1a45f3b08a01d2dd070d28d111c5fe6722' (SHA:ececac1a45f3b08a01d2dd070d28d111c5fe6722) 2025-11-03T16:57:12.4858117Z Download action repository 'aws-actions/amazon-ecr-login@062b18b96a7aff071d4dc91bc00c4c1a7945b076' (SHA:062b18b96a7aff071d4dc91bc00c4c1a7945b076) 2025-11-03T16:57:12.6739256Z Download action repository 'seemethere/download-artifact-s3@1da556a7aa0a088e3153970611f6c432d58e80e6' (SHA:1da556a7aa0a088e3153970611f6c432d58e80e6) 2025-11-03T16:57:12.9095059Z Download action repository 'seemethere/upload-artifact-s3@baba72d0712b404f646cebe0730933554ebce96a' (SHA:baba72d0712b404f646cebe0730933554ebce96a) 2025-11-03T16:57:13.2213102Z Getting action download info 2025-11-03T16:57:13.3651174Z Download action repository 'actions/checkout@v4' (SHA:08eba0b27e820071cde6df949e0beb9ba4906955) 2025-11-03T16:57:13.6564570Z Getting action download info 2025-11-03T16:57:13.7772706Z Download action repository 'nick-fields/retry@v3.0.0' (SHA:7152eba30c6575329ac0576536151aca5a72780e) 2025-11-03T16:57:13.9723962Z Getting action download info 2025-11-03T16:57:14.1341096Z Download action repository 'nick-fields/retry@3e91a01664abd3c5cd539100d10d33b9c5b68482' (SHA:3e91a01664abd3c5cd539100d10d33b9c5b68482) 2025-11-03T16:57:14.3081614Z Getting action download info 2025-11-03T16:57:14.4784494Z Uses: pytorch/pytorch/.github/workflows/_linux-test.yml@refs/heads/main (3f6538febd727b782e6e13cfd026a309fb14351d) 2025-11-03T16:57:14.4788567Z ##[group] Inputs 2025-11-03T16:57:14.4789072Z build-environment: linux-jammy-cuda12.8-py3.12-gcc11 2025-11-03T16:57:14.4794230Z test-matrix: {"include": [{"config": "vllm_basic_correctness_test", "shard": 1, "num_shards": 1, "runner": "linux.g6.4xlarge.experimental.nvidia.gpu"}, {"config": "vllm_basic_models_test", "shard": 1, "num_shards": 1, "runner": "linux.g6.4xlarge.experimental.nvidia.gpu"}, {"config": "vllm_entrypoints_test", "shard": 1, "num_shards": 1, "runner": "linux.g6.4xlarge.experimental.nvidia.gpu"}, {"config": "vllm_regression_test", "shard": 1, "num_shards": 1, "runner": "linux.g6.4xlarge.experimental.nvidia.gpu"}, {"config": "vllm_multi_model_processor_test", "shard": 1, "num_shards": 1, "runner": "linux.g6.4xlarge.experimental.nvidia.gpu"}, {"config": "vllm_pytorch_compilation_unit_tests", "shard": 1, "num_shards": 1, "runner": "linux.g6.4xlarge.experimental.nvidia.gpu"}, {"config": "vllm_lora_28_failure_test", "shard": 1, "num_shards": 1, "runner": "linux.g6.4xlarge.experimental.nvidia.gpu"}, {"config": "vllm_multi_model_test_28_failure_test", "shard": 1, "num_shards": 1, "runner": "linux.g6.4xlarge.experimental.nvidia.gpu"}, {"config": "vllm_language_model_test_extended_generation_28_failure_test", "shard": 1, "num_shards": 1, "runner": "linux.g6.4xlarge.experimental.nvidia.gpu"}, {"config": "vllm_distributed_test_2_gpu_28_failure_test", "shard": 1, "num_shards": 1, "runner": "linux.g6.4xlarge.experimental.nvidia.gpu"}, {"config": "vllm_lora_test", "shard": 0, "num_shards": 4, "runner": "linux.g6.4xlarge.experimental.nvidia.gpu"}, {"config": "vllm_lora_test", "shard": 1, "num_shards": 4, "runner": "linux.g6.4xlarge.experimental.nvidia.gpu"}, {"config": "vllm_lora_test", "shard": 2, "num_shards": 4, "runner": "linux.g6.4xlarge.experimental.nvidia.gpu"}, {"config": "vllm_lora_test", "shard": 3, "num_shards": 4, "runner": "linux.g6.4xlarge.experimental.nvidia.gpu"}, {"config": "vllm_lora_tp_test_distributed", "shard": 1, "num_shards": 1, "runner": "linux.g6.12xlarge.nvidia.gpu"}, {"config": "vllm_distributed_test_28_failure_test", "shard": 1, "num_shards": 1, "runner": "linux.g6.12xlarge.nvidia.gpu"}]} 2025-11-03T16:57:14.4800028Z docker-image: 308535385114.dkr.ecr.us-east-1.amazonaws.com/pytorch/ci-image:pytorch-linux-jammy-cuda12.8-cudnn9-py3.12-gcc11-vllm-30147fb242f88ab8f13e14b76ab186d189d8605c 2025-11-03T16:57:14.4816481Z sync-tag: 2025-11-03T16:57:14.4817395Z timeout-minutes: 240 2025-11-03T16:57:14.4817628Z use-gha: 2025-11-03T16:57:14.4817812Z dashboard-tag: 2025-11-03T16:57:14.4818008Z s3-bucket: gha-artifacts 2025-11-03T16:57:14.4818220Z aws-role-to-assume: 2025-11-03T16:57:14.4818759Z disable-monitor: false 2025-11-03T16:57:14.4819017Z monitor-log-interval: 5 2025-11-03T16:57:14.4819257Z monitor-data-collect-interval: 1 2025-11-03T16:57:14.4819505Z ##[endgroup] 2025-11-03T16:57:14.4819899Z Complete job name: ci-vllm-test / test (vllm_distributed_test_28_failure_test, 1, 1, linux.g6.12xlarge.nvidia.gpu) 2025-11-03T16:57:14.5662745Z A job started hook has been configured by the self-hosted runner administrator 2025-11-03T16:57:14.5758186Z ##[group]Run '/home/ec2-user/runner-scripts/before_job.sh' 2025-11-03T16:57:14.5768784Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2025-11-03T16:57:14.5769400Z ##[endgroup] 2025-11-03T16:57:16.0161422Z Runner Type: linux.g6.12xlarge.nvidia.gpu 2025-11-03T16:57:16.0161947Z Instance Type: g6.12xlarge 2025-11-03T16:57:16.0162151Z AMI Name: unknown 2025-11-03T16:57:16.0207815Z AMI ID: ami-08982f1c5bf93d976 2025-11-03T16:57:21.1362674Z ##[group]Run pytorch/test-infra/.github/actions/setup-ssh@main 2025-11-03T16:57:21.1363025Z with: 2025-11-03T16:57:21.1363494Z github-secret: *** 2025-11-03T16:57:21.1364015Z instructions: All testing is done inside the container, to start an interactive session run: docker exec -it $(docker container ps --format '{{.ID}}') bash 2025-11-03T16:57:21.1364555Z activate-with-label: false 2025-11-03T16:57:21.1364751Z label: with-ssh 2025-11-03T16:57:21.1364934Z remove-existing-keys: true 2025-11-03T16:57:21.1365141Z fail-silently: true 2025-11-03T16:57:21.1365317Z env: 2025-11-03T16:57:21.1365463Z GIT_DEFAULT_BRANCH: main 2025-11-03T16:57:21.1365662Z ##[endgroup] 2025-11-03T16:57:21.2773916Z Please see https://github.com/pytorch/pytorch/wiki/Debugging-using-with-ssh-for-Github-Actions for more info. 2025-11-03T16:57:21.2775105Z Not on pull request and ciflow reference could not be extracted, skipping adding ssh keys 2025-11-03T16:57:21.2947296Z ##[group]Run pytorch/pytorch/.github/actions/checkout-pytorch@main 2025-11-03T16:57:21.2947688Z with: 2025-11-03T16:57:21.2947849Z no-sudo: true 2025-11-03T16:57:21.2948029Z submodules: recursive 2025-11-03T16:57:21.2948230Z fetch-depth: 0 2025-11-03T16:57:21.2948415Z env: 2025-11-03T16:57:21.2948568Z GIT_DEFAULT_BRANCH: main 2025-11-03T16:57:21.2948758Z ##[endgroup] 2025-11-03T16:57:21.3050922Z ##[group]Run echo "IN_CONTAINER_RUNNER=$(if [ -f /.inarc ] || [ -f /.incontainer ]; then echo true ; else echo false; fi)" >> "$GITHUB_OUTPUT" 2025-11-03T16:57:21.3051643Z echo "IN_CONTAINER_RUNNER=$(if [ -f /.inarc ] || [ -f /.incontainer ]; then echo true ; else echo false; fi)" >> "$GITHUB_OUTPUT" 2025-11-03T16:57:21.3064178Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2025-11-03T16:57:21.3064466Z env: 2025-11-03T16:57:21.3064665Z GIT_DEFAULT_BRANCH: main 2025-11-03T16:57:21.3064881Z ##[endgroup] 2025-11-03T16:57:21.3171744Z ##[group]Run # Use all available CPUs for fetching 2025-11-03T16:57:21.3172372Z # Use all available CPUs for fetching 2025-11-03T16:57:21.3172652Z cd "${GITHUB_WORKSPACE}" 2025-11-03T16:57:21.3172917Z git config --global fetch.parallel 0 2025-11-03T16:57:21.3173221Z git config --global submodule.fetchJobs 0 2025-11-03T16:57:21.3173486Z  2025-11-03T16:57:21.3173828Z # Clean workspace. The default checkout action should also do this, but 2025-11-03T16:57:21.3174238Z # do it here as well just in case 2025-11-03T16:57:21.3174484Z if [[ -d .git ]]; then 2025-11-03T16:57:21.3174706Z  if [ -z "${NO_SUDO}" ]; then 2025-11-03T16:57:21.3174951Z  sudo git clean -ffdx 2025-11-03T16:57:21.3175165Z  else 2025-11-03T16:57:21.3175336Z  git clean -ffdx 2025-11-03T16:57:21.3175538Z  fi 2025-11-03T16:57:21.3175699Z fi 2025-11-03T16:57:21.3183953Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2025-11-03T16:57:21.3184262Z env: 2025-11-03T16:57:21.3184435Z GIT_DEFAULT_BRANCH: main 2025-11-03T16:57:21.3184641Z NO_SUDO: true 2025-11-03T16:57:21.3184810Z ##[endgroup] 2025-11-03T16:57:21.3331328Z ##[group]Run actions/checkout@v4 2025-11-03T16:57:21.3331542Z with: 2025-11-03T16:57:21.3331729Z ref: 3f6538febd727b782e6e13cfd026a309fb14351d 2025-11-03T16:57:21.3331978Z fetch-depth: 0 2025-11-03T16:57:21.3332155Z submodules: recursive 2025-11-03T16:57:21.3332343Z show-progress: false 2025-11-03T16:57:21.3332547Z repository: pytorch/pytorch 2025-11-03T16:57:21.3332870Z token: *** 2025-11-03T16:57:21.3333040Z ssh-strict: true 2025-11-03T16:57:21.3333206Z ssh-user: git 2025-11-03T16:57:21.3333387Z persist-credentials: true 2025-11-03T16:57:21.3333587Z clean: true 2025-11-03T16:57:21.3333773Z sparse-checkout-cone-mode: true 2025-11-03T16:57:21.3333986Z fetch-tags: false 2025-11-03T16:57:21.3334155Z lfs: false 2025-11-03T16:57:21.3334323Z set-safe-directory: true 2025-11-03T16:57:21.3334532Z env: 2025-11-03T16:57:21.3334694Z GIT_DEFAULT_BRANCH: main 2025-11-03T16:57:21.3334879Z ##[endgroup] 2025-11-03T16:57:21.4351903Z Syncing repository: pytorch/pytorch 2025-11-03T16:57:21.4353084Z ##[group]Getting Git version info 2025-11-03T16:57:21.4353483Z Working directory is '/home/ec2-user/actions-runner/_work/pytorch/pytorch' 2025-11-03T16:57:21.4354019Z [command]/usr/bin/git version 2025-11-03T16:57:21.4549682Z git version 2.50.1 2025-11-03T16:57:21.4610313Z ##[endgroup] 2025-11-03T16:57:21.4620763Z Copying '/home/ec2-user/.gitconfig' to '/home/ec2-user/actions-runner/_work/_temp/a47662a8-f8a3-43a9-a630-e4d910a68a68/.gitconfig' 2025-11-03T16:57:21.4642722Z Temporarily overriding HOME='/home/ec2-user/actions-runner/_work/_temp/a47662a8-f8a3-43a9-a630-e4d910a68a68' before making global git config changes 2025-11-03T16:57:21.4644727Z Adding repository directory to the temporary git global config as a safe directory 2025-11-03T16:57:21.4650025Z [command]/usr/bin/git config --global --add safe.directory /home/ec2-user/actions-runner/_work/pytorch/pytorch 2025-11-03T16:57:21.4709125Z Deleting the contents of '/home/ec2-user/actions-runner/_work/pytorch/pytorch' 2025-11-03T16:57:21.4714176Z ##[group]Initializing the repository 2025-11-03T16:57:21.4717806Z [command]/usr/bin/git init /home/ec2-user/actions-runner/_work/pytorch/pytorch 2025-11-03T16:57:21.4813853Z hint: Using 'master' as the name for the initial branch. This default branch name 2025-11-03T16:57:21.4814421Z hint: is subject to change. To configure the initial branch name to use in all 2025-11-03T16:57:21.4814961Z hint: of your new repositories, which will suppress this warning, call: 2025-11-03T16:57:21.4815335Z hint: 2025-11-03T16:57:21.4815614Z hint: git config --global init.defaultBranch 2025-11-03T16:57:21.4815936Z hint: 2025-11-03T16:57:21.4816233Z hint: Names commonly chosen instead of 'master' are 'main', 'trunk' and 2025-11-03T16:57:21.4816747Z hint: 'development'. The just-created branch can be renamed via this command: 2025-11-03T16:57:21.4817377Z hint: 2025-11-03T16:57:21.4817585Z hint: git branch -m 2025-11-03T16:57:21.4817819Z hint: 2025-11-03T16:57:21.4818191Z hint: Disable this message with "git config set advice.defaultBranchName false" 2025-11-03T16:57:21.4823921Z Initialized empty Git repository in /home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/ 2025-11-03T16:57:21.4835645Z [command]/usr/bin/git remote add origin https://github.com/pytorch/pytorch 2025-11-03T16:57:21.4891584Z ##[endgroup] 2025-11-03T16:57:21.4892424Z ##[group]Disabling automatic garbage collection 2025-11-03T16:57:21.4898792Z [command]/usr/bin/git config --local gc.auto 0 2025-11-03T16:57:21.4931106Z ##[endgroup] 2025-11-03T16:57:21.4931845Z ##[group]Setting up auth 2025-11-03T16:57:21.4936655Z [command]/usr/bin/git config --local --name-only --get-regexp core\.sshCommand 2025-11-03T16:57:21.4974603Z [command]/usr/bin/git submodule foreach --recursive sh -c "git config --local --name-only --get-regexp 'core\.sshCommand' && git config --local --unset-all 'core.sshCommand' || :" 2025-11-03T16:57:21.5457765Z [command]/usr/bin/git config --local --name-only --get-regexp http\.https\:\/\/github\.com\/\.extraheader 2025-11-03T16:57:21.5494512Z [command]/usr/bin/git submodule foreach --recursive sh -c "git config --local --name-only --get-regexp 'http\.https\:\/\/github\.com\/\.extraheader' && git config --local --unset-all 'http.https://github.com/.extraheader' || :" 2025-11-03T16:57:21.5961863Z [command]/usr/bin/git config --local http.https://github.com/.extraheader AUTHORIZATION: basic *** 2025-11-03T16:57:21.6017039Z ##[endgroup] 2025-11-03T16:57:21.6017480Z ##[group]Fetching the repository 2025-11-03T16:57:21.6025495Z [command]/usr/bin/git -c protocol.version=2 fetch --prune --no-recurse-submodules origin +refs/heads/*:refs/remotes/origin/* +refs/tags/*:refs/tags/* 2025-11-03T16:58:00.4029583Z From https://github.com/pytorch/pytorch 2025-11-03T16:58:00.4030290Z * [new branch] 2.6.0.dev20241004+ -> origin/2.6.0.dev20241004+ 2025-11-03T16:58:00.4030910Z * [new branch] AaronWang04_addmmfusion_perftest -> origin/AaronWang04_addmmfusion_perftest 2025-11-03T16:58:00.4031486Z * [new branch] DyVariaSourceRepr -> origin/DyVariaSourceRepr 2025-11-03T16:58:00.4032429Z * [new branch] DynamoFixGit -> origin/DynamoFixGit 2025-11-03T16:58:00.4034285Z * [new branch] DynamoVariaT -> origin/DynamoVariaT 2025-11-03T16:58:00.4036673Z * [new branch] Flamefire-misplaced-TestFailure -> origin/Flamefire-misplaced-TestFailure 2025-11-03T16:58:00.4038860Z * [new branch] HDCharles-2.6.0-release-notes -> origin/HDCharles-2.6.0-release-notes 2025-11-03T16:58:00.4041988Z * [new branch] IvanKobzarev/stack/1 -> origin/IvanKobzarev/stack/1 2025-11-03T16:58:00.4043764Z * [new branch] IvanKobzarev/stack/2 -> origin/IvanKobzarev/stack/2 2025-11-03T16:58:00.4046023Z * [new branch] NicoshevSVE128 -> origin/NicoshevSVE128 2025-11-03T16:58:00.4047582Z * [new branch] PR-AOTInductorNoneBug -> origin/PR-AOTInductorNoneBug 2025-11-03T16:58:00.4049527Z * [new branch] PR-AOTInductorNoneBugFix -> origin/PR-AOTInductorNoneBugFix 2025-11-03T16:58:00.4051308Z * [new branch] PR-FixConfigsIssue -> origin/PR-FixConfigsIssue 2025-11-03T16:58:00.4052958Z * [new branch] PR-NoneBugFix-viable -> origin/PR-NoneBugFix-viable 2025-11-03T16:58:00.4054891Z * [new branch] PR-ResetToZero -> origin/PR-ResetToZero 2025-11-03T16:58:00.4056752Z * [new branch] Update-Flash-Packaging -> origin/Update-Flash-Packaging 2025-11-03T16:58:00.4058392Z * [new branch] VLA_exp -> origin/VLA_exp 2025-11-03T16:58:00.4060228Z * [new branch] VariaT -> origin/VariaT 2025-11-03T16:58:00.4062114Z * [new branch] add-pyrefly-to-lintrunner -> origin/add-pyrefly-to-lintrunner 2025-11-03T16:58:00.4063885Z * [new branch] add_conv3d_ops -> origin/add_conv3d_ops 2025-11-03T16:58:00.4065657Z * [new branch] add_op_to_dashboard -> origin/add_op_to_dashboard 2025-11-03T16:58:00.4067562Z * [new branch] addmm-heuristic -> origin/addmm-heuristic 2025-11-03T16:58:00.4069355Z * [new branch] addmm10000rtx -> origin/addmm10000rtx 2025-11-03T16:58:00.4071800Z * [new branch] adi/onednn_aarch64 -> origin/adi/onednn_aarch64 2025-11-03T16:58:00.4073563Z * [new branch] adi/test -> origin/adi/test 2025-11-03T16:58:00.4075297Z * [new branch] adi/test_bgemm -> origin/adi/test_bgemm 2025-11-03T16:58:00.4077007Z * [new branch] adi/test_fusions -> origin/adi/test_fusions 2025-11-03T16:58:00.4078714Z * [new branch] adi/test_onednn -> origin/adi/test_onednn 2025-11-03T16:58:00.4081879Z * [new branch] adi/test_onednn_v3.9 -> origin/adi/test_onednn_v3.9 2025-11-03T16:58:00.4083036Z * [new branch] adi/test_presve_change -> origin/adi/test_presve_change 2025-11-03T16:58:00.4083889Z * [new branch] adi/test_timm -> origin/adi/test_timm 2025-11-03T16:58:00.4086174Z * [new branch] adi/testpresve_change -> origin/adi/testpresve_change 2025-11-03T16:58:00.4089166Z * [new branch] aditew01/test/vec_bf16 -> origin/aditew01/test/vec_bf16 2025-11-03T16:58:00.4091020Z * [new branch] ah-globalfeedback-hook -> origin/ah-globalfeedback-hook 2025-11-03T16:58:00.4092844Z * [new branch] alt-disable -> origin/alt-disable 2025-11-03T16:58:00.4095360Z * [new branch] angelayi/allow_fake -> origin/angelayi/allow_fake 2025-11-03T16:58:00.4097586Z * [new branch] angelayi/aoti_additional_files -> origin/angelayi/aoti_additional_files 2025-11-03T16:58:00.4099407Z * [new branch] angelayi/benchmark -> origin/angelayi/benchmark 2025-11-03T16:58:00.4101141Z * [new branch] angelayi/change_pytree_serialization -> origin/angelayi/change_pytree_serialization 2025-11-03T16:58:00.4102832Z * [new branch] angelayi/const_folder -> origin/angelayi/const_folder 2025-11-03T16:58:00.4104568Z * [new branch] angelayi/cpp_loader -> origin/angelayi/cpp_loader 2025-11-03T16:58:00.4106357Z * [new branch] angelayi/joint_kwargs -> origin/angelayi/joint_kwargs 2025-11-03T16:58:00.4108374Z * [new branch] angelayi/no_so_weight -> origin/angelayi/no_so_weight 2025-11-03T16:58:00.4110975Z * [new branch] angelayi/opaque_obj_v2 -> origin/angelayi/opaque_obj_v2 2025-11-03T16:58:00.4113027Z * [new branch] angelayi/post_grad -> origin/angelayi/post_grad 2025-11-03T16:58:00.4114616Z * [new branch] angelayi/pytree -> origin/angelayi/pytree 2025-11-03T16:58:00.4116346Z * [new branch] angelayi/scan_layers -> origin/angelayi/scan_layers 2025-11-03T16:58:00.4118156Z * [new branch] angelayi/state_dict -> origin/angelayi/state_dict 2025-11-03T16:58:00.4120082Z * [new branch] angelayi/static_input_indices -> origin/angelayi/static_input_indices 2025-11-03T16:58:00.4121844Z * [new branch] angelayi/symint_input -> origin/angelayi/symint_input 2025-11-03T16:58:00.4123598Z * [new branch] angelayi/symm_mem -> origin/angelayi/symm_mem 2025-11-03T16:58:00.4125429Z * [new branch] angelayi/test_cpp -> origin/angelayi/test_cpp 2025-11-03T16:58:00.4127304Z * [new branch] angelayi/torch_size -> origin/angelayi/torch_size 2025-11-03T16:58:00.4129242Z * [new branch] annotate_before_eliminate -> origin/annotate_before_eliminate 2025-11-03T16:58:00.4130939Z * [new branch] annotate_fn -> origin/annotate_fn 2025-11-03T16:58:00.4132642Z * [new branch] annotation_bw -> origin/annotation_bw 2025-11-03T16:58:00.4134437Z * [new branch] annotation_bw_1 -> origin/annotation_bw_1 2025-11-03T16:58:00.4136090Z * [new branch] annotation_dynamo -> origin/annotation_dynamo 2025-11-03T16:58:00.4137826Z * [new branch] annotation_log -> origin/annotation_log 2025-11-03T16:58:00.4139687Z * [new branch] annotation_replay -> origin/annotation_replay 2025-11-03T16:58:00.4141392Z * [new branch] annotation_submod -> origin/annotation_submod 2025-11-03T16:58:00.4143203Z * [new branch] aot_eager_stack_trace -> origin/aot_eager_stack_trace 2025-11-03T16:58:00.4145012Z * [new branch] aoti-cuda-alloc -> origin/aoti-cuda-alloc 2025-11-03T16:58:00.4146737Z * [new branch] aoti_fqn_name_interface -> origin/aoti_fqn_name_interface 2025-11-03T16:58:00.4148445Z * [new branch] aoti_metal_shimify -> origin/aoti_metal_shimify 2025-11-03T16:58:00.4150210Z * [new branch] aoti_package_weights_binary -> origin/aoti_package_weights_binary 2025-11-03T16:58:00.4151983Z * [new branch] aoti_shim_library_list -> origin/aoti_shim_library_list 2025-11-03T16:58:00.4153778Z * [new branch] aoti_target_windows -> origin/aoti_target_windows 2025-11-03T16:58:00.4155535Z * [new branch] aoti_weight_sharing -> origin/aoti_weight_sharing 2025-11-03T16:58:00.4158711Z * [new branch] arsh/feat/inductor_check_profiling -> origin/arsh/feat/inductor_check_profiling 2025-11-03T16:58:00.4160734Z * [new branch] async_tp -> origin/async_tp 2025-11-03T16:58:00.4162677Z * [new branch] atalman-inductor-perf-cu124 -> origin/atalman-inductor-perf-cu124 2025-11-03T16:58:00.4164517Z * [new branch] atalman-inductor-perf-cu124.1 -> origin/atalman-inductor-perf-cu124.1 2025-11-03T16:58:00.4166411Z * [new branch] atalman-patch-1 -> origin/atalman-patch-1 2025-11-03T16:58:00.4168173Z * [new branch] atalman-patch-2 -> origin/atalman-patch-2 2025-11-03T16:58:00.4170248Z * [new branch] atalman-patch-3 -> origin/atalman-patch-3 2025-11-03T16:58:00.4172105Z * [new branch] atalman-patch-5 -> origin/atalman-patch-5 2025-11-03T16:58:00.4174035Z * [new branch] atalman_inductor_2.3.0 -> origin/atalman_inductor_2.3.0 2025-11-03T16:58:00.4175859Z * [new branch] atalman_inductor_2.3.1 -> origin/atalman_inductor_2.3.1 2025-11-03T16:58:00.4177823Z * [new branch] atalman_inductor_2.4.0 -> origin/atalman_inductor_2.4.0 2025-11-03T16:58:00.4179580Z * [new branch] atalman_inductor_2.4.x -> origin/atalman_inductor_2.4.x 2025-11-03T16:58:00.4181366Z * [new branch] attention_benchmark -> origin/attention_benchmark 2025-11-03T16:58:00.4183231Z * [new branch] attention_benchmarking_clean -> origin/attention_benchmarking_clean 2025-11-03T16:58:00.4185639Z * [new branch] bahuang/dt_fix_scalar_add -> origin/bahuang/dt_fix_scalar_add 2025-11-03T16:58:00.4187389Z * [new branch] bahuang/fix_debug_mode -> origin/bahuang/fix_debug_mode 2025-11-03T16:58:00.4189044Z * [new branch] bahuang/fix_expand -> origin/bahuang/fix_expand 2025-11-03T16:58:00.4190755Z * [new branch] bahuang/test -> origin/bahuang/test 2025-11-03T16:58:00.4193387Z * [new branch] base/1.5 -> origin/base/1.5 2025-11-03T16:58:00.4195393Z * [new branch] batching_sdpa_efficient_attention -> origin/batching_sdpa_efficient_attention 2025-11-03T16:58:00.4197270Z * [new branch] bench_scaled_mm_ops -> origin/bench_scaled_mm_ops 2025-11-03T16:58:00.4199056Z * [new branch] benchmark-updates -> origin/benchmark-updates 2025-11-03T16:58:00.4200854Z * [new branch] benchmarking-script -> origin/benchmarking-script 2025-11-03T16:58:00.4203370Z * [new branch] bertmaher/pinbump26 -> origin/bertmaher/pinbump26 2025-11-03T16:58:00.4205788Z * [new branch] bertrand/cutlass -> origin/bertrand/cutlass 2025-11-03T16:58:00.4208155Z * [new branch] bf/cg-backend -> origin/bf/cg-backend 2025-11-03T16:58:00.4210319Z * [new branch] bf/cg-partition-custom-op-mutation -> origin/bf/cg-partition-custom-op-mutation 2025-11-03T16:58:00.4211901Z * [new branch] bf/cg-remove-check -> origin/bf/cg-remove-check 2025-11-03T16:58:00.4213599Z * [new branch] bf/clean-hf -> origin/bf/clean-hf 2025-11-03T16:58:00.4215204Z * [new branch] bf/clean-timm -> origin/bf/clean-timm 2025-11-03T16:58:00.4216903Z * [new branch] bf/clean-torchbench -> origin/bf/clean-torchbench 2025-11-03T16:58:00.4218941Z * [new branch] bf/clean-torchbench-hf -> origin/bf/clean-torchbench-hf 2025-11-03T16:58:00.4221018Z * [new branch] bf/cudagraph -> origin/bf/cudagraph 2025-11-03T16:58:00.4223234Z * [new branch] bf/cudagraph-disable-input-mutation -> origin/bf/cudagraph-disable-input-mutation 2025-11-03T16:58:00.4225289Z * [new branch] bf/cudagraph-enable-input-mutation-support-benchmark -> origin/bf/cudagraph-enable-input-mutation-support-benchmark 2025-11-03T16:58:00.4226994Z * [new branch] bf/cudagraph-partition -> origin/bf/cudagraph-partition 2025-11-03T16:58:00.4228747Z * [new branch] bf/donated-buffer-bench -> origin/bf/donated-buffer-bench 2025-11-03T16:58:00.4230553Z * [new branch] bf/dynamo-partition -> origin/bf/dynamo-partition 2025-11-03T16:58:00.4232338Z * [new branch] bf/pa-non-divisible -> origin/bf/pa-non-divisible 2025-11-03T16:58:00.4234197Z * [new branch] bf/partition-cache-free-symbols -> origin/bf/partition-cache-free-symbols 2025-11-03T16:58:00.4236017Z * [new branch] bf/partition-custom-rules -> origin/bf/partition-custom-rules 2025-11-03T16:58:00.4237830Z * [new branch] bf/partition-memory-plan -> origin/bf/partition-memory-plan 2025-11-03T16:58:00.4239688Z * [new branch] bf/partition-move-cpu -> origin/bf/partition-move-cpu 2025-11-03T16:58:00.4241466Z * [new branch] bf/partition-view-fallback -> origin/bf/partition-view-fallback 2025-11-03T16:58:00.4243387Z * [new branch] bf/remove-check-55b0c39d -> origin/bf/remove-check-55b0c39d 2025-11-03T16:58:00.4244989Z * [new branch] bf16_support_per_channel -> origin/bf16_support_per_channel 2025-11-03T16:58:00.4246764Z * [new branch] bisect_perf_hf_T5_3acc6eac492 -> origin/bisect_perf_hf_T5_3acc6eac492 2025-11-03T16:58:00.4248493Z * [new branch] bisect_perf_hf_T5_3fcf66f61fb -> origin/bisect_perf_hf_T5_3fcf66f61fb 2025-11-03T16:58:00.4250321Z * [new branch] bisect_perf_hf_T5_4009d154129 -> origin/bisect_perf_hf_T5_4009d154129 2025-11-03T16:58:00.4252059Z * [new branch] bisect_perf_hf_T5_40d0740e73d -> origin/bisect_perf_hf_T5_40d0740e73d 2025-11-03T16:58:00.4253732Z * [new branch] bisect_perf_hf_T5_5268754e -> origin/bisect_perf_hf_T5_5268754e 2025-11-03T16:58:00.4255448Z * [new branch] bisect_perf_hf_T5_7d89a8d385c -> origin/bisect_perf_hf_T5_7d89a8d385c 2025-11-03T16:58:00.4257271Z * [new branch] bisect_perf_hf_T5_b7a25c1ee7c -> origin/bisect_perf_hf_T5_b7a25c1ee7c 2025-11-03T16:58:00.4259058Z * [new branch] bisect_perf_hf_T5_c25b201583f -> origin/bisect_perf_hf_T5_c25b201583f 2025-11-03T16:58:00.4260876Z * [new branch] bisect_perf_hf_T5_c93e57efac0 -> origin/bisect_perf_hf_T5_c93e57efac0 2025-11-03T16:58:00.4262549Z * [new branch] bisect_perf_hf_T5_ca9813ea149 -> origin/bisect_perf_hf_T5_ca9813ea149 2025-11-03T16:58:00.4264275Z * [new branch] bisect_perf_hf_T5_d65f194a -> origin/bisect_perf_hf_T5_d65f194a 2025-11-03T16:58:00.4265990Z * [new branch] bisect_perf_hf_T5_da94ab0b -> origin/bisect_perf_hf_T5_da94ab0b 2025-11-03T16:58:00.4267736Z * [new branch] bisect_perf_hf_T5_da94ab0b_new -> origin/bisect_perf_hf_T5_da94ab0b_new 2025-11-03T16:58:00.4269493Z * [new branch] bisect_perf_hf_T5_db4e8a1d8a8 -> origin/bisect_perf_hf_T5_db4e8a1d8a8 2025-11-03T16:58:00.4271206Z * [new branch] bisect_perf_hf_T5_e0d97e936a2 -> origin/bisect_perf_hf_T5_e0d97e936a2 2025-11-03T16:58:00.4272926Z * [new branch] bisect_perf_hf_T5_f23621ec563 -> origin/bisect_perf_hf_T5_f23621ec563 2025-11-03T16:58:00.4275299Z * [new branch] bowbao/wip_prs -> origin/bowbao/wip_prs 2025-11-03T16:58:00.4277684Z * [new branch] brister/break_dynamic_scalar -> origin/brister/break_dynamic_scalar 2025-11-03T16:58:00.4279344Z * [new branch] brister/cast_rank0 -> origin/brister/cast_rank0 2025-11-03T16:58:00.4281089Z * [new branch] brister/item_fx -> origin/brister/item_fx 2025-11-03T16:58:00.4282826Z * [new branch] brister/test_inductor_all_fx -> origin/brister/test_inductor_all_fx 2025-11-03T16:58:00.4284571Z * [new branch] brister/tiled_reduction_no_numel_check -> origin/brister/tiled_reduction_no_numel_check 2025-11-03T16:58:00.4286315Z * [new branch] bwd-backup -> origin/bwd-backup 2025-11-03T16:58:00.4288366Z * [new branch] c57382a49 -> origin/c57382a49 2025-11-03T16:58:00.4290365Z * [new branch] ca_0431d47eaa -> origin/ca_0431d47eaa 2025-11-03T16:58:00.4292039Z * [new branch] ca_fix_0431d47eaa -> origin/ca_fix_0431d47eaa 2025-11-03T16:58:00.4294549Z * [new branch] camyllh/test_setup_hooks_push -> origin/camyllh/test_setup_hooks_push 2025-11-03T16:58:00.4296381Z * [new branch] cherry-pick-152361-by-pytorch_bot_bot_ -> origin/cherry-pick-152361-by-pytorch_bot_bot_ 2025-11-03T16:58:00.4298140Z * [new branch] cherry-pick-157453-by-pytorch_bot_bot_ -> origin/cherry-pick-157453-by-pytorch_bot_bot_ 2025-11-03T16:58:00.4299959Z * [new branch] cherry-pick-157513-by-pytorch_bot_bot_ -> origin/cherry-pick-157513-by-pytorch_bot_bot_ 2025-11-03T16:58:00.4301759Z * [new branch] cherry-pick-157695-by-pytorch_bot_bot_ -> origin/cherry-pick-157695-by-pytorch_bot_bot_ 2025-11-03T16:58:00.4303633Z * [new branch] cherry-pick-157732-by-pytorch_bot_bot_ -> origin/cherry-pick-157732-by-pytorch_bot_bot_ 2025-11-03T16:58:00.4305392Z * [new branch] cherry-pick-159969-by-pytorch_bot_bot_ -> origin/cherry-pick-159969-by-pytorch_bot_bot_ 2025-11-03T16:58:00.4307280Z * [new branch] cherry-pick-160586-by-pytorch_bot_bot_ -> origin/cherry-pick-160586-by-pytorch_bot_bot_ 2025-11-03T16:58:00.4309335Z * [new branch] cherry-pick-162194-by-pytorch_bot_bot_ -> origin/cherry-pick-162194-by-pytorch_bot_bot_ 2025-11-03T16:58:00.4311579Z * [new branch] cherry-pick-162693-by-pytorch_bot_bot_ -> origin/cherry-pick-162693-by-pytorch_bot_bot_ 2025-11-03T16:58:00.4313552Z * [new branch] cherry-pick-162764-by-pytorch_bot_bot_ -> origin/cherry-pick-162764-by-pytorch_bot_bot_ 2025-11-03T16:58:00.4315371Z * [new branch] cherry-pick-163029-by-pytorch_bot_bot_ -> origin/cherry-pick-163029-by-pytorch_bot_bot_ 2025-11-03T16:58:00.4317324Z * [new branch] cherry-pick-163776-by-pytorch_bot_bot_ -> origin/cherry-pick-163776-by-pytorch_bot_bot_ 2025-11-03T16:58:00.4319178Z * [new branch] cherry-pick-164774-by-pytorch_bot_bot_ -> origin/cherry-pick-164774-by-pytorch_bot_bot_ 2025-11-03T16:58:00.4321149Z * [new branch] cherry-pick-164870-by-pytorch_bot_bot_ -> origin/cherry-pick-164870-by-pytorch_bot_bot_ 2025-11-03T16:58:00.4323007Z * [new branch] cherry-pick-164946-by-pytorch_bot_bot_ -> origin/cherry-pick-164946-by-pytorch_bot_bot_ 2025-11-03T16:58:00.4324806Z * [new branch] cherry-pick-165013-by-pytorch_bot_bot_ -> origin/cherry-pick-165013-by-pytorch_bot_bot_ 2025-11-03T16:58:00.4326617Z * [new branch] cherry-pick-165465-by-pytorch_bot_bot_ -> origin/cherry-pick-165465-by-pytorch_bot_bot_ 2025-11-03T16:58:00.4328477Z * [new branch] cherry-pick-165665-by-pytorch_bot_bot_ -> origin/cherry-pick-165665-by-pytorch_bot_bot_ 2025-11-03T16:58:00.4330346Z * [new branch] cherry-pick-165794-by-pytorch_bot_bot_ -> origin/cherry-pick-165794-by-pytorch_bot_bot_ 2025-11-03T16:58:00.4332172Z * [new branch] cherry-pick-166181-by-pytorch_bot_bot_ -> origin/cherry-pick-166181-by-pytorch_bot_bot_ 2025-11-03T16:58:00.4333919Z * [new branch] chuanqi129-patch-1 -> origin/chuanqi129-patch-1 2025-11-03T16:58:00.4335683Z * [new branch] ci_attn -> origin/ci_attn 2025-11-03T16:58:00.4337570Z * [new branch] codegen_trace -> origin/codegen_trace 2025-11-03T16:58:00.4339894Z * [new branch] codex-testing -> origin/codex-testing 2025-11-03T16:58:00.4342482Z * [new branch] codex/add-check_memory_overlap-helper-functions -> origin/codex/add-check_memory_overlap-helper-functions 2025-11-03T16:58:00.4344303Z * [new branch] codex/create-test-for-tensor-memory-leak-in-cudagraph -> origin/codex/create-test-for-tensor-memory-leak-in-cudagraph 2025-11-03T16:58:00.4346282Z * [new branch] codex/enhance-cuda.matmul-with-allow_splitk-argument -> origin/codex/enhance-cuda.matmul-with-allow_splitk-argument 2025-11-03T16:58:00.4348265Z * [new branch] codex/fix-issue-121219-in-pytorch -> origin/codex/fix-issue-121219-in-pytorch 2025-11-03T16:58:00.4350258Z * [new branch] codex/refactor-dimension-handling-in-shape.cu -> origin/codex/refactor-dimension-handling-in-shape.cu 2025-11-03T16:58:00.4352093Z * [new branch] codex/refactor-lintrunner-config-to-use-uv-run -> origin/codex/refactor-lintrunner-config-to-use-uv-run 2025-11-03T16:58:00.4353902Z * [new branch] context_test -> origin/context_test 2025-11-03T16:58:00.4356510Z * [new branch] copilot/code-style-cleanup-python-pip -> origin/copilot/code-style-cleanup-python-pip 2025-11-03T16:58:00.4358771Z * [new branch] cpio/fix_new_ami_tests -> origin/cpio/fix_new_ami_tests 2025-11-03T16:58:00.4360725Z * [new branch] cpp-docs-dependency-upgrade -> origin/cpp-docs-dependency-upgrade 2025-11-03T16:58:00.4362379Z * [new branch] crcrpar-patch-1 -> origin/crcrpar-patch-1 2025-11-03T16:58:00.4364677Z * [new branch] csl/always_produce_xml -> origin/csl/always_produce_xml 2025-11-03T16:58:00.4366345Z * [new branch] csl/build_test_more_procs -> origin/csl/build_test_more_procs 2025-11-03T16:58:00.4367962Z * [new branch] csl/build_test_more_procs2 -> origin/csl/build_test_more_procs2 2025-11-03T16:58:00.4369932Z * [new branch] csl/debug_tests_larger_runner -> origin/csl/debug_tests_larger_runner 2025-11-03T16:58:00.4371827Z * [new branch] csl/disable_test_leaking_memory -> origin/csl/disable_test_leaking_memory 2025-11-03T16:58:00.4373904Z * [new branch] csl/failed_consistently_null -> origin/csl/failed_consistently_null 2025-11-03T16:58:00.4376023Z * [new branch] csl/fix_internal_graph_executor -> origin/csl/fix_internal_graph_executor 2025-11-03T16:58:00.4377823Z * [new branch] csl/fix_periodic_debug_1 -> origin/csl/fix_periodic_debug_1 2025-11-03T16:58:00.4379642Z * [new branch] csl/inductor_h100_nightly -> origin/csl/inductor_h100_nightly 2025-11-03T16:58:00.4381419Z * [new branch] csl/katex -> origin/csl/katex 2025-11-03T16:58:00.4383200Z * [new branch] csl/keep_going_manual -> origin/csl/keep_going_manual 2025-11-03T16:58:00.4384925Z * [new branch] csl/larger_runner -> origin/csl/larger_runner 2025-11-03T16:58:00.4386661Z * [new branch] csl/lint_testing -> origin/csl/lint_testing 2025-11-03T16:58:00.4388274Z * [new branch] csl/lint_thing -> origin/csl/lint_thing 2025-11-03T16:58:00.4390096Z * [new branch] csl/lintrunner_stuff -> origin/csl/lintrunner_stuff 2025-11-03T16:58:00.4391913Z * [new branch] csl/mps_sharding -> origin/csl/mps_sharding 2025-11-03T16:58:00.4393657Z * [new branch] csl/multistage_docker -> origin/csl/multistage_docker 2025-11-03T16:58:00.4395452Z * [new branch] csl/no_keep_goin_rocm -> origin/csl/no_keep_goin_rocm 2025-11-03T16:58:00.4397319Z * [new branch] csl/nogpu_inductor_caching_test -> origin/csl/nogpu_inductor_caching_test 2025-11-03T16:58:00.4399053Z * [new branch] csl/nogpu_on_no_gpu -> origin/csl/nogpu_on_no_gpu 2025-11-03T16:58:00.4400905Z * [new branch] csl/periodic_disable -> origin/csl/periodic_disable 2025-11-03T16:58:00.4402733Z * [new branch] csl/print_something_for_log_classifier -> origin/csl/print_something_for_log_classifier 2025-11-03T16:58:00.4404366Z * [new branch] csl/print_timing -> origin/csl/print_timing 2025-11-03T16:58:00.4406141Z * [new branch] csl/remove_maybe_unused_var -> origin/csl/remove_maybe_unused_var 2025-11-03T16:58:00.4407869Z * [new branch] csl/remove_run_parallel -> origin/csl/remove_run_parallel 2025-11-03T16:58:00.4409838Z * [new branch] csl/remove_unused_vars -> origin/csl/remove_unused_vars 2025-11-03T16:58:00.4411608Z * [new branch] csl/revert_no_rule -> origin/csl/revert_no_rule 2025-11-03T16:58:00.4413327Z * [new branch] csl/revert_open -> origin/csl/revert_open 2025-11-03T16:58:00.4415048Z * [new branch] csl/run_full_lint_trunk -> origin/csl/run_full_lint_trunk 2025-11-03T16:58:00.4417009Z * [new branch] csl/skip_build -> origin/csl/skip_build 2025-11-03T16:58:00.4418804Z * [new branch] csl/smaller_avx_amx_runenrs -> origin/csl/smaller_avx_amx_runenrs 2025-11-03T16:58:00.4420496Z * [new branch] csl/td_job_level -> origin/csl/td_job_level 2025-11-03T16:58:00.4422457Z * [new branch] csl/test_cuda_build_large_runner -> origin/csl/test_cuda_build_large_runner 2025-11-03T16:58:00.4424202Z * [new branch] csl/test_owners_autograd_dispatch_nn -> origin/csl/test_owners_autograd_dispatch_nn 2025-11-03T16:58:00.4425895Z * [new branch] csl/test_owners_higher_confidence -> origin/csl/test_owners_higher_confidence 2025-11-03T16:58:00.4427617Z * [new branch] csl/upload_all_data -> origin/csl/upload_all_data 2025-11-03T16:58:00.4429369Z * [new branch] csl/win_cpp_tests -> origin/csl/win_cpp_tests 2025-11-03T16:58:00.4431147Z * [new branch] csl/win_sccache -> origin/csl/win_sccache 2025-11-03T16:58:00.4433005Z * [new branch] csl/workflow_consistency_linter -> origin/csl/workflow_consistency_linter 2025-11-03T16:58:00.4434832Z * [new branch] csl/workflow_sync_more_files -> origin/csl/workflow_sync_more_files 2025-11-03T16:58:00.4436543Z * [new branch] csl/xml_stuff -> origin/csl/xml_stuff 2025-11-03T16:58:00.4438312Z * [new branch] cublasltrelax2 -> origin/cublasltrelax2 2025-11-03T16:58:00.4440072Z * [new branch] cublasrelax2 -> origin/cublasrelax2 2025-11-03T16:58:00.4441766Z * [new branch] custom_lowering_dict -> origin/custom_lowering_dict 2025-11-03T16:58:00.4444193Z * [new branch] daxia6/2.8o3 -> origin/daxia6/2.8o3 2025-11-03T16:58:00.4445938Z * [new branch] debug-guard -> origin/debug-guard 2025-11-03T16:58:00.4447839Z * [new branch] delete-quant-docs -> origin/delete-quant-docs 2025-11-03T16:58:00.4453195Z * [new branch] dependabot/pip/dot-ci/docker/ci_commit_pins/main/transformers-4.57.0 -> origin/dependabot/pip/dot-ci/docker/ci_commit_pins/main/transformers-4.57.0 2025-11-03T16:58:00.4455162Z * [new branch] dependabot/pip/dot-ci/docker/ci_commit_pins/main/transformers-4.57.1 -> origin/dependabot/pip/dot-ci/docker/ci_commit_pins/main/transformers-4.57.1 2025-11-03T16:58:00.4457229Z * [new branch] desertfire/test_cpp_wrapper -> origin/desertfire/test_cpp_wrapper 2025-11-03T16:58:00.4459024Z * [new branch] desertfire/triton-cpu-for-aarch64 -> origin/desertfire/triton-cpu-for-aarch64 2025-11-03T16:58:00.4461774Z * [new branch] dev/dhruva/flex_attn_opt -> origin/dev/dhruva/flex_attn_opt 2025-11-03T16:58:00.4464392Z * [new branch] dev/joona/MPSNDArrayAdd -> origin/dev/joona/MPSNDArrayAdd 2025-11-03T16:58:00.4466243Z * [new branch] dev/joona/Unranked -> origin/dev/joona/Unranked 2025-11-03T16:58:00.4468263Z * [new branch] dev/joona/cat -> origin/dev/joona/cat 2025-11-03T16:58:00.4470074Z * [new branch] dev/joona/embeddingbag -> origin/dev/joona/embeddingbag 2025-11-03T16:58:00.4472002Z * [new branch] dev/joona/getTensorsString -> origin/dev/joona/getTensorsString 2025-11-03T16:58:00.4473865Z * [new branch] dev/joona/mps_linear_macos14 -> origin/dev/joona/mps_linear_macos14 2025-11-03T16:58:00.4475702Z * [new branch] dev/joona/sdpa -> origin/dev/joona/sdpa 2025-11-03T16:58:00.4477566Z * [new branch] dev/joona/type_inf -> origin/dev/joona/type_inf 2025-11-03T16:58:00.4479440Z * [new branch] dev/joona/upsize3d -> origin/dev/joona/upsize3d 2025-11-03T16:58:00.4481140Z * [new branch] disp_counter -> origin/disp_counter 2025-11-03T16:58:00.4482848Z * [new branch] document -> origin/document 2025-11-03T16:58:00.4484709Z * [new branch] document-apis -> origin/document-apis 2025-11-03T16:58:00.4486423Z * [new branch] dynamo_stack_trace -> origin/dynamo_stack_trace 2025-11-03T16:58:00.4488201Z * [new branch] eager_model_benchmarks -> origin/eager_model_benchmarks 2025-11-03T16:58:00.4490687Z * [new branch] embg/test_inductor_ci_128B -> origin/embg/test_inductor_ci_128B 2025-11-03T16:58:00.4492391Z * [new branch] embg/test_inductor_ci_base -> origin/embg/test_inductor_ci_base 2025-11-03T16:58:00.4494121Z * [new branch] embg/test_inductor_ci_control -> origin/embg/test_inductor_ci_control 2025-11-03T16:58:00.4495744Z * [new branch] embg/triton_l2_prefetch_128B -> origin/embg/triton_l2_prefetch_128B 2025-11-03T16:58:00.4497747Z * [new branch] embg/triton_l2_prefetch_256B -> origin/embg/triton_l2_prefetch_256B 2025-11-03T16:58:00.4500156Z * [new branch] exclamaforte/amd-ma -> origin/exclamaforte/amd-ma 2025-11-03T16:58:00.4502014Z * [new branch] exclamaforte/combo-kernels-perf-run -> origin/exclamaforte/combo-kernels-perf-run 2025-11-03T16:58:00.4503656Z * [new branch] exclamaforte/do_bench_refactor -> origin/exclamaforte/do_bench_refactor 2025-11-03T16:58:00.4505390Z * [new branch] exclamaforte/enable-mem-dep-fusion -> origin/exclamaforte/enable-mem-dep-fusion 2025-11-03T16:58:00.4507095Z * [new branch] exclamaforte/fix-exhaustive-autotuning -> origin/exclamaforte/fix-exhaustive-autotuning 2025-11-03T16:58:00.4509150Z * [new branch] exclamaforte/fix-trace-parsing-fx-svg -> origin/exclamaforte/fix-trace-parsing-fx-svg 2025-11-03T16:58:00.4511698Z * [new branch] exclamaforte/force-pointwise-cat-perf-run -> origin/exclamaforte/force-pointwise-cat-perf-run 2025-11-03T16:58:00.4513371Z * [new branch] exclamaforte/fusion-data -> origin/exclamaforte/fusion-data 2025-11-03T16:58:00.4515201Z * [new branch] exclamaforte/gemm-benchmark-run -> origin/exclamaforte/gemm-benchmark-run 2025-11-03T16:58:00.4516937Z * [new branch] exclamaforte/gemm-export-model -> origin/exclamaforte/gemm-export-model 2025-11-03T16:58:00.4518770Z * [new branch] exclamaforte/gemm-model -> origin/exclamaforte/gemm-model 2025-11-03T16:58:00.4520663Z * [new branch] exclamaforte/gemm-model-all-data-collection -> origin/exclamaforte/gemm-model-all-data-collection 2025-11-03T16:58:00.4522284Z * [new branch] exclamaforte/gemm-to-amd -> origin/exclamaforte/gemm-to-amd 2025-11-03T16:58:00.4524040Z * [new branch] exclamaforte/just-gemm-model -> origin/exclamaforte/just-gemm-model 2025-11-03T16:58:00.4525911Z * [new branch] exclamaforte/just-gemm-model-no-refactor -> origin/exclamaforte/just-gemm-model-no-refactor 2025-11-03T16:58:00.4527597Z * [new branch] exclamaforte/profile-diff-algo -> origin/exclamaforte/profile-diff-algo 2025-11-03T16:58:00.4529486Z * [new branch] exclamaforte/profiler-visualization -> origin/exclamaforte/profiler-visualization 2025-11-03T16:58:00.4531277Z * [new branch] exclamaforte/test_cpp_wrapper_mode -> origin/exclamaforte/test_cpp_wrapper_mode 2025-11-03T16:58:00.4533140Z * [new branch] exclamaforte/update-autotune-configs -> origin/exclamaforte/update-autotune-configs 2025-11-03T16:58:00.4534947Z * [new branch] exclamaforte/update-autotune-configs-2 -> origin/exclamaforte/update-autotune-configs-2 2025-11-03T16:58:00.4537201Z * [new branch] exclamforte/gemm-model-final -> origin/exclamforte/gemm-model-final 2025-11-03T16:58:00.4538945Z * [new branch] exec -> origin/exec 2025-11-03T16:58:00.4540949Z * [new branch] experimental-mosaic -> origin/experimental-mosaic 2025-11-03T16:58:00.4542679Z * [new branch] export-D58091437 -> origin/export-D58091437 2025-11-03T16:58:00.4544515Z * [new branch] export-D61047529 -> origin/export-D61047529 2025-11-03T16:58:00.4546300Z * [new branch] export-D71412006 -> origin/export-D71412006 2025-11-03T16:58:00.4548591Z * [new branch] export-D73042989 -> origin/export-D73042989 2025-11-03T16:58:00.4550191Z * [new branch] export-D76797250 -> origin/export-D76797250 2025-11-03T16:58:00.4551957Z * [new branch] export-D78375400 -> origin/export-D78375400 2025-11-03T16:58:00.4553670Z * [new branch] export-D78431305 -> origin/export-D78431305 2025-11-03T16:58:00.4555452Z * [new branch] export-D78822171 -> origin/export-D78822171 2025-11-03T16:58:00.4557220Z * [new branch] export-D78822351 -> origin/export-D78822351 2025-11-03T16:58:00.4558894Z * [new branch] export-D78822507 -> origin/export-D78822507 2025-11-03T16:58:00.4560621Z * [new branch] export-D78826994 -> origin/export-D78826994 2025-11-03T16:58:00.4562377Z * [new branch] export-D78894324 -> origin/export-D78894324 2025-11-03T16:58:00.4564138Z * [new branch] export-D78929245 -> origin/export-D78929245 2025-11-03T16:58:00.4565960Z * [new branch] export-D78934925 -> origin/export-D78934925 2025-11-03T16:58:00.4567758Z * [new branch] export-D78953203 -> origin/export-D78953203 2025-11-03T16:58:00.4569654Z * [new branch] export-D78953229 -> origin/export-D78953229 2025-11-03T16:58:00.4571270Z * [new branch] export-D78957093 -> origin/export-D78957093 2025-11-03T16:58:00.4573072Z * [new branch] export-D78957389 -> origin/export-D78957389 2025-11-03T16:58:00.4574891Z * [new branch] export-D78996107 -> origin/export-D78996107 2025-11-03T16:58:00.4576648Z * [new branch] export-D79026433 -> origin/export-D79026433 2025-11-03T16:58:00.4578343Z * [new branch] export-D79319835 -> origin/export-D79319835 2025-11-03T16:58:00.4580031Z * [new branch] export-D79328456 -> origin/export-D79328456 2025-11-03T16:58:00.4581832Z * [new branch] export-D80823877 -> origin/export-D80823877 2025-11-03T16:58:00.4583708Z * [new branch] export-D80958642 -> origin/export-D80958642 2025-11-03T16:58:00.4585476Z * [new branch] export-D81054193 -> origin/export-D81054193 2025-11-03T16:58:00.4587314Z * [new branch] export-D81204584 -> origin/export-D81204584 2025-11-03T16:58:00.4589035Z * [new branch] export-D81429090 -> origin/export-D81429090 2025-11-03T16:58:00.4590868Z * [new branch] export-D81698719 -> origin/export-D81698719 2025-11-03T16:58:00.4592706Z * [new branch] export-D82250826 -> origin/export-D82250826 2025-11-03T16:58:00.4594438Z * [new branch] export-D82253817 -> origin/export-D82253817 2025-11-03T16:58:00.4596256Z * [new branch] export-D83390563 -> origin/export-D83390563 2025-11-03T16:58:00.4598075Z * [new branch] export-D83391942 -> origin/export-D83391942 2025-11-03T16:58:00.4599881Z * [new branch] export-D83395610 -> origin/export-D83395610 2025-11-03T16:58:00.4601596Z * [new branch] export-D83541846 -> origin/export-D83541846 2025-11-03T16:58:00.4603469Z * [new branch] export-D83591083 -> origin/export-D83591083 2025-11-03T16:58:00.4605417Z * [new branch] export-D83609850 -> origin/export-D83609850 2025-11-03T16:58:00.4607136Z * [new branch] export-D83627170 -> origin/export-D83627170 2025-11-03T16:58:00.4609155Z * [new branch] export-D83714690 -> origin/export-D83714690 2025-11-03T16:58:00.4611264Z * [new branch] export-D83766701 -> origin/export-D83766701 2025-11-03T16:58:00.4613024Z * [new branch] export-D83768878 -> origin/export-D83768878 2025-11-03T16:58:00.4614719Z * [new branch] export-D83769447 -> origin/export-D83769447 2025-11-03T16:58:00.4616444Z * [new branch] export-D84009392 -> origin/export-D84009392 2025-11-03T16:58:00.4618282Z * [new branch] export-D84025878 -> origin/export-D84025878 2025-11-03T16:58:00.4620053Z * [new branch] export-D84089824 -> origin/export-D84089824 2025-11-03T16:58:00.4621800Z * [new branch] export-D84098898 -> origin/export-D84098898 2025-11-03T16:58:00.4623556Z * [new branch] export-D84103213 -> origin/export-D84103213 2025-11-03T16:58:00.4625906Z * [new branch] export-D84169910 -> origin/export-D84169910 2025-11-03T16:58:00.4628179Z * [new branch] export-D84213020 -> origin/export-D84213020 2025-11-03T16:58:00.4630320Z * [new branch] export-D84373821 -> origin/export-D84373821 2025-11-03T16:58:00.4632146Z * [new branch] export-D84522373 -> origin/export-D84522373 2025-11-03T16:58:00.4633936Z * [new branch] export-D84612194 -> origin/export-D84612194 2025-11-03T16:58:00.4635692Z * [new branch] export-D84890985 -> origin/export-D84890985 2025-11-03T16:58:00.4637471Z * [new branch] export-D85122326 -> origin/export-D85122326 2025-11-03T16:58:00.4639211Z * [new branch] export-D85745809 -> origin/export-D85745809 2025-11-03T16:58:00.4641174Z * [new branch] exported-model-train-idempotent -> origin/exported-model-train-idempotent 2025-11-03T16:58:00.4642903Z * [new branch] ezyang-titan-october -> origin/ezyang-titan-october 2025-11-03T16:58:00.4644762Z * [new branch] ezyang-titan-october2 -> origin/ezyang-titan-october2 2025-11-03T16:58:00.4646481Z * [new branch] ezyang-war -> origin/ezyang-war 2025-11-03T16:58:00.4648835Z * [new branch] ezyang/wip-aot-descriptors -> origin/ezyang/wip-aot-descriptors 2025-11-03T16:58:00.4650539Z * [new branch] fa_u8_brgemm -> origin/fa_u8_brgemm 2025-11-03T16:58:00.4652558Z * [new branch] fadeputr-fix-fbgemm_genai-build -> origin/fadeputr-fix-fbgemm_genai-build 2025-11-03T16:58:00.4654845Z * [new branch] fadeputr/sequence_fbgemm -> origin/fadeputr/sequence_fbgemm 2025-11-03T16:58:00.4656661Z * [new branch] fastmath_baseline -> origin/fastmath_baseline 2025-11-03T16:58:00.4659106Z * [new branch] fbcode/warm -> origin/fbcode/warm 2025-11-03T16:58:00.4660988Z * [new branch] fca -> origin/fca 2025-11-03T16:58:00.4662724Z * [new branch] fca2_ca5984c -> origin/fca2_ca5984c 2025-11-03T16:58:00.4664451Z * [new branch] fca5 -> origin/fca5 2025-11-03T16:58:00.4666924Z * [new branch] feature/forkserver-numa -> origin/feature/forkserver-numa 2025-11-03T16:58:00.4668527Z * [new branch] feature/justknobs-cpp -> origin/feature/justknobs-cpp 2025-11-03T16:58:00.4670270Z * [new branch] feature/numa-forkserver -> origin/feature/numa-forkserver 2025-11-03T16:58:00.4672441Z * [new branch] ffast_math_baseline -> origin/ffast_math_baseline 2025-11-03T16:58:00.4674335Z * [new branch] ffast_math_target -> origin/ffast_math_target 2025-11-03T16:58:00.4676824Z * [new branch] findhao/base_commit -> origin/findhao/base_commit 2025-11-03T16:58:00.4678504Z * [new branch] findhao/base_commit1 -> origin/findhao/base_commit1 2025-11-03T16:58:00.4680104Z * [new branch] findhao/multistream2 -> origin/findhao/multistream2 2025-11-03T16:58:00.4681876Z * [new branch] findhao/multistream5 -> origin/findhao/multistream5 2025-11-03T16:58:00.4683406Z * [new branch] findhao/multistream6 -> origin/findhao/multistream6 2025-11-03T16:58:00.4685088Z * [new branch] findhao/operatorbench3 -> origin/findhao/operatorbench3 2025-11-03T16:58:00.4686946Z * [new branch] findhao/operatorbench5 -> origin/findhao/operatorbench5 2025-11-03T16:58:00.4688647Z * [new branch] findhao/tritonparse -> origin/findhao/tritonparse 2025-11-03T16:58:00.4690538Z * [new branch] fix-ck-gemm-template-format -> origin/fix-ck-gemm-template-format 2025-11-03T16:58:00.4692280Z * [new branch] fix-config-ignore -> origin/fix-config-ignore 2025-11-03T16:58:00.4694228Z * [new branch] fix-dict-guard -> origin/fix-dict-guard 2025-11-03T16:58:00.4696076Z * [new branch] fix-rlease-feature-template -> origin/fix-rlease-feature-template 2025-11-03T16:58:00.4697982Z * [new branch] fix_addmm_issue -> origin/fix_addmm_issue 2025-11-03T16:58:00.4699831Z * [new branch] fix_fx_graph_print_space -> origin/fix_fx_graph_print_space 2025-11-03T16:58:00.4701561Z * [new branch] fix_nvrtc_discovery -> origin/fix_nvrtc_discovery 2025-11-03T16:58:00.4718200Z * [new branch] fix_op_runner -> origin/fix_op_runner 2025-11-03T16:58:00.4718967Z * [new branch] fix_ubn_159469 -> origin/fix_ubn_159469 2025-11-03T16:58:00.4719368Z * [new branch] fixbugh100 -> origin/fixbugh100 2025-11-03T16:58:00.4719736Z * [new branch] fixes-triage -> origin/fixes-triage 2025-11-03T16:58:00.4720096Z * [new branch] fixflashinfer -> origin/fixflashinfer 2025-11-03T16:58:00.4720452Z * [new branch] fixvllmoct -> origin/fixvllmoct 2025-11-03T16:58:00.4720826Z * [new branch] flash_decoding_cpu -> origin/flash_decoding_cpu 2025-11-03T16:58:00.4721206Z * [new branch] flex-flash -> origin/flex-flash 2025-11-03T16:58:00.4721610Z * [new branch] flex_attention_functorch_grad -> origin/flex_attention_functorch_grad 2025-11-03T16:58:00.4722001Z * [new branch] flex_flash -> origin/flex_flash 2025-11-03T16:58:00.4722629Z * [new branch] fmassa/fix_memeff_sharding_rule -> origin/fmassa/fix_memeff_sharding_rule 2025-11-03T16:58:00.4724449Z * [new branch] forkserver_fix -> origin/forkserver_fix 2025-11-03T16:58:00.4726220Z * [new branch] fsdp2_trace_rules -> origin/fsdp2_trace_rules 2025-11-03T16:58:00.4728000Z * [new branch] fx_cpp -> origin/fx_cpp 2025-11-03T16:58:00.4730504Z * [new branch] fy/fix-win -> origin/fy/fix-win 2025-11-03T16:58:00.4732931Z * [new branch] georgehong/cmakelists-patch -> origin/georgehong/cmakelists-patch 2025-11-03T16:58:00.4736350Z * [new branch] gh/AlnisM/1/base -> origin/gh/AlnisM/1/base 2025-11-03T16:58:00.4738068Z * [new branch] gh/AlnisM/1/head -> origin/gh/AlnisM/1/head 2025-11-03T16:58:00.4741401Z * [new branch] gh/ColinPeppler/94/base -> origin/gh/ColinPeppler/94/base 2025-11-03T16:58:00.4743336Z * [new branch] gh/ColinPeppler/94/head -> origin/gh/ColinPeppler/94/head 2025-11-03T16:58:00.4745074Z * [new branch] gh/ColinPeppler/94/orig -> origin/gh/ColinPeppler/94/orig 2025-11-03T16:58:00.4747749Z * [new branch] gh/ColinPeppler/95/base -> origin/gh/ColinPeppler/95/base 2025-11-03T16:58:00.4749462Z * [new branch] gh/ColinPeppler/95/head -> origin/gh/ColinPeppler/95/head 2025-11-03T16:58:00.4751463Z * [new branch] gh/ColinPeppler/95/orig -> origin/gh/ColinPeppler/95/orig 2025-11-03T16:58:00.4754157Z * [new branch] gh/EikanWang/67/base -> origin/gh/EikanWang/67/base 2025-11-03T16:58:00.4755903Z * [new branch] gh/EikanWang/67/head -> origin/gh/EikanWang/67/head 2025-11-03T16:58:00.4758179Z * [new branch] gh/EikanWang/68/base -> origin/gh/EikanWang/68/base 2025-11-03T16:58:00.4759879Z * [new branch] gh/EikanWang/68/head -> origin/gh/EikanWang/68/head 2025-11-03T16:58:00.4761584Z * [new branch] gh/EikanWang/68/orig -> origin/gh/EikanWang/68/orig 2025-11-03T16:58:00.4764626Z * [new branch] gh/Gasoonjia/1/base -> origin/gh/Gasoonjia/1/base 2025-11-03T16:58:00.4766379Z * [new branch] gh/Gasoonjia/1/head -> origin/gh/Gasoonjia/1/head 2025-11-03T16:58:00.4769322Z * [new branch] gh/H-Huang/131/base -> origin/gh/H-Huang/131/base 2025-11-03T16:58:00.4771020Z * [new branch] gh/H-Huang/131/head -> origin/gh/H-Huang/131/head 2025-11-03T16:58:00.4772761Z * [new branch] gh/H-Huang/131/orig -> origin/gh/H-Huang/131/orig 2025-11-03T16:58:00.4775148Z * [new branch] gh/H-Huang/132/base -> origin/gh/H-Huang/132/base 2025-11-03T16:58:00.4776867Z * [new branch] gh/H-Huang/132/head -> origin/gh/H-Huang/132/head 2025-11-03T16:58:00.4778618Z * [new branch] gh/H-Huang/132/orig -> origin/gh/H-Huang/132/orig 2025-11-03T16:58:00.4781010Z * [new branch] gh/H-Huang/180/base -> origin/gh/H-Huang/180/base 2025-11-03T16:58:00.4782716Z * [new branch] gh/H-Huang/180/head -> origin/gh/H-Huang/180/head 2025-11-03T16:58:00.4784396Z * [new branch] gh/H-Huang/180/orig -> origin/gh/H-Huang/180/orig 2025-11-03T16:58:00.4786631Z * [new branch] gh/H-Huang/182/base -> origin/gh/H-Huang/182/base 2025-11-03T16:58:00.4788352Z * [new branch] gh/H-Huang/182/head -> origin/gh/H-Huang/182/head 2025-11-03T16:58:00.4790086Z * [new branch] gh/H-Huang/182/orig -> origin/gh/H-Huang/182/orig 2025-11-03T16:58:00.4792439Z * [new branch] gh/H-Huang/212/base -> origin/gh/H-Huang/212/base 2025-11-03T16:58:00.4794211Z * [new branch] gh/H-Huang/212/head -> origin/gh/H-Huang/212/head 2025-11-03T16:58:00.4795936Z * [new branch] gh/H-Huang/212/orig -> origin/gh/H-Huang/212/orig 2025-11-03T16:58:00.4798237Z * [new branch] gh/H-Huang/214/base -> origin/gh/H-Huang/214/base 2025-11-03T16:58:00.4799939Z * [new branch] gh/H-Huang/214/head -> origin/gh/H-Huang/214/head 2025-11-03T16:58:00.4801640Z * [new branch] gh/H-Huang/214/orig -> origin/gh/H-Huang/214/orig 2025-11-03T16:58:00.4803943Z * [new branch] gh/H-Huang/216/base -> origin/gh/H-Huang/216/base 2025-11-03T16:58:00.4805678Z * [new branch] gh/H-Huang/216/head -> origin/gh/H-Huang/216/head 2025-11-03T16:58:00.4807390Z * [new branch] gh/H-Huang/216/orig -> origin/gh/H-Huang/216/orig 2025-11-03T16:58:00.4810300Z * [new branch] gh/H-Huang/218/base -> origin/gh/H-Huang/218/base 2025-11-03T16:58:00.4811931Z * [new branch] gh/H-Huang/218/head -> origin/gh/H-Huang/218/head 2025-11-03T16:58:00.4813647Z * [new branch] gh/H-Huang/218/orig -> origin/gh/H-Huang/218/orig 2025-11-03T16:58:00.4815950Z * [new branch] gh/H-Huang/219/base -> origin/gh/H-Huang/219/base 2025-11-03T16:58:00.4817792Z * [new branch] gh/H-Huang/219/head -> origin/gh/H-Huang/219/head 2025-11-03T16:58:00.4819507Z * [new branch] gh/H-Huang/219/orig -> origin/gh/H-Huang/219/orig 2025-11-03T16:58:00.4822009Z * [new branch] gh/H-Huang/220/base -> origin/gh/H-Huang/220/base 2025-11-03T16:58:00.4823553Z * [new branch] gh/H-Huang/220/head -> origin/gh/H-Huang/220/head 2025-11-03T16:58:00.4825199Z * [new branch] gh/H-Huang/220/orig -> origin/gh/H-Huang/220/orig 2025-11-03T16:58:00.4827528Z * [new branch] gh/H-Huang/221/base -> origin/gh/H-Huang/221/base 2025-11-03T16:58:00.4829263Z * [new branch] gh/H-Huang/221/head -> origin/gh/H-Huang/221/head 2025-11-03T16:58:00.4830985Z * [new branch] gh/H-Huang/221/orig -> origin/gh/H-Huang/221/orig 2025-11-03T16:58:00.4833347Z * [new branch] gh/H-Huang/222/base -> origin/gh/H-Huang/222/base 2025-11-03T16:58:00.4835040Z * [new branch] gh/H-Huang/222/head -> origin/gh/H-Huang/222/head 2025-11-03T16:58:00.4836741Z * [new branch] gh/H-Huang/222/orig -> origin/gh/H-Huang/222/orig 2025-11-03T16:58:00.4839199Z * [new branch] gh/H-Huang/223/base -> origin/gh/H-Huang/223/base 2025-11-03T16:58:00.4840895Z * [new branch] gh/H-Huang/223/head -> origin/gh/H-Huang/223/head 2025-11-03T16:58:00.4842569Z * [new branch] gh/H-Huang/223/orig -> origin/gh/H-Huang/223/orig 2025-11-03T16:58:00.4844915Z * [new branch] gh/H-Huang/224/base -> origin/gh/H-Huang/224/base 2025-11-03T16:58:00.4846599Z * [new branch] gh/H-Huang/224/head -> origin/gh/H-Huang/224/head 2025-11-03T16:58:00.4848350Z * [new branch] gh/H-Huang/224/orig -> origin/gh/H-Huang/224/orig 2025-11-03T16:58:00.4850781Z * [new branch] gh/H-Huang/225/base -> origin/gh/H-Huang/225/base 2025-11-03T16:58:00.4852440Z * [new branch] gh/H-Huang/225/head -> origin/gh/H-Huang/225/head 2025-11-03T16:58:00.4854144Z * [new branch] gh/H-Huang/225/orig -> origin/gh/H-Huang/225/orig 2025-11-03T16:58:00.4856460Z * [new branch] gh/H-Huang/226/base -> origin/gh/H-Huang/226/base 2025-11-03T16:58:00.4858258Z * [new branch] gh/H-Huang/226/head -> origin/gh/H-Huang/226/head 2025-11-03T16:58:00.4859985Z * [new branch] gh/H-Huang/226/orig -> origin/gh/H-Huang/226/orig 2025-11-03T16:58:00.4862843Z * [new branch] gh/H-Huang/227/base -> origin/gh/H-Huang/227/base 2025-11-03T16:58:00.4864566Z * [new branch] gh/H-Huang/227/head -> origin/gh/H-Huang/227/head 2025-11-03T16:58:00.4866258Z * [new branch] gh/H-Huang/227/orig -> origin/gh/H-Huang/227/orig 2025-11-03T16:58:00.4868809Z * [new branch] gh/H-Huang/228/base -> origin/gh/H-Huang/228/base 2025-11-03T16:58:00.4870541Z * [new branch] gh/H-Huang/228/head -> origin/gh/H-Huang/228/head 2025-11-03T16:58:00.4872247Z * [new branch] gh/H-Huang/228/orig -> origin/gh/H-Huang/228/orig 2025-11-03T16:58:00.4875114Z * [new branch] gh/IvanKobzarev/150/base -> origin/gh/IvanKobzarev/150/base 2025-11-03T16:58:00.4876997Z * [new branch] gh/IvanKobzarev/150/head -> origin/gh/IvanKobzarev/150/head 2025-11-03T16:58:00.4878841Z * [new branch] gh/IvanKobzarev/150/orig -> origin/gh/IvanKobzarev/150/orig 2025-11-03T16:58:00.4881271Z * [new branch] gh/IvanKobzarev/151/base -> origin/gh/IvanKobzarev/151/base 2025-11-03T16:58:00.4883044Z * [new branch] gh/IvanKobzarev/151/head -> origin/gh/IvanKobzarev/151/head 2025-11-03T16:58:00.4884787Z * [new branch] gh/IvanKobzarev/151/orig -> origin/gh/IvanKobzarev/151/orig 2025-11-03T16:58:00.4887206Z * [new branch] gh/IvanKobzarev/156/base -> origin/gh/IvanKobzarev/156/base 2025-11-03T16:58:00.4889023Z * [new branch] gh/IvanKobzarev/156/head -> origin/gh/IvanKobzarev/156/head 2025-11-03T16:58:00.4890946Z * [new branch] gh/IvanKobzarev/156/orig -> origin/gh/IvanKobzarev/156/orig 2025-11-03T16:58:00.4893182Z * [new branch] gh/IvanKobzarev/157/base -> origin/gh/IvanKobzarev/157/base 2025-11-03T16:58:00.4894948Z * [new branch] gh/IvanKobzarev/157/head -> origin/gh/IvanKobzarev/157/head 2025-11-03T16:58:00.4896693Z * [new branch] gh/IvanKobzarev/157/orig -> origin/gh/IvanKobzarev/157/orig 2025-11-03T16:58:00.4898999Z * [new branch] gh/IvanKobzarev/158/base -> origin/gh/IvanKobzarev/158/base 2025-11-03T16:58:00.4900812Z * [new branch] gh/IvanKobzarev/158/head -> origin/gh/IvanKobzarev/158/head 2025-11-03T16:58:00.4902521Z * [new branch] gh/IvanKobzarev/158/orig -> origin/gh/IvanKobzarev/158/orig 2025-11-03T16:58:00.4904816Z * [new branch] gh/IvanKobzarev/159/base -> origin/gh/IvanKobzarev/159/base 2025-11-03T16:58:00.4906561Z * [new branch] gh/IvanKobzarev/159/head -> origin/gh/IvanKobzarev/159/head 2025-11-03T16:58:00.4908296Z * [new branch] gh/IvanKobzarev/159/orig -> origin/gh/IvanKobzarev/159/orig 2025-11-03T16:58:00.4912357Z * [new branch] gh/IvanKobzarev/161/base -> origin/gh/IvanKobzarev/161/base 2025-11-03T16:58:00.4914051Z * [new branch] gh/IvanKobzarev/161/head -> origin/gh/IvanKobzarev/161/head 2025-11-03T16:58:00.4915723Z * [new branch] gh/IvanKobzarev/161/orig -> origin/gh/IvanKobzarev/161/orig 2025-11-03T16:58:00.4918168Z * [new branch] gh/IvanKobzarev/162/base -> origin/gh/IvanKobzarev/162/base 2025-11-03T16:58:00.4920026Z * [new branch] gh/IvanKobzarev/162/head -> origin/gh/IvanKobzarev/162/head 2025-11-03T16:58:00.4921797Z * [new branch] gh/IvanKobzarev/162/orig -> origin/gh/IvanKobzarev/162/orig 2025-11-03T16:58:00.4924128Z * [new branch] gh/IvanKobzarev/163/base -> origin/gh/IvanKobzarev/163/base 2025-11-03T16:58:00.4925968Z * [new branch] gh/IvanKobzarev/163/head -> origin/gh/IvanKobzarev/163/head 2025-11-03T16:58:00.4927674Z * [new branch] gh/IvanKobzarev/163/orig -> origin/gh/IvanKobzarev/163/orig 2025-11-03T16:58:00.4930425Z * [new branch] gh/IvanKobzarev/164/base -> origin/gh/IvanKobzarev/164/base 2025-11-03T16:58:00.4932052Z * [new branch] gh/IvanKobzarev/164/head -> origin/gh/IvanKobzarev/164/head 2025-11-03T16:58:00.4933767Z * [new branch] gh/IvanKobzarev/164/orig -> origin/gh/IvanKobzarev/164/orig 2025-11-03T16:58:00.4936109Z * [new branch] gh/IvanKobzarev/165/base -> origin/gh/IvanKobzarev/165/base 2025-11-03T16:58:00.4937872Z * [new branch] gh/IvanKobzarev/165/head -> origin/gh/IvanKobzarev/165/head 2025-11-03T16:58:00.4939562Z * [new branch] gh/IvanKobzarev/165/orig -> origin/gh/IvanKobzarev/165/orig 2025-11-03T16:58:00.4941917Z * [new branch] gh/IvanKobzarev/166/base -> origin/gh/IvanKobzarev/166/base 2025-11-03T16:58:00.4943704Z * [new branch] gh/IvanKobzarev/166/head -> origin/gh/IvanKobzarev/166/head 2025-11-03T16:58:00.4945583Z * [new branch] gh/IvanKobzarev/166/orig -> origin/gh/IvanKobzarev/166/orig 2025-11-03T16:58:00.4948528Z * [new branch] gh/NikhilAPatel/1/base -> origin/gh/NikhilAPatel/1/base 2025-11-03T16:58:00.4950293Z * [new branch] gh/NikhilAPatel/1/head -> origin/gh/NikhilAPatel/1/head 2025-11-03T16:58:00.4952543Z * [new branch] gh/NikhilAPatel/2/base -> origin/gh/NikhilAPatel/2/base 2025-11-03T16:58:00.4954231Z * [new branch] gh/NikhilAPatel/2/head -> origin/gh/NikhilAPatel/2/head 2025-11-03T16:58:00.4956797Z * [new branch] gh/NikhilAPatel/4/base -> origin/gh/NikhilAPatel/4/base 2025-11-03T16:58:00.4958499Z * [new branch] gh/NikhilAPatel/4/head -> origin/gh/NikhilAPatel/4/head 2025-11-03T16:58:00.4961397Z * [new branch] gh/PaliC/17/base -> origin/gh/PaliC/17/base 2025-11-03T16:58:00.4963021Z * [new branch] gh/PaliC/17/head -> origin/gh/PaliC/17/head 2025-11-03T16:58:00.4964682Z * [new branch] gh/PaliC/17/orig -> origin/gh/PaliC/17/orig 2025-11-03T16:58:00.4967049Z * [new branch] gh/PaliC/18/base -> origin/gh/PaliC/18/base 2025-11-03T16:58:00.4968839Z * [new branch] gh/PaliC/18/head -> origin/gh/PaliC/18/head 2025-11-03T16:58:00.4970581Z * [new branch] gh/PaliC/18/orig -> origin/gh/PaliC/18/orig 2025-11-03T16:58:00.4972890Z * [new branch] gh/PaliC/20/base -> origin/gh/PaliC/20/base 2025-11-03T16:58:00.4974568Z * [new branch] gh/PaliC/20/head -> origin/gh/PaliC/20/head 2025-11-03T16:58:00.4976296Z * [new branch] gh/PaliC/20/orig -> origin/gh/PaliC/20/orig 2025-11-03T16:58:00.4978703Z * [new branch] gh/PaliC/21/base -> origin/gh/PaliC/21/base 2025-11-03T16:58:00.4980402Z * [new branch] gh/PaliC/21/head -> origin/gh/PaliC/21/head 2025-11-03T16:58:00.4982173Z * [new branch] gh/PaliC/21/orig -> origin/gh/PaliC/21/orig 2025-11-03T16:58:00.4984399Z * [new branch] gh/PaliC/23/base -> origin/gh/PaliC/23/base 2025-11-03T16:58:00.4986118Z * [new branch] gh/PaliC/23/head -> origin/gh/PaliC/23/head 2025-11-03T16:58:00.4987822Z * [new branch] gh/PaliC/23/orig -> origin/gh/PaliC/23/orig 2025-11-03T16:58:00.4990122Z * [new branch] gh/PaliC/24/base -> origin/gh/PaliC/24/base 2025-11-03T16:58:00.4991870Z * [new branch] gh/PaliC/24/head -> origin/gh/PaliC/24/head 2025-11-03T16:58:00.4993572Z * [new branch] gh/PaliC/24/orig -> origin/gh/PaliC/24/orig 2025-11-03T16:58:00.4995847Z * [new branch] gh/PaliC/25/head -> origin/gh/PaliC/25/head 2025-11-03T16:58:00.4997538Z * [new branch] gh/PaliC/25/next -> origin/gh/PaliC/25/next 2025-11-03T16:58:00.4999317Z * [new branch] gh/PaliC/25/orig -> origin/gh/PaliC/25/orig 2025-11-03T16:58:00.5001622Z * [new branch] gh/PaliC/26/head -> origin/gh/PaliC/26/head 2025-11-03T16:58:00.5003310Z * [new branch] gh/PaliC/26/next -> origin/gh/PaliC/26/next 2025-11-03T16:58:00.5005030Z * [new branch] gh/PaliC/26/orig -> origin/gh/PaliC/26/orig 2025-11-03T16:58:00.5007306Z * [new branch] gh/PaliC/27/next -> origin/gh/PaliC/27/next 2025-11-03T16:58:00.5009967Z * [new branch] gh/PaliC/28/head -> origin/gh/PaliC/28/head 2025-11-03T16:58:00.5011607Z * [new branch] gh/PaliC/28/next -> origin/gh/PaliC/28/next 2025-11-03T16:58:00.5013290Z * [new branch] gh/PaliC/28/orig -> origin/gh/PaliC/28/orig 2025-11-03T16:58:00.5015751Z * [new branch] gh/PaliC/29/head -> origin/gh/PaliC/29/head 2025-11-03T16:58:00.5017390Z * [new branch] gh/PaliC/29/next -> origin/gh/PaliC/29/next 2025-11-03T16:58:00.5019088Z * [new branch] gh/PaliC/29/orig -> origin/gh/PaliC/29/orig 2025-11-03T16:58:00.5021479Z * [new branch] gh/PaliC/30/head -> origin/gh/PaliC/30/head 2025-11-03T16:58:00.5023050Z * [new branch] gh/PaliC/30/next -> origin/gh/PaliC/30/next 2025-11-03T16:58:00.5024789Z * [new branch] gh/PaliC/30/orig -> origin/gh/PaliC/30/orig 2025-11-03T16:58:00.5027065Z * [new branch] gh/PaliC/31/head -> origin/gh/PaliC/31/head 2025-11-03T16:58:00.5028643Z * [new branch] gh/PaliC/31/next -> origin/gh/PaliC/31/next 2025-11-03T16:58:00.5030339Z * [new branch] gh/PaliC/31/orig -> origin/gh/PaliC/31/orig 2025-11-03T16:58:00.5033361Z * [new branch] gh/PaulZhang12/22/base -> origin/gh/PaulZhang12/22/base 2025-11-03T16:58:00.5034916Z * [new branch] gh/PaulZhang12/22/head -> origin/gh/PaulZhang12/22/head 2025-11-03T16:58:00.5036630Z * [new branch] gh/PaulZhang12/22/orig -> origin/gh/PaulZhang12/22/orig 2025-11-03T16:58:00.5038926Z * [new branch] gh/PaulZhang12/25/base -> origin/gh/PaulZhang12/25/base 2025-11-03T16:58:00.5040649Z * [new branch] gh/PaulZhang12/25/head -> origin/gh/PaulZhang12/25/head 2025-11-03T16:58:00.5042372Z * [new branch] gh/PaulZhang12/25/orig -> origin/gh/PaulZhang12/25/orig 2025-11-03T16:58:00.5044789Z * [new branch] gh/PaulZhang12/26/base -> origin/gh/PaulZhang12/26/base 2025-11-03T16:58:00.5046483Z * [new branch] gh/PaulZhang12/26/head -> origin/gh/PaulZhang12/26/head 2025-11-03T16:58:00.5048171Z * [new branch] gh/PaulZhang12/26/orig -> origin/gh/PaulZhang12/26/orig 2025-11-03T16:58:00.5050705Z * [new branch] gh/PaulZhang12/27/base -> origin/gh/PaulZhang12/27/base 2025-11-03T16:58:00.5052438Z * [new branch] gh/PaulZhang12/27/head -> origin/gh/PaulZhang12/27/head 2025-11-03T16:58:00.5054147Z * [new branch] gh/PaulZhang12/27/orig -> origin/gh/PaulZhang12/27/orig 2025-11-03T16:58:00.5056470Z * [new branch] gh/PaulZhang12/28/base -> origin/gh/PaulZhang12/28/base 2025-11-03T16:58:00.5058322Z * [new branch] gh/PaulZhang12/28/head -> origin/gh/PaulZhang12/28/head 2025-11-03T16:58:00.5059997Z * [new branch] gh/PaulZhang12/28/orig -> origin/gh/PaulZhang12/28/orig 2025-11-03T16:58:00.5062430Z * [new branch] gh/PaulZhang12/30/base -> origin/gh/PaulZhang12/30/base 2025-11-03T16:58:00.5064235Z * [new branch] gh/PaulZhang12/30/head -> origin/gh/PaulZhang12/30/head 2025-11-03T16:58:00.5065954Z * [new branch] gh/PaulZhang12/30/orig -> origin/gh/PaulZhang12/30/orig 2025-11-03T16:58:00.5068409Z * [new branch] gh/PaulZhang12/31/base -> origin/gh/PaulZhang12/31/base 2025-11-03T16:58:00.5070218Z * [new branch] gh/PaulZhang12/31/head -> origin/gh/PaulZhang12/31/head 2025-11-03T16:58:00.5071940Z * [new branch] gh/PaulZhang12/31/orig -> origin/gh/PaulZhang12/31/orig 2025-11-03T16:58:00.5074375Z * [new branch] gh/PaulZhang12/32/base -> origin/gh/PaulZhang12/32/base 2025-11-03T16:58:00.5076028Z * [new branch] gh/PaulZhang12/32/head -> origin/gh/PaulZhang12/32/head 2025-11-03T16:58:00.5077701Z * [new branch] gh/PaulZhang12/32/orig -> origin/gh/PaulZhang12/32/orig 2025-11-03T16:58:00.5081150Z * [new branch] gh/PaulZhang12/33/base -> origin/gh/PaulZhang12/33/base 2025-11-03T16:58:00.5083526Z * [new branch] gh/PaulZhang12/33/head -> origin/gh/PaulZhang12/33/head 2025-11-03T16:58:00.5085280Z * [new branch] gh/PaulZhang12/33/orig -> origin/gh/PaulZhang12/33/orig 2025-11-03T16:58:00.5087655Z * [new branch] gh/PaulZhang12/34/base -> origin/gh/PaulZhang12/34/base 2025-11-03T16:58:00.5089461Z * [new branch] gh/PaulZhang12/34/head -> origin/gh/PaulZhang12/34/head 2025-11-03T16:58:00.5091204Z * [new branch] gh/PaulZhang12/34/orig -> origin/gh/PaulZhang12/34/orig 2025-11-03T16:58:00.5095016Z * [new branch] gh/PaulZhang12/35/base -> origin/gh/PaulZhang12/35/base 2025-11-03T16:58:00.5097247Z * [new branch] gh/PaulZhang12/35/head -> origin/gh/PaulZhang12/35/head 2025-11-03T16:58:00.5098986Z * [new branch] gh/PaulZhang12/35/orig -> origin/gh/PaulZhang12/35/orig 2025-11-03T16:58:00.5101349Z * [new branch] gh/PaulZhang12/36/base -> origin/gh/PaulZhang12/36/base 2025-11-03T16:58:00.5103130Z * [new branch] gh/PaulZhang12/36/head -> origin/gh/PaulZhang12/36/head 2025-11-03T16:58:00.5104764Z * [new branch] gh/PaulZhang12/36/orig -> origin/gh/PaulZhang12/36/orig 2025-11-03T16:58:00.5107256Z * [new branch] gh/PaulZhang12/37/base -> origin/gh/PaulZhang12/37/base 2025-11-03T16:58:00.5108935Z * [new branch] gh/PaulZhang12/37/head -> origin/gh/PaulZhang12/37/head 2025-11-03T16:58:00.5110940Z * [new branch] gh/PaulZhang12/37/orig -> origin/gh/PaulZhang12/37/orig 2025-11-03T16:58:00.5113052Z * [new branch] gh/PaulZhang12/38/base -> origin/gh/PaulZhang12/38/base 2025-11-03T16:58:00.5114853Z * [new branch] gh/PaulZhang12/38/head -> origin/gh/PaulZhang12/38/head 2025-11-03T16:58:00.5116535Z * [new branch] gh/PaulZhang12/38/orig -> origin/gh/PaulZhang12/38/orig 2025-11-03T16:58:00.5118878Z * [new branch] gh/PaulZhang12/39/base -> origin/gh/PaulZhang12/39/base 2025-11-03T16:58:00.5120637Z * [new branch] gh/PaulZhang12/39/head -> origin/gh/PaulZhang12/39/head 2025-11-03T16:58:00.5122375Z * [new branch] gh/PaulZhang12/39/orig -> origin/gh/PaulZhang12/39/orig 2025-11-03T16:58:00.5124672Z * [new branch] gh/PaulZhang12/40/base -> origin/gh/PaulZhang12/40/base 2025-11-03T16:58:00.5126365Z * [new branch] gh/PaulZhang12/40/head -> origin/gh/PaulZhang12/40/head 2025-11-03T16:58:00.5128167Z * [new branch] gh/PaulZhang12/40/orig -> origin/gh/PaulZhang12/40/orig 2025-11-03T16:58:00.5130636Z * [new branch] gh/PaulZhang12/41/base -> origin/gh/PaulZhang12/41/base 2025-11-03T16:58:00.5132352Z * [new branch] gh/PaulZhang12/41/head -> origin/gh/PaulZhang12/41/head 2025-11-03T16:58:00.5134034Z * [new branch] gh/PaulZhang12/41/orig -> origin/gh/PaulZhang12/41/orig 2025-11-03T16:58:00.5136805Z * [new branch] gh/SamGinzburg/11/base -> origin/gh/SamGinzburg/11/base 2025-11-03T16:58:00.5138508Z * [new branch] gh/SamGinzburg/11/head -> origin/gh/SamGinzburg/11/head 2025-11-03T16:58:00.5141915Z * [new branch] gh/SherlockNoMad/1/base -> origin/gh/SherlockNoMad/1/base 2025-11-03T16:58:00.5143625Z * [new branch] gh/SherlockNoMad/1/head -> origin/gh/SherlockNoMad/1/head 2025-11-03T16:58:00.5145955Z * [new branch] gh/SherlockNoMad/10/base -> origin/gh/SherlockNoMad/10/base 2025-11-03T16:58:00.5147701Z * [new branch] gh/SherlockNoMad/10/head -> origin/gh/SherlockNoMad/10/head 2025-11-03T16:58:00.5149470Z * [new branch] gh/SherlockNoMad/10/orig -> origin/gh/SherlockNoMad/10/orig 2025-11-03T16:58:00.5151639Z * [new branch] gh/SherlockNoMad/11/base -> origin/gh/SherlockNoMad/11/base 2025-11-03T16:58:00.5153513Z * [new branch] gh/SherlockNoMad/11/head -> origin/gh/SherlockNoMad/11/head 2025-11-03T16:58:00.5157377Z * [new branch] gh/SherlockNoMad/11/orig -> origin/gh/SherlockNoMad/11/orig 2025-11-03T16:58:00.5158248Z * [new branch] gh/SherlockNoMad/12/base -> origin/gh/SherlockNoMad/12/base 2025-11-03T16:58:00.5159615Z * [new branch] gh/SherlockNoMad/12/head -> origin/gh/SherlockNoMad/12/head 2025-11-03T16:58:00.5160649Z * [new branch] gh/SherlockNoMad/12/orig -> origin/gh/SherlockNoMad/12/orig 2025-11-03T16:58:00.5162887Z * [new branch] gh/SherlockNoMad/13/base -> origin/gh/SherlockNoMad/13/base 2025-11-03T16:58:00.5164670Z * [new branch] gh/SherlockNoMad/13/head -> origin/gh/SherlockNoMad/13/head 2025-11-03T16:58:00.5166387Z * [new branch] gh/SherlockNoMad/13/orig -> origin/gh/SherlockNoMad/13/orig 2025-11-03T16:58:00.5169058Z * [new branch] gh/SherlockNoMad/14/base -> origin/gh/SherlockNoMad/14/base 2025-11-03T16:58:00.5171124Z * [new branch] gh/SherlockNoMad/14/head -> origin/gh/SherlockNoMad/14/head 2025-11-03T16:58:00.5172621Z * [new branch] gh/SherlockNoMad/14/orig -> origin/gh/SherlockNoMad/14/orig 2025-11-03T16:58:00.5175059Z * [new branch] gh/SherlockNoMad/15/base -> origin/gh/SherlockNoMad/15/base 2025-11-03T16:58:00.5176689Z * [new branch] gh/SherlockNoMad/15/head -> origin/gh/SherlockNoMad/15/head 2025-11-03T16:58:00.5178389Z * [new branch] gh/SherlockNoMad/15/orig -> origin/gh/SherlockNoMad/15/orig 2025-11-03T16:58:00.5180741Z * [new branch] gh/SherlockNoMad/16/base -> origin/gh/SherlockNoMad/16/base 2025-11-03T16:58:00.5182429Z * [new branch] gh/SherlockNoMad/16/head -> origin/gh/SherlockNoMad/16/head 2025-11-03T16:58:00.5184200Z * [new branch] gh/SherlockNoMad/16/orig -> origin/gh/SherlockNoMad/16/orig 2025-11-03T16:58:00.5186402Z * [new branch] gh/SherlockNoMad/17/base -> origin/gh/SherlockNoMad/17/base 2025-11-03T16:58:00.5188251Z * [new branch] gh/SherlockNoMad/17/head -> origin/gh/SherlockNoMad/17/head 2025-11-03T16:58:00.5190074Z * [new branch] gh/SherlockNoMad/17/orig -> origin/gh/SherlockNoMad/17/orig 2025-11-03T16:58:00.5192288Z * [new branch] gh/SherlockNoMad/2/base -> origin/gh/SherlockNoMad/2/base 2025-11-03T16:58:00.5193974Z * [new branch] gh/SherlockNoMad/2/head -> origin/gh/SherlockNoMad/2/head 2025-11-03T16:58:00.5196106Z * [new branch] gh/SherlockNoMad/3/base -> origin/gh/SherlockNoMad/3/base 2025-11-03T16:58:00.5197797Z * [new branch] gh/SherlockNoMad/3/head -> origin/gh/SherlockNoMad/3/head 2025-11-03T16:58:00.5199973Z * [new branch] gh/SherlockNoMad/4/base -> origin/gh/SherlockNoMad/4/base 2025-11-03T16:58:00.5201636Z * [new branch] gh/SherlockNoMad/4/head -> origin/gh/SherlockNoMad/4/head 2025-11-03T16:58:00.5203859Z * [new branch] gh/SherlockNoMad/5/base -> origin/gh/SherlockNoMad/5/base 2025-11-03T16:58:00.5205502Z * [new branch] gh/SherlockNoMad/5/head -> origin/gh/SherlockNoMad/5/head 2025-11-03T16:58:00.5208482Z * [new branch] gh/Sidharth123-cpu/24/base -> origin/gh/Sidharth123-cpu/24/base 2025-11-03T16:58:00.5210961Z * [new branch] gh/Sidharth123-cpu/25/base -> origin/gh/Sidharth123-cpu/25/base 2025-11-03T16:58:00.5213084Z * [new branch] gh/Sidharth123-cpu/26/base -> origin/gh/Sidharth123-cpu/26/base 2025-11-03T16:58:00.5215504Z * [new branch] gh/Sidharth123-cpu/27/base -> origin/gh/Sidharth123-cpu/27/base 2025-11-03T16:58:00.5218369Z * [new branch] gh/StrongerXi/1/base -> origin/gh/StrongerXi/1/base 2025-11-03T16:58:00.5220111Z * [new branch] gh/StrongerXi/1/head -> origin/gh/StrongerXi/1/head 2025-11-03T16:58:00.5222618Z * [new branch] gh/StrongerXi/136/base -> origin/gh/StrongerXi/136/base 2025-11-03T16:58:00.5224344Z * [new branch] gh/StrongerXi/136/head -> origin/gh/StrongerXi/136/head 2025-11-03T16:58:00.5226063Z * [new branch] gh/StrongerXi/136/orig -> origin/gh/StrongerXi/136/orig 2025-11-03T16:58:00.5228337Z * [new branch] gh/StrongerXi/137/base -> origin/gh/StrongerXi/137/base 2025-11-03T16:58:00.5230027Z * [new branch] gh/StrongerXi/137/head -> origin/gh/StrongerXi/137/head 2025-11-03T16:58:00.5231773Z * [new branch] gh/StrongerXi/137/orig -> origin/gh/StrongerXi/137/orig 2025-11-03T16:58:00.5234078Z * [new branch] gh/StrongerXi/138/base -> origin/gh/StrongerXi/138/base 2025-11-03T16:58:00.5235765Z * [new branch] gh/StrongerXi/138/head -> origin/gh/StrongerXi/138/head 2025-11-03T16:58:00.5237486Z * [new branch] gh/StrongerXi/138/orig -> origin/gh/StrongerXi/138/orig 2025-11-03T16:58:00.5240035Z * [new branch] gh/StrongerXi/71/base -> origin/gh/StrongerXi/71/base 2025-11-03T16:58:00.5241649Z * [new branch] gh/StrongerXi/71/head -> origin/gh/StrongerXi/71/head 2025-11-03T16:58:00.5243869Z * [new branch] gh/StrongerXi/72/base -> origin/gh/StrongerXi/72/base 2025-11-03T16:58:00.5245579Z * [new branch] gh/StrongerXi/72/head -> origin/gh/StrongerXi/72/head 2025-11-03T16:58:00.5248621Z * [new branch] gh/XilunWu/151/base -> origin/gh/XilunWu/151/base 2025-11-03T16:58:00.5250380Z * [new branch] gh/XilunWu/151/head -> origin/gh/XilunWu/151/head 2025-11-03T16:58:00.5252559Z * [new branch] gh/XilunWu/151/orig -> origin/gh/XilunWu/151/orig 2025-11-03T16:58:00.5255349Z * [new branch] gh/XilunWu/152/base -> origin/gh/XilunWu/152/base 2025-11-03T16:58:00.5257096Z * [new branch] gh/XilunWu/152/head -> origin/gh/XilunWu/152/head 2025-11-03T16:58:00.5258740Z * [new branch] gh/XilunWu/152/orig -> origin/gh/XilunWu/152/orig 2025-11-03T16:58:00.5261237Z * [new branch] gh/XilunWu/160/base -> origin/gh/XilunWu/160/base 2025-11-03T16:58:00.5262910Z * [new branch] gh/XilunWu/160/head -> origin/gh/XilunWu/160/head 2025-11-03T16:58:00.5264604Z * [new branch] gh/XilunWu/160/orig -> origin/gh/XilunWu/160/orig 2025-11-03T16:58:00.5266966Z * [new branch] gh/XilunWu/163/base -> origin/gh/XilunWu/163/base 2025-11-03T16:58:00.5268607Z * [new branch] gh/XilunWu/163/head -> origin/gh/XilunWu/163/head 2025-11-03T16:58:00.5270295Z * [new branch] gh/XilunWu/163/orig -> origin/gh/XilunWu/163/orig 2025-11-03T16:58:00.5272935Z * [new branch] gh/XilunWu/166/base -> origin/gh/XilunWu/166/base 2025-11-03T16:58:00.5274695Z * [new branch] gh/XilunWu/166/head -> origin/gh/XilunWu/166/head 2025-11-03T16:58:00.5276390Z * [new branch] gh/XilunWu/166/orig -> origin/gh/XilunWu/166/orig 2025-11-03T16:58:00.5278890Z * [new branch] gh/XilunWu/168/base -> origin/gh/XilunWu/168/base 2025-11-03T16:58:00.5280553Z * [new branch] gh/XilunWu/168/head -> origin/gh/XilunWu/168/head 2025-11-03T16:58:00.5282259Z * [new branch] gh/XilunWu/168/orig -> origin/gh/XilunWu/168/orig 2025-11-03T16:58:00.5284583Z * [new branch] gh/XilunWu/169/base -> origin/gh/XilunWu/169/base 2025-11-03T16:58:00.5286339Z * [new branch] gh/XilunWu/169/head -> origin/gh/XilunWu/169/head 2025-11-03T16:58:00.5288082Z * [new branch] gh/XilunWu/169/orig -> origin/gh/XilunWu/169/orig 2025-11-03T16:58:00.5290614Z * [new branch] gh/XilunWu/170/base -> origin/gh/XilunWu/170/base 2025-11-03T16:58:00.5292262Z * [new branch] gh/XilunWu/170/head -> origin/gh/XilunWu/170/head 2025-11-03T16:58:00.5293978Z * [new branch] gh/XilunWu/170/orig -> origin/gh/XilunWu/170/orig 2025-11-03T16:58:00.5296394Z * [new branch] gh/XilunWu/171/base -> origin/gh/XilunWu/171/base 2025-11-03T16:58:00.5298107Z * [new branch] gh/XilunWu/171/head -> origin/gh/XilunWu/171/head 2025-11-03T16:58:00.5299823Z * [new branch] gh/XilunWu/171/orig -> origin/gh/XilunWu/171/orig 2025-11-03T16:58:00.5302232Z * [new branch] gh/XilunWu/172/base -> origin/gh/XilunWu/172/base 2025-11-03T16:58:00.5304067Z * [new branch] gh/XilunWu/172/head -> origin/gh/XilunWu/172/head 2025-11-03T16:58:00.5305814Z * [new branch] gh/XilunWu/172/orig -> origin/gh/XilunWu/172/orig 2025-11-03T16:58:00.5308041Z * [new branch] gh/XilunWu/173/base -> origin/gh/XilunWu/173/base 2025-11-03T16:58:00.5309941Z * [new branch] gh/XilunWu/173/head -> origin/gh/XilunWu/173/head 2025-11-03T16:58:00.5313285Z * [new branch] gh/XilunWu/173/orig -> origin/gh/XilunWu/173/orig 2025-11-03T16:58:00.5315665Z * [new branch] gh/XilunWu/175/base -> origin/gh/XilunWu/175/base 2025-11-03T16:58:00.5317518Z * [new branch] gh/XilunWu/175/head -> origin/gh/XilunWu/175/head 2025-11-03T16:58:00.5319274Z * [new branch] gh/XilunWu/175/orig -> origin/gh/XilunWu/175/orig 2025-11-03T16:58:00.5321679Z * [new branch] gh/XilunWu/176/base -> origin/gh/XilunWu/176/base 2025-11-03T16:58:00.5323472Z * [new branch] gh/XilunWu/176/head -> origin/gh/XilunWu/176/head 2025-11-03T16:58:00.5325321Z * [new branch] gh/XilunWu/176/orig -> origin/gh/XilunWu/176/orig 2025-11-03T16:58:00.5327596Z * [new branch] gh/XilunWu/177/base -> origin/gh/XilunWu/177/base 2025-11-03T16:58:00.5329563Z * [new branch] gh/XilunWu/177/head -> origin/gh/XilunWu/177/head 2025-11-03T16:58:00.5331326Z * [new branch] gh/XilunWu/177/orig -> origin/gh/XilunWu/177/orig 2025-11-03T16:58:00.5334211Z * [new branch] gh/XuehaiPan/14/base -> origin/gh/XuehaiPan/14/base 2025-11-03T16:58:00.5335962Z * [new branch] gh/XuehaiPan/14/head -> origin/gh/XuehaiPan/14/head 2025-11-03T16:58:00.5337844Z * [new branch] gh/XuehaiPan/14/orig -> origin/gh/XuehaiPan/14/orig 2025-11-03T16:58:00.5340283Z * [new branch] gh/XuehaiPan/179/base -> origin/gh/XuehaiPan/179/base 2025-11-03T16:58:00.5342017Z * [new branch] gh/XuehaiPan/179/head -> origin/gh/XuehaiPan/179/head 2025-11-03T16:58:00.5343919Z * [new branch] gh/XuehaiPan/179/orig -> origin/gh/XuehaiPan/179/orig 2025-11-03T16:58:00.5346379Z * [new branch] gh/XuehaiPan/189/base -> origin/gh/XuehaiPan/189/base 2025-11-03T16:58:00.5348137Z * [new branch] gh/XuehaiPan/189/head -> origin/gh/XuehaiPan/189/head 2025-11-03T16:58:00.5349923Z * [new branch] gh/XuehaiPan/189/orig -> origin/gh/XuehaiPan/189/orig 2025-11-03T16:58:00.5352258Z * [new branch] gh/XuehaiPan/249/base -> origin/gh/XuehaiPan/249/base 2025-11-03T16:58:00.5354042Z * [new branch] gh/XuehaiPan/249/head -> origin/gh/XuehaiPan/249/head 2025-11-03T16:58:00.5355806Z * [new branch] gh/XuehaiPan/249/orig -> origin/gh/XuehaiPan/249/orig 2025-11-03T16:58:00.5358079Z * [new branch] gh/XuehaiPan/253/base -> origin/gh/XuehaiPan/253/base 2025-11-03T16:58:00.5359953Z * [new branch] gh/XuehaiPan/253/head -> origin/gh/XuehaiPan/253/head 2025-11-03T16:58:00.5361686Z * [new branch] gh/XuehaiPan/253/orig -> origin/gh/XuehaiPan/253/orig 2025-11-03T16:58:00.5364115Z * [new branch] gh/XuehaiPan/254/base -> origin/gh/XuehaiPan/254/base 2025-11-03T16:58:00.5365845Z * [new branch] gh/XuehaiPan/254/head -> origin/gh/XuehaiPan/254/head 2025-11-03T16:58:00.5367883Z * [new branch] gh/XuehaiPan/254/orig -> origin/gh/XuehaiPan/254/orig 2025-11-03T16:58:00.5370248Z * [new branch] gh/XuehaiPan/255/base -> origin/gh/XuehaiPan/255/base 2025-11-03T16:58:00.5371974Z * [new branch] gh/XuehaiPan/255/head -> origin/gh/XuehaiPan/255/head 2025-11-03T16:58:00.5373642Z * [new branch] gh/XuehaiPan/255/orig -> origin/gh/XuehaiPan/255/orig 2025-11-03T16:58:00.5376002Z * [new branch] gh/XuehaiPan/257/base -> origin/gh/XuehaiPan/257/base 2025-11-03T16:58:00.5377780Z * [new branch] gh/XuehaiPan/257/head -> origin/gh/XuehaiPan/257/head 2025-11-03T16:58:00.5379479Z * [new branch] gh/XuehaiPan/257/orig -> origin/gh/XuehaiPan/257/orig 2025-11-03T16:58:00.5381899Z * [new branch] gh/XuehaiPan/271/base -> origin/gh/XuehaiPan/271/base 2025-11-03T16:58:00.5383551Z * [new branch] gh/XuehaiPan/271/head -> origin/gh/XuehaiPan/271/head 2025-11-03T16:58:00.5385287Z * [new branch] gh/XuehaiPan/271/orig -> origin/gh/XuehaiPan/271/orig 2025-11-03T16:58:00.5387812Z * [new branch] gh/XuehaiPan/290/base -> origin/gh/XuehaiPan/290/base 2025-11-03T16:58:00.5389579Z * [new branch] gh/XuehaiPan/290/head -> origin/gh/XuehaiPan/290/head 2025-11-03T16:58:00.5391362Z * [new branch] gh/XuehaiPan/290/orig -> origin/gh/XuehaiPan/290/orig 2025-11-03T16:58:00.5393591Z * [new branch] gh/XuehaiPan/343/base -> origin/gh/XuehaiPan/343/base 2025-11-03T16:58:00.5395335Z * [new branch] gh/XuehaiPan/343/head -> origin/gh/XuehaiPan/343/head 2025-11-03T16:58:00.5397094Z * [new branch] gh/XuehaiPan/343/orig -> origin/gh/XuehaiPan/343/orig 2025-11-03T16:58:00.5399492Z * [new branch] gh/XuehaiPan/347/base -> origin/gh/XuehaiPan/347/base 2025-11-03T16:58:00.5401295Z * [new branch] gh/XuehaiPan/347/head -> origin/gh/XuehaiPan/347/head 2025-11-03T16:58:00.5402980Z * [new branch] gh/XuehaiPan/347/orig -> origin/gh/XuehaiPan/347/orig 2025-11-03T16:58:00.5405474Z * [new branch] gh/XuehaiPan/348/base -> origin/gh/XuehaiPan/348/base 2025-11-03T16:58:00.5407165Z * [new branch] gh/XuehaiPan/348/head -> origin/gh/XuehaiPan/348/head 2025-11-03T16:58:00.5409056Z * [new branch] gh/XuehaiPan/348/orig -> origin/gh/XuehaiPan/348/orig 2025-11-03T16:58:00.5411801Z * [new branch] gh/XuehaiPan/350/base -> origin/gh/XuehaiPan/350/base 2025-11-03T16:58:00.5413501Z * [new branch] gh/XuehaiPan/350/head -> origin/gh/XuehaiPan/350/head 2025-11-03T16:58:00.5415214Z * [new branch] gh/XuehaiPan/350/orig -> origin/gh/XuehaiPan/350/orig 2025-11-03T16:58:00.5417576Z * [new branch] gh/XuehaiPan/356/base -> origin/gh/XuehaiPan/356/base 2025-11-03T16:58:00.5419353Z * [new branch] gh/XuehaiPan/356/head -> origin/gh/XuehaiPan/356/head 2025-11-03T16:58:00.5421047Z * [new branch] gh/XuehaiPan/356/orig -> origin/gh/XuehaiPan/356/orig 2025-11-03T16:58:00.5423387Z * [new branch] gh/XuehaiPan/357/base -> origin/gh/XuehaiPan/357/base 2025-11-03T16:58:00.5425158Z * [new branch] gh/XuehaiPan/357/head -> origin/gh/XuehaiPan/357/head 2025-11-03T16:58:00.5426902Z * [new branch] gh/XuehaiPan/357/orig -> origin/gh/XuehaiPan/357/orig 2025-11-03T16:58:00.5429309Z * [new branch] gh/XuehaiPan/358/base -> origin/gh/XuehaiPan/358/base 2025-11-03T16:58:00.5431082Z * [new branch] gh/XuehaiPan/358/head -> origin/gh/XuehaiPan/358/head 2025-11-03T16:58:00.5432860Z * [new branch] gh/XuehaiPan/358/orig -> origin/gh/XuehaiPan/358/orig 2025-11-03T16:58:00.5435820Z * [new branch] gh/XuehaiPan/359/base -> origin/gh/XuehaiPan/359/base 2025-11-03T16:58:00.5437595Z * [new branch] gh/XuehaiPan/359/head -> origin/gh/XuehaiPan/359/head 2025-11-03T16:58:00.5439333Z * [new branch] gh/XuehaiPan/359/orig -> origin/gh/XuehaiPan/359/orig 2025-11-03T16:58:00.5441689Z * [new branch] gh/XuehaiPan/360/base -> origin/gh/XuehaiPan/360/base 2025-11-03T16:58:00.5443476Z * [new branch] gh/XuehaiPan/360/head -> origin/gh/XuehaiPan/360/head 2025-11-03T16:58:00.5445187Z * [new branch] gh/XuehaiPan/360/orig -> origin/gh/XuehaiPan/360/orig 2025-11-03T16:58:00.5447566Z * [new branch] gh/XuehaiPan/365/base -> origin/gh/XuehaiPan/365/base 2025-11-03T16:58:00.5449465Z * [new branch] gh/XuehaiPan/365/head -> origin/gh/XuehaiPan/365/head 2025-11-03T16:58:00.5451322Z * [new branch] gh/XuehaiPan/365/orig -> origin/gh/XuehaiPan/365/orig 2025-11-03T16:58:00.5453639Z * [new branch] gh/XuehaiPan/366/base -> origin/gh/XuehaiPan/366/base 2025-11-03T16:58:00.5455322Z * [new branch] gh/XuehaiPan/366/head -> origin/gh/XuehaiPan/366/head 2025-11-03T16:58:00.5457963Z * [new branch] gh/XuehaiPan/370/base -> origin/gh/XuehaiPan/370/base 2025-11-03T16:58:00.5459675Z * [new branch] gh/XuehaiPan/370/head -> origin/gh/XuehaiPan/370/head 2025-11-03T16:58:00.5461399Z * [new branch] gh/XuehaiPan/370/orig -> origin/gh/XuehaiPan/370/orig 2025-11-03T16:58:00.5463868Z * [new branch] gh/XuehaiPan/384/base -> origin/gh/XuehaiPan/384/base 2025-11-03T16:58:00.5465619Z * [new branch] gh/XuehaiPan/384/head -> origin/gh/XuehaiPan/384/head 2025-11-03T16:58:00.5467354Z * [new branch] gh/XuehaiPan/384/orig -> origin/gh/XuehaiPan/384/orig 2025-11-03T16:58:00.5469777Z * [new branch] gh/XuehaiPan/390/base -> origin/gh/XuehaiPan/390/base 2025-11-03T16:58:00.5471540Z * [new branch] gh/XuehaiPan/390/head -> origin/gh/XuehaiPan/390/head 2025-11-03T16:58:00.5473421Z * [new branch] gh/XuehaiPan/390/orig -> origin/gh/XuehaiPan/390/orig 2025-11-03T16:58:00.5475821Z * [new branch] gh/XuehaiPan/391/base -> origin/gh/XuehaiPan/391/base 2025-11-03T16:58:00.5477584Z * [new branch] gh/XuehaiPan/391/head -> origin/gh/XuehaiPan/391/head 2025-11-03T16:58:00.5479276Z * [new branch] gh/XuehaiPan/391/orig -> origin/gh/XuehaiPan/391/orig 2025-11-03T16:58:00.5481779Z * [new branch] gh/XuehaiPan/392/base -> origin/gh/XuehaiPan/392/base 2025-11-03T16:58:00.5483485Z * [new branch] gh/XuehaiPan/392/head -> origin/gh/XuehaiPan/392/head 2025-11-03T16:58:00.5485270Z * [new branch] gh/XuehaiPan/392/orig -> origin/gh/XuehaiPan/392/orig 2025-11-03T16:58:00.5487564Z * [new branch] gh/XuehaiPan/393/base -> origin/gh/XuehaiPan/393/base 2025-11-03T16:58:00.5489450Z * [new branch] gh/XuehaiPan/393/head -> origin/gh/XuehaiPan/393/head 2025-11-03T16:58:00.5491183Z * [new branch] gh/XuehaiPan/393/orig -> origin/gh/XuehaiPan/393/orig 2025-11-03T16:58:00.5493542Z * [new branch] gh/XuehaiPan/394/base -> origin/gh/XuehaiPan/394/base 2025-11-03T16:58:00.5495266Z * [new branch] gh/XuehaiPan/394/head -> origin/gh/XuehaiPan/394/head 2025-11-03T16:58:00.5496990Z * [new branch] gh/XuehaiPan/394/orig -> origin/gh/XuehaiPan/394/orig 2025-11-03T16:58:00.5499490Z * [new branch] gh/XuehaiPan/395/base -> origin/gh/XuehaiPan/395/base 2025-11-03T16:58:00.5501162Z * [new branch] gh/XuehaiPan/395/head -> origin/gh/XuehaiPan/395/head 2025-11-03T16:58:00.5502904Z * [new branch] gh/XuehaiPan/395/orig -> origin/gh/XuehaiPan/395/orig 2025-11-03T16:58:00.5505413Z * [new branch] gh/XuehaiPan/396/base -> origin/gh/XuehaiPan/396/base 2025-11-03T16:58:00.5507148Z * [new branch] gh/XuehaiPan/396/orig -> origin/gh/XuehaiPan/396/orig 2025-11-03T16:58:00.5510213Z * [new branch] gh/ZhiweiYan-96/39/base -> origin/gh/ZhiweiYan-96/39/base 2025-11-03T16:58:00.5512011Z * [new branch] gh/ZhiweiYan-96/39/head -> origin/gh/ZhiweiYan-96/39/head 2025-11-03T16:58:00.5513870Z * [new branch] gh/ZhiweiYan-96/39/orig -> origin/gh/ZhiweiYan-96/39/orig 2025-11-03T16:58:00.5516157Z * [new branch] gh/ZhiweiYan-96/44/base -> origin/gh/ZhiweiYan-96/44/base 2025-11-03T16:58:00.5517889Z * [new branch] gh/ZhiweiYan-96/44/head -> origin/gh/ZhiweiYan-96/44/head 2025-11-03T16:58:00.5520168Z * [new branch] gh/ZhiweiYan-96/45/base -> origin/gh/ZhiweiYan-96/45/base 2025-11-03T16:58:00.5521943Z * [new branch] gh/ZhiweiYan-96/45/head -> origin/gh/ZhiweiYan-96/45/head 2025-11-03T16:58:00.5524500Z * [new branch] gh/ZhiweiYan-96/49/base -> origin/gh/ZhiweiYan-96/49/base 2025-11-03T16:58:00.5526013Z * [new branch] gh/ZhiweiYan-96/49/head -> origin/gh/ZhiweiYan-96/49/head 2025-11-03T16:58:00.5528574Z * [new branch] gh/ZhiweiYan-96/62/base -> origin/gh/ZhiweiYan-96/62/base 2025-11-03T16:58:00.5530356Z * [new branch] gh/ZhiweiYan-96/62/head -> origin/gh/ZhiweiYan-96/62/head 2025-11-03T16:58:00.5532668Z * [new branch] gh/ZhiweiYan-96/66/base -> origin/gh/ZhiweiYan-96/66/base 2025-11-03T16:58:00.5534422Z * [new branch] gh/ZhiweiYan-96/66/head -> origin/gh/ZhiweiYan-96/66/head 2025-11-03T16:58:00.5536704Z * [new branch] gh/ZhiweiYan-96/67/base -> origin/gh/ZhiweiYan-96/67/base 2025-11-03T16:58:00.5538513Z * [new branch] gh/ZhiweiYan-96/67/head -> origin/gh/ZhiweiYan-96/67/head 2025-11-03T16:58:00.5540800Z * [new branch] gh/ZhiweiYan-96/68/base -> origin/gh/ZhiweiYan-96/68/base 2025-11-03T16:58:00.5542634Z * [new branch] gh/ZhiweiYan-96/68/head -> origin/gh/ZhiweiYan-96/68/head 2025-11-03T16:58:00.5544390Z * [new branch] gh/ZhiweiYan-96/68/orig -> origin/gh/ZhiweiYan-96/68/orig 2025-11-03T16:58:00.5547332Z * [new branch] gh/aakhundov/1/base -> origin/gh/aakhundov/1/base 2025-11-03T16:58:00.5549228Z * [new branch] gh/aakhundov/1/head -> origin/gh/aakhundov/1/head 2025-11-03T16:58:00.5551447Z * [new branch] gh/aakhundov/2/base -> origin/gh/aakhundov/2/base 2025-11-03T16:58:00.5553269Z * [new branch] gh/aakhundov/2/head -> origin/gh/aakhundov/2/head 2025-11-03T16:58:00.5555593Z * [new branch] gh/aakhundov/3/base -> origin/gh/aakhundov/3/base 2025-11-03T16:58:00.5557339Z * [new branch] gh/aakhundov/3/head -> origin/gh/aakhundov/3/head 2025-11-03T16:58:00.5559059Z * [new branch] gh/aakhundov/3/orig -> origin/gh/aakhundov/3/orig 2025-11-03T16:58:00.5561445Z * [new branch] gh/aditew01/openblas -> origin/gh/aditew01/openblas 2025-11-03T16:58:00.5563140Z * [new branch] gh/aditew01/sbgemm -> origin/gh/aditew01/sbgemm 2025-11-03T16:58:00.5564915Z * [new branch] gh/aditew01/vecbf16 -> origin/gh/aditew01/vecbf16 2025-11-03T16:58:00.5568276Z * [new branch] gh/albanD/1/base -> origin/gh/albanD/1/base 2025-11-03T16:58:00.5570178Z * [new branch] gh/albanD/1/head -> origin/gh/albanD/1/head 2025-11-03T16:58:00.5572003Z * [new branch] gh/albanD/1/orig -> origin/gh/albanD/1/orig 2025-11-03T16:58:00.5574335Z * [new branch] gh/albanD/2/base -> origin/gh/albanD/2/base 2025-11-03T16:58:00.5576062Z * [new branch] gh/albanD/2/head -> origin/gh/albanD/2/head 2025-11-03T16:58:00.5577868Z * [new branch] gh/albanD/2/orig -> origin/gh/albanD/2/orig 2025-11-03T16:58:00.5580195Z * [new branch] gh/albanD/3/base -> origin/gh/albanD/3/base 2025-11-03T16:58:00.5581895Z * [new branch] gh/albanD/3/head -> origin/gh/albanD/3/head 2025-11-03T16:58:00.5583628Z * [new branch] gh/albanD/3/orig -> origin/gh/albanD/3/orig 2025-11-03T16:58:00.5585959Z * [new branch] gh/albanD/4/base -> origin/gh/albanD/4/base 2025-11-03T16:58:00.5587700Z * [new branch] gh/albanD/4/head -> origin/gh/albanD/4/head 2025-11-03T16:58:00.5589428Z * [new branch] gh/albanD/4/orig -> origin/gh/albanD/4/orig 2025-11-03T16:58:00.5592119Z * [new branch] gh/alexbrauckmann/paddedtensor_faketensor_init -> origin/gh/alexbrauckmann/paddedtensor_faketensor_init 2025-11-03T16:58:00.5594947Z * [new branch] gh/alexsamardzic/11/base -> origin/gh/alexsamardzic/11/base 2025-11-03T16:58:00.5596635Z * [new branch] gh/alexsamardzic/11/head -> origin/gh/alexsamardzic/11/head 2025-11-03T16:58:00.5598277Z * [new branch] gh/alexsamardzic/11/orig -> origin/gh/alexsamardzic/11/orig 2025-11-03T16:58:00.5600611Z * [new branch] gh/alexsamardzic/12/base -> origin/gh/alexsamardzic/12/base 2025-11-03T16:58:00.5602401Z * [new branch] gh/alexsamardzic/12/head -> origin/gh/alexsamardzic/12/head 2025-11-03T16:58:00.5604136Z * [new branch] gh/alexsamardzic/12/orig -> origin/gh/alexsamardzic/12/orig 2025-11-03T16:58:00.5606537Z * [new branch] gh/alexsamardzic/13/base -> origin/gh/alexsamardzic/13/base 2025-11-03T16:58:00.5608344Z * [new branch] gh/alexsamardzic/13/head -> origin/gh/alexsamardzic/13/head 2025-11-03T16:58:00.5610252Z * [new branch] gh/alexsamardzic/13/orig -> origin/gh/alexsamardzic/13/orig 2025-11-03T16:58:00.5613130Z * [new branch] gh/alexsamardzic/14/base -> origin/gh/alexsamardzic/14/base 2025-11-03T16:58:00.5614902Z * [new branch] gh/alexsamardzic/14/head -> origin/gh/alexsamardzic/14/head 2025-11-03T16:58:00.5616657Z * [new branch] gh/alexsamardzic/14/orig -> origin/gh/alexsamardzic/14/orig 2025-11-03T16:58:00.5619576Z * [new branch] gh/amjames/18/base -> origin/gh/amjames/18/base 2025-11-03T16:58:00.5621360Z * [new branch] gh/amjames/18/head -> origin/gh/amjames/18/head 2025-11-03T16:58:00.5623053Z * [new branch] gh/amjames/18/orig -> origin/gh/amjames/18/orig 2025-11-03T16:58:00.5626582Z * [new branch] gh/andrewor14/35/base -> origin/gh/andrewor14/35/base 2025-11-03T16:58:00.5628521Z * [new branch] gh/andrewor14/35/head -> origin/gh/andrewor14/35/head 2025-11-03T16:58:00.5630283Z * [new branch] gh/andrewor14/35/orig -> origin/gh/andrewor14/35/orig 2025-11-03T16:58:00.5632740Z * [new branch] gh/andrewor14/50/base -> origin/gh/andrewor14/50/base 2025-11-03T16:58:00.5634546Z * [new branch] gh/andrewor14/50/head -> origin/gh/andrewor14/50/head 2025-11-03T16:58:00.5636444Z * [new branch] gh/andrewor14/50/orig -> origin/gh/andrewor14/50/orig 2025-11-03T16:58:00.5639292Z * [new branch] gh/andyanwang/30/base -> origin/gh/andyanwang/30/base 2025-11-03T16:58:00.5641185Z * [new branch] gh/andyanwang/30/orig -> origin/gh/andyanwang/30/orig 2025-11-03T16:58:00.5643611Z * [new branch] gh/andyanwang/31/base -> origin/gh/andyanwang/31/base 2025-11-03T16:58:00.5645480Z * [new branch] gh/andyanwang/31/orig -> origin/gh/andyanwang/31/orig 2025-11-03T16:58:00.5648328Z * [new branch] gh/andyanwang/32/base -> origin/gh/andyanwang/32/base 2025-11-03T16:58:00.5650267Z * [new branch] gh/andyanwang/32/head -> origin/gh/andyanwang/32/head 2025-11-03T16:58:00.5652032Z * [new branch] gh/andyanwang/32/orig -> origin/gh/andyanwang/32/orig 2025-11-03T16:58:00.5654508Z * [new branch] gh/andyanwang/39/base -> origin/gh/andyanwang/39/base 2025-11-03T16:58:00.5656277Z * [new branch] gh/andyanwang/39/head -> origin/gh/andyanwang/39/head 2025-11-03T16:58:00.5658079Z * [new branch] gh/andyanwang/39/orig -> origin/gh/andyanwang/39/orig 2025-11-03T16:58:00.5660507Z * [new branch] gh/andyanwang/40/base -> origin/gh/andyanwang/40/base 2025-11-03T16:58:00.5662468Z * [new branch] gh/andyanwang/40/head -> origin/gh/andyanwang/40/head 2025-11-03T16:58:00.5664276Z * [new branch] gh/andyanwang/40/orig -> origin/gh/andyanwang/40/orig 2025-11-03T16:58:00.5666960Z * [new branch] gh/andyanwang/41/base -> origin/gh/andyanwang/41/base 2025-11-03T16:58:00.5668563Z * [new branch] gh/andyanwang/41/head -> origin/gh/andyanwang/41/head 2025-11-03T16:58:00.5670274Z * [new branch] gh/andyanwang/41/orig -> origin/gh/andyanwang/41/orig 2025-11-03T16:58:00.5672688Z * [new branch] gh/andyanwang/42/base -> origin/gh/andyanwang/42/base 2025-11-03T16:58:00.5674502Z * [new branch] gh/andyanwang/42/head -> origin/gh/andyanwang/42/head 2025-11-03T16:58:00.5676335Z * [new branch] gh/andyanwang/42/orig -> origin/gh/andyanwang/42/orig 2025-11-03T16:58:00.5678713Z * [new branch] gh/andyanwang/43/base -> origin/gh/andyanwang/43/base 2025-11-03T16:58:00.5680493Z * [new branch] gh/andyanwang/43/head -> origin/gh/andyanwang/43/head 2025-11-03T16:58:00.5682266Z * [new branch] gh/andyanwang/43/orig -> origin/gh/andyanwang/43/orig 2025-11-03T16:58:00.5684809Z * [new branch] gh/andyanwang/44/base -> origin/gh/andyanwang/44/base 2025-11-03T16:58:00.5686643Z * [new branch] gh/andyanwang/44/head -> origin/gh/andyanwang/44/head 2025-11-03T16:58:00.5688457Z * [new branch] gh/andyanwang/44/orig -> origin/gh/andyanwang/44/orig 2025-11-03T16:58:00.5691521Z * [new branch] gh/angelayi/107/base -> origin/gh/angelayi/107/base 2025-11-03T16:58:00.5693225Z * [new branch] gh/angelayi/107/head -> origin/gh/angelayi/107/head 2025-11-03T16:58:00.5695617Z * [new branch] gh/angelayi/114/base -> origin/gh/angelayi/114/base 2025-11-03T16:58:00.5697511Z * [new branch] gh/angelayi/114/head -> origin/gh/angelayi/114/head 2025-11-03T16:58:00.5699259Z * [new branch] gh/angelayi/114/orig -> origin/gh/angelayi/114/orig 2025-11-03T16:58:00.5701629Z * [new branch] gh/angelayi/116/base -> origin/gh/angelayi/116/base 2025-11-03T16:58:00.5703368Z * [new branch] gh/angelayi/116/head -> origin/gh/angelayi/116/head 2025-11-03T16:58:00.5705176Z * [new branch] gh/angelayi/116/orig -> origin/gh/angelayi/116/orig 2025-11-03T16:58:00.5707539Z * [new branch] gh/angelayi/119/base -> origin/gh/angelayi/119/base 2025-11-03T16:58:00.5709286Z * [new branch] gh/angelayi/119/head -> origin/gh/angelayi/119/head 2025-11-03T16:58:00.5711633Z * [new branch] gh/angelayi/119/orig -> origin/gh/angelayi/119/orig 2025-11-03T16:58:00.5713955Z * [new branch] gh/angelayi/120/base -> origin/gh/angelayi/120/base 2025-11-03T16:58:00.5715971Z * [new branch] gh/angelayi/120/head -> origin/gh/angelayi/120/head 2025-11-03T16:58:00.5717502Z * [new branch] gh/angelayi/120/orig -> origin/gh/angelayi/120/orig 2025-11-03T16:58:00.5719935Z * [new branch] gh/angelayi/121/base -> origin/gh/angelayi/121/base 2025-11-03T16:58:00.5721874Z * [new branch] gh/angelayi/121/head -> origin/gh/angelayi/121/head 2025-11-03T16:58:00.5723404Z * [new branch] gh/angelayi/121/orig -> origin/gh/angelayi/121/orig 2025-11-03T16:58:00.5725819Z * [new branch] gh/angelayi/122/base -> origin/gh/angelayi/122/base 2025-11-03T16:58:00.5727482Z * [new branch] gh/angelayi/122/head -> origin/gh/angelayi/122/head 2025-11-03T16:58:00.5729340Z * [new branch] gh/angelayi/122/orig -> origin/gh/angelayi/122/orig 2025-11-03T16:58:00.5731772Z * [new branch] gh/angelayi/124/base -> origin/gh/angelayi/124/base 2025-11-03T16:58:00.5733638Z * [new branch] gh/angelayi/124/head -> origin/gh/angelayi/124/head 2025-11-03T16:58:00.5735401Z * [new branch] gh/angelayi/124/orig -> origin/gh/angelayi/124/orig 2025-11-03T16:58:00.5738151Z * [new branch] gh/angelayi/127/base -> origin/gh/angelayi/127/base 2025-11-03T16:58:00.5739788Z * [new branch] gh/angelayi/127/head -> origin/gh/angelayi/127/head 2025-11-03T16:58:00.5741514Z * [new branch] gh/angelayi/127/orig -> origin/gh/angelayi/127/orig 2025-11-03T16:58:00.5743820Z * [new branch] gh/angelayi/128/base -> origin/gh/angelayi/128/base 2025-11-03T16:58:00.5745473Z * [new branch] gh/angelayi/128/head -> origin/gh/angelayi/128/head 2025-11-03T16:58:00.5747172Z * [new branch] gh/angelayi/128/orig -> origin/gh/angelayi/128/orig 2025-11-03T16:58:00.5749777Z * [new branch] gh/angelayi/129/base -> origin/gh/angelayi/129/base 2025-11-03T16:58:00.5751500Z * [new branch] gh/angelayi/129/head -> origin/gh/angelayi/129/head 2025-11-03T16:58:00.5753520Z * [new branch] gh/angelayi/129/orig -> origin/gh/angelayi/129/orig 2025-11-03T16:58:00.5755935Z * [new branch] gh/angelayi/130/base -> origin/gh/angelayi/130/base 2025-11-03T16:58:00.5757594Z * [new branch] gh/angelayi/130/head -> origin/gh/angelayi/130/head 2025-11-03T16:58:00.5759334Z * [new branch] gh/angelayi/130/orig -> origin/gh/angelayi/130/orig 2025-11-03T16:58:00.5762424Z * [new branch] gh/anijain2305/753/base -> origin/gh/anijain2305/753/base 2025-11-03T16:58:00.5764128Z * [new branch] gh/anijain2305/753/head -> origin/gh/anijain2305/753/head 2025-11-03T16:58:00.5765916Z * [new branch] gh/anijain2305/753/orig -> origin/gh/anijain2305/753/orig 2025-11-03T16:58:00.5768509Z * [new branch] gh/anijain2305/792/base -> origin/gh/anijain2305/792/base 2025-11-03T16:58:00.5770313Z * [new branch] gh/anijain2305/792/head -> origin/gh/anijain2305/792/head 2025-11-03T16:58:00.5772085Z * [new branch] gh/anijain2305/792/orig -> origin/gh/anijain2305/792/orig 2025-11-03T16:58:00.5774367Z * [new branch] gh/anijain2305/805/base -> origin/gh/anijain2305/805/base 2025-11-03T16:58:00.5776179Z * [new branch] gh/anijain2305/805/head -> origin/gh/anijain2305/805/head 2025-11-03T16:58:00.5777912Z * [new branch] gh/anijain2305/805/orig -> origin/gh/anijain2305/805/orig 2025-11-03T16:58:00.5780339Z * [new branch] gh/anijain2305/810/base -> origin/gh/anijain2305/810/base 2025-11-03T16:58:00.5782157Z * [new branch] gh/anijain2305/810/head -> origin/gh/anijain2305/810/head 2025-11-03T16:58:00.5783873Z * [new branch] gh/anijain2305/810/orig -> origin/gh/anijain2305/810/orig 2025-11-03T16:58:00.5786288Z * [new branch] gh/anijain2305/812/base -> origin/gh/anijain2305/812/base 2025-11-03T16:58:00.5788082Z * [new branch] gh/anijain2305/812/head -> origin/gh/anijain2305/812/head 2025-11-03T16:58:00.5789854Z * [new branch] gh/anijain2305/812/orig -> origin/gh/anijain2305/812/orig 2025-11-03T16:58:00.5792240Z * [new branch] gh/anijain2305/854/base -> origin/gh/anijain2305/854/base 2025-11-03T16:58:00.5794008Z * [new branch] gh/anijain2305/854/head -> origin/gh/anijain2305/854/head 2025-11-03T16:58:00.5795756Z * [new branch] gh/anijain2305/854/orig -> origin/gh/anijain2305/854/orig 2025-11-03T16:58:00.5798209Z * [new branch] gh/anijain2305/864/base -> origin/gh/anijain2305/864/base 2025-11-03T16:58:00.5800015Z * [new branch] gh/anijain2305/864/head -> origin/gh/anijain2305/864/head 2025-11-03T16:58:00.5801680Z * [new branch] gh/anijain2305/864/orig -> origin/gh/anijain2305/864/orig 2025-11-03T16:58:00.5804722Z * [new branch] gh/anijain2305/870/base -> origin/gh/anijain2305/870/base 2025-11-03T16:58:00.5805914Z * [new branch] gh/anijain2305/870/head -> origin/gh/anijain2305/870/head 2025-11-03T16:58:00.5807642Z * [new branch] gh/anijain2305/870/orig -> origin/gh/anijain2305/870/orig 2025-11-03T16:58:00.5810557Z * [new branch] gh/anijain2305/873/base -> origin/gh/anijain2305/873/base 2025-11-03T16:58:00.5812195Z * [new branch] gh/anijain2305/873/head -> origin/gh/anijain2305/873/head 2025-11-03T16:58:00.5813952Z * [new branch] gh/anijain2305/873/orig -> origin/gh/anijain2305/873/orig 2025-11-03T16:58:00.5816670Z * [new branch] gh/anijain2305/891/base -> origin/gh/anijain2305/891/base 2025-11-03T16:58:00.5818455Z * [new branch] gh/anijain2305/891/head -> origin/gh/anijain2305/891/head 2025-11-03T16:58:00.5820290Z * [new branch] gh/anijain2305/891/orig -> origin/gh/anijain2305/891/orig 2025-11-03T16:58:00.5822885Z * [new branch] gh/anijain2305/894/base -> origin/gh/anijain2305/894/base 2025-11-03T16:58:00.5825108Z * [new branch] gh/anijain2305/894/head -> origin/gh/anijain2305/894/head 2025-11-03T16:58:00.5826590Z * [new branch] gh/anijain2305/894/orig -> origin/gh/anijain2305/894/orig 2025-11-03T16:58:00.5829065Z * [new branch] gh/anijain2305/895/base -> origin/gh/anijain2305/895/base 2025-11-03T16:58:00.5830895Z * [new branch] gh/anijain2305/895/head -> origin/gh/anijain2305/895/head 2025-11-03T16:58:00.5832523Z * [new branch] gh/anijain2305/895/orig -> origin/gh/anijain2305/895/orig 2025-11-03T16:58:00.5834982Z * [new branch] gh/anijain2305/896/base -> origin/gh/anijain2305/896/base 2025-11-03T16:58:00.5836723Z * [new branch] gh/anijain2305/896/head -> origin/gh/anijain2305/896/head 2025-11-03T16:58:00.5838491Z * [new branch] gh/anijain2305/896/orig -> origin/gh/anijain2305/896/orig 2025-11-03T16:58:00.5841106Z * [new branch] gh/anijain2305/897/base -> origin/gh/anijain2305/897/base 2025-11-03T16:58:00.5842892Z * [new branch] gh/anijain2305/897/head -> origin/gh/anijain2305/897/head 2025-11-03T16:58:00.5844767Z * [new branch] gh/anijain2305/897/orig -> origin/gh/anijain2305/897/orig 2025-11-03T16:58:00.5847288Z * [new branch] gh/anijain2305/898/base -> origin/gh/anijain2305/898/base 2025-11-03T16:58:00.5849251Z * [new branch] gh/anijain2305/898/head -> origin/gh/anijain2305/898/head 2025-11-03T16:58:00.5851083Z * [new branch] gh/anijain2305/898/orig -> origin/gh/anijain2305/898/orig 2025-11-03T16:58:00.5853584Z * [new branch] gh/anijain2305/899/base -> origin/gh/anijain2305/899/base 2025-11-03T16:58:00.5855443Z * [new branch] gh/anijain2305/899/head -> origin/gh/anijain2305/899/head 2025-11-03T16:58:00.5857193Z * [new branch] gh/anijain2305/899/orig -> origin/gh/anijain2305/899/orig 2025-11-03T16:58:00.5860506Z * [new branch] gh/anijain2305/900/base -> origin/gh/anijain2305/900/base 2025-11-03T16:58:00.5862392Z * [new branch] gh/anijain2305/900/head -> origin/gh/anijain2305/900/head 2025-11-03T16:58:00.5864179Z * [new branch] gh/anijain2305/900/orig -> origin/gh/anijain2305/900/orig 2025-11-03T16:58:00.5866487Z * [new branch] gh/anijain2305/901/base -> origin/gh/anijain2305/901/base 2025-11-03T16:58:00.5868256Z * [new branch] gh/anijain2305/901/head -> origin/gh/anijain2305/901/head 2025-11-03T16:58:00.5870067Z * [new branch] gh/anijain2305/901/orig -> origin/gh/anijain2305/901/orig 2025-11-03T16:58:00.5872528Z * [new branch] gh/anijain2305/902/base -> origin/gh/anijain2305/902/base 2025-11-03T16:58:00.5874300Z * [new branch] gh/anijain2305/902/head -> origin/gh/anijain2305/902/head 2025-11-03T16:58:00.5876067Z * [new branch] gh/anijain2305/902/orig -> origin/gh/anijain2305/902/orig 2025-11-03T16:58:00.5878694Z * [new branch] gh/anijain2305/903/base -> origin/gh/anijain2305/903/base 2025-11-03T16:58:00.5880152Z * [new branch] gh/anijain2305/903/head -> origin/gh/anijain2305/903/head 2025-11-03T16:58:00.5882000Z * [new branch] gh/anijain2305/903/orig -> origin/gh/anijain2305/903/orig 2025-11-03T16:58:00.5884578Z * [new branch] gh/anijain2305/904/base -> origin/gh/anijain2305/904/base 2025-11-03T16:58:00.5886314Z * [new branch] gh/anijain2305/904/head -> origin/gh/anijain2305/904/head 2025-11-03T16:58:00.5888067Z * [new branch] gh/anijain2305/904/orig -> origin/gh/anijain2305/904/orig 2025-11-03T16:58:00.5890703Z * [new branch] gh/anijain2305/905/base -> origin/gh/anijain2305/905/base 2025-11-03T16:58:00.5892753Z * [new branch] gh/anijain2305/905/head -> origin/gh/anijain2305/905/head 2025-11-03T16:58:00.5894381Z * [new branch] gh/anijain2305/905/orig -> origin/gh/anijain2305/905/orig 2025-11-03T16:58:00.5896923Z * [new branch] gh/anijain2305/906/base -> origin/gh/anijain2305/906/base 2025-11-03T16:58:00.5898684Z * [new branch] gh/anijain2305/906/head -> origin/gh/anijain2305/906/head 2025-11-03T16:58:00.5900391Z * [new branch] gh/anijain2305/906/orig -> origin/gh/anijain2305/906/orig 2025-11-03T16:58:00.5902902Z * [new branch] gh/anijain2305/907/base -> origin/gh/anijain2305/907/base 2025-11-03T16:58:00.5904686Z * [new branch] gh/anijain2305/907/head -> origin/gh/anijain2305/907/head 2025-11-03T16:58:00.5906409Z * [new branch] gh/anijain2305/907/orig -> origin/gh/anijain2305/907/orig 2025-11-03T16:58:00.5908899Z * [new branch] gh/anijain2305/908/base -> origin/gh/anijain2305/908/base 2025-11-03T16:58:00.5910953Z * [new branch] gh/anijain2305/908/head -> origin/gh/anijain2305/908/head 2025-11-03T16:58:00.5912643Z * [new branch] gh/anijain2305/908/orig -> origin/gh/anijain2305/908/orig 2025-11-03T16:58:00.5915217Z * [new branch] gh/anijain2305/909/base -> origin/gh/anijain2305/909/base 2025-11-03T16:58:00.5916948Z * [new branch] gh/anijain2305/909/head -> origin/gh/anijain2305/909/head 2025-11-03T16:58:00.5918715Z * [new branch] gh/anijain2305/909/orig -> origin/gh/anijain2305/909/orig 2025-11-03T16:58:00.5921256Z * [new branch] gh/anijain2305/910/base -> origin/gh/anijain2305/910/base 2025-11-03T16:58:00.5923002Z * [new branch] gh/anijain2305/910/head -> origin/gh/anijain2305/910/head 2025-11-03T16:58:00.5924740Z * [new branch] gh/anijain2305/910/orig -> origin/gh/anijain2305/910/orig 2025-11-03T16:58:00.5927256Z * [new branch] gh/anijain2305/911/base -> origin/gh/anijain2305/911/base 2025-11-03T16:58:00.5929223Z * [new branch] gh/anijain2305/911/head -> origin/gh/anijain2305/911/head 2025-11-03T16:58:00.5931023Z * [new branch] gh/anijain2305/911/orig -> origin/gh/anijain2305/911/orig 2025-11-03T16:58:00.5933434Z * [new branch] gh/anijain2305/912/base -> origin/gh/anijain2305/912/base 2025-11-03T16:58:00.5935218Z * [new branch] gh/anijain2305/912/head -> origin/gh/anijain2305/912/head 2025-11-03T16:58:00.5936943Z * [new branch] gh/anijain2305/912/orig -> origin/gh/anijain2305/912/orig 2025-11-03T16:58:00.5939390Z * [new branch] gh/anijain2305/913/base -> origin/gh/anijain2305/913/base 2025-11-03T16:58:00.5941201Z * [new branch] gh/anijain2305/913/head -> origin/gh/anijain2305/913/head 2025-11-03T16:58:00.5942933Z * [new branch] gh/anijain2305/913/orig -> origin/gh/anijain2305/913/orig 2025-11-03T16:58:00.5945603Z * [new branch] gh/anijain2305/914/base -> origin/gh/anijain2305/914/base 2025-11-03T16:58:00.5947579Z * [new branch] gh/anijain2305/914/head -> origin/gh/anijain2305/914/head 2025-11-03T16:58:00.5949179Z * [new branch] gh/anijain2305/914/orig -> origin/gh/anijain2305/914/orig 2025-11-03T16:58:00.5951576Z * [new branch] gh/anijain2305/915/base -> origin/gh/anijain2305/915/base 2025-11-03T16:58:00.5953454Z * [new branch] gh/anijain2305/915/head -> origin/gh/anijain2305/915/head 2025-11-03T16:58:00.5955162Z * [new branch] gh/anijain2305/915/orig -> origin/gh/anijain2305/915/orig 2025-11-03T16:58:00.5957400Z * [new branch] gh/anijain2305/916/base -> origin/gh/anijain2305/916/base 2025-11-03T16:58:00.5959217Z * [new branch] gh/anijain2305/916/head -> origin/gh/anijain2305/916/head 2025-11-03T16:58:00.5961172Z * [new branch] gh/anijain2305/916/orig -> origin/gh/anijain2305/916/orig 2025-11-03T16:58:00.5963451Z * [new branch] gh/anijain2305/917/base -> origin/gh/anijain2305/917/base 2025-11-03T16:58:00.5965393Z * [new branch] gh/anijain2305/917/head -> origin/gh/anijain2305/917/head 2025-11-03T16:58:00.5967127Z * [new branch] gh/anijain2305/917/orig -> origin/gh/anijain2305/917/orig 2025-11-03T16:58:00.5969697Z * [new branch] gh/anijain2305/918/base -> origin/gh/anijain2305/918/base 2025-11-03T16:58:00.5971408Z * [new branch] gh/anijain2305/918/head -> origin/gh/anijain2305/918/head 2025-11-03T16:58:00.5973160Z * [new branch] gh/anijain2305/918/orig -> origin/gh/anijain2305/918/orig 2025-11-03T16:58:00.5976098Z * [new branch] gh/anijain2305/919/base -> origin/gh/anijain2305/919/base 2025-11-03T16:58:00.5977424Z * [new branch] gh/anijain2305/919/head -> origin/gh/anijain2305/919/head 2025-11-03T16:58:00.5979267Z * [new branch] gh/anijain2305/919/orig -> origin/gh/anijain2305/919/orig 2025-11-03T16:58:00.5981787Z * [new branch] gh/anijain2305/920/base -> origin/gh/anijain2305/920/base 2025-11-03T16:58:00.5983703Z * [new branch] gh/anijain2305/920/head -> origin/gh/anijain2305/920/head 2025-11-03T16:58:00.5985341Z * [new branch] gh/anijain2305/920/orig -> origin/gh/anijain2305/920/orig 2025-11-03T16:58:00.5987641Z * [new branch] gh/anijain2305/921/base -> origin/gh/anijain2305/921/base 2025-11-03T16:58:00.5989483Z * [new branch] gh/anijain2305/921/head -> origin/gh/anijain2305/921/head 2025-11-03T16:58:00.5991260Z * [new branch] gh/anijain2305/921/orig -> origin/gh/anijain2305/921/orig 2025-11-03T16:58:00.5993580Z * [new branch] gh/anijain2305/922/base -> origin/gh/anijain2305/922/base 2025-11-03T16:58:00.5995346Z * [new branch] gh/anijain2305/922/head -> origin/gh/anijain2305/922/head 2025-11-03T16:58:00.5997083Z * [new branch] gh/anijain2305/922/orig -> origin/gh/anijain2305/922/orig 2025-11-03T16:58:00.5999541Z * [new branch] gh/anijain2305/923/base -> origin/gh/anijain2305/923/base 2025-11-03T16:58:00.6001360Z * [new branch] gh/anijain2305/923/head -> origin/gh/anijain2305/923/head 2025-11-03T16:58:00.6003105Z * [new branch] gh/anijain2305/923/orig -> origin/gh/anijain2305/923/orig 2025-11-03T16:58:00.6005411Z * [new branch] gh/anijain2305/924/base -> origin/gh/anijain2305/924/base 2025-11-03T16:58:00.6007249Z * [new branch] gh/anijain2305/924/head -> origin/gh/anijain2305/924/head 2025-11-03T16:58:00.6009135Z * [new branch] gh/anijain2305/924/orig -> origin/gh/anijain2305/924/orig 2025-11-03T16:58:00.6012151Z * [new branch] gh/anijain2305/925/base -> origin/gh/anijain2305/925/base 2025-11-03T16:58:00.6013946Z * [new branch] gh/anijain2305/925/head -> origin/gh/anijain2305/925/head 2025-11-03T16:58:00.6015865Z * [new branch] gh/anijain2305/925/orig -> origin/gh/anijain2305/925/orig 2025-11-03T16:58:00.6018152Z * [new branch] gh/anijain2305/926/base -> origin/gh/anijain2305/926/base 2025-11-03T16:58:00.6019973Z * [new branch] gh/anijain2305/926/head -> origin/gh/anijain2305/926/head 2025-11-03T16:58:00.6021700Z * [new branch] gh/anijain2305/926/orig -> origin/gh/anijain2305/926/orig 2025-11-03T16:58:00.6024261Z * [new branch] gh/anijain2305/927/base -> origin/gh/anijain2305/927/base 2025-11-03T16:58:00.6026022Z * [new branch] gh/anijain2305/927/head -> origin/gh/anijain2305/927/head 2025-11-03T16:58:00.6027766Z * [new branch] gh/anijain2305/927/orig -> origin/gh/anijain2305/927/orig 2025-11-03T16:58:00.6030340Z * [new branch] gh/anijain2305/928/base -> origin/gh/anijain2305/928/base 2025-11-03T16:58:00.6032077Z * [new branch] gh/anijain2305/928/head -> origin/gh/anijain2305/928/head 2025-11-03T16:58:00.6033698Z * [new branch] gh/anijain2305/928/orig -> origin/gh/anijain2305/928/orig 2025-11-03T16:58:00.6036191Z * [new branch] gh/anijain2305/929/base -> origin/gh/anijain2305/929/base 2025-11-03T16:58:00.6037935Z * [new branch] gh/anijain2305/929/head -> origin/gh/anijain2305/929/head 2025-11-03T16:58:00.6039697Z * [new branch] gh/anijain2305/929/orig -> origin/gh/anijain2305/929/orig 2025-11-03T16:58:00.6042177Z * [new branch] gh/anijain2305/930/base -> origin/gh/anijain2305/930/base 2025-11-03T16:58:00.6043918Z * [new branch] gh/anijain2305/930/head -> origin/gh/anijain2305/930/head 2025-11-03T16:58:00.6045685Z * [new branch] gh/anijain2305/930/orig -> origin/gh/anijain2305/930/orig 2025-11-03T16:58:00.6048137Z * [new branch] gh/anijain2305/931/base -> origin/gh/anijain2305/931/base 2025-11-03T16:58:00.6050035Z * [new branch] gh/anijain2305/931/head -> origin/gh/anijain2305/931/head 2025-11-03T16:58:00.6051723Z * [new branch] gh/anijain2305/931/orig -> origin/gh/anijain2305/931/orig 2025-11-03T16:58:00.6054187Z * [new branch] gh/anijain2305/932/base -> origin/gh/anijain2305/932/base 2025-11-03T16:58:00.6055986Z * [new branch] gh/anijain2305/932/head -> origin/gh/anijain2305/932/head 2025-11-03T16:58:00.6057822Z * [new branch] gh/anijain2305/932/orig -> origin/gh/anijain2305/932/orig 2025-11-03T16:58:00.6060305Z * [new branch] gh/anijain2305/933/base -> origin/gh/anijain2305/933/base 2025-11-03T16:58:00.6062092Z * [new branch] gh/anijain2305/933/head -> origin/gh/anijain2305/933/head 2025-11-03T16:58:00.6063837Z * [new branch] gh/anijain2305/933/orig -> origin/gh/anijain2305/933/orig 2025-11-03T16:58:00.6066302Z * [new branch] gh/anijain2305/934/base -> origin/gh/anijain2305/934/base 2025-11-03T16:58:00.6068063Z * [new branch] gh/anijain2305/934/head -> origin/gh/anijain2305/934/head 2025-11-03T16:58:00.6069786Z * [new branch] gh/anijain2305/934/orig -> origin/gh/anijain2305/934/orig 2025-11-03T16:58:00.6072352Z * [new branch] gh/anijain2305/935/base -> origin/gh/anijain2305/935/base 2025-11-03T16:58:00.6074166Z * [new branch] gh/anijain2305/935/head -> origin/gh/anijain2305/935/head 2025-11-03T16:58:00.6075883Z * [new branch] gh/anijain2305/935/orig -> origin/gh/anijain2305/935/orig 2025-11-03T16:58:00.6078252Z * [new branch] gh/anijain2305/936/base -> origin/gh/anijain2305/936/base 2025-11-03T16:58:00.6079948Z * [new branch] gh/anijain2305/936/head -> origin/gh/anijain2305/936/head 2025-11-03T16:58:00.6081700Z * [new branch] gh/anijain2305/936/orig -> origin/gh/anijain2305/936/orig 2025-11-03T16:58:00.6084341Z * [new branch] gh/anijain2305/937/base -> origin/gh/anijain2305/937/base 2025-11-03T16:58:00.6085787Z * [new branch] gh/anijain2305/937/head -> origin/gh/anijain2305/937/head 2025-11-03T16:58:00.6087758Z * [new branch] gh/anijain2305/937/orig -> origin/gh/anijain2305/937/orig 2025-11-03T16:58:00.6090199Z * [new branch] gh/anijain2305/938/base -> origin/gh/anijain2305/938/base 2025-11-03T16:58:00.6091947Z * [new branch] gh/anijain2305/938/head -> origin/gh/anijain2305/938/head 2025-11-03T16:58:00.6093694Z * [new branch] gh/anijain2305/938/orig -> origin/gh/anijain2305/938/orig 2025-11-03T16:58:00.6095996Z * [new branch] gh/anijain2305/939/base -> origin/gh/anijain2305/939/base 2025-11-03T16:58:00.6097952Z * [new branch] gh/anijain2305/939/head -> origin/gh/anijain2305/939/head 2025-11-03T16:58:00.6099756Z * [new branch] gh/anijain2305/939/orig -> origin/gh/anijain2305/939/orig 2025-11-03T16:58:00.6102122Z * [new branch] gh/anijain2305/940/base -> origin/gh/anijain2305/940/base 2025-11-03T16:58:00.6103880Z * [new branch] gh/anijain2305/940/head -> origin/gh/anijain2305/940/head 2025-11-03T16:58:00.6105649Z * [new branch] gh/anijain2305/940/orig -> origin/gh/anijain2305/940/orig 2025-11-03T16:58:00.6108585Z * [new branch] gh/anjali411/216/base -> origin/gh/anjali411/216/base 2025-11-03T16:58:00.6121086Z * [new branch] gh/anjali411/216/head -> origin/gh/anjali411/216/head 2025-11-03T16:58:00.6121664Z * [new branch] gh/anjali411/216/orig -> origin/gh/anjali411/216/orig 2025-11-03T16:58:00.6122218Z * [new branch] gh/ankitageorge/17/base -> origin/gh/ankitageorge/17/base 2025-11-03T16:58:00.6122955Z * [new branch] gh/ankitageorge/17/head -> origin/gh/ankitageorge/17/head 2025-11-03T16:58:00.6123666Z * [new branch] gh/ankitageorge/17/orig -> origin/gh/ankitageorge/17/orig 2025-11-03T16:58:00.6124260Z * [new branch] gh/anshul-si/1/base -> origin/gh/anshul-si/1/base 2025-11-03T16:58:00.6124712Z * [new branch] gh/anshul-si/1/head -> origin/gh/anshul-si/1/head 2025-11-03T16:58:00.6126944Z * [new branch] gh/anshul-si/2/base -> origin/gh/anshul-si/2/base 2025-11-03T16:58:00.6128665Z * [new branch] gh/anshul-si/2/head -> origin/gh/anshul-si/2/head 2025-11-03T16:58:00.6130985Z * [new branch] gh/anshul-si/3/base -> origin/gh/anshul-si/3/base 2025-11-03T16:58:00.6132624Z * [new branch] gh/anshul-si/3/head -> origin/gh/anshul-si/3/head 2025-11-03T16:58:00.6134831Z * [new branch] gh/anshul-si/4/base -> origin/gh/anshul-si/4/base 2025-11-03T16:58:00.6136649Z * [new branch] gh/anshul-si/4/head -> origin/gh/anshul-si/4/head 2025-11-03T16:58:00.6138820Z * [new branch] gh/anshul-si/5/base -> origin/gh/anshul-si/5/base 2025-11-03T16:58:00.6140601Z * [new branch] gh/anshul-si/5/head -> origin/gh/anshul-si/5/head 2025-11-03T16:58:00.6143021Z * [new branch] gh/anshul-si/51/base -> origin/gh/anshul-si/51/base 2025-11-03T16:58:00.6144774Z * [new branch] gh/anshul-si/51/head -> origin/gh/anshul-si/51/head 2025-11-03T16:58:00.6146535Z * [new branch] gh/anshul-si/51/orig -> origin/gh/anshul-si/51/orig 2025-11-03T16:58:00.6148716Z * [new branch] gh/anshul-si/52/base -> origin/gh/anshul-si/52/base 2025-11-03T16:58:00.6150563Z * [new branch] gh/anshul-si/52/head -> origin/gh/anshul-si/52/head 2025-11-03T16:58:00.6152314Z * [new branch] gh/anshul-si/52/orig -> origin/gh/anshul-si/52/orig 2025-11-03T16:58:00.6154856Z * [new branch] gh/anshul-si/53/base -> origin/gh/anshul-si/53/base 2025-11-03T16:58:00.6156811Z * [new branch] gh/anshul-si/53/head -> origin/gh/anshul-si/53/head 2025-11-03T16:58:00.6158884Z * [new branch] gh/anshul-si/54/base -> origin/gh/anshul-si/54/base 2025-11-03T16:58:00.6160510Z * [new branch] gh/anshul-si/54/head -> origin/gh/anshul-si/54/head 2025-11-03T16:58:00.6162281Z * [new branch] gh/anshul-si/54/orig -> origin/gh/anshul-si/54/orig 2025-11-03T16:58:00.6164549Z * [new branch] gh/anshul-si/55/base -> origin/gh/anshul-si/55/base 2025-11-03T16:58:00.6166250Z * [new branch] gh/anshul-si/55/head -> origin/gh/anshul-si/55/head 2025-11-03T16:58:00.6168390Z * [new branch] gh/anshul-si/55/orig -> origin/gh/anshul-si/55/orig 2025-11-03T16:58:00.6170999Z * [new branch] gh/anshul-si/56/base -> origin/gh/anshul-si/56/base 2025-11-03T16:58:00.6172640Z * [new branch] gh/anshul-si/56/head -> origin/gh/anshul-si/56/head 2025-11-03T16:58:00.6174376Z * [new branch] gh/anshul-si/56/orig -> origin/gh/anshul-si/56/orig 2025-11-03T16:58:00.6176741Z * [new branch] gh/anshul-si/57/base -> origin/gh/anshul-si/57/base 2025-11-03T16:58:00.6178464Z * [new branch] gh/anshul-si/57/head -> origin/gh/anshul-si/57/head 2025-11-03T16:58:00.6180202Z * [new branch] gh/anshul-si/57/orig -> origin/gh/anshul-si/57/orig 2025-11-03T16:58:00.6182419Z * [new branch] gh/anshul-si/58/base -> origin/gh/anshul-si/58/base 2025-11-03T16:58:00.6184238Z * [new branch] gh/anshul-si/58/head -> origin/gh/anshul-si/58/head 2025-11-03T16:58:00.6186446Z * [new branch] gh/anshul-si/59/base -> origin/gh/anshul-si/59/base 2025-11-03T16:58:00.6188299Z * [new branch] gh/anshul-si/59/head -> origin/gh/anshul-si/59/head 2025-11-03T16:58:00.6190062Z * [new branch] gh/anshul-si/59/orig -> origin/gh/anshul-si/59/orig 2025-11-03T16:58:00.6192458Z * [new branch] gh/anshul-si/60/base -> origin/gh/anshul-si/60/base 2025-11-03T16:58:00.6194203Z * [new branch] gh/anshul-si/60/head -> origin/gh/anshul-si/60/head 2025-11-03T16:58:00.6195986Z * [new branch] gh/anshul-si/60/orig -> origin/gh/anshul-si/60/orig 2025-11-03T16:58:00.6198390Z * [new branch] gh/anshul-si/61/base -> origin/gh/anshul-si/61/base 2025-11-03T16:58:00.6200352Z * [new branch] gh/anshul-si/61/head -> origin/gh/anshul-si/61/head 2025-11-03T16:58:00.6202131Z * [new branch] gh/anshul-si/61/orig -> origin/gh/anshul-si/61/orig 2025-11-03T16:58:00.6204314Z * [new branch] gh/anshul-si/62/base -> origin/gh/anshul-si/62/base 2025-11-03T16:58:00.6206089Z * [new branch] gh/anshul-si/62/head -> origin/gh/anshul-si/62/head 2025-11-03T16:58:00.6207803Z * [new branch] gh/anshul-si/62/orig -> origin/gh/anshul-si/62/orig 2025-11-03T16:58:00.6210214Z * [new branch] gh/anshul-si/63/base -> origin/gh/anshul-si/63/base 2025-11-03T16:58:00.6212736Z * [new branch] gh/anshul-si/63/head -> origin/gh/anshul-si/63/head 2025-11-03T16:58:00.6214116Z * [new branch] gh/anshul-si/63/orig -> origin/gh/anshul-si/63/orig 2025-11-03T16:58:00.6216301Z * [new branch] gh/anshul-si/64/base -> origin/gh/anshul-si/64/base 2025-11-03T16:58:00.6218047Z * [new branch] gh/anshul-si/64/head -> origin/gh/anshul-si/64/head 2025-11-03T16:58:00.6219758Z * [new branch] gh/anshul-si/64/orig -> origin/gh/anshul-si/64/orig 2025-11-03T16:58:00.6222561Z * [new branch] gh/anshul-si/65/base -> origin/gh/anshul-si/65/base 2025-11-03T16:58:00.6224254Z * [new branch] gh/anshul-si/65/head -> origin/gh/anshul-si/65/head 2025-11-03T16:58:00.6226051Z * [new branch] gh/anshul-si/65/orig -> origin/gh/anshul-si/65/orig 2025-11-03T16:58:00.6229010Z * [new branch] gh/aorenste/132/base -> origin/gh/aorenste/132/base 2025-11-03T16:58:00.6230746Z * [new branch] gh/aorenste/132/head -> origin/gh/aorenste/132/head 2025-11-03T16:58:00.6233221Z * [new branch] gh/aorenste/133/base -> origin/gh/aorenste/133/base 2025-11-03T16:58:00.6235032Z * [new branch] gh/aorenste/133/head -> origin/gh/aorenste/133/head 2025-11-03T16:58:00.6237076Z * [new branch] gh/aorenste/133/orig -> origin/gh/aorenste/133/orig 2025-11-03T16:58:00.6239561Z * [new branch] gh/aorenste/134/base -> origin/gh/aorenste/134/base 2025-11-03T16:58:00.6241432Z * [new branch] gh/aorenste/134/head -> origin/gh/aorenste/134/head 2025-11-03T16:58:00.6243171Z * [new branch] gh/aorenste/134/orig -> origin/gh/aorenste/134/orig 2025-11-03T16:58:00.6245638Z * [new branch] gh/aorenste/135/base -> origin/gh/aorenste/135/base 2025-11-03T16:58:00.6247354Z * [new branch] gh/aorenste/135/head -> origin/gh/aorenste/135/head 2025-11-03T16:58:00.6249135Z * [new branch] gh/aorenste/135/orig -> origin/gh/aorenste/135/orig 2025-11-03T16:58:00.6251754Z * [new branch] gh/aorenste/136/base -> origin/gh/aorenste/136/base 2025-11-03T16:58:00.6253477Z * [new branch] gh/aorenste/136/head -> origin/gh/aorenste/136/head 2025-11-03T16:58:00.6255207Z * [new branch] gh/aorenste/136/orig -> origin/gh/aorenste/136/orig 2025-11-03T16:58:00.6257684Z * [new branch] gh/aorenste/137/base -> origin/gh/aorenste/137/base 2025-11-03T16:58:00.6259554Z * [new branch] gh/aorenste/137/head -> origin/gh/aorenste/137/head 2025-11-03T16:58:00.6261299Z * [new branch] gh/aorenste/137/orig -> origin/gh/aorenste/137/orig 2025-11-03T16:58:00.6263811Z * [new branch] gh/aorenste/138/base -> origin/gh/aorenste/138/base 2025-11-03T16:58:00.6265499Z * [new branch] gh/aorenste/138/head -> origin/gh/aorenste/138/head 2025-11-03T16:58:00.6267288Z * [new branch] gh/aorenste/138/orig -> origin/gh/aorenste/138/orig 2025-11-03T16:58:00.6269719Z * [new branch] gh/aorenste/139/base -> origin/gh/aorenste/139/base 2025-11-03T16:58:00.6271446Z * [new branch] gh/aorenste/139/head -> origin/gh/aorenste/139/head 2025-11-03T16:58:00.6273208Z * [new branch] gh/aorenste/139/orig -> origin/gh/aorenste/139/orig 2025-11-03T16:58:00.6275623Z * [new branch] gh/aorenste/140/base -> origin/gh/aorenste/140/base 2025-11-03T16:58:00.6277375Z * [new branch] gh/aorenste/140/head -> origin/gh/aorenste/140/head 2025-11-03T16:58:00.6279171Z * [new branch] gh/aorenste/140/orig -> origin/gh/aorenste/140/orig 2025-11-03T16:58:00.6281545Z * [new branch] gh/aorenste/141/base -> origin/gh/aorenste/141/base 2025-11-03T16:58:00.6283355Z * [new branch] gh/aorenste/141/head -> origin/gh/aorenste/141/head 2025-11-03T16:58:00.6285791Z * [new branch] gh/aorenste/142/base -> origin/gh/aorenste/142/base 2025-11-03T16:58:00.6287682Z * [new branch] gh/aorenste/142/head -> origin/gh/aorenste/142/head 2025-11-03T16:58:00.6289515Z * [new branch] gh/aorenste/142/orig -> origin/gh/aorenste/142/orig 2025-11-03T16:58:00.6292085Z * [new branch] gh/aorenste/143/base -> origin/gh/aorenste/143/base 2025-11-03T16:58:00.6293709Z * [new branch] gh/aorenste/143/head -> origin/gh/aorenste/143/head 2025-11-03T16:58:00.6295470Z * [new branch] gh/aorenste/143/orig -> origin/gh/aorenste/143/orig 2025-11-03T16:58:00.6297898Z * [new branch] gh/aorenste/144/base -> origin/gh/aorenste/144/base 2025-11-03T16:58:00.6299579Z * [new branch] gh/aorenste/144/head -> origin/gh/aorenste/144/head 2025-11-03T16:58:00.6301361Z * [new branch] gh/aorenste/144/orig -> origin/gh/aorenste/144/orig 2025-11-03T16:58:00.6303753Z * [new branch] gh/aorenste/145/base -> origin/gh/aorenste/145/base 2025-11-03T16:58:00.6305704Z * [new branch] gh/aorenste/145/head -> origin/gh/aorenste/145/head 2025-11-03T16:58:00.6307614Z * [new branch] gh/aorenste/145/orig -> origin/gh/aorenste/145/orig 2025-11-03T16:58:00.6310828Z * [new branch] gh/avikchaudhuri/1/base -> origin/gh/avikchaudhuri/1/base 2025-11-03T16:58:00.6312513Z * [new branch] gh/avikchaudhuri/1/head -> origin/gh/avikchaudhuri/1/head 2025-11-03T16:58:00.6314725Z * [new branch] gh/avikchaudhuri/2/base -> origin/gh/avikchaudhuri/2/base 2025-11-03T16:58:00.6316455Z * [new branch] gh/avikchaudhuri/2/head -> origin/gh/avikchaudhuri/2/head 2025-11-03T16:58:00.6318075Z * [new branch] gh/avikchaudhuri/2/orig -> origin/gh/avikchaudhuri/2/orig 2025-11-03T16:58:00.6321203Z * [new branch] gh/bdhirsh/666/base -> origin/gh/bdhirsh/666/base 2025-11-03T16:58:00.6322913Z * [new branch] gh/bdhirsh/666/head -> origin/gh/bdhirsh/666/head 2025-11-03T16:58:00.6324734Z * [new branch] gh/bdhirsh/666/orig -> origin/gh/bdhirsh/666/orig 2025-11-03T16:58:00.6327294Z * [new branch] gh/bdhirsh/668/base -> origin/gh/bdhirsh/668/base 2025-11-03T16:58:00.6329242Z * [new branch] gh/bdhirsh/668/head -> origin/gh/bdhirsh/668/head 2025-11-03T16:58:00.6330972Z * [new branch] gh/bdhirsh/668/orig -> origin/gh/bdhirsh/668/orig 2025-11-03T16:58:00.6333479Z * [new branch] gh/bdhirsh/669/base -> origin/gh/bdhirsh/669/base 2025-11-03T16:58:00.6335208Z * [new branch] gh/bdhirsh/669/head -> origin/gh/bdhirsh/669/head 2025-11-03T16:58:00.6337015Z * [new branch] gh/bdhirsh/669/orig -> origin/gh/bdhirsh/669/orig 2025-11-03T16:58:00.6339556Z * [new branch] gh/bdhirsh/670/base -> origin/gh/bdhirsh/670/base 2025-11-03T16:58:00.6341392Z * [new branch] gh/bdhirsh/670/head -> origin/gh/bdhirsh/670/head 2025-11-03T16:58:00.6343150Z * [new branch] gh/bdhirsh/670/orig -> origin/gh/bdhirsh/670/orig 2025-11-03T16:58:00.6345654Z * [new branch] gh/bdhirsh/671/base -> origin/gh/bdhirsh/671/base 2025-11-03T16:58:00.6347447Z * [new branch] gh/bdhirsh/671/head -> origin/gh/bdhirsh/671/head 2025-11-03T16:58:00.6349164Z * [new branch] gh/bdhirsh/671/orig -> origin/gh/bdhirsh/671/orig 2025-11-03T16:58:00.6351595Z * [new branch] gh/bdhirsh/672/base -> origin/gh/bdhirsh/672/base 2025-11-03T16:58:00.6353414Z * [new branch] gh/bdhirsh/672/head -> origin/gh/bdhirsh/672/head 2025-11-03T16:58:00.6355555Z * [new branch] gh/bdhirsh/672/orig -> origin/gh/bdhirsh/672/orig 2025-11-03T16:58:00.6357614Z * [new branch] gh/bdhirsh/673/base -> origin/gh/bdhirsh/673/base 2025-11-03T16:58:00.6359327Z * [new branch] gh/bdhirsh/673/head -> origin/gh/bdhirsh/673/head 2025-11-03T16:58:00.6361182Z * [new branch] gh/bdhirsh/673/orig -> origin/gh/bdhirsh/673/orig 2025-11-03T16:58:00.6363622Z * [new branch] gh/bdhirsh/674/base -> origin/gh/bdhirsh/674/base 2025-11-03T16:58:00.6365443Z * [new branch] gh/bdhirsh/674/head -> origin/gh/bdhirsh/674/head 2025-11-03T16:58:00.6367194Z * [new branch] gh/bdhirsh/674/orig -> origin/gh/bdhirsh/674/orig 2025-11-03T16:58:00.6369808Z * [new branch] gh/bdhirsh/675/base -> origin/gh/bdhirsh/675/base 2025-11-03T16:58:00.6371441Z * [new branch] gh/bdhirsh/675/head -> origin/gh/bdhirsh/675/head 2025-11-03T16:58:00.6373145Z * [new branch] gh/bdhirsh/675/orig -> origin/gh/bdhirsh/675/orig 2025-11-03T16:58:00.6376233Z * [new branch] gh/benjaminglass1/101/base -> origin/gh/benjaminglass1/101/base 2025-11-03T16:58:00.6378057Z * [new branch] gh/benjaminglass1/101/head -> origin/gh/benjaminglass1/101/head 2025-11-03T16:58:00.6379846Z * [new branch] gh/benjaminglass1/101/orig -> origin/gh/benjaminglass1/101/orig 2025-11-03T16:58:00.6382201Z * [new branch] gh/benjaminglass1/102/base -> origin/gh/benjaminglass1/102/base 2025-11-03T16:58:00.6383911Z * [new branch] gh/benjaminglass1/102/head -> origin/gh/benjaminglass1/102/head 2025-11-03T16:58:00.6385664Z * [new branch] gh/benjaminglass1/102/orig -> origin/gh/benjaminglass1/102/orig 2025-11-03T16:58:00.6387911Z * [new branch] gh/benjaminglass1/106/base -> origin/gh/benjaminglass1/106/base 2025-11-03T16:58:00.6389761Z * [new branch] gh/benjaminglass1/106/head -> origin/gh/benjaminglass1/106/head 2025-11-03T16:58:00.6391908Z * [new branch] gh/benjaminglass1/106/orig -> origin/gh/benjaminglass1/106/orig 2025-11-03T16:58:00.6393868Z * [new branch] gh/benjaminglass1/107/base -> origin/gh/benjaminglass1/107/base 2025-11-03T16:58:00.6395558Z * [new branch] gh/benjaminglass1/107/head -> origin/gh/benjaminglass1/107/head 2025-11-03T16:58:00.6397302Z * [new branch] gh/benjaminglass1/107/orig -> origin/gh/benjaminglass1/107/orig 2025-11-03T16:58:00.6399735Z * [new branch] gh/benjaminglass1/108/base -> origin/gh/benjaminglass1/108/base 2025-11-03T16:58:00.6401421Z * [new branch] gh/benjaminglass1/108/head -> origin/gh/benjaminglass1/108/head 2025-11-03T16:58:00.6403155Z * [new branch] gh/benjaminglass1/108/orig -> origin/gh/benjaminglass1/108/orig 2025-11-03T16:58:00.6405528Z * [new branch] gh/benjaminglass1/79/base -> origin/gh/benjaminglass1/79/base 2025-11-03T16:58:00.6407238Z * [new branch] gh/benjaminglass1/79/head -> origin/gh/benjaminglass1/79/head 2025-11-03T16:58:00.6409076Z * [new branch] gh/benjaminglass1/79/orig -> origin/gh/benjaminglass1/79/orig 2025-11-03T16:58:00.6411746Z * [new branch] gh/benjaminglass1/86/base -> origin/gh/benjaminglass1/86/base 2025-11-03T16:58:00.6413451Z * [new branch] gh/benjaminglass1/86/head -> origin/gh/benjaminglass1/86/head 2025-11-03T16:58:00.6415245Z * [new branch] gh/benjaminglass1/86/orig -> origin/gh/benjaminglass1/86/orig 2025-11-03T16:58:00.6417584Z * [new branch] gh/benjaminglass1/97/base -> origin/gh/benjaminglass1/97/base 2025-11-03T16:58:00.6419403Z * [new branch] gh/benjaminglass1/97/head -> origin/gh/benjaminglass1/97/head 2025-11-03T16:58:00.6421149Z * [new branch] gh/benjaminglass1/97/orig -> origin/gh/benjaminglass1/97/orig 2025-11-03T16:58:00.6423977Z * [new branch] gh/bobrenjc93/570/base -> origin/gh/bobrenjc93/570/base 2025-11-03T16:58:00.6425769Z * [new branch] gh/bobrenjc93/570/head -> origin/gh/bobrenjc93/570/head 2025-11-03T16:58:00.6427578Z * [new branch] gh/bobrenjc93/570/orig -> origin/gh/bobrenjc93/570/orig 2025-11-03T16:58:00.6429863Z * [new branch] gh/bobrenjc93/604/base -> origin/gh/bobrenjc93/604/base 2025-11-03T16:58:00.6431616Z * [new branch] gh/bobrenjc93/604/head -> origin/gh/bobrenjc93/604/head 2025-11-03T16:58:00.6433347Z * [new branch] gh/bobrenjc93/604/orig -> origin/gh/bobrenjc93/604/orig 2025-11-03T16:58:00.6435908Z * [new branch] gh/bobrenjc93/625/base -> origin/gh/bobrenjc93/625/base 2025-11-03T16:58:00.6437737Z * [new branch] gh/bobrenjc93/625/head -> origin/gh/bobrenjc93/625/head 2025-11-03T16:58:00.6439330Z * [new branch] gh/bobrenjc93/625/orig -> origin/gh/bobrenjc93/625/orig 2025-11-03T16:58:00.6441636Z * [new branch] gh/bobrenjc93/626/base -> origin/gh/bobrenjc93/626/base 2025-11-03T16:58:00.6443537Z * [new branch] gh/bobrenjc93/626/head -> origin/gh/bobrenjc93/626/head 2025-11-03T16:58:00.6445198Z * [new branch] gh/bobrenjc93/626/orig -> origin/gh/bobrenjc93/626/orig 2025-11-03T16:58:00.6447686Z * [new branch] gh/bobrenjc93/627/base -> origin/gh/bobrenjc93/627/base 2025-11-03T16:58:00.6449577Z * [new branch] gh/bobrenjc93/627/head -> origin/gh/bobrenjc93/627/head 2025-11-03T16:58:00.6452477Z * [new branch] gh/bobrenjc93/627/orig -> origin/gh/bobrenjc93/627/orig 2025-11-03T16:58:00.6454498Z * [new branch] gh/bobrenjc93/630/base -> origin/gh/bobrenjc93/630/base 2025-11-03T16:58:00.6455326Z * [new branch] gh/bobrenjc93/630/head -> origin/gh/bobrenjc93/630/head 2025-11-03T16:58:00.6457116Z * [new branch] gh/bobrenjc93/630/orig -> origin/gh/bobrenjc93/630/orig 2025-11-03T16:58:00.6459479Z * [new branch] gh/bobrenjc93/631/base -> origin/gh/bobrenjc93/631/base 2025-11-03T16:58:00.6461207Z * [new branch] gh/bobrenjc93/631/head -> origin/gh/bobrenjc93/631/head 2025-11-03T16:58:00.6462915Z * [new branch] gh/bobrenjc93/631/orig -> origin/gh/bobrenjc93/631/orig 2025-11-03T16:58:00.6465261Z * [new branch] gh/bobrenjc93/632/base -> origin/gh/bobrenjc93/632/base 2025-11-03T16:58:00.6466986Z * [new branch] gh/bobrenjc93/632/head -> origin/gh/bobrenjc93/632/head 2025-11-03T16:58:00.6468673Z * [new branch] gh/bobrenjc93/632/orig -> origin/gh/bobrenjc93/632/orig 2025-11-03T16:58:00.6471064Z * [new branch] gh/bobrenjc93/633/base -> origin/gh/bobrenjc93/633/base 2025-11-03T16:58:00.6472832Z * [new branch] gh/bobrenjc93/633/head -> origin/gh/bobrenjc93/633/head 2025-11-03T16:58:00.6474556Z * [new branch] gh/bobrenjc93/633/orig -> origin/gh/bobrenjc93/633/orig 2025-11-03T16:58:00.6476845Z * [new branch] gh/bobrenjc93/634/base -> origin/gh/bobrenjc93/634/base 2025-11-03T16:58:00.6478609Z * [new branch] gh/bobrenjc93/634/head -> origin/gh/bobrenjc93/634/head 2025-11-03T16:58:00.6480308Z * [new branch] gh/bobrenjc93/634/orig -> origin/gh/bobrenjc93/634/orig 2025-11-03T16:58:00.6482598Z * [new branch] gh/bobrenjc93/635/base -> origin/gh/bobrenjc93/635/base 2025-11-03T16:58:00.6484322Z * [new branch] gh/bobrenjc93/635/head -> origin/gh/bobrenjc93/635/head 2025-11-03T16:58:00.6486064Z * [new branch] gh/bobrenjc93/635/orig -> origin/gh/bobrenjc93/635/orig 2025-11-03T16:58:00.6488366Z * [new branch] gh/bobrenjc93/636/base -> origin/gh/bobrenjc93/636/base 2025-11-03T16:58:00.6490254Z * [new branch] gh/bobrenjc93/636/head -> origin/gh/bobrenjc93/636/head 2025-11-03T16:58:00.6491981Z * [new branch] gh/bobrenjc93/636/orig -> origin/gh/bobrenjc93/636/orig 2025-11-03T16:58:00.6494420Z * [new branch] gh/bobrenjc93/637/base -> origin/gh/bobrenjc93/637/base 2025-11-03T16:58:00.6496112Z * [new branch] gh/bobrenjc93/637/head -> origin/gh/bobrenjc93/637/head 2025-11-03T16:58:00.6497868Z * [new branch] gh/bobrenjc93/637/orig -> origin/gh/bobrenjc93/637/orig 2025-11-03T16:58:00.6500204Z * [new branch] gh/bobrenjc93/638/base -> origin/gh/bobrenjc93/638/base 2025-11-03T16:58:00.6501933Z * [new branch] gh/bobrenjc93/638/head -> origin/gh/bobrenjc93/638/head 2025-11-03T16:58:00.6503761Z * [new branch] gh/bobrenjc93/638/orig -> origin/gh/bobrenjc93/638/orig 2025-11-03T16:58:00.6506035Z * [new branch] gh/bobrenjc93/639/base -> origin/gh/bobrenjc93/639/base 2025-11-03T16:58:00.6507807Z * [new branch] gh/bobrenjc93/639/head -> origin/gh/bobrenjc93/639/head 2025-11-03T16:58:00.6510894Z * [new branch] gh/bobrenjc93/639/orig -> origin/gh/bobrenjc93/639/orig 2025-11-03T16:58:00.6514663Z * [new branch] gh/bobrenjc93/640/base -> origin/gh/bobrenjc93/640/base 2025-11-03T16:58:00.6515790Z * [new branch] gh/bobrenjc93/640/head -> origin/gh/bobrenjc93/640/head 2025-11-03T16:58:00.6517318Z * [new branch] gh/bobrenjc93/640/orig -> origin/gh/bobrenjc93/640/orig 2025-11-03T16:58:00.6520206Z * [new branch] gh/bobrenjc93/641/base -> origin/gh/bobrenjc93/641/base 2025-11-03T16:58:00.6521908Z * [new branch] gh/bobrenjc93/641/head -> origin/gh/bobrenjc93/641/head 2025-11-03T16:58:00.6523648Z * [new branch] gh/bobrenjc93/641/orig -> origin/gh/bobrenjc93/641/orig 2025-11-03T16:58:00.6526087Z * [new branch] gh/bobrenjc93/642/base -> origin/gh/bobrenjc93/642/base 2025-11-03T16:58:00.6527866Z * [new branch] gh/bobrenjc93/642/head -> origin/gh/bobrenjc93/642/head 2025-11-03T16:58:00.6529728Z * [new branch] gh/bobrenjc93/642/orig -> origin/gh/bobrenjc93/642/orig 2025-11-03T16:58:00.6532099Z * [new branch] gh/bobrenjc93/643/base -> origin/gh/bobrenjc93/643/base 2025-11-03T16:58:00.6533810Z * [new branch] gh/bobrenjc93/643/head -> origin/gh/bobrenjc93/643/head 2025-11-03T16:58:00.6535528Z * [new branch] gh/bobrenjc93/643/orig -> origin/gh/bobrenjc93/643/orig 2025-11-03T16:58:00.6537895Z * [new branch] gh/bobrenjc93/644/base -> origin/gh/bobrenjc93/644/base 2025-11-03T16:58:00.6539653Z * [new branch] gh/bobrenjc93/644/head -> origin/gh/bobrenjc93/644/head 2025-11-03T16:58:00.6541375Z * [new branch] gh/bobrenjc93/644/orig -> origin/gh/bobrenjc93/644/orig 2025-11-03T16:58:00.6543882Z * [new branch] gh/bobrenjc93/645/base -> origin/gh/bobrenjc93/645/base 2025-11-03T16:58:00.6545658Z * [new branch] gh/bobrenjc93/645/head -> origin/gh/bobrenjc93/645/head 2025-11-03T16:58:00.6547394Z * [new branch] gh/bobrenjc93/645/orig -> origin/gh/bobrenjc93/645/orig 2025-11-03T16:58:00.6549848Z * [new branch] gh/bobrenjc93/646/base -> origin/gh/bobrenjc93/646/base 2025-11-03T16:58:00.6551569Z * [new branch] gh/bobrenjc93/646/head -> origin/gh/bobrenjc93/646/head 2025-11-03T16:58:00.6553319Z * [new branch] gh/bobrenjc93/646/orig -> origin/gh/bobrenjc93/646/orig 2025-11-03T16:58:00.6555916Z * [new branch] gh/bobrenjc93/647/base -> origin/gh/bobrenjc93/647/base 2025-11-03T16:58:00.6557706Z * [new branch] gh/bobrenjc93/647/head -> origin/gh/bobrenjc93/647/head 2025-11-03T16:58:00.6559412Z * [new branch] gh/bobrenjc93/647/orig -> origin/gh/bobrenjc93/647/orig 2025-11-03T16:58:00.6561705Z * [new branch] gh/bobrenjc93/648/base -> origin/gh/bobrenjc93/648/base 2025-11-03T16:58:00.6563379Z * [new branch] gh/bobrenjc93/648/head -> origin/gh/bobrenjc93/648/head 2025-11-03T16:58:00.6565027Z * [new branch] gh/bobrenjc93/648/orig -> origin/gh/bobrenjc93/648/orig 2025-11-03T16:58:00.6568148Z * [new branch] gh/bobrenjc93/649/base -> origin/gh/bobrenjc93/649/base 2025-11-03T16:58:00.6569674Z * [new branch] gh/bobrenjc93/649/head -> origin/gh/bobrenjc93/649/head 2025-11-03T16:58:00.6571320Z * [new branch] gh/bobrenjc93/649/orig -> origin/gh/bobrenjc93/649/orig 2025-11-03T16:58:00.6574150Z * [new branch] gh/bobrenjc93/650/base -> origin/gh/bobrenjc93/650/base 2025-11-03T16:58:00.6575680Z * [new branch] gh/bobrenjc93/650/head -> origin/gh/bobrenjc93/650/head 2025-11-03T16:58:00.6577288Z * [new branch] gh/bobrenjc93/650/orig -> origin/gh/bobrenjc93/650/orig 2025-11-03T16:58:00.6579791Z * [new branch] gh/bobrenjc93/651/base -> origin/gh/bobrenjc93/651/base 2025-11-03T16:58:00.6581789Z * [new branch] gh/bobrenjc93/651/head -> origin/gh/bobrenjc93/651/head 2025-11-03T16:58:00.6583496Z * [new branch] gh/bobrenjc93/651/orig -> origin/gh/bobrenjc93/651/orig 2025-11-03T16:58:00.6585772Z * [new branch] gh/bobrenjc93/652/base -> origin/gh/bobrenjc93/652/base 2025-11-03T16:58:00.6588190Z * [new branch] gh/bobrenjc93/652/head -> origin/gh/bobrenjc93/652/head 2025-11-03T16:58:00.6589380Z * [new branch] gh/bobrenjc93/652/orig -> origin/gh/bobrenjc93/652/orig 2025-11-03T16:58:00.6592927Z * [new branch] gh/bobrenjc93/653/base -> origin/gh/bobrenjc93/653/base 2025-11-03T16:58:00.6594194Z * [new branch] gh/bobrenjc93/653/head -> origin/gh/bobrenjc93/653/head 2025-11-03T16:58:00.6595868Z * [new branch] gh/bobrenjc93/653/orig -> origin/gh/bobrenjc93/653/orig 2025-11-03T16:58:00.6598245Z * [new branch] gh/bobrenjc93/654/base -> origin/gh/bobrenjc93/654/base 2025-11-03T16:58:00.6600002Z * [new branch] gh/bobrenjc93/654/head -> origin/gh/bobrenjc93/654/head 2025-11-03T16:58:00.6601756Z * [new branch] gh/bobrenjc93/654/orig -> origin/gh/bobrenjc93/654/orig 2025-11-03T16:58:00.6604175Z * [new branch] gh/bobrenjc93/655/base -> origin/gh/bobrenjc93/655/base 2025-11-03T16:58:00.6605908Z * [new branch] gh/bobrenjc93/655/head -> origin/gh/bobrenjc93/655/head 2025-11-03T16:58:00.6607612Z * [new branch] gh/bobrenjc93/655/orig -> origin/gh/bobrenjc93/655/orig 2025-11-03T16:58:00.6610278Z * [new branch] gh/bobrenjc93/656/base -> origin/gh/bobrenjc93/656/base 2025-11-03T16:58:00.6612133Z * [new branch] gh/bobrenjc93/656/head -> origin/gh/bobrenjc93/656/head 2025-11-03T16:58:00.6613827Z * [new branch] gh/bobrenjc93/656/orig -> origin/gh/bobrenjc93/656/orig 2025-11-03T16:58:00.6616335Z * [new branch] gh/bobrenjc93/657/base -> origin/gh/bobrenjc93/657/base 2025-11-03T16:58:00.6618115Z * [new branch] gh/bobrenjc93/657/head -> origin/gh/bobrenjc93/657/head 2025-11-03T16:58:00.6619823Z * [new branch] gh/bobrenjc93/657/orig -> origin/gh/bobrenjc93/657/orig 2025-11-03T16:58:00.6622207Z * [new branch] gh/bobrenjc93/658/base -> origin/gh/bobrenjc93/658/base 2025-11-03T16:58:00.6623979Z * [new branch] gh/bobrenjc93/658/head -> origin/gh/bobrenjc93/658/head 2025-11-03T16:58:00.6625648Z * [new branch] gh/bobrenjc93/658/orig -> origin/gh/bobrenjc93/658/orig 2025-11-03T16:58:00.6628097Z * [new branch] gh/bobrenjc93/659/base -> origin/gh/bobrenjc93/659/base 2025-11-03T16:58:00.6629879Z * [new branch] gh/bobrenjc93/659/head -> origin/gh/bobrenjc93/659/head 2025-11-03T16:58:00.6631569Z * [new branch] gh/bobrenjc93/659/orig -> origin/gh/bobrenjc93/659/orig 2025-11-03T16:58:00.6634125Z * [new branch] gh/bobrenjc93/660/base -> origin/gh/bobrenjc93/660/base 2025-11-03T16:58:00.6635829Z * [new branch] gh/bobrenjc93/660/head -> origin/gh/bobrenjc93/660/head 2025-11-03T16:58:00.6637639Z * [new branch] gh/bobrenjc93/660/orig -> origin/gh/bobrenjc93/660/orig 2025-11-03T16:58:00.6640155Z * [new branch] gh/bobrenjc93/661/base -> origin/gh/bobrenjc93/661/base 2025-11-03T16:58:00.6641917Z * [new branch] gh/bobrenjc93/661/head -> origin/gh/bobrenjc93/661/head 2025-11-03T16:58:00.6644018Z * [new branch] gh/bobrenjc93/661/orig -> origin/gh/bobrenjc93/661/orig 2025-11-03T16:58:00.6646334Z * [new branch] gh/bobrenjc93/662/base -> origin/gh/bobrenjc93/662/base 2025-11-03T16:58:00.6648082Z * [new branch] gh/bobrenjc93/662/head -> origin/gh/bobrenjc93/662/head 2025-11-03T16:58:00.6650175Z * [new branch] gh/bobrenjc93/662/orig -> origin/gh/bobrenjc93/662/orig 2025-11-03T16:58:00.6652580Z * [new branch] gh/bobrenjc93/663/base -> origin/gh/bobrenjc93/663/base 2025-11-03T16:58:00.6654267Z * [new branch] gh/bobrenjc93/663/head -> origin/gh/bobrenjc93/663/head 2025-11-03T16:58:00.6656053Z * [new branch] gh/bobrenjc93/663/orig -> origin/gh/bobrenjc93/663/orig 2025-11-03T16:58:00.6658510Z * [new branch] gh/bobrenjc93/664/base -> origin/gh/bobrenjc93/664/base 2025-11-03T16:58:00.6660853Z * [new branch] gh/bobrenjc93/664/head -> origin/gh/bobrenjc93/664/head 2025-11-03T16:58:00.6662670Z * [new branch] gh/bobrenjc93/664/orig -> origin/gh/bobrenjc93/664/orig 2025-11-03T16:58:00.6665187Z * [new branch] gh/bobrenjc93/665/base -> origin/gh/bobrenjc93/665/base 2025-11-03T16:58:00.6666966Z * [new branch] gh/bobrenjc93/665/head -> origin/gh/bobrenjc93/665/head 2025-11-03T16:58:00.6668706Z * [new branch] gh/bobrenjc93/665/orig -> origin/gh/bobrenjc93/665/orig 2025-11-03T16:58:00.6671111Z * [new branch] gh/bobrenjc93/666/base -> origin/gh/bobrenjc93/666/base 2025-11-03T16:58:00.6672872Z * [new branch] gh/bobrenjc93/666/head -> origin/gh/bobrenjc93/666/head 2025-11-03T16:58:00.6674593Z * [new branch] gh/bobrenjc93/666/orig -> origin/gh/bobrenjc93/666/orig 2025-11-03T16:58:00.6677034Z * [new branch] gh/bobrenjc93/667/base -> origin/gh/bobrenjc93/667/base 2025-11-03T16:58:00.6678760Z * [new branch] gh/bobrenjc93/667/head -> origin/gh/bobrenjc93/667/head 2025-11-03T16:58:00.6680449Z * [new branch] gh/bobrenjc93/667/orig -> origin/gh/bobrenjc93/667/orig 2025-11-03T16:58:00.6682833Z * [new branch] gh/bobrenjc93/668/base -> origin/gh/bobrenjc93/668/base 2025-11-03T16:58:00.6684585Z * [new branch] gh/bobrenjc93/668/head -> origin/gh/bobrenjc93/668/head 2025-11-03T16:58:00.6686317Z * [new branch] gh/bobrenjc93/668/orig -> origin/gh/bobrenjc93/668/orig 2025-11-03T16:58:00.6688934Z * [new branch] gh/bobrenjc93/669/base -> origin/gh/bobrenjc93/669/base 2025-11-03T16:58:00.6690753Z * [new branch] gh/bobrenjc93/669/head -> origin/gh/bobrenjc93/669/head 2025-11-03T16:58:00.6692472Z * [new branch] gh/bobrenjc93/669/orig -> origin/gh/bobrenjc93/669/orig 2025-11-03T16:58:00.6694892Z * [new branch] gh/bobrenjc93/670/base -> origin/gh/bobrenjc93/670/base 2025-11-03T16:58:00.6696599Z * [new branch] gh/bobrenjc93/670/head -> origin/gh/bobrenjc93/670/head 2025-11-03T16:58:00.6698820Z * [new branch] gh/bobrenjc93/670/orig -> origin/gh/bobrenjc93/670/orig 2025-11-03T16:58:00.6701326Z * [new branch] gh/bobrenjc93/671/base -> origin/gh/bobrenjc93/671/base 2025-11-03T16:58:00.6703014Z * [new branch] gh/bobrenjc93/671/head -> origin/gh/bobrenjc93/671/head 2025-11-03T16:58:00.6704807Z * [new branch] gh/bobrenjc93/671/orig -> origin/gh/bobrenjc93/671/orig 2025-11-03T16:58:00.6707253Z * [new branch] gh/bobrenjc93/672/base -> origin/gh/bobrenjc93/672/base 2025-11-03T16:58:00.6708919Z * [new branch] gh/bobrenjc93/672/head -> origin/gh/bobrenjc93/672/head 2025-11-03T16:58:00.6710961Z * [new branch] gh/bobrenjc93/672/orig -> origin/gh/bobrenjc93/672/orig 2025-11-03T16:58:00.6713573Z * [new branch] gh/bobrenjc93/673/base -> origin/gh/bobrenjc93/673/base 2025-11-03T16:58:00.6715213Z * [new branch] gh/bobrenjc93/673/head -> origin/gh/bobrenjc93/673/head 2025-11-03T16:58:00.6716925Z * [new branch] gh/bobrenjc93/673/orig -> origin/gh/bobrenjc93/673/orig 2025-11-03T16:58:00.6719402Z * [new branch] gh/bobrenjc93/674/base -> origin/gh/bobrenjc93/674/base 2025-11-03T16:58:00.6721131Z * [new branch] gh/bobrenjc93/674/head -> origin/gh/bobrenjc93/674/head 2025-11-03T16:58:00.6722878Z * [new branch] gh/bobrenjc93/674/orig -> origin/gh/bobrenjc93/674/orig 2025-11-03T16:58:00.6725153Z * [new branch] gh/bobrenjc93/675/base -> origin/gh/bobrenjc93/675/base 2025-11-03T16:58:00.6726875Z * [new branch] gh/bobrenjc93/675/head -> origin/gh/bobrenjc93/675/head 2025-11-03T16:58:00.6728699Z * [new branch] gh/bobrenjc93/675/orig -> origin/gh/bobrenjc93/675/orig 2025-11-03T16:58:00.6731058Z * [new branch] gh/bobrenjc93/676/base -> origin/gh/bobrenjc93/676/base 2025-11-03T16:58:00.6732812Z * [new branch] gh/bobrenjc93/676/head -> origin/gh/bobrenjc93/676/head 2025-11-03T16:58:00.6734516Z * [new branch] gh/bobrenjc93/676/orig -> origin/gh/bobrenjc93/676/orig 2025-11-03T16:58:00.6736936Z * [new branch] gh/bobrenjc93/677/base -> origin/gh/bobrenjc93/677/base 2025-11-03T16:58:00.6738966Z * [new branch] gh/bobrenjc93/677/head -> origin/gh/bobrenjc93/677/head 2025-11-03T16:58:00.6740432Z * [new branch] gh/bobrenjc93/677/orig -> origin/gh/bobrenjc93/677/orig 2025-11-03T16:58:00.6742646Z * [new branch] gh/bobrenjc93/678/base -> origin/gh/bobrenjc93/678/base 2025-11-03T16:58:00.6744556Z * [new branch] gh/bobrenjc93/678/head -> origin/gh/bobrenjc93/678/head 2025-11-03T16:58:00.6746132Z * [new branch] gh/bobrenjc93/678/orig -> origin/gh/bobrenjc93/678/orig 2025-11-03T16:58:00.6748576Z * [new branch] gh/bobrenjc93/679/base -> origin/gh/bobrenjc93/679/base 2025-11-03T16:58:00.6750476Z * [new branch] gh/bobrenjc93/679/head -> origin/gh/bobrenjc93/679/head 2025-11-03T16:58:00.6752226Z * [new branch] gh/bobrenjc93/679/orig -> origin/gh/bobrenjc93/679/orig 2025-11-03T16:58:00.6755180Z * [new branch] gh/c00w/23/base -> origin/gh/c00w/23/base 2025-11-03T16:58:00.6756938Z * [new branch] gh/c00w/23/head -> origin/gh/c00w/23/head 2025-11-03T16:58:00.6759385Z * [new branch] gh/c00w/53/base -> origin/gh/c00w/53/base 2025-11-03T16:58:00.6761241Z * [new branch] gh/c00w/53/head -> origin/gh/c00w/53/head 2025-11-03T16:58:00.6762902Z * [new branch] gh/c00w/53/orig -> origin/gh/c00w/53/orig 2025-11-03T16:58:00.6765125Z * [new branch] gh/c00w/54/base -> origin/gh/c00w/54/base 2025-11-03T16:58:00.6767467Z * [new branch] gh/c00w/54/head -> origin/gh/c00w/54/head 2025-11-03T16:58:00.6769011Z * [new branch] gh/c00w/54/orig -> origin/gh/c00w/54/orig 2025-11-03T16:58:00.6771277Z * [new branch] gh/c00w/55/base -> origin/gh/c00w/55/base 2025-11-03T16:58:00.6773013Z * [new branch] gh/c00w/55/head -> origin/gh/c00w/55/head 2025-11-03T16:58:00.6774706Z * [new branch] gh/c00w/55/orig -> origin/gh/c00w/55/orig 2025-11-03T16:58:00.6777034Z * [new branch] gh/c00w/56/base -> origin/gh/c00w/56/base 2025-11-03T16:58:00.6779229Z * [new branch] gh/c00w/56/head -> origin/gh/c00w/56/head 2025-11-03T16:58:00.6781006Z * [new branch] gh/c00w/56/orig -> origin/gh/c00w/56/orig 2025-11-03T16:58:00.6783437Z * [new branch] gh/c00w/57/base -> origin/gh/c00w/57/base 2025-11-03T16:58:00.6785294Z * [new branch] gh/c00w/57/head -> origin/gh/c00w/57/head 2025-11-03T16:58:00.6787640Z * [new branch] gh/c00w/57/orig -> origin/gh/c00w/57/orig 2025-11-03T16:58:00.6789484Z * [new branch] gh/c00w/58/base -> origin/gh/c00w/58/base 2025-11-03T16:58:00.6791189Z * [new branch] gh/c00w/58/head -> origin/gh/c00w/58/head 2025-11-03T16:58:00.6792877Z * [new branch] gh/c00w/58/orig -> origin/gh/c00w/58/orig 2025-11-03T16:58:00.6796189Z * [new branch] gh/clee2000/1/base -> origin/gh/clee2000/1/base 2025-11-03T16:58:00.6797432Z * [new branch] gh/clee2000/1/head -> origin/gh/clee2000/1/head 2025-11-03T16:58:00.6799177Z * [new branch] gh/clee2000/1/orig -> origin/gh/clee2000/1/orig 2025-11-03T16:58:00.6802062Z * [new branch] gh/coconutruben/1/base -> origin/gh/coconutruben/1/base 2025-11-03T16:58:00.6803931Z * [new branch] gh/coconutruben/1/head -> origin/gh/coconutruben/1/head 2025-11-03T16:58:00.6806531Z * [new branch] gh/coconutruben/20/base -> origin/gh/coconutruben/20/base 2025-11-03T16:58:00.6808372Z * [new branch] gh/coconutruben/20/head -> origin/gh/coconutruben/20/head 2025-11-03T16:58:00.6810313Z * [new branch] gh/coconutruben/20/orig -> origin/gh/coconutruben/20/orig 2025-11-03T16:58:00.6813082Z * [new branch] gh/coconutruben/55/base -> origin/gh/coconutruben/55/base 2025-11-03T16:58:00.6814758Z * [new branch] gh/coconutruben/55/head -> origin/gh/coconutruben/55/head 2025-11-03T16:58:00.6816526Z * [new branch] gh/coconutruben/55/orig -> origin/gh/coconutruben/55/orig 2025-11-03T16:58:00.6818987Z * [new branch] gh/coconutruben/57/base -> origin/gh/coconutruben/57/base 2025-11-03T16:58:00.6820840Z * [new branch] gh/coconutruben/57/head -> origin/gh/coconutruben/57/head 2025-11-03T16:58:00.6822647Z * [new branch] gh/coconutruben/57/orig -> origin/gh/coconutruben/57/orig 2025-11-03T16:58:00.6825081Z * [new branch] gh/coconutruben/70/base -> origin/gh/coconutruben/70/base 2025-11-03T16:58:00.6827476Z * [new branch] gh/coconutruben/70/head -> origin/gh/coconutruben/70/head 2025-11-03T16:58:00.6828877Z * [new branch] gh/coconutruben/70/orig -> origin/gh/coconutruben/70/orig 2025-11-03T16:58:00.6831049Z * [new branch] gh/coconutruben/71/base -> origin/gh/coconutruben/71/base 2025-11-03T16:58:00.6832796Z * [new branch] gh/coconutruben/71/head -> origin/gh/coconutruben/71/head 2025-11-03T16:58:00.6834752Z * [new branch] gh/coconutruben/71/orig -> origin/gh/coconutruben/71/orig 2025-11-03T16:58:00.6836807Z * [new branch] gh/coconutruben/72/base -> origin/gh/coconutruben/72/base 2025-11-03T16:58:00.6838588Z * [new branch] gh/coconutruben/72/head -> origin/gh/coconutruben/72/head 2025-11-03T16:58:00.6840335Z * [new branch] gh/coconutruben/72/orig -> origin/gh/coconutruben/72/orig 2025-11-03T16:58:00.6842555Z * [new branch] gh/coconutruben/73/base -> origin/gh/coconutruben/73/base 2025-11-03T16:58:00.6844558Z * [new branch] gh/coconutruben/73/head -> origin/gh/coconutruben/73/head 2025-11-03T16:58:00.6846061Z * [new branch] gh/coconutruben/73/orig -> origin/gh/coconutruben/73/orig 2025-11-03T16:58:00.6848579Z * [new branch] gh/coconutruben/74/base -> origin/gh/coconutruben/74/base 2025-11-03T16:58:00.6850444Z * [new branch] gh/coconutruben/74/head -> origin/gh/coconutruben/74/head 2025-11-03T16:58:00.6852328Z * [new branch] gh/coconutruben/74/orig -> origin/gh/coconutruben/74/orig 2025-11-03T16:58:00.6854849Z * [new branch] gh/coconutruben/79/base -> origin/gh/coconutruben/79/base 2025-11-03T16:58:00.6856771Z * [new branch] gh/coconutruben/79/head -> origin/gh/coconutruben/79/head 2025-11-03T16:58:00.6858849Z * [new branch] gh/coconutruben/79/orig -> origin/gh/coconutruben/79/orig 2025-11-03T16:58:00.6860856Z * [new branch] gh/coconutruben/80/base -> origin/gh/coconutruben/80/base 2025-11-03T16:58:00.6862663Z * [new branch] gh/coconutruben/80/head -> origin/gh/coconutruben/80/head 2025-11-03T16:58:00.6864433Z * [new branch] gh/coconutruben/80/orig -> origin/gh/coconutruben/80/orig 2025-11-03T16:58:00.6867088Z * [new branch] gh/coconutruben/81/base -> origin/gh/coconutruben/81/base 2025-11-03T16:58:00.6868792Z * [new branch] gh/coconutruben/81/head -> origin/gh/coconutruben/81/head 2025-11-03T16:58:00.6870489Z * [new branch] gh/coconutruben/81/orig -> origin/gh/coconutruben/81/orig 2025-11-03T16:58:00.6873050Z * [new branch] gh/coconutruben/82/base -> origin/gh/coconutruben/82/base 2025-11-03T16:58:00.6874813Z * [new branch] gh/coconutruben/82/head -> origin/gh/coconutruben/82/head 2025-11-03T16:58:00.6876623Z * [new branch] gh/coconutruben/82/orig -> origin/gh/coconutruben/82/orig 2025-11-03T16:58:00.6879089Z * [new branch] gh/coconutruben/83/base -> origin/gh/coconutruben/83/base 2025-11-03T16:58:00.6880871Z * [new branch] gh/coconutruben/83/head -> origin/gh/coconutruben/83/head 2025-11-03T16:58:00.6882836Z * [new branch] gh/coconutruben/83/orig -> origin/gh/coconutruben/83/orig 2025-11-03T16:58:00.6885475Z * [new branch] gh/colinchan15/1/base -> origin/gh/colinchan15/1/base 2025-11-03T16:58:00.6887772Z * [new branch] gh/colinchan15/1/head -> origin/gh/colinchan15/1/head 2025-11-03T16:58:00.6890237Z * [new branch] gh/colinchan15/2/base -> origin/gh/colinchan15/2/base 2025-11-03T16:58:00.6891923Z * [new branch] gh/colinchan15/2/head -> origin/gh/colinchan15/2/head 2025-11-03T16:58:00.6894166Z * [new branch] gh/colinchan15/3/base -> origin/gh/colinchan15/3/base 2025-11-03T16:58:00.6895811Z * [new branch] gh/colinchan15/3/head -> origin/gh/colinchan15/3/head 2025-11-03T16:58:00.6898008Z * [new branch] gh/colinchan15/6/base -> origin/gh/colinchan15/6/base 2025-11-03T16:58:00.6899798Z * [new branch] gh/colinchan15/6/head -> origin/gh/colinchan15/6/head 2025-11-03T16:58:00.6902800Z * [new branch] gh/davidberard98/382/base -> origin/gh/davidberard98/382/base 2025-11-03T16:58:00.6905151Z * [new branch] gh/davidberard98/382/head -> origin/gh/davidberard98/382/head 2025-11-03T16:58:00.6907477Z * [new branch] gh/davidberard98/382/orig -> origin/gh/davidberard98/382/orig 2025-11-03T16:58:00.6909384Z * [new branch] gh/davidberard98/386/base -> origin/gh/davidberard98/386/base 2025-11-03T16:58:00.6911449Z * [new branch] gh/davidberard98/386/head -> origin/gh/davidberard98/386/head 2025-11-03T16:58:00.6913124Z * [new branch] gh/davidberard98/386/orig -> origin/gh/davidberard98/386/orig 2025-11-03T16:58:00.6915500Z * [new branch] gh/davidberard98/391/base -> origin/gh/davidberard98/391/base 2025-11-03T16:58:00.6917305Z * [new branch] gh/davidberard98/391/head -> origin/gh/davidberard98/391/head 2025-11-03T16:58:00.6918995Z * [new branch] gh/davidberard98/391/orig -> origin/gh/davidberard98/391/orig 2025-11-03T16:58:00.6921322Z * [new branch] gh/davidberard98/392/base -> origin/gh/davidberard98/392/base 2025-11-03T16:58:00.6923163Z * [new branch] gh/davidberard98/392/head -> origin/gh/davidberard98/392/head 2025-11-03T16:58:00.6925049Z * [new branch] gh/davidberard98/392/orig -> origin/gh/davidberard98/392/orig 2025-11-03T16:58:00.6927618Z * [new branch] gh/davidberard98/399/base -> origin/gh/davidberard98/399/base 2025-11-03T16:58:00.6929504Z * [new branch] gh/davidberard98/399/head -> origin/gh/davidberard98/399/head 2025-11-03T16:58:00.6931278Z * [new branch] gh/davidberard98/399/orig -> origin/gh/davidberard98/399/orig 2025-11-03T16:58:00.6934103Z * [new branch] gh/desertfire/594/base -> origin/gh/desertfire/594/base 2025-11-03T16:58:00.6935850Z * [new branch] gh/desertfire/594/head -> origin/gh/desertfire/594/head 2025-11-03T16:58:00.6937586Z * [new branch] gh/desertfire/594/orig -> origin/gh/desertfire/594/orig 2025-11-03T16:58:00.6939890Z * [new branch] gh/desertfire/595/base -> origin/gh/desertfire/595/base 2025-11-03T16:58:00.6941676Z * [new branch] gh/desertfire/595/head -> origin/gh/desertfire/595/head 2025-11-03T16:58:00.6943442Z * [new branch] gh/desertfire/595/orig -> origin/gh/desertfire/595/orig 2025-11-03T16:58:00.6945955Z * [new branch] gh/desertfire/597/base -> origin/gh/desertfire/597/base 2025-11-03T16:58:00.6947713Z * [new branch] gh/desertfire/597/head -> origin/gh/desertfire/597/head 2025-11-03T16:58:00.6949472Z * [new branch] gh/desertfire/597/orig -> origin/gh/desertfire/597/orig 2025-11-03T16:58:00.6951835Z * [new branch] gh/desertfire/600/base -> origin/gh/desertfire/600/base 2025-11-03T16:58:00.6953588Z * [new branch] gh/desertfire/600/head -> origin/gh/desertfire/600/head 2025-11-03T16:58:00.6955524Z * [new branch] gh/desertfire/600/orig -> origin/gh/desertfire/600/orig 2025-11-03T16:58:00.6958410Z * [new branch] gh/desertfire/601/base -> origin/gh/desertfire/601/base 2025-11-03T16:58:00.6959800Z * [new branch] gh/desertfire/601/head -> origin/gh/desertfire/601/head 2025-11-03T16:58:00.6961591Z * [new branch] gh/desertfire/601/orig -> origin/gh/desertfire/601/orig 2025-11-03T16:58:00.6963980Z * [new branch] gh/desertfire/602/base -> origin/gh/desertfire/602/base 2025-11-03T16:58:00.6965718Z * [new branch] gh/desertfire/602/head -> origin/gh/desertfire/602/head 2025-11-03T16:58:00.6967383Z * [new branch] gh/desertfire/602/orig -> origin/gh/desertfire/602/orig 2025-11-03T16:58:00.6970019Z * [new branch] gh/desertfire/603/base -> origin/gh/desertfire/603/base 2025-11-03T16:58:00.6971801Z * [new branch] gh/desertfire/603/head -> origin/gh/desertfire/603/head 2025-11-03T16:58:00.6973522Z * [new branch] gh/desertfire/603/orig -> origin/gh/desertfire/603/orig 2025-11-03T16:58:00.6975681Z * [new branch] gh/desertfire/604/base -> origin/gh/desertfire/604/base 2025-11-03T16:58:00.6978004Z * [new branch] gh/desertfire/604/head -> origin/gh/desertfire/604/head 2025-11-03T16:58:00.6979273Z * [new branch] gh/desertfire/604/orig -> origin/gh/desertfire/604/orig 2025-11-03T16:58:00.6982553Z * [new branch] gh/dharakk/1/base -> origin/gh/dharakk/1/base 2025-11-03T16:58:00.6983922Z * [new branch] gh/dharakk/1/head -> origin/gh/dharakk/1/head 2025-11-03T16:58:00.6987194Z * [new branch] gh/drisspg/170/base -> origin/gh/drisspg/170/base 2025-11-03T16:58:00.6988189Z * [new branch] gh/drisspg/170/head -> origin/gh/drisspg/170/head 2025-11-03T16:58:00.6990522Z * [new branch] gh/drisspg/170/orig -> origin/gh/drisspg/170/orig 2025-11-03T16:58:00.6992935Z * [new branch] gh/drisspg/182/base -> origin/gh/drisspg/182/base 2025-11-03T16:58:00.6994684Z * [new branch] gh/drisspg/182/head -> origin/gh/drisspg/182/head 2025-11-03T16:58:00.6997135Z * [new branch] gh/drisspg/183/base -> origin/gh/drisspg/183/base 2025-11-03T16:58:00.6998731Z * [new branch] gh/drisspg/183/head -> origin/gh/drisspg/183/head 2025-11-03T16:58:00.7000925Z * [new branch] gh/drisspg/184/base -> origin/gh/drisspg/184/base 2025-11-03T16:58:00.7002578Z * [new branch] gh/drisspg/184/head -> origin/gh/drisspg/184/head 2025-11-03T16:58:00.7004967Z * [new branch] gh/drisspg/185/base -> origin/gh/drisspg/185/base 2025-11-03T16:58:00.7006678Z * [new branch] gh/drisspg/185/head -> origin/gh/drisspg/185/head 2025-11-03T16:58:00.7009611Z * [new branch] gh/drisspg/187/base -> origin/gh/drisspg/187/base 2025-11-03T16:58:00.7012455Z * [new branch] gh/drisspg/187/head -> origin/gh/drisspg/187/head 2025-11-03T16:58:00.7014146Z * [new branch] gh/drisspg/187/orig -> origin/gh/drisspg/187/orig 2025-11-03T16:58:00.7016413Z * [new branch] gh/drisspg/194/base -> origin/gh/drisspg/194/base 2025-11-03T16:58:00.7018221Z * [new branch] gh/drisspg/194/head -> origin/gh/drisspg/194/head 2025-11-03T16:58:00.7020379Z * [new branch] gh/drisspg/194/orig -> origin/gh/drisspg/194/orig 2025-11-03T16:58:00.7022621Z * [new branch] gh/drisspg/197/base -> origin/gh/drisspg/197/base 2025-11-03T16:58:00.7024364Z * [new branch] gh/drisspg/197/head -> origin/gh/drisspg/197/head 2025-11-03T16:58:00.7026100Z * [new branch] gh/drisspg/197/orig -> origin/gh/drisspg/197/orig 2025-11-03T16:58:00.7028409Z * [new branch] gh/drisspg/199/base -> origin/gh/drisspg/199/base 2025-11-03T16:58:00.7030096Z * [new branch] gh/drisspg/199/head -> origin/gh/drisspg/199/head 2025-11-03T16:58:00.7031874Z * [new branch] gh/drisspg/199/orig -> origin/gh/drisspg/199/orig 2025-11-03T16:58:00.7034178Z * [new branch] gh/drisspg/200/base -> origin/gh/drisspg/200/base 2025-11-03T16:58:00.7035902Z * [new branch] gh/drisspg/200/head -> origin/gh/drisspg/200/head 2025-11-03T16:58:00.7037571Z * [new branch] gh/drisspg/200/orig -> origin/gh/drisspg/200/orig 2025-11-03T16:58:00.7039883Z * [new branch] gh/drisspg/210/base -> origin/gh/drisspg/210/base 2025-11-03T16:58:00.7041554Z * [new branch] gh/drisspg/210/head -> origin/gh/drisspg/210/head 2025-11-03T16:58:00.7043245Z * [new branch] gh/drisspg/210/orig -> origin/gh/drisspg/210/orig 2025-11-03T16:58:00.7045514Z * [new branch] gh/drisspg/211/base -> origin/gh/drisspg/211/base 2025-11-03T16:58:00.7047230Z * [new branch] gh/drisspg/211/head -> origin/gh/drisspg/211/head 2025-11-03T16:58:00.7048992Z * [new branch] gh/drisspg/211/orig -> origin/gh/drisspg/211/orig 2025-11-03T16:58:00.7051408Z * [new branch] gh/drisspg/212/base -> origin/gh/drisspg/212/base 2025-11-03T16:58:00.7053118Z * [new branch] gh/drisspg/212/head -> origin/gh/drisspg/212/head 2025-11-03T16:58:00.7054806Z * [new branch] gh/drisspg/212/orig -> origin/gh/drisspg/212/orig 2025-11-03T16:58:00.7057270Z * [new branch] gh/drisspg/213/base -> origin/gh/drisspg/213/base 2025-11-03T16:58:00.7059255Z * [new branch] gh/drisspg/213/head -> origin/gh/drisspg/213/head 2025-11-03T16:58:00.7060714Z * [new branch] gh/drisspg/213/orig -> origin/gh/drisspg/213/orig 2025-11-03T16:58:00.7063025Z * [new branch] gh/drisspg/214/base -> origin/gh/drisspg/214/base 2025-11-03T16:58:00.7064697Z * [new branch] gh/drisspg/214/head -> origin/gh/drisspg/214/head 2025-11-03T16:58:00.7066500Z * [new branch] gh/drisspg/214/orig -> origin/gh/drisspg/214/orig 2025-11-03T16:58:00.7068760Z * [new branch] gh/drisspg/215/base -> origin/gh/drisspg/215/base 2025-11-03T16:58:00.7070414Z * [new branch] gh/drisspg/215/head -> origin/gh/drisspg/215/head 2025-11-03T16:58:00.7072220Z * [new branch] gh/drisspg/215/orig -> origin/gh/drisspg/215/orig 2025-11-03T16:58:00.7075088Z * [new branch] gh/drisspg/216/base -> origin/gh/drisspg/216/base 2025-11-03T16:58:00.7077146Z * [new branch] gh/drisspg/216/head -> origin/gh/drisspg/216/head 2025-11-03T16:58:00.7078519Z * [new branch] gh/drisspg/216/orig -> origin/gh/drisspg/216/orig 2025-11-03T16:58:00.7080944Z * [new branch] gh/drisspg/217/base -> origin/gh/drisspg/217/base 2025-11-03T16:58:00.7082643Z * [new branch] gh/drisspg/217/head -> origin/gh/drisspg/217/head 2025-11-03T16:58:00.7084349Z * [new branch] gh/drisspg/217/orig -> origin/gh/drisspg/217/orig 2025-11-03T16:58:00.7087279Z * [new branch] gh/dsjohns2/1/base -> origin/gh/dsjohns2/1/base 2025-11-03T16:58:00.7089124Z * [new branch] gh/dsjohns2/1/head -> origin/gh/dsjohns2/1/head 2025-11-03T16:58:00.7092028Z * [new branch] gh/dzmitry-huba/1/base -> origin/gh/dzmitry-huba/1/base 2025-11-03T16:58:00.7093787Z * [new branch] gh/dzmitry-huba/1/head -> origin/gh/dzmitry-huba/1/head 2025-11-03T16:58:00.7096207Z * [new branch] gh/dzmitry-huba/10/base -> origin/gh/dzmitry-huba/10/base 2025-11-03T16:58:00.7098251Z * [new branch] gh/dzmitry-huba/10/head -> origin/gh/dzmitry-huba/10/head 2025-11-03T16:58:00.7099787Z * [new branch] gh/dzmitry-huba/10/orig -> origin/gh/dzmitry-huba/10/orig 2025-11-03T16:58:00.7102505Z * [new branch] gh/dzmitry-huba/11/base -> origin/gh/dzmitry-huba/11/base 2025-11-03T16:58:00.7103803Z * [new branch] gh/dzmitry-huba/11/head -> origin/gh/dzmitry-huba/11/head 2025-11-03T16:58:00.7105603Z * [new branch] gh/dzmitry-huba/11/orig -> origin/gh/dzmitry-huba/11/orig 2025-11-03T16:58:00.7107992Z * [new branch] gh/dzmitry-huba/12/base -> origin/gh/dzmitry-huba/12/base 2025-11-03T16:58:00.7109835Z * [new branch] gh/dzmitry-huba/12/head -> origin/gh/dzmitry-huba/12/head 2025-11-03T16:58:00.7111779Z * [new branch] gh/dzmitry-huba/12/orig -> origin/gh/dzmitry-huba/12/orig 2025-11-03T16:58:00.7113960Z * [new branch] gh/dzmitry-huba/2/base -> origin/gh/dzmitry-huba/2/base 2025-11-03T16:58:00.7115578Z * [new branch] gh/dzmitry-huba/2/head -> origin/gh/dzmitry-huba/2/head 2025-11-03T16:58:00.7117846Z * [new branch] gh/dzmitry-huba/3/base -> origin/gh/dzmitry-huba/3/base 2025-11-03T16:58:00.7119503Z * [new branch] gh/dzmitry-huba/3/head -> origin/gh/dzmitry-huba/3/head 2025-11-03T16:58:00.7121727Z * [new branch] gh/dzmitry-huba/4/base -> origin/gh/dzmitry-huba/4/base 2025-11-03T16:58:00.7123280Z * [new branch] gh/dzmitry-huba/4/head -> origin/gh/dzmitry-huba/4/head 2025-11-03T16:58:00.7125132Z * [new branch] gh/dzmitry-huba/4/orig -> origin/gh/dzmitry-huba/4/orig 2025-11-03T16:58:00.7127798Z * [new branch] gh/dzmitry-huba/5/base -> origin/gh/dzmitry-huba/5/base 2025-11-03T16:58:00.7129652Z * [new branch] gh/dzmitry-huba/5/head -> origin/gh/dzmitry-huba/5/head 2025-11-03T16:58:00.7131350Z * [new branch] gh/dzmitry-huba/5/orig -> origin/gh/dzmitry-huba/5/orig 2025-11-03T16:58:00.7133518Z * [new branch] gh/dzmitry-huba/6/base -> origin/gh/dzmitry-huba/6/base 2025-11-03T16:58:00.7135290Z * [new branch] gh/dzmitry-huba/6/head -> origin/gh/dzmitry-huba/6/head 2025-11-03T16:58:00.7137169Z * [new branch] gh/dzmitry-huba/6/orig -> origin/gh/dzmitry-huba/6/orig 2025-11-03T16:58:00.7139662Z * [new branch] gh/dzmitry-huba/7/base -> origin/gh/dzmitry-huba/7/base 2025-11-03T16:58:00.7141391Z * [new branch] gh/dzmitry-huba/7/head -> origin/gh/dzmitry-huba/7/head 2025-11-03T16:58:00.7143076Z * [new branch] gh/dzmitry-huba/7/orig -> origin/gh/dzmitry-huba/7/orig 2025-11-03T16:58:00.7145462Z * [new branch] gh/dzmitry-huba/8/base -> origin/gh/dzmitry-huba/8/base 2025-11-03T16:58:00.7147251Z * [new branch] gh/dzmitry-huba/8/head -> origin/gh/dzmitry-huba/8/head 2025-11-03T16:58:00.7148989Z * [new branch] gh/dzmitry-huba/8/orig -> origin/gh/dzmitry-huba/8/orig 2025-11-03T16:58:00.7151373Z * [new branch] gh/dzmitry-huba/9/base -> origin/gh/dzmitry-huba/9/base 2025-11-03T16:58:00.7153176Z * [new branch] gh/dzmitry-huba/9/head -> origin/gh/dzmitry-huba/9/head 2025-11-03T16:58:00.7154969Z * [new branch] gh/dzmitry-huba/9/orig -> origin/gh/dzmitry-huba/9/orig 2025-11-03T16:58:00.7157850Z * [new branch] gh/eellison/808/base -> origin/gh/eellison/808/base 2025-11-03T16:58:00.7159614Z * [new branch] gh/eellison/808/head -> origin/gh/eellison/808/head 2025-11-03T16:58:00.7161333Z * [new branch] gh/eellison/808/orig -> origin/gh/eellison/808/orig 2025-11-03T16:58:00.7163681Z * [new branch] gh/eellison/809/base -> origin/gh/eellison/809/base 2025-11-03T16:58:00.7165405Z * [new branch] gh/eellison/809/head -> origin/gh/eellison/809/head 2025-11-03T16:58:00.7167178Z * [new branch] gh/eellison/809/orig -> origin/gh/eellison/809/orig 2025-11-03T16:58:00.7170111Z * [new branch] gh/eellison/822/base -> origin/gh/eellison/822/base 2025-11-03T16:58:00.7171637Z * [new branch] gh/eellison/822/head -> origin/gh/eellison/822/head 2025-11-03T16:58:00.7173482Z * [new branch] gh/eellison/822/orig -> origin/gh/eellison/822/orig 2025-11-03T16:58:00.7176177Z * [new branch] gh/eellison/823/base -> origin/gh/eellison/823/base 2025-11-03T16:58:00.7177756Z * [new branch] gh/eellison/823/head -> origin/gh/eellison/823/head 2025-11-03T16:58:00.7179929Z * [new branch] gh/eellison/823/orig -> origin/gh/eellison/823/orig 2025-11-03T16:58:00.7181933Z * [new branch] gh/eellison/836/base -> origin/gh/eellison/836/base 2025-11-03T16:58:00.7183658Z * [new branch] gh/eellison/836/head -> origin/gh/eellison/836/head 2025-11-03T16:58:00.7185342Z * [new branch] gh/eellison/836/orig -> origin/gh/eellison/836/orig 2025-11-03T16:58:00.7187801Z * [new branch] gh/eellison/837/base -> origin/gh/eellison/837/base 2025-11-03T16:58:00.7189423Z * [new branch] gh/eellison/837/head -> origin/gh/eellison/837/head 2025-11-03T16:58:00.7191103Z * [new branch] gh/eellison/837/orig -> origin/gh/eellison/837/orig 2025-11-03T16:58:00.7193639Z * [new branch] gh/eellison/838/base -> origin/gh/eellison/838/base 2025-11-03T16:58:00.7195366Z * [new branch] gh/eellison/838/head -> origin/gh/eellison/838/head 2025-11-03T16:58:00.7197820Z * [new branch] gh/eellison/838/orig -> origin/gh/eellison/838/orig 2025-11-03T16:58:00.7199500Z * [new branch] gh/eellison/839/base -> origin/gh/eellison/839/base 2025-11-03T16:58:00.7201297Z * [new branch] gh/eellison/839/head -> origin/gh/eellison/839/head 2025-11-03T16:58:00.7204062Z * [new branch] gh/eellison/839/orig -> origin/gh/eellison/839/orig 2025-11-03T16:58:00.7205948Z * [new branch] gh/eellison/840/base -> origin/gh/eellison/840/base 2025-11-03T16:58:00.7207529Z * [new branch] gh/eellison/840/head -> origin/gh/eellison/840/head 2025-11-03T16:58:00.7209605Z * [new branch] gh/eellison/840/orig -> origin/gh/eellison/840/orig 2025-11-03T16:58:00.7212134Z * [new branch] gh/eellison/841/base -> origin/gh/eellison/841/base 2025-11-03T16:58:00.7213827Z * [new branch] gh/eellison/841/head -> origin/gh/eellison/841/head 2025-11-03T16:58:00.7215522Z * [new branch] gh/eellison/841/orig -> origin/gh/eellison/841/orig 2025-11-03T16:58:00.7217735Z * [new branch] gh/eellison/842/base -> origin/gh/eellison/842/base 2025-11-03T16:58:00.7219434Z * [new branch] gh/eellison/842/head -> origin/gh/eellison/842/head 2025-11-03T16:58:00.7221141Z * [new branch] gh/eellison/842/orig -> origin/gh/eellison/842/orig 2025-11-03T16:58:00.7223483Z * [new branch] gh/eellison/843/base -> origin/gh/eellison/843/base 2025-11-03T16:58:00.7225140Z * [new branch] gh/eellison/843/head -> origin/gh/eellison/843/head 2025-11-03T16:58:00.7227002Z * [new branch] gh/eellison/843/orig -> origin/gh/eellison/843/orig 2025-11-03T16:58:00.7229938Z * [new branch] gh/eellison/844/base -> origin/gh/eellison/844/base 2025-11-03T16:58:00.7231571Z * [new branch] gh/eellison/844/head -> origin/gh/eellison/844/head 2025-11-03T16:58:00.7233180Z * [new branch] gh/eellison/844/orig -> origin/gh/eellison/844/orig 2025-11-03T16:58:00.7236099Z * [new branch] gh/eellison/845/base -> origin/gh/eellison/845/base 2025-11-03T16:58:00.7237339Z * [new branch] gh/eellison/845/head -> origin/gh/eellison/845/head 2025-11-03T16:58:00.7239410Z * [new branch] gh/eellison/845/orig -> origin/gh/eellison/845/orig 2025-11-03T16:58:00.7241644Z * [new branch] gh/eellison/846/base -> origin/gh/eellison/846/base 2025-11-03T16:58:00.7243419Z * [new branch] gh/eellison/846/head -> origin/gh/eellison/846/head 2025-11-03T16:58:00.7245005Z * [new branch] gh/eellison/846/orig -> origin/gh/eellison/846/orig 2025-11-03T16:58:00.7247315Z * [new branch] gh/eellison/847/base -> origin/gh/eellison/847/base 2025-11-03T16:58:00.7248969Z * [new branch] gh/eellison/847/head -> origin/gh/eellison/847/head 2025-11-03T16:58:00.7250792Z * [new branch] gh/eellison/847/orig -> origin/gh/eellison/847/orig 2025-11-03T16:58:00.7253716Z * [new branch] gh/eellison/848/base -> origin/gh/eellison/848/base 2025-11-03T16:58:00.7255408Z * [new branch] gh/eellison/848/head -> origin/gh/eellison/848/head 2025-11-03T16:58:00.7257399Z * [new branch] gh/eellison/848/orig -> origin/gh/eellison/848/orig 2025-11-03T16:58:00.7260303Z * [new branch] gh/eellison/849/base -> origin/gh/eellison/849/base 2025-11-03T16:58:00.7262162Z * [new branch] gh/eellison/849/head -> origin/gh/eellison/849/head 2025-11-03T16:58:00.7263834Z * [new branch] gh/eellison/849/orig -> origin/gh/eellison/849/orig 2025-11-03T16:58:00.7266764Z * [new branch] gh/eellison/850/base -> origin/gh/eellison/850/base 2025-11-03T16:58:00.7267787Z * [new branch] gh/eellison/850/head -> origin/gh/eellison/850/head 2025-11-03T16:58:00.7269792Z * [new branch] gh/eellison/850/orig -> origin/gh/eellison/850/orig 2025-11-03T16:58:00.7272134Z * [new branch] gh/eellison/851/base -> origin/gh/eellison/851/base 2025-11-03T16:58:00.7273851Z * [new branch] gh/eellison/851/head -> origin/gh/eellison/851/head 2025-11-03T16:58:00.7275530Z * [new branch] gh/eellison/851/orig -> origin/gh/eellison/851/orig 2025-11-03T16:58:00.7278261Z * [new branch] gh/eellison/852/base -> origin/gh/eellison/852/base 2025-11-03T16:58:00.7279813Z * [new branch] gh/eellison/852/head -> origin/gh/eellison/852/head 2025-11-03T16:58:00.7281304Z * [new branch] gh/eellison/852/orig -> origin/gh/eellison/852/orig 2025-11-03T16:58:00.7284164Z * [new branch] gh/eellison/853/base -> origin/gh/eellison/853/base 2025-11-03T16:58:00.7285429Z * [new branch] gh/eellison/853/head -> origin/gh/eellison/853/head 2025-11-03T16:58:00.7287125Z * [new branch] gh/eellison/853/orig -> origin/gh/eellison/853/orig 2025-11-03T16:58:00.7289729Z * [new branch] gh/eellison/854/base -> origin/gh/eellison/854/base 2025-11-03T16:58:00.7291475Z * [new branch] gh/eellison/854/head -> origin/gh/eellison/854/head 2025-11-03T16:58:00.7293221Z * [new branch] gh/eellison/854/orig -> origin/gh/eellison/854/orig 2025-11-03T16:58:00.7295593Z * [new branch] gh/eellison/855/base -> origin/gh/eellison/855/base 2025-11-03T16:58:00.7297257Z * [new branch] gh/eellison/855/head -> origin/gh/eellison/855/head 2025-11-03T16:58:00.7298949Z * [new branch] gh/eellison/855/orig -> origin/gh/eellison/855/orig 2025-11-03T16:58:00.7301332Z * [new branch] gh/eellison/856/base -> origin/gh/eellison/856/base 2025-11-03T16:58:00.7303037Z * [new branch] gh/eellison/856/head -> origin/gh/eellison/856/head 2025-11-03T16:58:00.7304747Z * [new branch] gh/eellison/856/orig -> origin/gh/eellison/856/orig 2025-11-03T16:58:00.7307230Z * [new branch] gh/eellison/857/base -> origin/gh/eellison/857/base 2025-11-03T16:58:00.7308974Z * [new branch] gh/eellison/857/head -> origin/gh/eellison/857/head 2025-11-03T16:58:00.7311157Z * [new branch] gh/eellison/857/orig -> origin/gh/eellison/857/orig 2025-11-03T16:58:00.7313743Z * [new branch] gh/eellison/858/base -> origin/gh/eellison/858/base 2025-11-03T16:58:00.7315403Z * [new branch] gh/eellison/858/head -> origin/gh/eellison/858/head 2025-11-03T16:58:00.7317669Z * [new branch] gh/eellison/858/orig -> origin/gh/eellison/858/orig 2025-11-03T16:58:00.7319673Z * [new branch] gh/eellison/859/base -> origin/gh/eellison/859/base 2025-11-03T16:58:00.7321397Z * [new branch] gh/eellison/859/head -> origin/gh/eellison/859/head 2025-11-03T16:58:00.7323101Z * [new branch] gh/eellison/859/orig -> origin/gh/eellison/859/orig 2025-11-03T16:58:00.7325960Z * [new branch] gh/eellison/860/base -> origin/gh/eellison/860/base 2025-11-03T16:58:00.7327139Z * [new branch] gh/eellison/860/head -> origin/gh/eellison/860/head 2025-11-03T16:58:00.7329126Z * [new branch] gh/eellison/860/orig -> origin/gh/eellison/860/orig 2025-11-03T16:58:00.7331450Z * [new branch] gh/eellison/861/base -> origin/gh/eellison/861/base 2025-11-03T16:58:00.7333211Z * [new branch] gh/eellison/861/head -> origin/gh/eellison/861/head 2025-11-03T16:58:00.7334961Z * [new branch] gh/eellison/861/orig -> origin/gh/eellison/861/orig 2025-11-03T16:58:00.7337800Z * [new branch] gh/etaf/147/base -> origin/gh/etaf/147/base 2025-11-03T16:58:00.7339863Z * [new branch] gh/etaf/147/head -> origin/gh/etaf/147/head 2025-11-03T16:58:00.7342366Z * [new branch] gh/etaf/154/base -> origin/gh/etaf/154/base 2025-11-03T16:58:00.7344176Z * [new branch] gh/etaf/154/head -> origin/gh/etaf/154/head 2025-11-03T16:58:00.7345601Z * [new branch] gh/etaf/154/orig -> origin/gh/etaf/154/orig 2025-11-03T16:58:00.7347943Z * [new branch] gh/etaf/156/base -> origin/gh/etaf/156/base 2025-11-03T16:58:00.7349573Z * [new branch] gh/etaf/156/head -> origin/gh/etaf/156/head 2025-11-03T16:58:00.7351295Z * [new branch] gh/etaf/156/orig -> origin/gh/etaf/156/orig 2025-11-03T16:58:00.7353731Z * [new branch] gh/etaf/157/base -> origin/gh/etaf/157/base 2025-11-03T16:58:00.7355445Z * [new branch] gh/etaf/157/head -> origin/gh/etaf/157/head 2025-11-03T16:58:00.7357205Z * [new branch] gh/etaf/157/orig -> origin/gh/etaf/157/orig 2025-11-03T16:58:00.7359896Z * [new branch] gh/etaf/158/base -> origin/gh/etaf/158/base 2025-11-03T16:58:00.7361190Z * [new branch] gh/etaf/158/head -> origin/gh/etaf/158/head 2025-11-03T16:58:00.7362821Z * [new branch] gh/etaf/158/orig -> origin/gh/etaf/158/orig 2025-11-03T16:58:00.7365673Z * [new branch] gh/etaf/159/base -> origin/gh/etaf/159/base 2025-11-03T16:58:00.7367064Z * [new branch] gh/etaf/159/head -> origin/gh/etaf/159/head 2025-11-03T16:58:00.7368804Z * [new branch] gh/etaf/159/orig -> origin/gh/etaf/159/orig 2025-11-03T16:58:00.7371577Z * [new branch] gh/etaf/160/base -> origin/gh/etaf/160/base 2025-11-03T16:58:00.7372828Z * [new branch] gh/etaf/160/head -> origin/gh/etaf/160/head 2025-11-03T16:58:00.7374516Z * [new branch] gh/etaf/160/orig -> origin/gh/etaf/160/orig 2025-11-03T16:58:00.7377453Z * [new branch] gh/etaf/161/base -> origin/gh/etaf/161/base 2025-11-03T16:58:00.7378703Z * [new branch] gh/etaf/161/head -> origin/gh/etaf/161/head 2025-11-03T16:58:00.7380382Z * [new branch] gh/etaf/161/orig -> origin/gh/etaf/161/orig 2025-11-03T16:58:00.7382659Z * [new branch] gh/etaf/162/base -> origin/gh/etaf/162/base 2025-11-03T16:58:00.7384358Z * [new branch] gh/etaf/162/head -> origin/gh/etaf/162/head 2025-11-03T16:58:00.7386057Z * [new branch] gh/etaf/162/orig -> origin/gh/etaf/162/orig 2025-11-03T16:58:00.7388587Z * [new branch] gh/etaf/166/base -> origin/gh/etaf/166/base 2025-11-03T16:58:00.7390204Z * [new branch] gh/etaf/166/head -> origin/gh/etaf/166/head 2025-11-03T16:58:00.7391902Z * [new branch] gh/etaf/166/orig -> origin/gh/etaf/166/orig 2025-11-03T16:58:00.7394238Z * [new branch] gh/etaf/167/base -> origin/gh/etaf/167/base 2025-11-03T16:58:00.7396003Z * [new branch] gh/etaf/167/head -> origin/gh/etaf/167/head 2025-11-03T16:58:00.7397855Z * [new branch] gh/etaf/167/orig -> origin/gh/etaf/167/orig 2025-11-03T16:58:00.7400318Z * [new branch] gh/etaf/168/base -> origin/gh/etaf/168/base 2025-11-03T16:58:00.7402048Z * [new branch] gh/etaf/168/head -> origin/gh/etaf/168/head 2025-11-03T16:58:00.7403822Z * [new branch] gh/etaf/168/orig -> origin/gh/etaf/168/orig 2025-11-03T16:58:00.7406136Z * [new branch] gh/etaf/171/base -> origin/gh/etaf/171/base 2025-11-03T16:58:00.7407845Z * [new branch] gh/etaf/171/head -> origin/gh/etaf/171/head 2025-11-03T16:58:00.7409848Z * [new branch] gh/etaf/171/orig -> origin/gh/etaf/171/orig 2025-11-03T16:58:00.7412332Z * [new branch] gh/etaf/172/base -> origin/gh/etaf/172/base 2025-11-03T16:58:00.7414021Z * [new branch] gh/etaf/172/head -> origin/gh/etaf/172/head 2025-11-03T16:58:00.7415750Z * [new branch] gh/etaf/172/orig -> origin/gh/etaf/172/orig 2025-11-03T16:58:00.7418745Z * [new branch] gh/exclamaforte/1/base -> origin/gh/exclamaforte/1/base 2025-11-03T16:58:00.7420881Z * [new branch] gh/exclamaforte/1/head -> origin/gh/exclamaforte/1/head 2025-11-03T16:58:00.7422624Z * [new branch] gh/exclamaforte/2/base -> origin/gh/exclamaforte/2/base 2025-11-03T16:58:00.7424224Z * [new branch] gh/exclamaforte/2/head -> origin/gh/exclamaforte/2/head 2025-11-03T16:58:00.7426524Z * [new branch] gh/exclamaforte/3/base -> origin/gh/exclamaforte/3/base 2025-11-03T16:58:00.7428217Z * [new branch] gh/exclamaforte/3/head -> origin/gh/exclamaforte/3/head 2025-11-03T16:58:00.7430483Z * [new branch] gh/exclamaforte/4/base -> origin/gh/exclamaforte/4/base 2025-11-03T16:58:00.7432174Z * [new branch] gh/exclamaforte/4/head -> origin/gh/exclamaforte/4/head 2025-11-03T16:58:00.7435097Z * [new branch] gh/ezyang/2374/base -> origin/gh/ezyang/2374/base 2025-11-03T16:58:00.7437252Z * [new branch] gh/ezyang/2374/head -> origin/gh/ezyang/2374/head 2025-11-03T16:58:00.7438754Z * [new branch] gh/ezyang/2374/orig -> origin/gh/ezyang/2374/orig 2025-11-03T16:58:00.7441392Z * [new branch] gh/ezyang/2973/base -> origin/gh/ezyang/2973/base 2025-11-03T16:58:00.7442731Z * [new branch] gh/ezyang/2973/head -> origin/gh/ezyang/2973/head 2025-11-03T16:58:00.7444389Z * [new branch] gh/ezyang/2973/orig -> origin/gh/ezyang/2973/orig 2025-11-03T16:58:00.7447251Z * [new branch] gh/ezyang/2974/base -> origin/gh/ezyang/2974/base 2025-11-03T16:58:00.7448140Z * [new branch] gh/ezyang/2974/head -> origin/gh/ezyang/2974/head 2025-11-03T16:58:00.7450125Z * [new branch] gh/ezyang/2974/orig -> origin/gh/ezyang/2974/orig 2025-11-03T16:58:00.7452536Z * [new branch] gh/ezyang/3120/base -> origin/gh/ezyang/3120/base 2025-11-03T16:58:00.7454243Z * [new branch] gh/ezyang/3120/head -> origin/gh/ezyang/3120/head 2025-11-03T16:58:00.7455974Z * [new branch] gh/ezyang/3120/orig -> origin/gh/ezyang/3120/orig 2025-11-03T16:58:00.7458252Z * [new branch] gh/ezyang/3131/base -> origin/gh/ezyang/3131/base 2025-11-03T16:58:00.7459986Z * [new branch] gh/ezyang/3131/head -> origin/gh/ezyang/3131/head 2025-11-03T16:58:00.7461667Z * [new branch] gh/ezyang/3131/orig -> origin/gh/ezyang/3131/orig 2025-11-03T16:58:00.7463940Z * [new branch] gh/ezyang/3139/base -> origin/gh/ezyang/3139/base 2025-11-03T16:58:00.7465831Z * [new branch] gh/ezyang/3139/head -> origin/gh/ezyang/3139/head 2025-11-03T16:58:00.7467563Z * [new branch] gh/ezyang/3139/orig -> origin/gh/ezyang/3139/orig 2025-11-03T16:58:00.7469856Z * [new branch] gh/ezyang/3140/base -> origin/gh/ezyang/3140/base 2025-11-03T16:58:00.7471523Z * [new branch] gh/ezyang/3140/head -> origin/gh/ezyang/3140/head 2025-11-03T16:58:00.7473281Z * [new branch] gh/ezyang/3140/orig -> origin/gh/ezyang/3140/orig 2025-11-03T16:58:00.7475548Z * [new branch] gh/ezyang/3143/base -> origin/gh/ezyang/3143/base 2025-11-03T16:58:00.7477217Z * [new branch] gh/ezyang/3143/head -> origin/gh/ezyang/3143/head 2025-11-03T16:58:00.7478980Z * [new branch] gh/ezyang/3143/orig -> origin/gh/ezyang/3143/orig 2025-11-03T16:58:00.7481288Z * [new branch] gh/ezyang/3144/base -> origin/gh/ezyang/3144/base 2025-11-03T16:58:00.7482965Z * [new branch] gh/ezyang/3144/head -> origin/gh/ezyang/3144/head 2025-11-03T16:58:00.7484665Z * [new branch] gh/ezyang/3144/orig -> origin/gh/ezyang/3144/orig 2025-11-03T16:58:00.7487038Z * [new branch] gh/ezyang/3148/base -> origin/gh/ezyang/3148/base 2025-11-03T16:58:00.7488749Z * [new branch] gh/ezyang/3148/head -> origin/gh/ezyang/3148/head 2025-11-03T16:58:00.7490497Z * [new branch] gh/ezyang/3148/orig -> origin/gh/ezyang/3148/orig 2025-11-03T16:58:00.7492792Z * [new branch] gh/ezyang/3158/base -> origin/gh/ezyang/3158/base 2025-11-03T16:58:00.7494504Z * [new branch] gh/ezyang/3158/head -> origin/gh/ezyang/3158/head 2025-11-03T16:58:00.7496207Z * [new branch] gh/ezyang/3158/orig -> origin/gh/ezyang/3158/orig 2025-11-03T16:58:00.7498474Z * [new branch] gh/ezyang/3165/base -> origin/gh/ezyang/3165/base 2025-11-03T16:58:00.7500149Z * [new branch] gh/ezyang/3165/head -> origin/gh/ezyang/3165/head 2025-11-03T16:58:00.7501831Z * [new branch] gh/ezyang/3165/orig -> origin/gh/ezyang/3165/orig 2025-11-03T16:58:00.7504092Z * [new branch] gh/ezyang/3167/base -> origin/gh/ezyang/3167/base 2025-11-03T16:58:00.7505834Z * [new branch] gh/ezyang/3167/head -> origin/gh/ezyang/3167/head 2025-11-03T16:58:00.7507498Z * [new branch] gh/ezyang/3167/orig -> origin/gh/ezyang/3167/orig 2025-11-03T16:58:00.7509999Z * [new branch] gh/ezyang/3168/base -> origin/gh/ezyang/3168/base 2025-11-03T16:58:00.7511950Z * [new branch] gh/ezyang/3168/head -> origin/gh/ezyang/3168/head 2025-11-03T16:58:00.7513638Z * [new branch] gh/ezyang/3168/orig -> origin/gh/ezyang/3168/orig 2025-11-03T16:58:00.7515927Z * [new branch] gh/ezyang/3169/base -> origin/gh/ezyang/3169/base 2025-11-03T16:58:00.7517600Z * [new branch] gh/ezyang/3169/head -> origin/gh/ezyang/3169/head 2025-11-03T16:58:00.7519290Z * [new branch] gh/ezyang/3169/orig -> origin/gh/ezyang/3169/orig 2025-11-03T16:58:00.7522303Z * [new branch] gh/ezyang/3170/base -> origin/gh/ezyang/3170/base 2025-11-03T16:58:00.7524037Z * [new branch] gh/ezyang/3170/head -> origin/gh/ezyang/3170/head 2025-11-03T16:58:00.7525769Z * [new branch] gh/ezyang/3170/orig -> origin/gh/ezyang/3170/orig 2025-11-03T16:58:00.7528158Z * [new branch] gh/ezyang/3171/base -> origin/gh/ezyang/3171/base 2025-11-03T16:58:00.7529970Z * [new branch] gh/ezyang/3171/head -> origin/gh/ezyang/3171/head 2025-11-03T16:58:00.7531819Z * [new branch] gh/ezyang/3171/orig -> origin/gh/ezyang/3171/orig 2025-11-03T16:58:00.7534200Z * [new branch] gh/ezyang/3172/base -> origin/gh/ezyang/3172/base 2025-11-03T16:58:00.7535964Z * [new branch] gh/ezyang/3172/head -> origin/gh/ezyang/3172/head 2025-11-03T16:58:00.7537725Z * [new branch] gh/ezyang/3172/orig -> origin/gh/ezyang/3172/orig 2025-11-03T16:58:00.7540108Z * [new branch] gh/ezyang/3173/base -> origin/gh/ezyang/3173/base 2025-11-03T16:58:00.7541775Z * [new branch] gh/ezyang/3173/head -> origin/gh/ezyang/3173/head 2025-11-03T16:58:00.7543415Z * [new branch] gh/ezyang/3173/orig -> origin/gh/ezyang/3173/orig 2025-11-03T16:58:00.7545828Z * [new branch] gh/ezyang/3174/base -> origin/gh/ezyang/3174/base 2025-11-03T16:58:00.7547533Z * [new branch] gh/ezyang/3174/head -> origin/gh/ezyang/3174/head 2025-11-03T16:58:00.7549214Z * [new branch] gh/ezyang/3174/orig -> origin/gh/ezyang/3174/orig 2025-11-03T16:58:00.7551638Z * [new branch] gh/ezyang/3175/base -> origin/gh/ezyang/3175/base 2025-11-03T16:58:00.7553370Z * [new branch] gh/ezyang/3175/head -> origin/gh/ezyang/3175/head 2025-11-03T16:58:00.7555229Z * [new branch] gh/ezyang/3175/orig -> origin/gh/ezyang/3175/orig 2025-11-03T16:58:00.7558043Z * [new branch] gh/ezyang/3176/base -> origin/gh/ezyang/3176/base 2025-11-03T16:58:00.7559317Z * [new branch] gh/ezyang/3176/head -> origin/gh/ezyang/3176/head 2025-11-03T16:58:00.7560897Z * [new branch] gh/ezyang/3176/orig -> origin/gh/ezyang/3176/orig 2025-11-03T16:58:00.7563835Z * [new branch] gh/ezyang/3177/base -> origin/gh/ezyang/3177/base 2025-11-03T16:58:00.7565078Z * [new branch] gh/ezyang/3177/head -> origin/gh/ezyang/3177/head 2025-11-03T16:58:00.7566750Z * [new branch] gh/ezyang/3177/orig -> origin/gh/ezyang/3177/orig 2025-11-03T16:58:00.7569160Z * [new branch] gh/ezyang/3178/base -> origin/gh/ezyang/3178/base 2025-11-03T16:58:00.7570888Z * [new branch] gh/ezyang/3178/head -> origin/gh/ezyang/3178/head 2025-11-03T16:58:00.7572668Z * [new branch] gh/ezyang/3178/orig -> origin/gh/ezyang/3178/orig 2025-11-03T16:58:00.7575052Z * [new branch] gh/ezyang/3179/base -> origin/gh/ezyang/3179/base 2025-11-03T16:58:00.7576667Z * [new branch] gh/ezyang/3179/head -> origin/gh/ezyang/3179/head 2025-11-03T16:58:00.7578679Z * [new branch] gh/ezyang/3179/orig -> origin/gh/ezyang/3179/orig 2025-11-03T16:58:00.7580742Z * [new branch] gh/ezyang/3180/base -> origin/gh/ezyang/3180/base 2025-11-03T16:58:00.7582506Z * [new branch] gh/ezyang/3180/head -> origin/gh/ezyang/3180/head 2025-11-03T16:58:00.7584211Z * [new branch] gh/ezyang/3180/orig -> origin/gh/ezyang/3180/orig 2025-11-03T16:58:00.7586506Z * [new branch] gh/ezyang/3181/base -> origin/gh/ezyang/3181/base 2025-11-03T16:58:00.7588204Z * [new branch] gh/ezyang/3181/head -> origin/gh/ezyang/3181/head 2025-11-03T16:58:00.7589921Z * [new branch] gh/ezyang/3181/orig -> origin/gh/ezyang/3181/orig 2025-11-03T16:58:00.7592271Z * [new branch] gh/ezyang/3182/base -> origin/gh/ezyang/3182/base 2025-11-03T16:58:00.7593966Z * [new branch] gh/ezyang/3182/head -> origin/gh/ezyang/3182/head 2025-11-03T16:58:00.7595702Z * [new branch] gh/ezyang/3182/orig -> origin/gh/ezyang/3182/orig 2025-11-03T16:58:00.7598248Z * [new branch] gh/ezyang/3183/base -> origin/gh/ezyang/3183/base 2025-11-03T16:58:00.7599970Z * [new branch] gh/ezyang/3183/head -> origin/gh/ezyang/3183/head 2025-11-03T16:58:00.7601636Z * [new branch] gh/ezyang/3183/orig -> origin/gh/ezyang/3183/orig 2025-11-03T16:58:00.7603972Z * [new branch] gh/ezyang/3184/base -> origin/gh/ezyang/3184/base 2025-11-03T16:58:00.7606185Z * [new branch] gh/ezyang/3184/head -> origin/gh/ezyang/3184/head 2025-11-03T16:58:00.7608293Z * [new branch] gh/ezyang/3184/orig -> origin/gh/ezyang/3184/orig 2025-11-03T16:58:00.7611070Z * [new branch] gh/ezyang/3185/base -> origin/gh/ezyang/3185/base 2025-11-03T16:58:00.7612862Z * [new branch] gh/ezyang/3185/head -> origin/gh/ezyang/3185/head 2025-11-03T16:58:00.7614454Z * [new branch] gh/ezyang/3185/orig -> origin/gh/ezyang/3185/orig 2025-11-03T16:58:00.7616748Z * [new branch] gh/ezyang/3186/base -> origin/gh/ezyang/3186/base 2025-11-03T16:58:00.7618316Z * [new branch] gh/ezyang/3186/head -> origin/gh/ezyang/3186/head 2025-11-03T16:58:00.7619945Z * [new branch] gh/ezyang/3186/orig -> origin/gh/ezyang/3186/orig 2025-11-03T16:58:00.7622312Z * [new branch] gh/ezyang/3187/base -> origin/gh/ezyang/3187/base 2025-11-03T16:58:00.7624005Z * [new branch] gh/ezyang/3187/head -> origin/gh/ezyang/3187/head 2025-11-03T16:58:00.7626000Z * [new branch] gh/ezyang/3187/orig -> origin/gh/ezyang/3187/orig 2025-11-03T16:58:00.7628719Z * [new branch] gh/ezyang/3188/base -> origin/gh/ezyang/3188/base 2025-11-03T16:58:00.7630528Z * [new branch] gh/ezyang/3188/head -> origin/gh/ezyang/3188/head 2025-11-03T16:58:00.7632062Z * [new branch] gh/ezyang/3188/orig -> origin/gh/ezyang/3188/orig 2025-11-03T16:58:00.7634966Z * [new branch] gh/ezyang/3189/base -> origin/gh/ezyang/3189/base 2025-11-03T16:58:00.7636265Z * [new branch] gh/ezyang/3189/head -> origin/gh/ezyang/3189/head 2025-11-03T16:58:00.7637746Z * [new branch] gh/ezyang/3189/orig -> origin/gh/ezyang/3189/orig 2025-11-03T16:58:00.7640436Z * [new branch] gh/ezyang/3190/base -> origin/gh/ezyang/3190/base 2025-11-03T16:58:00.7642161Z * [new branch] gh/ezyang/3190/head -> origin/gh/ezyang/3190/head 2025-11-03T16:58:00.7643822Z * [new branch] gh/ezyang/3190/orig -> origin/gh/ezyang/3190/orig 2025-11-03T16:58:00.7646631Z * [new branch] gh/fadara01/1/base -> origin/gh/fadara01/1/base 2025-11-03T16:58:00.7648535Z * [new branch] gh/fadara01/1/head -> origin/gh/fadara01/1/head 2025-11-03T16:58:00.7650121Z * [new branch] gh/fadara01/1/orig -> origin/gh/fadara01/1/orig 2025-11-03T16:58:00.7654007Z * [new branch] gh/fadara01/2/base -> origin/gh/fadara01/2/base 2025-11-03T16:58:00.7655239Z * [new branch] gh/fadara01/2/head -> origin/gh/fadara01/2/head 2025-11-03T16:58:00.7656918Z * [new branch] gh/fadara01/2/orig -> origin/gh/fadara01/2/orig 2025-11-03T16:58:00.7659161Z * [new branch] gh/fadara01/3/base -> origin/gh/fadara01/3/base 2025-11-03T16:58:00.7660827Z * [new branch] gh/fadara01/3/head -> origin/gh/fadara01/3/head 2025-11-03T16:58:00.7662534Z * [new branch] gh/fadara01/3/orig -> origin/gh/fadara01/3/orig 2025-11-03T16:58:00.7664730Z * [new branch] gh/fadara01/4/base -> origin/gh/fadara01/4/base 2025-11-03T16:58:00.7666590Z * [new branch] gh/fadara01/4/head -> origin/gh/fadara01/4/head 2025-11-03T16:58:00.7668267Z * [new branch] gh/fadara01/4/orig -> origin/gh/fadara01/4/orig 2025-11-03T16:58:00.7670615Z * [new branch] gh/fadara01/5/base -> origin/gh/fadara01/5/base 2025-11-03T16:58:00.7672308Z * [new branch] gh/fadara01/5/head -> origin/gh/fadara01/5/head 2025-11-03T16:58:00.7674013Z * [new branch] gh/fadara01/5/orig -> origin/gh/fadara01/5/orig 2025-11-03T16:58:00.7676435Z * [new branch] gh/fadara01/6/base -> origin/gh/fadara01/6/base 2025-11-03T16:58:00.7678126Z * [new branch] gh/fadara01/6/head -> origin/gh/fadara01/6/head 2025-11-03T16:58:00.7679834Z * [new branch] gh/fadara01/6/orig -> origin/gh/fadara01/6/orig 2025-11-03T16:58:00.7682088Z * [new branch] gh/fadara01/7/base -> origin/gh/fadara01/7/base 2025-11-03T16:58:00.7683778Z * [new branch] gh/fadara01/7/head -> origin/gh/fadara01/7/head 2025-11-03T16:58:00.7685498Z * [new branch] gh/fadara01/7/orig -> origin/gh/fadara01/7/orig 2025-11-03T16:58:00.7689041Z * [new branch] gh/fduwjj/182/base -> origin/gh/fduwjj/182/base 2025-11-03T16:58:00.7690212Z * [new branch] gh/fduwjj/182/head -> origin/gh/fduwjj/182/head 2025-11-03T16:58:00.7691830Z * [new branch] gh/fduwjj/182/orig -> origin/gh/fduwjj/182/orig 2025-11-03T16:58:00.7694830Z * [new branch] gh/fduwjj/185/base -> origin/gh/fduwjj/185/base 2025-11-03T16:58:00.7696323Z * [new branch] gh/fduwjj/185/head -> origin/gh/fduwjj/185/head 2025-11-03T16:58:00.7697517Z * [new branch] gh/fduwjj/185/orig -> origin/gh/fduwjj/185/orig 2025-11-03T16:58:00.7700064Z * [new branch] gh/fduwjj/206/base -> origin/gh/fduwjj/206/base 2025-11-03T16:58:00.7701850Z * [new branch] gh/fduwjj/206/head -> origin/gh/fduwjj/206/head 2025-11-03T16:58:00.7703570Z * [new branch] gh/fduwjj/206/orig -> origin/gh/fduwjj/206/orig 2025-11-03T16:58:00.7705862Z * [new branch] gh/fduwjj/211/base -> origin/gh/fduwjj/211/base 2025-11-03T16:58:00.7708231Z * [new branch] gh/fduwjj/211/head -> origin/gh/fduwjj/211/head 2025-11-03T16:58:00.7709804Z * [new branch] gh/fduwjj/211/orig -> origin/gh/fduwjj/211/orig 2025-11-03T16:58:00.7714376Z * [new branch] gh/fduwjj/212/base -> origin/gh/fduwjj/212/base 2025-11-03T16:58:00.7714776Z * [new branch] gh/fduwjj/212/head -> origin/gh/fduwjj/212/head 2025-11-03T16:58:00.7716283Z * [new branch] gh/fduwjj/212/orig -> origin/gh/fduwjj/212/orig 2025-11-03T16:58:00.7718609Z * [new branch] gh/fduwjj/213/base -> origin/gh/fduwjj/213/base 2025-11-03T16:58:00.7720079Z * [new branch] gh/fduwjj/213/head -> origin/gh/fduwjj/213/head 2025-11-03T16:58:00.7721758Z * [new branch] gh/fduwjj/213/orig -> origin/gh/fduwjj/213/orig 2025-11-03T16:58:00.7724254Z * [new branch] gh/fduwjj/217/base -> origin/gh/fduwjj/217/base 2025-11-03T16:58:00.7725993Z * [new branch] gh/fduwjj/217/head -> origin/gh/fduwjj/217/head 2025-11-03T16:58:00.7727671Z * [new branch] gh/fduwjj/217/orig -> origin/gh/fduwjj/217/orig 2025-11-03T16:58:00.7730235Z * [new branch] gh/fduwjj/219/base -> origin/gh/fduwjj/219/base 2025-11-03T16:58:00.7732059Z * [new branch] gh/fduwjj/219/head -> origin/gh/fduwjj/219/head 2025-11-03T16:58:00.7733894Z * [new branch] gh/fduwjj/219/orig -> origin/gh/fduwjj/219/orig 2025-11-03T16:58:00.7736212Z * [new branch] gh/fduwjj/220/base -> origin/gh/fduwjj/220/base 2025-11-03T16:58:00.7737990Z * [new branch] gh/fduwjj/220/head -> origin/gh/fduwjj/220/head 2025-11-03T16:58:00.7739975Z * [new branch] gh/fduwjj/220/orig -> origin/gh/fduwjj/220/orig 2025-11-03T16:58:00.7742276Z * [new branch] gh/fduwjj/221/base -> origin/gh/fduwjj/221/base 2025-11-03T16:58:00.7744035Z * [new branch] gh/fduwjj/221/head -> origin/gh/fduwjj/221/head 2025-11-03T16:58:00.7745701Z * [new branch] gh/fduwjj/221/orig -> origin/gh/fduwjj/221/orig 2025-11-03T16:58:00.7747749Z * [new branch] gh/fduwjj/222/base -> origin/gh/fduwjj/222/base 2025-11-03T16:58:00.7749578Z * [new branch] gh/fduwjj/222/head -> origin/gh/fduwjj/222/head 2025-11-03T16:58:00.7751144Z * [new branch] gh/fduwjj/222/orig -> origin/gh/fduwjj/222/orig 2025-11-03T16:58:00.7753987Z * [new branch] gh/fduwjj/223/base -> origin/gh/fduwjj/223/base 2025-11-03T16:58:00.7755099Z * [new branch] gh/fduwjj/223/head -> origin/gh/fduwjj/223/head 2025-11-03T16:58:00.7756823Z * [new branch] gh/fduwjj/223/orig -> origin/gh/fduwjj/223/orig 2025-11-03T16:58:00.7759136Z * [new branch] gh/fduwjj/224/base -> origin/gh/fduwjj/224/base 2025-11-03T16:58:00.7760863Z * [new branch] gh/fduwjj/224/head -> origin/gh/fduwjj/224/head 2025-11-03T16:58:00.7762579Z * [new branch] gh/fduwjj/224/orig -> origin/gh/fduwjj/224/orig 2025-11-03T16:58:00.7764862Z * [new branch] gh/fduwjj/225/base -> origin/gh/fduwjj/225/base 2025-11-03T16:58:00.7766687Z * [new branch] gh/fduwjj/225/head -> origin/gh/fduwjj/225/head 2025-11-03T16:58:00.7768239Z * [new branch] gh/fduwjj/225/orig -> origin/gh/fduwjj/225/orig 2025-11-03T16:58:00.7770794Z * [new branch] gh/fduwjj/226/base -> origin/gh/fduwjj/226/base 2025-11-03T16:58:00.7772439Z * [new branch] gh/fduwjj/226/head -> origin/gh/fduwjj/226/head 2025-11-03T16:58:00.7774097Z * [new branch] gh/fduwjj/226/orig -> origin/gh/fduwjj/226/orig 2025-11-03T16:58:00.7776548Z * [new branch] gh/fduwjj/227/base -> origin/gh/fduwjj/227/base 2025-11-03T16:58:00.7778269Z * [new branch] gh/fduwjj/227/head -> origin/gh/fduwjj/227/head 2025-11-03T16:58:00.7779942Z * [new branch] gh/fduwjj/227/orig -> origin/gh/fduwjj/227/orig 2025-11-03T16:58:00.7783365Z * [new branch] gh/fduwjj/228/base -> origin/gh/fduwjj/228/base 2025-11-03T16:58:00.7785059Z * [new branch] gh/fduwjj/228/head -> origin/gh/fduwjj/228/head 2025-11-03T16:58:00.7787053Z * [new branch] gh/fduwjj/228/orig -> origin/gh/fduwjj/228/orig 2025-11-03T16:58:00.7789534Z * [new branch] gh/fduwjj/229/base -> origin/gh/fduwjj/229/base 2025-11-03T16:58:00.7791477Z * [new branch] gh/fduwjj/229/head -> origin/gh/fduwjj/229/head 2025-11-03T16:58:00.7793040Z * [new branch] gh/fduwjj/229/orig -> origin/gh/fduwjj/229/orig 2025-11-03T16:58:00.7795775Z * [new branch] gh/fduwjj/230/base -> origin/gh/fduwjj/230/base 2025-11-03T16:58:00.7797224Z * [new branch] gh/fduwjj/230/head -> origin/gh/fduwjj/230/head 2025-11-03T16:58:00.7798862Z * [new branch] gh/fduwjj/230/orig -> origin/gh/fduwjj/230/orig 2025-11-03T16:58:00.7801484Z * [new branch] gh/fduwjj/231/base -> origin/gh/fduwjj/231/base 2025-11-03T16:58:00.7803247Z * [new branch] gh/fduwjj/231/head -> origin/gh/fduwjj/231/head 2025-11-03T16:58:00.7804932Z * [new branch] gh/fduwjj/231/orig -> origin/gh/fduwjj/231/orig 2025-11-03T16:58:00.7807186Z * [new branch] gh/fduwjj/232/base -> origin/gh/fduwjj/232/base 2025-11-03T16:58:00.7809050Z * [new branch] gh/fduwjj/232/head -> origin/gh/fduwjj/232/head 2025-11-03T16:58:00.7811338Z * [new branch] gh/fduwjj/232/orig -> origin/gh/fduwjj/232/orig 2025-11-03T16:58:00.7813399Z * [new branch] gh/fduwjj/233/base -> origin/gh/fduwjj/233/base 2025-11-03T16:58:00.7815082Z * [new branch] gh/fduwjj/233/head -> origin/gh/fduwjj/233/head 2025-11-03T16:58:00.7816778Z * [new branch] gh/fduwjj/233/orig -> origin/gh/fduwjj/233/orig 2025-11-03T16:58:00.7820240Z * [new branch] gh/fegin/318/base -> origin/gh/fegin/318/base 2025-11-03T16:58:00.7821457Z * [new branch] gh/fegin/318/head -> origin/gh/fegin/318/head 2025-11-03T16:58:00.7823286Z * [new branch] gh/fegin/318/orig -> origin/gh/fegin/318/orig 2025-11-03T16:58:00.7826558Z * [new branch] gh/fegin/321/base -> origin/gh/fegin/321/base 2025-11-03T16:58:00.7827614Z * [new branch] gh/fegin/321/head -> origin/gh/fegin/321/head 2025-11-03T16:58:00.7829530Z * [new branch] gh/fegin/321/orig -> origin/gh/fegin/321/orig 2025-11-03T16:58:00.7831814Z * [new branch] gh/fegin/325/base -> origin/gh/fegin/325/base 2025-11-03T16:58:00.7834037Z * [new branch] gh/fegin/325/head -> origin/gh/fegin/325/head 2025-11-03T16:58:00.7835767Z * [new branch] gh/fegin/325/orig -> origin/gh/fegin/325/orig 2025-11-03T16:58:00.7838523Z * [new branch] gh/fegin/326/base -> origin/gh/fegin/326/base 2025-11-03T16:58:00.7840558Z * [new branch] gh/fegin/326/head -> origin/gh/fegin/326/head 2025-11-03T16:58:00.7841875Z * [new branch] gh/fegin/326/orig -> origin/gh/fegin/326/orig 2025-11-03T16:58:00.7844209Z * [new branch] gh/fegin/327/base -> origin/gh/fegin/327/base 2025-11-03T16:58:00.7846013Z * [new branch] gh/fegin/327/head -> origin/gh/fegin/327/head 2025-11-03T16:58:00.7847437Z * [new branch] gh/fegin/327/orig -> origin/gh/fegin/327/orig 2025-11-03T16:58:00.7849786Z * [new branch] gh/fegin/328/base -> origin/gh/fegin/328/base 2025-11-03T16:58:00.7851465Z * [new branch] gh/fegin/328/head -> origin/gh/fegin/328/head 2025-11-03T16:58:00.7853329Z * [new branch] gh/fegin/328/orig -> origin/gh/fegin/328/orig 2025-11-03T16:58:00.7855636Z * [new branch] gh/fegin/329/base -> origin/gh/fegin/329/base 2025-11-03T16:58:00.7857371Z * [new branch] gh/fegin/329/head -> origin/gh/fegin/329/head 2025-11-03T16:58:00.7859281Z * [new branch] gh/fegin/329/orig -> origin/gh/fegin/329/orig 2025-11-03T16:58:00.7861750Z * [new branch] gh/fegin/330/base -> origin/gh/fegin/330/base 2025-11-03T16:58:00.7863109Z * [new branch] gh/fegin/330/head -> origin/gh/fegin/330/head 2025-11-03T16:58:00.7864966Z * [new branch] gh/fegin/330/orig -> origin/gh/fegin/330/orig 2025-11-03T16:58:00.7867554Z * [new branch] gh/fegin/331/base -> origin/gh/fegin/331/base 2025-11-03T16:58:00.7869092Z * [new branch] gh/fegin/331/head -> origin/gh/fegin/331/head 2025-11-03T16:58:00.7870731Z * [new branch] gh/fegin/331/orig -> origin/gh/fegin/331/orig 2025-11-03T16:58:00.7873526Z * [new branch] gh/fffrog/137/base -> origin/gh/fffrog/137/base 2025-11-03T16:58:00.7875198Z * [new branch] gh/fffrog/137/head -> origin/gh/fffrog/137/head 2025-11-03T16:58:00.7877439Z * [new branch] gh/fffrog/137/orig -> origin/gh/fffrog/137/orig 2025-11-03T16:58:00.7879330Z * [new branch] gh/fffrog/160/base -> origin/gh/fffrog/160/base 2025-11-03T16:58:00.7881048Z * [new branch] gh/fffrog/160/head -> origin/gh/fffrog/160/head 2025-11-03T16:58:00.7883804Z * [new branch] gh/fffrog/171/base -> origin/gh/fffrog/171/base 2025-11-03T16:58:00.7885056Z * [new branch] gh/fffrog/171/head -> origin/gh/fffrog/171/head 2025-11-03T16:58:00.7886825Z * [new branch] gh/fffrog/171/orig -> origin/gh/fffrog/171/orig 2025-11-03T16:58:00.7889115Z * [new branch] gh/fffrog/175/base -> origin/gh/fffrog/175/base 2025-11-03T16:58:00.7890856Z * [new branch] gh/fffrog/175/head -> origin/gh/fffrog/175/head 2025-11-03T16:58:00.7892506Z * [new branch] gh/fffrog/175/orig -> origin/gh/fffrog/175/orig 2025-11-03T16:58:00.7894802Z * [new branch] gh/fffrog/176/base -> origin/gh/fffrog/176/base 2025-11-03T16:58:00.7896449Z * [new branch] gh/fffrog/176/head -> origin/gh/fffrog/176/head 2025-11-03T16:58:00.7898202Z * [new branch] gh/fffrog/176/orig -> origin/gh/fffrog/176/orig 2025-11-03T16:58:00.7900483Z * [new branch] gh/fffrog/177/base -> origin/gh/fffrog/177/base 2025-11-03T16:58:00.7902227Z * [new branch] gh/fffrog/177/head -> origin/gh/fffrog/177/head 2025-11-03T16:58:00.7903925Z * [new branch] gh/fffrog/177/orig -> origin/gh/fffrog/177/orig 2025-11-03T16:58:00.7906259Z * [new branch] gh/fffrog/178/base -> origin/gh/fffrog/178/base 2025-11-03T16:58:00.7908060Z * [new branch] gh/fffrog/178/head -> origin/gh/fffrog/178/head 2025-11-03T16:58:00.7909806Z * [new branch] gh/fffrog/178/orig -> origin/gh/fffrog/178/orig 2025-11-03T16:58:00.7912284Z * [new branch] gh/fffrog/179/base -> origin/gh/fffrog/179/base 2025-11-03T16:58:00.7913940Z * [new branch] gh/fffrog/179/head -> origin/gh/fffrog/179/head 2025-11-03T16:58:00.7915646Z * [new branch] gh/fffrog/179/orig -> origin/gh/fffrog/179/orig 2025-11-03T16:58:00.7918459Z * [new branch] gh/fffrog/180/base -> origin/gh/fffrog/180/base 2025-11-03T16:58:00.7919784Z * [new branch] gh/fffrog/180/head -> origin/gh/fffrog/180/head 2025-11-03T16:58:00.7921445Z * [new branch] gh/fffrog/180/orig -> origin/gh/fffrog/180/orig 2025-11-03T16:58:00.7924124Z * [new branch] gh/fffrog/181/base -> origin/gh/fffrog/181/base 2025-11-03T16:58:00.7925428Z * [new branch] gh/fffrog/181/head -> origin/gh/fffrog/181/head 2025-11-03T16:58:00.7927268Z * [new branch] gh/fffrog/181/orig -> origin/gh/fffrog/181/orig 2025-11-03T16:58:00.7929538Z * [new branch] gh/fffrog/182/base -> origin/gh/fffrog/182/base 2025-11-03T16:58:00.7931181Z * [new branch] gh/fffrog/182/head -> origin/gh/fffrog/182/head 2025-11-03T16:58:00.7932959Z * [new branch] gh/fffrog/182/orig -> origin/gh/fffrog/182/orig 2025-11-03T16:58:00.7936039Z * [new branch] gh/fxdawnn/1/base -> origin/gh/fxdawnn/1/base 2025-11-03T16:58:00.7938359Z * [new branch] gh/fxdawnn/1/head -> origin/gh/fxdawnn/1/head 2025-11-03T16:58:00.7939635Z * [new branch] gh/fxdawnn/1/orig -> origin/gh/fxdawnn/1/orig 2025-11-03T16:58:00.7942153Z * [new branch] gh/fxdawnn/2/base -> origin/gh/fxdawnn/2/base 2025-11-03T16:58:00.7943567Z * [new branch] gh/fxdawnn/2/head -> origin/gh/fxdawnn/2/head 2025-11-03T16:58:00.7945181Z * [new branch] gh/fxdawnn/2/orig -> origin/gh/fxdawnn/2/orig 2025-11-03T16:58:00.7947628Z * [new branch] gh/fxdawnn/3/base -> origin/gh/fxdawnn/3/base 2025-11-03T16:58:00.7948966Z * [new branch] gh/fxdawnn/3/head -> origin/gh/fxdawnn/3/head 2025-11-03T16:58:00.7950645Z * [new branch] gh/fxdawnn/3/orig -> origin/gh/fxdawnn/3/orig 2025-11-03T16:58:00.7952972Z * [new branch] gh/fxdawnn/4/base -> origin/gh/fxdawnn/4/base 2025-11-03T16:58:00.7954746Z * [new branch] gh/fxdawnn/4/orig -> origin/gh/fxdawnn/4/orig 2025-11-03T16:58:00.7957271Z * [new branch] gh/fxdawnn/5/base -> origin/gh/fxdawnn/5/base 2025-11-03T16:58:00.7958712Z * [new branch] gh/fxdawnn/5/head -> origin/gh/fxdawnn/5/head 2025-11-03T16:58:00.7960398Z * [new branch] gh/fxdawnn/5/orig -> origin/gh/fxdawnn/5/orig 2025-11-03T16:58:00.7962793Z * [new branch] gh/fxdawnn/6/base -> origin/gh/fxdawnn/6/base 2025-11-03T16:58:00.7964493Z * [new branch] gh/fxdawnn/6/head -> origin/gh/fxdawnn/6/head 2025-11-03T16:58:00.7966191Z * [new branch] gh/fxdawnn/6/orig -> origin/gh/fxdawnn/6/orig 2025-11-03T16:58:00.7982458Z * [new branch] gh/fxdawnn/7/base -> origin/gh/fxdawnn/7/base 2025-11-03T16:58:00.7982877Z * [new branch] gh/fxdawnn/7/head -> origin/gh/fxdawnn/7/head 2025-11-03T16:58:00.7983084Z * [new branch] gh/fxdawnn/7/orig -> origin/gh/fxdawnn/7/orig 2025-11-03T16:58:00.7983268Z * [new branch] gh/gmagogsfm/2/base -> origin/gh/gmagogsfm/2/base 2025-11-03T16:58:00.7983441Z * [new branch] gh/gmagogsfm/2/head -> origin/gh/gmagogsfm/2/head 2025-11-03T16:58:00.7983803Z * [new branch] gh/gmagogsfm/2/orig -> origin/gh/gmagogsfm/2/orig 2025-11-03T16:58:00.7984129Z * [new branch] gh/gmagogsfm/3/base -> origin/gh/gmagogsfm/3/base 2025-11-03T16:58:00.7984416Z * [new branch] gh/gmagogsfm/3/head -> origin/gh/gmagogsfm/3/head 2025-11-03T16:58:00.7984621Z * [new branch] gh/gmagogsfm/3/orig -> origin/gh/gmagogsfm/3/orig 2025-11-03T16:58:00.7987513Z * [new branch] gh/guangyey/134/base -> origin/gh/guangyey/134/base 2025-11-03T16:58:00.7989221Z * [new branch] gh/guangyey/134/head -> origin/gh/guangyey/134/head 2025-11-03T16:58:00.7990883Z * [new branch] gh/guangyey/134/orig -> origin/gh/guangyey/134/orig 2025-11-03T16:58:00.7993158Z * [new branch] gh/guangyey/163/base -> origin/gh/guangyey/163/base 2025-11-03T16:58:00.7994863Z * [new branch] gh/guangyey/163/head -> origin/gh/guangyey/163/head 2025-11-03T16:58:00.7996576Z * [new branch] gh/guangyey/163/orig -> origin/gh/guangyey/163/orig 2025-11-03T16:58:00.7998939Z * [new branch] gh/guangyey/168/base -> origin/gh/guangyey/168/base 2025-11-03T16:58:00.8000615Z * [new branch] gh/guangyey/168/head -> origin/gh/guangyey/168/head 2025-11-03T16:58:00.8002662Z * [new branch] gh/guangyey/168/orig -> origin/gh/guangyey/168/orig 2025-11-03T16:58:00.8004890Z * [new branch] gh/guangyey/169/base -> origin/gh/guangyey/169/base 2025-11-03T16:58:00.8006647Z * [new branch] gh/guangyey/169/head -> origin/gh/guangyey/169/head 2025-11-03T16:58:00.8008378Z * [new branch] gh/guangyey/169/orig -> origin/gh/guangyey/169/orig 2025-11-03T16:58:00.8011028Z * [new branch] gh/guangyey/170/base -> origin/gh/guangyey/170/base 2025-11-03T16:58:00.8012695Z * [new branch] gh/guangyey/170/head -> origin/gh/guangyey/170/head 2025-11-03T16:58:00.8014367Z * [new branch] gh/guangyey/170/orig -> origin/gh/guangyey/170/orig 2025-11-03T16:58:00.8016687Z * [new branch] gh/guangyey/171/base -> origin/gh/guangyey/171/base 2025-11-03T16:58:00.8018388Z * [new branch] gh/guangyey/171/head -> origin/gh/guangyey/171/head 2025-11-03T16:58:00.8020067Z * [new branch] gh/guangyey/171/orig -> origin/gh/guangyey/171/orig 2025-11-03T16:58:00.8022477Z * [new branch] gh/guangyey/176/base -> origin/gh/guangyey/176/base 2025-11-03T16:58:00.8024125Z * [new branch] gh/guangyey/176/head -> origin/gh/guangyey/176/head 2025-11-03T16:58:00.8025833Z * [new branch] gh/guangyey/176/orig -> origin/gh/guangyey/176/orig 2025-11-03T16:58:00.8028161Z * [new branch] gh/guangyey/178/base -> origin/gh/guangyey/178/base 2025-11-03T16:58:00.8030238Z * [new branch] gh/guangyey/178/head -> origin/gh/guangyey/178/head 2025-11-03T16:58:00.8031648Z * [new branch] gh/guangyey/178/orig -> origin/gh/guangyey/178/orig 2025-11-03T16:58:00.8034540Z * [new branch] gh/guangyey/181/base -> origin/gh/guangyey/181/base 2025-11-03T16:58:00.8035740Z * [new branch] gh/guangyey/181/head -> origin/gh/guangyey/181/head 2025-11-03T16:58:00.8037346Z * [new branch] gh/guangyey/181/orig -> origin/gh/guangyey/181/orig 2025-11-03T16:58:00.8039700Z * [new branch] gh/guangyey/182/base -> origin/gh/guangyey/182/base 2025-11-03T16:58:00.8041667Z * [new branch] gh/guangyey/182/head -> origin/gh/guangyey/182/head 2025-11-03T16:58:00.8043457Z * [new branch] gh/guangyey/182/orig -> origin/gh/guangyey/182/orig 2025-11-03T16:58:00.8045737Z * [new branch] gh/guangyey/183/base -> origin/gh/guangyey/183/base 2025-11-03T16:58:00.8047576Z * [new branch] gh/guangyey/183/head -> origin/gh/guangyey/183/head 2025-11-03T16:58:00.8049344Z * [new branch] gh/guangyey/183/orig -> origin/gh/guangyey/183/orig 2025-11-03T16:58:00.8051446Z * [new branch] gh/guangyey/185/base -> origin/gh/guangyey/185/base 2025-11-03T16:58:00.8053167Z * [new branch] gh/guangyey/185/head -> origin/gh/guangyey/185/head 2025-11-03T16:58:00.8054842Z * [new branch] gh/guangyey/185/orig -> origin/gh/guangyey/185/orig 2025-11-03T16:58:00.8057445Z * [new branch] gh/guangyey/186/base -> origin/gh/guangyey/186/base 2025-11-03T16:58:00.8058713Z * [new branch] gh/guangyey/186/head -> origin/gh/guangyey/186/head 2025-11-03T16:58:00.8060309Z * [new branch] gh/guangyey/186/orig -> origin/gh/guangyey/186/orig 2025-11-03T16:58:00.8062784Z * [new branch] gh/guangyey/187/base -> origin/gh/guangyey/187/base 2025-11-03T16:58:00.8064425Z * [new branch] gh/guangyey/187/head -> origin/gh/guangyey/187/head 2025-11-03T16:58:00.8066168Z * [new branch] gh/guangyey/187/orig -> origin/gh/guangyey/187/orig 2025-11-03T16:58:00.8068682Z * [new branch] gh/guangyey/188/base -> origin/gh/guangyey/188/base 2025-11-03T16:58:00.8070414Z * [new branch] gh/guangyey/188/head -> origin/gh/guangyey/188/head 2025-11-03T16:58:00.8072092Z * [new branch] gh/guangyey/188/orig -> origin/gh/guangyey/188/orig 2025-11-03T16:58:00.8074468Z * [new branch] gh/guangyey/190/base -> origin/gh/guangyey/190/base 2025-11-03T16:58:00.8076220Z * [new branch] gh/guangyey/190/head -> origin/gh/guangyey/190/head 2025-11-03T16:58:00.8077878Z * [new branch] gh/guangyey/190/orig -> origin/gh/guangyey/190/orig 2025-11-03T16:58:00.8080738Z * [new branch] gh/guangyey/195/base -> origin/gh/guangyey/195/base 2025-11-03T16:58:00.8082542Z * [new branch] gh/guangyey/195/head -> origin/gh/guangyey/195/head 2025-11-03T16:58:00.8084237Z * [new branch] gh/guangyey/195/orig -> origin/gh/guangyey/195/orig 2025-11-03T16:58:00.8086678Z * [new branch] gh/guangyey/208/base -> origin/gh/guangyey/208/base 2025-11-03T16:58:00.8088383Z * [new branch] gh/guangyey/208/head -> origin/gh/guangyey/208/head 2025-11-03T16:58:00.8090199Z * [new branch] gh/guangyey/208/orig -> origin/gh/guangyey/208/orig 2025-11-03T16:58:00.8092649Z * [new branch] gh/guangyey/210/base -> origin/gh/guangyey/210/base 2025-11-03T16:58:00.8094359Z * [new branch] gh/guangyey/210/head -> origin/gh/guangyey/210/head 2025-11-03T16:58:00.8096074Z * [new branch] gh/guangyey/210/orig -> origin/gh/guangyey/210/orig 2025-11-03T16:58:00.8098450Z * [new branch] gh/guangyey/212/base -> origin/gh/guangyey/212/base 2025-11-03T16:58:00.8100139Z * [new branch] gh/guangyey/212/head -> origin/gh/guangyey/212/head 2025-11-03T16:58:00.8101876Z * [new branch] gh/guangyey/212/orig -> origin/gh/guangyey/212/orig 2025-11-03T16:58:00.8104299Z * [new branch] gh/guangyey/213/base -> origin/gh/guangyey/213/base 2025-11-03T16:58:00.8107817Z * [new branch] gh/guangyey/213/head -> origin/gh/guangyey/213/head 2025-11-03T16:58:00.8108359Z * [new branch] gh/guangyey/213/orig -> origin/gh/guangyey/213/orig 2025-11-03T16:58:00.8112113Z * [new branch] gh/guangyey/214/base -> origin/gh/guangyey/214/base 2025-11-03T16:58:00.8113112Z * [new branch] gh/guangyey/214/head -> origin/gh/guangyey/214/head 2025-11-03T16:58:00.8114883Z * [new branch] gh/guangyey/214/orig -> origin/gh/guangyey/214/orig 2025-11-03T16:58:00.8117711Z * [new branch] gh/guangyey/215/base -> origin/gh/guangyey/215/base 2025-11-03T16:58:00.8119209Z * [new branch] gh/guangyey/215/head -> origin/gh/guangyey/215/head 2025-11-03T16:58:00.8120870Z * [new branch] gh/guangyey/215/orig -> origin/gh/guangyey/215/orig 2025-11-03T16:58:00.8123412Z * [new branch] gh/guangyey/216/base -> origin/gh/guangyey/216/base 2025-11-03T16:58:00.8124993Z * [new branch] gh/guangyey/216/head -> origin/gh/guangyey/216/head 2025-11-03T16:58:00.8126592Z * [new branch] gh/guangyey/216/orig -> origin/gh/guangyey/216/orig 2025-11-03T16:58:00.8129097Z * [new branch] gh/guangyey/217/base -> origin/gh/guangyey/217/base 2025-11-03T16:58:00.8130811Z * [new branch] gh/guangyey/217/head -> origin/gh/guangyey/217/head 2025-11-03T16:58:00.8132542Z * [new branch] gh/guangyey/217/orig -> origin/gh/guangyey/217/orig 2025-11-03T16:58:00.8134901Z * [new branch] gh/guangyey/218/base -> origin/gh/guangyey/218/base 2025-11-03T16:58:00.8136775Z * [new branch] gh/guangyey/218/head -> origin/gh/guangyey/218/head 2025-11-03T16:58:00.8138410Z * [new branch] gh/guangyey/218/orig -> origin/gh/guangyey/218/orig 2025-11-03T16:58:00.8140908Z * [new branch] gh/guangyey/219/base -> origin/gh/guangyey/219/base 2025-11-03T16:58:00.8142559Z * [new branch] gh/guangyey/219/head -> origin/gh/guangyey/219/head 2025-11-03T16:58:00.8144269Z * [new branch] gh/guangyey/219/orig -> origin/gh/guangyey/219/orig 2025-11-03T16:58:00.8146597Z * [new branch] gh/guangyey/220/base -> origin/gh/guangyey/220/base 2025-11-03T16:58:00.8148432Z * [new branch] gh/guangyey/220/head -> origin/gh/guangyey/220/head 2025-11-03T16:58:00.8150085Z * [new branch] gh/guangyey/220/orig -> origin/gh/guangyey/220/orig 2025-11-03T16:58:00.8152994Z * [new branch] gh/guangyey/221/base -> origin/gh/guangyey/221/base 2025-11-03T16:58:00.8154733Z * [new branch] gh/guangyey/221/head -> origin/gh/guangyey/221/head 2025-11-03T16:58:00.8156401Z * [new branch] gh/guangyey/221/orig -> origin/gh/guangyey/221/orig 2025-11-03T16:58:00.8158765Z * [new branch] gh/guangyey/222/base -> origin/gh/guangyey/222/base 2025-11-03T16:58:00.8160486Z * [new branch] gh/guangyey/222/head -> origin/gh/guangyey/222/head 2025-11-03T16:58:00.8162770Z * [new branch] gh/guangyey/222/orig -> origin/gh/guangyey/222/orig 2025-11-03T16:58:00.8165226Z * [new branch] gh/guangyey/223/base -> origin/gh/guangyey/223/base 2025-11-03T16:58:00.8166943Z * [new branch] gh/guangyey/223/head -> origin/gh/guangyey/223/head 2025-11-03T16:58:00.8168937Z * [new branch] gh/guangyey/223/orig -> origin/gh/guangyey/223/orig 2025-11-03T16:58:00.8171630Z * [new branch] gh/guangyey/224/base -> origin/gh/guangyey/224/base 2025-11-03T16:58:00.8172903Z * [new branch] gh/guangyey/224/head -> origin/gh/guangyey/224/head 2025-11-03T16:58:00.8174591Z * [new branch] gh/guangyey/224/orig -> origin/gh/guangyey/224/orig 2025-11-03T16:58:00.8177347Z * [new branch] gh/guangyey/225/base -> origin/gh/guangyey/225/base 2025-11-03T16:58:00.8178579Z * [new branch] gh/guangyey/225/head -> origin/gh/guangyey/225/head 2025-11-03T16:58:00.8180342Z * [new branch] gh/guangyey/225/orig -> origin/gh/guangyey/225/orig 2025-11-03T16:58:00.8182703Z * [new branch] gh/guangyey/226/base -> origin/gh/guangyey/226/base 2025-11-03T16:58:00.8184393Z * [new branch] gh/guangyey/226/head -> origin/gh/guangyey/226/head 2025-11-03T16:58:00.8186363Z * [new branch] gh/guangyey/226/orig -> origin/gh/guangyey/226/orig 2025-11-03T16:58:00.8188579Z * [new branch] gh/guangyey/227/base -> origin/gh/guangyey/227/base 2025-11-03T16:58:00.8190310Z * [new branch] gh/guangyey/227/head -> origin/gh/guangyey/227/head 2025-11-03T16:58:00.8192015Z * [new branch] gh/guangyey/227/orig -> origin/gh/guangyey/227/orig 2025-11-03T16:58:00.8194490Z * [new branch] gh/guangyey/228/base -> origin/gh/guangyey/228/base 2025-11-03T16:58:00.8196243Z * [new branch] gh/guangyey/228/head -> origin/gh/guangyey/228/head 2025-11-03T16:58:00.8197936Z * [new branch] gh/guangyey/228/orig -> origin/gh/guangyey/228/orig 2025-11-03T16:58:00.8200331Z * [new branch] gh/guangyey/229/base -> origin/gh/guangyey/229/base 2025-11-03T16:58:00.8202053Z * [new branch] gh/guangyey/229/head -> origin/gh/guangyey/229/head 2025-11-03T16:58:00.8203956Z * [new branch] gh/guangyey/229/orig -> origin/gh/guangyey/229/orig 2025-11-03T16:58:00.8206355Z * [new branch] gh/guangyey/230/base -> origin/gh/guangyey/230/base 2025-11-03T16:58:00.8208091Z * [new branch] gh/guangyey/230/head -> origin/gh/guangyey/230/head 2025-11-03T16:58:00.8210522Z * [new branch] gh/guangyey/230/orig -> origin/gh/guangyey/230/orig 2025-11-03T16:58:00.8213332Z * [new branch] gh/guangyey/231/base -> origin/gh/guangyey/231/base 2025-11-03T16:58:00.8214486Z * [new branch] gh/guangyey/231/head -> origin/gh/guangyey/231/head 2025-11-03T16:58:00.8216311Z * [new branch] gh/guangyey/231/orig -> origin/gh/guangyey/231/orig 2025-11-03T16:58:00.8218472Z * [new branch] gh/guangyey/232/base -> origin/gh/guangyey/232/base 2025-11-03T16:58:00.8220611Z * [new branch] gh/guangyey/232/head -> origin/gh/guangyey/232/head 2025-11-03T16:58:00.8222441Z * [new branch] gh/guangyey/232/orig -> origin/gh/guangyey/232/orig 2025-11-03T16:58:00.8224847Z * [new branch] gh/guangyey/233/base -> origin/gh/guangyey/233/base 2025-11-03T16:58:00.8226578Z * [new branch] gh/guangyey/233/head -> origin/gh/guangyey/233/head 2025-11-03T16:58:00.8228355Z * [new branch] gh/guangyey/233/orig -> origin/gh/guangyey/233/orig 2025-11-03T16:58:00.8230762Z * [new branch] gh/guangyey/234/base -> origin/gh/guangyey/234/base 2025-11-03T16:58:00.8232493Z * [new branch] gh/guangyey/234/head -> origin/gh/guangyey/234/head 2025-11-03T16:58:00.8234160Z * [new branch] gh/guangyey/234/orig -> origin/gh/guangyey/234/orig 2025-11-03T16:58:00.8236554Z * [new branch] gh/guangyey/235/base -> origin/gh/guangyey/235/base 2025-11-03T16:58:00.8239020Z * [new branch] gh/guangyey/235/head -> origin/gh/guangyey/235/head 2025-11-03T16:58:00.8240172Z * [new branch] gh/guangyey/235/orig -> origin/gh/guangyey/235/orig 2025-11-03T16:58:00.8243468Z * [new branch] gh/guilhermeleobas/107/base -> origin/gh/guilhermeleobas/107/base 2025-11-03T16:58:00.8244726Z * [new branch] gh/guilhermeleobas/107/head -> origin/gh/guilhermeleobas/107/head 2025-11-03T16:58:00.8246887Z * [new branch] gh/guilhermeleobas/107/orig -> origin/gh/guilhermeleobas/107/orig 2025-11-03T16:58:00.8249065Z * [new branch] gh/guilhermeleobas/108/base -> origin/gh/guilhermeleobas/108/base 2025-11-03T16:58:00.8250800Z * [new branch] gh/guilhermeleobas/108/head -> origin/gh/guilhermeleobas/108/head 2025-11-03T16:58:00.8252474Z * [new branch] gh/guilhermeleobas/108/orig -> origin/gh/guilhermeleobas/108/orig 2025-11-03T16:58:00.8254837Z * [new branch] gh/guilhermeleobas/147/base -> origin/gh/guilhermeleobas/147/base 2025-11-03T16:58:00.8256732Z * [new branch] gh/guilhermeleobas/147/head -> origin/gh/guilhermeleobas/147/head 2025-11-03T16:58:00.8258278Z * [new branch] gh/guilhermeleobas/147/orig -> origin/gh/guilhermeleobas/147/orig 2025-11-03T16:58:00.8260616Z * [new branch] gh/guilhermeleobas/150/base -> origin/gh/guilhermeleobas/150/base 2025-11-03T16:58:00.8262278Z * [new branch] gh/guilhermeleobas/150/head -> origin/gh/guilhermeleobas/150/head 2025-11-03T16:58:00.8264015Z * [new branch] gh/guilhermeleobas/150/orig -> origin/gh/guilhermeleobas/150/orig 2025-11-03T16:58:00.8266281Z * [new branch] gh/guilhermeleobas/166/base -> origin/gh/guilhermeleobas/166/base 2025-11-03T16:58:00.8267996Z * [new branch] gh/guilhermeleobas/166/head -> origin/gh/guilhermeleobas/166/head 2025-11-03T16:58:00.8269734Z * [new branch] gh/guilhermeleobas/166/orig -> origin/gh/guilhermeleobas/166/orig 2025-11-03T16:58:00.8272272Z * [new branch] gh/guilhermeleobas/168/base -> origin/gh/guilhermeleobas/168/base 2025-11-03T16:58:00.8274028Z * [new branch] gh/guilhermeleobas/168/head -> origin/gh/guilhermeleobas/168/head 2025-11-03T16:58:00.8275720Z * [new branch] gh/guilhermeleobas/168/orig -> origin/gh/guilhermeleobas/168/orig 2025-11-03T16:58:00.8278551Z * [new branch] gh/guilhermeleobas/169/base -> origin/gh/guilhermeleobas/169/base 2025-11-03T16:58:00.8279948Z * [new branch] gh/guilhermeleobas/169/head -> origin/gh/guilhermeleobas/169/head 2025-11-03T16:58:00.8281508Z * [new branch] gh/guilhermeleobas/169/orig -> origin/gh/guilhermeleobas/169/orig 2025-11-03T16:58:00.8284351Z * [new branch] gh/guilhermeleobas/170/base -> origin/gh/guilhermeleobas/170/base 2025-11-03T16:58:00.8285697Z * [new branch] gh/guilhermeleobas/170/head -> origin/gh/guilhermeleobas/170/head 2025-11-03T16:58:00.8287361Z * [new branch] gh/guilhermeleobas/170/orig -> origin/gh/guilhermeleobas/170/orig 2025-11-03T16:58:00.8289539Z * [new branch] gh/guilhermeleobas/171/base -> origin/gh/guilhermeleobas/171/base 2025-11-03T16:58:00.8291262Z * [new branch] gh/guilhermeleobas/171/head -> origin/gh/guilhermeleobas/171/head 2025-11-03T16:58:00.8292982Z * [new branch] gh/guilhermeleobas/171/orig -> origin/gh/guilhermeleobas/171/orig 2025-11-03T16:58:00.8295282Z * [new branch] gh/guilhermeleobas/173/base -> origin/gh/guilhermeleobas/173/base 2025-11-03T16:58:00.8297082Z * [new branch] gh/guilhermeleobas/173/head -> origin/gh/guilhermeleobas/173/head 2025-11-03T16:58:00.8298735Z * [new branch] gh/guilhermeleobas/173/orig -> origin/gh/guilhermeleobas/173/orig 2025-11-03T16:58:00.8301069Z * [new branch] gh/guilhermeleobas/193/base -> origin/gh/guilhermeleobas/193/base 2025-11-03T16:58:00.8302788Z * [new branch] gh/guilhermeleobas/193/head -> origin/gh/guilhermeleobas/193/head 2025-11-03T16:58:00.8305158Z * [new branch] gh/guilhermeleobas/193/orig -> origin/gh/guilhermeleobas/193/orig 2025-11-03T16:58:00.8308171Z * [new branch] gh/guilhermeleobas/204/base -> origin/gh/guilhermeleobas/204/base 2025-11-03T16:58:00.8309728Z * [new branch] gh/guilhermeleobas/204/head -> origin/gh/guilhermeleobas/204/head 2025-11-03T16:58:00.8311451Z * [new branch] gh/guilhermeleobas/204/orig -> origin/gh/guilhermeleobas/204/orig 2025-11-03T16:58:00.8314163Z * [new branch] gh/guilhermeleobas/211/base -> origin/gh/guilhermeleobas/211/base 2025-11-03T16:58:00.8315442Z * [new branch] gh/guilhermeleobas/211/head -> origin/gh/guilhermeleobas/211/head 2025-11-03T16:58:00.8317092Z * [new branch] gh/guilhermeleobas/211/orig -> origin/gh/guilhermeleobas/211/orig 2025-11-03T16:58:00.8319275Z * [new branch] gh/guilhermeleobas/226/base -> origin/gh/guilhermeleobas/226/base 2025-11-03T16:58:00.8321111Z * [new branch] gh/guilhermeleobas/226/head -> origin/gh/guilhermeleobas/226/head 2025-11-03T16:58:00.8322744Z * [new branch] gh/guilhermeleobas/226/orig -> origin/gh/guilhermeleobas/226/orig 2025-11-03T16:58:00.8325005Z * [new branch] gh/guilhermeleobas/236/base -> origin/gh/guilhermeleobas/236/base 2025-11-03T16:58:00.8326706Z * [new branch] gh/guilhermeleobas/236/head -> origin/gh/guilhermeleobas/236/head 2025-11-03T16:58:00.8328442Z * [new branch] gh/guilhermeleobas/236/orig -> origin/gh/guilhermeleobas/236/orig 2025-11-03T16:58:00.8330872Z * [new branch] gh/guilhermeleobas/239/base -> origin/gh/guilhermeleobas/239/base 2025-11-03T16:58:00.8332556Z * [new branch] gh/guilhermeleobas/239/head -> origin/gh/guilhermeleobas/239/head 2025-11-03T16:58:00.8334297Z * [new branch] gh/guilhermeleobas/239/orig -> origin/gh/guilhermeleobas/239/orig 2025-11-03T16:58:00.8337217Z * [new branch] gh/guilhermeleobas/246/base -> origin/gh/guilhermeleobas/246/base 2025-11-03T16:58:00.8339099Z * [new branch] gh/guilhermeleobas/246/head -> origin/gh/guilhermeleobas/246/head 2025-11-03T16:58:00.8340779Z * [new branch] gh/guilhermeleobas/246/orig -> origin/gh/guilhermeleobas/246/orig 2025-11-03T16:58:00.8343113Z * [new branch] gh/guilhermeleobas/247/base -> origin/gh/guilhermeleobas/247/base 2025-11-03T16:58:00.8344857Z * [new branch] gh/guilhermeleobas/247/head -> origin/gh/guilhermeleobas/247/head 2025-11-03T16:58:00.8346577Z * [new branch] gh/guilhermeleobas/247/orig -> origin/gh/guilhermeleobas/247/orig 2025-11-03T16:58:00.8349005Z * [new branch] gh/guilhermeleobas/248/base -> origin/gh/guilhermeleobas/248/base 2025-11-03T16:58:00.8350677Z * [new branch] gh/guilhermeleobas/248/head -> origin/gh/guilhermeleobas/248/head 2025-11-03T16:58:00.8352466Z * [new branch] gh/guilhermeleobas/248/orig -> origin/gh/guilhermeleobas/248/orig 2025-11-03T16:58:00.8354926Z * [new branch] gh/guilhermeleobas/250/base -> origin/gh/guilhermeleobas/250/base 2025-11-03T16:58:00.8356754Z * [new branch] gh/guilhermeleobas/250/head -> origin/gh/guilhermeleobas/250/head 2025-11-03T16:58:00.8358486Z * [new branch] gh/guilhermeleobas/250/orig -> origin/gh/guilhermeleobas/250/orig 2025-11-03T16:58:00.8360853Z * [new branch] gh/guilhermeleobas/251/base -> origin/gh/guilhermeleobas/251/base 2025-11-03T16:58:00.8362638Z * [new branch] gh/guilhermeleobas/251/head -> origin/gh/guilhermeleobas/251/head 2025-11-03T16:58:00.8364333Z * [new branch] gh/guilhermeleobas/251/orig -> origin/gh/guilhermeleobas/251/orig 2025-11-03T16:58:00.8366713Z * [new branch] gh/guilhermeleobas/252/base -> origin/gh/guilhermeleobas/252/base 2025-11-03T16:58:00.8368964Z * [new branch] gh/guilhermeleobas/252/head -> origin/gh/guilhermeleobas/252/head 2025-11-03T16:58:00.8370777Z * [new branch] gh/guilhermeleobas/252/orig -> origin/gh/guilhermeleobas/252/orig 2025-11-03T16:58:00.8374157Z * [new branch] gh/henrylhtsang/150/base -> origin/gh/henrylhtsang/150/base 2025-11-03T16:58:00.8375844Z * [new branch] gh/henrylhtsang/150/head -> origin/gh/henrylhtsang/150/head 2025-11-03T16:58:00.8377715Z * [new branch] gh/henrylhtsang/150/orig -> origin/gh/henrylhtsang/150/orig 2025-11-03T16:58:00.8380969Z * [new branch] gh/huydhn/1/next -> origin/gh/huydhn/1/next 2025-11-03T16:58:00.8383179Z * [new branch] gh/huydhn/2/next -> origin/gh/huydhn/2/next 2025-11-03T16:58:00.8385441Z * [new branch] gh/huydhn/3/next -> origin/gh/huydhn/3/next 2025-11-03T16:58:00.8388304Z * [new branch] gh/huydhn/4/next -> origin/gh/huydhn/4/next 2025-11-03T16:58:00.8390595Z * [new branch] gh/huydhn/5/next -> origin/gh/huydhn/5/next 2025-11-03T16:58:00.8392616Z * [new branch] gh/huydhn/6/next -> origin/gh/huydhn/6/next 2025-11-03T16:58:00.8395812Z * [new branch] gh/int3/97/base -> origin/gh/int3/97/base 2025-11-03T16:58:00.8397537Z * [new branch] gh/int3/97/head -> origin/gh/int3/97/head 2025-11-03T16:58:00.8400750Z * [new branch] gh/isuruf/101/base -> origin/gh/isuruf/101/base 2025-11-03T16:58:00.8402550Z * [new branch] gh/isuruf/101/head -> origin/gh/isuruf/101/head 2025-11-03T16:58:00.8405680Z * [new branch] gh/isuruf/146/base -> origin/gh/isuruf/146/base 2025-11-03T16:58:00.8407039Z * [new branch] gh/isuruf/146/head -> origin/gh/isuruf/146/head 2025-11-03T16:58:00.8408808Z * [new branch] gh/isuruf/146/orig -> origin/gh/isuruf/146/orig 2025-11-03T16:58:00.8411840Z * [new branch] gh/isuruf/157/base -> origin/gh/isuruf/157/base 2025-11-03T16:58:00.8413272Z * [new branch] gh/isuruf/157/head -> origin/gh/isuruf/157/head 2025-11-03T16:58:00.8414864Z * [new branch] gh/isuruf/157/orig -> origin/gh/isuruf/157/orig 2025-11-03T16:58:00.8417492Z * [new branch] gh/isuruf/81/base -> origin/gh/isuruf/81/base 2025-11-03T16:58:00.8418830Z * [new branch] gh/isuruf/81/head -> origin/gh/isuruf/81/head 2025-11-03T16:58:00.8420526Z * [new branch] gh/isuruf/81/orig -> origin/gh/isuruf/81/orig 2025-11-03T16:58:00.8423428Z * [new branch] gh/jamesjwu/176/base -> origin/gh/jamesjwu/176/base 2025-11-03T16:58:00.8425137Z * [new branch] gh/jamesjwu/176/head -> origin/gh/jamesjwu/176/head 2025-11-03T16:58:00.8426853Z * [new branch] gh/jamesjwu/176/orig -> origin/gh/jamesjwu/176/orig 2025-11-03T16:58:00.8429089Z * [new branch] gh/jamesjwu/187/base -> origin/gh/jamesjwu/187/base 2025-11-03T16:58:00.8430788Z * [new branch] gh/jamesjwu/187/head -> origin/gh/jamesjwu/187/head 2025-11-03T16:58:00.8432467Z * [new branch] gh/jamesjwu/187/orig -> origin/gh/jamesjwu/187/orig 2025-11-03T16:58:00.8434883Z * [new branch] gh/jamesjwu/195/base -> origin/gh/jamesjwu/195/base 2025-11-03T16:58:00.8436512Z * [new branch] gh/jamesjwu/195/head -> origin/gh/jamesjwu/195/head 2025-11-03T16:58:00.8438257Z * [new branch] gh/jamesjwu/195/orig -> origin/gh/jamesjwu/195/orig 2025-11-03T16:58:00.8440585Z * [new branch] gh/jamesjwu/196/base -> origin/gh/jamesjwu/196/base 2025-11-03T16:58:00.8442352Z * [new branch] gh/jamesjwu/196/head -> origin/gh/jamesjwu/196/head 2025-11-03T16:58:00.8444028Z * [new branch] gh/jamesjwu/196/orig -> origin/gh/jamesjwu/196/orig 2025-11-03T16:58:00.8446349Z * [new branch] gh/jamesjwu/197/base -> origin/gh/jamesjwu/197/base 2025-11-03T16:58:00.8448532Z * [new branch] gh/jamesjwu/197/head -> origin/gh/jamesjwu/197/head 2025-11-03T16:58:00.8450247Z * [new branch] gh/jamesjwu/197/orig -> origin/gh/jamesjwu/197/orig 2025-11-03T16:58:00.8452439Z * [new branch] gh/jamesjwu/198/base -> origin/gh/jamesjwu/198/base 2025-11-03T16:58:00.8454242Z * [new branch] gh/jamesjwu/198/head -> origin/gh/jamesjwu/198/head 2025-11-03T16:58:00.8455948Z * [new branch] gh/jamesjwu/198/orig -> origin/gh/jamesjwu/198/orig 2025-11-03T16:58:00.8457997Z * [new branch] gh/jamesjwu/199/base -> origin/gh/jamesjwu/199/base 2025-11-03T16:58:00.8459840Z * [new branch] gh/jamesjwu/199/head -> origin/gh/jamesjwu/199/head 2025-11-03T16:58:00.8461391Z * [new branch] gh/jamesjwu/199/orig -> origin/gh/jamesjwu/199/orig 2025-11-03T16:58:00.8464474Z * [new branch] gh/jamesjwu/200/base -> origin/gh/jamesjwu/200/base 2025-11-03T16:58:00.8465497Z * [new branch] gh/jamesjwu/200/head -> origin/gh/jamesjwu/200/head 2025-11-03T16:58:00.8467172Z * [new branch] gh/jamesjwu/200/orig -> origin/gh/jamesjwu/200/orig 2025-11-03T16:58:00.8469346Z * [new branch] gh/jamesjwu/201/base -> origin/gh/jamesjwu/201/base 2025-11-03T16:58:00.8471069Z * [new branch] gh/jamesjwu/201/head -> origin/gh/jamesjwu/201/head 2025-11-03T16:58:00.8473279Z * [new branch] gh/jamesjwu/201/orig -> origin/gh/jamesjwu/201/orig 2025-11-03T16:58:00.8475793Z * [new branch] gh/jamesjwu/202/base -> origin/gh/jamesjwu/202/base 2025-11-03T16:58:00.8477474Z * [new branch] gh/jamesjwu/202/head -> origin/gh/jamesjwu/202/head 2025-11-03T16:58:00.8479667Z * [new branch] gh/jamesjwu/202/orig -> origin/gh/jamesjwu/202/orig 2025-11-03T16:58:00.8482229Z * [new branch] gh/jamesjwu/203/base -> origin/gh/jamesjwu/203/base 2025-11-03T16:58:00.8483871Z * [new branch] gh/jamesjwu/203/head -> origin/gh/jamesjwu/203/head 2025-11-03T16:58:00.8485508Z * [new branch] gh/jamesjwu/203/orig -> origin/gh/jamesjwu/203/orig 2025-11-03T16:58:00.8487819Z * [new branch] gh/jamesjwu/204/base -> origin/gh/jamesjwu/204/base 2025-11-03T16:58:00.8489764Z * [new branch] gh/jamesjwu/204/head -> origin/gh/jamesjwu/204/head 2025-11-03T16:58:00.8491334Z * [new branch] gh/jamesjwu/204/orig -> origin/gh/jamesjwu/204/orig 2025-11-03T16:58:00.8494141Z * [new branch] gh/jamesjwu/205/base -> origin/gh/jamesjwu/205/base 2025-11-03T16:58:00.8495502Z * [new branch] gh/jamesjwu/205/head -> origin/gh/jamesjwu/205/head 2025-11-03T16:58:00.8497194Z * [new branch] gh/jamesjwu/205/orig -> origin/gh/jamesjwu/205/orig 2025-11-03T16:58:00.8499270Z * [new branch] gh/jamesjwu/206/base -> origin/gh/jamesjwu/206/base 2025-11-03T16:58:00.8501015Z * [new branch] gh/jamesjwu/206/head -> origin/gh/jamesjwu/206/head 2025-11-03T16:58:00.8502766Z * [new branch] gh/jamesjwu/206/orig -> origin/gh/jamesjwu/206/orig 2025-11-03T16:58:00.8505120Z * [new branch] gh/jamesjwu/52/base -> origin/gh/jamesjwu/52/base 2025-11-03T16:58:00.8506839Z * [new branch] gh/jamesjwu/52/head -> origin/gh/jamesjwu/52/head 2025-11-03T16:58:00.8508994Z * [new branch] gh/jamesjwu/53/base -> origin/gh/jamesjwu/53/base 2025-11-03T16:58:00.8512104Z * [new branch] gh/jamesjwu/53/head -> origin/gh/jamesjwu/53/head 2025-11-03T16:58:00.8514449Z * [new branch] gh/jamesjwu/54/base -> origin/gh/jamesjwu/54/base 2025-11-03T16:58:00.8516184Z * [new branch] gh/jamesjwu/54/head -> origin/gh/jamesjwu/54/head 2025-11-03T16:58:00.8519085Z * [new branch] gh/jamesjwu/55/base -> origin/gh/jamesjwu/55/base 2025-11-03T16:58:00.8520794Z * [new branch] gh/jamesjwu/55/head -> origin/gh/jamesjwu/55/head 2025-11-03T16:58:00.8523059Z * [new branch] gh/jamesjwu/56/base -> origin/gh/jamesjwu/56/base 2025-11-03T16:58:00.8524752Z * [new branch] gh/jamesjwu/56/head -> origin/gh/jamesjwu/56/head 2025-11-03T16:58:00.8527023Z * [new branch] gh/jamesjwu/57/base -> origin/gh/jamesjwu/57/base 2025-11-03T16:58:00.8528747Z * [new branch] gh/jamesjwu/57/head -> origin/gh/jamesjwu/57/head 2025-11-03T16:58:00.8531121Z * [new branch] gh/jamesjwu/58/base -> origin/gh/jamesjwu/58/base 2025-11-03T16:58:00.8532811Z * [new branch] gh/jamesjwu/58/head -> origin/gh/jamesjwu/58/head 2025-11-03T16:58:00.8535334Z * [new branch] gh/jamesjwu/59/base -> origin/gh/jamesjwu/59/base 2025-11-03T16:58:00.8536910Z * [new branch] gh/jamesjwu/59/head -> origin/gh/jamesjwu/59/head 2025-11-03T16:58:00.8539132Z * [new branch] gh/jamesjwu/60/base -> origin/gh/jamesjwu/60/base 2025-11-03T16:58:00.8540825Z * [new branch] gh/jamesjwu/60/head -> origin/gh/jamesjwu/60/head 2025-11-03T16:58:00.8543126Z * [new branch] gh/jamesjwu/61/base -> origin/gh/jamesjwu/61/base 2025-11-03T16:58:00.8544839Z * [new branch] gh/jamesjwu/61/head -> origin/gh/jamesjwu/61/head 2025-11-03T16:58:00.8547082Z * [new branch] gh/jamesjwu/62/base -> origin/gh/jamesjwu/62/base 2025-11-03T16:58:00.8548745Z * [new branch] gh/jamesjwu/62/head -> origin/gh/jamesjwu/62/head 2025-11-03T16:58:00.8551048Z * [new branch] gh/jamesjwu/63/base -> origin/gh/jamesjwu/63/base 2025-11-03T16:58:00.8553249Z * [new branch] gh/jamesjwu/63/head -> origin/gh/jamesjwu/63/head 2025-11-03T16:58:00.8555824Z * [new branch] gh/jamesjwu/64/base -> origin/gh/jamesjwu/64/base 2025-11-03T16:58:00.8558130Z * [new branch] gh/jamesjwu/64/head -> origin/gh/jamesjwu/64/head 2025-11-03T16:58:00.8560068Z * [new branch] gh/jamesjwu/65/base -> origin/gh/jamesjwu/65/base 2025-11-03T16:58:00.8561690Z * [new branch] gh/jamesjwu/65/head -> origin/gh/jamesjwu/65/head 2025-11-03T16:58:00.8565091Z * [new branch] gh/janeyx99/165/base -> origin/gh/janeyx99/165/base 2025-11-03T16:58:00.8566552Z * [new branch] gh/janeyx99/165/head -> origin/gh/janeyx99/165/head 2025-11-03T16:58:00.8568225Z * [new branch] gh/janeyx99/165/orig -> origin/gh/janeyx99/165/orig 2025-11-03T16:58:00.8571082Z * [new branch] gh/janeyx99/201/base -> origin/gh/janeyx99/201/base 2025-11-03T16:58:00.8572791Z * [new branch] gh/janeyx99/201/head -> origin/gh/janeyx99/201/head 2025-11-03T16:58:00.8574485Z * [new branch] gh/janeyx99/201/orig -> origin/gh/janeyx99/201/orig 2025-11-03T16:58:00.8577542Z * [new branch] gh/janeyx99/225/base -> origin/gh/janeyx99/225/base 2025-11-03T16:58:00.8579295Z * [new branch] gh/janeyx99/225/head -> origin/gh/janeyx99/225/head 2025-11-03T16:58:00.8581023Z * [new branch] gh/janeyx99/225/orig -> origin/gh/janeyx99/225/orig 2025-11-03T16:58:00.8583517Z * [new branch] gh/janeyx99/299/base -> origin/gh/janeyx99/299/base 2025-11-03T16:58:00.8585256Z * [new branch] gh/janeyx99/299/head -> origin/gh/janeyx99/299/head 2025-11-03T16:58:00.8587560Z * [new branch] gh/janeyx99/299/orig -> origin/gh/janeyx99/299/orig 2025-11-03T16:58:00.8590193Z * [new branch] gh/janeyx99/302/base -> origin/gh/janeyx99/302/base 2025-11-03T16:58:00.8591509Z * [new branch] gh/janeyx99/302/head -> origin/gh/janeyx99/302/head 2025-11-03T16:58:00.8594126Z * [new branch] gh/janeyx99/303/base -> origin/gh/janeyx99/303/base 2025-11-03T16:58:00.8595430Z * [new branch] gh/janeyx99/303/head -> origin/gh/janeyx99/303/head 2025-11-03T16:58:00.8597691Z * [new branch] gh/janeyx99/305/base -> origin/gh/janeyx99/305/base 2025-11-03T16:58:00.8599237Z * [new branch] gh/janeyx99/305/head -> origin/gh/janeyx99/305/head 2025-11-03T16:58:00.8601413Z * [new branch] gh/janeyx99/306/base -> origin/gh/janeyx99/306/base 2025-11-03T16:58:00.8603074Z * [new branch] gh/janeyx99/306/head -> origin/gh/janeyx99/306/head 2025-11-03T16:58:00.8605911Z * [new branch] gh/janeyx99/310/base -> origin/gh/janeyx99/310/base 2025-11-03T16:58:00.8607812Z * [new branch] gh/janeyx99/310/head -> origin/gh/janeyx99/310/head 2025-11-03T16:58:00.8609952Z * [new branch] gh/janeyx99/310/orig -> origin/gh/janeyx99/310/orig 2025-11-03T16:58:00.8612603Z * [new branch] gh/janeyx99/311/base -> origin/gh/janeyx99/311/base 2025-11-03T16:58:00.8614232Z * [new branch] gh/janeyx99/311/head -> origin/gh/janeyx99/311/head 2025-11-03T16:58:00.8616022Z * [new branch] gh/janeyx99/311/orig -> origin/gh/janeyx99/311/orig 2025-11-03T16:58:00.8618079Z * [new branch] gh/janeyx99/312/base -> origin/gh/janeyx99/312/base 2025-11-03T16:58:00.8620050Z * [new branch] gh/janeyx99/312/head -> origin/gh/janeyx99/312/head 2025-11-03T16:58:00.8621552Z * [new branch] gh/janeyx99/312/orig -> origin/gh/janeyx99/312/orig 2025-11-03T16:58:00.8624180Z * [new branch] gh/janeyx99/313/base -> origin/gh/janeyx99/313/base 2025-11-03T16:58:00.8625538Z * [new branch] gh/janeyx99/313/head -> origin/gh/janeyx99/313/head 2025-11-03T16:58:00.8627344Z * [new branch] gh/janeyx99/313/orig -> origin/gh/janeyx99/313/orig 2025-11-03T16:58:00.8629604Z * [new branch] gh/janeyx99/314/base -> origin/gh/janeyx99/314/base 2025-11-03T16:58:00.8631312Z * [new branch] gh/janeyx99/314/head -> origin/gh/janeyx99/314/head 2025-11-03T16:58:00.8633079Z * [new branch] gh/janeyx99/314/orig -> origin/gh/janeyx99/314/orig 2025-11-03T16:58:00.8635389Z * [new branch] gh/janeyx99/315/base -> origin/gh/janeyx99/315/base 2025-11-03T16:58:00.8637310Z * [new branch] gh/janeyx99/315/head -> origin/gh/janeyx99/315/head 2025-11-03T16:58:00.8639067Z * [new branch] gh/janeyx99/315/orig -> origin/gh/janeyx99/315/orig 2025-11-03T16:58:00.8641496Z * [new branch] gh/janeyx99/316/base -> origin/gh/janeyx99/316/base 2025-11-03T16:58:00.8643208Z * [new branch] gh/janeyx99/316/head -> origin/gh/janeyx99/316/head 2025-11-03T16:58:00.8644917Z * [new branch] gh/janeyx99/316/orig -> origin/gh/janeyx99/316/orig 2025-11-03T16:58:00.8647097Z * [new branch] gh/janeyx99/317/base -> origin/gh/janeyx99/317/base 2025-11-03T16:58:00.8648852Z * [new branch] gh/janeyx99/317/head -> origin/gh/janeyx99/317/head 2025-11-03T16:58:00.8650592Z * [new branch] gh/janeyx99/317/orig -> origin/gh/janeyx99/317/orig 2025-11-03T16:58:00.8653777Z * [new branch] gh/janeyx99/318/base -> origin/gh/janeyx99/318/base 2025-11-03T16:58:00.8655505Z * [new branch] gh/janeyx99/318/head -> origin/gh/janeyx99/318/head 2025-11-03T16:58:00.8657805Z * [new branch] gh/janeyx99/318/orig -> origin/gh/janeyx99/318/orig 2025-11-03T16:58:00.8660379Z * [new branch] gh/janeyx99/319/base -> origin/gh/janeyx99/319/base 2025-11-03T16:58:00.8661885Z * [new branch] gh/janeyx99/319/head -> origin/gh/janeyx99/319/head 2025-11-03T16:58:00.8663537Z * [new branch] gh/janeyx99/319/orig -> origin/gh/janeyx99/319/orig 2025-11-03T16:58:00.8666366Z * [new branch] gh/janeyx99/320/base -> origin/gh/janeyx99/320/base 2025-11-03T16:58:00.8667620Z * [new branch] gh/janeyx99/320/head -> origin/gh/janeyx99/320/head 2025-11-03T16:58:00.8669427Z * [new branch] gh/janeyx99/320/orig -> origin/gh/janeyx99/320/orig 2025-11-03T16:58:00.8671888Z * [new branch] gh/janeyx99/321/base -> origin/gh/janeyx99/321/base 2025-11-03T16:58:00.8673617Z * [new branch] gh/janeyx99/321/head -> origin/gh/janeyx99/321/head 2025-11-03T16:58:00.8675332Z * [new branch] gh/janeyx99/321/orig -> origin/gh/janeyx99/321/orig 2025-11-03T16:58:00.8678569Z * [new branch] gh/janeyx99/322/base -> origin/gh/janeyx99/322/base 2025-11-03T16:58:00.8679750Z * [new branch] gh/janeyx99/322/head -> origin/gh/janeyx99/322/head 2025-11-03T16:58:00.8681362Z * [new branch] gh/janeyx99/322/orig -> origin/gh/janeyx99/322/orig 2025-11-03T16:58:00.8684082Z * [new branch] gh/janeyx99/323/base -> origin/gh/janeyx99/323/base 2025-11-03T16:58:00.8685381Z * [new branch] gh/janeyx99/323/head -> origin/gh/janeyx99/323/head 2025-11-03T16:58:00.8687195Z * [new branch] gh/janeyx99/323/orig -> origin/gh/janeyx99/323/orig 2025-11-03T16:58:00.8689567Z * [new branch] gh/janeyx99/324/base -> origin/gh/janeyx99/324/base 2025-11-03T16:58:00.8691248Z * [new branch] gh/janeyx99/324/head -> origin/gh/janeyx99/324/head 2025-11-03T16:58:00.8692966Z * [new branch] gh/janeyx99/324/orig -> origin/gh/janeyx99/324/orig 2025-11-03T16:58:00.8695505Z * [new branch] gh/janeyx99/325/base -> origin/gh/janeyx99/325/base 2025-11-03T16:58:00.8697206Z * [new branch] gh/janeyx99/325/head -> origin/gh/janeyx99/325/head 2025-11-03T16:58:00.8698934Z * [new branch] gh/janeyx99/325/orig -> origin/gh/janeyx99/325/orig 2025-11-03T16:58:00.8701488Z * [new branch] gh/janeyx99/326/base -> origin/gh/janeyx99/326/base 2025-11-03T16:58:00.8703381Z * [new branch] gh/janeyx99/326/head -> origin/gh/janeyx99/326/head 2025-11-03T16:58:00.8705165Z * [new branch] gh/janeyx99/326/orig -> origin/gh/janeyx99/326/orig 2025-11-03T16:58:00.8707650Z * [new branch] gh/janeyx99/88/base -> origin/gh/janeyx99/88/base 2025-11-03T16:58:00.8709903Z * [new branch] gh/janeyx99/88/head -> origin/gh/janeyx99/88/head 2025-11-03T16:58:00.8712067Z * [new branch] gh/janeyx99/88/orig -> origin/gh/janeyx99/88/orig 2025-11-03T16:58:00.8714808Z * [new branch] gh/jansel/360/base -> origin/gh/jansel/360/base 2025-11-03T16:58:00.8716546Z * [new branch] gh/jansel/360/head -> origin/gh/jansel/360/head 2025-11-03T16:58:00.8718844Z * [new branch] gh/jansel/451/base -> origin/gh/jansel/451/base 2025-11-03T16:58:00.8720555Z * [new branch] gh/jansel/451/head -> origin/gh/jansel/451/head 2025-11-03T16:58:00.8722230Z * [new branch] gh/jansel/451/orig -> origin/gh/jansel/451/orig 2025-11-03T16:58:00.8724600Z * [new branch] gh/jansel/462/base -> origin/gh/jansel/462/base 2025-11-03T16:58:00.8726397Z * [new branch] gh/jansel/462/head -> origin/gh/jansel/462/head 2025-11-03T16:58:00.8728111Z * [new branch] gh/jansel/462/orig -> origin/gh/jansel/462/orig 2025-11-03T16:58:00.8730540Z * [new branch] gh/jansel/531/base -> origin/gh/jansel/531/base 2025-11-03T16:58:00.8732233Z * [new branch] gh/jansel/531/head -> origin/gh/jansel/531/head 2025-11-03T16:58:00.8733986Z * [new branch] gh/jansel/531/orig -> origin/gh/jansel/531/orig 2025-11-03T16:58:00.8736251Z * [new branch] gh/jansel/533/base -> origin/gh/jansel/533/base 2025-11-03T16:58:00.8738944Z * [new branch] gh/jansel/533/head -> origin/gh/jansel/533/head 2025-11-03T16:58:00.8740195Z * [new branch] gh/jansel/533/orig -> origin/gh/jansel/533/orig 2025-11-03T16:58:00.8742876Z * [new branch] gh/jansel/549/base -> origin/gh/jansel/549/base 2025-11-03T16:58:00.8744289Z * [new branch] gh/jansel/549/head -> origin/gh/jansel/549/head 2025-11-03T16:58:00.8746089Z * [new branch] gh/jansel/549/orig -> origin/gh/jansel/549/orig 2025-11-03T16:58:00.8748563Z * [new branch] gh/jansel/550/base -> origin/gh/jansel/550/base 2025-11-03T16:58:00.8750197Z * [new branch] gh/jansel/550/head -> origin/gh/jansel/550/head 2025-11-03T16:58:00.8752138Z * [new branch] gh/jansel/550/orig -> origin/gh/jansel/550/orig 2025-11-03T16:58:00.8754546Z * [new branch] gh/jansel/551/base -> origin/gh/jansel/551/base 2025-11-03T16:58:00.8756247Z * [new branch] gh/jansel/551/head -> origin/gh/jansel/551/head 2025-11-03T16:58:00.8757876Z * [new branch] gh/jansel/551/orig -> origin/gh/jansel/551/orig 2025-11-03T16:58:00.8760639Z * [new branch] gh/jbschlosser/247/base -> origin/gh/jbschlosser/247/base 2025-11-03T16:58:00.8762337Z * [new branch] gh/jbschlosser/247/head -> origin/gh/jbschlosser/247/head 2025-11-03T16:58:00.8764071Z * [new branch] gh/jbschlosser/247/orig -> origin/gh/jbschlosser/247/orig 2025-11-03T16:58:00.8766612Z * [new branch] gh/jbschlosser/250/base -> origin/gh/jbschlosser/250/base 2025-11-03T16:58:00.8768702Z * [new branch] gh/jbschlosser/250/head -> origin/gh/jbschlosser/250/head 2025-11-03T16:58:00.8770478Z * [new branch] gh/jbschlosser/250/orig -> origin/gh/jbschlosser/250/orig 2025-11-03T16:58:00.8773205Z * [new branch] gh/jgong5/1/base -> origin/gh/jgong5/1/base 2025-11-03T16:58:00.8775294Z * [new branch] gh/jgong5/1/head -> origin/gh/jgong5/1/head 2025-11-03T16:58:00.8776934Z * [new branch] gh/jgong5/1/orig -> origin/gh/jgong5/1/orig 2025-11-03T16:58:00.8780202Z * [new branch] gh/jiayisunx/59/base -> origin/gh/jiayisunx/59/base 2025-11-03T16:58:00.8781259Z * [new branch] gh/jiayisunx/59/head -> origin/gh/jiayisunx/59/head 2025-11-03T16:58:00.8782918Z * [new branch] gh/jiayisunx/59/orig -> origin/gh/jiayisunx/59/orig 2025-11-03T16:58:00.8785628Z * [new branch] gh/jiayisunx/61/base -> origin/gh/jiayisunx/61/base 2025-11-03T16:58:00.8787006Z * [new branch] gh/jiayisunx/61/head -> origin/gh/jiayisunx/61/head 2025-11-03T16:58:00.8788592Z * [new branch] gh/jiayisunx/61/orig -> origin/gh/jiayisunx/61/orig 2025-11-03T16:58:00.8791299Z * [new branch] gh/jiayisunx/68/base -> origin/gh/jiayisunx/68/base 2025-11-03T16:58:00.8793060Z * [new branch] gh/jiayisunx/68/head -> origin/gh/jiayisunx/68/head 2025-11-03T16:58:00.8795205Z * [new branch] gh/jiayisunx/68/orig -> origin/gh/jiayisunx/68/orig 2025-11-03T16:58:00.8797438Z * [new branch] gh/jiayisunx/71/base -> origin/gh/jiayisunx/71/base 2025-11-03T16:58:00.8799231Z * [new branch] gh/jiayisunx/71/head -> origin/gh/jiayisunx/71/head 2025-11-03T16:58:00.8800909Z * [new branch] gh/jiayisunx/71/orig -> origin/gh/jiayisunx/71/orig 2025-11-03T16:58:00.8803280Z * [new branch] gh/jiayisunx/77/base -> origin/gh/jiayisunx/77/base 2025-11-03T16:58:00.8804974Z * [new branch] gh/jiayisunx/77/head -> origin/gh/jiayisunx/77/head 2025-11-03T16:58:00.8806708Z * [new branch] gh/jiayisunx/77/orig -> origin/gh/jiayisunx/77/orig 2025-11-03T16:58:00.8808996Z * [new branch] gh/jiayisunx/78/base -> origin/gh/jiayisunx/78/base 2025-11-03T16:58:00.8811206Z * [new branch] gh/jiayisunx/78/head -> origin/gh/jiayisunx/78/head 2025-11-03T16:58:00.8812917Z * [new branch] gh/jiayisunx/78/orig -> origin/gh/jiayisunx/78/orig 2025-11-03T16:58:00.8815269Z * [new branch] gh/jiayisunx/79/base -> origin/gh/jiayisunx/79/base 2025-11-03T16:58:00.8816963Z * [new branch] gh/jiayisunx/79/head -> origin/gh/jiayisunx/79/head 2025-11-03T16:58:00.8819331Z * [new branch] gh/jiayisunx/79/orig -> origin/gh/jiayisunx/79/orig 2025-11-03T16:58:00.8821251Z * [new branch] gh/jiayisunx/80/base -> origin/gh/jiayisunx/80/base 2025-11-03T16:58:00.8822885Z * [new branch] gh/jiayisunx/80/head -> origin/gh/jiayisunx/80/head 2025-11-03T16:58:00.8824595Z * [new branch] gh/jiayisunx/80/orig -> origin/gh/jiayisunx/80/orig 2025-11-03T16:58:00.8827348Z * [new branch] gh/jiayisunx/81/base -> origin/gh/jiayisunx/81/base 2025-11-03T16:58:00.8828613Z * [new branch] gh/jiayisunx/81/head -> origin/gh/jiayisunx/81/head 2025-11-03T16:58:00.8830349Z * [new branch] gh/jiayisunx/81/orig -> origin/gh/jiayisunx/81/orig 2025-11-03T16:58:00.8832733Z * [new branch] gh/jiayisunx/82/base -> origin/gh/jiayisunx/82/base 2025-11-03T16:58:00.8834419Z * [new branch] gh/jiayisunx/82/head -> origin/gh/jiayisunx/82/head 2025-11-03T16:58:00.8836189Z * [new branch] gh/jiayisunx/82/orig -> origin/gh/jiayisunx/82/orig 2025-11-03T16:58:00.8838626Z * [new branch] gh/jiayisunx/83/base -> origin/gh/jiayisunx/83/base 2025-11-03T16:58:00.8840366Z * [new branch] gh/jiayisunx/83/head -> origin/gh/jiayisunx/83/head 2025-11-03T16:58:00.8842752Z * [new branch] gh/jiayisunx/83/orig -> origin/gh/jiayisunx/83/orig 2025-11-03T16:58:00.8845037Z * [new branch] gh/jiayisunx/84/base -> origin/gh/jiayisunx/84/base 2025-11-03T16:58:00.8846774Z * [new branch] gh/jiayisunx/84/head -> origin/gh/jiayisunx/84/head 2025-11-03T16:58:00.8848852Z * [new branch] gh/jiayisunx/84/orig -> origin/gh/jiayisunx/84/orig 2025-11-03T16:58:00.8851257Z * [new branch] gh/jiayisunx/85/base -> origin/gh/jiayisunx/85/base 2025-11-03T16:58:00.8853081Z * [new branch] gh/jiayisunx/85/head -> origin/gh/jiayisunx/85/head 2025-11-03T16:58:00.8854856Z * [new branch] gh/jiayisunx/85/orig -> origin/gh/jiayisunx/85/orig 2025-11-03T16:58:00.8857013Z * [new branch] gh/jiayisunx/86/base -> origin/gh/jiayisunx/86/base 2025-11-03T16:58:00.8858631Z * [new branch] gh/jiayisunx/86/head -> origin/gh/jiayisunx/86/head 2025-11-03T16:58:00.8860321Z * [new branch] gh/jiayisunx/86/orig -> origin/gh/jiayisunx/86/orig 2025-11-03T16:58:00.8862700Z * [new branch] gh/jiayisunx/87/base -> origin/gh/jiayisunx/87/base 2025-11-03T16:58:00.8864449Z * [new branch] gh/jiayisunx/87/head -> origin/gh/jiayisunx/87/head 2025-11-03T16:58:00.8866176Z * [new branch] gh/jiayisunx/87/orig -> origin/gh/jiayisunx/87/orig 2025-11-03T16:58:00.8868950Z * [new branch] gh/jjwu@meta.com/1/base -> origin/gh/jjwu@meta.com/1/base 2025-11-03T16:58:00.8870623Z * [new branch] gh/jjwu@meta.com/1/head -> origin/gh/jjwu@meta.com/1/head 2025-11-03T16:58:00.8874340Z * [new branch] gh/karthickai/6/base -> origin/gh/karthickai/6/base 2025-11-03T16:58:00.8876598Z * [new branch] gh/karthickai/6/head -> origin/gh/karthickai/6/head 2025-11-03T16:58:00.8878405Z * [new branch] gh/karthickai/6/orig -> origin/gh/karthickai/6/orig 2025-11-03T16:58:00.8880656Z * [new branch] gh/karthickai/7/base -> origin/gh/karthickai/7/base 2025-11-03T16:58:00.8882513Z * [new branch] gh/karthickai/7/head -> origin/gh/karthickai/7/head 2025-11-03T16:58:00.8884284Z * [new branch] gh/karthickai/7/orig -> origin/gh/karthickai/7/orig 2025-11-03T16:58:00.8886845Z * [new branch] gh/karthickai/8/base -> origin/gh/karthickai/8/base 2025-11-03T16:58:00.8888664Z * [new branch] gh/karthickai/8/head -> origin/gh/karthickai/8/head 2025-11-03T16:58:00.8890625Z * [new branch] gh/karthickai/8/orig -> origin/gh/karthickai/8/orig 2025-11-03T16:58:00.8892963Z * [new branch] gh/karthickai/9/base -> origin/gh/karthickai/9/base 2025-11-03T16:58:00.8894676Z * [new branch] gh/karthickai/9/head -> origin/gh/karthickai/9/head 2025-11-03T16:58:00.8896424Z * [new branch] gh/karthickai/9/orig -> origin/gh/karthickai/9/orig 2025-11-03T16:58:00.8899204Z * [new branch] gh/kurtamohler/32/base -> origin/gh/kurtamohler/32/base 2025-11-03T16:58:00.8900932Z * [new branch] gh/kurtamohler/32/head -> origin/gh/kurtamohler/32/head 2025-11-03T16:58:00.8902676Z * [new branch] gh/kurtamohler/32/orig -> origin/gh/kurtamohler/32/orig 2025-11-03T16:58:00.8904947Z * [new branch] gh/kurtamohler/33/base -> origin/gh/kurtamohler/33/base 2025-11-03T16:58:00.8906680Z * [new branch] gh/kurtamohler/33/head -> origin/gh/kurtamohler/33/head 2025-11-03T16:58:00.8908508Z * [new branch] gh/kurtamohler/33/orig -> origin/gh/kurtamohler/33/orig 2025-11-03T16:58:00.8911862Z * [new branch] gh/kurtamohler/34/base -> origin/gh/kurtamohler/34/base 2025-11-03T16:58:00.8913731Z * [new branch] gh/kurtamohler/34/head -> origin/gh/kurtamohler/34/head 2025-11-03T16:58:00.8915485Z * [new branch] gh/kurtamohler/34/orig -> origin/gh/kurtamohler/34/orig 2025-11-03T16:58:00.8917884Z * [new branch] gh/kurtamohler/55/base -> origin/gh/kurtamohler/55/base 2025-11-03T16:58:00.8919843Z * [new branch] gh/kurtamohler/55/head -> origin/gh/kurtamohler/55/head 2025-11-03T16:58:00.8921321Z * [new branch] gh/kurtamohler/55/orig -> origin/gh/kurtamohler/55/orig 2025-11-03T16:58:00.8924142Z * [new branch] gh/kurtamohler/56/base -> origin/gh/kurtamohler/56/base 2025-11-03T16:58:00.8925361Z * [new branch] gh/kurtamohler/56/head -> origin/gh/kurtamohler/56/head 2025-11-03T16:58:00.8927192Z * [new branch] gh/kurtamohler/56/orig -> origin/gh/kurtamohler/56/orig 2025-11-03T16:58:00.8929456Z * [new branch] gh/kurtamohler/57/base -> origin/gh/kurtamohler/57/base 2025-11-03T16:58:00.8931157Z * [new branch] gh/kurtamohler/57/head -> origin/gh/kurtamohler/57/head 2025-11-03T16:58:00.8932955Z * [new branch] gh/kurtamohler/57/orig -> origin/gh/kurtamohler/57/orig 2025-11-03T16:58:00.8935308Z * [new branch] gh/kurtamohler/58/base -> origin/gh/kurtamohler/58/base 2025-11-03T16:58:00.8937532Z * [new branch] gh/kurtamohler/58/head -> origin/gh/kurtamohler/58/head 2025-11-03T16:58:00.8938838Z * [new branch] gh/kurtamohler/58/orig -> origin/gh/kurtamohler/58/orig 2025-11-03T16:58:00.8941565Z * [new branch] gh/kurtamohler/59/base -> origin/gh/kurtamohler/59/base 2025-11-03T16:58:00.8942871Z * [new branch] gh/kurtamohler/59/head -> origin/gh/kurtamohler/59/head 2025-11-03T16:58:00.8944525Z * [new branch] gh/kurtamohler/59/orig -> origin/gh/kurtamohler/59/orig 2025-11-03T16:58:00.8947703Z * [new branch] gh/kwen2501/130/base -> origin/gh/kwen2501/130/base 2025-11-03T16:58:00.8949275Z * [new branch] gh/kwen2501/130/head -> origin/gh/kwen2501/130/head 2025-11-03T16:58:00.8951032Z * [new branch] gh/kwen2501/130/orig -> origin/gh/kwen2501/130/orig 2025-11-03T16:58:00.8953446Z * [new branch] gh/kwen2501/170/base -> origin/gh/kwen2501/170/base 2025-11-03T16:58:00.8955228Z * [new branch] gh/kwen2501/170/head -> origin/gh/kwen2501/170/head 2025-11-03T16:58:00.8957673Z * [new branch] gh/kwen2501/187/base -> origin/gh/kwen2501/187/base 2025-11-03T16:58:00.8959484Z * [new branch] gh/kwen2501/187/head -> origin/gh/kwen2501/187/head 2025-11-03T16:58:00.8961392Z * [new branch] gh/kwen2501/187/orig -> origin/gh/kwen2501/187/orig 2025-11-03T16:58:00.8963684Z * [new branch] gh/kwen2501/188/base -> origin/gh/kwen2501/188/base 2025-11-03T16:58:00.8965377Z * [new branch] gh/kwen2501/188/head -> origin/gh/kwen2501/188/head 2025-11-03T16:58:00.8967715Z * [new branch] gh/kwen2501/188/orig -> origin/gh/kwen2501/188/orig 2025-11-03T16:58:00.8969767Z * [new branch] gh/kwen2501/211/base -> origin/gh/kwen2501/211/base 2025-11-03T16:58:00.8971462Z * [new branch] gh/kwen2501/211/head -> origin/gh/kwen2501/211/head 2025-11-03T16:58:00.8974361Z * [new branch] gh/kwen2501/224/base -> origin/gh/kwen2501/224/base 2025-11-03T16:58:00.8975999Z * [new branch] gh/kwen2501/224/head -> origin/gh/kwen2501/224/head 2025-11-03T16:58:00.8977603Z * [new branch] gh/kwen2501/224/orig -> origin/gh/kwen2501/224/orig 2025-11-03T16:58:00.8980315Z * [new branch] gh/kwen2501/228/base -> origin/gh/kwen2501/228/base 2025-11-03T16:58:00.8982270Z * [new branch] gh/kwen2501/228/head -> origin/gh/kwen2501/228/head 2025-11-03T16:58:00.8983786Z * [new branch] gh/kwen2501/228/orig -> origin/gh/kwen2501/228/orig 2025-11-03T16:58:00.8986242Z * [new branch] gh/kwen2501/231/base -> origin/gh/kwen2501/231/base 2025-11-03T16:58:00.8988041Z * [new branch] gh/kwen2501/231/head -> origin/gh/kwen2501/231/head 2025-11-03T16:58:00.8989947Z * [new branch] gh/kwen2501/231/orig -> origin/gh/kwen2501/231/orig 2025-11-03T16:58:00.8992335Z * [new branch] gh/kwen2501/234/base -> origin/gh/kwen2501/234/base 2025-11-03T16:58:00.8993851Z * [new branch] gh/kwen2501/234/head -> origin/gh/kwen2501/234/head 2025-11-03T16:58:00.8995595Z * [new branch] gh/kwen2501/234/orig -> origin/gh/kwen2501/234/orig 2025-11-03T16:58:00.8997890Z * [new branch] gh/kwen2501/235/base -> origin/gh/kwen2501/235/base 2025-11-03T16:58:00.8999616Z * [new branch] gh/kwen2501/235/head -> origin/gh/kwen2501/235/head 2025-11-03T16:58:00.9001166Z * [new branch] gh/kwen2501/235/orig -> origin/gh/kwen2501/235/orig 2025-11-03T16:58:00.9003585Z * [new branch] gh/kwen2501/236/base -> origin/gh/kwen2501/236/base 2025-11-03T16:58:00.9005345Z * [new branch] gh/kwen2501/236/head -> origin/gh/kwen2501/236/head 2025-11-03T16:58:00.9007585Z * [new branch] gh/kwen2501/236/orig -> origin/gh/kwen2501/236/orig 2025-11-03T16:58:00.9009934Z * [new branch] gh/kwen2501/237/base -> origin/gh/kwen2501/237/base 2025-11-03T16:58:00.9011666Z * [new branch] gh/kwen2501/237/head -> origin/gh/kwen2501/237/head 2025-11-03T16:58:00.9013307Z * [new branch] gh/kwen2501/237/orig -> origin/gh/kwen2501/237/orig 2025-11-03T16:58:00.9016062Z * [new branch] gh/kwen2501/238/base -> origin/gh/kwen2501/238/base 2025-11-03T16:58:00.9017341Z * [new branch] gh/kwen2501/238/head -> origin/gh/kwen2501/238/head 2025-11-03T16:58:00.9018945Z * [new branch] gh/kwen2501/238/orig -> origin/gh/kwen2501/238/orig 2025-11-03T16:58:00.9021269Z * [new branch] gh/kwen2501/240/base -> origin/gh/kwen2501/240/base 2025-11-03T16:58:00.9022950Z * [new branch] gh/kwen2501/240/head -> origin/gh/kwen2501/240/head 2025-11-03T16:58:00.9024652Z * [new branch] gh/kwen2501/240/orig -> origin/gh/kwen2501/240/orig 2025-11-03T16:58:00.9027070Z * [new branch] gh/kwen2501/241/base -> origin/gh/kwen2501/241/base 2025-11-03T16:58:00.9028748Z * [new branch] gh/kwen2501/241/head -> origin/gh/kwen2501/241/head 2025-11-03T16:58:00.9030616Z * [new branch] gh/kwen2501/241/orig -> origin/gh/kwen2501/241/orig 2025-11-03T16:58:00.9032899Z * [new branch] gh/kwen2501/247/base -> origin/gh/kwen2501/247/base 2025-11-03T16:58:00.9034581Z * [new branch] gh/kwen2501/247/head -> origin/gh/kwen2501/247/head 2025-11-03T16:58:00.9036291Z * [new branch] gh/kwen2501/247/orig -> origin/gh/kwen2501/247/orig 2025-11-03T16:58:00.9039160Z * [new branch] gh/kwen2501/252/base -> origin/gh/kwen2501/252/base 2025-11-03T16:58:00.9040870Z * [new branch] gh/kwen2501/252/head -> origin/gh/kwen2501/252/head 2025-11-03T16:58:00.9042585Z * [new branch] gh/kwen2501/252/orig -> origin/gh/kwen2501/252/orig 2025-11-03T16:58:00.9045138Z * [new branch] gh/kwen2501/259/base -> origin/gh/kwen2501/259/base 2025-11-03T16:58:00.9046866Z * [new branch] gh/kwen2501/259/head -> origin/gh/kwen2501/259/head 2025-11-03T16:58:00.9049125Z * [new branch] gh/kwen2501/259/orig -> origin/gh/kwen2501/259/orig 2025-11-03T16:58:00.9052159Z * [new branch] gh/kwen2501/260/base -> origin/gh/kwen2501/260/base 2025-11-03T16:58:00.9053902Z * [new branch] gh/kwen2501/260/head -> origin/gh/kwen2501/260/head 2025-11-03T16:58:00.9055625Z * [new branch] gh/kwen2501/260/orig -> origin/gh/kwen2501/260/orig 2025-11-03T16:58:00.9057772Z * [new branch] gh/kwen2501/268/base -> origin/gh/kwen2501/268/base 2025-11-03T16:58:00.9059703Z * [new branch] gh/kwen2501/268/head -> origin/gh/kwen2501/268/head 2025-11-03T16:58:00.9061288Z * [new branch] gh/kwen2501/268/orig -> origin/gh/kwen2501/268/orig 2025-11-03T16:58:00.9064173Z * [new branch] gh/kwen2501/269/base -> origin/gh/kwen2501/269/base 2025-11-03T16:58:00.9065528Z * [new branch] gh/kwen2501/269/head -> origin/gh/kwen2501/269/head 2025-11-03T16:58:00.9067202Z * [new branch] gh/kwen2501/269/orig -> origin/gh/kwen2501/269/orig 2025-11-03T16:58:00.9069565Z * [new branch] gh/kwen2501/270/base -> origin/gh/kwen2501/270/base 2025-11-03T16:58:00.9071284Z * [new branch] gh/kwen2501/270/head -> origin/gh/kwen2501/270/head 2025-11-03T16:58:00.9073049Z * [new branch] gh/kwen2501/270/orig -> origin/gh/kwen2501/270/orig 2025-11-03T16:58:00.9075569Z * [new branch] gh/kwen2501/271/base -> origin/gh/kwen2501/271/base 2025-11-03T16:58:00.9077841Z * [new branch] gh/kwen2501/271/head -> origin/gh/kwen2501/271/head 2025-11-03T16:58:00.9079156Z * [new branch] gh/kwen2501/271/orig -> origin/gh/kwen2501/271/orig 2025-11-03T16:58:00.9082071Z * [new branch] gh/kwen2501/272/base -> origin/gh/kwen2501/272/base 2025-11-03T16:58:00.9083320Z * [new branch] gh/kwen2501/272/head -> origin/gh/kwen2501/272/head 2025-11-03T16:58:00.9084978Z * [new branch] gh/kwen2501/272/orig -> origin/gh/kwen2501/272/orig 2025-11-03T16:58:00.9087595Z * [new branch] gh/kwen2501/273/base -> origin/gh/kwen2501/273/base 2025-11-03T16:58:00.9089085Z * [new branch] gh/kwen2501/273/head -> origin/gh/kwen2501/273/head 2025-11-03T16:58:00.9090781Z * [new branch] gh/kwen2501/273/orig -> origin/gh/kwen2501/273/orig 2025-11-03T16:58:00.9093183Z * [new branch] gh/kwen2501/274/base -> origin/gh/kwen2501/274/base 2025-11-03T16:58:00.9094880Z * [new branch] gh/kwen2501/274/head -> origin/gh/kwen2501/274/head 2025-11-03T16:58:00.9096586Z * [new branch] gh/kwen2501/274/orig -> origin/gh/kwen2501/274/orig 2025-11-03T16:58:00.9099110Z * [new branch] gh/kwen2501/275/base -> origin/gh/kwen2501/275/base 2025-11-03T16:58:00.9100960Z * [new branch] gh/kwen2501/275/head -> origin/gh/kwen2501/275/head 2025-11-03T16:58:00.9102708Z * [new branch] gh/kwen2501/275/orig -> origin/gh/kwen2501/275/orig 2025-11-03T16:58:00.9105073Z * [new branch] gh/kwen2501/276/base -> origin/gh/kwen2501/276/base 2025-11-03T16:58:00.9106776Z * [new branch] gh/kwen2501/276/head -> origin/gh/kwen2501/276/head 2025-11-03T16:58:00.9108764Z * [new branch] gh/kwen2501/276/orig -> origin/gh/kwen2501/276/orig 2025-11-03T16:58:00.9111852Z * [new branch] gh/kwen2501/277/base -> origin/gh/kwen2501/277/base 2025-11-03T16:58:00.9113030Z * [new branch] gh/kwen2501/277/head -> origin/gh/kwen2501/277/head 2025-11-03T16:58:00.9114893Z * [new branch] gh/kwen2501/277/orig -> origin/gh/kwen2501/277/orig 2025-11-03T16:58:00.9117273Z * [new branch] gh/kwen2501/278/base -> origin/gh/kwen2501/278/base 2025-11-03T16:58:00.9119517Z * [new branch] gh/kwen2501/278/head -> origin/gh/kwen2501/278/head 2025-11-03T16:58:00.9121440Z * [new branch] gh/kwen2501/278/orig -> origin/gh/kwen2501/278/orig 2025-11-03T16:58:00.9123880Z * [new branch] gh/kwen2501/279/base -> origin/gh/kwen2501/279/base 2025-11-03T16:58:00.9125601Z * [new branch] gh/kwen2501/279/head -> origin/gh/kwen2501/279/head 2025-11-03T16:58:00.9127377Z * [new branch] gh/kwen2501/279/orig -> origin/gh/kwen2501/279/orig 2025-11-03T16:58:00.9131014Z * [new branch] gh/laithsakka/251/base -> origin/gh/laithsakka/251/base 2025-11-03T16:58:00.9132158Z * [new branch] gh/laithsakka/251/head -> origin/gh/laithsakka/251/head 2025-11-03T16:58:00.9133810Z * [new branch] gh/laithsakka/251/orig -> origin/gh/laithsakka/251/orig 2025-11-03T16:58:00.9136657Z * [new branch] gh/laithsakka/276/base -> origin/gh/laithsakka/276/base 2025-11-03T16:58:00.9137785Z * [new branch] gh/laithsakka/276/head -> origin/gh/laithsakka/276/head 2025-11-03T16:58:00.9139477Z * [new branch] gh/laithsakka/276/orig -> origin/gh/laithsakka/276/orig 2025-11-03T16:58:00.9141928Z * [new branch] gh/laithsakka/28/base -> origin/gh/laithsakka/28/base 2025-11-03T16:58:00.9144837Z * [new branch] gh/laithsakka/285/base -> origin/gh/laithsakka/285/base 2025-11-03T16:58:00.9146752Z * [new branch] gh/laithsakka/285/head -> origin/gh/laithsakka/285/head 2025-11-03T16:58:00.9148473Z * [new branch] gh/laithsakka/285/orig -> origin/gh/laithsakka/285/orig 2025-11-03T16:58:00.9151172Z * [new branch] gh/laithsakka/287/base -> origin/gh/laithsakka/287/base 2025-11-03T16:58:00.9152939Z * [new branch] gh/laithsakka/287/head -> origin/gh/laithsakka/287/head 2025-11-03T16:58:00.9154675Z * [new branch] gh/laithsakka/287/orig -> origin/gh/laithsakka/287/orig 2025-11-03T16:58:00.9157129Z * [new branch] gh/laithsakka/289/base -> origin/gh/laithsakka/289/base 2025-11-03T16:58:00.9158988Z * [new branch] gh/laithsakka/289/head -> origin/gh/laithsakka/289/head 2025-11-03T16:58:00.9160916Z * [new branch] gh/laithsakka/289/orig -> origin/gh/laithsakka/289/orig 2025-11-03T16:58:00.9163138Z * [new branch] gh/laithsakka/29/base -> origin/gh/laithsakka/29/base 2025-11-03T16:58:00.9165682Z * [new branch] gh/laithsakka/292/base -> origin/gh/laithsakka/292/base 2025-11-03T16:58:00.9168009Z * [new branch] gh/laithsakka/292/head -> origin/gh/laithsakka/292/head 2025-11-03T16:58:00.9169508Z * [new branch] gh/laithsakka/292/orig -> origin/gh/laithsakka/292/orig 2025-11-03T16:58:00.9172224Z * [new branch] gh/laithsakka/293/base -> origin/gh/laithsakka/293/base 2025-11-03T16:58:00.9173905Z * [new branch] gh/laithsakka/293/head -> origin/gh/laithsakka/293/head 2025-11-03T16:58:00.9175234Z * [new branch] gh/laithsakka/293/orig -> origin/gh/laithsakka/293/orig 2025-11-03T16:58:00.9178074Z * [new branch] gh/laithsakka/294/base -> origin/gh/laithsakka/294/base 2025-11-03T16:58:00.9179698Z * [new branch] gh/laithsakka/294/head -> origin/gh/laithsakka/294/head 2025-11-03T16:58:00.9181349Z * [new branch] gh/laithsakka/294/orig -> origin/gh/laithsakka/294/orig 2025-11-03T16:58:00.9183790Z * [new branch] gh/laithsakka/295/base -> origin/gh/laithsakka/295/base 2025-11-03T16:58:00.9185446Z * [new branch] gh/laithsakka/295/head -> origin/gh/laithsakka/295/head 2025-11-03T16:58:00.9187951Z * [new branch] gh/laithsakka/295/orig -> origin/gh/laithsakka/295/orig 2025-11-03T16:58:00.9189795Z * [new branch] gh/laithsakka/296/base -> origin/gh/laithsakka/296/base 2025-11-03T16:58:00.9191607Z * [new branch] gh/laithsakka/296/head -> origin/gh/laithsakka/296/head 2025-11-03T16:58:00.9193335Z * [new branch] gh/laithsakka/296/orig -> origin/gh/laithsakka/296/orig 2025-11-03T16:58:00.9196313Z * [new branch] gh/laithsakka/297/base -> origin/gh/laithsakka/297/base 2025-11-03T16:58:00.9197509Z * [new branch] gh/laithsakka/297/head -> origin/gh/laithsakka/297/head 2025-11-03T16:58:00.9199069Z * [new branch] gh/laithsakka/297/orig -> origin/gh/laithsakka/297/orig 2025-11-03T16:58:00.9201517Z * [new branch] gh/laithsakka/298/base -> origin/gh/laithsakka/298/base 2025-11-03T16:58:00.9203309Z * [new branch] gh/laithsakka/298/head -> origin/gh/laithsakka/298/head 2025-11-03T16:58:00.9204994Z * [new branch] gh/laithsakka/298/orig -> origin/gh/laithsakka/298/orig 2025-11-03T16:58:00.9207343Z * [new branch] gh/laithsakka/299/base -> origin/gh/laithsakka/299/base 2025-11-03T16:58:00.9209135Z * [new branch] gh/laithsakka/299/head -> origin/gh/laithsakka/299/head 2025-11-03T16:58:00.9211325Z * [new branch] gh/laithsakka/299/orig -> origin/gh/laithsakka/299/orig 2025-11-03T16:58:00.9213462Z * [new branch] gh/laithsakka/30/base -> origin/gh/laithsakka/30/base 2025-11-03T16:58:00.9215224Z * [new branch] gh/laithsakka/30/head -> origin/gh/laithsakka/30/head 2025-11-03T16:58:00.9217755Z * [new branch] gh/laithsakka/300/base -> origin/gh/laithsakka/300/base 2025-11-03T16:58:00.9219435Z * [new branch] gh/laithsakka/300/head -> origin/gh/laithsakka/300/head 2025-11-03T16:58:00.9221155Z * [new branch] gh/laithsakka/300/orig -> origin/gh/laithsakka/300/orig 2025-11-03T16:58:00.9224017Z * [new branch] gh/laithsakka/301/base -> origin/gh/laithsakka/301/base 2025-11-03T16:58:00.9225679Z * [new branch] gh/laithsakka/301/head -> origin/gh/laithsakka/301/head 2025-11-03T16:58:00.9227445Z * [new branch] gh/laithsakka/301/orig -> origin/gh/laithsakka/301/orig 2025-11-03T16:58:00.9229880Z * [new branch] gh/laithsakka/302/base -> origin/gh/laithsakka/302/base 2025-11-03T16:58:00.9231655Z * [new branch] gh/laithsakka/302/head -> origin/gh/laithsakka/302/head 2025-11-03T16:58:00.9233303Z * [new branch] gh/laithsakka/302/orig -> origin/gh/laithsakka/302/orig 2025-11-03T16:58:00.9235551Z * [new branch] gh/laithsakka/303/base -> origin/gh/laithsakka/303/base 2025-11-03T16:58:00.9237864Z * [new branch] gh/laithsakka/303/head -> origin/gh/laithsakka/303/head 2025-11-03T16:58:00.9239538Z * [new branch] gh/laithsakka/303/orig -> origin/gh/laithsakka/303/orig 2025-11-03T16:58:00.9242050Z * [new branch] gh/laithsakka/304/base -> origin/gh/laithsakka/304/base 2025-11-03T16:58:00.9243717Z * [new branch] gh/laithsakka/304/head -> origin/gh/laithsakka/304/head 2025-11-03T16:58:00.9245353Z * [new branch] gh/laithsakka/304/orig -> origin/gh/laithsakka/304/orig 2025-11-03T16:58:00.9247949Z * [new branch] gh/laithsakka/305/base -> origin/gh/laithsakka/305/base 2025-11-03T16:58:00.9249482Z * [new branch] gh/laithsakka/305/head -> origin/gh/laithsakka/305/head 2025-11-03T16:58:00.9251489Z * [new branch] gh/laithsakka/305/orig -> origin/gh/laithsakka/305/orig 2025-11-03T16:58:00.9253784Z * [new branch] gh/laithsakka/306/base -> origin/gh/laithsakka/306/base 2025-11-03T16:58:00.9255319Z * [new branch] gh/laithsakka/306/head -> origin/gh/laithsakka/306/head 2025-11-03T16:58:00.9257361Z * [new branch] gh/laithsakka/306/orig -> origin/gh/laithsakka/306/orig 2025-11-03T16:58:00.9259307Z * [new branch] gh/laithsakka/307/base -> origin/gh/laithsakka/307/base 2025-11-03T16:58:00.9261131Z * [new branch] gh/laithsakka/307/head -> origin/gh/laithsakka/307/head 2025-11-03T16:58:00.9262845Z * [new branch] gh/laithsakka/307/orig -> origin/gh/laithsakka/307/orig 2025-11-03T16:58:00.9265213Z * [new branch] gh/laithsakka/308/base -> origin/gh/laithsakka/308/base 2025-11-03T16:58:00.9266953Z * [new branch] gh/laithsakka/308/head -> origin/gh/laithsakka/308/head 2025-11-03T16:58:00.9268729Z * [new branch] gh/laithsakka/308/orig -> origin/gh/laithsakka/308/orig 2025-11-03T16:58:00.9271071Z * [new branch] gh/laithsakka/309/base -> origin/gh/laithsakka/309/base 2025-11-03T16:58:00.9272877Z * [new branch] gh/laithsakka/309/head -> origin/gh/laithsakka/309/head 2025-11-03T16:58:00.9274671Z * [new branch] gh/laithsakka/309/orig -> origin/gh/laithsakka/309/orig 2025-11-03T16:58:00.9277391Z * [new branch] gh/laithsakka/31/base -> origin/gh/laithsakka/31/base 2025-11-03T16:58:00.9279049Z * [new branch] gh/laithsakka/31/head -> origin/gh/laithsakka/31/head 2025-11-03T16:58:00.9281607Z * [new branch] gh/laithsakka/310/base -> origin/gh/laithsakka/310/base 2025-11-03T16:58:00.9283279Z * [new branch] gh/laithsakka/310/head -> origin/gh/laithsakka/310/head 2025-11-03T16:58:00.9284979Z * [new branch] gh/laithsakka/310/orig -> origin/gh/laithsakka/310/orig 2025-11-03T16:58:00.9287627Z * [new branch] gh/laithsakka/311/base -> origin/gh/laithsakka/311/base 2025-11-03T16:58:00.9289512Z * [new branch] gh/laithsakka/311/head -> origin/gh/laithsakka/311/head 2025-11-03T16:58:00.9291175Z * [new branch] gh/laithsakka/311/orig -> origin/gh/laithsakka/311/orig 2025-11-03T16:58:00.9293592Z * [new branch] gh/laithsakka/312/base -> origin/gh/laithsakka/312/base 2025-11-03T16:58:00.9295423Z * [new branch] gh/laithsakka/312/head -> origin/gh/laithsakka/312/head 2025-11-03T16:58:00.9297665Z * [new branch] gh/laithsakka/312/orig -> origin/gh/laithsakka/312/orig 2025-11-03T16:58:00.9299730Z * [new branch] gh/laithsakka/313/base -> origin/gh/laithsakka/313/base 2025-11-03T16:58:00.9301388Z * [new branch] gh/laithsakka/313/head -> origin/gh/laithsakka/313/head 2025-11-03T16:58:00.9303155Z * [new branch] gh/laithsakka/313/orig -> origin/gh/laithsakka/313/orig 2025-11-03T16:58:00.9306089Z * [new branch] gh/laithsakka/314/base -> origin/gh/laithsakka/314/base 2025-11-03T16:58:00.9307398Z * [new branch] gh/laithsakka/314/head -> origin/gh/laithsakka/314/head 2025-11-03T16:58:00.9309143Z * [new branch] gh/laithsakka/314/orig -> origin/gh/laithsakka/314/orig 2025-11-03T16:58:00.9313622Z * [new branch] gh/laithsakka/315/base -> origin/gh/laithsakka/315/base 2025-11-03T16:58:00.9315810Z * [new branch] gh/laithsakka/315/head -> origin/gh/laithsakka/315/head 2025-11-03T16:58:00.9317572Z * [new branch] gh/laithsakka/315/orig -> origin/gh/laithsakka/315/orig 2025-11-03T16:58:00.9320295Z * [new branch] gh/laithsakka/316/base -> origin/gh/laithsakka/316/base 2025-11-03T16:58:00.9322222Z * [new branch] gh/laithsakka/316/head -> origin/gh/laithsakka/316/head 2025-11-03T16:58:00.9324015Z * [new branch] gh/laithsakka/316/orig -> origin/gh/laithsakka/316/orig 2025-11-03T16:58:00.9326522Z * [new branch] gh/laithsakka/317/base -> origin/gh/laithsakka/317/base 2025-11-03T16:58:00.9328724Z * [new branch] gh/laithsakka/317/head -> origin/gh/laithsakka/317/head 2025-11-03T16:58:00.9329932Z * [new branch] gh/laithsakka/317/orig -> origin/gh/laithsakka/317/orig 2025-11-03T16:58:00.9332318Z * [new branch] gh/laithsakka/318/base -> origin/gh/laithsakka/318/base 2025-11-03T16:58:00.9334059Z * [new branch] gh/laithsakka/318/head -> origin/gh/laithsakka/318/head 2025-11-03T16:58:00.9335785Z * [new branch] gh/laithsakka/318/orig -> origin/gh/laithsakka/318/orig 2025-11-03T16:58:00.9338308Z * [new branch] gh/laithsakka/319/base -> origin/gh/laithsakka/319/base 2025-11-03T16:58:00.9340131Z * [new branch] gh/laithsakka/319/head -> origin/gh/laithsakka/319/head 2025-11-03T16:58:00.9341846Z * [new branch] gh/laithsakka/319/orig -> origin/gh/laithsakka/319/orig 2025-11-03T16:58:00.9344185Z * [new branch] gh/laithsakka/32/base -> origin/gh/laithsakka/32/base 2025-11-03T16:58:00.9346035Z * [new branch] gh/laithsakka/32/head -> origin/gh/laithsakka/32/head 2025-11-03T16:58:00.9348221Z * [new branch] gh/laithsakka/320/base -> origin/gh/laithsakka/320/base 2025-11-03T16:58:00.9350173Z * [new branch] gh/laithsakka/320/head -> origin/gh/laithsakka/320/head 2025-11-03T16:58:00.9351775Z * [new branch] gh/laithsakka/320/orig -> origin/gh/laithsakka/320/orig 2025-11-03T16:58:00.9354707Z * [new branch] gh/laithsakka/321/base -> origin/gh/laithsakka/321/base 2025-11-03T16:58:00.9355919Z * [new branch] gh/laithsakka/321/head -> origin/gh/laithsakka/321/head 2025-11-03T16:58:00.9357541Z * [new branch] gh/laithsakka/321/orig -> origin/gh/laithsakka/321/orig 2025-11-03T16:58:00.9360447Z * [new branch] gh/liangel-02/1/base -> origin/gh/liangel-02/1/base 2025-11-03T16:58:00.9362211Z * [new branch] gh/liangel-02/1/head -> origin/gh/liangel-02/1/head 2025-11-03T16:58:00.9363934Z * [new branch] gh/liangel-02/1/orig -> origin/gh/liangel-02/1/orig 2025-11-03T16:58:00.9366286Z * [new branch] gh/liangel-02/2/base -> origin/gh/liangel-02/2/base 2025-11-03T16:58:00.9368678Z * [new branch] gh/liangel-02/2/head -> origin/gh/liangel-02/2/head 2025-11-03T16:58:00.9370011Z * [new branch] gh/liangel-02/2/orig -> origin/gh/liangel-02/2/orig 2025-11-03T16:58:00.9373459Z * [new branch] gh/liangel/1/base -> origin/gh/liangel/1/base 2025-11-03T16:58:00.9374862Z * [new branch] gh/liangel/1/head -> origin/gh/liangel/1/head 2025-11-03T16:58:00.9376535Z * [new branch] gh/liangel/1/orig -> origin/gh/liangel/1/orig 2025-11-03T16:58:00.9378858Z * [new branch] gh/liangel/2/base -> origin/gh/liangel/2/base 2025-11-03T16:58:00.9380665Z * [new branch] gh/liangel/2/head -> origin/gh/liangel/2/head 2025-11-03T16:58:00.9382532Z * [new branch] gh/liangel/2/orig -> origin/gh/liangel/2/orig 2025-11-03T16:58:00.9385047Z * [new branch] gh/liangel/3/base -> origin/gh/liangel/3/base 2025-11-03T16:58:00.9386814Z * [new branch] gh/liangel/3/head -> origin/gh/liangel/3/head 2025-11-03T16:58:00.9388852Z * [new branch] gh/liangel/3/orig -> origin/gh/liangel/3/orig 2025-11-03T16:58:00.9391191Z * [new branch] gh/liangel/4/base -> origin/gh/liangel/4/base 2025-11-03T16:58:00.9393012Z * [new branch] gh/liangel/4/head -> origin/gh/liangel/4/head 2025-11-03T16:58:00.9395222Z * [new branch] gh/liangel/4/orig -> origin/gh/liangel/4/orig 2025-11-03T16:58:00.9399246Z * [new branch] gh/lucaskabela/1/base -> origin/gh/lucaskabela/1/base 2025-11-03T16:58:00.9401147Z * [new branch] gh/lucaskabela/1/head -> origin/gh/lucaskabela/1/head 2025-11-03T16:58:00.9403743Z * [new branch] gh/lucaskabela/10/base -> origin/gh/lucaskabela/10/base 2025-11-03T16:58:00.9405072Z * [new branch] gh/lucaskabela/10/head -> origin/gh/lucaskabela/10/head 2025-11-03T16:58:00.9406739Z * [new branch] gh/lucaskabela/10/orig -> origin/gh/lucaskabela/10/orig 2025-11-03T16:58:00.9410100Z * [new branch] gh/lucaskabela/2/base -> origin/gh/lucaskabela/2/base 2025-11-03T16:58:00.9411350Z * [new branch] gh/lucaskabela/2/head -> origin/gh/lucaskabela/2/head 2025-11-03T16:58:00.9413759Z * [new branch] gh/lucaskabela/3/base -> origin/gh/lucaskabela/3/base 2025-11-03T16:58:00.9415441Z * [new branch] gh/lucaskabela/3/head -> origin/gh/lucaskabela/3/head 2025-11-03T16:58:00.9417142Z * [new branch] gh/lucaskabela/3/orig -> origin/gh/lucaskabela/3/orig 2025-11-03T16:58:00.9419166Z * [new branch] gh/lucaskabela/4/base -> origin/gh/lucaskabela/4/base 2025-11-03T16:58:00.9421011Z * [new branch] gh/lucaskabela/4/head -> origin/gh/lucaskabela/4/head 2025-11-03T16:58:00.9423060Z * [new branch] gh/lucaskabela/4/orig -> origin/gh/lucaskabela/4/orig 2025-11-03T16:58:00.9425651Z * [new branch] gh/lucaskabela/5/base -> origin/gh/lucaskabela/5/base 2025-11-03T16:58:00.9427003Z * [new branch] gh/lucaskabela/5/head -> origin/gh/lucaskabela/5/head 2025-11-03T16:58:00.9428602Z * [new branch] gh/lucaskabela/5/orig -> origin/gh/lucaskabela/5/orig 2025-11-03T16:58:00.9430867Z * [new branch] gh/lucaskabela/6/base -> origin/gh/lucaskabela/6/base 2025-11-03T16:58:00.9432554Z * [new branch] gh/lucaskabela/6/head -> origin/gh/lucaskabela/6/head 2025-11-03T16:58:00.9434284Z * [new branch] gh/lucaskabela/6/orig -> origin/gh/lucaskabela/6/orig 2025-11-03T16:58:00.9436477Z * [new branch] gh/lucaskabela/7/base -> origin/gh/lucaskabela/7/base 2025-11-03T16:58:00.9438251Z * [new branch] gh/lucaskabela/7/head -> origin/gh/lucaskabela/7/head 2025-11-03T16:58:00.9439971Z * [new branch] gh/lucaskabela/7/orig -> origin/gh/lucaskabela/7/orig 2025-11-03T16:58:00.9442272Z * [new branch] gh/lucaskabela/8/base -> origin/gh/lucaskabela/8/base 2025-11-03T16:58:00.9443930Z * [new branch] gh/lucaskabela/8/head -> origin/gh/lucaskabela/8/head 2025-11-03T16:58:00.9445666Z * [new branch] gh/lucaskabela/8/orig -> origin/gh/lucaskabela/8/orig 2025-11-03T16:58:00.9447842Z * [new branch] gh/lucaskabela/9/base -> origin/gh/lucaskabela/9/base 2025-11-03T16:58:00.9449586Z * [new branch] gh/lucaskabela/9/head -> origin/gh/lucaskabela/9/head 2025-11-03T16:58:00.9451263Z * [new branch] gh/lucaskabela/9/orig -> origin/gh/lucaskabela/9/orig 2025-11-03T16:58:00.9454317Z * [new branch] gh/lw/10/base -> origin/gh/lw/10/base 2025-11-03T16:58:00.9456011Z * [new branch] gh/lw/10/head -> origin/gh/lw/10/head 2025-11-03T16:58:00.9457724Z * [new branch] gh/lw/10/orig -> origin/gh/lw/10/orig 2025-11-03T16:58:00.9460228Z * [new branch] gh/lw/3/base -> origin/gh/lw/3/base 2025-11-03T16:58:00.9461924Z * [new branch] gh/lw/3/head -> origin/gh/lw/3/head 2025-11-03T16:58:00.9463973Z * [new branch] gh/lw/3/orig -> origin/gh/lw/3/orig 2025-11-03T16:58:00.9466191Z * [new branch] gh/lw/4/base -> origin/gh/lw/4/base 2025-11-03T16:58:00.9467734Z * [new branch] gh/lw/4/head -> origin/gh/lw/4/head 2025-11-03T16:58:00.9469354Z * [new branch] gh/lw/4/orig -> origin/gh/lw/4/orig 2025-11-03T16:58:00.9471684Z * [new branch] gh/lw/5/base -> origin/gh/lw/5/base 2025-11-03T16:58:00.9473446Z * [new branch] gh/lw/5/head -> origin/gh/lw/5/head 2025-11-03T16:58:00.9475191Z * [new branch] gh/lw/5/orig -> origin/gh/lw/5/orig 2025-11-03T16:58:00.9477535Z * [new branch] gh/lw/6/base -> origin/gh/lw/6/base 2025-11-03T16:58:00.9479376Z * [new branch] gh/lw/6/head -> origin/gh/lw/6/head 2025-11-03T16:58:00.9480927Z * [new branch] gh/lw/6/orig -> origin/gh/lw/6/orig 2025-11-03T16:58:00.9483783Z * [new branch] gh/lw/7/base -> origin/gh/lw/7/base 2025-11-03T16:58:00.9484997Z * [new branch] gh/lw/7/head -> origin/gh/lw/7/head 2025-11-03T16:58:00.9486766Z * [new branch] gh/lw/7/orig -> origin/gh/lw/7/orig 2025-11-03T16:58:00.9489230Z * [new branch] gh/lw/8/base -> origin/gh/lw/8/base 2025-11-03T16:58:00.9490923Z * [new branch] gh/lw/8/head -> origin/gh/lw/8/head 2025-11-03T16:58:00.9492643Z * [new branch] gh/lw/8/orig -> origin/gh/lw/8/orig 2025-11-03T16:58:00.9495005Z * [new branch] gh/lw/9/base -> origin/gh/lw/9/base 2025-11-03T16:58:00.9496801Z * [new branch] gh/lw/9/head -> origin/gh/lw/9/head 2025-11-03T16:58:00.9498489Z * [new branch] gh/lw/9/orig -> origin/gh/lw/9/orig 2025-11-03T16:58:00.9501529Z * [new branch] gh/maggiemoss/1/base -> origin/gh/maggiemoss/1/base 2025-11-03T16:58:00.9503252Z * [new branch] gh/maggiemoss/1/head -> origin/gh/maggiemoss/1/head 2025-11-03T16:58:00.9505000Z * [new branch] gh/maggiemoss/1/orig -> origin/gh/maggiemoss/1/orig 2025-11-03T16:58:00.9507246Z * [new branch] gh/maggiemoss/2/base -> origin/gh/maggiemoss/2/base 2025-11-03T16:58:00.9508977Z * [new branch] gh/maggiemoss/2/head -> origin/gh/maggiemoss/2/head 2025-11-03T16:58:00.9510948Z * [new branch] gh/maggiemoss/2/orig -> origin/gh/maggiemoss/2/orig 2025-11-03T16:58:00.9513637Z * [new branch] gh/malfet/14/base -> origin/gh/malfet/14/base 2025-11-03T16:58:00.9515973Z * [new branch] gh/malfet/417/base -> origin/gh/malfet/417/base 2025-11-03T16:58:00.9518224Z * [new branch] gh/malfet/417/head -> origin/gh/malfet/417/head 2025-11-03T16:58:00.9519490Z * [new branch] gh/malfet/417/orig -> origin/gh/malfet/417/orig 2025-11-03T16:58:00.9522378Z * [new branch] gh/malfet/506/base -> origin/gh/malfet/506/base 2025-11-03T16:58:00.9523636Z * [new branch] gh/malfet/506/head -> origin/gh/malfet/506/head 2025-11-03T16:58:00.9525297Z * [new branch] gh/malfet/506/orig -> origin/gh/malfet/506/orig 2025-11-03T16:58:00.9527938Z * [new branch] gh/malfet/517/base -> origin/gh/malfet/517/base 2025-11-03T16:58:00.9529525Z * [new branch] gh/malfet/517/head -> origin/gh/malfet/517/head 2025-11-03T16:58:00.9532159Z * [new branch] gh/malfet/528/base -> origin/gh/malfet/528/base 2025-11-03T16:58:00.9533613Z * [new branch] gh/malfet/528/head -> origin/gh/malfet/528/head 2025-11-03T16:58:00.9535333Z * [new branch] gh/malfet/528/orig -> origin/gh/malfet/528/orig 2025-11-03T16:58:00.9538173Z * [new branch] gh/malfet/537/base -> origin/gh/malfet/537/base 2025-11-03T16:58:00.9539422Z * [new branch] gh/malfet/537/head -> origin/gh/malfet/537/head 2025-11-03T16:58:00.9541171Z * [new branch] gh/malfet/537/orig -> origin/gh/malfet/537/orig 2025-11-03T16:58:00.9544074Z * [new branch] gh/malfet/546/base -> origin/gh/malfet/546/base 2025-11-03T16:58:00.9545345Z * [new branch] gh/malfet/546/head -> origin/gh/malfet/546/head 2025-11-03T16:58:00.9547090Z * [new branch] gh/malfet/546/orig -> origin/gh/malfet/546/orig 2025-11-03T16:58:00.9549333Z * [new branch] gh/malfet/547/base -> origin/gh/malfet/547/base 2025-11-03T16:58:00.9550962Z * [new branch] gh/malfet/547/head -> origin/gh/malfet/547/head 2025-11-03T16:58:00.9552658Z * [new branch] gh/malfet/547/orig -> origin/gh/malfet/547/orig 2025-11-03T16:58:00.9554956Z * [new branch] gh/malfet/548/base -> origin/gh/malfet/548/base 2025-11-03T16:58:00.9556704Z * [new branch] gh/malfet/548/head -> origin/gh/malfet/548/head 2025-11-03T16:58:00.9558552Z * [new branch] gh/malfet/548/orig -> origin/gh/malfet/548/orig 2025-11-03T16:58:00.9560849Z * [new branch] gh/malfet/549/base -> origin/gh/malfet/549/base 2025-11-03T16:58:00.9562545Z * [new branch] gh/malfet/549/head -> origin/gh/malfet/549/head 2025-11-03T16:58:00.9564387Z * [new branch] gh/malfet/549/orig -> origin/gh/malfet/549/orig 2025-11-03T16:58:00.9566818Z * [new branch] gh/malfet/550/base -> origin/gh/malfet/550/base 2025-11-03T16:58:00.9568542Z * [new branch] gh/malfet/550/head -> origin/gh/malfet/550/head 2025-11-03T16:58:00.9570380Z * [new branch] gh/malfet/550/orig -> origin/gh/malfet/550/orig 2025-11-03T16:58:00.9572662Z * [new branch] gh/malfet/551/base -> origin/gh/malfet/551/base 2025-11-03T16:58:00.9574386Z * [new branch] gh/malfet/551/head -> origin/gh/malfet/551/head 2025-11-03T16:58:00.9576101Z * [new branch] gh/malfet/551/orig -> origin/gh/malfet/551/orig 2025-11-03T16:58:00.9578446Z * [new branch] gh/malfet/552/base -> origin/gh/malfet/552/base 2025-11-03T16:58:00.9580178Z * [new branch] gh/malfet/552/head -> origin/gh/malfet/552/head 2025-11-03T16:58:00.9581934Z * [new branch] gh/malfet/552/orig -> origin/gh/malfet/552/orig 2025-11-03T16:58:00.9584306Z * [new branch] gh/malfet/553/base -> origin/gh/malfet/553/base 2025-11-03T16:58:00.9586088Z * [new branch] gh/malfet/553/head -> origin/gh/malfet/553/head 2025-11-03T16:58:00.9587832Z * [new branch] gh/malfet/553/orig -> origin/gh/malfet/553/orig 2025-11-03T16:58:00.9590219Z * [new branch] gh/malfet/554/base -> origin/gh/malfet/554/base 2025-11-03T16:58:00.9591920Z * [new branch] gh/malfet/554/head -> origin/gh/malfet/554/head 2025-11-03T16:58:00.9593639Z * [new branch] gh/malfet/554/orig -> origin/gh/malfet/554/orig 2025-11-03T16:58:00.9596235Z * [new branch] gh/malfet/555/base -> origin/gh/malfet/555/base 2025-11-03T16:58:00.9597838Z * [new branch] gh/malfet/555/head -> origin/gh/malfet/555/head 2025-11-03T16:58:00.9599564Z * [new branch] gh/malfet/555/orig -> origin/gh/malfet/555/orig 2025-11-03T16:58:00.9601986Z * [new branch] gh/malfet/556/base -> origin/gh/malfet/556/base 2025-11-03T16:58:00.9603561Z * [new branch] gh/malfet/556/head -> origin/gh/malfet/556/head 2025-11-03T16:58:00.9605316Z * [new branch] gh/malfet/556/orig -> origin/gh/malfet/556/orig 2025-11-03T16:58:00.9608864Z * [new branch] gh/malfet/557/base -> origin/gh/malfet/557/base 2025-11-03T16:58:00.9610232Z * [new branch] gh/malfet/557/head -> origin/gh/malfet/557/head 2025-11-03T16:58:00.9612148Z * [new branch] gh/malfet/557/orig -> origin/gh/malfet/557/orig 2025-11-03T16:58:00.9615075Z * [new branch] gh/malfet/558/base -> origin/gh/malfet/558/base 2025-11-03T16:58:00.9616457Z * [new branch] gh/malfet/558/head -> origin/gh/malfet/558/head 2025-11-03T16:58:00.9617868Z * [new branch] gh/malfet/558/orig -> origin/gh/malfet/558/orig 2025-11-03T16:58:00.9620516Z * [new branch] gh/malfet/559/base -> origin/gh/malfet/559/base 2025-11-03T16:58:00.9622150Z * [new branch] gh/malfet/559/head -> origin/gh/malfet/559/head 2025-11-03T16:58:00.9623852Z * [new branch] gh/malfet/559/orig -> origin/gh/malfet/559/orig 2025-11-03T16:58:00.9626234Z * [new branch] gh/malfet/560/base -> origin/gh/malfet/560/base 2025-11-03T16:58:00.9628525Z * [new branch] gh/malfet/560/head -> origin/gh/malfet/560/head 2025-11-03T16:58:00.9629893Z * [new branch] gh/malfet/560/orig -> origin/gh/malfet/560/orig 2025-11-03T16:58:00.9632763Z * [new branch] gh/malfet/561/base -> origin/gh/malfet/561/base 2025-11-03T16:58:00.9634427Z * [new branch] gh/malfet/561/head -> origin/gh/malfet/561/head 2025-11-03T16:58:00.9636210Z * [new branch] gh/malfet/561/orig -> origin/gh/malfet/561/orig 2025-11-03T16:58:00.9638626Z * [new branch] gh/malfet/562/base -> origin/gh/malfet/562/base 2025-11-03T16:58:00.9640376Z * [new branch] gh/malfet/562/head -> origin/gh/malfet/562/head 2025-11-03T16:58:00.9642080Z * [new branch] gh/malfet/562/orig -> origin/gh/malfet/562/orig 2025-11-03T16:58:00.9644444Z * [new branch] gh/malfet/563/base -> origin/gh/malfet/563/base 2025-11-03T16:58:00.9646122Z * [new branch] gh/malfet/563/head -> origin/gh/malfet/563/head 2025-11-03T16:58:00.9647903Z * [new branch] gh/malfet/563/orig -> origin/gh/malfet/563/orig 2025-11-03T16:58:00.9650399Z * [new branch] gh/malfet/564/base -> origin/gh/malfet/564/base 2025-11-03T16:58:00.9652214Z * [new branch] gh/malfet/564/head -> origin/gh/malfet/564/head 2025-11-03T16:58:00.9653943Z * [new branch] gh/malfet/564/orig -> origin/gh/malfet/564/orig 2025-11-03T16:58:00.9656282Z * [new branch] gh/malfet/565/base -> origin/gh/malfet/565/base 2025-11-03T16:58:00.9657952Z * [new branch] gh/malfet/565/head -> origin/gh/malfet/565/head 2025-11-03T16:58:00.9659715Z * [new branch] gh/malfet/565/orig -> origin/gh/malfet/565/orig 2025-11-03T16:58:00.9662118Z * [new branch] gh/malfet/566/base -> origin/gh/malfet/566/base 2025-11-03T16:58:00.9663814Z * [new branch] gh/malfet/566/head -> origin/gh/malfet/566/head 2025-11-03T16:58:00.9665704Z * [new branch] gh/malfet/566/orig -> origin/gh/malfet/566/orig 2025-11-03T16:58:00.9669014Z * [new branch] gh/malfet/567/base -> origin/gh/malfet/567/base 2025-11-03T16:58:00.9669974Z * [new branch] gh/malfet/567/head -> origin/gh/malfet/567/head 2025-11-03T16:58:00.9671630Z * [new branch] gh/malfet/567/orig -> origin/gh/malfet/567/orig 2025-11-03T16:58:00.9674502Z * [new branch] gh/malfet/568/base -> origin/gh/malfet/568/base 2025-11-03T16:58:00.9675857Z * [new branch] gh/malfet/568/head -> origin/gh/malfet/568/head 2025-11-03T16:58:00.9677450Z * [new branch] gh/malfet/568/orig -> origin/gh/malfet/568/orig 2025-11-03T16:58:00.9680253Z * [new branch] gh/malfet/569/base -> origin/gh/malfet/569/base 2025-11-03T16:58:00.9682021Z * [new branch] gh/malfet/569/head -> origin/gh/malfet/569/head 2025-11-03T16:58:00.9683798Z * [new branch] gh/malfet/569/orig -> origin/gh/malfet/569/orig 2025-11-03T16:58:00.9686197Z * [new branch] gh/malfet/570/base -> origin/gh/malfet/570/base 2025-11-03T16:58:00.9688527Z * [new branch] gh/malfet/570/head -> origin/gh/malfet/570/head 2025-11-03T16:58:00.9689982Z * [new branch] gh/malfet/570/orig -> origin/gh/malfet/570/orig 2025-11-03T16:58:00.9692915Z * [new branch] gh/malfet/571/base -> origin/gh/malfet/571/base 2025-11-03T16:58:00.9694105Z * [new branch] gh/malfet/571/head -> origin/gh/malfet/571/head 2025-11-03T16:58:00.9695809Z * [new branch] gh/malfet/571/orig -> origin/gh/malfet/571/orig 2025-11-03T16:58:00.9698208Z * [new branch] gh/malfet/572/base -> origin/gh/malfet/572/base 2025-11-03T16:58:00.9699908Z * [new branch] gh/malfet/572/head -> origin/gh/malfet/572/head 2025-11-03T16:58:00.9701688Z * [new branch] gh/malfet/572/orig -> origin/gh/malfet/572/orig 2025-11-03T16:58:00.9704070Z * [new branch] gh/malfet/573/base -> origin/gh/malfet/573/base 2025-11-03T16:58:00.9705821Z * [new branch] gh/malfet/573/head -> origin/gh/malfet/573/head 2025-11-03T16:58:00.9707998Z * [new branch] gh/malfet/573/orig -> origin/gh/malfet/573/orig 2025-11-03T16:58:00.9711644Z * [new branch] gh/malfet/574/base -> origin/gh/malfet/574/base 2025-11-03T16:58:00.9713459Z * [new branch] gh/malfet/574/head -> origin/gh/malfet/574/head 2025-11-03T16:58:00.9715306Z * [new branch] gh/malfet/574/orig -> origin/gh/malfet/574/orig 2025-11-03T16:58:00.9717462Z * [new branch] gh/malfet/575/base -> origin/gh/malfet/575/base 2025-11-03T16:58:00.9719202Z * [new branch] gh/malfet/575/head -> origin/gh/malfet/575/head 2025-11-03T16:58:00.9720760Z * [new branch] gh/malfet/575/orig -> origin/gh/malfet/575/orig 2025-11-03T16:58:00.9723651Z * [new branch] gh/malfet/576/base -> origin/gh/malfet/576/base 2025-11-03T16:58:00.9724936Z * [new branch] gh/malfet/576/head -> origin/gh/malfet/576/head 2025-11-03T16:58:00.9726744Z * [new branch] gh/malfet/576/orig -> origin/gh/malfet/576/orig 2025-11-03T16:58:00.9729199Z * [new branch] gh/malfet/577/base -> origin/gh/malfet/577/base 2025-11-03T16:58:00.9730965Z * [new branch] gh/malfet/577/head -> origin/gh/malfet/577/head 2025-11-03T16:58:00.9732824Z * [new branch] gh/malfet/577/orig -> origin/gh/malfet/577/orig 2025-11-03T16:58:00.9735433Z * [new branch] gh/malfet/578/base -> origin/gh/malfet/578/base 2025-11-03T16:58:00.9737677Z * [new branch] gh/malfet/578/head -> origin/gh/malfet/578/head 2025-11-03T16:58:00.9738954Z * [new branch] gh/malfet/578/orig -> origin/gh/malfet/578/orig 2025-11-03T16:58:00.9742189Z * [new branch] gh/malfet/579/base -> origin/gh/malfet/579/base 2025-11-03T16:58:00.9743182Z * [new branch] gh/malfet/579/head -> origin/gh/malfet/579/head 2025-11-03T16:58:00.9744827Z * [new branch] gh/malfet/579/orig -> origin/gh/malfet/579/orig 2025-11-03T16:58:00.9747574Z * [new branch] gh/malfet/580/base -> origin/gh/malfet/580/base 2025-11-03T16:58:00.9748820Z * [new branch] gh/malfet/580/head -> origin/gh/malfet/580/head 2025-11-03T16:58:00.9750498Z * [new branch] gh/malfet/580/orig -> origin/gh/malfet/580/orig 2025-11-03T16:58:00.9752903Z * [new branch] gh/malfet/581/base -> origin/gh/malfet/581/base 2025-11-03T16:58:00.9755108Z * [new branch] gh/malfet/581/head -> origin/gh/malfet/581/head 2025-11-03T16:58:00.9756852Z * [new branch] gh/malfet/581/orig -> origin/gh/malfet/581/orig 2025-11-03T16:58:00.9759600Z * [new branch] gh/malfet/582/base -> origin/gh/malfet/582/base 2025-11-03T16:58:00.9761254Z * [new branch] gh/malfet/582/head -> origin/gh/malfet/582/head 2025-11-03T16:58:00.9763163Z * [new branch] gh/malfet/582/orig -> origin/gh/malfet/582/orig 2025-11-03T16:58:00.9765400Z * [new branch] gh/malfet/583/base -> origin/gh/malfet/583/base 2025-11-03T16:58:00.9767091Z * [new branch] gh/malfet/583/head -> origin/gh/malfet/583/head 2025-11-03T16:58:00.9768901Z * [new branch] gh/malfet/583/orig -> origin/gh/malfet/583/orig 2025-11-03T16:58:00.9771612Z * [new branch] gh/malfet/584/base -> origin/gh/malfet/584/base 2025-11-03T16:58:00.9772827Z * [new branch] gh/malfet/584/head -> origin/gh/malfet/584/head 2025-11-03T16:58:00.9774589Z * [new branch] gh/malfet/584/orig -> origin/gh/malfet/584/orig 2025-11-03T16:58:00.9777438Z * [new branch] gh/malfet/585/base -> origin/gh/malfet/585/base 2025-11-03T16:58:00.9778707Z * [new branch] gh/malfet/585/head -> origin/gh/malfet/585/head 2025-11-03T16:58:00.9780420Z * [new branch] gh/malfet/585/orig -> origin/gh/malfet/585/orig 2025-11-03T16:58:00.9782854Z * [new branch] gh/malfet/586/base -> origin/gh/malfet/586/base 2025-11-03T16:58:00.9784543Z * [new branch] gh/malfet/586/head -> origin/gh/malfet/586/head 2025-11-03T16:58:00.9786288Z * [new branch] gh/malfet/586/orig -> origin/gh/malfet/586/orig 2025-11-03T16:58:00.9788668Z * [new branch] gh/malfet/587/base -> origin/gh/malfet/587/base 2025-11-03T16:58:00.9790474Z * [new branch] gh/malfet/587/head -> origin/gh/malfet/587/head 2025-11-03T16:58:00.9792197Z * [new branch] gh/malfet/587/orig -> origin/gh/malfet/587/orig 2025-11-03T16:58:00.9794674Z * [new branch] gh/malfet/64/base -> origin/gh/malfet/64/base 2025-11-03T16:58:00.9796380Z * [new branch] gh/malfet/64/head -> origin/gh/malfet/64/head 2025-11-03T16:58:00.9799265Z * [new branch] gh/manuelcandales/11/base -> origin/gh/manuelcandales/11/base 2025-11-03T16:58:00.9801049Z * [new branch] gh/manuelcandales/11/head -> origin/gh/manuelcandales/11/head 2025-11-03T16:58:00.9803013Z * [new branch] gh/manuelcandales/11/orig -> origin/gh/manuelcandales/11/orig 2025-11-03T16:58:00.9806017Z * [new branch] gh/markkm/1/base -> origin/gh/markkm/1/base 2025-11-03T16:58:00.9809865Z * [new branch] gh/masnesral/235/base -> origin/gh/masnesral/235/base 2025-11-03T16:58:00.9811170Z * [new branch] gh/masnesral/235/head -> origin/gh/masnesral/235/head 2025-11-03T16:58:00.9813163Z * [new branch] gh/masnesral/235/orig -> origin/gh/masnesral/235/orig 2025-11-03T16:58:00.9815154Z * [new branch] gh/masnesral/236/base -> origin/gh/masnesral/236/base 2025-11-03T16:58:00.9816819Z * [new branch] gh/masnesral/236/head -> origin/gh/masnesral/236/head 2025-11-03T16:58:00.9818473Z * [new branch] gh/masnesral/236/orig -> origin/gh/masnesral/236/orig 2025-11-03T16:58:00.9820864Z * [new branch] gh/masnesral/237/base -> origin/gh/masnesral/237/base 2025-11-03T16:58:00.9822729Z * [new branch] gh/masnesral/237/head -> origin/gh/masnesral/237/head 2025-11-03T16:58:00.9824621Z * [new branch] gh/masnesral/237/orig -> origin/gh/masnesral/237/orig 2025-11-03T16:58:00.9827066Z * [new branch] gh/masnesral/238/base -> origin/gh/masnesral/238/base 2025-11-03T16:58:00.9828818Z * [new branch] gh/masnesral/238/head -> origin/gh/masnesral/238/head 2025-11-03T16:58:00.9830590Z * [new branch] gh/masnesral/238/orig -> origin/gh/masnesral/238/orig 2025-11-03T16:58:00.9833498Z * [new branch] gh/mhorowitz/0/base -> origin/gh/mhorowitz/0/base 2025-11-03T16:58:00.9835178Z * [new branch] gh/mhorowitz/0/head -> origin/gh/mhorowitz/0/head 2025-11-03T16:58:00.9837361Z * [new branch] gh/mhorowitz/1/base -> origin/gh/mhorowitz/1/base 2025-11-03T16:58:00.9839098Z * [new branch] gh/mhorowitz/1/head -> origin/gh/mhorowitz/1/head 2025-11-03T16:58:00.9841260Z * [new branch] gh/mhorowitz/2/base -> origin/gh/mhorowitz/2/base 2025-11-03T16:58:00.9842977Z * [new branch] gh/mhorowitz/2/head -> origin/gh/mhorowitz/2/head 2025-11-03T16:58:00.9845668Z * [new branch] gh/mhorowitz/3/base -> origin/gh/mhorowitz/3/base 2025-11-03T16:58:00.9847495Z * [new branch] gh/mhorowitz/3/head -> origin/gh/mhorowitz/3/head 2025-11-03T16:58:00.9850114Z * [new branch] gh/mhorowitz/4/base -> origin/gh/mhorowitz/4/base 2025-11-03T16:58:00.9851949Z * [new branch] gh/mhorowitz/4/head -> origin/gh/mhorowitz/4/head 2025-11-03T16:58:00.9854135Z * [new branch] gh/mhorowitz/5/base -> origin/gh/mhorowitz/5/base 2025-11-03T16:58:00.9855757Z * [new branch] gh/mhorowitz/5/head -> origin/gh/mhorowitz/5/head 2025-11-03T16:58:00.9857642Z * [new branch] gh/mhorowitz/6/base -> origin/gh/mhorowitz/6/base 2025-11-03T16:58:00.9859520Z * [new branch] gh/mhorowitz/6/head -> origin/gh/mhorowitz/6/head 2025-11-03T16:58:00.9862784Z * [new branch] gh/mikaylagawarecki/234/base -> origin/gh/mikaylagawarecki/234/base 2025-11-03T16:58:00.9864051Z * [new branch] gh/mikaylagawarecki/234/head -> origin/gh/mikaylagawarecki/234/head 2025-11-03T16:58:00.9866597Z * [new branch] gh/mikaylagawarecki/235/base -> origin/gh/mikaylagawarecki/235/base 2025-11-03T16:58:00.9867757Z * [new branch] gh/mikaylagawarecki/235/head -> origin/gh/mikaylagawarecki/235/head 2025-11-03T16:58:00.9870280Z * [new branch] gh/mikaylagawarecki/236/base -> origin/gh/mikaylagawarecki/236/base 2025-11-03T16:58:00.9871984Z * [new branch] gh/mikaylagawarecki/236/head -> origin/gh/mikaylagawarecki/236/head 2025-11-03T16:58:00.9874262Z * [new branch] gh/mikaylagawarecki/237/base -> origin/gh/mikaylagawarecki/237/base 2025-11-03T16:58:00.9876069Z * [new branch] gh/mikaylagawarecki/237/head -> origin/gh/mikaylagawarecki/237/head 2025-11-03T16:58:00.9878419Z * [new branch] gh/mikaylagawarecki/238/base -> origin/gh/mikaylagawarecki/238/base 2025-11-03T16:58:00.9880127Z * [new branch] gh/mikaylagawarecki/238/head -> origin/gh/mikaylagawarecki/238/head 2025-11-03T16:58:00.9882572Z * [new branch] gh/mikaylagawarecki/336/base -> origin/gh/mikaylagawarecki/336/base 2025-11-03T16:58:00.9884275Z * [new branch] gh/mikaylagawarecki/336/head -> origin/gh/mikaylagawarecki/336/head 2025-11-03T16:58:00.9885911Z * [new branch] gh/mikaylagawarecki/336/orig -> origin/gh/mikaylagawarecki/336/orig 2025-11-03T16:58:00.9888759Z * [new branch] gh/mikaylagawarecki/337/base -> origin/gh/mikaylagawarecki/337/base 2025-11-03T16:58:00.9890640Z * [new branch] gh/mikaylagawarecki/337/head -> origin/gh/mikaylagawarecki/337/head 2025-11-03T16:58:00.9892525Z * [new branch] gh/mikaylagawarecki/337/orig -> origin/gh/mikaylagawarecki/337/orig 2025-11-03T16:58:00.9895045Z * [new branch] gh/mikaylagawarecki/341/base -> origin/gh/mikaylagawarecki/341/base 2025-11-03T16:58:00.9896709Z * [new branch] gh/mikaylagawarecki/341/head -> origin/gh/mikaylagawarecki/341/head 2025-11-03T16:58:00.9898420Z * [new branch] gh/mikaylagawarecki/341/orig -> origin/gh/mikaylagawarecki/341/orig 2025-11-03T16:58:00.9901128Z * [new branch] gh/mikaylagawarecki/342/base -> origin/gh/mikaylagawarecki/342/base 2025-11-03T16:58:00.9902739Z * [new branch] gh/mikaylagawarecki/342/head -> origin/gh/mikaylagawarecki/342/head 2025-11-03T16:58:00.9904434Z * [new branch] gh/mikaylagawarecki/342/orig -> origin/gh/mikaylagawarecki/342/orig 2025-11-03T16:58:00.9906772Z * [new branch] gh/mikaylagawarecki/343/base -> origin/gh/mikaylagawarecki/343/base 2025-11-03T16:58:00.9908499Z * [new branch] gh/mikaylagawarecki/343/head -> origin/gh/mikaylagawarecki/343/head 2025-11-03T16:58:00.9910284Z * [new branch] gh/mikaylagawarecki/343/orig -> origin/gh/mikaylagawarecki/343/orig 2025-11-03T16:58:00.9912829Z * [new branch] gh/mikaylagawarecki/344/base -> origin/gh/mikaylagawarecki/344/base 2025-11-03T16:58:00.9914517Z * [new branch] gh/mikaylagawarecki/344/head -> origin/gh/mikaylagawarecki/344/head 2025-11-03T16:58:00.9916711Z * [new branch] gh/mikaylagawarecki/344/orig -> origin/gh/mikaylagawarecki/344/orig 2025-11-03T16:58:00.9919268Z * [new branch] gh/mikaylagawarecki/345/base -> origin/gh/mikaylagawarecki/345/base 2025-11-03T16:58:00.9920789Z * [new branch] gh/mikaylagawarecki/345/head -> origin/gh/mikaylagawarecki/345/head 2025-11-03T16:58:00.9922568Z * [new branch] gh/mikaylagawarecki/345/orig -> origin/gh/mikaylagawarecki/345/orig 2025-11-03T16:58:00.9925103Z * [new branch] gh/mikaylagawarecki/346/base -> origin/gh/mikaylagawarecki/346/base 2025-11-03T16:58:00.9926792Z * [new branch] gh/mikaylagawarecki/346/head -> origin/gh/mikaylagawarecki/346/head 2025-11-03T16:58:00.9928572Z * [new branch] gh/mikaylagawarecki/346/orig -> origin/gh/mikaylagawarecki/346/orig 2025-11-03T16:58:00.9931041Z * [new branch] gh/mikaylagawarecki/347/base -> origin/gh/mikaylagawarecki/347/base 2025-11-03T16:58:00.9932767Z * [new branch] gh/mikaylagawarecki/347/head -> origin/gh/mikaylagawarecki/347/head 2025-11-03T16:58:00.9934458Z * [new branch] gh/mikaylagawarecki/347/orig -> origin/gh/mikaylagawarecki/347/orig 2025-11-03T16:58:00.9936657Z * [new branch] gh/mikaylagawarecki/348/base -> origin/gh/mikaylagawarecki/348/base 2025-11-03T16:58:00.9938878Z * [new branch] gh/mikaylagawarecki/348/head -> origin/gh/mikaylagawarecki/348/head 2025-11-03T16:58:00.9940825Z * [new branch] gh/mikaylagawarecki/348/orig -> origin/gh/mikaylagawarecki/348/orig 2025-11-03T16:58:00.9943277Z * [new branch] gh/mikaylagawarecki/349/base -> origin/gh/mikaylagawarecki/349/base 2025-11-03T16:58:00.9945075Z * [new branch] gh/mikaylagawarecki/349/head -> origin/gh/mikaylagawarecki/349/head 2025-11-03T16:58:00.9946928Z * [new branch] gh/mikaylagawarecki/349/orig -> origin/gh/mikaylagawarecki/349/orig 2025-11-03T16:58:00.9950016Z * [new branch] gh/mikaylagawarecki/350/base -> origin/gh/mikaylagawarecki/350/base 2025-11-03T16:58:00.9951682Z * [new branch] gh/mikaylagawarecki/350/head -> origin/gh/mikaylagawarecki/350/head 2025-11-03T16:58:00.9953336Z * [new branch] gh/mikaylagawarecki/350/orig -> origin/gh/mikaylagawarecki/350/orig 2025-11-03T16:58:00.9956153Z * [new branch] gh/mikaylagawarecki/351/base -> origin/gh/mikaylagawarecki/351/base 2025-11-03T16:58:00.9957874Z * [new branch] gh/mikaylagawarecki/351/head -> origin/gh/mikaylagawarecki/351/head 2025-11-03T16:58:00.9959850Z * [new branch] gh/mikaylagawarecki/351/orig -> origin/gh/mikaylagawarecki/351/orig 2025-11-03T16:58:00.9962826Z * [new branch] gh/mikaylagawarecki/352/base -> origin/gh/mikaylagawarecki/352/base 2025-11-03T16:58:00.9964163Z * [new branch] gh/mikaylagawarecki/352/head -> origin/gh/mikaylagawarecki/352/head 2025-11-03T16:58:00.9965951Z * [new branch] gh/mikaylagawarecki/352/orig -> origin/gh/mikaylagawarecki/352/orig 2025-11-03T16:58:00.9968466Z * [new branch] gh/mikaylagawarecki/353/base -> origin/gh/mikaylagawarecki/353/base 2025-11-03T16:58:00.9970460Z * [new branch] gh/mikaylagawarecki/353/head -> origin/gh/mikaylagawarecki/353/head 2025-11-03T16:58:00.9972170Z * [new branch] gh/mikaylagawarecki/353/orig -> origin/gh/mikaylagawarecki/353/orig 2025-11-03T16:58:00.9974481Z * [new branch] gh/mikaylagawarecki/354/base -> origin/gh/mikaylagawarecki/354/base 2025-11-03T16:58:00.9976200Z * [new branch] gh/mikaylagawarecki/354/head -> origin/gh/mikaylagawarecki/354/head 2025-11-03T16:58:00.9977953Z * [new branch] gh/mikaylagawarecki/354/orig -> origin/gh/mikaylagawarecki/354/orig 2025-11-03T16:58:00.9980342Z * [new branch] gh/mikaylagawarecki/355/base -> origin/gh/mikaylagawarecki/355/base 2025-11-03T16:58:00.9982146Z * [new branch] gh/mikaylagawarecki/355/head -> origin/gh/mikaylagawarecki/355/head 2025-11-03T16:58:00.9984441Z * [new branch] gh/mikaylagawarecki/355/orig -> origin/gh/mikaylagawarecki/355/orig 2025-11-03T16:58:00.9986893Z * [new branch] gh/mikaylagawarecki/356/base -> origin/gh/mikaylagawarecki/356/base 2025-11-03T16:58:00.9988652Z * [new branch] gh/mikaylagawarecki/356/head -> origin/gh/mikaylagawarecki/356/head 2025-11-03T16:58:00.9990469Z * [new branch] gh/mikaylagawarecki/356/orig -> origin/gh/mikaylagawarecki/356/orig 2025-11-03T16:58:00.9992794Z * [new branch] gh/mikaylagawarecki/357/base -> origin/gh/mikaylagawarecki/357/base 2025-11-03T16:58:00.9994497Z * [new branch] gh/mikaylagawarecki/357/head -> origin/gh/mikaylagawarecki/357/head 2025-11-03T16:58:00.9996238Z * [new branch] gh/mikaylagawarecki/357/orig -> origin/gh/mikaylagawarecki/357/orig 2025-11-03T16:58:00.9998667Z * [new branch] gh/mikaylagawarecki/358/base -> origin/gh/mikaylagawarecki/358/base 2025-11-03T16:58:01.0000404Z * [new branch] gh/mikaylagawarecki/358/head -> origin/gh/mikaylagawarecki/358/head 2025-11-03T16:58:01.0002110Z * [new branch] gh/mikaylagawarecki/358/orig -> origin/gh/mikaylagawarecki/358/orig 2025-11-03T16:58:01.0004412Z * [new branch] gh/mikaylagawarecki/359/base -> origin/gh/mikaylagawarecki/359/base 2025-11-03T16:58:01.0006163Z * [new branch] gh/mikaylagawarecki/359/head -> origin/gh/mikaylagawarecki/359/head 2025-11-03T16:58:01.0008468Z * [new branch] gh/mikaylagawarecki/359/orig -> origin/gh/mikaylagawarecki/359/orig 2025-11-03T16:58:01.0011280Z * [new branch] gh/mikaylagawarecki/360/base -> origin/gh/mikaylagawarecki/360/base 2025-11-03T16:58:01.0012976Z * [new branch] gh/mikaylagawarecki/360/head -> origin/gh/mikaylagawarecki/360/head 2025-11-03T16:58:01.0015047Z * [new branch] gh/mikaylagawarecki/360/orig -> origin/gh/mikaylagawarecki/360/orig 2025-11-03T16:58:01.0017364Z * [new branch] gh/mikaylagawarecki/361/base -> origin/gh/mikaylagawarecki/361/base 2025-11-03T16:58:01.0019111Z * [new branch] gh/mikaylagawarecki/361/head -> origin/gh/mikaylagawarecki/361/head 2025-11-03T16:58:01.0020869Z * [new branch] gh/mikaylagawarecki/361/orig -> origin/gh/mikaylagawarecki/361/orig 2025-11-03T16:58:01.0023752Z * [new branch] gh/mlazos/20/base -> origin/gh/mlazos/20/base 2025-11-03T16:58:01.0025560Z * [new branch] gh/mlazos/20/head -> origin/gh/mlazos/20/head 2025-11-03T16:58:01.0027815Z * [new branch] gh/mlazos/20/orig -> origin/gh/mlazos/20/orig 2025-11-03T16:58:01.0029802Z * [new branch] gh/mlazos/21/base -> origin/gh/mlazos/21/base 2025-11-03T16:58:01.0031519Z * [new branch] gh/mlazos/21/head -> origin/gh/mlazos/21/head 2025-11-03T16:58:01.0033268Z * [new branch] gh/mlazos/21/orig -> origin/gh/mlazos/21/orig 2025-11-03T16:58:01.0035625Z * [new branch] gh/mlazos/22/base -> origin/gh/mlazos/22/base 2025-11-03T16:58:01.0037472Z * [new branch] gh/mlazos/22/head -> origin/gh/mlazos/22/head 2025-11-03T16:58:01.0039134Z * [new branch] gh/mlazos/22/orig -> origin/gh/mlazos/22/orig 2025-11-03T16:58:01.0041468Z * [new branch] gh/mlazos/23/base -> origin/gh/mlazos/23/base 2025-11-03T16:58:01.0043234Z * [new branch] gh/mlazos/23/head -> origin/gh/mlazos/23/head 2025-11-03T16:58:01.0045475Z * [new branch] gh/mlazos/23/orig -> origin/gh/mlazos/23/orig 2025-11-03T16:58:01.0047814Z * [new branch] gh/mlazos/24/base -> origin/gh/mlazos/24/base 2025-11-03T16:58:01.0049654Z * [new branch] gh/mlazos/24/head -> origin/gh/mlazos/24/head 2025-11-03T16:58:01.0051340Z * [new branch] gh/mlazos/24/orig -> origin/gh/mlazos/24/orig 2025-11-03T16:58:01.0053812Z * [new branch] gh/mlazos/25/base -> origin/gh/mlazos/25/base 2025-11-03T16:58:01.0055596Z * [new branch] gh/mlazos/25/head -> origin/gh/mlazos/25/head 2025-11-03T16:58:01.0057884Z * [new branch] gh/mlazos/25/orig -> origin/gh/mlazos/25/orig 2025-11-03T16:58:01.0059775Z * [new branch] gh/mlazos/26/base -> origin/gh/mlazos/26/base 2025-11-03T16:58:01.0061518Z * [new branch] gh/mlazos/26/head -> origin/gh/mlazos/26/head 2025-11-03T16:58:01.0063363Z * [new branch] gh/mlazos/26/orig -> origin/gh/mlazos/26/orig 2025-11-03T16:58:01.0066826Z * [new branch] gh/mlazos/27/base -> origin/gh/mlazos/27/base 2025-11-03T16:58:01.0067895Z * [new branch] gh/mlazos/27/head -> origin/gh/mlazos/27/head 2025-11-03T16:58:01.0069798Z * [new branch] gh/mlazos/27/orig -> origin/gh/mlazos/27/orig 2025-11-03T16:58:01.0072216Z * [new branch] gh/mlazos/28/base -> origin/gh/mlazos/28/base 2025-11-03T16:58:01.0073891Z * [new branch] gh/mlazos/28/head -> origin/gh/mlazos/28/head 2025-11-03T16:58:01.0075606Z * [new branch] gh/mlazos/28/orig -> origin/gh/mlazos/28/orig 2025-11-03T16:58:01.0077963Z * [new branch] gh/mlazos/29/base -> origin/gh/mlazos/29/base 2025-11-03T16:58:01.0079714Z * [new branch] gh/mlazos/29/head -> origin/gh/mlazos/29/head 2025-11-03T16:58:01.0081399Z * [new branch] gh/mlazos/29/orig -> origin/gh/mlazos/29/orig 2025-11-03T16:58:01.0083838Z * [new branch] gh/mlazos/30/base -> origin/gh/mlazos/30/base 2025-11-03T16:58:01.0085655Z * [new branch] gh/mlazos/30/head -> origin/gh/mlazos/30/head 2025-11-03T16:58:01.0090592Z * [new branch] gh/mlazos/30/orig -> origin/gh/mlazos/30/orig 2025-11-03T16:58:01.0091008Z * [new branch] gh/mlazos/31/base -> origin/gh/mlazos/31/base 2025-11-03T16:58:01.0094124Z * [new branch] gh/mlazos/31/head -> origin/gh/mlazos/31/head 2025-11-03T16:58:01.0094477Z * [new branch] gh/mlazos/31/orig -> origin/gh/mlazos/31/orig 2025-11-03T16:58:01.0096384Z * [new branch] gh/mlazos/32/base -> origin/gh/mlazos/32/base 2025-11-03T16:58:01.0097798Z * [new branch] gh/mlazos/32/head -> origin/gh/mlazos/32/head 2025-11-03T16:58:01.0099713Z * [new branch] gh/mlazos/32/orig -> origin/gh/mlazos/32/orig 2025-11-03T16:58:01.0102053Z * [new branch] gh/mlazos/33/base -> origin/gh/mlazos/33/base 2025-11-03T16:58:01.0103831Z * [new branch] gh/mlazos/33/head -> origin/gh/mlazos/33/head 2025-11-03T16:58:01.0105483Z * [new branch] gh/mlazos/33/orig -> origin/gh/mlazos/33/orig 2025-11-03T16:58:01.0108410Z * [new branch] gh/mlazos/34/base -> origin/gh/mlazos/34/base 2025-11-03T16:58:01.0109924Z * [new branch] gh/mlazos/34/head -> origin/gh/mlazos/34/head 2025-11-03T16:58:01.0113120Z * [new branch] gh/mlazos/34/orig -> origin/gh/mlazos/34/orig 2025-11-03T16:58:01.0114770Z * [new branch] gh/mlazos/35/base -> origin/gh/mlazos/35/base 2025-11-03T16:58:01.0117280Z * [new branch] gh/mlazos/35/head -> origin/gh/mlazos/35/head 2025-11-03T16:58:01.0118606Z * [new branch] gh/mlazos/35/orig -> origin/gh/mlazos/35/orig 2025-11-03T16:58:01.0120942Z * [new branch] gh/mlazos/36/base -> origin/gh/mlazos/36/base 2025-11-03T16:58:01.0122718Z * [new branch] gh/mlazos/36/head -> origin/gh/mlazos/36/head 2025-11-03T16:58:01.0124442Z * [new branch] gh/mlazos/36/orig -> origin/gh/mlazos/36/orig 2025-11-03T16:58:01.0127280Z * [new branch] gh/mlazos/37/base -> origin/gh/mlazos/37/base 2025-11-03T16:58:01.0129077Z * [new branch] gh/mlazos/37/head -> origin/gh/mlazos/37/head 2025-11-03T16:58:01.0130825Z * [new branch] gh/mlazos/37/orig -> origin/gh/mlazos/37/orig 2025-11-03T16:58:01.0133693Z * [new branch] gh/mlazos/38/base -> origin/gh/mlazos/38/base 2025-11-03T16:58:01.0135495Z * [new branch] gh/mlazos/38/head -> origin/gh/mlazos/38/head 2025-11-03T16:58:01.0137699Z * [new branch] gh/mlazos/38/orig -> origin/gh/mlazos/38/orig 2025-11-03T16:58:01.0139676Z * [new branch] gh/mlazos/39/base -> origin/gh/mlazos/39/base 2025-11-03T16:58:01.0141426Z * [new branch] gh/mlazos/39/head -> origin/gh/mlazos/39/head 2025-11-03T16:58:01.0143134Z * [new branch] gh/mlazos/39/orig -> origin/gh/mlazos/39/orig 2025-11-03T16:58:01.0145921Z * [new branch] gh/mlazos/40/base -> origin/gh/mlazos/40/base 2025-11-03T16:58:01.0147099Z * [new branch] gh/mlazos/40/head -> origin/gh/mlazos/40/head 2025-11-03T16:58:01.0148739Z * [new branch] gh/mlazos/40/orig -> origin/gh/mlazos/40/orig 2025-11-03T16:58:01.0151233Z * [new branch] gh/mlazos/41/base -> origin/gh/mlazos/41/base 2025-11-03T16:58:01.0153258Z * [new branch] gh/mlazos/41/head -> origin/gh/mlazos/41/head 2025-11-03T16:58:01.0154997Z * [new branch] gh/mlazos/41/orig -> origin/gh/mlazos/41/orig 2025-11-03T16:58:01.0157458Z * [new branch] gh/mlazos/42/base -> origin/gh/mlazos/42/base 2025-11-03T16:58:01.0159211Z * [new branch] gh/mlazos/42/head -> origin/gh/mlazos/42/head 2025-11-03T16:58:01.0161131Z * [new branch] gh/mlazos/42/orig -> origin/gh/mlazos/42/orig 2025-11-03T16:58:01.0163335Z * [new branch] gh/mlazos/43/base -> origin/gh/mlazos/43/base 2025-11-03T16:58:01.0164973Z * [new branch] gh/mlazos/43/head -> origin/gh/mlazos/43/head 2025-11-03T16:58:01.0166651Z * [new branch] gh/mlazos/43/orig -> origin/gh/mlazos/43/orig 2025-11-03T16:58:01.0169505Z * [new branch] gh/mlazos/44/base -> origin/gh/mlazos/44/base 2025-11-03T16:58:01.0171217Z * [new branch] gh/mlazos/44/head -> origin/gh/mlazos/44/head 2025-11-03T16:58:01.0172991Z * [new branch] gh/mlazos/44/orig -> origin/gh/mlazos/44/orig 2025-11-03T16:58:01.0175440Z * [new branch] gh/mlazos/45/base -> origin/gh/mlazos/45/base 2025-11-03T16:58:01.0177142Z * [new branch] gh/mlazos/45/head -> origin/gh/mlazos/45/head 2025-11-03T16:58:01.0179107Z * [new branch] gh/mlazos/45/orig -> origin/gh/mlazos/45/orig 2025-11-03T16:58:01.0181474Z * [new branch] gh/mlazos/46/base -> origin/gh/mlazos/46/base 2025-11-03T16:58:01.0183306Z * [new branch] gh/mlazos/46/head -> origin/gh/mlazos/46/head 2025-11-03T16:58:01.0184983Z * [new branch] gh/mlazos/46/orig -> origin/gh/mlazos/46/orig 2025-11-03T16:58:01.0187776Z * [new branch] gh/mrmiywj/1/base -> origin/gh/mrmiywj/1/base 2025-11-03T16:58:01.0189581Z * [new branch] gh/mrmiywj/1/head -> origin/gh/mrmiywj/1/head 2025-11-03T16:58:01.0192470Z * [new branch] gh/muchulee8/68/base -> origin/gh/muchulee8/68/base 2025-11-03T16:58:01.0194180Z * [new branch] gh/muchulee8/68/head -> origin/gh/muchulee8/68/head 2025-11-03T16:58:01.0196370Z * [new branch] gh/muchulee8/68/orig -> origin/gh/muchulee8/68/orig 2025-11-03T16:58:01.0199708Z * [new branch] gh/muchulee8/69/base -> origin/gh/muchulee8/69/base 2025-11-03T16:58:01.0200551Z * [new branch] gh/muchulee8/69/head -> origin/gh/muchulee8/69/head 2025-11-03T16:58:01.0202158Z * [new branch] gh/muchulee8/69/orig -> origin/gh/muchulee8/69/orig 2025-11-03T16:58:01.0205303Z * [new branch] gh/muchulee8/70/base -> origin/gh/muchulee8/70/base 2025-11-03T16:58:01.0206625Z * [new branch] gh/muchulee8/70/head -> origin/gh/muchulee8/70/head 2025-11-03T16:58:01.0208485Z * [new branch] gh/muchulee8/70/orig -> origin/gh/muchulee8/70/orig 2025-11-03T16:58:01.0211258Z * [new branch] gh/muchulee8/71/base -> origin/gh/muchulee8/71/base 2025-11-03T16:58:01.0212985Z * [new branch] gh/muchulee8/71/head -> origin/gh/muchulee8/71/head 2025-11-03T16:58:01.0214734Z * [new branch] gh/muchulee8/71/orig -> origin/gh/muchulee8/71/orig 2025-11-03T16:58:01.0217068Z * [new branch] gh/muchulee8/72/base -> origin/gh/muchulee8/72/base 2025-11-03T16:58:01.0218823Z * [new branch] gh/muchulee8/72/head -> origin/gh/muchulee8/72/head 2025-11-03T16:58:01.0220594Z * [new branch] gh/muchulee8/72/orig -> origin/gh/muchulee8/72/orig 2025-11-03T16:58:01.0223044Z * [new branch] gh/muchulee8/73/base -> origin/gh/muchulee8/73/base 2025-11-03T16:58:01.0224940Z * [new branch] gh/muchulee8/73/head -> origin/gh/muchulee8/73/head 2025-11-03T16:58:01.0226724Z * [new branch] gh/muchulee8/73/orig -> origin/gh/muchulee8/73/orig 2025-11-03T16:58:01.0229654Z * [new branch] gh/naveenthangudu/1/base -> origin/gh/naveenthangudu/1/base 2025-11-03T16:58:01.0231343Z * [new branch] gh/naveenthangudu/1/head -> origin/gh/naveenthangudu/1/head 2025-11-03T16:58:01.0233417Z * [new branch] gh/naveenthangudu/1/orig -> origin/gh/naveenthangudu/1/orig 2025-11-03T16:58:01.0235631Z * [new branch] gh/naveenthangudu/2/base -> origin/gh/naveenthangudu/2/base 2025-11-03T16:58:01.0237340Z * [new branch] gh/naveenthangudu/2/head -> origin/gh/naveenthangudu/2/head 2025-11-03T16:58:01.0239038Z * [new branch] gh/naveenthangudu/2/orig -> origin/gh/naveenthangudu/2/orig 2025-11-03T16:58:01.0241516Z * [new branch] gh/naveenthangudu/3/base -> origin/gh/naveenthangudu/3/base 2025-11-03T16:58:01.0243212Z * [new branch] gh/naveenthangudu/3/head -> origin/gh/naveenthangudu/3/head 2025-11-03T16:58:01.0245411Z * [new branch] gh/naveenthangudu/3/orig -> origin/gh/naveenthangudu/3/orig 2025-11-03T16:58:01.0248353Z * [new branch] gh/naveenthangudu/4/base -> origin/gh/naveenthangudu/4/base 2025-11-03T16:58:01.0249758Z * [new branch] gh/naveenthangudu/4/head -> origin/gh/naveenthangudu/4/head 2025-11-03T16:58:01.0251587Z * [new branch] gh/naveenthangudu/4/orig -> origin/gh/naveenthangudu/4/orig 2025-11-03T16:58:01.0253984Z * [new branch] gh/naveenthangudu/5/base -> origin/gh/naveenthangudu/5/base 2025-11-03T16:58:01.0255664Z * [new branch] gh/naveenthangudu/5/head -> origin/gh/naveenthangudu/5/head 2025-11-03T16:58:01.0257569Z * [new branch] gh/naveenthangudu/5/orig -> origin/gh/naveenthangudu/5/orig 2025-11-03T16:58:01.0259790Z * [new branch] gh/naveenthangudu/6/base -> origin/gh/naveenthangudu/6/base 2025-11-03T16:58:01.0261487Z * [new branch] gh/naveenthangudu/6/head -> origin/gh/naveenthangudu/6/head 2025-11-03T16:58:01.0263154Z * [new branch] gh/naveenthangudu/6/orig -> origin/gh/naveenthangudu/6/orig 2025-11-03T16:58:01.0265511Z * [new branch] gh/naveenthangudu/7/base -> origin/gh/naveenthangudu/7/base 2025-11-03T16:58:01.0267847Z * [new branch] gh/naveenthangudu/7/head -> origin/gh/naveenthangudu/7/head 2025-11-03T16:58:01.0269032Z * [new branch] gh/naveenthangudu/7/orig -> origin/gh/naveenthangudu/7/orig 2025-11-03T16:58:01.0271873Z * [new branch] gh/naveenthangudu/8/base -> origin/gh/naveenthangudu/8/base 2025-11-03T16:58:01.0273161Z * [new branch] gh/naveenthangudu/8/head -> origin/gh/naveenthangudu/8/head 2025-11-03T16:58:01.0274875Z * [new branch] gh/naveenthangudu/8/orig -> origin/gh/naveenthangudu/8/orig 2025-11-03T16:58:01.0277636Z * [new branch] gh/naveenthangudu/9/base -> origin/gh/naveenthangudu/9/base 2025-11-03T16:58:01.0278879Z * [new branch] gh/naveenthangudu/9/head -> origin/gh/naveenthangudu/9/head 2025-11-03T16:58:01.0280632Z * [new branch] gh/naveenthangudu/9/orig -> origin/gh/naveenthangudu/9/orig 2025-11-03T16:58:01.0283403Z * [new branch] gh/nikitaved/1/base -> origin/gh/nikitaved/1/base 2025-11-03T16:58:01.0285179Z * [new branch] gh/nikitaved/1/head -> origin/gh/nikitaved/1/head 2025-11-03T16:58:01.0286928Z * [new branch] gh/nikitaved/1/orig -> origin/gh/nikitaved/1/orig 2025-11-03T16:58:01.0289333Z * [new branch] gh/nikitaved/2/base -> origin/gh/nikitaved/2/base 2025-11-03T16:58:01.0291181Z * [new branch] gh/nikitaved/2/head -> origin/gh/nikitaved/2/head 2025-11-03T16:58:01.0292900Z * [new branch] gh/nikitaved/2/orig -> origin/gh/nikitaved/2/orig 2025-11-03T16:58:01.0295341Z * [new branch] gh/nikitaved/4/base -> origin/gh/nikitaved/4/base 2025-11-03T16:58:01.0297166Z * [new branch] gh/nikitaved/4/head -> origin/gh/nikitaved/4/head 2025-11-03T16:58:01.0299031Z * [new branch] gh/nikitaved/4/orig -> origin/gh/nikitaved/4/orig 2025-11-03T16:58:01.0301531Z * [new branch] gh/nikitaved/5/base -> origin/gh/nikitaved/5/base 2025-11-03T16:58:01.0303368Z * [new branch] gh/nikitaved/5/head -> origin/gh/nikitaved/5/head 2025-11-03T16:58:01.0304963Z * [new branch] gh/nikitaved/5/orig -> origin/gh/nikitaved/5/orig 2025-11-03T16:58:01.0307223Z * [new branch] gh/nikitaved/6/base -> origin/gh/nikitaved/6/base 2025-11-03T16:58:01.0309017Z * [new branch] gh/nikitaved/6/head -> origin/gh/nikitaved/6/head 2025-11-03T16:58:01.0311029Z * [new branch] gh/nikitaved/6/orig -> origin/gh/nikitaved/6/orig 2025-11-03T16:58:01.0313294Z * [new branch] gh/nikitaved/7/base -> origin/gh/nikitaved/7/base 2025-11-03T16:58:01.0315070Z * [new branch] gh/nikitaved/7/head -> origin/gh/nikitaved/7/head 2025-11-03T16:58:01.0316775Z * [new branch] gh/nikitaved/7/orig -> origin/gh/nikitaved/7/orig 2025-11-03T16:58:01.0319215Z * [new branch] gh/nikitaved/8/base -> origin/gh/nikitaved/8/base 2025-11-03T16:58:01.0320959Z * [new branch] gh/nikitaved/8/head -> origin/gh/nikitaved/8/head 2025-11-03T16:58:01.0322723Z * [new branch] gh/nikitaved/8/orig -> origin/gh/nikitaved/8/orig 2025-11-03T16:58:01.0325507Z * [new branch] gh/oulgen/1/base -> origin/gh/oulgen/1/base 2025-11-03T16:58:01.0327280Z * [new branch] gh/oulgen/1/head -> origin/gh/oulgen/1/head 2025-11-03T16:58:01.0329070Z * [new branch] gh/oulgen/1/orig -> origin/gh/oulgen/1/orig 2025-11-03T16:58:01.0331382Z * [new branch] gh/oulgen/2/base -> origin/gh/oulgen/2/base 2025-11-03T16:58:01.0333042Z * [new branch] gh/oulgen/2/head -> origin/gh/oulgen/2/head 2025-11-03T16:58:01.0334749Z * [new branch] gh/oulgen/2/orig -> origin/gh/oulgen/2/orig 2025-11-03T16:58:01.0336945Z * [new branch] gh/oulgen/3/base -> origin/gh/oulgen/3/base 2025-11-03T16:58:01.0338650Z * [new branch] gh/oulgen/3/head -> origin/gh/oulgen/3/head 2025-11-03T16:58:01.0340353Z * [new branch] gh/oulgen/3/orig -> origin/gh/oulgen/3/orig 2025-11-03T16:58:01.0342724Z * [new branch] gh/oulgen/4/base -> origin/gh/oulgen/4/base 2025-11-03T16:58:01.0344484Z * [new branch] gh/oulgen/4/head -> origin/gh/oulgen/4/head 2025-11-03T16:58:01.0346167Z * [new branch] gh/oulgen/4/orig -> origin/gh/oulgen/4/orig 2025-11-03T16:58:01.0348474Z * [new branch] gh/oulgen/5/base -> origin/gh/oulgen/5/base 2025-11-03T16:58:01.0350190Z * [new branch] gh/oulgen/5/head -> origin/gh/oulgen/5/head 2025-11-03T16:58:01.0351905Z * [new branch] gh/oulgen/5/orig -> origin/gh/oulgen/5/orig 2025-11-03T16:58:01.0354320Z * [new branch] gh/oulgen/6/base -> origin/gh/oulgen/6/base 2025-11-03T16:58:01.0356004Z * [new branch] gh/oulgen/6/head -> origin/gh/oulgen/6/head 2025-11-03T16:58:01.0357847Z * [new branch] gh/oulgen/6/orig -> origin/gh/oulgen/6/orig 2025-11-03T16:58:01.0360162Z * [new branch] gh/oulgen/7/base -> origin/gh/oulgen/7/base 2025-11-03T16:58:01.0361866Z * [new branch] gh/oulgen/7/head -> origin/gh/oulgen/7/head 2025-11-03T16:58:01.0363583Z * [new branch] gh/oulgen/7/orig -> origin/gh/oulgen/7/orig 2025-11-03T16:58:01.0365994Z * [new branch] gh/patvig/mtia-serialization -> origin/gh/patvig/mtia-serialization 2025-11-03T16:58:01.0369111Z * [new branch] gh/pearu/108/base -> origin/gh/pearu/108/base 2025-11-03T16:58:01.0370852Z * [new branch] gh/pearu/108/head -> origin/gh/pearu/108/head 2025-11-03T16:58:01.0372846Z * [new branch] gh/pearu/108/orig -> origin/gh/pearu/108/orig 2025-11-03T16:58:01.0375088Z * [new branch] gh/pearu/109/base -> origin/gh/pearu/109/base 2025-11-03T16:58:01.0376726Z * [new branch] gh/pearu/109/head -> origin/gh/pearu/109/head 2025-11-03T16:58:01.0378424Z * [new branch] gh/pearu/109/orig -> origin/gh/pearu/109/orig 2025-11-03T16:58:01.0380784Z * [new branch] gh/pearu/110/base -> origin/gh/pearu/110/base 2025-11-03T16:58:01.0382515Z * [new branch] gh/pearu/110/head -> origin/gh/pearu/110/head 2025-11-03T16:58:01.0384221Z * [new branch] gh/pearu/110/orig -> origin/gh/pearu/110/orig 2025-11-03T16:58:01.0386592Z * [new branch] gh/pearu/111/base -> origin/gh/pearu/111/base 2025-11-03T16:58:01.0388340Z * [new branch] gh/pearu/111/head -> origin/gh/pearu/111/head 2025-11-03T16:58:01.0390196Z * [new branch] gh/pearu/111/orig -> origin/gh/pearu/111/orig 2025-11-03T16:58:01.0392584Z * [new branch] gh/pearu/112/base -> origin/gh/pearu/112/base 2025-11-03T16:58:01.0394302Z * [new branch] gh/pearu/112/head -> origin/gh/pearu/112/head 2025-11-03T16:58:01.0396062Z * [new branch] gh/pearu/112/orig -> origin/gh/pearu/112/orig 2025-11-03T16:58:01.0398406Z * [new branch] gh/pearu/115/base -> origin/gh/pearu/115/base 2025-11-03T16:58:01.0400158Z * [new branch] gh/pearu/115/head -> origin/gh/pearu/115/head 2025-11-03T16:58:01.0401880Z * [new branch] gh/pearu/115/orig -> origin/gh/pearu/115/orig 2025-11-03T16:58:01.0404098Z * [new branch] gh/pearu/116/base -> origin/gh/pearu/116/base 2025-11-03T16:58:01.0405883Z * [new branch] gh/pearu/116/head -> origin/gh/pearu/116/head 2025-11-03T16:58:01.0408183Z * [new branch] gh/pearu/116/orig -> origin/gh/pearu/116/orig 2025-11-03T16:58:01.0410383Z * [new branch] gh/pearu/117/base -> origin/gh/pearu/117/base 2025-11-03T16:58:01.0412107Z * [new branch] gh/pearu/117/head -> origin/gh/pearu/117/head 2025-11-03T16:58:01.0413904Z * [new branch] gh/pearu/117/orig -> origin/gh/pearu/117/orig 2025-11-03T16:58:01.0416683Z * [new branch] gh/pearu/118/base -> origin/gh/pearu/118/base 2025-11-03T16:58:01.0417801Z * [new branch] gh/pearu/118/head -> origin/gh/pearu/118/head 2025-11-03T16:58:01.0419500Z * [new branch] gh/pearu/118/orig -> origin/gh/pearu/118/orig 2025-11-03T16:58:01.0421785Z * [new branch] gh/pearu/119/base -> origin/gh/pearu/119/base 2025-11-03T16:58:01.0423534Z * [new branch] gh/pearu/119/head -> origin/gh/pearu/119/head 2025-11-03T16:58:01.0425394Z * [new branch] gh/pearu/119/orig -> origin/gh/pearu/119/orig 2025-11-03T16:58:01.0427721Z * [new branch] gh/pearu/120/base -> origin/gh/pearu/120/base 2025-11-03T16:58:01.0429393Z * [new branch] gh/pearu/120/head -> origin/gh/pearu/120/head 2025-11-03T16:58:01.0431082Z * [new branch] gh/pearu/120/orig -> origin/gh/pearu/120/orig 2025-11-03T16:58:01.0433831Z * [new branch] gh/pearu/121/base -> origin/gh/pearu/121/base 2025-11-03T16:58:01.0435602Z * [new branch] gh/pearu/121/head -> origin/gh/pearu/121/head 2025-11-03T16:58:01.0437950Z * [new branch] gh/pearu/121/orig -> origin/gh/pearu/121/orig 2025-11-03T16:58:01.0439914Z * [new branch] gh/pearu/122/base -> origin/gh/pearu/122/base 2025-11-03T16:58:01.0441614Z * [new branch] gh/pearu/122/head -> origin/gh/pearu/122/head 2025-11-03T16:58:01.0443677Z * [new branch] gh/pearu/122/orig -> origin/gh/pearu/122/orig 2025-11-03T16:58:01.0445702Z * [new branch] gh/pearu/138/base -> origin/gh/pearu/138/base 2025-11-03T16:58:01.0447334Z * [new branch] gh/pearu/138/head -> origin/gh/pearu/138/head 2025-11-03T16:58:01.0449102Z * [new branch] gh/pearu/138/orig -> origin/gh/pearu/138/orig 2025-11-03T16:58:01.0451466Z * [new branch] gh/pearu/139/base -> origin/gh/pearu/139/base 2025-11-03T16:58:01.0453104Z * [new branch] gh/pearu/139/head -> origin/gh/pearu/139/head 2025-11-03T16:58:01.0454831Z * [new branch] gh/pearu/139/orig -> origin/gh/pearu/139/orig 2025-11-03T16:58:01.0457158Z * [new branch] gh/pearu/140/base -> origin/gh/pearu/140/base 2025-11-03T16:58:01.0458898Z * [new branch] gh/pearu/140/head -> origin/gh/pearu/140/head 2025-11-03T16:58:01.0460644Z * [new branch] gh/pearu/140/orig -> origin/gh/pearu/140/orig 2025-11-03T16:58:01.0463489Z * [new branch] gh/pearu/141/base -> origin/gh/pearu/141/base 2025-11-03T16:58:01.0465284Z * [new branch] gh/pearu/141/head -> origin/gh/pearu/141/head 2025-11-03T16:58:01.0467018Z * [new branch] gh/pearu/141/orig -> origin/gh/pearu/141/orig 2025-11-03T16:58:01.0469689Z * [new branch] gh/pearu/142/base -> origin/gh/pearu/142/base 2025-11-03T16:58:01.0471498Z * [new branch] gh/pearu/142/head -> origin/gh/pearu/142/head 2025-11-03T16:58:01.0473512Z * [new branch] gh/pearu/142/orig -> origin/gh/pearu/142/orig 2025-11-03T16:58:01.0476063Z * [new branch] gh/pearu/143/base -> origin/gh/pearu/143/base 2025-11-03T16:58:01.0477726Z * [new branch] gh/pearu/143/head -> origin/gh/pearu/143/head 2025-11-03T16:58:01.0479489Z * [new branch] gh/pearu/143/orig -> origin/gh/pearu/143/orig 2025-11-03T16:58:01.0482230Z * [new branch] gh/pearu/144/base -> origin/gh/pearu/144/base 2025-11-03T16:58:01.0483490Z * [new branch] gh/pearu/144/head -> origin/gh/pearu/144/head 2025-11-03T16:58:01.0485256Z * [new branch] gh/pearu/144/orig -> origin/gh/pearu/144/orig 2025-11-03T16:58:01.0487852Z * [new branch] gh/pearu/145/base -> origin/gh/pearu/145/base 2025-11-03T16:58:01.0489555Z * [new branch] gh/pearu/145/head -> origin/gh/pearu/145/head 2025-11-03T16:58:01.0491134Z * [new branch] gh/pearu/145/orig -> origin/gh/pearu/145/orig 2025-11-03T16:58:01.0493701Z * [new branch] gh/pearu/146/base -> origin/gh/pearu/146/base 2025-11-03T16:58:01.0495896Z * [new branch] gh/pearu/146/head -> origin/gh/pearu/146/head 2025-11-03T16:58:01.0498901Z * [new branch] gh/pearu/147/base -> origin/gh/pearu/147/base 2025-11-03T16:58:01.0500203Z * [new branch] gh/pearu/147/head -> origin/gh/pearu/147/head 2025-11-03T16:58:01.0501807Z * [new branch] gh/pearu/147/orig -> origin/gh/pearu/147/orig 2025-11-03T16:58:01.0504724Z * [new branch] gh/pearu/148/base -> origin/gh/pearu/148/base 2025-11-03T16:58:01.0506065Z * [new branch] gh/pearu/148/head -> origin/gh/pearu/148/head 2025-11-03T16:58:01.0507650Z * [new branch] gh/pearu/148/orig -> origin/gh/pearu/148/orig 2025-11-03T16:58:01.0510129Z * [new branch] gh/pearu/149/base -> origin/gh/pearu/149/base 2025-11-03T16:58:01.0513328Z * [new branch] gh/pearu/149/head -> origin/gh/pearu/149/head 2025-11-03T16:58:01.0515025Z * [new branch] gh/pearu/149/orig -> origin/gh/pearu/149/orig 2025-11-03T16:58:01.0517653Z * [new branch] gh/pearu/150/base -> origin/gh/pearu/150/base 2025-11-03T16:58:01.0519343Z * [new branch] gh/pearu/150/head -> origin/gh/pearu/150/head 2025-11-03T16:58:01.0521426Z * [new branch] gh/pearu/150/orig -> origin/gh/pearu/150/orig 2025-11-03T16:58:01.0524164Z * [new branch] gh/pearu/56/base -> origin/gh/pearu/56/base 2025-11-03T16:58:01.0526131Z * [new branch] gh/pearu/56/head -> origin/gh/pearu/56/head 2025-11-03T16:58:01.0528498Z * [new branch] gh/pearu/56/orig -> origin/gh/pearu/56/orig 2025-11-03T16:58:01.0531016Z * [new branch] gh/pearu/97/base -> origin/gh/pearu/97/base 2025-11-03T16:58:01.0532374Z * [new branch] gh/pearu/97/head -> origin/gh/pearu/97/head 2025-11-03T16:58:01.0534105Z * [new branch] gh/pearu/97/orig -> origin/gh/pearu/97/orig 2025-11-03T16:58:01.0537453Z * [new branch] gh/pianpwk/1/base -> origin/gh/pianpwk/1/base 2025-11-03T16:58:01.0538524Z * [new branch] gh/pianpwk/1/head -> origin/gh/pianpwk/1/head 2025-11-03T16:58:01.0540640Z * [new branch] gh/pianpwk/1/orig -> origin/gh/pianpwk/1/orig 2025-11-03T16:58:01.0543032Z * [new branch] gh/pianpwk/10/base -> origin/gh/pianpwk/10/base 2025-11-03T16:58:01.0544860Z * [new branch] gh/pianpwk/10/head -> origin/gh/pianpwk/10/head 2025-11-03T16:58:01.0546544Z * [new branch] gh/pianpwk/10/orig -> origin/gh/pianpwk/10/orig 2025-11-03T16:58:01.0548996Z * [new branch] gh/pianpwk/11/base -> origin/gh/pianpwk/11/base 2025-11-03T16:58:01.0550859Z * [new branch] gh/pianpwk/11/head -> origin/gh/pianpwk/11/head 2025-11-03T16:58:01.0552733Z * [new branch] gh/pianpwk/11/orig -> origin/gh/pianpwk/11/orig 2025-11-03T16:58:01.0555197Z * [new branch] gh/pianpwk/12/base -> origin/gh/pianpwk/12/base 2025-11-03T16:58:01.0557004Z * [new branch] gh/pianpwk/12/head -> origin/gh/pianpwk/12/head 2025-11-03T16:58:01.0558839Z * [new branch] gh/pianpwk/12/orig -> origin/gh/pianpwk/12/orig 2025-11-03T16:58:01.0561228Z * [new branch] gh/pianpwk/13/base -> origin/gh/pianpwk/13/base 2025-11-03T16:58:01.0563018Z * [new branch] gh/pianpwk/13/head -> origin/gh/pianpwk/13/head 2025-11-03T16:58:01.0564911Z * [new branch] gh/pianpwk/13/orig -> origin/gh/pianpwk/13/orig 2025-11-03T16:58:01.0567205Z * [new branch] gh/pianpwk/14/base -> origin/gh/pianpwk/14/base 2025-11-03T16:58:01.0569014Z * [new branch] gh/pianpwk/14/head -> origin/gh/pianpwk/14/head 2025-11-03T16:58:01.0570846Z * [new branch] gh/pianpwk/14/orig -> origin/gh/pianpwk/14/orig 2025-11-03T16:58:01.0573156Z * [new branch] gh/pianpwk/15/base -> origin/gh/pianpwk/15/base 2025-11-03T16:58:01.0574846Z * [new branch] gh/pianpwk/15/head -> origin/gh/pianpwk/15/head 2025-11-03T16:58:01.0576574Z * [new branch] gh/pianpwk/15/orig -> origin/gh/pianpwk/15/orig 2025-11-03T16:58:01.0579003Z * [new branch] gh/pianpwk/16/base -> origin/gh/pianpwk/16/base 2025-11-03T16:58:01.0580779Z * [new branch] gh/pianpwk/16/head -> origin/gh/pianpwk/16/head 2025-11-03T16:58:01.0582594Z * [new branch] gh/pianpwk/16/orig -> origin/gh/pianpwk/16/orig 2025-11-03T16:58:01.0585271Z * [new branch] gh/pianpwk/17/base -> origin/gh/pianpwk/17/base 2025-11-03T16:58:01.0587221Z * [new branch] gh/pianpwk/17/head -> origin/gh/pianpwk/17/head 2025-11-03T16:58:01.0588960Z * [new branch] gh/pianpwk/17/orig -> origin/gh/pianpwk/17/orig 2025-11-03T16:58:01.0591513Z * [new branch] gh/pianpwk/18/base -> origin/gh/pianpwk/18/base 2025-11-03T16:58:01.0593131Z * [new branch] gh/pianpwk/18/head -> origin/gh/pianpwk/18/head 2025-11-03T16:58:01.0595265Z * [new branch] gh/pianpwk/19/base -> origin/gh/pianpwk/19/base 2025-11-03T16:58:01.0597371Z * [new branch] gh/pianpwk/19/head -> origin/gh/pianpwk/19/head 2025-11-03T16:58:01.0599234Z * [new branch] gh/pianpwk/19/orig -> origin/gh/pianpwk/19/orig 2025-11-03T16:58:01.0601270Z * [new branch] gh/pianpwk/2/base -> origin/gh/pianpwk/2/base 2025-11-03T16:58:01.0602948Z * [new branch] gh/pianpwk/2/head -> origin/gh/pianpwk/2/head 2025-11-03T16:58:01.0604680Z * [new branch] gh/pianpwk/2/orig -> origin/gh/pianpwk/2/orig 2025-11-03T16:58:01.0607674Z * [new branch] gh/pianpwk/20/base -> origin/gh/pianpwk/20/base 2025-11-03T16:58:01.0608921Z * [new branch] gh/pianpwk/20/head -> origin/gh/pianpwk/20/head 2025-11-03T16:58:01.0611002Z * [new branch] gh/pianpwk/20/orig -> origin/gh/pianpwk/20/orig 2025-11-03T16:58:01.0613138Z * [new branch] gh/pianpwk/21/base -> origin/gh/pianpwk/21/base 2025-11-03T16:58:01.0614862Z * [new branch] gh/pianpwk/21/head -> origin/gh/pianpwk/21/head 2025-11-03T16:58:01.0617066Z * [new branch] gh/pianpwk/22/base -> origin/gh/pianpwk/22/base 2025-11-03T16:58:01.0618879Z * [new branch] gh/pianpwk/22/head -> origin/gh/pianpwk/22/head 2025-11-03T16:58:01.0620599Z * [new branch] gh/pianpwk/22/orig -> origin/gh/pianpwk/22/orig 2025-11-03T16:58:01.0623040Z * [new branch] gh/pianpwk/23/base -> origin/gh/pianpwk/23/base 2025-11-03T16:58:01.0624782Z * [new branch] gh/pianpwk/23/head -> origin/gh/pianpwk/23/head 2025-11-03T16:58:01.0626527Z * [new branch] gh/pianpwk/23/orig -> origin/gh/pianpwk/23/orig 2025-11-03T16:58:01.0628802Z * [new branch] gh/pianpwk/24/base -> origin/gh/pianpwk/24/base 2025-11-03T16:58:01.0630637Z * [new branch] gh/pianpwk/24/head -> origin/gh/pianpwk/24/head 2025-11-03T16:58:01.0632452Z * [new branch] gh/pianpwk/24/orig -> origin/gh/pianpwk/24/orig 2025-11-03T16:58:01.0634692Z * [new branch] gh/pianpwk/25/base -> origin/gh/pianpwk/25/base 2025-11-03T16:58:01.0636369Z * [new branch] gh/pianpwk/25/head -> origin/gh/pianpwk/25/head 2025-11-03T16:58:01.0638059Z * [new branch] gh/pianpwk/25/orig -> origin/gh/pianpwk/25/orig 2025-11-03T16:58:01.0640738Z * [new branch] gh/pianpwk/26/base -> origin/gh/pianpwk/26/base 2025-11-03T16:58:01.0642514Z * [new branch] gh/pianpwk/26/head -> origin/gh/pianpwk/26/head 2025-11-03T16:58:01.0644226Z * [new branch] gh/pianpwk/26/orig -> origin/gh/pianpwk/26/orig 2025-11-03T16:58:01.0646685Z * [new branch] gh/pianpwk/27/base -> origin/gh/pianpwk/27/base 2025-11-03T16:58:01.0648941Z * [new branch] gh/pianpwk/27/head -> origin/gh/pianpwk/27/head 2025-11-03T16:58:01.0650764Z * [new branch] gh/pianpwk/27/orig -> origin/gh/pianpwk/27/orig 2025-11-03T16:58:01.0653100Z * [new branch] gh/pianpwk/28/base -> origin/gh/pianpwk/28/base 2025-11-03T16:58:01.0654870Z * [new branch] gh/pianpwk/28/head -> origin/gh/pianpwk/28/head 2025-11-03T16:58:01.0656503Z * [new branch] gh/pianpwk/28/orig -> origin/gh/pianpwk/28/orig 2025-11-03T16:58:01.0658684Z * [new branch] gh/pianpwk/3/base -> origin/gh/pianpwk/3/base 2025-11-03T16:58:01.0660366Z * [new branch] gh/pianpwk/3/head -> origin/gh/pianpwk/3/head 2025-11-03T16:58:01.0662272Z * [new branch] gh/pianpwk/3/orig -> origin/gh/pianpwk/3/orig 2025-11-03T16:58:01.0664501Z * [new branch] gh/pianpwk/4/base -> origin/gh/pianpwk/4/base 2025-11-03T16:58:01.0666211Z * [new branch] gh/pianpwk/4/head -> origin/gh/pianpwk/4/head 2025-11-03T16:58:01.0667932Z * [new branch] gh/pianpwk/4/orig -> origin/gh/pianpwk/4/orig 2025-11-03T16:58:01.0670708Z * [new branch] gh/pianpwk/5/base -> origin/gh/pianpwk/5/base 2025-11-03T16:58:01.0672451Z * [new branch] gh/pianpwk/5/head -> origin/gh/pianpwk/5/head 2025-11-03T16:58:01.0674183Z * [new branch] gh/pianpwk/5/orig -> origin/gh/pianpwk/5/orig 2025-11-03T16:58:01.0676654Z * [new branch] gh/pianpwk/6/base -> origin/gh/pianpwk/6/base 2025-11-03T16:58:01.0678333Z * [new branch] gh/pianpwk/6/head -> origin/gh/pianpwk/6/head 2025-11-03T16:58:01.0680099Z * [new branch] gh/pianpwk/6/orig -> origin/gh/pianpwk/6/orig 2025-11-03T16:58:01.0682391Z * [new branch] gh/pianpwk/7/base -> origin/gh/pianpwk/7/base 2025-11-03T16:58:01.0684055Z * [new branch] gh/pianpwk/7/head -> origin/gh/pianpwk/7/head 2025-11-03T16:58:01.0685767Z * [new branch] gh/pianpwk/7/orig -> origin/gh/pianpwk/7/orig 2025-11-03T16:58:01.0688173Z * [new branch] gh/pianpwk/8/base -> origin/gh/pianpwk/8/base 2025-11-03T16:58:01.0690253Z * [new branch] gh/pianpwk/8/head -> origin/gh/pianpwk/8/head 2025-11-03T16:58:01.0691780Z * [new branch] gh/pianpwk/8/orig -> origin/gh/pianpwk/8/orig 2025-11-03T16:58:01.0694690Z * [new branch] gh/pianpwk/9/base -> origin/gh/pianpwk/9/base 2025-11-03T16:58:01.0695962Z * [new branch] gh/pianpwk/9/head -> origin/gh/pianpwk/9/head 2025-11-03T16:58:01.0697550Z * [new branch] gh/pianpwk/9/orig -> origin/gh/pianpwk/9/orig 2025-11-03T16:58:01.0700640Z * [new branch] gh/qqaatw/29/head -> origin/gh/qqaatw/29/head 2025-11-03T16:58:01.0703328Z * [new branch] gh/rec/141/base -> origin/gh/rec/141/base 2025-11-03T16:58:01.0705065Z * [new branch] gh/rec/141/head -> origin/gh/rec/141/head 2025-11-03T16:58:01.0707426Z * [new branch] gh/rec/153/base -> origin/gh/rec/153/base 2025-11-03T16:58:01.0709282Z * [new branch] gh/rec/153/head -> origin/gh/rec/153/head 2025-11-03T16:58:01.0711297Z * [new branch] gh/rec/153/orig -> origin/gh/rec/153/orig 2025-11-03T16:58:01.0713415Z * [new branch] gh/rec/154/base -> origin/gh/rec/154/base 2025-11-03T16:58:01.0715190Z * [new branch] gh/rec/154/head -> origin/gh/rec/154/head 2025-11-03T16:58:01.0716856Z * [new branch] gh/rec/154/orig -> origin/gh/rec/154/orig 2025-11-03T16:58:01.0719278Z * [new branch] gh/rec/164/base -> origin/gh/rec/164/base 2025-11-03T16:58:01.0730567Z * [new branch] gh/rec/164/head -> origin/gh/rec/164/head 2025-11-03T16:58:01.0730900Z * [new branch] gh/rec/164/orig -> origin/gh/rec/164/orig 2025-11-03T16:58:01.0731073Z * [new branch] gh/rec/166/base -> origin/gh/rec/166/base 2025-11-03T16:58:01.0731223Z * [new branch] gh/rec/166/head -> origin/gh/rec/166/head 2025-11-03T16:58:01.0731374Z * [new branch] gh/rec/166/orig -> origin/gh/rec/166/orig 2025-11-03T16:58:01.0731575Z * [new branch] gh/robert-hardwick/3/base -> origin/gh/robert-hardwick/3/base 2025-11-03T16:58:01.0733324Z * [new branch] gh/robert-hardwick/3/head -> origin/gh/robert-hardwick/3/head 2025-11-03T16:58:01.0735405Z * [new branch] gh/robert-hardwick/3/orig -> origin/gh/robert-hardwick/3/orig 2025-11-03T16:58:01.0737448Z * [new branch] gh/robert-hardwick/4/base -> origin/gh/robert-hardwick/4/base 2025-11-03T16:58:01.0739249Z * [new branch] gh/robert-hardwick/4/head -> origin/gh/robert-hardwick/4/head 2025-11-03T16:58:01.0741385Z * [new branch] gh/robert-hardwick/4/orig -> origin/gh/robert-hardwick/4/orig 2025-11-03T16:58:01.0743691Z * [new branch] gh/robert-hardwick/5/base -> origin/gh/robert-hardwick/5/base 2025-11-03T16:58:01.0745388Z * [new branch] gh/robert-hardwick/5/head -> origin/gh/robert-hardwick/5/head 2025-11-03T16:58:01.0747236Z * [new branch] gh/robert-hardwick/5/orig -> origin/gh/robert-hardwick/5/orig 2025-11-03T16:58:01.0750444Z * [new branch] gh/rtimpe/1/base -> origin/gh/rtimpe/1/base 2025-11-03T16:58:01.0751719Z * [new branch] gh/rtimpe/1/head -> origin/gh/rtimpe/1/head 2025-11-03T16:58:01.0754139Z * [new branch] gh/rtimpe/11/base -> origin/gh/rtimpe/11/base 2025-11-03T16:58:01.0755743Z * [new branch] gh/rtimpe/11/head -> origin/gh/rtimpe/11/head 2025-11-03T16:58:01.0757419Z * [new branch] gh/rtimpe/11/orig -> origin/gh/rtimpe/11/orig 2025-11-03T16:58:01.0759690Z * [new branch] gh/rtimpe/16/base -> origin/gh/rtimpe/16/base 2025-11-03T16:58:01.0761396Z * [new branch] gh/rtimpe/16/head -> origin/gh/rtimpe/16/head 2025-11-03T16:58:01.0763130Z * [new branch] gh/rtimpe/16/orig -> origin/gh/rtimpe/16/orig 2025-11-03T16:58:01.0765524Z * [new branch] gh/rtimpe/19/base -> origin/gh/rtimpe/19/base 2025-11-03T16:58:01.0767278Z * [new branch] gh/rtimpe/19/head -> origin/gh/rtimpe/19/head 2025-11-03T16:58:01.0769338Z * [new branch] gh/rtimpe/19/orig -> origin/gh/rtimpe/19/orig 2025-11-03T16:58:01.0771574Z * [new branch] gh/rtimpe/2/base -> origin/gh/rtimpe/2/base 2025-11-03T16:58:01.0773257Z * [new branch] gh/rtimpe/2/head -> origin/gh/rtimpe/2/head 2025-11-03T16:58:01.0776028Z * [new branch] gh/rtimpe/20/base -> origin/gh/rtimpe/20/base 2025-11-03T16:58:01.0778818Z * [new branch] gh/rtimpe/20/head -> origin/gh/rtimpe/20/head 2025-11-03T16:58:01.0780181Z * [new branch] gh/rtimpe/20/orig -> origin/gh/rtimpe/20/orig 2025-11-03T16:58:01.0782766Z * [new branch] gh/rtimpe/21/base -> origin/gh/rtimpe/21/base 2025-11-03T16:58:01.0784186Z * [new branch] gh/rtimpe/21/head -> origin/gh/rtimpe/21/head 2025-11-03T16:58:01.0785911Z * [new branch] gh/rtimpe/21/orig -> origin/gh/rtimpe/21/orig 2025-11-03T16:58:01.0788466Z * [new branch] gh/rtimpe/3/base -> origin/gh/rtimpe/3/base 2025-11-03T16:58:01.0790450Z * [new branch] gh/rtimpe/3/head -> origin/gh/rtimpe/3/head 2025-11-03T16:58:01.0792753Z * [new branch] gh/rtimpe/4/base -> origin/gh/rtimpe/4/base 2025-11-03T16:58:01.0794631Z * [new branch] gh/rtimpe/4/head -> origin/gh/rtimpe/4/head 2025-11-03T16:58:01.0797324Z * [new branch] gh/ruisizhang123/1/base -> origin/gh/ruisizhang123/1/base 2025-11-03T16:58:01.0799012Z * [new branch] gh/ruisizhang123/1/head -> origin/gh/ruisizhang123/1/head 2025-11-03T16:58:01.0800742Z * [new branch] gh/ruisizhang123/1/orig -> origin/gh/ruisizhang123/1/orig 2025-11-03T16:58:01.0803075Z * [new branch] gh/ruisizhang123/4/base -> origin/gh/ruisizhang123/4/base 2025-11-03T16:58:01.0804763Z * [new branch] gh/ruisizhang123/4/head -> origin/gh/ruisizhang123/4/head 2025-11-03T16:58:01.0806703Z * [new branch] gh/ruisizhang123/4/orig -> origin/gh/ruisizhang123/4/orig 2025-11-03T16:58:01.0809028Z * [new branch] gh/ruisizhang123/5/base -> origin/gh/ruisizhang123/5/base 2025-11-03T16:58:01.0812647Z * [new branch] gh/ruisizhang123/5/head -> origin/gh/ruisizhang123/5/head 2025-11-03T16:58:01.0812850Z * [new branch] gh/ruisizhang123/5/orig -> origin/gh/ruisizhang123/5/orig 2025-11-03T16:58:01.0815152Z * [new branch] gh/ruisizhang123/6/base -> origin/gh/ruisizhang123/6/base 2025-11-03T16:58:01.0817357Z * [new branch] gh/ruisizhang123/6/head -> origin/gh/ruisizhang123/6/head 2025-11-03T16:58:01.0819126Z * [new branch] gh/ruisizhang123/6/orig -> origin/gh/ruisizhang123/6/orig 2025-11-03T16:58:01.0821407Z * [new branch] gh/ruisizhang123/7/base -> origin/gh/ruisizhang123/7/base 2025-11-03T16:58:01.0823160Z * [new branch] gh/ruisizhang123/7/head -> origin/gh/ruisizhang123/7/head 2025-11-03T16:58:01.0824597Z * [new branch] gh/ruisizhang123/7/orig -> origin/gh/ruisizhang123/7/orig 2025-11-03T16:58:01.0827183Z * [new branch] gh/ruisizhang123/8/base -> origin/gh/ruisizhang123/8/base 2025-11-03T16:58:01.0828745Z * [new branch] gh/ruisizhang123/8/head -> origin/gh/ruisizhang123/8/head 2025-11-03T16:58:01.0830419Z * [new branch] gh/ruisizhang123/8/orig -> origin/gh/ruisizhang123/8/orig 2025-11-03T16:58:01.0832745Z * [new branch] gh/ruisizhang123/9/base -> origin/gh/ruisizhang123/9/base 2025-11-03T16:58:01.0834429Z * [new branch] gh/ruisizhang123/9/head -> origin/gh/ruisizhang123/9/head 2025-11-03T16:58:01.0836254Z * [new branch] gh/ruisizhang123/9/orig -> origin/gh/ruisizhang123/9/orig 2025-11-03T16:58:01.0839258Z * [new branch] gh/seemethere/43/base -> origin/gh/seemethere/43/base 2025-11-03T16:58:01.0841136Z * [new branch] gh/seemethere/43/head -> origin/gh/seemethere/43/head 2025-11-03T16:58:01.0842827Z * [new branch] gh/seemethere/43/orig -> origin/gh/seemethere/43/orig 2025-11-03T16:58:01.0845175Z * [new branch] gh/seemethere/44/base -> origin/gh/seemethere/44/base 2025-11-03T16:58:01.0846789Z * [new branch] gh/seemethere/44/head -> origin/gh/seemethere/44/head 2025-11-03T16:58:01.0848575Z * [new branch] gh/seemethere/44/orig -> origin/gh/seemethere/44/orig 2025-11-03T16:58:01.0851048Z * [new branch] gh/seemethere/48/base -> origin/gh/seemethere/48/base 2025-11-03T16:58:01.0852675Z * [new branch] gh/seemethere/48/head -> origin/gh/seemethere/48/head 2025-11-03T16:58:01.0854406Z * [new branch] gh/seemethere/48/orig -> origin/gh/seemethere/48/orig 2025-11-03T16:58:01.0856788Z * [new branch] gh/seemethere/49/base -> origin/gh/seemethere/49/base 2025-11-03T16:58:01.0858629Z * [new branch] gh/seemethere/49/head -> origin/gh/seemethere/49/head 2025-11-03T16:58:01.0860791Z * [new branch] gh/seemethere/49/orig -> origin/gh/seemethere/49/orig 2025-11-03T16:58:01.0863262Z * [new branch] gh/seemethere/52/base -> origin/gh/seemethere/52/base 2025-11-03T16:58:01.0865131Z * [new branch] gh/seemethere/52/head -> origin/gh/seemethere/52/head 2025-11-03T16:58:01.0866880Z * [new branch] gh/seemethere/52/orig -> origin/gh/seemethere/52/orig 2025-11-03T16:58:01.0869167Z * [new branch] gh/seemethere/53/base -> origin/gh/seemethere/53/base 2025-11-03T16:58:01.0870856Z * [new branch] gh/seemethere/53/head -> origin/gh/seemethere/53/head 2025-11-03T16:58:01.0872629Z * [new branch] gh/seemethere/53/orig -> origin/gh/seemethere/53/orig 2025-11-03T16:58:01.0875083Z * [new branch] gh/seemethere/54/base -> origin/gh/seemethere/54/base 2025-11-03T16:58:01.0876765Z * [new branch] gh/seemethere/54/head -> origin/gh/seemethere/54/head 2025-11-03T16:58:01.0878491Z * [new branch] gh/seemethere/54/orig -> origin/gh/seemethere/54/orig 2025-11-03T16:58:01.0880762Z * [new branch] gh/seemethere/55/base -> origin/gh/seemethere/55/base 2025-11-03T16:58:01.0882436Z * [new branch] gh/seemethere/55/head -> origin/gh/seemethere/55/head 2025-11-03T16:58:01.0884208Z * [new branch] gh/seemethere/55/orig -> origin/gh/seemethere/55/orig 2025-11-03T16:58:01.0886469Z * [new branch] gh/seemethere/59/base -> origin/gh/seemethere/59/base 2025-11-03T16:58:01.0888206Z * [new branch] gh/seemethere/59/head -> origin/gh/seemethere/59/head 2025-11-03T16:58:01.0890023Z * [new branch] gh/seemethere/59/orig -> origin/gh/seemethere/59/orig 2025-11-03T16:58:01.0892350Z * [new branch] gh/seemethere/62/base -> origin/gh/seemethere/62/base 2025-11-03T16:58:01.0894075Z * [new branch] gh/seemethere/62/head -> origin/gh/seemethere/62/head 2025-11-03T16:58:01.0895797Z * [new branch] gh/seemethere/62/orig -> origin/gh/seemethere/62/orig 2025-11-03T16:58:01.0898085Z * [new branch] gh/seemethere/63/base -> origin/gh/seemethere/63/base 2025-11-03T16:58:01.0899850Z * [new branch] gh/seemethere/63/head -> origin/gh/seemethere/63/head 2025-11-03T16:58:01.0901562Z * [new branch] gh/seemethere/63/orig -> origin/gh/seemethere/63/orig 2025-11-03T16:58:01.0903881Z * [new branch] gh/seemethere/71/base -> origin/gh/seemethere/71/base 2025-11-03T16:58:01.0905897Z * [new branch] gh/seemethere/71/head -> origin/gh/seemethere/71/head 2025-11-03T16:58:01.0908135Z * [new branch] gh/seemethere/71/orig -> origin/gh/seemethere/71/orig 2025-11-03T16:58:01.0910991Z * [new branch] gh/seemethere/72/base -> origin/gh/seemethere/72/base 2025-11-03T16:58:01.0912729Z * [new branch] gh/seemethere/72/head -> origin/gh/seemethere/72/head 2025-11-03T16:58:01.0914458Z * [new branch] gh/seemethere/72/orig -> origin/gh/seemethere/72/orig 2025-11-03T16:58:01.0916750Z * [new branch] gh/seemethere/73/base -> origin/gh/seemethere/73/base 2025-11-03T16:58:01.0918250Z * [new branch] gh/seemethere/73/head -> origin/gh/seemethere/73/head 2025-11-03T16:58:01.0920187Z * [new branch] gh/seemethere/73/orig -> origin/gh/seemethere/73/orig 2025-11-03T16:58:01.0923117Z * [new branch] gh/shunting314/145/base -> origin/gh/shunting314/145/base 2025-11-03T16:58:01.0925030Z * [new branch] gh/shunting314/145/head -> origin/gh/shunting314/145/head 2025-11-03T16:58:01.0926792Z * [new branch] gh/shunting314/145/orig -> origin/gh/shunting314/145/orig 2025-11-03T16:58:01.0929895Z * [new branch] gh/shunting314/176/base -> origin/gh/shunting314/176/base 2025-11-03T16:58:01.0931820Z * [new branch] gh/shunting314/176/head -> origin/gh/shunting314/176/head 2025-11-03T16:58:01.0933709Z * [new branch] gh/shunting314/176/orig -> origin/gh/shunting314/176/orig 2025-11-03T16:58:01.0936128Z * [new branch] gh/shunting314/217/base -> origin/gh/shunting314/217/base 2025-11-03T16:58:01.0937817Z * [new branch] gh/shunting314/217/head -> origin/gh/shunting314/217/head 2025-11-03T16:58:01.0939475Z * [new branch] gh/shunting314/217/orig -> origin/gh/shunting314/217/orig 2025-11-03T16:58:01.0941995Z * [new branch] gh/shunting314/228/base -> origin/gh/shunting314/228/base 2025-11-03T16:58:01.0943718Z * [new branch] gh/shunting314/228/head -> origin/gh/shunting314/228/head 2025-11-03T16:58:01.0945551Z * [new branch] gh/shunting314/228/orig -> origin/gh/shunting314/228/orig 2025-11-03T16:58:01.0948371Z * [new branch] gh/shunting314/229/base -> origin/gh/shunting314/229/base 2025-11-03T16:58:01.0949666Z * [new branch] gh/shunting314/229/head -> origin/gh/shunting314/229/head 2025-11-03T16:58:01.0951383Z * [new branch] gh/shunting314/229/orig -> origin/gh/shunting314/229/orig 2025-11-03T16:58:01.0953765Z * [new branch] gh/shunting314/230/base -> origin/gh/shunting314/230/base 2025-11-03T16:58:01.0955531Z * [new branch] gh/shunting314/230/head -> origin/gh/shunting314/230/head 2025-11-03T16:58:01.0957271Z * [new branch] gh/shunting314/230/orig -> origin/gh/shunting314/230/orig 2025-11-03T16:58:01.0959541Z * [new branch] gh/shunting314/231/base -> origin/gh/shunting314/231/base 2025-11-03T16:58:01.0961320Z * [new branch] gh/shunting314/231/head -> origin/gh/shunting314/231/head 2025-11-03T16:58:01.0963118Z * [new branch] gh/shunting314/231/orig -> origin/gh/shunting314/231/orig 2025-11-03T16:58:01.0965444Z * [new branch] gh/shunting314/232/base -> origin/gh/shunting314/232/base 2025-11-03T16:58:01.0967197Z * [new branch] gh/shunting314/232/head -> origin/gh/shunting314/232/head 2025-11-03T16:58:01.0968911Z * [new branch] gh/shunting314/232/orig -> origin/gh/shunting314/232/orig 2025-11-03T16:58:01.0971187Z * [new branch] gh/shunting314/233/base -> origin/gh/shunting314/233/base 2025-11-03T16:58:01.0972906Z * [new branch] gh/shunting314/233/head -> origin/gh/shunting314/233/head 2025-11-03T16:58:01.0974841Z * [new branch] gh/shunting314/233/orig -> origin/gh/shunting314/233/orig 2025-11-03T16:58:01.0977567Z * [new branch] gh/shunting314/234/base -> origin/gh/shunting314/234/base 2025-11-03T16:58:01.0979490Z * [new branch] gh/shunting314/234/head -> origin/gh/shunting314/234/head 2025-11-03T16:58:01.0981351Z * [new branch] gh/shunting314/234/orig -> origin/gh/shunting314/234/orig 2025-11-03T16:58:01.0983646Z * [new branch] gh/shunting314/235/base -> origin/gh/shunting314/235/base 2025-11-03T16:58:01.0985388Z * [new branch] gh/shunting314/235/head -> origin/gh/shunting314/235/head 2025-11-03T16:58:01.0987113Z * [new branch] gh/shunting314/235/orig -> origin/gh/shunting314/235/orig 2025-11-03T16:58:01.0989995Z * [new branch] gh/shunting314/236/base -> origin/gh/shunting314/236/base 2025-11-03T16:58:01.0991707Z * [new branch] gh/shunting314/236/head -> origin/gh/shunting314/236/head 2025-11-03T16:58:01.0993468Z * [new branch] gh/shunting314/236/orig -> origin/gh/shunting314/236/orig 2025-11-03T16:58:01.0996079Z * [new branch] gh/shunting314/237/base -> origin/gh/shunting314/237/base 2025-11-03T16:58:01.0997894Z * [new branch] gh/shunting314/237/head -> origin/gh/shunting314/237/head 2025-11-03T16:58:01.0999650Z * [new branch] gh/shunting314/237/orig -> origin/gh/shunting314/237/orig 2025-11-03T16:58:01.1002170Z * [new branch] gh/shunting314/238/base -> origin/gh/shunting314/238/base 2025-11-03T16:58:01.1003961Z * [new branch] gh/shunting314/238/head -> origin/gh/shunting314/238/head 2025-11-03T16:58:01.1005747Z * [new branch] gh/shunting314/238/orig -> origin/gh/shunting314/238/orig 2025-11-03T16:58:01.1008847Z * [new branch] gh/shunting314/239/base -> origin/gh/shunting314/239/base 2025-11-03T16:58:01.1010296Z * [new branch] gh/shunting314/239/head -> origin/gh/shunting314/239/head 2025-11-03T16:58:01.1013888Z * [new branch] gh/shunting314/239/orig -> origin/gh/shunting314/239/orig 2025-11-03T16:58:01.1016663Z * [new branch] gh/shunting314/240/base -> origin/gh/shunting314/240/base 2025-11-03T16:58:01.1018025Z * [new branch] gh/shunting314/240/head -> origin/gh/shunting314/240/head 2025-11-03T16:58:01.1019670Z * [new branch] gh/shunting314/240/orig -> origin/gh/shunting314/240/orig 2025-11-03T16:58:01.1022355Z * [new branch] gh/shunting314/241/base -> origin/gh/shunting314/241/base 2025-11-03T16:58:01.1024205Z * [new branch] gh/shunting314/241/head -> origin/gh/shunting314/241/head 2025-11-03T16:58:01.1025923Z * [new branch] gh/shunting314/241/orig -> origin/gh/shunting314/241/orig 2025-11-03T16:58:01.1028359Z * [new branch] gh/shunting314/242/base -> origin/gh/shunting314/242/base 2025-11-03T16:58:01.1030136Z * [new branch] gh/shunting314/242/head -> origin/gh/shunting314/242/head 2025-11-03T16:58:01.1031867Z * [new branch] gh/shunting314/242/orig -> origin/gh/shunting314/242/orig 2025-11-03T16:58:01.1034120Z * [new branch] gh/shunting314/243/base -> origin/gh/shunting314/243/base 2025-11-03T16:58:01.1035910Z * [new branch] gh/shunting314/243/head -> origin/gh/shunting314/243/head 2025-11-03T16:58:01.1037644Z * [new branch] gh/shunting314/243/orig -> origin/gh/shunting314/243/orig 2025-11-03T16:58:01.1040167Z * [new branch] gh/shunting314/244/base -> origin/gh/shunting314/244/base 2025-11-03T16:58:01.1041753Z * [new branch] gh/shunting314/244/head -> origin/gh/shunting314/244/head 2025-11-03T16:58:01.1043442Z * [new branch] gh/shunting314/244/orig -> origin/gh/shunting314/244/orig 2025-11-03T16:58:01.1045669Z * [new branch] gh/shunting314/245/base -> origin/gh/shunting314/245/base 2025-11-03T16:58:01.1047476Z * [new branch] gh/shunting314/245/head -> origin/gh/shunting314/245/head 2025-11-03T16:58:01.1049552Z * [new branch] gh/shunting314/245/orig -> origin/gh/shunting314/245/orig 2025-11-03T16:58:01.1051788Z * [new branch] gh/shunting314/246/base -> origin/gh/shunting314/246/base 2025-11-03T16:58:01.1053464Z * [new branch] gh/shunting314/246/head -> origin/gh/shunting314/246/head 2025-11-03T16:58:01.1055183Z * [new branch] gh/shunting314/246/orig -> origin/gh/shunting314/246/orig 2025-11-03T16:58:01.1057601Z * [new branch] gh/shunting314/247/base -> origin/gh/shunting314/247/base 2025-11-03T16:58:01.1059406Z * [new branch] gh/shunting314/247/head -> origin/gh/shunting314/247/head 2025-11-03T16:58:01.1061210Z * [new branch] gh/shunting314/247/orig -> origin/gh/shunting314/247/orig 2025-11-03T16:58:01.1063723Z * [new branch] gh/shunting314/248/base -> origin/gh/shunting314/248/base 2025-11-03T16:58:01.1065892Z * [new branch] gh/shunting314/248/head -> origin/gh/shunting314/248/head 2025-11-03T16:58:01.1067608Z * [new branch] gh/shunting314/248/orig -> origin/gh/shunting314/248/orig 2025-11-03T16:58:01.1070069Z * [new branch] gh/shunting314/249/base -> origin/gh/shunting314/249/base 2025-11-03T16:58:01.1071845Z * [new branch] gh/shunting314/249/head -> origin/gh/shunting314/249/head 2025-11-03T16:58:01.1073644Z * [new branch] gh/shunting314/249/orig -> origin/gh/shunting314/249/orig 2025-11-03T16:58:01.1076109Z * [new branch] gh/shunting314/250/base -> origin/gh/shunting314/250/base 2025-11-03T16:58:01.1077765Z * [new branch] gh/shunting314/250/head -> origin/gh/shunting314/250/head 2025-11-03T16:58:01.1079480Z * [new branch] gh/shunting314/250/orig -> origin/gh/shunting314/250/orig 2025-11-03T16:58:01.1082066Z * [new branch] gh/shunting314/251/base -> origin/gh/shunting314/251/base 2025-11-03T16:58:01.1083862Z * [new branch] gh/shunting314/251/head -> origin/gh/shunting314/251/head 2025-11-03T16:58:01.1085507Z * [new branch] gh/shunting314/251/orig -> origin/gh/shunting314/251/orig 2025-11-03T16:58:01.1087836Z * [new branch] gh/shunting314/252/base -> origin/gh/shunting314/252/base 2025-11-03T16:58:01.1089734Z * [new branch] gh/shunting314/252/head -> origin/gh/shunting314/252/head 2025-11-03T16:58:01.1091513Z * [new branch] gh/shunting314/252/orig -> origin/gh/shunting314/252/orig 2025-11-03T16:58:01.1094003Z * [new branch] gh/shunting314/253/base -> origin/gh/shunting314/253/base 2025-11-03T16:58:01.1095722Z * [new branch] gh/shunting314/253/head -> origin/gh/shunting314/253/head 2025-11-03T16:58:01.1097433Z * [new branch] gh/shunting314/253/orig -> origin/gh/shunting314/253/orig 2025-11-03T16:58:01.1099995Z * [new branch] gh/shunting314/254/base -> origin/gh/shunting314/254/base 2025-11-03T16:58:01.1101782Z * [new branch] gh/shunting314/254/head -> origin/gh/shunting314/254/head 2025-11-03T16:58:01.1103465Z * [new branch] gh/shunting314/254/orig -> origin/gh/shunting314/254/orig 2025-11-03T16:58:01.1105747Z * [new branch] gh/shunting314/255/base -> origin/gh/shunting314/255/base 2025-11-03T16:58:01.1107565Z * [new branch] gh/shunting314/255/head -> origin/gh/shunting314/255/head 2025-11-03T16:58:01.1109348Z * [new branch] gh/shunting314/255/orig -> origin/gh/shunting314/255/orig 2025-11-03T16:58:01.1112511Z * [new branch] gh/silverguo/1/base -> origin/gh/silverguo/1/base 2025-11-03T16:58:01.1114192Z * [new branch] gh/silverguo/1/head -> origin/gh/silverguo/1/head 2025-11-03T16:58:01.1116343Z * [new branch] gh/silverguo/2/base -> origin/gh/silverguo/2/base 2025-11-03T16:58:01.1118192Z * [new branch] gh/silverguo/2/head -> origin/gh/silverguo/2/head 2025-11-03T16:58:01.1120363Z * [new branch] gh/silverguo/3/base -> origin/gh/silverguo/3/base 2025-11-03T16:58:01.1122103Z * [new branch] gh/silverguo/3/head -> origin/gh/silverguo/3/head 2025-11-03T16:58:01.1124298Z * [new branch] gh/silverguo/4/base -> origin/gh/silverguo/4/base 2025-11-03T16:58:01.1126452Z * [new branch] gh/silverguo/4/head -> origin/gh/silverguo/4/head 2025-11-03T16:58:01.1129446Z * [new branch] gh/sinhaanhsul/1/base -> origin/gh/sinhaanhsul/1/base 2025-11-03T16:58:01.1131163Z * [new branch] gh/sinhaanhsul/1/head -> origin/gh/sinhaanhsul/1/head 2025-11-03T16:58:01.1133923Z * [new branch] gh/slayton58/14/base -> origin/gh/slayton58/14/base 2025-11-03T16:58:01.1135623Z * [new branch] gh/slayton58/14/head -> origin/gh/slayton58/14/head 2025-11-03T16:58:01.1137923Z * [new branch] gh/slayton58/14/orig -> origin/gh/slayton58/14/orig 2025-11-03T16:58:01.1139826Z * [new branch] gh/slayton58/16/base -> origin/gh/slayton58/16/base 2025-11-03T16:58:01.1141448Z * [new branch] gh/slayton58/16/head -> origin/gh/slayton58/16/head 2025-11-03T16:58:01.1143237Z * [new branch] gh/slayton58/16/orig -> origin/gh/slayton58/16/orig 2025-11-03T16:58:01.1145584Z * [new branch] gh/slayton58/17/base -> origin/gh/slayton58/17/base 2025-11-03T16:58:01.1147254Z * [new branch] gh/slayton58/17/head -> origin/gh/slayton58/17/head 2025-11-03T16:58:01.1148954Z * [new branch] gh/slayton58/17/orig -> origin/gh/slayton58/17/orig 2025-11-03T16:58:01.1151408Z * [new branch] gh/slayton58/19/base -> origin/gh/slayton58/19/base 2025-11-03T16:58:01.1153122Z * [new branch] gh/slayton58/19/head -> origin/gh/slayton58/19/head 2025-11-03T16:58:01.1155337Z * [new branch] gh/slayton58/19/orig -> origin/gh/slayton58/19/orig 2025-11-03T16:58:01.1157289Z * [new branch] gh/slayton58/20/base -> origin/gh/slayton58/20/base 2025-11-03T16:58:01.1159048Z * [new branch] gh/slayton58/20/head -> origin/gh/slayton58/20/head 2025-11-03T16:58:01.1160824Z * [new branch] gh/slayton58/20/orig -> origin/gh/slayton58/20/orig 2025-11-03T16:58:01.1163136Z * [new branch] gh/slayton58/21/base -> origin/gh/slayton58/21/base 2025-11-03T16:58:01.1164844Z * [new branch] gh/slayton58/21/head -> origin/gh/slayton58/21/head 2025-11-03T16:58:01.1166564Z * [new branch] gh/slayton58/21/orig -> origin/gh/slayton58/21/orig 2025-11-03T16:58:01.1168830Z * [new branch] gh/slayton58/23/base -> origin/gh/slayton58/23/base 2025-11-03T16:58:01.1170570Z * [new branch] gh/slayton58/23/head -> origin/gh/slayton58/23/head 2025-11-03T16:58:01.1172250Z * [new branch] gh/slayton58/23/orig -> origin/gh/slayton58/23/orig 2025-11-03T16:58:01.1174670Z * [new branch] gh/slayton58/25/base -> origin/gh/slayton58/25/base 2025-11-03T16:58:01.1176355Z * [new branch] gh/slayton58/25/head -> origin/gh/slayton58/25/head 2025-11-03T16:58:01.1178109Z * [new branch] gh/slayton58/25/orig -> origin/gh/slayton58/25/orig 2025-11-03T16:58:01.1180483Z * [new branch] gh/slayton58/26/base -> origin/gh/slayton58/26/base 2025-11-03T16:58:01.1182155Z * [new branch] gh/slayton58/26/head -> origin/gh/slayton58/26/head 2025-11-03T16:58:01.1183836Z * [new branch] gh/slayton58/26/orig -> origin/gh/slayton58/26/orig 2025-11-03T16:58:01.1186366Z * [new branch] gh/slayton58/27/base -> origin/gh/slayton58/27/base 2025-11-03T16:58:01.1188125Z * [new branch] gh/slayton58/27/head -> origin/gh/slayton58/27/head 2025-11-03T16:58:01.1189865Z * [new branch] gh/slayton58/27/orig -> origin/gh/slayton58/27/orig 2025-11-03T16:58:01.1192168Z * [new branch] gh/slayton58/28/base -> origin/gh/slayton58/28/base 2025-11-03T16:58:01.1193846Z * [new branch] gh/slayton58/28/head -> origin/gh/slayton58/28/head 2025-11-03T16:58:01.1195640Z * [new branch] gh/slayton58/28/orig -> origin/gh/slayton58/28/orig 2025-11-03T16:58:01.1198484Z * [new branch] gh/slayton58/29/base -> origin/gh/slayton58/29/base 2025-11-03T16:58:01.1199777Z * [new branch] gh/slayton58/29/head -> origin/gh/slayton58/29/head 2025-11-03T16:58:01.1201735Z * [new branch] gh/slayton58/29/head-jeffdaily-patch-1 -> origin/gh/slayton58/29/head-jeffdaily-patch-1 2025-11-03T16:58:01.1203541Z * [new branch] gh/slayton58/29/head-jeffdaily-patch-2 -> origin/gh/slayton58/29/head-jeffdaily-patch-2 2025-11-03T16:58:01.1204976Z * [new branch] gh/slayton58/29/orig -> origin/gh/slayton58/29/orig 2025-11-03T16:58:01.1207452Z * [new branch] gh/slayton58/30/base -> origin/gh/slayton58/30/base 2025-11-03T16:58:01.1209198Z * [new branch] gh/slayton58/30/head -> origin/gh/slayton58/30/head 2025-11-03T16:58:01.1211396Z * [new branch] gh/slayton58/30/orig -> origin/gh/slayton58/30/orig 2025-11-03T16:58:01.1213852Z * [new branch] gh/slayton58/31/base -> origin/gh/slayton58/31/base 2025-11-03T16:58:01.1215455Z * [new branch] gh/slayton58/31/head -> origin/gh/slayton58/31/head 2025-11-03T16:58:01.1218140Z * [new branch] gh/slayton58/31/orig -> origin/gh/slayton58/31/orig 2025-11-03T16:58:01.1220171Z * [new branch] gh/slayton58/32/base -> origin/gh/slayton58/32/base 2025-11-03T16:58:01.1222231Z * [new branch] gh/slayton58/32/head -> origin/gh/slayton58/32/head 2025-11-03T16:58:01.1223551Z * [new branch] gh/slayton58/32/orig -> origin/gh/slayton58/32/orig 2025-11-03T16:58:01.1225854Z * [new branch] gh/slayton58/33/base -> origin/gh/slayton58/33/base 2025-11-03T16:58:01.1227706Z * [new branch] gh/slayton58/33/head -> origin/gh/slayton58/33/head 2025-11-03T16:58:01.1229827Z * [new branch] gh/slayton58/33/orig -> origin/gh/slayton58/33/orig 2025-11-03T16:58:01.1232009Z * [new branch] gh/slayton58/34/base -> origin/gh/slayton58/34/base 2025-11-03T16:58:01.1233833Z * [new branch] gh/slayton58/34/head -> origin/gh/slayton58/34/head 2025-11-03T16:58:01.1235541Z * [new branch] gh/slayton58/34/orig -> origin/gh/slayton58/34/orig 2025-11-03T16:58:01.1237669Z * [new branch] gh/slayton58/35/base -> origin/gh/slayton58/35/base 2025-11-03T16:58:01.1239327Z * [new branch] gh/slayton58/35/head -> origin/gh/slayton58/35/head 2025-11-03T16:58:01.1241084Z * [new branch] gh/slayton58/35/orig -> origin/gh/slayton58/35/orig 2025-11-03T16:58:01.1243925Z * [new branch] gh/slayton58/36/base -> origin/gh/slayton58/36/base 2025-11-03T16:58:01.1245760Z * [new branch] gh/slayton58/36/head -> origin/gh/slayton58/36/head 2025-11-03T16:58:01.1247548Z * [new branch] gh/slayton58/36/orig -> origin/gh/slayton58/36/orig 2025-11-03T16:58:01.1249996Z * [new branch] gh/slayton58/37/base -> origin/gh/slayton58/37/base 2025-11-03T16:58:01.1251883Z * [new branch] gh/slayton58/37/head -> origin/gh/slayton58/37/head 2025-11-03T16:58:01.1254253Z * [new branch] gh/slayton58/37/orig -> origin/gh/slayton58/37/orig 2025-11-03T16:58:01.1256626Z * [new branch] gh/slayton58/38/base -> origin/gh/slayton58/38/base 2025-11-03T16:58:01.1258302Z * [new branch] gh/slayton58/38/head -> origin/gh/slayton58/38/head 2025-11-03T16:58:01.1259935Z * [new branch] gh/slayton58/38/orig -> origin/gh/slayton58/38/orig 2025-11-03T16:58:01.1262407Z * [new branch] gh/slayton58/39/base -> origin/gh/slayton58/39/base 2025-11-03T16:58:01.1264036Z * [new branch] gh/slayton58/39/head -> origin/gh/slayton58/39/head 2025-11-03T16:58:01.1265722Z * [new branch] gh/slayton58/39/orig -> origin/gh/slayton58/39/orig 2025-11-03T16:58:01.1268073Z * [new branch] gh/slayton58/40/base -> origin/gh/slayton58/40/base 2025-11-03T16:58:01.1269885Z * [new branch] gh/slayton58/40/head -> origin/gh/slayton58/40/head 2025-11-03T16:58:01.1271626Z * [new branch] gh/slayton58/40/orig -> origin/gh/slayton58/40/orig 2025-11-03T16:58:01.1274038Z * [new branch] gh/slayton58/41/base -> origin/gh/slayton58/41/base 2025-11-03T16:58:01.1275778Z * [new branch] gh/slayton58/41/head -> origin/gh/slayton58/41/head 2025-11-03T16:58:01.1277586Z * [new branch] gh/slayton58/41/orig -> origin/gh/slayton58/41/orig 2025-11-03T16:58:01.1279956Z * [new branch] gh/slayton58/42/base -> origin/gh/slayton58/42/base 2025-11-03T16:58:01.1281637Z * [new branch] gh/slayton58/42/head -> origin/gh/slayton58/42/head 2025-11-03T16:58:01.1283421Z * [new branch] gh/slayton58/42/orig -> origin/gh/slayton58/42/orig 2025-11-03T16:58:01.1285659Z * [new branch] gh/slayton58/6/base -> origin/gh/slayton58/6/base 2025-11-03T16:58:01.1287408Z * [new branch] gh/slayton58/6/head -> origin/gh/slayton58/6/head 2025-11-03T16:58:01.1289741Z * [new branch] gh/slayton58/7/base -> origin/gh/slayton58/7/base 2025-11-03T16:58:01.1291487Z * [new branch] gh/slayton58/7/head -> origin/gh/slayton58/7/head 2025-11-03T16:58:01.1294530Z * [new branch] gh/soulitzer/269/base -> origin/gh/soulitzer/269/base 2025-11-03T16:58:01.1296179Z * [new branch] gh/soulitzer/269/head -> origin/gh/soulitzer/269/head 2025-11-03T16:58:01.1297930Z * [new branch] gh/soulitzer/269/orig -> origin/gh/soulitzer/269/orig 2025-11-03T16:58:01.1300322Z * [new branch] gh/soulitzer/276/base -> origin/gh/soulitzer/276/base 2025-11-03T16:58:01.1302080Z * [new branch] gh/soulitzer/276/head -> origin/gh/soulitzer/276/head 2025-11-03T16:58:01.1303766Z * [new branch] gh/soulitzer/276/orig -> origin/gh/soulitzer/276/orig 2025-11-03T16:58:01.1306394Z * [new branch] gh/soulitzer/287/base -> origin/gh/soulitzer/287/base 2025-11-03T16:58:01.1308125Z * [new branch] gh/soulitzer/287/head -> origin/gh/soulitzer/287/head 2025-11-03T16:58:01.1309910Z * [new branch] gh/soulitzer/287/orig -> origin/gh/soulitzer/287/orig 2025-11-03T16:58:01.1312681Z * [new branch] gh/soulitzer/296/base -> origin/gh/soulitzer/296/base 2025-11-03T16:58:01.1314430Z * [new branch] gh/soulitzer/296/head -> origin/gh/soulitzer/296/head 2025-11-03T16:58:01.1316204Z * [new branch] gh/soulitzer/296/orig -> origin/gh/soulitzer/296/orig 2025-11-03T16:58:01.1318505Z * [new branch] gh/soulitzer/299/base -> origin/gh/soulitzer/299/base 2025-11-03T16:58:01.1320253Z * [new branch] gh/soulitzer/299/head -> origin/gh/soulitzer/299/head 2025-11-03T16:58:01.1322164Z * [new branch] gh/soulitzer/299/orig -> origin/gh/soulitzer/299/orig 2025-11-03T16:58:01.1324456Z * [new branch] gh/soulitzer/300/base -> origin/gh/soulitzer/300/base 2025-11-03T16:58:01.1326236Z * [new branch] gh/soulitzer/300/head -> origin/gh/soulitzer/300/head 2025-11-03T16:58:01.1327950Z * [new branch] gh/soulitzer/300/orig -> origin/gh/soulitzer/300/orig 2025-11-03T16:58:01.1330527Z * [new branch] gh/soulitzer/301/base -> origin/gh/soulitzer/301/base 2025-11-03T16:58:01.1332245Z * [new branch] gh/soulitzer/301/head -> origin/gh/soulitzer/301/head 2025-11-03T16:58:01.1333914Z * [new branch] gh/soulitzer/301/orig -> origin/gh/soulitzer/301/orig 2025-11-03T16:58:01.1336244Z * [new branch] gh/soulitzer/313/base -> origin/gh/soulitzer/313/base 2025-11-03T16:58:01.1337962Z * [new branch] gh/soulitzer/313/head -> origin/gh/soulitzer/313/head 2025-11-03T16:58:01.1339753Z * [new branch] gh/soulitzer/313/orig -> origin/gh/soulitzer/313/orig 2025-11-03T16:58:01.1342118Z * [new branch] gh/soulitzer/319/base -> origin/gh/soulitzer/319/base 2025-11-03T16:58:01.1343772Z * [new branch] gh/soulitzer/319/head -> origin/gh/soulitzer/319/head 2025-11-03T16:58:01.1345473Z * [new branch] gh/soulitzer/319/orig -> origin/gh/soulitzer/319/orig 2025-11-03T16:58:01.1347787Z * [new branch] gh/soulitzer/320/base -> origin/gh/soulitzer/320/base 2025-11-03T16:58:01.1349395Z * [new branch] gh/soulitzer/320/head -> origin/gh/soulitzer/320/head 2025-11-03T16:58:01.1351042Z * [new branch] gh/soulitzer/320/orig -> origin/gh/soulitzer/320/orig 2025-11-03T16:58:01.1353435Z * [new branch] gh/soulitzer/336/base -> origin/gh/soulitzer/336/base 2025-11-03T16:58:01.1355083Z * [new branch] gh/soulitzer/336/head -> origin/gh/soulitzer/336/head 2025-11-03T16:58:01.1356746Z * [new branch] gh/soulitzer/336/orig -> origin/gh/soulitzer/336/orig 2025-11-03T16:58:01.1359123Z * [new branch] gh/soulitzer/347/base -> origin/gh/soulitzer/347/base 2025-11-03T16:58:01.1360908Z * [new branch] gh/soulitzer/347/head -> origin/gh/soulitzer/347/head 2025-11-03T16:58:01.1362658Z * [new branch] gh/soulitzer/347/orig -> origin/gh/soulitzer/347/orig 2025-11-03T16:58:01.1365073Z * [new branch] gh/soulitzer/349/base -> origin/gh/soulitzer/349/base 2025-11-03T16:58:01.1366841Z * [new branch] gh/soulitzer/349/head -> origin/gh/soulitzer/349/head 2025-11-03T16:58:01.1369189Z * [new branch] gh/soulitzer/349/orig -> origin/gh/soulitzer/349/orig 2025-11-03T16:58:01.1371419Z * [new branch] gh/soulitzer/350/base -> origin/gh/soulitzer/350/base 2025-11-03T16:58:01.1373051Z * [new branch] gh/soulitzer/350/head -> origin/gh/soulitzer/350/head 2025-11-03T16:58:01.1374822Z * [new branch] gh/soulitzer/350/orig -> origin/gh/soulitzer/350/orig 2025-11-03T16:58:01.1377224Z * [new branch] gh/soulitzer/351/base -> origin/gh/soulitzer/351/base 2025-11-03T16:58:01.1378951Z * [new branch] gh/soulitzer/351/head -> origin/gh/soulitzer/351/head 2025-11-03T16:58:01.1380689Z * [new branch] gh/soulitzer/351/orig -> origin/gh/soulitzer/351/orig 2025-11-03T16:58:01.1382942Z * [new branch] gh/soulitzer/353/base -> origin/gh/soulitzer/353/base 2025-11-03T16:58:01.1384738Z * [new branch] gh/soulitzer/353/head -> origin/gh/soulitzer/353/head 2025-11-03T16:58:01.1386942Z * [new branch] gh/soulitzer/353/orig -> origin/gh/soulitzer/353/orig 2025-11-03T16:58:01.1390043Z * [new branch] gh/soulitzer/358/base -> origin/gh/soulitzer/358/base 2025-11-03T16:58:01.1391855Z * [new branch] gh/soulitzer/358/head -> origin/gh/soulitzer/358/head 2025-11-03T16:58:01.1393591Z * [new branch] gh/soulitzer/358/orig -> origin/gh/soulitzer/358/orig 2025-11-03T16:58:01.1396299Z * [new branch] gh/soulitzer/359/base -> origin/gh/soulitzer/359/base 2025-11-03T16:58:01.1398546Z * [new branch] gh/soulitzer/359/head -> origin/gh/soulitzer/359/head 2025-11-03T16:58:01.1400008Z * [new branch] gh/soulitzer/359/orig -> origin/gh/soulitzer/359/orig 2025-11-03T16:58:01.1402212Z * [new branch] gh/soulitzer/374/base -> origin/gh/soulitzer/374/base 2025-11-03T16:58:01.1403900Z * [new branch] gh/soulitzer/374/head -> origin/gh/soulitzer/374/head 2025-11-03T16:58:01.1405570Z * [new branch] gh/soulitzer/374/orig -> origin/gh/soulitzer/374/orig 2025-11-03T16:58:01.1407732Z * [new branch] gh/soulitzer/375/base -> origin/gh/soulitzer/375/base 2025-11-03T16:58:01.1409679Z * [new branch] gh/soulitzer/375/head -> origin/gh/soulitzer/375/head 2025-11-03T16:58:01.1411878Z * [new branch] gh/soulitzer/375/orig -> origin/gh/soulitzer/375/orig 2025-11-03T16:58:01.1414236Z * [new branch] gh/soulitzer/380/base -> origin/gh/soulitzer/380/base 2025-11-03T16:58:01.1416092Z * [new branch] gh/soulitzer/380/head -> origin/gh/soulitzer/380/head 2025-11-03T16:58:01.1417767Z * [new branch] gh/soulitzer/380/orig -> origin/gh/soulitzer/380/orig 2025-11-03T16:58:01.1420084Z * [new branch] gh/soulitzer/381/base -> origin/gh/soulitzer/381/base 2025-11-03T16:58:01.1421786Z * [new branch] gh/soulitzer/381/head -> origin/gh/soulitzer/381/head 2025-11-03T16:58:01.1423489Z * [new branch] gh/soulitzer/381/orig -> origin/gh/soulitzer/381/orig 2025-11-03T16:58:01.1425851Z * [new branch] gh/soulitzer/384/base -> origin/gh/soulitzer/384/base 2025-11-03T16:58:01.1427559Z * [new branch] gh/soulitzer/384/head -> origin/gh/soulitzer/384/head 2025-11-03T16:58:01.1429189Z * [new branch] gh/soulitzer/384/orig -> origin/gh/soulitzer/384/orig 2025-11-03T16:58:01.1431782Z * [new branch] gh/soulitzer/385/base -> origin/gh/soulitzer/385/base 2025-11-03T16:58:01.1433339Z * [new branch] gh/soulitzer/385/head -> origin/gh/soulitzer/385/head 2025-11-03T16:58:01.1435059Z * [new branch] gh/soulitzer/385/orig -> origin/gh/soulitzer/385/orig 2025-11-03T16:58:01.1437455Z * [new branch] gh/soulitzer/386/base -> origin/gh/soulitzer/386/base 2025-11-03T16:58:01.1439181Z * [new branch] gh/soulitzer/386/head -> origin/gh/soulitzer/386/head 2025-11-03T16:58:01.1440871Z * [new branch] gh/soulitzer/386/orig -> origin/gh/soulitzer/386/orig 2025-11-03T16:58:01.1443251Z * [new branch] gh/soulitzer/387/base -> origin/gh/soulitzer/387/base 2025-11-03T16:58:01.1444962Z * [new branch] gh/soulitzer/387/head -> origin/gh/soulitzer/387/head 2025-11-03T16:58:01.1446701Z * [new branch] gh/soulitzer/387/orig -> origin/gh/soulitzer/387/orig 2025-11-03T16:58:01.1449828Z * [new branch] gh/swolchok/728/next -> origin/gh/swolchok/728/next 2025-11-03T16:58:01.1452111Z * [new branch] gh/swolchok/786/base -> origin/gh/swolchok/786/base 2025-11-03T16:58:01.1453746Z * [new branch] gh/swolchok/786/head -> origin/gh/swolchok/786/head 2025-11-03T16:58:01.1455425Z * [new branch] gh/swolchok/786/orig -> origin/gh/swolchok/786/orig 2025-11-03T16:58:01.1458395Z * [new branch] gh/swolchok/787/base -> origin/gh/swolchok/787/base 2025-11-03T16:58:01.1459686Z * [new branch] gh/swolchok/787/head -> origin/gh/swolchok/787/head 2025-11-03T16:58:01.1461354Z * [new branch] gh/swolchok/787/orig -> origin/gh/swolchok/787/orig 2025-11-03T16:58:01.1463864Z * [new branch] gh/swolchok/819/base -> origin/gh/swolchok/819/base 2025-11-03T16:58:01.1465520Z * [new branch] gh/swolchok/819/head -> origin/gh/swolchok/819/head 2025-11-03T16:58:01.1467232Z * [new branch] gh/swolchok/819/orig -> origin/gh/swolchok/819/orig 2025-11-03T16:58:01.1469422Z * [new branch] gh/swolchok/824/base -> origin/gh/swolchok/824/base 2025-11-03T16:58:01.1471162Z * [new branch] gh/swolchok/824/head -> origin/gh/swolchok/824/head 2025-11-03T16:58:01.1472846Z * [new branch] gh/swolchok/824/orig -> origin/gh/swolchok/824/orig 2025-11-03T16:58:01.1475179Z * [new branch] gh/swolchok/829/base -> origin/gh/swolchok/829/base 2025-11-03T16:58:01.1476810Z * [new branch] gh/swolchok/829/head -> origin/gh/swolchok/829/head 2025-11-03T16:58:01.1478597Z * [new branch] gh/swolchok/829/orig -> origin/gh/swolchok/829/orig 2025-11-03T16:58:01.1481533Z * [new branch] gh/swolchok/838/base -> origin/gh/swolchok/838/base 2025-11-03T16:58:01.1483257Z * [new branch] gh/swolchok/838/head -> origin/gh/swolchok/838/head 2025-11-03T16:58:01.1484973Z * [new branch] gh/swolchok/838/orig -> origin/gh/swolchok/838/orig 2025-11-03T16:58:01.1487326Z * [new branch] gh/swolchok/839/base -> origin/gh/swolchok/839/base 2025-11-03T16:58:01.1489070Z * [new branch] gh/swolchok/839/head -> origin/gh/swolchok/839/head 2025-11-03T16:58:01.1490771Z * [new branch] gh/swolchok/839/orig -> origin/gh/swolchok/839/orig 2025-11-03T16:58:01.1493016Z * [new branch] gh/swolchok/841/base -> origin/gh/swolchok/841/base 2025-11-03T16:58:01.1494713Z * [new branch] gh/swolchok/841/head -> origin/gh/swolchok/841/head 2025-11-03T16:58:01.1496478Z * [new branch] gh/swolchok/841/orig -> origin/gh/swolchok/841/orig 2025-11-03T16:58:01.1498770Z * [new branch] gh/swolchok/842/base -> origin/gh/swolchok/842/base 2025-11-03T16:58:01.1501043Z * [new branch] gh/swolchok/842/head -> origin/gh/swolchok/842/head 2025-11-03T16:58:01.1502661Z * [new branch] gh/swolchok/842/orig -> origin/gh/swolchok/842/orig 2025-11-03T16:58:01.1504970Z * [new branch] gh/swolchok/843/base -> origin/gh/swolchok/843/base 2025-11-03T16:58:01.1506703Z * [new branch] gh/swolchok/843/head -> origin/gh/swolchok/843/head 2025-11-03T16:58:01.1508439Z * [new branch] gh/swolchok/843/orig -> origin/gh/swolchok/843/orig 2025-11-03T16:58:01.1511184Z * [new branch] gh/swolchok/844/base -> origin/gh/swolchok/844/base 2025-11-03T16:58:01.1512903Z * [new branch] gh/swolchok/844/head -> origin/gh/swolchok/844/head 2025-11-03T16:58:01.1514714Z * [new branch] gh/swolchok/844/orig -> origin/gh/swolchok/844/orig 2025-11-03T16:58:01.1517050Z * [new branch] gh/swolchok/845/base -> origin/gh/swolchok/845/base 2025-11-03T16:58:01.1518676Z * [new branch] gh/swolchok/845/head -> origin/gh/swolchok/845/head 2025-11-03T16:58:01.1520371Z * [new branch] gh/swolchok/845/orig -> origin/gh/swolchok/845/orig 2025-11-03T16:58:01.1522751Z * [new branch] gh/swolchok/847/base -> origin/gh/swolchok/847/base 2025-11-03T16:58:01.1524664Z * [new branch] gh/swolchok/847/head -> origin/gh/swolchok/847/head 2025-11-03T16:58:01.1526418Z * [new branch] gh/swolchok/847/orig -> origin/gh/swolchok/847/orig 2025-11-03T16:58:01.1529397Z * [new branch] gh/swolchok/848/base -> origin/gh/swolchok/848/base 2025-11-03T16:58:01.1531136Z * [new branch] gh/swolchok/848/head -> origin/gh/swolchok/848/head 2025-11-03T16:58:01.1532813Z * [new branch] gh/swolchok/848/orig -> origin/gh/swolchok/848/orig 2025-11-03T16:58:01.1535158Z * [new branch] gh/swolchok/849/base -> origin/gh/swolchok/849/base 2025-11-03T16:58:01.1536940Z * [new branch] gh/swolchok/849/head -> origin/gh/swolchok/849/head 2025-11-03T16:58:01.1538669Z * [new branch] gh/swolchok/849/orig -> origin/gh/swolchok/849/orig 2025-11-03T16:58:01.1541451Z * [new branch] gh/swolchok/850/base -> origin/gh/swolchok/850/base 2025-11-03T16:58:01.1543188Z * [new branch] gh/swolchok/850/head -> origin/gh/swolchok/850/head 2025-11-03T16:58:01.1544931Z * [new branch] gh/swolchok/850/orig -> origin/gh/swolchok/850/orig 2025-11-03T16:58:01.1547182Z * [new branch] gh/swolchok/851/base -> origin/gh/swolchok/851/base 2025-11-03T16:58:01.1548962Z * [new branch] gh/swolchok/851/head -> origin/gh/swolchok/851/head 2025-11-03T16:58:01.1550671Z * [new branch] gh/swolchok/851/orig -> origin/gh/swolchok/851/orig 2025-11-03T16:58:01.1553249Z * [new branch] gh/swolchok/852/base -> origin/gh/swolchok/852/base 2025-11-03T16:58:01.1554970Z * [new branch] gh/swolchok/852/head -> origin/gh/swolchok/852/head 2025-11-03T16:58:01.1556661Z * [new branch] gh/swolchok/852/orig -> origin/gh/swolchok/852/orig 2025-11-03T16:58:01.1559096Z * [new branch] gh/swolchok/853/base -> origin/gh/swolchok/853/base 2025-11-03T16:58:01.1560852Z * [new branch] gh/swolchok/853/head -> origin/gh/swolchok/853/head 2025-11-03T16:58:01.1562559Z * [new branch] gh/swolchok/853/orig -> origin/gh/swolchok/853/orig 2025-11-03T16:58:01.1564928Z * [new branch] gh/swolchok/854/base -> origin/gh/swolchok/854/base 2025-11-03T16:58:01.1566726Z * [new branch] gh/swolchok/854/head -> origin/gh/swolchok/854/head 2025-11-03T16:58:01.1568509Z * [new branch] gh/swolchok/854/orig -> origin/gh/swolchok/854/orig 2025-11-03T16:58:01.1571141Z * [new branch] gh/swolchok/855/base -> origin/gh/swolchok/855/base 2025-11-03T16:58:01.1572750Z * [new branch] gh/swolchok/855/head -> origin/gh/swolchok/855/head 2025-11-03T16:58:01.1574446Z * [new branch] gh/swolchok/855/orig -> origin/gh/swolchok/855/orig 2025-11-03T16:58:01.1576919Z * [new branch] gh/swolchok/856/base -> origin/gh/swolchok/856/base 2025-11-03T16:58:01.1578588Z * [new branch] gh/swolchok/856/head -> origin/gh/swolchok/856/head 2025-11-03T16:58:01.1580253Z * [new branch] gh/swolchok/856/orig -> origin/gh/swolchok/856/orig 2025-11-03T16:58:01.1582759Z * [new branch] gh/swolchok/857/base -> origin/gh/swolchok/857/base 2025-11-03T16:58:01.1584550Z * [new branch] gh/swolchok/857/head -> origin/gh/swolchok/857/head 2025-11-03T16:58:01.1586315Z * [new branch] gh/swolchok/857/orig -> origin/gh/swolchok/857/orig 2025-11-03T16:58:01.1588678Z * [new branch] gh/swolchok/858/base -> origin/gh/swolchok/858/base 2025-11-03T16:58:01.1590398Z * [new branch] gh/swolchok/858/head -> origin/gh/swolchok/858/head 2025-11-03T16:58:01.1592099Z * [new branch] gh/swolchok/858/orig -> origin/gh/swolchok/858/orig 2025-11-03T16:58:01.1594650Z * [new branch] gh/swolchok/859/base -> origin/gh/swolchok/859/base 2025-11-03T16:58:01.1596362Z * [new branch] gh/swolchok/859/head -> origin/gh/swolchok/859/head 2025-11-03T16:58:01.1598096Z * [new branch] gh/swolchok/859/orig -> origin/gh/swolchok/859/orig 2025-11-03T16:58:01.1600669Z * [new branch] gh/swolchok/860/base -> origin/gh/swolchok/860/base 2025-11-03T16:58:01.1602502Z * [new branch] gh/swolchok/860/head -> origin/gh/swolchok/860/head 2025-11-03T16:58:01.1604244Z * [new branch] gh/swolchok/860/orig -> origin/gh/swolchok/860/orig 2025-11-03T16:58:01.1606785Z * [new branch] gh/swolchok/861/base -> origin/gh/swolchok/861/base 2025-11-03T16:58:01.1608570Z * [new branch] gh/swolchok/861/head -> origin/gh/swolchok/861/head 2025-11-03T16:58:01.1611271Z * [new branch] gh/swolchok/861/orig -> origin/gh/swolchok/861/orig 2025-11-03T16:58:01.1614078Z * [new branch] gh/swolchok/862/base -> origin/gh/swolchok/862/base 2025-11-03T16:58:01.1615856Z * [new branch] gh/swolchok/862/head -> origin/gh/swolchok/862/head 2025-11-03T16:58:01.1617974Z * [new branch] gh/swolchok/862/orig -> origin/gh/swolchok/862/orig 2025-11-03T16:58:01.1620574Z * [new branch] gh/swolchok/863/base -> origin/gh/swolchok/863/base 2025-11-03T16:58:01.1622312Z * [new branch] gh/swolchok/863/head -> origin/gh/swolchok/863/head 2025-11-03T16:58:01.1624028Z * [new branch] gh/swolchok/863/orig -> origin/gh/swolchok/863/orig 2025-11-03T16:58:01.1626651Z * [new branch] gh/swolchok/864/base -> origin/gh/swolchok/864/base 2025-11-03T16:58:01.1628477Z * [new branch] gh/swolchok/864/head -> origin/gh/swolchok/864/head 2025-11-03T16:58:01.1630202Z * [new branch] gh/swolchok/864/orig -> origin/gh/swolchok/864/orig 2025-11-03T16:58:01.1632437Z * [new branch] gh/swolchok/865/base -> origin/gh/swolchok/865/base 2025-11-03T16:58:01.1634178Z * [new branch] gh/swolchok/865/head -> origin/gh/swolchok/865/head 2025-11-03T16:58:01.1635914Z * [new branch] gh/swolchok/865/orig -> origin/gh/swolchok/865/orig 2025-11-03T16:58:01.1638724Z * [new branch] gh/teja-rao/4/base -> origin/gh/teja-rao/4/base 2025-11-03T16:58:01.1640412Z * [new branch] gh/teja-rao/4/head -> origin/gh/teja-rao/4/head 2025-11-03T16:58:01.1642291Z * [new branch] gh/teja-rao/4/orig -> origin/gh/teja-rao/4/orig 2025-11-03T16:58:01.1644996Z * [new branch] gh/tianyu-l/2/base -> origin/gh/tianyu-l/2/base 2025-11-03T16:58:01.1646774Z * [new branch] gh/tianyu-l/2/head -> origin/gh/tianyu-l/2/head 2025-11-03T16:58:01.1648493Z * [new branch] gh/tianyu-l/2/orig -> origin/gh/tianyu-l/2/orig 2025-11-03T16:58:01.1651678Z * [new branch] gh/tugsbayasgalan/10/base -> origin/gh/tugsbayasgalan/10/base 2025-11-03T16:58:01.1653412Z * [new branch] gh/tugsbayasgalan/10/head -> origin/gh/tugsbayasgalan/10/head 2025-11-03T16:58:01.1655097Z * [new branch] gh/tugsbayasgalan/10/orig -> origin/gh/tugsbayasgalan/10/orig 2025-11-03T16:58:01.1657406Z * [new branch] gh/tugsbayasgalan/13/base -> origin/gh/tugsbayasgalan/13/base 2025-11-03T16:58:01.1659111Z * [new branch] gh/tugsbayasgalan/13/head -> origin/gh/tugsbayasgalan/13/head 2025-11-03T16:58:01.1660851Z * [new branch] gh/tugsbayasgalan/13/orig -> origin/gh/tugsbayasgalan/13/orig 2025-11-03T16:58:01.1663209Z * [new branch] gh/tugsbayasgalan/17/base -> origin/gh/tugsbayasgalan/17/base 2025-11-03T16:58:01.1665055Z * [new branch] gh/tugsbayasgalan/17/head -> origin/gh/tugsbayasgalan/17/head 2025-11-03T16:58:01.1667206Z * [new branch] gh/tugsbayasgalan/17/orig -> origin/gh/tugsbayasgalan/17/orig 2025-11-03T16:58:01.1669683Z * [new branch] gh/tugsbayasgalan/2/base -> origin/gh/tugsbayasgalan/2/base 2025-11-03T16:58:01.1671364Z * [new branch] gh/tugsbayasgalan/2/head -> origin/gh/tugsbayasgalan/2/head 2025-11-03T16:58:01.1673132Z * [new branch] gh/tugsbayasgalan/2/orig -> origin/gh/tugsbayasgalan/2/orig 2025-11-03T16:58:01.1675692Z * [new branch] gh/tugsbayasgalan/28/base -> origin/gh/tugsbayasgalan/28/base 2025-11-03T16:58:01.1677450Z * [new branch] gh/tugsbayasgalan/28/head -> origin/gh/tugsbayasgalan/28/head 2025-11-03T16:58:01.1679117Z * [new branch] gh/tugsbayasgalan/28/orig -> origin/gh/tugsbayasgalan/28/orig 2025-11-03T16:58:01.1681303Z * [new branch] gh/tugsbayasgalan/3/base -> origin/gh/tugsbayasgalan/3/base 2025-11-03T16:58:01.1683088Z * [new branch] gh/tugsbayasgalan/3/head -> origin/gh/tugsbayasgalan/3/head 2025-11-03T16:58:01.1684776Z * [new branch] gh/tugsbayasgalan/3/orig -> origin/gh/tugsbayasgalan/3/orig 2025-11-03T16:58:01.1687521Z * [new branch] gh/tugsbayasgalan/32/base -> origin/gh/tugsbayasgalan/32/base 2025-11-03T16:58:01.1689399Z * [new branch] gh/tugsbayasgalan/32/head -> origin/gh/tugsbayasgalan/32/head 2025-11-03T16:58:01.1691240Z * [new branch] gh/tugsbayasgalan/32/orig -> origin/gh/tugsbayasgalan/32/orig 2025-11-03T16:58:01.1693672Z * [new branch] gh/tugsbayasgalan/35/base -> origin/gh/tugsbayasgalan/35/base 2025-11-03T16:58:01.1695440Z * [new branch] gh/tugsbayasgalan/35/head -> origin/gh/tugsbayasgalan/35/head 2025-11-03T16:58:01.1697162Z * [new branch] gh/tugsbayasgalan/35/orig -> origin/gh/tugsbayasgalan/35/orig 2025-11-03T16:58:01.1699411Z * [new branch] gh/tugsbayasgalan/36/base -> origin/gh/tugsbayasgalan/36/base 2025-11-03T16:58:01.1701084Z * [new branch] gh/tugsbayasgalan/36/head -> origin/gh/tugsbayasgalan/36/head 2025-11-03T16:58:01.1702760Z * [new branch] gh/tugsbayasgalan/36/orig -> origin/gh/tugsbayasgalan/36/orig 2025-11-03T16:58:01.1705055Z * [new branch] gh/tugsbayasgalan/37/base -> origin/gh/tugsbayasgalan/37/base 2025-11-03T16:58:01.1706811Z * [new branch] gh/tugsbayasgalan/37/head -> origin/gh/tugsbayasgalan/37/head 2025-11-03T16:58:01.1708540Z * [new branch] gh/tugsbayasgalan/37/orig -> origin/gh/tugsbayasgalan/37/orig 2025-11-03T16:58:01.1711356Z * [new branch] gh/tugsbayasgalan/39/base -> origin/gh/tugsbayasgalan/39/base 2025-11-03T16:58:01.1712949Z * [new branch] gh/tugsbayasgalan/39/head -> origin/gh/tugsbayasgalan/39/head 2025-11-03T16:58:01.1714848Z * [new branch] gh/tugsbayasgalan/39/orig -> origin/gh/tugsbayasgalan/39/orig 2025-11-03T16:58:01.1717209Z * [new branch] gh/tugsbayasgalan/43/base -> origin/gh/tugsbayasgalan/43/base 2025-11-03T16:58:01.1718940Z * [new branch] gh/tugsbayasgalan/43/head -> origin/gh/tugsbayasgalan/43/head 2025-11-03T16:58:01.1720647Z * [new branch] gh/tugsbayasgalan/43/orig -> origin/gh/tugsbayasgalan/43/orig 2025-11-03T16:58:01.1723104Z * [new branch] gh/tugsbayasgalan/46/base -> origin/gh/tugsbayasgalan/46/base 2025-11-03T16:58:01.1724750Z * [new branch] gh/tugsbayasgalan/46/head -> origin/gh/tugsbayasgalan/46/head 2025-11-03T16:58:01.1726461Z * [new branch] gh/tugsbayasgalan/46/orig -> origin/gh/tugsbayasgalan/46/orig 2025-11-03T16:58:01.1728840Z * [new branch] gh/tugsbayasgalan/47/base -> origin/gh/tugsbayasgalan/47/base 2025-11-03T16:58:01.1730684Z * [new branch] gh/tugsbayasgalan/47/head -> origin/gh/tugsbayasgalan/47/head 2025-11-03T16:58:01.1732468Z * [new branch] gh/tugsbayasgalan/47/orig -> origin/gh/tugsbayasgalan/47/orig 2025-11-03T16:58:01.1734611Z * [new branch] gh/tugsbayasgalan/48/base -> origin/gh/tugsbayasgalan/48/base 2025-11-03T16:58:01.1736324Z * [new branch] gh/tugsbayasgalan/48/head -> origin/gh/tugsbayasgalan/48/head 2025-11-03T16:58:01.1737997Z * [new branch] gh/tugsbayasgalan/48/orig -> origin/gh/tugsbayasgalan/48/orig 2025-11-03T16:58:01.1740242Z * [new branch] gh/tugsbayasgalan/49/base -> origin/gh/tugsbayasgalan/49/base 2025-11-03T16:58:01.1742044Z * [new branch] gh/tugsbayasgalan/49/head -> origin/gh/tugsbayasgalan/49/head 2025-11-03T16:58:01.1743935Z * [new branch] gh/tugsbayasgalan/49/orig -> origin/gh/tugsbayasgalan/49/orig 2025-11-03T16:58:01.1746205Z * [new branch] gh/tugsbayasgalan/50/base -> origin/gh/tugsbayasgalan/50/base 2025-11-03T16:58:01.1748004Z * [new branch] gh/tugsbayasgalan/50/head -> origin/gh/tugsbayasgalan/50/head 2025-11-03T16:58:01.1749688Z * [new branch] gh/tugsbayasgalan/50/orig -> origin/gh/tugsbayasgalan/50/orig 2025-11-03T16:58:01.1752361Z * [new branch] gh/tugsbayasgalan/51/base -> origin/gh/tugsbayasgalan/51/base 2025-11-03T16:58:01.1754083Z * [new branch] gh/tugsbayasgalan/51/head -> origin/gh/tugsbayasgalan/51/head 2025-11-03T16:58:01.1755744Z * [new branch] gh/tugsbayasgalan/51/orig -> origin/gh/tugsbayasgalan/51/orig 2025-11-03T16:58:01.1758425Z * [new branch] gh/tugsbayasgalan/52/base -> origin/gh/tugsbayasgalan/52/base 2025-11-03T16:58:01.1760153Z * [new branch] gh/tugsbayasgalan/52/head -> origin/gh/tugsbayasgalan/52/head 2025-11-03T16:58:01.1761856Z * [new branch] gh/tugsbayasgalan/52/orig -> origin/gh/tugsbayasgalan/52/orig 2025-11-03T16:58:01.1764381Z * [new branch] gh/tugsbayasgalan/53/base -> origin/gh/tugsbayasgalan/53/base 2025-11-03T16:58:01.1766058Z * [new branch] gh/tugsbayasgalan/53/head -> origin/gh/tugsbayasgalan/53/head 2025-11-03T16:58:01.1767763Z * [new branch] gh/tugsbayasgalan/53/orig -> origin/gh/tugsbayasgalan/53/orig 2025-11-03T16:58:01.1770139Z * [new branch] gh/tugsbayasgalan/54/base -> origin/gh/tugsbayasgalan/54/base 2025-11-03T16:58:01.1771866Z * [new branch] gh/tugsbayasgalan/54/head -> origin/gh/tugsbayasgalan/54/head 2025-11-03T16:58:01.1773539Z * [new branch] gh/tugsbayasgalan/54/orig -> origin/gh/tugsbayasgalan/54/orig 2025-11-03T16:58:01.1776226Z * [new branch] gh/tugsbayasgalan/55/base -> origin/gh/tugsbayasgalan/55/base 2025-11-03T16:58:01.1777849Z * [new branch] gh/tugsbayasgalan/55/head -> origin/gh/tugsbayasgalan/55/head 2025-11-03T16:58:01.1779544Z * [new branch] gh/tugsbayasgalan/55/orig -> origin/gh/tugsbayasgalan/55/orig 2025-11-03T16:58:01.1785320Z * [new branch] gh/tugsbayasgalan/56/base -> origin/gh/tugsbayasgalan/56/base 2025-11-03T16:58:01.1786974Z * [new branch] gh/tugsbayasgalan/56/head -> origin/gh/tugsbayasgalan/56/head 2025-11-03T16:58:01.1788657Z * [new branch] gh/tugsbayasgalan/56/orig -> origin/gh/tugsbayasgalan/56/orig 2025-11-03T16:58:01.1792480Z * [new branch] gh/tugsbayasgalan/57/base -> origin/gh/tugsbayasgalan/57/base 2025-11-03T16:58:01.1793476Z * [new branch] gh/tugsbayasgalan/57/head -> origin/gh/tugsbayasgalan/57/head 2025-11-03T16:58:01.1795322Z * [new branch] gh/tugsbayasgalan/57/orig -> origin/gh/tugsbayasgalan/57/orig 2025-11-03T16:58:01.1797628Z * [new branch] gh/tugsbayasgalan/58/base -> origin/gh/tugsbayasgalan/58/base 2025-11-03T16:58:01.1799012Z * [new branch] gh/tugsbayasgalan/58/head -> origin/gh/tugsbayasgalan/58/head 2025-11-03T16:58:01.1800934Z * [new branch] gh/tugsbayasgalan/58/orig -> origin/gh/tugsbayasgalan/58/orig 2025-11-03T16:58:01.1803429Z * [new branch] gh/tugsbayasgalan/59/base -> origin/gh/tugsbayasgalan/59/base 2025-11-03T16:58:01.1805288Z * [new branch] gh/tugsbayasgalan/59/head -> origin/gh/tugsbayasgalan/59/head 2025-11-03T16:58:01.1806925Z * [new branch] gh/tugsbayasgalan/59/orig -> origin/gh/tugsbayasgalan/59/orig 2025-11-03T16:58:01.1809727Z * [new branch] gh/tugsbayasgalan/6/base -> origin/gh/tugsbayasgalan/6/base 2025-11-03T16:58:01.1812883Z * [new branch] gh/tugsbayasgalan/6/head -> origin/gh/tugsbayasgalan/6/head 2025-11-03T16:58:01.1814571Z * [new branch] gh/tugsbayasgalan/6/orig -> origin/gh/tugsbayasgalan/6/orig 2025-11-03T16:58:01.1816794Z * [new branch] gh/tugsbayasgalan/60/base -> origin/gh/tugsbayasgalan/60/base 2025-11-03T16:58:01.1818590Z * [new branch] gh/tugsbayasgalan/60/head -> origin/gh/tugsbayasgalan/60/head 2025-11-03T16:58:01.1820368Z * [new branch] gh/tugsbayasgalan/60/orig -> origin/gh/tugsbayasgalan/60/orig 2025-11-03T16:58:01.1823319Z * [new branch] gh/tugsbayasgalan/61/base -> origin/gh/tugsbayasgalan/61/base 2025-11-03T16:58:01.1824990Z * [new branch] gh/tugsbayasgalan/61/head -> origin/gh/tugsbayasgalan/61/head 2025-11-03T16:58:01.1826687Z * [new branch] gh/tugsbayasgalan/61/orig -> origin/gh/tugsbayasgalan/61/orig 2025-11-03T16:58:01.1829158Z * [new branch] gh/tugsbayasgalan/62/base -> origin/gh/tugsbayasgalan/62/base 2025-11-03T16:58:01.1830950Z * [new branch] gh/tugsbayasgalan/62/head -> origin/gh/tugsbayasgalan/62/head 2025-11-03T16:58:01.1832701Z * [new branch] gh/tugsbayasgalan/62/orig -> origin/gh/tugsbayasgalan/62/orig 2025-11-03T16:58:01.1834998Z * [new branch] gh/tugsbayasgalan/63/base -> origin/gh/tugsbayasgalan/63/base 2025-11-03T16:58:01.1836696Z * [new branch] gh/tugsbayasgalan/63/head -> origin/gh/tugsbayasgalan/63/head 2025-11-03T16:58:01.1838460Z * [new branch] gh/tugsbayasgalan/63/orig -> origin/gh/tugsbayasgalan/63/orig 2025-11-03T16:58:01.1840970Z * [new branch] gh/tugsbayasgalan/64/base -> origin/gh/tugsbayasgalan/64/base 2025-11-03T16:58:01.1842687Z * [new branch] gh/tugsbayasgalan/64/head -> origin/gh/tugsbayasgalan/64/head 2025-11-03T16:58:01.1844535Z * [new branch] gh/tugsbayasgalan/64/orig -> origin/gh/tugsbayasgalan/64/orig 2025-11-03T16:58:01.1847436Z * [new branch] gh/tugsbayasgalan/65/base -> origin/gh/tugsbayasgalan/65/base 2025-11-03T16:58:01.1849487Z * [new branch] gh/tugsbayasgalan/65/head -> origin/gh/tugsbayasgalan/65/head 2025-11-03T16:58:01.1851056Z * [new branch] gh/tugsbayasgalan/65/orig -> origin/gh/tugsbayasgalan/65/orig 2025-11-03T16:58:01.1853406Z * [new branch] gh/tugsbayasgalan/66/base -> origin/gh/tugsbayasgalan/66/base 2025-11-03T16:58:01.1855116Z * [new branch] gh/tugsbayasgalan/66/head -> origin/gh/tugsbayasgalan/66/head 2025-11-03T16:58:01.1856837Z * [new branch] gh/tugsbayasgalan/66/orig -> origin/gh/tugsbayasgalan/66/orig 2025-11-03T16:58:01.1859086Z * [new branch] gh/tugsbayasgalan/67/base -> origin/gh/tugsbayasgalan/67/base 2025-11-03T16:58:01.1860844Z * [new branch] gh/tugsbayasgalan/67/head -> origin/gh/tugsbayasgalan/67/head 2025-11-03T16:58:01.1862606Z * [new branch] gh/tugsbayasgalan/67/orig -> origin/gh/tugsbayasgalan/67/orig 2025-11-03T16:58:01.1865111Z * [new branch] gh/tugsbayasgalan/68/base -> origin/gh/tugsbayasgalan/68/base 2025-11-03T16:58:01.1866909Z * [new branch] gh/tugsbayasgalan/68/head -> origin/gh/tugsbayasgalan/68/head 2025-11-03T16:58:01.1868595Z * [new branch] gh/tugsbayasgalan/68/orig -> origin/gh/tugsbayasgalan/68/orig 2025-11-03T16:58:01.1870915Z * [new branch] gh/tugsbayasgalan/69/base -> origin/gh/tugsbayasgalan/69/base 2025-11-03T16:58:01.1872698Z * [new branch] gh/tugsbayasgalan/69/head -> origin/gh/tugsbayasgalan/69/head 2025-11-03T16:58:01.1874412Z * [new branch] gh/tugsbayasgalan/69/orig -> origin/gh/tugsbayasgalan/69/orig 2025-11-03T16:58:01.1876944Z * [new branch] gh/tugsbayasgalan/7/base -> origin/gh/tugsbayasgalan/7/base 2025-11-03T16:58:01.1879783Z * [new branch] gh/tugsbayasgalan/7/head -> origin/gh/tugsbayasgalan/7/head 2025-11-03T16:58:01.1880687Z * [new branch] gh/tugsbayasgalan/7/orig -> origin/gh/tugsbayasgalan/7/orig 2025-11-03T16:58:01.1883383Z * [new branch] gh/tugsbayasgalan/70/base -> origin/gh/tugsbayasgalan/70/base 2025-11-03T16:58:01.1884952Z * [new branch] gh/tugsbayasgalan/70/head -> origin/gh/tugsbayasgalan/70/head 2025-11-03T16:58:01.1887037Z * [new branch] gh/tugsbayasgalan/70/orig -> origin/gh/tugsbayasgalan/70/orig 2025-11-03T16:58:01.1889596Z * [new branch] gh/tugsbayasgalan/71/base -> origin/gh/tugsbayasgalan/71/base 2025-11-03T16:58:01.1891546Z * [new branch] gh/tugsbayasgalan/71/head -> origin/gh/tugsbayasgalan/71/head 2025-11-03T16:58:01.1893262Z * [new branch] gh/tugsbayasgalan/71/orig -> origin/gh/tugsbayasgalan/71/orig 2025-11-03T16:58:01.1895701Z * [new branch] gh/tugsbayasgalan/8/base -> origin/gh/tugsbayasgalan/8/base 2025-11-03T16:58:01.1897376Z * [new branch] gh/tugsbayasgalan/8/head -> origin/gh/tugsbayasgalan/8/head 2025-11-03T16:58:01.1899238Z * [new branch] gh/tugsbayasgalan/8/orig -> origin/gh/tugsbayasgalan/8/orig 2025-11-03T16:58:01.1901593Z * [new branch] gh/tugsbayasgalan/9/base -> origin/gh/tugsbayasgalan/9/base 2025-11-03T16:58:01.1903178Z * [new branch] gh/tugsbayasgalan/9/head -> origin/gh/tugsbayasgalan/9/head 2025-11-03T16:58:01.1904895Z * [new branch] gh/tugsbayasgalan/9/orig -> origin/gh/tugsbayasgalan/9/orig 2025-11-03T16:58:01.1907792Z * [new branch] gh/v0i0/13/base -> origin/gh/v0i0/13/base 2025-11-03T16:58:01.1909641Z * [new branch] gh/v0i0/13/head -> origin/gh/v0i0/13/head 2025-11-03T16:58:01.1911561Z * [new branch] gh/v0i0/13/orig -> origin/gh/v0i0/13/orig 2025-11-03T16:58:01.1914059Z * [new branch] gh/v0i0/14/base -> origin/gh/v0i0/14/base 2025-11-03T16:58:01.1915742Z * [new branch] gh/v0i0/14/head -> origin/gh/v0i0/14/head 2025-11-03T16:58:01.1917710Z * [new branch] gh/v0i0/14/orig -> origin/gh/v0i0/14/orig 2025-11-03T16:58:01.1920965Z * [new branch] gh/vishal9-team/1/base -> origin/gh/vishal9-team/1/base 2025-11-03T16:58:01.1922200Z * [new branch] gh/vishal9-team/1/head -> origin/gh/vishal9-team/1/head 2025-11-03T16:58:01.1924783Z * [new branch] gh/vishal9-team/2/base -> origin/gh/vishal9-team/2/base 2025-11-03T16:58:01.1926222Z * [new branch] gh/vishal9-team/2/head -> origin/gh/vishal9-team/2/head 2025-11-03T16:58:01.1927820Z * [new branch] gh/vishal9-team/2/orig -> origin/gh/vishal9-team/2/orig 2025-11-03T16:58:01.1930337Z * [new branch] gh/vishal9-team/3/base -> origin/gh/vishal9-team/3/base 2025-11-03T16:58:01.1931982Z * [new branch] gh/vishal9-team/3/head -> origin/gh/vishal9-team/3/head 2025-11-03T16:58:01.1933639Z * [new branch] gh/vishal9-team/3/orig -> origin/gh/vishal9-team/3/orig 2025-11-03T16:58:01.1936414Z * [new branch] gh/vkuzo/1/next -> origin/gh/vkuzo/1/next 2025-11-03T16:58:01.1938732Z * [new branch] gh/vkuzo/2/next -> origin/gh/vkuzo/2/next 2025-11-03T16:58:01.1941201Z * [new branch] gh/vkuzo/3/next -> origin/gh/vkuzo/3/next 2025-11-03T16:58:01.1944434Z * [new branch] gh/wconstab/424/base -> origin/gh/wconstab/424/base 2025-11-03T16:58:01.1946193Z * [new branch] gh/wconstab/424/head -> origin/gh/wconstab/424/head 2025-11-03T16:58:01.1947882Z * [new branch] gh/wconstab/424/orig -> origin/gh/wconstab/424/orig 2025-11-03T16:58:01.1950597Z * [new branch] gh/wconstab/435/base -> origin/gh/wconstab/435/base 2025-11-03T16:58:01.1952356Z * [new branch] gh/wconstab/435/head -> origin/gh/wconstab/435/head 2025-11-03T16:58:01.1954137Z * [new branch] gh/wconstab/435/orig -> origin/gh/wconstab/435/orig 2025-11-03T16:58:01.1956946Z * [new branch] gh/wconstab/438/base -> origin/gh/wconstab/438/base 2025-11-03T16:58:01.1958007Z * [new branch] gh/wconstab/438/head -> origin/gh/wconstab/438/head 2025-11-03T16:58:01.1960000Z * [new branch] gh/wconstab/438/orig -> origin/gh/wconstab/438/orig 2025-11-03T16:58:01.1962245Z * [new branch] gh/wconstab/444/base -> origin/gh/wconstab/444/base 2025-11-03T16:58:01.1963969Z * [new branch] gh/wconstab/444/head -> origin/gh/wconstab/444/head 2025-11-03T16:58:01.1965741Z * [new branch] gh/wconstab/444/orig -> origin/gh/wconstab/444/orig 2025-11-03T16:58:01.1968640Z * [new branch] gh/wconstab/447/base -> origin/gh/wconstab/447/base 2025-11-03T16:58:01.1970719Z * [new branch] gh/wconstab/447/head -> origin/gh/wconstab/447/head 2025-11-03T16:58:01.1972274Z * [new branch] gh/wconstab/447/orig -> origin/gh/wconstab/447/orig 2025-11-03T16:58:01.1975169Z * [new branch] gh/wconstab/448/base -> origin/gh/wconstab/448/base 2025-11-03T16:58:01.1976440Z * [new branch] gh/wconstab/448/head -> origin/gh/wconstab/448/head 2025-11-03T16:58:01.1978588Z * [new branch] gh/wconstab/448/orig -> origin/gh/wconstab/448/orig 2025-11-03T16:58:01.1981646Z * [new branch] gh/weifengpy/34/base -> origin/gh/weifengpy/34/base 2025-11-03T16:58:01.1983451Z * [new branch] gh/weifengpy/34/head -> origin/gh/weifengpy/34/head 2025-11-03T16:58:01.1985356Z * [new branch] gh/weifengpy/34/orig -> origin/gh/weifengpy/34/orig 2025-11-03T16:58:01.1987710Z * [new branch] gh/weifengpy/35/base -> origin/gh/weifengpy/35/base 2025-11-03T16:58:01.1989588Z * [new branch] gh/weifengpy/35/head -> origin/gh/weifengpy/35/head 2025-11-03T16:58:01.1991574Z * [new branch] gh/weifengpy/35/orig -> origin/gh/weifengpy/35/orig 2025-11-03T16:58:01.1993925Z * [new branch] gh/weifengpy/36/base -> origin/gh/weifengpy/36/base 2025-11-03T16:58:01.1995560Z * [new branch] gh/weifengpy/36/head -> origin/gh/weifengpy/36/head 2025-11-03T16:58:01.1997410Z * [new branch] gh/weifengpy/36/orig -> origin/gh/weifengpy/36/orig 2025-11-03T16:58:01.2000620Z * [new branch] gh/weifengpy/37/base -> origin/gh/weifengpy/37/base 2025-11-03T16:58:01.2001910Z * [new branch] gh/weifengpy/37/head -> origin/gh/weifengpy/37/head 2025-11-03T16:58:01.2003579Z * [new branch] gh/weifengpy/37/orig -> origin/gh/weifengpy/37/orig 2025-11-03T16:58:01.2006448Z * [new branch] gh/weifengpy/38/base -> origin/gh/weifengpy/38/base 2025-11-03T16:58:01.2007830Z * [new branch] gh/weifengpy/38/head -> origin/gh/weifengpy/38/head 2025-11-03T16:58:01.2010037Z * [new branch] gh/weifengpy/38/orig -> origin/gh/weifengpy/38/orig 2025-11-03T16:58:01.2012567Z * [new branch] gh/weifengpy/39/base -> origin/gh/weifengpy/39/base 2025-11-03T16:58:01.2014249Z * [new branch] gh/weifengpy/39/head -> origin/gh/weifengpy/39/head 2025-11-03T16:58:01.2016115Z * [new branch] gh/weifengpy/39/orig -> origin/gh/weifengpy/39/orig 2025-11-03T16:58:01.2019537Z * [new branch] gh/williamwen42/250/base -> origin/gh/williamwen42/250/base 2025-11-03T16:58:01.2020764Z * [new branch] gh/williamwen42/250/head -> origin/gh/williamwen42/250/head 2025-11-03T16:58:01.2022500Z * [new branch] gh/williamwen42/250/orig -> origin/gh/williamwen42/250/orig 2025-11-03T16:58:01.2025543Z * [new branch] gh/williamwen42/278/base -> origin/gh/williamwen42/278/base 2025-11-03T16:58:01.2026770Z * [new branch] gh/williamwen42/278/head -> origin/gh/williamwen42/278/head 2025-11-03T16:58:01.2028193Z * [new branch] gh/williamwen42/278/orig -> origin/gh/williamwen42/278/orig 2025-11-03T16:58:01.2030891Z * [new branch] gh/williamwen42/279/base -> origin/gh/williamwen42/279/base 2025-11-03T16:58:01.2032792Z * [new branch] gh/williamwen42/279/head -> origin/gh/williamwen42/279/head 2025-11-03T16:58:01.2034498Z * [new branch] gh/williamwen42/279/orig -> origin/gh/williamwen42/279/orig 2025-11-03T16:58:01.2036816Z * [new branch] gh/williamwen42/282/base -> origin/gh/williamwen42/282/base 2025-11-03T16:58:01.2038647Z * [new branch] gh/williamwen42/282/head -> origin/gh/williamwen42/282/head 2025-11-03T16:58:01.2040347Z * [new branch] gh/williamwen42/282/orig -> origin/gh/williamwen42/282/orig 2025-11-03T16:58:01.2042734Z * [new branch] gh/williamwen42/287/base -> origin/gh/williamwen42/287/base 2025-11-03T16:58:01.2044610Z * [new branch] gh/williamwen42/287/head -> origin/gh/williamwen42/287/head 2025-11-03T16:58:01.2046356Z * [new branch] gh/williamwen42/287/orig -> origin/gh/williamwen42/287/orig 2025-11-03T16:58:01.2048766Z * [new branch] gh/williamwen42/288/base -> origin/gh/williamwen42/288/base 2025-11-03T16:58:01.2050518Z * [new branch] gh/williamwen42/288/head -> origin/gh/williamwen42/288/head 2025-11-03T16:58:01.2052331Z * [new branch] gh/williamwen42/288/orig -> origin/gh/williamwen42/288/orig 2025-11-03T16:58:01.2055416Z * [new branch] gh/williamwen42/290/base -> origin/gh/williamwen42/290/base 2025-11-03T16:58:01.2057134Z * [new branch] gh/williamwen42/290/head -> origin/gh/williamwen42/290/head 2025-11-03T16:58:01.2058865Z * [new branch] gh/williamwen42/290/orig -> origin/gh/williamwen42/290/orig 2025-11-03T16:58:01.2061677Z * [new branch] gh/williamwen42/296/base -> origin/gh/williamwen42/296/base 2025-11-03T16:58:01.2063315Z * [new branch] gh/williamwen42/296/head -> origin/gh/williamwen42/296/head 2025-11-03T16:58:01.2065072Z * [new branch] gh/williamwen42/296/orig -> origin/gh/williamwen42/296/orig 2025-11-03T16:58:01.2067322Z * [new branch] gh/williamwen42/297/base -> origin/gh/williamwen42/297/base 2025-11-03T16:58:01.2068862Z * [new branch] gh/williamwen42/297/head -> origin/gh/williamwen42/297/head 2025-11-03T16:58:01.2070588Z * [new branch] gh/williamwen42/297/orig -> origin/gh/williamwen42/297/orig 2025-11-03T16:58:01.2072952Z * [new branch] gh/williamwen42/305/base -> origin/gh/williamwen42/305/base 2025-11-03T16:58:01.2074688Z * [new branch] gh/williamwen42/305/head -> origin/gh/williamwen42/305/head 2025-11-03T16:58:01.2076438Z * [new branch] gh/williamwen42/305/orig -> origin/gh/williamwen42/305/orig 2025-11-03T16:58:01.2078739Z * [new branch] gh/williamwen42/306/base -> origin/gh/williamwen42/306/base 2025-11-03T16:58:01.2080557Z * [new branch] gh/williamwen42/306/head -> origin/gh/williamwen42/306/head 2025-11-03T16:58:01.2082216Z * [new branch] gh/williamwen42/306/orig -> origin/gh/williamwen42/306/orig 2025-11-03T16:58:01.2084514Z * [new branch] gh/williamwen42/307/base -> origin/gh/williamwen42/307/base 2025-11-03T16:58:01.2086331Z * [new branch] gh/williamwen42/307/head -> origin/gh/williamwen42/307/head 2025-11-03T16:58:01.2088014Z * [new branch] gh/williamwen42/307/orig -> origin/gh/williamwen42/307/orig 2025-11-03T16:58:01.2090522Z * [new branch] gh/williamwen42/308/base -> origin/gh/williamwen42/308/base 2025-11-03T16:58:01.2092214Z * [new branch] gh/williamwen42/308/head -> origin/gh/williamwen42/308/head 2025-11-03T16:58:01.2093993Z * [new branch] gh/williamwen42/308/orig -> origin/gh/williamwen42/308/orig 2025-11-03T16:58:01.2096751Z * [new branch] gh/williamwen42/309/base -> origin/gh/williamwen42/309/base 2025-11-03T16:58:01.2099252Z * [new branch] gh/williamwen42/309/head -> origin/gh/williamwen42/309/head 2025-11-03T16:58:01.2100504Z * [new branch] gh/williamwen42/309/orig -> origin/gh/williamwen42/309/orig 2025-11-03T16:58:01.2103253Z * [new branch] gh/williamwen42/310/base -> origin/gh/williamwen42/310/base 2025-11-03T16:58:01.2104491Z * [new branch] gh/williamwen42/310/head -> origin/gh/williamwen42/310/head 2025-11-03T16:58:01.2106237Z * [new branch] gh/williamwen42/310/orig -> origin/gh/williamwen42/310/orig 2025-11-03T16:58:01.2109388Z * [new branch] gh/williamwen42/311/base -> origin/gh/williamwen42/311/base 2025-11-03T16:58:01.2111618Z * [new branch] gh/williamwen42/311/head -> origin/gh/williamwen42/311/head 2025-11-03T16:58:01.2113300Z * [new branch] gh/williamwen42/311/orig -> origin/gh/williamwen42/311/orig 2025-11-03T16:58:01.2116123Z * [new branch] gh/williamwen42/312/base -> origin/gh/williamwen42/312/base 2025-11-03T16:58:01.2118615Z * [new branch] gh/williamwen42/312/head -> origin/gh/williamwen42/312/head 2025-11-03T16:58:01.2119868Z * [new branch] gh/williamwen42/312/orig -> origin/gh/williamwen42/312/orig 2025-11-03T16:58:01.2122902Z * [new branch] gh/williamwen42/313/base -> origin/gh/williamwen42/313/base 2025-11-03T16:58:01.2124395Z * [new branch] gh/williamwen42/313/head -> origin/gh/williamwen42/313/head 2025-11-03T16:58:01.2126172Z * [new branch] gh/williamwen42/313/orig -> origin/gh/williamwen42/313/orig 2025-11-03T16:58:01.2128524Z * [new branch] gh/williamwen42/314/base -> origin/gh/williamwen42/314/base 2025-11-03T16:58:01.2130514Z * [new branch] gh/williamwen42/314/head -> origin/gh/williamwen42/314/head 2025-11-03T16:58:01.2132076Z * [new branch] gh/williamwen42/314/orig -> origin/gh/williamwen42/314/orig 2025-11-03T16:58:01.2134819Z * [new branch] gh/williamwen42/315/base -> origin/gh/williamwen42/315/base 2025-11-03T16:58:01.2136697Z * [new branch] gh/williamwen42/315/head -> origin/gh/williamwen42/315/head 2025-11-03T16:58:01.2138384Z * [new branch] gh/williamwen42/315/orig -> origin/gh/williamwen42/315/orig 2025-11-03T16:58:01.2140802Z * [new branch] gh/williamwen42/316/base -> origin/gh/williamwen42/316/base 2025-11-03T16:58:01.2142644Z * [new branch] gh/williamwen42/316/head -> origin/gh/williamwen42/316/head 2025-11-03T16:58:01.2144375Z * [new branch] gh/williamwen42/316/orig -> origin/gh/williamwen42/316/orig 2025-11-03T16:58:01.2146805Z * [new branch] gh/williamwen42/317/base -> origin/gh/williamwen42/317/base 2025-11-03T16:58:01.2148560Z * [new branch] gh/williamwen42/317/head -> origin/gh/williamwen42/317/head 2025-11-03T16:58:01.2150248Z * [new branch] gh/williamwen42/317/orig -> origin/gh/williamwen42/317/orig 2025-11-03T16:58:01.2152531Z * [new branch] gh/williamwen42/318/base -> origin/gh/williamwen42/318/base 2025-11-03T16:58:01.2154276Z * [new branch] gh/williamwen42/318/head -> origin/gh/williamwen42/318/head 2025-11-03T16:58:01.2156014Z * [new branch] gh/williamwen42/318/orig -> origin/gh/williamwen42/318/orig 2025-11-03T16:58:01.2158918Z * [new branch] gh/williamwen42/319/base -> origin/gh/williamwen42/319/base 2025-11-03T16:58:01.2160277Z * [new branch] gh/williamwen42/319/head -> origin/gh/williamwen42/319/head 2025-11-03T16:58:01.2161925Z * [new branch] gh/williamwen42/319/orig -> origin/gh/williamwen42/319/orig 2025-11-03T16:58:01.2164914Z * [new branch] gh/williamwen42/320/base -> origin/gh/williamwen42/320/base 2025-11-03T16:58:01.2166105Z * [new branch] gh/williamwen42/320/head -> origin/gh/williamwen42/320/head 2025-11-03T16:58:01.2167785Z * [new branch] gh/williamwen42/320/orig -> origin/gh/williamwen42/320/orig 2025-11-03T16:58:01.2170364Z * [new branch] gh/williamwen42/321/base -> origin/gh/williamwen42/321/base 2025-11-03T16:58:01.2172055Z * [new branch] gh/williamwen42/321/head -> origin/gh/williamwen42/321/head 2025-11-03T16:58:01.2173836Z * [new branch] gh/williamwen42/321/orig -> origin/gh/williamwen42/321/orig 2025-11-03T16:58:01.2176221Z * [new branch] gh/williamwen42/322/base -> origin/gh/williamwen42/322/base 2025-11-03T16:58:01.2178599Z * [new branch] gh/williamwen42/322/head -> origin/gh/williamwen42/322/head 2025-11-03T16:58:01.2179809Z * [new branch] gh/williamwen42/322/orig -> origin/gh/williamwen42/322/orig 2025-11-03T16:58:01.2182754Z * [new branch] gh/williamwen42/323/base -> origin/gh/williamwen42/323/base 2025-11-03T16:58:01.2184071Z * [new branch] gh/williamwen42/323/head -> origin/gh/williamwen42/323/head 2025-11-03T16:58:01.2185796Z * [new branch] gh/williamwen42/323/orig -> origin/gh/williamwen42/323/orig 2025-11-03T16:58:01.2188175Z * [new branch] gh/williamwen42/324/base -> origin/gh/williamwen42/324/base 2025-11-03T16:58:01.2190024Z * [new branch] gh/williamwen42/324/head -> origin/gh/williamwen42/324/head 2025-11-03T16:58:01.2191890Z * [new branch] gh/williamwen42/324/orig -> origin/gh/williamwen42/324/orig 2025-11-03T16:58:01.2194283Z * [new branch] gh/williamwen42/325/base -> origin/gh/williamwen42/325/base 2025-11-03T16:58:01.2195985Z * [new branch] gh/williamwen42/325/head -> origin/gh/williamwen42/325/head 2025-11-03T16:58:01.2197840Z * [new branch] gh/williamwen42/325/orig -> origin/gh/williamwen42/325/orig 2025-11-03T16:58:01.2200639Z * [new branch] gh/williamwen42/326/base -> origin/gh/williamwen42/326/base 2025-11-03T16:58:01.2201919Z * [new branch] gh/williamwen42/326/head -> origin/gh/williamwen42/326/head 2025-11-03T16:58:01.2203641Z * [new branch] gh/williamwen42/326/orig -> origin/gh/williamwen42/326/orig 2025-11-03T16:58:01.2206492Z * [new branch] gh/williamwen42/327/base -> origin/gh/williamwen42/327/base 2025-11-03T16:58:01.2207856Z * [new branch] gh/williamwen42/327/head -> origin/gh/williamwen42/327/head 2025-11-03T16:58:01.2209759Z * [new branch] gh/williamwen42/327/orig -> origin/gh/williamwen42/327/orig 2025-11-03T16:58:01.2212410Z * [new branch] gh/williamwen42/328/base -> origin/gh/williamwen42/328/base 2025-11-03T16:58:01.2214173Z * [new branch] gh/williamwen42/328/head -> origin/gh/williamwen42/328/head 2025-11-03T16:58:01.2215934Z * [new branch] gh/williamwen42/328/orig -> origin/gh/williamwen42/328/orig 2025-11-03T16:58:01.2218324Z * [new branch] gh/williamwen42/329/base -> origin/gh/williamwen42/329/base 2025-11-03T16:58:01.2220198Z * [new branch] gh/williamwen42/329/head -> origin/gh/williamwen42/329/head 2025-11-03T16:58:01.2221882Z * [new branch] gh/williamwen42/329/orig -> origin/gh/williamwen42/329/orig 2025-11-03T16:58:01.2224603Z * [new branch] gh/xmfan/169/base -> origin/gh/xmfan/169/base 2025-11-03T16:58:01.2226307Z * [new branch] gh/xmfan/169/head -> origin/gh/xmfan/169/head 2025-11-03T16:58:01.2228528Z * [new branch] gh/xmfan/170/base -> origin/gh/xmfan/170/base 2025-11-03T16:58:01.2230229Z * [new branch] gh/xmfan/170/head -> origin/gh/xmfan/170/head 2025-11-03T16:58:01.2232580Z * [new branch] gh/xmfan/274/base -> origin/gh/xmfan/274/base 2025-11-03T16:58:01.2234356Z * [new branch] gh/xmfan/274/head -> origin/gh/xmfan/274/head 2025-11-03T16:58:01.2236114Z * [new branch] gh/xmfan/274/orig -> origin/gh/xmfan/274/orig 2025-11-03T16:58:01.2238470Z * [new branch] gh/xmfan/277/base -> origin/gh/xmfan/277/base 2025-11-03T16:58:01.2240612Z * [new branch] gh/xmfan/277/head -> origin/gh/xmfan/277/head 2025-11-03T16:58:01.2242343Z * [new branch] gh/xmfan/277/orig -> origin/gh/xmfan/277/orig 2025-11-03T16:58:01.2244745Z * [new branch] gh/xmfan/289/base -> origin/gh/xmfan/289/base 2025-11-03T16:58:01.2246560Z * [new branch] gh/xmfan/289/head -> origin/gh/xmfan/289/head 2025-11-03T16:58:01.2248284Z * [new branch] gh/xmfan/289/orig -> origin/gh/xmfan/289/orig 2025-11-03T16:58:01.2250888Z * [new branch] gh/xmfan/291/base -> origin/gh/xmfan/291/base 2025-11-03T16:58:01.2252627Z * [new branch] gh/xmfan/291/head -> origin/gh/xmfan/291/head 2025-11-03T16:58:01.2254438Z * [new branch] gh/xmfan/291/orig -> origin/gh/xmfan/291/orig 2025-11-03T16:58:01.2256857Z * [new branch] gh/xmfan/292/base -> origin/gh/xmfan/292/base 2025-11-03T16:58:01.2258692Z * [new branch] gh/xmfan/292/head -> origin/gh/xmfan/292/head 2025-11-03T16:58:01.2260588Z * [new branch] gh/xmfan/292/orig -> origin/gh/xmfan/292/orig 2025-11-03T16:58:01.2263022Z * [new branch] gh/xmfan/295/base -> origin/gh/xmfan/295/base 2025-11-03T16:58:01.2265041Z * [new branch] gh/xmfan/295/head -> origin/gh/xmfan/295/head 2025-11-03T16:58:01.2266778Z * [new branch] gh/xmfan/295/orig -> origin/gh/xmfan/295/orig 2025-11-03T16:58:01.2269253Z * [new branch] gh/xmfan/296/base -> origin/gh/xmfan/296/base 2025-11-03T16:58:01.2270903Z * [new branch] gh/xmfan/296/head -> origin/gh/xmfan/296/head 2025-11-03T16:58:01.2272583Z * [new branch] gh/xmfan/296/orig -> origin/gh/xmfan/296/orig 2025-11-03T16:58:01.2275030Z * [new branch] gh/xmfan/297/base -> origin/gh/xmfan/297/base 2025-11-03T16:58:01.2276887Z * [new branch] gh/xmfan/297/head -> origin/gh/xmfan/297/head 2025-11-03T16:58:01.2278668Z * [new branch] gh/xmfan/297/orig -> origin/gh/xmfan/297/orig 2025-11-03T16:58:01.2281035Z * [new branch] gh/xmfan/298/base -> origin/gh/xmfan/298/base 2025-11-03T16:58:01.2282814Z * [new branch] gh/xmfan/298/head -> origin/gh/xmfan/298/head 2025-11-03T16:58:01.2284588Z * [new branch] gh/xmfan/298/orig -> origin/gh/xmfan/298/orig 2025-11-03T16:58:01.2286868Z * [new branch] gh/xmfan/299/base -> origin/gh/xmfan/299/base 2025-11-03T16:58:01.2288775Z * [new branch] gh/xmfan/299/head -> origin/gh/xmfan/299/head 2025-11-03T16:58:01.2290500Z * [new branch] gh/xmfan/299/orig -> origin/gh/xmfan/299/orig 2025-11-03T16:58:01.2292731Z * [new branch] gh/xmfan/300/base -> origin/gh/xmfan/300/base 2025-11-03T16:58:01.2294446Z * [new branch] gh/xmfan/300/head -> origin/gh/xmfan/300/head 2025-11-03T16:58:01.2296153Z * [new branch] gh/xmfan/300/orig -> origin/gh/xmfan/300/orig 2025-11-03T16:58:01.2298474Z * [new branch] gh/xmfan/301/base -> origin/gh/xmfan/301/base 2025-11-03T16:58:01.2300160Z * [new branch] gh/xmfan/301/head -> origin/gh/xmfan/301/head 2025-11-03T16:58:01.2301964Z * [new branch] gh/xmfan/301/orig -> origin/gh/xmfan/301/orig 2025-11-03T16:58:01.2304289Z * [new branch] gh/xmfan/302/base -> origin/gh/xmfan/302/base 2025-11-03T16:58:01.2306010Z * [new branch] gh/xmfan/302/head -> origin/gh/xmfan/302/head 2025-11-03T16:58:01.2307803Z * [new branch] gh/xmfan/302/orig -> origin/gh/xmfan/302/orig 2025-11-03T16:58:01.2310944Z * [new branch] gh/xmfan/303/base -> origin/gh/xmfan/303/base 2025-11-03T16:58:01.2312368Z * [new branch] gh/xmfan/303/head -> origin/gh/xmfan/303/head 2025-11-03T16:58:01.2314086Z * [new branch] gh/xmfan/303/orig -> origin/gh/xmfan/303/orig 2025-11-03T16:58:01.2317457Z * [new branch] gh/xmfan/304/base -> origin/gh/xmfan/304/base 2025-11-03T16:58:01.2318440Z * [new branch] gh/xmfan/304/head -> origin/gh/xmfan/304/head 2025-11-03T16:58:01.2320357Z * [new branch] gh/xmfan/304/orig -> origin/gh/xmfan/304/orig 2025-11-03T16:58:01.2322661Z * [new branch] gh/xmfan/305/base -> origin/gh/xmfan/305/base 2025-11-03T16:58:01.2324373Z * [new branch] gh/xmfan/305/head -> origin/gh/xmfan/305/head 2025-11-03T16:58:01.2326210Z * [new branch] gh/xmfan/305/orig -> origin/gh/xmfan/305/orig 2025-11-03T16:58:01.2328756Z * [new branch] gh/xmfan/306/base -> origin/gh/xmfan/306/base 2025-11-03T16:58:01.2330664Z * [new branch] gh/xmfan/306/head -> origin/gh/xmfan/306/head 2025-11-03T16:58:01.2332360Z * [new branch] gh/xmfan/306/orig -> origin/gh/xmfan/306/orig 2025-11-03T16:58:01.2334763Z * [new branch] gh/xmfan/307/base -> origin/gh/xmfan/307/base 2025-11-03T16:58:01.2336880Z * [new branch] gh/xmfan/307/head -> origin/gh/xmfan/307/head 2025-11-03T16:58:01.2338609Z * [new branch] gh/xmfan/307/orig -> origin/gh/xmfan/307/orig 2025-11-03T16:58:01.2341722Z * [new branch] gh/xmfan/308/base -> origin/gh/xmfan/308/base 2025-11-03T16:58:01.2343571Z * [new branch] gh/xmfan/308/head -> origin/gh/xmfan/308/head 2025-11-03T16:58:01.2345305Z * [new branch] gh/xmfan/308/orig -> origin/gh/xmfan/308/orig 2025-11-03T16:58:01.2347696Z * [new branch] gh/xmfan/309/base -> origin/gh/xmfan/309/base 2025-11-03T16:58:01.2349607Z * [new branch] gh/xmfan/309/head -> origin/gh/xmfan/309/head 2025-11-03T16:58:01.2351336Z * [new branch] gh/xmfan/309/orig -> origin/gh/xmfan/309/orig 2025-11-03T16:58:01.2354214Z * [new branch] gh/xuanzhang816/22/base -> origin/gh/xuanzhang816/22/base 2025-11-03T16:58:01.2356046Z * [new branch] gh/xuanzhang816/22/head -> origin/gh/xuanzhang816/22/head 2025-11-03T16:58:01.2357498Z * [new branch] gh/xuanzhang816/22/orig -> origin/gh/xuanzhang816/22/orig 2025-11-03T16:58:01.2360460Z * [new branch] gh/xuanzhang816/23/base -> origin/gh/xuanzhang816/23/base 2025-11-03T16:58:01.2361707Z * [new branch] gh/xuanzhang816/23/head -> origin/gh/xuanzhang816/23/head 2025-11-03T16:58:01.2363411Z * [new branch] gh/xuanzhang816/23/orig -> origin/gh/xuanzhang816/23/orig 2025-11-03T16:58:01.2365992Z * [new branch] gh/xuanzhang816/27/base -> origin/gh/xuanzhang816/27/base 2025-11-03T16:58:01.2367390Z * [new branch] gh/xuanzhang816/27/head -> origin/gh/xuanzhang816/27/head 2025-11-03T16:58:01.2369132Z * [new branch] gh/xuanzhang816/27/orig -> origin/gh/xuanzhang816/27/orig 2025-11-03T16:58:01.2371605Z * [new branch] gh/xuanzhang816/32/base -> origin/gh/xuanzhang816/32/base 2025-11-03T16:58:01.2373368Z * [new branch] gh/xuanzhang816/32/head -> origin/gh/xuanzhang816/32/head 2025-11-03T16:58:01.2375110Z * [new branch] gh/xuanzhang816/32/orig -> origin/gh/xuanzhang816/32/orig 2025-11-03T16:58:01.2377431Z * [new branch] gh/xuanzhang816/33/base -> origin/gh/xuanzhang816/33/base 2025-11-03T16:58:01.2379151Z * [new branch] gh/xuanzhang816/33/head -> origin/gh/xuanzhang816/33/head 2025-11-03T16:58:01.2380852Z * [new branch] gh/xuanzhang816/33/orig -> origin/gh/xuanzhang816/33/orig 2025-11-03T16:58:01.2383682Z * [new branch] gh/yanbing-j/11/base -> origin/gh/yanbing-j/11/base 2025-11-03T16:58:01.2385446Z * [new branch] gh/yanbing-j/11/head -> origin/gh/yanbing-j/11/head 2025-11-03T16:58:01.2387203Z * [new branch] gh/yanbing-j/11/orig -> origin/gh/yanbing-j/11/orig 2025-11-03T16:58:01.2389669Z * [new branch] gh/yanbing-j/12/base -> origin/gh/yanbing-j/12/base 2025-11-03T16:58:01.2391383Z * [new branch] gh/yanbing-j/12/head -> origin/gh/yanbing-j/12/head 2025-11-03T16:58:01.2393020Z * [new branch] gh/yanbing-j/12/orig -> origin/gh/yanbing-j/12/orig 2025-11-03T16:58:01.2395426Z * [new branch] gh/yanbing-j/13/base -> origin/gh/yanbing-j/13/base 2025-11-03T16:58:01.2397261Z * [new branch] gh/yanbing-j/13/head -> origin/gh/yanbing-j/13/head 2025-11-03T16:58:01.2398880Z * [new branch] gh/yanbing-j/13/orig -> origin/gh/yanbing-j/13/orig 2025-11-03T16:58:01.2401305Z * [new branch] gh/yanbing-j/14/base -> origin/gh/yanbing-j/14/base 2025-11-03T16:58:01.2402999Z * [new branch] gh/yanbing-j/14/head -> origin/gh/yanbing-j/14/head 2025-11-03T16:58:01.2404691Z * [new branch] gh/yanbing-j/14/orig -> origin/gh/yanbing-j/14/orig 2025-11-03T16:58:01.2406909Z * [new branch] gh/yanbing-j/15/base -> origin/gh/yanbing-j/15/base 2025-11-03T16:58:01.2408764Z * [new branch] gh/yanbing-j/15/head -> origin/gh/yanbing-j/15/head 2025-11-03T16:58:01.2411031Z * [new branch] gh/yanbing-j/15/orig -> origin/gh/yanbing-j/15/orig 2025-11-03T16:58:01.2413101Z * [new branch] gh/yanbing-j/18/base -> origin/gh/yanbing-j/18/base 2025-11-03T16:58:01.2414827Z * [new branch] gh/yanbing-j/18/head -> origin/gh/yanbing-j/18/head 2025-11-03T16:58:01.2416609Z * [new branch] gh/yanbing-j/18/orig -> origin/gh/yanbing-j/18/orig 2025-11-03T16:58:01.2419033Z * [new branch] gh/yanbing-j/19/base -> origin/gh/yanbing-j/19/base 2025-11-03T16:58:01.2420754Z * [new branch] gh/yanbing-j/19/head -> origin/gh/yanbing-j/19/head 2025-11-03T16:58:01.2422468Z * [new branch] gh/yanbing-j/19/orig -> origin/gh/yanbing-j/19/orig 2025-11-03T16:58:01.2424750Z * [new branch] gh/yanbing-j/20/base -> origin/gh/yanbing-j/20/base 2025-11-03T16:58:01.2426454Z * [new branch] gh/yanbing-j/20/head -> origin/gh/yanbing-j/20/head 2025-11-03T16:58:01.2428213Z * [new branch] gh/yanbing-j/20/orig -> origin/gh/yanbing-j/20/orig 2025-11-03T16:58:01.2430445Z * [new branch] gh/yanbing-j/21/base -> origin/gh/yanbing-j/21/base 2025-11-03T16:58:01.2432195Z * [new branch] gh/yanbing-j/21/head -> origin/gh/yanbing-j/21/head 2025-11-03T16:58:01.2434489Z * [new branch] gh/yanbing-j/22/base -> origin/gh/yanbing-j/22/base 2025-11-03T16:58:01.2436222Z * [new branch] gh/yanbing-j/22/head -> origin/gh/yanbing-j/22/head 2025-11-03T16:58:01.2437967Z * [new branch] gh/yanbing-j/22/orig -> origin/gh/yanbing-j/22/orig 2025-11-03T16:58:01.2440293Z * [new branch] gh/yanbing-j/23/base -> origin/gh/yanbing-j/23/base 2025-11-03T16:58:01.2442082Z * [new branch] gh/yanbing-j/23/head -> origin/gh/yanbing-j/23/head 2025-11-03T16:58:01.2443782Z * [new branch] gh/yanbing-j/23/orig -> origin/gh/yanbing-j/23/orig 2025-11-03T16:58:01.2446072Z * [new branch] gh/yanbing-j/24/base -> origin/gh/yanbing-j/24/base 2025-11-03T16:58:01.2447858Z * [new branch] gh/yanbing-j/24/head -> origin/gh/yanbing-j/24/head 2025-11-03T16:58:01.2449972Z * [new branch] gh/yanbing-j/24/orig -> origin/gh/yanbing-j/24/orig 2025-11-03T16:58:01.2452535Z * [new branch] gh/yanbing-j/25/base -> origin/gh/yanbing-j/25/base 2025-11-03T16:58:01.2453909Z * [new branch] gh/yanbing-j/25/head -> origin/gh/yanbing-j/25/head 2025-11-03T16:58:01.2455550Z * [new branch] gh/yanbing-j/25/orig -> origin/gh/yanbing-j/25/orig 2025-11-03T16:58:01.2458133Z * [new branch] gh/yanbing-j/26/base -> origin/gh/yanbing-j/26/base 2025-11-03T16:58:01.2459727Z * [new branch] gh/yanbing-j/26/head -> origin/gh/yanbing-j/26/head 2025-11-03T16:58:01.2461448Z * [new branch] gh/yanbing-j/26/orig -> origin/gh/yanbing-j/26/orig 2025-11-03T16:58:01.2464454Z * [new branch] gh/yangw-dev/12/base -> origin/gh/yangw-dev/12/base 2025-11-03T16:58:01.2466150Z * [new branch] gh/yangw-dev/12/head -> origin/gh/yangw-dev/12/head 2025-11-03T16:58:01.2468115Z * [new branch] gh/yangw-dev/12/orig -> origin/gh/yangw-dev/12/orig 2025-11-03T16:58:01.2470450Z * [new branch] gh/yangw-dev/13/base -> origin/gh/yangw-dev/13/base 2025-11-03T16:58:01.2472277Z * [new branch] gh/yangw-dev/13/head -> origin/gh/yangw-dev/13/head 2025-11-03T16:58:01.2473953Z * [new branch] gh/yangw-dev/13/orig -> origin/gh/yangw-dev/13/orig 2025-11-03T16:58:01.2476266Z * [new branch] gh/yangw-dev/14/base -> origin/gh/yangw-dev/14/base 2025-11-03T16:58:01.2477979Z * [new branch] gh/yangw-dev/14/head -> origin/gh/yangw-dev/14/head 2025-11-03T16:58:01.2479852Z * [new branch] gh/yangw-dev/14/orig -> origin/gh/yangw-dev/14/orig 2025-11-03T16:58:01.2496239Z * [new branch] gh/yangw-dev/15/base -> origin/gh/yangw-dev/15/base 2025-11-03T16:58:01.2496610Z * [new branch] gh/yangw-dev/15/head -> origin/gh/yangw-dev/15/head 2025-11-03T16:58:01.2496821Z * [new branch] gh/yangw-dev/15/orig -> origin/gh/yangw-dev/15/orig 2025-11-03T16:58:01.2496989Z * [new branch] gh/yangw-dev/19/base -> origin/gh/yangw-dev/19/base 2025-11-03T16:58:01.2497145Z * [new branch] gh/yangw-dev/19/head -> origin/gh/yangw-dev/19/head 2025-11-03T16:58:01.2497290Z * [new branch] gh/yangw-dev/19/orig -> origin/gh/yangw-dev/19/orig 2025-11-03T16:58:01.2497430Z * [new branch] gh/yangw-dev/26/base -> origin/gh/yangw-dev/26/base 2025-11-03T16:58:01.2497596Z * [new branch] gh/yangw-dev/26/head -> origin/gh/yangw-dev/26/head 2025-11-03T16:58:01.2498116Z * [new branch] gh/yangw-dev/26/orig -> origin/gh/yangw-dev/26/orig 2025-11-03T16:58:01.2500291Z * [new branch] gh/yangw-dev/27/base -> origin/gh/yangw-dev/27/base 2025-11-03T16:58:01.2502102Z * [new branch] gh/yangw-dev/27/head -> origin/gh/yangw-dev/27/head 2025-11-03T16:58:01.2503832Z * [new branch] gh/yangw-dev/27/orig -> origin/gh/yangw-dev/27/orig 2025-11-03T16:58:01.2506812Z * [new branch] gh/ydwu4/283/base -> origin/gh/ydwu4/283/base 2025-11-03T16:58:01.2509148Z * [new branch] gh/ydwu4/283/head -> origin/gh/ydwu4/283/head 2025-11-03T16:58:01.2510534Z * [new branch] gh/ydwu4/283/orig -> origin/gh/ydwu4/283/orig 2025-11-03T16:58:01.2513514Z * [new branch] gh/ydwu4/292/base -> origin/gh/ydwu4/292/base 2025-11-03T16:58:01.2514768Z * [new branch] gh/ydwu4/292/head -> origin/gh/ydwu4/292/head 2025-11-03T16:58:01.2516464Z * [new branch] gh/ydwu4/292/orig -> origin/gh/ydwu4/292/orig 2025-11-03T16:58:01.2518728Z * [new branch] gh/ydwu4/294/base -> origin/gh/ydwu4/294/base 2025-11-03T16:58:01.2520422Z * [new branch] gh/ydwu4/294/head -> origin/gh/ydwu4/294/head 2025-11-03T16:58:01.2522120Z * [new branch] gh/ydwu4/294/orig -> origin/gh/ydwu4/294/orig 2025-11-03T16:58:01.2524545Z * [new branch] gh/ydwu4/295/base -> origin/gh/ydwu4/295/base 2025-11-03T16:58:01.2526351Z * [new branch] gh/ydwu4/295/head -> origin/gh/ydwu4/295/head 2025-11-03T16:58:01.2528102Z * [new branch] gh/ydwu4/295/orig -> origin/gh/ydwu4/295/orig 2025-11-03T16:58:01.2530478Z * [new branch] gh/ydwu4/296/base -> origin/gh/ydwu4/296/base 2025-11-03T16:58:01.2532341Z * [new branch] gh/ydwu4/296/head -> origin/gh/ydwu4/296/head 2025-11-03T16:58:01.2533995Z * [new branch] gh/ydwu4/296/orig -> origin/gh/ydwu4/296/orig 2025-11-03T16:58:01.2536504Z * [new branch] gh/ydwu4/306/base -> origin/gh/ydwu4/306/base 2025-11-03T16:58:01.2538274Z * [new branch] gh/ydwu4/306/head -> origin/gh/ydwu4/306/head 2025-11-03T16:58:01.2540015Z * [new branch] gh/ydwu4/306/orig -> origin/gh/ydwu4/306/orig 2025-11-03T16:58:01.2542392Z * [new branch] gh/ydwu4/312/base -> origin/gh/ydwu4/312/base 2025-11-03T16:58:01.2544118Z * [new branch] gh/ydwu4/312/head -> origin/gh/ydwu4/312/head 2025-11-03T16:58:01.2545879Z * [new branch] gh/ydwu4/312/orig -> origin/gh/ydwu4/312/orig 2025-11-03T16:58:01.2548697Z * [new branch] gh/ydwu4/322/base -> origin/gh/ydwu4/322/base 2025-11-03T16:58:01.2550462Z * [new branch] gh/ydwu4/322/head -> origin/gh/ydwu4/322/head 2025-11-03T16:58:01.2551854Z * [new branch] gh/ydwu4/322/orig -> origin/gh/ydwu4/322/orig 2025-11-03T16:58:01.2554570Z * [new branch] gh/ydwu4/326/base -> origin/gh/ydwu4/326/base 2025-11-03T16:58:01.2555857Z * [new branch] gh/ydwu4/326/head -> origin/gh/ydwu4/326/head 2025-11-03T16:58:01.2557572Z * [new branch] gh/ydwu4/326/orig -> origin/gh/ydwu4/326/orig 2025-11-03T16:58:01.2560258Z * [new branch] gh/ydwu4/327/base -> origin/gh/ydwu4/327/base 2025-11-03T16:58:01.2562042Z * [new branch] gh/ydwu4/327/head -> origin/gh/ydwu4/327/head 2025-11-03T16:58:01.2563739Z * [new branch] gh/ydwu4/327/orig -> origin/gh/ydwu4/327/orig 2025-11-03T16:58:01.2566208Z * [new branch] gh/ydwu4/328/base -> origin/gh/ydwu4/328/base 2025-11-03T16:58:01.2567938Z * [new branch] gh/ydwu4/328/head -> origin/gh/ydwu4/328/head 2025-11-03T16:58:01.2569825Z * [new branch] gh/ydwu4/328/orig -> origin/gh/ydwu4/328/orig 2025-11-03T16:58:01.2572014Z * [new branch] gh/ydwu4/329/base -> origin/gh/ydwu4/329/base 2025-11-03T16:58:01.2573788Z * [new branch] gh/ydwu4/329/head -> origin/gh/ydwu4/329/head 2025-11-03T16:58:01.2575946Z * [new branch] gh/ydwu4/329/orig -> origin/gh/ydwu4/329/orig 2025-11-03T16:58:01.2578782Z * [new branch] gh/ydwu4/330/base -> origin/gh/ydwu4/330/base 2025-11-03T16:58:01.2580608Z * [new branch] gh/ydwu4/330/head -> origin/gh/ydwu4/330/head 2025-11-03T16:58:01.2582203Z * [new branch] gh/ydwu4/330/orig -> origin/gh/ydwu4/330/orig 2025-11-03T16:58:01.2584434Z * [new branch] gh/ydwu4/331/base -> origin/gh/ydwu4/331/base 2025-11-03T16:58:01.2586177Z * [new branch] gh/ydwu4/331/head -> origin/gh/ydwu4/331/head 2025-11-03T16:58:01.2587825Z * [new branch] gh/ydwu4/331/orig -> origin/gh/ydwu4/331/orig 2025-11-03T16:58:01.2590657Z * [new branch] gh/ydwu4/332/base -> origin/gh/ydwu4/332/base 2025-11-03T16:58:01.2591971Z * [new branch] gh/ydwu4/332/head -> origin/gh/ydwu4/332/head 2025-11-03T16:58:01.2593645Z * [new branch] gh/ydwu4/332/orig -> origin/gh/ydwu4/332/orig 2025-11-03T16:58:01.2596227Z * [new branch] gh/ydwu4/333/base -> origin/gh/ydwu4/333/base 2025-11-03T16:58:01.2597673Z * [new branch] gh/ydwu4/333/head -> origin/gh/ydwu4/333/head 2025-11-03T16:58:01.2599315Z * [new branch] gh/ydwu4/333/orig -> origin/gh/ydwu4/333/orig 2025-11-03T16:58:01.2602003Z * [new branch] gh/ydwu4/334/base -> origin/gh/ydwu4/334/base 2025-11-03T16:58:01.2603947Z * [new branch] gh/ydwu4/334/head -> origin/gh/ydwu4/334/head 2025-11-03T16:58:01.2605688Z * [new branch] gh/ydwu4/334/orig -> origin/gh/ydwu4/334/orig 2025-11-03T16:58:01.2608649Z * [new branch] gh/ydwu4/335/base -> origin/gh/ydwu4/335/base 2025-11-03T16:58:01.2610358Z * [new branch] gh/ydwu4/335/head -> origin/gh/ydwu4/335/head 2025-11-03T16:58:01.2613514Z * [new branch] gh/ydwu4/335/orig -> origin/gh/ydwu4/335/orig 2025-11-03T16:58:01.2615382Z * [new branch] gh/ydwu4/336/base -> origin/gh/ydwu4/336/base 2025-11-03T16:58:01.2617138Z * [new branch] gh/ydwu4/336/head -> origin/gh/ydwu4/336/head 2025-11-03T16:58:01.2618745Z * [new branch] gh/ydwu4/336/orig -> origin/gh/ydwu4/336/orig 2025-11-03T16:58:01.2621575Z * [new branch] gh/ydwu4/337/base -> origin/gh/ydwu4/337/base 2025-11-03T16:58:01.2623502Z * [new branch] gh/ydwu4/337/head -> origin/gh/ydwu4/337/head 2025-11-03T16:58:01.2625074Z * [new branch] gh/ydwu4/337/orig -> origin/gh/ydwu4/337/orig 2025-11-03T16:58:01.2627938Z * [new branch] gh/ydwu4/338/base -> origin/gh/ydwu4/338/base 2025-11-03T16:58:01.2629765Z * [new branch] gh/ydwu4/338/head -> origin/gh/ydwu4/338/head 2025-11-03T16:58:01.2631554Z * [new branch] gh/ydwu4/338/orig -> origin/gh/ydwu4/338/orig 2025-11-03T16:58:01.2634065Z * [new branch] gh/ydwu4/339/base -> origin/gh/ydwu4/339/base 2025-11-03T16:58:01.2635797Z * [new branch] gh/ydwu4/339/head -> origin/gh/ydwu4/339/head 2025-11-03T16:58:01.2637961Z * [new branch] gh/ydwu4/339/orig -> origin/gh/ydwu4/339/orig 2025-11-03T16:58:01.2640009Z * [new branch] gh/ydwu4/340/base -> origin/gh/ydwu4/340/base 2025-11-03T16:58:01.2641743Z * [new branch] gh/ydwu4/340/head -> origin/gh/ydwu4/340/head 2025-11-03T16:58:01.2643638Z * [new branch] gh/ydwu4/340/orig -> origin/gh/ydwu4/340/orig 2025-11-03T16:58:01.2646893Z * [new branch] gh/yf225/133/base -> origin/gh/yf225/133/base 2025-11-03T16:58:01.2647968Z * [new branch] gh/yf225/133/head -> origin/gh/yf225/133/head 2025-11-03T16:58:01.2651136Z * [new branch] gh/yf225/93/base -> origin/gh/yf225/93/base 2025-11-03T16:58:01.2652740Z * [new branch] gh/yf225/93/head -> origin/gh/yf225/93/head 2025-11-03T16:58:01.2656144Z * [new branch] gh/yifuwang/152/base -> origin/gh/yifuwang/152/base 2025-11-03T16:58:01.2658080Z * [new branch] gh/yifuwang/152/head -> origin/gh/yifuwang/152/head 2025-11-03T16:58:01.2659830Z * [new branch] gh/yifuwang/152/orig -> origin/gh/yifuwang/152/orig 2025-11-03T16:58:01.2662206Z * [new branch] gh/yifuwang/195/base -> origin/gh/yifuwang/195/base 2025-11-03T16:58:01.2664400Z * [new branch] gh/yifuwang/195/head -> origin/gh/yifuwang/195/head 2025-11-03T16:58:01.2666147Z * [new branch] gh/yifuwang/195/orig -> origin/gh/yifuwang/195/orig 2025-11-03T16:58:01.2668949Z * [new branch] gh/yiming0416/1/base -> origin/gh/yiming0416/1/base 2025-11-03T16:58:01.2670991Z * [new branch] gh/yiming0416/1/head -> origin/gh/yiming0416/1/head 2025-11-03T16:58:01.2673677Z * [new branch] gh/yiming0416/2/base -> origin/gh/yiming0416/2/base 2025-11-03T16:58:01.2674959Z * [new branch] gh/yiming0416/2/head -> origin/gh/yiming0416/2/head 2025-11-03T16:58:01.2677945Z * [new branch] gh/yushangdi/1/base -> origin/gh/yushangdi/1/base 2025-11-03T16:58:01.2679524Z * [new branch] gh/yushangdi/1/head -> origin/gh/yushangdi/1/head 2025-11-03T16:58:01.2681749Z * [new branch] gh/yushangdi/2/base -> origin/gh/yushangdi/2/base 2025-11-03T16:58:01.2683586Z * [new branch] gh/yushangdi/2/head -> origin/gh/yushangdi/2/head 2025-11-03T16:58:01.2685806Z * [new branch] gh/yushangdi/3/base -> origin/gh/yushangdi/3/base 2025-11-03T16:58:01.2687541Z * [new branch] gh/yushangdi/3/head -> origin/gh/yushangdi/3/head 2025-11-03T16:58:01.2689421Z * [new branch] gh/yushangdi/3/orig -> origin/gh/yushangdi/3/orig 2025-11-03T16:58:01.2691818Z * [new branch] gh/yushangdi/4/base -> origin/gh/yushangdi/4/base 2025-11-03T16:58:01.2693559Z * [new branch] gh/yushangdi/4/head -> origin/gh/yushangdi/4/head 2025-11-03T16:58:01.2695305Z * [new branch] gh/yushangdi/4/orig -> origin/gh/yushangdi/4/orig 2025-11-03T16:58:01.2697600Z * [new branch] gh/yushangdi/5/base -> origin/gh/yushangdi/5/base 2025-11-03T16:58:01.2699573Z * [new branch] gh/yushangdi/5/head -> origin/gh/yushangdi/5/head 2025-11-03T16:58:01.2701297Z * [new branch] gh/yushangdi/5/orig -> origin/gh/yushangdi/5/orig 2025-11-03T16:58:01.2703351Z * [new branch] gh/yushangdi/6/base -> origin/gh/yushangdi/6/base 2025-11-03T16:58:01.2705269Z * [new branch] gh/yushangdi/6/head -> origin/gh/yushangdi/6/head 2025-11-03T16:58:01.2706947Z * [new branch] gh/yushangdi/6/orig -> origin/gh/yushangdi/6/orig 2025-11-03T16:58:01.2709299Z * [new branch] gh/yushangdi/7/base -> origin/gh/yushangdi/7/base 2025-11-03T16:58:01.2711733Z * [new branch] gh/yushangdi/7/head -> origin/gh/yushangdi/7/head 2025-11-03T16:58:01.2713307Z * [new branch] gh/yushangdi/7/orig -> origin/gh/yushangdi/7/orig 2025-11-03T16:58:01.2715651Z * [new branch] gh/yushangdi/8/base -> origin/gh/yushangdi/8/base 2025-11-03T16:58:01.2717291Z * [new branch] gh/yushangdi/8/head -> origin/gh/yushangdi/8/head 2025-11-03T16:58:01.2719345Z * [new branch] gh/yushangdi/8/orig -> origin/gh/yushangdi/8/orig 2025-11-03T16:58:01.2722417Z * [new branch] gh/zhxchen17/34/base -> origin/gh/zhxchen17/34/base 2025-11-03T16:58:01.2723889Z * [new branch] gh/zhxchen17/34/head -> origin/gh/zhxchen17/34/head 2025-11-03T16:58:01.2726515Z * [new branch] gh/zhxchen17/35/base -> origin/gh/zhxchen17/35/base 2025-11-03T16:58:01.2727837Z * [new branch] gh/zhxchen17/35/head -> origin/gh/zhxchen17/35/head 2025-11-03T16:58:01.2730372Z * [new branch] gh/zhxchen17/36/base -> origin/gh/zhxchen17/36/base 2025-11-03T16:58:01.2732279Z * [new branch] gh/zhxchen17/36/head -> origin/gh/zhxchen17/36/head 2025-11-03T16:58:01.2734000Z * [new branch] gh/zhxchen17/36/orig -> origin/gh/zhxchen17/36/orig 2025-11-03T16:58:01.2736214Z * [new branch] gh/zhxchen17/37/base -> origin/gh/zhxchen17/37/base 2025-11-03T16:58:01.2737989Z * [new branch] gh/zhxchen17/37/head -> origin/gh/zhxchen17/37/head 2025-11-03T16:58:01.2739800Z * [new branch] gh/zhxchen17/37/orig -> origin/gh/zhxchen17/37/orig 2025-11-03T16:58:01.2742478Z * [new branch] gh/zhxchen17/38/base -> origin/gh/zhxchen17/38/base 2025-11-03T16:58:01.2743868Z * [new branch] gh/zhxchen17/38/head -> origin/gh/zhxchen17/38/head 2025-11-03T16:58:01.2745769Z * [new branch] gh/zhxchen17/38/orig -> origin/gh/zhxchen17/38/orig 2025-11-03T16:58:01.2748411Z * [new branch] gh/zklaus/18/base -> origin/gh/zklaus/18/base 2025-11-03T16:58:01.2750091Z * [new branch] gh/zklaus/18/head -> origin/gh/zklaus/18/head 2025-11-03T16:58:01.2751895Z * [new branch] gh/zklaus/18/orig -> origin/gh/zklaus/18/orig 2025-11-03T16:58:01.2754805Z * [new branch] gh/zou3519/1197/base -> origin/gh/zou3519/1197/base 2025-11-03T16:58:01.2756434Z * [new branch] gh/zou3519/1197/head -> origin/gh/zou3519/1197/head 2025-11-03T16:58:01.2758192Z * [new branch] gh/zou3519/1197/orig -> origin/gh/zou3519/1197/orig 2025-11-03T16:58:01.2760521Z * [new branch] gh/zou3519/1198/base -> origin/gh/zou3519/1198/base 2025-11-03T16:58:01.2762353Z * [new branch] gh/zou3519/1198/head -> origin/gh/zou3519/1198/head 2025-11-03T16:58:01.2764096Z * [new branch] gh/zou3519/1198/orig -> origin/gh/zou3519/1198/orig 2025-11-03T16:58:01.2767006Z * [new branch] gh/zpcore/1/base -> origin/gh/zpcore/1/base 2025-11-03T16:58:01.2768929Z * [new branch] gh/zpcore/1/head -> origin/gh/zpcore/1/head 2025-11-03T16:58:01.2771660Z * [new branch] gh/zpcore/11/base -> origin/gh/zpcore/11/base 2025-11-03T16:58:01.2772996Z * [new branch] gh/zpcore/11/head -> origin/gh/zpcore/11/head 2025-11-03T16:58:01.2774815Z * [new branch] gh/zpcore/11/orig -> origin/gh/zpcore/11/orig 2025-11-03T16:58:01.2777891Z * [new branch] gh/zpcore/12/base -> origin/gh/zpcore/12/base 2025-11-03T16:58:01.2779643Z * [new branch] gh/zpcore/12/head -> origin/gh/zpcore/12/head 2025-11-03T16:58:01.2781381Z * [new branch] gh/zpcore/12/orig -> origin/gh/zpcore/12/orig 2025-11-03T16:58:01.2783768Z * [new branch] gh/zpcore/13/base -> origin/gh/zpcore/13/base 2025-11-03T16:58:01.2785410Z * [new branch] gh/zpcore/13/head -> origin/gh/zpcore/13/head 2025-11-03T16:58:01.2787126Z * [new branch] gh/zpcore/13/orig -> origin/gh/zpcore/13/orig 2025-11-03T16:58:01.2790218Z * [new branch] gh/zpcore/14/base -> origin/gh/zpcore/14/base 2025-11-03T16:58:01.2791553Z * [new branch] gh/zpcore/14/head -> origin/gh/zpcore/14/head 2025-11-03T16:58:01.2793162Z * [new branch] gh/zpcore/14/orig -> origin/gh/zpcore/14/orig 2025-11-03T16:58:01.2796077Z * [new branch] gh/zpcore/15/base -> origin/gh/zpcore/15/base 2025-11-03T16:58:01.2797389Z * [new branch] gh/zpcore/15/head -> origin/gh/zpcore/15/head 2025-11-03T16:58:01.2798990Z * [new branch] gh/zpcore/15/orig -> origin/gh/zpcore/15/orig 2025-11-03T16:58:01.2801581Z * [new branch] gh/zpcore/16/base -> origin/gh/zpcore/16/base 2025-11-03T16:58:01.2803349Z * [new branch] gh/zpcore/16/head -> origin/gh/zpcore/16/head 2025-11-03T16:58:01.2805093Z * [new branch] gh/zpcore/16/orig -> origin/gh/zpcore/16/orig 2025-11-03T16:58:01.2807926Z * [new branch] gh/zpcore/17/base -> origin/gh/zpcore/17/base 2025-11-03T16:58:01.2809895Z * [new branch] gh/zpcore/17/head -> origin/gh/zpcore/17/head 2025-11-03T16:58:01.2811969Z * [new branch] gh/zpcore/17/orig -> origin/gh/zpcore/17/orig 2025-11-03T16:58:01.2814407Z * [new branch] gh/zpcore/18/base -> origin/gh/zpcore/18/base 2025-11-03T16:58:01.2816161Z * [new branch] gh/zpcore/18/head -> origin/gh/zpcore/18/head 2025-11-03T16:58:01.2818453Z * [new branch] gh/zpcore/18/orig -> origin/gh/zpcore/18/orig 2025-11-03T16:58:01.2820349Z * [new branch] gh/zpcore/19/base -> origin/gh/zpcore/19/base 2025-11-03T16:58:01.2822075Z * [new branch] gh/zpcore/19/head -> origin/gh/zpcore/19/head 2025-11-03T16:58:01.2823815Z * [new branch] gh/zpcore/19/orig -> origin/gh/zpcore/19/orig 2025-11-03T16:58:01.2826647Z * [new branch] gh/zpcore/2/base -> origin/gh/zpcore/2/base 2025-11-03T16:58:01.2827988Z * [new branch] gh/zpcore/2/head -> origin/gh/zpcore/2/head 2025-11-03T16:58:01.2830354Z * [new branch] gh/zpcore/20/base -> origin/gh/zpcore/20/base 2025-11-03T16:58:01.2832233Z * [new branch] gh/zpcore/20/head -> origin/gh/zpcore/20/head 2025-11-03T16:58:01.2834009Z * [new branch] gh/zpcore/20/orig -> origin/gh/zpcore/20/orig 2025-11-03T16:58:01.2836729Z * [new branch] gh/zpcore/21/base -> origin/gh/zpcore/21/base 2025-11-03T16:58:01.2839060Z * [new branch] gh/zpcore/21/head -> origin/gh/zpcore/21/head 2025-11-03T16:58:01.2840499Z * [new branch] gh/zpcore/21/orig -> origin/gh/zpcore/21/orig 2025-11-03T16:58:01.2843345Z * [new branch] gh/zpcore/22/base -> origin/gh/zpcore/22/base 2025-11-03T16:58:01.2845054Z * [new branch] gh/zpcore/22/head -> origin/gh/zpcore/22/head 2025-11-03T16:58:01.2846508Z * [new branch] gh/zpcore/22/orig -> origin/gh/zpcore/22/orig 2025-11-03T16:58:01.2848625Z * [new branch] gh/zpcore/23/base -> origin/gh/zpcore/23/base 2025-11-03T16:58:01.2850577Z * [new branch] gh/zpcore/23/head -> origin/gh/zpcore/23/head 2025-11-03T16:58:01.2852221Z * [new branch] gh/zpcore/23/orig -> origin/gh/zpcore/23/orig 2025-11-03T16:58:01.2854413Z * [new branch] gh/zpcore/24/base -> origin/gh/zpcore/24/base 2025-11-03T16:58:01.2856107Z * [new branch] gh/zpcore/24/head -> origin/gh/zpcore/24/head 2025-11-03T16:58:01.2858331Z * [new branch] gh/zpcore/24/orig -> origin/gh/zpcore/24/orig 2025-11-03T16:58:01.2860243Z * [new branch] gh/zpcore/25/base -> origin/gh/zpcore/25/base 2025-11-03T16:58:01.2862078Z * [new branch] gh/zpcore/25/head -> origin/gh/zpcore/25/head 2025-11-03T16:58:01.2863861Z * [new branch] gh/zpcore/25/orig -> origin/gh/zpcore/25/orig 2025-11-03T16:58:01.2866506Z * [new branch] gh/zpcore/3/base -> origin/gh/zpcore/3/base 2025-11-03T16:58:01.2867722Z * [new branch] gh/zpcore/3/head -> origin/gh/zpcore/3/head 2025-11-03T16:58:01.2869811Z * [new branch] gh/zpcore/4/base -> origin/gh/zpcore/4/base 2025-11-03T16:58:01.2871587Z * [new branch] gh/zpcore/4/head -> origin/gh/zpcore/4/head 2025-11-03T16:58:01.2875105Z * [new branch] gh/zpcore/5/base -> origin/gh/zpcore/5/base 2025-11-03T16:58:01.2876836Z * [new branch] gh/zpcore/5/head -> origin/gh/zpcore/5/head 2025-11-03T16:58:01.2879193Z * [new branch] gh/zpcore/6/base -> origin/gh/zpcore/6/base 2025-11-03T16:58:01.2881055Z * [new branch] gh/zpcore/6/head -> origin/gh/zpcore/6/head 2025-11-03T16:58:01.2883323Z * [new branch] gh/zpcore/7/base -> origin/gh/zpcore/7/base 2025-11-03T16:58:01.2885049Z * [new branch] gh/zpcore/7/head -> origin/gh/zpcore/7/head 2025-11-03T16:58:01.2887346Z * [new branch] gh/zpcore/8/base -> origin/gh/zpcore/8/base 2025-11-03T16:58:01.2889127Z * [new branch] gh/zpcore/8/head -> origin/gh/zpcore/8/head 2025-11-03T16:58:01.2891138Z * [new branch] google-main -> origin/google-main 2025-11-03T16:58:01.2892892Z * [new branch] greencontext -> origin/greencontext 2025-11-03T16:58:01.2895312Z * [new branch] guangyey/config -> origin/guangyey/config 2025-11-03T16:58:01.2896994Z * [new branch] guangyey/external_stream -> origin/guangyey/external_stream 2025-11-03T16:58:01.2898603Z * [new branch] guangyey/reimport -> origin/guangyey/reimport 2025-11-03T16:58:01.2900240Z * [new branch] guangyey/test_2025 -> origin/guangyey/test_2025 2025-11-03T16:58:01.2902720Z * [new branch] guilhermeleobas/cherry-pick-55d87d9dfd9 -> origin/guilhermeleobas/cherry-pick-55d87d9dfd9 2025-11-03T16:58:01.2904986Z * [new branch] hameerabbasi/gradcheck-allclose -> origin/hameerabbasi/gradcheck-allclose 2025-11-03T16:58:01.2907401Z * [new branch] haozhe/bf16-dynamic-shape -> origin/haozhe/bf16-dynamic-shape 2025-11-03T16:58:01.2909173Z * [new branch] hc_baseline -> origin/hc_baseline 2025-11-03T16:58:01.2911306Z * [new branch] hhh_decomp_mul -> origin/hhh_decomp_mul 2025-11-03T16:58:01.2913033Z * [new branch] hhh_rand -> origin/hhh_rand 2025-11-03T16:58:01.2915324Z * [new branch] hoy/triton-PR3973 -> origin/hoy/triton-PR3973 2025-11-03T16:58:01.2918041Z * [new branch] huba/debug_mode -> origin/huba/debug_mode 2025-11-03T16:58:01.2919447Z * [new branch] huba/dtensor_equal -> origin/huba/dtensor_equal 2025-11-03T16:58:01.2921017Z * [new branch] huba/f1 -> origin/huba/f1 2025-11-03T16:58:01.2922920Z * [new branch] huba/local_tensor -> origin/huba/local_tensor 2025-11-03T16:58:01.2924835Z * [new branch] increase-asan-build-memory -> origin/increase-asan-build-memory 2025-11-03T16:58:01.2926677Z * [new branch] inductor-perf-increase-timeout -> origin/inductor-perf-increase-timeout 2025-11-03T16:58:01.2928565Z * [new branch] inductordecompfix -> origin/inductordecompfix 2025-11-03T16:58:01.2930453Z * [new branch] inlining -> origin/inlining 2025-11-03T16:58:01.2932299Z * [new branch] inlining-ezyang -> origin/inlining-ezyang 2025-11-03T16:58:01.2934200Z * [new branch] install-torchao-0.13.0 -> origin/install-torchao-0.13.0 2025-11-03T16:58:01.2936083Z * [new branch] install_free_tensors -> origin/install_free_tensors 2025-11-03T16:58:01.2937891Z * [new branch] invoke-subgraph -> origin/invoke-subgraph 2025-11-03T16:58:01.2939951Z * [new branch] issue#58739 -> origin/issue#58739 2025-11-03T16:58:01.2942118Z * [new branch] jathu/o3 -> origin/jathu/o3 2025-11-03T16:58:01.2943777Z * [new branch] jathu/sve -> origin/jathu/sve 2025-11-03T16:58:01.2946319Z * [new branch] jcaip/test-cusparselt-version-0.6.2 -> origin/jcaip/test-cusparselt-version-0.6.2 2025-11-03T16:58:01.2947909Z * [new branch] jcaip/update-cusparselt-0.6.2 -> origin/jcaip/update-cusparselt-0.6.2 2025-11-03T16:58:01.2951140Z * [new branch] jeanschmidt/autorevert_tag_cisev -> origin/jeanschmidt/autorevert_tag_cisev 2025-11-03T16:58:01.2952460Z * [new branch] jeanschmidt/autorevert_tag_cisev_2 -> origin/jeanschmidt/autorevert_tag_cisev_2 2025-11-03T16:58:01.2954281Z * [new branch] jithunnair-amd-patch-2 -> origin/jithunnair-amd-patch-2 2025-11-03T16:58:01.2956094Z * [new branch] jithunnair-amd-patch-3 -> origin/jithunnair-amd-patch-3 2025-11-03T16:58:01.2957903Z * [new branch] jithunnair-amd-patch-4 -> origin/jithunnair-amd-patch-4 2025-11-03T16:58:01.2960358Z * [new branch] justinchu/attention-tests -> origin/justinchu/attention-tests 2025-11-03T16:58:01.2962223Z * [new branch] justinchu/export-warning -> origin/justinchu/export-warning 2025-11-03T16:58:01.2963797Z * [new branch] justinchu/native-qdq -> origin/justinchu/native-qdq 2025-11-03T16:58:01.2966810Z * [new branch] justinchuby/onnx-deprecation-msg -> origin/justinchuby/onnx-deprecation-msg 2025-11-03T16:58:01.2969195Z * [new branch] kainan666/xlf_debug -> origin/kainan666/xlf_debug 2025-11-03T16:58:01.2970940Z * [new branch] kainan_test -> origin/kainan_test 2025-11-03T16:58:01.2973392Z * [new branch] leslie/test_group_gemm_epilogues -> origin/leslie/test_group_gemm_epilogues 2025-11-03T16:58:01.2975708Z * [new branch] lessw2020/fix_cutlass_cache_error -> origin/lessw2020/fix_cutlass_cache_error 2025-11-03T16:58:01.2978147Z * [new branch] liaoxuan/shm_all_reduce -> origin/liaoxuan/shm_all_reduce 2025-11-03T16:58:01.2979990Z * [new branch] liaoxuan/test_fa_disable_softmax -> origin/liaoxuan/test_fa_disable_softmax 2025-11-03T16:58:01.2981562Z * [new branch] liaoxuan/test_int8_sdpa -> origin/liaoxuan/test_int8_sdpa 2025-11-03T16:58:01.2983345Z * [new branch] lintbuilddocker -> origin/lintbuilddocker 2025-11-03T16:58:01.2985251Z * [new branch] llama4-stable -> origin/llama4-stable 2025-11-03T16:58:01.2986954Z * [new branch] logdetfix -> origin/logdetfix 2025-11-03T16:58:01.2989825Z * [new branch] lts/release/1.8 -> origin/lts/release/1.8 2025-11-03T16:58:01.2992179Z * [new branch] lucaskabela/#94773 -> origin/lucaskabela/#94773 2025-11-03T16:58:01.2993825Z * [new branch] lucaskabela/fix_164814 -> origin/lucaskabela/fix_164814 2025-11-03T16:58:01.2995445Z * [new branch] lucaskabela/fix_164823 -> origin/lucaskabela/fix_164823 2025-11-03T16:58:01.2997083Z * [new branch] lucaskabela/fix_164875 -> origin/lucaskabela/fix_164875 2025-11-03T16:58:01.2998924Z * [new branch] lucaskabela/fix_164876 -> origin/lucaskabela/fix_164876 2025-11-03T16:58:01.3000628Z * [new branch] lucaskabela/fix_error_for_named_tuples -> origin/lucaskabela/fix_error_for_named_tuples 2025-11-03T16:58:01.3002227Z * [new branch] lucaskabela/flop_counter -> origin/lucaskabela/flop_counter 2025-11-03T16:58:01.3004266Z * [new branch] lucaskabela/func_under_decomp -> origin/lucaskabela/func_under_decomp 2025-11-03T16:58:01.3006503Z * [new branch] lucaskabela/functional_in_dynamo -> origin/lucaskabela/functional_in_dynamo 2025-11-03T16:58:01.3008098Z * [new branch] lucaskabela/install_params_as_graph_attr -> origin/lucaskabela/install_params_as_graph_attr 2025-11-03T16:58:01.3009995Z * [new branch] lucaskabela/parameters_as_graph_attr -> origin/lucaskabela/parameters_as_graph_attr 2025-11-03T16:58:01.3012865Z * [new branch] lucaskabela/remove_aot_dispatcher_metadata -> origin/lucaskabela/remove_aot_dispatcher_metadata 2025-11-03T16:58:01.3014516Z * [new branch] lucaskabela/rnn_decomp -> origin/lucaskabela/rnn_decomp 2025-11-03T16:58:01.3016311Z * [new branch] lucaskabela/type_files_in_variables -> origin/lucaskabela/type_files_in_variables 2025-11-03T16:58:01.3018046Z * [new branch] lucaskabela/typing_backends -> origin/lucaskabela/typing_backends 2025-11-03T16:58:01.3020157Z * [new branch] lucaskabela/typing_variables -> origin/lucaskabela/typing_variables 2025-11-03T16:58:01.3021895Z * [new branch] lucaskabela/typinng_variables_builtin -> origin/lucaskabela/typinng_variables_builtin 2025-11-03T16:58:01.3024037Z * [new branch] lucaskablea/pickle_einops -> origin/lucaskablea/pickle_einops 2025-11-03T16:58:01.3025693Z * [new branch] main -> origin/main 2025-11-03T16:58:01.3028007Z * [new branch] main-enable-b200-distributed-tests -> origin/main-enable-b200-distributed-tests 2025-11-03T16:58:01.3029730Z * [new branch] malfet-patch-2 -> origin/malfet-patch-2 2025-11-03T16:58:01.3031586Z * [new branch] malfet-patch-3 -> origin/malfet-patch-3 2025-11-03T16:58:01.3033440Z * [new branch] malfet-patch-7 -> origin/malfet-patch-7 2025-11-03T16:58:01.3035226Z * [new branch] malfet-patch-8 -> origin/malfet-patch-8 2025-11-03T16:58:01.3037094Z * [new branch] malfet-patch-9 -> origin/malfet-patch-9 2025-11-03T16:58:01.3039359Z * [new branch] malfet/add-3.14-ci -> origin/malfet/add-3.14-ci 2025-11-03T16:58:01.3041089Z * [new branch] malfet/be-green-context -> origin/malfet/be-green-context 2025-11-03T16:58:01.3042895Z * [new branch] malfet/be-move-more-settings-to-checkout-pytorch -> origin/malfet/be-move-more-settings-to-checkout-pytorch 2025-11-03T16:58:01.3044967Z * [new branch] malfet/mps-implement-col2im -> origin/malfet/mps-implement-col2im 2025-11-03T16:58:01.3047323Z * [new branch] manuel/aoti_metal_shimify-thread_safe -> origin/manuel/aoti_metal_shimify-thread_safe 2025-11-03T16:58:01.3049211Z * [new branch] manuel/test-ops-common-allow-mps -> origin/manuel/test-ops-common-allow-mps 2025-11-03T16:58:01.3051648Z * [new branch] masnesral/metaconda -> origin/masnesral/metaconda 2025-11-03T16:58:01.3053150Z * [new branch] masnesral/pt2_internal_logging -> origin/masnesral/pt2_internal_logging 2025-11-03T16:58:01.3055298Z * [new branch] mem_profiler_stack_trace -> origin/mem_profiler_stack_trace 2025-11-03T16:58:01.3056921Z * [new branch] memory_profiler_stack -> origin/memory_profiler_stack 2025-11-03T16:58:01.3058674Z * [new branch] metascroy-patch-1 -> origin/metascroy-patch-1 2025-11-03T16:58:01.3060805Z * [new branch] mingw_posix -> origin/mingw_posix 2025-11-03T16:58:01.3063229Z * [new branch] mlazos/S429861-debug -> origin/mlazos/S429861-debug 2025-11-03T16:58:01.3064812Z * [new branch] mlazos/aa -> origin/mlazos/aa 2025-11-03T16:58:01.3066524Z * [new branch] mlazos/acts -> origin/mlazos/acts 2025-11-03T16:58:01.3068110Z * [new branch] mlazos/arg-renames -> origin/mlazos/arg-renames 2025-11-03T16:58:01.3070195Z * [new branch] mlazos/backup-test-branch -> origin/mlazos/backup-test-branch 2025-11-03T16:58:01.3071752Z * [new branch] mlazos/bad-cudagraphs -> origin/mlazos/bad-cudagraphs 2025-11-03T16:58:01.3073417Z * [new branch] mlazos/baseline -> origin/mlazos/baseline 2025-11-03T16:58:01.3075375Z * [new branch] mlazos/baseline-graph-breaks -> origin/mlazos/baseline-graph-breaks 2025-11-03T16:58:01.3077216Z * [new branch] mlazos/beta-tensor -> origin/mlazos/beta-tensor 2025-11-03T16:58:01.3079287Z * [new branch] mlazos/buffers -> origin/mlazos/buffers 2025-11-03T16:58:01.3080917Z * [new branch] mlazos/buffers2 -> origin/mlazos/buffers2 2025-11-03T16:58:01.3082674Z * [new branch] mlazos/buffers3 -> origin/mlazos/buffers3 2025-11-03T16:58:01.3084732Z * [new branch] mlazos/ck2 -> origin/mlazos/ck2 2025-11-03T16:58:01.3086562Z * [new branch] mlazos/combokernels -> origin/mlazos/combokernels 2025-11-03T16:58:01.3088865Z * [new branch] mlazos/ctx-cleanup -> origin/mlazos/ctx-cleanup 2025-11-03T16:58:01.3090753Z * [new branch] mlazos/cuda-cmd-log -> origin/mlazos/cuda-cmd-log 2025-11-03T16:58:01.3092529Z * [new branch] mlazos/cudagraph-tests -> origin/mlazos/cudagraph-tests 2025-11-03T16:58:01.3094358Z * [new branch] mlazos/cudagraphs-measurement -> origin/mlazos/cudagraphs-measurement 2025-11-03T16:58:01.3096078Z * [new branch] mlazos/cutlass-test -> origin/mlazos/cutlass-test 2025-11-03T16:58:01.3097823Z * [new branch] mlazos/cutlass-topo-bug -> origin/mlazos/cutlass-topo-bug 2025-11-03T16:58:01.3099646Z * [new branch] mlazos/dataclass-proxy -> origin/mlazos/dataclass-proxy 2025-11-03T16:58:01.3101338Z * [new branch] mlazos/dc-attrs -> origin/mlazos/dc-attrs 2025-11-03T16:58:01.3103124Z * [new branch] mlazos/dc-helion -> origin/mlazos/dc-helion 2025-11-03T16:58:01.3104790Z * [new branch] mlazos/dict-fix -> origin/mlazos/dict-fix 2025-11-03T16:58:01.3106551Z * [new branch] mlazos/disable-tf -> origin/mlazos/disable-tf 2025-11-03T16:58:01.3108237Z * [new branch] mlazos/dupe-fix -> origin/mlazos/dupe-fix 2025-11-03T16:58:01.3110279Z * [new branch] mlazos/dyn-batch -> origin/mlazos/dyn-batch 2025-11-03T16:58:01.3112047Z * [new branch] mlazos/evt -> origin/mlazos/evt 2025-11-03T16:58:01.3113913Z * [new branch] mlazos/extract-examples -> origin/mlazos/extract-examples 2025-11-03T16:58:01.3115490Z * [new branch] mlazos/foreach-op -> origin/mlazos/foreach-op 2025-11-03T16:58:01.3117313Z * [new branch] mlazos/fp8 -> origin/mlazos/fp8 2025-11-03T16:58:01.3119354Z * [new branch] mlazos/fp8-bias -> origin/mlazos/fp8-bias 2025-11-03T16:58:01.3120957Z * [new branch] mlazos/fp8-bias-fusion -> origin/mlazos/fp8-bias-fusion 2025-11-03T16:58:01.3122626Z * [new branch] mlazos/fp8-fixes -> origin/mlazos/fp8-fixes 2025-11-03T16:58:01.3124335Z * [new branch] mlazos/freezing -> origin/mlazos/freezing 2025-11-03T16:58:01.3126051Z * [new branch] mlazos/h-comp -> origin/mlazos/h-comp 2025-11-03T16:58:01.3127881Z * [new branch] mlazos/h-comp2 -> origin/mlazos/h-comp2 2025-11-03T16:58:01.3129756Z * [new branch] mlazos/hash-hop -> origin/mlazos/hash-hop 2025-11-03T16:58:01.3131753Z * [new branch] mlazos/hc -> origin/mlazos/hc 2025-11-03T16:58:01.3133496Z * [new branch] mlazos/hc-cycles -> origin/mlazos/hc-cycles 2025-11-03T16:58:01.3135263Z * [new branch] mlazos/hc-fixes -> origin/mlazos/hc-fixes 2025-11-03T16:58:01.3136957Z * [new branch] mlazos/hc-fixes3 -> origin/mlazos/hc-fixes3 2025-11-03T16:58:01.3138689Z * [new branch] mlazos/hc-fixes4 -> origin/mlazos/hc-fixes4 2025-11-03T16:58:01.3140366Z * [new branch] mlazos/hc-hf -> origin/mlazos/hc-hf 2025-11-03T16:58:01.3142074Z * [new branch] mlazos/hc-mut -> origin/mlazos/hc-mut 2025-11-03T16:58:01.3143805Z * [new branch] mlazos/hc10 -> origin/mlazos/hc10 2025-11-03T16:58:01.3145514Z * [new branch] mlazos/hc11 -> origin/mlazos/hc11 2025-11-03T16:58:01.3147228Z * [new branch] mlazos/hc12 -> origin/mlazos/hc12 2025-11-03T16:58:01.3148913Z * [new branch] mlazos/hc13 -> origin/mlazos/hc13 2025-11-03T16:58:01.3150597Z * [new branch] mlazos/hc14 -> origin/mlazos/hc14 2025-11-03T16:58:01.3152306Z * [new branch] mlazos/hc15 -> origin/mlazos/hc15 2025-11-03T16:58:01.3154097Z * [new branch] mlazos/hc2 -> origin/mlazos/hc2 2025-11-03T16:58:01.3155906Z * [new branch] mlazos/hc4 -> origin/mlazos/hc4 2025-11-03T16:58:01.3158111Z * [new branch] mlazos/hc5 -> origin/mlazos/hc5 2025-11-03T16:58:01.3159664Z * [new branch] mlazos/hc6 -> origin/mlazos/hc6 2025-11-03T16:58:01.3161289Z * [new branch] mlazos/hc7 -> origin/mlazos/hc7 2025-11-03T16:58:01.3162881Z * [new branch] mlazos/hc8 -> origin/mlazos/hc8 2025-11-03T16:58:01.3164705Z * [new branch] mlazos/hc9 -> origin/mlazos/hc9 2025-11-03T16:58:01.3167338Z * [new branch] mlazos/hc_baseline2 -> origin/mlazos/hc_baseline2 2025-11-03T16:58:01.3168364Z * [new branch] mlazos/inductor-streams -> origin/mlazos/inductor-streams 2025-11-03T16:58:01.3170444Z * [new branch] mlazos/lr-composibility -> origin/mlazos/lr-composibility 2025-11-03T16:58:01.3172010Z * [new branch] mlazos/main -> origin/mlazos/main 2025-11-03T16:58:01.3173878Z * [new branch] mlazos/main-test-enablement -> origin/mlazos/main-test-enablement 2025-11-03T16:58:01.3175577Z * [new branch] mlazos/mcg -> origin/mlazos/mcg 2025-11-03T16:58:01.3177360Z * [new branch] mlazos/mcg2 -> origin/mlazos/mcg2 2025-11-03T16:58:01.3179218Z * [new branch] mlazos/meta-guards -> origin/mlazos/meta-guards 2025-11-03T16:58:01.3181338Z * [new branch] mlazos/mlazos/ck2 -> origin/mlazos/mlazos/ck2 2025-11-03T16:58:01.3183159Z * [new branch] mlazos/mlazos/foreach-map-adam -> origin/mlazos/mlazos/foreach-map-adam 2025-11-03T16:58:01.3184887Z * [new branch] mlazos/mlazos/tf-mode-backup -> origin/mlazos/mlazos/tf-mode-backup 2025-11-03T16:58:01.3186578Z * [new branch] mlazos/mod-fix -> origin/mlazos/mod-fix 2025-11-03T16:58:01.3188404Z * [new branch] mlazos/mode-fix -> origin/mlazos/mode-fix 2025-11-03T16:58:01.3190114Z * [new branch] mlazos/more-tests -> origin/mlazos/more-tests 2025-11-03T16:58:01.3192008Z * [new branch] mlazos/offsets -> origin/mlazos/offsets 2025-11-03T16:58:01.3193810Z * [new branch] mlazos/proxy-ctors -> origin/mlazos/proxy-ctors 2025-11-03T16:58:01.3195593Z * [new branch] mlazos/quant-fix -> origin/mlazos/quant-fix 2025-11-03T16:58:01.3197325Z * [new branch] mlazos/resnet-fix -> origin/mlazos/resnet-fix 2025-11-03T16:58:01.3199245Z * [new branch] mlazos/rm-buf-names -> origin/mlazos/rm-buf-names 2025-11-03T16:58:01.3200894Z * [new branch] mlazos/rm-code -> origin/mlazos/rm-code 2025-11-03T16:58:01.3202637Z * [new branch] mlazos/rm-spam -> origin/mlazos/rm-spam 2025-11-03T16:58:01.3204362Z * [new branch] mlazos/rtp -> origin/mlazos/rtp 2025-11-03T16:58:01.3206112Z * [new branch] mlazos/static-idx-dbg -> origin/mlazos/static-idx-dbg 2025-11-03T16:58:01.3208241Z * [new branch] mlazos/static-inputs-log -> origin/mlazos/static-inputs-log 2025-11-03T16:58:01.3210323Z * [new branch] mlazos/td-fix2 -> origin/mlazos/td-fix2 2025-11-03T16:58:01.3212414Z * [new branch] mlazos/tensor-hasattr2 -> origin/mlazos/tensor-hasattr2 2025-11-03T16:58:01.3214068Z * [new branch] mlazos/test -> origin/mlazos/test 2025-11-03T16:58:01.3215838Z * [new branch] mlazos/tf-mode -> origin/mlazos/tf-mode 2025-11-03T16:58:01.3217648Z * [new branch] mlazos/tf-mode-backup2 -> origin/mlazos/tf-mode-backup2 2025-11-03T16:58:01.3219213Z * [new branch] mlazos/tf-mode-reland -> origin/mlazos/tf-mode-reland 2025-11-03T16:58:01.3221242Z * [new branch] mlazos/tf-mode-reland2 -> origin/mlazos/tf-mode-reland2 2025-11-03T16:58:01.3222936Z * [new branch] mlazos/tf-mode-reland3 -> origin/mlazos/tf-mode-reland3 2025-11-03T16:58:01.3224600Z * [new branch] mlazos/triton-no-epi -> origin/mlazos/triton-no-epi 2025-11-03T16:58:01.3226360Z * [new branch] mlazos/tune-proto -> origin/mlazos/tune-proto 2025-11-03T16:58:01.3228025Z * [new branch] mlazos/tuple-fixes -> origin/mlazos/tuple-fixes 2025-11-03T16:58:01.3230095Z * [new branch] mlazos/tuple-fixes2 -> origin/mlazos/tuple-fixes2 2025-11-03T16:58:01.3231741Z * [new branch] mlazos/tuple-handling -> origin/mlazos/tuple-handling 2025-11-03T16:58:01.3233413Z * [new branch] mlazos/user-stream-base -> origin/mlazos/user-stream-base 2025-11-03T16:58:01.3234980Z * [new branch] mlazos/user-streams -> origin/mlazos/user-streams 2025-11-03T16:58:01.3236836Z * [new branch] mlazos/user-streams-backup -> origin/mlazos/user-streams-backup 2025-11-03T16:58:01.3238213Z * [new branch] mlazos/user-streams-backup2 -> origin/mlazos/user-streams-backup2 2025-11-03T16:58:01.3240224Z * [new branch] mlazos/vary-beta -> origin/mlazos/vary-beta 2025-11-03T16:58:01.3242067Z * [new branch] mlazos/vary-beta2 -> origin/mlazos/vary-beta2 2025-11-03T16:58:01.3243711Z * [new branch] mlazos/weird-perf1 -> origin/mlazos/weird-perf1 2025-11-03T16:58:01.3245454Z * [new branch] mm_out_dtype_compile -> origin/mm_out_dtype_compile 2025-11-03T16:58:01.3247235Z * [new branch] module-shim -> origin/module-shim 2025-11-03T16:58:01.3249357Z * [new branch] move-theme-out-docker -> origin/move-theme-out-docker 2025-11-03T16:58:01.3251681Z * [new branch] move_aws_steps_inside_setup_rocm -> origin/move_aws_steps_inside_setup_rocm 2025-11-03T16:58:01.3253520Z * [new branch] msaroufim-patch-2 -> origin/msaroufim-patch-2 2025-11-03T16:58:01.3255839Z * [new branch] msaroufim/reduce -> origin/msaroufim/reduce 2025-11-03T16:58:01.3257955Z * [new branch] mtia/basic-cmake -> origin/mtia/basic-cmake 2025-11-03T16:58:01.3261033Z * [new branch] mwizak/fix-triton-block-shape -> origin/mwizak/fix-triton-block-shape 2025-11-03T16:58:01.3262235Z * [new branch] mwizak/restrict-test-mm-backend -> origin/mwizak/restrict-test-mm-backend 2025-11-03T16:58:01.3263810Z * [new branch] my_varlen_backup -> origin/my_varlen_backup 2025-11-03T16:58:01.3265643Z * [new branch] nativert_num_outputs -> origin/nativert_num_outputs 2025-11-03T16:58:01.3267405Z * [new branch] new-codegen -> origin/new-codegen 2025-11-03T16:58:01.3269172Z * [new branch] newtest-base -> origin/newtest-base 2025-11-03T16:58:01.3271727Z * [new branch] ngimel/discont_allgather -> origin/ngimel/discont_allgather 2025-11-03T16:58:01.3273270Z * [new branch] ngimel/error_index_list -> origin/ngimel/error_index_list 2025-11-03T16:58:01.3274831Z * [new branch] ngimel/gg_new -> origin/ngimel/gg_new 2025-11-03T16:58:01.3276501Z * [new branch] ngimel/nDim_fix -> origin/ngimel/nDim_fix 2025-11-03T16:58:01.3278109Z * [new branch] nightly -> origin/nightly 2025-11-03T16:58:01.3280594Z * [new branch] nikitaved/addmm_1_rowcol_lt_path_check -> origin/nikitaved/addmm_1_rowcol_lt_path_check 2025-11-03T16:58:01.3282195Z * [new branch] nikitaved/addmm_epilogue_fusions -> origin/nikitaved/addmm_epilogue_fusions 2025-11-03T16:58:01.3283844Z * [new branch] nikitaved/addmm_epilogue_fusions_2d_bias -> origin/nikitaved/addmm_epilogue_fusions_2d_bias 2025-11-03T16:58:01.3285804Z * [new branch] nikitaved/addmm_epilogue_fusions_inductor -> origin/nikitaved/addmm_epilogue_fusions_inductor 2025-11-03T16:58:01.3287784Z * [new branch] nikitaved/addmm_epilogue_fusions_scratch -> origin/nikitaved/addmm_epilogue_fusions_scratch 2025-11-03T16:58:01.3289922Z * [new branch] nikitaved/simpler_can_use_32bit_index -> origin/nikitaved/simpler_can_use_32bit_index 2025-11-03T16:58:01.3291595Z * [new branch] nikitaved/test -> origin/nikitaved/test 2025-11-03T16:58:01.3293677Z * [new branch] nmacchioni-40p-machine-balance -> origin/nmacchioni-40p-machine-balance 2025-11-03T16:58:01.3295511Z * [new branch] nmacchioni-60p-machine-balance -> origin/nmacchioni-60p-machine-balance 2025-11-03T16:58:01.3297348Z * [new branch] nmacchioni-padmm-no-compute-bound-check -> origin/nmacchioni-padmm-no-compute-bound-check 2025-11-03T16:58:01.3299133Z * [new branch] nmacchioni-patch-1 -> origin/nmacchioni-patch-1 2025-11-03T16:58:01.3300941Z * [new branch] no_distributed_log_spew -> origin/no_distributed_log_spew 2025-11-03T16:58:01.3302693Z * [new branch] nofun-hack -> origin/nofun-hack 2025-11-03T16:58:01.3305163Z * [new branch] nullplay/fuse_matmul -> origin/nullplay/fuse_matmul 2025-11-03T16:58:01.3306984Z * [new branch] nullplay_fuse_matmul -> origin/nullplay_fuse_matmul 2025-11-03T16:58:01.3309739Z * [new branch] orig/release/1.10 -> origin/orig/release/1.10 2025-11-03T16:58:01.3315450Z * [new branch] orig/release/1.11 -> origin/orig/release/1.11 2025-11-03T16:58:01.3319142Z * [new branch] orig/release/1.12 -> origin/orig/release/1.12 2025-11-03T16:58:01.3321730Z * [new branch] orig/release/1.13 -> origin/orig/release/1.13 2025-11-03T16:58:01.3323047Z * [new branch] orig/release/1.6 -> origin/orig/release/1.6 2025-11-03T16:58:01.3324938Z * [new branch] orig/release/1.7 -> origin/orig/release/1.7 2025-11-03T16:58:01.3327626Z * [new branch] orig/release/1.8 -> origin/orig/release/1.8 2025-11-03T16:58:01.3329151Z * [new branch] orig/release/1.9 -> origin/orig/release/1.9 2025-11-03T16:58:01.3330935Z * [new branch] orig/release/2.0 -> origin/orig/release/2.0 2025-11-03T16:58:01.3332689Z * [new branch] orig/release/2.1 -> origin/orig/release/2.1 2025-11-03T16:58:01.3334393Z * [new branch] orig/release/2.2 -> origin/orig/release/2.2 2025-11-03T16:58:01.3336111Z * [new branch] orig/release/2.3 -> origin/orig/release/2.3 2025-11-03T16:58:01.3338263Z * [new branch] orig/release/2.4 -> origin/orig/release/2.4 2025-11-03T16:58:01.3339937Z * [new branch] orig/release/2.5 -> origin/orig/release/2.5 2025-11-03T16:58:01.3341587Z * [new branch] orig/release/2.6 -> origin/orig/release/2.6 2025-11-03T16:58:01.3344158Z * [new branch] orig/release/2.7 -> origin/orig/release/2.7 2025-11-03T16:58:01.3345990Z * [new branch] orig/release/2.8 -> origin/orig/release/2.8 2025-11-03T16:58:01.3347775Z * [new branch] orig/release/2.9 -> origin/orig/release/2.9 2025-11-03T16:58:01.3352738Z * [new branch] origin/gh/fxdawnn/1/base -> origin/origin/gh/fxdawnn/1/base 2025-11-03T16:58:01.3353967Z * [new branch] origin/gh/fxdawnn/1/orig -> origin/origin/gh/fxdawnn/1/orig 2025-11-03T16:58:01.3357287Z * [new branch] origin/gh/zpcore/14/orig -> origin/origin/gh/zpcore/14/orig 2025-11-03T16:58:01.3358625Z * [new branch] padded-tensor -> origin/padded-tensor 2025-11-03T16:58:01.3360547Z * [new branch] pca2 -> origin/pca2 2025-11-03T16:58:01.3362290Z * [new branch] per_channel_backup -> origin/per_channel_backup 2025-11-03T16:58:01.3364094Z * [new branch] perf_ops -> origin/perf_ops 2025-11-03T16:58:01.3366002Z * [new branch] perf_ops_2_9 -> origin/perf_ops_2_9 2025-11-03T16:58:01.3369004Z * [new branch] pianpwk-patch-1 -> origin/pianpwk-patch-1 2025-11-03T16:58:01.3370641Z * [new branch] pianpwk-patch-2 -> origin/pianpwk-patch-2 2025-11-03T16:58:01.3372501Z * [new branch] pianpwk-patch-3 -> origin/pianpwk-patch-3 2025-11-03T16:58:01.3375277Z * [new branch] pianpwk/__draft_debug_mode -> origin/pianpwk/__draft_debug_mode 2025-11-03T16:58:01.3376770Z * [new branch] pianpwk/_debug_mode_for_triton_draft -> origin/pianpwk/_debug_mode_for_triton_draft 2025-11-03T16:58:01.3378486Z * [new branch] pianpwk/_profile_w_dispatch_keys -> origin/pianpwk/_profile_w_dispatch_keys 2025-11-03T16:58:01.3380085Z * [new branch] pianpwk/_super_draft_debug_mode -> origin/pianpwk/_super_draft_debug_mode 2025-11-03T16:58:01.3381646Z * [new branch] pianpwk/anomaly_tb -> origin/pianpwk/anomaly_tb 2025-11-03T16:58:01.3383869Z * [new branch] pianpwk/auto_fx_annotate -> origin/pianpwk/auto_fx_annotate 2025-11-03T16:58:01.3385898Z * [new branch] pianpwk/backed_size_oblivious_export -> origin/pianpwk/backed_size_oblivious_export 2025-11-03T16:58:01.3388591Z * [new branch] pianpwk/base_view_shape_key -> origin/pianpwk/base_view_shape_key 2025-11-03T16:58:01.3389850Z * [new branch] pianpwk/bert_dynamic_perf -> origin/pianpwk/bert_dynamic_perf 2025-11-03T16:58:01.3391700Z * [new branch] pianpwk/debug_fwd_stack_traces -> origin/pianpwk/debug_fwd_stack_traces 2025-11-03T16:58:01.3393275Z * [new branch] pianpwk/debug_mode_hacks -> origin/pianpwk/debug_mode_hacks 2025-11-03T16:58:01.3395287Z * [new branch] pianpwk/debug_mode_hooks -> origin/pianpwk/debug_mode_hooks 2025-11-03T16:58:01.3398023Z * [new branch] pianpwk/debug_mode_inductor -> origin/pianpwk/debug_mode_inductor 2025-11-03T16:58:01.3399608Z * [new branch] pianpwk/debug_mode_opcall_refactor -> origin/pianpwk/debug_mode_opcall_refactor 2025-11-03T16:58:01.3400840Z * [new branch] pianpwk/debug_mode_show_ids -> origin/pianpwk/debug_mode_show_ids 2025-11-03T16:58:01.3402505Z * [new branch] pianpwk/debugmode_compile_tf -> origin/pianpwk/debugmode_compile_tf 2025-11-03T16:58:01.3404228Z * [new branch] pianpwk/debugmode_show_ids -> origin/pianpwk/debugmode_show_ids 2025-11-03T16:58:01.3406067Z * [new branch] pianpwk/dispatch_key_debugging_for_debug -> origin/pianpwk/dispatch_key_debugging_for_debug 2025-11-03T16:58:01.3407980Z * [new branch] pianpwk/draft_debug_mode_tfcompile -> origin/pianpwk/draft_debug_mode_tfcompile 2025-11-03T16:58:01.3410715Z * [new branch] pianpwk/draft_multikernel_nn -> origin/pianpwk/draft_multikernel_nn 2025-11-03T16:58:01.3412219Z * [new branch] pianpwk/draft_multikernel_status_10_5 -> origin/pianpwk/draft_multikernel_status_10_5 2025-11-03T16:58:01.3413803Z * [new branch] pianpwk/event_list_tree -> origin/pianpwk/event_list_tree 2025-11-03T16:58:01.3415553Z * [new branch] pianpwk/false_numel_refs -> origin/pianpwk/false_numel_refs 2025-11-03T16:58:01.3417445Z * [new branch] pianpwk/fx_export_annotate -> origin/pianpwk/fx_export_annotate 2025-11-03T16:58:01.3419126Z * [new branch] pianpwk/maybe_guard_rel -> origin/pianpwk/maybe_guard_rel 2025-11-03T16:58:01.3420927Z * [new branch] pianpwk/multikernel_hints_draft -> origin/pianpwk/multikernel_hints_draft 2025-11-03T16:58:01.3422662Z * [new branch] pianpwk/no_size_oblivious_slice_scat -> origin/pianpwk/no_size_oblivious_slice_scat 2025-11-03T16:58:01.3424517Z * [new branch] pianpwk/oblivious_reshape_view_better -> origin/pianpwk/oblivious_reshape_view_better 2025-11-03T16:58:01.3426372Z * [new branch] pianpwk/pre_forward_hook -> origin/pianpwk/pre_forward_hook 2025-11-03T16:58:01.3428226Z * [new branch] pianpwk/skip_python_keys_alternate -> origin/pianpwk/skip_python_keys_alternate 2025-11-03T16:58:01.3430378Z * [new branch] pianpwk/skip_python_keys_in_guards -> origin/pianpwk/skip_python_keys_in_guards 2025-11-03T16:58:01.3431864Z * [new branch] pianpwk/sym_tokens_draft -> origin/pianpwk/sym_tokens_draft 2025-11-03T16:58:01.3433865Z * [new branch] pianpwk/test_pointwise_guard_or_false -> origin/pianpwk/test_pointwise_guard_or_false 2025-11-03T16:58:01.3435466Z * [new branch] pianpwk/totally_draft_sym_wrap -> origin/pianpwk/totally_draft_sym_wrap 2025-11-03T16:58:01.3437204Z * [new branch] pianpwk/try_dumb_stuff -> origin/pianpwk/try_dumb_stuff 2025-11-03T16:58:01.3438855Z * [new branch] pianpwk/try_dumb_stuff_2 -> origin/pianpwk/try_dumb_stuff_2 2025-11-03T16:58:01.3440687Z * [new branch] pianpwk/user_symints -> origin/pianpwk/user_symints 2025-11-03T16:58:01.3442750Z * [new branch] pianpwk/wan21_reshape -> origin/pianpwk/wan21_reshape 2025-11-03T16:58:01.3445028Z * [new branch] piz/fall_back_missing_0716 -> origin/piz/fall_back_missing_0716 2025-11-03T16:58:01.3446743Z * [new branch] piz/fix_static_shard_method -> origin/piz/fix_static_shard_method 2025-11-03T16:58:01.3448580Z * [new branch] pool-separate -> origin/pool-separate 2025-11-03T16:58:01.3450398Z * [new branch] pr-156087 -> origin/pr-156087 2025-11-03T16:58:01.3452752Z * [new branch] pr/131860 -> origin/pr/131860 2025-11-03T16:58:01.3454578Z * [new branch] pr165329 -> origin/pr165329 2025-11-03T16:58:01.3456324Z * [new branch] pr165330 -> origin/pr165330 2025-11-03T16:58:01.3458082Z * [new branch] pr165564 -> origin/pr165564 2025-11-03T16:58:01.3459871Z * [new branch] pr165636 -> origin/pr165636 2025-11-03T16:58:01.3461684Z * [new branch] pr165637 -> origin/pr165637 2025-11-03T16:58:01.3463416Z * [new branch] pr165638 -> origin/pr165638 2025-11-03T16:58:01.3465183Z * [new branch] pr165639 -> origin/pr165639 2025-11-03T16:58:01.3466958Z * [new branch] pr165640 -> origin/pr165640 2025-11-03T16:58:01.3468827Z * [new branch] pr165642 -> origin/pr165642 2025-11-03T16:58:01.3470780Z * [new branch] pr165994 -> origin/pr165994 2025-11-03T16:58:01.3472745Z * [new branch] pr166102 -> origin/pr166102 2025-11-03T16:58:01.3474522Z * [new branch] pr166103 -> origin/pr166103 2025-11-03T16:58:01.3476321Z * [new branch] pr166149 -> origin/pr166149 2025-11-03T16:58:01.3478162Z * [new branch] pr166182 -> origin/pr166182 2025-11-03T16:58:01.3479955Z * [new branch] predispatch_to -> origin/predispatch_to 2025-11-03T16:58:01.3481849Z * [new branch] prepare-perf-baseline-number-2.8 -> origin/prepare-perf-baseline-number-2.8 2025-11-03T16:58:01.3483937Z * [new branch] prepare-perf-number-2.9 -> origin/prepare-perf-number-2.9 2025-11-03T16:58:01.3485586Z * [new branch] pt-opt-cuda3 -> origin/pt-opt-cuda3 2025-11-03T16:58:01.3487660Z * [new branch] python_compiled_autograd -> origin/python_compiled_autograd 2025-11-03T16:58:01.3490542Z * [new branch] qchip/export-D54134695 -> origin/qchip/export-D54134695 2025-11-03T16:58:01.3492287Z * [new branch] quint-bits -> origin/quint-bits 2025-11-03T16:58:01.3494001Z * [new branch] reland-accgrad-stream-warn -> origin/reland-accgrad-stream-warn 2025-11-03T16:58:01.3495932Z * [new branch] reland-fx-annotate -> origin/reland-fx-annotate 2025-11-03T16:58:01.3498789Z * [new branch] release/1.10 -> origin/release/1.10 2025-11-03T16:58:01.3500134Z * [new branch] release/1.11 -> origin/release/1.11 2025-11-03T16:58:01.3501854Z * [new branch] release/1.12 -> origin/release/1.12 2025-11-03T16:58:01.3503572Z * [new branch] release/1.13 -> origin/release/1.13 2025-11-03T16:58:01.3505242Z * [new branch] release/1.4 -> origin/release/1.4 2025-11-03T16:58:01.3506716Z * [new branch] release/1.4.1 -> origin/release/1.4.1 2025-11-03T16:58:01.3508511Z * [new branch] release/1.5 -> origin/release/1.5 2025-11-03T16:58:01.3510437Z * [new branch] release/1.6 -> origin/release/1.6 2025-11-03T16:58:01.3512364Z * [new branch] release/1.7 -> origin/release/1.7 2025-11-03T16:58:01.3514493Z * [new branch] release/1.8 -> origin/release/1.8 2025-11-03T16:58:01.3516112Z * [new branch] release/1.9 -> origin/release/1.9 2025-11-03T16:58:01.3518165Z * [new branch] release/2.0 -> origin/release/2.0 2025-11-03T16:58:01.3520114Z * [new branch] release/2.1 -> origin/release/2.1 2025-11-03T16:58:01.3521945Z * [new branch] release/2.2 -> origin/release/2.2 2025-11-03T16:58:01.3524665Z * [new branch] release/2.3 -> origin/release/2.3 2025-11-03T16:58:01.3526827Z * [new branch] release/2.4 -> origin/release/2.4 2025-11-03T16:58:01.3529185Z * [new branch] release/2.5 -> origin/release/2.5 2025-11-03T16:58:01.3531084Z * [new branch] release/2.6 -> origin/release/2.6 2025-11-03T16:58:01.3533150Z * [new branch] release/2.7 -> origin/release/2.7 2025-11-03T16:58:01.3534913Z * [new branch] release/2.8 -> origin/release/2.8 2025-11-03T16:58:01.3537120Z * [new branch] release/2.9 -> origin/release/2.9 2025-11-03T16:58:01.3538785Z * [new branch] release_notes -> origin/release_notes 2025-11-03T16:58:01.3540574Z * [new branch] remove-meta-files -> origin/remove-meta-files 2025-11-03T16:58:01.3542331Z * [new branch] remove_pyinterpreter -> origin/remove_pyinterpreter 2025-11-03T16:58:01.3544295Z * [new branch] replace-pytorch-labs-20250812-195836 -> origin/replace-pytorch-labs-20250812-195836 2025-11-03T16:58:01.3545882Z * [new branch] replace-pytorch-labs-20250812-200248 -> origin/replace-pytorch-labs-20250812-200248 2025-11-03T16:58:01.3547469Z * [new branch] replace-pytorch-labs-20250812-200324 -> origin/replace-pytorch-labs-20250812-200324 2025-11-03T16:58:01.3549155Z * [new branch] replace-pytorch-labs-20250812-204020 -> origin/replace-pytorch-labs-20250812-204020 2025-11-03T16:58:01.3552798Z * [new branch] revert-131069-gh/krzysztofjordan/1/head -> origin/revert-131069-gh/krzysztofjordan/1/head 2025-11-03T16:58:01.3556356Z * [new branch] revert-131469-gh/andrewor14/51/head -> origin/revert-131469-gh/andrewor14/51/head 2025-11-03T16:58:01.3559679Z * [new branch] revert-152361-gh/fadara01/1/head -> origin/revert-152361-gh/fadara01/1/head 2025-11-03T16:58:01.3563764Z * [new branch] revert-156870-gh/skarjala/3/head -> origin/revert-156870-gh/skarjala/3/head 2025-11-03T16:58:01.3565456Z * [new branch] revert-157914-cherry-pick-157503-by-pytorch_bot_bot_ -> origin/revert-157914-cherry-pick-157503-by-pytorch_bot_bot_ 2025-11-03T16:58:01.3568328Z * [new branch] revert-165144-gh/fadara01/2/head -> origin/revert-165144-gh/fadara01/2/head 2025-11-03T16:58:01.3570239Z * [new branch] revert-cpp -> origin/revert-cpp 2025-11-03T16:58:01.3572487Z * [new branch] revert-failed -> origin/revert-failed 2025-11-03T16:58:01.3574308Z * [new branch] revert_always_build_distributed -> origin/revert_always_build_distributed 2025-11-03T16:58:01.3576030Z * [new branch] rms_norm_patch -> origin/rms_norm_patch 2025-11-03T16:58:01.3579355Z * [new branch] ruisi/aot_eager_pass -> origin/ruisi/aot_eager_pass 2025-11-03T16:58:01.3580657Z * [new branch] ruisi/manual_bucket_pass -> origin/ruisi/manual_bucket_pass 2025-11-03T16:58:01.3582278Z * [new branch] ruisi/placement_trace -> origin/ruisi/placement_trace 2025-11-03T16:58:01.3585596Z * [new branch] ryanguo99/cleanup-dynamo-expected-failures -> origin/ryanguo99/cleanup-dynamo-expected-failures 2025-11-03T16:58:01.3586333Z * [new branch] ryanguo99/fix-closure-var -> origin/ryanguo99/fix-closure-var 2025-11-03T16:58:01.3589018Z * [new branch] rzou/faketensor_bench -> origin/rzou/faketensor_bench 2025-11-03T16:58:01.3590474Z * [new branch] rzou/njt -> origin/rzou/njt 2025-11-03T16:58:01.3592160Z * [new branch] rzou/pca -> origin/rzou/pca 2025-11-03T16:58:01.3593778Z * [new branch] rzou/realprop -> origin/rzou/realprop 2025-11-03T16:58:01.3595662Z * [new branch] samplevllm -> origin/samplevllm 2025-11-03T16:58:01.3598289Z * [new branch] sanchitintel/weird_thing_with_test_cpu_select_algorithm -> origin/sanchitintel/weird_thing_with_test_cpu_select_algorithm 2025-11-03T16:58:01.3599995Z * [new branch] sapling-pr-archive-SS-JIA -> origin/sapling-pr-archive-SS-JIA 2025-11-03T16:58:01.3601977Z * [new branch] sapling-pr-archive-tushar00jain -> origin/sapling-pr-archive-tushar00jain 2025-11-03T16:58:01.3603649Z * [new branch] save -> origin/save 2025-11-03T16:58:01.3605623Z * [new branch] save_github_env_rocm -> origin/save_github_env_rocm 2025-11-03T16:58:01.3607278Z * [new branch] sdpa-bs-zero -> origin/sdpa-bs-zero 2025-11-03T16:58:01.3609877Z * [new branch] sdym/2.5.1 -> origin/sdym/2.5.1 2025-11-03T16:58:01.3611854Z * [new branch] sekyondaMeta-dynamoconfig-fix -> origin/sekyondaMeta-dynamoconfig-fix 2025-11-03T16:58:01.3614052Z * [new branch] shengf/fx-xform-perf -> origin/shengf/fx-xform-perf 2025-11-03T16:58:01.3615908Z * [new branch] shoumikhin-patch-1 -> origin/shoumikhin-patch-1 2025-11-03T16:58:01.3618505Z * [new branch] solve-accuracy-fix -> origin/solve-accuracy-fix 2025-11-03T16:58:01.3621183Z * [new branch] soulitzer/reland-codev-grad-dtype -> origin/soulitzer/reland-codev-grad-dtype 2025-11-03T16:58:01.3622231Z * [new branch] soulitzer/stash-tls-ac -> origin/soulitzer/stash-tls-ac 2025-11-03T16:58:01.3624145Z * [new branch] stablize-compilation-time -> origin/stablize-compilation-time 2025-11-03T16:58:01.3625740Z * [new branch] starterTaskUpdate -> origin/starterTaskUpdate 2025-11-03T16:58:01.3627760Z * [new branch] suo -> origin/suo 2025-11-03T16:58:01.3629581Z * [new branch] sve-poc -> origin/sve-poc 2025-11-03T16:58:01.3631396Z * [new branch] switch-bn -> origin/switch-bn 2025-11-03T16:58:01.3633252Z * [new branch] sympy-bottleneck-repro -> origin/sympy-bottleneck-repro 2025-11-03T16:58:01.3635114Z * [new branch] tensordict_integration -> origin/tensordict_integration 2025-11-03T16:58:01.3636948Z * [new branch] test-move-conda-builds -> origin/test-move-conda-builds 2025-11-03T16:58:01.3638904Z * [new branch] test-myst-markdown-docstring -> origin/test-myst-markdown-docstring 2025-11-03T16:58:01.3640670Z * [new branch] test-old -> origin/test-old 2025-11-03T16:58:01.3642549Z * [new branch] test-vec-migration-internally -> origin/test-vec-migration-internally 2025-11-03T16:58:01.3644789Z * [new branch] test/bmm_heur -> origin/test/bmm_heur 2025-11-03T16:58:01.3646478Z * [new branch] test/inductor -> origin/test/inductor 2025-11-03T16:58:01.3648961Z * [new branch] tianren/customOp_autotune -> origin/tianren/customOp_autotune 2025-11-03T16:58:01.3650662Z * [new branch] tianren/customOp_autotune_fix -> origin/tianren/customOp_autotune_fix 2025-11-03T16:58:01.3652625Z * [new branch] tianren/customOp_enable_max_autotune -> origin/tianren/customOp_enable_max_autotune 2025-11-03T16:58:01.3654048Z * [new branch] tianren/customOp_fusion -> origin/tianren/customOp_fusion 2025-11-03T16:58:01.3655650Z * [new branch] tianren/flex_paged_attn_fix_temp -> origin/tianren/flex_paged_attn_fix_temp 2025-11-03T16:58:01.3657500Z * [new branch] tianren/remove_repeate -> origin/tianren/remove_repeate 2025-11-03T16:58:01.3659368Z * [new branch] tianren/test -> origin/tianren/test 2025-11-03T16:58:01.3661439Z * [new branch] tidy_performance_cyy -> origin/tidy_performance_cyy 2025-11-03T16:58:01.3663032Z * [new branch] tmp -> origin/tmp 2025-11-03T16:58:01.3664942Z * [new branch] torchtitan_ep -> origin/torchtitan_ep 2025-11-03T16:58:01.3666730Z * [new branch] trace_fsdp_torchtune_lora -> origin/trace_fsdp_torchtune_lora 2025-11-03T16:58:01.3668508Z * [new branch] traceable_fsdp_unit_tests -> origin/traceable_fsdp_unit_tests 2025-11-03T16:58:01.3670311Z * [new branch] tree_loop_vec_base -> origin/tree_loop_vec_base 2025-11-03T16:58:01.3672151Z * [new branch] triton_kernel -> origin/triton_kernel 2025-11-03T16:58:01.3674304Z * [new branch] trunk-tagging-multi-commits -> origin/trunk-tagging-multi-commits 2025-11-03T16:58:01.3676174Z * [new branch] tt_pkg_1908 -> origin/tt_pkg_1908 2025-11-03T16:58:01.3678434Z * [new branch] type_dec -> origin/type_dec 2025-11-03T16:58:01.3680266Z * [new branch] udate-sphinx-dependancies -> origin/udate-sphinx-dependancies 2025-11-03T16:58:01.3683324Z * [new branch] update-audio-commit-hash/17630256502-1803-1 -> origin/update-audio-commit-hash/17630256502-1803-1 2025-11-03T16:58:01.3684590Z * [new branch] update-audio-commit-hash/18392707270-1874-1 -> origin/update-audio-commit-hash/18392707270-1874-1 2025-11-03T16:58:01.3686152Z * [new branch] update-audio-commit-hash/18513819375-1881-1 -> origin/update-audio-commit-hash/18513819375-1881-1 2025-11-03T16:58:01.3687981Z * [new branch] update-audio-commit-hash/18607927372-1885-1 -> origin/update-audio-commit-hash/18607927372-1885-1 2025-11-03T16:58:01.3689277Z * [new branch] update-audio-commit-hash/18925946320-1900-1 -> origin/update-audio-commit-hash/18925946320-1900-1 2025-11-03T16:58:01.3692678Z * [new branch] update-executorch-commit-hash/15694981040-1626-1 -> origin/update-executorch-commit-hash/15694981040-1626-1 2025-11-03T16:58:01.3694424Z * [new branch] update-triton-commit-hash/13663274526-1487-2 -> origin/update-triton-commit-hash/13663274526-1487-2 2025-11-03T16:58:01.3696794Z * [new branch] update-vision-commit-hash/15336342773-1607-1 -> origin/update-vision-commit-hash/15336342773-1607-1 2025-11-03T16:58:01.3698343Z * [new branch] update-vision-commit-hash/18361653903-1869-1 -> origin/update-vision-commit-hash/18361653903-1869-1 2025-11-03T16:58:01.3700602Z * [new branch] update-vision-commit-hash/18513819375-1881-1 -> origin/update-vision-commit-hash/18513819375-1881-1 2025-11-03T16:58:01.3702101Z * [new branch] update-vision-commit-hash/18701484675-1891-1 -> origin/update-vision-commit-hash/18701484675-1891-1 2025-11-03T16:58:01.3704126Z * [new branch] update-vision-commit-hash/18860003792-1897-1 -> origin/update-vision-commit-hash/18860003792-1897-1 2025-11-03T16:58:01.3706018Z * [new branch] update-vision-commit-hash/18988459983-1903-1 -> origin/update-vision-commit-hash/18988459983-1903-1 2025-11-03T16:58:01.3708870Z * [new branch] update-vllm-commit-hash/18236802781-1857-1 -> origin/update-vllm-commit-hash/18236802781-1857-1 2025-11-03T16:58:01.3710112Z * [new branch] update-vllm-commit-hash/18451675449-1879-1 -> origin/update-vllm-commit-hash/18451675449-1879-1 2025-11-03T16:58:01.3712769Z * [new branch] update-xla-commit-hash/18273597034-206-1 -> origin/update-xla-commit-hash/18273597034-206-1 2025-11-03T16:58:01.3714242Z * [new branch] update-xla-commit-hash/18458620648-207-1 -> origin/update-xla-commit-hash/18458620648-207-1 2025-11-03T16:58:01.3715870Z * [new branch] update-xla-commit-hash/18645596657-208-1 -> origin/update-xla-commit-hash/18645596657-208-1 2025-11-03T16:58:01.3717835Z * [new branch] update-xla-commit-hash/19027297646-210-1 -> origin/update-xla-commit-hash/19027297646-210-1 2025-11-03T16:58:01.3720017Z * [new branch] update_docs_torch_multinomial_issue#125388 -> origin/update_docs_torch_multinomial_issue#125388 2025-11-03T16:58:01.3721555Z * [new branch] update_slow_tests_1722488736 -> origin/update_slow_tests_1722488736 2025-11-03T16:58:01.3723418Z * [new branch] update_slow_tests_1722879173 -> origin/update_slow_tests_1722879173 2025-11-03T16:58:01.3725142Z * [new branch] update_slow_tests_1759736444 -> origin/update_slow_tests_1759736444 2025-11-03T16:58:01.3726982Z * [new branch] update_slow_tests_1760341258 -> origin/update_slow_tests_1760341258 2025-11-03T16:58:01.3728811Z * [new branch] update_slow_tests_1760946059 -> origin/update_slow_tests_1760946059 2025-11-03T16:58:01.3730762Z * [new branch] update_slow_tests_1762155677 -> origin/update_slow_tests_1762155677 2025-11-03T16:58:01.3732902Z * [new branch] update_submodule_FBGEMM -> origin/update_submodule_FBGEMM 2025-11-03T16:58:01.3734648Z * [new branch] update_submodule_kineto -> origin/update_submodule_kineto 2025-11-03T16:58:01.3736502Z * [new branch] update_submodule_tensorpipe -> origin/update_submodule_tensorpipe 2025-11-03T16:58:01.3738371Z * [new branch] v0.1.2 -> origin/v0.1.2 2025-11-03T16:58:01.3740432Z * [new branch] v1.0.1 -> origin/v1.0.1 2025-11-03T16:58:01.3742316Z * [new branch] v1.0.3 -> origin/v1.0.3 2025-11-03T16:58:01.3744090Z * [new branch] v1.1.0 -> origin/v1.1.0 2025-11-03T16:58:01.3745980Z * [new branch] v1.2.0 -> origin/v1.2.0 2025-11-03T16:58:01.3748603Z * [new branch] v1.3.0 -> origin/v1.3.0 2025-11-03T16:58:01.3750396Z * [new branch] v1.3.1 -> origin/v1.3.1 2025-11-03T16:58:01.3752178Z * [new branch] validate_fn -> origin/validate_fn 2025-11-03T16:58:01.3753945Z * [new branch] validations_2.6 -> origin/validations_2.6 2025-11-03T16:58:01.3755892Z * [new branch] validations_2.8 -> origin/validations_2.8 2025-11-03T16:58:01.3758036Z * [new branch] varlen-api -> origin/varlen-api 2025-11-03T16:58:01.3759818Z * [new branch] varlen-api-backup -> origin/varlen-api-backup 2025-11-03T16:58:01.3762174Z * [new branch] viable/strict -> origin/viable/strict 2025-11-03T16:58:01.3764668Z * [new branch] vishal9-team/dtensor_parallelism_toy -> origin/vishal9-team/dtensor_parallelism_toy 2025-11-03T16:58:01.3766353Z * [new branch] vllmbuildci -> origin/vllmbuildci 2025-11-03T16:58:01.3768635Z * [new branch] vllmpin -> origin/vllmpin 2025-11-03T16:58:01.3771119Z * [new branch] wdvr/iss_145259 -> origin/wdvr/iss_145259 2025-11-03T16:58:01.3773485Z * [new branch] whc/stage2 -> origin/whc/stage2 2025-11-03T16:58:01.3775570Z * [new branch] whc/uneven -> origin/whc/uneven 2025-11-03T16:58:01.3777849Z * [new branch] whc/uneven-merge -> origin/whc/uneven-merge 2025-11-03T16:58:01.3779713Z * [new branch] win_12_8_build -> origin/win_12_8_build 2025-11-03T16:58:01.3781339Z * [new branch] win_ci -> origin/win_ci 2025-11-03T16:58:01.3783211Z * [new branch] win_warnings -> origin/win_warnings 2025-11-03T16:58:01.3785422Z * [new branch] windows_libtorch_free -> origin/windows_libtorch_free 2025-11-03T16:58:01.3787416Z * [new branch] windows_mmap -> origin/windows_mmap 2025-11-03T16:58:01.3789123Z * [new branch] xmfan-war -> origin/xmfan-war 2025-11-03T16:58:01.3791449Z * [new branch] xmfan/ca_0516 -> origin/xmfan/ca_0516 2025-11-03T16:58:01.3793055Z * [new branch] xmfan/ca_1051b93192 -> origin/xmfan/ca_1051b93192 2025-11-03T16:58:01.3794759Z * [new branch] xmfan/ca_1a722f62c248391fc4a542e8851a5559aa356ae8 -> origin/xmfan/ca_1a722f62c248391fc4a542e8851a5559aa356ae8 2025-11-03T16:58:01.3796226Z * [new branch] xmfan/ca_5a2be192d1 -> origin/xmfan/ca_5a2be192d1 2025-11-03T16:58:01.3797852Z * [new branch] xmfan/ca_9d59b516e9 -> origin/xmfan/ca_9d59b516e9 2025-11-03T16:58:01.3799487Z * [new branch] xmfan/ca_api -> origin/xmfan/ca_api 2025-11-03T16:58:01.3801075Z * [new branch] xmfan/ca_apr8 -> origin/xmfan/ca_apr8 2025-11-03T16:58:01.3803159Z * [new branch] xmfan/ca_base -> origin/xmfan/ca_base 2025-11-03T16:58:01.3805669Z * [new branch] xmfan/ca_dynamic -> origin/xmfan/ca_dynamic 2025-11-03T16:58:01.3807637Z * [new branch] xmfan/ca_fix_dyn -> origin/xmfan/ca_fix_dyn 2025-11-03T16:58:01.3809587Z * [new branch] xmfan/ca_fix_lowering -> origin/xmfan/ca_fix_lowering 2025-11-03T16:58:01.3811732Z * [new branch] xmfan/ca_fix_polyfills -> origin/xmfan/ca_fix_polyfills 2025-11-03T16:58:01.3813519Z * [new branch] xmfan/ca_jan3 -> origin/xmfan/ca_jan3 2025-11-03T16:58:01.3815184Z * [new branch] xmfan/ca_jun18 -> origin/xmfan/ca_jun18 2025-11-03T16:58:01.3816799Z * [new branch] xmfan/ca_jun24 -> origin/xmfan/ca_jun24 2025-11-03T16:58:01.3818490Z * [new branch] xmfan/ca_nested -> origin/xmfan/ca_nested 2025-11-03T16:58:01.3820195Z * [new branch] xmfan/ca_overhead -> origin/xmfan/ca_overhead 2025-11-03T16:58:01.3822010Z * [new branch] xmfan/ca_overhead_0eba7e5451 -> origin/xmfan/ca_overhead_0eba7e5451 2025-11-03T16:58:01.3823665Z * [new branch] xmfan/cacu_jun18 -> origin/xmfan/cacu_jun18 2025-11-03T16:58:01.3825338Z * [new branch] xmfan/cacu_jun19 -> origin/xmfan/cacu_jun19 2025-11-03T16:58:01.3827046Z * [new branch] xmfan/cacu_jun4 -> origin/xmfan/cacu_jun4 2025-11-03T16:58:01.3828880Z * [new branch] xmfan/cacu_may27 -> origin/xmfan/cacu_may27 2025-11-03T16:58:01.3830630Z * [new branch] xmfan/disable_duck_shape -> origin/xmfan/disable_duck_shape 2025-11-03T16:58:01.3832492Z * [new branch] xmfan/fca_cpp_node_passthrough -> origin/xmfan/fca_cpp_node_passthrough 2025-11-03T16:58:01.3834199Z * [new branch] xmfan/post_3945954741e2d37023c5d6954f9483008e0892f9 -> origin/xmfan/post_3945954741e2d37023c5d6954f9483008e0892f9 2025-11-03T16:58:01.3836094Z * [new branch] xmfan/pre_3945954741e2d37023c5d6954f9483008e0892f9 -> origin/xmfan/pre_3945954741e2d37023c5d6954f9483008e0892f9 2025-11-03T16:58:01.3837646Z * [new branch] xmfan/single_step -> origin/xmfan/single_step 2025-11-03T16:58:01.3839226Z * [new branch] xmfan/sth_0829 -> origin/xmfan/sth_0829 2025-11-03T16:58:01.3841290Z * [new branch] xmfan/test -> origin/xmfan/test 2025-11-03T16:58:01.3843996Z * [new branch] yguo/debug-0226-constexpr -> origin/yguo/debug-0226-constexpr 2025-11-03T16:58:01.3845584Z * [new branch] yguo/new_latest_changes -> origin/yguo/new_latest_changes 2025-11-03T16:58:01.3847199Z * [new branch] yguo/patch_constexpr_changes -> origin/yguo/patch_constexpr_changes 2025-11-03T16:58:01.3848936Z * [new branch] yihan_quantization -> origin/yihan_quantization 2025-11-03T16:58:01.3851299Z * [new branch] yiming/add_timm_models -> origin/yiming/add_timm_models 2025-11-03T16:58:01.3852882Z * [new branch] yiming/bootcamp -> origin/yiming/bootcamp 2025-11-03T16:58:01.3854596Z * [new branch] yiming/fix_aot_joint_graph_capture_test -> origin/yiming/fix_aot_joint_graph_capture_test 2025-11-03T16:58:01.3856355Z * [new branch] yiming/flex_attention_export_with_kwargs -> origin/yiming/flex_attention_export_with_kwargs 2025-11-03T16:58:01.3858353Z * [new branch] yiming/improve_custom_op_stack_trace -> origin/yiming/improve_custom_op_stack_trace 2025-11-03T16:58:01.3861283Z * [new branch] yiming/improve_sharding_error_msg -> origin/yiming/improve_sharding_error_msg 2025-11-03T16:58:01.3862531Z * [new branch] yiming/moe_dtensor_region_annotation -> origin/yiming/moe_dtensor_region_annotation 2025-11-03T16:58:01.3864163Z * [new branch] yiming/precompile_benchmark -> origin/yiming/precompile_benchmark 2025-11-03T16:58:01.3866291Z * [new branch] yiming/remove_blockmask_pytree_global_registration -> origin/yiming/remove_blockmask_pytree_global_registration 2025-11-03T16:58:01.3867696Z * [new branch] yolo-llama3 -> origin/yolo-llama3 2025-11-03T16:58:01.3870809Z * [new branch] yuxin/fix-save-memleak -> origin/yuxin/fix-save-memleak 2025-11-03T16:58:01.3872740Z * [new branch] zainr/canary-test -> origin/zainr/canary-test 2025-11-03T16:58:01.3874577Z * [new branch] zainr/cleanup-gh-runners -> origin/zainr/cleanup-gh-runners 2025-11-03T16:58:01.3876180Z * [new branch] zainr/pull-migration-c -> origin/zainr/pull-migration-c 2025-11-03T16:58:01.3877665Z * [new branch] zainr/test2 -> origin/zainr/test2 2025-11-03T16:58:01.3879880Z * [new branch] zainr/unstable -> origin/zainr/unstable 2025-11-03T16:58:01.3881297Z * [new branch] zasdfgbnm-patch-3 -> origin/zasdfgbnm-patch-3 2025-11-03T16:58:01.3882993Z * [new branch] zb2p -> origin/zb2p 2025-11-03T16:58:01.3884944Z * [new branch] zeros-and-scatter-part2 -> origin/zeros-and-scatter-part2 2025-11-03T16:58:01.3887918Z * [new branch] zhxchen17/aot_compile_fix_load_guard_manager -> origin/zhxchen17/aot_compile_fix_load_guard_manager 2025-11-03T16:58:01.3889615Z * [new branch] zhxchen17/ci/vllm_pin -> origin/zhxchen17/ci/vllm_pin 2025-11-03T16:58:01.3891947Z * [new branch] zhxchen17/precompile/closure_serde -> origin/zhxchen17/precompile/closure_serde 2025-11-03T16:58:01.3893642Z * [new branch] zhxchen17/precompile/default_args -> origin/zhxchen17/precompile/default_args 2025-11-03T16:58:01.3895433Z * [new branch] zhxchen17/precompile/export -> origin/zhxchen17/precompile/export 2025-11-03T16:58:01.3897083Z * [new branch] zhxchen17/precompile/export_gm -> origin/zhxchen17/precompile/export_gm 2025-11-03T16:58:01.3898806Z * [new branch] zhxchen17/precompile_vllm_fix_0 -> origin/zhxchen17/precompile_vllm_fix_0 2025-11-03T16:58:01.3901198Z * [new branch] zhxchen17/scratch/0 -> origin/zhxchen17/scratch/0 2025-11-03T16:58:01.3903571Z * [new branch] zhxhcen17/moodycamel -> origin/zhxhcen17/moodycamel 2025-11-03T16:58:01.3905752Z * [new branch] zxiiro/build-times -> origin/zxiiro/build-times 2025-11-03T16:58:01.3907448Z * [new branch] zxiiro/c7i-docs-build -> origin/zxiiro/c7i-docs-build 2025-11-03T16:58:01.3909199Z * [new branch] zxiiro/c7i-linux-4xlarge -> origin/zxiiro/c7i-linux-4xlarge 2025-11-03T16:58:01.3911052Z * [new branch] zxiiro/c7i-linux-build-yaml -> origin/zxiiro/c7i-linux-build-yaml 2025-11-03T16:58:01.3912757Z * [new branch] zxiiro/c7i-pull -> origin/zxiiro/c7i-pull 2025-11-03T16:58:01.3914509Z * [new branch] zxiiro/main -> origin/zxiiro/main 2025-11-03T16:58:01.3916237Z * [new branch] zxiiro/test-multicloud-arc -> origin/zxiiro/test-multicloud-arc 2025-11-03T16:58:01.3917718Z * [new tag] bc2caa7fdf006894eff7af936babde69ab5a40f8-huydhn-debug -> bc2caa7fdf006894eff7af936babde69ab5a40f8-huydhn-debug 2025-11-03T16:58:01.3919500Z * [new tag] ci/binaries/77164 -> ci/binaries/77164 2025-11-03T16:58:01.3920959Z * [new tag] ciflow/b200/115316 -> ciflow/b200/115316 2025-11-03T16:58:01.3922047Z * [new tag] ciflow/b200/157553 -> ciflow/b200/157553 2025-11-03T16:58:01.3923147Z * [new tag] ciflow/b200/157635 -> ciflow/b200/157635 2025-11-03T16:58:01.3924283Z * [new tag] ciflow/b200/160685 -> ciflow/b200/160685 2025-11-03T16:58:01.3925410Z * [new tag] ciflow/b200/161404 -> ciflow/b200/161404 2025-11-03T16:58:01.3926590Z * [new tag] ciflow/b200/163369 -> ciflow/b200/163369 2025-11-03T16:58:01.3927820Z * [new tag] ciflow/b200/164657 -> ciflow/b200/164657 2025-11-03T16:58:01.3929150Z * [new tag] ciflow/b200/164938 -> ciflow/b200/164938 2025-11-03T16:58:01.3930274Z * [new tag] ciflow/b200/165132 -> ciflow/b200/165132 2025-11-03T16:58:01.3931514Z * [new tag] ciflow/b200/165541 -> ciflow/b200/165541 2025-11-03T16:58:01.3932642Z * [new tag] ciflow/b200/165548 -> ciflow/b200/165548 2025-11-03T16:58:01.3934127Z * [new tag] ciflow/b200/165952 -> ciflow/b200/165952 2025-11-03T16:58:01.3935665Z * [new tag] ciflow/b200/166250 -> ciflow/b200/166250 2025-11-03T16:58:01.3937151Z * [new tag] ciflow/b200/166526 -> ciflow/b200/166526 2025-11-03T16:58:01.3939095Z * [new tag] ciflow/b200/166744 -> ciflow/b200/166744 2025-11-03T16:58:01.3940696Z * [new tag] ciflow/b200/166752 -> ciflow/b200/166752 2025-11-03T16:58:01.3941876Z * [new tag] ciflow/b200/166834 -> ciflow/b200/166834 2025-11-03T16:58:01.3943307Z * [new tag] ciflow/binaries/157432 -> ciflow/binaries/157432 2025-11-03T16:58:01.3944427Z * [new tag] ciflow/binaries/158104 -> ciflow/binaries/158104 2025-11-03T16:58:01.3945730Z * [new tag] ciflow/binaries/165922 -> ciflow/binaries/165922 2025-11-03T16:58:01.3947199Z * [new tag] ciflow/binaries/166044 -> ciflow/binaries/166044 2025-11-03T16:58:01.3948210Z * [new tag] ciflow/binaries/166621 -> ciflow/binaries/166621 2025-11-03T16:58:01.3949630Z * [new tag] ciflow/binaries/166764 -> ciflow/binaries/166764 2025-11-03T16:58:01.3950924Z * [new tag] ciflow/binaries_libtorch/157432 -> ciflow/binaries_libtorch/157432 2025-11-03T16:58:01.3952468Z * [new tag] ciflow/binaries_wheel/155731 -> ciflow/binaries_wheel/155731 2025-11-03T16:58:01.3953621Z * [new tag] ciflow/binaries_wheel/157432 -> ciflow/binaries_wheel/157432 2025-11-03T16:58:01.3954940Z * [new tag] ciflow/binaries_wheel/166380 -> ciflow/binaries_wheel/166380 2025-11-03T16:58:01.3956129Z * [new tag] ciflow/binaries_wheel/166407 -> ciflow/binaries_wheel/166407 2025-11-03T16:58:01.3957598Z * [new tag] ciflow/binaries_wheel/166829 -> ciflow/binaries_wheel/166829 2025-11-03T16:58:01.3958821Z * [new tag] ciflow/h100-distributed/166367 -> ciflow/h100-distributed/166367 2025-11-03T16:58:01.3960281Z * [new tag] ciflow/h100-symm-mem/151845 -> ciflow/h100-symm-mem/151845 2025-11-03T16:58:01.3961449Z * [new tag] ciflow/h100-symm-mem/157635 -> ciflow/h100-symm-mem/157635 2025-11-03T16:58:01.3962608Z * [new tag] ciflow/h100-symm-mem/163814 -> ciflow/h100-symm-mem/163814 2025-11-03T16:58:01.3963777Z * [new tag] ciflow/h100-symm-mem/163815 -> ciflow/h100-symm-mem/163815 2025-11-03T16:58:01.3964796Z * [new tag] ciflow/h100-symm-mem/165548 -> ciflow/h100-symm-mem/165548 2025-11-03T16:58:01.3966004Z * [new tag] ciflow/h100-symm-mem/166772 -> ciflow/h100-symm-mem/166772 2025-11-03T16:58:01.3966985Z * [new tag] ciflow/h100-symm-mem/166773 -> ciflow/h100-symm-mem/166773 2025-11-03T16:58:01.3968268Z * [new tag] ciflow/h100-symm-mem/166774 -> ciflow/h100-symm-mem/166774 2025-11-03T16:58:01.3969765Z * [new tag] ciflow/h100/115316 -> ciflow/h100/115316 2025-11-03T16:58:01.3970870Z * [new tag] ciflow/h100/157553 -> ciflow/h100/157553 2025-11-03T16:58:01.3972008Z * [new tag] ciflow/h100/157635 -> ciflow/h100/157635 2025-11-03T16:58:01.3973162Z * [new tag] ciflow/h100/160685 -> ciflow/h100/160685 2025-11-03T16:58:01.3974265Z * [new tag] ciflow/h100/161404 -> ciflow/h100/161404 2025-11-03T16:58:01.3975409Z * [new tag] ciflow/h100/163369 -> ciflow/h100/163369 2025-11-03T16:58:01.3976552Z * [new tag] ciflow/h100/164657 -> ciflow/h100/164657 2025-11-03T16:58:01.3977790Z * [new tag] ciflow/h100/165029 -> ciflow/h100/165029 2025-11-03T16:58:01.3979004Z * [new tag] ciflow/h100/165132 -> ciflow/h100/165132 2025-11-03T16:58:01.3980348Z * [new tag] ciflow/h100/165541 -> ciflow/h100/165541 2025-11-03T16:58:01.3981085Z * [new tag] ciflow/h100/165548 -> ciflow/h100/165548 2025-11-03T16:58:01.3982988Z * [new tag] ciflow/h100/165952 -> ciflow/h100/165952 2025-11-03T16:58:01.3984316Z * [new tag] ciflow/h100/166250 -> ciflow/h100/166250 2025-11-03T16:58:01.3985898Z * [new tag] ciflow/h100/166515 -> ciflow/h100/166515 2025-11-03T16:58:01.3987497Z * [new tag] ciflow/h100/166526 -> ciflow/h100/166526 2025-11-03T16:58:01.3988842Z * [new tag] ciflow/h100/166744 -> ciflow/h100/166744 2025-11-03T16:58:01.3990078Z * [new tag] ciflow/h100/166752 -> ciflow/h100/166752 2025-11-03T16:58:01.3991299Z * [new tag] ciflow/h100/166834 -> ciflow/h100/166834 2025-11-03T16:58:01.3992795Z * [new tag] ciflow/inductor-perf-compare/165029 -> ciflow/inductor-perf-compare/165029 2025-11-03T16:58:01.3994692Z * [new tag] ciflow/inductor-perf-test-nightly-rocm-mi300/162053 -> ciflow/inductor-perf-test-nightly-rocm-mi300/162053 2025-11-03T16:58:01.3996093Z * [new tag] ciflow/inductor-perf-test-nightly-x86-zen/162954 -> ciflow/inductor-perf-test-nightly-x86-zen/162954 2025-11-03T16:58:01.3997094Z * [new tag] ciflow/inductor-perf-test-nightly-x86-zen/164126 -> ciflow/inductor-perf-test-nightly-x86-zen/164126 2025-11-03T16:58:01.3999330Z * [new tag] ciflow/inductor-perf-test-nightly/162053 -> ciflow/inductor-perf-test-nightly/162053 2025-11-03T16:58:01.4000371Z * [new tag] ciflow/inductor-perf-test-nightly/165029 -> ciflow/inductor-perf-test-nightly/165029 2025-11-03T16:58:01.4001373Z * [new tag] ciflow/inductor-periodic/165029 -> ciflow/inductor-periodic/165029 2025-11-03T16:58:01.4002638Z * [new tag] ciflow/inductor-periodic/166743 -> ciflow/inductor-periodic/166743 2025-11-03T16:58:01.4003976Z * [new tag] ciflow/inductor-rocm/151845 -> ciflow/inductor-rocm/151845 2025-11-03T16:58:01.4005121Z * [new tag] ciflow/inductor-rocm/162053 -> ciflow/inductor-rocm/162053 2025-11-03T16:58:01.4006202Z * [new tag] ciflow/inductor-rocm/166515 -> ciflow/inductor-rocm/166515 2025-11-03T16:58:01.4007507Z * [new tag] ciflow/inductor-rocm/166743 -> ciflow/inductor-rocm/166743 2025-11-03T16:58:01.4008953Z * [new tag] ciflow/inductor-rocm/166870 -> ciflow/inductor-rocm/166870 2025-11-03T16:58:01.4010737Z * [new tag] ciflow/inductor/137400 -> ciflow/inductor/137400 2025-11-03T16:58:01.4011857Z * [new tag] ciflow/inductor/148180 -> ciflow/inductor/148180 2025-11-03T16:58:01.4013126Z * [new tag] ciflow/inductor/148294 -> ciflow/inductor/148294 2025-11-03T16:58:01.4014381Z * [new tag] ciflow/inductor/148328 -> ciflow/inductor/148328 2025-11-03T16:58:01.4015480Z * [new tag] ciflow/inductor/148484 -> ciflow/inductor/148484 2025-11-03T16:58:01.4016606Z * [new tag] ciflow/inductor/148492 -> ciflow/inductor/148492 2025-11-03T16:58:01.4017698Z * [new tag] ciflow/inductor/151845 -> ciflow/inductor/151845 2025-11-03T16:58:01.4018864Z * [new tag] ciflow/inductor/152624 -> ciflow/inductor/152624 2025-11-03T16:58:01.4019973Z * [new tag] ciflow/inductor/157635 -> ciflow/inductor/157635 2025-11-03T16:58:01.4021093Z * [new tag] ciflow/inductor/158104 -> ciflow/inductor/158104 2025-11-03T16:58:01.4022748Z * [new tag] ciflow/inductor/159523 -> ciflow/inductor/159523 2025-11-03T16:58:01.4024229Z * [new tag] ciflow/inductor/160174 -> ciflow/inductor/160174 2025-11-03T16:58:01.4025716Z * [new tag] ciflow/inductor/160685 -> ciflow/inductor/160685 2025-11-03T16:58:01.4027138Z * [new tag] ciflow/inductor/160686 -> ciflow/inductor/160686 2025-11-03T16:58:01.4028495Z * [new tag] ciflow/inductor/160687 -> ciflow/inductor/160687 2025-11-03T16:58:01.4029762Z * [new tag] ciflow/inductor/160688 -> ciflow/inductor/160688 2025-11-03T16:58:01.4030977Z * [new tag] ciflow/inductor/160706 -> ciflow/inductor/160706 2025-11-03T16:58:01.4032196Z * [new tag] ciflow/inductor/160729 -> ciflow/inductor/160729 2025-11-03T16:58:01.4033411Z * [new tag] ciflow/inductor/161158 -> ciflow/inductor/161158 2025-11-03T16:58:01.4034793Z * [new tag] ciflow/inductor/161246 -> ciflow/inductor/161246 2025-11-03T16:58:01.4036054Z * [new tag] ciflow/inductor/161404 -> ciflow/inductor/161404 2025-11-03T16:58:01.4037282Z * [new tag] ciflow/inductor/161495 -> ciflow/inductor/161495 2025-11-03T16:58:01.4038557Z * [new tag] ciflow/inductor/161512 -> ciflow/inductor/161512 2025-11-03T16:58:01.4039792Z * [new tag] ciflow/inductor/161938 -> ciflow/inductor/161938 2025-11-03T16:58:01.4040996Z * [new tag] ciflow/inductor/161939 -> ciflow/inductor/161939 2025-11-03T16:58:01.4042240Z * [new tag] ciflow/inductor/161940 -> ciflow/inductor/161940 2025-11-03T16:58:01.4043430Z * [new tag] ciflow/inductor/162053 -> ciflow/inductor/162053 2025-11-03T16:58:01.4044685Z * [new tag] ciflow/inductor/162469 -> ciflow/inductor/162469 2025-11-03T16:58:01.4046001Z * [new tag] ciflow/inductor/162709 -> ciflow/inductor/162709 2025-11-03T16:58:01.4047163Z * [new tag] ciflow/inductor/162795 -> ciflow/inductor/162795 2025-11-03T16:58:01.4048277Z * [new tag] ciflow/inductor/162954 -> ciflow/inductor/162954 2025-11-03T16:58:01.4049737Z * [new tag] ciflow/inductor/163249 -> ciflow/inductor/163249 2025-11-03T16:58:01.4050928Z * [new tag] ciflow/inductor/163335 -> ciflow/inductor/163335 2025-11-03T16:58:01.4052135Z * [new tag] ciflow/inductor/163369 -> ciflow/inductor/163369 2025-11-03T16:58:01.4053338Z * [new tag] ciflow/inductor/163490 -> ciflow/inductor/163490 2025-11-03T16:58:01.4054634Z * [new tag] ciflow/inductor/163503 -> ciflow/inductor/163503 2025-11-03T16:58:01.4055840Z * [new tag] ciflow/inductor/163714 -> ciflow/inductor/163714 2025-11-03T16:58:01.4057051Z * [new tag] ciflow/inductor/163936 -> ciflow/inductor/163936 2025-11-03T16:58:01.4058245Z * [new tag] ciflow/inductor/164202 -> ciflow/inductor/164202 2025-11-03T16:58:01.4059490Z * [new tag] ciflow/inductor/164384 -> ciflow/inductor/164384 2025-11-03T16:58:01.4060654Z * [new tag] ciflow/inductor/164657 -> ciflow/inductor/164657 2025-11-03T16:58:01.4061870Z * [new tag] ciflow/inductor/164938 -> ciflow/inductor/164938 2025-11-03T16:58:01.4063115Z * [new tag] ciflow/inductor/164979 -> ciflow/inductor/164979 2025-11-03T16:58:01.4064353Z * [new tag] ciflow/inductor/164980 -> ciflow/inductor/164980 2025-11-03T16:58:01.4065654Z * [new tag] ciflow/inductor/165005 -> ciflow/inductor/165005 2025-11-03T16:58:01.4066843Z * [new tag] ciflow/inductor/165029 -> ciflow/inductor/165029 2025-11-03T16:58:01.4068049Z * [new tag] ciflow/inductor/165092 -> ciflow/inductor/165092 2025-11-03T16:58:01.4069276Z * [new tag] ciflow/inductor/165132 -> ciflow/inductor/165132 2025-11-03T16:58:01.4070465Z * [new tag] ciflow/inductor/165197 -> ciflow/inductor/165197 2025-11-03T16:58:01.4071677Z * [new tag] ciflow/inductor/165274 -> ciflow/inductor/165274 2025-11-03T16:58:01.4072936Z * [new tag] ciflow/inductor/165283 -> ciflow/inductor/165283 2025-11-03T16:58:01.4074173Z * [new tag] ciflow/inductor/165284 -> ciflow/inductor/165284 2025-11-03T16:58:01.4075506Z * [new tag] ciflow/inductor/165367 -> ciflow/inductor/165367 2025-11-03T16:58:01.4076841Z * [new tag] ciflow/inductor/165423 -> ciflow/inductor/165423 2025-11-03T16:58:01.4078087Z * [new tag] ciflow/inductor/165487 -> ciflow/inductor/165487 2025-11-03T16:58:01.4079305Z * [new tag] ciflow/inductor/165541 -> ciflow/inductor/165541 2025-11-03T16:58:01.4080528Z * [new tag] ciflow/inductor/165597 -> ciflow/inductor/165597 2025-11-03T16:58:01.4081944Z * [new tag] ciflow/inductor/165686 -> ciflow/inductor/165686 2025-11-03T16:58:01.4083334Z * [new tag] ciflow/inductor/165790 -> ciflow/inductor/165790 2025-11-03T16:58:01.4084614Z * [new tag] ciflow/inductor/165824 -> ciflow/inductor/165824 2025-11-03T16:58:01.4085833Z * [new tag] ciflow/inductor/165856 -> ciflow/inductor/165856 2025-11-03T16:58:01.4087054Z * [new tag] ciflow/inductor/165885 -> ciflow/inductor/165885 2025-11-03T16:58:01.4088258Z * [new tag] ciflow/inductor/165952 -> ciflow/inductor/165952 2025-11-03T16:58:01.4089591Z * [new tag] ciflow/inductor/165953 -> ciflow/inductor/165953 2025-11-03T16:58:01.4090779Z * [new tag] ciflow/inductor/166071 -> ciflow/inductor/166071 2025-11-03T16:58:01.4092205Z * [new tag] ciflow/inductor/166083 -> ciflow/inductor/166083 2025-11-03T16:58:01.4093376Z * [new tag] ciflow/inductor/166170 -> ciflow/inductor/166170 2025-11-03T16:58:01.4094753Z * [new tag] ciflow/inductor/166198 -> ciflow/inductor/166198 2025-11-03T16:58:01.4096094Z * [new tag] ciflow/inductor/166225 -> ciflow/inductor/166225 2025-11-03T16:58:01.4097316Z * [new tag] ciflow/inductor/166236 -> ciflow/inductor/166236 2025-11-03T16:58:01.4098581Z * [new tag] ciflow/inductor/166250 -> ciflow/inductor/166250 2025-11-03T16:58:01.4099876Z * [new tag] ciflow/inductor/166254 -> ciflow/inductor/166254 2025-11-03T16:58:01.4101095Z * [new tag] ciflow/inductor/166276 -> ciflow/inductor/166276 2025-11-03T16:58:01.4102324Z * [new tag] ciflow/inductor/166294 -> ciflow/inductor/166294 2025-11-03T16:58:01.4104025Z * [new tag] ciflow/inductor/166300 -> ciflow/inductor/166300 2025-11-03T16:58:01.4105285Z * [new tag] ciflow/inductor/166320 -> ciflow/inductor/166320 2025-11-03T16:58:01.4106455Z * [new tag] ciflow/inductor/166321 -> ciflow/inductor/166321 2025-11-03T16:58:01.4107683Z * [new tag] ciflow/inductor/166337 -> ciflow/inductor/166337 2025-11-03T16:58:01.4109104Z * [new tag] ciflow/inductor/166341 -> ciflow/inductor/166341 2025-11-03T16:58:01.4110489Z * [new tag] ciflow/inductor/166354 -> ciflow/inductor/166354 2025-11-03T16:58:01.4111731Z * [new tag] ciflow/inductor/166355 -> ciflow/inductor/166355 2025-11-03T16:58:01.4113090Z * [new tag] ciflow/inductor/166361 -> ciflow/inductor/166361 2025-11-03T16:58:01.4114369Z * [new tag] ciflow/inductor/166366 -> ciflow/inductor/166366 2025-11-03T16:58:01.4115565Z * [new tag] ciflow/inductor/166367 -> ciflow/inductor/166367 2025-11-03T16:58:01.4116808Z * [new tag] ciflow/inductor/166370 -> ciflow/inductor/166370 2025-11-03T16:58:01.4117986Z * [new tag] ciflow/inductor/166371 -> ciflow/inductor/166371 2025-11-03T16:58:01.4119235Z * [new tag] ciflow/inductor/166372 -> ciflow/inductor/166372 2025-11-03T16:58:01.4120548Z * [new tag] ciflow/inductor/166377 -> ciflow/inductor/166377 2025-11-03T16:58:01.4121772Z * [new tag] ciflow/inductor/166400 -> ciflow/inductor/166400 2025-11-03T16:58:01.4122973Z * [new tag] ciflow/inductor/166402 -> ciflow/inductor/166402 2025-11-03T16:58:01.4124233Z * [new tag] ciflow/inductor/166417 -> ciflow/inductor/166417 2025-11-03T16:58:01.4125456Z * [new tag] ciflow/inductor/166433 -> ciflow/inductor/166433 2025-11-03T16:58:01.4126837Z * [new tag] ciflow/inductor/166442 -> ciflow/inductor/166442 2025-11-03T16:58:01.4128100Z * [new tag] ciflow/inductor/166467 -> ciflow/inductor/166467 2025-11-03T16:58:01.4129407Z * [new tag] ciflow/inductor/166479 -> ciflow/inductor/166479 2025-11-03T16:58:01.4130691Z * [new tag] ciflow/inductor/166482 -> ciflow/inductor/166482 2025-11-03T16:58:01.4131869Z * [new tag] ciflow/inductor/166483 -> ciflow/inductor/166483 2025-11-03T16:58:01.4133098Z * [new tag] ciflow/inductor/166494 -> ciflow/inductor/166494 2025-11-03T16:58:01.4134439Z * [new tag] ciflow/inductor/166498 -> ciflow/inductor/166498 2025-11-03T16:58:01.4135854Z * [new tag] ciflow/inductor/166499 -> ciflow/inductor/166499 2025-11-03T16:58:01.4137083Z * [new tag] ciflow/inductor/166515 -> ciflow/inductor/166515 2025-11-03T16:58:01.4138474Z * [new tag] ciflow/inductor/166536 -> ciflow/inductor/166536 2025-11-03T16:58:01.4139730Z * [new tag] ciflow/inductor/166540 -> ciflow/inductor/166540 2025-11-03T16:58:01.4141020Z * [new tag] ciflow/inductor/166541 -> ciflow/inductor/166541 2025-11-03T16:58:01.4142222Z * [new tag] ciflow/inductor/166545 -> ciflow/inductor/166545 2025-11-03T16:58:01.4143466Z * [new tag] ciflow/inductor/166581 -> ciflow/inductor/166581 2025-11-03T16:58:01.4144685Z * [new tag] ciflow/inductor/166584 -> ciflow/inductor/166584 2025-11-03T16:58:01.4145906Z * [new tag] ciflow/inductor/166593 -> ciflow/inductor/166593 2025-11-03T16:58:01.4147309Z * [new tag] ciflow/inductor/166600 -> ciflow/inductor/166600 2025-11-03T16:58:01.4148586Z * [new tag] ciflow/inductor/166608 -> ciflow/inductor/166608 2025-11-03T16:58:01.4149822Z * [new tag] ciflow/inductor/166610 -> ciflow/inductor/166610 2025-11-03T16:58:01.4151011Z * [new tag] ciflow/inductor/166629 -> ciflow/inductor/166629 2025-11-03T16:58:01.4152405Z * [new tag] ciflow/inductor/166640 -> ciflow/inductor/166640 2025-11-03T16:58:01.4153775Z * [new tag] ciflow/inductor/166648 -> ciflow/inductor/166648 2025-11-03T16:58:01.4155162Z * [new tag] ciflow/inductor/166658 -> ciflow/inductor/166658 2025-11-03T16:58:01.4156402Z * [new tag] ciflow/inductor/166659 -> ciflow/inductor/166659 2025-11-03T16:58:01.4157838Z * [new tag] ciflow/inductor/166662 -> ciflow/inductor/166662 2025-11-03T16:58:01.4159047Z * [new tag] ciflow/inductor/166663 -> ciflow/inductor/166663 2025-11-03T16:58:01.4160278Z * [new tag] ciflow/inductor/166664 -> ciflow/inductor/166664 2025-11-03T16:58:01.4161508Z * [new tag] ciflow/inductor/166669 -> ciflow/inductor/166669 2025-11-03T16:58:01.4162742Z * [new tag] ciflow/inductor/166673 -> ciflow/inductor/166673 2025-11-03T16:58:01.4163994Z * [new tag] ciflow/inductor/166674 -> ciflow/inductor/166674 2025-11-03T16:58:01.4165238Z * [new tag] ciflow/inductor/166676 -> ciflow/inductor/166676 2025-11-03T16:58:01.4166456Z * [new tag] ciflow/inductor/166679 -> ciflow/inductor/166679 2025-11-03T16:58:01.4167821Z * [new tag] ciflow/inductor/166701 -> ciflow/inductor/166701 2025-11-03T16:58:01.4169102Z * [new tag] ciflow/inductor/166720 -> ciflow/inductor/166720 2025-11-03T16:58:01.4170364Z * [new tag] ciflow/inductor/166738 -> ciflow/inductor/166738 2025-11-03T16:58:01.4171574Z * [new tag] ciflow/inductor/166740 -> ciflow/inductor/166740 2025-11-03T16:58:01.4172811Z * [new tag] ciflow/inductor/166741 -> ciflow/inductor/166741 2025-11-03T16:58:01.4174153Z * [new tag] ciflow/inductor/166747 -> ciflow/inductor/166747 2025-11-03T16:58:01.4175410Z * [new tag] ciflow/inductor/166757 -> ciflow/inductor/166757 2025-11-03T16:58:01.4176707Z * [new tag] ciflow/inductor/166762 -> ciflow/inductor/166762 2025-11-03T16:58:01.4177995Z * [new tag] ciflow/inductor/166775 -> ciflow/inductor/166775 2025-11-03T16:58:01.4179242Z * [new tag] ciflow/inductor/166776 -> ciflow/inductor/166776 2025-11-03T16:58:01.4180446Z * [new tag] ciflow/inductor/166777 -> ciflow/inductor/166777 2025-11-03T16:58:01.4181656Z * [new tag] ciflow/inductor/166788 -> ciflow/inductor/166788 2025-11-03T16:58:01.4183016Z * [new tag] ciflow/inductor/166791 -> ciflow/inductor/166791 2025-11-03T16:58:01.4184294Z * [new tag] ciflow/inductor/166793 -> ciflow/inductor/166793 2025-11-03T16:58:01.4185995Z * [new tag] ciflow/inductor/166803 -> ciflow/inductor/166803 2025-11-03T16:58:01.4187193Z * [new tag] ciflow/inductor/166808 -> ciflow/inductor/166808 2025-11-03T16:58:01.4188454Z * [new tag] ciflow/inductor/166822 -> ciflow/inductor/166822 2025-11-03T16:58:01.4189805Z * [new tag] ciflow/inductor/166834 -> ciflow/inductor/166834 2025-11-03T16:58:01.4191077Z * [new tag] ciflow/inductor/166839 -> ciflow/inductor/166839 2025-11-03T16:58:01.4192408Z * [new tag] ciflow/inductor/166846 -> ciflow/inductor/166846 2025-11-03T16:58:01.4193621Z * [new tag] ciflow/inductor/166855 -> ciflow/inductor/166855 2025-11-03T16:58:01.4194905Z * [new tag] ciflow/inductor/166867 -> ciflow/inductor/166867 2025-11-03T16:58:01.4196101Z * [new tag] ciflow/inductor/166868 -> ciflow/inductor/166868 2025-11-03T16:58:01.4197349Z * [new tag] ciflow/inductor/166869 -> ciflow/inductor/166869 2025-11-03T16:58:01.4199013Z * [new tag] ciflow/inductor/3b9a386 -> ciflow/inductor/3b9a386 2025-11-03T16:58:01.4200433Z * [new tag] ciflow/inductor/3d4b92b -> ciflow/inductor/3d4b92b 2025-11-03T16:58:01.4201764Z * [new tag] ciflow/inductor/d224ac7 -> ciflow/inductor/d224ac7 2025-11-03T16:58:01.4203191Z * [new tag] ciflow/linux-aarch64/166380 -> ciflow/linux-aarch64/166380 2025-11-03T16:58:01.4204356Z * [new tag] ciflow/linux-aarch64/166441 -> ciflow/linux-aarch64/166441 2025-11-03T16:58:01.4205486Z * [new tag] ciflow/linux-aarch64/166549 -> ciflow/linux-aarch64/166549 2025-11-03T16:58:01.4206608Z * [new tag] ciflow/linux-aarch64/166640 -> ciflow/linux-aarch64/166640 2025-11-03T16:58:01.4207848Z * [new tag] ciflow/linux-aarch64/166691 -> ciflow/linux-aarch64/166691 2025-11-03T16:58:01.4209112Z * [new tag] ciflow/linux-aarch64/166849 -> ciflow/linux-aarch64/166849 2025-11-03T16:58:01.4211593Z * [new tag] ciflow/mps/157553 -> ciflow/mps/157553 2025-11-03T16:58:01.4212842Z * [new tag] ciflow/mps/157554 -> ciflow/mps/157554 2025-11-03T16:58:01.4214029Z * [new tag] ciflow/mps/157635 -> ciflow/mps/157635 2025-11-03T16:58:01.4215195Z * [new tag] ciflow/mps/165952 -> ciflow/mps/165952 2025-11-03T16:58:01.4216381Z * [new tag] ciflow/mps/166254 -> ciflow/mps/166254 2025-11-03T16:58:01.4217544Z * [new tag] ciflow/mps/166273 -> ciflow/mps/166273 2025-11-03T16:58:01.4218899Z * [new tag] ciflow/mps/166396 -> ciflow/mps/166396 2025-11-03T16:58:01.4220054Z * [new tag] ciflow/mps/166615 -> ciflow/mps/166615 2025-11-03T16:58:01.4221369Z * [new tag] ciflow/mps/166712 -> ciflow/mps/166712 2025-11-03T16:58:01.4222819Z * [new tag] ciflow/nightly/158104 -> ciflow/nightly/158104 2025-11-03T16:58:01.4223961Z * [new tag] ciflow/nightly/166727 -> ciflow/nightly/166727 2025-11-03T16:58:01.4225430Z * [new tag] ciflow/op-benchmark/165915 -> ciflow/op-benchmark/165915 2025-11-03T16:58:01.4226758Z * [new tag] ciflow/op-benchmark/166331 -> ciflow/op-benchmark/166331 2025-11-03T16:58:01.4227975Z * [new tag] ciflow/op-benchmark/166640 -> ciflow/op-benchmark/166640 2025-11-03T16:58:01.4229124Z * [new tag] ciflow/op-benchmark/166652 -> ciflow/op-benchmark/166652 2025-11-03T16:58:01.4230301Z * [new tag] ciflow/op-benchmark/166731 -> ciflow/op-benchmark/166731 2025-11-03T16:58:01.4231812Z * [new tag] ciflow/periodic-rocm-mi200/166743 -> ciflow/periodic-rocm-mi200/166743 2025-11-03T16:58:01.4233480Z * [new tag] ciflow/periodic-rocm-mi300/166517 -> ciflow/periodic-rocm-mi300/166517 2025-11-03T16:58:01.4234372Z * [new tag] ciflow/periodic-rocm-mi300/166743 -> ciflow/periodic-rocm-mi300/166743 2025-11-03T16:58:01.4236186Z * [new tag] ciflow/periodic/054a2fd -> ciflow/periodic/054a2fd 2025-11-03T16:58:01.4237867Z * [new tag] ciflow/periodic/06e71c85583d9857f716b441eef9db40d3f27a39 -> ciflow/periodic/06e71c85583d9857f716b441eef9db40d3f27a39 2025-11-03T16:58:01.4239213Z * [new tag] ciflow/periodic/0e46a10aa7cf69354a0a0cbd268506b53a7cc882 -> ciflow/periodic/0e46a10aa7cf69354a0a0cbd268506b53a7cc882 2025-11-03T16:58:01.4240276Z * [new tag] ciflow/periodic/163490 -> ciflow/periodic/163490 2025-11-03T16:58:01.4241445Z * [new tag] ciflow/periodic/164202 -> ciflow/periodic/164202 2025-11-03T16:58:01.4242569Z * [new tag] ciflow/periodic/164938 -> ciflow/periodic/164938 2025-11-03T16:58:01.4243727Z * [new tag] ciflow/periodic/165885 -> ciflow/periodic/165885 2025-11-03T16:58:01.4245378Z * [new tag] ciflow/periodic/166517 -> ciflow/periodic/166517 2025-11-03T16:58:01.4247275Z * [new tag] ciflow/periodic/1e836bc769f9cfabc7659a172b8c2edee7c375d3 -> ciflow/periodic/1e836bc769f9cfabc7659a172b8c2edee7c375d3 2025-11-03T16:58:01.4248519Z * [new tag] ciflow/periodic/2a6d37d -> ciflow/periodic/2a6d37d 2025-11-03T16:58:01.4249988Z * [new tag] ciflow/periodic/317eeb8 -> ciflow/periodic/317eeb8 2025-11-03T16:58:01.4251333Z * [new tag] ciflow/periodic/3c32 -> ciflow/periodic/3c32 2025-11-03T16:58:01.4252796Z * [new tag] ciflow/periodic/3e98831 -> ciflow/periodic/3e98831 2025-11-03T16:58:01.4254253Z * [new tag] ciflow/periodic/687c15c0b3f01118536413d21efcd052838fa10d -> ciflow/periodic/687c15c0b3f01118536413d21efcd052838fa10d 2025-11-03T16:58:01.4255683Z * [new tag] ciflow/periodic/7379972cc0e8a4b4d88b4bea5c8be0aeffdb076a -> ciflow/periodic/7379972cc0e8a4b4d88b4bea5c8be0aeffdb076a 2025-11-03T16:58:01.4257036Z * [new tag] ciflow/periodic/94512-point -> ciflow/periodic/94512-point 2025-11-03T16:58:01.4258637Z * [new tag] ciflow/periodic/96b61844a722d2ae8b1a2dc283a6c0060b40782e -> ciflow/periodic/96b61844a722d2ae8b1a2dc283a6c0060b40782e 2025-11-03T16:58:01.4260004Z * [new tag] ciflow/periodic/be283297100ab86123e74b7a8372995d32b140c8 -> ciflow/periodic/be283297100ab86123e74b7a8372995d32b140c8 2025-11-03T16:58:01.4261400Z * [new tag] ciflow/periodic/bfc2050db975e589795cd3eceaed2e83bf89ad35 -> ciflow/periodic/bfc2050db975e589795cd3eceaed2e83bf89ad35 2025-11-03T16:58:01.4262843Z * [new tag] ciflow/periodic/c5701d0ab5f55b23243dd437f57e856970c45bb6 -> ciflow/periodic/c5701d0ab5f55b23243dd437f57e856970c45bb6 2025-11-03T16:58:01.4264337Z * [new tag] ciflow/periodic/cde81e92b95eee9af2879c9c75f7b03699ca72ad -> ciflow/periodic/cde81e92b95eee9af2879c9c75f7b03699ca72ad 2025-11-03T16:58:01.4266432Z * [new tag] ciflow/periodic/csl/test87519 -> ciflow/periodic/csl/test87519 2025-11-03T16:58:01.4267844Z * [new tag] ciflow/periodic/csltest88275 -> ciflow/periodic/csltest88275 2025-11-03T16:58:01.4269249Z * [new tag] ciflow/periodic/csltest88761 -> ciflow/periodic/csltest88761 2025-11-03T16:58:01.4270685Z * [new tag] ciflow/periodic/release_1.12 -> ciflow/periodic/release_1.12 2025-11-03T16:58:01.4272215Z * [new tag] ciflow/periodic/release_1.12.0 -> ciflow/periodic/release_1.12.0 2025-11-03T16:58:01.4273745Z * [new tag] ciflow/periodic/sha-ec5b83 -> ciflow/periodic/sha-ec5b83 2025-11-03T16:58:01.4275145Z * [new tag] ciflow/pull/166321 -> ciflow/pull/166321 2025-11-03T16:58:01.4276351Z * [new tag] ciflow/pull/166768 -> ciflow/pull/166768 2025-11-03T16:58:01.4278038Z * [new tag] ciflow/pull/18f425962656f009e151944d12f76e2f856b85df -> ciflow/pull/18f425962656f009e151944d12f76e2f856b85df 2025-11-03T16:58:01.4279212Z * [new tag] ciflow/pull/76780b1a3d208b142f067bdd655e7d4c5615efe1 -> ciflow/pull/76780b1a3d208b142f067bdd655e7d4c5615efe1 2025-11-03T16:58:01.4281035Z * [new tag] ciflow/pull/79aee77381b21d41c77148e5ff84c4b351aaf144 -> ciflow/pull/79aee77381b21d41c77148e5ff84c4b351aaf144 2025-11-03T16:58:01.4282716Z * [new tag] ciflow/pull/bc03d7c97462abfd45fc1b7b73d5c7acbbd2c185 -> ciflow/pull/bc03d7c97462abfd45fc1b7b73d5c7acbbd2c185 2025-11-03T16:58:01.4284054Z * [new tag] ciflow/pull/cee03634dab2dbf259d84bc389bdea80de0f80ac -> ciflow/pull/cee03634dab2dbf259d84bc389bdea80de0f80ac 2025-11-03T16:58:01.4285413Z * [new tag] ciflow/pull/d962bed15798961ffd10e64068f86a2cec411ffb -> ciflow/pull/d962bed15798961ffd10e64068f86a2cec411ffb 2025-11-03T16:58:01.4286773Z * [new tag] ciflow/pull/e471800dceb8a6592562fc4fc96a5e80bc494c0a -> ciflow/pull/e471800dceb8a6592562fc4fc96a5e80bc494c0a 2025-11-03T16:58:01.4288247Z * [new tag] ciflow/pull/f013e804c876a34cbc86a6a475597e3e29812883 -> ciflow/pull/f013e804c876a34cbc86a6a475597e3e29812883 2025-11-03T16:58:01.4289766Z * [new tag] ciflow/pull/f5cb9a4c68d9271c58ef4d3257210984b8e85099 -> ciflow/pull/f5cb9a4c68d9271c58ef4d3257210984b8e85099 2025-11-03T16:58:01.4291013Z * [new tag] ciflow/rocm-mi300/162053 -> ciflow/rocm-mi300/162053 2025-11-03T16:58:01.4292168Z * [new tag] ciflow/rocm-mi300/165548 -> ciflow/rocm-mi300/165548 2025-11-03T16:58:01.4293469Z * [new tag] ciflow/rocm-mi300/166317 -> ciflow/rocm-mi300/166317 2025-11-03T16:58:01.4294657Z * [new tag] ciflow/rocm-mi300/166743 -> ciflow/rocm-mi300/166743 2025-11-03T16:58:01.4296034Z * [new tag] ciflow/rocm-mi355/162053 -> ciflow/rocm-mi355/162053 2025-11-03T16:58:01.4297212Z * [new tag] ciflow/rocm-mi355/166743 -> ciflow/rocm-mi355/166743 2025-11-03T16:58:01.4298574Z * [new tag] ciflow/rocm/115316 -> ciflow/rocm/115316 2025-11-03T16:58:01.4299749Z * [new tag] ciflow/rocm/148492 -> ciflow/rocm/148492 2025-11-03T16:58:01.4300892Z * [new tag] ciflow/rocm/151845 -> ciflow/rocm/151845 2025-11-03T16:58:01.4302544Z * [new tag] ciflow/rocm/157553 -> ciflow/rocm/157553 2025-11-03T16:58:01.4303673Z * [new tag] ciflow/rocm/157635 -> ciflow/rocm/157635 2025-11-03T16:58:01.4304814Z * [new tag] ciflow/rocm/160685 -> ciflow/rocm/160685 2025-11-03T16:58:01.4305930Z * [new tag] ciflow/rocm/161404 -> ciflow/rocm/161404 2025-11-03T16:58:01.4307094Z * [new tag] ciflow/rocm/162053 -> ciflow/rocm/162053 2025-11-03T16:58:01.4308209Z * [new tag] ciflow/rocm/163369 -> ciflow/rocm/163369 2025-11-03T16:58:01.4309358Z * [new tag] ciflow/rocm/164657 -> ciflow/rocm/164657 2025-11-03T16:58:01.4311059Z * [new tag] ciflow/rocm/165132 -> ciflow/rocm/165132 2025-11-03T16:58:01.4312524Z * [new tag] ciflow/rocm/165541 -> ciflow/rocm/165541 2025-11-03T16:58:01.4314043Z * [new tag] ciflow/rocm/165548 -> ciflow/rocm/165548 2025-11-03T16:58:01.4315482Z * [new tag] ciflow/rocm/165952 -> ciflow/rocm/165952 2025-11-03T16:58:01.4316893Z * [new tag] ciflow/rocm/165997 -> ciflow/rocm/165997 2025-11-03T16:58:01.4318092Z * [new tag] ciflow/rocm/166250 -> ciflow/rocm/166250 2025-11-03T16:58:01.4319545Z * [new tag] ciflow/rocm/166317 -> ciflow/rocm/166317 2025-11-03T16:58:01.4320923Z * [new tag] ciflow/rocm/166377 -> ciflow/rocm/166377 2025-11-03T16:58:01.4322283Z * [new tag] ciflow/rocm/166391 -> ciflow/rocm/166391 2025-11-03T16:58:01.4323402Z * [new tag] ciflow/rocm/166515 -> ciflow/rocm/166515 2025-11-03T16:58:01.4324641Z * [new tag] ciflow/rocm/166517 -> ciflow/rocm/166517 2025-11-03T16:58:01.4325863Z * [new tag] ciflow/rocm/166526 -> ciflow/rocm/166526 2025-11-03T16:58:01.4327096Z * [new tag] ciflow/rocm/166743 -> ciflow/rocm/166743 2025-11-03T16:58:01.4328309Z * [new tag] ciflow/rocm/166744 -> ciflow/rocm/166744 2025-11-03T16:58:01.4329648Z * [new tag] ciflow/rocm/166752 -> ciflow/rocm/166752 2025-11-03T16:58:01.4330850Z * [new tag] ciflow/rocm/166764 -> ciflow/rocm/166764 2025-11-03T16:58:01.4332074Z * [new tag] ciflow/rocm/166834 -> ciflow/rocm/166834 2025-11-03T16:58:01.4333288Z * [new tag] ciflow/rocm/166870 -> ciflow/rocm/166870 2025-11-03T16:58:01.4334914Z * [new tag] ciflow/slow/01c7106 -> ciflow/slow/01c7106 2025-11-03T16:58:01.4336609Z * [new tag] ciflow/slow/0577043 -> ciflow/slow/0577043 2025-11-03T16:58:01.4338377Z * [new tag] ciflow/slow/0d5b74da0cab798fbfdb9caa53fad816999c8386-sdym -> ciflow/slow/0d5b74da0cab798fbfdb9caa53fad816999c8386-sdym 2025-11-03T16:58:01.4339472Z * [new tag] ciflow/slow/0e81104 -> ciflow/slow/0e81104 2025-11-03T16:58:01.4340662Z * [new tag] ciflow/slow/166844 -> ciflow/slow/166844 2025-11-03T16:58:01.4342080Z * [new tag] ciflow/slow/1732077 -> ciflow/slow/1732077 2025-11-03T16:58:01.4343469Z * [new tag] ciflow/slow/187eb7c -> ciflow/slow/187eb7c 2025-11-03T16:58:01.4344749Z * [new tag] ciflow/slow/1faef89 -> ciflow/slow/1faef89 2025-11-03T16:58:01.4346350Z * [new tag] ciflow/slow/3920ec1 -> ciflow/slow/3920ec1 2025-11-03T16:58:01.4347940Z * [new tag] ciflow/slow/3b7c6b2 -> ciflow/slow/3b7c6b2 2025-11-03T16:58:01.4349360Z * [new tag] ciflow/slow/59a3759 -> ciflow/slow/59a3759 2025-11-03T16:58:01.4350709Z * [new tag] ciflow/slow/70ef0bb -> ciflow/slow/70ef0bb 2025-11-03T16:58:01.4352043Z * [new tag] ciflow/slow/788ff06 -> ciflow/slow/788ff06 2025-11-03T16:58:01.4353765Z * [new tag] ciflow/slow/8751002215790a3a88750faa8f4366933e296693-sdym -> ciflow/slow/8751002215790a3a88750faa8f4366933e296693-sdym 2025-11-03T16:58:01.4354915Z * [new tag] ciflow/slow/9d85864 -> ciflow/slow/9d85864 2025-11-03T16:58:01.4356283Z * [new tag] ciflow/slow/9ffad5b -> ciflow/slow/9ffad5b 2025-11-03T16:58:01.4357658Z * [new tag] ciflow/slow/a206e8b -> ciflow/slow/a206e8b 2025-11-03T16:58:01.4359048Z * [new tag] ciflow/slow/a837609 -> ciflow/slow/a837609 2025-11-03T16:58:01.4360484Z * [new tag] ciflow/slow/af841f3 -> ciflow/slow/af841f3 2025-11-03T16:58:01.4362251Z * [new tag] ciflow/slow/da3aba1e46157c4df504b067477cdf2b3c96b194-sdym -> ciflow/slow/da3aba1e46157c4df504b067477cdf2b3c96b194-sdym 2025-11-03T16:58:01.4363844Z * [new tag] ciflow/trunk/04d6a6f3392a87af8baf620dc0a323ffbb188c12 -> ciflow/trunk/04d6a6f3392a87af8baf620dc0a323ffbb188c12 2025-11-03T16:58:01.4365157Z * [new tag] ciflow/trunk/0573747b6af273c6ac8cf9f1d81dc56066a57445 -> ciflow/trunk/0573747b6af273c6ac8cf9f1d81dc56066a57445 2025-11-03T16:58:01.4366015Z * [new tag] ciflow/trunk/137400 -> ciflow/trunk/137400 2025-11-03T16:58:01.4367319Z * [new tag] ciflow/trunk/148180 -> ciflow/trunk/148180 2025-11-03T16:58:01.4368680Z * [new tag] ciflow/trunk/148328 -> ciflow/trunk/148328 2025-11-03T16:58:01.4369717Z * [new tag] ciflow/trunk/148492 -> ciflow/trunk/148492 2025-11-03T16:58:01.4370869Z * [new tag] ciflow/trunk/151845 -> ciflow/trunk/151845 2025-11-03T16:58:01.4372308Z * [new tag] ciflow/trunk/152624 -> ciflow/trunk/152624 2025-11-03T16:58:01.4373759Z * [new tag] ciflow/trunk/155731 -> ciflow/trunk/155731 2025-11-03T16:58:01.4375055Z * [new tag] ciflow/trunk/156812 -> ciflow/trunk/156812 2025-11-03T16:58:01.4376265Z * [new tag] ciflow/trunk/157432 -> ciflow/trunk/157432 2025-11-03T16:58:01.4377458Z * [new tag] ciflow/trunk/158104 -> ciflow/trunk/158104 2025-11-03T16:58:01.4378716Z * [new tag] ciflow/trunk/160417 -> ciflow/trunk/160417 2025-11-03T16:58:01.4380024Z * [new tag] ciflow/trunk/161035 -> ciflow/trunk/161035 2025-11-03T16:58:01.4381423Z * [new tag] ciflow/trunk/161771 -> ciflow/trunk/161771 2025-11-03T16:58:01.4382619Z * [new tag] ciflow/trunk/162053 -> ciflow/trunk/162053 2025-11-03T16:58:01.4383918Z * [new tag] ciflow/trunk/162454 -> ciflow/trunk/162454 2025-11-03T16:58:01.4385158Z * [new tag] ciflow/trunk/162564 -> ciflow/trunk/162564 2025-11-03T16:58:01.4386399Z * [new tag] ciflow/trunk/162709 -> ciflow/trunk/162709 2025-11-03T16:58:01.4387622Z * [new tag] ciflow/trunk/162795 -> ciflow/trunk/162795 2025-11-03T16:58:01.4389261Z * [new tag] ciflow/trunk/163249 -> ciflow/trunk/163249 2025-11-03T16:58:01.4390478Z * [new tag] ciflow/trunk/163490 -> ciflow/trunk/163490 2025-11-03T16:58:01.4391713Z * [new tag] ciflow/trunk/164202 -> ciflow/trunk/164202 2025-11-03T16:58:01.4392924Z * [new tag] ciflow/trunk/164938 -> ciflow/trunk/164938 2025-11-03T16:58:01.4394170Z * [new tag] ciflow/trunk/164979 -> ciflow/trunk/164979 2025-11-03T16:58:01.4395385Z * [new tag] ciflow/trunk/164980 -> ciflow/trunk/164980 2025-11-03T16:58:01.4396607Z * [new tag] ciflow/trunk/165005 -> ciflow/trunk/165005 2025-11-03T16:58:01.4397811Z * [new tag] ciflow/trunk/165029 -> ciflow/trunk/165029 2025-11-03T16:58:01.4399221Z * [new tag] ciflow/trunk/165049 -> ciflow/trunk/165049 2025-11-03T16:58:01.4400595Z * [new tag] ciflow/trunk/165067 -> ciflow/trunk/165067 2025-11-03T16:58:01.4401887Z * [new tag] ciflow/trunk/165132 -> ciflow/trunk/165132 2025-11-03T16:58:01.4403310Z * [new tag] ciflow/trunk/165216 -> ciflow/trunk/165216 2025-11-03T16:58:01.4404560Z * [new tag] ciflow/trunk/165274 -> ciflow/trunk/165274 2025-11-03T16:58:01.4405880Z * [new tag] ciflow/trunk/165431 -> ciflow/trunk/165431 2025-11-03T16:58:01.4407149Z * [new tag] ciflow/trunk/165548 -> ciflow/trunk/165548 2025-11-03T16:58:01.4408569Z * [new tag] ciflow/trunk/165642 -> ciflow/trunk/165642 2025-11-03T16:58:01.4409893Z * [new tag] ciflow/trunk/165646 -> ciflow/trunk/165646 2025-11-03T16:58:01.4411408Z * [new tag] ciflow/trunk/165728 -> ciflow/trunk/165728 2025-11-03T16:58:01.4412615Z * [new tag] ciflow/trunk/165824 -> ciflow/trunk/165824 2025-11-03T16:58:01.4413845Z * [new tag] ciflow/trunk/165885 -> ciflow/trunk/165885 2025-11-03T16:58:01.4415093Z * [new tag] ciflow/trunk/165922 -> ciflow/trunk/165922 2025-11-03T16:58:01.4416308Z * [new tag] ciflow/trunk/165972 -> ciflow/trunk/165972 2025-11-03T16:58:01.4417671Z * [new tag] ciflow/trunk/166010 -> ciflow/trunk/166010 2025-11-03T16:58:01.4418914Z * [new tag] ciflow/trunk/166071 -> ciflow/trunk/166071 2025-11-03T16:58:01.4420204Z * [new tag] ciflow/trunk/166160 -> ciflow/trunk/166160 2025-11-03T16:58:01.4421412Z * [new tag] ciflow/trunk/166170 -> ciflow/trunk/166170 2025-11-03T16:58:01.4422628Z * [new tag] ciflow/trunk/166198 -> ciflow/trunk/166198 2025-11-03T16:58:01.4423938Z * [new tag] ciflow/trunk/166294 -> ciflow/trunk/166294 2025-11-03T16:58:01.4425156Z * [new tag] ciflow/trunk/166299 -> ciflow/trunk/166299 2025-11-03T16:58:01.4426395Z * [new tag] ciflow/trunk/166317 -> ciflow/trunk/166317 2025-11-03T16:58:01.4427645Z * [new tag] ciflow/trunk/166321 -> ciflow/trunk/166321 2025-11-03T16:58:01.4428871Z * [new tag] ciflow/trunk/166361 -> ciflow/trunk/166361 2025-11-03T16:58:01.4430157Z * [new tag] ciflow/trunk/166367 -> ciflow/trunk/166367 2025-11-03T16:58:01.4431587Z * [new tag] ciflow/trunk/166376 -> ciflow/trunk/166376 2025-11-03T16:58:01.4432785Z * [new tag] ciflow/trunk/166379 -> ciflow/trunk/166379 2025-11-03T16:58:01.4433981Z * [new tag] ciflow/trunk/166380 -> ciflow/trunk/166380 2025-11-03T16:58:01.4435218Z * [new tag] ciflow/trunk/166402 -> ciflow/trunk/166402 2025-11-03T16:58:01.4436468Z * [new tag] ciflow/trunk/166416 -> ciflow/trunk/166416 2025-11-03T16:58:01.4437645Z * [new tag] ciflow/trunk/166431 -> ciflow/trunk/166431 2025-11-03T16:58:01.4438831Z * [new tag] ciflow/trunk/166433 -> ciflow/trunk/166433 2025-11-03T16:58:01.4440024Z * [new tag] ciflow/trunk/166459 -> ciflow/trunk/166459 2025-11-03T16:58:01.4441304Z * [new tag] ciflow/trunk/166469 -> ciflow/trunk/166469 2025-11-03T16:58:01.4442703Z * [new tag] ciflow/trunk/166493 -> ciflow/trunk/166493 2025-11-03T16:58:01.4443931Z * [new tag] ciflow/trunk/166526 -> ciflow/trunk/166526 2025-11-03T16:58:01.4445180Z * [new tag] ciflow/trunk/166536 -> ciflow/trunk/166536 2025-11-03T16:58:01.4446458Z * [new tag] ciflow/trunk/166541 -> ciflow/trunk/166541 2025-11-03T16:58:01.4447649Z * [new tag] ciflow/trunk/166549 -> ciflow/trunk/166549 2025-11-03T16:58:01.4449223Z * [new tag] ciflow/trunk/166560 -> ciflow/trunk/166560 2025-11-03T16:58:01.4450443Z * [new tag] ciflow/trunk/166608 -> ciflow/trunk/166608 2025-11-03T16:58:01.4451729Z * [new tag] ciflow/trunk/166610 -> ciflow/trunk/166610 2025-11-03T16:58:01.4453022Z * [new tag] ciflow/trunk/166613 -> ciflow/trunk/166613 2025-11-03T16:58:01.4454231Z * [new tag] ciflow/trunk/166640 -> ciflow/trunk/166640 2025-11-03T16:58:01.4455462Z * [new tag] ciflow/trunk/166688 -> ciflow/trunk/166688 2025-11-03T16:58:01.4456710Z * [new tag] ciflow/trunk/166689 -> ciflow/trunk/166689 2025-11-03T16:58:01.4458051Z * [new tag] ciflow/trunk/166715 -> ciflow/trunk/166715 2025-11-03T16:58:01.4459247Z * [new tag] ciflow/trunk/166727 -> ciflow/trunk/166727 2025-11-03T16:58:01.4460495Z * [new tag] ciflow/trunk/166744 -> ciflow/trunk/166744 2025-11-03T16:58:01.4461970Z * [new tag] ciflow/trunk/166751 -> ciflow/trunk/166751 2025-11-03T16:58:01.4463149Z * [new tag] ciflow/trunk/166768 -> ciflow/trunk/166768 2025-11-03T16:58:01.4464496Z * [new tag] ciflow/trunk/166791 -> ciflow/trunk/166791 2025-11-03T16:58:01.4465623Z * [new tag] ciflow/trunk/166793 -> ciflow/trunk/166793 2025-11-03T16:58:01.4466859Z * [new tag] ciflow/trunk/166805 -> ciflow/trunk/166805 2025-11-03T16:58:01.4468027Z * [new tag] ciflow/trunk/166806 -> ciflow/trunk/166806 2025-11-03T16:58:01.4469230Z * [new tag] ciflow/trunk/166813 -> ciflow/trunk/166813 2025-11-03T16:58:01.4470885Z * [new tag] ciflow/trunk/166830 -> ciflow/trunk/166830 2025-11-03T16:58:01.4472166Z * [new tag] ciflow/trunk/166839 -> ciflow/trunk/166839 2025-11-03T16:58:01.4473563Z * [new tag] ciflow/trunk/166840 -> ciflow/trunk/166840 2025-11-03T16:58:01.4474780Z * [new tag] ciflow/trunk/166844 -> ciflow/trunk/166844 2025-11-03T16:58:01.4475995Z * [new tag] ciflow/trunk/166847 -> ciflow/trunk/166847 2025-11-03T16:58:01.4477300Z * [new tag] ciflow/trunk/166861 -> ciflow/trunk/166861 2025-11-03T16:58:01.4478824Z * [new tag] ciflow/trunk/18f425962656f009e151944d12f76e2f856b85df -> ciflow/trunk/18f425962656f009e151944d12f76e2f856b85df 2025-11-03T16:58:01.4480010Z * [new tag] ciflow/trunk/76780b1a3d208b142f067bdd655e7d4c5615efe1 -> ciflow/trunk/76780b1a3d208b142f067bdd655e7d4c5615efe1 2025-11-03T16:58:01.4481566Z * [new tag] ciflow/trunk/a19e92d43355489c8d33325321e688a6e6182267 -> ciflow/trunk/a19e92d43355489c8d33325321e688a6e6182267 2025-11-03T16:58:01.4482843Z * [new tag] ciflow/trunk/bc03d7c97462abfd45fc1b7b73d5c7acbbd2c185 -> ciflow/trunk/bc03d7c97462abfd45fc1b7b73d5c7acbbd2c185 2025-11-03T16:58:01.4484405Z * [new tag] ciflow/trunk/c3dc0c7089f3204c36961acc227a3ff503530094 -> ciflow/trunk/c3dc0c7089f3204c36961acc227a3ff503530094 2025-11-03T16:58:01.4485682Z * [new tag] ciflow/trunk/cee03634dab2dbf259d84bc389bdea80de0f80ac -> ciflow/trunk/cee03634dab2dbf259d84bc389bdea80de0f80ac 2025-11-03T16:58:01.4486862Z * [new tag] ciflow/trunk/d962bed15798961ffd10e64068f86a2cec411ffb -> ciflow/trunk/d962bed15798961ffd10e64068f86a2cec411ffb 2025-11-03T16:58:01.4488125Z * [new tag] ciflow/trunk/e471800dceb8a6592562fc4fc96a5e80bc494c0a -> ciflow/trunk/e471800dceb8a6592562fc4fc96a5e80bc494c0a 2025-11-03T16:58:01.4489467Z * [new tag] ciflow/trunk/f013e804c876a34cbc86a6a475597e3e29812883 -> ciflow/trunk/f013e804c876a34cbc86a6a475597e3e29812883 2025-11-03T16:58:01.4491024Z * [new tag] ciflow/unstable/123 -> ciflow/unstable/123 2025-11-03T16:58:01.4492456Z * [new tag] ciflow/vllm/165274 -> ciflow/vllm/165274 2025-11-03T16:58:01.4493610Z * [new tag] ciflow/vllm/166494 -> ciflow/vllm/166494 2025-11-03T16:58:01.4495206Z * [new tag] ciflow/win-arm64/158104 -> ciflow/win-arm64/158104 2025-11-03T16:58:01.4496540Z * [new tag] ciflow/xpu/156812 -> ciflow/xpu/156812 2025-11-03T16:58:01.4497811Z * [new tag] ciflow/xpu/161246 -> ciflow/xpu/161246 2025-11-03T16:58:01.4498932Z * [new tag] ciflow/xpu/162454 -> ciflow/xpu/162454 2025-11-03T16:58:01.4500172Z * [new tag] ciflow/xpu/162475 -> ciflow/xpu/162475 2025-11-03T16:58:01.4501319Z * [new tag] ciflow/xpu/162564 -> ciflow/xpu/162564 2025-11-03T16:58:01.4502497Z * [new tag] ciflow/xpu/165049 -> ciflow/xpu/165049 2025-11-03T16:58:01.4503639Z * [new tag] ciflow/xpu/165423 -> ciflow/xpu/165423 2025-11-03T16:58:01.4504772Z * [new tag] ciflow/xpu/166292 -> ciflow/xpu/166292 2025-11-03T16:58:01.4505884Z * [new tag] ciflow/xpu/166299 -> ciflow/xpu/166299 2025-11-03T16:58:01.4507125Z * [new tag] ciflow/xpu/166376 -> ciflow/xpu/166376 2025-11-03T16:58:01.4507968Z * [new tag] ciflow/xpu/166396 -> ciflow/xpu/166396 2025-11-03T16:58:01.4509721Z * [new tag] ciflow/xpu/166424 -> ciflow/xpu/166424 2025-11-03T16:58:01.4511518Z * [new tag] ciflow/xpu/166436 -> ciflow/xpu/166436 2025-11-03T16:58:01.4512961Z * [new tag] ciflow/xpu/166495 -> ciflow/xpu/166495 2025-11-03T16:58:01.4514256Z * [new tag] ciflow/xpu/166504 -> ciflow/xpu/166504 2025-11-03T16:58:01.4515454Z * [new tag] ciflow/xpu/166613 -> ciflow/xpu/166613 2025-11-03T16:58:01.4516799Z * [new tag] ciflow/xpu/166684 -> ciflow/xpu/166684 2025-11-03T16:58:01.4518083Z * [new tag] ciflow/xpu/166830 -> ciflow/xpu/166830 2025-11-03T16:58:01.4519228Z * [new tag] ciflow/xpu/166834 -> ciflow/xpu/166834 2025-11-03T16:58:01.4520581Z * [new tag] ciflow/xpu/166839 -> ciflow/xpu/166839 2025-11-03T16:58:01.4521765Z * [new tag] ciflow/xpu/166840 -> ciflow/xpu/166840 2025-11-03T16:58:01.4523113Z * [new tag] ciflow/xpu/166847 -> ciflow/xpu/166847 2025-11-03T16:58:01.4524220Z * [new tag] ciflow/xpu/166861 -> ciflow/xpu/166861 2025-11-03T16:58:01.4525410Z * [new tag] cslpull75 -> cslpull75 2025-11-03T16:58:01.4526652Z * [new tag] cslpull76 -> cslpull76 2025-11-03T16:58:01.4527919Z * [new tag] cslpull77 -> cslpull77 2025-11-03T16:58:01.4529124Z * [new tag] cslpull78 -> cslpull78 2025-11-03T16:58:01.4530576Z * [new tag] cslpull79 -> cslpull79 2025-11-03T16:58:01.4532057Z * [new tag] cslpull80 -> cslpull80 2025-11-03T16:58:01.4533419Z * [new tag] cslpull81 -> cslpull81 2025-11-03T16:58:01.4534638Z * [new tag] cslpull82 -> cslpull82 2025-11-03T16:58:01.4535944Z * [new tag] cslpull83 -> cslpull83 2025-11-03T16:58:01.4537186Z * [new tag] cslpull84 -> cslpull84 2025-11-03T16:58:01.4538473Z * [new tag] cslpull85 -> cslpull85 2025-11-03T16:58:01.4539791Z * [new tag] cslpull86 -> cslpull86 2025-11-03T16:58:01.4541070Z * [new tag] cslpull87 -> cslpull87 2025-11-03T16:58:01.4542383Z * [new tag] cslpull88 -> cslpull88 2025-11-03T16:58:01.4543674Z * [new tag] cslpull89 -> cslpull89 2025-11-03T16:58:01.4545281Z * [new tag] cslpull90 -> cslpull90 2025-11-03T16:58:01.4546929Z * [new tag] cslpull91 -> cslpull91 2025-11-03T16:58:01.4548145Z * [new tag] cslpull92 -> cslpull92 2025-11-03T16:58:01.4549533Z * [new tag] flight_5 -> flight_5 2025-11-03T16:58:01.4550978Z * [new tag] flight_5.1 -> flight_5.1 2025-11-03T16:58:01.4552292Z * [new tag] flight_5.2 -> flight_5.2 2025-11-03T16:58:01.4553657Z * [new tag] flight_5.3 -> flight_5.3 2025-11-03T16:58:01.4554922Z * [new tag] forpull1 -> forpull1 2025-11-03T16:58:01.4556434Z * [new tag] malfet/tag-2ef5611 -> malfet/tag-2ef5611 2025-11-03T16:58:01.4557681Z * [new tag] malfet/tag-317b1a0 -> malfet/tag-317b1a0 2025-11-03T16:58:01.4558995Z * [new tag] malfet/tag-ec6f767 -> malfet/tag-ec6f767 2025-11-03T16:58:01.4560588Z * [new tag] nightly-binary -> nightly-binary 2025-11-03T16:58:01.4561846Z * [new tag] sqzhang_flight4_plus -> sqzhang_flight4_plus 2025-11-03T16:58:01.4563264Z * [new tag] sqzhang_flight_3 -> sqzhang_flight_3 2025-11-03T16:58:01.4565020Z * [new tag] trunk/000f49551bfda0683d395e1fe67f53cfb785868f -> trunk/000f49551bfda0683d395e1fe67f53cfb785868f 2025-11-03T16:58:01.4566013Z * [new tag] trunk/009ea77234788b29910b1205679036eaae3ee3ac -> trunk/009ea77234788b29910b1205679036eaae3ee3ac 2025-11-03T16:58:01.4567744Z * [new tag] trunk/0187db88d463905e5f7dcdf7689519e2807e5a12 -> trunk/0187db88d463905e5f7dcdf7689519e2807e5a12 2025-11-03T16:58:01.4569398Z * [new tag] trunk/02095cc09db01f1fecae436ac1a6ecdf2cfca0db -> trunk/02095cc09db01f1fecae436ac1a6ecdf2cfca0db 2025-11-03T16:58:01.4570748Z * [new tag] trunk/030de07affa347a80fa9f11ab8e9234fc6d7fd4d -> trunk/030de07affa347a80fa9f11ab8e9234fc6d7fd4d 2025-11-03T16:58:01.4572120Z * [new tag] trunk/034e951b0cfb02d7b55327cd482e58cf2695dca0 -> trunk/034e951b0cfb02d7b55327cd482e58cf2695dca0 2025-11-03T16:58:01.4573213Z * [new tag] trunk/04d6a6f3392a87af8baf620dc0a323ffbb188c12 -> trunk/04d6a6f3392a87af8baf620dc0a323ffbb188c12 2025-11-03T16:58:01.4574478Z * [new tag] trunk/0573747b6af273c6ac8cf9f1d81dc56066a57445 -> trunk/0573747b6af273c6ac8cf9f1d81dc56066a57445 2025-11-03T16:58:01.4575899Z * [new tag] trunk/061fa73c97dcfcb39077ce8fc4042e7f22824b22 -> trunk/061fa73c97dcfcb39077ce8fc4042e7f22824b22 2025-11-03T16:58:01.4577329Z * [new tag] trunk/0674e0a0f14775f920296e9dfb8b61e4960bf99d -> trunk/0674e0a0f14775f920296e9dfb8b61e4960bf99d 2025-11-03T16:58:01.4578289Z * [new tag] trunk/06e71c85583d9857f716b441eef9db40d3f27a39 -> trunk/06e71c85583d9857f716b441eef9db40d3f27a39 2025-11-03T16:58:01.4579782Z * [new tag] trunk/08ae55021eafba28b304e78ef2da9a76741ce039 -> trunk/08ae55021eafba28b304e78ef2da9a76741ce039 2025-11-03T16:58:01.4581105Z * [new tag] trunk/08b0a8f11a00e0442bd3b3066617a501dca29876 -> trunk/08b0a8f11a00e0442bd3b3066617a501dca29876 2025-11-03T16:58:01.4582501Z * [new tag] trunk/08f4535378b9ab5c7ba74828bb20ad535e6de8eb -> trunk/08f4535378b9ab5c7ba74828bb20ad535e6de8eb 2025-11-03T16:58:01.4583966Z * [new tag] trunk/0918bf321c2cfaebb91c5f9e5f0ebe462962328e -> trunk/0918bf321c2cfaebb91c5f9e5f0ebe462962328e 2025-11-03T16:58:01.4585299Z * [new tag] trunk/0947765eb9208996f221dbcb088df800be3953d7 -> trunk/0947765eb9208996f221dbcb088df800be3953d7 2025-11-03T16:58:01.4586592Z * [new tag] trunk/0a3ac47c0a36d7ed619e6577294648a7c14b1ae1 -> trunk/0a3ac47c0a36d7ed619e6577294648a7c14b1ae1 2025-11-03T16:58:01.4587986Z * [new tag] trunk/0ae3e306217a67362371267bbba0ac3fa0855fdc -> trunk/0ae3e306217a67362371267bbba0ac3fa0855fdc 2025-11-03T16:58:01.4589278Z * [new tag] trunk/0d3a4f7155508656bd928527840ed155e8b1a88b -> trunk/0d3a4f7155508656bd928527840ed155e8b1a88b 2025-11-03T16:58:01.4590628Z * [new tag] trunk/0d4992c1702d9d14bc73c29d44029adb7e698eba -> trunk/0d4992c1702d9d14bc73c29d44029adb7e698eba 2025-11-03T16:58:01.4591971Z * [new tag] trunk/0d50e5d8d4132cb09508b3e35701e7b06433420e -> trunk/0d50e5d8d4132cb09508b3e35701e7b06433420e 2025-11-03T16:58:01.4593281Z * [new tag] trunk/0d81bb7f9cef1b735ff373a245aaa2bc037b31b0 -> trunk/0d81bb7f9cef1b735ff373a245aaa2bc037b31b0 2025-11-03T16:58:01.4594587Z * [new tag] trunk/0e19561e235372a3fbed81338fa4e674116c0448 -> trunk/0e19561e235372a3fbed81338fa4e674116c0448 2025-11-03T16:58:01.4595815Z * [new tag] trunk/0e46a10aa7cf69354a0a0cbd268506b53a7cc882 -> trunk/0e46a10aa7cf69354a0a0cbd268506b53a7cc882 2025-11-03T16:58:01.4597470Z * [new tag] trunk/0eacd934bc1c1bee002fcd084300d596d789d166 -> trunk/0eacd934bc1c1bee002fcd084300d596d789d166 2025-11-03T16:58:01.4598385Z * [new tag] trunk/0ec054982304473b51e6ded40c2dc3993d4036be -> trunk/0ec054982304473b51e6ded40c2dc3993d4036be 2025-11-03T16:58:01.4599680Z * [new tag] trunk/104b86861862ccc47f6df6528347a017bef813ad -> trunk/104b86861862ccc47f6df6528347a017bef813ad 2025-11-03T16:58:01.4601061Z * [new tag] trunk/108bb224f77842593009214ebf6258030b934642 -> trunk/108bb224f77842593009214ebf6258030b934642 2025-11-03T16:58:01.4602230Z * [new tag] trunk/110efe4df4759006ce5aad4ae1b87d93dcf33c3b -> trunk/110efe4df4759006ce5aad4ae1b87d93dcf33c3b 2025-11-03T16:58:01.4603642Z * [new tag] trunk/112960541529ab0d27bb8aa8b487ef194e048494 -> trunk/112960541529ab0d27bb8aa8b487ef194e048494 2025-11-03T16:58:01.4605079Z * [new tag] trunk/121235956bab7430fb8d080cee209607f8387ead -> trunk/121235956bab7430fb8d080cee209607f8387ead 2025-11-03T16:58:01.4606582Z * [new tag] trunk/12577064dddfc6f5daf66c5b5a73cb418a588f20 -> trunk/12577064dddfc6f5daf66c5b5a73cb418a588f20 2025-11-03T16:58:01.4607968Z * [new tag] trunk/12c12466b082834d23c1f512130da9280badfd28 -> trunk/12c12466b082834d23c1f512130da9280badfd28 2025-11-03T16:58:01.4609559Z * [new tag] trunk/13413b3b07cc72fa9c2671b2535f7e54c1b30ca2 -> trunk/13413b3b07cc72fa9c2671b2535f7e54c1b30ca2 2025-11-03T16:58:01.4611160Z * [new tag] trunk/13549e0e105291721f6f80ceb8f520ca272b971d -> trunk/13549e0e105291721f6f80ceb8f520ca272b971d 2025-11-03T16:58:01.4612652Z * [new tag] trunk/14102fb1f337d01ca61584011f5f657d14da5de6 -> trunk/14102fb1f337d01ca61584011f5f657d14da5de6 2025-11-03T16:58:01.4614054Z * [new tag] trunk/1425b40f29ab11ce21f7a98de9e35551213efe27 -> trunk/1425b40f29ab11ce21f7a98de9e35551213efe27 2025-11-03T16:58:01.4615434Z * [new tag] trunk/14d4a77495dc80da9cd7c5c29b21aaf4613ddc5b -> trunk/14d4a77495dc80da9cd7c5c29b21aaf4613ddc5b 2025-11-03T16:58:01.4616740Z * [new tag] trunk/160ab53dd57e67b3574763615cf8b33249e9afa5 -> trunk/160ab53dd57e67b3574763615cf8b33249e9afa5 2025-11-03T16:58:01.4618122Z * [new tag] trunk/16212f0d6bc1c4fb6ac0498badbe2fb74d5ff47b -> trunk/16212f0d6bc1c4fb6ac0498badbe2fb74d5ff47b 2025-11-03T16:58:01.4620328Z * [new tag] trunk/1656b253c5fcc34a53a9991303170892c6d90c1d -> trunk/1656b253c5fcc34a53a9991303170892c6d90c1d 2025-11-03T16:58:01.4621270Z * [new tag] trunk/173bcda436e9a00aa8215a414efc0c06cf4840a8 -> trunk/173bcda436e9a00aa8215a414efc0c06cf4840a8 2025-11-03T16:58:01.4622625Z * [new tag] trunk/1764f3a9c8eea4cf0aba853348446948597b7b78 -> trunk/1764f3a9c8eea4cf0aba853348446948597b7b78 2025-11-03T16:58:01.4624023Z * [new tag] trunk/17bdb232e1b3df3435fc6f54d3d713404ba96726 -> trunk/17bdb232e1b3df3435fc6f54d3d713404ba96726 2025-11-03T16:58:01.4625329Z * [new tag] trunk/17d5aa47674c5195a3d4919169fb6a334598e2a6 -> trunk/17d5aa47674c5195a3d4919169fb6a334598e2a6 2025-11-03T16:58:01.4626639Z * [new tag] trunk/181ee3bd42447b71a1a8435bf16c0877c4bc3ae7 -> trunk/181ee3bd42447b71a1a8435bf16c0877c4bc3ae7 2025-11-03T16:58:01.4627649Z * [new tag] trunk/18f425962656f009e151944d12f76e2f856b85df -> trunk/18f425962656f009e151944d12f76e2f856b85df 2025-11-03T16:58:01.4629154Z * [new tag] trunk/1abfa5f70b9aa0dcbd83dc40b092dda1cdcc7805 -> trunk/1abfa5f70b9aa0dcbd83dc40b092dda1cdcc7805 2025-11-03T16:58:01.4630517Z * [new tag] trunk/1aef88c72d3aef629b20e97a188c9dc4bab46a1a -> trunk/1aef88c72d3aef629b20e97a188c9dc4bab46a1a 2025-11-03T16:58:01.4631946Z * [new tag] trunk/1b655a87ef137d2cc9603a982532c5e033432daa -> trunk/1b655a87ef137d2cc9603a982532c5e033432daa 2025-11-03T16:58:01.4633264Z * [new tag] trunk/1c4ced2eafa470832f2f4e24487785c5b7b188c9 -> trunk/1c4ced2eafa470832f2f4e24487785c5b7b188c9 2025-11-03T16:58:01.4634793Z * [new tag] trunk/1dd6b76914397d466a7d9bfff551b6d8299fdce4 -> trunk/1dd6b76914397d466a7d9bfff551b6d8299fdce4 2025-11-03T16:58:01.4636089Z * [new tag] trunk/1dec8a67a8dc88736e9c476a3eaee301684094cd -> trunk/1dec8a67a8dc88736e9c476a3eaee301684094cd 2025-11-03T16:58:01.4637410Z * [new tag] trunk/1e3600b5287346b29a835ad67f8b33945e0ec698 -> trunk/1e3600b5287346b29a835ad67f8b33945e0ec698 2025-11-03T16:58:01.4638387Z * [new tag] trunk/1e836bc769f9cfabc7659a172b8c2edee7c375d3 -> trunk/1e836bc769f9cfabc7659a172b8c2edee7c375d3 2025-11-03T16:58:01.4639937Z * [new tag] trunk/1fa520ea654f5fc0b3c65ce6e056dd73442dd65d -> trunk/1fa520ea654f5fc0b3c65ce6e056dd73442dd65d 2025-11-03T16:58:01.4641315Z * [new tag] trunk/1fdef664a5dffc3bd3d9eba9d78ed458a5f9dbcb -> trunk/1fdef664a5dffc3bd3d9eba9d78ed458a5f9dbcb 2025-11-03T16:58:01.4642823Z * [new tag] trunk/20be077085daf739af76325d04f649dcd908316c -> trunk/20be077085daf739af76325d04f649dcd908316c 2025-11-03T16:58:01.4643869Z * [new tag] trunk/21b48f8dfa7685699df4c97c0ba373d5364230d9 -> trunk/21b48f8dfa7685699df4c97c0ba373d5364230d9 2025-11-03T16:58:01.4645387Z * [new tag] trunk/23669d02a67023ef3192f7dca0ac0547436cae50 -> trunk/23669d02a67023ef3192f7dca0ac0547436cae50 2025-11-03T16:58:01.4646779Z * [new tag] trunk/236ce736a1a84352b21b564e3e5b0be833d7c10c -> trunk/236ce736a1a84352b21b564e3e5b0be833d7c10c 2025-11-03T16:58:01.4648120Z * [new tag] trunk/239e7b541a94d16a33a4a48d3edc9a561777d86a -> trunk/239e7b541a94d16a33a4a48d3edc9a561777d86a 2025-11-03T16:58:01.4649650Z * [new tag] trunk/23b57a445c6394d46a6b46e32b7787663f40d13b -> trunk/23b57a445c6394d46a6b46e32b7787663f40d13b 2025-11-03T16:58:01.4651332Z * [new tag] trunk/24b6eb7727f7eef078f9366302f2aa3e1fef1db4 -> trunk/24b6eb7727f7eef078f9366302f2aa3e1fef1db4 2025-11-03T16:58:01.4652726Z * [new tag] trunk/24e94e021a5a96bf16397ea66243070a7be4aaef -> trunk/24e94e021a5a96bf16397ea66243070a7be4aaef 2025-11-03T16:58:01.4654105Z * [new tag] trunk/26534e9809eb2f7cd804fde5152cdd13dda2293f -> trunk/26534e9809eb2f7cd804fde5152cdd13dda2293f 2025-11-03T16:58:01.4655380Z * [new tag] trunk/267d0197bfca0232488d51dd1ff735d619adc2cf -> trunk/267d0197bfca0232488d51dd1ff735d619adc2cf 2025-11-03T16:58:01.4656649Z * [new tag] trunk/2699f5410b6802277db724f145c281814d6c9ac4 -> trunk/2699f5410b6802277db724f145c281814d6c9ac4 2025-11-03T16:58:01.4658070Z * [new tag] trunk/2829d48bd113043a98e81b4ab7fa0f3ca9ae0072 -> trunk/2829d48bd113043a98e81b4ab7fa0f3ca9ae0072 2025-11-03T16:58:01.4659543Z * [new tag] trunk/284716a691580cf0508a7c5a4f9f7306a32092ad -> trunk/284716a691580cf0508a7c5a4f9f7306a32092ad 2025-11-03T16:58:01.4660910Z * [new tag] trunk/298666631baa6634666ac4dd4599f1d52bafa718 -> trunk/298666631baa6634666ac4dd4599f1d52bafa718 2025-11-03T16:58:01.4662248Z * [new tag] trunk/2a058bfecf4f4a47927b4eaae9324c4999a8a794 -> trunk/2a058bfecf4f4a47927b4eaae9324c4999a8a794 2025-11-03T16:58:01.4663608Z * [new tag] trunk/2a5f87decf34b3d0ea7670238e2fd4620ed19e9f -> trunk/2a5f87decf34b3d0ea7670238e2fd4620ed19e9f 2025-11-03T16:58:01.4664961Z * [new tag] trunk/2c9f877fa71c90a1313deeda4b4486d37e72a8a9 -> trunk/2c9f877fa71c90a1313deeda4b4486d37e72a8a9 2025-11-03T16:58:01.4666125Z * [new tag] trunk/2ce894bb1d3e9698306b87091c900a72acac75ea -> trunk/2ce894bb1d3e9698306b87091c900a72acac75ea 2025-11-03T16:58:01.4667552Z * [new tag] trunk/2dc56456cbb2645c37c3176b7d9a263933bdec27 -> trunk/2dc56456cbb2645c37c3176b7d9a263933bdec27 2025-11-03T16:58:01.4668868Z * [new tag] trunk/2de4cf2102fdc35da00cc8070fdbcaf57198a576 -> trunk/2de4cf2102fdc35da00cc8070fdbcaf57198a576 2025-11-03T16:58:01.4670338Z * [new tag] trunk/2df2c316e2929519a5bad029936e9e37ed637b12 -> trunk/2df2c316e2929519a5bad029936e9e37ed637b12 2025-11-03T16:58:01.4671684Z * [new tag] trunk/30157d30f0e0db6a4fa7a2a3ea75e515645e36de -> trunk/30157d30f0e0db6a4fa7a2a3ea75e515645e36de 2025-11-03T16:58:01.4673067Z * [new tag] trunk/3041ede082d3b0f33bec0cef5062eabc5173b446 -> trunk/3041ede082d3b0f33bec0cef5062eabc5173b446 2025-11-03T16:58:01.4674404Z * [new tag] trunk/311ea0dec0c50f395e6dac7b3875e81ee243fceb -> trunk/311ea0dec0c50f395e6dac7b3875e81ee243fceb 2025-11-03T16:58:01.4676027Z * [new tag] trunk/31e42eb732d32e71c9954d6b2104cf1d129370d5 -> trunk/31e42eb732d32e71c9954d6b2104cf1d129370d5 2025-11-03T16:58:01.4677379Z * [new tag] trunk/32066772b3dee643b1657b8957f32b5ac8b1390a -> trunk/32066772b3dee643b1657b8957f32b5ac8b1390a 2025-11-03T16:58:01.4678781Z * [new tag] trunk/32920926f07e573083ecf81a40c898f47f4df631 -> trunk/32920926f07e573083ecf81a40c898f47f4df631 2025-11-03T16:58:01.4680179Z * [new tag] trunk/32fe4f681e2bfb2fdecf20027e29e1aeb6ad5297 -> trunk/32fe4f681e2bfb2fdecf20027e29e1aeb6ad5297 2025-11-03T16:58:01.4681455Z * [new tag] trunk/335b5c7d4bf3295d517902370142f007ca024cd0 -> trunk/335b5c7d4bf3295d517902370142f007ca024cd0 2025-11-03T16:58:01.4682981Z * [new tag] trunk/34d6ef7022be252ce18a9673f75eed1732dff22a -> trunk/34d6ef7022be252ce18a9673f75eed1732dff22a 2025-11-03T16:58:01.4684394Z * [new tag] trunk/35f3572fa483a8edb101d5765564e1ae274f3d45 -> trunk/35f3572fa483a8edb101d5765564e1ae274f3d45 2025-11-03T16:58:01.4685725Z * [new tag] trunk/365ed62f61dde7fc14dca2199692555e5d1abbe1 -> trunk/365ed62f61dde7fc14dca2199692555e5d1abbe1 2025-11-03T16:58:01.4687083Z * [new tag] trunk/369f2d6951d4c12571eb03b95d74ef2f21296847 -> trunk/369f2d6951d4c12571eb03b95d74ef2f21296847 2025-11-03T16:58:01.4688615Z * [new tag] trunk/36a48e7e6d2d01cd47f002140589a1ac2580ef3f -> trunk/36a48e7e6d2d01cd47f002140589a1ac2580ef3f 2025-11-03T16:58:01.4689818Z * [new tag] trunk/3895ce093f3e71b768d032a28a949a9f390c0d00 -> trunk/3895ce093f3e71b768d032a28a949a9f390c0d00 2025-11-03T16:58:01.4691260Z * [new tag] trunk/392acee68a5ca90279109df023583edaef73d50f -> trunk/392acee68a5ca90279109df023583edaef73d50f 2025-11-03T16:58:01.4692681Z * [new tag] trunk/398775a43e9808205f75c81d36f5087117d3f3f4 -> trunk/398775a43e9808205f75c81d36f5087117d3f3f4 2025-11-03T16:58:01.4694123Z * [new tag] trunk/398fdd32bbf7b5f8a2b67e637c8a2e46572d9f28 -> trunk/398fdd32bbf7b5f8a2b67e637c8a2e46572d9f28 2025-11-03T16:58:01.4695298Z * [new tag] trunk/39e5cdddf7e57881c52473d1288a66f0222527e1 -> trunk/39e5cdddf7e57881c52473d1288a66f0222527e1 2025-11-03T16:58:01.4696750Z * [new tag] trunk/3b4315940d371385674f957f810cb0f1ca0060bd -> trunk/3b4315940d371385674f957f810cb0f1ca0060bd 2025-11-03T16:58:01.4698088Z * [new tag] trunk/3b5d38a3bc55c32232cee8cf87e492c965f276fd -> trunk/3b5d38a3bc55c32232cee8cf87e492c965f276fd 2025-11-03T16:58:01.4699557Z * [new tag] trunk/3ca216ae172e35adde34a319a1a01faaf218e7c5 -> trunk/3ca216ae172e35adde34a319a1a01faaf218e7c5 2025-11-03T16:58:01.4701049Z * [new tag] trunk/3cc5949dc2c6e588d42cff36121819a8539902a7 -> trunk/3cc5949dc2c6e588d42cff36121819a8539902a7 2025-11-03T16:58:01.4702391Z * [new tag] trunk/3d4a2d8a93ad57e9fd1c4e5a03dd431dd0123bb3 -> trunk/3d4a2d8a93ad57e9fd1c4e5a03dd431dd0123bb3 2025-11-03T16:58:01.4703929Z * [new tag] trunk/3d4ca228beb851e2f1cc24d8300804d5884973c1 -> trunk/3d4ca228beb851e2f1cc24d8300804d5884973c1 2025-11-03T16:58:01.4704811Z * [new tag] trunk/3dc92d69ed40fd952244e54bbda0240928756654 -> trunk/3dc92d69ed40fd952244e54bbda0240928756654 2025-11-03T16:58:01.4706473Z * [new tag] trunk/3eddf049221fc04c2ac9d4af53c00305484ef325 -> trunk/3eddf049221fc04c2ac9d4af53c00305484ef325 2025-11-03T16:58:01.4707812Z * [new tag] trunk/3f1824742cac2ffb9a3afd90953c492c6c7f2f50 -> trunk/3f1824742cac2ffb9a3afd90953c492c6c7f2f50 2025-11-03T16:58:01.4708816Z * [new tag] trunk/3f5401020b3fd66238a5eda31576b898d086e14e -> trunk/3f5401020b3fd66238a5eda31576b898d086e14e 2025-11-03T16:58:01.4711047Z * [new tag] trunk/3f6538febd727b782e6e13cfd026a309fb14351d -> trunk/3f6538febd727b782e6e13cfd026a309fb14351d 2025-11-03T16:58:01.4712333Z * [new tag] trunk/3f69b4d9b46b1de627896c457bce1937e53d82e5 -> trunk/3f69b4d9b46b1de627896c457bce1937e53d82e5 2025-11-03T16:58:01.4713664Z * [new tag] trunk/401c2f965726fda8f78b08f3745e11008b38e7bf -> trunk/401c2f965726fda8f78b08f3745e11008b38e7bf 2025-11-03T16:58:01.4714913Z * [new tag] trunk/4295a9a1587cc62a3ef4ad9b1e552c78471d5664 -> trunk/4295a9a1587cc62a3ef4ad9b1e552c78471d5664 2025-11-03T16:58:01.4716236Z * [new tag] trunk/4316df857c9e7f301142eb54d06a85a43f8d617b -> trunk/4316df857c9e7f301142eb54d06a85a43f8d617b 2025-11-03T16:58:01.4717600Z * [new tag] trunk/43c30f607eeca0d3e9a26911d9c2131fc250eadd -> trunk/43c30f607eeca0d3e9a26911d9c2131fc250eadd 2025-11-03T16:58:01.4719047Z * [new tag] trunk/45c3f02d69b0a8447381fbc3f3f8731ab0d48ed1 -> trunk/45c3f02d69b0a8447381fbc3f3f8731ab0d48ed1 2025-11-03T16:58:01.4720422Z * [new tag] trunk/467c21ad9ae4133c20a3c098a0355e9ac20d48aa -> trunk/467c21ad9ae4133c20a3c098a0355e9ac20d48aa 2025-11-03T16:58:01.4721803Z * [new tag] trunk/46d17e887150c5de8a74314b6d9fcba1feca259b -> trunk/46d17e887150c5de8a74314b6d9fcba1feca259b 2025-11-03T16:58:01.4723218Z * [new tag] trunk/476b149a00cf52ceba2b2d8b6e14989d2bf540e2 -> trunk/476b149a00cf52ceba2b2d8b6e14989d2bf540e2 2025-11-03T16:58:01.4724219Z * [new tag] trunk/47ec1e9990b89bf9705dc463cad587bd357bc848 -> trunk/47ec1e9990b89bf9705dc463cad587bd357bc848 2025-11-03T16:58:01.4725794Z * [new tag] trunk/47f00243106918eec3ef9119272977fdf2864ad8 -> trunk/47f00243106918eec3ef9119272977fdf2864ad8 2025-11-03T16:58:01.4727162Z * [new tag] trunk/47f50cfd456313d8b46fcc7a1f6de477aa0a5aee -> trunk/47f50cfd456313d8b46fcc7a1f6de477aa0a5aee 2025-11-03T16:58:01.4728505Z * [new tag] trunk/483845a9c42267a6a139b7382e5ee340502376f8 -> trunk/483845a9c42267a6a139b7382e5ee340502376f8 2025-11-03T16:58:01.4729891Z * [new tag] trunk/48c3b71ecc245cb422ec9ab4c7a2078f7ef0b3df -> trunk/48c3b71ecc245cb422ec9ab4c7a2078f7ef0b3df 2025-11-03T16:58:01.4731479Z * [new tag] trunk/48e672d14994c912caccba0c7f2f5df54c0d86b3 -> trunk/48e672d14994c912caccba0c7f2f5df54c0d86b3 2025-11-03T16:58:01.4732835Z * [new tag] trunk/4a7bc1d522bed4f3f792b4641372ec3d06b79b5d -> trunk/4a7bc1d522bed4f3f792b4641372ec3d06b79b5d 2025-11-03T16:58:01.4734206Z * [new tag] trunk/4a7fefd7c783913b811f49e5d6242086758dca6c -> trunk/4a7fefd7c783913b811f49e5d6242086758dca6c 2025-11-03T16:58:01.4735545Z * [new tag] trunk/4a94591321a62b31658f9259680de87e4be9e50d -> trunk/4a94591321a62b31658f9259680de87e4be9e50d 2025-11-03T16:58:01.4736974Z * [new tag] trunk/4acc66f1192ab7743abcc50383aefc5447447f9d -> trunk/4acc66f1192ab7743abcc50383aefc5447447f9d 2025-11-03T16:58:01.4738381Z * [new tag] trunk/4c38887346b6a6185bb3e170621dc13d5f34e9ce -> trunk/4c38887346b6a6185bb3e170621dc13d5f34e9ce 2025-11-03T16:58:01.4739802Z * [new tag] trunk/4cc64d62346edb2e13320f20740f41ba41f64c70 -> trunk/4cc64d62346edb2e13320f20740f41ba41f64c70 2025-11-03T16:58:01.4741134Z * [new tag] trunk/4e6afa8c07b73e0b339e3baacbc73102748e61dd -> trunk/4e6afa8c07b73e0b339e3baacbc73102748e61dd 2025-11-03T16:58:01.4742560Z * [new tag] trunk/4e7232c5daf753e04e8f4189229e3c33888a33e5 -> trunk/4e7232c5daf753e04e8f4189229e3c33888a33e5 2025-11-03T16:58:01.4744115Z * [new tag] trunk/4e8ba37ce3320cac033019131b817b4aca29699b -> trunk/4e8ba37ce3320cac033019131b817b4aca29699b 2025-11-03T16:58:01.4745411Z * [new tag] trunk/4fada51ada8b3190971f2f47d97346057320866a -> trunk/4fada51ada8b3190971f2f47d97346057320866a 2025-11-03T16:58:01.4746763Z * [new tag] trunk/5016e7b2eb73589e35c7a8f0f0ea3bfba221f17d -> trunk/5016e7b2eb73589e35c7a8f0f0ea3bfba221f17d 2025-11-03T16:58:01.4748109Z * [new tag] trunk/51667435f50c025ca3655ff0eeb917b4ef0ffb78 -> trunk/51667435f50c025ca3655ff0eeb917b4ef0ffb78 2025-11-03T16:58:01.4749514Z * [new tag] trunk/52db60170dd192afa48233f03574a8d8bcd39d35 -> trunk/52db60170dd192afa48233f03574a8d8bcd39d35 2025-11-03T16:58:01.4750902Z * [new tag] trunk/544b443ea1d1a9b19e65f981168a01cb87a2d333 -> trunk/544b443ea1d1a9b19e65f981168a01cb87a2d333 2025-11-03T16:58:01.4752218Z * [new tag] trunk/551921d484d209e9a95d77262b1472bc1d18422d -> trunk/551921d484d209e9a95d77262b1472bc1d18422d 2025-11-03T16:58:01.4753747Z * [new tag] trunk/56838bad5f197d32c0ee77ce2eaa7a0cb462af68 -> trunk/56838bad5f197d32c0ee77ce2eaa7a0cb462af68 2025-11-03T16:58:01.4767573Z * [new tag] trunk/56a809aa0765ec2edea57ab37f92b1e78a214dcc -> trunk/56a809aa0765ec2edea57ab37f92b1e78a214dcc 2025-11-03T16:58:01.4767906Z * [new tag] trunk/56afad4eb300be2008d8b24581bcddcd62021036 -> trunk/56afad4eb300be2008d8b24581bcddcd62021036 2025-11-03T16:58:01.4768197Z * [new tag] trunk/572cc12b42f6edf76affd45ea9a23f4c8db5c7e1 -> trunk/572cc12b42f6edf76affd45ea9a23f4c8db5c7e1 2025-11-03T16:58:01.4768524Z * [new tag] trunk/5849eea129b47556630aac4276529afe287b91e1 -> trunk/5849eea129b47556630aac4276529afe287b91e1 2025-11-03T16:58:01.4768801Z * [new tag] trunk/59ddfb69a731970bf5762fbea43ff2987425d3ea -> trunk/59ddfb69a731970bf5762fbea43ff2987425d3ea 2025-11-03T16:58:01.4769070Z * [new tag] trunk/5a3930abbc19eac9a179455df82e206e69765ed2 -> trunk/5a3930abbc19eac9a179455df82e206e69765ed2 2025-11-03T16:58:01.4769356Z * [new tag] trunk/5bcfdae71da2f405a89bace6b09d616107cc3965 -> trunk/5bcfdae71da2f405a89bace6b09d616107cc3965 2025-11-03T16:58:01.4769620Z * [new tag] trunk/5cbdade914d6bc9f67827ef917974980a74e9372 -> trunk/5cbdade914d6bc9f67827ef917974980a74e9372 2025-11-03T16:58:01.4769875Z * [new tag] trunk/5cdbcb52334222d1a2667dabd6da993d1e9ddc93 -> trunk/5cdbcb52334222d1a2667dabd6da993d1e9ddc93 2025-11-03T16:58:01.4770143Z * [new tag] trunk/5d0b3e28dc18fac3c09edb24078998371f878f82 -> trunk/5d0b3e28dc18fac3c09edb24078998371f878f82 2025-11-03T16:58:01.4770399Z * [new tag] trunk/5d288bc3f73873887f681e15af83c5525e6a60bd -> trunk/5d288bc3f73873887f681e15af83c5525e6a60bd 2025-11-03T16:58:01.4770705Z * [new tag] trunk/5d6230779d920e9c449c22d22fd0f69fe7b73632 -> trunk/5d6230779d920e9c449c22d22fd0f69fe7b73632 2025-11-03T16:58:01.4772435Z * [new tag] trunk/5e05a0ae99c320bc570a3a42c5fcd6bbf2b2e9b4 -> trunk/5e05a0ae99c320bc570a3a42c5fcd6bbf2b2e9b4 2025-11-03T16:58:01.4773454Z * [new tag] trunk/5e7272b60a5d7203fd1d048ecac6f202526cac63 -> trunk/5e7272b60a5d7203fd1d048ecac6f202526cac63 2025-11-03T16:58:01.4775070Z * [new tag] trunk/5e769ff86780a7ffd561615dbf4b0defe80cfbb9 -> trunk/5e769ff86780a7ffd561615dbf4b0defe80cfbb9 2025-11-03T16:58:01.4776498Z * [new tag] trunk/5ebf74a655b3335fb079ee03f41228a77bcc630e -> trunk/5ebf74a655b3335fb079ee03f41228a77bcc630e 2025-11-03T16:58:01.4777842Z * [new tag] trunk/5fc2c7a2a11444c31338da9134452810f3ca8acd -> trunk/5fc2c7a2a11444c31338da9134452810f3ca8acd 2025-11-03T16:58:01.4779242Z * [new tag] trunk/5fd1d41e62ce4436c2bb08c7f4702df0021e2c62 -> trunk/5fd1d41e62ce4436c2bb08c7f4702df0021e2c62 2025-11-03T16:58:01.4780833Z * [new tag] trunk/60333de85dec845d6a0e651764c16a730a27828d -> trunk/60333de85dec845d6a0e651764c16a730a27828d 2025-11-03T16:58:01.4782276Z * [new tag] trunk/6096c0fc746dc99189510a1845fd45c667178c03 -> trunk/6096c0fc746dc99189510a1845fd45c667178c03 2025-11-03T16:58:01.4783654Z * [new tag] trunk/60bcb4ee8844e7bac7278ff662c7d5735a326758 -> trunk/60bcb4ee8844e7bac7278ff662c7d5735a326758 2025-11-03T16:58:01.4785046Z * [new tag] trunk/610c09f8f47585e811945e48abc895f88e09bda3 -> trunk/610c09f8f47585e811945e48abc895f88e09bda3 2025-11-03T16:58:01.4786457Z * [new tag] trunk/61bad3c1ea718311d7c256b05614d207d73276e8 -> trunk/61bad3c1ea718311d7c256b05614d207d73276e8 2025-11-03T16:58:01.4787902Z * [new tag] trunk/61bcc8d75ad065a40b2d7953046616d5979233b9 -> trunk/61bcc8d75ad065a40b2d7953046616d5979233b9 2025-11-03T16:58:01.4789571Z * [new tag] trunk/6268883f9c081b88b611ef01be3821503ad180ca -> trunk/6268883f9c081b88b611ef01be3821503ad180ca 2025-11-03T16:58:01.4790962Z * [new tag] trunk/629293f568229adaf60cdaf57fe69306d7782c2d -> trunk/629293f568229adaf60cdaf57fe69306d7782c2d 2025-11-03T16:58:01.4792378Z * [new tag] trunk/639a0b12395e72a8a220dc8443317f4ab3b3f1a0 -> trunk/639a0b12395e72a8a220dc8443317f4ab3b3f1a0 2025-11-03T16:58:01.4793828Z * [new tag] trunk/6530bc70fbc7d600ab0e600c02ecd9b341ce8237 -> trunk/6530bc70fbc7d600ab0e600c02ecd9b341ce8237 2025-11-03T16:58:01.4795260Z * [new tag] trunk/657f8c3e21bd8901dd8ce79ca9a54a45b27f604f -> trunk/657f8c3e21bd8901dd8ce79ca9a54a45b27f604f 2025-11-03T16:58:01.4796649Z * [new tag] trunk/65868156c613521e9635eb3f8eb3f404826ac1f9 -> trunk/65868156c613521e9635eb3f8eb3f404826ac1f9 2025-11-03T16:58:01.4797939Z * [new tag] trunk/687c15c0b3f01118536413d21efcd052838fa10d -> trunk/687c15c0b3f01118536413d21efcd052838fa10d 2025-11-03T16:58:01.4799337Z * [new tag] trunk/68b3984b77a19ed2b2afcbcae201e8b06e20f8bc -> trunk/68b3984b77a19ed2b2afcbcae201e8b06e20f8bc 2025-11-03T16:58:01.4801180Z * [new tag] trunk/694d205143b03cb26626cb86ecb399df6d5e3dd8 -> trunk/694d205143b03cb26626cb86ecb399df6d5e3dd8 2025-11-03T16:58:01.4802598Z * [new tag] trunk/694db5f54927697c9e914d35029f7e5bd9b85b96 -> trunk/694db5f54927697c9e914d35029f7e5bd9b85b96 2025-11-03T16:58:01.4804074Z * [new tag] trunk/695cb0d342f67db563fb57ce1a5f3261b073b790 -> trunk/695cb0d342f67db563fb57ce1a5f3261b073b790 2025-11-03T16:58:01.4805436Z * [new tag] trunk/69be99ee51bd0439fead4c83d1ed3eea9e1e073e -> trunk/69be99ee51bd0439fead4c83d1ed3eea9e1e073e 2025-11-03T16:58:01.4806827Z * [new tag] trunk/69fb3ebb5dbec0bd4cfd73cc9a6c870b7f604840 -> trunk/69fb3ebb5dbec0bd4cfd73cc9a6c870b7f604840 2025-11-03T16:58:01.4808215Z * [new tag] trunk/6a5a436624d08feff1de191e61eb4d4773f73bdb -> trunk/6a5a436624d08feff1de191e61eb4d4773f73bdb 2025-11-03T16:58:01.4809957Z * [new tag] trunk/6c7cad697287dff7e47545ae4b1df1e407185868 -> trunk/6c7cad697287dff7e47545ae4b1df1e407185868 2025-11-03T16:58:01.4811449Z * [new tag] trunk/6d5e651a5024acdbaf9633d9d1d9dc33c6a283c5 -> trunk/6d5e651a5024acdbaf9633d9d1d9dc33c6a283c5 2025-11-03T16:58:01.4812678Z * [new tag] trunk/6ecd6b23b6d2ab17da269ceebe9f944153f67083 -> trunk/6ecd6b23b6d2ab17da269ceebe9f944153f67083 2025-11-03T16:58:01.4814060Z * [new tag] trunk/7045aab143b6d89c2e7de11bc8cbf919d8fd78d0 -> trunk/7045aab143b6d89c2e7de11bc8cbf919d8fd78d0 2025-11-03T16:58:01.4815460Z * [new tag] trunk/70aeb491987401a380914a72c1b9d19e4c882429 -> trunk/70aeb491987401a380914a72c1b9d19e4c882429 2025-11-03T16:58:01.4816727Z * [new tag] trunk/7379972cc0e8a4b4d88b4bea5c8be0aeffdb076a -> trunk/7379972cc0e8a4b4d88b4bea5c8be0aeffdb076a 2025-11-03T16:58:01.4818066Z * [new tag] trunk/73da7a40b6dd0509b179f0ca92d9fa79973ee306 -> trunk/73da7a40b6dd0509b179f0ca92d9fa79973ee306 2025-11-03T16:58:01.4819607Z * [new tag] trunk/74336f8c77f1a906314c274e43239306c8077e90 -> trunk/74336f8c77f1a906314c274e43239306c8077e90 2025-11-03T16:58:01.4820953Z * [new tag] trunk/748162223768fb9f7daaca1f7f93d0bc37be4de9 -> trunk/748162223768fb9f7daaca1f7f93d0bc37be4de9 2025-11-03T16:58:01.4822271Z * [new tag] trunk/753d9bd8063356c0e31517ef9aabd814efd754c5 -> trunk/753d9bd8063356c0e31517ef9aabd814efd754c5 2025-11-03T16:58:01.4823742Z * [new tag] trunk/7563f61cc8a40a5ba21a498a2d98895b4eec3f39 -> trunk/7563f61cc8a40a5ba21a498a2d98895b4eec3f39 2025-11-03T16:58:01.4825287Z * [new tag] trunk/75f798e05b35cbdf1814673e745b9b7817bb29cd -> trunk/75f798e05b35cbdf1814673e745b9b7817bb29cd 2025-11-03T16:58:01.4826679Z * [new tag] trunk/764c54ecaeb8b0da486b009f4d4ef1fa280e2517 -> trunk/764c54ecaeb8b0da486b009f4d4ef1fa280e2517 2025-11-03T16:58:01.4828020Z * [new tag] trunk/76780b1a3d208b142f067bdd655e7d4c5615efe1 -> trunk/76780b1a3d208b142f067bdd655e7d4c5615efe1 2025-11-03T16:58:01.4829203Z * [new tag] trunk/7692fa09cdd47fdbec66efca9edf42a318e6c6e4 -> trunk/7692fa09cdd47fdbec66efca9edf42a318e6c6e4 2025-11-03T16:58:01.4830828Z * [new tag] trunk/76b2c37045e52540ec51e967aa7b6436a6b9b174 -> trunk/76b2c37045e52540ec51e967aa7b6436a6b9b174 2025-11-03T16:58:01.4832162Z * [new tag] trunk/76bb27e248750fc558ec4ea9050e09a6f171e9bc -> trunk/76bb27e248750fc558ec4ea9050e09a6f171e9bc 2025-11-03T16:58:01.4833590Z * [new tag] trunk/774abb018eb1d44c07fda9fa78549755ac7d4367 -> trunk/774abb018eb1d44c07fda9fa78549755ac7d4367 2025-11-03T16:58:01.4835014Z * [new tag] trunk/791ca80d3a2ca72a77b5dedc75bbcbb35b859cfc -> trunk/791ca80d3a2ca72a77b5dedc75bbcbb35b859cfc 2025-11-03T16:58:01.4836417Z * [new tag] trunk/797cd80b2670a51601f997f8c67387bd30440a36 -> trunk/797cd80b2670a51601f997f8c67387bd30440a36 2025-11-03T16:58:01.4837824Z * [new tag] trunk/79aa88cc5dcc55597013cbe36f0e00557b185803 -> trunk/79aa88cc5dcc55597013cbe36f0e00557b185803 2025-11-03T16:58:01.4839051Z * [new tag] trunk/79aee77381b21d41c77148e5ff84c4b351aaf144 -> trunk/79aee77381b21d41c77148e5ff84c4b351aaf144 2025-11-03T16:58:01.4840578Z * [new tag] trunk/7a0cd8ed092a62123fcc1f4c40de2669f46fe2b1 -> trunk/7a0cd8ed092a62123fcc1f4c40de2669f46fe2b1 2025-11-03T16:58:01.4841970Z * [new tag] trunk/7ae8aaf4c02e7cc14b2e3cce0c88a7333dff439c -> trunk/7ae8aaf4c02e7cc14b2e3cce0c88a7333dff439c 2025-11-03T16:58:01.4843327Z * [new tag] trunk/7c203b8420d26bd707c6052f93a4ba77dca2b2c0 -> trunk/7c203b8420d26bd707c6052f93a4ba77dca2b2c0 2025-11-03T16:58:01.4844747Z * [new tag] trunk/7ce723d21cc2a1a874f988f3459f0da834c75daa -> trunk/7ce723d21cc2a1a874f988f3459f0da834c75daa 2025-11-03T16:58:01.4846129Z * [new tag] trunk/7d16fcf2df9b240895056107a015735774cbe4b5 -> trunk/7d16fcf2df9b240895056107a015735774cbe4b5 2025-11-03T16:58:01.4847555Z * [new tag] trunk/7d39401fa07eb2f7782d8f234aa19f8b2abff481 -> trunk/7d39401fa07eb2f7782d8f234aa19f8b2abff481 2025-11-03T16:58:01.4849168Z * [new tag] trunk/7d67a41db4801632d5862e99c79b766284dfd307 -> trunk/7d67a41db4801632d5862e99c79b766284dfd307 2025-11-03T16:58:01.4850584Z * [new tag] trunk/7e3b9d105e3359e31fb252becb01d789a72dc8d9 -> trunk/7e3b9d105e3359e31fb252becb01d789a72dc8d9 2025-11-03T16:58:01.4851987Z * [new tag] trunk/7f7a28046b5927ddbd8ef95d00aa1ca76f773bb1 -> trunk/7f7a28046b5927ddbd8ef95d00aa1ca76f773bb1 2025-11-03T16:58:01.4853412Z * [new tag] trunk/80ba6e458ff2bd178da9ed62e27a61fe46e4252a -> trunk/80ba6e458ff2bd178da9ed62e27a61fe46e4252a 2025-11-03T16:58:01.4854859Z * [new tag] trunk/8101fd46d4619d0f9ef69e5829d1ee6f77289b72 -> trunk/8101fd46d4619d0f9ef69e5829d1ee6f77289b72 2025-11-03T16:58:01.4856305Z * [new tag] trunk/8110ce02a2830f562ed60607e30e2f29aed119dd -> trunk/8110ce02a2830f562ed60607e30e2f29aed119dd 2025-11-03T16:58:01.4857658Z * [new tag] trunk/81fa4a204cf268bc55e96137fa803ae4938472cd -> trunk/81fa4a204cf268bc55e96137fa803ae4938472cd 2025-11-03T16:58:01.4859042Z * [new tag] trunk/8209a0506b399b40a48716126ee22737aa4607c5 -> trunk/8209a0506b399b40a48716126ee22737aa4607c5 2025-11-03T16:58:01.4860457Z * [new tag] trunk/8221ee6db9fd319c249087ad9bb9abaa7d56e113 -> trunk/8221ee6db9fd319c249087ad9bb9abaa7d56e113 2025-11-03T16:58:01.4861889Z * [new tag] trunk/82d86bacf3c34760d6ab8a8a1eafa22d0ea259d9 -> trunk/82d86bacf3c34760d6ab8a8a1eafa22d0ea259d9 2025-11-03T16:58:01.4863336Z * [new tag] trunk/82fafb3304f8baffdb2d54578926d8bb4aba621c -> trunk/82fafb3304f8baffdb2d54578926d8bb4aba621c 2025-11-03T16:58:01.4864874Z * [new tag] trunk/82ff07c7884d478ddd5d638bebbb938e55c9bebf -> trunk/82ff07c7884d478ddd5d638bebbb938e55c9bebf 2025-11-03T16:58:01.4866277Z * [new tag] trunk/83cc38d9c1b4371d4395bda8b6861b98986f527b -> trunk/83cc38d9c1b4371d4395bda8b6861b98986f527b 2025-11-03T16:58:01.4867390Z * [new tag] trunk/840d63c12d255dac1ae3c5e442c6ea6eb87a7256 -> trunk/840d63c12d255dac1ae3c5e442c6ea6eb87a7256 2025-11-03T16:58:01.4868977Z * [new tag] trunk/8417981c96b7577f0228226ccf76d23842173144 -> trunk/8417981c96b7577f0228226ccf76d23842173144 2025-11-03T16:58:01.4870322Z * [new tag] trunk/845da9c8172b0e6e8b1e326359379dc6e9ddbddf -> trunk/845da9c8172b0e6e8b1e326359379dc6e9ddbddf 2025-11-03T16:58:01.4871708Z * [new tag] trunk/84776e13744db6d59b41a063bb8714e2bffe7a06 -> trunk/84776e13744db6d59b41a063bb8714e2bffe7a06 2025-11-03T16:58:01.4873201Z * [new tag] trunk/84a2715d341f068a26a281a252c3907bbe827d9b -> trunk/84a2715d341f068a26a281a252c3907bbe827d9b 2025-11-03T16:58:01.4874636Z * [new tag] trunk/84fe84850383fc11e7150681742bf53f11cd425e -> trunk/84fe84850383fc11e7150681742bf53f11cd425e 2025-11-03T16:58:01.4876015Z * [new tag] trunk/856a7a529872aa2d222e2491708cb458c682bb93 -> trunk/856a7a529872aa2d222e2491708cb458c682bb93 2025-11-03T16:58:01.4877419Z * [new tag] trunk/85a7c745aa5a2428a7cc640ef4b304296fe07c42 -> trunk/85a7c745aa5a2428a7cc640ef4b304296fe07c42 2025-11-03T16:58:01.4878959Z * [new tag] trunk/85b035ca9c230199d91ff86d362def01ad36dc67 -> trunk/85b035ca9c230199d91ff86d362def01ad36dc67 2025-11-03T16:58:01.4880382Z * [new tag] trunk/85b85f6c2c7e6ca98abf9d9a38410d4b5703d64a -> trunk/85b85f6c2c7e6ca98abf9d9a38410d4b5703d64a 2025-11-03T16:58:01.4881771Z * [new tag] trunk/877f126e353f537417c53bff073fc56d3faecd60 -> trunk/877f126e353f537417c53bff073fc56d3faecd60 2025-11-03T16:58:01.4883275Z * [new tag] trunk/8887a33ede05c3e77286447a8ebd82b458878943 -> trunk/8887a33ede05c3e77286447a8ebd82b458878943 2025-11-03T16:58:01.4884595Z * [new tag] trunk/895795f07cc0348ebe0b7e369d555b545d76eb9f -> trunk/895795f07cc0348ebe0b7e369d555b545d76eb9f 2025-11-03T16:58:01.4885988Z * [new tag] trunk/8aa087a29dcb5eba4cffd6a452e7f327abfd14b2 -> trunk/8aa087a29dcb5eba4cffd6a452e7f327abfd14b2 2025-11-03T16:58:01.4887324Z * [new tag] trunk/8af9ed08243222853cf8ebfc9177c5b1ea364d00 -> trunk/8af9ed08243222853cf8ebfc9177c5b1ea364d00 2025-11-03T16:58:01.4888787Z * [new tag] trunk/8b188647cfdc1355070ccd5aaa18a8060d4f67bf -> trunk/8b188647cfdc1355070ccd5aaa18a8060d4f67bf 2025-11-03T16:58:01.4890268Z * [new tag] trunk/8d4e48831e5abb6cbca8d97a2df6c47b770c303b -> trunk/8d4e48831e5abb6cbca8d97a2df6c47b770c303b 2025-11-03T16:58:01.4892099Z * [new tag] trunk/8d599045cf4102e451a9e8a9ff215d053ebbe0e8 -> trunk/8d599045cf4102e451a9e8a9ff215d053ebbe0e8 2025-11-03T16:58:01.4893499Z * [new tag] trunk/8e1e4ee8e0c43202085b47c9d2b05b2c0d13b368 -> trunk/8e1e4ee8e0c43202085b47c9d2b05b2c0d13b368 2025-11-03T16:58:01.4895032Z * [new tag] trunk/8f40a0c634a65246889b478b7f14da7a448174a9 -> trunk/8f40a0c634a65246889b478b7f14da7a448174a9 2025-11-03T16:58:01.4896451Z * [new tag] trunk/8f51556daa347041fc7ac06a832d2fdc62ca594d -> trunk/8f51556daa347041fc7ac06a832d2fdc62ca594d 2025-11-03T16:58:01.4897866Z * [new tag] trunk/904abfc2ca285910b7b1458779bd9c63619f2f7d -> trunk/904abfc2ca285910b7b1458779bd9c63619f2f7d 2025-11-03T16:58:01.4899303Z * [new tag] trunk/90519402c2006237f891289a0afdec804515aa73 -> trunk/90519402c2006237f891289a0afdec804515aa73 2025-11-03T16:58:01.4900743Z * [new tag] trunk/90b30ebf7ecd7b0da9deeb35b165de9e28e90aff -> trunk/90b30ebf7ecd7b0da9deeb35b165de9e28e90aff 2025-11-03T16:58:01.4902206Z * [new tag] trunk/90d7be35e980a39cb4044f8fefea33a6167e5733 -> trunk/90d7be35e980a39cb4044f8fefea33a6167e5733 2025-11-03T16:58:01.4903590Z * [new tag] trunk/9139368b64ebcc32c8c1190fd114c7f4733e7c6d -> trunk/9139368b64ebcc32c8c1190fd114c7f4733e7c6d 2025-11-03T16:58:01.4905037Z * [new tag] trunk/92381a5aa7c27b740d375dad2ff9ca7be3c1bc24 -> trunk/92381a5aa7c27b740d375dad2ff9ca7be3c1bc24 2025-11-03T16:58:01.4906536Z * [new tag] trunk/924482a6f63305b0f00760170ad513305700b79c -> trunk/924482a6f63305b0f00760170ad513305700b79c 2025-11-03T16:58:01.4907923Z * [new tag] trunk/9261a1fb128412201ef009d30844a2417364d73b -> trunk/9261a1fb128412201ef009d30844a2417364d73b 2025-11-03T16:58:01.4909305Z * [new tag] trunk/93a70c717a5636505da0b4ccfbca0e111c88c6a8 -> trunk/93a70c717a5636505da0b4ccfbca0e111c88c6a8 2025-11-03T16:58:01.4911027Z * [new tag] trunk/94eaeb9cb812744251cef11d147e93b00a633823 -> trunk/94eaeb9cb812744251cef11d147e93b00a633823 2025-11-03T16:58:01.4912465Z * [new tag] trunk/94f2657c4b534136aa8958bc35d44ceac5ccd60c -> trunk/94f2657c4b534136aa8958bc35d44ceac5ccd60c 2025-11-03T16:58:01.4913913Z * [new tag] trunk/9501405de686068c5d7d0e4eab21e3ef0b95cc73 -> trunk/9501405de686068c5d7d0e4eab21e3ef0b95cc73 2025-11-03T16:58:01.4915379Z * [new tag] trunk/95b55347730b071d0b79df5ddd477db9c7f91fe4 -> trunk/95b55347730b071d0b79df5ddd477db9c7f91fe4 2025-11-03T16:58:01.4916654Z * [new tag] trunk/96b61844a722d2ae8b1a2dc283a6c0060b40782e -> trunk/96b61844a722d2ae8b1a2dc283a6c0060b40782e 2025-11-03T16:58:01.4918109Z * [new tag] trunk/972030fe2ef0a7500afb7acb86a5f971c30536b9 -> trunk/972030fe2ef0a7500afb7acb86a5f971c30536b9 2025-11-03T16:58:01.4919757Z * [new tag] trunk/984e64b2cddecfe43501a7e01a1cce0c25473f54 -> trunk/984e64b2cddecfe43501a7e01a1cce0c25473f54 2025-11-03T16:58:01.4921179Z * [new tag] trunk/98d640bb113a0b6a39c590bbe36873faeb71ceb0 -> trunk/98d640bb113a0b6a39c590bbe36873faeb71ceb0 2025-11-03T16:58:01.4922564Z * [new tag] trunk/9901d4441862cc0be881f913e1a7f932f7cb0177 -> trunk/9901d4441862cc0be881f913e1a7f932f7cb0177 2025-11-03T16:58:01.4924090Z * [new tag] trunk/9940e894ea9c4374691ec49eba96db0a3159a8ae -> trunk/9940e894ea9c4374691ec49eba96db0a3159a8ae 2025-11-03T16:58:01.4925519Z * [new tag] trunk/9970fb97fff5fb5ab0f0f3acfa73558cc5c1618a -> trunk/9970fb97fff5fb5ab0f0f3acfa73558cc5c1618a 2025-11-03T16:58:01.4926984Z * [new tag] trunk/99b05d1b786dc1f2448c23617bbb80d441a5d64a -> trunk/99b05d1b786dc1f2448c23617bbb80d441a5d64a 2025-11-03T16:58:01.4928386Z * [new tag] trunk/99e07c39ec3e9cf4573da6e6737c3d0362bc6b13 -> trunk/99e07c39ec3e9cf4573da6e6737c3d0362bc6b13 2025-11-03T16:58:01.4929808Z * [new tag] trunk/9a91486e453f300804c7031d0e904b0e2af7d802 -> trunk/9a91486e453f300804c7031d0e904b0e2af7d802 2025-11-03T16:58:01.4931285Z * [new tag] trunk/9c22bbb2dce31b854e3387db77eaff501434f352 -> trunk/9c22bbb2dce31b854e3387db77eaff501434f352 2025-11-03T16:58:01.4932875Z * [new tag] trunk/9d6597b1e9f95f4477e88c5bee473c2850c2e63f -> trunk/9d6597b1e9f95f4477e88c5bee473c2850c2e63f 2025-11-03T16:58:01.4934207Z * [new tag] trunk/9ee1afbf66a6533c712956a3ba5e8f60c5875a74 -> trunk/9ee1afbf66a6533c712956a3ba5e8f60c5875a74 2025-11-03T16:58:01.4935478Z * [new tag] trunk/9f9dbe0a9aa5e5c985f36ad6cc736b3fb3506056 -> trunk/9f9dbe0a9aa5e5c985f36ad6cc736b3fb3506056 2025-11-03T16:58:01.4936913Z * [new tag] trunk/a04edcb27acc6e1dd0fc27233738e6dec718b109 -> trunk/a04edcb27acc6e1dd0fc27233738e6dec718b109 2025-11-03T16:58:01.4938163Z * [new tag] trunk/a076b4d7ac09b7be59915a088b348bcbd952c8bd -> trunk/a076b4d7ac09b7be59915a088b348bcbd952c8bd 2025-11-03T16:58:01.4939563Z * [new tag] trunk/a186aa8d6cc197054deba309605f4b75ca59899c -> trunk/a186aa8d6cc197054deba309605f4b75ca59899c 2025-11-03T16:58:01.4940836Z * [new tag] trunk/a19e92d43355489c8d33325321e688a6e6182267 -> trunk/a19e92d43355489c8d33325321e688a6e6182267 2025-11-03T16:58:01.4942256Z * [new tag] trunk/a1eb6b55387b6ae21cf9bcad76a471a2820fa1b0 -> trunk/a1eb6b55387b6ae21cf9bcad76a471a2820fa1b0 2025-11-03T16:58:01.4943833Z * [new tag] trunk/a25818cf7ee2c0ed5c862dff214dc46a30211671 -> trunk/a25818cf7ee2c0ed5c862dff214dc46a30211671 2025-11-03T16:58:01.4945225Z * [new tag] trunk/a2da69385abb4800732e698afacdddd78e9993aa -> trunk/a2da69385abb4800732e698afacdddd78e9993aa 2025-11-03T16:58:01.4946588Z * [new tag] trunk/a3fe1825aa4d95707875ffa8f15fd00c3b94977d -> trunk/a3fe1825aa4d95707875ffa8f15fd00c3b94977d 2025-11-03T16:58:01.4947986Z * [new tag] trunk/a4077b568f891f8cad948cc70cbb14043791d762 -> trunk/a4077b568f891f8cad948cc70cbb14043791d762 2025-11-03T16:58:01.4949301Z * [new tag] trunk/a4a0378e6bb8afd774e4e253304cf3bf9d7e188a -> trunk/a4a0378e6bb8afd774e4e253304cf3bf9d7e188a 2025-11-03T16:58:01.4950785Z * [new tag] trunk/a51f8772873bc91de8aad1ce865fe8d98a58aec7 -> trunk/a51f8772873bc91de8aad1ce865fe8d98a58aec7 2025-11-03T16:58:01.4952099Z * [new tag] trunk/a5335263d32b5be2b2647661334d81225c3cc3fc -> trunk/a5335263d32b5be2b2647661334d81225c3cc3fc 2025-11-03T16:58:01.4953521Z * [new tag] trunk/a553ea9ea4e71b66a58a1c3c37f18d67ffa4ecd3 -> trunk/a553ea9ea4e71b66a58a1c3c37f18d67ffa4ecd3 2025-11-03T16:58:01.4954826Z * [new tag] trunk/a5c3c08d1099f96548398a637463efb0cbe115ae -> trunk/a5c3c08d1099f96548398a637463efb0cbe115ae 2025-11-03T16:58:01.4956272Z * [new tag] trunk/a5f00077fc9623c762c755c926b7eee60eb0ec3f -> trunk/a5f00077fc9623c762c755c926b7eee60eb0ec3f 2025-11-03T16:58:01.4957662Z * [new tag] trunk/a663eb9c80e92190832b3dd616f138f938f51525 -> trunk/a663eb9c80e92190832b3dd616f138f938f51525 2025-11-03T16:58:01.4959103Z * [new tag] trunk/a6b1ef17173f56ba93ac97ff4384fa4060b5e41e -> trunk/a6b1ef17173f56ba93ac97ff4384fa4060b5e41e 2025-11-03T16:58:01.4960682Z * [new tag] trunk/a76b59cc45a17b728d9da6f937c8a9ab537c657f -> trunk/a76b59cc45a17b728d9da6f937c8a9ab537c657f 2025-11-03T16:58:01.4962182Z * [new tag] trunk/a77f5d9a005a789804fc4d5064036490db5f48fb -> trunk/a77f5d9a005a789804fc4d5064036490db5f48fb 2025-11-03T16:58:01.4963590Z * [new tag] trunk/a7fd0b40018da3c692f74257fc9b1d8cc5c410ff -> trunk/a7fd0b40018da3c692f74257fc9b1d8cc5c410ff 2025-11-03T16:58:01.4965044Z * [new tag] trunk/a988510c3396a668e92801c61b93f7b06219ad0e -> trunk/a988510c3396a668e92801c61b93f7b06219ad0e 2025-11-03T16:58:01.4966515Z * [new tag] trunk/a9b29caeae77480abaee14c89cff19f2c23921a1 -> trunk/a9b29caeae77480abaee14c89cff19f2c23921a1 2025-11-03T16:58:01.4967915Z * [new tag] trunk/aa9c96af041b26c9c55adac490f3449b98f27d06 -> trunk/aa9c96af041b26c9c55adac490f3449b98f27d06 2025-11-03T16:58:01.4969598Z * [new tag] trunk/aab27b051a5e24e71931ef0386f9084d51347211 -> trunk/aab27b051a5e24e71931ef0386f9084d51347211 2025-11-03T16:58:01.4970849Z * [new tag] trunk/ac841267a197e102e1800786b3f6bfcd068b3d85 -> trunk/ac841267a197e102e1800786b3f6bfcd068b3d85 2025-11-03T16:58:01.4972284Z * [new tag] trunk/acd936cc1a53b3c97f1265b6e40e2c09ba0a182d -> trunk/acd936cc1a53b3c97f1265b6e40e2c09ba0a182d 2025-11-03T16:58:01.4973704Z * [new tag] trunk/ad02bd13dfa017f69def846b265a566c4ec5cb3f -> trunk/ad02bd13dfa017f69def846b265a566c4ec5cb3f 2025-11-03T16:58:01.4975276Z * [new tag] trunk/ad3a56ab9857905cdc8f62fb635d578982797c73 -> trunk/ad3a56ab9857905cdc8f62fb635d578982797c73 2025-11-03T16:58:01.4976795Z * [new tag] trunk/ad559072db0cc68aad277380d20bacfdf0d944ac -> trunk/ad559072db0cc68aad277380d20bacfdf0d944ac 2025-11-03T16:58:01.4978081Z * [new tag] trunk/add37bacda45d7de852dedee652346f2857df6fc -> trunk/add37bacda45d7de852dedee652346f2857df6fc 2025-11-03T16:58:01.4979519Z * [new tag] trunk/adedf26e21d1ce835750c99464b3ffa0f6b38fd0 -> trunk/adedf26e21d1ce835750c99464b3ffa0f6b38fd0 2025-11-03T16:58:01.4980946Z * [new tag] trunk/ae038f871b3cb91c36df3a9d4b0ffb4f35d20482 -> trunk/ae038f871b3cb91c36df3a9d4b0ffb4f35d20482 2025-11-03T16:58:01.4982414Z * [new tag] trunk/afaaaa314cc9358a10e9b1986642d49c00773560 -> trunk/afaaaa314cc9358a10e9b1986642d49c00773560 2025-11-03T16:58:01.4984177Z * [new tag] trunk/b060e5c131e9cf43ba6e9b5bda2cb5789c61de47 -> trunk/b060e5c131e9cf43ba6e9b5bda2cb5789c61de47 2025-11-03T16:58:01.4985622Z * [new tag] trunk/b0831930edaa732dd73dc81223fc0a9f5fe6b956 -> trunk/b0831930edaa732dd73dc81223fc0a9f5fe6b956 2025-11-03T16:58:01.4987101Z * [new tag] trunk/b09fb481e0fb00dc475f218e9b14660161b40a94 -> trunk/b09fb481e0fb00dc475f218e9b14660161b40a94 2025-11-03T16:58:01.4988496Z * [new tag] trunk/b2a0f90501dd3a16a6ccaf4c49e1c10f6df4ce1d -> trunk/b2a0f90501dd3a16a6ccaf4c49e1c10f6df4ce1d 2025-11-03T16:58:01.4989878Z * [new tag] trunk/b33762bd2f5f1b115f049d83bf553b08d1ffc97f -> trunk/b33762bd2f5f1b115f049d83bf553b08d1ffc97f 2025-11-03T16:58:01.4991358Z * [new tag] trunk/b3861ac8e70a2334278c17f834880e2d45bbe7a9 -> trunk/b3861ac8e70a2334278c17f834880e2d45bbe7a9 2025-11-03T16:58:01.4992749Z * [new tag] trunk/b4403bfc62ca97eec554cdf815baab1fe93057d9 -> trunk/b4403bfc62ca97eec554cdf815baab1fe93057d9 2025-11-03T16:58:01.4994198Z * [new tag] trunk/b44423bbb43860c1e340cbebc9d101dc18031ecb -> trunk/b44423bbb43860c1e340cbebc9d101dc18031ecb 2025-11-03T16:58:01.4995556Z * [new tag] trunk/b470e59c382649557c49d3c31b8dec1b036eb329 -> trunk/b470e59c382649557c49d3c31b8dec1b036eb329 2025-11-03T16:58:01.4997017Z * [new tag] trunk/b5189e269e899a6ad9ddc4e2ce6f33029290af82 -> trunk/b5189e269e899a6ad9ddc4e2ce6f33029290af82 2025-11-03T16:58:01.4998519Z * [new tag] trunk/b71966f67b2b688f9d71f1602bd3c52d413dacef -> trunk/b71966f67b2b688f9d71f1602bd3c52d413dacef 2025-11-03T16:58:01.4999792Z * [new tag] trunk/b7d348a9074af70f34848298a4e7a7985b197742 -> trunk/b7d348a9074af70f34848298a4e7a7985b197742 2025-11-03T16:58:01.5001173Z * [new tag] trunk/b903018c26217edca67c9023b104027aea377b83 -> trunk/b903018c26217edca67c9023b104027aea377b83 2025-11-03T16:58:01.5002647Z * [new tag] trunk/b939de26d115d33d3cbc17dcf49dbdc1a84e9727 -> trunk/b939de26d115d33d3cbc17dcf49dbdc1a84e9727 2025-11-03T16:58:01.5004175Z * [new tag] trunk/b9bcb37f40e72dc4c2fc1d9281175409fe474194 -> trunk/b9bcb37f40e72dc4c2fc1d9281175409fe474194 2025-11-03T16:58:01.5005577Z * [new tag] trunk/ba71e9ca9ac21e6570b20353d0a67089ad32e442 -> trunk/ba71e9ca9ac21e6570b20353d0a67089ad32e442 2025-11-03T16:58:01.5007006Z * [new tag] trunk/bb54296258854897058e8f158afa8efedc4fcf18 -> trunk/bb54296258854897058e8f158afa8efedc4fcf18 2025-11-03T16:58:01.5008533Z * [new tag] trunk/bbb7d2270b6f647b97f8a9c1d4e69ddaf48ef661 -> trunk/bbb7d2270b6f647b97f8a9c1d4e69ddaf48ef661 2025-11-03T16:58:01.5009959Z * [new tag] trunk/bc03d7c97462abfd45fc1b7b73d5c7acbbd2c185 -> trunk/bc03d7c97462abfd45fc1b7b73d5c7acbbd2c185 2025-11-03T16:58:01.5012721Z * [new tag] trunk/bc5111cd8dc5167cbcbcb2c0f23e53e61a6e8ea5 -> trunk/bc5111cd8dc5167cbcbcb2c0f23e53e61a6e8ea5 2025-11-03T16:58:01.5013991Z * [new tag] trunk/be283297100ab86123e74b7a8372995d32b140c8 -> trunk/be283297100ab86123e74b7a8372995d32b140c8 2025-11-03T16:58:01.5015518Z * [new tag] trunk/bea89d60603ca33f1903e57db5c734a67535c26d -> trunk/bea89d60603ca33f1903e57db5c734a67535c26d 2025-11-03T16:58:01.5016958Z * [new tag] trunk/bebabd7fce29ea49b9269aeaa9fe3f34a3e1127e -> trunk/bebabd7fce29ea49b9269aeaa9fe3f34a3e1127e 2025-11-03T16:58:01.5018241Z * [new tag] trunk/bfb47ec50e09dc2921ba47fedf84a1e28ee991a7 -> trunk/bfb47ec50e09dc2921ba47fedf84a1e28ee991a7 2025-11-03T16:58:01.5019384Z * [new tag] trunk/bfc2050db975e589795cd3eceaed2e83bf89ad35 -> trunk/bfc2050db975e589795cd3eceaed2e83bf89ad35 2025-11-03T16:58:01.5020820Z * [new tag] trunk/c01636e1bc7338eecbcfd2c17fd3fa2cad836983 -> trunk/c01636e1bc7338eecbcfd2c17fd3fa2cad836983 2025-11-03T16:58:01.5022255Z * [new tag] trunk/c0bbda37e8d74734b64421762eb365e620f258c1 -> trunk/c0bbda37e8d74734b64421762eb365e620f258c1 2025-11-03T16:58:01.5023673Z * [new tag] trunk/c201a1cab1439bfcfaff7e4fdab0d6eeeb880a70 -> trunk/c201a1cab1439bfcfaff7e4fdab0d6eeeb880a70 2025-11-03T16:58:01.5025131Z * [new tag] trunk/c2e3cc7aedb2e7d89443225c7cccd08a0f8a3587 -> trunk/c2e3cc7aedb2e7d89443225c7cccd08a0f8a3587 2025-11-03T16:58:01.5026534Z * [new tag] trunk/c3047938a01c844c752e5462b05d180a04d438fd -> trunk/c3047938a01c844c752e5462b05d180a04d438fd 2025-11-03T16:58:01.5027935Z * [new tag] trunk/c37802a8c4388160faa65f8bc9c29d98b4d2fb3c -> trunk/c37802a8c4388160faa65f8bc9c29d98b4d2fb3c 2025-11-03T16:58:01.5029379Z * [new tag] trunk/c3b71d54995a16fd09c18a5da5d598a23388bbd6 -> trunk/c3b71d54995a16fd09c18a5da5d598a23388bbd6 2025-11-03T16:58:01.5030657Z * [new tag] trunk/c3d205d5986824e72df353c9bf112dc9b2a3612a -> trunk/c3d205d5986824e72df353c9bf112dc9b2a3612a 2025-11-03T16:58:01.5031957Z * [new tag] trunk/c3dc0c7089f3204c36961acc227a3ff503530094 -> trunk/c3dc0c7089f3204c36961acc227a3ff503530094 2025-11-03T16:58:01.5033350Z * [new tag] trunk/c54e2c5b4138cb85412a74189d1f7f771ad1116e -> trunk/c54e2c5b4138cb85412a74189d1f7f771ad1116e 2025-11-03T16:58:01.5034606Z * [new tag] trunk/c5701d0ab5f55b23243dd437f57e856970c45bb6 -> trunk/c5701d0ab5f55b23243dd437f57e856970c45bb6 2025-11-03T16:58:01.5036225Z * [new tag] trunk/c58d0ad85dfbf026d5f21ee739c139407dc20e54 -> trunk/c58d0ad85dfbf026d5f21ee739c139407dc20e54 2025-11-03T16:58:01.5037630Z * [new tag] trunk/c594950e867e420a6086190a88e4b8fbd687af91 -> trunk/c594950e867e420a6086190a88e4b8fbd687af91 2025-11-03T16:58:01.5039072Z * [new tag] trunk/c6a02eae5b77758ee7ee0b1dfbcf20a8341914a7 -> trunk/c6a02eae5b77758ee7ee0b1dfbcf20a8341914a7 2025-11-03T16:58:01.5040350Z * [new tag] trunk/c8adc08b3b19b838dcdc71f2055600f420e6a91b -> trunk/c8adc08b3b19b838dcdc71f2055600f420e6a91b 2025-11-03T16:58:01.5041892Z * [new tag] trunk/c9eabadc5e964678556a6bfcde455a5e11dc0785 -> trunk/c9eabadc5e964678556a6bfcde455a5e11dc0785 2025-11-03T16:58:01.5043311Z * [new tag] trunk/cb6966704cf70ebac6d2b36b65e9a75e11c29e71 -> trunk/cb6966704cf70ebac6d2b36b65e9a75e11c29e71 2025-11-03T16:58:01.5044607Z * [new tag] trunk/cde81e92b95eee9af2879c9c75f7b03699ca72ad -> trunk/cde81e92b95eee9af2879c9c75f7b03699ca72ad 2025-11-03T16:58:01.5046024Z * [new tag] trunk/cee03634dab2dbf259d84bc389bdea80de0f80ac -> trunk/cee03634dab2dbf259d84bc389bdea80de0f80ac 2025-11-03T16:58:01.5047410Z * [new tag] trunk/cf7756da385e03c169f7e6a5b5ff2c0e29b2814a -> trunk/cf7756da385e03c169f7e6a5b5ff2c0e29b2814a 2025-11-03T16:58:01.5048839Z * [new tag] trunk/cf9a834f39d6c3b769381cc1321bcb5c54132415 -> trunk/cf9a834f39d6c3b769381cc1321bcb5c54132415 2025-11-03T16:58:01.5050496Z * [new tag] trunk/d049ed2cb1619c44279cb716b8a1d94e4df3b372 -> trunk/d049ed2cb1619c44279cb716b8a1d94e4df3b372 2025-11-03T16:58:01.5051776Z * [new tag] trunk/d17790072381f0e76e177ae4093504a6908a5a4d -> trunk/d17790072381f0e76e177ae4093504a6908a5a4d 2025-11-03T16:58:01.5053183Z * [new tag] trunk/d1a6e006e090b1365e6653853982f175cca452e8 -> trunk/d1a6e006e090b1365e6653853982f175cca452e8 2025-11-03T16:58:01.5054444Z * [new tag] trunk/d2be06f673ff11e59b234156dccaad2dc22c7746 -> trunk/d2be06f673ff11e59b234156dccaad2dc22c7746 2025-11-03T16:58:01.5056388Z * [new tag] trunk/d2eff5d454ab2cb0a5ccdfb5eb6e7d6dcc75e097 -> trunk/d2eff5d454ab2cb0a5ccdfb5eb6e7d6dcc75e097 2025-11-03T16:58:01.5058006Z * [new tag] trunk/d3be06cbdca4b2a9cd8ae35926b3f20b1bed1fe2 -> trunk/d3be06cbdca4b2a9cd8ae35926b3f20b1bed1fe2 2025-11-03T16:58:01.5059951Z * [new tag] trunk/d3e511f07c0b1fc6df22347aec7f8e700bdb0d24 -> trunk/d3e511f07c0b1fc6df22347aec7f8e700bdb0d24 2025-11-03T16:58:01.5061312Z * [new tag] trunk/d401e4e70a1dd19733d3e3c84affc4884605efb4 -> trunk/d401e4e70a1dd19733d3e3c84affc4884605efb4 2025-11-03T16:58:01.5062639Z * [new tag] trunk/d46d8d6f54b15ded4f2483c7bde31be124281ab8 -> trunk/d46d8d6f54b15ded4f2483c7bde31be124281ab8 2025-11-03T16:58:01.5063910Z * [new tag] trunk/d6d6fa26f540c10c57ac80547a9475e9f4c201f2 -> trunk/d6d6fa26f540c10c57ac80547a9475e9f4c201f2 2025-11-03T16:58:01.5065354Z * [new tag] trunk/d7040e6d7515cea485824d2b810bea94e5958dea -> trunk/d7040e6d7515cea485824d2b810bea94e5958dea 2025-11-03T16:58:01.5066795Z * [new tag] trunk/d80ae738c9c1810b30a88233bdcd34cae16c5c6e -> trunk/d80ae738c9c1810b30a88233bdcd34cae16c5c6e 2025-11-03T16:58:01.5068212Z * [new tag] trunk/d8283a317a335c628c4541c6a60e102fdead57b7 -> trunk/d8283a317a335c628c4541c6a60e102fdead57b7 2025-11-03T16:58:01.5069597Z * [new tag] trunk/d9483d4c8d3f220ad505afd05606e068184896a4 -> trunk/d9483d4c8d3f220ad505afd05606e068184896a4 2025-11-03T16:58:01.5070917Z * [new tag] trunk/d962bed15798961ffd10e64068f86a2cec411ffb -> trunk/d962bed15798961ffd10e64068f86a2cec411ffb 2025-11-03T16:58:01.5072352Z * [new tag] trunk/d97144d31ef42e762ce65f467cbaf065edb241cc -> trunk/d97144d31ef42e762ce65f467cbaf065edb241cc 2025-11-03T16:58:01.5073861Z * [new tag] trunk/dc011d3203b24ebf1f78ab8dd64766205e64d8ec -> trunk/dc011d3203b24ebf1f78ab8dd64766205e64d8ec 2025-11-03T16:58:01.5075262Z * [new tag] trunk/dd1fe7c22f8eee3984a22663636b0d1403b079ba -> trunk/dd1fe7c22f8eee3984a22663636b0d1403b079ba 2025-11-03T16:58:01.5077106Z * [new tag] trunk/deb776319b12cad7fffb5cf9d8851a50e2b0e9ea -> trunk/deb776319b12cad7fffb5cf9d8851a50e2b0e9ea 2025-11-03T16:58:01.5078484Z * [new tag] trunk/defac66e39c60f867795346ded171366ed682ee3 -> trunk/defac66e39c60f867795346ded171366ed682ee3 2025-11-03T16:58:01.5079933Z * [new tag] trunk/df71b7072799c451a008cb36142dfdb1487f0d5e -> trunk/df71b7072799c451a008cb36142dfdb1487f0d5e 2025-11-03T16:58:01.5081238Z * [new tag] trunk/dfebdcab86acbaa0eaa996b47595e5f27a66492e -> trunk/dfebdcab86acbaa0eaa996b47595e5f27a66492e 2025-11-03T16:58:01.5082485Z * [new tag] trunk/e0604d3170f1123ebacbced66be400c7f162621f -> trunk/e0604d3170f1123ebacbced66be400c7f162621f 2025-11-03T16:58:01.5083852Z * [new tag] trunk/e0791fc11dc0024a828495985898b29120dcc4c1 -> trunk/e0791fc11dc0024a828495985898b29120dcc4c1 2025-11-03T16:58:01.5085369Z * [new tag] trunk/e0ca3049c0c7726747094f47f80cadcfad51320d -> trunk/e0ca3049c0c7726747094f47f80cadcfad51320d 2025-11-03T16:58:01.5086754Z * [new tag] trunk/e105a475756ae6dd45c8134bd4583afc57ee416e -> trunk/e105a475756ae6dd45c8134bd4583afc57ee416e 2025-11-03T16:58:01.5088128Z * [new tag] trunk/e137cd0a1046ead2ca9d631f427655adced606c2 -> trunk/e137cd0a1046ead2ca9d631f427655adced606c2 2025-11-03T16:58:01.5089676Z * [new tag] trunk/e1d011d6eb571cd98ec7c7ed8e8b518a5463ec97 -> trunk/e1d011d6eb571cd98ec7c7ed8e8b518a5463ec97 2025-11-03T16:58:01.5090939Z * [new tag] trunk/e214af6ae8e613485ac3dac3e2094cf53dfc3c1d -> trunk/e214af6ae8e613485ac3dac3e2094cf53dfc3c1d 2025-11-03T16:58:01.5092377Z * [new tag] trunk/e2dc32f4baf853e6171777aed530fc83bcdb35ae -> trunk/e2dc32f4baf853e6171777aed530fc83bcdb35ae 2025-11-03T16:58:01.5093799Z * [new tag] trunk/e380028a51c38de80b8888d027c63e290a773651 -> trunk/e380028a51c38de80b8888d027c63e290a773651 2025-11-03T16:58:01.5095250Z * [new tag] trunk/e3ae0594d16134632ff587c9ab400d4148c83e9f -> trunk/e3ae0594d16134632ff587c9ab400d4148c83e9f 2025-11-03T16:58:01.5096589Z * [new tag] trunk/e3e93c7107830c13f4139c3a62fda62c6b84bbf5 -> trunk/e3e93c7107830c13f4139c3a62fda62c6b84bbf5 2025-11-03T16:58:01.5097970Z * [new tag] trunk/e4043884c734a3947f4c8da0cfbba43fc6319f3d -> trunk/e4043884c734a3947f4c8da0cfbba43fc6319f3d 2025-11-03T16:58:01.5099258Z * [new tag] trunk/e471800dceb8a6592562fc4fc96a5e80bc494c0a -> trunk/e471800dceb8a6592562fc4fc96a5e80bc494c0a 2025-11-03T16:58:01.5100656Z * [new tag] trunk/e83be7042e809ef51c67f53d83ac45794447c73f -> trunk/e83be7042e809ef51c67f53d83ac45794447c73f 2025-11-03T16:58:01.5102080Z * [new tag] trunk/e8d887ae3fceb97283bc2d751621b9615524b4b7 -> trunk/e8d887ae3fceb97283bc2d751621b9615524b4b7 2025-11-03T16:58:01.5103640Z * [new tag] trunk/e8fadba28cc225782ddce1bf0ef8315ee1743761 -> trunk/e8fadba28cc225782ddce1bf0ef8315ee1743761 2025-11-03T16:58:01.5105036Z * [new tag] trunk/e95920e3e6f9d8218c6aed2b8b6ce9b838a49104 -> trunk/e95920e3e6f9d8218c6aed2b8b6ce9b838a49104 2025-11-03T16:58:01.5106412Z * [new tag] trunk/ea698e8bfcd2ec1820b3ba90b9c47416a33348de -> trunk/ea698e8bfcd2ec1820b3ba90b9c47416a33348de 2025-11-03T16:58:01.5107921Z * [new tag] trunk/eae701cad03feb87908107cb5246d9e13a9d426e -> trunk/eae701cad03feb87908107cb5246d9e13a9d426e 2025-11-03T16:58:01.5109339Z * [new tag] trunk/eb2bad5bb54df8a2563f2619819e9e689e0ba10e -> trunk/eb2bad5bb54df8a2563f2619819e9e689e0ba10e 2025-11-03T16:58:01.5111066Z * [new tag] trunk/ebb2b2e894a4ede8efc5f5fff068d4ac9972f77a -> trunk/ebb2b2e894a4ede8efc5f5fff068d4ac9972f77a 2025-11-03T16:58:01.5112465Z * [new tag] trunk/ee7434be822cf6e75b4566d8159f550ee233d8ae -> trunk/ee7434be822cf6e75b4566d8159f550ee233d8ae 2025-11-03T16:58:01.5113882Z * [new tag] trunk/ef8d97efcf05e1f3a05ad5ce3444e3af42481505 -> trunk/ef8d97efcf05e1f3a05ad5ce3444e3af42481505 2025-11-03T16:58:01.5115158Z * [new tag] trunk/f013e804c876a34cbc86a6a475597e3e29812883 -> trunk/f013e804c876a34cbc86a6a475597e3e29812883 2025-11-03T16:58:01.5116596Z * [new tag] trunk/f02708c2be66c7cf0edec52663de69f66ba4b1df -> trunk/f02708c2be66c7cf0edec52663de69f66ba4b1df 2025-11-03T16:58:01.5118032Z * [new tag] trunk/f0745ddb11ab66a8bb4de8031bd7e9dccba19379 -> trunk/f0745ddb11ab66a8bb4de8031bd7e9dccba19379 2025-11-03T16:58:01.5119487Z * [new tag] trunk/f167fd09fabf3f5be63cd31451f29d49796c14e8 -> trunk/f167fd09fabf3f5be63cd31451f29d49796c14e8 2025-11-03T16:58:01.5120865Z * [new tag] trunk/f1a34407152302812bf633360cd29e9fba8de71e -> trunk/f1a34407152302812bf633360cd29e9fba8de71e 2025-11-03T16:58:01.5122421Z * [new tag] trunk/f1af679270392c83e03808c8af5e2cbe3cdf16ce -> trunk/f1af679270392c83e03808c8af5e2cbe3cdf16ce 2025-11-03T16:58:01.5123806Z * [new tag] trunk/f1e4c42b6ef3d3cea08ab3babb693e3ce42cf08b -> trunk/f1e4c42b6ef3d3cea08ab3babb693e3ce42cf08b 2025-11-03T16:58:01.5125159Z * [new tag] trunk/f20bf778746a7662a9bdbd2fa49d326772ca90ed -> trunk/f20bf778746a7662a9bdbd2fa49d326772ca90ed 2025-11-03T16:58:01.5126558Z * [new tag] trunk/f2450798cd6f30b811f6d8256da0c271cad2285c -> trunk/f2450798cd6f30b811f6d8256da0c271cad2285c 2025-11-03T16:58:01.5127954Z * [new tag] trunk/f2c81635c85bf8d1c523d6b17f8016ad78ea02e2 -> trunk/f2c81635c85bf8d1c523d6b17f8016ad78ea02e2 2025-11-03T16:58:01.5129534Z * [new tag] trunk/f33abae6957b8ef56268ce22e660d3ed3111c7c3 -> trunk/f33abae6957b8ef56268ce22e660d3ed3111c7c3 2025-11-03T16:58:01.5130954Z * [new tag] trunk/f36f372acc28062e0988d84699c62689b0d89a6e -> trunk/f36f372acc28062e0988d84699c62689b0d89a6e 2025-11-03T16:58:01.5132452Z * [new tag] trunk/f452edd782fb7342e2a5740cea9b78135e714858 -> trunk/f452edd782fb7342e2a5740cea9b78135e714858 2025-11-03T16:58:01.5133907Z * [new tag] trunk/f4d05feb7a5f9191509e1e539ea210fcbe538b32 -> trunk/f4d05feb7a5f9191509e1e539ea210fcbe538b32 2025-11-03T16:58:01.5135300Z * [new tag] trunk/f5543e3741bbfba4579afc3b49d0720f2eee7e4d -> trunk/f5543e3741bbfba4579afc3b49d0720f2eee7e4d 2025-11-03T16:58:01.5136556Z * [new tag] trunk/f5cb9a4c68d9271c58ef4d3257210984b8e85099 -> trunk/f5cb9a4c68d9271c58ef4d3257210984b8e85099 2025-11-03T16:58:01.5138011Z * [new tag] trunk/f60751024ecac57840a2d25132d0c1562351e66e -> trunk/f60751024ecac57840a2d25132d0c1562351e66e 2025-11-03T16:58:01.5139422Z * [new tag] trunk/f6951cb8ea1ef96a33627734eb68b5dc7a04b6ed -> trunk/f6951cb8ea1ef96a33627734eb68b5dc7a04b6ed 2025-11-03T16:58:01.5140935Z * [new tag] trunk/f89a7e9fe80fb0f74af52bad892e1afafa57be9c -> trunk/f89a7e9fe80fb0f74af52bad892e1afafa57be9c 2025-11-03T16:58:01.5142314Z * [new tag] trunk/f8b4c002940d96ba2afc14ac4cf0b5277776bc02 -> trunk/f8b4c002940d96ba2afc14ac4cf0b5277776bc02 2025-11-03T16:58:01.5143692Z * [new tag] trunk/f911d64750efb22097eef5505552a7669a0c53de -> trunk/f911d64750efb22097eef5505552a7669a0c53de 2025-11-03T16:58:01.5145099Z * [new tag] trunk/f91899ca6c15d611c453591527e1be8b43e5d4ea -> trunk/f91899ca6c15d611c453591527e1be8b43e5d4ea 2025-11-03T16:58:01.5146549Z * [new tag] trunk/f93ea7dab190fa0953b7a4921b76993cadafc7c9 -> trunk/f93ea7dab190fa0953b7a4921b76993cadafc7c9 2025-11-03T16:58:01.5148011Z * [new tag] trunk/fa4cb918462d3cbe7cf48db7c1498824c1ca1cc4 -> trunk/fa4cb918462d3cbe7cf48db7c1498824c1ca1cc4 2025-11-03T16:58:01.5149379Z * [new tag] trunk/fa560e1158b496d1c37ca2a2443a0b7f5a08dad5 -> trunk/fa560e1158b496d1c37ca2a2443a0b7f5a08dad5 2025-11-03T16:58:01.5150811Z * [new tag] trunk/fa8e073a4e17f12e7e4bb0b26471179d2ca57471 -> trunk/fa8e073a4e17f12e7e4bb0b26471179d2ca57471 2025-11-03T16:58:01.5152414Z * [new tag] trunk/fb545fb0686c94cc017c15ca7e342996b49edb81 -> trunk/fb545fb0686c94cc017c15ca7e342996b49edb81 2025-11-03T16:58:01.5153900Z * [new tag] trunk/fc540cefd498f1001a5fc7a4f187080628b27839 -> trunk/fc540cefd498f1001a5fc7a4f187080628b27839 2025-11-03T16:58:01.5155301Z * [new tag] trunk/fc8ac1216c61d8736f45a58825abf58e7d51e088 -> trunk/fc8ac1216c61d8736f45a58825abf58e7d51e088 2025-11-03T16:58:01.5156795Z * [new tag] trunk/fcc10635660ceb619f78d06eaaa8b3c0bdd52ab8 -> trunk/fcc10635660ceb619f78d06eaaa8b3c0bdd52ab8 2025-11-03T16:58:01.5158175Z * [new tag] trunk/fcd5f8c352b5b75bd32e57fa044ec5df095032da -> trunk/fcd5f8c352b5b75bd32e57fa044ec5df095032da 2025-11-03T16:58:01.5159682Z * [new tag] trunk/fd5da81fdd76c0a78ef935d077a93b48b78fca3a -> trunk/fd5da81fdd76c0a78ef935d077a93b48b78fca3a 2025-11-03T16:58:01.5161209Z * [new tag] trunk/fd68d409ada709450ced3030bde89ec662a3f7b7 -> trunk/fd68d409ada709450ced3030bde89ec662a3f7b7 2025-11-03T16:58:01.5162457Z * [new tag] trunk/fea819ed080175f9a93205cc9074022bfa486325 -> trunk/fea819ed080175f9a93205cc9074022bfa486325 2025-11-03T16:58:01.5163825Z * [new tag] trunk/fee1ac927d66fce67f862311292dcc325b8def44 -> trunk/fee1ac927d66fce67f862311292dcc325b8def44 2025-11-03T16:58:01.5165276Z * [new tag] trunk/fee7624bd66e64f575c1396fcaaa54cabce49ca4 -> trunk/fee7624bd66e64f575c1396fcaaa54cabce49ca4 2025-11-03T16:58:01.5166677Z * [new tag] trunk/fefb546b9100614da1fa703e12651bc115059f02 -> trunk/fefb546b9100614da1fa703e12651bc115059f02 2025-11-03T16:58:01.5168700Z * [new tag] trunk/ff46d5a79bec09d9ae334d325f5f8cc89306b536 -> trunk/ff46d5a79bec09d9ae334d325f5f8cc89306b536 2025-11-03T16:58:01.5170182Z * [new tag] trunk/ffaa6578b7a641b1ac1fcf451f04cf5e1212c9fe -> trunk/ffaa6578b7a641b1ac1fcf451f04cf5e1212c9fe 2025-11-03T16:58:01.5171368Z * [new tag] v0.1.1 -> v0.1.1 2025-11-03T16:58:01.5172683Z * [new tag] v0.1.10 -> v0.1.10 2025-11-03T16:58:01.5173940Z * [new tag] v0.1.11 -> v0.1.11 2025-11-03T16:58:01.5175256Z * [new tag] v0.1.12 -> v0.1.12 2025-11-03T16:58:01.5176499Z * [new tag] v0.1.2 -> v0.1.2 2025-11-03T16:58:01.5177763Z * [new tag] v0.1.3 -> v0.1.3 2025-11-03T16:58:01.5179031Z * [new tag] v0.1.4 -> v0.1.4 2025-11-03T16:58:01.5180334Z * [new tag] v0.1.5 -> v0.1.5 2025-11-03T16:58:01.5181651Z * [new tag] v0.1.6 -> v0.1.6 2025-11-03T16:58:01.5182891Z * [new tag] v0.1.7 -> v0.1.7 2025-11-03T16:58:01.5184137Z * [new tag] v0.1.8 -> v0.1.8 2025-11-03T16:58:01.5185384Z * [new tag] v0.1.9 -> v0.1.9 2025-11-03T16:58:01.5186765Z * [new tag] v0.2.0 -> v0.2.0 2025-11-03T16:58:01.5188087Z * [new tag] v0.3.0 -> v0.3.0 2025-11-03T16:58:01.5189532Z * [new tag] v0.3.1 -> v0.3.1 2025-11-03T16:58:01.5190822Z * [new tag] v0.4.0 -> v0.4.0 2025-11-03T16:58:01.5192108Z * [new tag] v0.4.1 -> v0.4.1 2025-11-03T16:58:01.5193504Z * [new tag] v1.0.0 -> v1.0.0 2025-11-03T16:58:01.5194864Z * [new tag] v1.0.0a0 -> v1.0.0a0 2025-11-03T16:58:01.5196152Z * [new tag] v1.0.1 -> v1.0.1 2025-11-03T16:58:01.5197503Z * [new tag] v1.0rc0 -> v1.0rc0 2025-11-03T16:58:01.5198628Z * [new tag] v1.0rc1 -> v1.0rc1 2025-11-03T16:58:01.5199907Z * [new tag] v1.1.0 -> v1.1.0 2025-11-03T16:58:01.5201312Z * [new tag] v1.1.0a0 -> v1.1.0a0 2025-11-03T16:58:01.5202765Z * [new tag] v1.10.0 -> v1.10.0 2025-11-03T16:58:01.5204154Z * [new tag] v1.10.0-rc1 -> v1.10.0-rc1 2025-11-03T16:58:01.5205569Z * [new tag] v1.10.0-rc2 -> v1.10.0-rc2 2025-11-03T16:58:01.5206678Z * [new tag] v1.10.0-rc3 -> v1.10.0-rc3 2025-11-03T16:58:01.5207981Z * [new tag] v1.10.1 -> v1.10.1 2025-11-03T16:58:01.5209191Z * [new tag] v1.10.1-rc1 -> v1.10.1-rc1 2025-11-03T16:58:01.5210603Z * [new tag] v1.10.2 -> v1.10.2 2025-11-03T16:58:01.5211888Z * [new tag] v1.10.2-rc1 -> v1.10.2-rc1 2025-11-03T16:58:01.5213177Z * [new tag] v1.11.0 -> v1.11.0 2025-11-03T16:58:01.5214532Z * [new tag] v1.11.0-rc1 -> v1.11.0-rc1 2025-11-03T16:58:01.5215871Z * [new tag] v1.11.0-rc2 -> v1.11.0-rc2 2025-11-03T16:58:01.5217228Z * [new tag] v1.11.0-rc3 -> v1.11.0-rc3 2025-11-03T16:58:01.5218603Z * [new tag] v1.11.0-rc4 -> v1.11.0-rc4 2025-11-03T16:58:01.5219953Z * [new tag] v1.11.0-rc5 -> v1.11.0-rc5 2025-11-03T16:58:01.5221080Z * [new tag] v1.11.0-rc6 -> v1.11.0-rc6 2025-11-03T16:58:01.5222201Z * [new tag] v1.11.0-rc7 -> v1.11.0-rc7 2025-11-03T16:58:01.5223482Z * [new tag] v1.12.0 -> v1.12.0 2025-11-03T16:58:01.5224764Z * [new tag] v1.12.0-rc1 -> v1.12.0-rc1 2025-11-03T16:58:01.5226204Z * [new tag] v1.12.0-rc2 -> v1.12.0-rc2 2025-11-03T16:58:01.5227568Z * [new tag] v1.12.0-rc3 -> v1.12.0-rc3 2025-11-03T16:58:01.5228957Z * [new tag] v1.12.0-rc4 -> v1.12.0-rc4 2025-11-03T16:58:01.5230212Z * [new tag] v1.12.0-rc5 -> v1.12.0-rc5 2025-11-03T16:58:01.5231656Z * [new tag] v1.12.0-rc6 -> v1.12.0-rc6 2025-11-03T16:58:01.5232769Z * [new tag] v1.12.0-rc7 -> v1.12.0-rc7 2025-11-03T16:58:01.5233898Z * [new tag] v1.12.0-rc8 -> v1.12.0-rc8 2025-11-03T16:58:01.5235080Z * [new tag] v1.12.1 -> v1.12.1 2025-11-03T16:58:01.5236529Z * [new tag] v1.12.1-rc1 -> v1.12.1-rc1 2025-11-03T16:58:01.5237813Z * [new tag] v1.12.1-rc2 -> v1.12.1-rc2 2025-11-03T16:58:01.5239180Z * [new tag] v1.12.1-rc3 -> v1.12.1-rc3 2025-11-03T16:58:01.5240864Z * [new tag] v1.12.1-rc4 -> v1.12.1-rc4 2025-11-03T16:58:01.5242012Z * [new tag] v1.12.1-rc5 -> v1.12.1-rc5 2025-11-03T16:58:01.5243557Z * [new tag] v1.13.0 -> v1.13.0 2025-11-03T16:58:01.5244950Z * [new tag] v1.13.0-rc1 -> v1.13.0-rc1 2025-11-03T16:58:01.5246317Z * [new tag] v1.13.0-rc2 -> v1.13.0-rc2 2025-11-03T16:58:01.5247540Z * [new tag] v1.13.0-rc3 -> v1.13.0-rc3 2025-11-03T16:58:01.5249075Z * [new tag] v1.13.0-rc4 -> v1.13.0-rc4 2025-11-03T16:58:01.5250225Z * [new tag] v1.13.0-rc5 -> v1.13.0-rc5 2025-11-03T16:58:01.5251352Z * [new tag] v1.13.0-rc6 -> v1.13.0-rc6 2025-11-03T16:58:01.5252741Z * [new tag] v1.13.1 -> v1.13.1 2025-11-03T16:58:01.5253865Z * [new tag] v1.13.1-rc1 -> v1.13.1-rc1 2025-11-03T16:58:01.5255161Z * [new tag] v1.2.0 -> v1.2.0 2025-11-03T16:58:01.5256507Z * [new tag] v1.2.0a0 -> v1.2.0a0 2025-11-03T16:58:01.5257854Z * [new tag] v1.3.0 -> v1.3.0 2025-11-03T16:58:01.5259081Z * [new tag] v1.3.0a0 -> v1.3.0a0 2025-11-03T16:58:01.5260217Z * [new tag] v1.3.1 -> v1.3.1 2025-11-03T16:58:01.5261486Z * [new tag] v1.4.0 -> v1.4.0 2025-11-03T16:58:01.5262831Z * [new tag] v1.4.0a0 -> v1.4.0a0 2025-11-03T16:58:01.5263948Z * [new tag] v1.4.1 -> v1.4.1 2025-11-03T16:58:01.5265405Z * [new tag] v1.5.0 -> v1.5.0 2025-11-03T16:58:01.5266774Z * [new tag] v1.5.0-rc1 -> v1.5.0-rc1 2025-11-03T16:58:01.5268094Z * [new tag] v1.5.0-rc2 -> v1.5.0-rc2 2025-11-03T16:58:01.5269516Z * [new tag] v1.5.0-rc3 -> v1.5.0-rc3 2025-11-03T16:58:01.5270817Z * [new tag] v1.5.0-rc4 -> v1.5.0-rc4 2025-11-03T16:58:01.5271935Z * [new tag] v1.5.0-rc5 -> v1.5.0-rc5 2025-11-03T16:58:01.5273361Z * [new tag] v1.5.1 -> v1.5.1 2025-11-03T16:58:01.5274498Z * [new tag] v1.5.1-rc1 -> v1.5.1-rc1 2025-11-03T16:58:01.5275626Z * [new tag] v1.6.0 -> v1.6.0 2025-11-03T16:58:01.5277062Z * [new tag] v1.6.0-rc1 -> v1.6.0-rc1 2025-11-03T16:58:01.5278379Z * [new tag] v1.6.0-rc2 -> v1.6.0-rc2 2025-11-03T16:58:01.5279847Z * [new tag] v1.6.0-rc3 -> v1.6.0-rc3 2025-11-03T16:58:01.5281004Z * [new tag] v1.6.0-rc4 -> v1.6.0-rc4 2025-11-03T16:58:01.5282293Z * [new tag] v1.6.0-rc5 -> v1.6.0-rc5 2025-11-03T16:58:01.5283624Z * [new tag] v1.6.0-rc6 -> v1.6.0-rc6 2025-11-03T16:58:01.5284717Z * [new tag] v1.6.0-rc7 -> v1.6.0-rc7 2025-11-03T16:58:01.5286080Z * [new tag] v1.7.0 -> v1.7.0 2025-11-03T16:58:01.5287374Z * [new tag] v1.7.0-rc1 -> v1.7.0-rc1 2025-11-03T16:58:01.5288980Z * [new tag] v1.7.0-rc2 -> v1.7.0-rc2 2025-11-03T16:58:01.5290299Z * [new tag] v1.7.0-rc3 -> v1.7.0-rc3 2025-11-03T16:58:01.5291400Z * [new tag] v1.7.0-rc4 -> v1.7.0-rc4 2025-11-03T16:58:01.5292729Z * [new tag] v1.7.1 -> v1.7.1 2025-11-03T16:58:01.5294150Z * [new tag] v1.7.1-rc1 -> v1.7.1-rc1 2025-11-03T16:58:01.5295522Z * [new tag] v1.7.1-rc2 -> v1.7.1-rc2 2025-11-03T16:58:01.5296672Z * [new tag] v1.7.1-rc3 -> v1.7.1-rc3 2025-11-03T16:58:01.5298029Z * [new tag] v1.8.0 -> v1.8.0 2025-11-03T16:58:01.5299112Z * [new tag] v1.8.0-rc1 -> v1.8.0-rc1 2025-11-03T16:58:01.5300519Z * [new tag] v1.8.0-rc2 -> v1.8.0-rc2 2025-11-03T16:58:01.5301856Z * [new tag] v1.8.0-rc3 -> v1.8.0-rc3 2025-11-03T16:58:01.5303230Z * [new tag] v1.8.0-rc4 -> v1.8.0-rc4 2025-11-03T16:58:01.5304299Z * [new tag] v1.8.0-rc5 -> v1.8.0-rc5 2025-11-03T16:58:01.5305433Z * [new tag] v1.8.1 -> v1.8.1 2025-11-03T16:58:01.5306779Z * [new tag] v1.8.1-rc1 -> v1.8.1-rc1 2025-11-03T16:58:01.5307940Z * [new tag] v1.8.1-rc2 -> v1.8.1-rc2 2025-11-03T16:58:01.5309043Z * [new tag] v1.8.1-rc3 -> v1.8.1-rc3 2025-11-03T16:58:01.5311036Z * [new tag] v1.8.2 -> v1.8.2 2025-11-03T16:58:01.5312168Z * [new tag] v1.8.2-rc1 -> v1.8.2-rc1 2025-11-03T16:58:01.5313564Z * [new tag] v1.9.0 -> v1.9.0 2025-11-03T16:58:01.5314886Z * [new tag] v1.9.0-rc1 -> v1.9.0-rc1 2025-11-03T16:58:01.5316252Z * [new tag] v1.9.0-rc2 -> v1.9.0-rc2 2025-11-03T16:58:01.5317611Z * [new tag] v1.9.0-rc3 -> v1.9.0-rc3 2025-11-03T16:58:01.5319039Z * [new tag] v1.9.0-rc4 -> v1.9.0-rc4 2025-11-03T16:58:01.5320390Z * [new tag] v1.9.1 -> v1.9.1 2025-11-03T16:58:01.5321743Z * [new tag] v1.9.1-rc1 -> v1.9.1-rc1 2025-11-03T16:58:01.5322841Z * [new tag] v1.9.1-rc2 -> v1.9.1-rc2 2025-11-03T16:58:01.5324663Z * [new tag] v2.0.0 -> v2.0.0 2025-11-03T16:58:01.5325944Z * [new tag] v2.0.0-rc1 -> v2.0.0-rc1 2025-11-03T16:58:01.5327297Z * [new tag] v2.0.0-rc2 -> v2.0.0-rc2 2025-11-03T16:58:01.5328714Z * [new tag] v2.0.0-rc3 -> v2.0.0-rc3 2025-11-03T16:58:01.5330115Z * [new tag] v2.0.0-rc4 -> v2.0.0-rc4 2025-11-03T16:58:01.5331664Z * [new tag] v2.0.0-rc5 -> v2.0.0-rc5 2025-11-03T16:58:01.5332810Z * [new tag] v2.0.0-rc6 -> v2.0.0-rc6 2025-11-03T16:58:01.5334242Z * [new tag] v2.0.1 -> v2.0.1 2025-11-03T16:58:01.5335688Z * [new tag] v2.0.1-rc1 -> v2.0.1-rc1 2025-11-03T16:58:01.5336785Z * [new tag] v2.0.1-rc2 -> v2.0.1-rc2 2025-11-03T16:58:01.5338034Z * [new tag] v2.0.1-rc3 -> v2.0.1-rc3 2025-11-03T16:58:01.5339158Z * [new tag] v2.0.1-rc4 -> v2.0.1-rc4 2025-11-03T16:58:01.5340984Z * [new tag] v2.1.0 -> v2.1.0 2025-11-03T16:58:01.5342229Z * [new tag] v2.1.0-rc1 -> v2.1.0-rc1 2025-11-03T16:58:01.5343523Z * [new tag] v2.1.0-rc2 -> v2.1.0-rc2 2025-11-03T16:58:01.5344923Z * [new tag] v2.1.0-rc3 -> v2.1.0-rc3 2025-11-03T16:58:01.5346345Z * [new tag] v2.1.0-rc4 -> v2.1.0-rc4 2025-11-03T16:58:01.5347670Z * [new tag] v2.1.0-rc5 -> v2.1.0-rc5 2025-11-03T16:58:01.5348840Z * [new tag] v2.1.0-rc6 -> v2.1.0-rc6 2025-11-03T16:58:01.5350212Z * [new tag] v2.1.1 -> v2.1.1 2025-11-03T16:58:01.5351627Z * [new tag] v2.1.1-rc1 -> v2.1.1-rc1 2025-11-03T16:58:01.5352922Z * [new tag] v2.1.1-rc2 -> v2.1.1-rc2 2025-11-03T16:58:01.5354351Z * [new tag] v2.1.1-rc3 -> v2.1.1-rc3 2025-11-03T16:58:01.5355727Z * [new tag] v2.1.1-rc4 -> v2.1.1-rc4 2025-11-03T16:58:01.5357131Z * [new tag] v2.1.1-rc5 -> v2.1.1-rc5 2025-11-03T16:58:01.5358185Z * [new tag] v2.1.1-rc6 -> v2.1.1-rc6 2025-11-03T16:58:01.5359498Z * [new tag] v2.1.2 -> v2.1.2 2025-11-03T16:58:01.5360902Z * [new tag] v2.1.2-rc1 -> v2.1.2-rc1 2025-11-03T16:58:01.5362422Z * [new tag] v2.1.2-rc2 -> v2.1.2-rc2 2025-11-03T16:58:01.5363549Z * [new tag] v2.1.2-rc3 -> v2.1.2-rc3 2025-11-03T16:58:01.5364946Z * [new tag] v2.2.0 -> v2.2.0 2025-11-03T16:58:01.5366216Z * [new tag] v2.2.0-rc1 -> v2.2.0-rc1 2025-11-03T16:58:01.5367512Z * [new tag] v2.2.0-rc2 -> v2.2.0-rc2 2025-11-03T16:58:01.5368878Z * [new tag] v2.2.0-rc3 -> v2.2.0-rc3 2025-11-03T16:58:01.5370194Z * [new tag] v2.2.0-rc4 -> v2.2.0-rc4 2025-11-03T16:58:01.5371508Z * [new tag] v2.2.0-rc5 -> v2.2.0-rc5 2025-11-03T16:58:01.5372897Z * [new tag] v2.2.0-rc6 -> v2.2.0-rc6 2025-11-03T16:58:01.5374110Z * [new tag] v2.2.0-rc7 -> v2.2.0-rc7 2025-11-03T16:58:01.5375129Z * [new tag] v2.2.0-rc8 -> v2.2.0-rc8 2025-11-03T16:58:01.5376502Z * [new tag] v2.2.1 -> v2.2.1 2025-11-03T16:58:01.5377881Z * [new tag] v2.2.1-rc1 -> v2.2.1-rc1 2025-11-03T16:58:01.5378988Z * [new tag] v2.2.1-rc2 -> v2.2.1-rc2 2025-11-03T16:58:01.5380078Z * [new tag] v2.2.1-rc3 -> v2.2.1-rc3 2025-11-03T16:58:01.5381230Z * [new tag] v2.2.2 -> v2.2.2 2025-11-03T16:58:01.5382739Z * [new tag] v2.2.2-rc1 -> v2.2.2-rc1 2025-11-03T16:58:01.5383842Z * [new tag] v2.2.2-rc2 -> v2.2.2-rc2 2025-11-03T16:58:01.5384973Z * [new tag] v2.2.2-rc3 -> v2.2.2-rc3 2025-11-03T16:58:01.5386353Z * [new tag] v2.3.0 -> v2.3.0 2025-11-03T16:58:01.5387710Z * [new tag] v2.3.0-rc1 -> v2.3.0-rc1 2025-11-03T16:58:01.5389164Z * [new tag] v2.3.0-rc10 -> v2.3.0-rc10 2025-11-03T16:58:01.5390612Z * [new tag] v2.3.0-rc11 -> v2.3.0-rc11 2025-11-03T16:58:01.5391774Z * [new tag] v2.3.0-rc12 -> v2.3.0-rc12 2025-11-03T16:58:01.5393124Z * [new tag] v2.3.0-rc2 -> v2.3.0-rc2 2025-11-03T16:58:01.5394468Z * [new tag] v2.3.0-rc3 -> v2.3.0-rc3 2025-11-03T16:58:01.5395794Z * [new tag] v2.3.0-rc4 -> v2.3.0-rc4 2025-11-03T16:58:01.5397129Z * [new tag] v2.3.0-rc5 -> v2.3.0-rc5 2025-11-03T16:58:01.5398312Z * [new tag] v2.3.0-rc6 -> v2.3.0-rc6 2025-11-03T16:58:01.5399673Z * [new tag] v2.3.0-rc7 -> v2.3.0-rc7 2025-11-03T16:58:01.5401030Z * [new tag] v2.3.0-rc8 -> v2.3.0-rc8 2025-11-03T16:58:01.5402139Z * [new tag] v2.3.0-rc9 -> v2.3.0-rc9 2025-11-03T16:58:01.5403359Z * [new tag] v2.3.1 -> v2.3.1 2025-11-03T16:58:01.5404670Z * [new tag] v2.3.1-rc1 -> v2.3.1-rc1 2025-11-03T16:58:01.5405985Z * [new tag] v2.3.1-rc2 -> v2.3.1-rc2 2025-11-03T16:58:01.5407358Z * [new tag] v2.3.1-rc3 -> v2.3.1-rc3 2025-11-03T16:58:01.5409375Z * [new tag] v2.4.0 -> v2.4.0 2025-11-03T16:58:01.5412134Z * [new tag] v2.4.0-rc1 -> v2.4.0-rc1 2025-11-03T16:58:01.5413467Z * [new tag] v2.4.0-rc2 -> v2.4.0-rc2 2025-11-03T16:58:01.5414932Z * [new tag] v2.4.0-rc3 -> v2.4.0-rc3 2025-11-03T16:58:01.5416274Z * [new tag] v2.4.0-rc4 -> v2.4.0-rc4 2025-11-03T16:58:01.5417738Z * [new tag] v2.4.0-rc5 -> v2.4.0-rc5 2025-11-03T16:58:01.5419114Z * [new tag] v2.4.0-rc6 -> v2.4.0-rc6 2025-11-03T16:58:01.5420488Z * [new tag] v2.4.0-rc7 -> v2.4.0-rc7 2025-11-03T16:58:01.5421820Z * [new tag] v2.4.0-rc8 -> v2.4.0-rc8 2025-11-03T16:58:01.5423163Z * [new tag] v2.4.0-rc9 -> v2.4.0-rc9 2025-11-03T16:58:01.5424313Z * [new tag] v2.4.1 -> v2.4.1 2025-11-03T16:58:01.5425726Z * [new tag] v2.4.1-rc1 -> v2.4.1-rc1 2025-11-03T16:58:01.5427106Z * [new tag] v2.4.1-rc2 -> v2.4.1-rc2 2025-11-03T16:58:01.5428398Z * [new tag] v2.4.1-rc3 -> v2.4.1-rc3 2025-11-03T16:58:01.5429866Z * [new tag] v2.5.0 -> v2.5.0 2025-11-03T16:58:01.5431164Z * [new tag] v2.5.0-rc1 -> v2.5.0-rc1 2025-11-03T16:58:01.5432210Z * [new tag] v2.5.0-rc10 -> v2.5.0-rc10 2025-11-03T16:58:01.5433565Z * [new tag] v2.5.0-rc2 -> v2.5.0-rc2 2025-11-03T16:58:01.5434848Z * [new tag] v2.5.0-rc3 -> v2.5.0-rc3 2025-11-03T16:58:01.5436160Z * [new tag] v2.5.0-rc4 -> v2.5.0-rc4 2025-11-03T16:58:01.5437592Z * [new tag] v2.5.0-rc5 -> v2.5.0-rc5 2025-11-03T16:58:01.5438985Z * [new tag] v2.5.0-rc6 -> v2.5.0-rc6 2025-11-03T16:58:01.5440324Z * [new tag] v2.5.0-rc7 -> v2.5.0-rc7 2025-11-03T16:58:01.5441648Z * [new tag] v2.5.0-rc8 -> v2.5.0-rc8 2025-11-03T16:58:01.5443017Z * [new tag] v2.5.0-rc9 -> v2.5.0-rc9 2025-11-03T16:58:01.5444157Z * [new tag] v2.5.1 -> v2.5.1 2025-11-03T16:58:01.5445171Z * [new tag] v2.5.1-rc1 -> v2.5.1-rc1 2025-11-03T16:58:01.5446461Z * [new tag] v2.6.0 -> v2.6.0 2025-11-03T16:58:01.5447827Z * [new tag] v2.6.0-rc1 -> v2.6.0-rc1 2025-11-03T16:58:01.5449297Z * [new tag] v2.6.0-rc2 -> v2.6.0-rc2 2025-11-03T16:58:01.5450697Z * [new tag] v2.6.0-rc3 -> v2.6.0-rc3 2025-11-03T16:58:01.5451947Z * [new tag] v2.6.0-rc4 -> v2.6.0-rc4 2025-11-03T16:58:01.5453531Z * [new tag] v2.6.0-rc5 -> v2.6.0-rc5 2025-11-03T16:58:01.5454944Z * [new tag] v2.6.0-rc6 -> v2.6.0-rc6 2025-11-03T16:58:01.5456278Z * [new tag] v2.6.0-rc7 -> v2.6.0-rc7 2025-11-03T16:58:01.5457750Z * [new tag] v2.6.0-rc8 -> v2.6.0-rc8 2025-11-03T16:58:01.5459113Z * [new tag] v2.6.0-rc9 -> v2.6.0-rc9 2025-11-03T16:58:01.5460651Z * [new tag] v2.7.0 -> v2.7.0 2025-11-03T16:58:01.5461974Z * [new tag] v2.7.0-rc1 -> v2.7.0-rc1 2025-11-03T16:58:01.5463141Z * [new tag] v2.7.0-rc10 -> v2.7.0-rc10 2025-11-03T16:58:01.5464562Z * [new tag] v2.7.0-rc2 -> v2.7.0-rc2 2025-11-03T16:58:01.5465922Z * [new tag] v2.7.0-rc3 -> v2.7.0-rc3 2025-11-03T16:58:01.5467242Z * [new tag] v2.7.0-rc4 -> v2.7.0-rc4 2025-11-03T16:58:01.5468517Z * [new tag] v2.7.0-rc5 -> v2.7.0-rc5 2025-11-03T16:58:01.5469849Z * [new tag] v2.7.0-rc6 -> v2.7.0-rc6 2025-11-03T16:58:01.5471218Z * [new tag] v2.7.0-rc7 -> v2.7.0-rc7 2025-11-03T16:58:01.5472608Z * [new tag] v2.7.0-rc8 -> v2.7.0-rc8 2025-11-03T16:58:01.5474025Z * [new tag] v2.7.0-rc9 -> v2.7.0-rc9 2025-11-03T16:58:01.5475196Z * [new tag] v2.7.1 -> v2.7.1 2025-11-03T16:58:01.5476567Z * [new tag] v2.7.1-rc1 -> v2.7.1-rc1 2025-11-03T16:58:01.5477999Z * [new tag] v2.7.1-rc2 -> v2.7.1-rc2 2025-11-03T16:58:01.5479456Z * [new tag] v2.7.1-rc3 -> v2.7.1-rc3 2025-11-03T16:58:01.5480893Z * [new tag] v2.7.1-rc4 -> v2.7.1-rc4 2025-11-03T16:58:01.5482217Z * [new tag] v2.7.1-rc5 -> v2.7.1-rc5 2025-11-03T16:58:01.5483392Z * [new tag] v2.8.0 -> v2.8.0 2025-11-03T16:58:01.5484744Z * [new tag] v2.8.0-rc1 -> v2.8.0-rc1 2025-11-03T16:58:01.5486259Z * [new tag] v2.8.0-rc2 -> v2.8.0-rc2 2025-11-03T16:58:01.5487549Z * [new tag] v2.8.0-rc3 -> v2.8.0-rc3 2025-11-03T16:58:01.5489230Z * [new tag] v2.8.0-rc4 -> v2.8.0-rc4 2025-11-03T16:58:01.5490607Z * [new tag] v2.8.0-rc5 -> v2.8.0-rc5 2025-11-03T16:58:01.5491989Z * [new tag] v2.8.0-rc6 -> v2.8.0-rc6 2025-11-03T16:58:01.5493354Z * [new tag] v2.8.0-rc7 -> v2.8.0-rc7 2025-11-03T16:58:01.5494684Z * [new tag] v2.8.0-rc8 -> v2.8.0-rc8 2025-11-03T16:58:01.5496066Z * [new tag] v2.9.0 -> v2.9.0 2025-11-03T16:58:01.5497892Z * [new tag] v2.9.0-rc1 -> v2.9.0-rc1 2025-11-03T16:58:01.5499217Z * [new tag] v2.9.0-rc10 -> v2.9.0-rc10 2025-11-03T16:58:01.5500392Z * [new tag] v2.9.0-rc11 -> v2.9.0-rc11 2025-11-03T16:58:01.5501963Z * [new tag] v2.9.0-rc2 -> v2.9.0-rc2 2025-11-03T16:58:01.5503357Z * [new tag] v2.9.0-rc3 -> v2.9.0-rc3 2025-11-03T16:58:01.5504834Z * [new tag] v2.9.0-rc4 -> v2.9.0-rc4 2025-11-03T16:58:01.5506234Z * [new tag] v2.9.0-rc5 -> v2.9.0-rc5 2025-11-03T16:58:01.5507797Z * [new tag] v2.9.0-rc6 -> v2.9.0-rc6 2025-11-03T16:58:01.5509106Z * [new tag] v2.9.0-rc7 -> v2.9.0-rc7 2025-11-03T16:58:01.5510953Z * [new tag] v2.9.0-rc8 -> v2.9.0-rc8 2025-11-03T16:58:01.5512096Z * [new tag] v2.9.0-rc9 -> v2.9.0-rc9 2025-11-03T16:58:01.5513993Z * [new tag] viable/strict/1759343184 -> viable/strict/1759343184 2025-11-03T16:58:01.5515394Z * [new tag] viable/strict/1759346540 -> viable/strict/1759346540 2025-11-03T16:58:01.5516555Z * [new tag] viable/strict/1759348181 -> viable/strict/1759348181 2025-11-03T16:58:01.5517803Z * [new tag] viable/strict/1759350324 -> viable/strict/1759350324 2025-11-03T16:58:01.5519034Z * [new tag] viable/strict/1759351793 -> viable/strict/1759351793 2025-11-03T16:58:01.5520337Z * [new tag] viable/strict/1759353844 -> viable/strict/1759353844 2025-11-03T16:58:01.5521588Z * [new tag] viable/strict/1759355374 -> viable/strict/1759355374 2025-11-03T16:58:01.5522713Z * [new tag] viable/strict/1759357472 -> viable/strict/1759357472 2025-11-03T16:58:01.5524003Z * [new tag] viable/strict/1759361002 -> viable/strict/1759361002 2025-11-03T16:58:01.5525653Z * [new tag] viable/strict/1759362585 -> viable/strict/1759362585 2025-11-03T16:58:01.5527204Z * [new tag] viable/strict/1759365359 -> viable/strict/1759365359 2025-11-03T16:58:01.5528748Z * [new tag] viable/strict/1759370089 -> viable/strict/1759370089 2025-11-03T16:58:01.5530155Z * [new tag] viable/strict/1759377554 -> viable/strict/1759377554 2025-11-03T16:58:01.5531397Z * [new tag] viable/strict/1759379133 -> viable/strict/1759379133 2025-11-03T16:58:01.5532701Z * [new tag] viable/strict/1759389871 -> viable/strict/1759389871 2025-11-03T16:58:01.5534080Z * [new tag] viable/strict/1759393562 -> viable/strict/1759393562 2025-11-03T16:58:01.5535433Z * [new tag] viable/strict/1759395076 -> viable/strict/1759395076 2025-11-03T16:58:01.5536872Z * [new tag] viable/strict/1759398579 -> viable/strict/1759398579 2025-11-03T16:58:01.5538312Z * [new tag] viable/strict/1759404142 -> viable/strict/1759404142 2025-11-03T16:58:01.5539803Z * [new tag] viable/strict/1759405773 -> viable/strict/1759405773 2025-11-03T16:58:01.5541013Z * [new tag] viable/strict/1759408041 -> viable/strict/1759408041 2025-11-03T16:58:01.5542380Z * [new tag] viable/strict/1759411593 -> viable/strict/1759411593 2025-11-03T16:58:01.5543662Z * [new tag] viable/strict/1759427395 -> viable/strict/1759427395 2025-11-03T16:58:01.5544991Z * [new tag] viable/strict/1759434582 -> viable/strict/1759434582 2025-11-03T16:58:01.5546474Z * [new tag] viable/strict/1759436720 -> viable/strict/1759436720 2025-11-03T16:58:01.5547817Z * [new tag] viable/strict/1759440219 -> viable/strict/1759440219 2025-11-03T16:58:01.5549148Z * [new tag] viable/strict/1759441948 -> viable/strict/1759441948 2025-11-03T16:58:01.5550480Z * [new tag] viable/strict/1759443860 -> viable/strict/1759443860 2025-11-03T16:58:01.5551809Z * [new tag] viable/strict/1759445377 -> viable/strict/1759445377 2025-11-03T16:58:01.5553313Z * [new tag] viable/strict/1759447415 -> viable/strict/1759447415 2025-11-03T16:58:01.5554625Z * [new tag] viable/strict/1759451750 -> viable/strict/1759451750 2025-11-03T16:58:01.5555975Z * [new tag] viable/strict/1759453910 -> viable/strict/1759453910 2025-11-03T16:58:01.5557344Z * [new tag] viable/strict/1759456483 -> viable/strict/1759456483 2025-11-03T16:58:01.5558738Z * [new tag] viable/strict/1759459279 -> viable/strict/1759459279 2025-11-03T16:58:01.5560058Z * [new tag] viable/strict/1759460742 -> viable/strict/1759460742 2025-11-03T16:58:01.5561406Z * [new tag] viable/strict/1759462025 -> viable/strict/1759462025 2025-11-03T16:58:01.5562736Z * [new tag] viable/strict/1759469086 -> viable/strict/1759469086 2025-11-03T16:58:01.5564082Z * [new tag] viable/strict/1759470581 -> viable/strict/1759470581 2025-11-03T16:58:01.5565422Z * [new tag] viable/strict/1759472786 -> viable/strict/1759472786 2025-11-03T16:58:01.5566811Z * [new tag] viable/strict/1759476294 -> viable/strict/1759476294 2025-11-03T16:58:01.5568200Z * [new tag] viable/strict/1759479963 -> viable/strict/1759479963 2025-11-03T16:58:01.5569648Z * [new tag] viable/strict/1759492177 -> viable/strict/1759492177 2025-11-03T16:58:01.5570977Z * [new tag] viable/strict/1759519278 -> viable/strict/1759519278 2025-11-03T16:58:01.5572742Z * [new tag] viable/strict/1759524580 -> viable/strict/1759524580 2025-11-03T16:58:01.5574107Z * [new tag] viable/strict/1759528193 -> viable/strict/1759528193 2025-11-03T16:58:01.5575325Z * [new tag] viable/strict/1759533797 -> viable/strict/1759533797 2025-11-03T16:58:01.5576698Z * [new tag] viable/strict/1759542780 -> viable/strict/1759542780 2025-11-03T16:58:01.5578036Z * [new tag] viable/strict/1759549779 -> viable/strict/1759549779 2025-11-03T16:58:01.5579525Z * [new tag] viable/strict/1759555455 -> viable/strict/1759555455 2025-11-03T16:58:01.5580862Z * [new tag] viable/strict/1759559176 -> viable/strict/1759559176 2025-11-03T16:58:01.5582355Z * [new tag] viable/strict/1759560629 -> viable/strict/1759560629 2025-11-03T16:58:01.5583700Z * [new tag] viable/strict/1759569848 -> viable/strict/1759569848 2025-11-03T16:58:01.5585219Z * [new tag] viable/strict/1759571382 -> viable/strict/1759571382 2025-11-03T16:58:01.5586542Z * [new tag] viable/strict/1759573474 -> viable/strict/1759573474 2025-11-03T16:58:01.5587915Z * [new tag] viable/strict/1759618187 -> viable/strict/1759618187 2025-11-03T16:58:01.5589406Z * [new tag] viable/strict/1759626742 -> viable/strict/1759626742 2025-11-03T16:58:01.5590788Z * [new tag] viable/strict/1759632427 -> viable/strict/1759632427 2025-11-03T16:58:01.5592106Z * [new tag] viable/strict/1759634971 -> viable/strict/1759634971 2025-11-03T16:58:01.5593533Z * [new tag] viable/strict/1759661382 -> viable/strict/1759661382 2025-11-03T16:58:01.5594906Z * [new tag] viable/strict/1759663294 -> viable/strict/1759663294 2025-11-03T16:58:01.5596119Z * [new tag] viable/strict/1759708178 -> viable/strict/1759708178 2025-11-03T16:58:01.5597525Z * [new tag] viable/strict/1759715695 -> viable/strict/1759715695 2025-11-03T16:58:01.5598816Z * [new tag] viable/strict/1759728293 -> viable/strict/1759728293 2025-11-03T16:58:01.5600237Z * [new tag] viable/strict/1759735513 -> viable/strict/1759735513 2025-11-03T16:58:01.5601600Z * [new tag] viable/strict/1759739177 -> viable/strict/1759739177 2025-11-03T16:58:01.5602962Z * [new tag] viable/strict/1759758635 -> viable/strict/1759758635 2025-11-03T16:58:01.5604295Z * [new tag] viable/strict/1759765784 -> viable/strict/1759765784 2025-11-03T16:58:01.5605622Z * [new tag] viable/strict/1759767948 -> viable/strict/1759767948 2025-11-03T16:58:01.5607017Z * [new tag] viable/strict/1759771461 -> viable/strict/1759771461 2025-11-03T16:58:01.5608223Z * [new tag] viable/strict/1759776706 -> viable/strict/1759776706 2025-11-03T16:58:01.5609875Z * [new tag] viable/strict/1759782317 -> viable/strict/1759782317 2025-11-03T16:58:01.5611526Z * [new tag] viable/strict/1759783777 -> viable/strict/1759783777 2025-11-03T16:58:01.5612971Z * [new tag] viable/strict/1759785815 -> viable/strict/1759785815 2025-11-03T16:58:01.5614171Z * [new tag] viable/strict/1759789459 -> viable/strict/1759789459 2025-11-03T16:58:01.5615401Z * [new tag] viable/strict/1759790974 -> viable/strict/1759790974 2025-11-03T16:58:01.5616615Z * [new tag] viable/strict/1759794583 -> viable/strict/1759794583 2025-11-03T16:58:01.5618025Z * [new tag] viable/strict/1759797408 -> viable/strict/1759797408 2025-11-03T16:58:01.5619388Z * [new tag] viable/strict/1759799518 -> viable/strict/1759799518 2025-11-03T16:58:01.5620731Z * [new tag] viable/strict/1759804909 -> viable/strict/1759804909 2025-11-03T16:58:01.5622141Z * [new tag] viable/strict/1759807643 -> viable/strict/1759807643 2025-11-03T16:58:01.5623352Z * [new tag] viable/strict/1759809089 -> viable/strict/1759809089 2025-11-03T16:58:01.5624679Z * [new tag] viable/strict/1759811145 -> viable/strict/1759811145 2025-11-03T16:58:01.5626081Z * [new tag] viable/strict/1759812581 -> viable/strict/1759812581 2025-11-03T16:58:01.5627502Z * [new tag] viable/strict/1759814683 -> viable/strict/1759814683 2025-11-03T16:58:01.5628929Z * [new tag] viable/strict/1759821889 -> viable/strict/1759821889 2025-11-03T16:58:01.5630330Z * [new tag] viable/strict/1759823376 -> viable/strict/1759823376 2025-11-03T16:58:01.5631674Z * [new tag] viable/strict/1759827107 -> viable/strict/1759827107 2025-11-03T16:58:01.5633010Z * [new tag] viable/strict/1759830577 -> viable/strict/1759830577 2025-11-03T16:58:01.5634537Z * [new tag] viable/strict/1759832720 -> viable/strict/1759832720 2025-11-03T16:58:01.5635853Z * [new tag] viable/strict/1759842063 -> viable/strict/1759842063 2025-11-03T16:58:01.5637216Z * [new tag] viable/strict/1759847121 -> viable/strict/1759847121 2025-11-03T16:58:01.5638895Z * [new tag] viable/strict/1759850721 -> viable/strict/1759850721 2025-11-03T16:58:01.5640034Z * [new tag] viable/strict/1759857870 -> viable/strict/1759857870 2025-11-03T16:58:01.5641215Z * [new tag] viable/strict/1759863143 -> viable/strict/1759863143 2025-11-03T16:58:01.5642594Z * [new tag] viable/strict/1759875874 -> viable/strict/1759875874 2025-11-03T16:58:01.5643815Z * [new tag] viable/strict/1759877385 -> viable/strict/1759877385 2025-11-03T16:58:01.5645182Z * [new tag] viable/strict/1759883801 -> viable/strict/1759883801 2025-11-03T16:58:01.5646537Z * [new tag] viable/strict/1759885922 -> viable/strict/1759885922 2025-11-03T16:58:01.5647744Z * [new tag] viable/strict/1759888488 -> viable/strict/1759888488 2025-11-03T16:58:01.5649235Z * [new tag] viable/strict/1759895471 -> viable/strict/1759895471 2025-11-03T16:58:01.5650629Z * [new tag] viable/strict/1759904803 -> viable/strict/1759904803 2025-11-03T16:58:01.5652286Z * [new tag] viable/strict/1759908300 -> viable/strict/1759908300 2025-11-03T16:58:01.5653547Z * [new tag] viable/strict/1759915520 -> viable/strict/1759915520 2025-11-03T16:58:01.5654910Z * [new tag] viable/strict/1759916978 -> viable/strict/1759916978 2025-11-03T16:58:01.5656143Z * [new tag] viable/strict/1759930024 -> viable/strict/1759930024 2025-11-03T16:58:01.5657611Z * [new tag] viable/strict/1759948122 -> viable/strict/1759948122 2025-11-03T16:58:01.5659003Z * [new tag] viable/strict/1759952983 -> viable/strict/1759952983 2025-11-03T16:58:01.5660656Z * [new tag] viable/strict/1759955121 -> viable/strict/1759955121 2025-11-03T16:58:01.5661976Z * [new tag] viable/strict/1759962298 -> viable/strict/1759962298 2025-11-03T16:58:01.5663192Z * [new tag] viable/strict/1759965837 -> viable/strict/1759965837 2025-11-03T16:58:01.5664557Z * [new tag] viable/strict/1759970213 -> viable/strict/1759970213 2025-11-03T16:58:01.5665929Z * [new tag] viable/strict/1759974894 -> viable/strict/1759974894 2025-11-03T16:58:01.5667248Z * [new tag] viable/strict/1759977763 -> viable/strict/1759977763 2025-11-03T16:58:01.5668613Z * [new tag] viable/strict/1759979241 -> viable/strict/1759979241 2025-11-03T16:58:01.5670031Z * [new tag] viable/strict/1759985417 -> viable/strict/1759985417 2025-11-03T16:58:01.5671398Z * [new tag] viable/strict/1759987490 -> viable/strict/1759987490 2025-11-03T16:58:01.5672737Z * [new tag] viable/strict/1759996180 -> viable/strict/1759996180 2025-11-03T16:58:01.5674048Z * [new tag] viable/strict/1760065682 -> viable/strict/1760065682 2025-11-03T16:58:01.5675652Z * [new tag] viable/strict/1760066894 -> viable/strict/1760066894 2025-11-03T16:58:01.5677069Z * [new tag] viable/strict/1760070345 -> viable/strict/1760070345 2025-11-03T16:58:01.5678488Z * [new tag] viable/strict/1760089782 -> viable/strict/1760089782 2025-11-03T16:58:01.5679895Z * [new tag] viable/strict/1760091921 -> viable/strict/1760091921 2025-11-03T16:58:01.5681319Z * [new tag] viable/strict/1760127924 -> viable/strict/1760127924 2025-11-03T16:58:01.5682707Z * [new tag] viable/strict/1760129489 -> viable/strict/1760129489 2025-11-03T16:58:01.5684105Z * [new tag] viable/strict/1760132980 -> viable/strict/1760132980 2025-11-03T16:58:01.5685521Z * [new tag] viable/strict/1760135060 -> viable/strict/1760135060 2025-11-03T16:58:01.5686866Z * [new tag] viable/strict/1760215782 -> viable/strict/1760215782 2025-11-03T16:58:01.5688496Z * [new tag] viable/strict/1760273849 -> viable/strict/1760273849 2025-11-03T16:58:01.5689856Z * [new tag] viable/strict/1760275517 -> viable/strict/1760275517 2025-11-03T16:58:01.5691086Z * [new tag] viable/strict/1760276979 -> viable/strict/1760276979 2025-11-03T16:58:01.5692327Z * [new tag] viable/strict/1760279007 -> viable/strict/1760279007 2025-11-03T16:58:01.5693560Z * [new tag] viable/strict/1760286328 -> viable/strict/1760286328 2025-11-03T16:58:01.5694765Z * [new tag] viable/strict/1760493304 -> viable/strict/1760493304 2025-11-03T16:58:01.5695990Z * [new tag] viable/strict/1760496298 -> viable/strict/1760496298 2025-11-03T16:58:01.5697410Z * [new tag] viable/strict/1760518396 -> viable/strict/1760518396 2025-11-03T16:58:01.5698740Z * [new tag] viable/strict/1760534864 -> viable/strict/1760534864 2025-11-03T16:58:01.5699981Z * [new tag] viable/strict/1760549062 -> viable/strict/1760549062 2025-11-03T16:58:01.5701468Z * [new tag] viable/strict/1760552799 -> viable/strict/1760552799 2025-11-03T16:58:01.5702820Z * [new tag] viable/strict/1760554355 -> viable/strict/1760554355 2025-11-03T16:58:01.5704184Z * [new tag] viable/strict/1760556275 -> viable/strict/1760556275 2025-11-03T16:58:01.5705572Z * [new tag] viable/strict/1760564979 -> viable/strict/1760564979 2025-11-03T16:58:01.5706991Z * [new tag] viable/strict/1760567049 -> viable/strict/1760567049 2025-11-03T16:58:01.5708707Z * [new tag] viable/strict/1760568585 -> viable/strict/1760568585 2025-11-03T16:58:01.5710052Z * [new tag] viable/strict/1760570630 -> viable/strict/1760570630 2025-11-03T16:58:01.5711708Z * [new tag] viable/strict/1760572180 -> viable/strict/1760572180 2025-11-03T16:58:01.5713099Z * [new tag] viable/strict/1760575094 -> viable/strict/1760575094 2025-11-03T16:58:01.5714587Z * [new tag] viable/strict/1760579709 -> viable/strict/1760579709 2025-11-03T16:58:01.5716296Z * [new tag] viable/strict/1760582614 -> viable/strict/1760582614 2025-11-03T16:58:01.5717645Z * [new tag] viable/strict/1760586815 -> viable/strict/1760586815 2025-11-03T16:58:01.5718877Z * [new tag] viable/strict/1760588829 -> viable/strict/1760588829 2025-11-03T16:58:01.5720228Z * [new tag] viable/strict/1760590200 -> viable/strict/1760590200 2025-11-03T16:58:01.5721722Z * [new tag] viable/strict/1760592311 -> viable/strict/1760592311 2025-11-03T16:58:01.5723705Z * [new tag] viable/strict/1760619733 -> viable/strict/1760619733 2025-11-03T16:58:01.5724795Z * [new tag] viable/strict/1760628335 -> viable/strict/1760628335 2025-11-03T16:58:01.5725537Z * [new tag] viable/strict/1760635490 -> viable/strict/1760635490 2025-11-03T16:58:01.5727093Z * [new tag] viable/strict/1760640743 -> viable/strict/1760640743 2025-11-03T16:58:01.5728286Z * [new tag] viable/strict/1760642528 -> viable/strict/1760642528 2025-11-03T16:58:01.5729784Z * [new tag] viable/strict/1760646330 -> viable/strict/1760646330 2025-11-03T16:58:01.5731103Z * [new tag] viable/strict/1760666101 -> viable/strict/1760666101 2025-11-03T16:58:01.5732570Z * [new tag] viable/strict/1760668990 -> viable/strict/1760668990 2025-11-03T16:58:01.5733920Z * [new tag] viable/strict/1760670600 -> viable/strict/1760670600 2025-11-03T16:58:01.5735295Z * [new tag] viable/strict/1760671704 -> viable/strict/1760671704 2025-11-03T16:58:01.5736634Z * [new tag] viable/strict/1760673121 -> viable/strict/1760673121 2025-11-03T16:58:01.5738267Z * [new tag] viable/strict/1760675352 -> viable/strict/1760675352 2025-11-03T16:58:01.5739444Z * [new tag] viable/strict/1760696731 -> viable/strict/1760696731 2025-11-03T16:58:01.5741936Z * [new tag] viable/strict/1760723515 -> viable/strict/1760723515 2025-11-03T16:58:01.5743330Z * [new tag] viable/strict/1760727234 -> viable/strict/1760727234 2025-11-03T16:58:01.5744777Z * [new tag] viable/strict/1760730578 -> viable/strict/1760730578 2025-11-03T16:58:01.5745987Z * [new tag] viable/strict/1760732726 -> viable/strict/1760732726 2025-11-03T16:58:01.5747345Z * [new tag] viable/strict/1760734180 -> viable/strict/1760734180 2025-11-03T16:58:01.5748647Z * [new tag] viable/strict/1760736251 -> viable/strict/1760736251 2025-11-03T16:58:01.5750535Z * [new tag] viable/strict/1760737772 -> viable/strict/1760737772 2025-11-03T16:58:01.5751939Z * [new tag] viable/strict/1760758005 -> viable/strict/1760758005 2025-11-03T16:58:01.5753335Z * [new tag] viable/strict/1760761532 -> viable/strict/1760761532 2025-11-03T16:58:01.5754674Z * [new tag] viable/strict/1760802581 -> viable/strict/1760802581 2025-11-03T16:58:01.5756101Z * [new tag] viable/strict/1760827772 -> viable/strict/1760827772 2025-11-03T16:58:01.5757424Z * [new tag] viable/strict/1760834524 -> viable/strict/1760834524 2025-11-03T16:58:01.5758628Z * [new tag] viable/strict/1760845009 -> viable/strict/1760845009 2025-11-03T16:58:01.5760075Z * [new tag] viable/strict/1760876836 -> viable/strict/1760876836 2025-11-03T16:58:01.5761412Z * [new tag] viable/strict/1760880329 -> viable/strict/1760880329 2025-11-03T16:58:01.5762809Z * [new tag] viable/strict/1760888987 -> viable/strict/1760888987 2025-11-03T16:58:01.5764144Z * [new tag] viable/strict/1760912664 -> viable/strict/1760912664 2025-11-03T16:58:01.5765489Z * [new tag] viable/strict/1760925321 -> viable/strict/1760925321 2025-11-03T16:58:01.5766823Z * [new tag] viable/strict/1760931488 -> viable/strict/1760931488 2025-11-03T16:58:01.5768258Z * [new tag] viable/strict/1760932693 -> viable/strict/1760932693 2025-11-03T16:58:01.5769957Z * [new tag] viable/strict/1761004184 -> viable/strict/1761004184 2025-11-03T16:58:01.5771310Z * [new tag] viable/strict/1761014748 -> viable/strict/1761014748 2025-11-03T16:58:01.5772578Z * [new tag] viable/strict/1761017491 -> viable/strict/1761017491 2025-11-03T16:58:01.5773983Z * [new tag] viable/strict/1761018806 -> viable/strict/1761018806 2025-11-03T16:58:01.5775258Z * [new tag] viable/strict/1761020754 -> viable/strict/1761020754 2025-11-03T16:58:01.5776595Z * [new tag] viable/strict/1761024303 -> viable/strict/1761024303 2025-11-03T16:58:01.5777976Z * [new tag] viable/strict/1761029582 -> viable/strict/1761029582 2025-11-03T16:58:01.5779299Z * [new tag] viable/strict/1761031535 -> viable/strict/1761031535 2025-11-03T16:58:01.5780682Z * [new tag] viable/strict/1761035196 -> viable/strict/1761035196 2025-11-03T16:58:01.5782116Z * [new tag] viable/strict/1761045825 -> viable/strict/1761045825 2025-11-03T16:58:01.5783620Z * [new tag] viable/strict/1761054796 -> viable/strict/1761054796 2025-11-03T16:58:01.5784971Z * [new tag] viable/strict/1761060314 -> viable/strict/1761060314 2025-11-03T16:58:01.5786407Z * [new tag] viable/strict/1761071198 -> viable/strict/1761071198 2025-11-03T16:58:01.5787866Z * [new tag] viable/strict/1761074628 -> viable/strict/1761074628 2025-11-03T16:58:01.5789333Z * [new tag] viable/strict/1761078351 -> viable/strict/1761078351 2025-11-03T16:58:01.5790643Z * [new tag] viable/strict/1761079822 -> viable/strict/1761079822 2025-11-03T16:58:01.5791832Z * [new tag] viable/strict/1761081873 -> viable/strict/1761081873 2025-11-03T16:58:01.5793242Z * [new tag] viable/strict/1761083392 -> viable/strict/1761083392 2025-11-03T16:58:01.5794631Z * [new tag] viable/strict/1761085465 -> viable/strict/1761085465 2025-11-03T16:58:01.5796046Z * [new tag] viable/strict/1761089099 -> viable/strict/1761089099 2025-11-03T16:58:01.5797380Z * [new tag] viable/strict/1761095535 -> viable/strict/1761095535 2025-11-03T16:58:01.5798700Z * [new tag] viable/strict/1761098119 -> viable/strict/1761098119 2025-11-03T16:58:01.5800453Z * [new tag] viable/strict/1761101330 -> viable/strict/1761101330 2025-11-03T16:58:01.5801825Z * [new tag] viable/strict/1761114425 -> viable/strict/1761114425 2025-11-03T16:58:01.5803253Z * [new tag] viable/strict/1761116036 -> viable/strict/1761116036 2025-11-03T16:58:01.5804618Z * [new tag] viable/strict/1761119379 -> viable/strict/1761119379 2025-11-03T16:58:01.5805991Z * [new tag] viable/strict/1761121601 -> viable/strict/1761121601 2025-11-03T16:58:01.5807314Z * [new tag] viable/strict/1761123234 -> viable/strict/1761123234 2025-11-03T16:58:01.5808740Z * [new tag] viable/strict/1761126621 -> viable/strict/1761126621 2025-11-03T16:58:01.5810404Z * [new tag] viable/strict/1761132259 -> viable/strict/1761132259 2025-11-03T16:58:01.5812117Z * [new tag] viable/strict/1761146746 -> viable/strict/1761146746 2025-11-03T16:58:01.5813658Z * [new tag] viable/strict/1761164752 -> viable/strict/1761164752 2025-11-03T16:58:01.5815071Z * [new tag] viable/strict/1761166198 -> viable/strict/1761166198 2025-11-03T16:58:01.5816478Z * [new tag] viable/strict/1761175424 -> viable/strict/1761175424 2025-11-03T16:58:01.5817826Z * [new tag] viable/strict/1761176983 -> viable/strict/1761176983 2025-11-03T16:58:01.5819194Z * [new tag] viable/strict/1761179891 -> viable/strict/1761179891 2025-11-03T16:58:01.5820453Z * [new tag] viable/strict/1761181930 -> viable/strict/1761181930 2025-11-03T16:58:01.5821879Z * [new tag] viable/strict/1761184516 -> viable/strict/1761184516 2025-11-03T16:58:01.5823269Z * [new tag] viable/strict/1761190179 -> viable/strict/1761190179 2025-11-03T16:58:01.5824615Z * [new tag] viable/strict/1761193558 -> viable/strict/1761193558 2025-11-03T16:58:01.5826000Z * [new tag] viable/strict/1761207990 -> viable/strict/1761207990 2025-11-03T16:58:01.5827475Z * [new tag] viable/strict/1761229539 -> viable/strict/1761229539 2025-11-03T16:58:01.5828892Z * [new tag] viable/strict/1761244031 -> viable/strict/1761244031 2025-11-03T16:58:01.5830229Z * [new tag] viable/strict/1761248986 -> viable/strict/1761248986 2025-11-03T16:58:01.5831452Z * [new tag] viable/strict/1761259791 -> viable/strict/1761259791 2025-11-03T16:58:01.5832807Z * [new tag] viable/strict/1761266139 -> viable/strict/1761266139 2025-11-03T16:58:01.5834188Z * [new tag] viable/strict/1761268316 -> viable/strict/1761268316 2025-11-03T16:58:01.5835501Z * [new tag] viable/strict/1761273805 -> viable/strict/1761273805 2025-11-03T16:58:01.5836858Z * [new tag] viable/strict/1761275261 -> viable/strict/1761275261 2025-11-03T16:58:01.5838652Z * [new tag] viable/strict/1761277913 -> viable/strict/1761277913 2025-11-03T16:58:01.5840324Z * [new tag] viable/strict/1761290701 -> viable/strict/1761290701 2025-11-03T16:58:01.5841664Z * [new tag] viable/strict/1761294396 -> viable/strict/1761294396 2025-11-03T16:58:01.5843029Z * [new tag] viable/strict/1761303047 -> viable/strict/1761303047 2025-11-03T16:58:01.5844434Z * [new tag] viable/strict/1761335388 -> viable/strict/1761335388 2025-11-03T16:58:01.5845820Z * [new tag] viable/strict/1761337551 -> viable/strict/1761337551 2025-11-03T16:58:01.5847210Z * [new tag] viable/strict/1761339007 -> viable/strict/1761339007 2025-11-03T16:58:01.5848479Z * [new tag] viable/strict/1761341050 -> viable/strict/1761341050 2025-11-03T16:58:01.5849946Z * [new tag] viable/strict/1761346188 -> viable/strict/1761346188 2025-11-03T16:58:01.5851362Z * [new tag] viable/strict/1761349792 -> viable/strict/1761349792 2025-11-03T16:58:01.5852578Z * [new tag] viable/strict/1761352620 -> viable/strict/1761352620 2025-11-03T16:58:01.5853679Z * [new tag] viable/strict/1761354730 -> viable/strict/1761354730 2025-11-03T16:58:01.5855297Z * [new tag] viable/strict/1761357298 -> viable/strict/1761357298 2025-11-03T16:58:01.5856422Z * [new tag] viable/strict/1761360201 -> viable/strict/1761360201 2025-11-03T16:58:01.5858099Z * [new tag] viable/strict/1761361753 -> viable/strict/1761361753 2025-11-03T16:58:01.5859261Z * [new tag] viable/strict/1761364351 -> viable/strict/1761364351 2025-11-03T16:58:01.5860747Z * [new tag] viable/strict/1761366338 -> viable/strict/1761366338 2025-11-03T16:58:01.5862447Z * [new tag] viable/strict/1761367802 -> viable/strict/1761367802 2025-11-03T16:58:01.5863988Z * [new tag] viable/strict/1761369889 -> viable/strict/1761369889 2025-11-03T16:58:01.5865021Z * [new tag] viable/strict/1761371385 -> viable/strict/1761371385 2025-11-03T16:58:01.5866580Z * [new tag] viable/strict/1761373581 -> viable/strict/1761373581 2025-11-03T16:58:01.5868072Z * [new tag] viable/strict/1761375054 -> viable/strict/1761375054 2025-11-03T16:58:01.5869488Z * [new tag] viable/strict/1761421785 -> viable/strict/1761421785 2025-11-03T16:58:01.5870990Z * [new tag] viable/strict/1761434614 -> viable/strict/1761434614 2025-11-03T16:58:01.5872599Z * [new tag] viable/strict/1761439254 -> viable/strict/1761439254 2025-11-03T16:58:01.5873654Z * [new tag] viable/strict/1761454187 -> viable/strict/1761454187 2025-11-03T16:58:01.5875262Z * [new tag] viable/strict/1761459991 -> viable/strict/1761459991 2025-11-03T16:58:01.5876694Z * [new tag] viable/strict/1761470668 -> viable/strict/1761470668 2025-11-03T16:58:01.5878391Z * [new tag] viable/strict/1761472188 -> viable/strict/1761472188 2025-11-03T16:58:01.5879770Z * [new tag] viable/strict/1761503178 -> viable/strict/1761503178 2025-11-03T16:58:01.5881130Z * [new tag] viable/strict/1761517492 -> viable/strict/1761517492 2025-11-03T16:58:01.5882503Z * [new tag] viable/strict/1761518981 -> viable/strict/1761518981 2025-11-03T16:58:01.5883856Z * [new tag] viable/strict/1761533609 -> viable/strict/1761533609 2025-11-03T16:58:01.5884927Z * [new tag] viable/strict/1761546438 -> viable/strict/1761546438 2025-11-03T16:58:01.5886344Z * [new tag] viable/strict/1761548133 -> viable/strict/1761548133 2025-11-03T16:58:01.5887560Z * [new tag] viable/strict/1761555186 -> viable/strict/1761555186 2025-11-03T16:58:01.5889003Z * [new tag] viable/strict/1761557178 -> viable/strict/1761557178 2025-11-03T16:58:01.5890096Z * [new tag] viable/strict/1761560772 -> viable/strict/1761560772 2025-11-03T16:58:01.5891672Z * [new tag] viable/strict/1761562266 -> viable/strict/1761562266 2025-11-03T16:58:01.5892602Z * [new tag] viable/strict/1761564260 -> viable/strict/1761564260 2025-11-03T16:58:01.5894036Z * [new tag] viable/strict/1761568072 -> viable/strict/1761568072 2025-11-03T16:58:01.5895080Z * [new tag] viable/strict/1761571683 -> viable/strict/1761571683 2025-11-03T16:58:01.5896521Z * [new tag] viable/strict/1761580199 -> viable/strict/1761580199 2025-11-03T16:58:01.5897562Z * [new tag] viable/strict/1761587383 -> viable/strict/1761587383 2025-11-03T16:58:01.5899020Z * [new tag] viable/strict/1761591165 -> viable/strict/1761591165 2025-11-03T16:58:01.5900056Z * [new tag] viable/strict/1761594575 -> viable/strict/1761594575 2025-11-03T16:58:01.5901576Z * [new tag] viable/strict/1761596710 -> viable/strict/1761596710 2025-11-03T16:58:01.5902972Z * [new tag] viable/strict/1761598189 -> viable/strict/1761598189 2025-11-03T16:58:01.5903906Z * [new tag] viable/strict/1761600254 -> viable/strict/1761600254 2025-11-03T16:58:01.5905313Z * [new tag] viable/strict/1761603879 -> viable/strict/1761603879 2025-11-03T16:58:01.5906381Z * [new tag] viable/strict/1761605429 -> viable/strict/1761605429 2025-11-03T16:58:01.5907874Z * [new tag] viable/strict/1761607468 -> viable/strict/1761607468 2025-11-03T16:58:01.5908906Z * [new tag] viable/strict/1761608983 -> viable/strict/1761608983 2025-11-03T16:58:01.5910598Z * [new tag] viable/strict/1761611846 -> viable/strict/1761611846 2025-11-03T16:58:01.5911667Z * [new tag] viable/strict/1761613922 -> viable/strict/1761613922 2025-11-03T16:58:01.5913140Z * [new tag] viable/strict/1761616504 -> viable/strict/1761616504 2025-11-03T16:58:01.5914161Z * [new tag] viable/strict/1761619599 -> viable/strict/1761619599 2025-11-03T16:58:01.5915582Z * [new tag] viable/strict/1761686693 -> viable/strict/1761686693 2025-11-03T16:58:01.5916635Z * [new tag] viable/strict/1761688179 -> viable/strict/1761688179 2025-11-03T16:58:01.5918458Z * [new tag] viable/strict/1761691973 -> viable/strict/1761691973 2025-11-03T16:58:01.5919114Z * [new tag] viable/strict/1761693884 -> viable/strict/1761693884 2025-11-03T16:58:01.5920652Z * [new tag] viable/strict/1761695389 -> viable/strict/1761695389 2025-11-03T16:58:01.5921621Z * [new tag] viable/strict/1761698408 -> viable/strict/1761698408 2025-11-03T16:58:01.5923170Z * [new tag] viable/strict/1761702931 -> viable/strict/1761702931 2025-11-03T16:58:01.5924151Z * [new tag] viable/strict/1761706307 -> viable/strict/1761706307 2025-11-03T16:58:01.5926095Z * [new tag] viable/strict/1761709065 -> viable/strict/1761709065 2025-11-03T16:58:01.5927136Z * [new tag] viable/strict/1761710285 -> viable/strict/1761710285 2025-11-03T16:58:01.5928706Z * [new tag] viable/strict/1761711983 -> viable/strict/1761711983 2025-11-03T16:58:01.5929943Z * [new tag] viable/strict/1761713514 -> viable/strict/1761713514 2025-11-03T16:58:01.5931315Z * [new tag] viable/strict/1761715523 -> viable/strict/1761715523 2025-11-03T16:58:01.5932395Z * [new tag] viable/strict/1761727973 -> viable/strict/1761727973 2025-11-03T16:58:01.5933897Z * [new tag] viable/strict/1761751558 -> viable/strict/1761751558 2025-11-03T16:58:01.5935439Z * [new tag] viable/strict/1761755187 -> viable/strict/1761755187 2025-11-03T16:58:01.5936367Z * [new tag] viable/strict/1761756826 -> viable/strict/1761756826 2025-11-03T16:58:01.5937992Z * [new tag] viable/strict/1761769551 -> viable/strict/1761769551 2025-11-03T16:58:01.5938985Z * [new tag] viable/strict/1761771032 -> viable/strict/1761771032 2025-11-03T16:58:01.5940435Z * [new tag] viable/strict/1761773101 -> viable/strict/1761773101 2025-11-03T16:58:01.5941905Z * [new tag] viable/strict/1761781792 -> viable/strict/1761781792 2025-11-03T16:58:01.5942930Z * [new tag] viable/strict/1761784788 -> viable/strict/1761784788 2025-11-03T16:58:01.5944400Z * [new tag] viable/strict/1761786740 -> viable/strict/1761786740 2025-11-03T16:58:01.5945632Z * [new tag] viable/strict/1761789332 -> viable/strict/1761789332 2025-11-03T16:58:01.5946981Z * [new tag] viable/strict/1761792569 -> viable/strict/1761792569 2025-11-03T16:58:01.5948044Z * [new tag] viable/strict/1761795289 -> viable/strict/1761795289 2025-11-03T16:58:01.5949636Z * [new tag] viable/strict/1761798345 -> viable/strict/1761798345 2025-11-03T16:58:01.5950633Z * [new tag] viable/strict/1761799827 -> viable/strict/1761799827 2025-11-03T16:58:01.5952407Z * [new tag] viable/strict/1761805604 -> viable/strict/1761805604 2025-11-03T16:58:01.5953414Z * [new tag] viable/strict/1761807202 -> viable/strict/1761807202 2025-11-03T16:58:01.5954968Z * [new tag] viable/strict/1761809094 -> viable/strict/1761809094 2025-11-03T16:58:01.5955983Z * [new tag] viable/strict/1761810576 -> viable/strict/1761810576 2025-11-03T16:58:01.5957520Z * [new tag] viable/strict/1761812771 -> viable/strict/1761812771 2025-11-03T16:58:01.5958587Z * [new tag] viable/strict/1761814363 -> viable/strict/1761814363 2025-11-03T16:58:01.5960047Z * [new tag] viable/strict/1761857410 -> viable/strict/1761857410 2025-11-03T16:58:01.5961279Z * [new tag] viable/strict/1761860985 -> viable/strict/1761860985 2025-11-03T16:58:01.5962731Z * [new tag] viable/strict/1761863094 -> viable/strict/1761863094 2025-11-03T16:58:01.5964069Z * [new tag] viable/strict/1761864590 -> viable/strict/1761864590 2025-11-03T16:58:01.5965143Z * [new tag] viable/strict/1761866675 -> viable/strict/1761866675 2025-11-03T16:58:01.5966590Z * [new tag] viable/strict/1761868178 -> viable/strict/1761868178 2025-11-03T16:58:01.5967912Z * [new tag] viable/strict/1761871111 -> viable/strict/1761871111 2025-11-03T16:58:01.5969327Z * [new tag] viable/strict/1761873126 -> viable/strict/1761873126 2025-11-03T16:58:01.5970407Z * [new tag] viable/strict/1761875714 -> viable/strict/1761875714 2025-11-03T16:58:01.5971858Z * [new tag] viable/strict/1761878924 -> viable/strict/1761878924 2025-11-03T16:58:01.5973175Z * [new tag] viable/strict/1761881727 -> viable/strict/1761881727 2025-11-03T16:58:01.5974394Z * [new tag] viable/strict/1761882959 -> viable/strict/1761882959 2025-11-03T16:58:01.5975752Z * [new tag] viable/strict/1761886268 -> viable/strict/1761886268 2025-11-03T16:58:01.5976959Z * [new tag] viable/strict/1761893641 -> viable/strict/1761893641 2025-11-03T16:58:01.5978402Z * [new tag] viable/strict/1761931517 -> viable/strict/1761931517 2025-11-03T16:58:01.5979422Z * [new tag] viable/strict/1761933080 -> viable/strict/1761933080 2025-11-03T16:58:01.5980998Z * [new tag] viable/strict/1761935217 -> viable/strict/1761935217 2025-11-03T16:58:01.5982132Z * [new tag] viable/strict/1761938533 -> viable/strict/1761938533 2025-11-03T16:58:01.5983676Z * [new tag] viable/strict/1761940184 -> viable/strict/1761940184 2025-11-03T16:58:01.5984717Z * [new tag] viable/strict/1761942338 -> viable/strict/1761942338 2025-11-03T16:58:01.5986271Z * [new tag] viable/strict/1761946100 -> viable/strict/1761946100 2025-11-03T16:58:01.5987261Z * [new tag] viable/strict/1761947374 -> viable/strict/1761947374 2025-11-03T16:58:01.5988752Z * [new tag] viable/strict/1761950978 -> viable/strict/1761950978 2025-11-03T16:58:01.5990160Z * [new tag] viable/strict/1761957727 -> viable/strict/1761957727 2025-11-03T16:58:01.5991200Z * [new tag] viable/strict/1761959532 -> viable/strict/1761959532 2025-11-03T16:58:01.5992661Z * [new tag] viable/strict/1761965366 -> viable/strict/1761965366 2025-11-03T16:58:01.5993881Z * [new tag] viable/strict/1761968066 -> viable/strict/1761968066 2025-11-03T16:58:01.5995254Z * [new tag] viable/strict/1761969322 -> viable/strict/1761969322 2025-11-03T16:58:01.5996323Z * [new tag] viable/strict/1761974723 -> viable/strict/1761974723 2025-11-03T16:58:01.5997796Z * [new tag] viable/strict/1761981837 -> viable/strict/1761981837 2025-11-03T16:58:01.5998961Z * [new tag] viable/strict/1761985546 -> viable/strict/1761985546 2025-11-03T16:58:01.6000443Z * [new tag] viable/strict/1761987030 -> viable/strict/1761987030 2025-11-03T16:58:01.6001500Z * [new tag] viable/strict/1762003554 -> viable/strict/1762003554 2025-11-03T16:58:01.6002939Z * [new tag] viable/strict/1762021560 -> viable/strict/1762021560 2025-11-03T16:58:01.6004269Z * [new tag] viable/strict/1762032190 -> viable/strict/1762032190 2025-11-03T16:58:01.6005611Z * [new tag] viable/strict/1762040981 -> viable/strict/1762040981 2025-11-03T16:58:01.6006651Z * [new tag] viable/strict/1762048525 -> viable/strict/1762048525 2025-11-03T16:58:01.6008193Z * [new tag] viable/strict/1762104223 -> viable/strict/1762104223 2025-11-03T16:58:01.6010155Z * [new tag] viable/strict/1762105778 -> viable/strict/1762105778 2025-11-03T16:58:01.6011657Z * [new tag] viable/strict/1762115109 -> viable/strict/1762115109 2025-11-03T16:58:01.6012600Z * [new tag] viable/strict/1762125840 -> viable/strict/1762125840 2025-11-03T16:58:01.6014165Z * [new tag] viable/strict/1762127377 -> viable/strict/1762127377 2025-11-03T16:58:01.6015164Z * [new tag] viable/strict/1762134925 -> viable/strict/1762134925 2025-11-03T16:58:01.6016764Z * [new tag] viable/strict/1762138338 -> viable/strict/1762138338 2025-11-03T16:58:01.6017784Z * [new tag] viable/strict/1762148993 -> viable/strict/1762148993 2025-11-03T16:58:01.6019243Z * [new tag] viable/strict/1762152871 -> viable/strict/1762152871 2025-11-03T16:58:01.6020475Z * [new tag] viable/strict/1762156183 -> viable/strict/1762156183 2025-11-03T16:58:01.6021858Z * [new tag] viable/strict/1762163457 -> viable/strict/1762163457 2025-11-03T16:58:01.6023132Z * [new tag] viable/strict/1762165569 -> viable/strict/1762165569 2025-11-03T16:58:01.6024337Z * [new tag] viable/strict/1762169035 -> viable/strict/1762169035 2025-11-03T16:58:01.6025717Z * [new tag] viable/strict/1762174936 -> viable/strict/1762174936 2025-11-03T16:58:01.6027123Z * [new tag] whc_flight_1 -> whc_flight_1 2025-11-03T16:58:01.6028564Z * [new tag] whc_flight_2 -> whc_flight_2 2025-11-03T16:58:01.6030149Z * [new tag] whc_flight_4 -> whc_flight_4 2025-11-03T16:58:01.7094115Z [command]/usr/bin/git rev-parse --verify --quiet 3f6538febd727b782e6e13cfd026a309fb14351d^{object} 2025-11-03T16:58:01.7131824Z 3f6538febd727b782e6e13cfd026a309fb14351d 2025-11-03T16:58:01.7137277Z ##[endgroup] 2025-11-03T16:58:01.7138139Z ##[group]Determining the checkout info 2025-11-03T16:58:01.7139197Z ##[endgroup] 2025-11-03T16:58:01.7143017Z [command]/usr/bin/git sparse-checkout disable 2025-11-03T16:58:01.7193075Z [command]/usr/bin/git config --local --unset-all extensions.worktreeConfig 2025-11-03T16:58:01.7226287Z ##[group]Checking out the ref 2025-11-03T16:58:01.7229905Z [command]/usr/bin/git checkout --progress --force 3f6538febd727b782e6e13cfd026a309fb14351d 2025-11-03T16:58:02.7605971Z Updating files: 74% (14861/20004) 2025-11-03T16:58:02.8067768Z Updating files: 75% (15003/20004) 2025-11-03T16:58:02.8261304Z Updating files: 76% (15204/20004) 2025-11-03T16:58:02.8421718Z Updating files: 77% (15404/20004) 2025-11-03T16:58:02.8596118Z Updating files: 78% (15604/20004) 2025-11-03T16:58:02.8883589Z Updating files: 79% (15804/20004) 2025-11-03T16:58:02.9180441Z Updating files: 80% (16004/20004) 2025-11-03T16:58:02.9467416Z Updating files: 81% (16204/20004) 2025-11-03T16:58:02.9712444Z Updating files: 82% (16404/20004) 2025-11-03T16:58:02.9895953Z Updating files: 83% (16604/20004) 2025-11-03T16:58:03.0051721Z Updating files: 84% (16804/20004) 2025-11-03T16:58:03.0235325Z Updating files: 85% (17004/20004) 2025-11-03T16:58:03.0411680Z Updating files: 86% (17204/20004) 2025-11-03T16:58:03.0567945Z Updating files: 87% (17404/20004) 2025-11-03T16:58:03.0712002Z Updating files: 88% (17604/20004) 2025-11-03T16:58:03.0865025Z Updating files: 89% (17804/20004) 2025-11-03T16:58:03.1051443Z Updating files: 90% (18004/20004) 2025-11-03T16:58:03.1198582Z Updating files: 91% (18204/20004) 2025-11-03T16:58:03.1360491Z Updating files: 92% (18404/20004) 2025-11-03T16:58:03.1562084Z Updating files: 93% (18604/20004) 2025-11-03T16:58:03.1769357Z Updating files: 94% (18804/20004) 2025-11-03T16:58:03.1965113Z Updating files: 95% (19004/20004) 2025-11-03T16:58:03.2143103Z Updating files: 96% (19204/20004) 2025-11-03T16:58:03.2324304Z Updating files: 97% (19404/20004) 2025-11-03T16:58:03.2593045Z Updating files: 98% (19604/20004) 2025-11-03T16:58:03.2784148Z Updating files: 99% (19804/20004) 2025-11-03T16:58:03.2784677Z Updating files: 100% (20004/20004) 2025-11-03T16:58:03.2785196Z Updating files: 100% (20004/20004), done. 2025-11-03T16:58:03.3020879Z Note: switching to '3f6538febd727b782e6e13cfd026a309fb14351d'. 2025-11-03T16:58:03.3021431Z 2025-11-03T16:58:03.3021845Z You are in 'detached HEAD' state. You can look around, make experimental 2025-11-03T16:58:03.3022760Z changes and commit them, and you can discard any commits you make in this 2025-11-03T16:58:03.3023665Z state without impacting any branches by switching back to a branch. 2025-11-03T16:58:03.3024198Z 2025-11-03T16:58:03.3024546Z If you want to create a new branch to retain commits you create, you may 2025-11-03T16:58:03.3025397Z do so (now or later) by using -c with the switch command. Example: 2025-11-03T16:58:03.3025885Z 2025-11-03T16:58:03.3026088Z git switch -c 2025-11-03T16:58:03.3026421Z 2025-11-03T16:58:03.3026605Z Or undo this operation with: 2025-11-03T16:58:03.3026919Z 2025-11-03T16:58:03.3027071Z git switch - 2025-11-03T16:58:03.3027330Z 2025-11-03T16:58:03.3027735Z Turn off this advice by setting config variable advice.detachedHead to false 2025-11-03T16:58:03.3028338Z 2025-11-03T16:58:03.3028742Z HEAD is now at 3f6538febd7 Remove tools from BC linter (#166858) 2025-11-03T16:58:03.3153698Z ##[endgroup] 2025-11-03T16:58:03.3154450Z ##[group]Setting up auth for fetching submodules 2025-11-03T16:58:03.3164662Z [command]/usr/bin/git config --global http.https://github.com/.extraheader AUTHORIZATION: basic *** 2025-11-03T16:58:03.3226676Z [command]/usr/bin/git config --global --unset-all url.https://github.com/.insteadOf 2025-11-03T16:58:03.3272272Z [command]/usr/bin/git config --global --add url.https://github.com/.insteadOf git@github.com: 2025-11-03T16:58:03.3315312Z [command]/usr/bin/git config --global --add url.https://github.com/.insteadOf org-21003710@github.com: 2025-11-03T16:58:03.3351247Z ##[endgroup] 2025-11-03T16:58:03.3351969Z ##[group]Fetching submodules 2025-11-03T16:58:03.3353793Z [command]/usr/bin/git submodule sync --recursive 2025-11-03T16:58:03.3836089Z [command]/usr/bin/git -c protocol.version=2 submodule update --init --force --recursive 2025-11-03T16:58:03.4300243Z Submodule 'android/libs/fbjni' (https://github.com/facebookincubator/fbjni.git) registered for path 'android/libs/fbjni' 2025-11-03T16:58:03.4302248Z Submodule 'third_party/NNPACK_deps/FP16' (https://github.com/Maratyszcza/FP16.git) registered for path 'third_party/FP16' 2025-11-03T16:58:03.4306383Z Submodule 'third_party/NNPACK_deps/FXdiv' (https://github.com/Maratyszcza/FXdiv.git) registered for path 'third_party/FXdiv' 2025-11-03T16:58:03.4309873Z Submodule 'third_party/NNPACK' (https://github.com/Maratyszcza/NNPACK.git) registered for path 'third_party/NNPACK' 2025-11-03T16:58:03.4313574Z Submodule 'third_party/NVTX' (https://github.com/NVIDIA/NVTX.git) registered for path 'third_party/NVTX' 2025-11-03T16:58:03.4317556Z Submodule 'third_party/VulkanMemoryAllocator' (https://github.com/GPUOpen-LibrariesAndSDKs/VulkanMemoryAllocator.git) registered for path 'third_party/VulkanMemoryAllocator' 2025-11-03T16:58:03.4321302Z Submodule 'third_party/XNNPACK' (https://github.com/google/XNNPACK.git) registered for path 'third_party/XNNPACK' 2025-11-03T16:58:03.4325308Z Submodule 'third_party/aiter' (https://github.com/ROCm/aiter.git) registered for path 'third_party/aiter' 2025-11-03T16:58:03.4330763Z Submodule 'third_party/benchmark' (https://github.com/google/benchmark.git) registered for path 'third_party/benchmark' 2025-11-03T16:58:03.4335142Z Submodule 'third_party/composable_kernel' (https://github.com/ROCm/composable_kernel.git) registered for path 'third_party/composable_kernel' 2025-11-03T16:58:03.4339483Z Submodule 'third_party/cpp-httplib' (https://github.com/yhirose/cpp-httplib.git) registered for path 'third_party/cpp-httplib' 2025-11-03T16:58:03.4343590Z Submodule 'third_party/cpuinfo' (https://github.com/pytorch/cpuinfo.git) registered for path 'third_party/cpuinfo' 2025-11-03T16:58:03.4348086Z Submodule 'third_party/cudnn_frontend' (https://github.com/NVIDIA/cudnn-frontend.git) registered for path 'third_party/cudnn_frontend' 2025-11-03T16:58:03.4353637Z Submodule 'third_party/cutlass' (https://github.com/NVIDIA/cutlass.git) registered for path 'third_party/cutlass' 2025-11-03T16:58:03.4358114Z Submodule 'third_party/fbgemm' (https://github.com/pytorch/fbgemm) registered for path 'third_party/fbgemm' 2025-11-03T16:58:03.4362937Z Submodule 'third_party/flash-attention' (https://github.com/Dao-AILab/flash-attention.git) registered for path 'third_party/flash-attention' 2025-11-03T16:58:03.4370670Z Submodule 'third_party/flatbuffers' (https://github.com/google/flatbuffers.git) registered for path 'third_party/flatbuffers' 2025-11-03T16:58:03.4375313Z Submodule 'third_party/fmt' (https://github.com/fmtlib/fmt.git) registered for path 'third_party/fmt' 2025-11-03T16:58:03.4380583Z Submodule 'third_party/gemmlowp/gemmlowp' (https://github.com/google/gemmlowp.git) registered for path 'third_party/gemmlowp/gemmlowp' 2025-11-03T16:58:03.4385005Z Submodule 'third_party/gloo' (https://github.com/pytorch/gloo) registered for path 'third_party/gloo' 2025-11-03T16:58:03.4391191Z Submodule 'third_party/googletest' (https://github.com/google/googletest.git) registered for path 'third_party/googletest' 2025-11-03T16:58:03.4396175Z Submodule 'third_party/ideep' (https://github.com/intel/ideep) registered for path 'third_party/ideep' 2025-11-03T16:58:03.4401134Z Submodule 'third_party/ittapi' (https://github.com/intel/ittapi.git) registered for path 'third_party/ittapi' 2025-11-03T16:58:03.4406462Z Submodule 'third_party/kineto' (https://github.com/pytorch/kineto) registered for path 'third_party/kineto' 2025-11-03T16:58:03.4412577Z Submodule 'third_party/kleidiai' (https://github.com/ARM-software/kleidiai.git) registered for path 'third_party/kleidiai' 2025-11-03T16:58:03.4417814Z Submodule 'third_party/mimalloc' (https://github.com/microsoft/mimalloc.git) registered for path 'third_party/mimalloc' 2025-11-03T16:58:03.4423974Z Submodule 'third_party/nlohmann' (https://github.com/nlohmann/json.git) registered for path 'third_party/nlohmann' 2025-11-03T16:58:03.4429456Z Submodule 'third_party/onnx' (https://github.com/onnx/onnx.git) registered for path 'third_party/onnx' 2025-11-03T16:58:03.4435553Z Submodule 'third_party/opentelemetry-cpp' (https://github.com/open-telemetry/opentelemetry-cpp.git) registered for path 'third_party/opentelemetry-cpp' 2025-11-03T16:58:03.4440994Z Submodule 'third_party/pocketfft' (https://github.com/mreineck/pocketfft) registered for path 'third_party/pocketfft' 2025-11-03T16:58:03.4446778Z Submodule 'third_party/protobuf' (https://github.com/protocolbuffers/protobuf.git) registered for path 'third_party/protobuf' 2025-11-03T16:58:03.4452898Z Submodule 'third_party/NNPACK_deps/psimd' (https://github.com/Maratyszcza/psimd.git) registered for path 'third_party/psimd' 2025-11-03T16:58:03.4458973Z Submodule 'third_party/NNPACK_deps/pthreadpool' (https://github.com/Maratyszcza/pthreadpool.git) registered for path 'third_party/pthreadpool' 2025-11-03T16:58:03.4468812Z Submodule 'third_party/pybind11' (https://github.com/pybind/pybind11.git) registered for path 'third_party/pybind11' 2025-11-03T16:58:03.4475065Z Submodule 'third_party/python-peachpy' (https://github.com/malfet/PeachPy.git) registered for path 'third_party/python-peachpy' 2025-11-03T16:58:03.4480863Z Submodule 'third_party/sleef' (https://github.com/shibatch/sleef) registered for path 'third_party/sleef' 2025-11-03T16:58:03.4487192Z Submodule 'third_party/tensorpipe' (https://github.com/pytorch/tensorpipe.git) registered for path 'third_party/tensorpipe' 2025-11-03T16:58:03.4531719Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/android/libs/fbjni'... 2025-11-03T16:58:03.7530740Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/psimd'... 2025-11-03T16:58:03.7531881Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/FXdiv'... 2025-11-03T16:58:03.7532700Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/FP16'... 2025-11-03T16:58:03.7533345Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/pocketfft'... 2025-11-03T16:58:03.7533994Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/pthreadpool'... 2025-11-03T16:58:03.7534638Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/NNPACK'... 2025-11-03T16:58:03.8076778Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/NVTX'... 2025-11-03T16:58:03.9406104Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/python-peachpy'... 2025-11-03T16:58:03.9406982Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/gloo'... 2025-11-03T16:58:03.9407535Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/ideep'... 2025-11-03T16:58:03.9408369Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/gemmlowp/gemmlowp'... 2025-11-03T16:58:03.9409010Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/benchmark'... 2025-11-03T16:58:03.9555021Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/VulkanMemoryAllocator'... 2025-11-03T16:58:05.0595140Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/tensorpipe'... 2025-11-03T16:58:05.0595939Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/ittapi'... 2025-11-03T16:58:05.0596631Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/kleidiai'... 2025-11-03T16:58:05.0597573Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/cpuinfo'... 2025-11-03T16:58:05.0598173Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/flash-attention'... 2025-11-03T16:58:05.0598862Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/cpp-httplib'... 2025-11-03T16:58:05.0599604Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/sleef'... 2025-11-03T16:58:05.0600123Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/mimalloc'... 2025-11-03T16:58:05.0600763Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/googletest'... 2025-11-03T16:58:05.0601421Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/pybind11'... 2025-11-03T16:58:05.0601919Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/fmt'... 2025-11-03T16:58:05.0602533Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/cudnn_frontend'... 2025-11-03T16:58:05.0603475Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/kineto'... 2025-11-03T16:58:05.1596031Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/XNNPACK'... 2025-11-03T16:58:15.3843969Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/flatbuffers'... 2025-11-03T16:58:15.3845279Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/onnx'... 2025-11-03T16:58:15.3846429Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/cutlass'... 2025-11-03T16:58:15.3847566Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/fbgemm'... 2025-11-03T16:58:15.3849014Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/composable_kernel'... 2025-11-03T16:58:15.3850448Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/aiter'... 2025-11-03T16:58:15.3851930Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/opentelemetry-cpp'... 2025-11-03T16:58:15.3853416Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/nlohmann'... 2025-11-03T16:58:15.3854552Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/protobuf'... 2025-11-03T16:58:15.4077046Z Submodule path 'android/libs/fbjni': checked out '7e1e1fe3858c63c251c637ae41a20de425dde96f' 2025-11-03T16:58:15.4283220Z Submodule path 'third_party/FP16': checked out '4dfe081cf6bcd15db339cf2680b9281b8451eeb3' 2025-11-03T16:58:15.4438061Z Submodule path 'third_party/FXdiv': checked out 'b408327ac2a15ec3e43352421954f5b1967701d1' 2025-11-03T16:58:15.4795074Z Submodule path 'third_party/NNPACK': checked out 'c07e3a0400713d546e0dea2d5466dd22ea389c73' 2025-11-03T16:58:15.5703717Z Submodule path 'third_party/NVTX': checked out '2942f167cc30c5e3a44a2aecd5b0d9c07ff61a07' 2025-11-03T16:58:15.6297592Z Submodule path 'third_party/VulkanMemoryAllocator': checked out '1d8f600fd424278486eade7ed3e877c99f0846b1' 2025-11-03T16:58:16.5667380Z Submodule path 'third_party/XNNPACK': checked out '51a0103656eff6fc9bfd39a4597923c4b542c883' 2025-11-03T16:58:16.7505117Z Submodule path 'third_party/aiter': checked out '01aae101b9e5e94d6c16a9514c9fb8df99c93150' 2025-11-03T16:58:16.7539565Z Submodule '3rdparty/composable_kernel' (https://github.com/ROCm/composable_kernel.git) registered for path 'third_party/aiter/3rdparty/composable_kernel' 2025-11-03T16:58:16.7585503Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/aiter/3rdparty/composable_kernel'... 2025-11-03T16:58:20.3112613Z Submodule path 'third_party/aiter/3rdparty/composable_kernel': checked out 'cffe8fa2a442ac8e80dd236a1a5d24fe3d7e0cbf' 2025-11-03T16:58:20.3439799Z Submodule path 'third_party/benchmark': checked out '299e5928955cc62af9968370293b916f5130916f' 2025-11-03T16:58:20.7755834Z Submodule path 'third_party/composable_kernel': checked out '7fe50dc3da2069d6645d9deb8c017a876472a977' 2025-11-03T16:58:20.8320731Z Submodule path 'third_party/cpp-httplib': checked out '89c932f313c6437c38f2982869beacc89c2f2246' 2025-11-03T16:58:20.9401819Z Submodule path 'third_party/cpuinfo': checked out '5e3d2445e6a84d9599bee2bf78edbb4d80865e1d' 2025-11-03T16:58:20.9978925Z Submodule path 'third_party/cudnn_frontend': checked out '0b1577c8c83401237d601d0d0db5210506705396' 2025-11-03T16:58:21.7394154Z Submodule path 'third_party/cutlass': checked out 'f3fde58372d33e9a5650ba7b80fc48b3b49d40c8' 2025-11-03T16:58:21.9250168Z Submodule path 'third_party/fbgemm': checked out 'c0b988d39a9e47c794d699f29930ed4d7c7e13a4' 2025-11-03T16:58:21.9286811Z Submodule 'external/asmjit' (https://github.com/asmjit/asmjit.git) registered for path 'third_party/fbgemm/external/asmjit' 2025-11-03T16:58:21.9289165Z Submodule 'external/composable_kernel' (https://github.com/ROCm/composable_kernel.git) registered for path 'third_party/fbgemm/external/composable_kernel' 2025-11-03T16:58:21.9293918Z Submodule 'external/cpuinfo' (https://github.com/pytorch/cpuinfo) registered for path 'third_party/fbgemm/external/cpuinfo' 2025-11-03T16:58:21.9297001Z Submodule 'external/cutlass' (https://github.com/jwfromm/cutlass) registered for path 'third_party/fbgemm/external/cutlass' 2025-11-03T16:58:21.9300728Z Submodule 'external/googletest' (https://github.com/google/googletest) registered for path 'third_party/fbgemm/external/googletest' 2025-11-03T16:58:21.9304971Z Submodule 'external/hipify_torch' (https://github.com/ROCmSoftwarePlatform/hipify_torch.git) registered for path 'third_party/fbgemm/external/hipify_torch' 2025-11-03T16:58:21.9310020Z Submodule 'external/json' (https://github.com/nlohmann/json.git) registered for path 'third_party/fbgemm/external/json' 2025-11-03T16:58:21.9353395Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/fbgemm/external/asmjit'... 2025-11-03T16:58:22.9911537Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/fbgemm/external/hipify_torch'... 2025-11-03T16:58:22.9913268Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/fbgemm/external/cpuinfo'... 2025-11-03T16:58:22.9914821Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/fbgemm/external/googletest'... 2025-11-03T16:58:23.0911548Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/fbgemm/external/composable_kernel'... 2025-11-03T16:58:25.1877921Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/fbgemm/external/cutlass'... 2025-11-03T16:58:25.2876411Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/fbgemm/external/json'... 2025-11-03T16:58:29.7823615Z Submodule path 'third_party/fbgemm/external/asmjit': checked out 'a3199e8857792cd10b7589ff5d58343d2c9008ea' 2025-11-03T16:58:30.2190608Z Submodule path 'third_party/fbgemm/external/composable_kernel': checked out '7fe50dc3da2069d6645d9deb8c017a876472a977' 2025-11-03T16:58:30.3286395Z Submodule path 'third_party/fbgemm/external/cpuinfo': checked out '6543fec09b2f04ac4a666882998b534afc9c1349' 2025-11-03T16:58:31.0630437Z Submodule path 'third_party/fbgemm/external/cutlass': checked out '98125ce499b0fdf7ffbe0e3052f5b8709f4840f8' 2025-11-03T16:58:31.1162814Z Submodule path 'third_party/fbgemm/external/googletest': checked out '52eb8108c5bdec04579160ae17225d66034bd723' 2025-11-03T16:58:31.1344773Z Submodule path 'third_party/fbgemm/external/hipify_torch': checked out '63b6a7b541fa7f08f8475ca7d74054db36ff2691' 2025-11-03T16:58:31.2649505Z Submodule path 'third_party/fbgemm/external/json': checked out '9cca280a4d0ccf0c08f47a99aa71d1b0e52f8d03' 2025-11-03T16:58:31.3576557Z Submodule path 'third_party/flash-attention': checked out '979702c87a8713a8e0a5e9fee122b90d2ef13be5' 2025-11-03T16:58:31.3606094Z Submodule 'csrc/composable_kernel' (https://github.com/ROCm/composable_kernel.git) registered for path 'third_party/flash-attention/csrc/composable_kernel' 2025-11-03T16:58:31.3610364Z Submodule 'csrc/cutlass' (https://github.com/NVIDIA/cutlass.git) registered for path 'third_party/flash-attention/csrc/cutlass' 2025-11-03T16:58:31.3649502Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/flash-attention/csrc/composable_kernel'... 2025-11-03T16:58:34.6765872Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/flash-attention/csrc/cutlass'... 2025-11-03T16:58:34.9920012Z Submodule path 'third_party/flash-attention/csrc/composable_kernel': checked out '888317e698e9803c62bd38568abc9e05d7709f33' 2025-11-03T16:58:35.6503555Z Submodule path 'third_party/flash-attention/csrc/cutlass': checked out 'c506e16788cb08416a4a57e11a9067beeee29420' 2025-11-03T16:58:35.8292817Z Submodule path 'third_party/flatbuffers': checked out 'a2cd1ea3b6d3fee220106b5fed3f7ce8da9eb757' 2025-11-03T16:58:35.8660639Z Submodule path 'third_party/fmt': checked out 'e424e3f2e607da02742f73db84873b8084fc714c' 2025-11-03T16:58:35.9116667Z Submodule path 'third_party/gemmlowp/gemmlowp': checked out '3fb5c176c17c765a3492cd2f0321b0dab712f350' 2025-11-03T16:58:35.9468748Z Submodule path 'third_party/gloo': checked out '54cbae0d3a67fa890b4c3d9ee162b7860315e341' 2025-11-03T16:58:35.9993137Z Submodule path 'third_party/googletest': checked out '52eb8108c5bdec04579160ae17225d66034bd723' 2025-11-03T16:58:36.0191371Z Submodule path 'third_party/ideep': checked out '719d8e6cd7f7a0e01b155657526d693acf97c2b3' 2025-11-03T16:58:36.0213305Z Submodule 'mkl-dnn' (https://github.com/intel/mkl-dnn.git) registered for path 'third_party/ideep/mkl-dnn' 2025-11-03T16:58:36.0259554Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/ideep/mkl-dnn'... 2025-11-03T16:58:50.7991101Z Submodule path 'third_party/ideep/mkl-dnn': checked out '8d263e693366ef8db40acc569cc7d8edf644556d' 2025-11-03T16:58:50.8277649Z Submodule path 'third_party/ittapi': checked out 'dec1d23ca65ab069d225dfe40dea14f455170959' 2025-11-03T16:58:50.9205165Z Submodule path 'third_party/kineto': checked out '6fcbc53d33dd275c0aba1e5d7701d471b7f6eeb3' 2025-11-03T16:58:50.9233301Z Submodule 'libkineto/third_party/dynolog' (https://github.com/facebookincubator/dynolog.git) registered for path 'third_party/kineto/libkineto/third_party/dynolog' 2025-11-03T16:58:50.9239119Z Submodule 'libkineto/third_party/fmt' (https://github.com/fmtlib/fmt.git) registered for path 'third_party/kineto/libkineto/third_party/fmt' 2025-11-03T16:58:50.9241445Z Submodule 'libkineto/third_party/googletest' (https://github.com/google/googletest.git) registered for path 'third_party/kineto/libkineto/third_party/googletest' 2025-11-03T16:58:50.9281713Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/kineto/libkineto/third_party/dynolog'... 2025-11-03T16:58:51.6148234Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/kineto/libkineto/third_party/fmt'... 2025-11-03T16:58:51.9263802Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/kineto/libkineto/third_party/googletest'... 2025-11-03T16:58:52.0287560Z Submodule path 'third_party/kineto/libkineto/third_party/dynolog': checked out 'd2ffe0a4e3acace628db49974246b66fc3e85fb1' 2025-11-03T16:58:52.0315266Z Submodule 'third_party/DCGM' (https://github.com/NVIDIA/DCGM.git) registered for path 'third_party/kineto/libkineto/third_party/dynolog/third_party/DCGM' 2025-11-03T16:58:52.0317761Z Submodule 'third_party/cpr' (https://github.com/libcpr/cpr.git) registered for path 'third_party/kineto/libkineto/third_party/dynolog/third_party/cpr' 2025-11-03T16:58:52.0322454Z Submodule 'third_party/fmt' (https://github.com/fmtlib/fmt.git) registered for path 'third_party/kineto/libkineto/third_party/dynolog/third_party/fmt' 2025-11-03T16:58:52.0326369Z Submodule 'third_party/gflags' (https://github.com/gflags/gflags.git) registered for path 'third_party/kineto/libkineto/third_party/dynolog/third_party/gflags' 2025-11-03T16:58:52.0331309Z Submodule 'third_party/glog' (https://github.com/google/glog.git) registered for path 'third_party/kineto/libkineto/third_party/dynolog/third_party/glog' 2025-11-03T16:58:52.0337153Z Submodule 'third_party/googletest' (https://github.com/google/googletest.git) registered for path 'third_party/kineto/libkineto/third_party/dynolog/third_party/googletest' 2025-11-03T16:58:52.0341274Z Submodule 'third_party/json' (https://github.com/nlohmann/json.git) registered for path 'third_party/kineto/libkineto/third_party/dynolog/third_party/json' 2025-11-03T16:58:52.0345555Z Submodule 'third_party/pfs' (https://github.com/dtrugman/pfs.git) registered for path 'third_party/kineto/libkineto/third_party/dynolog/third_party/pfs' 2025-11-03T16:58:52.0350322Z Submodule 'third_party/prometheus-cpp' (https://github.com/jupp0r/prometheus-cpp.git) registered for path 'third_party/kineto/libkineto/third_party/dynolog/third_party/prometheus-cpp' 2025-11-03T16:58:52.0395751Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/kineto/libkineto/third_party/dynolog/third_party/DCGM'... 2025-11-03T16:58:53.9866027Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/kineto/libkineto/third_party/dynolog/third_party/pfs'... 2025-11-03T16:58:53.9867510Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/kineto/libkineto/third_party/dynolog/third_party/gflags'... 2025-11-03T16:58:53.9868501Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/kineto/libkineto/third_party/dynolog/third_party/cpr'... 2025-11-03T16:58:53.9869529Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/kineto/libkineto/third_party/dynolog/third_party/prometheus-cpp'... 2025-11-03T16:58:53.9870532Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/kineto/libkineto/third_party/dynolog/third_party/glog'... 2025-11-03T16:58:53.9871506Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/kineto/libkineto/third_party/dynolog/third_party/googletest'... 2025-11-03T16:58:53.9872495Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/kineto/libkineto/third_party/dynolog/third_party/fmt'... 2025-11-03T16:58:54.0866042Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/kineto/libkineto/third_party/dynolog/third_party/json'... 2025-11-03T16:58:58.1548173Z Submodule path 'third_party/kineto/libkineto/third_party/dynolog/third_party/DCGM': checked out 'ffde4e54bc7249a6039a5e6b45b395141e1217f9' 2025-11-03T16:58:58.1824820Z Submodule path 'third_party/kineto/libkineto/third_party/dynolog/third_party/cpr': checked out '871ed52d350214a034f6ef8a3b8f51c5ce1bd400' 2025-11-03T16:58:58.2276127Z Submodule path 'third_party/kineto/libkineto/third_party/dynolog/third_party/fmt': checked out 'cd4af11efc9c622896a3e4cb599fa28668ca3d05' 2025-11-03T16:58:58.2483795Z Submodule path 'third_party/kineto/libkineto/third_party/dynolog/third_party/gflags': checked out 'e171aa2d15ed9eb17054558e0b3a6a413bb01067' 2025-11-03T16:58:58.2515073Z Submodule 'doc' (https://github.com/gflags/gflags.git) registered for path 'third_party/kineto/libkineto/third_party/dynolog/third_party/gflags/doc' 2025-11-03T16:58:58.2560639Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/kineto/libkineto/third_party/dynolog/third_party/gflags/doc'... 2025-11-03T16:58:58.5556842Z Submodule path 'third_party/kineto/libkineto/third_party/dynolog/third_party/gflags/doc': checked out '8411df715cf522606e3b1aca386ddfc0b63d34b4' 2025-11-03T16:58:58.5843386Z Submodule path 'third_party/kineto/libkineto/third_party/dynolog/third_party/glog': checked out 'b33e3bad4c46c8a6345525fd822af355e5ef9446' 2025-11-03T16:58:58.6393790Z Submodule path 'third_party/kineto/libkineto/third_party/dynolog/third_party/googletest': checked out '52eb8108c5bdec04579160ae17225d66034bd723' 2025-11-03T16:58:58.7588377Z Submodule path 'third_party/kineto/libkineto/third_party/dynolog/third_party/json': checked out '4f8fba14066156b73f1189a2b8bd568bde5284c5' 2025-11-03T16:58:58.7827450Z Submodule path 'third_party/kineto/libkineto/third_party/dynolog/third_party/pfs': checked out 'f68a2fa8ea36c783bdd760371411fcb495aa3150' 2025-11-03T16:58:58.8089926Z Submodule path 'third_party/kineto/libkineto/third_party/dynolog/third_party/prometheus-cpp': checked out 'b1234816facfdda29845c46696a02998a4af115a' 2025-11-03T16:58:58.8118204Z Submodule 'civetweb' (https://github.com/civetweb/civetweb.git) registered for path 'third_party/kineto/libkineto/third_party/dynolog/third_party/prometheus-cpp/3rdparty/civetweb' 2025-11-03T16:58:58.8120871Z Submodule 'googletest' (https://github.com/google/googletest.git) registered for path 'third_party/kineto/libkineto/third_party/dynolog/third_party/prometheus-cpp/3rdparty/googletest' 2025-11-03T16:58:58.8162527Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/kineto/libkineto/third_party/dynolog/third_party/prometheus-cpp/3rdparty/civetweb'... 2025-11-03T16:59:00.6779084Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/kineto/libkineto/third_party/dynolog/third_party/prometheus-cpp/3rdparty/googletest'... 2025-11-03T16:59:00.9431221Z Submodule path 'third_party/kineto/libkineto/third_party/dynolog/third_party/prometheus-cpp/3rdparty/civetweb': checked out 'd7ba35bbb649209c66e582d5a0244ba988a15159' 2025-11-03T16:59:00.9976625Z Submodule path 'third_party/kineto/libkineto/third_party/dynolog/third_party/prometheus-cpp/3rdparty/googletest': checked out 'e2239ee6043f73722e7aa812a459f54a28552929' 2025-11-03T16:59:01.0365757Z Submodule path 'third_party/kineto/libkineto/third_party/fmt': checked out '40626af88bd7df9a5fb80be7b25ac85b122d6c21' 2025-11-03T16:59:01.0898963Z Submodule path 'third_party/kineto/libkineto/third_party/googletest': checked out '52eb8108c5bdec04579160ae17225d66034bd723' 2025-11-03T16:59:01.1591868Z Submodule path 'third_party/kleidiai': checked out 'd7770c89632329a9914ef1a90289917597639cbe' 2025-11-03T16:59:01.2088552Z Submodule path 'third_party/mimalloc': checked out 'fbd8b99c2b828428947d70fdc046bb55609be93e' 2025-11-03T16:59:01.3449757Z Submodule path 'third_party/nlohmann': checked out '55f93686c01528224f448c19128836e7df245f72' 2025-11-03T16:59:01.9143534Z Submodule path 'third_party/onnx': checked out 'e709452ef2bbc1d113faf678c24e6d3467696e83' 2025-11-03T16:59:01.9187186Z Submodule 'third_party/pybind11' (https://github.com/pybind/pybind11.git) registered for path 'third_party/onnx/third_party/pybind11' 2025-11-03T16:59:01.9232376Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/onnx/third_party/pybind11'... 2025-11-03T16:59:03.0211565Z Submodule path 'third_party/onnx/third_party/pybind11': checked out 'a2e59f0e7065404b44dfe92a28aca47ba1378dc4' 2025-11-03T16:59:03.1171655Z Submodule path 'third_party/opentelemetry-cpp': checked out 'a799f4aed9c94b765dcdaabaeab7d5e7e2310878' 2025-11-03T16:59:03.1205074Z Submodule 'third_party/benchmark' (https://github.com/google/benchmark) registered for path 'third_party/opentelemetry-cpp/third_party/benchmark' 2025-11-03T16:59:03.1207214Z Submodule 'third_party/googletest' (https://github.com/google/googletest) registered for path 'third_party/opentelemetry-cpp/third_party/googletest' 2025-11-03T16:59:03.1212155Z Submodule 'third_party/ms-gsl' (https://github.com/microsoft/GSL) registered for path 'third_party/opentelemetry-cpp/third_party/ms-gsl' 2025-11-03T16:59:03.1215804Z Submodule 'third_party/nlohmann-json' (https://github.com/nlohmann/json) registered for path 'third_party/opentelemetry-cpp/third_party/nlohmann-json' 2025-11-03T16:59:03.1219947Z Submodule 'third_party/opentelemetry-proto' (https://github.com/open-telemetry/opentelemetry-proto) registered for path 'third_party/opentelemetry-cpp/third_party/opentelemetry-proto' 2025-11-03T16:59:03.1225183Z Submodule 'third_party/opentracing-cpp' (https://github.com/opentracing/opentracing-cpp.git) registered for path 'third_party/opentelemetry-cpp/third_party/opentracing-cpp' 2025-11-03T16:59:03.1228756Z Submodule 'third_party/prometheus-cpp' (https://github.com/jupp0r/prometheus-cpp) registered for path 'third_party/opentelemetry-cpp/third_party/prometheus-cpp' 2025-11-03T16:59:03.1232914Z Submodule 'tools/vcpkg' (https://github.com/Microsoft/vcpkg) registered for path 'third_party/opentelemetry-cpp/tools/vcpkg' 2025-11-03T16:59:03.1272388Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/opentelemetry-cpp/third_party/benchmark'... 2025-11-03T16:59:03.5555758Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/opentelemetry-cpp/third_party/opentracing-cpp'... 2025-11-03T16:59:03.5556815Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/opentelemetry-cpp/third_party/opentelemetry-proto'... 2025-11-03T16:59:03.5557728Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/opentelemetry-cpp/third_party/ms-gsl'... 2025-11-03T16:59:03.5558605Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/opentelemetry-cpp/third_party/prometheus-cpp'... 2025-11-03T16:59:03.6556362Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/opentelemetry-cpp/third_party/googletest'... 2025-11-03T16:59:04.0185327Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/opentelemetry-cpp/third_party/nlohmann-json'... 2025-11-03T16:59:09.5573441Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/opentelemetry-cpp/tools/vcpkg'... 2025-11-03T16:59:10.3533785Z Submodule path 'third_party/opentelemetry-cpp/third_party/benchmark': checked out 'd572f4777349d43653b21d6c2fc63020ab326db2' 2025-11-03T16:59:10.4018584Z Submodule path 'third_party/opentelemetry-cpp/third_party/googletest': checked out 'b796f7d44681514f58a683a3a71ff17c94edb0c1' 2025-11-03T16:59:10.4251544Z Submodule path 'third_party/opentelemetry-cpp/third_party/ms-gsl': checked out '6f4529395c5b7c2d661812257cd6780c67e54afa' 2025-11-03T16:59:10.5531192Z Submodule path 'third_party/opentelemetry-cpp/third_party/nlohmann-json': checked out 'bc889afb4c5bf1c0d8ee29ef35eaaf4c8bef8a5d' 2025-11-03T16:59:10.5743249Z Submodule path 'third_party/opentelemetry-cpp/third_party/opentelemetry-proto': checked out '4ca4f0335c63cda7ab31ea7ed70d6553aee14dce' 2025-11-03T16:59:10.5966857Z Submodule path 'third_party/opentelemetry-cpp/third_party/opentracing-cpp': checked out '06b57f48ded1fa3bdd3d4346f6ef29e40e08eaf5' 2025-11-03T16:59:10.6226201Z Submodule path 'third_party/opentelemetry-cpp/third_party/prometheus-cpp': checked out 'c9ffcdda9086ffd9e1283ea7a0276d831f3c8a8d' 2025-11-03T16:59:10.6255250Z Submodule 'civetweb' (https://github.com/civetweb/civetweb.git) registered for path 'third_party/opentelemetry-cpp/third_party/prometheus-cpp/3rdparty/civetweb' 2025-11-03T16:59:10.6258444Z Submodule 'googletest' (https://github.com/google/googletest.git) registered for path 'third_party/opentelemetry-cpp/third_party/prometheus-cpp/3rdparty/googletest' 2025-11-03T16:59:10.6302296Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/opentelemetry-cpp/third_party/prometheus-cpp/3rdparty/civetweb'... 2025-11-03T16:59:12.6075765Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/opentelemetry-cpp/third_party/prometheus-cpp/3rdparty/googletest'... 2025-11-03T16:59:12.8723314Z Submodule path 'third_party/opentelemetry-cpp/third_party/prometheus-cpp/3rdparty/civetweb': checked out 'eefb26f82b233268fc98577d265352720d477ba4' 2025-11-03T16:59:12.9277610Z Submodule path 'third_party/opentelemetry-cpp/third_party/prometheus-cpp/3rdparty/googletest': checked out 'e2239ee6043f73722e7aa812a459f54a28552929' 2025-11-03T16:59:13.6328773Z Submodule path 'third_party/opentelemetry-cpp/tools/vcpkg': checked out '8eb57355a4ffb410a2e94c07b4dca2dffbee8e50' 2025-11-03T16:59:13.6504973Z Submodule path 'third_party/pocketfft': checked out '0fa0ef591e38c2758e3184c6c23e497b9f732ffa' 2025-11-03T16:59:13.9603372Z Submodule path 'third_party/protobuf': checked out 'd1eca4e4b421cd2997495c4b4e65cea6be4e9b8a' 2025-11-03T16:59:13.9636896Z Submodule 'third_party/benchmark' (https://github.com/google/benchmark.git) registered for path 'third_party/protobuf/third_party/benchmark' 2025-11-03T16:59:13.9640936Z Submodule 'third_party/googletest' (https://github.com/google/googletest.git) registered for path 'third_party/protobuf/third_party/googletest' 2025-11-03T16:59:13.9684994Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/protobuf/third_party/benchmark'... 2025-11-03T16:59:14.4880739Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/protobuf/third_party/googletest'... 2025-11-03T16:59:14.7604909Z Submodule path 'third_party/protobuf/third_party/benchmark': checked out '5b7683f49e1e9223cf9927b24f6fd3d6bd82e3f8' 2025-11-03T16:59:14.8411208Z Submodule path 'third_party/protobuf/third_party/googletest': checked out '5ec7f0c4a113e2f18ac2c6cc7df51ad6afc24081' 2025-11-03T16:59:14.8568144Z Submodule path 'third_party/psimd': checked out '072586a71b55b7f8c584153d223e95687148a900' 2025-11-03T16:59:14.8755358Z Submodule path 'third_party/pthreadpool': checked out '4fe0e1e183925bf8cfa6aae24237e724a96479b8' 2025-11-03T16:59:14.9270310Z Submodule path 'third_party/pybind11': checked out 'f5fbe867d2d26e4a0a9177a51f6e568868ad3dc8' 2025-11-03T16:59:14.9636944Z Submodule path 'third_party/python-peachpy': checked out 'f45429b087dd7d5bc78bb40dc7cf06425c252d67' 2025-11-03T16:59:15.0159241Z Submodule path 'third_party/sleef': checked out '5a1d179df9cf652951b59010a2d2075372d67f68' 2025-11-03T16:59:15.0547787Z Submodule path 'third_party/tensorpipe': checked out 'af0118d13e52f5a08841464a768e01a0bf3e3075' 2025-11-03T16:59:15.0578107Z Submodule 'third_party/googletest' (https://github.com/google/googletest.git) registered for path 'third_party/tensorpipe/third_party/googletest' 2025-11-03T16:59:15.0580187Z Submodule 'third_party/libnop' (https://github.com/google/libnop.git) registered for path 'third_party/tensorpipe/third_party/libnop' 2025-11-03T16:59:15.0583829Z Submodule 'third_party/libuv' (https://github.com/libuv/libuv.git) registered for path 'third_party/tensorpipe/third_party/libuv' 2025-11-03T16:59:15.0587668Z Submodule 'third_party/pybind11' (https://github.com/pybind/pybind11.git) registered for path 'third_party/tensorpipe/third_party/pybind11' 2025-11-03T16:59:15.0630527Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/tensorpipe/third_party/googletest'... 2025-11-03T16:59:15.8526399Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/tensorpipe/third_party/libnop'... 2025-11-03T16:59:15.9525766Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/tensorpipe/third_party/libuv'... 2025-11-03T16:59:16.1912593Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/tensorpipe/third_party/pybind11'... 2025-11-03T16:59:16.2559039Z Submodule path 'third_party/tensorpipe/third_party/googletest': checked out 'aee0f9d9b5b87796ee8a0ab26b7587ec30e8858e' 2025-11-03T16:59:16.2786937Z Submodule path 'third_party/tensorpipe/third_party/libnop': checked out '910b55815be16109f04f4180e9adee14fb4ce281' 2025-11-03T16:59:16.3637049Z Submodule path 'third_party/tensorpipe/third_party/libuv': checked out '5152db2cbfeb5582e9c27c5ea1dba2cd9e10759b' 2025-11-03T16:59:16.4013681Z Submodule path 'third_party/tensorpipe/third_party/pybind11': checked out 'a23996fce38ff6ccfbcdc09f1e63f2c4be5ea2ef' 2025-11-03T16:59:16.4042659Z Submodule 'tools/clang' (https://github.com/wjakob/clang-cindex-python3) registered for path 'third_party/tensorpipe/third_party/pybind11/tools/clang' 2025-11-03T16:59:16.4087533Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/tensorpipe/third_party/pybind11/tools/clang'... 2025-11-03T16:59:16.6418210Z Submodule path 'third_party/tensorpipe/third_party/pybind11/tools/clang': checked out '6a00cbc4a9b8e68b71caf7f774b3f9c753ae84d5' 2025-11-03T16:59:16.6489919Z [command]/usr/bin/git submodule foreach --recursive git config --local gc.auto 0 2025-11-03T16:59:16.6980531Z Entering 'android/libs/fbjni' 2025-11-03T16:59:16.7050646Z Entering 'third_party/FP16' 2025-11-03T16:59:16.7120823Z Entering 'third_party/FXdiv' 2025-11-03T16:59:16.7193070Z Entering 'third_party/NNPACK' 2025-11-03T16:59:16.7255953Z Entering 'third_party/NVTX' 2025-11-03T16:59:16.7322473Z Entering 'third_party/VulkanMemoryAllocator' 2025-11-03T16:59:16.7388428Z Entering 'third_party/XNNPACK' 2025-11-03T16:59:16.7465471Z Entering 'third_party/aiter' 2025-11-03T16:59:16.7529707Z Entering 'third_party/aiter/3rdparty/composable_kernel' 2025-11-03T16:59:16.7614942Z Entering 'third_party/benchmark' 2025-11-03T16:59:16.7682799Z Entering 'third_party/composable_kernel' 2025-11-03T16:59:16.7756140Z Entering 'third_party/cpp-httplib' 2025-11-03T16:59:16.7832745Z Entering 'third_party/cpuinfo' 2025-11-03T16:59:16.7903514Z Entering 'third_party/cudnn_frontend' 2025-11-03T16:59:16.7968140Z Entering 'third_party/cutlass' 2025-11-03T16:59:16.8039614Z Entering 'third_party/fbgemm' 2025-11-03T16:59:16.8110267Z Entering 'third_party/fbgemm/external/asmjit' 2025-11-03T16:59:16.8179380Z Entering 'third_party/fbgemm/external/composable_kernel' 2025-11-03T16:59:16.8257257Z Entering 'third_party/fbgemm/external/cpuinfo' 2025-11-03T16:59:16.8317663Z Entering 'third_party/fbgemm/external/cutlass' 2025-11-03T16:59:16.8395870Z Entering 'third_party/fbgemm/external/googletest' 2025-11-03T16:59:16.8462580Z Entering 'third_party/fbgemm/external/hipify_torch' 2025-11-03T16:59:16.8529102Z Entering 'third_party/fbgemm/external/json' 2025-11-03T16:59:16.8609351Z Entering 'third_party/flash-attention' 2025-11-03T16:59:16.8676687Z Entering 'third_party/flash-attention/csrc/composable_kernel' 2025-11-03T16:59:16.8748881Z Entering 'third_party/flash-attention/csrc/cutlass' 2025-11-03T16:59:16.8825791Z Entering 'third_party/flatbuffers' 2025-11-03T16:59:16.8893472Z Entering 'third_party/fmt' 2025-11-03T16:59:16.8961727Z Entering 'third_party/gemmlowp/gemmlowp' 2025-11-03T16:59:16.9031562Z Entering 'third_party/gloo' 2025-11-03T16:59:16.9099423Z Entering 'third_party/googletest' 2025-11-03T16:59:16.9169822Z Entering 'third_party/ideep' 2025-11-03T16:59:16.9232461Z Entering 'third_party/ideep/mkl-dnn' 2025-11-03T16:59:16.9312369Z Entering 'third_party/ittapi' 2025-11-03T16:59:16.9384563Z Entering 'third_party/kineto' 2025-11-03T16:59:16.9447039Z Entering 'third_party/kineto/libkineto/third_party/dynolog' 2025-11-03T16:59:16.9504642Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/DCGM' 2025-11-03T16:59:16.9573446Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/cpr' 2025-11-03T16:59:16.9637377Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/fmt' 2025-11-03T16:59:16.9701574Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/gflags' 2025-11-03T16:59:16.9766964Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/gflags/doc' 2025-11-03T16:59:16.9832634Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/glog' 2025-11-03T16:59:16.9903435Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/googletest' 2025-11-03T16:59:16.9963249Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/json' 2025-11-03T16:59:17.0028764Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/pfs' 2025-11-03T16:59:17.0099228Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/prometheus-cpp' 2025-11-03T16:59:17.0158793Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/prometheus-cpp/3rdparty/civetweb' 2025-11-03T16:59:17.0220123Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/prometheus-cpp/3rdparty/googletest' 2025-11-03T16:59:17.0297329Z Entering 'third_party/kineto/libkineto/third_party/fmt' 2025-11-03T16:59:17.0361854Z Entering 'third_party/kineto/libkineto/third_party/googletest' 2025-11-03T16:59:17.0437734Z Entering 'third_party/kleidiai' 2025-11-03T16:59:17.0501599Z Entering 'third_party/mimalloc' 2025-11-03T16:59:17.0565268Z Entering 'third_party/nlohmann' 2025-11-03T16:59:17.0633367Z Entering 'third_party/onnx' 2025-11-03T16:59:17.0716610Z Entering 'third_party/onnx/third_party/pybind11' 2025-11-03T16:59:17.0786934Z Entering 'third_party/opentelemetry-cpp' 2025-11-03T16:59:17.0858816Z Entering 'third_party/opentelemetry-cpp/third_party/benchmark' 2025-11-03T16:59:17.0930353Z Entering 'third_party/opentelemetry-cpp/third_party/googletest' 2025-11-03T16:59:17.1002209Z Entering 'third_party/opentelemetry-cpp/third_party/ms-gsl' 2025-11-03T16:59:17.1068411Z Entering 'third_party/opentelemetry-cpp/third_party/nlohmann-json' 2025-11-03T16:59:17.1128001Z Entering 'third_party/opentelemetry-cpp/third_party/opentelemetry-proto' 2025-11-03T16:59:17.1187297Z Entering 'third_party/opentelemetry-cpp/third_party/opentracing-cpp' 2025-11-03T16:59:17.1246876Z Entering 'third_party/opentelemetry-cpp/third_party/prometheus-cpp' 2025-11-03T16:59:17.1299978Z Entering 'third_party/opentelemetry-cpp/third_party/prometheus-cpp/3rdparty/civetweb' 2025-11-03T16:59:17.1363495Z Entering 'third_party/opentelemetry-cpp/third_party/prometheus-cpp/3rdparty/googletest' 2025-11-03T16:59:17.1424802Z Entering 'third_party/opentelemetry-cpp/tools/vcpkg' 2025-11-03T16:59:17.1501003Z Entering 'third_party/pocketfft' 2025-11-03T16:59:17.1568945Z Entering 'third_party/protobuf' 2025-11-03T16:59:17.1639036Z Entering 'third_party/protobuf/third_party/benchmark' 2025-11-03T16:59:17.1707968Z Entering 'third_party/protobuf/third_party/googletest' 2025-11-03T16:59:17.1770241Z Entering 'third_party/psimd' 2025-11-03T16:59:17.1838384Z Entering 'third_party/pthreadpool' 2025-11-03T16:59:17.1906874Z Entering 'third_party/pybind11' 2025-11-03T16:59:17.1965339Z Entering 'third_party/python-peachpy' 2025-11-03T16:59:17.2028371Z Entering 'third_party/sleef' 2025-11-03T16:59:17.2097773Z Entering 'third_party/tensorpipe' 2025-11-03T16:59:17.2153876Z Entering 'third_party/tensorpipe/third_party/googletest' 2025-11-03T16:59:17.2214879Z Entering 'third_party/tensorpipe/third_party/libnop' 2025-11-03T16:59:17.2268655Z Entering 'third_party/tensorpipe/third_party/libuv' 2025-11-03T16:59:17.2325578Z Entering 'third_party/tensorpipe/third_party/pybind11' 2025-11-03T16:59:17.2384210Z Entering 'third_party/tensorpipe/third_party/pybind11/tools/clang' 2025-11-03T16:59:17.2478202Z ##[endgroup] 2025-11-03T16:59:17.2479009Z ##[group]Persisting credentials for submodules 2025-11-03T16:59:17.2489239Z [command]/usr/bin/git submodule foreach --recursive sh -c "git config --local --name-only --get-regexp 'url\.https\:\/\/github\.com\/\.insteadOf' && git config --local --unset-all 'url.https://github.com/.insteadOf' || :" 2025-11-03T16:59:17.2949834Z Entering 'android/libs/fbjni' 2025-11-03T16:59:17.3039632Z Entering 'third_party/FP16' 2025-11-03T16:59:17.3126796Z Entering 'third_party/FXdiv' 2025-11-03T16:59:17.3204752Z Entering 'third_party/NNPACK' 2025-11-03T16:59:17.3287966Z Entering 'third_party/NVTX' 2025-11-03T16:59:17.3370117Z Entering 'third_party/VulkanMemoryAllocator' 2025-11-03T16:59:17.3459685Z Entering 'third_party/XNNPACK' 2025-11-03T16:59:17.3561026Z Entering 'third_party/aiter' 2025-11-03T16:59:17.3648266Z Entering 'third_party/aiter/3rdparty/composable_kernel' 2025-11-03T16:59:17.3742495Z Entering 'third_party/benchmark' 2025-11-03T16:59:17.3828789Z Entering 'third_party/composable_kernel' 2025-11-03T16:59:17.3926357Z Entering 'third_party/cpp-httplib' 2025-11-03T16:59:17.4007929Z Entering 'third_party/cpuinfo' 2025-11-03T16:59:17.4102615Z Entering 'third_party/cudnn_frontend' 2025-11-03T16:59:17.4198496Z Entering 'third_party/cutlass' 2025-11-03T16:59:17.4285970Z Entering 'third_party/fbgemm' 2025-11-03T16:59:17.4373901Z Entering 'third_party/fbgemm/external/asmjit' 2025-11-03T16:59:17.4460441Z Entering 'third_party/fbgemm/external/composable_kernel' 2025-11-03T16:59:17.4554642Z Entering 'third_party/fbgemm/external/cpuinfo' 2025-11-03T16:59:17.4640622Z Entering 'third_party/fbgemm/external/cutlass' 2025-11-03T16:59:17.4738740Z Entering 'third_party/fbgemm/external/googletest' 2025-11-03T16:59:17.4819618Z Entering 'third_party/fbgemm/external/hipify_torch' 2025-11-03T16:59:17.4900067Z Entering 'third_party/fbgemm/external/json' 2025-11-03T16:59:17.4996864Z Entering 'third_party/flash-attention' 2025-11-03T16:59:17.5076819Z Entering 'third_party/flash-attention/csrc/composable_kernel' 2025-11-03T16:59:17.5165954Z Entering 'third_party/flash-attention/csrc/cutlass' 2025-11-03T16:59:17.5269227Z Entering 'third_party/flatbuffers' 2025-11-03T16:59:17.5361982Z Entering 'third_party/fmt' 2025-11-03T16:59:17.5449899Z Entering 'third_party/gemmlowp/gemmlowp' 2025-11-03T16:59:17.5544969Z Entering 'third_party/gloo' 2025-11-03T16:59:17.5634904Z Entering 'third_party/googletest' 2025-11-03T16:59:17.5719484Z Entering 'third_party/ideep' 2025-11-03T16:59:17.5804476Z Entering 'third_party/ideep/mkl-dnn' 2025-11-03T16:59:17.5900695Z Entering 'third_party/ittapi' 2025-11-03T16:59:17.5988572Z Entering 'third_party/kineto' 2025-11-03T16:59:17.6065518Z Entering 'third_party/kineto/libkineto/third_party/dynolog' 2025-11-03T16:59:17.6144997Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/DCGM' 2025-11-03T16:59:17.6232954Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/cpr' 2025-11-03T16:59:17.6321853Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/fmt' 2025-11-03T16:59:17.6408877Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/gflags' 2025-11-03T16:59:17.6491278Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/gflags/doc' 2025-11-03T16:59:17.6587562Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/glog' 2025-11-03T16:59:17.6674352Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/googletest' 2025-11-03T16:59:17.6759147Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/json' 2025-11-03T16:59:17.6842364Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/pfs' 2025-11-03T16:59:17.6926205Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/prometheus-cpp' 2025-11-03T16:59:17.7003708Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/prometheus-cpp/3rdparty/civetweb' 2025-11-03T16:59:17.7103694Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/prometheus-cpp/3rdparty/googletest' 2025-11-03T16:59:17.7212172Z Entering 'third_party/kineto/libkineto/third_party/fmt' 2025-11-03T16:59:17.7302619Z Entering 'third_party/kineto/libkineto/third_party/googletest' 2025-11-03T16:59:17.7392408Z Entering 'third_party/kleidiai' 2025-11-03T16:59:17.7480101Z Entering 'third_party/mimalloc' 2025-11-03T16:59:17.7568961Z Entering 'third_party/nlohmann' 2025-11-03T16:59:17.7657518Z Entering 'third_party/onnx' 2025-11-03T16:59:17.7754337Z Entering 'third_party/onnx/third_party/pybind11' 2025-11-03T16:59:17.7848438Z Entering 'third_party/opentelemetry-cpp' 2025-11-03T16:59:17.7941320Z Entering 'third_party/opentelemetry-cpp/third_party/benchmark' 2025-11-03T16:59:17.8032919Z Entering 'third_party/opentelemetry-cpp/third_party/googletest' 2025-11-03T16:59:17.8120921Z Entering 'third_party/opentelemetry-cpp/third_party/ms-gsl' 2025-11-03T16:59:17.8209676Z Entering 'third_party/opentelemetry-cpp/third_party/nlohmann-json' 2025-11-03T16:59:17.8292506Z Entering 'third_party/opentelemetry-cpp/third_party/opentelemetry-proto' 2025-11-03T16:59:17.8377344Z Entering 'third_party/opentelemetry-cpp/third_party/opentracing-cpp' 2025-11-03T16:59:17.8462214Z Entering 'third_party/opentelemetry-cpp/third_party/prometheus-cpp' 2025-11-03T16:59:17.8547804Z Entering 'third_party/opentelemetry-cpp/third_party/prometheus-cpp/3rdparty/civetweb' 2025-11-03T16:59:17.8639910Z Entering 'third_party/opentelemetry-cpp/third_party/prometheus-cpp/3rdparty/googletest' 2025-11-03T16:59:17.8735111Z Entering 'third_party/opentelemetry-cpp/tools/vcpkg' 2025-11-03T16:59:17.8844980Z Entering 'third_party/pocketfft' 2025-11-03T16:59:17.8938775Z Entering 'third_party/protobuf' 2025-11-03T16:59:17.9026134Z Entering 'third_party/protobuf/third_party/benchmark' 2025-11-03T16:59:17.9109993Z Entering 'third_party/protobuf/third_party/googletest' 2025-11-03T16:59:17.9203259Z Entering 'third_party/psimd' 2025-11-03T16:59:17.9290610Z Entering 'third_party/pthreadpool' 2025-11-03T16:59:17.9378418Z Entering 'third_party/pybind11' 2025-11-03T16:59:17.9460680Z Entering 'third_party/python-peachpy' 2025-11-03T16:59:17.9547784Z Entering 'third_party/sleef' 2025-11-03T16:59:17.9629223Z Entering 'third_party/tensorpipe' 2025-11-03T16:59:17.9718598Z Entering 'third_party/tensorpipe/third_party/googletest' 2025-11-03T16:59:17.9797629Z Entering 'third_party/tensorpipe/third_party/libnop' 2025-11-03T16:59:17.9880783Z Entering 'third_party/tensorpipe/third_party/libuv' 2025-11-03T16:59:17.9967727Z Entering 'third_party/tensorpipe/third_party/pybind11' 2025-11-03T16:59:18.0045177Z Entering 'third_party/tensorpipe/third_party/pybind11/tools/clang' 2025-11-03T16:59:18.0167634Z [command]/usr/bin/git submodule foreach --recursive sh -c "git config --local 'http.https://github.com/.extraheader' 'AUTHORIZATION: basic ***' && git config --local --show-origin --name-only --get-regexp remote.origin.url" 2025-11-03T16:59:18.0622854Z Entering 'android/libs/fbjni' 2025-11-03T16:59:18.0699364Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/android/libs/fbjni/config remote.origin.url 2025-11-03T16:59:18.0729586Z Entering 'third_party/FP16' 2025-11-03T16:59:18.0809197Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/NNPACK_deps/FP16/config remote.origin.url 2025-11-03T16:59:18.0835746Z Entering 'third_party/FXdiv' 2025-11-03T16:59:18.0918981Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/NNPACK_deps/FXdiv/config remote.origin.url 2025-11-03T16:59:18.0946027Z Entering 'third_party/NNPACK' 2025-11-03T16:59:18.1019514Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/NNPACK/config remote.origin.url 2025-11-03T16:59:18.1050345Z Entering 'third_party/NVTX' 2025-11-03T16:59:18.1130865Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/NVTX/config remote.origin.url 2025-11-03T16:59:18.1159983Z Entering 'third_party/VulkanMemoryAllocator' 2025-11-03T16:59:18.1236524Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/VulkanMemoryAllocator/config remote.origin.url 2025-11-03T16:59:18.1268350Z Entering 'third_party/XNNPACK' 2025-11-03T16:59:18.1349729Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/XNNPACK/config remote.origin.url 2025-11-03T16:59:18.1395969Z Entering 'third_party/aiter' 2025-11-03T16:59:18.1476552Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/aiter/config remote.origin.url 2025-11-03T16:59:18.1507298Z Entering 'third_party/aiter/3rdparty/composable_kernel' 2025-11-03T16:59:18.1589704Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/aiter/modules/3rdparty/composable_kernel/config remote.origin.url 2025-11-03T16:59:18.1622024Z Entering 'third_party/benchmark' 2025-11-03T16:59:18.1699838Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/benchmark/config remote.origin.url 2025-11-03T16:59:18.1730558Z Entering 'third_party/composable_kernel' 2025-11-03T16:59:18.1811279Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/composable_kernel/config remote.origin.url 2025-11-03T16:59:18.1844076Z Entering 'third_party/cpp-httplib' 2025-11-03T16:59:18.1930575Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/cpp-httplib/config remote.origin.url 2025-11-03T16:59:18.1960864Z Entering 'third_party/cpuinfo' 2025-11-03T16:59:18.2037994Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/cpuinfo/config remote.origin.url 2025-11-03T16:59:18.2072263Z Entering 'third_party/cudnn_frontend' 2025-11-03T16:59:18.2154737Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/cudnn_frontend/config remote.origin.url 2025-11-03T16:59:18.2185393Z Entering 'third_party/cutlass' 2025-11-03T16:59:18.2269273Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/cutlass/config remote.origin.url 2025-11-03T16:59:18.2308206Z Entering 'third_party/fbgemm' 2025-11-03T16:59:18.2389919Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/fbgemm/config remote.origin.url 2025-11-03T16:59:18.2419675Z Entering 'third_party/fbgemm/external/asmjit' 2025-11-03T16:59:18.2499062Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/fbgemm/modules/external/asmjit/config remote.origin.url 2025-11-03T16:59:18.2525620Z Entering 'third_party/fbgemm/external/composable_kernel' 2025-11-03T16:59:18.2610656Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/fbgemm/modules/external/composable_kernel/config remote.origin.url 2025-11-03T16:59:18.2643801Z Entering 'third_party/fbgemm/external/cpuinfo' 2025-11-03T16:59:18.2726405Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/fbgemm/modules/external/cpuinfo/config remote.origin.url 2025-11-03T16:59:18.2750699Z Entering 'third_party/fbgemm/external/cutlass' 2025-11-03T16:59:18.2830503Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/fbgemm/modules/external/cutlass/config remote.origin.url 2025-11-03T16:59:18.2862417Z Entering 'third_party/fbgemm/external/googletest' 2025-11-03T16:59:18.2938305Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/fbgemm/modules/external/googletest/config remote.origin.url 2025-11-03T16:59:18.2969248Z Entering 'third_party/fbgemm/external/hipify_torch' 2025-11-03T16:59:18.3049204Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/fbgemm/modules/external/hipify_torch/config remote.origin.url 2025-11-03T16:59:18.3078065Z Entering 'third_party/fbgemm/external/json' 2025-11-03T16:59:18.3150687Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/fbgemm/modules/external/json/config remote.origin.url 2025-11-03T16:59:18.3180840Z Entering 'third_party/flash-attention' 2025-11-03T16:59:18.3258378Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/flash-attention/config remote.origin.url 2025-11-03T16:59:18.3288149Z Entering 'third_party/flash-attention/csrc/composable_kernel' 2025-11-03T16:59:18.3369782Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/flash-attention/modules/csrc/composable_kernel/config remote.origin.url 2025-11-03T16:59:18.3405746Z Entering 'third_party/flash-attention/csrc/cutlass' 2025-11-03T16:59:18.3489882Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/flash-attention/modules/csrc/cutlass/config remote.origin.url 2025-11-03T16:59:18.3531234Z Entering 'third_party/flatbuffers' 2025-11-03T16:59:18.3611785Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/flatbuffers/config remote.origin.url 2025-11-03T16:59:18.3639161Z Entering 'third_party/fmt' 2025-11-03T16:59:18.3720746Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/fmt/config remote.origin.url 2025-11-03T16:59:18.3746817Z Entering 'third_party/gemmlowp/gemmlowp' 2025-11-03T16:59:18.3820194Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/gemmlowp/gemmlowp/config remote.origin.url 2025-11-03T16:59:18.3847859Z Entering 'third_party/gloo' 2025-11-03T16:59:18.3928252Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/gloo/config remote.origin.url 2025-11-03T16:59:18.3959544Z Entering 'third_party/googletest' 2025-11-03T16:59:18.4039509Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/googletest/config remote.origin.url 2025-11-03T16:59:18.4072452Z Entering 'third_party/ideep' 2025-11-03T16:59:18.4151129Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/ideep/config remote.origin.url 2025-11-03T16:59:18.4178055Z Entering 'third_party/ideep/mkl-dnn' 2025-11-03T16:59:18.4249541Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/ideep/modules/mkl-dnn/config remote.origin.url 2025-11-03T16:59:18.4286475Z Entering 'third_party/ittapi' 2025-11-03T16:59:18.4369259Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/ittapi/config remote.origin.url 2025-11-03T16:59:18.4401474Z Entering 'third_party/kineto' 2025-11-03T16:59:18.4479955Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/kineto/config remote.origin.url 2025-11-03T16:59:18.4510832Z Entering 'third_party/kineto/libkineto/third_party/dynolog' 2025-11-03T16:59:18.4590130Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/kineto/modules/libkineto/third_party/dynolog/config remote.origin.url 2025-11-03T16:59:18.4617914Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/DCGM' 2025-11-03T16:59:18.4694625Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/kineto/modules/libkineto/third_party/dynolog/modules/third_party/DCGM/config remote.origin.url 2025-11-03T16:59:18.4721274Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/cpr' 2025-11-03T16:59:18.4796713Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/kineto/modules/libkineto/third_party/dynolog/modules/third_party/cpr/config remote.origin.url 2025-11-03T16:59:18.4833600Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/fmt' 2025-11-03T16:59:18.4913150Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/kineto/modules/libkineto/third_party/dynolog/modules/third_party/fmt/config remote.origin.url 2025-11-03T16:59:18.4940923Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/gflags' 2025-11-03T16:59:18.5021824Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/kineto/modules/libkineto/third_party/dynolog/modules/third_party/gflags/config remote.origin.url 2025-11-03T16:59:18.5043571Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/gflags/doc' 2025-11-03T16:59:18.5121616Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/kineto/modules/libkineto/third_party/dynolog/modules/third_party/gflags/modules/doc/config remote.origin.url 2025-11-03T16:59:18.5155403Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/glog' 2025-11-03T16:59:18.5240495Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/kineto/modules/libkineto/third_party/dynolog/modules/third_party/glog/config remote.origin.url 2025-11-03T16:59:18.5265505Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/googletest' 2025-11-03T16:59:18.5340572Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/kineto/modules/libkineto/third_party/dynolog/modules/third_party/googletest/config remote.origin.url 2025-11-03T16:59:18.5364866Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/json' 2025-11-03T16:59:18.5440562Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/kineto/modules/libkineto/third_party/dynolog/modules/third_party/json/config remote.origin.url 2025-11-03T16:59:18.5471533Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/pfs' 2025-11-03T16:59:18.5560042Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/kineto/modules/libkineto/third_party/dynolog/modules/third_party/pfs/config remote.origin.url 2025-11-03T16:59:18.5590586Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/prometheus-cpp' 2025-11-03T16:59:18.5661843Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/kineto/modules/libkineto/third_party/dynolog/modules/third_party/prometheus-cpp/config remote.origin.url 2025-11-03T16:59:18.5691290Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/prometheus-cpp/3rdparty/civetweb' 2025-11-03T16:59:18.5771886Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/kineto/modules/libkineto/third_party/dynolog/modules/third_party/prometheus-cpp/modules/civetweb/config remote.origin.url 2025-11-03T16:59:18.5799023Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/prometheus-cpp/3rdparty/googletest' 2025-11-03T16:59:18.5871324Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/kineto/modules/libkineto/third_party/dynolog/modules/third_party/prometheus-cpp/modules/googletest/config remote.origin.url 2025-11-03T16:59:18.5912503Z Entering 'third_party/kineto/libkineto/third_party/fmt' 2025-11-03T16:59:18.5992013Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/kineto/modules/libkineto/third_party/fmt/config remote.origin.url 2025-11-03T16:59:18.6018830Z Entering 'third_party/kineto/libkineto/third_party/googletest' 2025-11-03T16:59:18.6091325Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/kineto/modules/libkineto/third_party/googletest/config remote.origin.url 2025-11-03T16:59:18.6120797Z Entering 'third_party/kleidiai' 2025-11-03T16:59:18.6190590Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/kleidiai/config remote.origin.url 2025-11-03T16:59:18.6223632Z Entering 'third_party/mimalloc' 2025-11-03T16:59:18.6300920Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/mimalloc/config remote.origin.url 2025-11-03T16:59:18.6331572Z Entering 'third_party/nlohmann' 2025-11-03T16:59:18.6412915Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/nlohmann/config remote.origin.url 2025-11-03T16:59:18.6443358Z Entering 'third_party/onnx' 2025-11-03T16:59:18.6520715Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/onnx/config remote.origin.url 2025-11-03T16:59:18.6565321Z Entering 'third_party/onnx/third_party/pybind11' 2025-11-03T16:59:18.6650979Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/onnx/modules/third_party/pybind11/config remote.origin.url 2025-11-03T16:59:18.6682654Z Entering 'third_party/opentelemetry-cpp' 2025-11-03T16:59:18.6760515Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/opentelemetry-cpp/config remote.origin.url 2025-11-03T16:59:18.6790242Z Entering 'third_party/opentelemetry-cpp/third_party/benchmark' 2025-11-03T16:59:18.6869510Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/opentelemetry-cpp/modules/third_party/benchmark/config remote.origin.url 2025-11-03T16:59:18.6899214Z Entering 'third_party/opentelemetry-cpp/third_party/googletest' 2025-11-03T16:59:18.6970912Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/opentelemetry-cpp/modules/third_party/googletest/config remote.origin.url 2025-11-03T16:59:18.6994424Z Entering 'third_party/opentelemetry-cpp/third_party/ms-gsl' 2025-11-03T16:59:18.7071699Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/opentelemetry-cpp/modules/third_party/ms-gsl/config remote.origin.url 2025-11-03T16:59:18.7102641Z Entering 'third_party/opentelemetry-cpp/third_party/nlohmann-json' 2025-11-03T16:59:18.7181633Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/opentelemetry-cpp/modules/third_party/nlohmann-json/config remote.origin.url 2025-11-03T16:59:18.7215078Z Entering 'third_party/opentelemetry-cpp/third_party/opentelemetry-proto' 2025-11-03T16:59:18.7290186Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/opentelemetry-cpp/modules/third_party/opentelemetry-proto/config remote.origin.url 2025-11-03T16:59:18.7319793Z Entering 'third_party/opentelemetry-cpp/third_party/opentracing-cpp' 2025-11-03T16:59:18.7392624Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/opentelemetry-cpp/modules/third_party/opentracing-cpp/config remote.origin.url 2025-11-03T16:59:18.7424123Z Entering 'third_party/opentelemetry-cpp/third_party/prometheus-cpp' 2025-11-03T16:59:18.7505770Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/opentelemetry-cpp/modules/third_party/prometheus-cpp/config remote.origin.url 2025-11-03T16:59:18.7531707Z Entering 'third_party/opentelemetry-cpp/third_party/prometheus-cpp/3rdparty/civetweb' 2025-11-03T16:59:18.7612551Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/opentelemetry-cpp/modules/third_party/prometheus-cpp/modules/civetweb/config remote.origin.url 2025-11-03T16:59:18.7639754Z Entering 'third_party/opentelemetry-cpp/third_party/prometheus-cpp/3rdparty/googletest' 2025-11-03T16:59:18.7712198Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/opentelemetry-cpp/modules/third_party/prometheus-cpp/modules/googletest/config remote.origin.url 2025-11-03T16:59:18.7741189Z Entering 'third_party/opentelemetry-cpp/tools/vcpkg' 2025-11-03T16:59:18.7826716Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/opentelemetry-cpp/modules/tools/vcpkg/config remote.origin.url 2025-11-03T16:59:18.7869332Z Entering 'third_party/pocketfft' 2025-11-03T16:59:18.7958581Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/pocketfft/config remote.origin.url 2025-11-03T16:59:18.7990206Z Entering 'third_party/protobuf' 2025-11-03T16:59:18.8067946Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/protobuf/config remote.origin.url 2025-11-03T16:59:18.8100754Z Entering 'third_party/protobuf/third_party/benchmark' 2025-11-03T16:59:18.8178119Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/protobuf/modules/third_party/benchmark/config remote.origin.url 2025-11-03T16:59:18.8201283Z Entering 'third_party/protobuf/third_party/googletest' 2025-11-03T16:59:18.8280198Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/protobuf/modules/third_party/googletest/config remote.origin.url 2025-11-03T16:59:18.8320266Z Entering 'third_party/psimd' 2025-11-03T16:59:18.8400641Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/NNPACK_deps/psimd/config remote.origin.url 2025-11-03T16:59:18.8432519Z Entering 'third_party/pthreadpool' 2025-11-03T16:59:18.8518166Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/NNPACK_deps/pthreadpool/config remote.origin.url 2025-11-03T16:59:18.8542707Z Entering 'third_party/pybind11' 2025-11-03T16:59:18.8629544Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/pybind11/config remote.origin.url 2025-11-03T16:59:18.8661438Z Entering 'third_party/python-peachpy' 2025-11-03T16:59:18.8741055Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/python-peachpy/config remote.origin.url 2025-11-03T16:59:18.8774740Z Entering 'third_party/sleef' 2025-11-03T16:59:18.8850752Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/sleef/config remote.origin.url 2025-11-03T16:59:18.8881613Z Entering 'third_party/tensorpipe' 2025-11-03T16:59:18.8955427Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/tensorpipe/config remote.origin.url 2025-11-03T16:59:18.8981293Z Entering 'third_party/tensorpipe/third_party/googletest' 2025-11-03T16:59:18.9061202Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/tensorpipe/modules/third_party/googletest/config remote.origin.url 2025-11-03T16:59:18.9091497Z Entering 'third_party/tensorpipe/third_party/libnop' 2025-11-03T16:59:18.9170979Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/tensorpipe/modules/third_party/libnop/config remote.origin.url 2025-11-03T16:59:18.9199755Z Entering 'third_party/tensorpipe/third_party/libuv' 2025-11-03T16:59:18.9272527Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/tensorpipe/modules/third_party/libuv/config remote.origin.url 2025-11-03T16:59:18.9305933Z Entering 'third_party/tensorpipe/third_party/pybind11' 2025-11-03T16:59:18.9380797Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/tensorpipe/modules/third_party/pybind11/config remote.origin.url 2025-11-03T16:59:18.9402482Z Entering 'third_party/tensorpipe/third_party/pybind11/tools/clang' 2025-11-03T16:59:18.9481815Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/tensorpipe/modules/third_party/pybind11/modules/tools/clang/config remote.origin.url 2025-11-03T16:59:19.0340314Z [command]/usr/bin/git submodule foreach --recursive git config --local --add 'url.https://github.com/.insteadOf' 'git@github.com:' 2025-11-03T16:59:19.0783662Z Entering 'android/libs/fbjni' 2025-11-03T16:59:19.0854300Z Entering 'third_party/FP16' 2025-11-03T16:59:19.0916667Z Entering 'third_party/FXdiv' 2025-11-03T16:59:19.0981084Z Entering 'third_party/NNPACK' 2025-11-03T16:59:19.1054607Z Entering 'third_party/NVTX' 2025-11-03T16:59:19.1119870Z Entering 'third_party/VulkanMemoryAllocator' 2025-11-03T16:59:19.1178526Z Entering 'third_party/XNNPACK' 2025-11-03T16:59:19.1252288Z Entering 'third_party/aiter' 2025-11-03T16:59:19.1309192Z Entering 'third_party/aiter/3rdparty/composable_kernel' 2025-11-03T16:59:19.1383417Z Entering 'third_party/benchmark' 2025-11-03T16:59:19.1454656Z Entering 'third_party/composable_kernel' 2025-11-03T16:59:19.1528088Z Entering 'third_party/cpp-httplib' 2025-11-03T16:59:19.1589357Z Entering 'third_party/cpuinfo' 2025-11-03T16:59:19.1652570Z Entering 'third_party/cudnn_frontend' 2025-11-03T16:59:19.1710892Z Entering 'third_party/cutlass' 2025-11-03T16:59:19.1782982Z Entering 'third_party/fbgemm' 2025-11-03T16:59:19.1852837Z Entering 'third_party/fbgemm/external/asmjit' 2025-11-03T16:59:19.1920343Z Entering 'third_party/fbgemm/external/composable_kernel' 2025-11-03T16:59:19.1989361Z Entering 'third_party/fbgemm/external/cpuinfo' 2025-11-03T16:59:19.2064560Z Entering 'third_party/fbgemm/external/cutlass' 2025-11-03T16:59:19.2140604Z Entering 'third_party/fbgemm/external/googletest' 2025-11-03T16:59:19.2199702Z Entering 'third_party/fbgemm/external/hipify_torch' 2025-11-03T16:59:19.2259647Z Entering 'third_party/fbgemm/external/json' 2025-11-03T16:59:19.2328276Z Entering 'third_party/flash-attention' 2025-11-03T16:59:19.2391931Z Entering 'third_party/flash-attention/csrc/composable_kernel' 2025-11-03T16:59:19.2468386Z Entering 'third_party/flash-attention/csrc/cutlass' 2025-11-03T16:59:19.2540127Z Entering 'third_party/flatbuffers' 2025-11-03T16:59:19.2613467Z Entering 'third_party/fmt' 2025-11-03T16:59:19.2680654Z Entering 'third_party/gemmlowp/gemmlowp' 2025-11-03T16:59:19.2750689Z Entering 'third_party/gloo' 2025-11-03T16:59:19.2813605Z Entering 'third_party/googletest' 2025-11-03T16:59:19.2880298Z Entering 'third_party/ideep' 2025-11-03T16:59:19.2947188Z Entering 'third_party/ideep/mkl-dnn' 2025-11-03T16:59:19.3021792Z Entering 'third_party/ittapi' 2025-11-03T16:59:19.3092110Z Entering 'third_party/kineto' 2025-11-03T16:59:19.3146392Z Entering 'third_party/kineto/libkineto/third_party/dynolog' 2025-11-03T16:59:19.3206553Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/DCGM' 2025-11-03T16:59:19.3274530Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/cpr' 2025-11-03T16:59:19.3338161Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/fmt' 2025-11-03T16:59:19.3401144Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/gflags' 2025-11-03T16:59:19.3457938Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/gflags/doc' 2025-11-03T16:59:19.3523742Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/glog' 2025-11-03T16:59:19.3587280Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/googletest' 2025-11-03T16:59:19.3654006Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/json' 2025-11-03T16:59:19.3716917Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/pfs' 2025-11-03T16:59:19.3781267Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/prometheus-cpp' 2025-11-03T16:59:19.3837556Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/prometheus-cpp/3rdparty/civetweb' 2025-11-03T16:59:19.3903010Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/prometheus-cpp/3rdparty/googletest' 2025-11-03T16:59:19.3983626Z Entering 'third_party/kineto/libkineto/third_party/fmt' 2025-11-03T16:59:19.4051522Z Entering 'third_party/kineto/libkineto/third_party/googletest' 2025-11-03T16:59:19.4122910Z Entering 'third_party/kleidiai' 2025-11-03T16:59:19.4189318Z Entering 'third_party/mimalloc' 2025-11-03T16:59:19.4261750Z Entering 'third_party/nlohmann' 2025-11-03T16:59:19.4334443Z Entering 'third_party/onnx' 2025-11-03T16:59:19.4415968Z Entering 'third_party/onnx/third_party/pybind11' 2025-11-03T16:59:19.4489080Z Entering 'third_party/opentelemetry-cpp' 2025-11-03T16:59:19.4548051Z Entering 'third_party/opentelemetry-cpp/third_party/benchmark' 2025-11-03T16:59:19.4616035Z Entering 'third_party/opentelemetry-cpp/third_party/googletest' 2025-11-03T16:59:19.4681218Z Entering 'third_party/opentelemetry-cpp/third_party/ms-gsl' 2025-11-03T16:59:19.4749496Z Entering 'third_party/opentelemetry-cpp/third_party/nlohmann-json' 2025-11-03T16:59:19.4813924Z Entering 'third_party/opentelemetry-cpp/third_party/opentelemetry-proto' 2025-11-03T16:59:19.4872854Z Entering 'third_party/opentelemetry-cpp/third_party/opentracing-cpp' 2025-11-03T16:59:19.4931536Z Entering 'third_party/opentelemetry-cpp/third_party/prometheus-cpp' 2025-11-03T16:59:19.4998353Z Entering 'third_party/opentelemetry-cpp/third_party/prometheus-cpp/3rdparty/civetweb' 2025-11-03T16:59:19.5066601Z Entering 'third_party/opentelemetry-cpp/third_party/prometheus-cpp/3rdparty/googletest' 2025-11-03T16:59:19.5139292Z Entering 'third_party/opentelemetry-cpp/tools/vcpkg' 2025-11-03T16:59:19.5220172Z Entering 'third_party/pocketfft' 2025-11-03T16:59:19.5292682Z Entering 'third_party/protobuf' 2025-11-03T16:59:19.5360602Z Entering 'third_party/protobuf/third_party/benchmark' 2025-11-03T16:59:19.5421666Z Entering 'third_party/protobuf/third_party/googletest' 2025-11-03T16:59:19.5500644Z Entering 'third_party/psimd' 2025-11-03T16:59:19.5574297Z Entering 'third_party/pthreadpool' 2025-11-03T16:59:19.5632362Z Entering 'third_party/pybind11' 2025-11-03T16:59:19.5701701Z Entering 'third_party/python-peachpy' 2025-11-03T16:59:19.5771936Z Entering 'third_party/sleef' 2025-11-03T16:59:19.5838626Z Entering 'third_party/tensorpipe' 2025-11-03T16:59:19.5897307Z Entering 'third_party/tensorpipe/third_party/googletest' 2025-11-03T16:59:19.5960231Z Entering 'third_party/tensorpipe/third_party/libnop' 2025-11-03T16:59:19.6022954Z Entering 'third_party/tensorpipe/third_party/libuv' 2025-11-03T16:59:19.6091237Z Entering 'third_party/tensorpipe/third_party/pybind11' 2025-11-03T16:59:19.6153434Z Entering 'third_party/tensorpipe/third_party/pybind11/tools/clang' 2025-11-03T16:59:19.6260307Z [command]/usr/bin/git submodule foreach --recursive git config --local --add 'url.https://github.com/.insteadOf' 'org-21003710@github.com:' 2025-11-03T16:59:19.6723253Z Entering 'android/libs/fbjni' 2025-11-03T16:59:19.6792217Z Entering 'third_party/FP16' 2025-11-03T16:59:19.6855676Z Entering 'third_party/FXdiv' 2025-11-03T16:59:19.6916540Z Entering 'third_party/NNPACK' 2025-11-03T16:59:19.6981740Z Entering 'third_party/NVTX' 2025-11-03T16:59:19.7042658Z Entering 'third_party/VulkanMemoryAllocator' 2025-11-03T16:59:19.7103140Z Entering 'third_party/XNNPACK' 2025-11-03T16:59:19.7180323Z Entering 'third_party/aiter' 2025-11-03T16:59:19.7239647Z Entering 'third_party/aiter/3rdparty/composable_kernel' 2025-11-03T16:59:19.7318088Z Entering 'third_party/benchmark' 2025-11-03T16:59:19.7381619Z Entering 'third_party/composable_kernel' 2025-11-03T16:59:19.7450654Z Entering 'third_party/cpp-httplib' 2025-11-03T16:59:19.7514249Z Entering 'third_party/cpuinfo' 2025-11-03T16:59:19.7586495Z Entering 'third_party/cudnn_frontend' 2025-11-03T16:59:19.7649449Z Entering 'third_party/cutlass' 2025-11-03T16:59:19.7723186Z Entering 'third_party/fbgemm' 2025-11-03T16:59:19.7792343Z Entering 'third_party/fbgemm/external/asmjit' 2025-11-03T16:59:19.7863987Z Entering 'third_party/fbgemm/external/composable_kernel' 2025-11-03T16:59:19.7934243Z Entering 'third_party/fbgemm/external/cpuinfo' 2025-11-03T16:59:19.8002788Z Entering 'third_party/fbgemm/external/cutlass' 2025-11-03T16:59:19.8080330Z Entering 'third_party/fbgemm/external/googletest' 2025-11-03T16:59:19.8144863Z Entering 'third_party/fbgemm/external/hipify_torch' 2025-11-03T16:59:19.8206089Z Entering 'third_party/fbgemm/external/json' 2025-11-03T16:59:19.8280534Z Entering 'third_party/flash-attention' 2025-11-03T16:59:19.8343685Z Entering 'third_party/flash-attention/csrc/composable_kernel' 2025-11-03T16:59:19.8422252Z Entering 'third_party/flash-attention/csrc/cutlass' 2025-11-03T16:59:19.8503090Z Entering 'third_party/flatbuffers' 2025-11-03T16:59:19.8572253Z Entering 'third_party/fmt' 2025-11-03T16:59:19.8635489Z Entering 'third_party/gemmlowp/gemmlowp' 2025-11-03T16:59:19.8702061Z Entering 'third_party/gloo' 2025-11-03T16:59:19.8761335Z Entering 'third_party/googletest' 2025-11-03T16:59:19.8818821Z Entering 'third_party/ideep' 2025-11-03T16:59:19.8875388Z Entering 'third_party/ideep/mkl-dnn' 2025-11-03T16:59:19.8949465Z Entering 'third_party/ittapi' 2025-11-03T16:59:19.9016662Z Entering 'third_party/kineto' 2025-11-03T16:59:19.9079005Z Entering 'third_party/kineto/libkineto/third_party/dynolog' 2025-11-03T16:59:19.9140299Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/DCGM' 2025-11-03T16:59:19.9200978Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/cpr' 2025-11-03T16:59:19.9257593Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/fmt' 2025-11-03T16:59:19.9315483Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/gflags' 2025-11-03T16:59:19.9369260Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/gflags/doc' 2025-11-03T16:59:19.9437487Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/glog' 2025-11-03T16:59:19.9493005Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/googletest' 2025-11-03T16:59:19.9554667Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/json' 2025-11-03T16:59:19.9625534Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/pfs' 2025-11-03T16:59:19.9687294Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/prometheus-cpp' 2025-11-03T16:59:19.9752927Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/prometheus-cpp/3rdparty/civetweb' 2025-11-03T16:59:19.9824270Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/prometheus-cpp/3rdparty/googletest' 2025-11-03T16:59:19.9906490Z Entering 'third_party/kineto/libkineto/third_party/fmt' 2025-11-03T16:59:19.9974055Z Entering 'third_party/kineto/libkineto/third_party/googletest' 2025-11-03T16:59:20.0048798Z Entering 'third_party/kleidiai' 2025-11-03T16:59:20.0116862Z Entering 'third_party/mimalloc' 2025-11-03T16:59:20.0180062Z Entering 'third_party/nlohmann' 2025-11-03T16:59:20.0245076Z Entering 'third_party/onnx' 2025-11-03T16:59:20.0322623Z Entering 'third_party/onnx/third_party/pybind11' 2025-11-03T16:59:20.0401097Z Entering 'third_party/opentelemetry-cpp' 2025-11-03T16:59:20.0462034Z Entering 'third_party/opentelemetry-cpp/third_party/benchmark' 2025-11-03T16:59:20.0531731Z Entering 'third_party/opentelemetry-cpp/third_party/googletest' 2025-11-03T16:59:20.0593826Z Entering 'third_party/opentelemetry-cpp/third_party/ms-gsl' 2025-11-03T16:59:20.0665946Z Entering 'third_party/opentelemetry-cpp/third_party/nlohmann-json' 2025-11-03T16:59:20.0730589Z Entering 'third_party/opentelemetry-cpp/third_party/opentelemetry-proto' 2025-11-03T16:59:20.0801239Z Entering 'third_party/opentelemetry-cpp/third_party/opentracing-cpp' 2025-11-03T16:59:20.0862667Z Entering 'third_party/opentelemetry-cpp/third_party/prometheus-cpp' 2025-11-03T16:59:20.0932048Z Entering 'third_party/opentelemetry-cpp/third_party/prometheus-cpp/3rdparty/civetweb' 2025-11-03T16:59:20.1004552Z Entering 'third_party/opentelemetry-cpp/third_party/prometheus-cpp/3rdparty/googletest' 2025-11-03T16:59:20.1080164Z Entering 'third_party/opentelemetry-cpp/tools/vcpkg' 2025-11-03T16:59:20.1160144Z Entering 'third_party/pocketfft' 2025-11-03T16:59:20.1224286Z Entering 'third_party/protobuf' 2025-11-03T16:59:20.1295465Z Entering 'third_party/protobuf/third_party/benchmark' 2025-11-03T16:59:20.1363098Z Entering 'third_party/protobuf/third_party/googletest' 2025-11-03T16:59:20.1431615Z Entering 'third_party/psimd' 2025-11-03T16:59:20.1494026Z Entering 'third_party/pthreadpool' 2025-11-03T16:59:20.1562216Z Entering 'third_party/pybind11' 2025-11-03T16:59:20.1623982Z Entering 'third_party/python-peachpy' 2025-11-03T16:59:20.1681328Z Entering 'third_party/sleef' 2025-11-03T16:59:20.1752626Z Entering 'third_party/tensorpipe' 2025-11-03T16:59:20.1812972Z Entering 'third_party/tensorpipe/third_party/googletest' 2025-11-03T16:59:20.1875036Z Entering 'third_party/tensorpipe/third_party/libnop' 2025-11-03T16:59:20.1936208Z Entering 'third_party/tensorpipe/third_party/libuv' 2025-11-03T16:59:20.2003400Z Entering 'third_party/tensorpipe/third_party/pybind11' 2025-11-03T16:59:20.2069736Z Entering 'third_party/tensorpipe/third_party/pybind11/tools/clang' 2025-11-03T16:59:20.2167201Z ##[endgroup] 2025-11-03T16:59:20.2222555Z [command]/usr/bin/git log -1 --format=%H 2025-11-03T16:59:20.2254105Z 3f6538febd727b782e6e13cfd026a309fb14351d 2025-11-03T16:59:20.2378639Z ##[group]Run cd "${GITHUB_WORKSPACE}" 2025-11-03T16:59:20.2378916Z cd "${GITHUB_WORKSPACE}" 2025-11-03T16:59:20.2379144Z # Clean stale submodule dirs 2025-11-03T16:59:20.2379381Z if [ -z "${NO_SUDO}" ]; then 2025-11-03T16:59:20.2379673Z  sudo git submodule foreach --recursive git clean -ffdx 2025-11-03T16:59:20.2379971Z else 2025-11-03T16:59:20.2380190Z  git submodule foreach --recursive git clean -ffdx 2025-11-03T16:59:20.2380443Z fi 2025-11-03T16:59:20.2390504Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2025-11-03T16:59:20.2390776Z env: 2025-11-03T16:59:20.2390940Z GIT_DEFAULT_BRANCH: main 2025-11-03T16:59:20.2391128Z NO_SUDO: true 2025-11-03T16:59:20.2391285Z ##[endgroup] 2025-11-03T16:59:20.2889258Z Entering 'android/libs/fbjni' 2025-11-03T16:59:20.2939730Z Entering 'third_party/FP16' 2025-11-03T16:59:20.2989849Z Entering 'third_party/FXdiv' 2025-11-03T16:59:20.3037585Z Entering 'third_party/NNPACK' 2025-11-03T16:59:20.3090643Z Entering 'third_party/NVTX' 2025-11-03T16:59:20.3153453Z Entering 'third_party/VulkanMemoryAllocator' 2025-11-03T16:59:20.3210243Z Entering 'third_party/XNNPACK' 2025-11-03T16:59:20.3370101Z Entering 'third_party/aiter' 2025-11-03T16:59:20.3427345Z Entering 'third_party/aiter/3rdparty/composable_kernel' 2025-11-03T16:59:20.3584526Z Entering 'third_party/benchmark' 2025-11-03T16:59:20.3642066Z Entering 'third_party/composable_kernel' 2025-11-03T16:59:20.3804287Z Entering 'third_party/cpp-httplib' 2025-11-03T16:59:20.3862132Z Entering 'third_party/cpuinfo' 2025-11-03T16:59:20.3923908Z Entering 'third_party/cudnn_frontend' 2025-11-03T16:59:20.3983638Z Entering 'third_party/cutlass' 2025-11-03T16:59:20.4112263Z Entering 'third_party/fbgemm' 2025-11-03T16:59:20.4203459Z Entering 'third_party/fbgemm/external/asmjit' 2025-11-03T16:59:20.4260187Z Entering 'third_party/fbgemm/external/composable_kernel' 2025-11-03T16:59:20.4403299Z Entering 'third_party/fbgemm/external/cpuinfo' 2025-11-03T16:59:20.4463858Z Entering 'third_party/fbgemm/external/cutlass' 2025-11-03T16:59:20.4599597Z Entering 'third_party/fbgemm/external/googletest' 2025-11-03T16:59:20.4653577Z Entering 'third_party/fbgemm/external/hipify_torch' 2025-11-03T16:59:20.4701139Z Entering 'third_party/fbgemm/external/json' 2025-11-03T16:59:20.4770241Z Entering 'third_party/flash-attention' 2025-11-03T16:59:20.4833532Z Entering 'third_party/flash-attention/csrc/composable_kernel' 2025-11-03T16:59:20.4959254Z Entering 'third_party/flash-attention/csrc/cutlass' 2025-11-03T16:59:20.5077505Z Entering 'third_party/flatbuffers' 2025-11-03T16:59:20.5184372Z Entering 'third_party/fmt' 2025-11-03T16:59:20.5240471Z Entering 'third_party/gemmlowp/gemmlowp' 2025-11-03T16:59:20.5291706Z Entering 'third_party/gloo' 2025-11-03T16:59:20.5355448Z Entering 'third_party/googletest' 2025-11-03T16:59:20.5414435Z Entering 'third_party/ideep' 2025-11-03T16:59:20.5468048Z Entering 'third_party/ideep/mkl-dnn' 2025-11-03T16:59:20.5586536Z Entering 'third_party/ittapi' 2025-11-03T16:59:20.5642099Z Entering 'third_party/kineto' 2025-11-03T16:59:20.5700459Z Entering 'third_party/kineto/libkineto/third_party/dynolog' 2025-11-03T16:59:20.5759278Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/DCGM' 2025-11-03T16:59:20.5824767Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/cpr' 2025-11-03T16:59:20.5883973Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/fmt' 2025-11-03T16:59:20.5943701Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/gflags' 2025-11-03T16:59:20.5991658Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/gflags/doc' 2025-11-03T16:59:20.6047938Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/glog' 2025-11-03T16:59:20.6092087Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/googletest' 2025-11-03T16:59:20.6151891Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/json' 2025-11-03T16:59:20.6211816Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/pfs' 2025-11-03T16:59:20.6270097Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/prometheus-cpp' 2025-11-03T16:59:20.6316178Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/prometheus-cpp/3rdparty/civetweb' 2025-11-03T16:59:20.6394895Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/prometheus-cpp/3rdparty/googletest' 2025-11-03T16:59:20.6456162Z Entering 'third_party/kineto/libkineto/third_party/fmt' 2025-11-03T16:59:20.6512709Z Entering 'third_party/kineto/libkineto/third_party/googletest' 2025-11-03T16:59:20.6569642Z Entering 'third_party/kleidiai' 2025-11-03T16:59:20.6632703Z Entering 'third_party/mimalloc' 2025-11-03T16:59:20.6687133Z Entering 'third_party/nlohmann' 2025-11-03T16:59:20.6754927Z Entering 'third_party/onnx' 2025-11-03T16:59:20.7197889Z Entering 'third_party/onnx/third_party/pybind11' 2025-11-03T16:59:20.7261469Z Entering 'third_party/opentelemetry-cpp' 2025-11-03T16:59:20.7349707Z Entering 'third_party/opentelemetry-cpp/third_party/benchmark' 2025-11-03T16:59:20.7400034Z Entering 'third_party/opentelemetry-cpp/third_party/googletest' 2025-11-03T16:59:20.7446681Z Entering 'third_party/opentelemetry-cpp/third_party/ms-gsl' 2025-11-03T16:59:20.7487308Z Entering 'third_party/opentelemetry-cpp/third_party/nlohmann-json' 2025-11-03T16:59:20.7550322Z Entering 'third_party/opentelemetry-cpp/third_party/opentelemetry-proto' 2025-11-03T16:59:20.7600158Z Entering 'third_party/opentelemetry-cpp/third_party/opentracing-cpp' 2025-11-03T16:59:20.7651969Z Entering 'third_party/opentelemetry-cpp/third_party/prometheus-cpp' 2025-11-03T16:59:20.7710542Z Entering 'third_party/opentelemetry-cpp/third_party/prometheus-cpp/3rdparty/civetweb' 2025-11-03T16:59:20.7775967Z Entering 'third_party/opentelemetry-cpp/third_party/prometheus-cpp/3rdparty/googletest' 2025-11-03T16:59:20.7832791Z Entering 'third_party/opentelemetry-cpp/tools/vcpkg' 2025-11-03T16:59:20.8170033Z Entering 'third_party/pocketfft' 2025-11-03T16:59:20.8218771Z Entering 'third_party/protobuf' 2025-11-03T16:59:20.8324236Z Entering 'third_party/protobuf/third_party/benchmark' 2025-11-03T16:59:20.8382732Z Entering 'third_party/protobuf/third_party/googletest' 2025-11-03T16:59:20.8441391Z Entering 'third_party/psimd' 2025-11-03T16:59:20.8491438Z Entering 'third_party/pthreadpool' 2025-11-03T16:59:20.8544421Z Entering 'third_party/pybind11' 2025-11-03T16:59:20.8602850Z Entering 'third_party/python-peachpy' 2025-11-03T16:59:20.8654695Z Entering 'third_party/sleef' 2025-11-03T16:59:20.8719005Z Entering 'third_party/tensorpipe' 2025-11-03T16:59:20.8769537Z Entering 'third_party/tensorpipe/third_party/googletest' 2025-11-03T16:59:20.8823296Z Entering 'third_party/tensorpipe/third_party/libnop' 2025-11-03T16:59:20.8876080Z Entering 'third_party/tensorpipe/third_party/libuv' 2025-11-03T16:59:20.8925297Z Entering 'third_party/tensorpipe/third_party/pybind11' 2025-11-03T16:59:20.8980577Z Entering 'third_party/tensorpipe/third_party/pybind11/tools/clang' 2025-11-03T16:59:20.9162661Z Prepare all required actions 2025-11-03T16:59:20.9163163Z Getting action download info 2025-11-03T16:59:21.1018546Z ##[group]Run ./.github/actions/setup-linux 2025-11-03T16:59:21.1018784Z env: 2025-11-03T16:59:21.1018949Z GIT_DEFAULT_BRANCH: main 2025-11-03T16:59:21.1019332Z ##[endgroup] 2025-11-03T16:59:21.1061478Z ##[group]Run set -euo pipefail 2025-11-03T16:59:21.1061732Z set -euo pipefail 2025-11-03T16:59:21.1061949Z function get_ec2_metadata() { 2025-11-03T16:59:21.1062241Z  # Pulled from instance metadata endpoint for EC2 2025-11-03T16:59:21.1062697Z  # see https://docs.aws.amazon.com/AWSEC2/latest/UserGuide/instancedata-data-retrieval.html 2025-11-03T16:59:21.1063088Z  category=$1 2025-11-03T16:59:21.1063348Z  # If it is GCP runner (runner name contains gcp), do not run this 2025-11-03T16:59:21.1063653Z  runner_name_str=i-05b4d4af410353347 2025-11-03T16:59:21.1063946Z  if [[ -f /.inarc ]]; then 2025-11-03T16:59:21.1064196Z  echo "ARC Runner, no info on ec2 metadata" 2025-11-03T16:59:21.1064468Z  elif [[ $runner_name_str == *"gcp"* ]]; then 2025-11-03T16:59:21.1064799Z  echo "Runner is from Google Cloud Platform, No info on ec2 metadata" 2025-11-03T16:59:21.1065107Z  else 2025-11-03T16:59:21.1065700Z  curl -H "X-aws-ec2-metadata-token: $(curl -s -X PUT "http://169.254.169.254/latest/api/token" -H "X-aws-ec2-metadata-token-ttl-seconds: 30")" -fsSL "http://169.254.169.254/latest/meta-data/${category}" 2025-11-03T16:59:21.1066319Z  fi 2025-11-03T16:59:21.1066486Z } 2025-11-03T16:59:21.1066681Z echo "ami-id: $(get_ec2_metadata ami-id)" 2025-11-03T16:59:21.1066986Z echo "instance-id: $(get_ec2_metadata instance-id)" 2025-11-03T16:59:21.1067314Z echo "instance-type: $(get_ec2_metadata instance-type)" 2025-11-03T16:59:21.1067606Z echo "system info $(uname -a)" 2025-11-03T16:59:21.1076316Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2025-11-03T16:59:21.1076605Z env: 2025-11-03T16:59:21.1076770Z GIT_DEFAULT_BRANCH: main 2025-11-03T16:59:21.1076960Z ##[endgroup] 2025-11-03T16:59:21.1264835Z ami-id: ami-08982f1c5bf93d976 2025-11-03T16:59:21.1399453Z instance-id: i-05b4d4af410353347 2025-11-03T16:59:21.1538295Z instance-type: g6.12xlarge 2025-11-03T16:59:21.1559632Z system info Linux ip-10-0-8-115.ec2.internal 6.1.150-174.273.amzn2023.x86_64 #1 SMP PREEMPT_DYNAMIC Tue Sep 9 12:21:26 UTC 2025 x86_64 x86_64 x86_64 GNU/Linux 2025-11-03T16:59:21.1587102Z ##[group]Run if [ -f /usr/bin/nvidia-smi ]; then nvidia-smi; fi 2025-11-03T16:59:21.1587449Z if [ -f /usr/bin/nvidia-smi ]; then nvidia-smi; fi 2025-11-03T16:59:21.1594644Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2025-11-03T16:59:21.1594917Z env: 2025-11-03T16:59:21.1595090Z GIT_DEFAULT_BRANCH: main 2025-11-03T16:59:21.1595280Z ##[endgroup] 2025-11-03T16:59:23.4582470Z Mon Nov 3 16:59:23 2025 2025-11-03T16:59:23.4583381Z +-----------------------------------------------------------------------------------------+ 2025-11-03T16:59:23.4584358Z | NVIDIA-SMI 580.82.07 Driver Version: 580.82.07 CUDA Version: 13.0 | 2025-11-03T16:59:23.4585255Z +-----------------------------------------+------------------------+----------------------+ 2025-11-03T16:59:23.4586160Z | GPU Name Persistence-M | Bus-Id Disp.A | Volatile Uncorr. ECC | 2025-11-03T16:59:23.4587106Z | Fan Temp Perf Pwr:Usage/Cap | Memory-Usage | GPU-Util Compute M. | 2025-11-03T16:59:23.4588003Z | | | MIG M. | 2025-11-03T16:59:23.4588737Z |=========================================+========================+======================| 2025-11-03T16:59:23.4869143Z | 0 NVIDIA L4 Off | 00000000:38:00.0 Off | 0 | 2025-11-03T16:59:23.4870608Z | N/A 43C P0 28W / 72W | 0MiB / 23034MiB | 0% Default | 2025-11-03T16:59:23.4871339Z | | | N/A | 2025-11-03T16:59:23.4872014Z +-----------------------------------------+------------------------+----------------------+ 2025-11-03T16:59:23.4873065Z | 1 NVIDIA L4 Off | 00000000:3A:00.0 Off | 0 | 2025-11-03T16:59:23.4873812Z | N/A 40C P0 28W / 72W | 0MiB / 23034MiB | 0% Default | 2025-11-03T16:59:23.4874490Z | | | N/A | 2025-11-03T16:59:23.4875162Z +-----------------------------------------+------------------------+----------------------+ 2025-11-03T16:59:23.4875952Z | 2 NVIDIA L4 Off | 00000000:3C:00.0 Off | 0 | 2025-11-03T16:59:23.4876737Z | N/A 43C P0 28W / 72W | 0MiB / 23034MiB | 3% Default | 2025-11-03T16:59:23.4877470Z | | | N/A | 2025-11-03T16:59:23.4878272Z +-----------------------------------------+------------------------+----------------------+ 2025-11-03T16:59:23.4879071Z | 3 NVIDIA L4 Off | 00000000:3E:00.0 Off | 0 | 2025-11-03T16:59:23.4879426Z | N/A 41C P0 27W / 72W | 0MiB / 23034MiB | 0% Default | 2025-11-03T16:59:23.4879712Z | | | N/A | 2025-11-03T16:59:23.4879984Z +-----------------------------------------+------------------------+----------------------+ 2025-11-03T16:59:23.4880196Z 2025-11-03T16:59:23.4880320Z +-----------------------------------------------------------------------------------------+ 2025-11-03T16:59:23.4880632Z | Processes: | 2025-11-03T16:59:23.4880955Z | GPU GI CI PID Type Process name GPU Memory | 2025-11-03T16:59:23.4881254Z | ID ID Usage | 2025-11-03T16:59:23.4881492Z |=========================================================================================| 2025-11-03T16:59:23.4894422Z | No running processes found | 2025-11-03T16:59:23.4895311Z +-----------------------------------------------------------------------------------------+ 2025-11-03T16:59:24.7903301Z ##[group]Run echo "IN_CONTAINER_RUNNER=$(if [ -f /.inarc ] || [ -f /.incontainer ]; then echo true ; else echo false; fi)" >> "$GITHUB_OUTPUT" 2025-11-03T16:59:24.7904003Z echo "IN_CONTAINER_RUNNER=$(if [ -f /.inarc ] || [ -f /.incontainer ]; then echo true ; else echo false; fi)" >> "$GITHUB_OUTPUT" 2025-11-03T16:59:24.7916139Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2025-11-03T16:59:24.7916435Z env: 2025-11-03T16:59:24.7916607Z GIT_DEFAULT_BRANCH: main 2025-11-03T16:59:24.7916791Z ##[endgroup] 2025-11-03T16:59:24.8006508Z ##[group]Run if systemctl is-active --quiet docker; then 2025-11-03T16:59:24.8006837Z if systemctl is-active --quiet docker; then 2025-11-03T16:59:24.8007139Z  echo "Docker daemon is running..."; 2025-11-03T16:59:24.8007375Z else 2025-11-03T16:59:24.8007640Z  echo "Starting docker daemon..." && sudo systemctl start docker; 2025-11-03T16:59:24.8007940Z fi 2025-11-03T16:59:24.8015519Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2025-11-03T16:59:24.8015809Z env: 2025-11-03T16:59:24.8015982Z GIT_DEFAULT_BRANCH: main 2025-11-03T16:59:24.8016181Z ##[endgroup] 2025-11-03T16:59:24.8128686Z Docker daemon is running... 2025-11-03T16:59:24.8180134Z ##[group]Run nick-fields/retry@v3.0.0 2025-11-03T16:59:24.8180411Z with: 2025-11-03T16:59:24.8180578Z shell: bash 2025-11-03T16:59:24.8180967Z timeout_minutes: 5 2025-11-03T16:59:24.8181172Z max_attempts: 3 2025-11-03T16:59:24.8181350Z retry_wait_seconds: 30 2025-11-03T16:59:24.8182979Z command: AWS_ACCOUNT_ID=$(aws sts get-caller-identity|grep Account|cut -f4 -d\") aws ecr get-login-password --region "$AWS_DEFAULT_REGION" | docker login --username AWS \ --password-stdin "$AWS_ACCOUNT_ID.dkr.ecr.$AWS_DEFAULT_REGION.amazonaws.com" # For LF Runners we need to make sure we also login to Meta's ECR docker registry too. META_AWS_ACCOUNT_ID=308535385114 if [ "$AWS_ACCOUNT_ID" != "$META_AWS_ACCOUNT_ID" ] ; then aws ecr get-login-password --region "$AWS_DEFAULT_REGION" | docker login --username AWS \ --password-stdin "$META_AWS_ACCOUNT_ID.dkr.ecr.$AWS_DEFAULT_REGION.amazonaws.com" fi 2025-11-03T16:59:24.8184762Z polling_interval_seconds: 1 2025-11-03T16:59:24.8184969Z warning_on_retry: true 2025-11-03T16:59:24.8185162Z continue_on_error: false 2025-11-03T16:59:24.8185349Z env: 2025-11-03T16:59:24.8185505Z GIT_DEFAULT_BRANCH: main 2025-11-03T16:59:24.8185703Z AWS_RETRY_MODE: standard 2025-11-03T16:59:24.8185891Z AWS_MAX_ATTEMPTS: 5 2025-11-03T16:59:24.8186078Z AWS_DEFAULT_REGION: us-east-1 2025-11-03T16:59:24.8186273Z ##[endgroup] 2025-11-03T16:59:25.9055223Z WARNING! Your password will be stored unencrypted in /home/ec2-user/.docker/config.json. 2025-11-03T16:59:25.9055857Z Configure a credential helper to remove this warning. See 2025-11-03T16:59:25.9056397Z https://docs.docker.com/engine/reference/commandline/login/#credentials-store 2025-11-03T16:59:25.9056749Z 2025-11-03T16:59:25.9056842Z Login Succeeded 2025-11-03T16:59:26.8962897Z Command completed after 1 attempt(s). 2025-11-03T16:59:26.9049531Z ##[group]Run env | grep '^GITHUB' >> "/tmp/github_env_${GITHUB_RUN_ID}" 2025-11-03T16:59:26.9049933Z env | grep '^GITHUB' >> "/tmp/github_env_${GITHUB_RUN_ID}" 2025-11-03T16:59:26.9050260Z env | grep '^CI' >> "/tmp/github_env_${GITHUB_RUN_ID}" 2025-11-03T16:59:26.9059738Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2025-11-03T16:59:26.9060023Z env: 2025-11-03T16:59:26.9060188Z GIT_DEFAULT_BRANCH: main 2025-11-03T16:59:26.9060394Z ##[endgroup] 2025-11-03T16:59:26.9186868Z ##[group]Run # ignore expansion of "docker ps -q" since it could be empty 2025-11-03T16:59:26.9187324Z # ignore expansion of "docker ps -q" since it could be empty 2025-11-03T16:59:26.9187639Z # shellcheck disable=SC2046 2025-11-03T16:59:26.9187893Z docker stop $(docker ps -q) || true 2025-11-03T16:59:26.9188145Z # Prune all of the docker images 2025-11-03T16:59:26.9188392Z docker system prune -af 2025-11-03T16:59:26.9196846Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2025-11-03T16:59:26.9197126Z env: 2025-11-03T16:59:26.9197285Z GIT_DEFAULT_BRANCH: main 2025-11-03T16:59:26.9197471Z ##[endgroup] 2025-11-03T16:59:26.9532321Z "docker stop" requires at least 1 argument. 2025-11-03T16:59:26.9533036Z See 'docker stop --help'. 2025-11-03T16:59:26.9533385Z 2025-11-03T16:59:26.9533686Z Usage: docker stop [OPTIONS] CONTAINER [CONTAINER...] 2025-11-03T16:59:26.9534152Z 2025-11-03T16:59:26.9534341Z Stop one or more running containers 2025-11-03T16:59:26.9825153Z Total reclaimed space: 0B 2025-11-03T16:59:26.9981558Z ##[group]Run pytorch/test-infra/.github/actions/calculate-docker-image@main 2025-11-03T16:59:26.9981950Z with: 2025-11-03T16:59:26.9982560Z docker-image-name: 308535385114.dkr.ecr.us-east-1.amazonaws.com/pytorch/ci-image:pytorch-linux-jammy-cuda12.8-cudnn9-py3.12-gcc11-vllm-30147fb242f88ab8f13e14b76ab186d189d8605c 2025-11-03T16:59:26.9983234Z use-custom-docker-registry: true 2025-11-03T16:59:26.9983469Z docker-build-dir: .ci/docker 2025-11-03T16:59:26.9983675Z docker-build-script: ./build.sh 2025-11-03T16:59:26.9983890Z working-directory: . 2025-11-03T16:59:26.9984144Z docker-registry: 308535385114.dkr.ecr.us-east-1.amazonaws.com 2025-11-03T16:59:26.9984425Z force-push: false 2025-11-03T16:59:26.9984596Z env: 2025-11-03T16:59:26.9984743Z GIT_DEFAULT_BRANCH: main 2025-11-03T16:59:26.9984928Z ##[endgroup] 2025-11-03T16:59:27.0006104Z ##[group]Run set -ex 2025-11-03T16:59:27.0006335Z set -ex 2025-11-03T16:59:27.0006490Z  2025-11-03T16:59:27.0006806Z # If the docker build directory or the build script doesn't exist, the action will 2025-11-03T16:59:27.0007454Z # gracefully return the docker image name as it is. Pulling docker image in Linux 2025-11-03T16:59:27.0007856Z # job could then download the pre-built image as usual 2025-11-03T16:59:27.0008340Z if [[ -d "${DOCKER_BUILD_DIR}" ]] && [[ -f "${DOCKER_BUILD_DIR}/${DOCKER_BUILD_SCRIPT}" ]] && [[ "${USE_CUSTOM_DOCKER_REGISTRY}" == "true" ]]; then 2025-11-03T16:59:27.0008888Z  echo "skip=false" >> "${GITHUB_OUTPUT}" 2025-11-03T16:59:27.0009119Z else 2025-11-03T16:59:27.0009304Z  echo "skip=true" >> "${GITHUB_OUTPUT}" 2025-11-03T16:59:27.0009862Z  echo "docker-image=${DOCKER_IMAGE_NAME}" >> "${GITHUB_OUTPUT}" 2025-11-03T16:59:27.0010144Z  2025-11-03T16:59:27.0010540Z  echo "Not using custom ECR registry. Either it was not requested or there is no Docker build script in the ${REPO_NAME} repo..." 2025-11-03T16:59:27.0010994Z  exit 0 2025-11-03T16:59:27.0011143Z fi 2025-11-03T16:59:27.0011289Z  2025-11-03T16:59:27.0011524Z if [[ "${DOCKER_IMAGE_NAME}" == *"${DOCKER_REGISTRY}/${REPO_NAME}"* ]]; then 2025-11-03T16:59:27.0011956Z  # The docker image name already includes the ECR prefix and tag, so we can just 2025-11-03T16:59:27.0012330Z  # use it as it is, but first let's extract the tag 2025-11-03T16:59:27.0012658Z  DOCKER_TAG=$(echo "${DOCKER_IMAGE_NAME}" | awk -F '[:,]' '{print $2}') 2025-11-03T16:59:27.0013016Z  echo "docker-tag=${DOCKER_TAG}" >> "${GITHUB_OUTPUT}" 2025-11-03T16:59:27.0013360Z  echo "docker-image=${DOCKER_IMAGE_NAME}" >> "${GITHUB_OUTPUT}" 2025-11-03T16:59:27.0013637Z else 2025-11-03T16:59:27.0013823Z  if [[ "${DOCKER_IMAGE_NAME}" == *:* ]]; then 2025-11-03T16:59:27.0014088Z  CUSTOM_TAG_PREFIX=${DOCKER_IMAGE_NAME#*:} 2025-11-03T16:59:27.0014370Z  DOCKER_IMAGE_NAME=${DOCKER_IMAGE_NAME%%:*} 2025-11-03T16:59:27.0014615Z  fi 2025-11-03T16:59:27.0014936Z  DOCKER_TAG=${CUSTOM_TAG_PREFIX:+${CUSTOM_TAG_PREFIX}-}$(git rev-parse HEAD:"${DOCKER_BUILD_DIR}") 2025-11-03T16:59:27.0015350Z  echo "docker-tag=${DOCKER_TAG}" >> "${GITHUB_OUTPUT}" 2025-11-03T16:59:27.0015794Z  echo "docker-image=${DOCKER_REGISTRY}/${REPO_NAME}/${DOCKER_IMAGE_NAME}:${DOCKER_TAG}" >> "${GITHUB_OUTPUT}" 2025-11-03T16:59:27.0016283Z  echo "custom-tag-prefix=${CUSTOM_TAG_PREFIX}" >> "${GITHUB_OUTPUT}" 2025-11-03T16:59:27.0016573Z fi 2025-11-03T16:59:27.0023970Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2025-11-03T16:59:27.0024244Z env: 2025-11-03T16:59:27.0024413Z GIT_DEFAULT_BRANCH: main 2025-11-03T16:59:27.0024606Z REPO_NAME: pytorch 2025-11-03T16:59:27.0025384Z DOCKER_IMAGE_NAME: 308535385114.dkr.ecr.us-east-1.amazonaws.com/pytorch/ci-image:pytorch-linux-jammy-cuda12.8-cudnn9-py3.12-gcc11-vllm-30147fb242f88ab8f13e14b76ab186d189d8605c 2025-11-03T16:59:27.0026051Z DOCKER_BUILD_DIR: .ci/docker 2025-11-03T16:59:27.0026268Z DOCKER_BUILD_SCRIPT: ./build.sh 2025-11-03T16:59:27.0026542Z DOCKER_REGISTRY: 308535385114.dkr.ecr.us-east-1.amazonaws.com 2025-11-03T16:59:27.0026831Z USE_CUSTOM_DOCKER_REGISTRY: true 2025-11-03T16:59:27.0027036Z CUSTOM_TAG_PREFIX: 2025-11-03T16:59:27.0027199Z ##[endgroup] 2025-11-03T16:59:27.0060209Z + [[ -d .ci/docker ]] 2025-11-03T16:59:27.0060505Z + [[ -f .ci/docker/./build.sh ]] 2025-11-03T16:59:27.0060786Z + [[ true == \t\r\u\e ]] 2025-11-03T16:59:27.0061039Z + echo skip=false 2025-11-03T16:59:27.0062030Z + [[ 308535385114.dkr.ecr.us-east-1.amazonaws.com/pytorch/ci-image:pytorch-linux-jammy-cuda12.8-cudnn9-py3.12-gcc11-vllm-30147fb242f88ab8f13e14b76ab186d189d8605c == *\3\0\8\5\3\5\3\8\5\1\1\4\.\d\k\r\.\e\c\r\.\u\s\-\e\a\s\t\-\1\.\a\m\a\z\o\n\a\w\s\.\c\o\m\/\p\y\t\o\r\c\h* ]] 2025-11-03T16:59:27.0069570Z ++ echo 308535385114.dkr.ecr.us-east-1.amazonaws.com/pytorch/ci-image:pytorch-linux-jammy-cuda12.8-cudnn9-py3.12-gcc11-vllm-30147fb242f88ab8f13e14b76ab186d189d8605c 2025-11-03T16:59:27.0070669Z ++ awk -F '[:,]' '{print $2}' 2025-11-03T16:59:27.0103564Z + DOCKER_TAG=pytorch-linux-jammy-cuda12.8-cudnn9-py3.12-gcc11-vllm-30147fb242f88ab8f13e14b76ab186d189d8605c 2025-11-03T16:59:27.0105275Z + echo docker-tag=pytorch-linux-jammy-cuda12.8-cudnn9-py3.12-gcc11-vllm-30147fb242f88ab8f13e14b76ab186d189d8605c 2025-11-03T16:59:27.0107471Z + echo docker-image=308535385114.dkr.ecr.us-east-1.amazonaws.com/pytorch/ci-image:pytorch-linux-jammy-cuda12.8-cudnn9-py3.12-gcc11-vllm-30147fb242f88ab8f13e14b76ab186d189d8605c 2025-11-03T16:59:27.0149140Z ##[group]Run set +e 2025-11-03T16:59:27.0149355Z set +e 2025-11-03T16:59:27.0149523Z set -x 2025-11-03T16:59:27.0149688Z  2025-11-03T16:59:27.0149837Z login() { 2025-11-03T16:59:27.0150188Z  aws ecr get-login-password --region us-east-1 | docker login -u AWS --password-stdin "$1" 2025-11-03T16:59:27.0150558Z } 2025-11-03T16:59:27.0150703Z  2025-11-03T16:59:27.0150849Z retry () { 2025-11-03T16:59:27.0151041Z  $* || (sleep 1 && $*) || (sleep 2 && $*) 2025-11-03T16:59:27.0151275Z } 2025-11-03T16:59:27.0151423Z  2025-11-03T16:59:27.0151579Z retry login "${DOCKER_REGISTRY}" 2025-11-03T16:59:27.0151798Z  2025-11-03T16:59:27.0151952Z START_TIME=$(date +%s) 2025-11-03T16:59:27.0152160Z # Wait up to 120 minutes 2025-11-03T16:59:27.0152422Z while [[ $(( $(date +%s) - 7200 )) -lt $START_TIME ]]; do 2025-11-03T16:59:27.0152762Z  # Check if image already exists, if it does then skip building it 2025-11-03T16:59:27.0153122Z  if docker manifest inspect "${DOCKER_IMAGE}"; then 2025-11-03T16:59:27.0153378Z  exit 0 2025-11-03T16:59:27.0153544Z  fi 2025-11-03T16:59:27.0153688Z  2025-11-03T16:59:27.0153963Z  # NB: This flag is used by Docker build workflow to push the image to ECR, so we can 2025-11-03T16:59:27.0154429Z  # use this to differentiate between the Docker build and regular build jobs. For the 2025-11-03T16:59:27.0154884Z  # latter, it will wait for the Docker images to become available before continuing 2025-11-03T16:59:27.0155240Z  if [ "${DOCKER_PUSH:-false}" == "true" ]; then 2025-11-03T16:59:27.0155519Z  # It's a Docker build job, let's build the image 2025-11-03T16:59:27.0155764Z  break 2025-11-03T16:59:27.0155929Z  else 2025-11-03T16:59:27.0156184Z  # It's a regular build job, wait for the image to become available 2025-11-03T16:59:27.0156470Z  sleep 300 2025-11-03T16:59:27.0156649Z  fi 2025-11-03T16:59:27.0156800Z done 2025-11-03T16:59:27.0156945Z  2025-11-03T16:59:27.0157352Z # NB: This part requires a full checkout. Otherwise, the merge base will 2025-11-03T16:59:27.0157782Z # be empty. The default action would be to continue rebuild the image 2025-11-03T16:59:27.0158148Z if [[ "$BASE_REVISION" = "$(git rev-parse HEAD)" ]]; then 2025-11-03T16:59:27.0158474Z  # if we're on the base branch then use the parent commit 2025-11-03T16:59:27.0158761Z  MERGE_BASE=$(git rev-parse HEAD~) 2025-11-03T16:59:27.0158975Z else 2025-11-03T16:59:27.0159205Z  # otherwise we're on a PR, so use the most recent base commit 2025-11-03T16:59:27.0159536Z  MERGE_BASE=$(git merge-base HEAD "$BASE_REVISION") 2025-11-03T16:59:27.0159791Z fi 2025-11-03T16:59:27.0159938Z  2025-11-03T16:59:27.0160105Z if [[ -z "${MERGE_BASE}" ]]; then 2025-11-03T16:59:27.0160368Z  echo "rebuild=true" >> "${GITHUB_OUTPUT}" 2025-11-03T16:59:27.0160603Z  2025-11-03T16:59:27.0160928Z  echo "Finding merge base only works with full checkout, please set fetch-depth to 0, continuing ..." 2025-11-03T16:59:27.0161450Z  exit 0 2025-11-03T16:59:27.0161612Z fi 2025-11-03T16:59:27.0161758Z  2025-11-03T16:59:27.0161987Z if ! git rev-parse "${MERGE_BASE}:${DOCKER_BUILD_DIR}"; then 2025-11-03T16:59:27.0162462Z  echo "Directory '${DOCKER_BUILD_DIR}' not found in commit $MERGE_BASE, you should rebase onto a more recent commit" 2025-11-03T16:59:27.0162862Z  exit 1 2025-11-03T16:59:27.0163011Z fi 2025-11-03T16:59:27.0175962Z  2025-11-03T16:59:27.0176261Z PREVIOUS_DOCKER_TAG=$(git rev-parse "${MERGE_BASE}:${DOCKER_BUILD_DIR}") 2025-11-03T16:59:27.0176738Z # If no image exists but the hash is the same as the previous hash then we should error out here 2025-11-03T16:59:27.0177157Z if [[ "${PREVIOUS_DOCKER_TAG}" == "${DOCKER_TAG}" ]]; then 2025-11-03T16:59:27.0177664Z  echo "WARNING: Something has gone wrong and the previous image isn't available for the merge-base of your branch" 2025-11-03T16:59:27.0178212Z  echo " Will re-build docker image to store in local cache, TTS may be longer" 2025-11-03T16:59:27.0178529Z fi 2025-11-03T16:59:27.0178681Z  2025-11-03T16:59:27.0178867Z echo "rebuild=true" >> "${GITHUB_OUTPUT}" 2025-11-03T16:59:27.0186236Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2025-11-03T16:59:27.0186521Z env: 2025-11-03T16:59:27.0186691Z GIT_DEFAULT_BRANCH: main 2025-11-03T16:59:27.0186895Z DOCKER_BUILD_DIR: .ci/docker 2025-11-03T16:59:27.0187152Z BASE_REVISION: 3f6538febd727b782e6e13cfd026a309fb14351d 2025-11-03T16:59:27.0187842Z DOCKER_IMAGE: 308535385114.dkr.ecr.us-east-1.amazonaws.com/pytorch/ci-image:pytorch-linux-jammy-cuda12.8-cudnn9-py3.12-gcc11-vllm-30147fb242f88ab8f13e14b76ab186d189d8605c 2025-11-03T16:59:27.0188706Z DOCKER_TAG: pytorch-linux-jammy-cuda12.8-cudnn9-py3.12-gcc11-vllm-30147fb242f88ab8f13e14b76ab186d189d8605c 2025-11-03T16:59:27.0189234Z DOCKER_REGISTRY: 308535385114.dkr.ecr.us-east-1.amazonaws.com 2025-11-03T16:59:27.0189523Z DOCKER_PUSH: 2025-11-03T16:59:27.0189688Z ##[endgroup] 2025-11-03T16:59:27.0222166Z + retry login 308535385114.dkr.ecr.us-east-1.amazonaws.com 2025-11-03T16:59:27.0223004Z + login 308535385114.dkr.ecr.us-east-1.amazonaws.com 2025-11-03T16:59:27.0227000Z + docker login -u AWS --password-stdin 308535385114.dkr.ecr.us-east-1.amazonaws.com 2025-11-03T16:59:27.0227934Z + aws ecr get-login-password --region us-east-1 2025-11-03T16:59:27.4860558Z WARNING! Your password will be stored unencrypted in /home/ec2-user/.docker/config.json. 2025-11-03T16:59:27.4861555Z Configure a credential helper to remove this warning. See 2025-11-03T16:59:27.4862462Z https://docs.docker.com/engine/reference/commandline/login/#credentials-store 2025-11-03T16:59:27.4864979Z 2025-11-03T16:59:27.4865120Z Login Succeeded 2025-11-03T16:59:27.4893482Z ++ date +%s 2025-11-03T16:59:27.4911530Z + START_TIME=1762189167 2025-11-03T16:59:27.4916798Z ++ date +%s 2025-11-03T16:59:27.4929132Z + [[ 1762181967 -lt 1762189167 ]] 2025-11-03T16:59:27.4930741Z + docker manifest inspect 308535385114.dkr.ecr.us-east-1.amazonaws.com/pytorch/ci-image:pytorch-linux-jammy-cuda12.8-cudnn9-py3.12-gcc11-vllm-30147fb242f88ab8f13e14b76ab186d189d8605c 2025-11-03T16:59:27.7833458Z { 2025-11-03T16:59:27.7833701Z "schemaVersion": 2, 2025-11-03T16:59:27.7834310Z "mediaType": "application/vnd.docker.distribution.manifest.v2+json", 2025-11-03T16:59:27.7834775Z "config": { 2025-11-03T16:59:27.7835100Z "mediaType": "application/vnd.docker.container.image.v1+json", 2025-11-03T16:59:27.7835478Z "size": 32985, 2025-11-03T16:59:27.7835880Z "digest": "sha256:123a0a8d2d8124dd2cca825f1276a192d04ee365eb7de45a4829163bae42bcc5" 2025-11-03T16:59:27.7836294Z }, 2025-11-03T16:59:27.7836472Z "layers": [ 2025-11-03T16:59:27.7836654Z { 2025-11-03T16:59:27.7837123Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-11-03T16:59:27.7837509Z "size": 30447990, 2025-11-03T16:59:27.7837910Z "digest": "sha256:828c1365039a657352c737a62d13e1932951b5658eb6bd9b9096ea9b73562453" 2025-11-03T16:59:27.7838708Z }, 2025-11-03T16:59:27.7838869Z { 2025-11-03T16:59:27.7839165Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-11-03T16:59:27.7839529Z "size": 1552, 2025-11-03T16:59:27.7839887Z "digest": "sha256:1c4ee0d43392f96acf52d2e8e2f6257611a1334051823ff62d69c216dbc7daca" 2025-11-03T16:59:27.7840284Z }, 2025-11-03T16:59:27.7840459Z { 2025-11-03T16:59:27.7840738Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-11-03T16:59:27.7841106Z "size": 313626101, 2025-11-03T16:59:27.7841480Z "digest": "sha256:728590afcb21193f16499a4744dada5e256b1413ed8ced7560c4eb4543953d2b" 2025-11-03T16:59:27.7841963Z }, 2025-11-03T16:59:27.7842104Z { 2025-11-03T16:59:27.7842332Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-11-03T16:59:27.7842613Z "size": 791, 2025-11-03T16:59:27.7842914Z "digest": "sha256:9a925ed696bfcc8ebda91a3c220252d69635e24adf4fc2835afd1a4aa9e601f6" 2025-11-03T16:59:27.7843250Z }, 2025-11-03T16:59:27.7843378Z { 2025-11-03T16:59:27.7843594Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-11-03T16:59:27.7843875Z "size": 106, 2025-11-03T16:59:27.7844158Z "digest": "sha256:e7ab1efdd1bd8cfda48e301193b8e1eb058d2c83851044818db8099ff597ce35" 2025-11-03T16:59:27.7844483Z }, 2025-11-03T16:59:27.7844606Z { 2025-11-03T16:59:27.7844822Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-11-03T16:59:27.7845103Z "size": 703, 2025-11-03T16:59:27.7845375Z "digest": "sha256:33a0c5e93563ea73282e573325b1527d301f618dbec76bb05c07228cf1c69230" 2025-11-03T16:59:27.7845682Z }, 2025-11-03T16:59:27.7845814Z { 2025-11-03T16:59:27.7846044Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-11-03T16:59:27.7846326Z "size": 1219, 2025-11-03T16:59:27.7846612Z "digest": "sha256:7401e108f2b86e7530e0d9f8fa1dea4ae5f79c68810edceb2a7abdd1630ee9d6" 2025-11-03T16:59:27.7846942Z }, 2025-11-03T16:59:27.7847071Z { 2025-11-03T16:59:27.7847291Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-11-03T16:59:27.7847566Z "size": 484, 2025-11-03T16:59:27.7847847Z "digest": "sha256:9d9c7be466f87846f48df125fcf7b25bc5d633cefbf0584ed645d3ab90c043bd" 2025-11-03T16:59:27.7848170Z }, 2025-11-03T16:59:27.7848302Z { 2025-11-03T16:59:27.7848614Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-11-03T16:59:27.7848895Z "size": 110363285, 2025-11-03T16:59:27.7849200Z "digest": "sha256:faa9a1190c2e97d043159143a313b91d39d5a6068aee31e2f2708a4eac04afa1" 2025-11-03T16:59:27.7849523Z }, 2025-11-03T16:59:27.7849648Z { 2025-11-03T16:59:27.7849865Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-11-03T16:59:27.7850145Z "size": 5132, 2025-11-03T16:59:27.7850432Z "digest": "sha256:7c2887fa8aa0685c83bbb51790f6890622425f1e8bfc2aa5ae9a6156505842a9" 2025-11-03T16:59:27.7850910Z }, 2025-11-03T16:59:27.7851064Z { 2025-11-03T16:59:27.7851280Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-11-03T16:59:27.7851560Z "size": 1755, 2025-11-03T16:59:27.7851841Z "digest": "sha256:95bb0abcaee9af0c3aafb8021c45ce9381c0a1e2c8b4c6a394a49b98107f8102" 2025-11-03T16:59:27.7852166Z }, 2025-11-03T16:59:27.7852293Z { 2025-11-03T16:59:27.7852513Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-11-03T16:59:27.7852833Z "size": 724, 2025-11-03T16:59:27.7853110Z "digest": "sha256:b3a429f4faee83e3d7e616cbcca1b458319d213977b4807ea863fc416abb2229" 2025-11-03T16:59:27.7853429Z }, 2025-11-03T16:59:27.7853563Z { 2025-11-03T16:59:27.7853781Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-11-03T16:59:27.7854061Z "size": 544, 2025-11-03T16:59:27.7854338Z "digest": "sha256:f4cb10b91a16ac44a669c09f1037854ceb9884ac4c6a118da74679c3f2629f24" 2025-11-03T16:59:27.7854653Z }, 2025-11-03T16:59:27.7854780Z { 2025-11-03T16:59:27.7855096Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-11-03T16:59:27.7855384Z "size": 3365931269, 2025-11-03T16:59:27.7855677Z "digest": "sha256:48618db2b1a6e401f30542bcf94fe34cdac711cdc8811c56800d9b013220cf85" 2025-11-03T16:59:27.7855988Z }, 2025-11-03T16:59:27.7856121Z { 2025-11-03T16:59:27.7856339Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-11-03T16:59:27.7856614Z "size": 32, 2025-11-03T16:59:27.7856887Z "digest": "sha256:4f4fb700ef54461cfa02571ae0db9a0dc1e0cdb5577484a6d75e68dc38e8acc1" 2025-11-03T16:59:27.7857214Z }, 2025-11-03T16:59:27.7857340Z { 2025-11-03T16:59:27.7857560Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-11-03T16:59:27.7857847Z "size": 380, 2025-11-03T16:59:27.7858139Z "digest": "sha256:2ec7ba999b860bc831c8e5b4e42a1087c8cfde1d02d8f2b3b0a1a297def702b1" 2025-11-03T16:59:27.7858470Z }, 2025-11-03T16:59:27.7858604Z { 2025-11-03T16:59:27.7858829Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-11-03T16:59:27.7859127Z "size": 236652, 2025-11-03T16:59:27.7859423Z "digest": "sha256:85d9004dcaf2a14dc2db8fb4caf12cc30e40747d20860a246e7c99fce1029e90" 2025-11-03T16:59:27.7859749Z }, 2025-11-03T16:59:27.7859901Z { 2025-11-03T16:59:27.7860125Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-11-03T16:59:27.7860413Z "size": 230, 2025-11-03T16:59:27.7860696Z "digest": "sha256:df8f8108c158cd5519b0ea50bd22522f354b19bf2388184e23faa3733b3fd6e3" 2025-11-03T16:59:27.7861015Z }, 2025-11-03T16:59:27.7861148Z { 2025-11-03T16:59:27.7861370Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-11-03T16:59:27.7861645Z "size": 3267656, 2025-11-03T16:59:27.7861931Z "digest": "sha256:329fb7ac2c9ea86ede71036735c0b2f3b2f388129235a33e002409205f4514f1" 2025-11-03T16:59:27.7862250Z }, 2025-11-03T16:59:27.7862379Z { 2025-11-03T16:59:27.7862595Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-11-03T16:59:27.7862879Z "size": 1478, 2025-11-03T16:59:27.7863152Z "digest": "sha256:397c5718353c9270969b965bfddacd659394a4189be78c230558a674cbde1a2c" 2025-11-03T16:59:27.7863459Z }, 2025-11-03T16:59:27.7863580Z { 2025-11-03T16:59:27.7863795Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-11-03T16:59:27.7864071Z "size": 481, 2025-11-03T16:59:27.7864338Z "digest": "sha256:099139c9d8b455920b3e71991a0af82a00315f2708e1b8398089dad4889d0c82" 2025-11-03T16:59:27.7864699Z }, 2025-11-03T16:59:27.7864925Z { 2025-11-03T16:59:27.7865321Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-11-03T16:59:27.7865760Z "size": 198, 2025-11-03T16:59:27.7866048Z "digest": "sha256:c094f15f3938d9123ed1f3a086212ccf36d3a0471de7dbfdbe14a185e16cb362" 2025-11-03T16:59:27.7866375Z }, 2025-11-03T16:59:27.7866507Z { 2025-11-03T16:59:27.7866731Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-11-03T16:59:27.7867124Z "size": 607, 2025-11-03T16:59:27.7867426Z "digest": "sha256:a99b38a3dcdd8fe005f4f28ca62e73a4c465ac67ea320f9cfee2fcae35797245" 2025-11-03T16:59:27.7867754Z }, 2025-11-03T16:59:27.7867884Z { 2025-11-03T16:59:27.7868102Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-11-03T16:59:27.7868388Z "size": 7871224158, 2025-11-03T16:59:27.7868686Z "digest": "sha256:bc183ac25ebc0a82c062b5aa9d512109c3349be9dada7361068edf5814ac9a76" 2025-11-03T16:59:27.7869024Z }, 2025-11-03T16:59:27.7869164Z { 2025-11-03T16:59:27.7869380Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-11-03T16:59:27.7869661Z "size": 830, 2025-11-03T16:59:27.7869934Z "digest": "sha256:1a6a81cb899b0f9a55045ca3f89823d22816308be9440327b64990ae8108de42" 2025-11-03T16:59:27.7870245Z }, 2025-11-03T16:59:27.7870368Z { 2025-11-03T16:59:27.7870586Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-11-03T16:59:27.7870875Z "size": 33451730, 2025-11-03T16:59:27.7871176Z "digest": "sha256:41ec23fc0915ee04c6a7f6142343834517dcca5c446aa66557f147eafc52de95" 2025-11-03T16:59:27.7871582Z }, 2025-11-03T16:59:27.7871712Z { 2025-11-03T16:59:27.7871933Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-11-03T16:59:27.7872218Z "size": 104, 2025-11-03T16:59:27.7872496Z "digest": "sha256:540c0b6aa5f3cd25baec56a97c4cad8bb1cfaee48f742276c41a24fa07ff619a" 2025-11-03T16:59:27.7872823Z }, 2025-11-03T16:59:27.7872955Z { 2025-11-03T16:59:27.7873172Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-11-03T16:59:27.7873447Z "size": 1496, 2025-11-03T16:59:27.7873745Z "digest": "sha256:5f740b394ea310b0c2f28c055d3ba824427d5b2b55a9226ac4039be54b839698" 2025-11-03T16:59:27.7874065Z }, 2025-11-03T16:59:27.7874192Z { 2025-11-03T16:59:27.7874405Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-11-03T16:59:27.7874690Z "size": 458410283, 2025-11-03T16:59:27.7874991Z "digest": "sha256:69c4fff3ff1a24255d59f1b56a0f00a70f25b3b1f627fe4d5c330d8c15c83b1b" 2025-11-03T16:59:27.7875326Z }, 2025-11-03T16:59:27.7875452Z { 2025-11-03T16:59:27.7875672Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-11-03T16:59:27.7875957Z "size": 163, 2025-11-03T16:59:27.7876230Z "digest": "sha256:ac00f880c769556a073909f55054fea6299ed9137a1bcf4478f23acfd73e2fdf" 2025-11-03T16:59:27.7876540Z }, 2025-11-03T16:59:27.7876669Z { 2025-11-03T16:59:27.7876890Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-11-03T16:59:27.7877165Z "size": 346, 2025-11-03T16:59:27.7877430Z "digest": "sha256:c084e8e1076b1f971aad4348e38c6204c20c0d3b768959f05a5f7e298c600d46" 2025-11-03T16:59:27.7877744Z }, 2025-11-03T16:59:27.7877875Z { 2025-11-03T16:59:27.7878107Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-11-03T16:59:27.7878380Z "size": 32, 2025-11-03T16:59:27.7878664Z "digest": "sha256:4f4fb700ef54461cfa02571ae0db9a0dc1e0cdb5577484a6d75e68dc38e8acc1" 2025-11-03T16:59:27.7878994Z }, 2025-11-03T16:59:27.7879128Z { 2025-11-03T16:59:27.7879345Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-11-03T16:59:27.7879628Z "size": 106, 2025-11-03T16:59:27.7879904Z "digest": "sha256:a300a654b9ecbe91618f11f1bf727a7e1df8e1819158bc0c647fa85654f25a9b" 2025-11-03T16:59:27.7880223Z }, 2025-11-03T16:59:27.7880357Z { 2025-11-03T16:59:27.7880580Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-11-03T16:59:27.7880858Z "size": 425, 2025-11-03T16:59:27.7881131Z "digest": "sha256:6345aec6eb298d06b276130fd8c432205e01f3090bf30bce281256fb484b6f4c" 2025-11-03T16:59:27.7881439Z }, 2025-11-03T16:59:27.7881570Z { 2025-11-03T16:59:27.7881787Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-11-03T16:59:27.7882068Z "size": 19309365, 2025-11-03T16:59:27.7882354Z "digest": "sha256:e7298879fae82b1621bb2ec3afaec102f5d3d19babf009f2af49f96fd1cefe6a" 2025-11-03T16:59:27.7882783Z }, 2025-11-03T16:59:27.7882923Z { 2025-11-03T16:59:27.7883145Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-11-03T16:59:27.7883419Z "size": 108, 2025-11-03T16:59:27.7883698Z "digest": "sha256:943364a6285e1bc14c6e7f9a03efb0b26a1d40da8d4ef89a6284ad066ad3aec7" 2025-11-03T16:59:27.7884021Z }, 2025-11-03T16:59:27.7884155Z { 2025-11-03T16:59:27.7884365Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-11-03T16:59:27.7884646Z "size": 639, 2025-11-03T16:59:27.7884916Z "digest": "sha256:14f0993d5468420d7d0c2454a7b63ebd7a0a1c8a27674982e3a4427f49ee1b4c" 2025-11-03T16:59:27.7885232Z }, 2025-11-03T16:59:27.7885356Z { 2025-11-03T16:59:27.7885574Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-11-03T16:59:27.7885858Z "size": 724, 2025-11-03T16:59:27.7886139Z "digest": "sha256:b3a429f4faee83e3d7e616cbcca1b458319d213977b4807ea863fc416abb2229" 2025-11-03T16:59:27.7886455Z }, 2025-11-03T16:59:27.7886586Z { 2025-11-03T16:59:27.7886889Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-11-03T16:59:27.7887165Z "size": 149, 2025-11-03T16:59:27.7887432Z "digest": "sha256:d6a558a7de5e9dd5633745ba04365d261f0c5ef255d68335a1e34d2393acabd6" 2025-11-03T16:59:27.7887746Z }, 2025-11-03T16:59:27.7887878Z { 2025-11-03T16:59:27.7888097Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-11-03T16:59:27.7888466Z "size": 135, 2025-11-03T16:59:27.7888743Z "digest": "sha256:05ee86267141af52f7acaa389053961b9b3af580b4cfa52635fb11c3cc9ac77a" 2025-11-03T16:59:27.7889061Z }, 2025-11-03T16:59:27.7889191Z { 2025-11-03T16:59:27.7889400Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-11-03T16:59:27.7889679Z "size": 140, 2025-11-03T16:59:27.7889963Z "digest": "sha256:f0ccf17afa340e506e78617ef552a5e1bcde2d3a8f329d0741997cabdd68b3df" 2025-11-03T16:59:27.7890298Z }, 2025-11-03T16:59:27.7890426Z { 2025-11-03T16:59:27.7890648Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-11-03T16:59:27.7890932Z "size": 32, 2025-11-03T16:59:27.7891218Z "digest": "sha256:4f4fb700ef54461cfa02571ae0db9a0dc1e0cdb5577484a6d75e68dc38e8acc1" 2025-11-03T16:59:27.7891534Z }, 2025-11-03T16:59:27.7891666Z { 2025-11-03T16:59:27.7891886Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-11-03T16:59:27.7892165Z "size": 221, 2025-11-03T16:59:27.7892439Z "digest": "sha256:a68d4d73adc3e7837c49e151841fa4d2ac9ad874ffae2ae9f6582d4da63ee803" 2025-11-03T16:59:27.7892765Z }, 2025-11-03T16:59:27.7892893Z { 2025-11-03T16:59:27.7893122Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-11-03T16:59:27.7893399Z "size": 255, 2025-11-03T16:59:27.7893677Z "digest": "sha256:1ef3a6930902936a8a4d2f964a465c0b4ceecec2fd03a78c0e5b4a4174a6dcc7" 2025-11-03T16:59:27.7893999Z }, 2025-11-03T16:59:27.7894127Z { 2025-11-03T16:59:27.7894342Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-11-03T16:59:27.7894625Z "size": 32, 2025-11-03T16:59:27.7894907Z "digest": "sha256:4f4fb700ef54461cfa02571ae0db9a0dc1e0cdb5577484a6d75e68dc38e8acc1" 2025-11-03T16:59:27.7895232Z }, 2025-11-03T16:59:27.7895358Z { 2025-11-03T16:59:27.7895592Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-11-03T16:59:27.7895876Z "size": 106, 2025-11-03T16:59:27.7896151Z "digest": "sha256:ba2105a84ce6992ef1d41f379ee51bbd5b34872691f4924d2ca3a46aee8ffe61" 2025-11-03T16:59:27.7896467Z }, 2025-11-03T16:59:27.7896598Z { 2025-11-03T16:59:27.7896819Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-11-03T16:59:27.7897106Z "size": 312468923, 2025-11-03T16:59:27.7897387Z "digest": "sha256:85c483845f299fec36365e6fb5d71ad97d8f0e1478233c63d51699e16212a287" 2025-11-03T16:59:27.7897712Z }, 2025-11-03T16:59:27.7897838Z { 2025-11-03T16:59:27.7898053Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-11-03T16:59:27.7898456Z "size": 3222059909, 2025-11-03T16:59:27.7898770Z "digest": "sha256:ce611413a46e5e0df2bb13c62f1fcaa11d5e2e474b803d0df2f5ef3daa50f9e9" 2025-11-03T16:59:27.7899097Z }, 2025-11-03T16:59:27.7899223Z { 2025-11-03T16:59:27.7899437Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-11-03T16:59:27.7899714Z "size": 128, 2025-11-03T16:59:27.7899985Z "digest": "sha256:2b12a35a9605735c61f0f93dbabfe19540a2849f46f257f8eb5bf2ff0e2f37e2" 2025-11-03T16:59:27.7900303Z }, 2025-11-03T16:59:27.7900439Z { 2025-11-03T16:59:27.7900656Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-11-03T16:59:27.7900934Z "size": 880, 2025-11-03T16:59:27.7901204Z "digest": "sha256:2a86c078c73d1d0224e1046a3e2ca94663ca04652ec9ac44700e7fd6a4d9fc9c" 2025-11-03T16:59:27.7901516Z }, 2025-11-03T16:59:27.7901647Z { 2025-11-03T16:59:27.7901871Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-11-03T16:59:27.7902148Z "size": 724, 2025-11-03T16:59:27.7902421Z "digest": "sha256:b3a429f4faee83e3d7e616cbcca1b458319d213977b4807ea863fc416abb2229" 2025-11-03T16:59:27.7902825Z }, 2025-11-03T16:59:27.7902953Z { 2025-11-03T16:59:27.7903170Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-11-03T16:59:27.7903444Z "size": 139, 2025-11-03T16:59:27.7903719Z "digest": "sha256:44e5c47c9ce4024f40e82cfb276df42771be0a6991ba260f37a9789f3caee04a" 2025-11-03T16:59:27.7904040Z }, 2025-11-03T16:59:27.7904170Z { 2025-11-03T16:59:27.7904394Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-11-03T16:59:27.7904676Z "size": 32, 2025-11-03T16:59:27.7904957Z "digest": "sha256:4f4fb700ef54461cfa02571ae0db9a0dc1e0cdb5577484a6d75e68dc38e8acc1" 2025-11-03T16:59:27.7905281Z }, 2025-11-03T16:59:27.7905408Z { 2025-11-03T16:59:27.7905628Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-11-03T16:59:27.7905909Z "size": 159, 2025-11-03T16:59:27.7906205Z "digest": "sha256:92d25d2ab4f36bf32d7e163eeaf2c24a582062777d0bc5231fbb32482184cfd7" 2025-11-03T16:59:27.7906533Z }, 2025-11-03T16:59:27.7906668Z { 2025-11-03T16:59:27.7906889Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-11-03T16:59:27.7907171Z "size": 1011, 2025-11-03T16:59:27.7907442Z "digest": "sha256:07c8d78a73e65a5453a1045bb24c0080b003ffc3168f56f16fea4c30f9581f5d" 2025-11-03T16:59:27.7907762Z }, 2025-11-03T16:59:27.7907892Z { 2025-11-03T16:59:27.7908113Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-11-03T16:59:27.7908392Z "size": 724, 2025-11-03T16:59:27.7908670Z "digest": "sha256:b3a429f4faee83e3d7e616cbcca1b458319d213977b4807ea863fc416abb2229" 2025-11-03T16:59:27.7908996Z }, 2025-11-03T16:59:27.7909126Z { 2025-11-03T16:59:27.7909335Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-11-03T16:59:27.7909952Z "size": 134, 2025-11-03T16:59:27.7910239Z "digest": "sha256:66372f032f924ed974c345878c60ae95400e3826d05cadb854b5b378c86fef10" 2025-11-03T16:59:27.7910552Z }, 2025-11-03T16:59:27.7910680Z { 2025-11-03T16:59:27.7910917Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-11-03T16:59:27.7911199Z "size": 32, 2025-11-03T16:59:27.7911477Z "digest": "sha256:4f4fb700ef54461cfa02571ae0db9a0dc1e0cdb5577484a6d75e68dc38e8acc1" 2025-11-03T16:59:27.7911802Z }, 2025-11-03T16:59:27.7911925Z { 2025-11-03T16:59:27.7912143Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-11-03T16:59:27.7912422Z "size": 157, 2025-11-03T16:59:27.7912683Z "digest": "sha256:60194b590b00d0238b1e90ea860d06fc775fe687151dc536f5d082458f0a386c" 2025-11-03T16:59:27.7912997Z }, 2025-11-03T16:59:27.7913123Z { 2025-11-03T16:59:27.7913339Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-11-03T16:59:27.7913630Z "size": 1371, 2025-11-03T16:59:27.7913913Z "digest": "sha256:b41ffa5a6a4588148bf64c1e6d509af9d49b9ff7f27f4f7be606e4c0a48f3a12" 2025-11-03T16:59:27.7914241Z }, 2025-11-03T16:59:27.7914518Z { 2025-11-03T16:59:27.7914741Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-11-03T16:59:27.7915023Z "size": 32, 2025-11-03T16:59:27.7915301Z "digest": "sha256:4f4fb700ef54461cfa02571ae0db9a0dc1e0cdb5577484a6d75e68dc38e8acc1" 2025-11-03T16:59:27.7915624Z }, 2025-11-03T16:59:27.7915753Z { 2025-11-03T16:59:27.7915977Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-11-03T16:59:27.7916262Z "size": 136, 2025-11-03T16:59:27.7916539Z "digest": "sha256:f165304e20915f2cdcc03f0e977c2db943a17feb53a13343371f12bae7fd74d3" 2025-11-03T16:59:27.7916862Z }, 2025-11-03T16:59:27.7916984Z { 2025-11-03T16:59:27.7917202Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-11-03T16:59:27.7917482Z "size": 528, 2025-11-03T16:59:27.7917758Z "digest": "sha256:16ec736149b0011270995bb0bab5befaafbc716cfdf9f280ff25f75a2072c90d" 2025-11-03T16:59:27.7918075Z }, 2025-11-03T16:59:27.7918204Z { 2025-11-03T16:59:27.7918427Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-11-03T16:59:27.7918830Z "size": 32, 2025-11-03T16:59:27.7919104Z "digest": "sha256:4f4fb700ef54461cfa02571ae0db9a0dc1e0cdb5577484a6d75e68dc38e8acc1" 2025-11-03T16:59:27.7919429Z }, 2025-11-03T16:59:27.7919559Z { 2025-11-03T16:59:27.7919778Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-11-03T16:59:27.7920055Z "size": 104, 2025-11-03T16:59:27.7920342Z "digest": "sha256:ea0ef0105e5f56368e0dbe0d4cea9d772ddf42bb0e3f6b5066fbe970c014a684" 2025-11-03T16:59:27.7920671Z }, 2025-11-03T16:59:27.7920799Z { 2025-11-03T16:59:27.7921032Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-11-03T16:59:27.7921377Z "size": 428, 2025-11-03T16:59:27.7921653Z "digest": "sha256:b294669af2b5786b9bc5601bffdc874e4673b0fcb879e1d77c17e674e7954b10" 2025-11-03T16:59:27.7921974Z }, 2025-11-03T16:59:27.7922095Z { 2025-11-03T16:59:27.7922313Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-11-03T16:59:27.7922598Z "size": 32, 2025-11-03T16:59:27.7922884Z "digest": "sha256:4f4fb700ef54461cfa02571ae0db9a0dc1e0cdb5577484a6d75e68dc38e8acc1" 2025-11-03T16:59:27.7923200Z }, 2025-11-03T16:59:27.7923327Z { 2025-11-03T16:59:27.7923545Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-11-03T16:59:27.7923825Z "size": 107, 2025-11-03T16:59:27.7924094Z "digest": "sha256:726179d41040afffbfeb80d353b6032e11543ff3d0337dc5c67a9238cebefc27" 2025-11-03T16:59:27.7924412Z }, 2025-11-03T16:59:27.7924543Z { 2025-11-03T16:59:27.7924758Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-11-03T16:59:27.7925031Z "size": 1896, 2025-11-03T16:59:27.7925312Z "digest": "sha256:375f6cf74921eec948c69077e8280fd9d9c2a6ccf83f8a33cccc413fc36edca8" 2025-11-03T16:59:27.7925636Z }, 2025-11-03T16:59:27.7925765Z { 2025-11-03T16:59:27.7925977Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-11-03T16:59:27.7926260Z "size": 245097719, 2025-11-03T16:59:27.7926560Z "digest": "sha256:0cc2718afe447c4d220e3db62787c33345e73283935f307a2033aeb8c875b6af" 2025-11-03T16:59:27.7926878Z }, 2025-11-03T16:59:27.7927002Z { 2025-11-03T16:59:27.7927219Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-11-03T16:59:27.7927501Z "size": 106, 2025-11-03T16:59:27.7927768Z "digest": "sha256:cc98ed46781d140520a78924755c1054fee09f9303a3611a63fea0ff97d5b478" 2025-11-03T16:59:27.7928080Z }, 2025-11-03T16:59:27.7928208Z { 2025-11-03T16:59:27.7928477Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-11-03T16:59:27.7928763Z "size": 165, 2025-11-03T16:59:27.7929026Z "digest": "sha256:83ecc1143684a6535b14ce79395dec432eca66768055598312e14a7d4522ad8b" 2025-11-03T16:59:27.7929341Z }, 2025-11-03T16:59:27.7929471Z { 2025-11-03T16:59:27.7929695Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-11-03T16:59:27.7929970Z "size": 7944, 2025-11-03T16:59:27.7930341Z "digest": "sha256:84d7e8a5d95e595a910304a466fdf855f4ffc34878881b1e1a3c9124207fdb5d" 2025-11-03T16:59:27.7930666Z }, 2025-11-03T16:59:27.7930796Z { 2025-11-03T16:59:27.7931011Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-11-03T16:59:27.7931304Z "size": 8070, 2025-11-03T16:59:27.7931576Z "digest": "sha256:9913742e4da406263abf521b5f54b52b61c82f5523433268948ae207565470e4" 2025-11-03T16:59:27.7931892Z }, 2025-11-03T16:59:27.7932015Z { 2025-11-03T16:59:27.7932249Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-11-03T16:59:27.7932533Z "size": 303, 2025-11-03T16:59:27.7932803Z "digest": "sha256:a5d33870f153d309e9c3150f196979fdaee17378d91b1ed8257bad66194c99c6" 2025-11-03T16:59:27.7933115Z }, 2025-11-03T16:59:27.7933243Z { 2025-11-03T16:59:27.7933458Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-11-03T16:59:27.7933743Z "size": 13360444, 2025-11-03T16:59:27.7934020Z "digest": "sha256:50b140b19ce99832006da1669a1feb8202de98a2960e50cbc1d2107212f6de30" 2025-11-03T16:59:27.7934411Z }, 2025-11-03T16:59:27.7934548Z { 2025-11-03T16:59:27.7934767Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-11-03T16:59:27.7935040Z "size": 108, 2025-11-03T16:59:27.7935324Z "digest": "sha256:e7227b2cf3c5ba811ee334ab6ae30e3894df5c07ace96da7bd9d79fa63569e97" 2025-11-03T16:59:27.7935649Z }, 2025-11-03T16:59:27.7935776Z { 2025-11-03T16:59:27.7935982Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-11-03T16:59:27.7936262Z "size": 54145699, 2025-11-03T16:59:27.7936550Z "digest": "sha256:4807408eff38f15afc1aad24567cabdde845bbc9b1f79a1d63509dac466a197b" 2025-11-03T16:59:27.7936875Z }, 2025-11-03T16:59:27.7937001Z { 2025-11-03T16:59:27.7937217Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-11-03T16:59:27.7937500Z "size": 32, 2025-11-03T16:59:27.7937782Z "digest": "sha256:4f4fb700ef54461cfa02571ae0db9a0dc1e0cdb5577484a6d75e68dc38e8acc1" 2025-11-03T16:59:27.7938119Z } 2025-11-03T16:59:27.7938249Z ] 2025-11-03T16:59:27.7938382Z } 2025-11-03T16:59:27.7938633Z + exit 0 2025-11-03T16:59:27.7967428Z ##[group]Run set -eux 2025-11-03T16:59:27.7967652Z set -eux 2025-11-03T16:59:27.7967944Z # It's ok if this steps fails, it would then be an anonymous user like what we used to have 2025-11-03T16:59:27.7968859Z aws secretsmanager get-secret-value --secret-id docker_hub_readonly_token | jq --raw-output '.SecretString' | jq -r .docker_hub_readonly_token | docker login --username pytorchbot --password-stdin || true 2025-11-03T16:59:27.7977356Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2025-11-03T16:59:27.7977641Z env: 2025-11-03T16:59:27.7977802Z GIT_DEFAULT_BRANCH: main 2025-11-03T16:59:27.7977987Z ##[endgroup] 2025-11-03T16:59:27.8017725Z + jq --raw-output .SecretString 2025-11-03T16:59:27.8018458Z + aws secretsmanager get-secret-value --secret-id docker_hub_readonly_token 2025-11-03T16:59:27.8018995Z + jq -r .docker_hub_readonly_token 2025-11-03T16:59:27.8021876Z + docker login --username pytorchbot --password-stdin 2025-11-03T16:59:28.3369062Z WARNING! Your password will be stored unencrypted in /home/ec2-user/.docker/config.json. 2025-11-03T16:59:28.3370143Z Configure a credential helper to remove this warning. See 2025-11-03T16:59:28.3371152Z https://docs.docker.com/engine/reference/commandline/login/#credentials-store 2025-11-03T16:59:28.3372369Z 2025-11-03T16:59:28.3372715Z Login Succeeded 2025-11-03T16:59:28.3481127Z ##[group]Run tag=${ECR_DOCKER_IMAGE##*:} 2025-11-03T16:59:28.3481477Z tag=${ECR_DOCKER_IMAGE##*:} 2025-11-03T16:59:28.3481775Z echo "docker pull ghcr.io/pytorch/ci-image:${tag/:/-}" 2025-11-03T16:59:28.3490144Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2025-11-03T16:59:28.3490423Z env: 2025-11-03T16:59:28.3490597Z GIT_DEFAULT_BRANCH: main 2025-11-03T16:59:28.3491224Z ECR_DOCKER_IMAGE: 308535385114.dkr.ecr.us-east-1.amazonaws.com/pytorch/ci-image:pytorch-linux-jammy-cuda12.8-cudnn9-py3.12-gcc11-vllm-30147fb242f88ab8f13e14b76ab186d189d8605c 2025-11-03T16:59:28.3491874Z ##[endgroup] 2025-11-03T16:59:28.3526235Z docker pull ghcr.io/pytorch/ci-image:pytorch-linux-jammy-cuda12.8-cudnn9-py3.12-gcc11-vllm-30147fb242f88ab8f13e14b76ab186d189d8605c 2025-11-03T16:59:28.3590510Z ##[group]Run pytorch/test-infra/.github/actions/pull-docker-image@main 2025-11-03T16:59:28.3590895Z with: 2025-11-03T16:59:28.3591491Z docker-image: 308535385114.dkr.ecr.us-east-1.amazonaws.com/pytorch/ci-image:pytorch-linux-jammy-cuda12.8-cudnn9-py3.12-gcc11-vllm-30147fb242f88ab8f13e14b76ab186d189d8605c 2025-11-03T16:59:28.3592206Z docker-registry: 308535385114.dkr.ecr.us-east-1.amazonaws.com 2025-11-03T16:59:28.3592474Z env: 2025-11-03T16:59:28.3592631Z GIT_DEFAULT_BRANCH: main 2025-11-03T16:59:28.3592825Z ##[endgroup] 2025-11-03T16:59:28.3611661Z ##[group]Run set -x 2025-11-03T16:59:28.3611859Z set -x 2025-11-03T16:59:28.3612034Z set +e 2025-11-03T16:59:28.3612389Z  2025-11-03T16:59:28.3612547Z login() { 2025-11-03T16:59:28.3612904Z  aws ecr get-login-password --region us-east-1 | docker login -u AWS --password-stdin "$1" 2025-11-03T16:59:28.3613275Z } 2025-11-03T16:59:28.3613435Z  2025-11-03T16:59:28.3613631Z retry () { 2025-11-03T16:59:28.3613839Z  $* || (sleep 1 && $*) || (sleep 2 && $*) 2025-11-03T16:59:28.3614061Z } 2025-11-03T16:59:28.3614217Z  2025-11-03T16:59:28.3614396Z retry login "${DOCKER_REGISTRY}" 2025-11-03T16:59:28.3614627Z  2025-11-03T16:59:28.3614988Z IMAGE_SIZE=$(docker manifest inspect "${DOCKER_IMAGE}" | jq '[.layers[].size, .config.size] | add / 1024 / 1024') 2025-11-03T16:59:28.3615468Z echo "Compressed size of image in MB: ${IMAGE_SIZE}" 2025-11-03T16:59:28.3615736Z  2025-11-03T16:59:28.3615886Z set -e 2025-11-03T16:59:28.3616127Z # ignore output since only exit code is used for conditional 2025-11-03T16:59:28.3616480Z # only pull docker image if it's not available locally 2025-11-03T16:59:28.3616862Z if ! docker inspect --type=image "${DOCKER_IMAGE}" >/dev/null 2>/dev/null; then 2025-11-03T16:59:28.3617217Z  retry docker pull "${DOCKER_IMAGE}" 2025-11-03T16:59:28.3617446Z fi 2025-11-03T16:59:28.3624862Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2025-11-03T16:59:28.3625146Z env: 2025-11-03T16:59:28.3625315Z GIT_DEFAULT_BRANCH: main 2025-11-03T16:59:28.3625946Z DOCKER_IMAGE: 308535385114.dkr.ecr.us-east-1.amazonaws.com/pytorch/ci-image:pytorch-linux-jammy-cuda12.8-cudnn9-py3.12-gcc11-vllm-30147fb242f88ab8f13e14b76ab186d189d8605c 2025-11-03T16:59:28.3626649Z DOCKER_REGISTRY: 308535385114.dkr.ecr.us-east-1.amazonaws.com 2025-11-03T16:59:28.3626946Z ##[endgroup] 2025-11-03T16:59:28.3662238Z + set +e 2025-11-03T16:59:28.3663382Z + retry login 308535385114.dkr.ecr.us-east-1.amazonaws.com 2025-11-03T16:59:28.3664261Z + login 308535385114.dkr.ecr.us-east-1.amazonaws.com 2025-11-03T16:59:28.3666161Z + aws ecr get-login-password --region us-east-1 2025-11-03T16:59:28.3667594Z + docker login -u AWS --password-stdin 308535385114.dkr.ecr.us-east-1.amazonaws.com 2025-11-03T16:59:28.8386283Z WARNING! Your password will be stored unencrypted in /home/ec2-user/.docker/config.json. 2025-11-03T16:59:28.8387380Z Configure a credential helper to remove this warning. See 2025-11-03T16:59:28.8388373Z https://docs.docker.com/engine/reference/commandline/login/#credentials-store 2025-11-03T16:59:28.8389053Z 2025-11-03T16:59:28.8389218Z Login Succeeded 2025-11-03T16:59:28.8418627Z ++ jq '[.layers[].size, .config.size] | add / 1024 / 1024' 2025-11-03T16:59:28.8420366Z ++ docker manifest inspect 308535385114.dkr.ecr.us-east-1.amazonaws.com/pytorch/ci-image:pytorch-linux-jammy-cuda12.8-cudnn9-py3.12-gcc11-vllm-30147fb242f88ab8f13e14b76ab186d189d8605c 2025-11-03T16:59:29.1173851Z + IMAGE_SIZE=15309.793606758118 2025-11-03T16:59:29.1174239Z + echo 'Compressed size of image in MB: 15309.793606758118' 2025-11-03T16:59:29.1174577Z + set -e 2025-11-03T16:59:29.1177707Z + docker inspect --type=image 308535385114.dkr.ecr.us-east-1.amazonaws.com/pytorch/ci-image:pytorch-linux-jammy-cuda12.8-cudnn9-py3.12-gcc11-vllm-30147fb242f88ab8f13e14b76ab186d189d8605c 2025-11-03T16:59:29.1178654Z Compressed size of image in MB: 15309.793606758118 2025-11-03T16:59:29.1339079Z + retry docker pull 308535385114.dkr.ecr.us-east-1.amazonaws.com/pytorch/ci-image:pytorch-linux-jammy-cuda12.8-cudnn9-py3.12-gcc11-vllm-30147fb242f88ab8f13e14b76ab186d189d8605c 2025-11-03T16:59:29.1341609Z + docker pull 308535385114.dkr.ecr.us-east-1.amazonaws.com/pytorch/ci-image:pytorch-linux-jammy-cuda12.8-cudnn9-py3.12-gcc11-vllm-30147fb242f88ab8f13e14b76ab186d189d8605c 2025-11-03T16:59:29.3852974Z pytorch-linux-jammy-cuda12.8-cudnn9-py3.12-gcc11-vllm-30147fb242f88ab8f13e14b76ab186d189d8605c: Pulling from pytorch/ci-image 2025-11-03T16:59:29.3855047Z 828c1365039a: Pulling fs layer 2025-11-03T16:59:29.3856089Z 1c4ee0d43392: Pulling fs layer 2025-11-03T16:59:29.3856559Z 728590afcb21: Pulling fs layer 2025-11-03T16:59:29.3857009Z 9a925ed696bf: Pulling fs layer 2025-11-03T16:59:29.3857505Z e7ab1efdd1bd: Pulling fs layer 2025-11-03T16:59:29.3858013Z 33a0c5e93563: Pulling fs layer 2025-11-03T16:59:29.3858445Z 7401e108f2b8: Pulling fs layer 2025-11-03T16:59:29.3858891Z 9d9c7be466f8: Pulling fs layer 2025-11-03T16:59:29.3859349Z faa9a1190c2e: Pulling fs layer 2025-11-03T16:59:29.3859804Z 7c2887fa8aa0: Pulling fs layer 2025-11-03T16:59:29.3860231Z 95bb0abcaee9: Pulling fs layer 2025-11-03T16:59:29.3860675Z b3a429f4faee: Pulling fs layer 2025-11-03T16:59:29.3861119Z f4cb10b91a16: Pulling fs layer 2025-11-03T16:59:29.3861551Z 48618db2b1a6: Pulling fs layer 2025-11-03T16:59:29.3861978Z 4f4fb700ef54: Pulling fs layer 2025-11-03T16:59:29.3862439Z e7ab1efdd1bd: Waiting 2025-11-03T16:59:29.3862942Z 2ec7ba999b86: Pulling fs layer 2025-11-03T16:59:29.3863418Z 85d9004dcaf2: Pulling fs layer 2025-11-03T16:59:29.3863882Z df8f8108c158: Pulling fs layer 2025-11-03T16:59:29.3864198Z 33a0c5e93563: Waiting 2025-11-03T16:59:29.3864388Z 329fb7ac2c9e: Pulling fs layer 2025-11-03T16:59:29.3864574Z 397c5718353c: Pulling fs layer 2025-11-03T16:59:29.3864755Z 9d9c7be466f8: Waiting 2025-11-03T16:59:29.3864922Z b3a429f4faee: Waiting 2025-11-03T16:59:29.3865089Z 7c2887fa8aa0: Waiting 2025-11-03T16:59:29.3865268Z 099139c9d8b4: Pulling fs layer 2025-11-03T16:59:29.3865453Z 95bb0abcaee9: Waiting 2025-11-03T16:59:29.3865616Z c094f15f3938: Pulling fs layer 2025-11-03T16:59:29.3865807Z a99b38a3dcdd: Pulling fs layer 2025-11-03T16:59:29.3865992Z f4cb10b91a16: Waiting 2025-11-03T16:59:29.3866164Z bc183ac25ebc: Pulling fs layer 2025-11-03T16:59:29.3866353Z 1a6a81cb899b: Pulling fs layer 2025-11-03T16:59:29.3866547Z 41ec23fc0915: Pulling fs layer 2025-11-03T16:59:29.3866728Z 329fb7ac2c9e: Waiting 2025-11-03T16:59:29.3866895Z 540c0b6aa5f3: Pulling fs layer 2025-11-03T16:59:29.3867068Z 397c5718353c: Waiting 2025-11-03T16:59:29.3867227Z 85d9004dcaf2: Waiting 2025-11-03T16:59:29.3867395Z df8f8108c158: Waiting 2025-11-03T16:59:29.3867566Z 5f740b394ea3: Pulling fs layer 2025-11-03T16:59:29.3867753Z 69c4fff3ff1a: Pulling fs layer 2025-11-03T16:59:29.3867947Z 2ec7ba999b86: Waiting 2025-11-03T16:59:29.3868109Z 7401e108f2b8: Waiting 2025-11-03T16:59:29.3868273Z 9a925ed696bf: Waiting 2025-11-03T16:59:29.3868423Z 099139c9d8b4: Waiting 2025-11-03T16:59:29.3868593Z ac00f880c769: Pulling fs layer 2025-11-03T16:59:29.3868788Z c084e8e1076b: Pulling fs layer 2025-11-03T16:59:29.3868968Z c094f15f3938: Waiting 2025-11-03T16:59:29.3869123Z a99b38a3dcdd: Waiting 2025-11-03T16:59:29.3869291Z 69c4fff3ff1a: Waiting 2025-11-03T16:59:29.3869457Z 5f740b394ea3: Waiting 2025-11-03T16:59:29.3869626Z 41ec23fc0915: Waiting 2025-11-03T16:59:29.3869804Z a300a654b9ec: Pulling fs layer 2025-11-03T16:59:29.3869995Z ac00f880c769: Waiting 2025-11-03T16:59:29.3870172Z 6345aec6eb29: Pulling fs layer 2025-11-03T16:59:29.3870365Z e7298879fae8: Pulling fs layer 2025-11-03T16:59:29.3870549Z 943364a6285e: Pulling fs layer 2025-11-03T16:59:29.3870747Z 14f0993d5468: Pulling fs layer 2025-11-03T16:59:29.3870930Z a300a654b9ec: Waiting 2025-11-03T16:59:29.3871092Z 943364a6285e: Waiting 2025-11-03T16:59:29.3871248Z e7298879fae8: Waiting 2025-11-03T16:59:29.3871565Z c084e8e1076b: Waiting 2025-11-03T16:59:29.3871729Z 14f0993d5468: Waiting 2025-11-03T16:59:29.3871880Z 6345aec6eb29: Waiting 2025-11-03T16:59:29.3872046Z d6a558a7de5e: Pulling fs layer 2025-11-03T16:59:29.3872233Z 05ee86267141: Pulling fs layer 2025-11-03T16:59:29.3872432Z f0ccf17afa34: Pulling fs layer 2025-11-03T16:59:29.3872612Z 1a6a81cb899b: Waiting 2025-11-03T16:59:29.3872783Z a68d4d73adc3: Pulling fs layer 2025-11-03T16:59:29.3872974Z 1ef3a6930902: Pulling fs layer 2025-11-03T16:59:29.3873153Z d6a558a7de5e: Waiting 2025-11-03T16:59:29.3873310Z f0ccf17afa34: Waiting 2025-11-03T16:59:29.3873471Z a68d4d73adc3: Waiting 2025-11-03T16:59:29.3873629Z 05ee86267141: Waiting 2025-11-03T16:59:29.3873785Z 1ef3a6930902: Waiting 2025-11-03T16:59:29.3874041Z ba2105a84ce6: Pulling fs layer 2025-11-03T16:59:29.3874229Z 85c483845f29: Pulling fs layer 2025-11-03T16:59:29.3874420Z ce611413a46e: Pulling fs layer 2025-11-03T16:59:29.3874605Z 2b12a35a9605: Pulling fs layer 2025-11-03T16:59:29.3874782Z 2a86c078c73d: Pulling fs layer 2025-11-03T16:59:29.3874970Z faa9a1190c2e: Waiting 2025-11-03T16:59:29.3875140Z 44e5c47c9ce4: Pulling fs layer 2025-11-03T16:59:29.3875327Z 85c483845f29: Waiting 2025-11-03T16:59:29.3875480Z ce611413a46e: Waiting 2025-11-03T16:59:29.3875640Z 2b12a35a9605: Waiting 2025-11-03T16:59:29.3875796Z 2a86c078c73d: Waiting 2025-11-03T16:59:29.3875961Z 92d25d2ab4f3: Pulling fs layer 2025-11-03T16:59:29.3876138Z 44e5c47c9ce4: Waiting 2025-11-03T16:59:29.3876308Z 07c8d78a73e6: Pulling fs layer 2025-11-03T16:59:29.3876490Z 92d25d2ab4f3: Waiting 2025-11-03T16:59:29.3876656Z 66372f032f92: Pulling fs layer 2025-11-03T16:59:29.3876834Z 60194b590b00: Pulling fs layer 2025-11-03T16:59:29.3877018Z 66372f032f92: Waiting 2025-11-03T16:59:29.3877187Z b41ffa5a6a45: Pulling fs layer 2025-11-03T16:59:29.3877395Z 60194b590b00: Waiting 2025-11-03T16:59:29.3877550Z b41ffa5a6a45: Waiting 2025-11-03T16:59:29.3877718Z f165304e2091: Pulling fs layer 2025-11-03T16:59:29.3877908Z 16ec736149b0: Pulling fs layer 2025-11-03T16:59:29.3878102Z ea0ef0105e5f: Pulling fs layer 2025-11-03T16:59:29.3878294Z f165304e2091: Waiting 2025-11-03T16:59:29.3878464Z b294669af2b5: Pulling fs layer 2025-11-03T16:59:29.3878652Z ba2105a84ce6: Waiting 2025-11-03T16:59:29.3878820Z 726179d41040: Pulling fs layer 2025-11-03T16:59:29.3879004Z 16ec736149b0: Waiting 2025-11-03T16:59:29.3879174Z 375f6cf74921: Pulling fs layer 2025-11-03T16:59:29.3879358Z b294669af2b5: Waiting 2025-11-03T16:59:29.3879522Z ea0ef0105e5f: Waiting 2025-11-03T16:59:29.3879675Z 726179d41040: Waiting 2025-11-03T16:59:29.3879841Z 0cc2718afe44: Pulling fs layer 2025-11-03T16:59:29.3880034Z cc98ed46781d: Pulling fs layer 2025-11-03T16:59:29.3880214Z 375f6cf74921: Waiting 2025-11-03T16:59:29.3880372Z 83ecc1143684: Pulling fs layer 2025-11-03T16:59:29.3880572Z 84d7e8a5d95e: Pulling fs layer 2025-11-03T16:59:29.3880761Z bc183ac25ebc: Waiting 2025-11-03T16:59:29.3880918Z 0cc2718afe44: Waiting 2025-11-03T16:59:29.3881078Z 9913742e4da4: Pulling fs layer 2025-11-03T16:59:29.3881259Z 83ecc1143684: Waiting 2025-11-03T16:59:29.3881428Z 84d7e8a5d95e: Waiting 2025-11-03T16:59:29.3881593Z a5d33870f153: Pulling fs layer 2025-11-03T16:59:29.3881772Z 50b140b19ce9: Pulling fs layer 2025-11-03T16:59:29.3881959Z e7227b2cf3c5: Pulling fs layer 2025-11-03T16:59:29.3882145Z 4807408eff38: Pulling fs layer 2025-11-03T16:59:29.3882325Z a5d33870f153: Waiting 2025-11-03T16:59:29.3882480Z 9913742e4da4: Waiting 2025-11-03T16:59:29.3882637Z cc98ed46781d: Waiting 2025-11-03T16:59:29.3882805Z 4f4fb700ef54: Waiting 2025-11-03T16:59:29.3882965Z 50b140b19ce9: Waiting 2025-11-03T16:59:29.3883115Z 4807408eff38: Waiting 2025-11-03T16:59:29.3883274Z e7227b2cf3c5: Waiting 2025-11-03T16:59:29.4807517Z 1c4ee0d43392: Verifying Checksum 2025-11-03T16:59:29.4807821Z 1c4ee0d43392: Download complete 2025-11-03T16:59:29.5833919Z 9a925ed696bf: Verifying Checksum 2025-11-03T16:59:29.5834496Z 9a925ed696bf: Download complete 2025-11-03T16:59:29.6843216Z e7ab1efdd1bd: Verifying Checksum 2025-11-03T16:59:29.6843802Z e7ab1efdd1bd: Download complete 2025-11-03T16:59:29.7552215Z 828c1365039a: Verifying Checksum 2025-11-03T16:59:29.7552759Z 828c1365039a: Download complete 2025-11-03T16:59:29.7595791Z 33a0c5e93563: Verifying Checksum 2025-11-03T16:59:29.7596084Z 33a0c5e93563: Download complete 2025-11-03T16:59:29.8419542Z 7401e108f2b8: Download complete 2025-11-03T16:59:29.9002469Z 9d9c7be466f8: Verifying Checksum 2025-11-03T16:59:29.9002826Z 9d9c7be466f8: Download complete 2025-11-03T16:59:29.9760283Z 7c2887fa8aa0: Verifying Checksum 2025-11-03T16:59:29.9760629Z 7c2887fa8aa0: Download complete 2025-11-03T16:59:30.0792758Z 95bb0abcaee9: Verifying Checksum 2025-11-03T16:59:30.0793221Z 95bb0abcaee9: Download complete 2025-11-03T16:59:30.1526869Z b3a429f4faee: Download complete 2025-11-03T16:59:30.2691612Z f4cb10b91a16: Download complete 2025-11-03T16:59:30.7165362Z 828c1365039a: Pull complete 2025-11-03T16:59:30.7294856Z 1c4ee0d43392: Pull complete 2025-11-03T16:59:31.3854728Z faa9a1190c2e: Verifying Checksum 2025-11-03T16:59:31.3855028Z faa9a1190c2e: Download complete 2025-11-03T16:59:31.3964747Z 4f4fb700ef54: Verifying Checksum 2025-11-03T16:59:31.3965540Z 4f4fb700ef54: Download complete 2025-11-03T16:59:31.4939165Z 2ec7ba999b86: Download complete 2025-11-03T16:59:31.5987469Z 85d9004dcaf2: Verifying Checksum 2025-11-03T16:59:31.5987792Z 85d9004dcaf2: Download complete 2025-11-03T16:59:31.6835530Z df8f8108c158: Verifying Checksum 2025-11-03T16:59:31.6835844Z df8f8108c158: Download complete 2025-11-03T16:59:31.8893999Z 329fb7ac2c9e: Verifying Checksum 2025-11-03T16:59:31.8894298Z 329fb7ac2c9e: Download complete 2025-11-03T16:59:31.9843163Z 397c5718353c: Verifying Checksum 2025-11-03T16:59:31.9843436Z 397c5718353c: Download complete 2025-11-03T16:59:32.0745443Z 099139c9d8b4: Verifying Checksum 2025-11-03T16:59:32.0746026Z 099139c9d8b4: Download complete 2025-11-03T16:59:32.1624186Z c094f15f3938: Download complete 2025-11-03T16:59:32.3110583Z a99b38a3dcdd: Verifying Checksum 2025-11-03T16:59:32.3111226Z a99b38a3dcdd: Download complete 2025-11-03T16:59:33.7406333Z 728590afcb21: Verifying Checksum 2025-11-03T16:59:33.7406720Z 728590afcb21: Download complete 2025-11-03T16:59:33.8104298Z 1a6a81cb899b: Download complete 2025-11-03T16:59:34.2190422Z 41ec23fc0915: Verifying Checksum 2025-11-03T16:59:34.2190901Z 41ec23fc0915: Download complete 2025-11-03T16:59:34.2885114Z 540c0b6aa5f3: Verifying Checksum 2025-11-03T16:59:34.2885672Z 540c0b6aa5f3: Download complete 2025-11-03T16:59:34.3754109Z 5f740b394ea3: Verifying Checksum 2025-11-03T16:59:34.3754422Z 5f740b394ea3: Download complete 2025-11-03T16:59:40.1834583Z 69c4fff3ff1a: Verifying Checksum 2025-11-03T16:59:40.1834926Z 69c4fff3ff1a: Download complete 2025-11-03T16:59:40.2875620Z ac00f880c769: Verifying Checksum 2025-11-03T16:59:40.2876199Z ac00f880c769: Download complete 2025-11-03T16:59:40.4017206Z c084e8e1076b: Download complete 2025-11-03T16:59:40.5234030Z a300a654b9ec: Verifying Checksum 2025-11-03T16:59:40.5234339Z a300a654b9ec: Download complete 2025-11-03T16:59:40.6150632Z 6345aec6eb29: Download complete 2025-11-03T16:59:40.8628316Z e7298879fae8: Verifying Checksum 2025-11-03T16:59:40.8629032Z e7298879fae8: Download complete 2025-11-03T16:59:40.9420693Z 943364a6285e: Verifying Checksum 2025-11-03T16:59:40.9421361Z 943364a6285e: Download complete 2025-11-03T16:59:41.0303366Z 14f0993d5468: Verifying Checksum 2025-11-03T16:59:41.0303667Z 14f0993d5468: Download complete 2025-11-03T16:59:41.1066132Z d6a558a7de5e: Verifying Checksum 2025-11-03T16:59:41.1066996Z d6a558a7de5e: Download complete 2025-11-03T16:59:41.1956198Z 05ee86267141: Verifying Checksum 2025-11-03T16:59:41.1956760Z 05ee86267141: Download complete 2025-11-03T16:59:41.2832008Z f0ccf17afa34: Download complete 2025-11-03T16:59:41.3818407Z a68d4d73adc3: Verifying Checksum 2025-11-03T16:59:41.3819062Z a68d4d73adc3: Download complete 2025-11-03T16:59:41.4809179Z 1ef3a6930902: Verifying Checksum 2025-11-03T16:59:41.4810042Z 1ef3a6930902: Download complete 2025-11-03T16:59:41.5792189Z ba2105a84ce6: Verifying Checksum 2025-11-03T16:59:41.5792753Z ba2105a84ce6: Download complete 2025-11-03T16:59:42.3690069Z 728590afcb21: Pull complete 2025-11-03T16:59:42.7756152Z 9a925ed696bf: Pull complete 2025-11-03T16:59:43.1428958Z e7ab1efdd1bd: Pull complete 2025-11-03T16:59:43.2233089Z 33a0c5e93563: Pull complete 2025-11-03T16:59:43.3221892Z 7401e108f2b8: Pull complete 2025-11-03T16:59:43.5894675Z 9d9c7be466f8: Pull complete 2025-11-03T16:59:45.3954536Z 85c483845f29: Verifying Checksum 2025-11-03T16:59:45.3954919Z 85c483845f29: Download complete 2025-11-03T16:59:46.1000730Z faa9a1190c2e: Pull complete 2025-11-03T16:59:46.4655009Z 7c2887fa8aa0: Pull complete 2025-11-03T16:59:46.7449980Z 95bb0abcaee9: Pull complete 2025-11-03T16:59:47.0772777Z b3a429f4faee: Pull complete 2025-11-03T16:59:47.3434909Z f4cb10b91a16: Pull complete 2025-11-03T17:00:08.6371696Z 48618db2b1a6: Verifying Checksum 2025-11-03T17:00:08.6372584Z 48618db2b1a6: Download complete 2025-11-03T17:00:08.7417065Z 2b12a35a9605: Verifying Checksum 2025-11-03T17:00:08.7417808Z 2b12a35a9605: Download complete 2025-11-03T17:00:08.8293285Z 2a86c078c73d: Download complete 2025-11-03T17:00:08.9143657Z 44e5c47c9ce4: Verifying Checksum 2025-11-03T17:00:08.9143955Z 44e5c47c9ce4: Download complete 2025-11-03T17:00:09.0091761Z 92d25d2ab4f3: Verifying Checksum 2025-11-03T17:00:09.0092052Z 92d25d2ab4f3: Download complete 2025-11-03T17:00:09.1087121Z 07c8d78a73e6: Verifying Checksum 2025-11-03T17:00:09.1087598Z 07c8d78a73e6: Download complete 2025-11-03T17:00:09.2047930Z 66372f032f92: Verifying Checksum 2025-11-03T17:00:09.2048250Z 66372f032f92: Download complete 2025-11-03T17:00:09.3307022Z 60194b590b00: Verifying Checksum 2025-11-03T17:00:09.3307320Z 60194b590b00: Download complete 2025-11-03T17:00:09.4573327Z b41ffa5a6a45: Verifying Checksum 2025-11-03T17:00:09.4573848Z b41ffa5a6a45: Download complete 2025-11-03T17:00:09.5400339Z f165304e2091: Verifying Checksum 2025-11-03T17:00:09.5400878Z f165304e2091: Download complete 2025-11-03T17:00:09.6333090Z 16ec736149b0: Verifying Checksum 2025-11-03T17:00:09.6333657Z 16ec736149b0: Download complete 2025-11-03T17:00:09.7142171Z ea0ef0105e5f: Download complete 2025-11-03T17:00:09.8071087Z b294669af2b5: Download complete 2025-11-03T17:00:09.8851275Z 726179d41040: Verifying Checksum 2025-11-03T17:00:09.8851784Z 726179d41040: Download complete 2025-11-03T17:00:09.9664215Z 375f6cf74921: Verifying Checksum 2025-11-03T17:00:09.9664764Z 375f6cf74921: Download complete 2025-11-03T17:00:12.9411269Z 0cc2718afe44: Verifying Checksum 2025-11-03T17:00:12.9411661Z 0cc2718afe44: Download complete 2025-11-03T17:00:13.0559869Z cc98ed46781d: Verifying Checksum 2025-11-03T17:00:13.0560465Z cc98ed46781d: Download complete 2025-11-03T17:00:13.1589223Z 83ecc1143684: Verifying Checksum 2025-11-03T17:00:13.1589847Z 83ecc1143684: Download complete 2025-11-03T17:00:13.2539337Z 84d7e8a5d95e: Verifying Checksum 2025-11-03T17:00:13.2539647Z 84d7e8a5d95e: Download complete 2025-11-03T17:00:13.3466763Z 9913742e4da4: Verifying Checksum 2025-11-03T17:00:13.3467163Z 9913742e4da4: Download complete 2025-11-03T17:00:13.4519918Z a5d33870f153: Verifying Checksum 2025-11-03T17:00:13.4520324Z a5d33870f153: Download complete 2025-11-03T17:00:13.6295395Z 50b140b19ce9: Verifying Checksum 2025-11-03T17:00:13.6295789Z 50b140b19ce9: Download complete 2025-11-03T17:00:13.7484361Z e7227b2cf3c5: Verifying Checksum 2025-11-03T17:00:13.7484679Z e7227b2cf3c5: Download complete 2025-11-03T17:00:14.3759093Z 4807408eff38: Verifying Checksum 2025-11-03T17:00:14.3759428Z 4807408eff38: Download complete 2025-11-03T17:00:21.0143604Z ce611413a46e: Verifying Checksum 2025-11-03T17:00:21.0144290Z ce611413a46e: Download complete 2025-11-03T17:00:49.7947248Z 48618db2b1a6: Pull complete 2025-11-03T17:00:50.2313711Z 4f4fb700ef54: Pull complete 2025-11-03T17:00:50.6073791Z 2ec7ba999b86: Pull complete 2025-11-03T17:00:51.0221530Z 85d9004dcaf2: Pull complete 2025-11-03T17:00:51.4528015Z df8f8108c158: Pull complete 2025-11-03T17:00:51.8349269Z 329fb7ac2c9e: Pull complete 2025-11-03T17:00:51.9814961Z 397c5718353c: Pull complete 2025-11-03T17:00:52.2918000Z 099139c9d8b4: Pull complete 2025-11-03T17:00:52.7435860Z c094f15f3938: Pull complete 2025-11-03T17:00:53.1277045Z a99b38a3dcdd: Pull complete 2025-11-03T17:01:00.8079175Z bc183ac25ebc: Verifying Checksum 2025-11-03T17:01:00.8082321Z bc183ac25ebc: Download complete 2025-11-03T17:01:53.7756972Z bc183ac25ebc: Pull complete 2025-11-03T17:01:53.9494440Z 1a6a81cb899b: Pull complete 2025-11-03T17:01:54.6760132Z 41ec23fc0915: Pull complete 2025-11-03T17:01:55.1446773Z 540c0b6aa5f3: Pull complete 2025-11-03T17:01:55.4925574Z 5f740b394ea3: Pull complete 2025-11-03T17:02:03.3891701Z 69c4fff3ff1a: Pull complete 2025-11-03T17:02:03.7969953Z ac00f880c769: Pull complete 2025-11-03T17:02:04.1477193Z c084e8e1076b: Pull complete 2025-11-03T17:02:04.8533089Z a300a654b9ec: Pull complete 2025-11-03T17:02:05.2644543Z 6345aec6eb29: Pull complete 2025-11-03T17:02:05.7961443Z e7298879fae8: Pull complete 2025-11-03T17:02:06.1649190Z 943364a6285e: Pull complete 2025-11-03T17:02:06.5384536Z 14f0993d5468: Pull complete 2025-11-03T17:02:06.9819090Z d6a558a7de5e: Pull complete 2025-11-03T17:02:07.2673573Z 05ee86267141: Pull complete 2025-11-03T17:02:07.2941299Z f0ccf17afa34: Pull complete 2025-11-03T17:02:07.3441786Z a68d4d73adc3: Pull complete 2025-11-03T17:02:07.3680099Z 1ef3a6930902: Pull complete 2025-11-03T17:02:07.4169510Z ba2105a84ce6: Pull complete 2025-11-03T17:02:08.3576216Z 85c483845f29: Pull complete 2025-11-03T17:02:56.3629038Z ce611413a46e: Pull complete 2025-11-03T17:02:56.8977593Z 2b12a35a9605: Pull complete 2025-11-03T17:02:57.3220856Z 2a86c078c73d: Pull complete 2025-11-03T17:02:58.1434454Z 44e5c47c9ce4: Pull complete 2025-11-03T17:02:59.0119359Z 92d25d2ab4f3: Pull complete 2025-11-03T17:02:59.4222509Z 07c8d78a73e6: Pull complete 2025-11-03T17:03:00.1926106Z 66372f032f92: Pull complete 2025-11-03T17:03:01.0669076Z 60194b590b00: Pull complete 2025-11-03T17:03:01.5114937Z b41ffa5a6a45: Pull complete 2025-11-03T17:03:02.4094579Z f165304e2091: Pull complete 2025-11-03T17:03:02.8046853Z 16ec736149b0: Pull complete 2025-11-03T17:03:03.6579323Z ea0ef0105e5f: Pull complete 2025-11-03T17:03:04.1121149Z b294669af2b5: Pull complete 2025-11-03T17:03:04.9609037Z 726179d41040: Pull complete 2025-11-03T17:03:05.4495117Z 375f6cf74921: Pull complete 2025-11-03T17:03:12.0227093Z 0cc2718afe44: Pull complete 2025-11-03T17:03:12.4037476Z cc98ed46781d: Pull complete 2025-11-03T17:03:12.7806300Z 83ecc1143684: Pull complete 2025-11-03T17:03:13.1698297Z 84d7e8a5d95e: Pull complete 2025-11-03T17:03:13.5607144Z 9913742e4da4: Pull complete 2025-11-03T17:03:13.9098737Z a5d33870f153: Pull complete 2025-11-03T17:03:15.7732165Z 50b140b19ce9: Pull complete 2025-11-03T17:03:16.1601874Z e7227b2cf3c5: Pull complete 2025-11-03T17:03:17.7103829Z 4807408eff38: Pull complete 2025-11-03T17:03:18.4332164Z Digest: sha256:655b37e47c7ab0fab00339d6162ef718600c6b4e76f1601297e6a07026860230 2025-11-03T17:03:18.5214340Z Status: Downloaded newer image for 308535385114.dkr.ecr.us-east-1.amazonaws.com/pytorch/ci-image:pytorch-linux-jammy-cuda12.8-cudnn9-py3.12-gcc11-vllm-30147fb242f88ab8f13e14b76ab186d189d8605c 2025-11-03T17:03:18.5455796Z 308535385114.dkr.ecr.us-east-1.amazonaws.com/pytorch/ci-image:pytorch-linux-jammy-cuda12.8-cudnn9-py3.12-gcc11-vllm-30147fb242f88ab8f13e14b76ab186d189d8605c 2025-11-03T17:03:18.5523373Z ##[group]Run echo "IN_CONTAINER_RUNNER=$(if [ -f /.inarc ] || [ -f /.incontainer ]; then echo true ; else echo false; fi)" >> "$GITHUB_OUTPUT" 2025-11-03T17:03:18.5524147Z echo "IN_CONTAINER_RUNNER=$(if [ -f /.inarc ] || [ -f /.incontainer ]; then echo true ; else echo false; fi)" >> "$GITHUB_OUTPUT" 2025-11-03T17:03:18.5535626Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2025-11-03T17:03:18.5535912Z env: 2025-11-03T17:03:18.5536083Z GIT_DEFAULT_BRANCH: main 2025-11-03T17:03:18.5536277Z ##[endgroup] 2025-11-03T17:03:18.5761882Z ##[group]Run pytorch/test-infra/.github/actions/setup-nvidia@main 2025-11-03T17:03:18.5762231Z with: 2025-11-03T17:03:18.5762400Z driver-version: 580.82.07 2025-11-03T17:03:18.5762590Z env: 2025-11-03T17:03:18.5762735Z GIT_DEFAULT_BRANCH: main 2025-11-03T17:03:18.5762916Z ##[endgroup] 2025-11-03T17:03:18.5932983Z ##[group]Run nick-fields/retry@3e91a01664abd3c5cd539100d10d33b9c5b68482 2025-11-03T17:03:18.5933283Z with: 2025-11-03T17:03:18.5933435Z timeout_minutes: 10 2025-11-03T17:03:18.5933613Z max_attempts: 3 2025-11-03T17:03:18.5952904Z command: # Is it disgusting to have a full shell script here in this github action? Sure # But is it the best way to make it so that this action relies on nothing else? Absolutely set -eou pipefail DISTRIBUTION=$(. /etc/os-release;echo $ID$VERSION_ID) DRIVER_FN="NVIDIA-Linux-x86_64-${DRIVER_VERSION}.run" install_nvidia_docker2_amzn2() { ( set -x # Needed for yum-config-manager sudo yum install -y yum-utils if [[ "${DISTRIBUTION}" == "amzn2023" ]] ; then YUM_REPO_URL="https://nvidia.github.io/libnvidia-container/stable/rpm/nvidia-container-toolkit.repo" else # Amazon Linux 2 YUM_REPO_URL="https://nvidia.github.io/nvidia-docker/${DISTRIBUTION}/nvidia-docker.repo" fi sudo yum-config-manager --add-repo "${YUM_REPO_URL}" sudo yum install -y \ nvidia-container-toolkit-1.17.8 \ libnvidia-container-tools-1.17.8 \ libnvidia-container1-1.17.8 \ nvidia-container-toolkit-base-1.17.8 sudo systemctl restart docker ) } install_nvidia_docker2_ubuntu20() { ( set -x # Install nvidia-driver package if not installed status="$(dpkg-query -W --showformat='${db:Status-Status}' nvidia-docker2 2>&1)" if [ ! $? = 0 ] || [ ! "$status" = installed ]; then sudo apt-get install -y nvidia-container-toolkit-1.17.8 sudo systemctl restart docker fi ) } pre_install_nvidia_driver_amzn2() { ( # Purge any nvidia driver installed from RHEL repo sudo yum remove -y nvidia-driver-latest-dkms ) } install_nvidia_driver_common() { ( # Try to gather more information about the runner and its existing NVIDIA driver if any echo "Before installing NVIDIA driver" lspci lsmod modinfo nvidia || true HAS_NVIDIA_DRIVER=0 # Check if NVIDIA driver has already been installed if [ -x "$(command -v nvidia-smi)" ]; then set +e # The driver exists, check its version next. Also check only the first GPU if there are more than one of them # so that the same driver version is not print over multiple lines INSTALLED_DRIVER_VERSION=$(nvidia-smi --query-gpu=driver_version --format=csv,noheader --id=0) NVIDIA_SMI_STATUS=$? if [ "$NVIDIA_SMI_STATUS" -ne 0 ] && [ "$NVIDIA_SMI_STATUS" -ne 14 ]; then echo "Failed to get NVIDIA driver version ($INSTALLED_DRIVER_VERSION). Continuing" elif [ "$INSTALLED_DRIVER_VERSION" != "$DRIVER_VERSION" ]; then echo "NVIDIA driver ($INSTALLED_DRIVER_VERSION) has been installed, but we expect to have $DRIVER_VERSION instead. Continuing" # Turn off persistent mode so that the installation script can unload the kernel module sudo killall nvidia-persistenced || true else HAS_NVIDIA_DRIVER=1 echo "NVIDIA driver ($INSTALLED_DRIVER_VERSION) has already been installed. Skipping NVIDIA driver installation" fi set -e fi if [ "$HAS_NVIDIA_DRIVER" -eq 0 ]; then # CAUTION: this may need to be updated in future if [ "${DISTRIBUTION}" != ubuntu20.04 ]; then sudo yum groupinstall -y "Development Tools" # ensure our kernel install is the same as our underlying kernel, # groupinstall "Development Tools" has a habit of mismatching kernel headers sudo yum install -y "kernel-devel-uname-r == $(uname -r)" sudo modprobe backlight fi sudo curl -fsL -o /tmp/nvidia_driver "https://s3.amazonaws.com/ossci-linux/nvidia_driver/$DRIVER_FN" set +e sudo /bin/bash /tmp/nvidia_driver -s --no-drm NVIDIA_INSTALLATION_STATUS=$? RESET_GPU=0 if [ "$NVIDIA_INSTALLATION_STATUS" -ne 0 ]; then sudo cat /var/log/nvidia-installer.log # Fail to install NVIDIA driver, try to reset the GPU RESET_GPU=1 elif [ -x "$(command -v nvidia-smi)" ]; then # Check again if nvidia-smi works even if the driver installation completes successfully INSTALLED_DRIVER_VERSION=$(nvidia-smi --query-gpu=driver_version --format=csv,noheader --id=0) NVIDIA_SMI_STATUS=$? if [ "$NVIDIA_SMI_STATUS" -ne 0 ] && [ "$NVIDIA_SMI_STATUS" -ne 14 ]; then RESET_GPU=1 fi fi if [ "$RESET_GPU" -eq 1 ]; then NVIDIA_DEVICES=$(lspci -D | grep -i NVIDIA | cut -d' ' -f1) # The GPU can get stuck in a failure state if somehow the test crashs the GPU microcode. When this # happens, we'll try to reset all NVIDIA devices https://github.com/pytorch/pytorch/issues/88388 for PCI_ID in $NVIDIA_DEVICES; do DEVICE_ENABLED=$(cat /sys/bus/pci/devices/$PCI_ID/enable) echo "Reseting $PCI_ID (enabled state: $DEVICE_ENABLED)" # This requires sudo permission of course echo "1" | sudo tee /sys/bus/pci/devices/$PCI_ID/reset sleep 1 done fi sudo rm -fv /tmp/nvidia_driver set -e fi ) } post_install_nvidia_driver_common() { ( sudo modprobe nvidia || true echo "After installing NVIDIA driver" lspci lsmod modinfo nvidia || true ( set +e nvidia-smi # NB: Annoyingly, nvidia-smi command returns successfully with return code 0 even in # the case where the driver has already crashed as it still can get the driver version # and some basic information like the bus ID. However, the rest of the information # would be missing (ERR!), for example: # # +-----------------------------------------------------------------------------+ # | NVIDIA-SMI 525.89.02 Driver Version: 525.89.02 CUDA Version: 12.0 | # |-------------------------------+----------------------+----------------------+ # | GPU Name Persistence-M| Bus-Id Disp.A | Volatile Uncorr. ECC | # | Fan Temp Perf Pwr:Usage/Cap| Memory-Usage | GPU-Util Compute M. | # | | | MIG M. | # |===============================+======================+======================| # | 0 ERR! Off | 00000000:00:1E.0 Off | ERR! | # |ERR! ERR! ERR! ERR! / ERR! | 4184MiB / 23028MiB | ERR! Default | # | | | ERR! | # +-------------------------------+----------------------+----------------------+ # # +-----------------------------------------------------------------------------+ # | Processes: | # | GPU GI CI PID Type Process name GPU Memory | # | ID ID Usage | # |=============================================================================| # +-----------------------------------------------------------------------------+ # # This should be reported as a failure instead as it will guarantee to fail when # Docker tries to run with --gpus all # # So, the correct check here is to query one of the missing piece of info like # GPU name, so that the command can fail accordingly nvidia-smi --query-gpu=gpu_name --format=csv,noheader --id=0 NVIDIA_SMI_STATUS=$? # Allowable exit statuses for nvidia-smi, see: https://github.com/NVIDIA/gpu-operator/issues/285 if [ "$NVIDIA_SMI_STATUS" -eq 0 ] || [ "$NVIDIA_SMI_STATUS" -eq 14 ]; then echo "INFO: Ignoring allowed status ${NVIDIA_SMI_STATUS}" else echo "ERROR: nvidia-smi exited with unresolved status ${NVIDIA_SMI_STATUS}" exit ${NVIDIA_SMI_STATUS} fi set -e ) ) } install_nvidia_driver_amzn2() { ( set -x pre_install_nvidia_driver_amzn2 install_nvidia_driver_common post_install_nvidia_driver_common ) } install_nvidia_driver_ubuntu20() { ( set -x install_nvidia_driver_common post_install_nvidia_driver_common ) } echo "== Installing nvidia driver ${DRIVER_FN} ==" case "${DISTRIBUTION}" in amzn*) install_nvidia_driver_amzn2 ;; ubuntu20.04) install_nvidia_driver_ubuntu20 ;; *) echo "ERROR: Unknown distribution ${DISTRIBUTION}" exit 1 ;; esac # Install container toolkit based on distribution echo "== Installing nvidia container toolkit for ${DISTRIBUTION} ==" case "${DISTRIBUTION}" in amzn*) install_nvidia_docker2_amzn2 ;; ubuntu20.04) install_nvidia_docker2_ubuntu20 ;; *) echo "ERROR: Unknown distribution ${DISTRIBUTION}" exit 1 ;; esac echo "GPU_FLAG=--gpus all -e NVIDIA_DRIVER_CAPABILITIES=all" >> "${GITHUB_ENV}" # Fix https://github.com/NVIDIA/nvidia-docker/issues/1648 on runners with # more than one GPUs. This just needs to be run once. The command fails # on subsequent runs and complains that the mode is already on, but that's # ok sudo nvidia-persistenced || true # This should show persistence mode ON nvidia-smi # check if the container-toolkit is correctly installed and CUDA is available inside a container docker run --rm -t --gpus=all public.ecr.aws/docker/library/python:3.13 nvidia-smi 2025-11-03T17:03:18.5972402Z retry_wait_seconds: 10 2025-11-03T17:03:18.5972604Z polling_interval_seconds: 1 2025-11-03T17:03:18.5972805Z warning_on_retry: true 2025-11-03T17:03:18.5972985Z continue_on_error: false 2025-11-03T17:03:18.5973172Z env: 2025-11-03T17:03:18.5973322Z GIT_DEFAULT_BRANCH: main 2025-11-03T17:03:18.5973511Z DRIVER_VERSION: 580.82.07 2025-11-03T17:03:18.5973691Z ##[endgroup] 2025-11-03T17:03:18.6888944Z == Installing nvidia driver NVIDIA-Linux-x86_64-580.82.07.run == 2025-11-03T17:03:18.6890301Z + pre_install_nvidia_driver_amzn2 2025-11-03T17:03:18.6895965Z + sudo yum remove -y nvidia-driver-latest-dkms 2025-11-03T17:03:19.3397777Z No match for argument: nvidia-driver-latest-dkms 2025-11-03T17:03:19.3398363Z No packages marked for removal. 2025-11-03T17:03:19.3454684Z Dependencies resolved. 2025-11-03T17:03:19.3463759Z Nothing to do. 2025-11-03T17:03:19.3464429Z Complete! 2025-11-03T17:03:19.4349044Z + install_nvidia_driver_common 2025-11-03T17:03:19.4355437Z + echo 'Before installing NVIDIA driver' 2025-11-03T17:03:19.4355737Z + lspci 2025-11-03T17:03:19.4359073Z Before installing NVIDIA driver 2025-11-03T17:03:19.5463026Z 00:00.0 Host bridge: Intel Corporation 440FX - 82441FX PMC [Natoma] 2025-11-03T17:03:19.5464059Z 00:01.0 ISA bridge: Intel Corporation 82371SB PIIX3 ISA [Natoma/Triton II] 2025-11-03T17:03:19.5465084Z 00:01.3 Non-VGA unclassified device: Intel Corporation 82371AB/EB/MB PIIX4 ACPI (rev 08) 2025-11-03T17:03:19.5466488Z 00:03.0 VGA compatible controller: Amazon.com, Inc. Device 1111 2025-11-03T17:03:19.5467432Z 00:04.0 Non-Volatile memory controller: Amazon.com, Inc. NVMe EBS Controller 2025-11-03T17:03:19.5468245Z 01:00.0 PCI bridge: Amazon.com, Inc. Device 0200 2025-11-03T17:03:19.5468866Z 02:00.0 PCI bridge: Amazon.com, Inc. Device 0200 2025-11-03T17:03:19.5469459Z 03:00.0 PCI bridge: Amazon.com, Inc. Device 0200 2025-11-03T17:03:19.5470091Z 03:00.1 PCI bridge: Amazon.com, Inc. Device 0200 2025-11-03T17:03:19.5470758Z 03:00.2 PCI bridge: Amazon.com, Inc. Device 0200 2025-11-03T17:03:19.5471430Z 03:00.3 PCI bridge: Amazon.com, Inc. Device 0200 2025-11-03T17:03:19.5471860Z 03:00.4 PCI bridge: Amazon.com, Inc. Device 0200 2025-11-03T17:03:19.5472139Z 03:00.5 PCI bridge: Amazon.com, Inc. Device 0200 2025-11-03T17:03:19.5472394Z 03:00.6 PCI bridge: Amazon.com, Inc. Device 0200 2025-11-03T17:03:19.5472633Z 03:00.7 PCI bridge: Amazon.com, Inc. Device 0200 2025-11-03T17:03:19.5472878Z 03:01.0 PCI bridge: Amazon.com, Inc. Device 0200 2025-11-03T17:03:19.5473118Z 03:01.1 PCI bridge: Amazon.com, Inc. Device 0200 2025-11-03T17:03:19.5473520Z 03:01.2 PCI bridge: Amazon.com, Inc. Device 0200 2025-11-03T17:03:19.5473768Z 03:01.3 PCI bridge: Amazon.com, Inc. Device 0200 2025-11-03T17:03:19.5474010Z 03:01.4 PCI bridge: Amazon.com, Inc. Device 0200 2025-11-03T17:03:19.5474249Z 03:01.5 PCI bridge: Amazon.com, Inc. Device 0200 2025-11-03T17:03:19.5474495Z 03:01.6 PCI bridge: Amazon.com, Inc. Device 0200 2025-11-03T17:03:19.5474737Z 03:01.7 PCI bridge: Amazon.com, Inc. Device 0200 2025-11-03T17:03:19.5474978Z 03:02.0 PCI bridge: Amazon.com, Inc. Device 0200 2025-11-03T17:03:19.5475218Z 03:02.1 PCI bridge: Amazon.com, Inc. Device 0200 2025-11-03T17:03:19.5475451Z 03:02.2 PCI bridge: Amazon.com, Inc. Device 0200 2025-11-03T17:03:19.5475689Z 03:02.3 PCI bridge: Amazon.com, Inc. Device 0200 2025-11-03T17:03:19.5475930Z 03:02.4 PCI bridge: Amazon.com, Inc. Device 0200 2025-11-03T17:03:19.5476170Z 03:02.5 PCI bridge: Amazon.com, Inc. Device 0200 2025-11-03T17:03:19.5476412Z 03:02.6 PCI bridge: Amazon.com, Inc. Device 0200 2025-11-03T17:03:19.5476668Z 03:02.7 PCI bridge: Amazon.com, Inc. Device 0200 2025-11-03T17:03:19.5476914Z 03:03.0 PCI bridge: Amazon.com, Inc. Device 0200 2025-11-03T17:03:19.5477162Z 03:03.1 PCI bridge: Amazon.com, Inc. Device 0200 2025-11-03T17:03:19.5477396Z 03:03.2 PCI bridge: Amazon.com, Inc. Device 0200 2025-11-03T17:03:19.5477639Z 03:03.3 PCI bridge: Amazon.com, Inc. Device 0200 2025-11-03T17:03:19.5477881Z 03:03.4 PCI bridge: Amazon.com, Inc. Device 0200 2025-11-03T17:03:19.5478123Z 03:03.5 PCI bridge: Amazon.com, Inc. Device 0200 2025-11-03T17:03:19.5478364Z 03:03.6 PCI bridge: Amazon.com, Inc. Device 0200 2025-11-03T17:03:19.5478602Z 03:03.7 PCI bridge: Amazon.com, Inc. Device 0200 2025-11-03T17:03:19.5478848Z 24:00.0 PCI bridge: Amazon.com, Inc. Device 0200 2025-11-03T17:03:19.5479093Z 25:00.0 PCI bridge: Amazon.com, Inc. Device 0200 2025-11-03T17:03:19.5479335Z 26:00.0 PCI bridge: Amazon.com, Inc. Device 0200 2025-11-03T17:03:19.5479569Z 26:00.1 PCI bridge: Amazon.com, Inc. Device 0200 2025-11-03T17:03:19.5479812Z 26:00.2 PCI bridge: Amazon.com, Inc. Device 0200 2025-11-03T17:03:19.5480060Z 26:00.3 PCI bridge: Amazon.com, Inc. Device 0200 2025-11-03T17:03:19.5480298Z 26:00.4 PCI bridge: Amazon.com, Inc. Device 0200 2025-11-03T17:03:19.5480527Z 26:00.5 PCI bridge: Amazon.com, Inc. Device 0200 2025-11-03T17:03:19.5480765Z 26:00.6 PCI bridge: Amazon.com, Inc. Device 0200 2025-11-03T17:03:19.5481016Z 26:00.7 PCI bridge: Amazon.com, Inc. Device 0200 2025-11-03T17:03:19.5481254Z 26:01.0 PCI bridge: Amazon.com, Inc. Device 0200 2025-11-03T17:03:19.5481566Z 27:00.0 Ethernet controller: Amazon.com, Inc. Elastic Network Adapter (ENA) 2025-11-03T17:03:19.5481893Z 30:00.0 PCI bridge: Amazon.com, Inc. Device 0200 2025-11-03T17:03:19.5482153Z 31:00.0 PCI bridge: Amazon.com, Inc. Device 0200 2025-11-03T17:03:19.5482397Z 32:00.0 PCI bridge: Amazon.com, Inc. Device 0200 2025-11-03T17:03:19.5482639Z 32:00.1 PCI bridge: Amazon.com, Inc. Device 0200 2025-11-03T17:03:19.5482966Z 32:00.2 PCI bridge: Amazon.com, Inc. Device 0200 2025-11-03T17:03:19.5483217Z 32:00.3 PCI bridge: Amazon.com, Inc. Device 0200 2025-11-03T17:03:19.5483532Z 33:00.0 Non-Volatile memory controller: Amazon.com, Inc. NVMe SSD Controller 2025-11-03T17:03:19.5483924Z 34:00.0 Non-Volatile memory controller: Amazon.com, Inc. NVMe SSD Controller 2025-11-03T17:03:19.5484305Z 35:00.0 Non-Volatile memory controller: Amazon.com, Inc. NVMe SSD Controller 2025-11-03T17:03:19.5484685Z 36:00.0 Non-Volatile memory controller: Amazon.com, Inc. NVMe SSD Controller 2025-11-03T17:03:19.5484998Z 37:00.0 PCI bridge: Amazon.com, Inc. Device 0200 2025-11-03T17:03:19.5485285Z 38:00.0 3D controller: NVIDIA Corporation AD104GL [L4] (rev a1) 2025-11-03T17:03:19.5485562Z 39:00.0 PCI bridge: Amazon.com, Inc. Device 0200 2025-11-03T17:03:19.5485834Z 3a:00.0 3D controller: NVIDIA Corporation AD104GL [L4] (rev a1) 2025-11-03T17:03:19.5486111Z 3b:00.0 PCI bridge: Amazon.com, Inc. Device 0200 2025-11-03T17:03:19.5486388Z 3c:00.0 3D controller: NVIDIA Corporation AD104GL [L4] (rev a1) 2025-11-03T17:03:19.5486738Z 3d:00.0 PCI bridge: Amazon.com, Inc. Device 0200 2025-11-03T17:03:19.5486999Z 3e:00.0 3D controller: NVIDIA Corporation AD104GL [L4] (rev a1) 2025-11-03T17:03:19.5487254Z + lsmod 2025-11-03T17:03:19.5516640Z Module Size Used by 2025-11-03T17:03:19.5517259Z nvidia_uvm 1925120 0 2025-11-03T17:03:19.5517772Z nvidia 14286848 1 nvidia_uvm 2025-11-03T17:03:19.5518285Z drm 602112 1 nvidia 2025-11-03T17:03:19.5518834Z drm_panel_orientation_quirks 32768 1 drm 2025-11-03T17:03:19.5519403Z backlight 24576 1 drm 2025-11-03T17:03:19.5519907Z i2c_core 110592 2 nvidia,drm 2025-11-03T17:03:19.5520400Z xt_conntrack 16384 1 2025-11-03T17:03:19.5520854Z nft_chain_nat 16384 3 2025-11-03T17:03:19.5521261Z xt_MASQUERADE 20480 1 2025-11-03T17:03:19.5521512Z nf_nat 57344 2 nft_chain_nat,xt_MASQUERADE 2025-11-03T17:03:19.5521759Z nf_conntrack_netlink 57344 0 2025-11-03T17:03:19.5522064Z nf_conntrack 184320 4 xt_conntrack,nf_nat,nf_conntrack_netlink,xt_MASQUERADE 2025-11-03T17:03:19.5522400Z nf_defrag_ipv6 24576 1 nf_conntrack 2025-11-03T17:03:19.5522652Z nf_defrag_ipv4 16384 1 nf_conntrack 2025-11-03T17:03:19.5522863Z xfrm_user 57344 1 2025-11-03T17:03:19.5523058Z xfrm_algo 16384 1 xfrm_user 2025-11-03T17:03:19.5523270Z xt_addrtype 16384 2 2025-11-03T17:03:19.5523455Z nft_compat 20480 4 2025-11-03T17:03:19.5523671Z nf_tables 311296 57 nft_compat,nft_chain_nat 2025-11-03T17:03:19.5523979Z nfnetlink 20480 4 nft_compat,nf_conntrack_netlink,nf_tables 2025-11-03T17:03:19.5524260Z br_netfilter 36864 0 2025-11-03T17:03:19.5524456Z bridge 323584 1 br_netfilter 2025-11-03T17:03:19.5524667Z stp 16384 1 bridge 2025-11-03T17:03:19.5524873Z llc 16384 2 bridge,stp 2025-11-03T17:03:19.5525085Z overlay 167936 0 2025-11-03T17:03:19.5525264Z tls 139264 0 2025-11-03T17:03:19.5525432Z nls_ascii 16384 1 2025-11-03T17:03:19.5525609Z nls_cp437 20480 1 2025-11-03T17:03:19.5525782Z vfat 24576 1 2025-11-03T17:03:19.5525973Z fat 86016 1 vfat 2025-11-03T17:03:19.5526164Z sunrpc 700416 1 2025-11-03T17:03:19.5526350Z ghash_clmulni_intel 16384 0 2025-11-03T17:03:19.5526533Z i8042 45056 0 2025-11-03T17:03:19.5526706Z ena 184320 0 2025-11-03T17:03:19.5526884Z serio 28672 3 i8042 2025-11-03T17:03:19.5527084Z button 24576 0 2025-11-03T17:03:19.5527261Z sch_fq_codel 20480 9 2025-11-03T17:03:19.5527440Z fuse 184320 1 2025-11-03T17:03:19.5527609Z dm_mod 188416 0 2025-11-03T17:03:19.5527786Z configfs 57344 1 2025-11-03T17:03:19.5528128Z loop 36864 0 2025-11-03T17:03:19.5528338Z dmi_sysfs 20480 0 2025-11-03T17:03:19.5528589Z crc32_pclmul 16384 0 2025-11-03T17:03:19.5528772Z crc32c_intel 24576 0 2025-11-03T17:03:19.5528951Z efivarfs 24576 1 2025-11-03T17:03:19.5529139Z + modinfo nvidia 2025-11-03T17:03:19.5537205Z filename: /lib/modules/6.1.150-174.273.amzn2023.x86_64/kernel/drivers/video/nvidia.ko 2025-11-03T17:03:19.5537598Z import_ns: DMA_BUF 2025-11-03T17:03:19.5537807Z alias: char-major-195-* 2025-11-03T17:03:19.5538016Z version: 580.82.07 2025-11-03T17:03:19.5538203Z supported: external 2025-11-03T17:03:19.5538395Z license: Dual MIT/GPL 2025-11-03T17:03:19.5538611Z firmware: nvidia/580.82.07/gsp_tu10x.bin 2025-11-03T17:03:19.5538874Z firmware: nvidia/580.82.07/gsp_ga10x.bin 2025-11-03T17:03:19.5539117Z srcversion: BA7240A71DCF7DC6FE88C1D 2025-11-03T17:03:19.5539366Z alias: of:N*T*Cnvidia,tegra264-displayC* 2025-11-03T17:03:19.5539799Z alias: of:N*T*Cnvidia,tegra264-display 2025-11-03T17:03:19.5540064Z alias: of:N*T*Cnvidia,tegra234-displayC* 2025-11-03T17:03:19.5540322Z alias: of:N*T*Cnvidia,tegra234-display 2025-11-03T17:03:19.5540568Z alias: pci:v000010DEd*sv*sd*bc06sc80i00* 2025-11-03T17:03:19.5540808Z alias: pci:v000010DEd*sv*sd*bc03sc02i00* 2025-11-03T17:03:19.5541055Z alias: pci:v000010DEd*sv*sd*bc03sc00i00* 2025-11-03T17:03:19.5541287Z depends: i2c-core,drm 2025-11-03T17:03:19.5541474Z retpoline: Y 2025-11-03T17:03:19.5541626Z name: nvidia 2025-11-03T17:03:19.5541910Z vermagic: 6.1.150-174.273.amzn2023.x86_64 SMP preempt mod_unload modversions 2025-11-03T17:03:19.5542265Z parm: NvSwitchRegDwords:NvSwitch regkey (charp) 2025-11-03T17:03:19.5542596Z parm: NvSwitchBlacklist:NvSwitchBlacklist=uuid[,uuid...] (charp) 2025-11-03T17:03:19.5542897Z parm: NVreg_ResmanDebugLevel:int 2025-11-03T17:03:19.5543125Z parm: NVreg_RmLogonRC:int 2025-11-03T17:03:19.5543351Z parm: NVreg_ModifyDeviceFiles:int 2025-11-03T17:03:19.5543579Z parm: NVreg_DeviceFileUID:int 2025-11-03T17:03:19.5543792Z parm: NVreg_DeviceFileGID:int 2025-11-03T17:03:19.5544013Z parm: NVreg_DeviceFileMode:int 2025-11-03T17:03:19.5544280Z parm: NVreg_InitializeSystemMemoryAllocations:int 2025-11-03T17:03:19.5544564Z parm: NVreg_UsePageAttributeTable:int 2025-11-03T17:03:19.5544796Z parm: NVreg_EnablePCIeGen3:int 2025-11-03T17:03:19.5545010Z parm: NVreg_EnableMSI:int 2025-11-03T17:03:19.5545233Z parm: NVreg_EnableStreamMemOPs:int 2025-11-03T17:03:19.5545501Z parm: NVreg_RestrictProfilingToAdminUsers:int 2025-11-03T17:03:19.5545787Z parm: NVreg_PreserveVideoMemoryAllocations:int 2025-11-03T17:03:19.5546065Z parm: NVreg_EnableS0ixPowerManagement:int 2025-11-03T17:03:19.5546365Z parm: NVreg_S0ixPowerManagementVideoMemoryThreshold:int 2025-11-03T17:03:19.5546675Z parm: NVreg_DynamicPowerManagement:int 2025-11-03T17:03:19.5546996Z parm: NVreg_DynamicPowerManagementVideoMemoryThreshold:int 2025-11-03T17:03:19.5547292Z parm: NVreg_EnableGpuFirmware:int 2025-11-03T17:03:19.5547541Z parm: NVreg_EnableGpuFirmwareLogs:int 2025-11-03T17:03:19.5547813Z parm: NVreg_OpenRmEnableUnsupportedGpus:int 2025-11-03T17:03:19.5548091Z parm: NVreg_EnableUserNUMAManagement:int 2025-11-03T17:03:19.5548336Z parm: NVreg_MemoryPoolSize:int 2025-11-03T17:03:19.5548579Z parm: NVreg_KMallocHeapMaxSize:int 2025-11-03T17:03:19.5548820Z parm: NVreg_VMallocHeapMaxSize:int 2025-11-03T17:03:19.5549055Z parm: NVreg_IgnoreMMIOCheck:int 2025-11-03T17:03:19.5549277Z parm: NVreg_NvLinkDisable:int 2025-11-03T17:03:19.5549534Z parm: NVreg_EnablePCIERelaxedOrderingMode:int 2025-11-03T17:03:19.5549796Z parm: NVreg_RegisterPCIDriver:int 2025-11-03T17:03:19.5550181Z parm: NVreg_RegisterPlatformDeviceDriver:int 2025-11-03T17:03:19.5550461Z parm: NVreg_EnableResizableBar:int 2025-11-03T17:03:19.5550717Z parm: NVreg_EnableDbgBreakpoint:int 2025-11-03T17:03:19.5550972Z parm: NVreg_EnableNonblockingOpen:int 2025-11-03T17:03:19.5551230Z parm: NVreg_CoherentGPUMemoryMode:charp 2025-11-03T17:03:19.5551477Z parm: NVreg_RegistryDwords:charp 2025-11-03T17:03:19.5551729Z parm: NVreg_RegistryDwordsPerDevice:charp 2025-11-03T17:03:19.5551972Z parm: NVreg_RmMsg:charp 2025-11-03T17:03:19.5552190Z parm: NVreg_GpuBlacklist:charp 2025-11-03T17:03:19.5552422Z parm: NVreg_TemporaryFilePath:charp 2025-11-03T17:03:19.5552670Z parm: NVreg_ExcludedGpus:charp 2025-11-03T17:03:19.5552910Z parm: NVreg_DmaRemapPeerMmio:int 2025-11-03T17:03:19.5553157Z parm: NVreg_RmNvlinkBandwidth:charp 2025-11-03T17:03:19.5553419Z parm: NVreg_RmNvlinkBandwidthLinkCount:int 2025-11-03T17:03:19.5553780Z parm: NVreg_ImexChannelCount:int 2025-11-03T17:03:19.5554038Z parm: NVreg_CreateImexChannel0:int 2025-11-03T17:03:19.5554294Z parm: NVreg_GrdmaPciTopoCheckOverride:int 2025-11-03T17:03:19.5554541Z parm: rm_firmware_active:charp 2025-11-03T17:03:19.5554745Z + HAS_NVIDIA_DRIVER=0 2025-11-03T17:03:19.5554927Z ++ command -v nvidia-smi 2025-11-03T17:03:19.5555121Z + '[' -x /usr/bin/nvidia-smi ']' 2025-11-03T17:03:19.5555308Z + set +e 2025-11-03T17:03:19.5557184Z ++ nvidia-smi --query-gpu=driver_version --format=csv,noheader --id=0 2025-11-03T17:03:22.9559342Z + INSTALLED_DRIVER_VERSION=580.82.07 2025-11-03T17:03:22.9559880Z + NVIDIA_SMI_STATUS=0 2025-11-03T17:03:22.9560216Z + '[' 0 -ne 0 ']' 2025-11-03T17:03:22.9560530Z + '[' 580.82.07 '!=' 580.82.07 ']' 2025-11-03T17:03:22.9560898Z + HAS_NVIDIA_DRIVER=1 2025-11-03T17:03:22.9561514Z + echo 'NVIDIA driver (580.82.07) has already been installed. Skipping NVIDIA driver installation' 2025-11-03T17:03:22.9562239Z + set -e 2025-11-03T17:03:22.9562516Z + '[' 1 -eq 0 ']' 2025-11-03T17:03:22.9563051Z NVIDIA driver (580.82.07) has already been installed. Skipping NVIDIA driver installation 2025-11-03T17:03:22.9564092Z + post_install_nvidia_driver_common 2025-11-03T17:03:22.9567775Z + sudo modprobe nvidia 2025-11-03T17:03:23.1021730Z + echo 'After installing NVIDIA driver' 2025-11-03T17:03:23.1022018Z + lspci 2025-11-03T17:03:23.1022199Z After installing NVIDIA driver 2025-11-03T17:03:23.1225790Z 00:00.0 Host bridge: Intel Corporation 440FX - 82441FX PMC [Natoma] 2025-11-03T17:03:23.1226774Z 00:01.0 ISA bridge: Intel Corporation 82371SB PIIX3 ISA [Natoma/Triton II] 2025-11-03T17:03:23.1227776Z 00:01.3 Non-VGA unclassified device: Intel Corporation 82371AB/EB/MB PIIX4 ACPI (rev 08) 2025-11-03T17:03:23.1228765Z 00:03.0 VGA compatible controller: Amazon.com, Inc. Device 1111 2025-11-03T17:03:23.1229635Z 00:04.0 Non-Volatile memory controller: Amazon.com, Inc. NVMe EBS Controller 2025-11-03T17:03:23.1230490Z 01:00.0 PCI bridge: Amazon.com, Inc. Device 0200 2025-11-03T17:03:23.1231101Z 02:00.0 PCI bridge: Amazon.com, Inc. Device 0200 2025-11-03T17:03:23.1231722Z 03:00.0 PCI bridge: Amazon.com, Inc. Device 0200 2025-11-03T17:03:23.1232312Z 03:00.1 PCI bridge: Amazon.com, Inc. Device 0200 2025-11-03T17:03:23.1232875Z 03:00.2 PCI bridge: Amazon.com, Inc. Device 0200 2025-11-03T17:03:23.1233428Z 03:00.3 PCI bridge: Amazon.com, Inc. Device 0200 2025-11-03T17:03:23.1233991Z 03:00.4 PCI bridge: Amazon.com, Inc. Device 0200 2025-11-03T17:03:23.1234292Z 03:00.5 PCI bridge: Amazon.com, Inc. Device 0200 2025-11-03T17:03:23.1234535Z 03:00.6 PCI bridge: Amazon.com, Inc. Device 0200 2025-11-03T17:03:23.1234766Z 03:00.7 PCI bridge: Amazon.com, Inc. Device 0200 2025-11-03T17:03:23.1235007Z 03:01.0 PCI bridge: Amazon.com, Inc. Device 0200 2025-11-03T17:03:23.1235244Z 03:01.1 PCI bridge: Amazon.com, Inc. Device 0200 2025-11-03T17:03:23.1235485Z 03:01.2 PCI bridge: Amazon.com, Inc. Device 0200 2025-11-03T17:03:23.1236016Z 03:01.3 PCI bridge: Amazon.com, Inc. Device 0200 2025-11-03T17:03:23.1236301Z 03:01.4 PCI bridge: Amazon.com, Inc. Device 0200 2025-11-03T17:03:23.1236546Z 03:01.5 PCI bridge: Amazon.com, Inc. Device 0200 2025-11-03T17:03:23.1236785Z 03:01.6 PCI bridge: Amazon.com, Inc. Device 0200 2025-11-03T17:03:23.1237019Z 03:01.7 PCI bridge: Amazon.com, Inc. Device 0200 2025-11-03T17:03:23.1237260Z 03:02.0 PCI bridge: Amazon.com, Inc. Device 0200 2025-11-03T17:03:23.1237504Z 03:02.1 PCI bridge: Amazon.com, Inc. Device 0200 2025-11-03T17:03:23.1237744Z 03:02.2 PCI bridge: Amazon.com, Inc. Device 0200 2025-11-03T17:03:23.1237987Z 03:02.3 PCI bridge: Amazon.com, Inc. Device 0200 2025-11-03T17:03:23.1238221Z 03:02.4 PCI bridge: Amazon.com, Inc. Device 0200 2025-11-03T17:03:23.1238460Z 03:02.5 PCI bridge: Amazon.com, Inc. Device 0200 2025-11-03T17:03:23.1238705Z 03:02.6 PCI bridge: Amazon.com, Inc. Device 0200 2025-11-03T17:03:23.1238946Z 03:02.7 PCI bridge: Amazon.com, Inc. Device 0200 2025-11-03T17:03:23.1239183Z 03:03.0 PCI bridge: Amazon.com, Inc. Device 0200 2025-11-03T17:03:23.1239615Z 03:03.1 PCI bridge: Amazon.com, Inc. Device 0200 2025-11-03T17:03:23.1239855Z 03:03.2 PCI bridge: Amazon.com, Inc. Device 0200 2025-11-03T17:03:23.1240101Z 03:03.3 PCI bridge: Amazon.com, Inc. Device 0200 2025-11-03T17:03:23.1240346Z 03:03.4 PCI bridge: Amazon.com, Inc. Device 0200 2025-11-03T17:03:23.1240585Z 03:03.5 PCI bridge: Amazon.com, Inc. Device 0200 2025-11-03T17:03:23.1240826Z 03:03.6 PCI bridge: Amazon.com, Inc. Device 0200 2025-11-03T17:03:23.1241064Z 03:03.7 PCI bridge: Amazon.com, Inc. Device 0200 2025-11-03T17:03:23.1241303Z 24:00.0 PCI bridge: Amazon.com, Inc. Device 0200 2025-11-03T17:03:23.1241549Z 25:00.0 PCI bridge: Amazon.com, Inc. Device 0200 2025-11-03T17:03:23.1241797Z 26:00.0 PCI bridge: Amazon.com, Inc. Device 0200 2025-11-03T17:03:23.1242039Z 26:00.1 PCI bridge: Amazon.com, Inc. Device 0200 2025-11-03T17:03:23.1242277Z 26:00.2 PCI bridge: Amazon.com, Inc. Device 0200 2025-11-03T17:03:23.1242524Z 26:00.3 PCI bridge: Amazon.com, Inc. Device 0200 2025-11-03T17:03:23.1242769Z 26:00.4 PCI bridge: Amazon.com, Inc. Device 0200 2025-11-03T17:03:23.1243010Z 26:00.5 PCI bridge: Amazon.com, Inc. Device 0200 2025-11-03T17:03:23.1243252Z 26:00.6 PCI bridge: Amazon.com, Inc. Device 0200 2025-11-03T17:03:23.1243495Z 26:00.7 PCI bridge: Amazon.com, Inc. Device 0200 2025-11-03T17:03:23.1243739Z 26:01.0 PCI bridge: Amazon.com, Inc. Device 0200 2025-11-03T17:03:23.1244057Z 27:00.0 Ethernet controller: Amazon.com, Inc. Elastic Network Adapter (ENA) 2025-11-03T17:03:23.1244406Z 30:00.0 PCI bridge: Amazon.com, Inc. Device 0200 2025-11-03T17:03:23.1244654Z 31:00.0 PCI bridge: Amazon.com, Inc. Device 0200 2025-11-03T17:03:23.1244907Z 32:00.0 PCI bridge: Amazon.com, Inc. Device 0200 2025-11-03T17:03:23.1245156Z 32:00.1 PCI bridge: Amazon.com, Inc. Device 0200 2025-11-03T17:03:23.1245401Z 32:00.2 PCI bridge: Amazon.com, Inc. Device 0200 2025-11-03T17:03:23.1245644Z 32:00.3 PCI bridge: Amazon.com, Inc. Device 0200 2025-11-03T17:03:23.1245967Z 33:00.0 Non-Volatile memory controller: Amazon.com, Inc. NVMe SSD Controller 2025-11-03T17:03:23.1246360Z 34:00.0 Non-Volatile memory controller: Amazon.com, Inc. NVMe SSD Controller 2025-11-03T17:03:23.1246743Z 35:00.0 Non-Volatile memory controller: Amazon.com, Inc. NVMe SSD Controller 2025-11-03T17:03:23.1247117Z 36:00.0 Non-Volatile memory controller: Amazon.com, Inc. NVMe SSD Controller 2025-11-03T17:03:23.1247436Z 37:00.0 PCI bridge: Amazon.com, Inc. Device 0200 2025-11-03T17:03:23.1247718Z 38:00.0 3D controller: NVIDIA Corporation AD104GL [L4] (rev a1) 2025-11-03T17:03:23.1248011Z 39:00.0 PCI bridge: Amazon.com, Inc. Device 0200 2025-11-03T17:03:23.1248299Z 3a:00.0 3D controller: NVIDIA Corporation AD104GL [L4] (rev a1) 2025-11-03T17:03:23.1248644Z 3b:00.0 PCI bridge: Amazon.com, Inc. Device 0200 2025-11-03T17:03:23.1248928Z 3c:00.0 3D controller: NVIDIA Corporation AD104GL [L4] (rev a1) 2025-11-03T17:03:23.1249207Z 3d:00.0 PCI bridge: Amazon.com, Inc. Device 0200 2025-11-03T17:03:23.1249561Z 3e:00.0 3D controller: NVIDIA Corporation AD104GL [L4] (rev a1) 2025-11-03T17:03:23.1249832Z + lsmod 2025-11-03T17:03:23.1261511Z Module Size Used by 2025-11-03T17:03:23.1262196Z nvidia_uvm 1925120 0 2025-11-03T17:03:23.1262714Z nvidia 14286848 1 nvidia_uvm 2025-11-03T17:03:23.1263252Z drm 602112 1 nvidia 2025-11-03T17:03:23.1263814Z drm_panel_orientation_quirks 32768 1 drm 2025-11-03T17:03:23.1264299Z backlight 24576 1 drm 2025-11-03T17:03:23.1264519Z i2c_core 110592 2 nvidia,drm 2025-11-03T17:03:23.1264741Z xt_conntrack 16384 1 2025-11-03T17:03:23.1264937Z nft_chain_nat 16384 3 2025-11-03T17:03:23.1265126Z xt_MASQUERADE 20480 1 2025-11-03T17:03:23.1265348Z nf_nat 57344 2 nft_chain_nat,xt_MASQUERADE 2025-11-03T17:03:23.1265594Z nf_conntrack_netlink 57344 0 2025-11-03T17:03:23.1265901Z nf_conntrack 184320 4 xt_conntrack,nf_nat,nf_conntrack_netlink,xt_MASQUERADE 2025-11-03T17:03:23.1266367Z nf_defrag_ipv6 24576 1 nf_conntrack 2025-11-03T17:03:23.1266603Z nf_defrag_ipv4 16384 1 nf_conntrack 2025-11-03T17:03:23.1266816Z xfrm_user 57344 1 2025-11-03T17:03:23.1267022Z xfrm_algo 16384 1 xfrm_user 2025-11-03T17:03:23.1267242Z xt_addrtype 16384 2 2025-11-03T17:03:23.1267433Z nft_compat 20480 4 2025-11-03T17:03:23.1267712Z nf_tables 311296 57 nft_compat,nft_chain_nat 2025-11-03T17:03:23.1268082Z nfnetlink 20480 4 nft_compat,nf_conntrack_netlink,nf_tables 2025-11-03T17:03:23.1268370Z br_netfilter 36864 0 2025-11-03T17:03:23.1268583Z bridge 323584 1 br_netfilter 2025-11-03T17:03:23.1268808Z stp 16384 1 bridge 2025-11-03T17:03:23.1269015Z llc 16384 2 bridge,stp 2025-11-03T17:03:23.1269228Z overlay 167936 0 2025-11-03T17:03:23.1269429Z tls 139264 0 2025-11-03T17:03:23.1269612Z nls_ascii 16384 1 2025-11-03T17:03:23.1269790Z nls_cp437 20480 1 2025-11-03T17:03:23.1269974Z vfat 24576 1 2025-11-03T17:03:23.1270154Z fat 86016 1 vfat 2025-11-03T17:03:23.1270354Z sunrpc 700416 1 2025-11-03T17:03:23.1270534Z ghash_clmulni_intel 16384 0 2025-11-03T17:03:23.1270726Z i8042 45056 0 2025-11-03T17:03:23.1270903Z ena 184320 0 2025-11-03T17:03:23.1271101Z serio 28672 3 i8042 2025-11-03T17:03:23.1271307Z button 24576 0 2025-11-03T17:03:23.1271492Z sch_fq_codel 20480 9 2025-11-03T17:03:23.1271671Z fuse 184320 1 2025-11-03T17:03:23.1271842Z dm_mod 188416 0 2025-11-03T17:03:23.1272010Z configfs 57344 1 2025-11-03T17:03:23.1272188Z loop 36864 0 2025-11-03T17:03:23.1272367Z dmi_sysfs 20480 0 2025-11-03T17:03:23.1272549Z crc32_pclmul 16384 0 2025-11-03T17:03:23.1272730Z crc32c_intel 24576 0 2025-11-03T17:03:23.1272913Z efivarfs 24576 1 2025-11-03T17:03:23.1273088Z + modinfo nvidia 2025-11-03T17:03:23.1283077Z filename: /lib/modules/6.1.150-174.273.amzn2023.x86_64/kernel/drivers/video/nvidia.ko 2025-11-03T17:03:23.1283707Z import_ns: DMA_BUF 2025-11-03T17:03:23.1284008Z alias: char-major-195-* 2025-11-03T17:03:23.1284316Z version: 580.82.07 2025-11-03T17:03:23.1284590Z supported: external 2025-11-03T17:03:23.1284850Z license: Dual MIT/GPL 2025-11-03T17:03:23.1285119Z firmware: nvidia/580.82.07/gsp_tu10x.bin 2025-11-03T17:03:23.1285422Z firmware: nvidia/580.82.07/gsp_ga10x.bin 2025-11-03T17:03:23.1285718Z srcversion: BA7240A71DCF7DC6FE88C1D 2025-11-03T17:03:23.1286030Z alias: of:N*T*Cnvidia,tegra264-displayC* 2025-11-03T17:03:23.1286359Z alias: of:N*T*Cnvidia,tegra264-display 2025-11-03T17:03:23.1286671Z alias: of:N*T*Cnvidia,tegra234-displayC* 2025-11-03T17:03:23.1287164Z alias: of:N*T*Cnvidia,tegra234-display 2025-11-03T17:03:23.1287501Z alias: pci:v000010DEd*sv*sd*bc06sc80i00* 2025-11-03T17:03:23.1287812Z alias: pci:v000010DEd*sv*sd*bc03sc02i00* 2025-11-03T17:03:23.1288103Z alias: pci:v000010DEd*sv*sd*bc03sc00i00* 2025-11-03T17:03:23.1288491Z depends: i2c-core,drm 2025-11-03T17:03:23.1288735Z retpoline: Y 2025-11-03T17:03:23.1288944Z name: nvidia 2025-11-03T17:03:23.1289291Z vermagic: 6.1.150-174.273.amzn2023.x86_64 SMP preempt mod_unload modversions 2025-11-03T17:03:23.1289737Z parm: NvSwitchRegDwords:NvSwitch regkey (charp) 2025-11-03T17:03:23.1290147Z parm: NvSwitchBlacklist:NvSwitchBlacklist=uuid[,uuid...] (charp) 2025-11-03T17:03:23.1290536Z parm: NVreg_ResmanDebugLevel:int 2025-11-03T17:03:23.1290819Z parm: NVreg_RmLogonRC:int 2025-11-03T17:03:23.1291101Z parm: NVreg_ModifyDeviceFiles:int 2025-11-03T17:03:23.1291408Z parm: NVreg_DeviceFileUID:int 2025-11-03T17:03:23.1291790Z parm: NVreg_DeviceFileGID:int 2025-11-03T17:03:23.1292058Z parm: NVreg_DeviceFileMode:int 2025-11-03T17:03:23.1292392Z parm: NVreg_InitializeSystemMemoryAllocations:int 2025-11-03T17:03:23.1292751Z parm: NVreg_UsePageAttributeTable:int 2025-11-03T17:03:23.1293085Z parm: NVreg_EnablePCIeGen3:int 2025-11-03T17:03:23.1293401Z parm: NVreg_EnableMSI:int 2025-11-03T17:03:23.1293726Z parm: NVreg_EnableStreamMemOPs:int 2025-11-03T17:03:23.1294121Z parm: NVreg_RestrictProfilingToAdminUsers:int 2025-11-03T17:03:23.1294506Z parm: NVreg_PreserveVideoMemoryAllocations:int 2025-11-03T17:03:23.1294841Z parm: NVreg_EnableS0ixPowerManagement:int 2025-11-03T17:03:23.1295208Z parm: NVreg_S0ixPowerManagementVideoMemoryThreshold:int 2025-11-03T17:03:23.1295550Z parm: NVreg_DynamicPowerManagement:int 2025-11-03T17:03:23.1295867Z parm: NVreg_DynamicPowerManagementVideoMemoryThreshold:int 2025-11-03T17:03:23.1296174Z parm: NVreg_EnableGpuFirmware:int 2025-11-03T17:03:23.1296417Z parm: NVreg_EnableGpuFirmwareLogs:int 2025-11-03T17:03:23.1296690Z parm: NVreg_OpenRmEnableUnsupportedGpus:int 2025-11-03T17:03:23.1296966Z parm: NVreg_EnableUserNUMAManagement:int 2025-11-03T17:03:23.1297226Z parm: NVreg_MemoryPoolSize:int 2025-11-03T17:03:23.1297462Z parm: NVreg_KMallocHeapMaxSize:int 2025-11-03T17:03:23.1297709Z parm: NVreg_VMallocHeapMaxSize:int 2025-11-03T17:03:23.1297950Z parm: NVreg_IgnoreMMIOCheck:int 2025-11-03T17:03:23.1298180Z parm: NVreg_NvLinkDisable:int 2025-11-03T17:03:23.1298433Z parm: NVreg_EnablePCIERelaxedOrderingMode:int 2025-11-03T17:03:23.1298699Z parm: NVreg_RegisterPCIDriver:int 2025-11-03T17:03:23.1298958Z parm: NVreg_RegisterPlatformDeviceDriver:int 2025-11-03T17:03:23.1299234Z parm: NVreg_EnableResizableBar:int 2025-11-03T17:03:23.1299497Z parm: NVreg_EnableDbgBreakpoint:int 2025-11-03T17:03:23.1299745Z parm: NVreg_EnableNonblockingOpen:int 2025-11-03T17:03:23.1300006Z parm: NVreg_CoherentGPUMemoryMode:charp 2025-11-03T17:03:23.1300261Z parm: NVreg_RegistryDwords:charp 2025-11-03T17:03:23.1300515Z parm: NVreg_RegistryDwordsPerDevice:charp 2025-11-03T17:03:23.1300755Z parm: NVreg_RmMsg:charp 2025-11-03T17:03:23.1300971Z parm: NVreg_GpuBlacklist:charp 2025-11-03T17:03:23.1301209Z parm: NVreg_TemporaryFilePath:charp 2025-11-03T17:03:23.1301445Z parm: NVreg_ExcludedGpus:charp 2025-11-03T17:03:23.1301668Z parm: NVreg_DmaRemapPeerMmio:int 2025-11-03T17:03:23.1301907Z parm: NVreg_RmNvlinkBandwidth:charp 2025-11-03T17:03:23.1302179Z parm: NVreg_RmNvlinkBandwidthLinkCount:int 2025-11-03T17:03:23.1302434Z parm: NVreg_ImexChannelCount:int 2025-11-03T17:03:23.1302668Z parm: NVreg_CreateImexChannel0:int 2025-11-03T17:03:23.1303019Z parm: NVreg_GrdmaPciTopoCheckOverride:int 2025-11-03T17:03:23.1303288Z parm: rm_firmware_active:charp 2025-11-03T17:03:23.1303505Z + set +e 2025-11-03T17:03:23.1303641Z + nvidia-smi 2025-11-03T17:03:25.7790251Z Mon Nov 3 17:03:25 2025 2025-11-03T17:03:25.7791121Z +-----------------------------------------------------------------------------------------+ 2025-11-03T17:03:25.7791695Z | NVIDIA-SMI 580.82.07 Driver Version: 580.82.07 CUDA Version: 13.0 | 2025-11-03T17:03:25.7792152Z +-----------------------------------------+------------------------+----------------------+ 2025-11-03T17:03:25.7792605Z | GPU Name Persistence-M | Bus-Id Disp.A | Volatile Uncorr. ECC | 2025-11-03T17:03:25.7793105Z | Fan Temp Perf Pwr:Usage/Cap | Memory-Usage | GPU-Util Compute M. | 2025-11-03T17:03:25.7793494Z | | | MIG M. | 2025-11-03T17:03:25.7794144Z |=========================================+========================+======================| 2025-11-03T17:03:25.8068199Z | 0 NVIDIA L4 Off | 00000000:38:00.0 Off | 0 | 2025-11-03T17:03:25.8068641Z | N/A 41C P0 28W / 72W | 0MiB / 23034MiB | 0% Default | 2025-11-03T17:03:25.8069015Z | | | N/A | 2025-11-03T17:03:25.8069379Z +-----------------------------------------+------------------------+----------------------+ 2025-11-03T17:03:25.8069773Z | 1 NVIDIA L4 Off | 00000000:3A:00.0 Off | 0 | 2025-11-03T17:03:25.8070153Z | N/A 39C P0 28W / 72W | 0MiB / 23034MiB | 0% Default | 2025-11-03T17:03:25.8070479Z | | | N/A | 2025-11-03T17:03:25.8070836Z +-----------------------------------------+------------------------+----------------------+ 2025-11-03T17:03:25.8071231Z | 2 NVIDIA L4 Off | 00000000:3C:00.0 Off | 0 | 2025-11-03T17:03:25.8071599Z | N/A 42C P0 27W / 72W | 0MiB / 23034MiB | 0% Default | 2025-11-03T17:03:25.8071924Z | | | N/A | 2025-11-03T17:03:25.8072264Z +-----------------------------------------+------------------------+----------------------+ 2025-11-03T17:03:25.8072645Z | 3 NVIDIA L4 Off | 00000000:3E:00.0 Off | 0 | 2025-11-03T17:03:25.8073016Z | N/A 40C P0 28W / 72W | 0MiB / 23034MiB | 3% Default | 2025-11-03T17:03:25.8073346Z | | | N/A | 2025-11-03T17:03:25.8073703Z +-----------------------------------------+------------------------+----------------------+ 2025-11-03T17:03:25.8075154Z 2025-11-03T17:03:25.8075442Z +-----------------------------------------------------------------------------------------+ 2025-11-03T17:03:25.8075848Z | Processes: | 2025-11-03T17:03:25.8076200Z | GPU GI CI PID Type Process name GPU Memory | 2025-11-03T17:03:25.8076505Z | ID ID Usage | 2025-11-03T17:03:25.8076750Z |=========================================================================================| 2025-11-03T17:03:25.8094964Z | No running processes found | 2025-11-03T17:03:25.8096302Z +-----------------------------------------------------------------------------------------+ 2025-11-03T17:03:27.1078290Z + nvidia-smi --query-gpu=gpu_name --format=csv,noheader --id=0 2025-11-03T17:03:29.7091168Z NVIDIA L4 2025-11-03T17:03:30.4483244Z + NVIDIA_SMI_STATUS=0 2025-11-03T17:03:30.4483686Z + '[' 0 -eq 0 ']' 2025-11-03T17:03:30.4484112Z + echo 'INFO: Ignoring allowed status 0' 2025-11-03T17:03:30.4484810Z + set -e 2025-11-03T17:03:30.4485296Z INFO: Ignoring allowed status 0 2025-11-03T17:03:30.4497373Z == Installing nvidia container toolkit for amzn2023 == 2025-11-03T17:03:30.4505922Z + sudo yum install -y yum-utils 2025-11-03T17:03:30.8798638Z Last metadata expiration check: 0:21:07 ago on Mon Nov 3 16:42:23 2025. 2025-11-03T17:03:30.9030840Z Package dnf-utils-4.3.0-13.amzn2023.0.5.noarch is already installed. 2025-11-03T17:03:30.9446877Z Dependencies resolved. 2025-11-03T17:03:30.9687064Z Nothing to do. 2025-11-03T17:03:30.9687490Z Complete! 2025-11-03T17:03:31.1196741Z + [[ amzn2023 == \a\m\z\n\2\0\2\3 ]] 2025-11-03T17:03:31.1197794Z + YUM_REPO_URL=https://nvidia.github.io/libnvidia-container/stable/rpm/nvidia-container-toolkit.repo 2025-11-03T17:03:31.1199448Z + sudo yum-config-manager --add-repo https://nvidia.github.io/libnvidia-container/stable/rpm/nvidia-container-toolkit.repo 2025-11-03T17:03:31.4338097Z Adding repo from: https://nvidia.github.io/libnvidia-container/stable/rpm/nvidia-container-toolkit.repo 2025-11-03T17:03:31.4863783Z + sudo yum install -y nvidia-container-toolkit-1.17.8 libnvidia-container-tools-1.17.8 libnvidia-container1-1.17.8 nvidia-container-toolkit-base-1.17.8 2025-11-03T17:03:32.0090412Z nvidia-container-toolkit 19 kB/s | 833 B 00:00 2025-11-03T17:03:32.0754922Z Dependencies resolved. 2025-11-03T17:03:32.0981833Z ================================================================================ 2025-11-03T17:03:32.0982623Z Package Arch Version Repository Size 2025-11-03T17:03:32.0983316Z ================================================================================ 2025-11-03T17:03:32.0983858Z Downgrading: 2025-11-03T17:03:32.0984523Z libnvidia-container-tools x86_64 1.17.8-1 nvidia-container-toolkit 40 k 2025-11-03T17:03:32.0985587Z libnvidia-container1 x86_64 1.17.8-1 nvidia-container-toolkit 1.0 M 2025-11-03T17:03:32.0986578Z nvidia-container-toolkit x86_64 1.17.8-1 nvidia-container-toolkit 1.2 M 2025-11-03T17:03:32.0987594Z nvidia-container-toolkit-base x86_64 1.17.8-1 nvidia-container-toolkit 5.8 M 2025-11-03T17:03:32.0988223Z 2025-11-03T17:03:32.0988383Z Transaction Summary 2025-11-03T17:03:32.0988818Z ================================================================================ 2025-11-03T17:03:32.0989346Z Downgrade 4 Packages 2025-11-03T17:03:32.0989611Z 2025-11-03T17:03:32.0989792Z Total download size: 8.0 M 2025-11-03T17:03:32.0990251Z Downloading Packages: 2025-11-03T17:03:32.1840201Z (1/4): libnvidia-container-tools-1.17.8-1.x86_6 482 kB/s | 40 kB 00:00 2025-11-03T17:03:32.2533140Z (2/4): libnvidia-container1-1.17.8-1.x86_64.rpm 6.4 MB/s | 1.0 MB 00:00 2025-11-03T17:03:32.3206747Z (3/4): nvidia-container-toolkit-1.17.8-1.x86_64 5.6 MB/s | 1.2 MB 00:00 2025-11-03T17:03:32.4483836Z (4/4): nvidia-container-toolkit-base-1.17.8-1.x 22 MB/s | 5.8 MB 00:00 2025-11-03T17:03:32.4503022Z -------------------------------------------------------------------------------- 2025-11-03T17:03:32.4506818Z Total 23 MB/s | 8.0 MB 00:00 2025-11-03T17:03:32.4509853Z Running transaction check 2025-11-03T17:03:32.4640516Z Transaction check succeeded. 2025-11-03T17:03:32.4640971Z Running transaction test 2025-11-03T17:03:32.5004481Z Transaction test succeeded. 2025-11-03T17:03:32.5007640Z Running transaction 2025-11-03T17:03:33.2941593Z Preparing : 1/1 2025-11-03T17:03:33.4261558Z Downgrading : nvidia-container-toolkit-base-1.17.8-1.x86_64 1/8 2025-11-03T17:03:33.4688906Z Downgrading : libnvidia-container1-1.17.8-1.x86_64 2/8 2025-11-03T17:03:33.6186176Z Running scriptlet: libnvidia-container1-1.17.8-1.x86_64 2/8 2025-11-03T17:03:33.7589660Z Downgrading : libnvidia-container-tools-1.17.8-1.x86_64 3/8 2025-11-03T17:03:33.8284706Z Downgrading : nvidia-container-toolkit-1.17.8-1.x86_64 4/8 2025-11-03T17:03:33.9200590Z Running scriptlet: nvidia-container-toolkit-1.17.8-1.x86_64 4/8 2025-11-03T17:03:33.9201657Z Cleanup : nvidia-container-toolkit-1.18.0-1.x86_64 5/8 2025-11-03T17:03:33.9631801Z Running scriptlet: nvidia-container-toolkit-1.18.0-1.x86_64 5/8 2025-11-03T17:03:34.0311060Z Cleanup : libnvidia-container-tools-1.18.0-1.x86_64 6/8 2025-11-03T17:03:34.0324219Z Cleanup : libnvidia-container1-1.18.0-1.x86_64 7/8 2025-11-03T17:03:34.1679368Z Running scriptlet: libnvidia-container1-1.18.0-1.x86_64 7/8 2025-11-03T17:03:34.2006267Z Cleanup : nvidia-container-toolkit-base-1.18.0-1.x86_64 8/8 2025-11-03T17:03:34.2476439Z Running scriptlet: nvidia-container-toolkit-1.17.8-1.x86_64 8/8 2025-11-03T17:05:08.8493225Z Running scriptlet: nvidia-container-toolkit-base-1.18.0-1.x86_64 8/8 2025-11-03T17:05:08.8497589Z Verifying : libnvidia-container-tools-1.17.8-1.x86_64 1/8 2025-11-03T17:05:08.8498715Z Verifying : libnvidia-container-tools-1.18.0-1.x86_64 2/8 2025-11-03T17:05:08.8499668Z Verifying : libnvidia-container1-1.17.8-1.x86_64 3/8 2025-11-03T17:05:08.8500582Z Verifying : libnvidia-container1-1.18.0-1.x86_64 4/8 2025-11-03T17:05:08.8501516Z Verifying : nvidia-container-toolkit-1.17.8-1.x86_64 5/8 2025-11-03T17:05:08.8502420Z Verifying : nvidia-container-toolkit-1.18.0-1.x86_64 6/8 2025-11-03T17:05:08.8503329Z Verifying : nvidia-container-toolkit-base-1.17.8-1.x86_64 7/8 2025-11-03T17:05:08.9831553Z Verifying : nvidia-container-toolkit-base-1.18.0-1.x86_64 8/8================================================================================ 2025-11-03T17:05:08.9832603Z WARNING: 2025-11-03T17:05:08.9833091Z A newer release of "Amazon Linux" is available. 2025-11-03T17:05:08.9833527Z 2025-11-03T17:05:08.9833698Z Available Versions: 2025-11-03T17:05:08.9833968Z 2025-11-03T17:05:08.9834129Z Version 2023.9.20250929: 2025-11-03T17:05:08.9834695Z Run the following command to upgrade to 2023.9.20250929: 2025-11-03T17:05:08.9835177Z 2025-11-03T17:05:08.9835402Z dnf upgrade --releasever=2023.9.20250929 2025-11-03T17:05:08.9835620Z 2025-11-03T17:05:08.9835690Z Release notes: 2025-11-03T17:05:08.9836017Z https://docs.aws.amazon.com/linux/al2023/release-notes/relnotes-2023.9.20250929.html 2025-11-03T17:05:08.9836306Z 2025-11-03T17:05:08.9836377Z Version 2023.9.20251014: 2025-11-03T17:05:08.9836617Z Run the following command to upgrade to 2023.9.20251014: 2025-11-03T17:05:08.9836815Z 2025-11-03T17:05:08.9836922Z dnf upgrade --releasever=2023.9.20251014 2025-11-03T17:05:08.9837083Z 2025-11-03T17:05:08.9837151Z Release notes: 2025-11-03T17:05:08.9837451Z https://docs.aws.amazon.com/linux/al2023/release-notes/relnotes-2023.9.20251014.html 2025-11-03T17:05:08.9837737Z 2025-11-03T17:05:08.9837801Z Version 2023.9.20251020: 2025-11-03T17:05:08.9838050Z Run the following command to upgrade to 2023.9.20251020: 2025-11-03T17:05:08.9838239Z 2025-11-03T17:05:08.9838338Z dnf upgrade --releasever=2023.9.20251020 2025-11-03T17:05:08.9838500Z 2025-11-03T17:05:08.9838570Z Release notes: 2025-11-03T17:05:08.9838870Z https://docs.aws.amazon.com/linux/al2023/release-notes/relnotes-2023.9.20251020.html 2025-11-03T17:05:08.9839159Z 2025-11-03T17:05:08.9839226Z Version 2023.9.20251027: 2025-11-03T17:05:08.9839462Z Run the following command to upgrade to 2023.9.20251027: 2025-11-03T17:05:08.9839650Z 2025-11-03T17:05:08.9839742Z dnf upgrade --releasever=2023.9.20251027 2025-11-03T17:05:08.9840302Z 2025-11-03T17:05:08.9840379Z Release notes: 2025-11-03T17:05:08.9840691Z https://docs.aws.amazon.com/linux/al2023/release-notes/relnotes-2023.9.20251027.html 2025-11-03T17:05:08.9840978Z 2025-11-03T17:05:08.9841068Z ================================================================================ 2025-11-03T17:05:09.0293291Z 2025-11-03T17:05:09.0293507Z 2025-11-03T17:05:09.0294272Z Downgraded: 2025-11-03T17:05:09.0295039Z libnvidia-container-tools-1.17.8-1.x86_64 2025-11-03T17:05:09.0296071Z libnvidia-container1-1.17.8-1.x86_64 2025-11-03T17:05:09.0297041Z nvidia-container-toolkit-1.17.8-1.x86_64 2025-11-03T17:05:09.0298038Z nvidia-container-toolkit-base-1.17.8-1.x86_64 2025-11-03T17:05:09.0298641Z 2025-11-03T17:05:09.0298789Z Complete! 2025-11-03T17:05:09.0750963Z + sudo systemctl restart docker 2025-11-03T17:05:20.3183406Z Mon Nov 3 17:05:20 2025 2025-11-03T17:05:20.3183854Z +-----------------------------------------------------------------------------------------+ 2025-11-03T17:05:20.3184379Z | NVIDIA-SMI 580.82.07 Driver Version: 580.82.07 CUDA Version: 13.0 | 2025-11-03T17:05:20.3184849Z +-----------------------------------------+------------------------+----------------------+ 2025-11-03T17:05:20.3185320Z | GPU Name Persistence-M | Bus-Id Disp.A | Volatile Uncorr. ECC | 2025-11-03T17:05:20.3185841Z | Fan Temp Perf Pwr:Usage/Cap | Memory-Usage | GPU-Util Compute M. | 2025-11-03T17:05:20.3186229Z | | | MIG M. | 2025-11-03T17:05:20.3186529Z |=========================================+========================+======================| 2025-11-03T17:05:20.3470459Z | 0 NVIDIA L4 On | 00000000:38:00.0 Off | 0 | 2025-11-03T17:05:20.3470922Z | N/A 41C P0 27W / 72W | 0MiB / 23034MiB | 0% Default | 2025-11-03T17:05:20.3471286Z | | | N/A | 2025-11-03T17:05:20.3471642Z +-----------------------------------------+------------------------+----------------------+ 2025-11-03T17:05:20.3472039Z | 1 NVIDIA L4 On | 00000000:3A:00.0 Off | 0 | 2025-11-03T17:05:20.3472422Z | N/A 39C P0 27W / 72W | 0MiB / 23034MiB | 0% Default | 2025-11-03T17:05:20.3472766Z | | | N/A | 2025-11-03T17:05:20.3473135Z +-----------------------------------------+------------------------+----------------------+ 2025-11-03T17:05:20.3473522Z | 2 NVIDIA L4 On | 00000000:3C:00.0 Off | 0 | 2025-11-03T17:05:20.3473919Z | N/A 41C P0 27W / 72W | 0MiB / 23034MiB | 0% Default | 2025-11-03T17:05:20.3474219Z | | | N/A | 2025-11-03T17:05:20.3474504Z +-----------------------------------------+------------------------+----------------------+ 2025-11-03T17:05:20.3474822Z | 3 NVIDIA L4 On | 00000000:3E:00.0 Off | 0 | 2025-11-03T17:05:20.3475113Z | N/A 39C P0 30W / 72W | 0MiB / 23034MiB | 4% Default | 2025-11-03T17:05:20.3475383Z | | | N/A | 2025-11-03T17:05:20.3475668Z +-----------------------------------------+------------------------+----------------------+ 2025-11-03T17:05:20.3477415Z 2025-11-03T17:05:20.3477582Z +-----------------------------------------------------------------------------------------+ 2025-11-03T17:05:20.3478220Z | Processes: | 2025-11-03T17:05:20.3478567Z | GPU GI CI PID Type Process name GPU Memory | 2025-11-03T17:05:20.3478874Z | ID ID Usage | 2025-11-03T17:05:20.3479125Z |=========================================================================================| 2025-11-03T17:05:20.3495577Z | No running processes found | 2025-11-03T17:05:20.3496025Z +-----------------------------------------------------------------------------------------+ 2025-11-03T17:05:20.9449470Z Unable to find image 'public.ecr.aws/docker/library/python:3.13' locally 2025-11-03T17:05:21.1756537Z 3.13: Pulling from docker/library/python 2025-11-03T17:05:21.2644902Z 795dbedde24d: Pulling fs layer 2025-11-03T17:05:21.2645284Z 89d573bf42b3: Pulling fs layer 2025-11-03T17:05:21.2645598Z 26dfe2fac1c4: Pulling fs layer 2025-11-03T17:05:21.2645920Z 79d5bd8a8d26: Pulling fs layer 2025-11-03T17:05:21.2646560Z d53b4b1b9ebd: Pulling fs layer 2025-11-03T17:05:21.2646875Z 73ea62972d18: Pulling fs layer 2025-11-03T17:05:21.2647130Z eb23eeea1475: Pulling fs layer 2025-11-03T17:05:21.2647354Z 73ea62972d18: Waiting 2025-11-03T17:05:21.2647568Z d53b4b1b9ebd: Waiting 2025-11-03T17:05:21.2647770Z eb23eeea1475: Waiting 2025-11-03T17:05:21.2647979Z 79d5bd8a8d26: Waiting 2025-11-03T17:05:21.3880800Z 89d573bf42b3: Verifying Checksum 2025-11-03T17:05:21.3881107Z 89d573bf42b3: Download complete 2025-11-03T17:05:21.4464910Z 795dbedde24d: Verifying Checksum 2025-11-03T17:05:21.4465295Z 795dbedde24d: Download complete 2025-11-03T17:05:21.5122281Z d53b4b1b9ebd: Verifying Checksum 2025-11-03T17:05:21.5122566Z d53b4b1b9ebd: Download complete 2025-11-03T17:05:21.5824429Z 26dfe2fac1c4: Verifying Checksum 2025-11-03T17:05:21.5825073Z 26dfe2fac1c4: Download complete 2025-11-03T17:05:21.6359648Z eb23eeea1475: Verifying Checksum 2025-11-03T17:05:21.6360154Z eb23eeea1475: Download complete 2025-11-03T17:05:21.6508338Z 73ea62972d18: Verifying Checksum 2025-11-03T17:05:21.6508642Z 73ea62972d18: Download complete 2025-11-03T17:05:22.3137056Z 79d5bd8a8d26: Verifying Checksum 2025-11-03T17:05:22.3137360Z 79d5bd8a8d26: Download complete 2025-11-03T17:05:22.8883217Z 795dbedde24d: Pull complete 2025-11-03T17:05:23.4911417Z 89d573bf42b3: Pull complete 2025-11-03T17:05:25.5429923Z 26dfe2fac1c4: Pull complete 2025-11-03T17:05:30.9820571Z 79d5bd8a8d26: Pull complete 2025-11-03T17:05:31.2300013Z d53b4b1b9ebd: Pull complete 2025-11-03T17:05:32.1614635Z 73ea62972d18: Pull complete 2025-11-03T17:05:32.4165278Z eb23eeea1475: Pull complete 2025-11-03T17:05:32.5568087Z Digest: sha256:12513c633252a28bcfee85839aa384e1af322f11275779c6645076c6cd0cfe52 2025-11-03T17:05:32.6225484Z Status: Downloaded newer image for public.ecr.aws/docker/library/python:3.13 2025-11-03T17:05:40.2622357Z Mon Nov 3 17:05:40 2025 2025-11-03T17:05:40.2623291Z +-----------------------------------------------------------------------------------------+ 2025-11-03T17:05:40.2624303Z | NVIDIA-SMI 580.82.07 Driver Version: 580.82.07 CUDA Version: 13.0 | 2025-11-03T17:05:40.2625248Z +-----------------------------------------+------------------------+----------------------+ 2025-11-03T17:05:40.2626153Z | GPU Name Persistence-M | Bus-Id Disp.A | Volatile Uncorr. ECC | 2025-11-03T17:05:40.2627092Z | Fan Temp Perf Pwr:Usage/Cap | Memory-Usage | GPU-Util Compute M. | 2025-11-03T17:05:40.2627820Z | | | MIG M. | 2025-11-03T17:05:40.2628373Z |=========================================+========================+======================| 2025-11-03T17:05:40.3066971Z | 0 NVIDIA L4 On | 00000000:38:00.0 Off | 0 | 2025-11-03T17:05:40.3067836Z | N/A 39C P8 13W / 72W | 0MiB / 23034MiB | 0% Default | 2025-11-03T17:05:40.3069062Z | | | N/A | 2025-11-03T17:05:40.3069801Z +-----------------------------------------+------------------------+----------------------+ 2025-11-03T17:05:40.3070574Z | 1 NVIDIA L4 On | 00000000:3A:00.0 Off | 0 | 2025-11-03T17:05:40.3071336Z | N/A 37C P8 12W / 72W | 0MiB / 23034MiB | 0% Default | 2025-11-03T17:05:40.3072018Z | | | N/A | 2025-11-03T17:05:40.3072393Z +-----------------------------------------+------------------------+----------------------+ 2025-11-03T17:05:40.3072706Z | 2 NVIDIA L4 On | 00000000:3C:00.0 Off | 0 | 2025-11-03T17:05:40.3073027Z | N/A 40C P8 12W / 72W | 0MiB / 23034MiB | 0% Default | 2025-11-03T17:05:40.3073307Z | | | N/A | 2025-11-03T17:05:40.3073763Z +-----------------------------------------+------------------------+----------------------+ 2025-11-03T17:05:40.3074077Z | 3 NVIDIA L4 On | 00000000:3E:00.0 Off | 0 | 2025-11-03T17:05:40.3074389Z | N/A 38C P8 12W / 72W | 0MiB / 23034MiB | 0% Default | 2025-11-03T17:05:40.3074662Z | | | N/A | 2025-11-03T17:05:40.3074949Z +-----------------------------------------+------------------------+----------------------+ 2025-11-03T17:05:40.3085513Z 2025-11-03T17:05:40.3085741Z +-----------------------------------------------------------------------------------------+ 2025-11-03T17:05:40.3086204Z | Processes: | 2025-11-03T17:05:40.3086646Z | GPU GI CI PID Type Process name GPU Memory | 2025-11-03T17:05:40.3087062Z | ID ID Usage | 2025-11-03T17:05:40.3087388Z |=========================================================================================| 2025-11-03T17:05:40.3118370Z | No running processes found | 2025-11-03T17:05:40.3119297Z +-----------------------------------------------------------------------------------------+ 2025-11-03T17:05:42.8028592Z Command completed after 1 attempt(s). 2025-11-03T17:05:42.8123078Z Prepare all required actions 2025-11-03T17:05:42.8147394Z ##[group]Run ./.github/actions/get-workflow-job-id 2025-11-03T17:05:42.8147648Z with: 2025-11-03T17:05:42.8148184Z github-token: *** 2025-11-03T17:05:42.8148368Z env: 2025-11-03T17:05:42.8148515Z GIT_DEFAULT_BRANCH: main 2025-11-03T17:05:42.8148759Z GPU_FLAG: --gpus all -e NVIDIA_DRIVER_CAPABILITIES=all 2025-11-03T17:05:42.8149020Z ##[endgroup] 2025-11-03T17:05:42.8179901Z ##[group]Run set -eux 2025-11-03T17:05:42.8180110Z set -eux 2025-11-03T17:05:42.8180426Z python3 .github/scripts/get_workflow_job_id.py "${GITHUB_RUN_ID}" "${RUNNER_NAME}" 2025-11-03T17:05:42.8193171Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2025-11-03T17:05:42.8193446Z env: 2025-11-03T17:05:42.8193618Z GIT_DEFAULT_BRANCH: main 2025-11-03T17:05:42.8193864Z GPU_FLAG: --gpus all -e NVIDIA_DRIVER_CAPABILITIES=all 2025-11-03T17:05:42.8194237Z GITHUB_TOKEN: *** 2025-11-03T17:05:42.8194410Z ##[endgroup] 2025-11-03T17:05:42.8234720Z + python3 .github/scripts/get_workflow_job_id.py 19040285420 i-05b4d4af410353347 2025-11-03T17:05:43.6484663Z Setting output job-id=54383051348 2025-11-03T17:05:43.6485847Z Setting output job-name=ci-vllm-test / test (vllm_distributed_test_28_failure_test, 1, 1, linux.g6.12xlarge.nvidia.gpu) 2025-11-03T17:05:43.6602448Z ##[group]Run python3 -m pip install psutil==5.9.8 dataclasses_json==0.6.7 nvidia-ml-py==11.525.84 2025-11-03T17:05:43.6603032Z python3 -m pip install psutil==5.9.8 dataclasses_json==0.6.7 nvidia-ml-py==11.525.84 2025-11-03T17:05:43.6603757Z python3 -m tools.stats.monitor --log-interval "$MONITOR_LOG_INTERVAL" --data-collect-interval "$MONITOR_DATA_COLLECT_INTERVAL" > usage_log.txt 2>&1 & 2025-11-03T17:05:43.6604360Z echo "monitor-script-pid=${!}" >> "${GITHUB_OUTPUT}" 2025-11-03T17:05:43.6612937Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2025-11-03T17:05:43.6613230Z env: 2025-11-03T17:05:43.6613400Z GIT_DEFAULT_BRANCH: main 2025-11-03T17:05:43.6613651Z GPU_FLAG: --gpus all -e NVIDIA_DRIVER_CAPABILITIES=all 2025-11-03T17:05:43.6613915Z JOB_ID: 54383051348 2025-11-03T17:05:43.6614279Z JOB_NAME: ci-vllm-test / test (vllm_distributed_test_28_failure_test, 1, 1, linux.g6.12xlarge.nvidia.gpu) 2025-11-03T17:05:43.6614703Z WORKFLOW_NAME: vllm-test 2025-11-03T17:05:43.6614903Z WORKFLOW_RUN_ID: 19040285420 2025-11-03T17:05:43.6615099Z MONITOR_LOG_INTERVAL: 5 2025-11-03T17:05:43.6615478Z MONITOR_DATA_COLLECT_INTERVAL: 1 2025-11-03T17:05:43.6615690Z ##[endgroup] 2025-11-03T17:05:43.9429209Z Defaulting to user installation because normal site-packages is not writeable 2025-11-03T17:05:44.2899631Z Collecting psutil==5.9.8 2025-11-03T17:05:44.3082072Z Downloading psutil-5.9.8-cp36-abi3-manylinux_2_12_x86_64.manylinux2010_x86_64.manylinux_2_17_x86_64.manylinux2014_x86_64.whl (288 kB) 2025-11-03T17:05:44.3800904Z Collecting dataclasses_json==0.6.7 2025-11-03T17:05:44.3943475Z Downloading dataclasses_json-0.6.7-py3-none-any.whl (28 kB) 2025-11-03T17:05:44.4217593Z Collecting nvidia-ml-py==11.525.84 2025-11-03T17:05:44.4258997Z Downloading nvidia_ml_py-11.525.84-py3-none-any.whl (34 kB) 2025-11-03T17:05:44.4556518Z Collecting typing-inspect<1,>=0.4.0 2025-11-03T17:05:44.4598767Z Downloading typing_inspect-0.9.0-py3-none-any.whl (8.8 kB) 2025-11-03T17:05:44.5576979Z Collecting marshmallow<4.0.0,>=3.18.0 2025-11-03T17:05:44.5618293Z Downloading marshmallow-3.26.1-py3-none-any.whl (50 kB) 2025-11-03T17:05:44.6142965Z Collecting packaging>=17.0 2025-11-03T17:05:44.6197800Z Downloading packaging-25.0-py3-none-any.whl (66 kB) 2025-11-03T17:05:44.6690578Z Collecting typing-extensions>=3.7.4 2025-11-03T17:05:44.6724750Z Downloading typing_extensions-4.15.0-py3-none-any.whl (44 kB) 2025-11-03T17:05:44.6928880Z Collecting mypy-extensions>=0.3.0 2025-11-03T17:05:44.6968716Z Downloading mypy_extensions-1.1.0-py3-none-any.whl (5.0 kB) 2025-11-03T17:05:44.7832676Z Installing collected packages: typing-extensions, packaging, mypy-extensions, typing-inspect, marshmallow, psutil, nvidia-ml-py, dataclasses-json 2025-11-03T17:05:45.0292587Z Successfully installed dataclasses-json-0.6.7 marshmallow-3.26.1 mypy-extensions-1.1.0 nvidia-ml-py-11.525.84 packaging-25.0 psutil-5.9.8 typing-extensions-4.15.0 typing-inspect-0.9.0 2025-11-03T17:05:45.1950610Z Prepare all required actions 2025-11-03T17:05:45.1950985Z Getting action download info 2025-11-03T17:05:45.3784082Z Download action repository 'seemethere/download-artifact-s3@v4' (SHA:1da556a7aa0a088e3153970611f6c432d58e80e6) 2025-11-03T17:05:45.7247161Z Download action repository 'actions/download-artifact@v4' (SHA:d3f86a106a0bac45b974a628896c90dbdf5c8093) 2025-11-03T17:05:46.0568937Z ##[group]Run ./.github/actions/download-build-artifacts 2025-11-03T17:05:46.0569205Z with: 2025-11-03T17:05:46.0569388Z name: linux-jammy-cuda12.8-py3.12-gcc11 2025-11-03T17:05:46.0569631Z s3-bucket: gha-artifacts 2025-11-03T17:05:46.0569811Z env: 2025-11-03T17:05:46.0569960Z GIT_DEFAULT_BRANCH: main 2025-11-03T17:05:46.0570215Z GPU_FLAG: --gpus all -e NVIDIA_DRIVER_CAPABILITIES=all 2025-11-03T17:05:46.0570471Z ##[endgroup] 2025-11-03T17:05:46.0603627Z ##[group]Run seemethere/download-artifact-s3@v4 2025-11-03T17:05:46.0603872Z with: 2025-11-03T17:05:46.0604091Z name: linux-jammy-cuda12.8-py3.12-gcc11 2025-11-03T17:05:46.0604332Z s3-bucket: gha-artifacts 2025-11-03T17:05:46.0604530Z region: us-east-1 2025-11-03T17:05:46.0604706Z env: 2025-11-03T17:05:46.0604861Z GIT_DEFAULT_BRANCH: main 2025-11-03T17:05:46.0605111Z GPU_FLAG: --gpus all -e NVIDIA_DRIVER_CAPABILITIES=all 2025-11-03T17:05:46.0605366Z ##[endgroup] 2025-11-03T17:05:46.4756801Z (node:67447) NOTE: We are formalizing our plans to enter AWS SDK for JavaScript (v2) into maintenance mode in 2023. 2025-11-03T17:05:46.4757261Z 2025-11-03T17:05:46.4757447Z Please migrate your code to use AWS SDK for JavaScript (v3). 2025-11-03T17:05:46.4757937Z For more information, check the migration guide at https://a.co/7PzMCcy 2025-11-03T17:05:46.4758459Z (Use `node --trace-warnings ...` to show where the warning was created) 2025-11-03T17:05:46.7294616Z Found 1 objects with prefix pytorch/pytorch/19040285420/linux-jammy-cuda12.8-py3.12-gcc11/ 2025-11-03T17:05:46.7295929Z Starting download (1/1): /home/ec2-user/actions-runner/_work/pytorch/pytorch/artifacts.zip 2025-11-03T17:06:15.0530485Z Finished download (1/1): /home/ec2-user/actions-runner/_work/pytorch/pytorch/artifacts.zip 2025-11-03T17:06:15.0536187Z Artifact download has finished successfully 2025-11-03T17:06:15.0798354Z ##[group]Run unzip -o artifacts.zip 2025-11-03T17:06:15.0798624Z unzip -o artifacts.zip 2025-11-03T17:06:15.0807703Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2025-11-03T17:06:15.0807984Z env: 2025-11-03T17:06:15.0808164Z GIT_DEFAULT_BRANCH: main 2025-11-03T17:06:15.0808517Z GPU_FLAG: --gpus all -e NVIDIA_DRIVER_CAPABILITIES=all 2025-11-03T17:06:15.0808777Z ##[endgroup] 2025-11-03T17:06:15.2065765Z Archive: artifacts.zip 2025-11-03T17:06:15.2068845Z creating: dist/ 2025-11-03T17:06:17.7125803Z inflating: dist/torch-2.10.0a0+git3f6538f-cp312-cp312-linux_x86_64.whl 2025-11-03T17:06:17.7128806Z creating: dist/vision/ 2025-11-03T17:06:17.7254297Z inflating: dist/vision/torchvision-0.25.0a0+cfbc5c2-cp312-cp312-linux_x86_64.whl 2025-11-03T17:06:17.7254806Z creating: dist/audio/ 2025-11-03T17:06:17.7327757Z inflating: dist/audio/torchaudio-2.10.0a0+3b0e7a6-cp312-cp312-linux_x86_64.whl 2025-11-03T17:06:17.7443401Z inflating: dist/.ninja_log 2025-11-03T17:06:17.7443705Z creating: dist/external/ 2025-11-03T17:06:17.7443991Z creating: dist/external/vllm/ 2025-11-03T17:06:17.7444645Z creating: dist/external/vllm/wheels/ 2025-11-03T17:06:17.7445569Z inflating: dist/external/vllm/wheels/build_summary.txt 2025-11-03T17:06:17.7446260Z creating: dist/external/vllm/wheels/flashinfer-python/ 2025-11-03T17:06:18.3596202Z inflating: dist/external/vllm/wheels/flashinfer-python/flashinfer_python-0.2.14.post1-cp39-abi3-linux_x86_64.whl 2025-11-03T17:06:18.3596865Z creating: dist/external/vllm/wheels/vllm/ 2025-11-03T17:06:21.0108928Z inflating: dist/external/vllm/wheels/vllm/vllm-0.11.0rc2.dev389+ge51928192.d20251103-cp38-abi3-linux_x86_64.whl 2025-11-03T17:06:21.0110476Z creating: dist/external/vllm/wheels/xformers/ 2025-11-03T17:06:22.5757546Z inflating: dist/external/vllm/wheels/xformers/xformers-0.0.33+5d4b92a5.d20251103-cp39-abi3-linux_x86_64.whl 2025-11-03T17:06:22.5758619Z creating: build/custom_test_artifacts/ 2025-11-03T17:06:22.5759214Z creating: build/custom_test_artifacts/custom-op-build/ 2025-11-03T17:06:22.5759818Z creating: build/custom_test_artifacts/custom-op-build/CMakeFiles/ 2025-11-03T17:06:22.5760548Z creating: build/custom_test_artifacts/custom-op-build/CMakeFiles/pkgRedirects/ 2025-11-03T17:06:22.5766752Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/CMakeConfigureLog.yaml 2025-11-03T17:06:22.5767645Z creating: build/custom_test_artifacts/custom-op-build/CMakeFiles/3.31.6/ 2025-11-03T17:06:22.5768567Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/3.31.6/CMakeSystem.cmake 2025-11-03T17:06:22.5769583Z creating: build/custom_test_artifacts/custom-op-build/CMakeFiles/3.31.6/CompilerIdC/ 2025-11-03T17:06:22.5770457Z creating: build/custom_test_artifacts/custom-op-build/CMakeFiles/3.31.6/CompilerIdC/tmp/ 2025-11-03T17:06:22.5772117Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/3.31.6/CompilerIdC/CMakeCCompilerId.c 2025-11-03T17:06:22.5773654Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/3.31.6/CompilerIdC/a.out 2025-11-03T17:06:22.5774727Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/3.31.6/CMakeCCompiler.cmake 2025-11-03T17:06:22.5775768Z creating: build/custom_test_artifacts/custom-op-build/CMakeFiles/3.31.6/CompilerIdCXX/ 2025-11-03T17:06:22.5776660Z creating: build/custom_test_artifacts/custom-op-build/CMakeFiles/3.31.6/CompilerIdCXX/tmp/ 2025-11-03T17:06:22.5778126Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/3.31.6/CompilerIdCXX/CMakeCXXCompilerId.cpp 2025-11-03T17:06:22.5779877Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/3.31.6/CompilerIdCXX/a.out 2025-11-03T17:06:22.5780914Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/3.31.6/CMakeCXXCompiler.cmake 2025-11-03T17:06:22.5782582Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/3.31.6/CMakeDetermineCompilerABI_C.bin 2025-11-03T17:06:22.5784448Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/3.31.6/CMakeDetermineCompilerABI_CXX.bin 2025-11-03T17:06:22.5785371Z creating: build/custom_test_artifacts/custom-op-build/CMakeFiles/3.31.6/CompilerIdCUDA/ 2025-11-03T17:06:22.5786086Z creating: build/custom_test_artifacts/custom-op-build/CMakeFiles/3.31.6/CompilerIdCUDA/tmp/ 2025-11-03T17:06:22.5837895Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/3.31.6/CompilerIdCUDA/tmp/CMakeCUDACompilerId.cpp4.ii 2025-11-03T17:06:22.5890367Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/3.31.6/CompilerIdCUDA/tmp/CMakeCUDACompilerId.cudafe1.cpp 2025-11-03T17:06:22.5891727Z extracting: build/custom_test_artifacts/custom-op-build/CMakeFiles/3.31.6/CompilerIdCUDA/tmp/CMakeCUDACompilerId.module_id 2025-11-03T17:06:22.5947404Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/3.31.6/CompilerIdCUDA/tmp/CMakeCUDACompilerId.cpp1.ii 2025-11-03T17:06:22.5948822Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/3.31.6/CompilerIdCUDA/tmp/CMakeCUDACompilerId.cudafe1.c 2025-11-03T17:06:22.5950369Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/3.31.6/CompilerIdCUDA/tmp/CMakeCUDACompilerId.cudafe1.gpu 2025-11-03T17:06:22.5951843Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/3.31.6/CompilerIdCUDA/tmp/CMakeCUDACompilerId.cudafe1.stub.c 2025-11-03T17:06:22.5953248Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/3.31.6/CompilerIdCUDA/tmp/CMakeCUDACompilerId.ptx 2025-11-03T17:06:22.5954569Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/3.31.6/CompilerIdCUDA/tmp/CMakeCUDACompilerId.sm_52.cubin 2025-11-03T17:06:22.5956152Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/3.31.6/CompilerIdCUDA/tmp/CMakeCUDACompilerId.fatbin 2025-11-03T17:06:22.5957565Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/3.31.6/CompilerIdCUDA/tmp/CMakeCUDACompilerId.fatbin.c 2025-11-03T17:06:22.5958670Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/3.31.6/CompilerIdCUDA/tmp/CMakeCUDACompilerId.o 2025-11-03T17:06:22.5959612Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/3.31.6/CompilerIdCUDA/tmp/a_dlink.sm_52.cubin 2025-11-03T17:06:22.5960599Z extracting: build/custom_test_artifacts/custom-op-build/CMakeFiles/3.31.6/CompilerIdCUDA/tmp/a_dlink.reg.c 2025-11-03T17:06:22.5961597Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/3.31.6/CompilerIdCUDA/tmp/a_dlink.fatbin 2025-11-03T17:06:22.5962484Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/3.31.6/CompilerIdCUDA/tmp/a_dlink.fatbin.c 2025-11-03T17:06:22.5963336Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/3.31.6/CompilerIdCUDA/tmp/a_dlink.o 2025-11-03T17:06:22.5964231Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/3.31.6/CompilerIdCUDA/CMakeCUDACompilerId.cu 2025-11-03T17:06:22.6028532Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/3.31.6/CompilerIdCUDA/a.out 2025-11-03T17:06:22.6029643Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/3.31.6/CMakeCUDACompiler.cmake 2025-11-03T17:06:22.6094137Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/3.31.6/CMakeDetermineCompilerABI_CUDA.bin 2025-11-03T17:06:22.6095149Z creating: build/custom_test_artifacts/custom-op-build/CMakeFiles/CMakeScratch/ 2025-11-03T17:06:22.6096052Z creating: build/custom_test_artifacts/custom-op-build/CMakeFiles/CMakeTmp/ 2025-11-03T17:06:22.6096971Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/cmake.check_cache 2025-11-03T17:06:22.6097964Z creating: build/custom_test_artifacts/custom-op-build/CMakeFiles/custom_ops.dir/ 2025-11-03T17:06:22.6098787Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/custom_ops.dir/compiler_depend.ts 2025-11-03T17:06:22.6099863Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/custom_ops.dir/compiler_depend.make 2025-11-03T17:06:22.6100716Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/custom_ops.dir/depend.make 2025-11-03T17:06:22.6101552Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/custom_ops.dir/link.txt 2025-11-03T17:06:22.6102396Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/custom_ops.dir/cmake_clean.cmake 2025-11-03T17:06:22.6103242Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/custom_ops.dir/build.make 2025-11-03T17:06:22.6104048Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/custom_ops.dir/DependInfo.cmake 2025-11-03T17:06:22.6104815Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/custom_ops.dir/flags.make 2025-11-03T17:06:22.6105579Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/custom_ops.dir/progress.make 2025-11-03T17:06:22.6123049Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/custom_ops.dir/op.cpp.o.d 2025-11-03T17:06:22.6298089Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/custom_ops.dir/op.cpp.o 2025-11-03T17:06:22.6299128Z creating: build/custom_test_artifacts/custom-op-build/CMakeFiles/test_custom_ops.dir/ 2025-11-03T17:06:22.6300397Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/test_custom_ops.dir/compiler_depend.ts 2025-11-03T17:06:22.6301613Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/test_custom_ops.dir/compiler_depend.make 2025-11-03T17:06:22.6302787Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/test_custom_ops.dir/depend.make 2025-11-03T17:06:22.6304069Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/test_custom_ops.dir/link.txt 2025-11-03T17:06:22.6305175Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/test_custom_ops.dir/cmake_clean.cmake 2025-11-03T17:06:22.6306293Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/test_custom_ops.dir/build.make 2025-11-03T17:06:22.6307286Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/test_custom_ops.dir/DependInfo.cmake 2025-11-03T17:06:22.6308102Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/test_custom_ops.dir/flags.make 2025-11-03T17:06:22.6308899Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/test_custom_ops.dir/progress.make 2025-11-03T17:06:22.6325350Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/test_custom_ops.dir/test_custom_ops.cpp.o.d 2025-11-03T17:06:22.6395109Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/test_custom_ops.dir/test_custom_ops.cpp.o 2025-11-03T17:06:22.6396443Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/CMakeDirectoryInformation.cmake 2025-11-03T17:06:22.6397603Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/TargetDirectories.txt 2025-11-03T17:06:22.6398589Z extracting: build/custom_test_artifacts/custom-op-build/CMakeFiles/progress.marks 2025-11-03T17:06:22.6399412Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/Makefile2 2025-11-03T17:06:22.6400274Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/Makefile.cmake 2025-11-03T17:06:22.6401115Z inflating: build/custom_test_artifacts/custom-op-build/detect_cuda_version.cc 2025-11-03T17:06:22.6403576Z inflating: build/custom_test_artifacts/custom-op-build/CMakeCache.txt 2025-11-03T17:06:22.6404372Z inflating: build/custom_test_artifacts/custom-op-build/Makefile 2025-11-03T17:06:22.6405294Z inflating: build/custom_test_artifacts/custom-op-build/cmake_install.cmake 2025-11-03T17:06:22.6554903Z inflating: build/custom_test_artifacts/custom-op-build/libcustom_ops.so 2025-11-03T17:06:22.6603411Z inflating: build/custom_test_artifacts/custom-op-build/test_custom_ops 2025-11-03T17:06:22.6604414Z creating: build/custom_test_artifacts/jit-hook-build/ 2025-11-03T17:06:22.6604980Z creating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/ 2025-11-03T17:06:22.6605669Z creating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/pkgRedirects/ 2025-11-03T17:06:22.6612303Z inflating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/CMakeConfigureLog.yaml 2025-11-03T17:06:22.6612960Z creating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/3.31.6/ 2025-11-03T17:06:22.6613744Z inflating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/3.31.6/CMakeSystem.cmake 2025-11-03T17:06:22.6614550Z creating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/3.31.6/CompilerIdC/ 2025-11-03T17:06:22.6615215Z creating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/3.31.6/CompilerIdC/tmp/ 2025-11-03T17:06:22.6616938Z inflating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/3.31.6/CompilerIdC/CMakeCCompilerId.c 2025-11-03T17:06:22.6618312Z inflating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/3.31.6/CompilerIdC/a.out 2025-11-03T17:06:22.6619286Z inflating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/3.31.6/CMakeCCompiler.cmake 2025-11-03T17:06:22.6620169Z creating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/3.31.6/CompilerIdCXX/ 2025-11-03T17:06:22.6620893Z creating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/3.31.6/CompilerIdCXX/tmp/ 2025-11-03T17:06:22.6623013Z inflating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/3.31.6/CompilerIdCXX/CMakeCXXCompilerId.cpp 2025-11-03T17:06:22.6624372Z inflating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/3.31.6/CompilerIdCXX/a.out 2025-11-03T17:06:22.6625792Z inflating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/3.31.6/CMakeCXXCompiler.cmake 2025-11-03T17:06:22.6627282Z inflating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/3.31.6/CMakeDetermineCompilerABI_C.bin 2025-11-03T17:06:22.6629399Z inflating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/3.31.6/CMakeDetermineCompilerABI_CXX.bin 2025-11-03T17:06:22.6630303Z creating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/3.31.6/CompilerIdCUDA/ 2025-11-03T17:06:22.6630997Z creating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/3.31.6/CompilerIdCUDA/tmp/ 2025-11-03T17:06:22.6682423Z inflating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/3.31.6/CompilerIdCUDA/tmp/CMakeCUDACompilerId.cpp4.ii 2025-11-03T17:06:22.6735289Z inflating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/3.31.6/CompilerIdCUDA/tmp/CMakeCUDACompilerId.cudafe1.cpp 2025-11-03T17:06:22.6736772Z extracting: build/custom_test_artifacts/jit-hook-build/CMakeFiles/3.31.6/CompilerIdCUDA/tmp/CMakeCUDACompilerId.module_id 2025-11-03T17:06:22.6792329Z inflating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/3.31.6/CompilerIdCUDA/tmp/CMakeCUDACompilerId.cpp1.ii 2025-11-03T17:06:22.6793331Z inflating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/3.31.6/CompilerIdCUDA/tmp/CMakeCUDACompilerId.cudafe1.c 2025-11-03T17:06:22.6794277Z inflating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/3.31.6/CompilerIdCUDA/tmp/CMakeCUDACompilerId.cudafe1.gpu 2025-11-03T17:06:22.6795244Z inflating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/3.31.6/CompilerIdCUDA/tmp/CMakeCUDACompilerId.cudafe1.stub.c 2025-11-03T17:06:22.6796165Z inflating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/3.31.6/CompilerIdCUDA/tmp/CMakeCUDACompilerId.ptx 2025-11-03T17:06:22.6797054Z inflating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/3.31.6/CompilerIdCUDA/tmp/CMakeCUDACompilerId.sm_52.cubin 2025-11-03T17:06:22.6797896Z inflating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/3.31.6/CompilerIdCUDA/tmp/CMakeCUDACompilerId.fatbin 2025-11-03T17:06:22.6798654Z inflating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/3.31.6/CompilerIdCUDA/tmp/CMakeCUDACompilerId.fatbin.c 2025-11-03T17:06:22.6800046Z inflating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/3.31.6/CompilerIdCUDA/tmp/CMakeCUDACompilerId.o 2025-11-03T17:06:22.6800892Z inflating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/3.31.6/CompilerIdCUDA/tmp/a_dlink.sm_52.cubin 2025-11-03T17:06:22.6801643Z extracting: build/custom_test_artifacts/jit-hook-build/CMakeFiles/3.31.6/CompilerIdCUDA/tmp/a_dlink.reg.c 2025-11-03T17:06:22.6802602Z inflating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/3.31.6/CompilerIdCUDA/tmp/a_dlink.fatbin 2025-11-03T17:06:22.6803696Z inflating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/3.31.6/CompilerIdCUDA/tmp/a_dlink.fatbin.c 2025-11-03T17:06:22.6804757Z inflating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/3.31.6/CompilerIdCUDA/tmp/a_dlink.o 2025-11-03T17:06:22.6807465Z inflating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/3.31.6/CompilerIdCUDA/CMakeCUDACompilerId.cu 2025-11-03T17:06:22.6871977Z inflating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/3.31.6/CompilerIdCUDA/a.out 2025-11-03T17:06:22.6872716Z inflating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/3.31.6/CMakeCUDACompiler.cmake 2025-11-03T17:06:22.6937833Z inflating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/3.31.6/CMakeDetermineCompilerABI_CUDA.bin 2025-11-03T17:06:22.6938558Z creating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/CMakeScratch/ 2025-11-03T17:06:22.6939112Z creating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/CMakeTmp/ 2025-11-03T17:06:22.6939692Z inflating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/cmake.check_cache 2025-11-03T17:06:22.6940303Z creating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/test_jit_hooks.dir/ 2025-11-03T17:06:22.6941180Z inflating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/test_jit_hooks.dir/compiler_depend.ts 2025-11-03T17:06:22.6941973Z inflating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/test_jit_hooks.dir/compiler_depend.make 2025-11-03T17:06:22.6942726Z inflating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/test_jit_hooks.dir/depend.make 2025-11-03T17:06:22.6943424Z inflating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/test_jit_hooks.dir/link.txt 2025-11-03T17:06:22.6944142Z inflating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/test_jit_hooks.dir/cmake_clean.cmake 2025-11-03T17:06:22.6945108Z inflating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/test_jit_hooks.dir/build.make 2025-11-03T17:06:22.6946142Z inflating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/test_jit_hooks.dir/DependInfo.cmake 2025-11-03T17:06:22.6946952Z inflating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/test_jit_hooks.dir/flags.make 2025-11-03T17:06:22.6947804Z inflating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/test_jit_hooks.dir/progress.make 2025-11-03T17:06:22.6966432Z inflating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/test_jit_hooks.dir/test_jit_hooks.cpp.o.d 2025-11-03T17:06:22.7021619Z inflating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/test_jit_hooks.dir/test_jit_hooks.cpp.o 2025-11-03T17:06:22.7022534Z inflating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/CMakeDirectoryInformation.cmake 2025-11-03T17:06:22.7023268Z inflating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/TargetDirectories.txt 2025-11-03T17:06:22.7024031Z extracting: build/custom_test_artifacts/jit-hook-build/CMakeFiles/progress.marks 2025-11-03T17:06:22.7025109Z inflating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/Makefile2 2025-11-03T17:06:22.7026887Z inflating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/Makefile.cmake 2025-11-03T17:06:22.7027530Z inflating: build/custom_test_artifacts/jit-hook-build/detect_cuda_version.cc 2025-11-03T17:06:22.7030040Z inflating: build/custom_test_artifacts/jit-hook-build/CMakeCache.txt 2025-11-03T17:06:22.7031129Z inflating: build/custom_test_artifacts/jit-hook-build/Makefile 2025-11-03T17:06:22.7031993Z inflating: build/custom_test_artifacts/jit-hook-build/cmake_install.cmake 2025-11-03T17:06:22.7065672Z inflating: build/custom_test_artifacts/jit-hook-build/test_jit_hooks 2025-11-03T17:06:22.7066168Z creating: build/custom_test_artifacts/custom-backend-build/ 2025-11-03T17:06:22.7066638Z creating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/ 2025-11-03T17:06:22.7067186Z creating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/pkgRedirects/ 2025-11-03T17:06:22.7074138Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/CMakeConfigureLog.yaml 2025-11-03T17:06:22.7074800Z creating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/3.31.6/ 2025-11-03T17:06:22.7075720Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/3.31.6/CMakeSystem.cmake 2025-11-03T17:06:22.7076382Z creating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/3.31.6/CompilerIdC/ 2025-11-03T17:06:22.7077035Z creating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/3.31.6/CompilerIdC/tmp/ 2025-11-03T17:06:22.7079250Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/3.31.6/CompilerIdC/CMakeCCompilerId.c 2025-11-03T17:06:22.7080825Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/3.31.6/CompilerIdC/a.out 2025-11-03T17:06:22.7081770Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/3.31.6/CMakeCCompiler.cmake 2025-11-03T17:06:22.7082505Z creating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/3.31.6/CompilerIdCXX/ 2025-11-03T17:06:22.7083382Z creating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/3.31.6/CompilerIdCXX/tmp/ 2025-11-03T17:06:22.7085443Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/3.31.6/CompilerIdCXX/CMakeCXXCompilerId.cpp 2025-11-03T17:06:22.7086833Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/3.31.6/CompilerIdCXX/a.out 2025-11-03T17:06:22.7087927Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/3.31.6/CMakeCXXCompiler.cmake 2025-11-03T17:06:22.7089847Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/3.31.6/CMakeDetermineCompilerABI_C.bin 2025-11-03T17:06:22.7091807Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/3.31.6/CMakeDetermineCompilerABI_CXX.bin 2025-11-03T17:06:22.7092461Z creating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/3.31.6/CompilerIdCUDA/ 2025-11-03T17:06:22.7093038Z creating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/3.31.6/CompilerIdCUDA/tmp/ 2025-11-03T17:06:22.7146152Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/3.31.6/CompilerIdCUDA/tmp/CMakeCUDACompilerId.cpp4.ii 2025-11-03T17:06:22.7198827Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/3.31.6/CompilerIdCUDA/tmp/CMakeCUDACompilerId.cudafe1.cpp 2025-11-03T17:06:22.7199870Z extracting: build/custom_test_artifacts/custom-backend-build/CMakeFiles/3.31.6/CompilerIdCUDA/tmp/CMakeCUDACompilerId.module_id 2025-11-03T17:06:22.7256042Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/3.31.6/CompilerIdCUDA/tmp/CMakeCUDACompilerId.cpp1.ii 2025-11-03T17:06:22.7257035Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/3.31.6/CompilerIdCUDA/tmp/CMakeCUDACompilerId.cudafe1.c 2025-11-03T17:06:22.7258175Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/3.31.6/CompilerIdCUDA/tmp/CMakeCUDACompilerId.cudafe1.gpu 2025-11-03T17:06:22.7259183Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/3.31.6/CompilerIdCUDA/tmp/CMakeCUDACompilerId.cudafe1.stub.c 2025-11-03T17:06:22.7260158Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/3.31.6/CompilerIdCUDA/tmp/CMakeCUDACompilerId.ptx 2025-11-03T17:06:22.7261345Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/3.31.6/CompilerIdCUDA/tmp/CMakeCUDACompilerId.sm_52.cubin 2025-11-03T17:06:22.7262335Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/3.31.6/CompilerIdCUDA/tmp/CMakeCUDACompilerId.fatbin 2025-11-03T17:06:22.7263314Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/3.31.6/CompilerIdCUDA/tmp/CMakeCUDACompilerId.fatbin.c 2025-11-03T17:06:22.7264383Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/3.31.6/CompilerIdCUDA/tmp/CMakeCUDACompilerId.o 2025-11-03T17:06:22.7265270Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/3.31.6/CompilerIdCUDA/tmp/a_dlink.sm_52.cubin 2025-11-03T17:06:22.7266140Z extracting: build/custom_test_artifacts/custom-backend-build/CMakeFiles/3.31.6/CompilerIdCUDA/tmp/a_dlink.reg.c 2025-11-03T17:06:22.7267089Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/3.31.6/CompilerIdCUDA/tmp/a_dlink.fatbin 2025-11-03T17:06:22.7267948Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/3.31.6/CompilerIdCUDA/tmp/a_dlink.fatbin.c 2025-11-03T17:06:22.7268976Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/3.31.6/CompilerIdCUDA/tmp/a_dlink.o 2025-11-03T17:06:22.7271515Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/3.31.6/CompilerIdCUDA/CMakeCUDACompilerId.cu 2025-11-03T17:06:22.7336162Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/3.31.6/CompilerIdCUDA/a.out 2025-11-03T17:06:22.7337073Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/3.31.6/CMakeCUDACompiler.cmake 2025-11-03T17:06:22.7401829Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/3.31.6/CMakeDetermineCompilerABI_CUDA.bin 2025-11-03T17:06:22.7402617Z creating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/CMakeScratch/ 2025-11-03T17:06:22.7403245Z creating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/CMakeTmp/ 2025-11-03T17:06:22.7403886Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/cmake.check_cache 2025-11-03T17:06:22.7404553Z creating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/custom_backend.dir/ 2025-11-03T17:06:22.7405311Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/custom_backend.dir/compiler_depend.ts 2025-11-03T17:06:22.7406163Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/custom_backend.dir/compiler_depend.make 2025-11-03T17:06:22.7406984Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/custom_backend.dir/depend.make 2025-11-03T17:06:22.7407872Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/custom_backend.dir/link.txt 2025-11-03T17:06:22.7408845Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/custom_backend.dir/cmake_clean.cmake 2025-11-03T17:06:22.7409745Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/custom_backend.dir/build.make 2025-11-03T17:06:22.7410390Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/custom_backend.dir/DependInfo.cmake 2025-11-03T17:06:22.7411190Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/custom_backend.dir/flags.make 2025-11-03T17:06:22.7412154Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/custom_backend.dir/progress.make 2025-11-03T17:06:22.7416403Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/custom_backend.dir/custom_backend.cpp.o.d 2025-11-03T17:06:22.7521983Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/custom_backend.dir/custom_backend.cpp.o 2025-11-03T17:06:22.7522765Z creating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/test_custom_backend.dir/ 2025-11-03T17:06:22.7523737Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/test_custom_backend.dir/compiler_depend.ts 2025-11-03T17:06:22.7524619Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/test_custom_backend.dir/compiler_depend.make 2025-11-03T17:06:22.7525491Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/test_custom_backend.dir/depend.make 2025-11-03T17:06:22.7526280Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/test_custom_backend.dir/link.txt 2025-11-03T17:06:22.7527152Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/test_custom_backend.dir/cmake_clean.cmake 2025-11-03T17:06:22.7528118Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/test_custom_backend.dir/build.make 2025-11-03T17:06:22.7529041Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/test_custom_backend.dir/DependInfo.cmake 2025-11-03T17:06:22.7529749Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/test_custom_backend.dir/flags.make 2025-11-03T17:06:22.7530711Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/test_custom_backend.dir/progress.make 2025-11-03T17:06:22.7549838Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/test_custom_backend.dir/test_custom_backend.cpp.o.d 2025-11-03T17:06:22.7597053Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/test_custom_backend.dir/test_custom_backend.cpp.o 2025-11-03T17:06:22.7597929Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/CMakeDirectoryInformation.cmake 2025-11-03T17:06:22.7598696Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/TargetDirectories.txt 2025-11-03T17:06:22.7600035Z extracting: build/custom_test_artifacts/custom-backend-build/CMakeFiles/progress.marks 2025-11-03T17:06:22.7600981Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/Makefile2 2025-11-03T17:06:22.7602825Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/Makefile.cmake 2025-11-03T17:06:22.7603473Z inflating: build/custom_test_artifacts/custom-backend-build/detect_cuda_version.cc 2025-11-03T17:06:22.7606036Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeCache.txt 2025-11-03T17:06:22.7607017Z inflating: build/custom_test_artifacts/custom-backend-build/Makefile 2025-11-03T17:06:22.7607948Z inflating: build/custom_test_artifacts/custom-backend-build/cmake_install.cmake 2025-11-03T17:06:22.7695486Z inflating: build/custom_test_artifacts/custom-backend-build/libcustom_backend.so 2025-11-03T17:06:22.7729773Z inflating: build/custom_test_artifacts/custom-backend-build/test_custom_backend 2025-11-03T17:06:22.7730215Z creating: build/lib/ 2025-11-03T17:06:22.7801830Z inflating: build/lib/libprotobuf-lite.a 2025-11-03T17:06:22.7810994Z inflating: build/lib/libpthreadpool.a 2025-11-03T17:06:22.8195545Z inflating: build/lib/libprotobuf.a 2025-11-03T17:06:22.8203083Z inflating: build/lib/libcpuinfo.a 2025-11-03T17:06:22.8632850Z inflating: build/lib/libprotoc.a 2025-11-03T17:06:22.8640204Z inflating: build/lib/libcpuinfo_internals.a 2025-11-03T17:06:22.8641151Z inflating: build/lib/libclog.a 2025-11-03T17:06:22.8643737Z inflating: build/lib/libnnpack_reference_layers.a 2025-11-03T17:06:22.8660955Z inflating: build/lib/libpytorch_qnnpack.a 2025-11-03T17:06:22.8676700Z inflating: build/lib/libnnpack.a 2025-11-03T17:06:22.8838035Z inflating: build/lib/libmicrokernels-prod.a 2025-11-03T17:06:22.9598324Z inflating: build/lib/libmicrokernels-all.a 2025-11-03T17:06:22.9659587Z inflating: build/lib/libgtest.a 2025-11-03T17:06:22.9674887Z inflating: build/lib/libgmock.a 2025-11-03T17:06:22.9675742Z inflating: build/lib/libgtest_main.a 2025-11-03T17:06:22.9741947Z inflating: build/lib/libbenchmark.a 2025-11-03T17:06:22.9742770Z inflating: build/lib/libgmock_main.a 2025-11-03T17:06:22.9743780Z inflating: build/lib/libbenchmark_main.a 2025-11-03T17:06:22.9822429Z inflating: build/lib/libXNNPACK.a 2025-11-03T17:06:22.9823335Z inflating: build/lib/libjitprofiling.a 2025-11-03T17:06:22.9830786Z inflating: build/lib/libittnotify.a 2025-11-03T17:06:22.9887618Z inflating: build/lib/libasmjit.a 2025-11-03T17:06:23.0915346Z inflating: build/lib/libfbgemm.a 2025-11-03T17:06:23.0942112Z inflating: build/lib/libtensorpipe_uv.a 2025-11-03T17:06:23.1421048Z inflating: build/lib/libtensorpipe.a 2025-11-03T17:06:23.1635814Z inflating: build/lib/libtensorpipe_cuda.a 2025-11-03T17:06:23.1752504Z inflating: build/lib/libgloo.a 2025-11-03T17:06:23.1793647Z inflating: build/lib/libonnx_proto.a 2025-11-03T17:06:23.2415226Z inflating: build/lib/libonnx.a 2025-11-03T17:06:23.2798177Z inflating: build/lib/libgloo_cuda.a 2025-11-03T17:06:23.2815503Z inflating: build/lib/libfmt.a 2025-11-03T17:06:24.1655741Z inflating: build/lib/libdnnl.a 2025-11-03T17:06:24.2068910Z inflating: build/lib/libkineto.a 2025-11-03T17:06:24.2070490Z inflating: build/lib/libtorch_global_deps.so 2025-11-03T17:06:24.2174784Z inflating: build/lib/libc10.so 2025-11-03T17:06:24.2176603Z inflating: build/lib/libcaffe2_nvrtc.so 2025-11-03T17:06:24.2219523Z inflating: build/lib/libc10_cuda.so 2025-11-03T17:06:26.8967739Z inflating: build/lib/libtorch_cpu.so 2025-11-03T17:06:26.9654359Z inflating: build/lib/libtorch_nvshmem.so 2025-11-03T17:06:31.0199261Z inflating: build/lib/libtorch_cuda.so 2025-11-03T17:06:31.0203229Z inflating: build/lib/libshm.so 2025-11-03T17:06:31.0205944Z inflating: build/lib/libc10d_cuda_test.so 2025-11-03T17:06:31.0207417Z inflating: build/lib/libtorch.so 2025-11-03T17:06:31.0252781Z inflating: build/lib/libtorch_cuda_linalg.so 2025-11-03T17:06:31.0313978Z inflating: build/lib/libtorchbind_test.so 2025-11-03T17:06:31.0330232Z inflating: build/lib/libjitbackend_test.so 2025-11-03T17:06:31.0351259Z inflating: build/lib/libbackend_with_compiler.so 2025-11-03T17:06:31.0374247Z inflating: build/lib/libaoti_custom_ops.so 2025-11-03T17:06:31.2441528Z inflating: build/lib/libtorch_python.so 2025-11-03T17:06:31.2472339Z inflating: build/lib/libnnapi_backend.so 2025-11-03T17:06:31.2472682Z creating: build/bin/ 2025-11-03T17:06:31.2868776Z inflating: build/bin/protoc-3.13.0.0 2025-11-03T17:06:31.3262796Z inflating: build/bin/protoc 2025-11-03T17:06:31.3312808Z inflating: build/bin/c10_Device_test 2025-11-03T17:06:31.3360198Z inflating: build/bin/c10_StreamGuard_test 2025-11-03T17:06:31.3414936Z inflating: build/bin/c10_SymInt_test 2025-11-03T17:06:31.3467043Z inflating: build/bin/c10_InlineDeviceGuard_test 2025-11-03T17:06:31.3520501Z inflating: build/bin/c10_Metaprogramming_test 2025-11-03T17:06:31.3574103Z inflating: build/bin/c10_InlineStreamGuard_test 2025-11-03T17:06:31.3627415Z inflating: build/bin/c10_ThreadLocal_test 2025-11-03T17:06:31.3675293Z inflating: build/bin/c10_Semaphore_test 2025-11-03T17:06:31.3726903Z inflating: build/bin/c10_Bitset_test 2025-11-03T17:06:31.3778932Z inflating: build/bin/c10_Scalar_test 2025-11-03T17:06:31.3828150Z inflating: build/bin/c10_TypeList_test 2025-11-03T17:06:31.3884752Z inflating: build/bin/c10_DispatchKeySet_test 2025-11-03T17:06:31.3936267Z inflating: build/bin/c10_AllocatorConfig_test 2025-11-03T17:06:31.3984302Z inflating: build/bin/c10_DeadlockDetection_test 2025-11-03T17:06:31.4032519Z inflating: build/bin/c10_Synchronized_test 2025-11-03T17:06:31.4082554Z inflating: build/bin/c10_TypeIndex_test 2025-11-03T17:06:31.4136790Z inflating: build/bin/c10_LeftRight_test 2025-11-03T17:06:31.4190582Z inflating: build/bin/c10_SizesAndStrides_test 2025-11-03T17:06:31.4238152Z inflating: build/bin/c10_ConstexprCrc_test 2025-11-03T17:06:31.4288919Z inflating: build/bin/c10_IntrusiveList_test 2025-11-03T17:06:31.4338934Z inflating: build/bin/c10_accumulate_test 2025-11-03T17:06:31.4393435Z inflating: build/bin/c10_Enumerate_test 2025-11-03T17:06:31.4441903Z inflating: build/bin/c10_CompileTimeFunctionPointer_test 2025-11-03T17:06:31.4490850Z inflating: build/bin/c10_Half_test 2025-11-03T17:06:31.4542298Z inflating: build/bin/c10_NetworkFlow_test 2025-11-03T17:06:31.4596528Z inflating: build/bin/c10_complex_math_test 2025-11-03T17:06:31.4645073Z inflating: build/bin/c10_flags_test 2025-11-03T17:06:31.4692541Z inflating: build/bin/c10_TypeTraits_test 2025-11-03T17:06:31.4742996Z inflating: build/bin/c10_exception_test 2025-11-03T17:06:31.4794104Z inflating: build/bin/c10_registry_test 2025-11-03T17:06:31.4853405Z inflating: build/bin/c10_ordered_preserving_dict_test 2025-11-03T17:06:31.4907730Z inflating: build/bin/c10_logging_test 2025-11-03T17:06:31.4956588Z inflating: build/bin/c10_generic_math_test 2025-11-03T17:06:31.5005403Z inflating: build/bin/c10_tempfile_test 2025-11-03T17:06:31.5059993Z inflating: build/bin/c10_string_util_test 2025-11-03T17:06:31.5107353Z inflating: build/bin/c10_string_view_test 2025-11-03T17:06:31.5157253Z inflating: build/bin/c10_DeviceGuard_test 2025-11-03T17:06:31.5205005Z inflating: build/bin/c10_ArrayRef_test 2025-11-03T17:06:31.5253678Z inflating: build/bin/c10_bit_cast_test 2025-11-03T17:06:31.5320273Z inflating: build/bin/c10_cow_test 2025-11-03T17:06:31.5461463Z inflating: build/bin/c10_small_vector_test 2025-11-03T17:06:31.5510658Z inflating: build/bin/c10_irange_test 2025-11-03T17:06:31.5560249Z inflating: build/bin/c10_ssize_test 2025-11-03T17:06:31.5611692Z inflating: build/bin/c10_lazy_test 2025-11-03T17:06:31.5659427Z inflating: build/bin/c10_error_test 2025-11-03T17:06:31.5713144Z inflating: build/bin/c10_bfloat16_test 2025-11-03T17:06:31.5765749Z inflating: build/bin/c10_complex_test 2025-11-03T17:06:31.5917261Z inflating: build/bin/c10_intrusive_ptr_test 2025-11-03T17:06:31.5959235Z inflating: build/bin/c10_intrusive_ptr_benchmark 2025-11-03T17:06:31.6029931Z inflating: build/bin/c10_optional_test 2025-11-03T17:06:31.6083794Z inflating: build/bin/c10_typeid_test 2025-11-03T17:06:31.6131745Z inflating: build/bin/c10_cuda_CUDATest 2025-11-03T17:06:31.6179781Z inflating: build/bin/test_vec_half_AVX2 2025-11-03T17:06:31.6253752Z inflating: build/bin/test_aoti_abi_check 2025-11-03T17:06:31.6301787Z inflating: build/bin/test_vec_half_AVX512 2025-11-03T17:06:31.6349519Z inflating: build/bin/test_vec_half_DEFAULT 2025-11-03T17:06:31.6876372Z inflating: build/bin/vec_test_all_types_DEFAULT 2025-11-03T17:06:31.7417146Z inflating: build/bin/vec_test_all_types_AVX512 2025-11-03T17:06:31.7965899Z inflating: build/bin/vec_test_all_types_AVX2 2025-11-03T17:06:31.8016853Z inflating: build/bin/c10_cuda_CUDAAssertionsTest_catches_thread_and_block_and_device 2025-11-03T17:06:31.8067322Z inflating: build/bin/c10_cuda_CUDAAssertionsTest_multiple_writes_from_same_block 2025-11-03T17:06:31.8118422Z inflating: build/bin/c10_cuda_CUDAAssertionsTest_multiple_writes_from_multiple_blocks 2025-11-03T17:06:31.8168747Z inflating: build/bin/c10_cuda_CUDAAssertionsTest_from_2_processes 2025-11-03T17:06:31.8219696Z inflating: build/bin/c10_cuda_CUDAAssertionsTest_1_var_test 2025-11-03T17:06:31.8270344Z inflating: build/bin/c10_cuda_CUDAAssertionsTest_multiple_writes_from_blocks_and_threads 2025-11-03T17:06:31.8321487Z inflating: build/bin/c10_cuda_CUDAAssertionsTest_catches_stream 2025-11-03T17:06:31.8372024Z inflating: build/bin/BackoffTest 2025-11-03T17:06:31.8423487Z inflating: build/bin/FileStoreTest 2025-11-03T17:06:31.8477940Z inflating: build/bin/TCPStoreTest 2025-11-03T17:06:31.8529885Z inflating: build/bin/HashStoreTest 2025-11-03T17:06:31.8542364Z inflating: build/bin/ProcessGroupMPITest 2025-11-03T17:06:31.8545885Z inflating: build/bin/torch_shm_manager 2025-11-03T17:06:31.8548853Z inflating: build/bin/example_allreduce 2025-11-03T17:06:31.8603193Z inflating: build/bin/ProcessGroupGlooAsyncTest 2025-11-03T17:06:31.8664757Z inflating: build/bin/ProcessGroupNCCLTest 2025-11-03T17:06:31.8723682Z inflating: build/bin/ProcessGroupNCCLErrorsTest 2025-11-03T17:06:31.8786767Z inflating: build/bin/ProcessGroupGlooTest 2025-11-03T17:06:31.8847878Z inflating: build/bin/basic 2025-11-03T17:06:31.8900944Z inflating: build/bin/broadcast_test 2025-11-03T17:06:31.8955359Z inflating: build/bin/NamedTensor_test 2025-11-03T17:06:31.9005897Z inflating: build/bin/Dimname_test 2025-11-03T17:06:31.9057134Z inflating: build/bin/cpu_profiling_allocator_test 2025-11-03T17:06:31.9110422Z inflating: build/bin/half_test 2025-11-03T17:06:31.9166876Z inflating: build/bin/apply_utils_test 2025-11-03T17:06:31.9219620Z inflating: build/bin/math_kernel_test 2025-11-03T17:06:31.9309087Z inflating: build/bin/ivalue_test 2025-11-03T17:06:31.9363205Z inflating: build/bin/native_test 2025-11-03T17:06:31.9425278Z inflating: build/bin/MaybeOwned_test 2025-11-03T17:06:31.9489059Z inflating: build/bin/pow_test 2025-11-03T17:06:31.9558746Z inflating: build/bin/Dict_test 2025-11-03T17:06:31.9611962Z inflating: build/bin/memory_format_test 2025-11-03T17:06:31.9666606Z inflating: build/bin/scalar_tensor_test 2025-11-03T17:06:31.9715714Z inflating: build/bin/operators_test 2025-11-03T17:06:31.9772211Z inflating: build/bin/atest 2025-11-03T17:06:31.9821199Z inflating: build/bin/thread_init_test 2025-11-03T17:06:31.9907062Z inflating: build/bin/cpu_rng_test 2025-11-03T17:06:31.9956343Z inflating: build/bin/cpu_allocator_test 2025-11-03T17:06:32.0012367Z inflating: build/bin/cpu_generator_test 2025-11-03T17:06:32.0060469Z inflating: build/bin/verify_api_visibility 2025-11-03T17:06:32.0108450Z inflating: build/bin/reduce_ops_test 2025-11-03T17:06:32.0160536Z inflating: build/bin/memory_overlapping_test 2025-11-03T17:06:32.0236336Z inflating: build/bin/tensor_iterator_test 2025-11-03T17:06:32.0287303Z inflating: build/bin/stride_properties_test 2025-11-03T17:06:32.0337424Z inflating: build/bin/StorageUtils_test 2025-11-03T17:06:32.0392447Z inflating: build/bin/extension_backend_test 2025-11-03T17:06:32.0504166Z inflating: build/bin/kernel_function_legacy_test 2025-11-03T17:06:32.0552520Z inflating: build/bin/lazy_tensor_test 2025-11-03T17:06:32.0604039Z inflating: build/bin/mobile_memory_cleanup 2025-11-03T17:06:32.0659236Z inflating: build/bin/quantized_test 2025-11-03T17:06:32.0747284Z inflating: build/bin/make_boxed_from_unboxed_functor_test 2025-11-03T17:06:32.0797542Z inflating: build/bin/packedtensoraccessor_test 2025-11-03T17:06:32.0846530Z inflating: build/bin/operator_name_test 2025-11-03T17:06:32.0893866Z inflating: build/bin/op_allowlist_test 2025-11-03T17:06:32.0950066Z inflating: build/bin/scalar_test 2025-11-03T17:06:32.1018360Z inflating: build/bin/cuda_complex_math_test 2025-11-03T17:06:32.1074525Z inflating: build/bin/type_test 2025-11-03T17:06:32.1131234Z inflating: build/bin/IListRef_test 2025-11-03T17:06:32.1180519Z inflating: build/bin/reportMemoryUsage_test 2025-11-03T17:06:32.1230254Z inflating: build/bin/weakref_test 2025-11-03T17:06:32.1281014Z inflating: build/bin/cuda_apply_test 2025-11-03T17:06:32.1331790Z inflating: build/bin/undefined_tensor_test 2025-11-03T17:06:32.1394040Z inflating: build/bin/cuda_cub_test 2025-11-03T17:06:32.1443299Z inflating: build/bin/CppSignature_test 2025-11-03T17:06:32.1496095Z inflating: build/bin/type_ptr_test 2025-11-03T17:06:32.1584724Z inflating: build/bin/kernel_function_test 2025-11-03T17:06:32.1634778Z inflating: build/bin/cuda_dlconvertor_test 2025-11-03T17:06:32.1687175Z inflating: build/bin/test_parallel 2025-11-03T17:06:32.1735238Z inflating: build/bin/cuda_device_test 2025-11-03T17:06:32.1829551Z inflating: build/bin/kernel_lambda_test 2025-11-03T17:06:32.1878497Z inflating: build/bin/dlconvertor_test 2025-11-03T17:06:32.1941872Z inflating: build/bin/inline_container_test 2025-11-03T17:06:32.2000356Z inflating: build/bin/cuda_atomic_ops_test 2025-11-03T17:06:32.2051103Z inflating: build/bin/cuda_vectorized_test 2025-11-03T17:06:32.2102761Z inflating: build/bin/cuda_allocator_test 2025-11-03T17:06:32.2152763Z inflating: build/bin/xla_tensor_test 2025-11-03T17:06:32.2209793Z inflating: build/bin/cuda_complex_test 2025-11-03T17:06:32.2272768Z inflating: build/bin/KernelFunction_test 2025-11-03T17:06:32.2322296Z inflating: build/bin/wrapdim_test 2025-11-03T17:06:32.2377408Z inflating: build/bin/cuda_generator_test 2025-11-03T17:06:32.2428142Z inflating: build/bin/cuda_reportMemoryUsage_test 2025-11-03T17:06:32.2485486Z inflating: build/bin/kernel_stackbased_test 2025-11-03T17:06:32.2543888Z inflating: build/bin/cuda_stream_test 2025-11-03T17:06:32.2592203Z inflating: build/bin/cuda_half_test 2025-11-03T17:06:32.2640160Z inflating: build/bin/cuda_exchange_device_test 2025-11-03T17:06:32.2693251Z inflating: build/bin/backend_fallback_test 2025-11-03T17:06:32.2741246Z inflating: build/bin/cuda_cudnn_test 2025-11-03T17:06:32.2790905Z inflating: build/bin/cuda_integer_divider_test 2025-11-03T17:06:32.2870770Z inflating: build/bin/cuda_distributions_test 2025-11-03T17:06:32.2986177Z inflating: build/bin/kernel_lambda_legacy_test 2025-11-03T17:06:32.3034165Z inflating: build/bin/cuda_optional_test 2025-11-03T17:06:32.3088099Z inflating: build/bin/cuda_caching_host_allocator_test 2025-11-03T17:06:32.3367255Z inflating: build/bin/op_registration_test 2025-11-03T17:06:32.3465643Z inflating: build/bin/List_test 2025-11-03T17:06:32.3532012Z inflating: build/bin/legacy_vmap_test 2025-11-03T17:06:32.3580431Z inflating: build/bin/cuda_allocatorTraceTracker_test 2025-11-03T17:06:32.3644830Z inflating: build/bin/test_cpp_rpc 2025-11-03T17:06:32.3697544Z inflating: build/bin/test_dist_autograd 2025-11-03T17:06:32.3699976Z inflating: build/bin/parallel_benchmark 2025-11-03T17:06:32.4728899Z inflating: build/bin/test_api 2025-11-03T17:06:32.4779367Z inflating: build/bin/cuda_packedtensoraccessor_test 2025-11-03T17:06:32.5090733Z inflating: build/bin/test_lazy 2025-11-03T17:06:32.6068637Z inflating: build/bin/test_jit 2025-11-03T17:06:32.6081128Z creating: .additional_ci_files/ 2025-11-03T17:06:32.6119526Z inflating: .additional_ci_files/test-times.json 2025-11-03T17:06:32.6303594Z inflating: .additional_ci_files/test-class-times.json 2025-11-03T17:06:32.6357861Z ##[group]Run rm artifacts.zip 2025-11-03T17:06:32.6358119Z rm artifacts.zip 2025-11-03T17:06:32.6367063Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2025-11-03T17:06:32.6367361Z env: 2025-11-03T17:06:32.6367533Z GIT_DEFAULT_BRANCH: main 2025-11-03T17:06:32.6367793Z GPU_FLAG: --gpus all -e NVIDIA_DRIVER_CAPABILITIES=all 2025-11-03T17:06:32.6368060Z ##[endgroup] 2025-11-03T17:06:32.8355839Z ##[group]Run df -H 2025-11-03T17:06:32.8356132Z df -H 2025-11-03T17:06:32.8364300Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2025-11-03T17:06:32.8364594Z env: 2025-11-03T17:06:32.8364764Z GIT_DEFAULT_BRANCH: main 2025-11-03T17:06:32.8365013Z GPU_FLAG: --gpus all -e NVIDIA_DRIVER_CAPABILITIES=all 2025-11-03T17:06:32.8365263Z ##[endgroup] 2025-11-03T17:06:32.8426426Z Filesystem Size Used Avail Use% Mounted on 2025-11-03T17:06:32.8427180Z devtmpfs 4.2M 0 4.2M 0% /dev 2025-11-03T17:06:32.8427784Z tmpfs 98G 0 98G 0% /dev/shm 2025-11-03T17:06:32.8428369Z tmpfs 40G 984k 40G 1% /run 2025-11-03T17:06:32.8428920Z /dev/nvme0n1p1 161G 56G 106G 35% / 2025-11-03T17:06:32.8429564Z tmpfs 98G 13k 98G 1% /tmp 2025-11-03T17:06:32.8430203Z /dev/nvme0n1p128 11M 1.4M 9.2M 13% /boot/efi 2025-11-03T17:06:32.8430848Z tmpfs 20G 0 20G 0% /run/user/0 2025-11-03T17:06:32.8462808Z Prepare all required actions 2025-11-03T17:06:32.8463354Z Getting action download info 2025-11-03T17:06:33.0482361Z ##[group]Run ./.github/actions/download-td-artifacts 2025-11-03T17:06:33.0482621Z with: 2025-11-03T17:06:33.0482771Z env: 2025-11-03T17:06:33.0482919Z GIT_DEFAULT_BRANCH: main 2025-11-03T17:06:33.0483160Z GPU_FLAG: --gpus all -e NVIDIA_DRIVER_CAPABILITIES=all 2025-11-03T17:06:33.0483415Z ##[endgroup] 2025-11-03T17:06:33.0517941Z ##[group]Run seemethere/download-artifact-s3@v4 2025-11-03T17:06:33.0518184Z with: 2025-11-03T17:06:33.0518332Z name: td_results 2025-11-03T17:06:33.0518508Z s3-bucket: gha-artifacts 2025-11-03T17:06:33.0518700Z region: us-east-1 2025-11-03T17:06:33.0518868Z env: 2025-11-03T17:06:33.0519016Z GIT_DEFAULT_BRANCH: main 2025-11-03T17:06:33.0519253Z GPU_FLAG: --gpus all -e NVIDIA_DRIVER_CAPABILITIES=all 2025-11-03T17:06:33.0519504Z ##[endgroup] 2025-11-03T17:06:33.5015109Z (node:67485) NOTE: We are formalizing our plans to enter AWS SDK for JavaScript (v2) into maintenance mode in 2023. 2025-11-03T17:06:33.5016012Z 2025-11-03T17:06:33.5016420Z Please migrate your code to use AWS SDK for JavaScript (v3). 2025-11-03T17:06:33.5017498Z For more information, check the migration guide at https://a.co/7PzMCcy 2025-11-03T17:06:33.5018382Z (Use `node --trace-warnings ...` to show where the warning was created) 2025-11-03T17:06:33.5931069Z Found 0 objects with prefix pytorch/pytorch/19040285420/td_results/ 2025-11-03T17:06:33.5936754Z Artifact download has finished successfully 2025-11-03T17:06:33.6201164Z ##[group]Run mkdir -p .additional_ci_files 2025-11-03T17:06:33.6201490Z mkdir -p .additional_ci_files 2025-11-03T17:06:33.6201815Z mv td_results.json .additional_ci_files/td_results.json || true 2025-11-03T17:06:33.6212013Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2025-11-03T17:06:33.6212316Z env: 2025-11-03T17:06:33.6212500Z GIT_DEFAULT_BRANCH: main 2025-11-03T17:06:33.6212781Z GPU_FLAG: --gpus all -e NVIDIA_DRIVER_CAPABILITIES=all 2025-11-03T17:06:33.6213064Z ##[endgroup] 2025-11-03T17:06:33.6285455Z mv: cannot stat 'td_results.json': No such file or directory 2025-11-03T17:06:33.6341969Z ##[group]Run .github/scripts/parse_ref.py 2025-11-03T17:06:33.6342282Z .github/scripts/parse_ref.py 2025-11-03T17:06:33.6349476Z shell: /usr/bin/bash -e {0} 2025-11-03T17:06:33.6349686Z env: 2025-11-03T17:06:33.6349856Z GIT_DEFAULT_BRANCH: main 2025-11-03T17:06:33.6350110Z GPU_FLAG: --gpus all -e NVIDIA_DRIVER_CAPABILITIES=all 2025-11-03T17:06:33.6350377Z ##[endgroup] 2025-11-03T17:06:33.6560718Z Setting output branch=main 2025-11-03T17:06:33.6661713Z Prepare all required actions 2025-11-03T17:06:33.6662036Z Getting action download info 2025-11-03T17:06:33.8037761Z ##[group]Run ./.github/actions/filter-test-configs 2025-11-03T17:06:33.8038012Z with: 2025-11-03T17:06:33.8038353Z github-token: *** 2025-11-03T17:06:33.8043039Z test-matrix: {"include": [{"config": "vllm_basic_correctness_test", "shard": 1, "num_shards": 1, "runner": "linux.g6.4xlarge.experimental.nvidia.gpu"}, {"config": "vllm_basic_models_test", "shard": 1, "num_shards": 1, "runner": "linux.g6.4xlarge.experimental.nvidia.gpu"}, {"config": "vllm_entrypoints_test", "shard": 1, "num_shards": 1, "runner": "linux.g6.4xlarge.experimental.nvidia.gpu"}, {"config": "vllm_regression_test", "shard": 1, "num_shards": 1, "runner": "linux.g6.4xlarge.experimental.nvidia.gpu"}, {"config": "vllm_multi_model_processor_test", "shard": 1, "num_shards": 1, "runner": "linux.g6.4xlarge.experimental.nvidia.gpu"}, {"config": "vllm_pytorch_compilation_unit_tests", "shard": 1, "num_shards": 1, "runner": "linux.g6.4xlarge.experimental.nvidia.gpu"}, {"config": "vllm_lora_28_failure_test", "shard": 1, "num_shards": 1, "runner": "linux.g6.4xlarge.experimental.nvidia.gpu"}, {"config": "vllm_multi_model_test_28_failure_test", "shard": 1, "num_shards": 1, "runner": "linux.g6.4xlarge.experimental.nvidia.gpu"}, {"config": "vllm_language_model_test_extended_generation_28_failure_test", "shard": 1, "num_shards": 1, "runner": "linux.g6.4xlarge.experimental.nvidia.gpu"}, {"config": "vllm_distributed_test_2_gpu_28_failure_test", "shard": 1, "num_shards": 1, "runner": "linux.g6.4xlarge.experimental.nvidia.gpu"}, {"config": "vllm_lora_test", "shard": 0, "num_shards": 4, "runner": "linux.g6.4xlarge.experimental.nvidia.gpu"}, {"config": "vllm_lora_test", "shard": 1, "num_shards": 4, "runner": "linux.g6.4xlarge.experimental.nvidia.gpu"}, {"config": "vllm_lora_test", "shard": 2, "num_shards": 4, "runner": "linux.g6.4xlarge.experimental.nvidia.gpu"}, {"config": "vllm_lora_test", "shard": 3, "num_shards": 4, "runner": "linux.g6.4xlarge.experimental.nvidia.gpu"}, {"config": "vllm_lora_tp_test_distributed", "shard": 1, "num_shards": 1, "runner": "linux.g6.12xlarge.nvidia.gpu"}, {"config": "vllm_distributed_test_28_failure_test", "shard": 1, "num_shards": 1, "runner": "linux.g6.12xlarge.nvidia.gpu"}]} 2025-11-03T17:06:33.8048325Z job-name: ci-vllm-test / test (vllm_distributed_test_28_failure_test, 1, 1, linux.g6.12xlarge.nvidia.gpu) 2025-11-03T17:06:33.8048791Z env: 2025-11-03T17:06:33.8048951Z GIT_DEFAULT_BRANCH: main 2025-11-03T17:06:33.8049191Z GPU_FLAG: --gpus all -e NVIDIA_DRIVER_CAPABILITIES=all 2025-11-03T17:06:33.8049444Z ##[endgroup] 2025-11-03T17:06:33.8107515Z ##[group]Run nick-fields/retry@v3.0.0 2025-11-03T17:06:33.8107748Z with: 2025-11-03T17:06:33.8107904Z shell: bash 2025-11-03T17:06:33.8108065Z timeout_minutes: 10 2025-11-03T17:06:33.8108263Z max_attempts: 5 2025-11-03T17:06:33.8108444Z retry_wait_seconds: 30 2025-11-03T17:06:33.8109022Z command: set -eux # PyYAML 6.0 doesn't work with MacOS x86 anymore # This must run on Python-3.7 (AmazonLinux2) so can't use request=3.32.2 python3 -m pip install requests==2.27.1 pyyaml==6.0.2 2025-11-03T17:06:33.8109802Z polling_interval_seconds: 1 2025-11-03T17:06:33.8110014Z warning_on_retry: true 2025-11-03T17:06:33.8110203Z continue_on_error: false 2025-11-03T17:06:33.8110386Z env: 2025-11-03T17:06:33.8110537Z GIT_DEFAULT_BRANCH: main 2025-11-03T17:06:33.8110781Z GPU_FLAG: --gpus all -e NVIDIA_DRIVER_CAPABILITIES=all 2025-11-03T17:06:33.8111163Z GITHUB_TOKEN: *** 2025-11-03T17:06:33.8111343Z ##[endgroup] 2025-11-03T17:06:33.9185284Z + python3 -m pip install requests==2.27.1 pyyaml==6.0.2 2025-11-03T17:06:34.1326567Z Defaulting to user installation because normal site-packages is not writeable 2025-11-03T17:06:34.2464968Z Collecting requests==2.27.1 2025-11-03T17:06:34.2633921Z Downloading requests-2.27.1-py2.py3-none-any.whl (63 kB) 2025-11-03T17:06:34.4226967Z Collecting pyyaml==6.0.2 2025-11-03T17:06:34.4267918Z Downloading PyYAML-6.0.2-cp39-cp39-manylinux_2_17_x86_64.manylinux2014_x86_64.whl (737 kB) 2025-11-03T17:06:34.4507332Z Requirement already satisfied: idna<4,>=2.5 in /usr/lib/python3.9/site-packages (from requests==2.27.1) (2.10) 2025-11-03T17:06:34.8017624Z Collecting charset-normalizer~=2.0.0 2025-11-03T17:06:34.8065460Z Downloading charset_normalizer-2.0.12-py3-none-any.whl (39 kB) 2025-11-03T17:06:34.8544078Z Collecting certifi>=2017.4.17 2025-11-03T17:06:34.8587133Z Downloading certifi-2025.10.5-py3-none-any.whl (163 kB) 2025-11-03T17:06:34.8655217Z Requirement already satisfied: urllib3<1.27,>=1.21.1 in /usr/lib/python3.9/site-packages (from requests==2.27.1) (1.25.10) 2025-11-03T17:06:34.9370570Z Installing collected packages: charset-normalizer, certifi, requests, pyyaml 2025-11-03T17:06:35.0536921Z Successfully installed certifi-2025.10.5 charset-normalizer-2.0.12 pyyaml-6.0.2 requests-2.27.1 2025-11-03T17:06:35.8885605Z Command completed after 1 attempt(s). 2025-11-03T17:06:35.8964510Z ##[group]Run set -x 2025-11-03T17:06:35.8964815Z set -x 2025-11-03T17:06:35.8964983Z  2025-11-03T17:06:35.8965262Z # Use relative path here as this could be checked out anywhere, not necessarily 2025-11-03T17:06:35.8965616Z # in runner workspace 2025-11-03T17:06:35.8965896Z python3 "${GITHUB_ACTION_PATH}/../../scripts/parse_ref.py" 2025-11-03T17:06:35.8975306Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2025-11-03T17:06:35.8975593Z env: 2025-11-03T17:06:35.8975750Z GIT_DEFAULT_BRANCH: main 2025-11-03T17:06:35.8975998Z GPU_FLAG: --gpus all -e NVIDIA_DRIVER_CAPABILITIES=all 2025-11-03T17:06:35.8976252Z ##[endgroup] 2025-11-03T17:06:35.9012164Z + python3 /home/ec2-user/actions-runner/_work/pytorch/pytorch/./.github/actions/filter-test-configs/../../scripts/parse_ref.py 2025-11-03T17:06:35.9195254Z Setting output branch=main 2025-11-03T17:06:35.9256568Z ##[group]Run echo "Workflow: ${GITHUB_WORKFLOW}" 2025-11-03T17:06:35.9256923Z echo "Workflow: ${GITHUB_WORKFLOW}" 2025-11-03T17:06:35.9257167Z echo "Job name: ${JOB_NAME}" 2025-11-03T17:06:35.9257377Z  2025-11-03T17:06:35.9257647Z # Use relative path here as this could be checked out anywhere, not necessarily 2025-11-03T17:06:35.9257981Z # in runner workspace 2025-11-03T17:06:35.9258289Z python3 "${GITHUB_ACTION_PATH}/../../scripts/filter_test_configs.py" \ 2025-11-03T17:06:35.9258661Z  --workflow "${GITHUB_WORKFLOW}" \ 2025-11-03T17:06:35.9258900Z  --job-name "${JOB_NAME}" \ 2025-11-03T17:06:35.9263797Z  --test-matrix "{"include": [{"config": "vllm_basic_correctness_test", "shard": 1, "num_shards": 1, "runner": "linux.g6.4xlarge.experimental.nvidia.gpu"}, {"config": "vllm_basic_models_test", "shard": 1, "num_shards": 1, "runner": "linux.g6.4xlarge.experimental.nvidia.gpu"}, {"config": "vllm_entrypoints_test", "shard": 1, "num_shards": 1, "runner": "linux.g6.4xlarge.experimental.nvidia.gpu"}, {"config": "vllm_regression_test", "shard": 1, "num_shards": 1, "runner": "linux.g6.4xlarge.experimental.nvidia.gpu"}, {"config": "vllm_multi_model_processor_test", "shard": 1, "num_shards": 1, "runner": "linux.g6.4xlarge.experimental.nvidia.gpu"}, {"config": "vllm_pytorch_compilation_unit_tests", "shard": 1, "num_shards": 1, "runner": "linux.g6.4xlarge.experimental.nvidia.gpu"}, {"config": "vllm_lora_28_failure_test", "shard": 1, "num_shards": 1, "runner": "linux.g6.4xlarge.experimental.nvidia.gpu"}, {"config": "vllm_multi_model_test_28_failure_test", "shard": 1, "num_shards": 1, "runner": "linux.g6.4xlarge.experimental.nvidia.gpu"}, {"config": "vllm_language_model_test_extended_generation_28_failure_test", "shard": 1, "num_shards": 1, "runner": "linux.g6.4xlarge.experimental.nvidia.gpu"}, {"config": "vllm_distributed_test_2_gpu_28_failure_test", "shard": 1, "num_shards": 1, "runner": "linux.g6.4xlarge.experimental.nvidia.gpu"}, {"config": "vllm_lora_test", "shard": 0, "num_shards": 4, "runner": "linux.g6.4xlarge.experimental.nvidia.gpu"}, {"config": "vllm_lora_test", "shard": 1, "num_shards": 4, "runner": "linux.g6.4xlarge.experimental.nvidia.gpu"}, {"config": "vllm_lora_test", "shard": 2, "num_shards": 4, "runner": "linux.g6.4xlarge.experimental.nvidia.gpu"}, {"config": "vllm_lora_test", "shard": 3, "num_shards": 4, "runner": "linux.g6.4xlarge.experimental.nvidia.gpu"}, {"config": "vllm_lora_tp_test_distributed", "shard": 1, "num_shards": 1, "runner": "linux.g6.12xlarge.nvidia.gpu"}, {"config": "vllm_distributed_test_28_failure_test", "shard": 1, "num_shards": 1, "runner": "linux.g6.12xlarge.nvidia.gpu"}]}" \ 2025-11-03T17:06:35.9268931Z  --selected-test-configs "" \ 2025-11-03T17:06:35.9269183Z  --pr-number "${PR_NUMBER}" \ 2025-11-03T17:06:35.9269405Z  --tag "${TAG}" \ 2025-11-03T17:06:35.9269620Z  --event-name "${EVENT_NAME}" \ 2025-11-03T17:06:35.9269842Z  --schedule "${SCHEDULE}" \ 2025-11-03T17:06:35.9270064Z  --branch "${HEAD_BRANCH}" 2025-11-03T17:06:35.9278324Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2025-11-03T17:06:35.9278676Z env: 2025-11-03T17:06:35.9278864Z GIT_DEFAULT_BRANCH: main 2025-11-03T17:06:35.9279147Z GPU_FLAG: --gpus all -e NVIDIA_DRIVER_CAPABILITIES=all 2025-11-03T17:06:35.9279805Z GITHUB_TOKEN: *** 2025-11-03T17:06:35.9280229Z JOB_NAME: ci-vllm-test / test (vllm_distributed_test_28_failure_test, 1, 1, linux.g6.12xlarge.nvidia.gpu) 2025-11-03T17:06:35.9280809Z PR_NUMBER: 2025-11-03T17:06:35.9280977Z TAG: 2025-11-03T17:06:35.9281131Z EVENT_NAME: push 2025-11-03T17:06:35.9281300Z SCHEDULE: 2025-11-03T17:06:35.9281454Z HEAD_BRANCH: main 2025-11-03T17:06:35.9281612Z ##[endgroup] 2025-11-03T17:06:35.9313246Z Workflow: vllm-test 2025-11-03T17:06:35.9314205Z Job name: ci-vllm-test / test (vllm_distributed_test_28_failure_test, 1, 1, linux.g6.12xlarge.nvidia.gpu) 2025-11-03T17:06:36.1252363Z Setting output keep-going=True 2025-11-03T17:06:36.1253143Z Setting output ci-verbose-test-logs=False 2025-11-03T17:06:36.1253909Z Setting output ci-test-showlocals=False 2025-11-03T17:06:36.1254627Z Setting output ci-no-test-timeout=False 2025-11-03T17:06:36.1255298Z Setting output ci-no-td=False 2025-11-03T17:06:36.1255911Z Setting output ci-td-distributed=False 2025-11-03T17:06:36.1256561Z Setting output is-unstable=False 2025-11-03T17:06:36.1257159Z Setting output reenabled-issues= 2025-11-03T17:06:36.1266154Z Setting output test-matrix={"include": [{"config": "vllm_basic_correctness_test", "shard": 1, "num_shards": 1, "runner": "linux.g6.4xlarge.experimental.nvidia.gpu"}, {"config": "vllm_basic_models_test", "shard": 1, "num_shards": 1, "runner": "linux.g6.4xlarge.experimental.nvidia.gpu"}, {"config": "vllm_entrypoints_test", "shard": 1, "num_shards": 1, "runner": "linux.g6.4xlarge.experimental.nvidia.gpu"}, {"config": "vllm_regression_test", "shard": 1, "num_shards": 1, "runner": "linux.g6.4xlarge.experimental.nvidia.gpu"}, {"config": "vllm_multi_model_processor_test", "shard": 1, "num_shards": 1, "runner": "linux.g6.4xlarge.experimental.nvidia.gpu"}, {"config": "vllm_pytorch_compilation_unit_tests", "shard": 1, "num_shards": 1, "runner": "linux.g6.4xlarge.experimental.nvidia.gpu"}, {"config": "vllm_lora_28_failure_test", "shard": 1, "num_shards": 1, "runner": "linux.g6.4xlarge.experimental.nvidia.gpu"}, {"config": "vllm_multi_model_test_28_failure_test", "shard": 1, "num_shards": 1, "runner": "linux.g6.4xlarge.experimental.nvidia.gpu"}, {"config": "vllm_language_model_test_extended_generation_28_failure_test", "shard": 1, "num_shards": 1, "runner": "linux.g6.4xlarge.experimental.nvidia.gpu"}, {"config": "vllm_distributed_test_2_gpu_28_failure_test", "shard": 1, "num_shards": 1, "runner": "linux.g6.4xlarge.experimental.nvidia.gpu"}, {"config": "vllm_lora_test", "shard": 0, "num_shards": 4, "runner": "linux.g6.4xlarge.experimental.nvidia.gpu"}, {"config": "vllm_lora_test", "shard": 1, "num_shards": 4, "runner": "linux.g6.4xlarge.experimental.nvidia.gpu"}, {"config": "vllm_lora_test", "shard": 2, "num_shards": 4, "runner": "linux.g6.4xlarge.experimental.nvidia.gpu"}, {"config": "vllm_lora_test", "shard": 3, "num_shards": 4, "runner": "linux.g6.4xlarge.experimental.nvidia.gpu"}, {"config": "vllm_lora_tp_test_distributed", "shard": 1, "num_shards": 1, "runner": "linux.g6.12xlarge.nvidia.gpu"}, {"config": "vllm_distributed_test_28_failure_test", "shard": 1, "num_shards": 1, "runner": "linux.g6.12xlarge.nvidia.gpu"}]} 2025-11-03T17:06:36.1273737Z Setting output is-test-matrix-empty=False 2025-11-03T17:06:36.1377208Z ##[group]Run echo "Filtered matrix:" 2025-11-03T17:06:36.1377492Z echo "Filtered matrix:" 2025-11-03T17:06:36.1382235Z echo "{"include": [{"config": "vllm_basic_correctness_test", "shard": 1, "num_shards": 1, "runner": "linux.g6.4xlarge.experimental.nvidia.gpu"}, {"config": "vllm_basic_models_test", "shard": 1, "num_shards": 1, "runner": "linux.g6.4xlarge.experimental.nvidia.gpu"}, {"config": "vllm_entrypoints_test", "shard": 1, "num_shards": 1, "runner": "linux.g6.4xlarge.experimental.nvidia.gpu"}, {"config": "vllm_regression_test", "shard": 1, "num_shards": 1, "runner": "linux.g6.4xlarge.experimental.nvidia.gpu"}, {"config": "vllm_multi_model_processor_test", "shard": 1, "num_shards": 1, "runner": "linux.g6.4xlarge.experimental.nvidia.gpu"}, {"config": "vllm_pytorch_compilation_unit_tests", "shard": 1, "num_shards": 1, "runner": "linux.g6.4xlarge.experimental.nvidia.gpu"}, {"config": "vllm_lora_28_failure_test", "shard": 1, "num_shards": 1, "runner": "linux.g6.4xlarge.experimental.nvidia.gpu"}, {"config": "vllm_multi_model_test_28_failure_test", "shard": 1, "num_shards": 1, "runner": "linux.g6.4xlarge.experimental.nvidia.gpu"}, {"config": "vllm_language_model_test_extended_generation_28_failure_test", "shard": 1, "num_shards": 1, "runner": "linux.g6.4xlarge.experimental.nvidia.gpu"}, {"config": "vllm_distributed_test_2_gpu_28_failure_test", "shard": 1, "num_shards": 1, "runner": "linux.g6.4xlarge.experimental.nvidia.gpu"}, {"config": "vllm_lora_test", "shard": 0, "num_shards": 4, "runner": "linux.g6.4xlarge.experimental.nvidia.gpu"}, {"config": "vllm_lora_test", "shard": 1, "num_shards": 4, "runner": "linux.g6.4xlarge.experimental.nvidia.gpu"}, {"config": "vllm_lora_test", "shard": 2, "num_shards": 4, "runner": "linux.g6.4xlarge.experimental.nvidia.gpu"}, {"config": "vllm_lora_test", "shard": 3, "num_shards": 4, "runner": "linux.g6.4xlarge.experimental.nvidia.gpu"}, {"config": "vllm_lora_tp_test_distributed", "shard": 1, "num_shards": 1, "runner": "linux.g6.12xlarge.nvidia.gpu"}, {"config": "vllm_distributed_test_28_failure_test", "shard": 1, "num_shards": 1, "runner": "linux.g6.12xlarge.nvidia.gpu"}]}" 2025-11-03T17:06:36.1387292Z  2025-11-03T17:06:36.1387444Z echo 2025-11-03T17:06:36.1387646Z echo "Is the current job unstable? False" 2025-11-03T17:06:36.1387881Z  2025-11-03T17:06:36.1388028Z echo 2025-11-03T17:06:36.1388213Z echo "Is keep-going label set? True" 2025-11-03T17:06:36.1388492Z  2025-11-03T17:06:36.1388660Z echo 2025-11-03T17:06:36.1388874Z echo "Reenabled issues? " 2025-11-03T17:06:36.1396054Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2025-11-03T17:06:36.1396338Z env: 2025-11-03T17:06:36.1396504Z GIT_DEFAULT_BRANCH: main 2025-11-03T17:06:36.1396763Z GPU_FLAG: --gpus all -e NVIDIA_DRIVER_CAPABILITIES=all 2025-11-03T17:06:36.1397025Z ##[endgroup] 2025-11-03T17:06:36.1428916Z Filtered matrix: 2025-11-03T17:06:36.1439191Z {include: [{config: vllm_basic_correctness_test, shard: 1, num_shards: 1, runner: linux.g6.4xlarge.experimental.nvidia.gpu}, {config: vllm_basic_models_test, shard: 1, num_shards: 1, runner: linux.g6.4xlarge.experimental.nvidia.gpu}, {config: vllm_entrypoints_test, shard: 1, num_shards: 1, runner: linux.g6.4xlarge.experimental.nvidia.gpu}, {config: vllm_regression_test, shard: 1, num_shards: 1, runner: linux.g6.4xlarge.experimental.nvidia.gpu}, {config: vllm_multi_model_processor_test, shard: 1, num_shards: 1, runner: linux.g6.4xlarge.experimental.nvidia.gpu}, {config: vllm_pytorch_compilation_unit_tests, shard: 1, num_shards: 1, runner: linux.g6.4xlarge.experimental.nvidia.gpu}, {config: vllm_lora_28_failure_test, shard: 1, num_shards: 1, runner: linux.g6.4xlarge.experimental.nvidia.gpu}, {config: vllm_multi_model_test_28_failure_test, shard: 1, num_shards: 1, runner: linux.g6.4xlarge.experimental.nvidia.gpu}, {config: vllm_language_model_test_extended_generation_28_failure_test, shard: 1, num_shards: 1, runner: linux.g6.4xlarge.experimental.nvidia.gpu}, {config: vllm_distributed_test_2_gpu_28_failure_test, shard: 1, num_shards: 1, runner: linux.g6.4xlarge.experimental.nvidia.gpu}, {config: vllm_lora_test, shard: 0, num_shards: 4, runner: linux.g6.4xlarge.experimental.nvidia.gpu}, {config: vllm_lora_test, shard: 1, num_shards: 4, runner: linux.g6.4xlarge.experimental.nvidia.gpu}, {config: vllm_lora_test, shard: 2, num_shards: 4, runner: linux.g6.4xlarge.experimental.nvidia.gpu}, {config: vllm_lora_test, shard: 3, num_shards: 4, runner: linux.g6.4xlarge.experimental.nvidia.gpu}, {config: vllm_lora_tp_test_distributed, shard: 1, num_shards: 1, runner: linux.g6.12xlarge.nvidia.gpu}, {config: vllm_distributed_test_28_failure_test, shard: 1, num_shards: 1, runner: linux.g6.12xlarge.nvidia.gpu}]} 2025-11-03T17:06:36.1446426Z 2025-11-03T17:06:36.1446534Z Is the current job unstable? False 2025-11-03T17:06:36.1446724Z 2025-11-03T17:06:36.1446803Z Is keep-going label set? True 2025-11-03T17:06:36.1447134Z 2025-11-03T17:06:36.1447199Z Reenabled issues? 2025-11-03T17:06:36.1496463Z ##[group]Run echo "timeout=$((JOB_TIMEOUT-30))" >> "${GITHUB_OUTPUT}" 2025-11-03T17:06:36.1496901Z echo "timeout=$((JOB_TIMEOUT-30))" >> "${GITHUB_OUTPUT}" 2025-11-03T17:06:36.1504266Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2025-11-03T17:06:36.1504543Z env: 2025-11-03T17:06:36.1504704Z GIT_DEFAULT_BRANCH: main 2025-11-03T17:06:36.1504943Z GPU_FLAG: --gpus all -e NVIDIA_DRIVER_CAPABILITIES=all 2025-11-03T17:06:36.1505189Z JOB_TIMEOUT: 240 2025-11-03T17:06:36.1505354Z ##[endgroup] 2025-11-03T17:06:36.1581288Z ##[group]Run env | grep '^GITHUB' >> "/tmp/github_env_${GITHUB_RUN_ID}" 2025-11-03T17:06:36.1581716Z env | grep '^GITHUB' >> "/tmp/github_env_${GITHUB_RUN_ID}" 2025-11-03T17:06:36.1582033Z env | grep '^CI' >> "/tmp/github_env_${GITHUB_RUN_ID}" 2025-11-03T17:06:36.1589023Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2025-11-03T17:06:36.1589308Z env: 2025-11-03T17:06:36.1589478Z GIT_DEFAULT_BRANCH: main 2025-11-03T17:06:36.1589715Z GPU_FLAG: --gpus all -e NVIDIA_DRIVER_CAPABILITIES=all 2025-11-03T17:06:36.1589963Z ##[endgroup] 2025-11-03T17:06:36.1710084Z ##[group]Run set -x 2025-11-03T17:06:36.1710383Z set -x 2025-11-03T17:06:36.1710554Z  2025-11-03T17:06:36.1710747Z if [[ $TEST_CONFIG == 'multigpu' ]]; then 2025-11-03T17:06:36.1711030Z  TEST_COMMAND=.ci/pytorch/multigpu-test.sh 2025-11-03T17:06:36.1711319Z elif [[ $BUILD_ENVIRONMENT == *onnx* ]]; then 2025-11-03T17:06:36.1711576Z  TEST_COMMAND=.ci/onnx/test.sh 2025-11-03T17:06:36.1711779Z else 2025-11-03T17:06:36.1711974Z  TEST_COMMAND=.ci/pytorch/test.sh 2025-11-03T17:06:36.1712196Z fi 2025-11-03T17:06:36.1712338Z  2025-11-03T17:06:36.1712517Z # Leaving 1GB for the runner and other things 2025-11-03T17:06:36.1712934Z TOTAL_AVAILABLE_MEMORY_IN_GB=$(awk '/MemTotal/ { printf "%.3f \n", $2/1024/1024 - 1 }' /proc/meminfo) 2025-11-03T17:06:36.1713557Z # https://docs.docker.com/engine/containers/resource_constraints/#--memory-swap-details, the 3GB swap 2025-11-03T17:06:36.1714055Z # comes from https://github.com/pytorch/test-infra/pull/6058 2025-11-03T17:06:36.1714431Z TOTAL_MEMORY_WITH_SWAP=$(("${TOTAL_AVAILABLE_MEMORY_IN_GB%.*}" + 3)) 2025-11-03T17:06:36.1714719Z  2025-11-03T17:06:36.1714910Z if [[ ${BUILD_ENVIRONMENT} == *"s390x"* ]]; then 2025-11-03T17:06:36.1715155Z  SHM_OPTS= 2025-11-03T17:06:36.1715326Z  JENKINS_USER= 2025-11-03T17:06:36.1715581Z  # ensure that docker container cleanly exits in 12 hours 2025-11-03T17:06:36.1715919Z  # if for some reason cleanup action doesn't stop container 2025-11-03T17:06:36.1716195Z  # when job is cancelled 2025-11-03T17:06:36.1716413Z  DOCKER_SHELL_CMD="sleep 12h" 2025-11-03T17:06:36.1716622Z else 2025-11-03T17:06:36.1716804Z  SHM_OPTS="--shm-size=${SHM_SIZE}" 2025-11-03T17:06:36.1717040Z  JENKINS_USER="--user jenkins" 2025-11-03T17:06:36.1717262Z  DOCKER_SHELL_CMD= 2025-11-03T17:06:36.1717438Z fi 2025-11-03T17:06:36.1717583Z  2025-11-03T17:06:36.1717816Z # detached container should get cleaned up by teardown_ec2_linux 2025-11-03T17:06:36.1718185Z # TODO: Stop building test binaries as part of the build phase 2025-11-03T17:06:36.1718608Z # Used for GPU_FLAG, SHM_OPTS, JENKINS_USER and DOCKER_SHELL_CMD since that doesn't play nice 2025-11-03T17:06:36.1718991Z # shellcheck disable=SC2086,SC2090 2025-11-03T17:06:36.1719229Z container_name=$(docker run \ 2025-11-03T17:06:36.1719455Z  ${GPU_FLAG:-} \ 2025-11-03T17:06:36.1719677Z  ${SCCACHE_SERVER_PORT_DOCKER_FLAG:-} \ 2025-11-03T17:06:36.1719925Z  -e BUILD_ENVIRONMENT \ 2025-11-03T17:06:36.1720306Z  -e PR_NUMBER \ 2025-11-03T17:06:36.1720522Z  -e GITHUB_ACTIONS \ 2025-11-03T17:06:36.1720733Z  -e GITHUB_REPOSITORY \ 2025-11-03T17:06:36.1720936Z  -e GITHUB_WORKFLOW \ 2025-11-03T17:06:36.1721141Z  -e GITHUB_JOB \ 2025-11-03T17:06:36.1721332Z  -e GITHUB_RUN_ID \ 2025-11-03T17:06:36.1721528Z  -e GITHUB_RUN_NUMBER \ 2025-11-03T17:06:36.1721725Z  -e GITHUB_RUN_ATTEMPT \ 2025-11-03T17:06:36.1721926Z  -e JOB_ID \ 2025-11-03T17:06:36.1722102Z  -e JOB_NAME \ 2025-11-03T17:06:36.1722282Z  -e BASE_SHA \ 2025-11-03T17:06:36.1722448Z  -e BRANCH \ 2025-11-03T17:06:36.1722616Z  -e SHA1 \ 2025-11-03T17:06:36.1722808Z  -e AWS_DEFAULT_REGION \ 2025-11-03T17:06:36.1723013Z  -e IN_WHEEL_TEST \ 2025-11-03T17:06:36.1723194Z  -e SHARD_NUMBER \ 2025-11-03T17:06:36.1723386Z  -e TEST_CONFIG \ 2025-11-03T17:06:36.1723583Z  -e NUM_TEST_SHARDS \ 2025-11-03T17:06:36.1723782Z  -e REENABLED_ISSUES \ 2025-11-03T17:06:36.1723985Z  -e CONTINUE_THROUGH_ERROR \ 2025-11-03T17:06:36.1724338Z  -e VERBOSE_TEST_LOGS \ 2025-11-03T17:06:36.1724555Z  -e TEST_SHOWLOCALS \ 2025-11-03T17:06:36.1724753Z  -e NO_TEST_TIMEOUT \ 2025-11-03T17:06:36.1724933Z  -e NO_TD \ 2025-11-03T17:06:36.1725109Z  -e TD_DISTRIBUTED \ 2025-11-03T17:06:36.1725305Z  -e PR_LABELS \ 2025-11-03T17:06:36.1725516Z  -e MAX_JOBS="$(nproc --ignore=2)" \ 2025-11-03T17:06:36.1725745Z  -e SCCACHE_BUCKET \ 2025-11-03T17:06:36.1725941Z  -e SCCACHE_REGION \ 2025-11-03T17:06:36.1726136Z  -e XLA_CUDA \ 2025-11-03T17:06:36.1726340Z  -e XLA_CLANG_CACHE_S3_BUCKET_NAME \ 2025-11-03T17:06:36.1726590Z  -e PYTORCH_TEST_CUDA_MEM_LEAK_CHECK \ 2025-11-03T17:06:36.1726841Z  -e PYTORCH_TEST_RERUN_DISABLED_TESTS \ 2025-11-03T17:06:36.1727118Z  -e SKIP_SCCACHE_INITIALIZATION=1 \ 2025-11-03T17:06:36.1727356Z  -e HUGGING_FACE_HUB_TOKEN \ 2025-11-03T17:06:36.1727589Z  -e VLLM_TEST_HUGGING_FACE_TOKEN \ 2025-11-03T17:06:36.1727822Z  -e SCRIBE_GRAPHQL_ACCESS_TOKEN \ 2025-11-03T17:06:36.1728048Z  -e DASHBOARD_TAG \ 2025-11-03T17:06:36.1728259Z  -e ARTIFACTS_FILE_SUFFIX \ 2025-11-03T17:06:36.1728583Z  --memory="${TOTAL_AVAILABLE_MEMORY_IN_GB%.*}g" \ 2025-11-03T17:06:36.1728885Z  --memory-swap="${TOTAL_MEMORY_WITH_SWAP}g" \ 2025-11-03T17:06:36.1729178Z  --env-file="/tmp/github_env_${GITHUB_RUN_ID}" \ 2025-11-03T17:06:36.1729461Z  --security-opt seccomp=unconfined \ 2025-11-03T17:06:36.1729707Z  --cap-add=SYS_PTRACE \ 2025-11-03T17:06:36.1729918Z  --ipc=host \ 2025-11-03T17:06:36.1730096Z  ${SHM_OPTS} \ 2025-11-03T17:06:36.1730277Z  --tty \ 2025-11-03T17:06:36.1730450Z  --detach \ 2025-11-03T17:06:36.1730641Z  --name="${container_name}" \ 2025-11-03T17:06:36.1730855Z  ${JENKINS_USER} \ 2025-11-03T17:06:36.1731107Z  -v "${GITHUB_WORKSPACE}:/var/lib/jenkins/workspace" \ 2025-11-03T17:06:36.1731390Z  -w /var/lib/jenkins/workspace \ 2025-11-03T17:06:36.1731616Z  "${DOCKER_IMAGE}" \ 2025-11-03T17:06:36.1731809Z  ${DOCKER_SHELL_CMD} 2025-11-03T17:06:36.1731987Z ) 2025-11-03T17:06:36.1732221Z echo "DOCKER_CONTAINER_ID=${container_name}" >> "${GITHUB_ENV}" 2025-11-03T17:06:36.1732508Z  2025-11-03T17:06:36.1732686Z if [[ ${BUILD_ENVIRONMENT} == *"s390x"* ]]; then 2025-11-03T17:06:36.1733115Z  docker exec -t "${container_name}" sh -c "python3 -m pip install -r .ci/docker/requirements-ci.txt" 2025-11-03T17:06:36.1733484Z fi 2025-11-03T17:06:36.1733627Z  2025-11-03T17:06:36.1733971Z docker exec -t "${container_name}" sh -c "python3 -m pip install $(echo dist/*.whl)[opt-einsum] && ${TEST_COMMAND}" 2025-11-03T17:06:36.1741204Z shell: /usr/bin/bash -e {0} 2025-11-03T17:06:36.1741419Z env: 2025-11-03T17:06:36.1741582Z GIT_DEFAULT_BRANCH: main 2025-11-03T17:06:36.1741821Z GPU_FLAG: --gpus all -e NVIDIA_DRIVER_CAPABILITIES=all 2025-11-03T17:06:36.1742124Z BUILD_ENVIRONMENT: linux-jammy-cuda12.8-py3.12-gcc11 2025-11-03T17:06:36.1742375Z PR_NUMBER: 2025-11-03T17:06:36.1742549Z GITHUB_REPOSITORY: pytorch/pytorch 2025-11-03T17:06:36.1742768Z GITHUB_WORKFLOW: vllm-test 2025-11-03T17:06:36.1742960Z GITHUB_JOB: test 2025-11-03T17:06:36.1743130Z GITHUB_RUN_ID: 19040285420 2025-11-03T17:06:36.1743317Z GITHUB_RUN_NUMBER: 3325 2025-11-03T17:06:36.1743503Z GITHUB_RUN_ATTEMPT: 1 2025-11-03T17:06:36.1743669Z JOB_ID: 54383051348 2025-11-03T17:06:36.1744014Z JOB_NAME: ci-vllm-test / test (vllm_distributed_test_28_failure_test, 1, 1, linux.g6.12xlarge.nvidia.gpu) 2025-11-03T17:06:36.1744406Z BRANCH: main 2025-11-03T17:06:36.1744593Z SHA1: 3f6538febd727b782e6e13cfd026a309fb14351d 2025-11-03T17:06:36.1744869Z BASE_SHA: 3f6538febd727b782e6e13cfd026a309fb14351d 2025-11-03T17:06:36.1745282Z TEST_CONFIG: vllm_distributed_test_28_failure_test 2025-11-03T17:06:36.1745518Z SHARD_NUMBER: 1 2025-11-03T17:06:36.1745693Z NUM_TEST_SHARDS: 1 2025-11-03T17:06:36.1745854Z EXTRA_FLAGS: 2025-11-03T17:06:36.1746019Z OP_BENCHMARK_TESTS: 2025-11-03T17:06:36.1746186Z REENABLED_ISSUES: 2025-11-03T17:06:36.1746361Z CONTINUE_THROUGH_ERROR: True 2025-11-03T17:06:36.1746563Z VERBOSE_TEST_LOGS: False 2025-11-03T17:06:36.1746751Z TEST_SHOWLOCALS: False 2025-11-03T17:06:36.1746925Z NO_TEST_TIMEOUT: False 2025-11-03T17:06:36.1747096Z NO_TD: False 2025-11-03T17:06:36.1747257Z TD_DISTRIBUTED: False 2025-11-03T17:06:36.1747481Z SCCACHE_BUCKET: ossci-compiler-cache-circleci-v2 2025-11-03T17:06:36.1747728Z SCCACHE_REGION: us-east-1 2025-11-03T17:06:36.1747905Z SHM_SIZE: 2g 2025-11-03T17:06:36.1748482Z DOCKER_IMAGE: 308535385114.dkr.ecr.us-east-1.amazonaws.com/pytorch/ci-image:pytorch-linux-jammy-cuda12.8-cudnn9-py3.12-gcc11-vllm-30147fb242f88ab8f13e14b76ab186d189d8605c 2025-11-03T17:06:36.1749105Z XLA_CUDA: 2025-11-03T17:06:36.1749368Z XLA_CLANG_CACHE_S3_BUCKET_NAME: ossci-compiler-clang-cache-circleci-xla 2025-11-03T17:06:36.1749685Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK: 0 2025-11-03T17:06:36.1749922Z PYTORCH_TEST_RERUN_DISABLED_TESTS: 0 2025-11-03T17:06:36.1750130Z DASHBOARD_TAG: 2025-11-03T17:06:36.1750468Z VLLM_TEST_HUGGING_FACE_TOKEN: *** 2025-11-03T17:06:36.1750769Z HUGGING_FACE_HUB_TOKEN: *** 2025-11-03T17:06:36.1751065Z SCRIBE_GRAPHQL_ACCESS_TOKEN: *** 2025-11-03T17:06:36.1751468Z ARTIFACTS_FILE_SUFFIX: test-vllm_distributed_test_28_failure_test-1-1-linux.g6.12xlarge.nvidia.gpu_54383051348 2025-11-03T17:06:36.1751874Z ##[endgroup] 2025-11-03T17:06:36.1796711Z + [[ vllm_distributed_test_28_failure_test == \m\u\l\t\i\g\p\u ]] 2025-11-03T17:06:36.1797188Z + [[ linux-jammy-cuda12.8-py3.12-gcc11 == *onnx* ]] 2025-11-03T17:06:36.1797570Z + TEST_COMMAND=.ci/pytorch/test.sh 2025-11-03T17:06:36.1803485Z ++ awk '/MemTotal/ { printf "%.3f \n", $2/1024/1024 - 1 }' /proc/meminfo 2025-11-03T17:06:36.1831893Z + TOTAL_AVAILABLE_MEMORY_IN_GB='180.760 ' 2025-11-03T17:06:36.1832202Z + TOTAL_MEMORY_WITH_SWAP=183 2025-11-03T17:06:36.1832491Z + [[ linux-jammy-cuda12.8-py3.12-gcc11 == *\s\3\9\0\x* ]] 2025-11-03T17:06:36.1832776Z + SHM_OPTS=--shm-size=2g 2025-11-03T17:06:36.1832991Z + JENKINS_USER='--user jenkins' 2025-11-03T17:06:36.1833201Z + DOCKER_SHELL_CMD= 2025-11-03T17:06:36.1848857Z +++ nproc --ignore=2 2025-11-03T17:06:36.1884124Z ++ docker run --gpus all -e NVIDIA_DRIVER_CAPABILITIES=all -e BUILD_ENVIRONMENT -e PR_NUMBER -e GITHUB_ACTIONS -e GITHUB_REPOSITORY -e GITHUB_WORKFLOW -e GITHUB_JOB -e GITHUB_RUN_ID -e GITHUB_RUN_NUMBER -e GITHUB_RUN_ATTEMPT -e JOB_ID -e JOB_NAME -e BASE_SHA -e BRANCH -e SHA1 -e AWS_DEFAULT_REGION -e IN_WHEEL_TEST -e SHARD_NUMBER -e TEST_CONFIG -e NUM_TEST_SHARDS -e REENABLED_ISSUES -e CONTINUE_THROUGH_ERROR -e VERBOSE_TEST_LOGS -e TEST_SHOWLOCALS -e NO_TEST_TIMEOUT -e NO_TD -e TD_DISTRIBUTED -e PR_LABELS -e MAX_JOBS=46 -e SCCACHE_BUCKET -e SCCACHE_REGION -e XLA_CUDA -e XLA_CLANG_CACHE_S3_BUCKET_NAME -e PYTORCH_TEST_CUDA_MEM_LEAK_CHECK -e PYTORCH_TEST_RERUN_DISABLED_TESTS -e SKIP_SCCACHE_INITIALIZATION=1 -e HUGGING_FACE_HUB_TOKEN -e VLLM_TEST_HUGGING_FACE_TOKEN -e SCRIBE_GRAPHQL_ACCESS_TOKEN -e DASHBOARD_TAG -e ARTIFACTS_FILE_SUFFIX --memory=180g --memory-swap=183g --env-file=/tmp/github_env_19040285420 --security-opt seccomp=unconfined --cap-add=SYS_PTRACE --ipc=host --shm-size=2g --tty --detach --name= --user jenkins -v /home/ec2-user/actions-runner/_work/pytorch/pytorch:/var/lib/jenkins/workspace -w /var/lib/jenkins/workspace 308535385114.dkr.ecr.us-east-1.amazonaws.com/pytorch/ci-image:pytorch-linux-jammy-cuda12.8-cudnn9-py3.12-gcc11-vllm-30147fb242f88ab8f13e14b76ab186d189d8605c 2025-11-03T17:06:58.0722533Z + container_name=99e00294b4bd7ffe64017e12af3ca26302f354938a9d519b091e205874b31030 2025-11-03T17:06:58.0724315Z + echo DOCKER_CONTAINER_ID=99e00294b4bd7ffe64017e12af3ca26302f354938a9d519b091e205874b31030 2025-11-03T17:06:58.0725208Z + [[ linux-jammy-cuda12.8-py3.12-gcc11 == *\s\3\9\0\x* ]] 2025-11-03T17:06:58.0732042Z ++ echo dist/torch-2.10.0a0+git3f6538f-cp312-cp312-linux_x86_64.whl 2025-11-03T17:06:58.0735266Z + docker exec -t 99e00294b4bd7ffe64017e12af3ca26302f354938a9d519b091e205874b31030 sh -c 'python3 -m pip install dist/torch-2.10.0a0+git3f6538f-cp312-cp312-linux_x86_64.whl[opt-einsum] && .ci/pytorch/test.sh' 2025-11-03T17:06:58.5846426Z Processing ./dist/torch-2.10.0a0+git3f6538f-cp312-cp312-linux_x86_64.whl (from torch==2.10.0a0+git3f6538f) 2025-11-03T17:06:58.9817381Z Requirement already satisfied: filelock in /opt/conda/envs/py_3.12/lib/python3.12/site-packages (from torch==2.10.0a0+git3f6538f->torch==2.10.0a0+git3f6538f) (3.18.0) 2025-11-03T17:06:58.9819469Z Requirement already satisfied: typing-extensions>=4.10.0 in /opt/conda/envs/py_3.12/lib/python3.12/site-packages (from torch==2.10.0a0+git3f6538f->torch==2.10.0a0+git3f6538f) (4.12.2) 2025-11-03T17:06:58.9823162Z Requirement already satisfied: setuptools in /opt/conda/envs/py_3.12/lib/python3.12/site-packages (from torch==2.10.0a0+git3f6538f->torch==2.10.0a0+git3f6538f) (78.1.1) 2025-11-03T17:06:58.9826632Z Requirement already satisfied: sympy>=1.13.3 in /opt/conda/envs/py_3.12/lib/python3.12/site-packages (from torch==2.10.0a0+git3f6538f->torch==2.10.0a0+git3f6538f) (1.13.3) 2025-11-03T17:06:58.9831190Z Requirement already satisfied: networkx>=2.5.1 in /opt/conda/envs/py_3.12/lib/python3.12/site-packages (from torch==2.10.0a0+git3f6538f->torch==2.10.0a0+git3f6538f) (2.8.8) 2025-11-03T17:06:58.9834679Z Requirement already satisfied: jinja2 in /opt/conda/envs/py_3.12/lib/python3.12/site-packages (from torch==2.10.0a0+git3f6538f->torch==2.10.0a0+git3f6538f) (3.1.6) 2025-11-03T17:06:58.9838153Z Requirement already satisfied: fsspec>=0.8.5 in /opt/conda/envs/py_3.12/lib/python3.12/site-packages (from torch==2.10.0a0+git3f6538f->torch==2.10.0a0+git3f6538f) (2025.9.0) 2025-11-03T17:06:58.9850821Z Requirement already satisfied: opt-einsum>=3.3 in /opt/conda/envs/py_3.12/lib/python3.12/site-packages (from torch==2.10.0a0+git3f6538f->torch==2.10.0a0+git3f6538f) (3.3.0) 2025-11-03T17:06:58.9947361Z Requirement already satisfied: numpy>=1.7 in /opt/conda/envs/py_3.12/lib/python3.12/site-packages (from opt-einsum>=3.3->torch==2.10.0a0+git3f6538f->torch==2.10.0a0+git3f6538f) (1.26.2) 2025-11-03T17:06:58.9986801Z Requirement already satisfied: mpmath<1.4,>=1.1.0 in /opt/conda/envs/py_3.12/lib/python3.12/site-packages (from sympy>=1.13.3->torch==2.10.0a0+git3f6538f->torch==2.10.0a0+git3f6538f) (1.3.0) 2025-11-03T17:06:59.0030675Z Requirement already satisfied: MarkupSafe>=2.0 in /opt/conda/envs/py_3.12/lib/python3.12/site-packages (from jinja2->torch==2.10.0a0+git3f6538f->torch==2.10.0a0+git3f6538f) (3.0.3) 2025-11-03T17:06:59.1996942Z Installing collected packages: torch 2025-11-03T17:07:11.3597511Z Successfully installed torch-2.10.0a0+git3f6538f 2025-11-03T17:07:11.4378063Z + export TERM=vt100 2025-11-03T17:07:11.4378860Z + TERM=vt100 2025-11-03T17:07:11.4382968Z ++ dirname .ci/pytorch/test.sh 2025-11-03T17:07:11.4396118Z + source .ci/pytorch/common.sh 2025-11-03T17:07:11.4401788Z +++ dirname .ci/pytorch/common.sh 2025-11-03T17:07:11.4411147Z ++ source .ci/pytorch/common_utils.sh 2025-11-03T17:07:11.4413232Z +++ declare -f -t trap_add 2025-11-03T17:07:11.4417579Z ++ set -ex -o pipefail 2025-11-03T17:07:11.4417876Z ++ [[ linux-jammy-cuda12.8-py3.12-gcc11 == *rocm* ]] 2025-11-03T17:07:11.4421834Z ++ BUILD_TEST_LIBTORCH=0 2025-11-03T17:07:11.4422046Z ++ dirname .ci/pytorch/test.sh 2025-11-03T17:07:11.4430458Z + source .ci/pytorch/common-build.sh 2025-11-03T17:07:11.4431305Z ++ [[ linux-jammy-cuda12.8-py3.12-gcc11 != *win-* ]] 2025-11-03T17:07:11.4442538Z ++++ dirname .ci/pytorch/common-build.sh 2025-11-03T17:07:11.4452023Z +++ cd .ci/pytorch 2025-11-03T17:07:11.4452258Z +++ pwd -P 2025-11-03T17:07:11.4700038Z ++ script_dir=/var/lib/jenkins/workspace/.ci/pytorch 2025-11-03T17:07:11.4701045Z ++ [[ linux-jammy-cuda12.8-py3.12-gcc11 == *-pch* ]] 2025-11-03T17:07:11.4701731Z ++ which sccache 2025-11-03T17:07:11.4772804Z ++ [[ -z ossci-compiler-cache-circleci-v2 ]] 2025-11-03T17:07:11.4773424Z ++ sccache --stop-server 2025-11-03T17:07:11.4804314Z ++ true 2025-11-03T17:07:11.4804858Z ++ rm -f /var/lib/jenkins/sccache_error.log 2025-11-03T17:07:11.4820287Z ++ trap_add sccache_epilogue EXIT 2025-11-03T17:07:11.4820640Z ++ trap_add_cmd=sccache_epilogue 2025-11-03T17:07:11.4820844Z ++ shift 2025-11-03T17:07:11.4821004Z ++ for trap_add_name in "$@" 2025-11-03T17:07:11.4832775Z ++++ trap -p EXIT 2025-11-03T17:07:11.4840194Z +++ eval 'extract_trap_cmd ' 2025-11-03T17:07:11.4840547Z ++++ extract_trap_cmd 2025-11-03T17:07:11.4840715Z ++++ printf '%s\n' '' 2025-11-03T17:07:11.4840957Z +++ printf '%s\n' sccache_epilogue 2025-11-03T17:07:11.4841506Z ++ trap -- ' 2025-11-03T17:07:11.4841747Z sccache_epilogue' EXIT 2025-11-03T17:07:11.4841985Z ++ [[ -n 1 ]] 2025-11-03T17:07:11.4842349Z ++ echo 'Skipping sccache server initialization, setting environment variables' 2025-11-03T17:07:11.4842882Z Skipping sccache server initialization, setting environment variables 2025-11-03T17:07:11.4843271Z ++ export SCCACHE_IDLE_TIMEOUT=0 2025-11-03T17:07:11.4843528Z ++ SCCACHE_IDLE_TIMEOUT=0 2025-11-03T17:07:11.4843840Z ++ export SCCACHE_ERROR_LOG=/var/lib/jenkins/sccache_error.log 2025-11-03T17:07:11.4844259Z ++ SCCACHE_ERROR_LOG=/var/lib/jenkins/sccache_error.log 2025-11-03T17:07:11.4851384Z ++ export RUST_LOG=sccache::server=error 2025-11-03T17:07:11.4851680Z ++ RUST_LOG=sccache::server=error 2025-11-03T17:07:11.4851917Z ++ sccache --zero-stats 2025-11-03T17:07:11.7958253Z Statistics zeroed. 2025-11-03T17:07:11.7972256Z ++ which ccache 2025-11-03T17:07:11.8037532Z + [[ linux-jammy-cuda12.8-py3.12-gcc11 != *rocm* ]] 2025-11-03T17:07:11.8037950Z + [[ linux-jammy-cuda12.8-py3.12-gcc11 != *s390x* ]] 2025-11-03T17:07:11.8038311Z + [[ -d /var/lib/jenkins/workspace ]] 2025-11-03T17:07:11.8042217Z ++ stat -c %u /var/lib/jenkins/workspace 2025-11-03T17:07:11.8062127Z + WORKSPACE_ORIGINAL_OWNER_ID=1000 2025-11-03T17:07:11.8062414Z + trap_add cleanup_workspace EXIT 2025-11-03T17:07:11.8062713Z + trap_add_cmd=cleanup_workspace 2025-11-03T17:07:11.8062958Z + shift 2025-11-03T17:07:11.8063151Z + for trap_add_name in "$@" 2025-11-03T17:07:11.8073044Z +++ trap -p EXIT 2025-11-03T17:07:11.8078443Z ++ eval 'extract_trap_cmd trap -- '\'' 2025-11-03T17:07:11.8079021Z sccache_epilogue'\'' EXIT' 2025-11-03T17:07:11.8079603Z +++ extract_trap_cmd trap -- ' 2025-11-03T17:07:11.8080266Z sccache_epilogue' EXIT 2025-11-03T17:07:11.8080677Z +++ printf '%s\n' ' 2025-11-03T17:07:11.8081018Z sccache_epilogue' 2025-11-03T17:07:11.8081249Z ++ printf '%s\n' cleanup_workspace 2025-11-03T17:07:11.8081517Z + trap -- ' 2025-11-03T17:07:11.8081702Z sccache_epilogue 2025-11-03T17:07:11.8081911Z cleanup_workspace' EXIT 2025-11-03T17:07:11.8082181Z + sudo chown -R jenkins /var/lib/jenkins/workspace 2025-11-03T17:07:12.7789471Z + git config --global --add safe.directory /var/lib/jenkins/workspace 2025-11-03T17:07:12.7816278Z + [[ linux-jammy-cuda12.8-py3.12-gcc11 == *cuda* ]] 2025-11-03T17:07:12.7820416Z ++ python -c 'import os;import numba.cuda; print(os.path.dirname(numba.cuda.__file__))' 2025-11-03T17:07:13.2519487Z + NUMBA_CUDA_DIR=/opt/conda/envs/py_3.12/lib/python3.12/site-packages/numba/cuda 2025-11-03T17:07:13.2520596Z + '[' -n /opt/conda/envs/py_3.12/lib/python3.12/site-packages/numba/cuda ']' 2025-11-03T17:07:13.2527802Z +++ realpath .ci/pytorch/test.sh 2025-11-03T17:07:13.2542489Z ++ dirname /var/lib/jenkins/workspace/.ci/pytorch/test.sh 2025-11-03T17:07:13.2757742Z + NUMBA_PATCH=/var/lib/jenkins/workspace/.ci/pytorch/numba-cuda-13.patch 2025-11-03T17:07:13.2759060Z + pushd /opt/conda/envs/py_3.12/lib/python3.12/site-packages/numba/cuda 2025-11-03T17:07:13.2760466Z /opt/conda/envs/py_3.12/lib/python3.12/site-packages/numba/cuda ~/workspace 2025-11-03T17:07:13.2761454Z + patch -p4 2025-11-03T17:07:13.2777538Z patching file cudadrv/driver.py 2025-11-03T17:07:13.2782052Z + popd 2025-11-03T17:07:13.2782569Z ~/workspace 2025-11-03T17:07:13.2782953Z + echo 'Environment variables:' 2025-11-03T17:07:13.2783448Z Environment variables: 2025-11-03T17:07:13.2783654Z + env 2025-11-03T17:07:13.2797857Z GITHUB_WORKSPACE=/home/ec2-user/actions-runner/_work/pytorch/pytorch 2025-11-03T17:07:13.2798329Z CONTINUE_THROUGH_ERROR=True 2025-11-03T17:07:13.2798778Z BUILD_ENVIRONMENT=linux-jammy-cuda12.8-py3.12-gcc11 2025-11-03T17:07:13.2799545Z VLLM_TEST_HUGGING_FACE_TOKEN=*** 2025-11-03T17:07:13.2799931Z HOSTNAME=99e00294b4bd 2025-11-03T17:07:13.2800508Z GITHUB_PATH=/home/ec2-user/actions-runner/_work/_temp/_runner_file_commands/add_path_a033d513-bd65-4bfe-a9ed-90f849b1fda8 2025-11-03T17:07:13.2801172Z GITHUB_ACTION=__run_3 2025-11-03T17:07:13.2801488Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=0 2025-11-03T17:07:13.2801790Z GITHUB_RUN_NUMBER=3325 2025-11-03T17:07:13.2802042Z TEST_CONFIG=vllm_distributed_test_28_failure_test 2025-11-03T17:07:13.2802364Z GITHUB_REPOSITORY_OWNER_ID=21003710 2025-11-03T17:07:13.2802652Z TORCH_NVCC_FLAGS=-Xfatbin -compress-all 2025-11-03T17:07:13.2802884Z SCCACHE_IDLE_TIMEOUT=0 2025-11-03T17:07:13.2803184Z SCRIBE_GRAPHQL_ACCESS_TOKEN=*** 2025-11-03T17:07:13.2803414Z GITHUB_TRIGGERING_ACTOR=pytorchmergebot 2025-11-03T17:07:13.2803645Z GITHUB_REF_TYPE=branch 2025-11-03T17:07:13.2803859Z BASE_SHA=3f6538febd727b782e6e13cfd026a309fb14351d 2025-11-03T17:07:13.2804090Z XLA_CUDA= 2025-11-03T17:07:13.2804251Z NCCL_LIB_DIR=/usr/local/cuda/lib64/ 2025-11-03T17:07:13.2804544Z HUGGING_FACE_HUB_TOKEN=*** 2025-11-03T17:07:13.2804959Z *** 2025-11-03T17:07:13.2805114Z GITHUB_REPOSITORY_ID=65600975 2025-11-03T17:07:13.2805323Z GITHUB_ACTIONS=true 2025-11-03T17:07:13.2805506Z NVIDIA_DRIVER_CAPABILITIES=all 2025-11-03T17:07:13.2805750Z SCCACHE_ERROR_LOG=/var/lib/jenkins/sccache_error.log 2025-11-03T17:07:13.2806027Z SHA1=3f6538febd727b782e6e13cfd026a309fb14351d 2025-11-03T17:07:13.2806397Z GITHUB_SHA=3f6538febd727b782e6e13cfd026a309fb14351d 2025-11-03T17:07:13.2806863Z GITHUB_WORKFLOW_REF=pytorch/pytorch/.github/workflows/vllm.yml@refs/heads/main 2025-11-03T17:07:13.2807190Z UCC_HOME=/usr 2025-11-03T17:07:13.2807349Z VERBOSE_TEST_LOGS=False 2025-11-03T17:07:13.2807539Z GITHUB_REF=refs/heads/main 2025-11-03T17:07:13.2807720Z SHARD_NUMBER=1 2025-11-03T17:07:13.2807899Z GITHUB_REF_PROTECTED=true 2025-11-03T17:07:13.2808078Z HOME=/var/lib/jenkins 2025-11-03T17:07:13.2808281Z GITHUB_API_URL=https://api.github.com 2025-11-03T17:07:13.2808626Z PYTORCH_TEST_RERUN_DISABLED_TESTS=0 2025-11-03T17:07:13.2808873Z UCX_COMMIT=7836b165abdbe468a2f607e7254011c07d788152 2025-11-03T17:07:13.2809106Z USE_SYSTEM_NCCL=1 2025-11-03T17:07:13.2809272Z NUM_TEST_SHARDS=1 2025-11-03T17:07:13.2809635Z UCX_HOME=/usr 2025-11-03T17:07:13.2810115Z GITHUB_STATE=/home/ec2-user/actions-runner/_work/_temp/_runner_file_commands/save_state_a033d513-bd65-4bfe-a9ed-90f849b1fda8 2025-11-03T17:07:13.2810996Z JOB_NAME=ci-vllm-test / test (vllm_distributed_test_28_failure_test, 1, 1, linux.g6.12xlarge.nvidia.gpu) 2025-11-03T17:07:13.2811880Z GITHUB_ENV=/home/ec2-user/actions-runner/_work/_temp/_runner_file_commands/set_env_a033d513-bd65-4bfe-a9ed-90f849b1fda8 2025-11-03T17:07:13.2812461Z GITHUB_EVENT_PATH=/home/ec2-user/actions-runner/_work/_temp/_github_workflow/event.json 2025-11-03T17:07:13.2812813Z GITHUB_EVENT_NAME=push 2025-11-03T17:07:13.2812994Z DASHBOARD_TAG= 2025-11-03T17:07:13.2813155Z GITHUB_RUN_ID=19040285420 2025-11-03T17:07:13.2813344Z INSTALLED_OPENBLAS= 2025-11-03T17:07:13.2813779Z GITHUB_STEP_SUMMARY=/home/ec2-user/actions-runner/_work/_temp/_runner_file_commands/step_summary_a033d513-bd65-4bfe-a9ed-90f849b1fda8 2025-11-03T17:07:13.2814260Z GITHUB_ACTOR=pytorchmergebot 2025-11-03T17:07:13.2814447Z PR_NUMBER= 2025-11-03T17:07:13.2814599Z DESIRED_CUDA=12.8.1 2025-11-03T17:07:13.2814768Z GITHUB_RUN_ATTEMPT=1 2025-11-03T17:07:13.2814951Z ANACONDA_PYTHON_VERSION=3.12 2025-11-03T17:07:13.2815180Z GITHUB_GRAPHQL_URL=https://api.github.com/graphql 2025-11-03T17:07:13.2815423Z TERM=vt100 2025-11-03T17:07:13.2815580Z INSTALLED_VISION=yes 2025-11-03T17:07:13.2815753Z BRANCH=main 2025-11-03T17:07:13.2815908Z SCCACHE_REGION=us-east-1 2025-11-03T17:07:13.2816100Z OPENSSL_ROOT_DIR=/opt/openssl 2025-11-03T17:07:13.2816445Z BUILD_AOT_INDUCTOR_TEST= 2025-11-03T17:07:13.2816639Z CUDA_PATH=/usr/local/cuda 2025-11-03T17:07:13.2816999Z GITHUB_ACTION_PATH=/home/ec2-user/actions-runner/_work/pytorch/pytorch/./.github/actions/setup-linux 2025-11-03T17:07:13.2817407Z GITHUB_SERVER_URL=https://github.com 2025-11-03T17:07:13.2817660Z UCC_COMMIT=430e241bf5d38cbc73fc7a6b89155397232e3f96 2025-11-03T17:07:13.2817901Z REENABLED_ISSUES= 2025-11-03T17:07:13.2818053Z DOCS= 2025-11-03T17:07:13.2818192Z SHLVL=1 2025-11-03T17:07:13.2818331Z MAX_JOBS=46 2025-11-03T17:07:13.2818485Z GITHUB_ACTOR_ID=97764156 2025-11-03T17:07:13.2818721Z GITHUB_WORKFLOW_SHA=3f6538febd727b782e6e13cfd026a309fb14351d 2025-11-03T17:07:13.2818989Z GITHUB_REF_NAME=main 2025-11-03T17:07:13.2819258Z XLA_CLANG_CACHE_S3_BUCKET_NAME=ossci-compiler-clang-cache-circleci-xla 2025-11-03T17:07:13.2819559Z GITHUB_JOB=test 2025-11-03T17:07:13.2819714Z NO_TEST_TIMEOUT=False 2025-11-03T17:07:13.2819886Z TD_DISTRIBUTED=False 2025-11-03T17:07:13.2820077Z GITHUB_REPOSITORY=pytorch/pytorch 2025-11-03T17:07:13.2820290Z GITHUB_RETENTION_DAYS=90 2025-11-03T17:07:13.2820468Z OPENSSL_DIR=/opt/openssl 2025-11-03T17:07:13.2820655Z GITHUB_ACTION_REPOSITORY= 2025-11-03T17:07:13.2821201Z PATH=/opt/cache/bin:/usr/local/nvidia/bin:/usr/local/cuda/bin:/opt/conda/envs/py_3.12/bin:/opt/conda/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin 2025-11-03T17:07:13.2821752Z GITHUB_BASE_REF= 2025-11-03T17:07:13.2821909Z INSTALLED_ACL= 2025-11-03T17:07:13.2822275Z ARTIFACTS_FILE_SUFFIX=test-vllm_distributed_test_28_failure_test-1-1-linux.g6.12xlarge.nvidia.gpu_54383051348 2025-11-03T17:07:13.2822691Z CI=true 2025-11-03T17:07:13.2822854Z GITHUB_REPOSITORY_OWNER=pytorch 2025-11-03T17:07:13.2823084Z RUST_LOG=sccache::server=error 2025-11-03T17:07:13.2823278Z JOB_ID=54383051348 2025-11-03T17:07:13.2823444Z GITHUB_HEAD_REF= 2025-11-03T17:07:13.2823605Z GITHUB_ACTION_REF= 2025-11-03T17:07:13.2823808Z SCCACHE_BUCKET=ossci-compiler-cache-circleci-v2 2025-11-03T17:07:13.2824060Z TEST_SHOWLOCALS=False 2025-11-03T17:07:13.2824243Z GITHUB_WORKFLOW=vllm-test 2025-11-03T17:07:13.2824438Z DEBIAN_FRONTEND=noninteractive 2025-11-03T17:07:13.2824873Z GITHUB_OUTPUT=/home/ec2-user/actions-runner/_work/_temp/_runner_file_commands/set_output_a033d513-bd65-4bfe-a9ed-90f849b1fda8 2025-11-03T17:07:13.2825316Z NO_TD=False 2025-11-03T17:07:13.2825480Z SKIP_SCCACHE_INITIALIZATION=1 2025-11-03T17:07:13.2825696Z NCCL_INCLUDE_DIR=/usr/local/cuda/include/ 2025-11-03T17:07:13.2825911Z _=/usr/bin/env 2025-11-03T17:07:13.2826177Z OLDPWD=/opt/conda/envs/py_3.12/lib/python3.12/site-packages/numba/cuda 2025-11-03T17:07:13.2826552Z ++ python -c 'import site; print(site.getsitepackages()[0])' 2025-11-03T17:07:13.2933384Z + TORCH_INSTALL_DIR=/opt/conda/envs/py_3.12/lib/python3.12/site-packages/torch 2025-11-03T17:07:13.2935272Z + TORCH_BIN_DIR=/opt/conda/envs/py_3.12/lib/python3.12/site-packages/torch/bin 2025-11-03T17:07:13.2936500Z + TORCH_LIB_DIR=/opt/conda/envs/py_3.12/lib/python3.12/site-packages/torch/lib 2025-11-03T17:07:13.2937740Z + TORCH_TEST_DIR=/opt/conda/envs/py_3.12/lib/python3.12/site-packages/torch/test 2025-11-03T17:07:13.2938508Z + BUILD_DIR=build 2025-11-03T17:07:13.2938925Z + BUILD_RENAMED_DIR=build_renamed 2025-11-03T17:07:13.2939414Z + BUILD_BIN_DIR=build/bin 2025-11-03T17:07:13.2939837Z + SHARD_NUMBER=1 2025-11-03T17:07:13.2940211Z + NUM_TEST_SHARDS=1 2025-11-03T17:07:13.2940625Z + export TORCH_SERIALIZATION_DEBUG=1 2025-11-03T17:07:13.2941123Z + TORCH_SERIALIZATION_DEBUG=1 2025-11-03T17:07:13.2941568Z + export VALGRIND=ON 2025-11-03T17:07:13.2941791Z + VALGRIND=ON 2025-11-03T17:07:13.2941995Z + [[ linux-jammy-cuda12.8-py3.12-gcc11 == *clang9* ]] 2025-11-03T17:07:13.2942281Z + [[ linux-jammy-cuda12.8-py3.12-gcc11 == *xpu* ]] 2025-11-03T17:07:13.2942524Z + detect_cuda_arch 2025-11-03T17:07:13.2942730Z + [[ linux-jammy-cuda12.8-py3.12-gcc11 == *cuda* ]] 2025-11-03T17:07:13.2942976Z + command -v nvidia-smi 2025-11-03T17:07:13.2943161Z /usr/bin/nvidia-smi 2025-11-03T17:07:13.2946175Z ++ nvidia-smi --query-gpu=compute_cap --format=csv 2025-11-03T17:07:13.2946570Z ++ tail -n 1 2025-11-03T17:07:13.3507069Z + TORCH_CUDA_ARCH_LIST=8.9 2025-11-03T17:07:13.3507584Z + export TORCH_CUDA_ARCH_LIST 2025-11-03T17:07:13.3508177Z + [[ linux-jammy-cuda12.8-py3.12-gcc11 == *s390x* ]] 2025-11-03T17:07:13.3508765Z + [[ 0 == \1 ]] 2025-11-03T17:07:13.3509136Z + [[ True == \1 ]] 2025-11-03T17:07:13.3511006Z + [[ linux-jammy-cuda12.8-py3.12-gcc11 != *bazel* ]] 2025-11-03T17:07:13.3512009Z ++ realpath build/custom_test_artifacts 2025-11-03T17:07:13.3656579Z + CUSTOM_TEST_ARTIFACT_BUILD_DIR=/var/lib/jenkins/workspace/build/custom_test_artifacts 2025-11-03T17:07:13.3657457Z + [[ -n '' ]] 2025-11-03T17:07:13.3657874Z + echo 'Environment variables' 2025-11-03T17:07:13.3658343Z Environment variables 2025-11-03T17:07:13.3658726Z + env 2025-11-03T17:07:13.3789627Z GITHUB_WORKSPACE=/home/ec2-user/actions-runner/_work/pytorch/pytorch 2025-11-03T17:07:13.3790609Z CONTINUE_THROUGH_ERROR=True 2025-11-03T17:07:13.3791277Z BUILD_ENVIRONMENT=linux-jammy-cuda12.8-py3.12-gcc11 2025-11-03T17:07:13.3792803Z VLLM_TEST_HUGGING_FACE_TOKEN=*** 2025-11-03T17:07:13.3793314Z HOSTNAME=99e00294b4bd 2025-11-03T17:07:13.3793899Z GITHUB_PATH=/home/ec2-user/actions-runner/_work/_temp/_runner_file_commands/add_path_a033d513-bd65-4bfe-a9ed-90f849b1fda8 2025-11-03T17:07:13.3794451Z GITHUB_ACTION=__run_3 2025-11-03T17:07:13.3794682Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=0 2025-11-03T17:07:13.3794964Z GITHUB_RUN_NUMBER=3325 2025-11-03T17:07:13.3795229Z TEST_CONFIG=vllm_distributed_test_28_failure_test 2025-11-03T17:07:13.3795546Z GITHUB_REPOSITORY_OWNER_ID=21003710 2025-11-03T17:07:13.3795827Z TORCH_NVCC_FLAGS=-Xfatbin -compress-all 2025-11-03T17:07:13.3796101Z SCCACHE_IDLE_TIMEOUT=0 2025-11-03T17:07:13.3796464Z SCRIBE_GRAPHQL_ACCESS_TOKEN=*** 2025-11-03T17:07:13.3796752Z GITHUB_TRIGGERING_ACTOR=pytorchmergebot 2025-11-03T17:07:13.3797030Z GITHUB_REF_TYPE=branch 2025-11-03T17:07:13.3797261Z TORCH_CUDA_ARCH_LIST=8.9 2025-11-03T17:07:13.3797703Z BASE_SHA=3f6538febd727b782e6e13cfd026a309fb14351d 2025-11-03T17:07:13.3798088Z XLA_CUDA= 2025-11-03T17:07:13.3798292Z NCCL_LIB_DIR=/usr/local/cuda/lib64/ 2025-11-03T17:07:13.3799124Z HUGGING_FACE_HUB_TOKEN=*** 2025-11-03T17:07:13.3799548Z *** 2025-11-03T17:07:13.3799826Z GITHUB_REPOSITORY_ID=65600975 2025-11-03T17:07:13.3800208Z GITHUB_ACTIONS=true 2025-11-03T17:07:13.3800434Z NVIDIA_DRIVER_CAPABILITIES=all 2025-11-03T17:07:13.3800730Z SCCACHE_ERROR_LOG=/var/lib/jenkins/sccache_error.log 2025-11-03T17:07:13.3801087Z SHA1=3f6538febd727b782e6e13cfd026a309fb14351d 2025-11-03T17:07:13.3801355Z GITHUB_SHA=3f6538febd727b782e6e13cfd026a309fb14351d 2025-11-03T17:07:13.3801738Z GITHUB_WORKFLOW_REF=pytorch/pytorch/.github/workflows/vllm.yml@refs/heads/main 2025-11-03T17:07:13.3802057Z UCC_HOME=/usr 2025-11-03T17:07:13.3802223Z TORCH_SERIALIZATION_DEBUG=1 2025-11-03T17:07:13.3802612Z VERBOSE_TEST_LOGS=False 2025-11-03T17:07:13.3802801Z GITHUB_REF=refs/heads/main 2025-11-03T17:07:13.3802987Z SHARD_NUMBER=1 2025-11-03T17:07:13.3803147Z GITHUB_REF_PROTECTED=true 2025-11-03T17:07:13.3803337Z HOME=/var/lib/jenkins 2025-11-03T17:07:13.3803536Z GITHUB_API_URL=https://api.github.com 2025-11-03T17:07:13.3803770Z PYTORCH_TEST_RERUN_DISABLED_TESTS=0 2025-11-03T17:07:13.3804016Z UCX_COMMIT=7836b165abdbe468a2f607e7254011c07d788152 2025-11-03T17:07:13.3804255Z USE_SYSTEM_NCCL=1 2025-11-03T17:07:13.3804416Z NUM_TEST_SHARDS=1 2025-11-03T17:07:13.3804567Z UCX_HOME=/usr 2025-11-03T17:07:13.3804977Z GITHUB_STATE=/home/ec2-user/actions-runner/_work/_temp/_runner_file_commands/save_state_a033d513-bd65-4bfe-a9ed-90f849b1fda8 2025-11-03T17:07:13.3805600Z JOB_NAME=ci-vllm-test / test (vllm_distributed_test_28_failure_test, 1, 1, linux.g6.12xlarge.nvidia.gpu) 2025-11-03T17:07:13.3806202Z GITHUB_ENV=/home/ec2-user/actions-runner/_work/_temp/_runner_file_commands/set_env_a033d513-bd65-4bfe-a9ed-90f849b1fda8 2025-11-03T17:07:13.3806770Z GITHUB_EVENT_PATH=/home/ec2-user/actions-runner/_work/_temp/_github_workflow/event.json 2025-11-03T17:07:13.3807110Z GITHUB_EVENT_NAME=push 2025-11-03T17:07:13.3807425Z DASHBOARD_TAG= 2025-11-03T17:07:13.3807598Z GITHUB_RUN_ID=19040285420 2025-11-03T17:07:13.3807785Z INSTALLED_OPENBLAS= 2025-11-03T17:07:13.3808214Z GITHUB_STEP_SUMMARY=/home/ec2-user/actions-runner/_work/_temp/_runner_file_commands/step_summary_a033d513-bd65-4bfe-a9ed-90f849b1fda8 2025-11-03T17:07:13.3808807Z GITHUB_ACTOR=pytorchmergebot 2025-11-03T17:07:13.3809007Z PR_NUMBER= 2025-11-03T17:07:13.3809160Z DESIRED_CUDA=12.8.1 2025-11-03T17:07:13.3809322Z GITHUB_RUN_ATTEMPT=1 2025-11-03T17:07:13.3809721Z VALGRIND=ON 2025-11-03T17:07:13.3809894Z ANACONDA_PYTHON_VERSION=3.12 2025-11-03T17:07:13.3810135Z GITHUB_GRAPHQL_URL=https://api.github.com/graphql 2025-11-03T17:07:13.3810374Z TERM=vt100 2025-11-03T17:07:13.3810526Z INSTALLED_VISION=yes 2025-11-03T17:07:13.3810703Z BRANCH=main 2025-11-03T17:07:13.3810869Z SCCACHE_REGION=us-east-1 2025-11-03T17:07:13.3811057Z OPENSSL_ROOT_DIR=/opt/openssl 2025-11-03T17:07:13.3811261Z BUILD_AOT_INDUCTOR_TEST= 2025-11-03T17:07:13.3811442Z CUDA_PATH=/usr/local/cuda 2025-11-03T17:07:13.3811810Z GITHUB_ACTION_PATH=/home/ec2-user/actions-runner/_work/pytorch/pytorch/./.github/actions/setup-linux 2025-11-03T17:07:13.3812209Z GITHUB_SERVER_URL=https://github.com 2025-11-03T17:07:13.3812461Z UCC_COMMIT=430e241bf5d38cbc73fc7a6b89155397232e3f96 2025-11-03T17:07:13.3812699Z REENABLED_ISSUES= 2025-11-03T17:07:13.3812853Z DOCS= 2025-11-03T17:07:13.3812987Z SHLVL=1 2025-11-03T17:07:13.3813126Z MAX_JOBS=46 2025-11-03T17:07:13.3813277Z GITHUB_ACTOR_ID=97764156 2025-11-03T17:07:13.3813515Z GITHUB_WORKFLOW_SHA=3f6538febd727b782e6e13cfd026a309fb14351d 2025-11-03T17:07:13.3813776Z GITHUB_REF_NAME=main 2025-11-03T17:07:13.3814044Z XLA_CLANG_CACHE_S3_BUCKET_NAME=ossci-compiler-clang-cache-circleci-xla 2025-11-03T17:07:13.3814340Z GITHUB_JOB=test 2025-11-03T17:07:13.3814500Z NO_TEST_TIMEOUT=False 2025-11-03T17:07:13.3814673Z TD_DISTRIBUTED=False 2025-11-03T17:07:13.3814856Z GITHUB_REPOSITORY=pytorch/pytorch 2025-11-03T17:07:13.3815067Z GITHUB_RETENTION_DAYS=90 2025-11-03T17:07:13.3815254Z OPENSSL_DIR=/opt/openssl 2025-11-03T17:07:13.3815434Z GITHUB_ACTION_REPOSITORY= 2025-11-03T17:07:13.3815969Z PATH=/opt/cache/bin:/usr/local/nvidia/bin:/usr/local/cuda/bin:/opt/conda/envs/py_3.12/bin:/opt/conda/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin 2025-11-03T17:07:13.3816511Z GITHUB_BASE_REF= 2025-11-03T17:07:13.3816677Z INSTALLED_ACL= 2025-11-03T17:07:13.3817032Z ARTIFACTS_FILE_SUFFIX=test-vllm_distributed_test_28_failure_test-1-1-linux.g6.12xlarge.nvidia.gpu_54383051348 2025-11-03T17:07:13.3817437Z CI=true 2025-11-03T17:07:13.3817595Z GITHUB_REPOSITORY_OWNER=pytorch 2025-11-03T17:07:13.3817818Z RUST_LOG=sccache::server=error 2025-11-03T17:07:13.3818002Z JOB_ID=54383051348 2025-11-03T17:07:13.3818169Z GITHUB_HEAD_REF= 2025-11-03T17:07:13.3818334Z GITHUB_ACTION_REF= 2025-11-03T17:07:13.3818690Z SCCACHE_BUCKET=ossci-compiler-cache-circleci-v2 2025-11-03T17:07:13.3818939Z TEST_SHOWLOCALS=False 2025-11-03T17:07:13.3819130Z GITHUB_WORKFLOW=vllm-test 2025-11-03T17:07:13.3819331Z DEBIAN_FRONTEND=noninteractive 2025-11-03T17:07:13.3819776Z GITHUB_OUTPUT=/home/ec2-user/actions-runner/_work/_temp/_runner_file_commands/set_output_a033d513-bd65-4bfe-a9ed-90f849b1fda8 2025-11-03T17:07:13.3820214Z NO_TD=False 2025-11-03T17:07:13.3820381Z SKIP_SCCACHE_INITIALIZATION=1 2025-11-03T17:07:13.3820597Z NCCL_INCLUDE_DIR=/usr/local/cuda/include/ 2025-11-03T17:07:13.3820921Z OLDPWD=/opt/conda/envs/py_3.12/lib/python3.12/site-packages/numba/cuda 2025-11-03T17:07:13.3821207Z _=/usr/bin/env 2025-11-03T17:07:13.3821373Z + echo 'Testing pytorch' 2025-11-03T17:07:13.3821554Z Testing pytorch 2025-11-03T17:07:13.3821719Z + export LANG=C.UTF-8 2025-11-03T17:07:13.3821877Z + LANG=C.UTF-8 2025-11-03T17:07:13.3822040Z + PR_NUMBER= 2025-11-03T17:07:13.3822252Z + [[ vllm_distributed_test_28_failure_test == \d\e\f\a\u\l\t ]] 2025-11-03T17:07:13.3822598Z + [[ vllm_distributed_test_28_failure_test == \d\i\s\t\r\i\b\u\t\e\d ]] 2025-11-03T17:07:13.3822916Z + [[ vllm_distributed_test_28_failure_test == \s\l\o\w ]] 2025-11-03T17:07:13.3823347Z + [[ linux-jammy-cuda12.8-py3.12-gcc11 == *slow-gradcheck* ]] 2025-11-03T17:07:13.3823661Z + [[ linux-jammy-cuda12.8-py3.12-gcc11 == *cuda* ]] 2025-11-03T17:07:13.3823928Z + export PYTORCH_TESTING_DEVICE_ONLY_FOR=cuda 2025-11-03T17:07:13.3824183Z + PYTORCH_TESTING_DEVICE_ONLY_FOR=cuda 2025-11-03T17:07:13.3824437Z + [[ vllm_distributed_test_28_failure_test == *crossref* ]] 2025-11-03T17:07:13.3824721Z + [[ linux-jammy-cuda12.8-py3.12-gcc11 == *rocm* ]] 2025-11-03T17:07:13.3824986Z + [[ linux-jammy-cuda12.8-py3.12-gcc11 == *xpu* ]] 2025-11-03T17:07:13.3825263Z + [[ linux-jammy-cuda12.8-py3.12-gcc11 != *-bazel-* ]] 2025-11-03T17:07:13.3825507Z + pip_install ninja==1.10.2 2025-11-03T17:07:13.3825758Z + pip_install_pkg='python3 -m pip install --progress-bar off' 2025-11-03T17:07:13.3826085Z + python3 -m pip install --progress-bar off ninja==1.10.2 2025-11-03T17:07:14.1203793Z Collecting ninja==1.10.2 2025-11-03T17:07:14.1390647Z Downloading ninja-1.10.2-py2.py3-none-manylinux_2_5_x86_64.manylinux1_x86_64.whl.metadata (5.0 kB) 2025-11-03T17:07:14.2139874Z Downloading ninja-1.10.2-py2.py3-none-manylinux_2_5_x86_64.manylinux1_x86_64.whl (108 kB) 2025-11-03T17:07:14.4762986Z Installing collected packages: ninja 2025-11-03T17:07:14.4763367Z Attempting uninstall: ninja 2025-11-03T17:07:14.4790482Z Found existing installation: ninja 1.11.1.4 2025-11-03T17:07:14.4813096Z Uninstalling ninja-1.11.1.4: 2025-11-03T17:07:14.4907516Z Successfully uninstalled ninja-1.11.1.4 2025-11-03T17:07:14.5711115Z Successfully installed ninja-1.10.2 2025-11-03T17:07:14.6196247Z + export PATH=/var/lib/jenkins/.local/bin:/opt/cache/bin:/usr/local/nvidia/bin:/usr/local/cuda/bin:/opt/conda/envs/py_3.12/bin:/opt/conda/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin 2025-11-03T17:07:14.6198970Z + PATH=/var/lib/jenkins/.local/bin:/opt/cache/bin:/usr/local/nvidia/bin:/usr/local/cuda/bin:/opt/conda/envs/py_3.12/bin:/opt/conda/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin 2025-11-03T17:07:14.6200634Z + [[ linux-jammy-cuda12.8-py3.12-gcc11 == *aarch64* ]] 2025-11-03T17:07:14.6201698Z + [[ linux-jammy-cuda12.8-py3.12-gcc11 == *asan* ]] 2025-11-03T17:07:14.6202559Z + [[ linux-jammy-cuda12.8-py3.12-gcc11 == *-debug* ]] 2025-11-03T17:07:14.6203667Z + [[ linux-jammy-cuda12.8-py3.12-gcc11 != *-bazel-* ]] 2025-11-03T17:07:14.6204202Z + echo 'We are not in debug mode: linux-jammy-cuda12.8-py3.12-gcc11. Expect the assertion to pass' 2025-11-03T17:07:14.6204722Z We are not in debug mode: linux-jammy-cuda12.8-py3.12-gcc11. Expect the assertion to pass 2025-11-03T17:07:14.6205063Z + cd test 2025-11-03T17:07:14.6205319Z + python -c 'import torch; torch._C._crash_if_debug_asserts_fail(424242)' 2025-11-03T17:07:16.0956950Z + [[ vllm_distributed_test_28_failure_test == \n\o\g\p\u\_\N\O\_\A\V\X\2 ]] 2025-11-03T17:07:16.0957973Z + [[ vllm_distributed_test_28_failure_test == \n\o\g\p\u\_\A\V\X\5\1\2 ]] 2025-11-03T17:07:16.0958486Z + [[ vllm_distributed_test_28_failure_test == \l\e\g\a\c\y\_\n\v\i\d\i\a\_\d\r\i\v\e\r ]] 2025-11-03T17:07:16.0964072Z + DYNAMO_BENCHMARK_FLAGS=() 2025-11-03T17:07:16.0965259Z + [[ vllm_distributed_test_28_failure_test == *pr_time_benchmarks* ]] 2025-11-03T17:07:16.0965772Z + [[ vllm_distributed_test_28_failure_test == *dynamo_eager* ]] 2025-11-03T17:07:16.0966190Z + [[ vllm_distributed_test_28_failure_test == *aot_eager* ]] 2025-11-03T17:07:16.0966600Z + [[ vllm_distributed_test_28_failure_test == *aot_inductor* ]] 2025-11-03T17:07:16.0967029Z + [[ vllm_distributed_test_28_failure_test == *max_autotune_inductor* ]] 2025-11-03T17:07:16.0967474Z + [[ vllm_distributed_test_28_failure_test == *inductor* ]] 2025-11-03T17:07:16.0967867Z + [[ vllm_distributed_test_28_failure_test == *dynamic* ]] 2025-11-03T17:07:16.0968242Z + [[ vllm_distributed_test_28_failure_test == *cpu* ]] 2025-11-03T17:07:16.0968730Z + [[ vllm_distributed_test_28_failure_test == *xpu* ]] 2025-11-03T17:07:16.0969069Z + DYNAMO_BENCHMARK_FLAGS+=(--device cuda) 2025-11-03T17:07:16.1233024Z + [[ linux-jammy-cuda12.8-py3.12-gcc11 == *libtorch* ]] 2025-11-03T17:07:16.1233989Z + [[ linux-jammy-cuda12.8-py3.12-gcc11 == *-bazel-* ]] 2025-11-03T17:07:16.1236808Z + cd test 2025-11-03T17:07:16.1237134Z + python -c 'import torch; print(torch.__config__.show())' 2025-11-03T17:07:18.4737765Z PyTorch built with: 2025-11-03T17:07:18.4738084Z - GCC 11.4 2025-11-03T17:07:18.4738291Z - C++ Version: 201703 2025-11-03T17:07:18.4738829Z - Intel(R) oneAPI Math Kernel Library Version 2024.2-Product Build 20240605 for Intel(R) 64 architecture applications 2025-11-03T17:07:18.4739493Z - Intel(R) MKL-DNN v3.7.1 (Git Hash 8d263e693366ef8db40acc569cc7d8edf644556d) 2025-11-03T17:07:18.4739903Z - OpenMP 201511 (a.k.a. OpenMP 4.5) 2025-11-03T17:07:18.4740224Z - LAPACK is enabled (usually provided by MKL) 2025-11-03T17:07:18.4740509Z - NNPACK is enabled 2025-11-03T17:07:18.4740785Z - CPU capability usage: AVX2 2025-11-03T17:07:18.4741025Z - CUDA Runtime 12.8 2025-11-03T17:07:18.4741592Z - NVCC architecture flags: -gencode;arch=compute_80,code=sm_80;-gencode;arch=compute_89,code=sm_89;-gencode;arch=compute_90,code=sm_90 2025-11-03T17:07:18.4742153Z - CuDNN 90.8 2025-11-03T17:07:18.4746063Z - Build settings: BLAS_INFO=mkl, BUILD_TYPE=Release, COMMIT_SHA=3f6538febd727b782e6e13cfd026a309fb14351d, CUDA_VERSION=12.8, CUDNN_VERSION=9.8.0, CXX_COMPILER=/opt/cache/bin/c++, CXX_FLAGS= -fvisibility-inlines-hidden -DUSE_PTHREADPOOL -DNDEBUG -DUSE_KINETO -DLIBKINETO_NOROCTRACER -DLIBKINETO_NOXPUPTI=ON -DUSE_FBGEMM -DUSE_PYTORCH_QNNPACK -DUSE_XNNPACK -DSYMBOLICATE_MOBILE_DEBUG_HANDLE -O2 -fPIC -DC10_NODEPRECATED -Wall -Wextra -Werror=return-type -Werror=non-virtual-dtor -Werror=range-loop-construct -Werror=bool-operation -Wnarrowing -Wno-missing-field-initializers -Wno-unknown-pragmas -Wno-unused-parameter -Wno-strict-overflow -Wno-strict-aliasing -Wno-stringop-overflow -Wsuggest-override -Wno-psabi -Wno-error=old-style-cast -faligned-new -Werror -Wno-maybe-uninitialized -fno-math-errno -fno-trapping-math -Werror=format -Wno-stringop-overflow, FORCE_FALLBACK_CUDA_MPI=1, LAPACK_INFO=mkl, PERF_WITH_AVX=1, PERF_WITH_AVX2=1, TORCH_VERSION=2.10.0, USE_CUDA=ON, USE_CUDNN=ON, USE_CUSPARSELT=ON, USE_GFLAGS=OFF, USE_GLOG=OFF, USE_GLOO=ON, USE_MKL=ON, USE_MKLDNN=ON, USE_MPI=ON, USE_NCCL=ON, USE_NNPACK=ON, USE_OPENMP=ON, USE_ROCM=OFF, USE_ROCM_KERNEL_ASSERT=OFF, USE_XCCL=OFF, USE_XPU=OFF, 2025-11-03T17:07:18.4749521Z 2025-11-03T17:07:18.8114004Z + cd test 2025-11-03T17:07:18.8114371Z + python -c 'import torch; print(torch.__config__.parallel_info())' 2025-11-03T17:07:20.0306102Z ATen/Parallel: 2025-11-03T17:07:20.0306423Z at::get_num_threads() : 24 2025-11-03T17:07:20.0306719Z at::get_num_interop_threads() : 24 2025-11-03T17:07:20.0307012Z OpenMP 201511 (a.k.a. OpenMP 4.5) 2025-11-03T17:07:20.0307276Z omp_get_max_threads() : 24 2025-11-03T17:07:20.0308218Z Intel(R) oneAPI Math Kernel Library Version 2024.2-Product Build 20240605 for Intel(R) 64 architecture applications 2025-11-03T17:07:20.0308753Z mkl_get_max_threads() : 24 2025-11-03T17:07:20.0309120Z Intel(R) MKL-DNN v3.7.1 (Git Hash 8d263e693366ef8db40acc569cc7d8edf644556d) 2025-11-03T17:07:20.0309796Z std::thread::hardware_concurrency() : 48 2025-11-03T17:07:20.0310086Z Environment variables: 2025-11-03T17:07:20.0310313Z OMP_NUM_THREADS : [not set] 2025-11-03T17:07:20.0310549Z MKL_NUM_THREADS : [not set] 2025-11-03T17:07:20.0310780Z ATen parallel backend: OpenMP 2025-11-03T17:07:20.0310944Z 2025-11-03T17:07:20.2674389Z + [[ vllm_distributed_test_28_failure_test == *numpy_2* ]] 2025-11-03T17:07:20.2674836Z + [[ linux-jammy-cuda12.8-py3.12-gcc11 == *aarch64* ]] 2025-11-03T17:07:20.2675238Z + [[ vllm_distributed_test_28_failure_test == *backward* ]] 2025-11-03T17:07:20.2675626Z + [[ vllm_distributed_test_28_failure_test == *xla* ]] 2025-11-03T17:07:20.2675977Z + [[ vllm_distributed_test_28_failure_test == *vllm* ]] 2025-11-03T17:07:20.2676332Z + echo 'vLLM CI uses TORCH_CUDA_ARCH_LIST: 8.9' 2025-11-03T17:07:20.2676639Z vLLM CI uses TORCH_CUDA_ARCH_LIST: 8.9 2025-11-03T17:07:20.2679661Z + cd .ci/lumen_cli 2025-11-03T17:07:20.2680484Z + python -m pip install -e . 2025-11-03T17:07:20.5620930Z Obtaining file:///var/lib/jenkins/workspace/.ci/lumen_cli 2025-11-03T17:07:21.2986797Z Installing build dependencies ... [?25l- \ done 2025-11-03T17:07:21.5145313Z [?25h Checking if build backend supports build_editable ... [?25ldone 2025-11-03T17:07:21.7887005Z [?25h Getting requirements to build editable ... [?25l- done 2025-11-03T17:07:21.9560758Z [?25h Preparing editable metadata (pyproject.toml) ... [?25l- done 2025-11-03T17:07:22.0135837Z [?25hCollecting pyyaml==6.0.2 (from lumen-ci==0.1.0) 2025-11-03T17:07:22.0322946Z Downloading PyYAML-6.0.2-cp312-cp312-manylinux_2_17_x86_64.manylinux2014_x86_64.whl.metadata (2.1 kB) 2025-11-03T17:07:22.0548282Z Collecting GitPython==3.1.45 (from lumen-ci==0.1.0) 2025-11-03T17:07:22.0592478Z Downloading gitpython-3.1.45-py3-none-any.whl.metadata (13 kB) 2025-11-03T17:07:22.0811182Z Collecting docker==7.1.0 (from lumen-ci==0.1.0) 2025-11-03T17:07:22.0867565Z Downloading docker-7.1.0-py3-none-any.whl.metadata (3.8 kB) 2025-11-03T17:07:22.0936056Z Requirement already satisfied: pytest==7.3.2 in /opt/conda/envs/py_3.12/lib/python3.12/site-packages (from lumen-ci==0.1.0) (7.3.2) 2025-11-03T17:07:22.2794161Z Collecting uv==0.9.6 (from lumen-ci==0.1.0) 2025-11-03T17:07:22.2845598Z Downloading uv-0.9.6-py3-none-manylinux_2_17_x86_64.manylinux2014_x86_64.whl.metadata (11 kB) 2025-11-03T17:07:22.2917933Z Requirement already satisfied: requests>=2.26.0 in /opt/conda/envs/py_3.12/lib/python3.12/site-packages (from docker==7.1.0->lumen-ci==0.1.0) (2.32.5) 2025-11-03T17:07:22.2921116Z Requirement already satisfied: urllib3>=1.26.0 in /opt/conda/envs/py_3.12/lib/python3.12/site-packages (from docker==7.1.0->lumen-ci==0.1.0) (2.5.0) 2025-11-03T17:07:22.3042138Z Collecting gitdb<5,>=4.0.1 (from GitPython==3.1.45->lumen-ci==0.1.0) 2025-11-03T17:07:22.3082725Z Downloading gitdb-4.0.12-py3-none-any.whl.metadata (1.2 kB) 2025-11-03T17:07:22.3150858Z Requirement already satisfied: iniconfig in /opt/conda/envs/py_3.12/lib/python3.12/site-packages (from pytest==7.3.2->lumen-ci==0.1.0) (2.3.0) 2025-11-03T17:07:22.3154128Z Requirement already satisfied: packaging in /opt/conda/envs/py_3.12/lib/python3.12/site-packages (from pytest==7.3.2->lumen-ci==0.1.0) (23.1) 2025-11-03T17:07:22.3159123Z Requirement already satisfied: pluggy<2.0,>=0.12 in /opt/conda/envs/py_3.12/lib/python3.12/site-packages (from pytest==7.3.2->lumen-ci==0.1.0) (1.6.0) 2025-11-03T17:07:22.3289384Z Collecting smmap<6,>=3.0.1 (from gitdb<5,>=4.0.1->GitPython==3.1.45->lumen-ci==0.1.0) 2025-11-03T17:07:22.3335120Z Downloading smmap-5.0.2-py3-none-any.whl.metadata (4.3 kB) 2025-11-03T17:07:22.3410222Z Requirement already satisfied: charset_normalizer<4,>=2 in /opt/conda/envs/py_3.12/lib/python3.12/site-packages (from requests>=2.26.0->docker==7.1.0->lumen-ci==0.1.0) (3.4.4) 2025-11-03T17:07:22.3415508Z Requirement already satisfied: idna<4,>=2.5 in /opt/conda/envs/py_3.12/lib/python3.12/site-packages (from requests>=2.26.0->docker==7.1.0->lumen-ci==0.1.0) (3.11) 2025-11-03T17:07:22.3421080Z Requirement already satisfied: certifi>=2017.4.17 in /opt/conda/envs/py_3.12/lib/python3.12/site-packages (from requests>=2.26.0->docker==7.1.0->lumen-ci==0.1.0) (2025.10.5) 2025-11-03T17:07:22.3525829Z Downloading docker-7.1.0-py3-none-any.whl (147 kB) 2025-11-03T17:07:22.3670315Z Downloading gitpython-3.1.45-py3-none-any.whl (208 kB) 2025-11-03T17:07:22.3773470Z Downloading PyYAML-6.0.2-cp312-cp312-manylinux_2_17_x86_64.manylinux2014_x86_64.whl (767 kB) 2025-11-03T17:07:22.3865821Z [?25l ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 0.0/767.5 kB ? eta -:--:-- 2025-11-03T17:07:22.3866606Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 767.5/767.5 kB 92.6 MB/s 0:00:00 2025-11-03T17:07:22.3926297Z [?25hDownloading uv-0.9.6-py3-none-manylinux_2_17_x86_64.manylinux2014_x86_64.whl (21.4 MB) 2025-11-03T17:07:22.5613812Z [?25l ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 0.0/21.4 MB [31m? eta -:--:-- 2025-11-03T17:07:22.5614532Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 21.4/21.4 MB 127.6 MB/s 0:00:00 2025-11-03T17:07:22.5661433Z [?25hDownloading gitdb-4.0.12-py3-none-any.whl (62 kB) 2025-11-03T17:07:22.5741626Z Downloading smmap-5.0.2-py3-none-any.whl (24 kB) 2025-11-03T17:07:22.5990852Z Building wheels for collected packages: lumen-ci 2025-11-03T17:07:22.7807851Z Building editable for lumen-ci (pyproject.toml) ... [?25l- done 2025-11-03T17:07:22.7813821Z [?25h Created wheel for lumen-ci: filename=lumen_ci-0.1.0-0.editable-py3-none-any.whl size=2707 sha256=7b32567db24a576f8c1aa28a9577b4c215a3c05ed929aeb73fbc0a658d10afc8 2025-11-03T17:07:22.7814853Z Stored in directory: /tmp/pip-ephem-wheel-cache-u7p297lo/wheels/6d/29/6c/82214c85262c5bdb1d615291039f4ec9f974034103e7a303fa 2025-11-03T17:07:22.7828692Z Successfully built lumen-ci 2025-11-03T17:07:22.9454971Z Installing collected packages: uv, smmap, pyyaml, gitdb, docker, GitPython, lumen-ci 2025-11-03T17:07:23.1136104Z [?25l 2025-11-03T17:07:23.1901459Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 0/7 [uv] 2025-11-03T17:07:23.1902545Z  Attempting uninstall: pyyaml 2025-11-03T17:07:23.1954411Z ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 0/7 [uv] 2025-11-03T17:07:23.1955260Z  Found existing installation: PyYAML 6.0.3 2025-11-03T17:07:23.1990143Z ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 0/7 [uv] 2025-11-03T17:07:23.1990886Z  Uninstalling PyYAML-6.0.3: 2025-11-03T17:07:23.2134178Z ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 0/7 [uv] 2025-11-03T17:07:23.2134984Z  Successfully uninstalled PyYAML-6.0.3 2025-11-03T17:07:23.2850424Z ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 0/7 [uv] 2025-11-03T17:07:23.4568844Z  ━━━━━━━━━━━━━━━━━╺━━━━━━━━━━━━━━━━━━━━━━ 3/7 [gitdb] 2025-11-03T17:07:23.4714284Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━╸━━━━━━━━━━━ 5/7 [GitPython] 2025-11-03T17:07:23.4715371Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 7/7 [lumen-ci] 2025-11-03T17:07:23.4716030Z [?25h 2025-11-03T17:07:23.4817804Z Successfully installed GitPython-3.1.45 docker-7.1.0 gitdb-4.0.12 lumen-ci-0.1.0 pyyaml-6.0.2 smmap-5.0.2 uv-0.9.6 2025-11-03T17:07:23.5743462Z + python -m cli.run test external vllm --test-plan vllm_distributed_test_28_failure_test --shard-id 1 --num-shards 1 2025-11-03T17:07:23.7816004Z 2025-11-03 17:07:23,781 [INFO] cli.lib.core.vllm.vllm_test: Display VllmTestParameters VllmTestParameters(torch_whls_path=PosixPath('/var/lib/jenkins/workspace/dist'), vllm_whls_path=PosixPath('/var/lib/jenkins/workspace/dist/external/vllm/wheels'), torch_cuda_arch_list='8.9', cleaning_script=PosixPath('/var/lib/jenkins/workspace/.github/ci_configs/vllm/use_existing_torch.py')) 2025-11-03T17:07:23.7817976Z 2025-11-03 17:07:23,781 [INFO] cli.lib.common.git_helper: Cloning vllm to vllm 2025-11-03T17:07:23.9898363Z 2025-11-03 17:07:23,989 [INFO] cli.lib.common.git_helper: Progress: 5% - remote: Counting objects: 5% (2/39) 2025-11-03T17:07:23.9900181Z 2025-11-03 17:07:23,989 [INFO] cli.lib.common.git_helper: Progress: 10% - remote: Counting objects: 10% (4/39) 2025-11-03T17:07:23.9901951Z 2025-11-03 17:07:23,989 [INFO] cli.lib.common.git_helper: Progress: 15% - remote: Counting objects: 15% (6/39) 2025-11-03T17:07:23.9903726Z 2025-11-03 17:07:23,989 [INFO] cli.lib.common.git_helper: Progress: 20% - remote: Counting objects: 20% (8/39) 2025-11-03T17:07:23.9905108Z 2025-11-03 17:07:23,989 [INFO] cli.lib.common.git_helper: Progress: 25% - remote: Counting objects: 25% (10/39) 2025-11-03T17:07:23.9906413Z 2025-11-03 17:07:23,989 [INFO] cli.lib.common.git_helper: Progress: 30% - remote: Counting objects: 30% (12/39) 2025-11-03T17:07:23.9906996Z 2025-11-03 17:07:23,989 [INFO] cli.lib.common.git_helper: Progress: 35% - remote: Counting objects: 35% (14/39) 2025-11-03T17:07:23.9907833Z 2025-11-03 17:07:23,989 [INFO] cli.lib.common.git_helper: Progress: 100% - remote: Counting objects: 100% (39/39) 2025-11-03T17:07:23.9908365Z 2025-11-03 17:07:23,990 [INFO] cli.lib.common.git_helper: Progress: 5% - remote: Compressing objects: 5% (2/39) 2025-11-03T17:07:23.9924719Z 2025-11-03 17:07:23,992 [INFO] cli.lib.common.git_helper: Progress: 10% - remote: Compressing objects: 10% (4/39) 2025-11-03T17:07:23.9930175Z 2025-11-03 17:07:23,992 [INFO] cli.lib.common.git_helper: Progress: 15% - remote: Compressing objects: 15% (6/39) 2025-11-03T17:07:23.9945489Z 2025-11-03 17:07:23,994 [INFO] cli.lib.common.git_helper: Progress: 20% - remote: Compressing objects: 20% (8/39) 2025-11-03T17:07:23.9946043Z 2025-11-03 17:07:23,994 [INFO] cli.lib.common.git_helper: Progress: 25% - remote: Compressing objects: 25% (10/39) 2025-11-03T17:07:23.9946578Z 2025-11-03 17:07:23,994 [INFO] cli.lib.common.git_helper: Progress: 30% - remote: Compressing objects: 30% (12/39) 2025-11-03T17:07:23.9947111Z 2025-11-03 17:07:23,994 [INFO] cli.lib.common.git_helper: Progress: 35% - remote: Compressing objects: 35% (14/39) 2025-11-03T17:07:23.9995551Z 2025-11-03 17:07:23,999 [INFO] cli.lib.common.git_helper: Progress: 100% - remote: Compressing objects: 100% (39/39) 2025-11-03T17:07:24.0193702Z 2025-11-03 17:07:24,019 [INFO] cli.lib.common.git_helper: Progress: 0% - Receiving objects: 0% (1/136572) 2025-11-03T17:07:24.1162053Z 2025-11-03 17:07:24,115 [INFO] cli.lib.common.git_helper: Progress: 5% - Receiving objects: 5% (6829/136572) 2025-11-03T17:07:24.3491352Z 2025-11-03 17:07:24,348 [INFO] cli.lib.common.git_helper: Progress: 10% - Receiving objects: 10% (13658/136572) 2025-11-03T17:07:24.5547256Z 2025-11-03 17:07:24,554 [INFO] cli.lib.common.git_helper: Progress: 15% - Receiving objects: 15% (20486/136572), 28.06 MiB | 56.11 MiB/s 2025-11-03T17:07:24.7002683Z 2025-11-03 17:07:24,699 [INFO] cli.lib.common.git_helper: Progress: 20% - Receiving objects: 20% (27315/136572), 28.06 MiB | 56.11 MiB/s 2025-11-03T17:07:24.8183599Z 2025-11-03 17:07:24,818 [INFO] cli.lib.common.git_helper: Progress: 25% - Receiving objects: 25% (34143/136572), 28.06 MiB | 56.11 MiB/s 2025-11-03T17:07:24.9521985Z 2025-11-03 17:07:24,951 [INFO] cli.lib.common.git_helper: Progress: 30% - Receiving objects: 30% (40972/136572), 28.06 MiB | 56.11 MiB/s 2025-11-03T17:07:25.0790005Z 2025-11-03 17:07:25,078 [INFO] cli.lib.common.git_helper: Progress: 35% - Receiving objects: 35% (47801/136572), 56.22 MiB | 56.21 MiB/s 2025-11-03T17:07:25.1932400Z 2025-11-03 17:07:25,192 [INFO] cli.lib.common.git_helper: Progress: 40% - Receiving objects: 40% (54629/136572), 56.22 MiB | 56.21 MiB/s 2025-11-03T17:07:25.2900326Z 2025-11-03 17:07:25,289 [INFO] cli.lib.common.git_helper: Progress: 45% - Receiving objects: 45% (61458/136572), 56.22 MiB | 56.21 MiB/s 2025-11-03T17:07:25.3823968Z 2025-11-03 17:07:25,381 [INFO] cli.lib.common.git_helper: Progress: 50% - Receiving objects: 50% (68286/136572), 56.22 MiB | 56.21 MiB/s 2025-11-03T17:07:25.4257838Z 2025-11-03 17:07:25,425 [INFO] cli.lib.common.git_helper: Progress: 55% - Receiving objects: 55% (75115/136572), 56.22 MiB | 56.21 MiB/s 2025-11-03T17:07:25.4706306Z 2025-11-03 17:07:25,470 [INFO] cli.lib.common.git_helper: Progress: 60% - Receiving objects: 60% (81944/136572), 56.22 MiB | 56.21 MiB/s 2025-11-03T17:07:25.5058923Z 2025-11-03 17:07:25,505 [INFO] cli.lib.common.git_helper: Progress: 65% - Receiving objects: 65% (88772/136572), 56.22 MiB | 56.21 MiB/s 2025-11-03T17:07:25.5457962Z 2025-11-03 17:07:25,545 [INFO] cli.lib.common.git_helper: Progress: 70% - Receiving objects: 70% (95601/136572), 86.64 MiB | 57.75 MiB/s 2025-11-03T17:07:25.6054745Z 2025-11-03 17:07:25,605 [INFO] cli.lib.common.git_helper: Progress: 75% - Receiving objects: 75% (102429/136572), 86.64 MiB | 57.75 MiB/s 2025-11-03T17:07:25.6664512Z 2025-11-03 17:07:25,666 [INFO] cli.lib.common.git_helper: Progress: 80% - Receiving objects: 80% (109258/136572), 86.64 MiB | 57.75 MiB/s 2025-11-03T17:07:25.7275689Z 2025-11-03 17:07:25,727 [INFO] cli.lib.common.git_helper: Progress: 85% - Receiving objects: 85% (116087/136572), 86.64 MiB | 57.75 MiB/s 2025-11-03T17:07:25.7601006Z 2025-11-03 17:07:25,759 [INFO] cli.lib.common.git_helper: Progress: 90% - Receiving objects: 90% (122915/136572), 86.64 MiB | 57.75 MiB/s 2025-11-03T17:07:25.8070723Z 2025-11-03 17:07:25,806 [INFO] cli.lib.common.git_helper: Progress: 95% - Receiving objects: 95% (129744/136572), 86.64 MiB | 57.75 MiB/s 2025-11-03T17:07:25.8401692Z 2025-11-03 17:07:25,839 [INFO] cli.lib.common.git_helper: Progress: 100% - Receiving objects: 100% (136572/136572), 86.64 MiB | 57.75 MiB/s 2025-11-03T17:07:25.8601585Z 2025-11-03 17:07:25,859 [INFO] cli.lib.common.git_helper: Resolving deltas: 0% (0/108697) 2025-11-03T17:07:25.8786333Z 2025-11-03 17:07:25,878 [INFO] cli.lib.common.git_helper: Progress: 5% - Resolving deltas: 5% (5435/108697) 2025-11-03T17:07:25.9029485Z 2025-11-03 17:07:25,902 [INFO] cli.lib.common.git_helper: Progress: 10% - Resolving deltas: 10% (10870/108697) 2025-11-03T17:07:25.9271487Z 2025-11-03 17:07:25,926 [INFO] cli.lib.common.git_helper: Progress: 15% - Resolving deltas: 15% (16305/108697) 2025-11-03T17:07:25.9478467Z 2025-11-03 17:07:25,947 [INFO] cli.lib.common.git_helper: Progress: 20% - Resolving deltas: 20% (21740/108697) 2025-11-03T17:07:25.9717752Z 2025-11-03 17:07:25,971 [INFO] cli.lib.common.git_helper: Progress: 25% - Resolving deltas: 25% (27176/108697) 2025-11-03T17:07:26.0012899Z 2025-11-03 17:07:26,000 [INFO] cli.lib.common.git_helper: Progress: 30% - Resolving deltas: 30% (32610/108697) 2025-11-03T17:07:26.0330050Z 2025-11-03 17:07:26,032 [INFO] cli.lib.common.git_helper: Progress: 35% - Resolving deltas: 35% (38044/108697) 2025-11-03T17:07:26.0568751Z 2025-11-03 17:07:26,056 [INFO] cli.lib.common.git_helper: Progress: 40% - Resolving deltas: 40% (43479/108697) 2025-11-03T17:07:26.0813035Z 2025-11-03 17:07:26,081 [INFO] cli.lib.common.git_helper: Progress: 45% - Resolving deltas: 45% (48914/108697) 2025-11-03T17:07:26.1143391Z 2025-11-03 17:07:26,114 [INFO] cli.lib.common.git_helper: Progress: 50% - Resolving deltas: 50% (54349/108697) 2025-11-03T17:07:26.1403441Z 2025-11-03 17:07:26,140 [INFO] cli.lib.common.git_helper: Progress: 55% - Resolving deltas: 55% (59784/108697) 2025-11-03T17:07:26.1732980Z 2025-11-03 17:07:26,173 [INFO] cli.lib.common.git_helper: Progress: 60% - Resolving deltas: 60% (65219/108697) 2025-11-03T17:07:26.2112656Z 2025-11-03 17:07:26,210 [INFO] cli.lib.common.git_helper: Progress: 65% - Resolving deltas: 65% (70654/108697) 2025-11-03T17:07:26.2326621Z 2025-11-03 17:07:26,232 [INFO] cli.lib.common.git_helper: Progress: 70% - Resolving deltas: 70% (76089/108697) 2025-11-03T17:07:26.2575546Z 2025-11-03 17:07:26,257 [INFO] cli.lib.common.git_helper: Progress: 75% - Resolving deltas: 75% (81523/108697) 2025-11-03T17:07:26.2984408Z 2025-11-03 17:07:26,298 [INFO] cli.lib.common.git_helper: Progress: 80% - Resolving deltas: 80% (86958/108697) 2025-11-03T17:07:26.3388747Z 2025-11-03 17:07:26,338 [INFO] cli.lib.common.git_helper: Progress: 85% - Resolving deltas: 85% (92393/108697) 2025-11-03T17:07:26.3729615Z 2025-11-03 17:07:26,372 [INFO] cli.lib.common.git_helper: Progress: 90% - Resolving deltas: 90% (97828/108697) 2025-11-03T17:07:26.4056677Z 2025-11-03 17:07:26,405 [INFO] cli.lib.common.git_helper: Progress: 95% - Resolving deltas: 95% (103263/108697) 2025-11-03T17:07:26.4247773Z 2025-11-03 17:07:26,424 [INFO] cli.lib.common.git_helper: Progress: 100% - Resolving deltas: 100% (108697/108697) 2025-11-03T17:07:27.1560529Z 2025-11-03 17:07:27,155 [INFO] cli.lib.common.git_helper: Checking out pinned vllm commit e5192819208c4d68194844b7dfafbc00020d0dea 2025-11-03T17:07:27.5250777Z 2025-11-03 17:07:27,524 [INFO] cli.lib.common.git_helper: Successfully cloned vllm 2025-11-03T17:07:27.5654422Z 2025-11-03 17:07:27,564 [INFO] cli.lib.core.vllm.vllm_test: Running vllm test with inputs: VllmTestParameters(torch_whls_path=PosixPath('/var/lib/jenkins/workspace/dist'), vllm_whls_path=PosixPath('/var/lib/jenkins/workspace/dist/external/vllm/wheels'), torch_cuda_arch_list='8.9', cleaning_script=PosixPath('/var/lib/jenkins/workspace/.github/ci_configs/vllm/use_existing_torch.py')) 2025-11-03T17:07:27.5673214Z 2025-11-03 17:07:27,566 [INFO] cli.lib.common.pip_helper: torch already exist with version: 2.10.0a0+git3f6538f 2025-11-03T17:07:27.5675217Z 2025-11-03 17:07:27,567 [INFO] cli.lib.common.pip_helper: Installing /var/lib/jenkins/workspace/dist/vision/torchvision-0.25.0a0+cfbc5c2-cp312-cp312-linux_x86_64.whl... 2025-11-03T17:07:27.5676955Z 2025-11-03 17:07:27,567 [INFO] cli.lib.common.pip_helper: pip installing packages: /opt/conda/envs/py_3.12/bin/python -m pip install /var/lib/jenkins/workspace/dist/vision/torchvision-0.25.0a0+cfbc5c2-cp312-cp312-linux_x86_64.whl 2025-11-03T17:07:27.5678410Z 2025-11-03 17:07:27,567 [INFO] cli.lib.common.utils: [cmd] /opt/conda/envs/py_3.12/bin/python -m pip install /var/lib/jenkins/workspace/dist/vision/torchvision-0.25.0a0+cfbc5c2-cp312-cp312-linux_x86_64.whl 2025-11-03T17:07:27.8662414Z Processing /var/lib/jenkins/workspace/dist/vision/torchvision-0.25.0a0+cfbc5c2-cp312-cp312-linux_x86_64.whl 2025-11-03T17:07:27.8740401Z Requirement already satisfied: numpy in /opt/conda/envs/py_3.12/lib/python3.12/site-packages (from torchvision==0.25.0a0+cfbc5c2) (1.26.2) 2025-11-03T17:07:27.8742767Z Requirement already satisfied: torch in /opt/conda/envs/py_3.12/lib/python3.12/site-packages (from torchvision==0.25.0a0+cfbc5c2) (2.10.0a0+git3f6538f) 2025-11-03T17:07:27.8745207Z Requirement already satisfied: pillow!=8.3.*,>=5.3.0 in /opt/conda/envs/py_3.12/lib/python3.12/site-packages (from torchvision==0.25.0a0+cfbc5c2) (11.0.0) 2025-11-03T17:07:27.8796167Z Requirement already satisfied: filelock in /opt/conda/envs/py_3.12/lib/python3.12/site-packages (from torch->torchvision==0.25.0a0+cfbc5c2) (3.18.0) 2025-11-03T17:07:27.8799647Z Requirement already satisfied: typing-extensions>=4.10.0 in /opt/conda/envs/py_3.12/lib/python3.12/site-packages (from torch->torchvision==0.25.0a0+cfbc5c2) (4.12.2) 2025-11-03T17:07:27.8804248Z Requirement already satisfied: setuptools in /opt/conda/envs/py_3.12/lib/python3.12/site-packages (from torch->torchvision==0.25.0a0+cfbc5c2) (78.1.1) 2025-11-03T17:07:27.8806233Z Requirement already satisfied: sympy>=1.13.3 in /opt/conda/envs/py_3.12/lib/python3.12/site-packages (from torch->torchvision==0.25.0a0+cfbc5c2) (1.13.3) 2025-11-03T17:07:27.8815071Z Requirement already satisfied: networkx>=2.5.1 in /opt/conda/envs/py_3.12/lib/python3.12/site-packages (from torch->torchvision==0.25.0a0+cfbc5c2) (2.8.8) 2025-11-03T17:07:27.8817774Z Requirement already satisfied: jinja2 in /opt/conda/envs/py_3.12/lib/python3.12/site-packages (from torch->torchvision==0.25.0a0+cfbc5c2) (3.1.6) 2025-11-03T17:07:27.8821420Z Requirement already satisfied: fsspec>=0.8.5 in /opt/conda/envs/py_3.12/lib/python3.12/site-packages (from torch->torchvision==0.25.0a0+cfbc5c2) (2025.9.0) 2025-11-03T17:07:27.8922659Z Requirement already satisfied: mpmath<1.4,>=1.1.0 in /opt/conda/envs/py_3.12/lib/python3.12/site-packages (from sympy>=1.13.3->torch->torchvision==0.25.0a0+cfbc5c2) (1.3.0) 2025-11-03T17:07:27.8963953Z Requirement already satisfied: MarkupSafe>=2.0 in /opt/conda/envs/py_3.12/lib/python3.12/site-packages (from jinja2->torch->torchvision==0.25.0a0+cfbc5c2) (3.0.3) 2025-11-03T17:07:28.0594705Z Installing collected packages: torchvision 2025-11-03T17:07:28.4891144Z Successfully installed torchvision-0.25.0a0+cfbc5c2 2025-11-03T17:07:28.5383611Z 2025-11-03 17:07:28,537 [INFO] cli.lib.common.pip_helper: Installing /var/lib/jenkins/workspace/dist/audio/torchaudio-2.10.0a0+3b0e7a6-cp312-cp312-linux_x86_64.whl... 2025-11-03T17:07:28.5385954Z 2025-11-03 17:07:28,537 [INFO] cli.lib.common.pip_helper: pip installing packages: /opt/conda/envs/py_3.12/bin/python -m pip install /var/lib/jenkins/workspace/dist/audio/torchaudio-2.10.0a0+3b0e7a6-cp312-cp312-linux_x86_64.whl 2025-11-03T17:07:28.5387731Z 2025-11-03 17:07:28,538 [INFO] cli.lib.common.utils: [cmd] /opt/conda/envs/py_3.12/bin/python -m pip install /var/lib/jenkins/workspace/dist/audio/torchaudio-2.10.0a0+3b0e7a6-cp312-cp312-linux_x86_64.whl 2025-11-03T17:07:28.8380043Z Processing /var/lib/jenkins/workspace/dist/audio/torchaudio-2.10.0a0+3b0e7a6-cp312-cp312-linux_x86_64.whl 2025-11-03T17:07:28.8441708Z Requirement already satisfied: torch in /opt/conda/envs/py_3.12/lib/python3.12/site-packages (from torchaudio==2.10.0a0+3b0e7a6) (2.10.0a0+git3f6538f) 2025-11-03T17:07:28.8450472Z Requirement already satisfied: filelock in /opt/conda/envs/py_3.12/lib/python3.12/site-packages (from torch->torchaudio==2.10.0a0+3b0e7a6) (3.18.0) 2025-11-03T17:07:28.8454743Z Requirement already satisfied: typing-extensions>=4.10.0 in /opt/conda/envs/py_3.12/lib/python3.12/site-packages (from torch->torchaudio==2.10.0a0+3b0e7a6) (4.12.2) 2025-11-03T17:07:28.8460101Z Requirement already satisfied: setuptools in /opt/conda/envs/py_3.12/lib/python3.12/site-packages (from torch->torchaudio==2.10.0a0+3b0e7a6) (78.1.1) 2025-11-03T17:07:28.8463981Z Requirement already satisfied: sympy>=1.13.3 in /opt/conda/envs/py_3.12/lib/python3.12/site-packages (from torch->torchaudio==2.10.0a0+3b0e7a6) (1.13.3) 2025-11-03T17:07:28.8467579Z Requirement already satisfied: networkx>=2.5.1 in /opt/conda/envs/py_3.12/lib/python3.12/site-packages (from torch->torchaudio==2.10.0a0+3b0e7a6) (2.8.8) 2025-11-03T17:07:28.8473525Z Requirement already satisfied: jinja2 in /opt/conda/envs/py_3.12/lib/python3.12/site-packages (from torch->torchaudio==2.10.0a0+3b0e7a6) (3.1.6) 2025-11-03T17:07:28.8475231Z Requirement already satisfied: fsspec>=0.8.5 in /opt/conda/envs/py_3.12/lib/python3.12/site-packages (from torch->torchaudio==2.10.0a0+3b0e7a6) (2025.9.0) 2025-11-03T17:07:28.8583725Z Requirement already satisfied: mpmath<1.4,>=1.1.0 in /opt/conda/envs/py_3.12/lib/python3.12/site-packages (from sympy>=1.13.3->torch->torchaudio==2.10.0a0+3b0e7a6) (1.3.0) 2025-11-03T17:07:28.8624662Z Requirement already satisfied: MarkupSafe>=2.0 in /opt/conda/envs/py_3.12/lib/python3.12/site-packages (from jinja2->torch->torchaudio==2.10.0a0+3b0e7a6) (3.0.3) 2025-11-03T17:07:29.0300877Z Installing collected packages: torchaudio 2025-11-03T17:07:29.1797576Z Successfully installed torchaudio-2.10.0a0+3b0e7a6 2025-11-03T17:07:29.2265329Z 2025-11-03 17:07:29,226 [INFO] cli.lib.core.vllm.vllm_test: Done. Installed torch and other torch-related wheels 2025-11-03T17:07:29.2266021Z 2025-11-03 17:07:29,226 [INFO] cli.lib.core.vllm.vllm_test: Installing vllm wheels 2025-11-03T17:07:29.2267281Z 2025-11-03 17:07:29,226 [INFO] cli.lib.common.pip_helper: Installing /var/lib/jenkins/workspace/dist/external/vllm/wheels/xformers/xformers-0.0.33+5d4b92a5.d20251103-cp39-abi3-linux_x86_64.whl... 2025-11-03T17:07:29.2269226Z 2025-11-03 17:07:29,226 [INFO] cli.lib.common.pip_helper: pip installing packages: /opt/conda/envs/py_3.12/bin/python -m pip install /var/lib/jenkins/workspace/dist/external/vllm/wheels/xformers/xformers-0.0.33+5d4b92a5.d20251103-cp39-abi3-linux_x86_64.whl 2025-11-03T17:07:29.2271639Z 2025-11-03 17:07:29,226 [INFO] cli.lib.common.utils: [cmd] /opt/conda/envs/py_3.12/bin/python -m pip install /var/lib/jenkins/workspace/dist/external/vllm/wheels/xformers/xformers-0.0.33+5d4b92a5.d20251103-cp39-abi3-linux_x86_64.whl 2025-11-03T17:07:29.5261415Z Processing /var/lib/jenkins/workspace/dist/external/vllm/wheels/xformers/xformers-0.0.33+5d4b92a5.d20251103-cp39-abi3-linux_x86_64.whl 2025-11-03T17:07:29.7159053Z Requirement already satisfied: torch>=2.8 in /opt/conda/envs/py_3.12/lib/python3.12/site-packages (from xformers==0.0.33+5d4b92a5.d20251103) (2.10.0a0+git3f6538f) 2025-11-03T17:07:29.7161080Z Requirement already satisfied: numpy in /opt/conda/envs/py_3.12/lib/python3.12/site-packages (from xformers==0.0.33+5d4b92a5.d20251103) (1.26.2) 2025-11-03T17:07:29.7167791Z Requirement already satisfied: filelock in /opt/conda/envs/py_3.12/lib/python3.12/site-packages (from torch>=2.8->xformers==0.0.33+5d4b92a5.d20251103) (3.18.0) 2025-11-03T17:07:29.7172183Z Requirement already satisfied: typing-extensions>=4.10.0 in /opt/conda/envs/py_3.12/lib/python3.12/site-packages (from torch>=2.8->xformers==0.0.33+5d4b92a5.d20251103) (4.12.2) 2025-11-03T17:07:29.7177367Z Requirement already satisfied: setuptools in /opt/conda/envs/py_3.12/lib/python3.12/site-packages (from torch>=2.8->xformers==0.0.33+5d4b92a5.d20251103) (78.1.1) 2025-11-03T17:07:29.7181037Z Requirement already satisfied: sympy>=1.13.3 in /opt/conda/envs/py_3.12/lib/python3.12/site-packages (from torch>=2.8->xformers==0.0.33+5d4b92a5.d20251103) (1.13.3) 2025-11-03T17:07:29.7185179Z Requirement already satisfied: networkx>=2.5.1 in /opt/conda/envs/py_3.12/lib/python3.12/site-packages (from torch>=2.8->xformers==0.0.33+5d4b92a5.d20251103) (2.8.8) 2025-11-03T17:07:29.7188003Z Requirement already satisfied: jinja2 in /opt/conda/envs/py_3.12/lib/python3.12/site-packages (from torch>=2.8->xformers==0.0.33+5d4b92a5.d20251103) (3.1.6) 2025-11-03T17:07:29.7191657Z Requirement already satisfied: fsspec>=0.8.5 in /opt/conda/envs/py_3.12/lib/python3.12/site-packages (from torch>=2.8->xformers==0.0.33+5d4b92a5.d20251103) (2025.9.0) 2025-11-03T17:07:29.7300665Z Requirement already satisfied: mpmath<1.4,>=1.1.0 in /opt/conda/envs/py_3.12/lib/python3.12/site-packages (from sympy>=1.13.3->torch>=2.8->xformers==0.0.33+5d4b92a5.d20251103) (1.3.0) 2025-11-03T17:07:29.7343101Z Requirement already satisfied: MarkupSafe>=2.0 in /opt/conda/envs/py_3.12/lib/python3.12/site-packages (from jinja2->torch>=2.8->xformers==0.0.33+5d4b92a5.d20251103) (3.0.3) 2025-11-03T17:07:29.9016699Z Installing collected packages: xformers 2025-11-03T17:07:33.5442554Z Successfully installed xformers-0.0.33+5d4b92a5.d20251103 2025-11-03T17:07:33.5932474Z 2025-11-03 17:07:33,592 [INFO] cli.lib.common.pip_helper: Installing /var/lib/jenkins/workspace/dist/external/vllm/wheels/vllm/vllm-0.11.0rc2.dev389+ge51928192.d20251103-cp38-abi3-linux_x86_64.whl... 2025-11-03T17:07:33.5934157Z 2025-11-03 17:07:33,592 [INFO] cli.lib.common.pip_helper: pip installing packages: /opt/conda/envs/py_3.12/bin/python -m pip install /var/lib/jenkins/workspace/dist/external/vllm/wheels/vllm/vllm-0.11.0rc2.dev389+ge51928192.d20251103-cp38-abi3-linux_x86_64.whl 2025-11-03T17:07:33.5935901Z 2025-11-03 17:07:33,593 [INFO] cli.lib.common.utils: [cmd] /opt/conda/envs/py_3.12/bin/python -m pip install /var/lib/jenkins/workspace/dist/external/vllm/wheels/vllm/vllm-0.11.0rc2.dev389+ge51928192.d20251103-cp38-abi3-linux_x86_64.whl 2025-11-03T17:07:33.8969212Z Processing /var/lib/jenkins/workspace/dist/external/vllm/wheels/vllm/vllm-0.11.0rc2.dev389+ge51928192.d20251103-cp38-abi3-linux_x86_64.whl 2025-11-03T17:07:34.5094206Z Collecting regex (from vllm==0.11.0rc2.dev389+ge51928192.d20251103) 2025-11-03T17:07:34.5288251Z Downloading regex-2025.10.23-cp312-cp312-manylinux2014_x86_64.manylinux_2_17_x86_64.manylinux_2_28_x86_64.whl.metadata (40 kB) 2025-11-03T17:07:34.5369820Z Requirement already satisfied: cachetools in /opt/conda/envs/py_3.12/lib/python3.12/site-packages (from vllm==0.11.0rc2.dev389+ge51928192.d20251103) (6.2.1) 2025-11-03T17:07:34.5372366Z Requirement already satisfied: psutil in /opt/conda/envs/py_3.12/lib/python3.12/site-packages (from vllm==0.11.0rc2.dev389+ge51928192.d20251103) (7.1.2) 2025-11-03T17:07:34.5746198Z Collecting sentencepiece (from vllm==0.11.0rc2.dev389+ge51928192.d20251103) 2025-11-03T17:07:34.5803310Z Downloading sentencepiece-0.2.1-cp312-cp312-manylinux_2_27_x86_64.manylinux_2_28_x86_64.whl.metadata (10 kB) 2025-11-03T17:07:34.5868318Z Requirement already satisfied: numpy in /opt/conda/envs/py_3.12/lib/python3.12/site-packages (from vllm==0.11.0rc2.dev389+ge51928192.d20251103) (1.26.2) 2025-11-03T17:07:34.5870547Z Requirement already satisfied: requests>=2.26.0 in /opt/conda/envs/py_3.12/lib/python3.12/site-packages (from vllm==0.11.0rc2.dev389+ge51928192.d20251103) (2.32.5) 2025-11-03T17:07:34.5871922Z Requirement already satisfied: tqdm in /opt/conda/envs/py_3.12/lib/python3.12/site-packages (from vllm==0.11.0rc2.dev389+ge51928192.d20251103) (4.67.1) 2025-11-03T17:07:34.6377998Z Collecting blake3 (from vllm==0.11.0rc2.dev389+ge51928192.d20251103) 2025-11-03T17:07:34.6429142Z Downloading blake3-1.0.8-cp312-cp312-manylinux_2_17_x86_64.manylinux2014_x86_64.whl.metadata (4.6 kB) 2025-11-03T17:07:34.6574554Z Collecting py-cpuinfo (from vllm==0.11.0rc2.dev389+ge51928192.d20251103) 2025-11-03T17:07:34.6612412Z Downloading py_cpuinfo-9.0.0-py3-none-any.whl.metadata (794 bytes) 2025-11-03T17:07:34.6963110Z Collecting transformers>=4.55.2 (from vllm==0.11.0rc2.dev389+ge51928192.d20251103) 2025-11-03T17:07:34.7008005Z Downloading transformers-4.57.1-py3-none-any.whl.metadata (43 kB) 2025-11-03T17:07:34.9563419Z Collecting tokenizers>=0.21.1 (from vllm==0.11.0rc2.dev389+ge51928192.d20251103) 2025-11-03T17:07:34.9601022Z Downloading tokenizers-0.22.1-cp39-abi3-manylinux_2_17_x86_64.manylinux2014_x86_64.whl.metadata (6.8 kB) 2025-11-03T17:07:34.9667567Z Requirement already satisfied: protobuf in /opt/conda/envs/py_3.12/lib/python3.12/site-packages (from vllm==0.11.0rc2.dev389+ge51928192.d20251103) (5.29.5) 2025-11-03T17:07:35.0164737Z Collecting fastapi>=0.115.0 (from fastapi[standard]>=0.115.0->vllm==0.11.0rc2.dev389+ge51928192.d20251103) 2025-11-03T17:07:35.0204458Z Downloading fastapi-0.121.0-py3-none-any.whl.metadata (28 kB) 2025-11-03T17:07:35.0308365Z Requirement already satisfied: aiohttp in /opt/conda/envs/py_3.12/lib/python3.12/site-packages (from vllm==0.11.0rc2.dev389+ge51928192.d20251103) (3.13.2) 2025-11-03T17:07:35.0734535Z Collecting openai>=1.99.1 (from vllm==0.11.0rc2.dev389+ge51928192.d20251103) 2025-11-03T17:07:35.0771913Z Downloading openai-2.6.1-py3-none-any.whl.metadata (29 kB) 2025-11-03T17:07:35.1737675Z Collecting pydantic>=2.12.0 (from vllm==0.11.0rc2.dev389+ge51928192.d20251103) 2025-11-03T17:07:35.1770843Z Downloading pydantic-2.12.3-py3-none-any.whl.metadata (87 kB) 2025-11-03T17:07:35.1954748Z Collecting prometheus_client>=0.18.0 (from vllm==0.11.0rc2.dev389+ge51928192.d20251103) 2025-11-03T17:07:35.1990772Z Downloading prometheus_client-0.23.1-py3-none-any.whl.metadata (1.9 kB) 2025-11-03T17:07:35.2045242Z Requirement already satisfied: pillow in /opt/conda/envs/py_3.12/lib/python3.12/site-packages (from vllm==0.11.0rc2.dev389+ge51928192.d20251103) (11.0.0) 2025-11-03T17:07:35.2167902Z Collecting prometheus-fastapi-instrumentator>=7.0.0 (from vllm==0.11.0rc2.dev389+ge51928192.d20251103) 2025-11-03T17:07:35.2217377Z Downloading prometheus_fastapi_instrumentator-7.1.0-py3-none-any.whl.metadata (13 kB) 2025-11-03T17:07:35.2579865Z Collecting tiktoken>=0.6.0 (from vllm==0.11.0rc2.dev389+ge51928192.d20251103) 2025-11-03T17:07:35.2620882Z Downloading tiktoken-0.12.0-cp312-cp312-manylinux_2_28_x86_64.whl.metadata (6.7 kB) 2025-11-03T17:07:35.2826685Z Collecting lm-format-enforcer==0.11.3 (from vllm==0.11.0rc2.dev389+ge51928192.d20251103) 2025-11-03T17:07:35.2883180Z Downloading lm_format_enforcer-0.11.3-py3-none-any.whl.metadata (17 kB) 2025-11-03T17:07:35.3296654Z Collecting llguidance<0.8.0,>=0.7.11 (from vllm==0.11.0rc2.dev389+ge51928192.d20251103) 2025-11-03T17:07:35.3347981Z Downloading llguidance-0.7.30-cp39-abi3-manylinux_2_17_x86_64.manylinux2014_x86_64.whl.metadata (10 kB) 2025-11-03T17:07:35.3643494Z Collecting outlines_core==0.2.11 (from vllm==0.11.0rc2.dev389+ge51928192.d20251103) 2025-11-03T17:07:35.3720220Z Downloading outlines_core-0.2.11-cp312-cp312-manylinux_2_17_x86_64.manylinux2014_x86_64.whl.metadata (5.8 kB) 2025-11-03T17:07:35.3945448Z Collecting diskcache==5.6.3 (from vllm==0.11.0rc2.dev389+ge51928192.d20251103) 2025-11-03T17:07:35.3989218Z Downloading diskcache-5.6.3-py3-none-any.whl.metadata (20 kB) 2025-11-03T17:07:35.4142075Z Collecting lark==1.2.2 (from vllm==0.11.0rc2.dev389+ge51928192.d20251103) 2025-11-03T17:07:35.4201606Z Downloading lark-1.2.2-py3-none-any.whl.metadata (1.8 kB) 2025-11-03T17:07:35.4569667Z Collecting xgrammar==0.1.25 (from vllm==0.11.0rc2.dev389+ge51928192.d20251103) 2025-11-03T17:07:35.4621005Z Downloading xgrammar-0.1.25-cp312-cp312-manylinux_2_17_x86_64.manylinux2014_x86_64.whl.metadata (6.5 kB) 2025-11-03T17:07:35.4686091Z Requirement already satisfied: typing_extensions>=4.10 in /opt/conda/envs/py_3.12/lib/python3.12/site-packages (from vllm==0.11.0rc2.dev389+ge51928192.d20251103) (4.12.2) 2025-11-03T17:07:35.4688284Z Requirement already satisfied: filelock>=3.16.1 in /opt/conda/envs/py_3.12/lib/python3.12/site-packages (from vllm==0.11.0rc2.dev389+ge51928192.d20251103) (3.18.0) 2025-11-03T17:07:35.4792606Z Collecting partial-json-parser (from vllm==0.11.0rc2.dev389+ge51928192.d20251103) 2025-11-03T17:07:35.4849715Z Downloading partial_json_parser-0.2.1.1.post6-py3-none-any.whl.metadata (6.1 kB) 2025-11-03T17:07:35.6765164Z Collecting pyzmq>=25.0.0 (from vllm==0.11.0rc2.dev389+ge51928192.d20251103) 2025-11-03T17:07:35.6810108Z Downloading pyzmq-27.1.0-cp312-abi3-manylinux_2_26_x86_64.manylinux_2_28_x86_64.whl.metadata (6.0 kB) 2025-11-03T17:07:35.7287816Z Collecting msgspec (from vllm==0.11.0rc2.dev389+ge51928192.d20251103) 2025-11-03T17:07:35.7338586Z Downloading msgspec-0.19.0-cp312-cp312-manylinux_2_17_x86_64.manylinux2014_x86_64.whl.metadata (6.9 kB) 2025-11-03T17:07:35.7526188Z Collecting gguf>=0.13.0 (from vllm==0.11.0rc2.dev389+ge51928192.d20251103) 2025-11-03T17:07:35.7577732Z Downloading gguf-0.17.1-py3-none-any.whl.metadata (4.3 kB) 2025-11-03T17:07:35.8041749Z Collecting mistral_common>=1.8.5 (from mistral_common[audio,image]>=1.8.5->vllm==0.11.0rc2.dev389+ge51928192.d20251103) 2025-11-03T17:07:35.8120208Z Downloading mistral_common-1.8.5-py3-none-any.whl.metadata (5.1 kB) 2025-11-03T17:07:35.8670802Z Collecting opencv-python-headless>=4.11.0 (from vllm==0.11.0rc2.dev389+ge51928192.d20251103) 2025-11-03T17:07:35.8710608Z Downloading opencv_python_headless-4.12.0.88-cp37-abi3-manylinux2014_x86_64.manylinux_2_17_x86_64.whl.metadata (19 kB) 2025-11-03T17:07:35.8766961Z Requirement already satisfied: pyyaml in /opt/conda/envs/py_3.12/lib/python3.12/site-packages (from vllm==0.11.0rc2.dev389+ge51928192.d20251103) (6.0.2) 2025-11-03T17:07:35.8771713Z Requirement already satisfied: six>=1.16.0 in /opt/conda/envs/py_3.12/lib/python3.12/site-packages (from vllm==0.11.0rc2.dev389+ge51928192.d20251103) (1.17.0) 2025-11-03T17:07:35.8775163Z Requirement already satisfied: setuptools<80,>=77.0.3 in /opt/conda/envs/py_3.12/lib/python3.12/site-packages (from vllm==0.11.0rc2.dev389+ge51928192.d20251103) (78.1.1) 2025-11-03T17:07:35.8879769Z Collecting einops (from vllm==0.11.0rc2.dev389+ge51928192.d20251103) 2025-11-03T17:07:35.8918794Z Downloading einops-0.8.1-py3-none-any.whl.metadata (13 kB) 2025-11-03T17:07:35.9154351Z Collecting compressed-tensors==0.11.0 (from vllm==0.11.0rc2.dev389+ge51928192.d20251103) 2025-11-03T17:07:35.9207125Z Downloading compressed_tensors-0.11.0-py3-none-any.whl.metadata (7.0 kB) 2025-11-03T17:07:35.9361277Z Collecting depyf==0.19.0 (from vllm==0.11.0rc2.dev389+ge51928192.d20251103) 2025-11-03T17:07:35.9418095Z Downloading depyf-0.19.0-py3-none-any.whl.metadata (7.3 kB) 2025-11-03T17:07:35.9611190Z Collecting cloudpickle (from vllm==0.11.0rc2.dev389+ge51928192.d20251103) 2025-11-03T17:07:35.9644363Z Downloading cloudpickle-3.1.2-py3-none-any.whl.metadata (7.1 kB) 2025-11-03T17:07:36.0279610Z Collecting watchfiles (from vllm==0.11.0rc2.dev389+ge51928192.d20251103) 2025-11-03T17:07:36.0321724Z Downloading watchfiles-1.1.1-cp312-cp312-manylinux_2_17_x86_64.manylinux2014_x86_64.whl.metadata (4.9 kB) 2025-11-03T17:07:36.0495279Z Collecting python-json-logger (from vllm==0.11.0rc2.dev389+ge51928192.d20251103) 2025-11-03T17:07:36.0539250Z Downloading python_json_logger-4.0.0-py3-none-any.whl.metadata (4.0 kB) 2025-11-03T17:07:36.0607855Z Requirement already satisfied: scipy in /opt/conda/envs/py_3.12/lib/python3.12/site-packages (from vllm==0.11.0rc2.dev389+ge51928192.d20251103) (1.14.1) 2025-11-03T17:07:36.0610489Z Requirement already satisfied: ninja in /opt/conda/envs/py_3.12/lib/python3.12/site-packages (from vllm==0.11.0rc2.dev389+ge51928192.d20251103) (1.10.2) 2025-11-03T17:07:36.1306853Z Collecting pybase64 (from vllm==0.11.0rc2.dev389+ge51928192.d20251103) 2025-11-03T17:07:36.1360707Z Downloading pybase64-1.4.2-cp312-cp312-manylinux1_x86_64.manylinux2014_x86_64.manylinux_2_17_x86_64.manylinux_2_5_x86_64.whl.metadata (8.7 kB) 2025-11-03T17:07:36.1746397Z Collecting cbor2 (from vllm==0.11.0rc2.dev389+ge51928192.d20251103) 2025-11-03T17:07:36.1789854Z Downloading cbor2-5.7.1-cp312-cp312-manylinux2014_x86_64.manylinux_2_17_x86_64.manylinux_2_28_x86_64.whl.metadata (5.4 kB) 2025-11-03T17:07:36.2236113Z Collecting setproctitle (from vllm==0.11.0rc2.dev389+ge51928192.d20251103) 2025-11-03T17:07:36.2289502Z Downloading setproctitle-1.3.7-cp312-cp312-manylinux1_x86_64.manylinux_2_28_x86_64.manylinux_2_5_x86_64.whl.metadata (10 kB) 2025-11-03T17:07:36.2481944Z Collecting openai-harmony>=0.0.3 (from vllm==0.11.0rc2.dev389+ge51928192.d20251103) 2025-11-03T17:07:36.2529792Z Downloading openai_harmony-0.0.4-cp38-abi3-manylinux_2_17_x86_64.manylinux2014_x86_64.whl.metadata (8.0 kB) 2025-11-03T17:07:36.3274682Z Collecting numba==0.61.2 (from vllm==0.11.0rc2.dev389+ge51928192.d20251103) 2025-11-03T17:07:36.3328105Z Downloading numba-0.61.2-cp312-cp312-manylinux2014_x86_64.manylinux_2_17_x86_64.whl.metadata (2.8 kB) 2025-11-03T17:07:36.4269417Z Collecting ray>=2.48.0 (from ray[cgraph]>=2.48.0->vllm==0.11.0rc2.dev389+ge51928192.d20251103) 2025-11-03T17:07:36.4311079Z Downloading ray-2.51.1-cp312-cp312-manylinux2014_x86_64.whl.metadata (21 kB) 2025-11-03T17:07:36.4943374Z Collecting flashinfer-python==0.4.0 (from vllm==0.11.0rc2.dev389+ge51928192.d20251103) 2025-11-03T17:07:36.5027591Z Downloading flashinfer_python-0.4.0.tar.gz (4.5 MB) 2025-11-03T17:07:36.5373978Z [?25l ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 0.0/4.5 MB ? eta -:--:-- 2025-11-03T17:07:36.5375299Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 4.5/4.5 MB 136.3 MB/s 0:00:00 2025-11-03T17:07:38.2347221Z [?25h Installing build dependencies ... [?25l- \ | / done 2025-11-03T17:07:38.8102857Z [?25h Getting requirements to build wheel ... [?25l- done 2025-11-03T17:07:39.6181881Z [?25h Preparing metadata (pyproject.toml) ... [?25l- done 2025-11-03T17:07:39.6210359Z [?25hRequirement already satisfied: torch>=1.7.0 in /opt/conda/envs/py_3.12/lib/python3.12/site-packages (from compressed-tensors==0.11.0->vllm==0.11.0rc2.dev389+ge51928192.d20251103) (2.10.0a0+git3f6538f) 2025-11-03T17:07:39.7065977Z Collecting frozendict (from compressed-tensors==0.11.0->vllm==0.11.0rc2.dev389+ge51928192.d20251103) 2025-11-03T17:07:39.7117770Z Downloading frozendict-2.4.6-py312-none-any.whl.metadata (23 kB) 2025-11-03T17:07:39.7265530Z Collecting astor (from depyf==0.19.0->vllm==0.11.0rc2.dev389+ge51928192.d20251103) 2025-11-03T17:07:39.7302669Z Downloading astor-0.8.1-py2.py3-none-any.whl.metadata (4.2 kB) 2025-11-03T17:07:39.7347667Z Requirement already satisfied: dill in /opt/conda/envs/py_3.12/lib/python3.12/site-packages (from depyf==0.19.0->vllm==0.11.0rc2.dev389+ge51928192.d20251103) (0.3.7) 2025-11-03T17:07:39.7599065Z Collecting apache-tvm-ffi==0.1.0b15 (from flashinfer-python==0.4.0->vllm==0.11.0rc2.dev389+ge51928192.d20251103) 2025-11-03T17:07:39.7606756Z Using cached apache_tvm_ffi-0.1.0b15-cp312-abi3-manylinux_2_24_x86_64.manylinux_2_28_x86_64.whl.metadata (2.0 kB) 2025-11-03T17:07:39.7626321Z Requirement already satisfied: click in /opt/conda/envs/py_3.12/lib/python3.12/site-packages (from flashinfer-python==0.4.0->vllm==0.11.0rc2.dev389+ge51928192.d20251103) (8.3.0) 2025-11-03T17:07:39.7789001Z Collecting nvidia-cudnn-frontend>=1.13.0 (from flashinfer-python==0.4.0->vllm==0.11.0rc2.dev389+ge51928192.d20251103) 2025-11-03T17:07:39.7832585Z Downloading nvidia_cudnn_frontend-1.15.0-cp312-cp312-manylinux_2_27_x86_64.manylinux_2_28_x86_64.whl.metadata (6.7 kB) 2025-11-03T17:07:39.8021423Z Collecting nvidia-cutlass-dsl>=4.2.1 (from flashinfer-python==0.4.0->vllm==0.11.0rc2.dev389+ge51928192.d20251103) 2025-11-03T17:07:39.8072415Z Downloading nvidia_cutlass_dsl-4.2.1-cp312-cp312-manylinux_2_28_x86_64.whl.metadata (2.5 kB) 2025-11-03T17:07:39.8225799Z Collecting nvidia-ml-py (from flashinfer-python==0.4.0->vllm==0.11.0rc2.dev389+ge51928192.d20251103) 2025-11-03T17:07:39.8257911Z Downloading nvidia_ml_py-13.580.82-py3-none-any.whl.metadata (9.6 kB) 2025-11-03T17:07:39.8409274Z Collecting packaging>=24.2 (from flashinfer-python==0.4.0->vllm==0.11.0rc2.dev389+ge51928192.d20251103) 2025-11-03T17:07:39.8417487Z Using cached packaging-25.0-py3-none-any.whl.metadata (3.3 kB) 2025-11-03T17:07:39.8432478Z Requirement already satisfied: tabulate in /opt/conda/envs/py_3.12/lib/python3.12/site-packages (from flashinfer-python==0.4.0->vllm==0.11.0rc2.dev389+ge51928192.d20251103) (0.9.0) 2025-11-03T17:07:39.8539322Z Collecting interegular>=0.3.2 (from lm-format-enforcer==0.11.3->vllm==0.11.0rc2.dev389+ge51928192.d20251103) 2025-11-03T17:07:39.8584736Z Downloading interegular-0.3.3-py37-none-any.whl.metadata (3.0 kB) 2025-11-03T17:07:39.9170260Z Collecting llvmlite<0.45,>=0.44.0dev0 (from numba==0.61.2->vllm==0.11.0rc2.dev389+ge51928192.d20251103) 2025-11-03T17:07:39.9223151Z Downloading llvmlite-0.44.0-cp312-cp312-manylinux_2_17_x86_64.manylinux2014_x86_64.whl.metadata (5.0 kB) 2025-11-03T17:07:39.9360847Z Requirement already satisfied: triton in /opt/conda/envs/py_3.12/lib/python3.12/site-packages (from xgrammar==0.1.25->vllm==0.11.0rc2.dev389+ge51928192.d20251103) (3.5.0+git7416ffcb) 2025-11-03T17:07:39.9668353Z Collecting starlette<0.50.0,>=0.40.0 (from fastapi>=0.115.0->fastapi[standard]>=0.115.0->vllm==0.11.0rc2.dev389+ge51928192.d20251103) 2025-11-03T17:07:39.9699842Z Downloading starlette-0.49.3-py3-none-any.whl.metadata (6.4 kB) 2025-11-03T17:07:39.9967797Z Collecting annotated-doc>=0.0.2 (from fastapi>=0.115.0->fastapi[standard]>=0.115.0->vllm==0.11.0rc2.dev389+ge51928192.d20251103) 2025-11-03T17:07:40.0004594Z Downloading annotated_doc-0.0.3-py3-none-any.whl.metadata (6.6 kB) 2025-11-03T17:07:40.0139185Z Collecting annotated-types>=0.6.0 (from pydantic>=2.12.0->vllm==0.11.0rc2.dev389+ge51928192.d20251103) 2025-11-03T17:07:40.0172653Z Downloading annotated_types-0.7.0-py3-none-any.whl.metadata (15 kB) 2025-11-03T17:07:40.5918203Z Collecting pydantic-core==2.41.4 (from pydantic>=2.12.0->vllm==0.11.0rc2.dev389+ge51928192.d20251103) 2025-11-03T17:07:40.5949761Z Downloading pydantic_core-2.41.4-cp312-cp312-manylinux_2_17_x86_64.manylinux2014_x86_64.whl.metadata (7.3 kB) 2025-11-03T17:07:40.6120108Z Collecting typing_extensions>=4.10 (from vllm==0.11.0rc2.dev389+ge51928192.d20251103) 2025-11-03T17:07:40.6130258Z Using cached typing_extensions-4.15.0-py3-none-any.whl.metadata (3.3 kB) 2025-11-03T17:07:40.6233750Z Collecting typing-inspection>=0.4.2 (from pydantic>=2.12.0->vllm==0.11.0rc2.dev389+ge51928192.d20251103) 2025-11-03T17:07:40.6263488Z Downloading typing_inspection-0.4.2-py3-none-any.whl.metadata (2.6 kB) 2025-11-03T17:07:40.6487332Z Collecting anyio<5,>=3.6.2 (from starlette<0.50.0,>=0.40.0->fastapi>=0.115.0->fastapi[standard]>=0.115.0->vllm==0.11.0rc2.dev389+ge51928192.d20251103) 2025-11-03T17:07:40.6518338Z Downloading anyio-4.11.0-py3-none-any.whl.metadata (4.1 kB) 2025-11-03T17:07:40.6580116Z Requirement already satisfied: idna>=2.8 in /opt/conda/envs/py_3.12/lib/python3.12/site-packages (from anyio<5,>=3.6.2->starlette<0.50.0,>=0.40.0->fastapi>=0.115.0->fastapi[standard]>=0.115.0->vllm==0.11.0rc2.dev389+ge51928192.d20251103) (3.11) 2025-11-03T17:07:40.6655747Z Collecting sniffio>=1.1 (from anyio<5,>=3.6.2->starlette<0.50.0,>=0.40.0->fastapi>=0.115.0->fastapi[standard]>=0.115.0->vllm==0.11.0rc2.dev389+ge51928192.d20251103) 2025-11-03T17:07:40.6690540Z Downloading sniffio-1.3.1-py3-none-any.whl.metadata (3.9 kB) 2025-11-03T17:07:40.6855781Z Collecting fastapi-cli>=0.0.8 (from fastapi-cli[standard]>=0.0.8; extra == "standard"->fastapi[standard]>=0.115.0->vllm==0.11.0rc2.dev389+ge51928192.d20251103) 2025-11-03T17:07:40.6891399Z Downloading fastapi_cli-0.0.14-py3-none-any.whl.metadata (6.4 kB) 2025-11-03T17:07:40.7114526Z Collecting httpx<1.0.0,>=0.23.0 (from fastapi[standard]>=0.115.0->vllm==0.11.0rc2.dev389+ge51928192.d20251103) 2025-11-03T17:07:40.7145168Z Downloading httpx-0.28.1-py3-none-any.whl.metadata (7.1 kB) 2025-11-03T17:07:40.7207353Z Requirement already satisfied: jinja2>=3.1.5 in /opt/conda/envs/py_3.12/lib/python3.12/site-packages (from fastapi[standard]>=0.115.0->vllm==0.11.0rc2.dev389+ge51928192.d20251103) (3.1.6) 2025-11-03T17:07:40.7296923Z Collecting python-multipart>=0.0.18 (from fastapi[standard]>=0.115.0->vllm==0.11.0rc2.dev389+ge51928192.d20251103) 2025-11-03T17:07:40.7326932Z Downloading python_multipart-0.0.20-py3-none-any.whl.metadata (1.8 kB) 2025-11-03T17:07:40.7475630Z Collecting email-validator>=2.0.0 (from fastapi[standard]>=0.115.0->vllm==0.11.0rc2.dev389+ge51928192.d20251103) 2025-11-03T17:07:40.7701044Z Downloading email_validator-2.3.0-py3-none-any.whl.metadata (26 kB) 2025-11-03T17:07:40.7969789Z Collecting uvicorn>=0.12.0 (from uvicorn[standard]>=0.12.0; extra == "standard"->fastapi[standard]>=0.115.0->vllm==0.11.0rc2.dev389+ge51928192.d20251103) 2025-11-03T17:07:40.7999530Z Downloading uvicorn-0.38.0-py3-none-any.whl.metadata (6.8 kB) 2025-11-03T17:07:40.8070865Z Requirement already satisfied: certifi in /opt/conda/envs/py_3.12/lib/python3.12/site-packages (from httpx<1.0.0,>=0.23.0->fastapi[standard]>=0.115.0->vllm==0.11.0rc2.dev389+ge51928192.d20251103) (2025.10.5) 2025-11-03T17:07:40.8243224Z Collecting httpcore==1.* (from httpx<1.0.0,>=0.23.0->fastapi[standard]>=0.115.0->vllm==0.11.0rc2.dev389+ge51928192.d20251103) 2025-11-03T17:07:40.8281258Z Downloading httpcore-1.0.9-py3-none-any.whl.metadata (21 kB) 2025-11-03T17:07:40.8437435Z Collecting h11>=0.16 (from httpcore==1.*->httpx<1.0.0,>=0.23.0->fastapi[standard]>=0.115.0->vllm==0.11.0rc2.dev389+ge51928192.d20251103) 2025-11-03T17:07:40.8467748Z Downloading h11-0.16.0-py3-none-any.whl.metadata (8.3 kB) 2025-11-03T17:07:40.8619351Z Collecting dnspython>=2.0.0 (from email-validator>=2.0.0->fastapi[standard]>=0.115.0->vllm==0.11.0rc2.dev389+ge51928192.d20251103) 2025-11-03T17:07:40.8656707Z Downloading dnspython-2.8.0-py3-none-any.whl.metadata (5.7 kB) 2025-11-03T17:07:40.8902298Z Collecting typer>=0.15.1 (from fastapi-cli>=0.0.8->fastapi-cli[standard]>=0.0.8; extra == "standard"->fastapi[standard]>=0.115.0->vllm==0.11.0rc2.dev389+ge51928192.d20251103) 2025-11-03T17:07:40.8940050Z Downloading typer-0.20.0-py3-none-any.whl.metadata (16 kB) 2025-11-03T17:07:40.9187653Z Collecting rich-toolkit>=0.14.8 (from fastapi-cli>=0.0.8->fastapi-cli[standard]>=0.0.8; extra == "standard"->fastapi[standard]>=0.115.0->vllm==0.11.0rc2.dev389+ge51928192.d20251103) 2025-11-03T17:07:40.9224444Z Downloading rich_toolkit-0.15.1-py3-none-any.whl.metadata (1.0 kB) 2025-11-03T17:07:40.9362027Z Collecting fastapi-cloud-cli>=0.1.1 (from fastapi-cli[standard]>=0.0.8; extra == "standard"->fastapi[standard]>=0.115.0->vllm==0.11.0rc2.dev389+ge51928192.d20251103) 2025-11-03T17:07:40.9395823Z Downloading fastapi_cloud_cli-0.3.1-py3-none-any.whl.metadata (3.2 kB) 2025-11-03T17:07:41.0250892Z Collecting rignore>=0.5.1 (from fastapi-cloud-cli>=0.1.1->fastapi-cli[standard]>=0.0.8; extra == "standard"->fastapi[standard]>=0.115.0->vllm==0.11.0rc2.dev389+ge51928192.d20251103) 2025-11-03T17:07:41.0304173Z Downloading rignore-0.7.4-cp312-cp312-manylinux_2_17_x86_64.manylinux2014_x86_64.whl.metadata (4.2 kB) 2025-11-03T17:07:41.0982285Z Collecting sentry-sdk>=2.20.0 (from fastapi-cloud-cli>=0.1.1->fastapi-cli[standard]>=0.0.8; extra == "standard"->fastapi[standard]>=0.115.0->vllm==0.11.0rc2.dev389+ge51928192.d20251103) 2025-11-03T17:07:41.1022695Z Downloading sentry_sdk-2.43.0-py2.py3-none-any.whl.metadata (10 kB) 2025-11-03T17:07:41.1262612Z Requirement already satisfied: MarkupSafe>=2.0 in /opt/conda/envs/py_3.12/lib/python3.12/site-packages (from jinja2>=3.1.5->fastapi[standard]>=0.115.0->vllm==0.11.0rc2.dev389+ge51928192.d20251103) (3.0.3) 2025-11-03T17:07:41.1573169Z Collecting jsonschema>=4.21.1 (from mistral_common>=1.8.5->mistral_common[audio,image]>=1.8.5->vllm==0.11.0rc2.dev389+ge51928192.d20251103) 2025-11-03T17:07:41.1610653Z Downloading jsonschema-4.25.1-py3-none-any.whl.metadata (7.6 kB) 2025-11-03T17:07:41.1816540Z Collecting pydantic-extra-types>=2.10.5 (from pydantic-extra-types[pycountry]>=2.10.5->mistral_common>=1.8.5->mistral_common[audio,image]>=1.8.5->vllm==0.11.0rc2.dev389+ge51928192.d20251103) 2025-11-03T17:07:41.1853418Z Downloading pydantic_extra_types-2.10.6-py3-none-any.whl.metadata (4.0 kB) 2025-11-03T17:07:41.1941164Z Requirement already satisfied: attrs>=22.2.0 in /opt/conda/envs/py_3.12/lib/python3.12/site-packages (from jsonschema>=4.21.1->mistral_common>=1.8.5->mistral_common[audio,image]>=1.8.5->vllm==0.11.0rc2.dev389+ge51928192.d20251103) (25.4.0) 2025-11-03T17:07:41.2056964Z Collecting jsonschema-specifications>=2023.03.6 (from jsonschema>=4.21.1->mistral_common>=1.8.5->mistral_common[audio,image]>=1.8.5->vllm==0.11.0rc2.dev389+ge51928192.d20251103) 2025-11-03T17:07:41.2094298Z Downloading jsonschema_specifications-2025.9.1-py3-none-any.whl.metadata (2.9 kB) 2025-11-03T17:07:41.2319403Z Collecting referencing>=0.28.4 (from jsonschema>=4.21.1->mistral_common>=1.8.5->mistral_common[audio,image]>=1.8.5->vllm==0.11.0rc2.dev389+ge51928192.d20251103) 2025-11-03T17:07:41.2349827Z Downloading referencing-0.37.0-py3-none-any.whl.metadata (2.8 kB) 2025-11-03T17:07:41.5099102Z Collecting rpds-py>=0.7.1 (from jsonschema>=4.21.1->mistral_common>=1.8.5->mistral_common[audio,image]>=1.8.5->vllm==0.11.0rc2.dev389+ge51928192.d20251103) 2025-11-03T17:07:41.5143731Z Downloading rpds_py-0.28.0-cp312-cp312-manylinux_2_17_x86_64.manylinux2014_x86_64.whl.metadata (4.1 kB) 2025-11-03T17:07:41.5554944Z Collecting cuda-python>=12.8 (from nvidia-cutlass-dsl>=4.2.1->flashinfer-python==0.4.0->vllm==0.11.0rc2.dev389+ge51928192.d20251103) 2025-11-03T17:07:41.5603707Z Downloading cuda_python-13.0.3-py3-none-any.whl.metadata (4.7 kB) 2025-11-03T17:07:41.5866217Z Collecting cuda-bindings~=13.0.3 (from cuda-python>=12.8->nvidia-cutlass-dsl>=4.2.1->flashinfer-python==0.4.0->vllm==0.11.0rc2.dev389+ge51928192.d20251103) 2025-11-03T17:07:41.5917904Z Downloading cuda_bindings-13.0.3-cp312-cp312-manylinux_2_24_x86_64.manylinux_2_28_x86_64.whl.metadata (2.3 kB) 2025-11-03T17:07:41.5968948Z Requirement already satisfied: cuda-pathfinder~=1.1 in /opt/conda/envs/py_3.12/lib/python3.12/site-packages (from cuda-python>=12.8->nvidia-cutlass-dsl>=4.2.1->flashinfer-python==0.4.0->vllm==0.11.0rc2.dev389+ge51928192.d20251103) (1.3.2) 2025-11-03T17:07:41.6050994Z Requirement already satisfied: distro<2,>=1.7.0 in /opt/conda/envs/py_3.12/lib/python3.12/site-packages (from openai>=1.99.1->vllm==0.11.0rc2.dev389+ge51928192.d20251103) (1.9.0) 2025-11-03T17:07:41.6560430Z Collecting jiter<1,>=0.10.0 (from openai>=1.99.1->vllm==0.11.0rc2.dev389+ge51928192.d20251103) 2025-11-03T17:07:41.6599233Z Downloading jiter-0.11.1-cp312-cp312-manylinux_2_17_x86_64.manylinux2014_x86_64.whl.metadata (5.2 kB) 2025-11-03T17:07:41.8265958Z Collecting numpy (from vllm==0.11.0rc2.dev389+ge51928192.d20251103) 2025-11-03T17:07:41.8304660Z Downloading numpy-2.2.6-cp312-cp312-manylinux_2_17_x86_64.manylinux2014_x86_64.whl.metadata (62 kB) 2025-11-03T17:07:41.8873575Z Collecting pycountry>=23 (from pydantic-extra-types[pycountry]>=2.10.5->mistral_common>=1.8.5->mistral_common[audio,image]>=1.8.5->vllm==0.11.0rc2.dev389+ge51928192.d20251103) 2025-11-03T17:07:41.8903371Z Downloading pycountry-24.6.1-py3-none-any.whl.metadata (12 kB) 2025-11-03T17:07:41.9192118Z Collecting click (from flashinfer-python==0.4.0->vllm==0.11.0rc2.dev389+ge51928192.d20251103) 2025-11-03T17:07:41.9221233Z Downloading click-8.2.1-py3-none-any.whl.metadata (2.5 kB) 2025-11-03T17:07:41.9276172Z Requirement already satisfied: msgpack<2.0.0,>=1.0.0 in /opt/conda/envs/py_3.12/lib/python3.12/site-packages (from ray>=2.48.0->ray[cgraph]>=2.48.0->vllm==0.11.0rc2.dev389+ge51928192.d20251103) (1.1.2) 2025-11-03T17:07:41.9514966Z Collecting cupy-cuda12x (from ray[cgraph]>=2.48.0->vllm==0.11.0rc2.dev389+ge51928192.d20251103) 2025-11-03T17:07:41.9746566Z Downloading cupy_cuda12x-13.6.0-cp312-cp312-manylinux2014_x86_64.whl.metadata (2.4 kB) 2025-11-03T17:07:41.9952075Z Requirement already satisfied: charset_normalizer<4,>=2 in /opt/conda/envs/py_3.12/lib/python3.12/site-packages (from requests>=2.26.0->vllm==0.11.0rc2.dev389+ge51928192.d20251103) (3.4.4) 2025-11-03T17:07:41.9956963Z Requirement already satisfied: urllib3<3,>=1.21.1 in /opt/conda/envs/py_3.12/lib/python3.12/site-packages (from requests>=2.26.0->vllm==0.11.0rc2.dev389+ge51928192.d20251103) (2.5.0) 2025-11-03T17:07:42.0345355Z Collecting rich>=13.7.1 (from rich-toolkit>=0.14.8->fastapi-cli>=0.0.8->fastapi-cli[standard]>=0.0.8; extra == "standard"->fastapi[standard]>=0.115.0->vllm==0.11.0rc2.dev389+ge51928192.d20251103) 2025-11-03T17:07:42.0384482Z Downloading rich-14.2.0-py3-none-any.whl.metadata (18 kB) 2025-11-03T17:07:42.0623488Z Collecting markdown-it-py>=2.2.0 (from rich>=13.7.1->rich-toolkit>=0.14.8->fastapi-cli>=0.0.8->fastapi-cli[standard]>=0.0.8; extra == "standard"->fastapi[standard]>=0.115.0->vllm==0.11.0rc2.dev389+ge51928192.d20251103) 2025-11-03T17:07:42.0674621Z Downloading markdown_it_py-4.0.0-py3-none-any.whl.metadata (7.3 kB) 2025-11-03T17:07:42.0765119Z Requirement already satisfied: pygments<3.0.0,>=2.13.0 in /opt/conda/envs/py_3.12/lib/python3.12/site-packages (from rich>=13.7.1->rich-toolkit>=0.14.8->fastapi-cli>=0.0.8->fastapi-cli[standard]>=0.0.8; extra == "standard"->fastapi[standard]>=0.115.0->vllm==0.11.0rc2.dev389+ge51928192.d20251103) (2.15.0) 2025-11-03T17:07:42.0884325Z Collecting mdurl~=0.1 (from markdown-it-py>=2.2.0->rich>=13.7.1->rich-toolkit>=0.14.8->fastapi-cli>=0.0.8->fastapi-cli[standard]>=0.0.8; extra == "standard"->fastapi[standard]>=0.115.0->vllm==0.11.0rc2.dev389+ge51928192.d20251103) 2025-11-03T17:07:42.0919941Z Downloading mdurl-0.1.2-py3-none-any.whl.metadata (1.6 kB) 2025-11-03T17:07:42.1547222Z Collecting huggingface-hub<2.0,>=0.16.4 (from tokenizers>=0.21.1->vllm==0.11.0rc2.dev389+ge51928192.d20251103) 2025-11-03T17:07:42.1585291Z Downloading huggingface_hub-1.0.1-py3-none-any.whl.metadata (13 kB) 2025-11-03T17:07:42.1745457Z Requirement already satisfied: fsspec>=2023.5.0 in /opt/conda/envs/py_3.12/lib/python3.12/site-packages (from huggingface-hub<2.0,>=0.16.4->tokenizers>=0.21.1->vllm==0.11.0rc2.dev389+ge51928192.d20251103) (2025.9.0) 2025-11-03T17:07:42.1888625Z Collecting shellingham (from huggingface-hub<2.0,>=0.16.4->tokenizers>=0.21.1->vllm==0.11.0rc2.dev389+ge51928192.d20251103) 2025-11-03T17:07:42.1924767Z Downloading shellingham-1.5.4-py2.py3-none-any.whl.metadata (3.5 kB) 2025-11-03T17:07:42.2111413Z Collecting typer-slim (from huggingface-hub<2.0,>=0.16.4->tokenizers>=0.21.1->vllm==0.11.0rc2.dev389+ge51928192.d20251103) 2025-11-03T17:07:42.2154738Z Downloading typer_slim-0.20.0-py3-none-any.whl.metadata (16 kB) 2025-11-03T17:07:42.2558705Z Collecting hf-xet<2.0.0,>=1.2.0 (from huggingface-hub<2.0,>=0.16.4->tokenizers>=0.21.1->vllm==0.11.0rc2.dev389+ge51928192.d20251103) 2025-11-03T17:07:42.2596899Z Downloading hf_xet-1.2.0-cp37-abi3-manylinux_2_17_x86_64.manylinux2014_x86_64.whl.metadata (4.9 kB) 2025-11-03T17:07:42.2772380Z Requirement already satisfied: sympy>=1.13.3 in /opt/conda/envs/py_3.12/lib/python3.12/site-packages (from torch>=1.7.0->compressed-tensors==0.11.0->vllm==0.11.0rc2.dev389+ge51928192.d20251103) (1.13.3) 2025-11-03T17:07:42.2775515Z Requirement already satisfied: networkx>=2.5.1 in /opt/conda/envs/py_3.12/lib/python3.12/site-packages (from torch>=1.7.0->compressed-tensors==0.11.0->vllm==0.11.0rc2.dev389+ge51928192.d20251103) (2.8.8) 2025-11-03T17:07:42.2831516Z Requirement already satisfied: mpmath<1.4,>=1.1.0 in /opt/conda/envs/py_3.12/lib/python3.12/site-packages (from sympy>=1.13.3->torch>=1.7.0->compressed-tensors==0.11.0->vllm==0.11.0rc2.dev389+ge51928192.d20251103) (1.3.0) 2025-11-03T17:07:42.2972673Z Collecting huggingface-hub<2.0,>=0.16.4 (from tokenizers>=0.21.1->vllm==0.11.0rc2.dev389+ge51928192.d20251103) 2025-11-03T17:07:42.3009395Z Downloading huggingface_hub-0.36.0-py3-none-any.whl.metadata (14 kB) 2025-11-03T17:07:42.4249521Z Collecting safetensors>=0.4.3 (from transformers>=4.55.2->vllm==0.11.0rc2.dev389+ge51928192.d20251103) 2025-11-03T17:07:42.4286820Z Downloading safetensors-0.6.2-cp38-abi3-manylinux_2_17_x86_64.manylinux2014_x86_64.whl.metadata (4.1 kB) 2025-11-03T17:07:42.5037010Z Collecting httptools>=0.6.3 (from uvicorn[standard]>=0.12.0; extra == "standard"->fastapi[standard]>=0.115.0->vllm==0.11.0rc2.dev389+ge51928192.d20251103) 2025-11-03T17:07:42.5075405Z Downloading httptools-0.7.1-cp312-cp312-manylinux1_x86_64.manylinux_2_28_x86_64.manylinux_2_5_x86_64.whl.metadata (3.5 kB) 2025-11-03T17:07:42.5253510Z Collecting python-dotenv>=0.13 (from uvicorn[standard]>=0.12.0; extra == "standard"->fastapi[standard]>=0.115.0->vllm==0.11.0rc2.dev389+ge51928192.d20251103) 2025-11-03T17:07:42.5296777Z Downloading python_dotenv-1.2.1-py3-none-any.whl.metadata (25 kB) 2025-11-03T17:07:42.5692748Z Collecting uvloop>=0.15.1 (from uvicorn[standard]>=0.12.0; extra == "standard"->fastapi[standard]>=0.115.0->vllm==0.11.0rc2.dev389+ge51928192.d20251103) 2025-11-03T17:07:42.5728324Z Downloading uvloop-0.22.1-cp312-cp312-manylinux2014_x86_64.manylinux_2_17_x86_64.manylinux_2_28_x86_64.whl.metadata (4.9 kB) 2025-11-03T17:07:42.6510989Z Collecting websockets>=10.4 (from uvicorn[standard]>=0.12.0; extra == "standard"->fastapi[standard]>=0.115.0->vllm==0.11.0rc2.dev389+ge51928192.d20251103) 2025-11-03T17:07:42.6544333Z Downloading websockets-15.0.1-cp312-cp312-manylinux_2_5_x86_64.manylinux1_x86_64.manylinux_2_17_x86_64.manylinux2014_x86_64.whl.metadata (6.8 kB) 2025-11-03T17:07:42.6798380Z Requirement already satisfied: aiohappyeyeballs>=2.5.0 in /opt/conda/envs/py_3.12/lib/python3.12/site-packages (from aiohttp->vllm==0.11.0rc2.dev389+ge51928192.d20251103) (2.6.1) 2025-11-03T17:07:42.6801385Z Requirement already satisfied: aiosignal>=1.4.0 in /opt/conda/envs/py_3.12/lib/python3.12/site-packages (from aiohttp->vllm==0.11.0rc2.dev389+ge51928192.d20251103) (1.4.0) 2025-11-03T17:07:42.6807289Z Requirement already satisfied: frozenlist>=1.1.1 in /opt/conda/envs/py_3.12/lib/python3.12/site-packages (from aiohttp->vllm==0.11.0rc2.dev389+ge51928192.d20251103) (1.8.0) 2025-11-03T17:07:42.6810238Z Requirement already satisfied: multidict<7.0,>=4.5 in /opt/conda/envs/py_3.12/lib/python3.12/site-packages (from aiohttp->vllm==0.11.0rc2.dev389+ge51928192.d20251103) (6.7.0) 2025-11-03T17:07:42.6813901Z Requirement already satisfied: propcache>=0.2.0 in /opt/conda/envs/py_3.12/lib/python3.12/site-packages (from aiohttp->vllm==0.11.0rc2.dev389+ge51928192.d20251103) (0.4.1) 2025-11-03T17:07:42.6817729Z Requirement already satisfied: yarl<2.0,>=1.17.0 in /opt/conda/envs/py_3.12/lib/python3.12/site-packages (from aiohttp->vllm==0.11.0rc2.dev389+ge51928192.d20251103) (1.22.0) 2025-11-03T17:07:42.7412360Z Collecting fastrlock>=0.5 (from cupy-cuda12x->ray[cgraph]>=2.48.0->vllm==0.11.0rc2.dev389+ge51928192.d20251103) 2025-11-03T17:07:42.7466912Z Downloading fastrlock-0.8.3-cp312-cp312-manylinux_2_5_x86_64.manylinux1_x86_64.manylinux_2_28_x86_64.whl.metadata (7.7 kB) 2025-11-03T17:07:42.7654958Z Requirement already satisfied: soundfile>=0.12.1 in /opt/conda/envs/py_3.12/lib/python3.12/site-packages (from mistral_common>=1.8.5->mistral_common[audio,image]>=1.8.5->vllm==0.11.0rc2.dev389+ge51928192.d20251103) (0.13.1) 2025-11-03T17:07:42.7674283Z Requirement already satisfied: cffi>=1.0 in /opt/conda/envs/py_3.12/lib/python3.12/site-packages (from soundfile>=0.12.1->mistral_common>=1.8.5->mistral_common[audio,image]>=1.8.5->vllm==0.11.0rc2.dev389+ge51928192.d20251103) (2.0.0) 2025-11-03T17:07:42.7694623Z Requirement already satisfied: pycparser in /opt/conda/envs/py_3.12/lib/python3.12/site-packages (from cffi>=1.0->soundfile>=0.12.1->mistral_common>=1.8.5->mistral_common[audio,image]>=1.8.5->vllm==0.11.0rc2.dev389+ge51928192.d20251103) (2.23) 2025-11-03T17:07:42.7718780Z Requirement already satisfied: soxr>=0.5.0 in /opt/conda/envs/py_3.12/lib/python3.12/site-packages (from mistral_common>=1.8.5->mistral_common[audio,image]>=1.8.5->vllm==0.11.0rc2.dev389+ge51928192.d20251103) (1.0.0) 2025-11-03T17:07:42.8262889Z Downloading compressed_tensors-0.11.0-py3-none-any.whl (179 kB) 2025-11-03T17:07:42.8363598Z Downloading depyf-0.19.0-py3-none-any.whl (39 kB) 2025-11-03T17:07:42.8442642Z Downloading diskcache-5.6.3-py3-none-any.whl (45 kB) 2025-11-03T17:07:42.8497653Z Using cached apache_tvm_ffi-0.1.0b15-cp312-abi3-manylinux_2_24_x86_64.manylinux_2_28_x86_64.whl (1.6 MB) 2025-11-03T17:07:42.8543067Z Downloading lark-1.2.2-py3-none-any.whl (111 kB) 2025-11-03T17:07:42.8634535Z Downloading lm_format_enforcer-0.11.3-py3-none-any.whl (45 kB) 2025-11-03T17:07:42.8744507Z Downloading numba-0.61.2-cp312-cp312-manylinux2014_x86_64.manylinux_2_17_x86_64.whl (3.9 MB) 2025-11-03T17:07:42.8994248Z [?25l ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 0.0/3.9 MB ? eta -:--:-- 2025-11-03T17:07:42.8995649Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 3.9/3.9 MB 163.2 MB/s 0:00:00 2025-11-03T17:07:42.9064917Z [?25hDownloading outlines_core-0.2.11-cp312-cp312-manylinux_2_17_x86_64.manylinux2014_x86_64.whl (2.3 MB) 2025-11-03T17:07:42.9235900Z [?25l ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 0.0/2.3 MB ? eta -:--:-- 2025-11-03T17:07:42.9237059Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 2.3/2.3 MB 141.4 MB/s 0:00:00 2025-11-03T17:07:42.9308534Z [?25hDownloading xgrammar-0.1.25-cp312-cp312-manylinux_2_17_x86_64.manylinux2014_x86_64.whl (8.7 MB) 2025-11-03T17:07:42.9915023Z [?25l ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 0.0/8.7 MB ? eta -:--:-- 2025-11-03T17:07:42.9915668Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 8.7/8.7 MB 146.4 MB/s 0:00:00 2025-11-03T17:07:42.9986898Z [?25hDownloading llguidance-0.7.30-cp39-abi3-manylinux_2_17_x86_64.manylinux2014_x86_64.whl (15.0 MB) 2025-11-03T17:07:43.0866952Z [?25l ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 0.0/15.0 MB ? eta -:--:-- 2025-11-03T17:07:43.0867946Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 15.0/15.0 MB 173.9 MB/s 0:00:00 2025-11-03T17:07:43.0962548Z [?25hDownloading llvmlite-0.44.0-cp312-cp312-manylinux_2_17_x86_64.manylinux2014_x86_64.whl (42.4 MB) 2025-11-03T17:07:43.2994265Z [?25l ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 0.0/42.4 MB ? eta -:--:-- 2025-11-03T17:07:43.3930685Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╸━━━━ 37.7/42.4 MB 189.5 MB/s eta 0:00:01 2025-11-03T17:07:43.3932561Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 42.4/42.4 MB 143.4 MB/s 0:00:00 2025-11-03T17:07:43.3963616Z [?25hDownloading fastapi-0.121.0-py3-none-any.whl (109 kB) 2025-11-03T17:07:43.4040098Z Downloading pydantic-2.12.3-py3-none-any.whl (462 kB) 2025-11-03T17:07:43.4130634Z Downloading pydantic_core-2.41.4-cp312-cp312-manylinux_2_17_x86_64.manylinux2014_x86_64.whl (2.1 MB) 2025-11-03T17:07:43.4272275Z [?25l ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 0.0/2.1 MB ? eta -:--:-- 2025-11-03T17:07:43.4273545Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 2.1/2.1 MB 165.2 MB/s 0:00:00 2025-11-03T17:07:43.4307847Z [?25hDownloading starlette-0.49.3-py3-none-any.whl (74 kB) 2025-11-03T17:07:43.4371924Z Downloading anyio-4.11.0-py3-none-any.whl (109 kB) 2025-11-03T17:07:43.4441619Z Downloading annotated_doc-0.0.3-py3-none-any.whl (5.5 kB) 2025-11-03T17:07:43.4503674Z Downloading annotated_types-0.7.0-py3-none-any.whl (13 kB) 2025-11-03T17:07:43.4569208Z Downloading httpx-0.28.1-py3-none-any.whl (73 kB) 2025-11-03T17:07:43.4644975Z Downloading httpcore-1.0.9-py3-none-any.whl (78 kB) 2025-11-03T17:07:43.4710800Z Downloading email_validator-2.3.0-py3-none-any.whl (35 kB) 2025-11-03T17:07:43.4772283Z Downloading dnspython-2.8.0-py3-none-any.whl (331 kB) 2025-11-03T17:07:43.4852053Z Downloading fastapi_cli-0.0.14-py3-none-any.whl (11 kB) 2025-11-03T17:07:43.4918146Z Downloading fastapi_cloud_cli-0.3.1-py3-none-any.whl (19 kB) 2025-11-03T17:07:43.5002121Z Downloading gguf-0.17.1-py3-none-any.whl (96 kB) 2025-11-03T17:07:43.5081750Z Downloading h11-0.16.0-py3-none-any.whl (37 kB) 2025-11-03T17:07:43.5183875Z Downloading interegular-0.3.3-py37-none-any.whl (23 kB) 2025-11-03T17:07:43.5285506Z Downloading mistral_common-1.8.5-py3-none-any.whl (6.5 MB) 2025-11-03T17:07:43.5720588Z [?25l ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 0.0/6.5 MB ? eta -:--:-- 2025-11-03T17:07:43.5721752Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 6.5/6.5 MB 154.2 MB/s 0:00:00 2025-11-03T17:07:43.5751012Z [?25hDownloading jsonschema-4.25.1-py3-none-any.whl (90 kB) 2025-11-03T17:07:43.5819650Z Downloading jsonschema_specifications-2025.9.1-py3-none-any.whl (18 kB) 2025-11-03T17:07:43.5918296Z Downloading nvidia_cudnn_frontend-1.15.0-cp312-cp312-manylinux_2_27_x86_64.manylinux_2_28_x86_64.whl (1.9 MB) 2025-11-03T17:07:43.6060498Z [?25l ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 0.0/1.9 MB ? eta -:--:-- 2025-11-03T17:07:43.6061658Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 1.9/1.9 MB 142.1 MB/s 0:00:00 2025-11-03T17:07:43.6140304Z [?25hDownloading nvidia_cutlass_dsl-4.2.1-cp312-cp312-manylinux_2_28_x86_64.whl (62.2 MB) 2025-11-03T17:07:43.8168718Z [?25l ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 0.0/62.2 MB ? eta -:--:-- 2025-11-03T17:07:44.0190636Z  ━━━━━━━━━━━━━━━━━━━━━━━━╺━━━━━━━━━━━━━━━ 38.0/62.2 MB 189.4 MB/s eta 0:00:01 2025-11-03T17:07:44.0553226Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╸ 62.1/62.2 MB 187.2 MB/s eta 0:00:01 2025-11-03T17:07:44.0554532Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 62.2/62.2 MB 141.4 MB/s 0:00:00 2025-11-03T17:07:44.0627204Z [?25hDownloading cuda_python-13.0.3-py3-none-any.whl (7.6 kB) 2025-11-03T17:07:44.0743046Z Downloading cuda_bindings-13.0.3-cp312-cp312-manylinux_2_24_x86_64.manylinux_2_28_x86_64.whl (12.1 MB) 2025-11-03T17:07:44.1586910Z [?25l ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 0.0/12.1 MB ? eta -:--:-- 2025-11-03T17:07:44.1588108Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 12.1/12.1 MB 145.6 MB/s 0:00:00 2025-11-03T17:07:44.1626162Z [?25hDownloading openai-2.6.1-py3-none-any.whl (1.0 MB) 2025-11-03T17:07:44.1728143Z [?25l ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 0.0/1.0 MB ? eta -:--:-- 2025-11-03T17:07:44.1729487Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 1.0/1.0 MB 99.1 MB/s 0:00:00 2025-11-03T17:07:44.1764436Z [?25hDownloading jiter-0.11.1-cp312-cp312-manylinux_2_17_x86_64.manylinux2014_x86_64.whl (358 kB) 2025-11-03T17:07:44.1843241Z Using cached typing_extensions-4.15.0-py3-none-any.whl (44 kB) 2025-11-03T17:07:44.1913675Z Downloading openai_harmony-0.0.4-cp38-abi3-manylinux_2_17_x86_64.manylinux2014_x86_64.whl (3.0 MB) 2025-11-03T17:07:44.2219382Z [?25l ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 0.0/3.0 MB ? eta -:--:-- 2025-11-03T17:07:44.2220546Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 3.0/3.0 MB 100.0 MB/s 0:00:00 2025-11-03T17:07:44.2302831Z [?25hDownloading opencv_python_headless-4.12.0.88-cp37-abi3-manylinux2014_x86_64.manylinux_2_17_x86_64.whl (54.0 MB) 2025-11-03T17:07:44.4332465Z [?25l ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 0.0/54.0 MB ? eta -:--:-- 2025-11-03T17:07:44.5850508Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╸━━━━━━━━━ 41.4/54.0 MB 208.7 MB/s eta 0:00:01 2025-11-03T17:07:44.5852847Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 54.0/54.0 MB 153.7 MB/s 0:00:00 2025-11-03T17:07:44.5885525Z [?25hDownloading numpy-2.2.6-cp312-cp312-manylinux_2_17_x86_64.manylinux2014_x86_64.whl (16.5 MB) 2025-11-03T17:07:44.7167213Z [?25l ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 0.0/16.5 MB ? eta -:--:-- 2025-11-03T17:07:44.7168429Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 16.5/16.5 MB 130.1 MB/s 0:00:00 2025-11-03T17:07:44.7175496Z [?25hUsing cached packaging-25.0-py3-none-any.whl (66 kB) 2025-11-03T17:07:44.7211140Z Downloading prometheus_client-0.23.1-py3-none-any.whl (61 kB) 2025-11-03T17:07:44.7279506Z Downloading prometheus_fastapi_instrumentator-7.1.0-py3-none-any.whl (19 kB) 2025-11-03T17:07:44.7497114Z Downloading pydantic_extra_types-2.10.6-py3-none-any.whl (40 kB) 2025-11-03T17:07:44.7564050Z Downloading pycountry-24.6.1-py3-none-any.whl (6.3 MB) 2025-11-03T17:07:44.7853772Z [?25l ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 0.0/6.3 MB ? eta -:--:-- 2025-11-03T17:07:44.7855422Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 6.3/6.3 MB 231.6 MB/s 0:00:00 2025-11-03T17:07:44.7883554Z [?25hDownloading python_multipart-0.0.20-py3-none-any.whl (24 kB) 2025-11-03T17:07:44.7947683Z Downloading pyzmq-27.1.0-cp312-abi3-manylinux_2_26_x86_64.manylinux_2_28_x86_64.whl (840 kB) 2025-11-03T17:07:44.8034628Z [?25l ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 0.0/841.0 kB ? eta -:--:-- 2025-11-03T17:07:44.8035815Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 841.0/841.0 kB 94.1 MB/s 0:00:00 2025-11-03T17:07:44.8110039Z [?25hDownloading ray-2.51.1-cp312-cp312-manylinux2014_x86_64.whl (71.4 MB) 2025-11-03T17:07:45.0143799Z [?25l ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 0.0/71.4 MB ? eta -:--:-- 2025-11-03T17:07:45.2159708Z  ━━━━━━━━━━━━━━━━━╸━━━━━━━━━━━━━━━━━━━━━━ 31.5/71.4 MB 160.0 MB/s eta 0:00:01 2025-11-03T17:07:45.3563537Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╸━━━ 65.8/71.4 MB 163.1 MB/s eta 0:00:01 2025-11-03T17:07:45.3564897Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 71.4/71.4 MB 131.1 MB/s 0:00:00 2025-11-03T17:07:45.3613288Z [?25hDownloading click-8.2.1-py3-none-any.whl (102 kB) 2025-11-03T17:07:45.3698227Z Downloading referencing-0.37.0-py3-none-any.whl (26 kB) 2025-11-03T17:07:45.3789551Z Downloading rich_toolkit-0.15.1-py3-none-any.whl (29 kB) 2025-11-03T17:07:45.3879149Z Downloading rich-14.2.0-py3-none-any.whl (243 kB) 2025-11-03T17:07:45.3976124Z Downloading markdown_it_py-4.0.0-py3-none-any.whl (87 kB) 2025-11-03T17:07:45.4059695Z Downloading mdurl-0.1.2-py3-none-any.whl (10.0 kB) 2025-11-03T17:07:45.4170200Z Downloading rignore-0.7.4-cp312-cp312-manylinux_2_17_x86_64.manylinux2014_x86_64.whl (959 kB) 2025-11-03T17:07:45.4271874Z [?25l ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 0.0/959.8 kB ? eta -:--:-- 2025-11-03T17:07:45.4273133Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 959.8/959.8 kB 95.8 MB/s 0:00:00 2025-11-03T17:07:45.4313426Z [?25hDownloading rpds_py-0.28.0-cp312-cp312-manylinux_2_17_x86_64.manylinux2014_x86_64.whl (386 kB) 2025-11-03T17:07:45.4417858Z Downloading sentry_sdk-2.43.0-py2.py3-none-any.whl (400 kB) 2025-11-03T17:07:45.4527317Z Downloading sniffio-1.3.1-py3-none-any.whl (10 kB) 2025-11-03T17:07:45.4610529Z Downloading tiktoken-0.12.0-cp312-cp312-manylinux_2_28_x86_64.whl (1.2 MB) 2025-11-03T17:07:45.4742618Z [?25l ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 0.0/1.2 MB ? eta -:--:-- 2025-11-03T17:07:45.4743836Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 1.2/1.2 MB 89.7 MB/s 0:00:00 2025-11-03T17:07:45.4780909Z [?25hDownloading regex-2025.10.23-cp312-cp312-manylinux2014_x86_64.manylinux_2_17_x86_64.manylinux_2_28_x86_64.whl (803 kB) 2025-11-03T17:07:45.4877525Z [?25l ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 0.0/803.4 kB ? eta -:--:-- 2025-11-03T17:07:45.4878786Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 803.4/803.4 kB 82.9 MB/s 0:00:00 2025-11-03T17:07:45.4919426Z [?25hDownloading tokenizers-0.22.1-cp39-abi3-manylinux_2_17_x86_64.manylinux2014_x86_64.whl (3.3 MB) 2025-11-03T17:07:45.5103546Z [?25l ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 0.0/3.3 MB ? eta -:--:-- 2025-11-03T17:07:45.5104724Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 3.3/3.3 MB 189.2 MB/s 0:00:00 2025-11-03T17:07:45.5143836Z [?25hDownloading hf_xet-1.2.0-cp37-abi3-manylinux_2_17_x86_64.manylinux2014_x86_64.whl (3.3 MB) 2025-11-03T17:07:45.6214918Z [?25l ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 0.0/3.3 MB ? eta -:--:-- 2025-11-03T17:07:45.6216146Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 3.3/3.3 MB 193.3 MB/s 0:00:00 2025-11-03T17:07:45.6258310Z [?25hDownloading transformers-4.57.1-py3-none-any.whl (12.0 MB) 2025-11-03T17:07:45.6748824Z [?25l ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 0.0/12.0 MB ? eta -:--:-- 2025-11-03T17:07:45.6750003Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 12.0/12.0 MB 257.6 MB/s 0:00:00 2025-11-03T17:07:45.6788868Z [?25hDownloading huggingface_hub-0.36.0-py3-none-any.whl (566 kB) 2025-11-03T17:07:45.6873223Z [?25l ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 0.0/566.1 kB ? eta -:--:-- 2025-11-03T17:07:45.6874449Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 566.1/566.1 kB 53.0 MB/s 0:00:00 2025-11-03T17:07:45.6910974Z [?25hDownloading safetensors-0.6.2-cp38-abi3-manylinux_2_17_x86_64.manylinux2014_x86_64.whl (485 kB) 2025-11-03T17:07:45.7018833Z Downloading typer-0.20.0-py3-none-any.whl (47 kB) 2025-11-03T17:07:45.7099815Z Downloading shellingham-1.5.4-py2.py3-none-any.whl (9.8 kB) 2025-11-03T17:07:45.7180049Z Downloading typing_inspection-0.4.2-py3-none-any.whl (14 kB) 2025-11-03T17:07:45.7261794Z Downloading uvicorn-0.38.0-py3-none-any.whl (68 kB) 2025-11-03T17:07:45.7350407Z Downloading httptools-0.7.1-cp312-cp312-manylinux1_x86_64.manylinux_2_28_x86_64.manylinux_2_5_x86_64.whl (517 kB) 2025-11-03T17:07:45.7477167Z Downloading python_dotenv-1.2.1-py3-none-any.whl (21 kB) 2025-11-03T17:07:45.7563075Z Downloading uvloop-0.22.1-cp312-cp312-manylinux2014_x86_64.manylinux_2_17_x86_64.manylinux_2_28_x86_64.whl (4.4 MB) 2025-11-03T17:07:45.7988919Z [?25l ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 0.0/4.4 MB ? eta -:--:-- 2025-11-03T17:07:45.7990199Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 4.4/4.4 MB 106.5 MB/s 0:00:00 2025-11-03T17:07:45.8033882Z [?25hDownloading watchfiles-1.1.1-cp312-cp312-manylinux_2_17_x86_64.manylinux2014_x86_64.whl (456 kB) 2025-11-03T17:07:45.8167401Z Downloading websockets-15.0.1-cp312-cp312-manylinux_2_5_x86_64.manylinux1_x86_64.manylinux_2_17_x86_64.manylinux2014_x86_64.whl (182 kB) 2025-11-03T17:07:45.8257793Z Downloading astor-0.8.1-py2.py3-none-any.whl (27 kB) 2025-11-03T17:07:45.8359978Z Downloading blake3-1.0.8-cp312-cp312-manylinux_2_17_x86_64.manylinux2014_x86_64.whl (388 kB) 2025-11-03T17:07:45.8486390Z Downloading cbor2-5.7.1-cp312-cp312-manylinux2014_x86_64.manylinux_2_17_x86_64.manylinux_2_28_x86_64.whl (285 kB) 2025-11-03T17:07:45.8578178Z Downloading cloudpickle-3.1.2-py3-none-any.whl (22 kB) 2025-11-03T17:07:45.8701554Z Downloading cupy_cuda12x-13.6.0-cp312-cp312-manylinux2014_x86_64.whl (112.9 MB) 2025-11-03T17:07:46.0733183Z [?25l ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 0.0/112.9 MB ? eta -:--:-- 2025-11-03T17:07:46.2749447Z  ━━━━━━━━━━━━╸━━━━━━━━━━━━━━━━━━━━━━━━━━━ 35.7/112.9 MB 178.6 MB/s eta 0:00:01 2025-11-03T17:07:46.4768886Z  ━━━━━━━━━━━━━━━━━━━━━━━━━╸━━━━━━━━━━━━━━ 72.6/112.9 MB 180.4 MB/s eta 0:00:01 2025-11-03T17:07:46.6420297Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╺ 111.1/112.9 MB 184.8 MB/s eta 0:00:01 2025-11-03T17:07:46.6421607Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 112.9/112.9 MB 146.5 MB/s 0:00:00 2025-11-03T17:07:46.6488684Z [?25hDownloading fastrlock-0.8.3-cp312-cp312-manylinux_2_5_x86_64.manylinux1_x86_64.manylinux_2_28_x86_64.whl (53 kB) 2025-11-03T17:07:46.6583391Z Downloading einops-0.8.1-py3-none-any.whl (64 kB) 2025-11-03T17:07:46.6672550Z Downloading frozendict-2.4.6-py312-none-any.whl (16 kB) 2025-11-03T17:07:46.6757938Z Downloading msgspec-0.19.0-cp312-cp312-manylinux_2_17_x86_64.manylinux2014_x86_64.whl (213 kB) 2025-11-03T17:07:46.6870714Z Downloading nvidia_ml_py-13.580.82-py3-none-any.whl (49 kB) 2025-11-03T17:07:46.6966603Z Downloading partial_json_parser-0.2.1.1.post6-py3-none-any.whl (10 kB) 2025-11-03T17:07:46.7089015Z Downloading py_cpuinfo-9.0.0-py3-none-any.whl (22 kB) 2025-11-03T17:07:46.7400626Z Downloading pybase64-1.4.2-cp312-cp312-manylinux1_x86_64.manylinux2014_x86_64.manylinux_2_17_x86_64.manylinux_2_5_x86_64.whl (71 kB) 2025-11-03T17:07:46.7482669Z Downloading python_json_logger-4.0.0-py3-none-any.whl (15 kB) 2025-11-03T17:07:46.7598903Z Downloading sentencepiece-0.2.1-cp312-cp312-manylinux_2_27_x86_64.manylinux_2_28_x86_64.whl (1.4 MB) 2025-11-03T17:07:46.7712236Z [?25l ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 0.0/1.4 MB ? eta -:--:-- 2025-11-03T17:07:46.7713958Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 1.4/1.4 MB 130.6 MB/s 0:00:00 2025-11-03T17:07:46.7767834Z [?25hDownloading setproctitle-1.3.7-cp312-cp312-manylinux1_x86_64.manylinux_2_28_x86_64.manylinux_2_5_x86_64.whl (32 kB) 2025-11-03T17:07:47.1675744Z Building wheels for collected packages: flashinfer-python 2025-11-03T17:07:49.7852027Z Building wheel for flashinfer-python (pyproject.toml) ... [?25l- \ | / - \ | / - \ | / - \ | / done 2025-11-03T17:07:49.7905320Z [?25h Created wheel for flashinfer-python: filename=flashinfer_python-0.4.0-py3-none-any.whl size=6764886 sha256=abca777c73ec4e1b3e60ee787915d128c6561395e219d3a4d3a765eaff8a58cc 2025-11-03T17:07:49.7907940Z Stored in directory: /var/lib/jenkins/.cache/pip/wheels/86/5d/0d/2752ded4190ef42ea316eafbbf63e32fd02342fe0764a0dfc6 2025-11-03T17:07:49.7989813Z Successfully built flashinfer-python 2025-11-03T17:07:50.1136483Z Installing collected packages: py-cpuinfo, nvidia-ml-py, fastrlock, websockets, uvloop, typing_extensions, sniffio, shellingham, setproctitle, sentry-sdk, sentencepiece, safetensors, rpds-py, rignore, regex, pyzmq, python-multipart, python-json-logger, python-dotenv, pycountry, pybase64, prometheus_client, partial-json-parser, packaging, outlines_core, nvidia-cudnn-frontend, numpy, msgspec, mdurl, llvmlite, llguidance, lark, jiter, interegular, httptools, hf-xet, h11, frozendict, einops, dnspython, diskcache, cuda-bindings, cloudpickle, click, cbor2, blake3, astor, annotated-types, annotated-doc, uvicorn, typing-inspection, tiktoken, referencing, pydantic-core, opencv-python-headless, numba, markdown-it-py, huggingface-hub, httpcore, gguf, email-validator, depyf, cupy-cuda12x, cuda-python, apache-tvm-ffi, anyio, watchfiles, tokenizers, starlette, rich, pydantic, nvidia-cutlass-dsl, jsonschema-specifications, httpx, typer, transformers, rich-toolkit, pydantic-extra-types, prometheus-fastapi-instrumentator, openai-harmony, openai, lm-format-enforcer, jsonschema, flashinfer-python, fastapi, xgrammar, ray, fastapi-cloud-cli, fastapi-cli, compressed-tensors, mistral_common, vllm 2025-11-03T17:07:50.2820493Z [?25l 2025-11-03T17:07:50.3446800Z  ━╸━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━  4/92 [uvloop] 2025-11-03T17:07:50.3447772Z  Attempting uninstall: typing_extensions 2025-11-03T17:07:50.3503345Z ━╸━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━  4/92 [uvloop] 2025-11-03T17:07:50.3504152Z  Found existing installation: typing_extensions 4.12.2 2025-11-03T17:07:50.3524211Z ━╸━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━  4/92 [uvloop] 2025-11-03T17:07:50.3525120Z  Uninstalling typing_extensions-4.12.2: 2025-11-03T17:07:50.4500763Z ━╸━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━  4/92 [uvloop] 2025-11-03T17:07:50.6180161Z  ━━╺━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━  5/92 [typing_extensions] 2025-11-03T17:07:50.7859960Z  ━━╺━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━  5/92 [typing_extensions] 2025-11-03T17:07:50.8347154Z  ━━╺━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━  5/92 [typing_extensions] 2025-11-03T17:07:50.8348368Z  Successfully uninstalled typing_extensions-4.12.2 2025-11-03T17:07:50.9543451Z ━━╺━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━  5/92 [typing_extensions] 2025-11-03T17:07:51.1224418Z  ━━━╸━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━  9/92 [sentry-sdk] 2025-11-03T17:07:51.2900560Z  ━━━╸━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━  9/92 [sentry-sdk] 2025-11-03T17:07:51.4575931Z  ━━━━━━╸━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 15/92 [pyzmq] 2025-11-03T17:07:51.6250153Z  ━━━━━━━━╺━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 19/92 [pycountry] 2025-11-03T17:07:51.6711715Z  ━━━━━━━━━╺━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 21/92 [prometheus_client] 2025-11-03T17:07:51.6712671Z  Attempting uninstall: packaging 2025-11-03T17:07:51.6757301Z ━━━━━━━━━╺━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 21/92 [prometheus_client] 2025-11-03T17:07:51.6758363Z  Found existing installation: packaging 23.1 2025-11-03T17:07:51.6788878Z ━━━━━━━━━╺━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 21/92 [prometheus_client] 2025-11-03T17:07:51.6789778Z  Uninstalling packaging-23.1: 2025-11-03T17:07:51.6891152Z ━━━━━━━━━╺━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 21/92 [prometheus_client] 2025-11-03T17:07:51.6892117Z  Successfully uninstalled packaging-23.1 2025-11-03T17:07:51.7925127Z ━━━━━━━━━╺━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 21/92 [prometheus_client] 2025-11-03T17:07:51.8054641Z  ━━━━━━━━━━╸━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 25/92 [nvidia-cudnn-frontend] 2025-11-03T17:07:51.8055428Z  Attempting uninstall: numpy 2025-11-03T17:07:51.8135421Z ━━━━━━━━━━╸━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 25/92 [nvidia-cudnn-frontend] 2025-11-03T17:07:51.8136416Z  Found existing installation: numpy 1.26.2 2025-11-03T17:07:51.8679627Z ━━━━━━━━━━╸━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 25/92 [nvidia-cudnn-frontend] 2025-11-03T17:07:51.8680706Z  Uninstalling numpy-1.26.2: 2025-11-03T17:07:51.9600569Z ━━━━━━━━━━╸━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 25/92 [nvidia-cudnn-frontend] 2025-11-03T17:07:52.1275578Z  ━━━━━━━━━━━╺━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 26/92 [numpy] 2025-11-03T17:07:52.2073055Z  ━━━━━━━━━━━╺━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 26/92 [numpy] 2025-11-03T17:07:52.2074043Z  Successfully uninstalled numpy-1.26.2 2025-11-03T17:07:52.2949744Z ━━━━━━━━━━━╺━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 26/92 [numpy] 2025-11-03T17:07:52.4625586Z  ━━━━━━━━━━━╺━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 26/92 [numpy] 2025-11-03T17:07:52.6323926Z  ━━━━━━━━━━━╺━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 26/92 [numpy] 2025-11-03T17:07:52.8008453Z  ━━━━━━━━━━━╺━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 26/92 [numpy] 2025-11-03T17:07:52.9836207Z  ━━━━━━━━━━━╺━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 26/92 [numpy] 2025-11-03T17:07:53.1533196Z  ━━━━━━━━━━━╺━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 26/92 [numpy] 2025-11-03T17:07:53.3224282Z  ━━━━━━━━━━━╺━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 26/92 [numpy] 2025-11-03T17:07:53.4919035Z  ━━━━━━━━━━━╺━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 26/92 [numpy] 2025-11-03T17:07:53.6595164Z  ━━━━━━━━━━━╺━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 26/92 [numpy] 2025-11-03T17:07:53.7525583Z  ━━━━━━━━━━━╺━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 26/92 [numpy] 2025-11-03T17:07:53.7526216Z  Attempting uninstall: llvmlite 2025-11-03T17:07:53.7566093Z ━━━━━━━━━━━╺━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 26/92 [numpy] 2025-11-03T17:07:53.7567029Z  Found existing installation: llvmlite 0.43.0 2025-11-03T17:07:53.7617204Z ━━━━━━━━━━━╺━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 26/92 [numpy] 2025-11-03T17:07:53.7617822Z  Uninstalling llvmlite-0.43.0: 2025-11-03T17:07:53.8270028Z ━━━━━━━━━━━╺━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 26/92 [numpy] 2025-11-03T17:07:53.8511007Z  ━━━━━━━━━━━━╸━━━━━━━━━━━━━━━━━━━━━━━━━━━ 29/92 [llvmlite] 2025-11-03T17:07:53.8512101Z  Successfully uninstalled llvmlite-0.43.0 2025-11-03T17:07:53.9949058Z ━━━━━━━━━━━━╸━━━━━━━━━━━━━━━━━━━━━━━━━━━ 29/92 [llvmlite] 2025-11-03T17:07:54.1624691Z  ━━━━━━━━━━━━╸━━━━━━━━━━━━━━━━━━━━━━━━━━━ 29/92 [llvmlite] 2025-11-03T17:07:54.3301159Z  ━━━━━━━━━━━━╸━━━━━━━━━━━━━━━━━━━━━━━━━━━ 29/92 [llvmlite] 2025-11-03T17:07:54.4976737Z  ━━━━━━━━━━━━╸━━━━━━━━━━━━━━━━━━━━━━━━━━━ 29/92 [llvmlite] 2025-11-03T17:07:54.6213017Z  ━━━━━━━━━━━━━╺━━━━━━━━━━━━━━━━━━━━━━━━━━ 30/92 [llguidance] 2025-11-03T17:07:54.6214205Z  Attempting uninstall: lark 2025-11-03T17:07:54.6251071Z ━━━━━━━━━━━━━╺━━━━━━━━━━━━━━━━━━━━━━━━━━ 30/92 [llguidance] 2025-11-03T17:07:54.6251498Z  Found existing installation: lark 0.12.0 2025-11-03T17:07:54.6312326Z ━━━━━━━━━━━━━╺━━━━━━━━━━━━━━━━━━━━━━━━━━ 30/92 [llguidance] 2025-11-03T17:07:54.6314018Z  Uninstalling lark-0.12.0: 2025-11-03T17:07:54.6556980Z ━━━━━━━━━━━━━╺━━━━━━━━━━━━━━━━━━━━━━━━━━ 30/92 [llguidance] 2025-11-03T17:07:54.6557938Z  Successfully uninstalled lark-0.12.0 2025-11-03T17:07:54.6651676Z ━━━━━━━━━━━━━╺━━━━━━━━━━━━━━━━━━━━━━━━━━ 30/92 [llguidance] 2025-11-03T17:07:54.8331153Z  ━━━━━━━━━━━━━╺━━━━━━━━━━━━━━━━━━━━━━━━━━ 31/92 [lark] 2025-11-03T17:07:55.0006745Z  ━━━━━━━━━━━━━━━━╺━━━━━━━━━━━━━━━━━━━━━━━ 37/92 [frozendict] 2025-11-03T17:07:55.1095960Z  ━━━━━━━━━━━━━━━━╸━━━━━━━━━━━━━━━━━━━━━━━ 39/92 [dnspython] 2025-11-03T17:07:55.1096881Z  Attempting uninstall: cuda-bindings 2025-11-03T17:07:55.1135062Z ━━━━━━━━━━━━━━━━╸━━━━━━━━━━━━━━━━━━━━━━━ 39/92 [dnspython] 2025-11-03T17:07:55.1136485Z  Found existing installation: cuda-bindings 12.9.4 2025-11-03T17:07:55.1199146Z ━━━━━━━━━━━━━━━━╸━━━━━━━━━━━━━━━━━━━━━━━ 39/92 [dnspython] 2025-11-03T17:07:55.1200043Z  Uninstalling cuda-bindings-12.9.4: 2025-11-03T17:07:55.1687831Z ━━━━━━━━━━━━━━━━╸━━━━━━━━━━━━━━━━━━━━━━━ 39/92 [dnspython] 2025-11-03T17:07:55.3362957Z  ━━━━━━━━━━━━━━━━━╸━━━━━━━━━━━━━━━━━━━━━━ 41/92 [cuda-bindings] 2025-11-03T17:07:55.5037547Z  ━━━━━━━━━━━━━━━━━╸━━━━━━━━━━━━━━━━━━━━━━ 41/92 [cuda-bindings] 2025-11-03T17:07:55.6512467Z  ━━━━━━━━━━━━━━━━━╸━━━━━━━━━━━━━━━━━━━━━━ 41/92 [cuda-bindings] 2025-11-03T17:07:55.6513445Z  Successfully uninstalled cuda-bindings-12.9.4 2025-11-03T17:07:55.6714479Z ━━━━━━━━━━━━━━━━━╸━━━━━━━━━━━━━━━━━━━━━━ 41/92 [cuda-bindings] 2025-11-03T17:07:55.8390301Z  ━━━━━━━━━━━━━━━━━╸━━━━━━━━━━━━━━━━━━━━━━ 41/92 [cuda-bindings] 2025-11-03T17:07:55.9219455Z  ━━━━━━━━━━━━━━━━━╸━━━━━━━━━━━━━━━━━━━━━━ 41/92 [cuda-bindings] 2025-11-03T17:07:55.9220318Z  Attempting uninstall: click 2025-11-03T17:07:55.9252658Z ━━━━━━━━━━━━━━━━━╸━━━━━━━━━━━━━━━━━━━━━━ 41/92 [cuda-bindings] 2025-11-03T17:07:55.9253568Z  Found existing installation: click 8.3.0 2025-11-03T17:07:55.9285988Z ━━━━━━━━━━━━━━━━━╸━━━━━━━━━━━━━━━━━━━━━━ 41/92 [cuda-bindings] 2025-11-03T17:07:55.9286415Z  Uninstalling click-8.3.0: 2025-11-03T17:07:55.9406462Z ━━━━━━━━━━━━━━━━━╸━━━━━━━━━━━━━━━━━━━━━━ 41/92 [cuda-bindings] 2025-11-03T17:07:55.9407359Z  Successfully uninstalled click-8.3.0 2025-11-03T17:07:56.0094176Z ━━━━━━━━━━━━━━━━━╸━━━━━━━━━━━━━━━━━━━━━━ 41/92 [cuda-bindings] 2025-11-03T17:07:56.1770831Z  ━━━━━━━━━━━━━━━━━━━╺━━━━━━━━━━━━━━━━━━━━ 44/92 [cbor2] 2025-11-03T17:07:56.3445940Z  ━━━━━━━━━━━━━━━━━━━━━━━╺━━━━━━━━━━━━━━━━ 53/92 [pydantic-core] 2025-11-03T17:07:56.5121191Z  ━━━━━━━━━━━━━━━━━━━━━━━╺━━━━━━━━━━━━━━━━ 54/92 [opencv-python-headless] 2025-11-03T17:07:56.6796461Z  ━━━━━━━━━━━━━━━━━━━━━━━╺━━━━━━━━━━━━━━━━ 54/92 [opencv-python-headless] 2025-11-03T17:07:56.8058112Z  ━━━━━━━━━━━━━━━━━━━━━━━╺━━━━━━━━━━━━━━━━ 54/92 [opencv-python-headless] 2025-11-03T17:07:56.8058979Z  Attempting uninstall: numba 2025-11-03T17:07:56.8100707Z ━━━━━━━━━━━━━━━━━━━━━━━╺━━━━━━━━━━━━━━━━ 54/92 [opencv-python-headless] 2025-11-03T17:07:56.8101229Z  Found existing installation: numba 0.60.0 2025-11-03T17:07:56.8471302Z ━━━━━━━━━━━━━━━━━━━━━━━╺━━━━━━━━━━━━━━━━ 54/92 [opencv-python-headless] 2025-11-03T17:07:56.8835094Z  ━━━━━━━━━━━━━━━━━━━━━━━╸━━━━━━━━━━━━━━━━ 55/92 [numba] 2025-11-03T17:07:56.8835965Z  Uninstalling numba-0.60.0: 2025-11-03T17:07:57.0145933Z ━━━━━━━━━━━━━━━━━━━━━━━╸━━━━━━━━━━━━━━━━ 55/92 [numba] 2025-11-03T17:07:57.1821154Z  ━━━━━━━━━━━━━━━━━━━━━━━╸━━━━━━━━━━━━━━━━ 55/92 [numba] 2025-11-03T17:07:57.2653644Z  ━━━━━━━━━━━━━━━━━━━━━━━╸━━━━━━━━━━━━━━━━ 55/92 [numba] 2025-11-03T17:07:57.2654525Z  Successfully uninstalled numba-0.60.0 2025-11-03T17:07:57.3495807Z ━━━━━━━━━━━━━━━━━━━━━━━╸━━━━━━━━━━━━━━━━ 55/92 [numba] 2025-11-03T17:07:57.5173667Z  ━━━━━━━━━━━━━━━━━━━━━━━╸━━━━━━━━━━━━━━━━ 55/92 [numba] 2025-11-03T17:07:57.6847958Z  ━━━━━━━━━━━━━━━━━━━━━━━╸━━━━━━━━━━━━━━━━ 55/92 [numba] 2025-11-03T17:07:57.8613310Z  ━━━━━━━━━━━━━━━━━━━━━━━╸━━━━━━━━━━━━━━━━ 55/92 [numba] 2025-11-03T17:07:58.0296889Z  ━━━━━━━━━━━━━━━━━━━━━━━╸━━━━━━━━━━━━━━━━ 55/92 [numba] 2025-11-03T17:07:58.1988407Z  ━━━━━━━━━━━━━━━━━━━━━━━╸━━━━━━━━━━━━━━━━ 55/92 [numba] 2025-11-03T17:07:58.3752766Z  ━━━━━━━━━━━━━━━━━━━━━━━╸━━━━━━━━━━━━━━━━ 55/92 [numba] 2025-11-03T17:07:58.5472738Z  ━━━━━━━━━━━━━━━━━━━━━━━╸━━━━━━━━━━━━━━━━ 55/92 [numba] 2025-11-03T17:07:58.7176590Z  ━━━━━━━━━━━━━━━━━━━━━━━╸━━━━━━━━━━━━━━━━ 55/92 [numba] 2025-11-03T17:07:58.8924250Z  ━━━━━━━━━━━━━━━━━━━━━━━╸━━━━━━━━━━━━━━━━ 55/92 [numba] 2025-11-03T17:07:59.0621673Z  ━━━━━━━━━━━━━━━━━━━━━━━╸━━━━━━━━━━━━━━━━ 55/92 [numba] 2025-11-03T17:07:59.2303537Z  ━━━━━━━━━━━━━━━━━━━━━━━╸━━━━━━━━━━━━━━━━ 55/92 [numba] 2025-11-03T17:07:59.3978030Z  ━━━━━━━━━━━━━━━━━━━━━━━━╸━━━━━━━━━━━━━━━ 57/92 [huggingface-hub] 2025-11-03T17:07:59.5675742Z  ━━━━━━━━━━━━━━━━━━━━━━━━╸━━━━━━━━━━━━━━━ 57/92 [huggingface-hub] 2025-11-03T17:07:59.7351990Z  ━━━━━━━━━━━━━━━━━━━━━━━━━╸━━━━━━━━━━━━━━ 59/92 [gguf] 2025-11-03T17:07:59.9026709Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━╸━━━━━━━━━━━━━ 62/92 [cupy-cuda12x] 2025-11-03T17:08:00.0702139Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━╸━━━━━━━━━━━━━ 62/92 [cupy-cuda12x] 2025-11-03T17:08:00.2377603Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━╸━━━━━━━━━━━━━ 62/92 [cupy-cuda12x] 2025-11-03T17:08:00.4052611Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━╸━━━━━━━━━━━━━ 62/92 [cupy-cuda12x] 2025-11-03T17:08:00.5726540Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━╸━━━━━━━━━━━━━ 62/92 [cupy-cuda12x] 2025-11-03T17:08:00.7401440Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━╸━━━━━━━━━━━━━ 62/92 [cupy-cuda12x] 2025-11-03T17:08:00.9075918Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━╸━━━━━━━━━━━━━ 62/92 [cupy-cuda12x] 2025-11-03T17:08:01.0756102Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━╸━━━━━━━━━━━━━ 62/92 [cupy-cuda12x] 2025-11-03T17:08:01.2437427Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━╸━━━━━━━━━━━━━ 62/92 [cupy-cuda12x] 2025-11-03T17:08:01.4112738Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━╸━━━━━━━━━━━━━ 62/92 [cupy-cuda12x] 2025-11-03T17:08:01.5787093Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━╸━━━━━━━━━━━━ 64/92 [apache-tvm-ffi] 2025-11-03T17:08:01.7536396Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╺━━━━━━━━━━ 67/92 [tokenizers] 2025-11-03T17:08:01.9215744Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╺━━━━━━━━━ 69/92 [rich] 2025-11-03T17:08:02.0890257Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╺━━━━━━━━━ 70/92 [pydantic] 2025-11-03T17:08:02.2565142Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╸━━━━━━━━━ 71/92 [nvidia-cutlass-dsl] 2025-11-03T17:08:02.4239027Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╸━━━━━━━━━ 71/92 [nvidia-cutlass-dsl] 2025-11-03T17:08:02.5913963Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╸━━━━━━━━━ 71/92 [nvidia-cutlass-dsl] 2025-11-03T17:08:02.7588930Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╸━━━━━━━━━ 71/92 [nvidia-cutlass-dsl] 2025-11-03T17:08:02.9271503Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╸━━━━━━━━━ 71/92 [nvidia-cutlass-dsl] 2025-11-03T17:08:03.0952156Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╸━━━━━━━━━ 71/92 [nvidia-cutlass-dsl] 2025-11-03T17:08:03.2626028Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╸━━━━━━━━━ 71/92 [nvidia-cutlass-dsl] 2025-11-03T17:08:03.4301097Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╸━━━━━━━ 75/92 [transformers] 2025-11-03T17:08:03.5975575Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╸━━━━━━━ 75/92 [transformers] 2025-11-03T17:08:03.7655838Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╸━━━━━━━ 75/92 [transformers] 2025-11-03T17:08:03.9361189Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╸━━━━━━━ 75/92 [transformers] 2025-11-03T17:08:04.1064985Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╸━━━━━━━ 75/92 [transformers] 2025-11-03T17:08:04.2811702Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╸━━━━━━━ 75/92 [transformers] 2025-11-03T17:08:04.4510899Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╸━━━━━━━ 75/92 [transformers] 2025-11-03T17:08:04.6250590Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╸━━━━━━━ 75/92 [transformers] 2025-11-03T17:08:04.7925059Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╸━━━━━━━ 75/92 [transformers] 2025-11-03T17:08:04.9605481Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╸━━━━━━━ 75/92 [transformers] 2025-11-03T17:08:05.1285953Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╸━━━━━━━ 75/92 [transformers] 2025-11-03T17:08:05.2965810Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╸━━━━━━━ 75/92 [transformers] 2025-11-03T17:08:05.4640820Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╸━━━━━━━ 75/92 [transformers] 2025-11-03T17:08:05.6319960Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╸━━━━━━━ 75/92 [transformers] 2025-11-03T17:08:05.8020582Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╸━━━━━━━ 75/92 [transformers] 2025-11-03T17:08:05.9701626Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╸━━━━━━━ 75/92 [transformers] 2025-11-03T17:08:06.1379415Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╸━━━━━━━ 75/92 [transformers] 2025-11-03T17:08:06.3063206Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╸━━━━━━━ 75/92 [transformers] 2025-11-03T17:08:06.4744358Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╸━━━━━━━ 75/92 [transformers] 2025-11-03T17:08:06.6435313Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╸━━━━━━━ 75/92 [transformers] 2025-11-03T17:08:06.8126958Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╸━━━━━━━ 75/92 [transformers] 2025-11-03T17:08:06.9804906Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╸━━━━━━━ 75/92 [transformers] 2025-11-03T17:08:07.1491208Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╸━━━━━━━ 75/92 [transformers] 2025-11-03T17:08:07.3169002Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╸━━━━━━━ 75/92 [transformers] 2025-11-03T17:08:07.4883114Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╸━━━━━━━ 75/92 [transformers] 2025-11-03T17:08:07.6569289Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╸━━━━━━━ 75/92 [transformers] 2025-11-03T17:08:07.8295007Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╸━━━━━━━ 75/92 [transformers] 2025-11-03T17:08:07.9970660Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╸━━━━━━━ 75/92 [transformers] 2025-11-03T17:08:08.1669428Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╸━━━━━━━ 75/92 [transformers] 2025-11-03T17:08:08.3348846Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╸━━━━━━━ 75/92 [transformers] 2025-11-03T17:08:08.5058581Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╸━━━━━━━ 75/92 [transformers] 2025-11-03T17:08:08.6783609Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╸━━━━━━━ 75/92 [transformers] 2025-11-03T17:08:08.8483541Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╸━━━━━━━ 75/92 [transformers] 2025-11-03T17:08:09.0306548Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╸━━━━━━━ 75/92 [transformers] 2025-11-03T17:08:09.1985830Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╸━━━━━━━ 75/92 [transformers] 2025-11-03T17:08:09.3667812Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╸━━━━━━━ 75/92 [transformers] 2025-11-03T17:08:09.5348835Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╺━━━━━━ 76/92 [rich-toolkit] 2025-11-03T17:08:09.7030256Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╺━━━━━━ 77/92 [pydantic-extra-types] 2025-11-03T17:08:09.8710335Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╸━━━━━ 78/92 [prometheus-fastapi-instrumentator] 2025-11-03T17:08:10.0384369Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╺━━━━━ 79/92 [openai-harmony] 2025-11-03T17:08:10.2061257Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╸━━━━━ 80/92 [openai] 2025-11-03T17:08:10.3736147Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╸━━━━━ 80/92 [openai] 2025-11-03T17:08:10.5415935Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╸━━━━━ 80/92 [openai] 2025-11-03T17:08:10.7097956Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╸━━━━━ 80/92 [openai] 2025-11-03T17:08:10.8773191Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╸━━━━━ 80/92 [openai] 2025-11-03T17:08:11.0448089Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╺━━━ 83/92 [flashinfer-python] 2025-11-03T17:08:11.2122390Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╺━━━ 83/92 [flashinfer-python] 2025-11-03T17:08:11.3814478Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╺━━━ 83/92 [flashinfer-python] 2025-11-03T17:08:11.5493199Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╺━━━ 83/92 [flashinfer-python] 2025-11-03T17:08:11.7174592Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╺━━━ 83/92 [flashinfer-python] 2025-11-03T17:08:11.8848435Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╺━━━ 83/92 [flashinfer-python] 2025-11-03T17:08:12.0527272Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╺━━━ 83/92 [flashinfer-python] 2025-11-03T17:08:12.2201698Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╸━━━ 85/92 [xgrammar] 2025-11-03T17:08:12.3876733Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╺━━ 86/92 [ray] 2025-11-03T17:08:12.5551493Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╺━━ 86/92 [ray] 2025-11-03T17:08:12.7224221Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╺━━ 86/92 [ray] 2025-11-03T17:08:12.8901355Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╺━━ 86/92 [ray] 2025-11-03T17:08:13.0576774Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╺━━ 86/92 [ray] 2025-11-03T17:08:13.2267235Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╺━━ 86/92 [ray] 2025-11-03T17:08:13.3949985Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╺━━ 86/92 [ray] 2025-11-03T17:08:13.5625386Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╺━━ 86/92 [ray] 2025-11-03T17:08:13.7334241Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╺━━ 86/92 [ray] 2025-11-03T17:08:13.9018004Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╺━━ 86/92 [ray] 2025-11-03T17:08:14.0726474Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╺━━ 86/92 [ray] 2025-11-03T17:08:14.2412300Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╺━━ 86/92 [ray] 2025-11-03T17:08:14.4138691Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╺━━ 86/92 [ray] 2025-11-03T17:08:14.5817780Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╺━━ 86/92 [ray] 2025-11-03T17:08:14.7497882Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╺━━ 86/92 [ray] 2025-11-03T17:08:14.9199098Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╺━━ 86/92 [ray] 2025-11-03T17:08:15.0889881Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╺━━ 86/92 [ray] 2025-11-03T17:08:15.2564565Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╺━━ 86/92 [ray] 2025-11-03T17:08:15.4239818Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╺━━ 86/92 [ray] 2025-11-03T17:08:15.5916007Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╺━━ 86/92 [ray] 2025-11-03T17:08:15.7591118Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╺━━ 86/92 [ray] 2025-11-03T17:08:15.9275062Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╺━━ 86/92 [ray] 2025-11-03T17:08:16.0952433Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╸━ 89/92 [compressed-tensors] 2025-11-03T17:08:16.2629482Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╺ 90/92 [mistral_common] 2025-11-03T17:08:16.4305879Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╸ 91/92 [vllm] 2025-11-03T17:08:16.5981692Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╸ 91/92 [vllm] 2025-11-03T17:08:16.7657573Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╸ 91/92 [vllm] 2025-11-03T17:08:16.9333387Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╸ 91/92 [vllm] 2025-11-03T17:08:17.1009899Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╸ 91/92 [vllm] 2025-11-03T17:08:17.2685771Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╸ 91/92 [vllm] 2025-11-03T17:08:17.4361524Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╸ 91/92 [vllm] 2025-11-03T17:08:17.6037425Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╸ 91/92 [vllm] 2025-11-03T17:08:17.7714646Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╸ 91/92 [vllm] 2025-11-03T17:08:17.9390331Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╸ 91/92 [vllm] 2025-11-03T17:08:18.1065998Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╸ 91/92 [vllm] 2025-11-03T17:08:18.2741576Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╸ 91/92 [vllm] 2025-11-03T17:08:18.4417270Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╸ 91/92 [vllm] 2025-11-03T17:08:18.6093052Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╸ 91/92 [vllm] 2025-11-03T17:08:18.7768768Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╸ 91/92 [vllm] 2025-11-03T17:08:18.9443127Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╸ 91/92 [vllm] 2025-11-03T17:08:19.1118240Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╸ 91/92 [vllm] 2025-11-03T17:08:19.2792862Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╸ 91/92 [vllm] 2025-11-03T17:08:19.4468659Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╸ 91/92 [vllm] 2025-11-03T17:08:19.6143310Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╸ 91/92 [vllm] 2025-11-03T17:08:19.7819933Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╸ 91/92 [vllm] 2025-11-03T17:08:19.9495537Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╸ 91/92 [vllm] 2025-11-03T17:08:20.1170358Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╸ 91/92 [vllm] 2025-11-03T17:08:20.2844533Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╸ 91/92 [vllm] 2025-11-03T17:08:20.4520339Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╸ 91/92 [vllm] 2025-11-03T17:08:20.6194211Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╸ 91/92 [vllm] 2025-11-03T17:08:20.7869106Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╸ 91/92 [vllm] 2025-11-03T17:08:20.9543465Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╸ 91/92 [vllm] 2025-11-03T17:08:21.1218600Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╸ 91/92 [vllm] 2025-11-03T17:08:21.2894233Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╸ 91/92 [vllm] 2025-11-03T17:08:21.4583293Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╸ 91/92 [vllm] 2025-11-03T17:08:21.6263288Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╸ 91/92 [vllm] 2025-11-03T17:08:21.7949189Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╸ 91/92 [vllm] 2025-11-03T17:08:21.9624310Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╸ 91/92 [vllm] 2025-11-03T17:08:22.1316660Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╸ 91/92 [vllm] 2025-11-03T17:08:22.2996359Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╸ 91/92 [vllm] 2025-11-03T17:08:22.4676387Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╸ 91/92 [vllm] 2025-11-03T17:08:22.6352233Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╸ 91/92 [vllm] 2025-11-03T17:08:22.8027131Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╸ 91/92 [vllm] 2025-11-03T17:08:22.9819902Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╸ 91/92 [vllm] 2025-11-03T17:08:23.0351632Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╸ 91/92 [vllm] 2025-11-03T17:08:23.0352622Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 92/92 [vllm] 2025-11-03T17:08:23.0353256Z [?25h 2025-11-03T17:08:23.0504670Z Successfully installed annotated-doc-0.0.3 annotated-types-0.7.0 anyio-4.11.0 apache-tvm-ffi-0.1.0b15 astor-0.8.1 blake3-1.0.8 cbor2-5.7.1 click-8.2.1 cloudpickle-3.1.2 compressed-tensors-0.11.0 cuda-bindings-13.0.3 cuda-python-13.0.3 cupy-cuda12x-13.6.0 depyf-0.19.0 diskcache-5.6.3 dnspython-2.8.0 einops-0.8.1 email-validator-2.3.0 fastapi-0.121.0 fastapi-cli-0.0.14 fastapi-cloud-cli-0.3.1 fastrlock-0.8.3 flashinfer-python-0.4.0 frozendict-2.4.6 gguf-0.17.1 h11-0.16.0 hf-xet-1.2.0 httpcore-1.0.9 httptools-0.7.1 httpx-0.28.1 huggingface-hub-0.36.0 interegular-0.3.3 jiter-0.11.1 jsonschema-4.25.1 jsonschema-specifications-2025.9.1 lark-1.2.2 llguidance-0.7.30 llvmlite-0.44.0 lm-format-enforcer-0.11.3 markdown-it-py-4.0.0 mdurl-0.1.2 mistral_common-1.8.5 msgspec-0.19.0 numba-0.61.2 numpy-2.2.6 nvidia-cudnn-frontend-1.15.0 nvidia-cutlass-dsl-4.2.1 nvidia-ml-py-13.580.82 openai-2.6.1 openai-harmony-0.0.4 opencv-python-headless-4.12.0.88 outlines_core-0.2.11 packaging-25.0 partial-json-parser-0.2.1.1.post6 prometheus-fastapi-instrumentator-7.1.0 prometheus_client-0.23.1 py-cpuinfo-9.0.0 pybase64-1.4.2 pycountry-24.6.1 pydantic-2.12.3 pydantic-core-2.41.4 pydantic-extra-types-2.10.6 python-dotenv-1.2.1 python-json-logger-4.0.0 python-multipart-0.0.20 pyzmq-27.1.0 ray-2.51.1 referencing-0.37.0 regex-2025.10.23 rich-14.2.0 rich-toolkit-0.15.1 rignore-0.7.4 rpds-py-0.28.0 safetensors-0.6.2 sentencepiece-0.2.1 sentry-sdk-2.43.0 setproctitle-1.3.7 shellingham-1.5.4 sniffio-1.3.1 starlette-0.49.3 tiktoken-0.12.0 tokenizers-0.22.1 transformers-4.57.1 typer-0.20.0 typing-inspection-0.4.2 typing_extensions-4.15.0 uvicorn-0.38.0 uvloop-0.22.1 vllm-0.11.0rc2.dev389+ge51928192.d20251103 watchfiles-1.1.1 websockets-15.0.1 xgrammar-0.1.25 2025-11-03T17:08:23.9304894Z 2025-11-03 17:08:23,930 [INFO] cli.lib.common.pip_helper: Installing /var/lib/jenkins/workspace/dist/external/vllm/wheels/flashinfer-python/flashinfer_python-0.2.14.post1-cp39-abi3-linux_x86_64.whl... 2025-11-03T17:08:23.9307183Z 2025-11-03 17:08:23,930 [INFO] cli.lib.common.pip_helper: pip installing packages: /opt/conda/envs/py_3.12/bin/python -m pip install /var/lib/jenkins/workspace/dist/external/vllm/wheels/flashinfer-python/flashinfer_python-0.2.14.post1-cp39-abi3-linux_x86_64.whl 2025-11-03T17:08:23.9308993Z 2025-11-03 17:08:23,930 [INFO] cli.lib.common.utils: [cmd] /opt/conda/envs/py_3.12/bin/python -m pip install /var/lib/jenkins/workspace/dist/external/vllm/wheels/flashinfer-python/flashinfer_python-0.2.14.post1-cp39-abi3-linux_x86_64.whl 2025-11-03T17:08:24.2341920Z Processing /var/lib/jenkins/workspace/dist/external/vllm/wheels/flashinfer-python/flashinfer_python-0.2.14.post1-cp39-abi3-linux_x86_64.whl 2025-11-03T17:08:24.3095271Z Requirement already satisfied: numpy in /opt/conda/envs/py_3.12/lib/python3.12/site-packages (from flashinfer-python==0.2.14.post1) (2.2.6) 2025-11-03T17:08:24.3097867Z Requirement already satisfied: torch in /opt/conda/envs/py_3.12/lib/python3.12/site-packages (from flashinfer-python==0.2.14.post1) (2.10.0a0+git3f6538f) 2025-11-03T17:08:24.3100211Z Requirement already satisfied: ninja in /opt/conda/envs/py_3.12/lib/python3.12/site-packages (from flashinfer-python==0.2.14.post1) (1.10.2) 2025-11-03T17:08:24.3102146Z Requirement already satisfied: requests in /opt/conda/envs/py_3.12/lib/python3.12/site-packages (from flashinfer-python==0.2.14.post1) (2.32.5) 2025-11-03T17:08:24.3561056Z Collecting cuda-python<=12.9 (from flashinfer-python==0.2.14.post1) 2025-11-03T17:08:24.3750461Z Downloading cuda_python-12.9.0-py3-none-any.whl.metadata (4.6 kB) 2025-11-03T17:08:24.3972557Z Collecting pynvml (from flashinfer-python==0.2.14.post1) 2025-11-03T17:08:24.4020451Z Downloading pynvml-13.0.1-py3-none-any.whl.metadata (5.6 kB) 2025-11-03T17:08:24.5065346Z Requirement already satisfied: einops in /opt/conda/envs/py_3.12/lib/python3.12/site-packages (from flashinfer-python==0.2.14.post1) (0.8.1) 2025-11-03T17:08:24.5068639Z Requirement already satisfied: packaging>=24.2 in /opt/conda/envs/py_3.12/lib/python3.12/site-packages (from flashinfer-python==0.2.14.post1) (25.0) 2025-11-03T17:08:24.5071489Z Requirement already satisfied: nvidia-cudnn-frontend>=1.13.0 in /opt/conda/envs/py_3.12/lib/python3.12/site-packages (from flashinfer-python==0.2.14.post1) (1.15.0) 2025-11-03T17:08:24.5695407Z Collecting cuda-bindings~=12.9.0 (from cuda-python<=12.9->flashinfer-python==0.2.14.post1) 2025-11-03T17:08:24.5705857Z Using cached cuda_bindings-12.9.4-cp312-cp312-manylinux_2_24_x86_64.manylinux_2_28_x86_64.whl.metadata (2.6 kB) 2025-11-03T17:08:24.5733636Z Requirement already satisfied: cuda-pathfinder~=1.1 in /opt/conda/envs/py_3.12/lib/python3.12/site-packages (from cuda-bindings~=12.9.0->cuda-python<=12.9->flashinfer-python==0.2.14.post1) (1.3.2) 2025-11-03T17:08:24.5746797Z Requirement already satisfied: filelock in /opt/conda/envs/py_3.12/lib/python3.12/site-packages (from torch->flashinfer-python==0.2.14.post1) (3.18.0) 2025-11-03T17:08:24.5750757Z Requirement already satisfied: typing-extensions>=4.10.0 in /opt/conda/envs/py_3.12/lib/python3.12/site-packages (from torch->flashinfer-python==0.2.14.post1) (4.15.0) 2025-11-03T17:08:24.5755421Z Requirement already satisfied: setuptools in /opt/conda/envs/py_3.12/lib/python3.12/site-packages (from torch->flashinfer-python==0.2.14.post1) (78.1.1) 2025-11-03T17:08:24.5758982Z Requirement already satisfied: sympy>=1.13.3 in /opt/conda/envs/py_3.12/lib/python3.12/site-packages (from torch->flashinfer-python==0.2.14.post1) (1.13.3) 2025-11-03T17:08:24.5762456Z Requirement already satisfied: networkx>=2.5.1 in /opt/conda/envs/py_3.12/lib/python3.12/site-packages (from torch->flashinfer-python==0.2.14.post1) (2.8.8) 2025-11-03T17:08:24.5765274Z Requirement already satisfied: jinja2 in /opt/conda/envs/py_3.12/lib/python3.12/site-packages (from torch->flashinfer-python==0.2.14.post1) (3.1.6) 2025-11-03T17:08:24.5768791Z Requirement already satisfied: fsspec>=0.8.5 in /opt/conda/envs/py_3.12/lib/python3.12/site-packages (from torch->flashinfer-python==0.2.14.post1) (2025.9.0) 2025-11-03T17:08:24.5885864Z Requirement already satisfied: mpmath<1.4,>=1.1.0 in /opt/conda/envs/py_3.12/lib/python3.12/site-packages (from sympy>=1.13.3->torch->flashinfer-python==0.2.14.post1) (1.3.0) 2025-11-03T17:08:24.5933442Z Requirement already satisfied: MarkupSafe>=2.0 in /opt/conda/envs/py_3.12/lib/python3.12/site-packages (from jinja2->torch->flashinfer-python==0.2.14.post1) (3.0.3) 2025-11-03T17:08:24.5983209Z Requirement already satisfied: nvidia-ml-py>=12.0.0 in /opt/conda/envs/py_3.12/lib/python3.12/site-packages (from pynvml->flashinfer-python==0.2.14.post1) (13.580.82) 2025-11-03T17:08:24.5998007Z Requirement already satisfied: charset_normalizer<4,>=2 in /opt/conda/envs/py_3.12/lib/python3.12/site-packages (from requests->flashinfer-python==0.2.14.post1) (3.4.4) 2025-11-03T17:08:24.6001719Z Requirement already satisfied: idna<4,>=2.5 in /opt/conda/envs/py_3.12/lib/python3.12/site-packages (from requests->flashinfer-python==0.2.14.post1) (3.11) 2025-11-03T17:08:24.6004225Z Requirement already satisfied: urllib3<3,>=1.21.1 in /opt/conda/envs/py_3.12/lib/python3.12/site-packages (from requests->flashinfer-python==0.2.14.post1) (2.5.0) 2025-11-03T17:08:24.6007442Z Requirement already satisfied: certifi>=2017.4.17 in /opt/conda/envs/py_3.12/lib/python3.12/site-packages (from requests->flashinfer-python==0.2.14.post1) (2025.10.5) 2025-11-03T17:08:24.6148586Z Downloading cuda_python-12.9.0-py3-none-any.whl (7.5 kB) 2025-11-03T17:08:24.7225029Z Using cached cuda_bindings-12.9.4-cp312-cp312-manylinux_2_24_x86_64.manylinux_2_28_x86_64.whl (12.2 MB) 2025-11-03T17:08:24.7331681Z Downloading pynvml-13.0.1-py3-none-any.whl (28 kB) 2025-11-03T17:08:25.1576598Z Installing collected packages: pynvml, cuda-bindings, cuda-python, flashinfer-python 2025-11-03T17:08:25.3047997Z [?25l 2025-11-03T17:08:25.3048757Z  Attempting uninstall: cuda-bindings 2025-11-03T17:08:25.3120138Z 2025-11-03T17:08:25.3120634Z  Found existing installation: cuda-bindings 13.0.3 2025-11-03T17:08:25.3166130Z 2025-11-03T17:08:25.3166531Z  Uninstalling cuda-bindings-13.0.3: 2025-11-03T17:08:25.3179945Z 2025-11-03T17:08:25.3180154Z  Successfully uninstalled cuda-bindings-13.0.3 2025-11-03T17:08:25.3264938Z 2025-11-03T17:08:25.4946638Z  ━━━━━━━━━━╺━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 1/4 [cuda-bindings] 2025-11-03T17:08:25.6628789Z  ━━━━━━━━━━╺━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 1/4 [cuda-bindings] 2025-11-03T17:08:25.6774598Z  ━━━━━━━━━━╺━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 1/4 [cuda-bindings] 2025-11-03T17:08:25.6775094Z  Attempting uninstall: cuda-python 2025-11-03T17:08:25.6853581Z ━━━━━━━━━━╺━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 1/4 [cuda-bindings] 2025-11-03T17:08:25.6854540Z  Found existing installation: cuda-python 13.0.3 2025-11-03T17:08:25.6873333Z ━━━━━━━━━━╺━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 1/4 [cuda-bindings] 2025-11-03T17:08:25.6874581Z  Uninstalling cuda-python-13.0.3: 2025-11-03T17:08:25.6881977Z ━━━━━━━━━━╺━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 1/4 [cuda-bindings] 2025-11-03T17:08:25.6882889Z  Successfully uninstalled cuda-python-13.0.3 2025-11-03T17:08:25.7183618Z ━━━━━━━━━━╺━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 1/4 [cuda-bindings] 2025-11-03T17:08:25.7184122Z  Attempting uninstall: flashinfer-python 2025-11-03T17:08:25.7268118Z ━━━━━━━━━━╺━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 1/4 [cuda-bindings] 2025-11-03T17:08:25.7269117Z  Found existing installation: flashinfer-python 0.4.0 2025-11-03T17:08:25.8015813Z ━━━━━━━━━━╺━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 1/4 [cuda-bindings] 2025-11-03T17:08:25.8016789Z  Uninstalling flashinfer-python-0.4.0: 2025-11-03T17:08:25.8308896Z ━━━━━━━━━━╺━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 1/4 [cuda-bindings] 2025-11-03T17:08:25.9989231Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╺━━━━━━━━━ 3/4 [flashinfer-python] 2025-11-03T17:08:26.1669716Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╺━━━━━━━━━ 3/4 [flashinfer-python] 2025-11-03T17:08:26.3349862Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╺━━━━━━━━━ 3/4 [flashinfer-python] 2025-11-03T17:08:26.3728891Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╺━━━━━━━━━ 3/4 [flashinfer-python] 2025-11-03T17:08:26.3729905Z  Successfully uninstalled flashinfer-python-0.4.0 2025-11-03T17:08:26.5030799Z ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╺━━━━━━━━━ 3/4 [flashinfer-python] 2025-11-03T17:08:26.6710118Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╺━━━━━━━━━ 3/4 [flashinfer-python] 2025-11-03T17:08:26.8385442Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╺━━━━━━━━━ 3/4 [flashinfer-python] 2025-11-03T17:08:27.0062239Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╺━━━━━━━━━ 3/4 [flashinfer-python] 2025-11-03T17:08:27.1740724Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╺━━━━━━━━━ 3/4 [flashinfer-python] 2025-11-03T17:08:27.3418733Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╺━━━━━━━━━ 3/4 [flashinfer-python] 2025-11-03T17:08:27.5100696Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╺━━━━━━━━━ 3/4 [flashinfer-python] 2025-11-03T17:08:27.6780753Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╺━━━━━━━━━ 3/4 [flashinfer-python] 2025-11-03T17:08:27.8470842Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╺━━━━━━━━━ 3/4 [flashinfer-python] 2025-11-03T17:08:27.9679684Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╺━━━━━━━━━ 3/4 [flashinfer-python] 2025-11-03T17:08:27.9681301Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 4/4 [flashinfer-python] 2025-11-03T17:08:27.9682055Z [?25h 2025-11-03T17:08:27.9834909Z ERROR: pip's dependency resolver does not currently take into account all the packages that are installed. This behaviour is the source of the following dependency conflicts. 2025-11-03T17:08:27.9836492Z vllm 0.11.0rc2.dev389+ge51928192.d20251103 requires flashinfer-python==0.4.0, but you have flashinfer-python 0.2.14.post1 which is incompatible. 2025-11-03T17:08:27.9837523Z Successfully installed cuda-bindings-12.9.4 cuda-python-12.9.0 flashinfer-python-0.2.14.post1 pynvml-13.0.1 2025-11-03T17:08:28.0514602Z 2025-11-03 17:08:28,050 [INFO] cli.lib.core.vllm.vllm_test: Done. Installed vllm wheels 2025-11-03T17:08:28.0516923Z 2025-11-03 17:08:28,051 [INFO] cli.lib.common.pip_helper: pip installing packages: /opt/conda/envs/py_3.12/bin/python -m uv pip install -e tests/vllm_test_utils 2025-11-03T17:08:28.0517981Z 2025-11-03 17:08:28,051 [INFO] cli.lib.common.utils: [cmd] /opt/conda/envs/py_3.12/bin/python -m uv pip install -e tests/vllm_test_utils 2025-11-03T17:08:28.1719336Z Using Python 3.12.12 environment at: /opt/conda/envs/py_3.12 2025-11-03T17:08:28.8974695Z Resolved 1 package in 675ms 2025-11-03T17:08:28.8992599Z Building vllm-test-utils @ file:///var/lib/jenkins/workspace/vllm/tests/vllm_test_utils 2025-11-03T17:08:29.0817878Z Built vllm-test-utils @ file:///var/lib/jenkins/workspace/vllm/tests/vllm_test_utils 2025-11-03T17:08:29.0844605Z Prepared 1 package in 186ms 2025-11-03T17:08:29.0879469Z Installed 1 package in 3ms 2025-11-03T17:08:29.0881872Z + vllm-test-utils==0.1 (from file:///var/lib/jenkins/workspace/vllm/tests/vllm_test_utils) 2025-11-03T17:08:29.1101097Z 2025-11-03 17:08:29,109 [INFO] cli.lib.common.pip_helper: pip installing packages: /opt/conda/envs/py_3.12/bin/python -m uv pip install hf_transfer 2025-11-03T17:08:29.1102007Z 2025-11-03 17:08:29,109 [INFO] cli.lib.common.utils: [cmd] /opt/conda/envs/py_3.12/bin/python -m uv pip install hf_transfer 2025-11-03T17:08:29.1653657Z Using Python 3.12.12 environment at: /opt/conda/envs/py_3.12 2025-11-03T17:08:29.1975877Z Resolved 1 package in 24ms 2025-11-03T17:08:29.2039355Z Downloading hf-transfer (3.4MiB) 2025-11-03T17:08:29.2316025Z Downloading hf-transfer 2025-11-03T17:08:29.2318925Z Prepared 1 package in 34ms 2025-11-03T17:08:29.2441257Z Installed 1 package in 12ms 2025-11-03T17:08:29.2441675Z + hf-transfer==0.1.9 2025-11-03T17:08:29.2497514Z 2025-11-03 17:08:29,249 [INFO] cli.lib.common.utils: [cmd] /opt/conda/envs/py_3.12/bin/python use_existing_torch.py 2025-11-03T17:08:29.2896942Z >>> cleaning requirements/cpu.txt 2025-11-03T17:08:29.2897601Z removed: 2025-11-03T17:08:29.2897901Z <<< done cleaning requirements/cpu.txt 2025-11-03T17:08:29.2898162Z 2025-11-03T17:08:29.2898264Z >>> cleaning requirements/dev.txt 2025-11-03T17:08:29.2898638Z <<< done cleaning requirements/dev.txt 2025-11-03T17:08:29.2898832Z 2025-11-03T17:08:29.2898946Z >>> cleaning requirements/kv_connectors.txt 2025-11-03T17:08:29.2899391Z <<< done cleaning requirements/kv_connectors.txt 2025-11-03T17:08:29.2899614Z 2025-11-03T17:08:29.2899719Z >>> cleaning requirements/lint.txt 2025-11-03T17:08:29.2900271Z <<< done cleaning requirements/lint.txt 2025-11-03T17:08:29.2900620Z 2025-11-03T17:08:29.2900722Z >>> cleaning requirements/tpu.txt 2025-11-03T17:08:29.2900975Z removed: 2025-11-03T17:08:29.2901256Z <<< done cleaning requirements/tpu.txt 2025-11-03T17:08:29.2901504Z 2025-11-03T17:08:29.2901608Z >>> cleaning requirements/build.txt 2025-11-03T17:08:29.2901994Z removed: 2025-11-03T17:08:29.2902195Z <<< done cleaning requirements/build.txt 2025-11-03T17:08:29.2902380Z 2025-11-03T17:08:29.2902494Z >>> cleaning requirements/common.txt 2025-11-03T17:08:29.2903184Z <<< done cleaning requirements/common.txt 2025-11-03T17:08:29.2903506Z 2025-11-03T17:08:29.2903621Z >>> cleaning requirements/cpu-build.txt 2025-11-03T17:08:29.2903891Z removed: 2025-11-03T17:08:29.2904119Z <<< done cleaning requirements/cpu-build.txt 2025-11-03T17:08:29.2904480Z 2025-11-03T17:08:29.2904623Z >>> cleaning requirements/cuda.txt 2025-11-03T17:08:29.2904963Z removed: 2025-11-03T17:08:29.2905302Z <<< done cleaning requirements/cuda.txt 2025-11-03T17:08:29.2905586Z 2025-11-03T17:08:29.2905683Z >>> cleaning requirements/docs.txt 2025-11-03T17:08:29.2905931Z removed: 2025-11-03T17:08:29.2906120Z <<< done cleaning requirements/docs.txt 2025-11-03T17:08:29.2906308Z 2025-11-03T17:08:29.2906468Z >>> cleaning requirements/nightly_torch_test.txt 2025-11-03T17:08:29.2906763Z <<< done cleaning requirements/nightly_torch_test.txt 2025-11-03T17:08:29.2907060Z 2025-11-03T17:08:29.2907147Z >>> cleaning requirements/rocm-build.txt 2025-11-03T17:08:29.2907359Z removed: 2025-11-03T17:08:29.2907532Z <<< done cleaning requirements/rocm-build.txt 2025-11-03T17:08:29.2907804Z 2025-11-03T17:08:29.2907890Z >>> cleaning requirements/rocm-test.txt 2025-11-03T17:08:29.2908129Z <<< done cleaning requirements/rocm-test.txt 2025-11-03T17:08:29.2908281Z 2025-11-03T17:08:29.2908362Z >>> cleaning requirements/rocm.txt 2025-11-03T17:08:29.2908781Z <<< done cleaning requirements/rocm.txt 2025-11-03T17:08:29.2908991Z 2025-11-03T17:08:29.2909083Z >>> cleaning requirements/test.txt 2025-11-03T17:08:29.2910950Z removed: 2025-11-03T17:08:29.2914172Z <<< done cleaning requirements/test.txt 2025-11-03T17:08:29.2914348Z 2025-11-03T17:08:29.2914434Z >>> cleaning requirements/xpu.txt 2025-11-03T17:08:29.2914858Z removed: 2025-11-03T17:08:29.2916042Z <<< done cleaning requirements/xpu.txt 2025-11-03T17:08:29.2916225Z 2025-11-03T17:08:29.2916308Z >>> cleaning pyproject.toml 2025-11-03T17:08:29.2916889Z removed: 2025-11-03T17:08:29.2919737Z <<< done cleaning pyproject.toml 2025-11-03T17:08:29.2919886Z 2025-11-03T17:08:29.2977926Z 2025-11-03 17:08:29,297 [INFO] cli.lib.common.pip_helper: pip installing packages: /opt/conda/envs/py_3.12/bin/python -m uv pip install -r requirements/common.txt 2025-11-03T17:08:29.2978890Z 2025-11-03 17:08:29,297 [INFO] cli.lib.common.utils: [cmd] /opt/conda/envs/py_3.12/bin/python -m uv pip install -r requirements/common.txt 2025-11-03T17:08:29.3501435Z Using Python 3.12.12 environment at: /opt/conda/envs/py_3.12 2025-11-03T17:08:29.3670670Z Audited 50 packages in 19ms 2025-11-03T17:08:29.3695349Z 2025-11-03 17:08:29,369 [INFO] cli.lib.common.pip_helper: pip installing packages: /opt/conda/envs/py_3.12/bin/python -m uv pip install -r requirements/build.txt 2025-11-03T17:08:29.3696869Z 2025-11-03 17:08:29,369 [INFO] cli.lib.common.utils: [cmd] /opt/conda/envs/py_3.12/bin/python -m uv pip install -r requirements/build.txt 2025-11-03T17:08:29.4174818Z Using Python 3.12.12 environment at: /opt/conda/envs/py_3.12 2025-11-03T17:08:29.4968223Z Resolved 11 packages in 72ms 2025-11-03T17:08:29.5170573Z Prepared 1 package in 19ms 2025-11-03T17:08:29.5317047Z Installed 1 package in 14ms 2025-11-03T17:08:29.5317521Z + setuptools-scm==9.2.2 2025-11-03T17:08:29.5390750Z 2025-11-03 17:08:29,538 [INFO] cli.lib.core.vllm.vllm_test: generate test.txt from requirements/test.in with local torch whls 2025-11-03T17:08:29.9096968Z 2025-11-03 17:08:29,909 [INFO] cli.lib.core.vllm.vllm_test: [INFO] Updated requirements/test.in 2025-11-03T17:08:29.9102410Z 2025-11-03 17:08:29,909 [INFO] cli.lib.common.utils: [cmd] /opt/conda/envs/py_3.12/bin/python -m uv pip compile requirements/test.in -o test.txt --index-strategy unsafe-best-match --constraint snapshot_constraint.txt --torch-backend cu128 2025-11-03T17:08:29.9817482Z Updating https://github.com/IBM/terratorch.git (1.1.rc3) 2025-11-03T17:08:32.4544243Z Updated https://github.com/IBM/terratorch.git (07184fcf91a1324f831ff521dd238d97fe350e3e) 2025-11-03T17:08:35.4798633Z Resolved 329 packages in 5.51s 2025-11-03T17:08:35.4799719Z # This file was autogenerated by uv via the following command: 2025-11-03T17:08:35.5097085Z # uv pip compile requirements/test.in -o test.txt --index-strategy unsafe-best-match --constraint snapshot_constraint.txt --torch-backend cu128 2025-11-03T17:08:35.5502279Z absl-py==2.1.0 2025-11-03T17:08:35.5502845Z  # via 2025-11-03T17:08:35.5503240Z # -c snapshot_constraint.txt 2025-11-03T17:08:35.5503783Z # rouge-score 2025-11-03T17:08:35.5504179Z accelerate==1.0.1 2025-11-03T17:08:35.5504572Z  # via 2025-11-03T17:08:35.5504933Z # -c snapshot_constraint.txt 2025-11-03T17:08:35.5505418Z # lm-eval 2025-11-03T17:08:35.5505818Z # peft 2025-11-03T17:08:35.5506152Z aenum==3.1.16 2025-11-03T17:08:35.5506534Z  # via 2025-11-03T17:08:35.5506904Z # -c snapshot_constraint.txt 2025-11-03T17:08:35.5507223Z # lightly 2025-11-03T17:08:35.5507437Z affine==2.4.0 2025-11-03T17:08:35.5507651Z  # via 2025-11-03T17:08:35.5507862Z # -c snapshot_constraint.txt 2025-11-03T17:08:35.5508148Z # rasterio 2025-11-03T17:08:35.5508370Z aiohappyeyeballs==2.6.1 2025-11-03T17:08:35.5508614Z  # via 2025-11-03T17:08:35.5508815Z # -c snapshot_constraint.txt 2025-11-03T17:08:35.5509105Z # aiohttp 2025-11-03T17:08:35.5509317Z aiohttp==3.13.0 2025-11-03T17:08:35.5509755Z  # via 2025-11-03T17:08:35.5510071Z # -c snapshot_constraint.txt 2025-11-03T17:08:35.5510508Z # aiohttp-cors 2025-11-03T17:08:35.5510846Z # datasets 2025-11-03T17:08:35.5511179Z # fsspec 2025-11-03T17:08:35.5511481Z # gpt-oss 2025-11-03T17:08:35.5511746Z # lm-eval 2025-11-03T17:08:35.5512064Z # ray 2025-11-03T17:08:35.5512335Z aiohttp-cors==0.8.1 2025-11-03T17:08:35.5512637Z  # via 2025-11-03T17:08:35.5512929Z # -c snapshot_constraint.txt 2025-11-03T17:08:35.5513322Z # ray 2025-11-03T17:08:35.5513587Z aiosignal==1.4.0 2025-11-03T17:08:35.5513875Z  # via 2025-11-03T17:08:35.5514145Z # -c snapshot_constraint.txt 2025-11-03T17:08:35.5514532Z # aiohttp 2025-11-03T17:08:35.5514817Z albucore==0.0.16 2025-11-03T17:08:35.5515116Z  # via 2025-11-03T17:08:35.5515404Z # -c snapshot_constraint.txt 2025-11-03T17:08:35.5515797Z # terratorch 2025-11-03T17:08:35.5516093Z albumentations==1.4.6 2025-11-03T17:08:35.5516416Z  # via 2025-11-03T17:08:35.5516689Z # -c snapshot_constraint.txt 2025-11-03T17:08:35.5517079Z # terratorch 2025-11-03T17:08:35.5517388Z alembic==1.16.4 2025-11-03T17:08:35.5517693Z  # via 2025-11-03T17:08:35.5518345Z # -c snapshot_constraint.txt 2025-11-03T17:08:35.5518914Z # mlflow 2025-11-03T17:08:35.5519208Z annotated-types==0.7.0 2025-11-03T17:08:35.5519547Z  # via 2025-11-03T17:08:35.5519844Z # -c snapshot_constraint.txt 2025-11-03T17:08:35.5520256Z # pydantic 2025-11-03T17:08:35.5520501Z antlr4-python3-runtime==4.9.3 2025-11-03T17:08:35.5520862Z  # via 2025-11-03T17:08:35.5521139Z # -c snapshot_constraint.txt 2025-11-03T17:08:35.5521477Z # hydra-core 2025-11-03T17:08:35.5521789Z # omegaconf 2025-11-03T17:08:35.5522084Z anyio==4.6.2.post1 2025-11-03T17:08:35.5522388Z  # via 2025-11-03T17:08:35.5522588Z # -c snapshot_constraint.txt 2025-11-03T17:08:35.5522824Z # httpx 2025-11-03T17:08:35.5523138Z # starlette 2025-11-03T17:08:35.5523438Z argcomplete==3.5.1 2025-11-03T17:08:35.5523699Z  # via 2025-11-03T17:08:35.5523971Z # -c snapshot_constraint.txt 2025-11-03T17:08:35.5524415Z # datamodel-code-generator 2025-11-03T17:08:35.5524789Z arrow==1.3.0 2025-11-03T17:08:35.5525043Z  # via 2025-11-03T17:08:35.5525241Z # -c snapshot_constraint.txt 2025-11-03T17:08:35.5525856Z # isoduration 2025-11-03T17:08:35.5526178Z attrs==24.2.0 2025-11-03T17:08:35.5526489Z  # via 2025-11-03T17:08:35.5526753Z # -c snapshot_constraint.txt 2025-11-03T17:08:35.5527083Z # aiohttp 2025-11-03T17:08:35.5527335Z # fiona 2025-11-03T17:08:35.5527493Z # hypothesis 2025-11-03T17:08:35.5527746Z # jsonlines 2025-11-03T17:08:35.5527998Z # jsonschema 2025-11-03T17:08:35.5528274Z # pytest-subtests 2025-11-03T17:08:35.5528637Z # rasterio 2025-11-03T17:08:35.5528966Z # referencing 2025-11-03T17:08:35.5529289Z audioread==3.0.1 2025-11-03T17:08:35.5529558Z  # via 2025-11-03T17:08:35.5529820Z # -c snapshot_constraint.txt 2025-11-03T17:08:35.5530210Z # librosa 2025-11-03T17:08:35.5530487Z backoff==2.2.1 2025-11-03T17:08:35.5530800Z  # via 2025-11-03T17:08:35.5531072Z # -c snapshot_constraint.txt 2025-11-03T17:08:35.5531421Z # -r requirements/test.in 2025-11-03T17:08:35.5531826Z # schemathesis 2025-11-03T17:08:35.5532139Z bitsandbytes==0.46.1 2025-11-03T17:08:35.5532461Z  # via 2025-11-03T17:08:35.5532733Z # -c snapshot_constraint.txt 2025-11-03T17:08:35.5533109Z # -r requirements/test.in 2025-11-03T17:08:35.5533486Z # lightning 2025-11-03T17:08:35.5533768Z black==24.10.0 2025-11-03T17:08:35.5534054Z  # via 2025-11-03T17:08:35.5534221Z # -c snapshot_constraint.txt 2025-11-03T17:08:35.5534477Z # datamodel-code-generator 2025-11-03T17:08:35.5534684Z blinker==1.9.0 2025-11-03T17:08:35.5534859Z  # via 2025-11-03T17:08:35.5535030Z # -c snapshot_constraint.txt 2025-11-03T17:08:35.5535253Z # flask 2025-11-03T17:08:35.5535444Z blobfile==3.0.0 2025-11-03T17:08:35.5535752Z  # via 2025-11-03T17:08:35.5536047Z # -c snapshot_constraint.txt 2025-11-03T17:08:35.5536474Z # -r requirements/test.in 2025-11-03T17:08:35.5536807Z bm25s==0.2.13 2025-11-03T17:08:35.5537102Z  # via 2025-11-03T17:08:35.5537382Z # -c snapshot_constraint.txt 2025-11-03T17:08:35.5537778Z # mteb 2025-11-03T17:08:35.5538031Z boto3==1.35.57 2025-11-03T17:08:35.5538324Z  # via 2025-11-03T17:08:35.5538580Z # -c snapshot_constraint.txt 2025-11-03T17:08:35.5538935Z # runai-model-streamer-s3 2025-11-03T17:08:35.5539304Z # tensorizer 2025-11-03T17:08:35.5539589Z botocore==1.35.57 2025-11-03T17:08:35.5539888Z  # via 2025-11-03T17:08:35.5540074Z # -c snapshot_constraint.txt 2025-11-03T17:08:35.5540391Z # boto3 2025-11-03T17:08:35.5540668Z # s3transfer 2025-11-03T17:08:35.5540988Z bounded-pool-executor==0.0.3 2025-11-03T17:08:35.5541357Z  # via 2025-11-03T17:08:35.5541629Z # -c snapshot_constraint.txt 2025-11-03T17:08:35.5542016Z # pqdm 2025-11-03T17:08:35.5542515Z buildkite-test-collector==0.1.9 2025-11-03T17:08:35.5542897Z  # via 2025-11-03T17:08:35.5543177Z # -c snapshot_constraint.txt 2025-11-03T17:08:35.5543645Z # -r requirements/test.in 2025-11-03T17:08:35.5543912Z cachetools==5.5.2 2025-11-03T17:08:35.5544222Z  # via 2025-11-03T17:08:35.5544505Z # -c snapshot_constraint.txt 2025-11-03T17:08:35.5544878Z # google-auth 2025-11-03T17:08:35.5545217Z # mlflow-skinny 2025-11-03T17:08:35.5545524Z certifi==2024.8.30 2025-11-03T17:08:35.5545837Z  # via 2025-11-03T17:08:35.5546029Z # -c snapshot_constraint.txt 2025-11-03T17:08:35.5546227Z # fiona 2025-11-03T17:08:35.5546372Z # httpcore 2025-11-03T17:08:35.5546527Z # httpx 2025-11-03T17:08:35.5546670Z # lightly 2025-11-03T17:08:35.5546820Z # pyogrio 2025-11-03T17:08:35.5546965Z # pyproj 2025-11-03T17:08:35.5547113Z # rasterio 2025-11-03T17:08:35.5547305Z # requests 2025-11-03T17:08:35.5547477Z cffi==1.17.1 2025-11-03T17:08:35.5547642Z  # via 2025-11-03T17:08:35.5547806Z # -c snapshot_constraint.txt 2025-11-03T17:08:35.5548033Z # soundfile 2025-11-03T17:08:35.5548296Z chardet==5.2.0 2025-11-03T17:08:35.5548476Z  # via 2025-11-03T17:08:35.5548638Z # -c snapshot_constraint.txt 2025-11-03T17:08:35.5548869Z # mbstrdecoder 2025-11-03T17:08:35.5549053Z charset-normalizer==3.4.0 2025-11-03T17:08:35.5549251Z  # via 2025-11-03T17:08:35.5549410Z # -c snapshot_constraint.txt 2025-11-03T17:08:35.5549632Z # requests 2025-11-03T17:08:35.5549798Z chz==0.3.0 2025-11-03T17:08:35.5549964Z  # via 2025-11-03T17:08:35.5550122Z # -c snapshot_constraint.txt 2025-11-03T17:08:35.5550342Z # gpt-oss 2025-11-03T17:08:35.5550500Z click==8.1.7 2025-11-03T17:08:35.5550671Z  # via 2025-11-03T17:08:35.5550829Z # -c snapshot_constraint.txt 2025-11-03T17:08:35.5551029Z # black 2025-11-03T17:08:35.5551181Z # click-plugins 2025-11-03T17:08:35.5551338Z # cligj 2025-11-03T17:08:35.5551480Z # fiona 2025-11-03T17:08:35.5551620Z # flask 2025-11-03T17:08:35.5551754Z # jiwer 2025-11-03T17:08:35.5551906Z # mlflow-skinny 2025-11-03T17:08:35.5552072Z # nltk 2025-11-03T17:08:35.5552209Z # rasterio 2025-11-03T17:08:35.5552364Z # ray 2025-11-03T17:08:35.5552515Z # schemathesis 2025-11-03T17:08:35.5552676Z # typer 2025-11-03T17:08:35.5552847Z # uvicorn 2025-11-03T17:08:35.5553016Z click-plugins==1.1.1.2 2025-11-03T17:08:35.5553209Z  # via 2025-11-03T17:08:35.5553374Z # -c snapshot_constraint.txt 2025-11-03T17:08:35.5553575Z # fiona 2025-11-03T17:08:35.5553756Z # rasterio 2025-11-03T17:08:35.5553913Z cligj==0.7.2 2025-11-03T17:08:35.5554077Z  # via 2025-11-03T17:08:35.5554232Z # -c snapshot_constraint.txt 2025-11-03T17:08:35.5554416Z # fiona 2025-11-03T17:08:35.5554586Z # rasterio 2025-11-03T17:08:35.5554751Z cloudpickle==3.1.1 2025-11-03T17:08:35.5554929Z  # via 2025-11-03T17:08:35.5555085Z # -c snapshot_constraint.txt 2025-11-03T17:08:35.5555314Z # mlflow-skinny 2025-11-03T17:08:35.5555494Z colorama==0.4.6 2025-11-03T17:08:35.5555661Z  # via 2025-11-03T17:08:35.5555819Z # -c snapshot_constraint.txt 2025-11-03T17:08:35.5556008Z # sacrebleu 2025-11-03T17:08:35.5556162Z # schemathesis 2025-11-03T17:08:35.5556360Z # tqdm-multiprocess 2025-11-03T17:08:35.5556550Z colorful==0.5.6 2025-11-03T17:08:35.5556721Z  # via 2025-11-03T17:08:35.5556890Z # -c snapshot_constraint.txt 2025-11-03T17:08:35.5557106Z # ray 2025-11-03T17:08:35.5557262Z contourpy==1.3.0 2025-11-03T17:08:35.5557436Z  # via 2025-11-03T17:08:35.5557589Z # -c snapshot_constraint.txt 2025-11-03T17:08:35.5557813Z # matplotlib 2025-11-03T17:08:35.5557984Z coverage==7.10.6 2025-11-03T17:08:35.5558158Z  # via 2025-11-03T17:08:35.5558411Z # -c snapshot_constraint.txt 2025-11-03T17:08:35.5558637Z # pytest-cov 2025-11-03T17:08:35.5558811Z cramjam==2.9.0 2025-11-03T17:08:35.5558983Z  # via 2025-11-03T17:08:35.5559136Z # -c snapshot_constraint.txt 2025-11-03T17:08:35.5559365Z # fastparquet 2025-11-03T17:08:35.5559557Z cupy-cuda12x==13.6.0 2025-11-03T17:08:35.5559747Z  # via 2025-11-03T17:08:35.5559903Z # -c snapshot_constraint.txt 2025-11-03T17:08:35.5560116Z # ray 2025-11-03T17:08:35.5560267Z cycler==0.12.1 2025-11-03T17:08:35.5560437Z  # via 2025-11-03T17:08:35.5560590Z # -c snapshot_constraint.txt 2025-11-03T17:08:35.5560812Z # matplotlib 2025-11-03T17:08:35.5560997Z databricks-sdk==0.59.0 2025-11-03T17:08:35.5561186Z  # via 2025-11-03T17:08:35.5561338Z # -c snapshot_constraint.txt 2025-11-03T17:08:35.5561565Z # mlflow-skinny 2025-11-03T17:08:35.5561778Z datamodel-code-generator==0.26.3 2025-11-03T17:08:35.5562004Z  # via 2025-11-03T17:08:35.5562157Z # -c snapshot_constraint.txt 2025-11-03T17:08:35.5562394Z # -r requirements/test.in 2025-11-03T17:08:35.5562600Z dataproperty==1.0.1 2025-11-03T17:08:35.5562868Z  # via 2025-11-03T17:08:35.5563028Z # -c snapshot_constraint.txt 2025-11-03T17:08:35.5563222Z # pytablewriter 2025-11-03T17:08:35.5563417Z # tabledata 2025-11-03T17:08:35.5563584Z datasets==3.0.2 2025-11-03T17:08:35.5563755Z  # via 2025-11-03T17:08:35.5563910Z # -c snapshot_constraint.txt 2025-11-03T17:08:35.5564097Z # evaluate 2025-11-03T17:08:35.5564240Z # lm-eval 2025-11-03T17:08:35.5564408Z # mteb 2025-11-03T17:08:35.5564564Z decorator==5.1.1 2025-11-03T17:08:35.5564737Z  # via 2025-11-03T17:08:35.5564888Z # -c snapshot_constraint.txt 2025-11-03T17:08:35.5565101Z # librosa 2025-11-03T17:08:35.5565259Z decord==0.6.0 2025-11-03T17:08:35.5565424Z  # via 2025-11-03T17:08:35.5565581Z # -c snapshot_constraint.txt 2025-11-03T17:08:35.5565818Z # -r requirements/test.in 2025-11-03T17:08:35.5566018Z dill==0.3.8 2025-11-03T17:08:35.5566183Z  # via 2025-11-03T17:08:35.5566338Z # -c snapshot_constraint.txt 2025-11-03T17:08:35.5566528Z # datasets 2025-11-03T17:08:35.5566679Z # evaluate 2025-11-03T17:08:35.5566825Z # lm-eval 2025-11-03T17:08:35.5567002Z # multiprocess 2025-11-03T17:08:35.5567181Z distlib==0.3.9 2025-11-03T17:08:35.5567350Z  # via 2025-11-03T17:08:35.5567506Z # -c snapshot_constraint.txt 2025-11-03T17:08:35.5567721Z # virtualenv 2025-11-03T17:08:35.5567887Z dnspython==2.7.0 2025-11-03T17:08:35.5568056Z  # via 2025-11-03T17:08:35.5568212Z # -c snapshot_constraint.txt 2025-11-03T17:08:35.5568560Z # email-validator 2025-11-03T17:08:35.5568750Z docker==7.1.0 2025-11-03T17:08:35.5568921Z  # via 2025-11-03T17:08:35.5569071Z # -c snapshot_constraint.txt 2025-11-03T17:08:35.5569267Z # gpt-oss 2025-11-03T17:08:35.5569441Z # mlflow 2025-11-03T17:08:35.5569600Z docopt==0.6.2 2025-11-03T17:08:35.5569763Z  # via 2025-11-03T17:08:35.5569926Z # -c snapshot_constraint.txt 2025-11-03T17:08:35.5570146Z # num2words 2025-11-03T17:08:35.5570334Z docstring-parser==0.17.0 2025-11-03T17:08:35.5570530Z  # via 2025-11-03T17:08:35.5570687Z # -c snapshot_constraint.txt 2025-11-03T17:08:35.5570912Z # jsonargparse 2025-11-03T17:08:35.5571105Z efficientnet-pytorch==0.7.1 2025-11-03T17:08:35.5571362Z  # via segmentation-models-pytorch 2025-11-03T17:08:35.5571597Z einops==0.8.1 2025-11-03T17:08:35.5571763Z  # via 2025-11-03T17:08:35.5571926Z # -c snapshot_constraint.txt 2025-11-03T17:08:35.5572130Z # -r requirements/test.in 2025-11-03T17:08:35.5572314Z # encodec 2025-11-03T17:08:35.5572462Z # terratorch 2025-11-03T17:08:35.5572613Z # torchgeo 2025-11-03T17:08:35.5572866Z # vector-quantize-pytorch 2025-11-03T17:08:35.5573080Z # vocos 2025-11-03T17:08:35.5573233Z einx==0.3.0 2025-11-03T17:08:35.5573397Z  # via 2025-11-03T17:08:35.5573555Z # -c snapshot_constraint.txt 2025-11-03T17:08:35.5573812Z # vector-quantize-pytorch 2025-11-03T17:08:35.5574025Z email-validator==2.2.0 2025-11-03T17:08:35.5574207Z  # via 2025-11-03T17:08:35.5574365Z # -c snapshot_constraint.txt 2025-11-03T17:08:35.5574584Z # pydantic 2025-11-03T17:08:35.5574749Z encodec==0.1.1 2025-11-03T17:08:35.5574910Z  # via 2025-11-03T17:08:35.5575072Z # -c snapshot_constraint.txt 2025-11-03T17:08:35.5575286Z # vocos 2025-11-03T17:08:35.5575467Z eval-type-backport==0.2.2 2025-11-03T17:08:35.5575675Z  # via 2025-11-03T17:08:35.5575838Z # -c snapshot_constraint.txt 2025-11-03T17:08:35.5576056Z # mteb 2025-11-03T17:08:35.5576210Z evaluate==0.4.3 2025-11-03T17:08:35.5576375Z  # via 2025-11-03T17:08:35.5576538Z # -c snapshot_constraint.txt 2025-11-03T17:08:35.5576755Z # lm-eval 2025-11-03T17:08:35.5576916Z fastapi==0.116.1 2025-11-03T17:08:35.5577085Z  # via 2025-11-03T17:08:35.5577350Z # -c snapshot_constraint.txt 2025-11-03T17:08:35.5577547Z # gpt-oss 2025-11-03T17:08:35.5577739Z # mlflow-skinny 2025-11-03T17:08:35.5577919Z fastparquet==2024.11.0 2025-11-03T17:08:35.5578106Z  # via 2025-11-03T17:08:35.5578265Z # -c snapshot_constraint.txt 2025-11-03T17:08:35.5578488Z # genai-perf 2025-11-03T17:08:35.5578652Z fastrlock==0.8.2 2025-11-03T17:08:35.5578837Z  # via 2025-11-03T17:08:35.5578997Z # -c snapshot_constraint.txt 2025-11-03T17:08:35.5579224Z # cupy-cuda12x 2025-11-03T17:08:35.5579401Z fastsafetensors==0.1.10 2025-11-03T17:08:35.5579588Z  # via 2025-11-03T17:08:35.5579743Z # -c snapshot_constraint.txt 2025-11-03T17:08:35.5579982Z # -r requirements/test.in 2025-11-03T17:08:35.5580182Z filelock==3.16.1 2025-11-03T17:08:35.5580356Z  # via 2025-11-03T17:08:35.5580521Z # -c snapshot_constraint.txt 2025-11-03T17:08:35.5580701Z # blobfile 2025-11-03T17:08:35.5580847Z # datasets 2025-11-03T17:08:35.5581004Z # huggingface-hub 2025-11-03T17:08:35.5581176Z # ray 2025-11-03T17:08:35.5581310Z # torch 2025-11-03T17:08:35.5581457Z # transformers 2025-11-03T17:08:35.5581651Z # virtualenv 2025-11-03T17:08:35.5581814Z fiona==1.10.1 2025-11-03T17:08:35.5581975Z  # via 2025-11-03T17:08:35.5582132Z # -c snapshot_constraint.txt 2025-11-03T17:08:35.5582352Z # torchgeo 2025-11-03T17:08:35.5582513Z flask==3.1.1 2025-11-03T17:08:35.5582675Z  # via 2025-11-03T17:08:35.5582832Z # -c snapshot_constraint.txt 2025-11-03T17:08:35.5583057Z # mlflow 2025-11-03T17:08:35.5583222Z fonttools==4.55.0 2025-11-03T17:08:35.5583404Z  # via 2025-11-03T17:08:35.5583568Z # -c snapshot_constraint.txt 2025-11-03T17:08:35.5583796Z # matplotlib 2025-11-03T17:08:35.5583957Z fqdn==1.5.1 2025-11-03T17:08:35.5584115Z  # via 2025-11-03T17:08:35.5584271Z # -c snapshot_constraint.txt 2025-11-03T17:08:35.5584509Z # jsonschema 2025-11-03T17:08:35.5584677Z frozendict==2.4.6 2025-11-03T17:08:35.5584848Z  # via 2025-11-03T17:08:35.5585007Z # -c snapshot_constraint.txt 2025-11-03T17:08:35.5585222Z # einx 2025-11-03T17:08:35.5585373Z frozenlist==1.5.0 2025-11-03T17:08:35.5585549Z  # via 2025-11-03T17:08:35.5585707Z # -c snapshot_constraint.txt 2025-11-03T17:08:35.5585897Z # aiohttp 2025-11-03T17:08:35.5586068Z # aiosignal 2025-11-03T17:08:35.5586237Z fsspec==2024.9.0 2025-11-03T17:08:35.5586414Z  # via 2025-11-03T17:08:35.5586577Z # -c snapshot_constraint.txt 2025-11-03T17:08:35.5586760Z # datasets 2025-11-03T17:08:35.5586911Z # evaluate 2025-11-03T17:08:35.5587062Z # fastparquet 2025-11-03T17:08:35.5587319Z # huggingface-hub 2025-11-03T17:08:35.5587475Z # lightning 2025-11-03T17:08:35.5587635Z # pytorch-lightning 2025-11-03T17:08:35.5587836Z # torch 2025-11-03T17:08:35.5587988Z ftfy==6.3.1 2025-11-03T17:08:35.5588148Z  # via 2025-11-03T17:08:35.5588306Z # -c snapshot_constraint.txt 2025-11-03T17:08:35.5588550Z # open-clip-torch 2025-11-03T17:08:35.5588741Z genai-perf==0.0.8 2025-11-03T17:08:35.5588910Z  # via 2025-11-03T17:08:35.5589066Z # -c snapshot_constraint.txt 2025-11-03T17:08:35.5589303Z # -r requirements/test.in 2025-11-03T17:08:35.5589502Z genson==1.3.0 2025-11-03T17:08:35.5589662Z  # via 2025-11-03T17:08:35.5589818Z # -c snapshot_constraint.txt 2025-11-03T17:08:35.5590067Z # datamodel-code-generator 2025-11-03T17:08:35.5590271Z geopandas==1.0.1 2025-11-03T17:08:35.5590458Z  # via 2025-11-03T17:08:35.5590617Z # -c snapshot_constraint.txt 2025-11-03T17:08:35.5590842Z # terratorch 2025-11-03T17:08:35.5591012Z gitdb==4.0.12 2025-11-03T17:08:35.5591176Z  # via 2025-11-03T17:08:35.5591333Z # -c snapshot_constraint.txt 2025-11-03T17:08:35.5591551Z # gitpython 2025-11-03T17:08:35.5591798Z gitpython==3.1.44 2025-11-03T17:08:35.5591984Z  # via 2025-11-03T17:08:35.5592142Z # -c snapshot_constraint.txt 2025-11-03T17:08:35.5592369Z # mlflow-skinny 2025-11-03T17:08:35.5592552Z google-api-core==2.24.2 2025-11-03T17:08:35.5592754Z  # via 2025-11-03T17:08:35.5592916Z # -c snapshot_constraint.txt 2025-11-03T17:08:35.5593114Z # google-cloud-core 2025-11-03T17:08:35.5593290Z # google-cloud-storage 2025-11-03T17:08:35.5593508Z # opencensus 2025-11-03T17:08:35.5593680Z google-auth==2.40.2 2025-11-03T17:08:35.5593859Z  # via 2025-11-03T17:08:35.5594012Z # -c snapshot_constraint.txt 2025-11-03T17:08:35.5594207Z # databricks-sdk 2025-11-03T17:08:35.5594389Z # google-api-core 2025-11-03T17:08:35.5594569Z # google-cloud-core 2025-11-03T17:08:35.5594746Z # google-cloud-storage 2025-11-03T17:08:35.5594987Z # runai-model-streamer-gcs 2025-11-03T17:08:35.5595206Z google-cloud-core==2.4.3 2025-11-03T17:08:35.5595405Z  # via 2025-11-03T17:08:35.5595564Z # -c snapshot_constraint.txt 2025-11-03T17:08:35.5595806Z # google-cloud-storage 2025-11-03T17:08:35.5596019Z google-cloud-storage==3.4.0 2025-11-03T17:08:35.5596228Z  # via 2025-11-03T17:08:35.5596379Z # -c snapshot_constraint.txt 2025-11-03T17:08:35.5596628Z # runai-model-streamer-gcs 2025-11-03T17:08:35.5596841Z google-crc32c==1.7.1 2025-11-03T17:08:35.5597025Z  # via 2025-11-03T17:08:35.5597175Z # -c snapshot_constraint.txt 2025-11-03T17:08:35.5597373Z # google-cloud-storage 2025-11-03T17:08:35.5597602Z # google-resumable-media 2025-11-03T17:08:35.5597825Z google-resumable-media==2.7.2 2025-11-03T17:08:35.5598039Z  # via 2025-11-03T17:08:35.5598200Z # -c snapshot_constraint.txt 2025-11-03T17:08:35.5598436Z # google-cloud-storage 2025-11-03T17:08:35.5598662Z googleapis-common-protos==1.70.0 2025-11-03T17:08:35.5598878Z  # via 2025-11-03T17:08:35.5599037Z # -c snapshot_constraint.txt 2025-11-03T17:08:35.5599265Z # google-api-core 2025-11-03T17:08:35.5599451Z gpt-oss==0.0.8 2025-11-03T17:08:35.5599625Z  # via 2025-11-03T17:08:35.5599786Z # -c snapshot_constraint.txt 2025-11-03T17:08:35.5600021Z # -r requirements/test.in 2025-11-03T17:08:35.5600216Z graphene==3.4.3 2025-11-03T17:08:35.5600391Z  # via 2025-11-03T17:08:35.5600547Z # -c snapshot_constraint.txt 2025-11-03T17:08:35.5600762Z # mlflow 2025-11-03T17:08:35.5600918Z graphql-core==3.2.6 2025-11-03T17:08:35.5601098Z  # via 2025-11-03T17:08:35.5601255Z # -c snapshot_constraint.txt 2025-11-03T17:08:35.5601445Z # graphene 2025-11-03T17:08:35.5601597Z # graphql-relay 2025-11-03T17:08:35.5601908Z # hypothesis-graphql 2025-11-03T17:08:35.5602106Z graphql-relay==3.2.0 2025-11-03T17:08:35.5602290Z  # via 2025-11-03T17:08:35.5602443Z # -c snapshot_constraint.txt 2025-11-03T17:08:35.5602667Z # graphene 2025-11-03T17:08:35.5602831Z greenlet==3.2.3 2025-11-03T17:08:35.5603006Z  # via 2025-11-03T17:08:35.5603160Z # -c snapshot_constraint.txt 2025-11-03T17:08:35.5603380Z # sqlalchemy 2025-11-03T17:08:35.5603548Z grpcio==1.71.0 2025-11-03T17:08:35.5603720Z  # via 2025-11-03T17:08:35.5603871Z # -c snapshot_constraint.txt 2025-11-03T17:08:35.5604087Z # ray 2025-11-03T17:08:35.5604243Z gunicorn==23.0.0 2025-11-03T17:08:35.5604415Z  # via 2025-11-03T17:08:35.5604568Z # -c snapshot_constraint.txt 2025-11-03T17:08:35.5604786Z # mlflow 2025-11-03T17:08:35.5604940Z h11==0.14.0 2025-11-03T17:08:35.5605095Z  # via 2025-11-03T17:08:35.5605249Z # -c snapshot_constraint.txt 2025-11-03T17:08:35.5605449Z # httpcore 2025-11-03T17:08:35.5605627Z # uvicorn 2025-11-03T17:08:35.5605783Z h5py==3.13.0 2025-11-03T17:08:35.5605949Z  # via 2025-11-03T17:08:35.5618204Z # -c snapshot_constraint.txt 2025-11-03T17:08:35.5618553Z # terratorch 2025-11-03T17:08:35.5618749Z harfile==0.3.0 2025-11-03T17:08:35.5618940Z  # via 2025-11-03T17:08:35.5619112Z # -c snapshot_constraint.txt 2025-11-03T17:08:35.5619361Z # schemathesis 2025-11-03T17:08:35.5619554Z hf-xet==1.1.7 2025-11-03T17:08:35.5619728Z  # via 2025-11-03T17:08:35.5619900Z # -c snapshot_constraint.txt 2025-11-03T17:08:35.5620152Z # huggingface-hub 2025-11-03T17:08:35.5620349Z hiredis==3.0.0 2025-11-03T17:08:35.5620515Z  # via 2025-11-03T17:08:35.5620683Z # -c snapshot_constraint.txt 2025-11-03T17:08:35.5620916Z # tensorizer 2025-11-03T17:08:35.5621100Z html2text==2025.4.15 2025-11-03T17:08:35.5621281Z  # via 2025-11-03T17:08:35.5621453Z # -c snapshot_constraint.txt 2025-11-03T17:08:35.5621682Z # gpt-oss 2025-11-03T17:08:35.5621848Z httpcore==1.0.6 2025-11-03T17:08:35.5622021Z  # via 2025-11-03T17:08:35.5622190Z # -c snapshot_constraint.txt 2025-11-03T17:08:35.5622406Z # httpx 2025-11-03T17:08:35.5622560Z httpx==0.27.2 2025-11-03T17:08:35.5622731Z  # via 2025-11-03T17:08:35.5622895Z # -c snapshot_constraint.txt 2025-11-03T17:08:35.5623118Z # -r requirements/test.in 2025-11-03T17:08:35.5623341Z # schemathesis 2025-11-03T17:08:35.5623529Z huggingface-hub==0.34.3 2025-11-03T17:08:35.5623723Z  # via 2025-11-03T17:08:35.5623882Z # -c snapshot_constraint.txt 2025-11-03T17:08:35.5624073Z # accelerate 2025-11-03T17:08:35.5624229Z # datasets 2025-11-03T17:08:35.5624384Z # evaluate 2025-11-03T17:08:35.5624542Z # open-clip-torch 2025-11-03T17:08:35.5624729Z # peft 2025-11-03T17:08:35.5624908Z # segmentation-models-pytorch 2025-11-03T17:08:35.5625139Z # sentence-transformers 2025-11-03T17:08:35.5625330Z # terratorch 2025-11-03T17:08:35.5625488Z # timm 2025-11-03T17:08:35.5625637Z # tokenizers 2025-11-03T17:08:35.5625800Z # transformers 2025-11-03T17:08:35.5625987Z # vocos 2025-11-03T17:08:35.5626146Z humanize==4.11.0 2025-11-03T17:08:35.5626327Z  # via 2025-11-03T17:08:35.5626496Z # -c snapshot_constraint.txt 2025-11-03T17:08:35.5626741Z # runai-model-streamer 2025-11-03T17:08:35.5626943Z hydra-core==1.3.2 2025-11-03T17:08:35.5627121Z  # via 2025-11-03T17:08:35.5627286Z # -c snapshot_constraint.txt 2025-11-03T17:08:35.5627472Z # lightly 2025-11-03T17:08:35.5627655Z # lightning 2025-11-03T17:08:35.5627831Z hypothesis==6.131.0 2025-11-03T17:08:35.5628010Z  # via 2025-11-03T17:08:35.5628164Z # -c snapshot_constraint.txt 2025-11-03T17:08:35.5628365Z # hypothesis-graphql 2025-11-03T17:08:35.5628556Z # hypothesis-jsonschema 2025-11-03T17:08:35.5628926Z # schemathesis 2025-11-03T17:08:35.5629110Z hypothesis-graphql==0.11.1 2025-11-03T17:08:35.5629310Z  # via 2025-11-03T17:08:35.5629475Z # -c snapshot_constraint.txt 2025-11-03T17:08:35.5629702Z # schemathesis 2025-11-03T17:08:35.5629884Z hypothesis-jsonschema==0.23.1 2025-11-03T17:08:35.5630094Z  # via 2025-11-03T17:08:35.5630257Z # -c snapshot_constraint.txt 2025-11-03T17:08:35.5630480Z # schemathesis 2025-11-03T17:08:35.5630644Z idna==3.10 2025-11-03T17:08:35.5630809Z  # via 2025-11-03T17:08:35.5630969Z # -c snapshot_constraint.txt 2025-11-03T17:08:35.5631161Z # anyio 2025-11-03T17:08:35.5631310Z # email-validator 2025-11-03T17:08:35.5631475Z # httpx 2025-11-03T17:08:35.5631621Z # jsonschema 2025-11-03T17:08:35.5631767Z # requests 2025-11-03T17:08:35.5631941Z # yarl 2025-11-03T17:08:35.5632097Z imageio==2.37.0 2025-11-03T17:08:35.5632273Z  # via 2025-11-03T17:08:35.5632433Z # -c snapshot_constraint.txt 2025-11-03T17:08:35.5632659Z # scikit-image 2025-11-03T17:08:35.5632844Z importlib-metadata==8.7.0 2025-11-03T17:08:35.5633040Z  # via 2025-11-03T17:08:35.5633281Z # -c snapshot_constraint.txt 2025-11-03T17:08:35.5633487Z # mlflow-skinny 2025-11-03T17:08:35.5633703Z # opentelemetry-api 2025-11-03T17:08:35.5633905Z importlib-resources==6.5.2 2025-11-03T17:08:35.5634101Z  # via 2025-11-03T17:08:35.5634264Z # -c snapshot_constraint.txt 2025-11-03T17:08:35.5634498Z # typeshed-client 2025-11-03T17:08:35.5634685Z inflect==5.6.2 2025-11-03T17:08:35.5634853Z  # via 2025-11-03T17:08:35.5635015Z # -c snapshot_constraint.txt 2025-11-03T17:08:35.5635270Z # datamodel-code-generator 2025-11-03T17:08:35.5635488Z iniconfig==2.0.0 2025-11-03T17:08:35.5635659Z  # via 2025-11-03T17:08:35.5635831Z # -c snapshot_constraint.txt 2025-11-03T17:08:35.5636050Z # pytest 2025-11-03T17:08:35.5636218Z isoduration==20.11.0 2025-11-03T17:08:35.5636397Z  # via 2025-11-03T17:08:35.5636552Z # -c snapshot_constraint.txt 2025-11-03T17:08:35.5636773Z # jsonschema 2025-11-03T17:08:35.5636940Z isort==5.13.2 2025-11-03T17:08:35.5637119Z  # via 2025-11-03T17:08:35.5637282Z # -c snapshot_constraint.txt 2025-11-03T17:08:35.5637526Z # datamodel-code-generator 2025-11-03T17:08:35.5637728Z itsdangerous==2.2.0 2025-11-03T17:08:35.5637905Z  # via 2025-11-03T17:08:35.5638063Z # -c snapshot_constraint.txt 2025-11-03T17:08:35.5638279Z # flask 2025-11-03T17:08:35.5638427Z jinja2==3.1.6 2025-11-03T17:08:35.5638593Z  # via 2025-11-03T17:08:35.5638759Z # -c snapshot_constraint.txt 2025-11-03T17:08:35.5638967Z # datamodel-code-generator 2025-11-03T17:08:35.5639154Z # flask 2025-11-03T17:08:35.5639297Z # mlflow 2025-11-03T17:08:35.5639467Z # torch 2025-11-03T17:08:35.5639619Z jiwer==3.0.5 2025-11-03T17:08:35.5639784Z  # via 2025-11-03T17:08:35.5639944Z # -c snapshot_constraint.txt 2025-11-03T17:08:35.5640187Z # -r requirements/test.in 2025-11-03T17:08:35.5640401Z jmespath==1.0.1 2025-11-03T17:08:35.5640578Z  # via 2025-11-03T17:08:35.5640737Z # -c snapshot_constraint.txt 2025-11-03T17:08:35.5640926Z # boto3 2025-11-03T17:08:35.5641097Z # botocore 2025-11-03T17:08:35.5641256Z joblib==1.4.2 2025-11-03T17:08:35.5641424Z  # via 2025-11-03T17:08:35.5641586Z # -c snapshot_constraint.txt 2025-11-03T17:08:35.5641776Z # librosa 2025-11-03T17:08:35.5641918Z # nltk 2025-11-03T17:08:35.5642102Z # scikit-learn 2025-11-03T17:08:35.5642285Z jsonargparse==4.35.0 2025-11-03T17:08:35.5642465Z  # via 2025-11-03T17:08:35.5642625Z # -c snapshot_constraint.txt 2025-11-03T17:08:35.5642825Z # lightning 2025-11-03T17:08:35.5643016Z # terratorch 2025-11-03T17:08:35.5643181Z jsonlines==4.0.0 2025-11-03T17:08:35.5643469Z  # via 2025-11-03T17:08:35.5643632Z # -c snapshot_constraint.txt 2025-11-03T17:08:35.5643861Z # lm-eval 2025-11-03T17:08:35.5644024Z jsonpointer==3.0.0 2025-11-03T17:08:35.5644211Z  # via 2025-11-03T17:08:35.5644367Z # -c snapshot_constraint.txt 2025-11-03T17:08:35.5644588Z # jsonschema 2025-11-03T17:08:35.5644750Z jsonschema==4.23.0 2025-11-03T17:08:35.5644929Z  # via 2025-11-03T17:08:35.5645085Z # -c snapshot_constraint.txt 2025-11-03T17:08:35.5645290Z # hypothesis-jsonschema 2025-11-03T17:08:35.5645474Z # mistral-common 2025-11-03T17:08:35.5645638Z # ray 2025-11-03T17:08:35.5645816Z # schemathesis 2025-11-03T17:08:35.5646009Z jsonschema-specifications==2024.10.1 2025-11-03T17:08:35.5646230Z  # via 2025-11-03T17:08:35.5646390Z # -c snapshot_constraint.txt 2025-11-03T17:08:35.5646612Z # jsonschema 2025-11-03T17:08:35.5646781Z junit-xml==1.9 2025-11-03T17:08:35.5646960Z  # via 2025-11-03T17:08:35.5647121Z # -c snapshot_constraint.txt 2025-11-03T17:08:35.5647348Z # schemathesis 2025-11-03T17:08:35.5647520Z kaleido==0.2.1 2025-11-03T17:08:35.5647761Z  # via 2025-11-03T17:08:35.5647919Z # -c snapshot_constraint.txt 2025-11-03T17:08:35.5648147Z # genai-perf 2025-11-03T17:08:35.5648308Z kiwisolver==1.4.7 2025-11-03T17:08:35.5648548Z  # via 2025-11-03T17:08:35.5648698Z # -c snapshot_constraint.txt 2025-11-03T17:08:35.5648926Z # matplotlib 2025-11-03T17:08:35.5649091Z kornia==0.8.1 2025-11-03T17:08:35.5649251Z  # via 2025-11-03T17:08:35.5649399Z # -c snapshot_constraint.txt 2025-11-03T17:08:35.5649612Z # torchgeo 2025-11-03T17:08:35.5649772Z kornia-rs==0.1.9 2025-11-03T17:08:35.5649944Z  # via 2025-11-03T17:08:35.5650101Z # -c snapshot_constraint.txt 2025-11-03T17:08:35.5650311Z # kornia 2025-11-03T17:08:35.5650468Z lazy-loader==0.4 2025-11-03T17:08:35.5650637Z  # via 2025-11-03T17:08:35.5650788Z # -c snapshot_constraint.txt 2025-11-03T17:08:35.5650971Z # librosa 2025-11-03T17:08:35.5651158Z # scikit-image 2025-11-03T17:08:35.5651325Z libnacl==2.1.0 2025-11-03T17:08:35.5651488Z  # via 2025-11-03T17:08:35.5651643Z # -c snapshot_constraint.txt 2025-11-03T17:08:35.5651856Z # tensorizer 2025-11-03T17:08:35.5652016Z librosa==0.10.2.post1 2025-11-03T17:08:35.5652195Z  # via 2025-11-03T17:08:35.5652346Z # -c snapshot_constraint.txt 2025-11-03T17:08:35.5652579Z # -r requirements/test.in 2025-11-03T17:08:35.5652766Z lightly==1.5.20 2025-11-03T17:08:35.5652932Z  # via 2025-11-03T17:08:35.5653095Z # -c snapshot_constraint.txt 2025-11-03T17:08:35.5653282Z # terratorch 2025-11-03T17:08:35.5653455Z # torchgeo 2025-11-03T17:08:35.5653616Z lightly-utils==0.0.2 2025-11-03T17:08:35.5653793Z  # via 2025-11-03T17:08:35.5653944Z # -c snapshot_constraint.txt 2025-11-03T17:08:35.5654153Z # lightly 2025-11-03T17:08:35.5654313Z lightning==2.5.1.post0 2025-11-03T17:08:35.5654497Z  # via 2025-11-03T17:08:35.5654651Z # -c snapshot_constraint.txt 2025-11-03T17:08:35.5654841Z # terratorch 2025-11-03T17:08:35.5655019Z # torchgeo 2025-11-03T17:08:35.5655188Z lightning-utilities==0.14.3 2025-11-03T17:08:35.5655389Z  # via 2025-11-03T17:08:35.5655551Z # -c snapshot_constraint.txt 2025-11-03T17:08:35.5655738Z # lightning 2025-11-03T17:08:35.5655898Z # pytorch-lightning 2025-11-03T17:08:35.5656098Z # torchmetrics 2025-11-03T17:08:35.5656263Z llvmlite==0.44.0 2025-11-03T17:08:35.5656433Z  # via 2025-11-03T17:08:35.5656587Z # -c snapshot_constraint.txt 2025-11-03T17:08:35.5656793Z # numba 2025-11-03T17:08:35.5657187Z lm-eval @ git+https://github.com/EleutherAI/lm-evaluation-harness.git@206b7722158f58c35b7ffcd53b035fdbdda5126d 2025-11-03T17:08:35.5657626Z  # via 2025-11-03T17:08:35.5657871Z # -c snapshot_constraint.txt 2025-11-03T17:08:35.5658114Z # -r requirements/test.in 2025-11-03T17:08:35.5658307Z lxml==5.3.0 2025-11-03T17:08:35.5658468Z  # via 2025-11-03T17:08:35.5658627Z # -c snapshot_constraint.txt 2025-11-03T17:08:35.5658809Z # blobfile 2025-11-03T17:08:35.5658965Z # gpt-oss 2025-11-03T17:08:35.5659140Z # sacrebleu 2025-11-03T17:08:35.5659300Z mako==1.3.10 2025-11-03T17:08:35.5659457Z  # via 2025-11-03T17:08:35.5659613Z # -c snapshot_constraint.txt 2025-11-03T17:08:35.5659823Z # alembic 2025-11-03T17:08:35.5659978Z markdown==3.8.2 2025-11-03T17:08:35.5660263Z  # via 2025-11-03T17:08:35.5660418Z # -c snapshot_constraint.txt 2025-11-03T17:08:35.5660628Z # mlflow 2025-11-03T17:08:35.5660789Z markdown-it-py==3.0.0 2025-11-03T17:08:35.5660975Z  # via 2025-11-03T17:08:35.5661133Z # -c snapshot_constraint.txt 2025-11-03T17:08:35.5661356Z # rich 2025-11-03T17:08:35.5661513Z markupsafe==3.0.1 2025-11-03T17:08:35.5661680Z  # via 2025-11-03T17:08:35.5661834Z # -c snapshot_constraint.txt 2025-11-03T17:08:35.5662017Z # flask 2025-11-03T17:08:35.5662232Z # jinja2 2025-11-03T17:08:35.5662367Z # mako 2025-11-03T17:08:35.5662553Z # werkzeug 2025-11-03T17:08:35.5662718Z matplotlib==3.9.2 2025-11-03T17:08:35.5662889Z  # via 2025-11-03T17:08:35.5663037Z # -c snapshot_constraint.txt 2025-11-03T17:08:35.5663239Z # -r requirements/test.in 2025-11-03T17:08:35.5663428Z # lightning 2025-11-03T17:08:35.5663575Z # mlflow 2025-11-03T17:08:35.5663728Z # pycocotools 2025-11-03T17:08:35.5663922Z # torchgeo 2025-11-03T17:08:35.5664090Z mbstrdecoder==1.1.3 2025-11-03T17:08:35.5664259Z  # via 2025-11-03T17:08:35.5664417Z # -c snapshot_constraint.txt 2025-11-03T17:08:35.5664615Z # dataproperty 2025-11-03T17:08:35.5664782Z # pytablewriter 2025-11-03T17:08:35.5664968Z # typepy 2025-11-03T17:08:35.5665120Z mdurl==0.1.2 2025-11-03T17:08:35.5665283Z  # via 2025-11-03T17:08:35.5665444Z # -c snapshot_constraint.txt 2025-11-03T17:08:35.5665677Z # markdown-it-py 2025-11-03T17:08:35.5665862Z mistral-common==1.8.5 2025-11-03T17:08:35.5666041Z  # via 2025-11-03T17:08:35.5666203Z # -c snapshot_constraint.txt 2025-11-03T17:08:35.5666436Z # -r requirements/test.in 2025-11-03T17:08:35.5666632Z mlflow==2.22.0 2025-11-03T17:08:35.5666799Z  # via 2025-11-03T17:08:35.5666952Z # -c snapshot_constraint.txt 2025-11-03T17:08:35.5667165Z # terratorch 2025-11-03T17:08:35.5667336Z mlflow-skinny==2.22.0 2025-11-03T17:08:35.5667524Z  # via 2025-11-03T17:08:35.5667685Z # -c snapshot_constraint.txt 2025-11-03T17:08:35.5667894Z # mlflow 2025-11-03T17:08:35.5668049Z more-itertools==10.5.0 2025-11-03T17:08:35.5668237Z  # via 2025-11-03T17:08:35.5668391Z # -c snapshot_constraint.txt 2025-11-03T17:08:35.5668604Z # lm-eval 2025-11-03T17:08:35.5668762Z mpmath==1.3.0 2025-11-03T17:08:35.5668927Z  # via 2025-11-03T17:08:35.5669075Z # -c snapshot_constraint.txt 2025-11-03T17:08:35.5669290Z # sympy 2025-11-03T17:08:35.5669442Z msgpack==1.1.0 2025-11-03T17:08:35.5669605Z  # via 2025-11-03T17:08:35.5669753Z # -c snapshot_constraint.txt 2025-11-03T17:08:35.5669951Z # librosa 2025-11-03T17:08:35.5670124Z # ray 2025-11-03T17:08:35.5670275Z mteb==1.38.11 2025-11-03T17:08:35.5670430Z  # via 2025-11-03T17:08:35.5670583Z # -c snapshot_constraint.txt 2025-11-03T17:08:35.5670817Z # -r requirements/test.in 2025-11-03T17:08:35.5671024Z multidict==6.1.0 2025-11-03T17:08:35.5671189Z  # via 2025-11-03T17:08:35.5671340Z # -c snapshot_constraint.txt 2025-11-03T17:08:35.5671522Z # aiohttp 2025-11-03T17:08:35.5671693Z # yarl 2025-11-03T17:08:35.5671855Z multiprocess==0.70.16 2025-11-03T17:08:35.5672126Z  # via 2025-11-03T17:08:35.5672295Z # -c snapshot_constraint.txt 2025-11-03T17:08:35.5672481Z # datasets 2025-11-03T17:08:35.5672653Z # evaluate 2025-11-03T17:08:35.5672812Z munch==4.0.0 2025-11-03T17:08:35.5672977Z  # via 2025-11-03T17:08:35.5673130Z # -c snapshot_constraint.txt 2025-11-03T17:08:35.5673366Z # pretrainedmodels 2025-11-03T17:08:35.5673560Z mypy-extensions==1.0.0 2025-11-03T17:08:35.5673743Z  # via 2025-11-03T17:08:35.5673901Z # -c snapshot_constraint.txt 2025-11-03T17:08:35.5674108Z # black 2025-11-03T17:08:35.5674270Z networkx==3.2.1 2025-11-03T17:08:35.5674441Z  # via 2025-11-03T17:08:35.5674589Z # -c snapshot_constraint.txt 2025-11-03T17:08:35.5674775Z # scikit-image 2025-11-03T17:08:35.5674952Z # torch 2025-11-03T17:08:35.5675100Z nltk==3.9.1 2025-11-03T17:08:35.5675257Z  # via 2025-11-03T17:08:35.5675410Z # -c snapshot_constraint.txt 2025-11-03T17:08:35.5675633Z # rouge-score 2025-11-03T17:08:35.5675804Z num2words==0.5.14 2025-11-03T17:08:35.5675968Z  # via 2025-11-03T17:08:35.5676120Z # -c snapshot_constraint.txt 2025-11-03T17:08:35.5676440Z # -r requirements/test.in 2025-11-03T17:08:35.5676640Z numba==0.61.2 2025-11-03T17:08:35.5676813Z  # via 2025-11-03T17:08:35.5676964Z # -c snapshot_constraint.txt 2025-11-03T17:08:35.5677160Z # -r requirements/test.in 2025-11-03T17:08:35.5677367Z # librosa 2025-11-03T17:08:35.5677520Z numexpr==2.10.1 2025-11-03T17:08:35.5677698Z  # via 2025-11-03T17:08:35.5677855Z # -c snapshot_constraint.txt 2025-11-03T17:08:35.5678071Z # lm-eval 2025-11-03T17:08:35.5678227Z numpy==1.26.4 2025-11-03T17:08:35.5678391Z  # via 2025-11-03T17:08:35.5678553Z # -c snapshot_constraint.txt 2025-11-03T17:08:35.5678766Z # -r requirements/test.in 2025-11-03T17:08:35.5678949Z # accelerate 2025-11-03T17:08:35.5679104Z # albucore 2025-11-03T17:08:35.5679255Z # albumentations 2025-11-03T17:08:35.5679416Z # bitsandbytes 2025-11-03T17:08:35.5679563Z # bm25s 2025-11-03T17:08:35.5679708Z # contourpy 2025-11-03T17:08:35.5679865Z # cupy-cuda12x 2025-11-03T17:08:35.5680016Z # datasets 2025-11-03T17:08:35.5680164Z # decord 2025-11-03T17:08:35.5680302Z # einx 2025-11-03T17:08:35.5680437Z # encodec 2025-11-03T17:08:35.5680577Z # evaluate 2025-11-03T17:08:35.5680724Z # fastparquet 2025-11-03T17:08:35.5680877Z # genai-perf 2025-11-03T17:08:35.5681044Z # geopandas 2025-11-03T17:08:35.5681193Z # h5py 2025-11-03T17:08:35.5681333Z # imageio 2025-11-03T17:08:35.5681479Z # librosa 2025-11-03T17:08:35.5681617Z # lightly 2025-11-03T17:08:35.5681760Z # lightly-utils 2025-11-03T17:08:35.5681928Z # matplotlib 2025-11-03T17:08:35.5682093Z # mistral-common 2025-11-03T17:08:35.5682250Z # mlflow 2025-11-03T17:08:35.5682394Z # mteb 2025-11-03T17:08:35.5682534Z # numba 2025-11-03T17:08:35.5682676Z # numexpr 2025-11-03T17:08:35.5682832Z # opencv-python-headless 2025-11-03T17:08:35.5683019Z # pandas 2025-11-03T17:08:35.5683168Z # patsy 2025-11-03T17:08:35.5683307Z # peft 2025-11-03T17:08:35.5683445Z # pycocotools 2025-11-03T17:08:35.5683600Z # pyogrio 2025-11-03T17:08:35.5683742Z # rasterio 2025-11-03T17:08:35.5683885Z # rioxarray 2025-11-03T17:08:35.5684032Z # rouge-score 2025-11-03T17:08:35.5684206Z # runai-model-streamer 2025-11-03T17:08:35.5684390Z # sacrebleu 2025-11-03T17:08:35.5684539Z # scikit-image 2025-11-03T17:08:35.5684692Z # scikit-learn 2025-11-03T17:08:35.5684842Z # scipy 2025-11-03T17:08:35.5685009Z # segmentation-models-pytorch 2025-11-03T17:08:35.5685208Z # shapely 2025-11-03T17:08:35.5685343Z # soxr 2025-11-03T17:08:35.5685494Z # statsmodels 2025-11-03T17:08:35.5685648Z # tensorboardx 2025-11-03T17:08:35.5685895Z # tensorizer 2025-11-03T17:08:35.5686040Z # tifffile 2025-11-03T17:08:35.5686187Z # torchgeo 2025-11-03T17:08:35.5686337Z # torchmetrics 2025-11-03T17:08:35.5686490Z # torchvision 2025-11-03T17:08:35.5686640Z # transformers 2025-11-03T17:08:35.5686791Z # tritonclient 2025-11-03T17:08:35.5686939Z # vocos 2025-11-03T17:08:35.5687111Z # xarray 2025-11-03T17:08:35.5687266Z omegaconf==2.3.0 2025-11-03T17:08:35.5687433Z  # via 2025-11-03T17:08:35.5687594Z # -c snapshot_constraint.txt 2025-11-03T17:08:35.5687786Z # hydra-core 2025-11-03T17:08:35.5687966Z # lightning 2025-11-03T17:08:35.5688137Z open-clip-torch==2.32.0 2025-11-03T17:08:35.5688369Z  # via -r requirements/test.in 2025-11-03T17:08:35.5688646Z openai-harmony==0.0.4 2025-11-03T17:08:35.5688835Z  # via 2025-11-03T17:08:35.5688998Z # -c snapshot_constraint.txt 2025-11-03T17:08:35.5689217Z # gpt-oss 2025-11-03T17:08:35.5689379Z opencensus==0.11.4 2025-11-03T17:08:35.5689555Z  # via 2025-11-03T17:08:35.5689712Z # -c snapshot_constraint.txt 2025-11-03T17:08:35.5689927Z # ray 2025-11-03T17:08:35.5690185Z opencensus-context==0.1.3 2025-11-03T17:08:35.5690389Z  # via 2025-11-03T17:08:35.5690548Z # -c snapshot_constraint.txt 2025-11-03T17:08:35.5690768Z # opencensus 2025-11-03T17:08:35.5690962Z opencv-python-headless==4.11.0.86 2025-11-03T17:08:35.5691196Z  # via 2025-11-03T17:08:35.5691364Z # -c snapshot_constraint.txt 2025-11-03T17:08:35.5691563Z # -r requirements/test.in 2025-11-03T17:08:35.5691740Z # albucore 2025-11-03T17:08:35.5691890Z # albumentations 2025-11-03T17:08:35.5692091Z # mistral-common 2025-11-03T17:08:35.5692276Z opentelemetry-api==1.35.0 2025-11-03T17:08:35.5692472Z  # via 2025-11-03T17:08:35.5692628Z # -c snapshot_constraint.txt 2025-11-03T17:08:35.5692830Z # mlflow-skinny 2025-11-03T17:08:35.5693027Z # opentelemetry-exporter-prometheus 2025-11-03T17:08:35.5693256Z # opentelemetry-sdk 2025-11-03T17:08:35.5693506Z # opentelemetry-semantic-conventions 2025-11-03T17:08:35.5693774Z opentelemetry-exporter-prometheus==0.56b0 2025-11-03T17:08:35.5694013Z  # via 2025-11-03T17:08:35.5694175Z # -c snapshot_constraint.txt 2025-11-03T17:08:35.5694387Z # ray 2025-11-03T17:08:35.5694550Z opentelemetry-proto==1.36.0 2025-11-03T17:08:35.5694747Z  # via 2025-11-03T17:08:35.5694904Z # -c snapshot_constraint.txt 2025-11-03T17:08:35.5695115Z # ray 2025-11-03T17:08:35.5695276Z opentelemetry-sdk==1.35.0 2025-11-03T17:08:35.5695477Z  # via 2025-11-03T17:08:35.5695644Z # -c snapshot_constraint.txt 2025-11-03T17:08:35.5695838Z # mlflow-skinny 2025-11-03T17:08:35.5696028Z # opentelemetry-exporter-prometheus 2025-11-03T17:08:35.5696265Z # ray 2025-11-03T17:08:35.5696455Z opentelemetry-semantic-conventions==0.56b0 2025-11-03T17:08:35.5696701Z  # via 2025-11-03T17:08:35.5696877Z # -c snapshot_constraint.txt 2025-11-03T17:08:35.5697114Z # opentelemetry-sdk 2025-11-03T17:08:35.5697306Z packaging==24.2 2025-11-03T17:08:35.5697483Z  # via 2025-11-03T17:08:35.5697644Z # -c snapshot_constraint.txt 2025-11-03T17:08:35.5697831Z # accelerate 2025-11-03T17:08:35.5697987Z # black 2025-11-03T17:08:35.5698152Z # datamodel-code-generator 2025-11-03T17:08:35.5698343Z # datasets 2025-11-03T17:08:35.5698487Z # evaluate 2025-11-03T17:08:35.5698649Z # fastparquet 2025-11-03T17:08:35.5698810Z # geopandas 2025-11-03T17:08:35.5698960Z # gunicorn 2025-11-03T17:08:35.5699106Z # huggingface-hub 2025-11-03T17:08:35.5699277Z # hydra-core 2025-11-03T17:08:35.5699428Z # kornia 2025-11-03T17:08:35.5699566Z # lazy-loader 2025-11-03T17:08:35.5699716Z # lightning 2025-11-03T17:08:35.5699874Z # lightning-utilities 2025-11-03T17:08:35.5700150Z # matplotlib 2025-11-03T17:08:35.5700302Z # mlflow-skinny 2025-11-03T17:08:35.5700459Z # peft 2025-11-03T17:08:35.5700598Z # plotly 2025-11-03T17:08:35.5700738Z # pooch 2025-11-03T17:08:35.5700877Z # pyogrio 2025-11-03T17:08:35.5701023Z # pytest 2025-11-03T17:08:35.5701178Z # pytest-rerunfailures 2025-11-03T17:08:35.5701368Z # pytorch-lightning 2025-11-03T17:08:35.5701538Z # ray 2025-11-03T17:08:35.5701678Z # rioxarray 2025-11-03T17:08:35.5701838Z # scikit-image 2025-11-03T17:08:35.5701999Z # statsmodels 2025-11-03T17:08:35.5702147Z # tensorboardx 2025-11-03T17:08:35.5702307Z # torchmetrics 2025-11-03T17:08:35.5702462Z # transformers 2025-11-03T17:08:35.5702613Z # typepy 2025-11-03T17:08:35.5702781Z # xarray 2025-11-03T17:08:35.5702934Z pandas==2.2.3 2025-11-03T17:08:35.5703104Z  # via 2025-11-03T17:08:35.5703259Z # -c snapshot_constraint.txt 2025-11-03T17:08:35.5703451Z # datasets 2025-11-03T17:08:35.5703597Z # evaluate 2025-11-03T17:08:35.5703744Z # fastparquet 2025-11-03T17:08:35.5703892Z # genai-perf 2025-11-03T17:08:35.5704042Z # geopandas 2025-11-03T17:08:35.5704269Z # mlflow 2025-11-03T17:08:35.5704413Z # statsmodels 2025-11-03T17:08:35.5704556Z # torchgeo 2025-11-03T17:08:35.5704733Z # xarray 2025-11-03T17:08:35.5704887Z pathspec==0.12.1 2025-11-03T17:08:35.5705061Z  # via 2025-11-03T17:08:35.5705212Z # -c snapshot_constraint.txt 2025-11-03T17:08:35.5705439Z # black 2025-11-03T17:08:35.5705598Z pathvalidate==3.2.1 2025-11-03T17:08:35.5705772Z  # via 2025-11-03T17:08:35.5705926Z # -c snapshot_constraint.txt 2025-11-03T17:08:35.5706157Z # pytablewriter 2025-11-03T17:08:35.5706332Z patsy==1.0.1 2025-11-03T17:08:35.5706503Z  # via 2025-11-03T17:08:35.5706654Z # -c snapshot_constraint.txt 2025-11-03T17:08:35.5706874Z # statsmodels 2025-11-03T17:08:35.5707044Z peft==0.16.0 2025-11-03T17:08:35.5707207Z  # via 2025-11-03T17:08:35.5707360Z # -c snapshot_constraint.txt 2025-11-03T17:08:35.5707573Z # -r requirements/test.in 2025-11-03T17:08:35.5707790Z # lm-eval 2025-11-03T17:08:35.5707947Z pillow==10.4.0 2025-11-03T17:08:35.5708114Z  # via 2025-11-03T17:08:35.5708267Z # -c snapshot_constraint.txt 2025-11-03T17:08:35.5708455Z # genai-perf 2025-11-03T17:08:35.5708608Z # imageio 2025-11-03T17:08:35.5708762Z # lightly-utils 2025-11-03T17:08:35.5708935Z # matplotlib 2025-11-03T17:08:35.5709096Z # mistral-common 2025-11-03T17:08:35.5709259Z # scikit-image 2025-11-03T17:08:35.5709625Z # segmentation-models-pytorch 2025-11-03T17:08:35.5709859Z # sentence-transformers 2025-11-03T17:08:35.5710048Z # torchgeo 2025-11-03T17:08:35.5710235Z # torchvision 2025-11-03T17:08:35.5710419Z platformdirs==4.3.6 2025-11-03T17:08:35.5710601Z  # via 2025-11-03T17:08:35.5710769Z # -c snapshot_constraint.txt 2025-11-03T17:08:35.5710957Z # black 2025-11-03T17:08:35.5711103Z # pooch 2025-11-03T17:08:35.5711292Z # virtualenv 2025-11-03T17:08:35.5711469Z plotly==5.24.1 2025-11-03T17:08:35.5711638Z  # via 2025-11-03T17:08:35.5711800Z # -c snapshot_constraint.txt 2025-11-03T17:08:35.5712027Z # genai-perf 2025-11-03T17:08:35.5712193Z pluggy==1.5.0 2025-11-03T17:08:35.5712355Z  # via 2025-11-03T17:08:35.5712512Z # -c snapshot_constraint.txt 2025-11-03T17:08:35.5712702Z # pytest 2025-11-03T17:08:35.5712871Z # pytest-cov 2025-11-03T17:08:35.5713034Z polars==1.29.0 2025-11-03T17:08:35.5713202Z  # via 2025-11-03T17:08:35.5713358Z # -c snapshot_constraint.txt 2025-11-03T17:08:35.5713575Z # mteb 2025-11-03T17:08:35.5713737Z pooch==1.8.2 2025-11-03T17:08:35.5713909Z  # via 2025-11-03T17:08:35.5714075Z # -c snapshot_constraint.txt 2025-11-03T17:08:35.5714450Z # librosa 2025-11-03T17:08:35.5714627Z portalocker==2.10.1 2025-11-03T17:08:35.5714816Z  # via 2025-11-03T17:08:35.5714980Z # -c snapshot_constraint.txt 2025-11-03T17:08:35.5715203Z # sacrebleu 2025-11-03T17:08:35.5715375Z pqdm==0.2.0 2025-11-03T17:08:35.5715547Z  # via 2025-11-03T17:08:35.5715715Z # -c snapshot_constraint.txt 2025-11-03T17:08:35.5715964Z # -r requirements/test.in 2025-11-03T17:08:35.5716179Z pretrainedmodels==0.7.4 2025-11-03T17:08:35.5716375Z  # via 2025-11-03T17:08:35.5716535Z # -c snapshot_constraint.txt 2025-11-03T17:08:35.5716787Z # segmentation-models-pytorch 2025-11-03T17:08:35.5717022Z prometheus-client==0.22.0 2025-11-03T17:08:35.5717226Z  # via 2025-11-03T17:08:35.5717386Z # -c snapshot_constraint.txt 2025-11-03T17:08:35.5717608Z # opentelemetry-exporter-prometheus 2025-11-03T17:08:35.5717850Z # ray 2025-11-03T17:08:35.5718011Z propcache==0.2.0 2025-11-03T17:08:35.5718191Z  # via 2025-11-03T17:08:35.5718348Z # -c snapshot_constraint.txt 2025-11-03T17:08:35.5718414Z # aiohttp 2025-11-03T17:08:35.5718498Z # yarl 2025-11-03T17:08:35.5718679Z proto-plus==1.26.1 2025-11-03T17:08:35.5718769Z  # via 2025-11-03T17:08:35.5718854Z # -c snapshot_constraint.txt 2025-11-03T17:08:35.5718965Z # google-api-core 2025-11-03T17:08:35.5719030Z protobuf==5.28.3 2025-11-03T17:08:35.5719110Z  # via 2025-11-03T17:08:35.5719188Z # -c snapshot_constraint.txt 2025-11-03T17:08:35.5719256Z # google-api-core 2025-11-03T17:08:35.5719340Z # googleapis-common-protos 2025-11-03T17:08:35.5719406Z # mlflow-skinny 2025-11-03T17:08:35.5719479Z # opentelemetry-proto 2025-11-03T17:08:35.5719547Z # proto-plus 2025-11-03T17:08:35.5719606Z # ray 2025-11-03T17:08:35.5719666Z # tensorboardx 2025-11-03T17:08:35.5719765Z # tensorizer 2025-11-03T17:08:35.5719826Z psutil==6.1.0 2025-11-03T17:08:35.5719916Z  # via 2025-11-03T17:08:35.5719989Z # -c snapshot_constraint.txt 2025-11-03T17:08:35.5720048Z # accelerate 2025-11-03T17:08:35.5720112Z # peft 2025-11-03T17:08:35.5720207Z # tensorizer 2025-11-03T17:08:35.5720267Z py==1.11.0 2025-11-03T17:08:35.5720354Z  # via 2025-11-03T17:08:35.5720427Z # -c snapshot_constraint.txt 2025-11-03T17:08:35.5720531Z # pytest-forked 2025-11-03T17:08:35.5720593Z py-spy==0.4.0 2025-11-03T17:08:35.5720673Z  # via 2025-11-03T17:08:35.5720749Z # -c snapshot_constraint.txt 2025-11-03T17:08:35.5720831Z # ray 2025-11-03T17:08:35.5720899Z pyarrow==18.0.0 2025-11-03T17:08:35.5720978Z  # via 2025-11-03T17:08:35.5721050Z # -c snapshot_constraint.txt 2025-11-03T17:08:35.5721118Z # datasets 2025-11-03T17:08:35.5721180Z # genai-perf 2025-11-03T17:08:35.5721264Z # mlflow 2025-11-03T17:08:35.5721334Z pyasn1==0.6.1 2025-11-03T17:08:35.5721412Z  # via 2025-11-03T17:08:35.5721501Z # -c snapshot_constraint.txt 2025-11-03T17:08:35.5721570Z # pyasn1-modules 2025-11-03T17:08:35.5721655Z # rsa 2025-11-03T17:08:35.5721732Z pyasn1-modules==0.4.2 2025-11-03T17:08:35.5721815Z  # via 2025-11-03T17:08:35.5721895Z # -c snapshot_constraint.txt 2025-11-03T17:08:35.5721990Z # google-auth 2025-11-03T17:08:35.5722054Z pybind11==2.13.6 2025-11-03T17:08:35.5722142Z  # via 2025-11-03T17:08:35.5722214Z # -c snapshot_constraint.txt 2025-11-03T17:08:35.5722302Z # lm-eval 2025-11-03T17:08:35.5722374Z pycocotools==2.0.8 2025-11-03T17:08:35.5722456Z  # via 2025-11-03T17:08:35.5722537Z # -c snapshot_constraint.txt 2025-11-03T17:08:35.5722627Z # terratorch 2025-11-03T17:08:35.5722692Z pycountry==24.6.1 2025-11-03T17:08:35.5722776Z  # via 2025-11-03T17:08:35.5722849Z # -c snapshot_constraint.txt 2025-11-03T17:08:35.5722965Z # pydantic-extra-types 2025-11-03T17:08:35.5723116Z pycparser==2.22 2025-11-03T17:08:35.5723198Z  # via 2025-11-03T17:08:35.5723277Z # -c snapshot_constraint.txt 2025-11-03T17:08:35.5723361Z # cffi 2025-11-03T17:08:35.5723435Z pycryptodomex==3.22.0 2025-11-03T17:08:35.5723523Z  # via 2025-11-03T17:08:35.5723599Z # -c snapshot_constraint.txt 2025-11-03T17:08:35.5723696Z # blobfile 2025-11-03T17:08:35.5723761Z pydantic==2.12.0 2025-11-03T17:08:35.5723839Z  # via 2025-11-03T17:08:35.5723917Z # -c snapshot_constraint.txt 2025-11-03T17:08:35.5723988Z # -r requirements/test.in 2025-11-03T17:08:35.5724058Z # albumentations 2025-11-03T17:08:35.5724132Z # datamodel-code-generator 2025-11-03T17:08:35.5724190Z # fastapi 2025-11-03T17:08:35.5724253Z # gpt-oss 2025-11-03T17:08:35.5724319Z # lightly 2025-11-03T17:08:35.5724386Z # mistral-common 2025-11-03T17:08:35.5724455Z # mlflow-skinny 2025-11-03T17:08:35.5724512Z # mteb 2025-11-03T17:08:35.5724585Z # openai-harmony 2025-11-03T17:08:35.5724659Z # pydantic-extra-types 2025-11-03T17:08:35.5724742Z # ray 2025-11-03T17:08:35.5724814Z pydantic-core==2.41.1 2025-11-03T17:08:35.5724971Z  # via 2025-11-03T17:08:35.5725049Z # -c snapshot_constraint.txt 2025-11-03T17:08:35.5725139Z # pydantic 2025-11-03T17:08:35.5725215Z pydantic-extra-types==2.10.5 2025-11-03T17:08:35.5725299Z  # via 2025-11-03T17:08:35.5725370Z # -c snapshot_constraint.txt 2025-11-03T17:08:35.5725466Z # mistral-common 2025-11-03T17:08:35.5725535Z pygments==2.18.0 2025-11-03T17:08:35.5725614Z  # via 2025-11-03T17:08:35.5725700Z # -c snapshot_constraint.txt 2025-11-03T17:08:35.5725789Z # rich 2025-11-03T17:08:35.5725851Z pyogrio==0.11.0 2025-11-03T17:08:35.5725936Z  # via 2025-11-03T17:08:35.5726008Z # -c snapshot_constraint.txt 2025-11-03T17:08:35.5726102Z # geopandas 2025-11-03T17:08:35.5726166Z pyparsing==3.2.0 2025-11-03T17:08:35.5726248Z  # via 2025-11-03T17:08:35.5726332Z # -c snapshot_constraint.txt 2025-11-03T17:08:35.5726393Z # matplotlib 2025-11-03T17:08:35.5726481Z # rasterio 2025-11-03T17:08:35.5726551Z pyproj==3.7.1 2025-11-03T17:08:35.5726631Z  # via 2025-11-03T17:08:35.5726708Z # -c snapshot_constraint.txt 2025-11-03T17:08:35.5726767Z # geopandas 2025-11-03T17:08:35.5726824Z # rioxarray 2025-11-03T17:08:35.5726918Z # torchgeo 2025-11-03T17:08:35.5726986Z pyrate-limiter==3.7.0 2025-11-03T17:08:35.5727071Z  # via 2025-11-03T17:08:35.5727142Z # -c snapshot_constraint.txt 2025-11-03T17:08:35.5727239Z # schemathesis 2025-11-03T17:08:35.5727309Z pystemmer==3.0.0 2025-11-03T17:08:35.5727388Z  # via 2025-11-03T17:08:35.5727460Z # -c snapshot_constraint.txt 2025-11-03T17:08:35.5727549Z # mteb 2025-11-03T17:08:35.5727618Z pytablewriter==1.2.0 2025-11-03T17:08:35.5727703Z  # via 2025-11-03T17:08:35.5727778Z # -c snapshot_constraint.txt 2025-11-03T17:08:35.5727865Z # lm-eval 2025-11-03T17:08:35.5727931Z pytest==8.3.5 2025-11-03T17:08:35.5728011Z  # via 2025-11-03T17:08:35.5728089Z # -c snapshot_constraint.txt 2025-11-03T17:08:35.5728160Z # -r requirements/test.in 2025-11-03T17:08:35.5728234Z # buildkite-test-collector 2025-11-03T17:08:35.5728301Z # genai-perf 2025-11-03T17:08:35.5728365Z # pytest-asyncio 2025-11-03T17:08:35.5728484Z # pytest-cov 2025-11-03T17:08:35.5728556Z # pytest-forked 2025-11-03T17:08:35.5728621Z # pytest-mock 2025-11-03T17:08:35.5728697Z # pytest-rerunfailures 2025-11-03T17:08:35.5728759Z # pytest-shard 2025-11-03T17:08:35.5728826Z # pytest-subtests 2025-11-03T17:08:35.5728895Z # pytest-timeout 2025-11-03T17:08:35.5728954Z # schemathesis 2025-11-03T17:08:35.5729054Z # terratorch 2025-11-03T17:08:35.5729123Z pytest-asyncio==0.24.0 2025-11-03T17:08:35.5729200Z  # via 2025-11-03T17:08:35.5729367Z # -c snapshot_constraint.txt 2025-11-03T17:08:35.5729477Z # -r requirements/test.in 2025-11-03T17:08:35.5729539Z pytest-cov==6.3.0 2025-11-03T17:08:35.5729622Z  # via 2025-11-03T17:08:35.5729694Z # -c snapshot_constraint.txt 2025-11-03T17:08:35.5729800Z # -r requirements/test.in 2025-11-03T17:08:35.5729866Z pytest-forked==1.6.0 2025-11-03T17:08:35.5729947Z  # via 2025-11-03T17:08:35.5730022Z # -c snapshot_constraint.txt 2025-11-03T17:08:35.5730122Z # -r requirements/test.in 2025-11-03T17:08:35.5730191Z pytest-mock==3.14.0 2025-11-03T17:08:35.5730271Z  # via 2025-11-03T17:08:35.5730342Z # -c snapshot_constraint.txt 2025-11-03T17:08:35.5730439Z # genai-perf 2025-11-03T17:08:35.5730515Z pytest-rerunfailures==14.0 2025-11-03T17:08:35.5730595Z  # via 2025-11-03T17:08:35.5730667Z # -c snapshot_constraint.txt 2025-11-03T17:08:35.5730772Z # -r requirements/test.in 2025-11-03T17:08:35.5730846Z pytest-shard==0.1.2 2025-11-03T17:08:35.5730931Z  # via 2025-11-03T17:08:35.5731001Z # -c snapshot_constraint.txt 2025-11-03T17:08:35.5731109Z # -r requirements/test.in 2025-11-03T17:08:35.5731394Z pytest-subtests==0.14.1 2025-11-03T17:08:35.5731487Z  # via 2025-11-03T17:08:35.5731560Z # -c snapshot_constraint.txt 2025-11-03T17:08:35.5731659Z # schemathesis 2025-11-03T17:08:35.5731732Z pytest-timeout==2.3.1 2025-11-03T17:08:35.5731808Z  # via 2025-11-03T17:08:35.5731883Z # -c snapshot_constraint.txt 2025-11-03T17:08:35.5731993Z # -r requirements/test.in 2025-11-03T17:08:35.5732057Z python-box==7.3.2 2025-11-03T17:08:35.5732141Z  # via 2025-11-03T17:08:35.5732212Z # -c snapshot_constraint.txt 2025-11-03T17:08:35.5732301Z # terratorch 2025-11-03T17:08:35.5732375Z python-dateutil==2.9.0.post0 2025-11-03T17:08:35.5732453Z  # via 2025-11-03T17:08:35.5732528Z # -c snapshot_constraint.txt 2025-11-03T17:08:35.5732588Z # arrow 2025-11-03T17:08:35.5732645Z # botocore 2025-11-03T17:08:35.5732706Z # graphene 2025-11-03T17:08:35.5732762Z # lightly 2025-11-03T17:08:35.5732823Z # matplotlib 2025-11-03T17:08:35.5732886Z # pandas 2025-11-03T17:08:35.5732980Z # typepy 2025-11-03T17:08:35.5733056Z python-rapidjson==1.20 2025-11-03T17:08:35.5733136Z  # via 2025-11-03T17:08:35.5733208Z # -c snapshot_constraint.txt 2025-11-03T17:08:35.5733307Z # tritonclient 2025-11-03T17:08:35.5733379Z pytorch-lightning==2.5.5 2025-11-03T17:08:35.5733462Z  # via 2025-11-03T17:08:35.5733518Z # lightly 2025-11-03T17:08:35.5733605Z # lightning 2025-11-03T17:08:35.5733685Z pytrec-eval-terrier==0.5.7 2025-11-03T17:08:35.5733764Z  # via 2025-11-03T17:08:35.5733835Z # -c snapshot_constraint.txt 2025-11-03T17:08:35.5733921Z # mteb 2025-11-03T17:08:35.5733980Z pytz==2024.2 2025-11-03T17:08:35.5734066Z  # via 2025-11-03T17:08:35.5734146Z # -c snapshot_constraint.txt 2025-11-03T17:08:35.5734207Z # pandas 2025-11-03T17:08:35.5734299Z # typepy 2025-11-03T17:08:35.5734360Z pyyaml==6.0.2 2025-11-03T17:08:35.5734441Z  # via 2025-11-03T17:08:35.5734517Z # -c snapshot_constraint.txt 2025-11-03T17:08:35.5734575Z # accelerate 2025-11-03T17:08:35.5734642Z # albumentations 2025-11-03T17:08:35.5734716Z # datamodel-code-generator 2025-11-03T17:08:35.5734775Z # datasets 2025-11-03T17:08:35.5734836Z # genai-perf 2025-11-03T17:08:35.5734901Z # huggingface-hub 2025-11-03T17:08:35.5734967Z # jsonargparse 2025-11-03T17:08:35.5735026Z # lightning 2025-11-03T17:08:35.5735088Z # mlflow-skinny 2025-11-03T17:08:35.5735154Z # omegaconf 2025-11-03T17:08:35.5735211Z # peft 2025-11-03T17:08:35.5735278Z # pytorch-lightning 2025-11-03T17:08:35.5735338Z # ray 2025-11-03T17:08:35.5735398Z # responses 2025-11-03T17:08:35.5735560Z # schemathesis 2025-11-03T17:08:35.5735616Z # timm 2025-11-03T17:08:35.5735679Z # transformers 2025-11-03T17:08:35.5735769Z # vocos 2025-11-03T17:08:35.5735836Z rapidfuzz==3.12.1 2025-11-03T17:08:35.5735918Z  # via 2025-11-03T17:08:35.5735991Z # -c snapshot_constraint.txt 2025-11-03T17:08:35.5736073Z # jiwer 2025-11-03T17:08:35.5736139Z rasterio==1.4.3 2025-11-03T17:08:35.5736219Z  # via 2025-11-03T17:08:35.5736291Z # -c snapshot_constraint.txt 2025-11-03T17:08:35.5736357Z # rioxarray 2025-11-03T17:08:35.5736414Z # terratorch 2025-11-03T17:08:35.5736508Z # torchgeo 2025-11-03T17:08:35.5736566Z ray==2.48.0 2025-11-03T17:08:35.5736644Z  # via 2025-11-03T17:08:35.5736720Z # -c snapshot_constraint.txt 2025-11-03T17:08:35.5736827Z # -r requirements/test.in 2025-11-03T17:08:35.5736893Z redis==5.2.0 2025-11-03T17:08:35.5736970Z  # via 2025-11-03T17:08:35.5737041Z # -c snapshot_constraint.txt 2025-11-03T17:08:35.5737149Z # tensorizer 2025-11-03T17:08:35.5737216Z referencing==0.35.1 2025-11-03T17:08:35.5737295Z  # via 2025-11-03T17:08:35.5737452Z # -c snapshot_constraint.txt 2025-11-03T17:08:35.5737516Z # jsonschema 2025-11-03T17:08:35.5737640Z # jsonschema-specifications 2025-11-03T17:08:35.5737699Z regex==2024.9.11 2025-11-03T17:08:35.5737778Z  # via 2025-11-03T17:08:35.5737856Z # -c snapshot_constraint.txt 2025-11-03T17:08:35.5737911Z # nltk 2025-11-03T17:08:35.5737979Z # open-clip-torch 2025-11-03T17:08:35.5738037Z # sacrebleu 2025-11-03T17:08:35.5738093Z # tiktoken 2025-11-03T17:08:35.5738189Z # transformers 2025-11-03T17:08:35.5738247Z requests==2.32.3 2025-11-03T17:08:35.5738324Z  # via 2025-11-03T17:08:35.5738399Z # -c snapshot_constraint.txt 2025-11-03T17:08:35.5738471Z # buildkite-test-collector 2025-11-03T17:08:35.5738537Z # databricks-sdk 2025-11-03T17:08:35.5738595Z # datasets 2025-11-03T17:08:35.5738651Z # docker 2025-11-03T17:08:35.5738714Z # evaluate 2025-11-03T17:08:35.5738791Z # google-api-core 2025-11-03T17:08:35.5738869Z # google-cloud-storage 2025-11-03T17:08:35.5738929Z # gpt-oss 2025-11-03T17:08:35.5738993Z # huggingface-hub 2025-11-03T17:08:35.5739051Z # lightly 2025-11-03T17:08:35.5739108Z # lm-eval 2025-11-03T17:08:35.5739171Z # mistral-common 2025-11-03T17:08:35.5739238Z # mlflow-skinny 2025-11-03T17:08:35.5739293Z # mteb 2025-11-03T17:08:35.5739353Z # pooch 2025-11-03T17:08:35.5739408Z # ray 2025-11-03T17:08:35.5739469Z # responses 2025-11-03T17:08:35.5739532Z # schemathesis 2025-11-03T17:08:35.5739601Z # starlette-testclient 2025-11-03T17:08:35.5739659Z # tiktoken 2025-11-03T17:08:35.5739754Z # transformers 2025-11-03T17:08:35.5739816Z responses==0.25.3 2025-11-03T17:08:35.5739897Z  # via 2025-11-03T17:08:35.5739971Z # -c snapshot_constraint.txt 2025-11-03T17:08:35.5740061Z # genai-perf 2025-11-03T17:08:35.5740134Z rfc3339-validator==0.1.4 2025-11-03T17:08:35.5740212Z  # via 2025-11-03T17:08:35.5740291Z # -c snapshot_constraint.txt 2025-11-03T17:08:35.5740389Z # jsonschema 2025-11-03T17:08:35.5740448Z rfc3987==1.3.8 2025-11-03T17:08:35.5740533Z  # via 2025-11-03T17:08:35.5740608Z # -c snapshot_constraint.txt 2025-11-03T17:08:35.5740700Z # jsonschema 2025-11-03T17:08:35.5740763Z rich==13.9.4 2025-11-03T17:08:35.5740849Z  # via 2025-11-03T17:08:35.5740923Z # -c snapshot_constraint.txt 2025-11-03T17:08:35.5740983Z # genai-perf 2025-11-03T17:08:35.5741039Z # lightning 2025-11-03T17:08:35.5741100Z # mteb 2025-11-03T17:08:35.5741186Z # typer 2025-11-03T17:08:35.5741249Z rioxarray==0.19.0 2025-11-03T17:08:35.5741329Z  # via 2025-11-03T17:08:35.5741400Z # -c snapshot_constraint.txt 2025-11-03T17:08:35.5741581Z # terratorch 2025-11-03T17:08:35.5741648Z rouge-score==0.1.2 2025-11-03T17:08:35.5741728Z  # via 2025-11-03T17:08:35.5741806Z # -c snapshot_constraint.txt 2025-11-03T17:08:35.5741896Z # lm-eval 2025-11-03T17:08:35.5741961Z rpds-py==0.20.1 2025-11-03T17:08:35.5742039Z  # via 2025-11-03T17:08:35.5742110Z # -c snapshot_constraint.txt 2025-11-03T17:08:35.5742175Z # jsonschema 2025-11-03T17:08:35.5742266Z # referencing 2025-11-03T17:08:35.5742331Z rsa==4.9.1 2025-11-03T17:08:35.5742409Z  # via 2025-11-03T17:08:35.5742477Z # -c snapshot_constraint.txt 2025-11-03T17:08:35.5742573Z # google-auth 2025-11-03T17:08:35.5742633Z rtree==1.4.0 2025-11-03T17:08:35.5742713Z  # via 2025-11-03T17:08:35.5742790Z # -c snapshot_constraint.txt 2025-11-03T17:08:35.5742876Z # torchgeo 2025-11-03T17:08:35.5742955Z runai-model-streamer==0.14.0 2025-11-03T17:08:35.5743033Z  # via 2025-11-03T17:08:35.5743105Z # -c snapshot_constraint.txt 2025-11-03T17:08:35.5743229Z # -r requirements/test.in 2025-11-03T17:08:35.5743308Z runai-model-streamer-gcs==0.14.0 2025-11-03T17:08:35.5743394Z  # via 2025-11-03T17:08:35.5743540Z # -c snapshot_constraint.txt 2025-11-03T17:08:35.5743652Z # runai-model-streamer 2025-11-03T17:08:35.5743731Z runai-model-streamer-s3==0.14.0 2025-11-03T17:08:35.5743811Z  # via 2025-11-03T17:08:35.5743882Z # -c snapshot_constraint.txt 2025-11-03T17:08:35.5743990Z # runai-model-streamer 2025-11-03T17:08:35.5744053Z s3transfer==0.10.3 2025-11-03T17:08:35.5744135Z  # via 2025-11-03T17:08:35.5744208Z # -c snapshot_constraint.txt 2025-11-03T17:08:35.5744290Z # boto3 2025-11-03T17:08:35.5744359Z sacrebleu==2.4.3 2025-11-03T17:08:35.5744438Z  # via 2025-11-03T17:08:35.5744510Z # -c snapshot_constraint.txt 2025-11-03T17:08:35.5744597Z # lm-eval 2025-11-03T17:08:35.5744658Z safetensors==0.4.5 2025-11-03T17:08:35.5744745Z  # via 2025-11-03T17:08:35.5744814Z # -c snapshot_constraint.txt 2025-11-03T17:08:35.5744871Z # accelerate 2025-11-03T17:08:35.5744941Z # open-clip-torch 2025-11-03T17:08:35.5745001Z # peft 2025-11-03T17:08:35.5745064Z # timm 2025-11-03T17:08:35.5745157Z # transformers 2025-11-03T17:08:35.5745222Z schemathesis==3.39.15 2025-11-03T17:08:35.5745309Z  # via 2025-11-03T17:08:35.5745380Z # -c snapshot_constraint.txt 2025-11-03T17:08:35.5745488Z # -r requirements/test.in 2025-11-03T17:08:35.5745563Z scikit-image==0.25.2 2025-11-03T17:08:35.5745642Z  # via 2025-11-03T17:08:35.5745717Z # -c snapshot_constraint.txt 2025-11-03T17:08:35.5745813Z # albumentations 2025-11-03T17:08:35.5745879Z scikit-learn==1.5.2 2025-11-03T17:08:35.5745960Z  # via 2025-11-03T17:08:35.5746031Z # -c snapshot_constraint.txt 2025-11-03T17:08:35.5746096Z # albumentations 2025-11-03T17:08:35.5746159Z # librosa 2025-11-03T17:08:35.5746215Z # lm-eval 2025-11-03T17:08:35.5746277Z # mlflow 2025-11-03T17:08:35.5746334Z # mteb 2025-11-03T17:08:35.5746447Z # sentence-transformers 2025-11-03T17:08:35.5746512Z scipy==1.13.1 2025-11-03T17:08:35.5746591Z  # via 2025-11-03T17:08:35.5746668Z # -c snapshot_constraint.txt 2025-11-03T17:08:35.5746730Z # albumentations 2025-11-03T17:08:35.5746797Z # bm25s 2025-11-03T17:08:35.5746861Z # librosa 2025-11-03T17:08:35.5746917Z # mlflow 2025-11-03T17:08:35.5746979Z # mteb 2025-11-03T17:08:35.5747041Z # scikit-image 2025-11-03T17:08:35.5747105Z # scikit-learn 2025-11-03T17:08:35.5747187Z # sentence-transformers 2025-11-03T17:08:35.5755593Z # statsmodels 2025-11-03T17:08:35.5755746Z # vocos 2025-11-03T17:08:35.5755848Z segmentation-models-pytorch==0.4.0 2025-11-03T17:08:35.5755936Z  # via 2025-11-03T17:08:35.5756006Z # terratorch 2025-11-03T17:08:35.5756238Z # torchgeo 2025-11-03T17:08:35.5756321Z sentence-transformers==3.2.1 2025-11-03T17:08:35.5756407Z  # via 2025-11-03T17:08:35.5756487Z # -c snapshot_constraint.txt 2025-11-03T17:08:35.5756571Z # -r requirements/test.in 2025-11-03T17:08:35.5756656Z # mteb 2025-11-03T17:08:35.5756723Z setuptools==77.0.3 2025-11-03T17:08:35.5756808Z  # via 2025-11-03T17:08:35.5756883Z # -c snapshot_constraint.txt 2025-11-03T17:08:35.5756954Z # lightning-utilities 2025-11-03T17:08:35.5757024Z # pytablewriter 2025-11-03T17:08:35.5757108Z # torch 2025-11-03T17:08:35.5757175Z shapely==2.1.1 2025-11-03T17:08:35.5757253Z  # via 2025-11-03T17:08:35.5757326Z # -c snapshot_constraint.txt 2025-11-03T17:08:35.5757389Z # geopandas 2025-11-03T17:08:35.5757475Z # torchgeo 2025-11-03T17:08:35.5757538Z shellingham==1.5.4 2025-11-03T17:08:35.5757621Z  # via 2025-11-03T17:08:35.5757692Z # -c snapshot_constraint.txt 2025-11-03T17:08:35.5757784Z # typer 2025-11-03T17:08:35.5757842Z six==1.16.0 2025-11-03T17:08:35.5757919Z  # via 2025-11-03T17:08:35.5757995Z # -c snapshot_constraint.txt 2025-11-03T17:08:35.5758137Z # junit-xml 2025-11-03T17:08:35.5758201Z # lightly 2025-11-03T17:08:35.5758260Z # opencensus 2025-11-03T17:08:35.5758325Z # python-dateutil 2025-11-03T17:08:35.5758395Z # rfc3339-validator 2025-11-03T17:08:35.5758453Z # rouge-score 2025-11-03T17:08:35.5758582Z # segmentation-models-pytorch 2025-11-03T17:08:35.5758645Z smart-open==7.1.0 2025-11-03T17:08:35.5758731Z  # via 2025-11-03T17:08:35.5758804Z # -c snapshot_constraint.txt 2025-11-03T17:08:35.5758884Z # ray 2025-11-03T17:08:35.5758951Z smmap==5.0.2 2025-11-03T17:08:35.5759033Z  # via 2025-11-03T17:08:35.5759107Z # -c snapshot_constraint.txt 2025-11-03T17:08:35.5759190Z # gitdb 2025-11-03T17:08:35.5759249Z sniffio==1.3.1 2025-11-03T17:08:35.5759335Z  # via 2025-11-03T17:08:35.5759407Z # -c snapshot_constraint.txt 2025-11-03T17:08:35.5759471Z # anyio 2025-11-03T17:08:35.5759562Z # httpx 2025-11-03T17:08:35.5759633Z sortedcontainers==2.4.0 2025-11-03T17:08:35.5759716Z  # via 2025-11-03T17:08:35.5759786Z # -c snapshot_constraint.txt 2025-11-03T17:08:35.5759879Z # hypothesis 2025-11-03T17:08:35.5759945Z soundfile==0.12.1 2025-11-03T17:08:35.5760022Z  # via 2025-11-03T17:08:35.5760097Z # -c snapshot_constraint.txt 2025-11-03T17:08:35.5760169Z # -r requirements/test.in 2025-11-03T17:08:35.5760226Z # librosa 2025-11-03T17:08:35.5760327Z # mistral-common 2025-11-03T17:08:35.5760387Z soxr==0.5.0.post1 2025-11-03T17:08:35.5760468Z  # via 2025-11-03T17:08:35.5760536Z # -c snapshot_constraint.txt 2025-11-03T17:08:35.5760592Z # librosa 2025-11-03T17:08:35.5760689Z # mistral-common 2025-11-03T17:08:35.5760751Z sqlalchemy==2.0.41 2025-11-03T17:08:35.5760836Z  # via 2025-11-03T17:08:35.5760915Z # -c snapshot_constraint.txt 2025-11-03T17:08:35.5760980Z # alembic 2025-11-03T17:08:35.5761069Z # mlflow 2025-11-03T17:08:35.5761134Z sqlitedict==2.1.0 2025-11-03T17:08:35.5761212Z  # via 2025-11-03T17:08:35.5761287Z # -c snapshot_constraint.txt 2025-11-03T17:08:35.5761372Z # lm-eval 2025-11-03T17:08:35.5761438Z sqlparse==0.5.3 2025-11-03T17:08:35.5761519Z  # via 2025-11-03T17:08:35.5761592Z # -c snapshot_constraint.txt 2025-11-03T17:08:35.5761691Z # mlflow-skinny 2025-11-03T17:08:35.5761752Z starlette==0.46.2 2025-11-03T17:08:35.5761829Z  # via 2025-11-03T17:08:35.5761902Z # -c snapshot_constraint.txt 2025-11-03T17:08:35.5761959Z # fastapi 2025-11-03T17:08:35.5762023Z # schemathesis 2025-11-03T17:08:35.5762130Z # starlette-testclient 2025-11-03T17:08:35.5762202Z starlette-testclient==0.4.1 2025-11-03T17:08:35.5762283Z  # via 2025-11-03T17:08:35.5762457Z # -c snapshot_constraint.txt 2025-11-03T17:08:35.5762558Z # schemathesis 2025-11-03T17:08:35.5762622Z statsmodels==0.14.4 2025-11-03T17:08:35.5762699Z  # via 2025-11-03T17:08:35.5762775Z # -c snapshot_constraint.txt 2025-11-03T17:08:35.5762865Z # genai-perf 2025-11-03T17:08:35.5762925Z structlog==25.4.0 2025-11-03T17:08:35.5763005Z  # via 2025-11-03T17:08:35.5763074Z # -c snapshot_constraint.txt 2025-11-03T17:08:35.5763162Z # gpt-oss 2025-11-03T17:08:35.5763222Z sympy==1.13.3 2025-11-03T17:08:35.5763298Z  # via 2025-11-03T17:08:35.5763386Z # -c snapshot_constraint.txt 2025-11-03T17:08:35.5763443Z # einx 2025-11-03T17:08:35.5763529Z # torch 2025-11-03T17:08:35.5763592Z tabledata==1.3.3 2025-11-03T17:08:35.5763670Z  # via 2025-11-03T17:08:35.5763743Z # -c snapshot_constraint.txt 2025-11-03T17:08:35.5763837Z # pytablewriter 2025-11-03T17:08:35.5763896Z tabulate==0.9.0 2025-11-03T17:08:35.5763981Z  # via 2025-11-03T17:08:35.5764049Z # -c snapshot_constraint.txt 2025-11-03T17:08:35.5764140Z # sacrebleu 2025-11-03T17:08:35.5764198Z tblib==3.1.0 2025-11-03T17:08:35.5764355Z  # via 2025-11-03T17:08:35.5764428Z # -c snapshot_constraint.txt 2025-11-03T17:08:35.5764536Z # -r requirements/test.in 2025-11-03T17:08:35.5764596Z tcolorpy==0.1.6 2025-11-03T17:08:35.5764678Z  # via 2025-11-03T17:08:35.5764749Z # -c snapshot_constraint.txt 2025-11-03T17:08:35.5764847Z # pytablewriter 2025-11-03T17:08:35.5764907Z tenacity==9.1.2 2025-11-03T17:08:35.5764985Z  # via 2025-11-03T17:08:35.5765059Z # -c snapshot_constraint.txt 2025-11-03T17:08:35.5765116Z # gpt-oss 2025-11-03T17:08:35.5765175Z # lm-eval 2025-11-03T17:08:35.5765260Z # plotly 2025-11-03T17:08:35.5765320Z tensorboardx==2.6.4 2025-11-03T17:08:35.5765401Z  # via 2025-11-03T17:08:35.5765470Z # -c snapshot_constraint.txt 2025-11-03T17:08:35.5765561Z # lightning 2025-11-03T17:08:35.5765626Z tensorizer==2.10.1 2025-11-03T17:08:35.5765703Z  # via 2025-11-03T17:08:35.5765784Z # -c snapshot_constraint.txt 2025-11-03T17:08:35.5765887Z # -r requirements/test.in 2025-11-03T17:08:35.5765948Z termcolor==3.1.0 2025-11-03T17:08:35.5766030Z  # via 2025-11-03T17:08:35.5766098Z # -c snapshot_constraint.txt 2025-11-03T17:08:35.5766187Z # gpt-oss 2025-11-03T17:08:35.5766446Z terratorch @ git+https://github.com/IBM/terratorch.git@07184fcf91a1324f831ff521dd238d97fe350e3e 2025-11-03T17:08:35.5766573Z  # via -r requirements/test.in 2025-11-03T17:08:35.5766646Z threadpoolctl==3.5.0 2025-11-03T17:08:35.5766724Z  # via 2025-11-03T17:08:35.5766800Z # -c snapshot_constraint.txt 2025-11-03T17:08:35.5766894Z # scikit-learn 2025-11-03T17:08:35.5766955Z tifffile==2025.3.30 2025-11-03T17:08:35.5767034Z  # via 2025-11-03T17:08:35.5767119Z # -c snapshot_constraint.txt 2025-11-03T17:08:35.5767182Z # scikit-image 2025-11-03T17:08:35.5767278Z # terratorch 2025-11-03T17:08:35.5767339Z tiktoken==0.12.0 2025-11-03T17:08:35.5767422Z  # via 2025-11-03T17:08:35.5767493Z # -c snapshot_constraint.txt 2025-11-03T17:08:35.5767551Z # gpt-oss 2025-11-03T17:08:35.5767612Z # lm-eval 2025-11-03T17:08:35.5767708Z # mistral-common 2025-11-03T17:08:35.5767769Z timm==1.0.17 2025-11-03T17:08:35.5767847Z  # via 2025-11-03T17:08:35.5767919Z # -c snapshot_constraint.txt 2025-11-03T17:08:35.5767990Z # -r requirements/test.in 2025-11-03T17:08:35.5768056Z # open-clip-torch 2025-11-03T17:08:35.5768137Z # segmentation-models-pytorch 2025-11-03T17:08:35.5768200Z # terratorch 2025-11-03T17:08:35.5768286Z # torchgeo 2025-11-03T17:08:35.5768351Z tokenizers==0.22.0 2025-11-03T17:08:35.5768488Z  # via 2025-11-03T17:08:35.5768561Z # -c snapshot_constraint.txt 2025-11-03T17:08:35.5769218Z # -r requirements/test.in 2025-11-03T17:08:35.5769316Z # transformers 2025-11-03T17:08:35.5769380Z tomli==2.2.1 2025-11-03T17:08:35.5769459Z  # via 2025-11-03T17:08:35.5769534Z # -c snapshot_constraint.txt 2025-11-03T17:08:35.5769629Z # schemathesis 2025-11-03T17:08:35.5769690Z tomli-w==1.2.0 2025-11-03T17:08:35.5769768Z  # via 2025-11-03T17:08:35.5769842Z # -c snapshot_constraint.txt 2025-11-03T17:08:35.5769934Z # schemathesis 2025-11-03T17:08:35.5770423Z torch @ file:///var/lib/jenkins/workspace/dist/torch-2.10.0a0%2Bgit3f6538f-cp312-cp312-linux_x86_64.whl#sha256=288b4779bb4efa62cf5264c0eca9fdb02025e7a7fb44d42e3f512a68d8e60da0 2025-11-03T17:08:35.5770503Z  # via 2025-11-03T17:08:35.5770574Z # -r requirements/test.in 2025-11-03T17:08:35.5770638Z # accelerate 2025-11-03T17:08:35.5770699Z # bitsandbytes 2025-11-03T17:08:35.5770779Z # efficientnet-pytorch 2025-11-03T17:08:35.5770839Z # encodec 2025-11-03T17:08:35.5770902Z # fastsafetensors 2025-11-03T17:08:35.5770961Z # kornia 2025-11-03T17:08:35.5771017Z # lightly 2025-11-03T17:08:35.5771076Z # lightning 2025-11-03T17:08:35.5771208Z # lm-eval 2025-11-03T17:08:35.5771272Z # mteb 2025-11-03T17:08:35.5771353Z # open-clip-torch 2025-11-03T17:08:35.5771411Z # peft 2025-11-03T17:08:35.5771477Z # pretrainedmodels 2025-11-03T17:08:35.5771548Z # pytorch-lightning 2025-11-03T17:08:35.5771619Z # runai-model-streamer 2025-11-03T17:08:35.5771700Z # segmentation-models-pytorch 2025-11-03T17:08:35.5771770Z # sentence-transformers 2025-11-03T17:08:35.5771828Z # tensorizer 2025-11-03T17:08:35.5771890Z # terratorch 2025-11-03T17:08:35.5771944Z # timm 2025-11-03T17:08:35.5772005Z # torchaudio 2025-11-03T17:08:35.5772067Z # torchgeo 2025-11-03T17:08:35.5772126Z # torchmetrics 2025-11-03T17:08:35.5772187Z # torchvision 2025-11-03T17:08:35.5772261Z # vector-quantize-pytorch 2025-11-03T17:08:35.5772347Z # vocos 2025-11-03T17:08:35.5772874Z torchaudio @ file:///var/lib/jenkins/workspace/dist/audio/torchaudio-2.10.0a0%2B3b0e7a6-cp312-cp312-linux_x86_64.whl#sha256=70d7f19c17cbc4bb012226c17e5f5380b3bd6bda732a45f329984580e0bc17a6 2025-11-03T17:08:35.5772953Z  # via 2025-11-03T17:08:35.5773026Z # -r requirements/test.in 2025-11-03T17:08:35.5773081Z # encodec 2025-11-03T17:08:35.5773163Z # vocos 2025-11-03T17:08:35.5773226Z torchgeo==0.7.0 2025-11-03T17:08:35.5773323Z  # via terratorch 2025-11-03T17:08:35.5773389Z torchmetrics==1.8.2 2025-11-03T17:08:35.5773466Z  # via 2025-11-03T17:08:35.5773524Z # lightning 2025-11-03T17:08:35.5773596Z # pytorch-lightning 2025-11-03T17:08:35.5773653Z # terratorch 2025-11-03T17:08:35.5773739Z # torchgeo 2025-11-03T17:08:35.5774302Z torchvision @ file:///var/lib/jenkins/workspace/dist/vision/torchvision-0.25.0a0%2Bcfbc5c2-cp312-cp312-linux_x86_64.whl#sha256=7fa43396b4b9ce1b239050abea02e7fbabadf1cbdf180883becb1f6530983b15 2025-11-03T17:08:35.5774386Z  # via 2025-11-03T17:08:35.5774462Z # -r requirements/test.in 2025-11-03T17:08:35.5774521Z # lightly 2025-11-03T17:08:35.5774586Z # open-clip-torch 2025-11-03T17:08:35.5774660Z # pretrainedmodels 2025-11-03T17:08:35.5774740Z # segmentation-models-pytorch 2025-11-03T17:08:35.5774804Z # terratorch 2025-11-03T17:08:35.5774860Z # timm 2025-11-03T17:08:35.5774948Z # torchgeo 2025-11-03T17:08:35.5775010Z tqdm==4.66.6 2025-11-03T17:08:35.5775086Z  # via 2025-11-03T17:08:35.5775167Z # -c snapshot_constraint.txt 2025-11-03T17:08:35.5775227Z # datasets 2025-11-03T17:08:35.5775282Z # evaluate 2025-11-03T17:08:35.5775354Z # huggingface-hub 2025-11-03T17:08:35.5775418Z # lightly 2025-11-03T17:08:35.5775477Z # lightning 2025-11-03T17:08:35.5775538Z # lm-eval 2025-11-03T17:08:35.5775673Z # mteb 2025-11-03T17:08:35.5775731Z # nltk 2025-11-03T17:08:35.5775796Z # open-clip-torch 2025-11-03T17:08:35.5775852Z # peft 2025-11-03T17:08:35.5775912Z # pqdm 2025-11-03T17:08:35.5775981Z # pretrainedmodels 2025-11-03T17:08:35.5776047Z # pytorch-lightning 2025-11-03T17:08:35.5776129Z # segmentation-models-pytorch 2025-11-03T17:08:35.5776211Z # sentence-transformers 2025-11-03T17:08:35.5776281Z # tqdm-multiprocess 2025-11-03T17:08:35.5776377Z # transformers 2025-11-03T17:08:35.5776449Z tqdm-multiprocess==0.0.11 2025-11-03T17:08:35.5776533Z  # via 2025-11-03T17:08:35.5776606Z # -c snapshot_constraint.txt 2025-11-03T17:08:35.5776698Z # lm-eval 2025-11-03T17:08:35.5776761Z transformers==4.56.2 2025-11-03T17:08:35.5776838Z  # via 2025-11-03T17:08:35.5776916Z # -c snapshot_constraint.txt 2025-11-03T17:08:35.5776987Z # -r requirements/test.in 2025-11-03T17:08:35.5777052Z # genai-perf 2025-11-03T17:08:35.5777111Z # lm-eval 2025-11-03T17:08:35.5777169Z # peft 2025-11-03T17:08:35.5777241Z # sentence-transformers 2025-11-03T17:08:35.5777369Z # transformers-stream-generator 2025-11-03T17:08:35.5777526Z transformers-stream-generator==0.0.5 2025-11-03T17:08:35.5777612Z  # via 2025-11-03T17:08:35.5777684Z # -c snapshot_constraint.txt 2025-11-03T17:08:35.5777792Z # -r requirements/test.in 2025-11-03T17:08:35.5777855Z tritonclient==2.51.0 2025-11-03T17:08:35.5777933Z  # via 2025-11-03T17:08:35.5778010Z # -c snapshot_constraint.txt 2025-11-03T17:08:35.5778077Z # -r requirements/test.in 2025-11-03T17:08:35.5778176Z # genai-perf 2025-11-03T17:08:35.5778237Z typepy==1.3.2 2025-11-03T17:08:35.5778314Z  # via 2025-11-03T17:08:35.5778388Z # -c snapshot_constraint.txt 2025-11-03T17:08:35.5778459Z # dataproperty 2025-11-03T17:08:35.5778520Z # pytablewriter 2025-11-03T17:08:35.5778616Z # tabledata 2025-11-03T17:08:35.5778677Z typer==0.15.2 2025-11-03T17:08:35.5778760Z  # via 2025-11-03T17:08:35.5778830Z # -c snapshot_constraint.txt 2025-11-03T17:08:35.5778930Z # fastsafetensors 2025-11-03T17:08:35.5779021Z types-python-dateutil==2.9.0.20241206 2025-11-03T17:08:35.5779098Z  # via 2025-11-03T17:08:35.5779174Z # -c snapshot_constraint.txt 2025-11-03T17:08:35.5779256Z # arrow 2025-11-03T17:08:35.5779324Z typeshed-client==2.8.2 2025-11-03T17:08:35.5779406Z  # via 2025-11-03T17:08:35.5779477Z # -c snapshot_constraint.txt 2025-11-03T17:08:35.5779569Z # jsonargparse 2025-11-03T17:08:35.5779646Z typing-extensions==4.15.0 2025-11-03T17:08:35.5779722Z  # via 2025-11-03T17:08:35.5779796Z # -c snapshot_constraint.txt 2025-11-03T17:08:35.5779854Z # aiosignal 2025-11-03T17:08:35.5779914Z # albumentations 2025-11-03T17:08:35.5779974Z # alembic 2025-11-03T17:08:35.5780040Z # chz 2025-11-03T17:08:35.5780106Z # fastapi 2025-11-03T17:08:35.5780163Z # graphene 2025-11-03T17:08:35.5780227Z # huggingface-hub 2025-11-03T17:08:35.5780285Z # librosa 2025-11-03T17:08:35.5780342Z # lightning 2025-11-03T17:08:35.5780413Z # lightning-utilities 2025-11-03T17:08:35.5780479Z # mistral-common 2025-11-03T17:08:35.5780538Z # mlflow-skinny 2025-11-03T17:08:35.5780595Z # mteb 2025-11-03T17:08:35.5780662Z # opentelemetry-api 2025-11-03T17:08:35.5780727Z # opentelemetry-sdk 2025-11-03T17:08:35.5780823Z # opentelemetry-semantic-conventions 2025-11-03T17:08:35.5780878Z # pqdm 2025-11-03T17:08:35.5780941Z # pydantic 2025-11-03T17:08:35.5781002Z # pydantic-core 2025-11-03T17:08:35.5781075Z # pydantic-extra-types 2025-11-03T17:08:35.5781144Z # pytorch-lightning 2025-11-03T17:08:35.5781200Z # sqlalchemy 2025-11-03T17:08:35.5781256Z # torch 2025-11-03T17:08:35.5781316Z # torchgeo 2025-11-03T17:08:35.5781371Z # typer 2025-11-03T17:08:35.5781533Z # typeshed-client 2025-11-03T17:08:35.5781640Z # typing-inspection 2025-11-03T17:08:35.5781715Z typing-inspection==0.4.2 2025-11-03T17:08:35.5781792Z  # via 2025-11-03T17:08:35.5781867Z # -c snapshot_constraint.txt 2025-11-03T17:08:35.5781957Z # pydantic 2025-11-03T17:08:35.5782015Z tzdata==2024.2 2025-11-03T17:08:35.5782099Z  # via 2025-11-03T17:08:35.5782169Z # -c snapshot_constraint.txt 2025-11-03T17:08:35.5782252Z # pandas 2025-11-03T17:08:35.5782320Z uri-template==1.3.0 2025-11-03T17:08:35.5782397Z  # via 2025-11-03T17:08:35.5782470Z # -c snapshot_constraint.txt 2025-11-03T17:08:35.5782556Z # jsonschema 2025-11-03T17:08:35.5782613Z urllib3==2.2.3 2025-11-03T17:08:35.5782694Z  # via 2025-11-03T17:08:35.5782763Z # -c snapshot_constraint.txt 2025-11-03T17:08:35.5782818Z # blobfile 2025-11-03T17:08:35.5782878Z # botocore 2025-11-03T17:08:35.5782934Z # docker 2025-11-03T17:08:35.5782994Z # lightly 2025-11-03T17:08:35.5783048Z # requests 2025-11-03T17:08:35.5783105Z # responses 2025-11-03T17:08:35.5783201Z # tritonclient 2025-11-03T17:08:35.5783334Z uvicorn==0.35.0 2025-11-03T17:08:35.5783419Z  # via 2025-11-03T17:08:35.5783488Z # -c snapshot_constraint.txt 2025-11-03T17:08:35.5783544Z # gpt-oss 2025-11-03T17:08:35.5783642Z # mlflow-skinny 2025-11-03T17:08:35.5783720Z vector-quantize-pytorch==1.24.2 2025-11-03T17:08:35.5783840Z  # via -r requirements/test.in 2025-11-03T17:08:35.5783904Z virtualenv==20.31.2 2025-11-03T17:08:35.5783980Z  # via 2025-11-03T17:08:35.5784052Z # -c snapshot_constraint.txt 2025-11-03T17:08:35.5784131Z # ray 2025-11-03T17:08:35.5784188Z vocos==0.1.0 2025-11-03T17:08:35.5784266Z  # via 2025-11-03T17:08:35.5784335Z # -c snapshot_constraint.txt 2025-11-03T17:08:35.5784447Z # -r requirements/test.in 2025-11-03T17:08:35.5784519Z wcwidth==0.2.13 2025-11-03T17:08:35.5784597Z  # via 2025-11-03T17:08:35.5784669Z # -c snapshot_constraint.txt 2025-11-03T17:08:35.5784749Z # ftfy 2025-11-03T17:08:35.5784813Z webcolors==24.11.1 2025-11-03T17:08:35.5784893Z  # via 2025-11-03T17:08:35.5784962Z # -c snapshot_constraint.txt 2025-11-03T17:08:35.5785052Z # jsonschema 2025-11-03T17:08:35.5785110Z werkzeug==3.1.3 2025-11-03T17:08:35.5785185Z  # via 2025-11-03T17:08:35.5785261Z # -c snapshot_constraint.txt 2025-11-03T17:08:35.5785315Z # flask 2025-11-03T17:08:35.5785409Z # schemathesis 2025-11-03T17:08:35.5785470Z word2number==1.1 2025-11-03T17:08:35.5785546Z  # via 2025-11-03T17:08:35.5785617Z # -c snapshot_constraint.txt 2025-11-03T17:08:35.5785711Z # lm-eval 2025-11-03T17:08:35.5785770Z wrapt==1.17.2 2025-11-03T17:08:35.5785850Z  # via 2025-11-03T17:08:35.5785921Z # -c snapshot_constraint.txt 2025-11-03T17:08:35.5786018Z # smart-open 2025-11-03T17:08:35.5786077Z xarray==2025.7.1 2025-11-03T17:08:35.5786157Z  # via 2025-11-03T17:08:35.5786229Z # -c snapshot_constraint.txt 2025-11-03T17:08:35.5786318Z # rioxarray 2025-11-03T17:08:35.5786375Z xxhash==3.5.0 2025-11-03T17:08:35.5786452Z  # via 2025-11-03T17:08:35.5786520Z # -c snapshot_constraint.txt 2025-11-03T17:08:35.5786576Z # datasets 2025-11-03T17:08:35.5786662Z # evaluate 2025-11-03T17:08:35.5786719Z yarl==1.17.1 2025-11-03T17:08:35.5786798Z  # via 2025-11-03T17:08:35.5786868Z # -c snapshot_constraint.txt 2025-11-03T17:08:35.5786925Z # aiohttp 2025-11-03T17:08:35.5787015Z # schemathesis 2025-11-03T17:08:35.5787074Z zipp==3.23.0 2025-11-03T17:08:35.5787152Z  # via 2025-11-03T17:08:35.5787221Z # -c snapshot_constraint.txt 2025-11-03T17:08:35.5787334Z # importlib-metadata 2025-11-03T17:08:35.5787398Z zstandard==0.23.0 2025-11-03T17:08:35.5787475Z  # via 2025-11-03T17:08:35.5787645Z # -c snapshot_constraint.txt 2025-11-03T17:08:35.5787731Z # lm-eval 2025-11-03T17:08:35.6986858Z 2025-11-03 17:08:35,698 [INFO] cli.lib.common.pip_helper: pip installing packages: /opt/conda/envs/py_3.12/bin/python -m uv pip install -r test.txt 2025-11-03T17:08:35.6987146Z 2025-11-03 17:08:35,698 [INFO] cli.lib.common.utils: [cmd] /opt/conda/envs/py_3.12/bin/python -m uv pip install -r test.txt 2025-11-03T17:08:35.7501078Z Using Python 3.12.12 environment at: /opt/conda/envs/py_3.12 2025-11-03T17:08:35.8518455Z Resolved 329 packages in 94ms 2025-11-03T17:08:35.8758941Z Updating https://github.com/EleutherAI/lm-evaluation-harness.git (206b7722158f58c35b7ffcd53b035fdbdda5126d) 2025-11-03T17:08:35.9773875Z Building terratorch @ git+https://github.com/IBM/terratorch.git@07184fcf91a1324f831ff521dd238d97fe350e3e 2025-11-03T17:08:36.0850786Z Downloading numpy (17.1MiB) 2025-11-03T17:08:36.0851427Z Downloading pillow (4.3MiB) 2025-11-03T17:08:36.0924200Z Downloading sqlalchemy (3.2MiB) 2025-11-03T17:08:36.0926398Z Downloading networkx (1.6MiB) 2025-11-03T17:08:36.0984200Z Downloading polars (33.2MiB) 2025-11-03T17:08:36.0985910Z Downloading zstandard (5.2MiB) 2025-11-03T17:08:36.0988888Z Downloading kaleido (76.2MiB) 2025-11-03T17:08:36.0989817Z Downloading black (1.7MiB) 2025-11-03T17:08:36.0991560Z Downloading transformers (11.1MiB) 2025-11-03T17:08:36.0993526Z Downloading py-spy (2.6MiB) 2025-11-03T17:08:36.0995627Z Downloading fastparquet (1.7MiB) 2025-11-03T17:08:36.0998221Z Downloading virtualenv (5.8MiB) 2025-11-03T17:08:36.1118983Z Downloading kornia-rs (2.6MiB) 2025-11-03T17:08:36.1122190Z Downloading mteb (1.5MiB) 2025-11-03T17:08:36.1124550Z Downloading runai-model-streamer-s3 (5.6MiB) 2025-11-03T17:08:36.1127065Z Downloading tokenizers (3.2MiB) 2025-11-03T17:08:36.1129496Z Downloading scipy (36.4MiB) 2025-11-03T17:08:36.1132102Z Downloading hf-xet (3.0MiB) 2025-11-03T17:08:36.1134387Z Downloading runai-model-streamer-gcs (4.2MiB) 2025-11-03T17:08:36.1136602Z Downloading rasterio (21.2MiB) 2025-11-03T17:08:36.1139354Z Downloading python-box (4.1MiB) 2025-11-03T17:08:36.1141598Z Downloading pyogrio (26.4MiB) 2025-11-03T17:08:36.1143955Z Downloading pandas (12.1MiB) 2025-11-03T17:08:36.1147212Z Downloading pycryptodomex (2.2MiB) 2025-11-03T17:08:36.1149564Z Downloading scikit-learn (12.3MiB) 2025-11-03T17:08:36.1152031Z Downloading pydantic-core (2.0MiB) 2025-11-03T17:08:36.1155154Z Downloading bitsandbytes (69.5MiB) 2025-11-03T17:08:36.1157938Z Downloading pyproj (9.1MiB) 2025-11-03T17:08:36.1159841Z Downloading opencv-python-headless (47.7MiB) 2025-11-03T17:08:36.1162004Z Downloading shapely (3.0MiB) 2025-11-03T17:08:36.1163967Z Downloading statsmodels (10.2MiB) 2025-11-03T17:08:36.1165688Z Downloading botocore (12.1MiB) 2025-11-03T17:08:36.1168063Z Downloading grpcio (5.6MiB) 2025-11-03T17:08:36.1170267Z Downloading python-rapidjson (1.6MiB) 2025-11-03T17:08:36.1172603Z Downloading h5py (4.7MiB) 2025-11-03T17:08:36.1174641Z Downloading plotly (18.2MiB) 2025-11-03T17:08:36.1176474Z Downloading timm (2.4MiB) 2025-11-03T17:08:36.1178583Z Downloading fiona (16.4MiB) 2025-11-03T17:08:36.1180774Z Downloading mlflow (27.7MiB) 2025-11-03T17:08:36.1183245Z Downloading mlflow-skinny (6.0MiB) 2025-11-03T17:08:36.1185604Z Downloading tritonclient (13.3MiB) 2025-11-03T17:08:36.1187876Z Downloading aiohttp (1.7MiB) 2025-11-03T17:08:36.1189966Z Downloading pyarrow (38.2MiB) 2025-11-03T17:08:36.1192101Z Downloading decord (13.0MiB) 2025-11-03T17:08:36.1194894Z Downloading matplotlib (7.9MiB) 2025-11-03T17:08:36.1197079Z Downloading cramjam (2.2MiB) 2025-11-03T17:08:36.1199161Z Downloading rapidfuzz (3.0MiB) 2025-11-03T17:08:36.1201486Z Downloading fonttools (4.7MiB) 2025-11-03T17:08:36.1321833Z Downloading scikit-image (14.3MiB) 2025-11-03T17:08:36.1323036Z Downloading ray (66.9MiB) 2025-11-03T17:08:36.9137554Z Downloading python-rapidjson 2025-11-03T17:08:36.9249018Z Downloading open-clip-torch (1.5MiB) 2025-11-03T17:08:36.9298796Z Downloading aiohttp 2025-11-03T17:08:36.9386012Z Downloading fastparquet 2025-11-03T17:08:36.9403503Z Downloading nltk (1.4MiB) 2025-11-03T17:08:36.9588941Z Downloading kiwisolver (1.4MiB) 2025-11-03T17:08:36.9748263Z Downloading black 2025-11-03T17:08:36.9881267Z Downloading fastsafetensors (1.4MiB) 2025-11-03T17:08:37.0660918Z Downloading pydantic-core 2025-11-03T17:08:37.0790525Z Downloading xarray (1.3MiB) 2025-11-03T17:08:37.1540025Z Downloading cramjam 2025-11-03T17:08:37.1720375Z Downloading pycryptodomex 2025-11-03T17:08:37.1766439Z Downloading setuptools (1.2MiB) 2025-11-03T17:08:37.1868871Z Downloading pygments (1.1MiB) 2025-11-03T17:08:37.1873536Z Built terratorch @ git+https://github.com/IBM/terratorch.git@07184fcf91a1324f831ff521dd238d97fe350e3e 2025-11-03T17:08:37.2772052Z Downloading timm 2025-11-03T17:08:37.3011271Z Downloading soundfile (1.1MiB) 2025-11-03T17:08:37.3078700Z Downloading py-spy 2025-11-03T17:08:37.3185158Z Downloading kornia-rs 2025-11-03T17:08:37.3394817Z Downloading kornia (1.0MiB) 2025-11-03T17:08:37.4759869Z Downloading networkx 2025-11-03T17:08:37.5255730Z Downloading hf-xet 2025-11-03T17:08:37.5554139Z Downloading tokenizers 2025-11-03T17:08:37.5571755Z Downloading rapidfuzz 2025-11-03T17:08:37.5615512Z Downloading kiwisolver 2025-11-03T17:08:37.5965880Z Downloading sqlalchemy 2025-11-03T17:08:37.6080945Z Downloading fastsafetensors 2025-11-03T17:08:37.6290577Z Downloading shapely 2025-11-03T17:08:37.6660685Z Downloading xarray 2025-11-03T17:08:37.7885674Z Downloading open-clip-torch 2025-11-03T17:08:37.7936993Z Downloading soundfile 2025-11-03T17:08:37.8186596Z Downloading nltk 2025-11-03T17:08:37.8866859Z Downloading pygments 2025-11-03T17:08:37.9792576Z Downloading python-box 2025-11-03T17:08:38.0908266Z Downloading runai-model-streamer-gcs 2025-11-03T17:08:38.1373169Z Downloading pillow 2025-11-03T17:08:38.1966313Z Downloading setuptools 2025-11-03T17:08:38.2462839Z Downloading mteb 2025-11-03T17:08:38.2956595Z Downloading kornia 2025-11-03T17:08:38.3141948Z Downloading fonttools 2025-11-03T17:08:38.3574721Z Downloading h5py 2025-11-03T17:08:38.4208704Z Updated https://github.com/EleutherAI/lm-evaluation-harness.git (206b7722158f58c35b7ffcd53b035fdbdda5126d) 2025-11-03T17:08:38.4222793Z Building lm-eval @ git+https://github.com/EleutherAI/lm-evaluation-harness.git@206b7722158f58c35b7ffcd53b035fdbdda5126d 2025-11-03T17:08:38.5443490Z Downloading zstandard 2025-11-03T17:08:38.7808937Z Downloading grpcio 2025-11-03T17:08:38.7897542Z Downloading runai-model-streamer-s3 2025-11-03T17:08:38.8755362Z Downloading virtualenv 2025-11-03T17:08:39.2490247Z Building docopt==0.6.2 2025-11-03T17:08:39.2491069Z Building pretrainedmodels==0.7.4 2025-11-03T17:08:39.2493096Z Building antlr4-python3-runtime==4.9.3 2025-11-03T17:08:39.2495200Z Building encodec==0.1.1 2025-11-03T17:08:39.2560852Z Building sqlitedict==2.1.0 2025-11-03T17:08:39.2563370Z Building efficientnet-pytorch==0.7.1 2025-11-03T17:08:39.2567140Z Building rouge-score==0.1.2 2025-11-03T17:08:39.2569631Z Building transformers-stream-generator==0.0.5 2025-11-03T17:08:39.2601361Z Building word2number==1.1 2025-11-03T17:08:39.4431900Z Downloading mlflow-skinny 2025-11-03T17:08:39.6225480Z Downloading matplotlib 2025-11-03T17:08:39.6834445Z Downloading pyproj 2025-11-03T17:08:39.7688943Z Built docopt==0.6.2 2025-11-03T17:08:39.7726708Z Built transformers-stream-generator==0.0.5 2025-11-03T17:08:39.7741667Z Built sqlitedict==2.1.0 2025-11-03T17:08:39.7787212Z Built efficientnet-pytorch==0.7.1 2025-11-03T17:08:39.7827793Z Built word2number==1.1 2025-11-03T17:08:39.7846848Z Built rouge-score==0.1.2 2025-11-03T17:08:39.7959229Z Built encodec==0.1.1 2025-11-03T17:08:39.8045979Z Built pretrainedmodels==0.7.4 2025-11-03T17:08:39.8379281Z Built antlr4-python3-runtime==4.9.3 2025-11-03T17:08:39.9267188Z Downloading statsmodels 2025-11-03T17:08:40.1977627Z Downloading pandas 2025-11-03T17:08:40.2514502Z Downloading scikit-learn 2025-11-03T17:08:40.3868387Z Downloading decord 2025-11-03T17:08:40.4357273Z Downloading tritonclient 2025-11-03T17:08:40.4626266Z Downloading botocore 2025-11-03T17:08:40.4835644Z Downloading transformers 2025-11-03T17:08:40.5854397Z Downloading scikit-image 2025-11-03T17:08:40.7271002Z Downloading fiona 2025-11-03T17:08:40.7307432Z Downloading numpy 2025-11-03T17:08:41.0545883Z Downloading rasterio 2025-11-03T17:08:41.3430849Z Downloading pyogrio 2025-11-03T17:08:41.7056964Z Downloading polars 2025-11-03T17:08:41.8224286Z Downloading mlflow 2025-11-03T17:08:42.6289291Z Downloading opencv-python-headless 2025-11-03T17:08:42.6686883Z Downloading scipy 2025-11-03T17:08:42.7146435Z Downloading bitsandbytes 2025-11-03T17:08:42.7346463Z Downloading pyarrow 2025-11-03T17:08:42.9613727Z Downloading kaleido 2025-11-03T17:08:43.0861768Z Downloading ray 2025-11-03T17:08:45.3556241Z Downloading plotly 2025-11-03T17:08:53.4252557Z Built lm-eval @ git+https://github.com/EleutherAI/lm-evaluation-harness.git@206b7722158f58c35b7ffcd53b035fdbdda5126d 2025-11-03T17:08:53.8435385Z Prepared 290 packages in 17.98s 2025-11-03T17:08:57.3894893Z Uninstalled 94 packages in 3.54s 2025-11-03T17:09:01.2175915Z Installed 290 packages in 3.82s 2025-11-03T17:09:01.2182407Z - absl-py==2.3.1 2025-11-03T17:09:01.2182887Z + absl-py==2.1.0 2025-11-03T17:09:01.2183444Z + accelerate==1.0.1 2025-11-03T17:09:01.2183883Z + aenum==3.1.16 2025-11-03T17:09:01.2184246Z + affine==2.4.0 2025-11-03T17:09:01.2186357Z - aiohttp==3.13.2 2025-11-03T17:09:01.2186709Z + aiohttp==3.13.0 2025-11-03T17:09:01.2187016Z + aiohttp-cors==0.8.1 2025-11-03T17:09:01.2187346Z + albucore==0.0.16 2025-11-03T17:09:01.2187651Z + albumentations==1.4.6 2025-11-03T17:09:01.2188228Z + alembic==1.16.4 2025-11-03T17:09:01.2188571Z + antlr4-python3-runtime==4.9.3 2025-11-03T17:09:01.2188882Z - anyio==4.11.0 2025-11-03T17:09:01.2189769Z + anyio==4.6.2.post1 2025-11-03T17:09:01.2190770Z + argcomplete==3.5.1 2025-11-03T17:09:01.2191659Z + arrow==1.3.0 2025-11-03T17:09:01.2192190Z - attrs==25.4.0 2025-11-03T17:09:01.2193006Z + attrs==24.2.0 2025-11-03T17:09:01.2193520Z - audioread==3.1.0 2025-11-03T17:09:01.2194667Z + audioread==3.0.1 2025-11-03T17:09:01.2195301Z + backoff==2.2.1 2025-11-03T17:09:01.2196199Z + bitsandbytes==0.46.1 2025-11-03T17:09:01.2196984Z + black==24.10.0 2025-11-03T17:09:01.2197819Z + blinker==1.9.0 2025-11-03T17:09:01.2198672Z + blobfile==3.0.0 2025-11-03T17:09:01.2199221Z + bm25s==0.2.13 2025-11-03T17:09:01.2199826Z - boto3==1.35.42 2025-11-03T17:09:01.2200570Z + boto3==1.35.57 2025-11-03T17:09:01.2201513Z - botocore==1.35.99 2025-11-03T17:09:01.2202165Z + botocore==1.35.57 2025-11-03T17:09:01.2203348Z + bounded-pool-executor==0.0.3 2025-11-03T17:09:01.2203999Z + buildkite-test-collector==0.1.9 2025-11-03T17:09:01.2204462Z - cachetools==6.2.1 2025-11-03T17:09:01.2205075Z + cachetools==5.5.2 2025-11-03T17:09:01.2206632Z - certifi==2025.10.5 2025-11-03T17:09:01.2207119Z + certifi==2024.8.30 2025-11-03T17:09:01.2207904Z - cffi==2.0.0 2025-11-03T17:09:01.2209181Z + cffi==1.17.1 2025-11-03T17:09:01.2209947Z + chardet==5.2.0 2025-11-03T17:09:01.2214744Z - charset-normalizer==3.4.4 2025-11-03T17:09:01.2215236Z + charset-normalizer==3.4.0 2025-11-03T17:09:01.2215626Z + chz==0.3.0 2025-11-03T17:09:01.2216072Z - click==8.2.1 2025-11-03T17:09:01.2216478Z + click==8.1.7 2025-11-03T17:09:01.2216804Z + click-plugins==1.1.1.2 2025-11-03T17:09:01.2217271Z + cligj==0.7.2 2025-11-03T17:09:01.2217609Z - cloudpickle==3.1.2 2025-11-03T17:09:01.2218024Z + cloudpickle==3.1.1 2025-11-03T17:09:01.2218688Z + colorful==0.5.6 2025-11-03T17:09:01.2218967Z + contourpy==1.3.0 2025-11-03T17:09:01.2219438Z + coverage==7.10.6 2025-11-03T17:09:01.2220727Z + cramjam==2.9.0 2025-11-03T17:09:01.2221225Z + cycler==0.12.1 2025-11-03T17:09:01.2222032Z + databricks-sdk==0.59.0 2025-11-03T17:09:01.2222762Z + datamodel-code-generator==0.26.3 2025-11-03T17:09:01.2223577Z + dataproperty==1.0.1 2025-11-03T17:09:01.2224190Z + datasets==3.0.2 2025-11-03T17:09:01.2225283Z - decorator==5.2.1 2025-11-03T17:09:01.2225865Z + decorator==5.1.1 2025-11-03T17:09:01.2226447Z + decord==0.6.0 2025-11-03T17:09:01.2227566Z - dill==0.3.7 2025-11-03T17:09:01.2227944Z + dill==0.3.8 2025-11-03T17:09:01.2228849Z + distlib==0.3.9 2025-11-03T17:09:01.2229549Z - dnspython==2.8.0 2025-11-03T17:09:01.2230411Z + dnspython==2.7.0 2025-11-03T17:09:01.2231248Z + docopt==0.6.2 2025-11-03T17:09:01.2232121Z + docstring-parser==0.17.0 2025-11-03T17:09:01.2234984Z + efficientnet-pytorch==0.7.1 2025-11-03T17:09:01.2235550Z + einx==0.3.0 2025-11-03T17:09:01.2236063Z - email-validator==2.3.0 2025-11-03T17:09:01.2236613Z + email-validator==2.2.0 2025-11-03T17:09:01.2237137Z + encodec==0.1.1 2025-11-03T17:09:01.2237676Z + eval-type-backport==0.2.2 2025-11-03T17:09:01.2238193Z + evaluate==0.4.3 2025-11-03T17:09:01.2238679Z - fastapi==0.121.0 2025-11-03T17:09:01.2239206Z + fastapi==0.116.1 2025-11-03T17:09:01.2239791Z + fastparquet==2024.11.0 2025-11-03T17:09:01.2240674Z - fastrlock==0.8.3 2025-11-03T17:09:01.2241781Z + fastrlock==0.8.2 2025-11-03T17:09:01.2242514Z + fastsafetensors==0.1.10 2025-11-03T17:09:01.2243387Z - filelock==3.18.0 2025-11-03T17:09:01.2243945Z + filelock==3.16.1 2025-11-03T17:09:01.2244990Z + fiona==1.10.1 2025-11-03T17:09:01.2245381Z + flask==3.1.1 2025-11-03T17:09:01.2246286Z + fonttools==4.55.0 2025-11-03T17:09:01.2247299Z + fqdn==1.5.1 2025-11-03T17:09:01.2247699Z - frozenlist==1.8.0 2025-11-03T17:09:01.2248627Z + frozenlist==1.5.0 2025-11-03T17:09:01.2249039Z - fsspec==2025.9.0 2025-11-03T17:09:01.2249848Z + fsspec==2024.9.0 2025-11-03T17:09:01.2250683Z + ftfy==6.3.1 2025-11-03T17:09:01.2251274Z + genai-perf==0.0.8 2025-11-03T17:09:01.2252146Z + genson==1.3.0 2025-11-03T17:09:01.2252650Z + geopandas==1.0.1 2025-11-03T17:09:01.2253835Z - gitpython==3.1.45 2025-11-03T17:09:01.2254385Z + gitpython==3.1.44 2025-11-03T17:09:01.2255096Z + google-api-core==2.24.2 2025-11-03T17:09:01.2256263Z - google-auth==2.42.1 2025-11-03T17:09:01.2256789Z + google-auth==2.40.2 2025-11-03T17:09:01.2257717Z + google-cloud-core==2.4.3 2025-11-03T17:09:01.2258262Z + google-cloud-storage==3.4.0 2025-11-03T17:09:01.2258995Z + google-crc32c==1.7.1 2025-11-03T17:09:01.2259939Z + google-resumable-media==2.7.2 2025-11-03T17:09:01.2260547Z + googleapis-common-protos==1.70.0 2025-11-03T17:09:01.2261065Z + gpt-oss==0.0.8 2025-11-03T17:09:01.2261840Z + graphene==3.4.3 2025-11-03T17:09:01.2262680Z + graphql-core==3.2.6 2025-11-03T17:09:01.2263731Z + graphql-relay==3.2.0 2025-11-03T17:09:01.2264262Z + greenlet==3.2.3 2025-11-03T17:09:01.2265102Z - grpcio==1.76.0 2025-11-03T17:09:01.2265851Z + grpcio==1.71.0 2025-11-03T17:09:01.2266541Z + gunicorn==23.0.0 2025-11-03T17:09:01.2267544Z - h11==0.16.0 2025-11-03T17:09:01.2268096Z + h11==0.14.0 2025-11-03T17:09:01.2268769Z + h5py==3.13.0 2025-11-03T17:09:01.2269652Z + harfile==0.3.0 2025-11-03T17:09:01.2270488Z - hf-xet==1.2.0 2025-11-03T17:09:01.2271188Z + hf-xet==1.1.7 2025-11-03T17:09:01.2272017Z + hiredis==3.0.0 2025-11-03T17:09:01.2273008Z + html2text==2025.4.15 2025-11-03T17:09:01.2273642Z - httpcore==1.0.9 2025-11-03T17:09:01.2274503Z + httpcore==1.0.6 2025-11-03T17:09:01.2275060Z - httpx==0.28.1 2025-11-03T17:09:01.2275708Z + httpx==0.27.2 2025-11-03T17:09:01.2276688Z - huggingface-hub==0.36.0 2025-11-03T17:09:01.2277473Z + huggingface-hub==0.34.3 2025-11-03T17:09:01.2278060Z + humanize==4.11.0 2025-11-03T17:09:01.2279096Z + hydra-core==1.3.2 2025-11-03T17:09:01.2279780Z - hypothesis==6.56.4 2025-11-03T17:09:01.2280510Z + hypothesis==6.131.0 2025-11-03T17:09:01.2281233Z + hypothesis-graphql==0.11.1 2025-11-03T17:09:01.2282153Z + hypothesis-jsonschema==0.23.1 2025-11-03T17:09:01.2282685Z - idna==3.11 2025-11-03T17:09:01.2283526Z + idna==3.10 2025-11-03T17:09:01.2284267Z + importlib-metadata==8.7.0 2025-11-03T17:09:01.2284848Z + importlib-resources==6.5.2 2025-11-03T17:09:01.2285645Z + inflect==5.6.2 2025-11-03T17:09:01.2286737Z - iniconfig==2.3.0 2025-11-03T17:09:01.2287185Z + iniconfig==2.0.0 2025-11-03T17:09:01.2288029Z + isoduration==20.11.0 2025-11-03T17:09:01.2288681Z + isort==5.13.2 2025-11-03T17:09:01.2289522Z + itsdangerous==2.2.0 2025-11-03T17:09:01.2290596Z + jiwer==3.0.5 2025-11-03T17:09:01.2291009Z - joblib==1.5.2 2025-11-03T17:09:01.2291904Z + joblib==1.4.2 2025-11-03T17:09:01.2292727Z + jsonargparse==4.35.0 2025-11-03T17:09:01.2293151Z + jsonlines==4.0.0 2025-11-03T17:09:01.2294038Z + jsonpointer==3.0.0 2025-11-03T17:09:01.2294619Z - jsonschema==4.25.1 2025-11-03T17:09:01.2295825Z + jsonschema==4.23.0 2025-11-03T17:09:01.2296525Z - jsonschema-specifications==2025.9.1 2025-11-03T17:09:01.2297256Z + jsonschema-specifications==2024.10.1 2025-11-03T17:09:01.2298306Z + junit-xml==1.9 2025-11-03T17:09:01.2298787Z + kaleido==0.2.1 2025-11-03T17:09:01.2299508Z + kiwisolver==1.4.7 2025-11-03T17:09:01.2300447Z + kornia==0.8.1 2025-11-03T17:09:01.2300919Z + kornia-rs==0.1.9 2025-11-03T17:09:01.2302447Z + libnacl==2.1.0 2025-11-03T17:09:01.2302910Z - librosa==0.10.2 2025-11-03T17:09:01.2303814Z + librosa==0.10.2.post1 2025-11-03T17:09:01.2304623Z + lightly==1.5.20 2025-11-03T17:09:01.2305170Z + lightly-utils==0.0.2 2025-11-03T17:09:01.2306235Z + lightning==2.5.1.post0 2025-11-03T17:09:01.2307320Z + lightning-utilities==0.14.3 2025-11-03T17:09:01.2308543Z + lm-eval==0.4.9.1 (from git+https://github.com/EleutherAI/lm-evaluation-harness.git@206b7722158f58c35b7ffcd53b035fdbdda5126d) 2025-11-03T17:09:01.2309650Z + mako==1.3.10 2025-11-03T17:09:01.2310102Z - markdown==3.9 2025-11-03T17:09:01.2311029Z + markdown==3.8.2 2025-11-03T17:09:01.2311567Z - markdown-it-py==4.0.0 2025-11-03T17:09:01.2312411Z + markdown-it-py==3.0.0 2025-11-03T17:09:01.2313304Z - markupsafe==3.0.3 2025-11-03T17:09:01.2313877Z + markupsafe==3.0.1 2025-11-03T17:09:01.2314852Z + matplotlib==3.9.2 2025-11-03T17:09:01.2315427Z + mbstrdecoder==1.1.3 2025-11-03T17:09:01.2316594Z + mlflow==2.22.0 2025-11-03T17:09:01.2317311Z + mlflow-skinny==2.22.0 2025-11-03T17:09:01.2318008Z + more-itertools==10.5.0 2025-11-03T17:09:01.2318906Z - msgpack==1.1.2 2025-11-03T17:09:01.2319483Z + msgpack==1.1.0 2025-11-03T17:09:01.2320471Z + mteb==1.38.11 2025-11-03T17:09:01.2321221Z - multidict==6.7.0 2025-11-03T17:09:01.2321897Z + multidict==6.1.0 2025-11-03T17:09:01.2322817Z + multiprocess==0.70.16 2025-11-03T17:09:01.2323807Z + munch==4.0.0 2025-11-03T17:09:01.2324440Z - mypy-extensions==1.1.0 2025-11-03T17:09:01.2325373Z + mypy-extensions==1.0.0 2025-11-03T17:09:01.2326233Z - networkx==2.8.8 2025-11-03T17:09:01.2326802Z + networkx==3.2.1 2025-11-03T17:09:01.2327948Z + nltk==3.9.1 2025-11-03T17:09:01.2328656Z + num2words==0.5.14 2025-11-03T17:09:01.2329872Z + numexpr==2.10.1 2025-11-03T17:09:01.2330664Z - numpy==2.2.6 2025-11-03T17:09:01.2331378Z + numpy==1.26.4 2025-11-03T17:09:01.2332289Z + omegaconf==2.3.0 2025-11-03T17:09:01.2332994Z + open-clip-torch==2.32.0 2025-11-03T17:09:01.2334114Z + opencensus==0.11.4 2025-11-03T17:09:01.2334578Z + opencensus-context==0.1.3 2025-11-03T17:09:01.2336325Z - opencv-python-headless==4.12.0.88 2025-11-03T17:09:01.2336976Z + opencv-python-headless==4.11.0.86 2025-11-03T17:09:01.2337439Z + opentelemetry-api==1.35.0 2025-11-03T17:09:01.2338113Z + opentelemetry-exporter-prometheus==0.56b0 2025-11-03T17:09:01.2339108Z + opentelemetry-proto==1.36.0 2025-11-03T17:09:01.2339695Z + opentelemetry-sdk==1.35.0 2025-11-03T17:09:01.2340598Z + opentelemetry-semantic-conventions==0.56b0 2025-11-03T17:09:01.2341162Z - packaging==25.0 2025-11-03T17:09:01.2341646Z + packaging==24.2 2025-11-03T17:09:01.2342275Z - pandas==2.0.3 2025-11-03T17:09:01.2344126Z + pandas==2.2.3 2025-11-03T17:09:01.2344450Z + pathvalidate==3.2.1 2025-11-03T17:09:01.2345064Z + patsy==1.0.1 2025-11-03T17:09:01.2345978Z + peft==0.16.0 2025-11-03T17:09:01.2346450Z - pillow==11.0.0 2025-11-03T17:09:01.2347275Z + pillow==10.4.0 2025-11-03T17:09:01.2347870Z - platformdirs==4.5.0 2025-11-03T17:09:01.2348885Z + platformdirs==4.3.6 2025-11-03T17:09:01.2349513Z + plotly==5.24.1 2025-11-03T17:09:01.2350260Z - pluggy==1.6.0 2025-11-03T17:09:01.2350864Z + pluggy==1.5.0 2025-11-03T17:09:01.2351836Z + polars==1.29.0 2025-11-03T17:09:01.2352588Z + portalocker==2.10.1 2025-11-03T17:09:01.2353133Z + pqdm==0.2.0 2025-11-03T17:09:01.2354177Z + pretrainedmodels==0.7.4 2025-11-03T17:09:01.2354837Z - prometheus-client==0.23.1 2025-11-03T17:09:01.2358074Z + prometheus-client==0.22.0 2025-11-03T17:09:01.2358677Z - propcache==0.4.1 2025-11-03T17:09:01.2359176Z + propcache==0.2.0 2025-11-03T17:09:01.2359665Z + proto-plus==1.26.1 2025-11-03T17:09:01.2360176Z - protobuf==5.29.5 2025-11-03T17:09:01.2360653Z + protobuf==5.28.3 2025-11-03T17:09:01.2361122Z - psutil==7.1.2 2025-11-03T17:09:01.2361672Z + psutil==6.1.0 2025-11-03T17:09:01.2362184Z + py==1.11.0 2025-11-03T17:09:01.2362728Z + py-spy==0.4.0 2025-11-03T17:09:01.2363821Z + pyarrow==18.0.0 2025-11-03T17:09:01.2364372Z + pybind11==2.13.6 2025-11-03T17:09:01.2365153Z + pycocotools==2.0.8 2025-11-03T17:09:01.2365694Z - pycparser==2.23 2025-11-03T17:09:01.2366329Z + pycparser==2.22 2025-11-03T17:09:01.2367176Z + pycryptodomex==3.22.0 2025-11-03T17:09:01.2367706Z - pydantic==2.12.3 2025-11-03T17:09:01.2368927Z + pydantic==2.12.0 2025-11-03T17:09:01.2369812Z - pydantic-core==2.41.4 2025-11-03T17:09:01.2370743Z + pydantic-core==2.41.1 2025-11-03T17:09:01.2371325Z - pydantic-extra-types==2.10.6 2025-11-03T17:09:01.2372725Z + pydantic-extra-types==2.10.5 2025-11-03T17:09:01.2373278Z - pygments==2.15.0 2025-11-03T17:09:01.2373852Z + pygments==2.18.0 2025-11-03T17:09:01.2374547Z + pyogrio==0.11.0 2025-11-03T17:09:01.2375640Z + pyparsing==3.2.0 2025-11-03T17:09:01.2376525Z + pyproj==3.7.1 2025-11-03T17:09:01.2377176Z + pyrate-limiter==3.7.0 2025-11-03T17:09:01.2377748Z + pystemmer==3.0.0 2025-11-03T17:09:01.2378883Z + pytablewriter==1.2.0 2025-11-03T17:09:01.2379446Z - pytest==7.3.2 2025-11-03T17:09:01.2380597Z + pytest==8.3.5 2025-11-03T17:09:01.2381204Z + pytest-asyncio==0.24.0 2025-11-03T17:09:01.2381908Z + pytest-cov==6.3.0 2025-11-03T17:09:01.2382928Z + pytest-forked==1.6.0 2025-11-03T17:09:01.2383535Z + pytest-mock==3.14.0 2025-11-03T17:09:01.2384511Z + pytest-shard==0.1.2 2025-11-03T17:09:01.2385549Z - pytest-subtests==0.13.1 2025-11-03T17:09:01.2386119Z + pytest-subtests==0.14.1 2025-11-03T17:09:01.2387031Z + pytest-timeout==2.3.1 2025-11-03T17:09:01.2387822Z + python-box==7.3.2 2025-11-03T17:09:01.2388291Z + python-rapidjson==1.20 2025-11-03T17:09:01.2389482Z + pytorch-lightning==2.5.5 2025-11-03T17:09:01.2389944Z + pytrec-eval-terrier==0.5.7 2025-11-03T17:09:01.2390943Z - pytz==2025.2 2025-11-03T17:09:01.2391777Z + pytz==2024.2 2025-11-03T17:09:01.2392306Z + rapidfuzz==3.12.1 2025-11-03T17:09:01.2393187Z + rasterio==1.4.3 2025-11-03T17:09:01.2394240Z - ray==2.51.1 2025-11-03T17:09:01.2395061Z + ray==2.48.0 2025-11-03T17:09:01.2395580Z - redis==7.0.1 2025-11-03T17:09:01.2396126Z + redis==5.2.0 2025-11-03T17:09:01.2396859Z - referencing==0.37.0 2025-11-03T17:09:01.2397693Z + referencing==0.35.1 2025-11-03T17:09:01.2398652Z - regex==2025.10.23 2025-11-03T17:09:01.2399395Z + regex==2024.9.11 2025-11-03T17:09:01.2400227Z - requests==2.32.5 2025-11-03T17:09:01.2401029Z + requests==2.32.3 2025-11-03T17:09:01.2401747Z + responses==0.25.3 2025-11-03T17:09:01.2402643Z + rfc3339-validator==0.1.4 2025-11-03T17:09:01.2403526Z + rfc3987==1.3.8 2025-11-03T17:09:01.2404308Z - rich==14.2.0 2025-11-03T17:09:01.2404787Z + rich==13.9.4 2025-11-03T17:09:01.2405647Z + rioxarray==0.19.0 2025-11-03T17:09:01.2406170Z + rouge-score==0.1.2 2025-11-03T17:09:01.2407141Z - rpds-py==0.28.0 2025-11-03T17:09:01.2408029Z + rpds-py==0.20.1 2025-11-03T17:09:01.2408679Z + rtree==1.4.0 2025-11-03T17:09:01.2409685Z + runai-model-streamer==0.14.0 2025-11-03T17:09:01.2410455Z + runai-model-streamer-gcs==0.14.0 2025-11-03T17:09:01.2411622Z + runai-model-streamer-s3==0.14.0 2025-11-03T17:09:01.2412446Z - s3transfer==0.10.4 2025-11-03T17:09:01.2413162Z + s3transfer==0.10.3 2025-11-03T17:09:01.2414063Z + sacrebleu==2.4.3 2025-11-03T17:09:01.2415141Z - safetensors==0.6.2 2025-11-03T17:09:01.2415816Z + safetensors==0.4.5 2025-11-03T17:09:01.2416854Z + schemathesis==3.39.15 2025-11-03T17:09:01.2417444Z - scikit-image==0.22.0 2025-11-03T17:09:01.2418530Z + scikit-image==0.25.2 2025-11-03T17:09:01.2419614Z - scikit-learn==1.7.2 2025-11-03T17:09:01.2420357Z + scikit-learn==1.5.2 2025-11-03T17:09:01.2421095Z - scipy==1.14.1 2025-11-03T17:09:01.2421635Z + scipy==1.13.1 2025-11-03T17:09:01.2422920Z + segmentation-models-pytorch==0.4.0 2025-11-03T17:09:01.2423647Z + sentence-transformers==3.2.1 2025-11-03T17:09:01.2424762Z - setuptools==78.1.1 2025-11-03T17:09:01.2425277Z + setuptools==77.0.3 2025-11-03T17:09:01.2426195Z + shapely==2.1.1 2025-11-03T17:09:01.2427273Z - six==1.17.0 2025-11-03T17:09:01.2428392Z + six==1.16.0 2025-11-03T17:09:01.2429023Z + smart-open==7.1.0 2025-11-03T17:09:01.2429749Z - soundfile==0.13.1 2025-11-03T17:09:01.2431440Z + soundfile==0.12.1 2025-11-03T17:09:01.2431928Z - soxr==1.0.0 2025-11-03T17:09:01.2432565Z + soxr==0.5.0.post1 2025-11-03T17:09:01.2433939Z + sqlalchemy==2.0.41 2025-11-03T17:09:01.2434969Z + sqlitedict==2.1.0 2025-11-03T17:09:01.2435467Z + sqlparse==0.5.3 2025-11-03T17:09:01.2436456Z - starlette==0.49.3 2025-11-03T17:09:01.2437399Z + starlette==0.46.2 2025-11-03T17:09:01.2437855Z + starlette-testclient==0.4.1 2025-11-03T17:09:01.2439051Z + statsmodels==0.14.4 2025-11-03T17:09:01.2439923Z + structlog==25.4.0 2025-11-03T17:09:01.2441035Z + tabledata==1.3.3 2025-11-03T17:09:01.2441610Z + tblib==3.1.0 2025-11-03T17:09:01.2442735Z + tcolorpy==0.1.6 2025-11-03T17:09:01.2443397Z + tenacity==9.1.2 2025-11-03T17:09:01.2444100Z + tensorboardx==2.6.4 2025-11-03T17:09:01.2445177Z + tensorizer==2.10.1 2025-11-03T17:09:01.2445899Z + termcolor==3.1.0 2025-11-03T17:09:01.2447708Z + terratorch==1.0.2 (from git+https://github.com/IBM/terratorch.git@07184fcf91a1324f831ff521dd238d97fe350e3e) 2025-11-03T17:09:01.2448539Z - threadpoolctl==3.6.0 2025-11-03T17:09:01.2449332Z + threadpoolctl==3.5.0 2025-11-03T17:09:01.2449801Z - tifffile==2025.10.16 2025-11-03T17:09:01.2450408Z + tifffile==2025.3.30 2025-11-03T17:09:01.2451682Z + timm==1.0.17 2025-11-03T17:09:01.2452556Z - tokenizers==0.22.1 2025-11-03T17:09:01.2453362Z + tokenizers==0.22.0 2025-11-03T17:09:01.2467645Z + tomli==2.2.1 2025-11-03T17:09:01.2468236Z + tomli-w==1.2.0 2025-11-03T17:09:01.2469382Z - torch==2.10.0a0+git3f6538f (from file:///var/lib/jenkins/workspace/dist/torch-2.10.0a0%2Bgit3f6538f-cp312-cp312-linux_x86_64.whl) 2025-11-03T17:09:01.2470992Z + torch==2.10.0a0+git3f6538f (from file:///var/lib/jenkins/workspace/dist/torch-2.10.0a0+git3f6538f-cp312-cp312-linux_x86_64.whl#sha256=288b4779bb4efa62cf5264c0eca9fdb02025e7a7fb44d42e3f512a68d8e60da0) 2025-11-03T17:09:01.2472907Z - torchaudio==2.10.0a0+3b0e7a6 (from file:///var/lib/jenkins/workspace/dist/audio/torchaudio-2.10.0a0%2B3b0e7a6-cp312-cp312-linux_x86_64.whl) 2025-11-03T17:09:01.2474876Z + torchaudio==2.10.0a0+3b0e7a6 (from file:///var/lib/jenkins/workspace/dist/audio/torchaudio-2.10.0a0+3b0e7a6-cp312-cp312-linux_x86_64.whl#sha256=70d7f19c17cbc4bb012226c17e5f5380b3bd6bda732a45f329984580e0bc17a6) 2025-11-03T17:09:01.2476360Z + torchgeo==0.7.0 2025-11-03T17:09:01.2476872Z + torchmetrics==1.8.2 2025-11-03T17:09:01.2478108Z - torchvision==0.25.0a0+cfbc5c2 (from file:///var/lib/jenkins/workspace/dist/vision/torchvision-0.25.0a0%2Bcfbc5c2-cp312-cp312-linux_x86_64.whl) 2025-11-03T17:09:01.2480100Z + torchvision==0.25.0a0+cfbc5c2 (from file:///var/lib/jenkins/workspace/dist/vision/torchvision-0.25.0a0+cfbc5c2-cp312-cp312-linux_x86_64.whl#sha256=7fa43396b4b9ce1b239050abea02e7fbabadf1cbdf180883becb1f6530983b15) 2025-11-03T17:09:01.2481288Z - tqdm==4.67.1 2025-11-03T17:09:01.2481722Z + tqdm==4.66.6 2025-11-03T17:09:01.2482192Z + tqdm-multiprocess==0.0.11 2025-11-03T17:09:01.2482845Z - transformers==4.57.1 2025-11-03T17:09:01.2483265Z + transformers==4.56.2 2025-11-03T17:09:01.2483876Z + transformers-stream-generator==0.0.5 2025-11-03T17:09:01.2484354Z + tritonclient==2.51.0 2025-11-03T17:09:01.2484812Z + typepy==1.3.2 2025-11-03T17:09:01.2485160Z - typer==0.20.0 2025-11-03T17:09:01.2485513Z + typer==0.15.2 2025-11-03T17:09:01.2486020Z + types-python-dateutil==2.9.0.20241206 2025-11-03T17:09:01.2486488Z + typeshed-client==2.8.2 2025-11-03T17:09:01.2486894Z - tzdata==2025.2 2025-11-03T17:09:01.2487290Z + tzdata==2024.2 2025-11-03T17:09:01.2487637Z + uri-template==1.3.0 2025-11-03T17:09:01.2488102Z - urllib3==2.5.0 2025-11-03T17:09:01.2488657Z + urllib3==2.2.3 2025-11-03T17:09:01.2488996Z - uvicorn==0.38.0 2025-11-03T17:09:01.2489322Z + uvicorn==0.35.0 2025-11-03T17:09:01.2489746Z + vector-quantize-pytorch==1.24.2 2025-11-03T17:09:01.2490201Z + virtualenv==20.31.2 2025-11-03T17:09:01.2490581Z + vocos==0.1.0 2025-11-03T17:09:01.2490928Z + wcwidth==0.2.13 2025-11-03T17:09:01.2491242Z + webcolors==24.11.1 2025-11-03T17:09:01.2491518Z + word2number==1.1 2025-11-03T17:09:01.2491782Z - wrapt==2.0.0 2025-11-03T17:09:01.2492039Z + wrapt==1.17.2 2025-11-03T17:09:01.2492295Z + xarray==2025.7.1 2025-11-03T17:09:01.2492552Z + xxhash==3.5.0 2025-11-03T17:09:01.2492795Z - yarl==1.22.0 2025-11-03T17:09:01.2493049Z + yarl==1.17.1 2025-11-03T17:09:01.2493291Z + zipp==3.23.0 2025-11-03T17:09:01.2493554Z + zstandard==0.23.0 2025-11-03T17:09:01.4384533Z 2025-11-03 17:09:01,437 [INFO] cli.lib.core.vllm.vllm_test: Done. installed requirements for test dependencies 2025-11-03T17:09:01.4385205Z 2025-11-03 17:09:01,438 [INFO] cli.lib.core.vllm.vllm_test: Double check installed packages 2025-11-03T17:09:01.4433220Z 2025-11-03 17:09:01,443 [INFO] cli.lib.common.pip_helper: torch already exist with version: 2.10.0a0+git3f6538f 2025-11-03T17:09:01.4436484Z 2025-11-03 17:09:01,443 [INFO] cli.lib.common.pip_helper: xformers already exist with version: 0.0.33+5d4b92a5.d20251103 2025-11-03T17:09:01.4439989Z 2025-11-03 17:09:01,443 [INFO] cli.lib.common.pip_helper: torchvision already exist with version: 0.25.0a0+cfbc5c2 2025-11-03T17:09:01.4443634Z 2025-11-03 17:09:01,444 [INFO] cli.lib.common.pip_helper: torchaudio already exist with version: 2.10.0a0+3b0e7a6 2025-11-03T17:09:01.4448967Z 2025-11-03 17:09:01,444 [INFO] cli.lib.common.pip_helper: vllm already exist with version: 0.11.0rc2.dev389+ge51928192.d20251103 2025-11-03T17:09:01.4449522Z 2025-11-03 17:09:01,444 [INFO] cli.lib.core.vllm.vllm_test: Done. checked installed packages 2025-11-03T17:09:01.4450018Z 2025-11-03 17:09:01,444 [INFO] cli.lib.core.vllm.lib: run vllm tests..... 2025-11-03T17:09:01.4450489Z 2025-11-03 17:09:01,444 [INFO] cli.lib.core.vllm.lib: Running tests: Distributed Tests (2 GPUs) pytorch 2.8 release failure 2025-11-03T17:09:01.4451068Z 2025-11-03 17:09:01,444 [INFO] cli.lib.core.vllm.lib: Running step: pytest -v -s distributed/test_sequence_parallel.py 2025-11-03T17:09:01.4451618Z 2025-11-03 17:09:01,444 [INFO] cli.lib.common.utils: [shell] pytest -v -s distributed/test_sequence_parallel.py 2025-11-03T17:09:11.4373178Z INFO 11-03 17:09:11 [__init__.py:224] Automatically detected platform cuda. 2025-11-03T17:09:13.8620967Z /opt/conda/envs/py_3.12/lib/python3.12/site-packages/pytest_asyncio/plugin.py:208: PytestDeprecationWarning: The configuration option "asyncio_default_fixture_loop_scope" is unset. 2025-11-03T17:09:13.8625548Z The event loop scope for asynchronous fixtures will default to the fixture caching scope. Future versions of pytest-asyncio will default the loop scope for asynchronous fixtures to function scope. Set the default fixture loop scope explicitly in order to avoid unexpected behavior in the future. Valid fixture loop scopes are: "function", "class", "module", "package", "session" 2025-11-03T17:09:13.8626669Z 2025-11-03T17:09:13.8626861Z warnings.warn(PytestDeprecationWarning(_DEFAULT_FIXTURE_LOOP_SCOPE_UNSET)) 2025-11-03T17:09:13.8694833Z ============================= test session starts ============================== 2025-11-03T17:09:13.8695896Z platform linux -- Python 3.12.12, pytest-8.3.5, pluggy-1.5.0 -- /opt/conda/envs/py_3.12/bin/python 2025-11-03T17:09:13.8787841Z cachedir: .pytest_cache 2025-11-03T17:09:13.8788965Z hypothesis profile 'ci' -> database=None, deadline=None, print_blob=True, derandomize=True, suppress_health_check=[HealthCheck.too_slow] 2025-11-03T17:09:13.8790114Z rootdir: /var/lib/jenkins/workspace/vllm 2025-11-03T17:09:13.8791080Z configfile: pyproject.toml 2025-11-03T17:09:13.8793033Z plugins: cpp-2.3.0, flakefinder-1.1.0, rerunfailures-14.0, xdist-3.3.1, xdoctest-1.3.0, cov-6.3.0, timeout-2.3.1, hydra-core-1.3.2, buildkite-test-collector-0.1.9, forked-1.6.0, asyncio-0.24.0, shard-0.1.2, mock-3.14.0, schemathesis-3.39.15, hypothesis-6.131.0, subtests-0.14.1, anyio-4.6.2.post1 2025-11-03T17:09:13.8794817Z asyncio: mode=Mode.STRICT, default_loop_scope=None 2025-11-03T17:09:17.0681586Z collecting ...  2025-11-03T17:09:17.0684924Z collecting 20 items  2025-11-03T17:09:17.0685951Z collected 20 items  2025-11-03T17:09:17.0695506Z Running 20 items in this shard: tests/distributed/test_sequence_parallel.py::test_tp_sp_generation[meta-llama/Llama-3.2-1B-Instruct-parallel_setup0-mp-auto-test_options0], tests/distributed/test_sequence_parallel.py::test_tp_sp_generation[meta-llama/Llama-3.2-1B-Instruct-parallel_setup1-ray-auto-test_options1], tests/distributed/test_sequence_parallel.py::test_tp_sp_generation[meta-llama/Llama-3.2-1B-Instruct-parallel_setup2-mp-auto-test_options2], tests/distributed/test_sequence_parallel.py::test_tp_sp_generation[meta-llama/Llama-3.2-1B-Instruct-parallel_setup3-ray-auto-test_options3], tests/distributed/test_sequence_parallel.py::test_tp_sp_generation[meta-llama/Llama-3.2-1B-Instruct-parallel_setup4-mp-auto-test_options4], tests/distributed/test_sequence_parallel.py::test_tp_sp_generation[meta-llama/Llama-3.2-1B-Instruct-parallel_setup5-ray-auto-test_options5], tests/distributed/test_sequence_parallel.py::test_tp_sp_generation[meta-llama/Llama-3.2-1B-Instruct-parallel_setup6-mp-auto-test_options6], tests/distributed/test_sequence_parallel.py::test_tp_sp_generation[meta-llama/Llama-3.2-1B-Instruct-parallel_setup7-ray-auto-test_options7], tests/distributed/test_sequence_parallel.py::test_tp_sp_generation[meta-llama/Llama-3.2-1B-Instruct-parallel_setup8-mp-auto-test_options8], tests/distributed/test_sequence_parallel.py::test_tp_sp_generation[meta-llama/Llama-3.2-1B-Instruct-parallel_setup9-ray-auto-test_options9], tests/distributed/test_sequence_parallel.py::test_tp_sp_generation[meta-llama/Llama-3.2-1B-Instruct-parallel_setup10-mp-auto-test_options10], tests/distributed/test_sequence_parallel.py::test_tp_sp_generation[meta-llama/Llama-3.2-1B-Instruct-parallel_setup11-ray-auto-test_options11], tests/distributed/test_sequence_parallel.py::test_tp_sp_generation[meta-llama/Llama-3.2-1B-Instruct-parallel_setup12-mp-auto-test_options12], tests/distributed/test_sequence_parallel.py::test_tp_sp_generation[meta-llama/Llama-3.2-1B-Instruct-parallel_setup13-ray-auto-test_options13], tests/distributed/test_sequence_parallel.py::test_tp_sp_generation[meta-llama/Llama-3.2-1B-Instruct-parallel_setup14-mp-auto-test_options14], tests/distributed/test_sequence_parallel.py::test_tp_sp_generation[meta-llama/Llama-3.2-1B-Instruct-parallel_setup15-ray-auto-test_options15], tests/distributed/test_sequence_parallel.py::test_tp_sp_generation[RedHatAI/Meta-Llama-3.1-8B-Instruct-FP8-parallel_setup16-mp-auto-test_options16], tests/distributed/test_sequence_parallel.py::test_tp_sp_generation[RedHatAI/Meta-Llama-3.1-8B-Instruct-FP8-parallel_setup17-ray-auto-test_options17], tests/distributed/test_sequence_parallel.py::test_tp_sp_generation[RedHatAI/Meta-Llama-3.1-8B-Instruct-FP8-parallel_setup18-mp-auto-test_options18], tests/distributed/test_sequence_parallel.py::test_tp_sp_generation[RedHatAI/Meta-Llama-3.1-8B-Instruct-FP8-parallel_setup19-ray-auto-test_options19] 2025-11-03T17:09:17.0704134Z 2025-11-03T17:09:17.0839469Z distributed/test_sequence_parallel.py::test_tp_sp_generation[meta-llama/Llama-3.2-1B-Instruct-parallel_setup0-mp-auto-test_options0] Fork a new process to run a test 1224 2025-11-03T17:09:17.0852924Z Fork a new process to run a test 0 2025-11-03T17:09:17.1487417Z 2025-11-03T17:09:17.1491844Z tokenizer_config.json: 0% 0.00/54.5k [00:00, 'use_cudagraph': True, 'cudagraph_num_of_warmups': 1, 'cudagraph_capture_sizes': [16, 8, 4, 2], 'cudagraph_copy_inputs': False, 'full_cuda_graph': True, 'use_inductor_graph_partition': False, 'pass_config': {'enable_noop': True, 'enable_sequence_parallelism': True}, 'max_capture_size': 16, 'local_cache_dir': None} 2025-11-03T17:09:39.4598032Z (EngineCore_DP0 pid=1511) WARNING 11-03 17:09:39 [multiproc_executor.py:744] Reducing Torch parallelism from 24 threads to 1 to avoid unnecessary CPU contention. Set OMP_NUM_THREADS in the external environment to tune this value as needed. 2025-11-03T17:09:39.4602143Z (EngineCore_DP0 pid=1511) INFO 11-03 17:09:39 [shm_broadcast.py:302] vLLM message queue communication handle: Handle(local_reader_ranks=[0, 1], buffer_handle=(2, 16777216, 10, 'psm_98b10687'), local_subscribe_addr='ipc:///tmp/822fa556-8f53-4c97-a655-9ffd0f6bad2f', remote_subscribe_addr=None, remote_addr_ipv6=False) 2025-11-03T17:09:39.9470883Z /opt/conda/envs/py_3.12/lib/python3.12/site-packages/torch/cuda/__init__.py:65: FutureWarning: The pynvml package is deprecated. Please install nvidia-ml-py instead. If you did not install pynvml directly, please report this to the maintainers of the package that installed pynvml for you. 2025-11-03T17:09:39.9472475Z import pynvml # type: ignore[import] 2025-11-03T17:09:39.9602787Z /opt/conda/envs/py_3.12/lib/python3.12/site-packages/torch/cuda/__init__.py:65: FutureWarning: The pynvml package is deprecated. Please install nvidia-ml-py instead. If you did not install pynvml directly, please report this to the maintainers of the package that installed pynvml for you. 2025-11-03T17:09:39.9604208Z import pynvml # type: ignore[import] 2025-11-03T17:09:42.3224236Z INFO 11-03 17:09:42 [__init__.py:224] Automatically detected platform cuda. 2025-11-03T17:09:42.3780099Z INFO 11-03 17:09:42 [__init__.py:224] Automatically detected platform cuda. 2025-11-03T17:09:47.9330601Z INFO 11-03 17:09:47 [shm_broadcast.py:302] vLLM message queue communication handle: Handle(local_reader_ranks=[0], buffer_handle=(1, 10485760, 10, 'psm_e2d5173f'), local_subscribe_addr='ipc:///tmp/47350817-6a87-41fd-9b25-1d2abb365bfe', remote_subscribe_addr=None, remote_addr_ipv6=False) 2025-11-03T17:09:47.9332569Z INFO 11-03 17:09:47 [shm_broadcast.py:302] vLLM message queue communication handle: Handle(local_reader_ranks=[0], buffer_handle=(1, 10485760, 10, 'psm_10d25f7b'), local_subscribe_addr='ipc:///tmp/d0ee36de-6699-4ed1-a579-63cf6eab1d91', remote_subscribe_addr=None, remote_addr_ipv6=False) 2025-11-03T17:09:48.1280160Z [Gloo] Rank 1 is connected to 1 peer ranks. Expected number of connected peer ranks is : 1 2025-11-03T17:09:48.1280803Z [Gloo] Rank 0 is connected to 1 peer ranks. Expected number of connected peer ranks is : 1 2025-11-03T17:09:48.1324889Z [Gloo] Rank 1 is connected to 1 peer ranks. Expected number of connected peer ranks is : 1 2025-11-03T17:09:48.1325554Z [Gloo] Rank 0 is connected to 1 peer ranks. Expected number of connected peer ranks is : 1 2025-11-03T17:09:48.1458561Z INFO 11-03 17:09:48 [__init__.py:1146] Found nccl from library libnccl.so.2 2025-11-03T17:09:48.1459051Z INFO 11-03 17:09:48 [__init__.py:1146] Found nccl from library libnccl.so.2 2025-11-03T17:09:48.1460498Z INFO 11-03 17:09:48 [pynccl.py:109] vLLM is using nccl==2.27.5 2025-11-03T17:09:48.1460935Z INFO 11-03 17:09:48 [pynccl.py:109] vLLM is using nccl==2.27.5 2025-11-03T17:09:48.3467553Z WARNING 11-03 17:09:48 [symm_mem.py:60] SymmMemCommunicator: Device capability 8.9 not supported, communicator is not available. 2025-11-03T17:09:48.3468447Z WARNING 11-03 17:09:48 [symm_mem.py:60] SymmMemCommunicator: Device capability 8.9 not supported, communicator is not available. 2025-11-03T17:09:48.3489652Z INFO 11-03 17:09:48 [custom_all_reduce.py:37] Skipping P2P check and trusting the driver's P2P report. 2025-11-03T17:09:48.3490293Z INFO 11-03 17:09:48 [custom_all_reduce.py:37] Skipping P2P check and trusting the driver's P2P report. 2025-11-03T17:09:48.3491283Z WARNING 11-03 17:09:48 [custom_all_reduce.py:162] Custom allreduce is disabled because your platform lacks GPU P2P capability or P2P test failed. To silence this warning, specify disable_custom_all_reduce=True explicitly. 2025-11-03T17:09:48.3493145Z WARNING 11-03 17:09:48 [custom_all_reduce.py:162] Custom allreduce is disabled because your platform lacks GPU P2P capability or P2P test failed. To silence this warning, specify disable_custom_all_reduce=True explicitly. 2025-11-03T17:09:48.4124582Z INFO 11-03 17:09:48 [shm_broadcast.py:302] vLLM message queue communication handle: Handle(local_reader_ranks=[1], buffer_handle=(1, 4194304, 6, 'psm_3f3aefdc'), local_subscribe_addr='ipc:///tmp/cc7322e7-6a49-4bdf-8534-b20fa081a2b3', remote_subscribe_addr=None, remote_addr_ipv6=False) 2025-11-03T17:09:48.4163851Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:09:48.4166259Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:09:48.4172553Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:09:48.4180026Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:09:48.4721467Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:09:48.4731198Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:09:48.4742248Z [Gloo] Rank 0 is connected to 1 peer ranks. Expected number of connected peer ranks is : 1 2025-11-03T17:09:48.4742731Z [Gloo] Rank 1 is connected to 1 peer ranks. Expected number of connected peer ranks is : 1 2025-11-03T17:09:48.4745512Z INFO 11-03 17:09:48 [__init__.py:1146] Found nccl from library libnccl.so.2 2025-11-03T17:09:48.4745978Z INFO 11-03 17:09:48 [__init__.py:1146] Found nccl from library libnccl.so.2 2025-11-03T17:09:48.4746320Z INFO 11-03 17:09:48 [pynccl.py:109] vLLM is using nccl==2.27.5 2025-11-03T17:09:48.4746746Z INFO 11-03 17:09:48 [pynccl.py:109] vLLM is using nccl==2.27.5 2025-11-03T17:09:48.5065989Z INFO 11-03 17:09:48 [parallel_state.py:1231] rank 0 in world size 2 is assigned as DP rank 0, PP rank 0, TP rank 0, EP rank 0 2025-11-03T17:09:48.5066777Z INFO 11-03 17:09:48 [parallel_state.py:1231] rank 1 in world size 2 is assigned as DP rank 0, PP rank 0, TP rank 1, EP rank 1 2025-11-03T17:09:48.7444611Z INFO 11-03 17:09:48 [topk_topp_sampler.py:59] Using FlashInfer for top-p & top-k sampling. 2025-11-03T17:09:48.7656567Z INFO 11-03 17:09:48 [topk_topp_sampler.py:59] Using FlashInfer for top-p & top-k sampling. 2025-11-03T17:09:48.7770452Z (Worker_TP0 pid=1630) INFO 11-03 17:09:48 [gpu_model_runner.py:2840] Starting to load model meta-llama/Llama-3.2-1B-Instruct... 2025-11-03T17:09:48.7967932Z (Worker_TP1 pid=1631) INFO 11-03 17:09:48 [gpu_model_runner.py:2840] Starting to load model meta-llama/Llama-3.2-1B-Instruct... 2025-11-03T17:09:48.9914022Z (Worker_TP0 pid=1630) INFO 11-03 17:09:48 [gpu_model_runner.py:2870] Loading model from scratch... 2025-11-03T17:09:49.0068749Z (Worker_TP1 pid=1631) INFO 11-03 17:09:49 [gpu_model_runner.py:2870] Loading model from scratch... 2025-11-03T17:09:49.0814880Z (Worker_TP0 pid=1630) INFO 11-03 17:09:49 [cuda.py:404] Using Flash Attention backend on V1 engine. 2025-11-03T17:09:49.0965155Z (Worker_TP1 pid=1631) INFO 11-03 17:09:49 [cuda.py:404] Using Flash Attention backend on V1 engine. 2025-11-03T17:09:49.5820643Z (Worker_TP1 pid=1631) INFO 11-03 17:09:49 [gpu_model_runner.py:2902] Model loading took 1.1667 GiB and 0.168746 seconds 2025-11-03T17:09:49.5828294Z (Worker_TP0 pid=1630) INFO 11-03 17:09:49 [gpu_model_runner.py:2902] Model loading took 1.1667 GiB and 0.173434 seconds 2025-11-03T17:09:53.0434479Z (Worker_TP0 pid=1630) INFO 11-03 17:09:53 [backends.py:594] Using cache directory: /var/lib/jenkins/.cache/vllm/torch_compile_cache/618b881a70/rank_0_0/backbone for vLLM's torch.compile 2025-11-03T17:09:53.0437298Z (Worker_TP0 pid=1630) INFO 11-03 17:09:53 [backends.py:608] Dynamo bytecode transform time: 3.19 s 2025-11-03T17:09:53.0766258Z (Worker_TP1 pid=1631) INFO 11-03 17:09:53 [backends.py:594] Using cache directory: /var/lib/jenkins/.cache/vllm/torch_compile_cache/618b881a70/rank_1_0/backbone for vLLM's torch.compile 2025-11-03T17:09:53.0770763Z (Worker_TP1 pid=1631) INFO 11-03 17:09:53 [backends.py:608] Dynamo bytecode transform time: 3.22 s 2025-11-03T17:09:54.3000304Z (Worker_TP0 pid=1630) [rank0]:W1103 17:09:54.298000 1630 site-packages/torch/_inductor/utils.py:1665] [0/0] Not enough SMs to use max_autotune_gemm mode 2025-11-03T17:09:54.3043373Z (Worker_TP1 pid=1631) [rank1]:W1103 17:09:54.303000 1631 site-packages/torch/_inductor/utils.py:1665] [0/0] Not enough SMs to use max_autotune_gemm mode 2025-11-03T17:09:55.8906143Z (Worker_TP0 pid=1630) INFO 11-03 17:09:55 [backends.py:214] Cache the graph for dynamic shape for later use 2025-11-03T17:09:55.9034201Z (Worker_TP1 pid=1631) INFO 11-03 17:09:55 [backends.py:214] Cache the graph for dynamic shape for later use 2025-11-03T17:10:02.2672197Z (Worker_TP1 pid=1631) INFO 11-03 17:10:02 [backends.py:241] Compiling a graph for dynamic shape takes 8.92 s 2025-11-03T17:10:02.4939753Z (Worker_TP0 pid=1630) INFO 11-03 17:10:02 [backends.py:241] Compiling a graph for dynamic shape takes 9.18 s 2025-11-03T17:10:09.6721661Z (Worker_TP0 pid=1630) INFO 11-03 17:10:09 [gpu_worker.py:314] Available KV cache memory: 18.50 GiB 2025-11-03T17:10:09.6913369Z (Worker_TP1 pid=1631) INFO 11-03 17:10:09 [gpu_worker.py:314] Available KV cache memory: 18.50 GiB 2025-11-03T17:10:09.9684271Z (EngineCore_DP0 pid=1511) INFO 11-03 17:10:09 [kv_cache_utils.py:1199] GPU KV cache size: 1,212,160 tokens 2025-11-03T17:10:09.9685236Z (EngineCore_DP0 pid=1511) INFO 11-03 17:10:09 [kv_cache_utils.py:1204] Maximum concurrency for 2,048 tokens per request: 591.88x 2025-11-03T17:10:09.9686067Z (EngineCore_DP0 pid=1511) INFO 11-03 17:10:09 [kv_cache_utils.py:1199] GPU KV cache size: 1,212,160 tokens 2025-11-03T17:10:09.9687184Z (EngineCore_DP0 pid=1511) INFO 11-03 17:10:09 [kv_cache_utils.py:1204] Maximum concurrency for 2,048 tokens per request: 591.88x 2025-11-03T17:10:10.2690539Z (Worker_TP0 pid=1630) 2025-11-03T17:10:11.4127941Z Capturing CUDA graphs (mixed prefill-decode, PIECEWISE): 0% 0/4 [00:00, 'use_cudagraph': True, 'cudagraph_num_of_warmups': 1, 'cudagraph_capture_sizes': [16, 8, 4, 2, 1], 'cudagraph_copy_inputs': False, 'full_cuda_graph': True, 'use_inductor_graph_partition': False, 'pass_config': {}, 'max_capture_size': 16, 'local_cache_dir': None} 2025-11-03T17:10:50.3337989Z (EngineCore_DP0 pid=3619) WARNING 11-03 17:10:50 [multiproc_executor.py:744] Reducing Torch parallelism from 24 threads to 1 to avoid unnecessary CPU contention. Set OMP_NUM_THREADS in the external environment to tune this value as needed. 2025-11-03T17:10:50.3341150Z (EngineCore_DP0 pid=3619) INFO 11-03 17:10:50 [shm_broadcast.py:302] vLLM message queue communication handle: Handle(local_reader_ranks=[0, 1], buffer_handle=(2, 16777216, 10, 'psm_4d5b3002'), local_subscribe_addr='ipc:///tmp/b6b2b08b-6121-438c-86b3-25e5958cd44b', remote_subscribe_addr=None, remote_addr_ipv6=False) 2025-11-03T17:10:50.8174284Z /opt/conda/envs/py_3.12/lib/python3.12/site-packages/torch/cuda/__init__.py:65: FutureWarning: The pynvml package is deprecated. Please install nvidia-ml-py instead. If you did not install pynvml directly, please report this to the maintainers of the package that installed pynvml for you. 2025-11-03T17:10:50.8176695Z import pynvml # type: ignore[import] 2025-11-03T17:10:50.8377603Z /opt/conda/envs/py_3.12/lib/python3.12/site-packages/torch/cuda/__init__.py:65: FutureWarning: The pynvml package is deprecated. Please install nvidia-ml-py instead. If you did not install pynvml directly, please report this to the maintainers of the package that installed pynvml for you. 2025-11-03T17:10:50.8379081Z import pynvml # type: ignore[import] 2025-11-03T17:10:53.1481429Z INFO 11-03 17:10:53 [__init__.py:224] Automatically detected platform cuda. 2025-11-03T17:10:53.2147679Z INFO 11-03 17:10:53 [__init__.py:224] Automatically detected platform cuda. 2025-11-03T17:10:57.9628804Z INFO 11-03 17:10:57 [shm_broadcast.py:302] vLLM message queue communication handle: Handle(local_reader_ranks=[0], buffer_handle=(1, 10485760, 10, 'psm_244050c0'), local_subscribe_addr='ipc:///tmp/9d4beee9-bd7c-4fd2-9878-945c5db88412', remote_subscribe_addr=None, remote_addr_ipv6=False) 2025-11-03T17:10:58.0022626Z INFO 11-03 17:10:58 [shm_broadcast.py:302] vLLM message queue communication handle: Handle(local_reader_ranks=[0], buffer_handle=(1, 10485760, 10, 'psm_7dde6d32'), local_subscribe_addr='ipc:///tmp/65dc9b50-64cc-4c5f-a18d-486bd0e834b1', remote_subscribe_addr=None, remote_addr_ipv6=False) 2025-11-03T17:10:58.1755834Z [Gloo] Rank 0 is connected to 1 peer ranks. Expected number of connected peer ranks is : 1 2025-11-03T17:10:58.1757059Z [Gloo] Rank 1 is connected to 1 peer ranks. Expected number of connected peer ranks is : 1 2025-11-03T17:10:58.1797524Z [Gloo] Rank 1 is connected to 1 peer ranks. Expected number of connected peer ranks is : 1 2025-11-03T17:10:58.1798198Z [Gloo] Rank 0 is connected to 1 peer ranks. Expected number of connected peer ranks is : 1 2025-11-03T17:10:58.1841221Z INFO 11-03 17:10:58 [__init__.py:1146] Found nccl from library libnccl.so.2 2025-11-03T17:10:58.1842200Z INFO 11-03 17:10:58 [__init__.py:1146] Found nccl from library libnccl.so.2 2025-11-03T17:10:58.1843031Z INFO 11-03 17:10:58 [pynccl.py:109] vLLM is using nccl==2.27.5 2025-11-03T17:10:58.1843750Z INFO 11-03 17:10:58 [pynccl.py:109] vLLM is using nccl==2.27.5 2025-11-03T17:10:58.3572264Z WARNING 11-03 17:10:58 [symm_mem.py:60] SymmMemCommunicator: Device capability 8.9 not supported, communicator is not available. 2025-11-03T17:10:58.3573900Z WARNING 11-03 17:10:58 [symm_mem.py:60] SymmMemCommunicator: Device capability 8.9 not supported, communicator is not available. 2025-11-03T17:10:58.3597936Z INFO 11-03 17:10:58 [custom_all_reduce.py:37] Skipping P2P check and trusting the driver's P2P report. 2025-11-03T17:10:58.3598731Z INFO 11-03 17:10:58 [custom_all_reduce.py:37] Skipping P2P check and trusting the driver's P2P report. 2025-11-03T17:10:58.3600097Z WARNING 11-03 17:10:58 [custom_all_reduce.py:162] Custom allreduce is disabled because your platform lacks GPU P2P capability or P2P test failed. To silence this warning, specify disable_custom_all_reduce=True explicitly. 2025-11-03T17:10:58.3601478Z WARNING 11-03 17:10:58 [custom_all_reduce.py:162] Custom allreduce is disabled because your platform lacks GPU P2P capability or P2P test failed. To silence this warning, specify disable_custom_all_reduce=True explicitly. 2025-11-03T17:10:58.4325102Z INFO 11-03 17:10:58 [shm_broadcast.py:302] vLLM message queue communication handle: Handle(local_reader_ranks=[1], buffer_handle=(1, 4194304, 6, 'psm_5c614f24'), local_subscribe_addr='ipc:///tmp/c7687e11-11a4-4f6c-9c2d-7858c36a4ed1', remote_subscribe_addr=None, remote_addr_ipv6=False) 2025-11-03T17:10:58.4358579Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:10:58.4359730Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:10:58.4365981Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:10:58.4375267Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:10:58.4822121Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:10:58.4829188Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:10:58.4842758Z [Gloo] Rank 1 is connected to 1 peer ranks. Expected number of connected peer ranks is : 1 2025-11-03T17:10:58.4844124Z [Gloo] Rank 0 is connected to 1 peer ranks. Expected number of connected peer ranks is : 1 2025-11-03T17:10:58.4844918Z INFO 11-03 17:10:58 [__init__.py:1146] Found nccl from library libnccl.so.2 2025-11-03T17:10:58.4845479Z INFO 11-03 17:10:58 [pynccl.py:109] vLLM is using nccl==2.27.5 2025-11-03T17:10:58.4845906Z INFO 11-03 17:10:58 [__init__.py:1146] Found nccl from library libnccl.so.2 2025-11-03T17:10:58.4846478Z INFO 11-03 17:10:58 [pynccl.py:109] vLLM is using nccl==2.27.5 2025-11-03T17:10:58.5182830Z INFO 11-03 17:10:58 [parallel_state.py:1231] rank 0 in world size 2 is assigned as DP rank 0, PP rank 0, TP rank 0, EP rank 0 2025-11-03T17:10:58.5183659Z INFO 11-03 17:10:58 [parallel_state.py:1231] rank 1 in world size 2 is assigned as DP rank 0, PP rank 0, TP rank 1, EP rank 1 2025-11-03T17:10:58.7391303Z INFO 11-03 17:10:58 [topk_topp_sampler.py:59] Using FlashInfer for top-p & top-k sampling. 2025-11-03T17:10:58.7534508Z INFO 11-03 17:10:58 [topk_topp_sampler.py:59] Using FlashInfer for top-p & top-k sampling. 2025-11-03T17:10:58.7708558Z (Worker_TP1 pid=3739) INFO 11-03 17:10:58 [gpu_model_runner.py:2840] Starting to load model meta-llama/Llama-3.2-1B-Instruct... 2025-11-03T17:10:58.7852688Z (Worker_TP0 pid=3738) INFO 11-03 17:10:58 [gpu_model_runner.py:2840] Starting to load model meta-llama/Llama-3.2-1B-Instruct... 2025-11-03T17:10:58.9716370Z (Worker_TP1 pid=3739) INFO 11-03 17:10:58 [gpu_model_runner.py:2870] Loading model from scratch... 2025-11-03T17:10:58.9925502Z (Worker_TP0 pid=3738) INFO 11-03 17:10:58 [gpu_model_runner.py:2870] Loading model from scratch... 2025-11-03T17:10:59.0614344Z (Worker_TP1 pid=3739) INFO 11-03 17:10:59 [cuda.py:404] Using Flash Attention backend on V1 engine. 2025-11-03T17:10:59.0820689Z (Worker_TP0 pid=3738) INFO 11-03 17:10:59 [cuda.py:404] Using Flash Attention backend on V1 engine. 2025-11-03T17:10:59.5704653Z (Worker_TP0 pid=3738) INFO 11-03 17:10:59 [gpu_model_runner.py:2902] Model loading took 1.1667 GiB and 0.166447 seconds 2025-11-03T17:10:59.5883940Z (Worker_TP1 pid=3739) INFO 11-03 17:10:59 [gpu_model_runner.py:2902] Model loading took 1.1667 GiB and 0.168157 seconds 2025-11-03T17:11:03.2726819Z (Worker_TP1 pid=3739) INFO 11-03 17:11:03 [backends.py:594] Using cache directory: /var/lib/jenkins/.cache/vllm/torch_compile_cache/cdd8585f72/rank_1_0/backbone for vLLM's torch.compile 2025-11-03T17:11:03.2731843Z (Worker_TP1 pid=3739) INFO 11-03 17:11:03 [backends.py:608] Dynamo bytecode transform time: 3.38 s 2025-11-03T17:11:03.3359950Z (Worker_TP0 pid=3738) INFO 11-03 17:11:03 [backends.py:594] Using cache directory: /var/lib/jenkins/.cache/vllm/torch_compile_cache/cdd8585f72/rank_0_0/backbone for vLLM's torch.compile 2025-11-03T17:11:03.3364437Z (Worker_TP0 pid=3738) INFO 11-03 17:11:03 [backends.py:608] Dynamo bytecode transform time: 3.46 s 2025-11-03T17:11:04.3260452Z (Worker_TP1 pid=3739) [rank1]:W1103 17:11:04.325000 3739 site-packages/torch/_inductor/utils.py:1665] [0/0] Not enough SMs to use max_autotune_gemm mode 2025-11-03T17:11:04.3782530Z (Worker_TP0 pid=3738) [rank0]:W1103 17:11:04.377000 3738 site-packages/torch/_inductor/utils.py:1665] [0/0] Not enough SMs to use max_autotune_gemm mode 2025-11-03T17:11:05.7843616Z (Worker_TP1 pid=3739) INFO 11-03 17:11:05 [backends.py:214] Cache the graph for dynamic shape for later use 2025-11-03T17:11:05.7908256Z (Worker_TP0 pid=3738) INFO 11-03 17:11:05 [backends.py:214] Cache the graph for dynamic shape for later use 2025-11-03T17:11:14.0386348Z (Worker_TP1 pid=3739) INFO 11-03 17:11:14 [backends.py:241] Compiling a graph for dynamic shape takes 10.58 s 2025-11-03T17:11:14.0510337Z (Worker_TP0 pid=3738) INFO 11-03 17:11:14 [backends.py:241] Compiling a graph for dynamic shape takes 10.53 s 2025-11-03T17:11:20.7862135Z (Worker_TP1 pid=3739) INFO 11-03 17:11:20 [monitor.py:33] torch.compile takes 13.95 s in total 2025-11-03T17:11:20.7862904Z (Worker_TP0 pid=3738) INFO 11-03 17:11:20 [monitor.py:33] torch.compile takes 13.99 s in total 2025-11-03T17:11:21.4994354Z (Worker_TP1 pid=3739) INFO 11-03 17:11:21 [gpu_worker.py:314] Available KV cache memory: 18.49 GiB 2025-11-03T17:11:21.5149735Z (Worker_TP0 pid=3738) INFO 11-03 17:11:21 [gpu_worker.py:314] Available KV cache memory: 18.49 GiB 2025-11-03T17:11:21.8080060Z (EngineCore_DP0 pid=3619) INFO 11-03 17:11:21 [kv_cache_utils.py:1199] GPU KV cache size: 1,211,648 tokens 2025-11-03T17:11:21.8080965Z (EngineCore_DP0 pid=3619) INFO 11-03 17:11:21 [kv_cache_utils.py:1204] Maximum concurrency for 2,048 tokens per request: 591.62x 2025-11-03T17:11:21.8081792Z (EngineCore_DP0 pid=3619) INFO 11-03 17:11:21 [kv_cache_utils.py:1199] GPU KV cache size: 1,211,648 tokens 2025-11-03T17:11:21.8082596Z (EngineCore_DP0 pid=3619) INFO 11-03 17:11:21 [kv_cache_utils.py:1204] Maximum concurrency for 2,048 tokens per request: 591.62x 2025-11-03T17:11:22.1026667Z (Worker_TP0 pid=3738) 2025-11-03T17:11:22.2305071Z Capturing CUDA graphs (mixed prefill-decode, PIECEWISE): 0% 0/5 [00:00, 'use_cudagraph': True, 'cudagraph_num_of_warmups': 1, 'cudagraph_capture_sizes': [16, 8, 4, 2], 'cudagraph_copy_inputs': False, 'full_cuda_graph': True, 'use_inductor_graph_partition': False, 'pass_config': {'enable_noop': True, 'enable_sequence_parallelism': True}, 'max_capture_size': 16, 'local_cache_dir': None} 2025-11-03T17:11:44.7804656Z (EngineCore_DP0 pid=4593) 2025-11-03 17:11:44,779 INFO worker.py:1918 -- Started a local Ray instance. View the dashboard at http://127.0.0.1:8265  2025-11-03T17:11:45.6841667Z (EngineCore_DP0 pid=4593) INFO 11-03 17:11:45 [ray_utils.py:373] No current placement group found. Creating a new placement group. 2025-11-03T17:11:45.8613790Z (EngineCore_DP0 pid=4593) INFO 11-03 17:11:45 [ray_distributed_executor.py:179] use_ray_spmd_worker: True 2025-11-03T17:11:46.3987821Z (EngineCore_DP0 pid=4593) (pid=5487) /opt/conda/envs/py_3.12/lib/python3.12/site-packages/torch/cuda/__init__.py:65: FutureWarning: The pynvml package is deprecated. Please install nvidia-ml-py instead. If you did not install pynvml directly, please report this to the maintainers of the package that installed pynvml for you. 2025-11-03T17:11:46.3989152Z (EngineCore_DP0 pid=4593) (pid=5487) import pynvml # type: ignore[import] 2025-11-03T17:11:48.7489056Z (EngineCore_DP0 pid=4593) (pid=5487) INFO 11-03 17:11:48 [__init__.py:224] Automatically detected platform cuda. 2025-11-03T17:11:50.1929405Z (EngineCore_DP0 pid=4593) INFO 11-03 17:11:50 [ray_env.py:67] RAY_NON_CARRY_OVER_ENV_VARS from config: set() 2025-11-03T17:11:50.1930929Z (EngineCore_DP0 pid=4593) INFO 11-03 17:11:50 [ray_env.py:70] Copying the following environment variables to workers: ['HF_TOKEN', 'MAX_JOBS', 'VLLM_USE_RAY_SPMD_WORKER', 'VLLM_USE_RAY_COMPILED_DAG', 'LD_LIBRARY_PATH', 'HUGGING_FACE_HUB_TOKEN', 'VLLM_WORKER_MULTIPROC_METHOD', 'VLLM_USE_V1'] 2025-11-03T17:11:50.1932957Z (EngineCore_DP0 pid=4593) INFO 11-03 17:11:50 [ray_env.py:75] If certain env vars should NOT be copied, add them to /var/lib/jenkins/.config/vllm/ray_non_carry_over_env_vars.json file 2025-11-03T17:11:53.1360803Z (EngineCore_DP0 pid=4593) (RayWorkerWrapper pid=5487) WARNING 11-03 17:11:53 [worker_base.py:307] Missing `shared_worker_lock` argument from executor. This argument is needed for mm_processor_cache_type='shm'. 2025-11-03T17:11:54.0590636Z (EngineCore_DP0 pid=4593) (RayWorkerWrapper pid=5487) [Gloo] Rank 0 is connected to 1 peer ranks. Expected number of connected peer ranks is : 1 2025-11-03T17:11:54.0591767Z (EngineCore_DP0 pid=4593) (RayWorkerWrapper pid=5487) [Gloo] Rank 0 is connected to 1 peer ranks. Expected number of connected peer ranks is : 1 2025-11-03T17:11:54.0592952Z (EngineCore_DP0 pid=4593) (RayWorkerWrapper pid=5487) INFO 11-03 17:11:54 [__init__.py:1146] Found nccl from library libnccl.so.2 2025-11-03T17:11:54.0593750Z (EngineCore_DP0 pid=4593) (RayWorkerWrapper pid=5487) INFO 11-03 17:11:54 [pynccl.py:109] vLLM is using nccl==2.27.5 2025-11-03T17:11:54.0594453Z (EngineCore_DP0 pid=4593) (pid=5483) INFO 11-03 17:11:48 [__init__.py:224] Automatically detected platform cuda. 2025-11-03T17:11:54.3695036Z (EngineCore_DP0 pid=4593) (RayWorkerWrapper pid=5487) WARNING 11-03 17:11:54 [symm_mem.py:60] SymmMemCommunicator: Device capability 8.9 not supported, communicator is not available. 2025-11-03T17:11:54.3696385Z (EngineCore_DP0 pid=4593) (RayWorkerWrapper pid=5487) INFO 11-03 17:11:54 [custom_all_reduce.py:37] Skipping P2P check and trusting the driver's P2P report. 2025-11-03T17:11:54.3698046Z (EngineCore_DP0 pid=4593) (RayWorkerWrapper pid=5487) WARNING 11-03 17:11:54 [custom_all_reduce.py:162] Custom allreduce is disabled because your platform lacks GPU P2P capability or P2P test failed. To silence this warning, specify disable_custom_all_reduce=True explicitly. 2025-11-03T17:11:54.3700345Z (EngineCore_DP0 pid=4593) (RayWorkerWrapper pid=5487) INFO 11-03 17:11:54 [shm_broadcast.py:302] vLLM message queue communication handle: Handle(local_reader_ranks=[1], buffer_handle=(1, 4194304, 6, 'psm_460e14dc'), local_subscribe_addr='ipc:///tmp/86d445bb-5df8-44c4-b56b-4e1ecf7fecfc', remote_subscribe_addr=None, remote_addr_ipv6=False) 2025-11-03T17:11:54.4746986Z (EngineCore_DP0 pid=4593) (RayWorkerWrapper pid=5487) INFO 11-03 17:11:54 [parallel_state.py:1231] rank 0 in world size 2 is assigned as DP rank 0, PP rank 0, TP rank 0, EP rank 0 2025-11-03T17:11:54.6816913Z (EngineCore_DP0 pid=4593) (RayWorkerWrapper pid=5487) INFO 11-03 17:11:54 [topk_topp_sampler.py:59] Using FlashInfer for top-p & top-k sampling. 2025-11-03T17:11:54.7881247Z (EngineCore_DP0 pid=4593) (RayWorkerWrapper pid=5487) INFO 11-03 17:11:54 [gpu_model_runner.py:2840] Starting to load model meta-llama/Llama-3.2-1B-Instruct... 2025-11-03T17:11:54.9966399Z (EngineCore_DP0 pid=4593) (RayWorkerWrapper pid=5487) INFO 11-03 17:11:54 [gpu_model_runner.py:2870] Loading model from scratch... 2025-11-03T17:11:55.1025396Z (EngineCore_DP0 pid=4593) (RayWorkerWrapper pid=5487) INFO 11-03 17:11:55 [cuda.py:404] Using Flash Attention backend on V1 engine. 2025-11-03T17:11:55.6181594Z (EngineCore_DP0 pid=4593) (RayWorkerWrapper pid=5483) INFO 11-03 17:11:55 [gpu_model_runner.py:2902] Model loading took 1.1667 GiB and 0.166176 seconds 2025-11-03T17:11:58.9919476Z (EngineCore_DP0 pid=4593) (RayWorkerWrapper pid=5487) INFO 11-03 17:11:58 [backends.py:594] Using cache directory: /var/lib/jenkins/.cache/vllm/torch_compile_cache/618b881a70/rank_0_0/backbone for vLLM's torch.compile 2025-11-03T17:11:58.9921335Z (EngineCore_DP0 pid=4593) (RayWorkerWrapper pid=5487) INFO 11-03 17:11:58 [backends.py:608] Dynamo bytecode transform time: 3.08 s 2025-11-03T17:11:58.9922639Z (EngineCore_DP0 pid=4593) (RayWorkerWrapper pid=5483) WARNING 11-03 17:11:53 [worker_base.py:307] Missing `shared_worker_lock` argument from executor. This argument is needed for mm_processor_cache_type='shm'. 2025-11-03T17:12:00.1218032Z (EngineCore_DP0 pid=4593) (RayWorkerWrapper pid=5487) INFO 11-03 17:12:00 [backends.py:179] Directly load the compiled graph(s) for dynamic shape from the cache, took 0.833 s 2025-11-03T17:12:00.1220397Z (EngineCore_DP0 pid=4593) (RayWorkerWrapper pid=5483) [Gloo] Rank 1 is connected to 1 peer ranks. Expected number of connected peer ranks is : 1 [repeated 10x across cluster] (Ray deduplicates logs by default. Set RAY_DEDUP_LOGS=0 to disable log deduplication, or see https://docs.ray.io/en/master/ray-observability/user-guides/configure-logging.html#log-deduplication for more options.) 2025-11-03T17:12:00.1222806Z (EngineCore_DP0 pid=4593) (RayWorkerWrapper pid=5483) INFO 11-03 17:11:54 [__init__.py:1146] Found nccl from library libnccl.so.2 [repeated 3x across cluster] 2025-11-03T17:12:00.1224271Z (EngineCore_DP0 pid=4593) (RayWorkerWrapper pid=5483) INFO 11-03 17:11:54 [pynccl.py:109] vLLM is using nccl==2.27.5 [repeated 3x across cluster] 2025-11-03T17:12:00.1225634Z (EngineCore_DP0 pid=4593) (RayWorkerWrapper pid=5483) WARNING 11-03 17:11:54 [symm_mem.py:60] SymmMemCommunicator: Device capability 8.9 not supported, communicator is not available. 2025-11-03T17:12:00.1226732Z (EngineCore_DP0 pid=4593) (RayWorkerWrapper pid=5483) INFO 11-03 17:11:54 [custom_all_reduce.py:37] Skipping P2P check and trusting the driver's P2P report. 2025-11-03T17:12:00.1228039Z (EngineCore_DP0 pid=4593) (RayWorkerWrapper pid=5483) WARNING 11-03 17:11:54 [custom_all_reduce.py:162] Custom allreduce is disabled because your platform lacks GPU P2P capability or P2P test failed. To silence this warning, specify disable_custom_all_reduce=True explicitly. 2025-11-03T17:12:00.1229320Z (EngineCore_DP0 pid=4593) (RayWorkerWrapper pid=5483) INFO 11-03 17:11:54 [parallel_state.py:1231] rank 1 in world size 2 is assigned as DP rank 0, PP rank 0, TP rank 1, EP rank 1 2025-11-03T17:12:00.1230236Z (EngineCore_DP0 pid=4593) (RayWorkerWrapper pid=5483) INFO 11-03 17:11:54 [topk_topp_sampler.py:59] Using FlashInfer for top-p & top-k sampling. 2025-11-03T17:12:00.1231124Z (EngineCore_DP0 pid=4593) (RayWorkerWrapper pid=5483) INFO 11-03 17:11:54 [gpu_model_runner.py:2840] Starting to load model meta-llama/Llama-3.2-1B-Instruct... 2025-11-03T17:12:00.1231968Z (EngineCore_DP0 pid=4593) (RayWorkerWrapper pid=5483) INFO 11-03 17:11:54 [gpu_model_runner.py:2870] Loading model from scratch... 2025-11-03T17:12:00.1232739Z (EngineCore_DP0 pid=4593) (RayWorkerWrapper pid=5483) INFO 11-03 17:11:55 [cuda.py:404] Using Flash Attention backend on V1 engine. 2025-11-03T17:12:01.3479443Z (EngineCore_DP0 pid=4593) (RayWorkerWrapper pid=5487) INFO 11-03 17:12:01 [gpu_worker.py:314] Available KV cache memory: 18.54 GiB 2025-11-03T17:12:01.3480645Z (EngineCore_DP0 pid=4593) (RayWorkerWrapper pid=5487) INFO 11-03 17:11:55 [gpu_model_runner.py:2902] Model loading took 1.1667 GiB and 0.170900 seconds 2025-11-03T17:12:01.5589547Z (EngineCore_DP0 pid=4593) INFO 11-03 17:12:01 [kv_cache_utils.py:1199] GPU KV cache size: 1,215,360 tokens 2025-11-03T17:12:01.5590965Z (EngineCore_DP0 pid=4593) INFO 11-03 17:12:01 [kv_cache_utils.py:1204] Maximum concurrency for 2,048 tokens per request: 593.44x 2025-11-03T17:12:01.5592801Z (EngineCore_DP0 pid=4593) INFO 11-03 17:12:01 [kv_cache_utils.py:1199] GPU KV cache size: 1,215,360 tokens 2025-11-03T17:12:01.5594104Z (EngineCore_DP0 pid=4593) INFO 11-03 17:12:01 [kv_cache_utils.py:1204] Maximum concurrency for 2,048 tokens per request: 593.44x 2025-11-03T17:12:01.8600978Z (EngineCore_DP0 pid=4593) (RayWorkerWrapper pid=5487) 2025-11-03T17:12:01.8602202Z Capturing CUDA graphs (mixed prefill-decode, PIECEWISE): 0%| | 0/4 [00:00, 'use_cudagraph': True, 'cudagraph_num_of_warmups': 1, 'cudagraph_capture_sizes': [16, 8, 4, 2, 1], 'cudagraph_copy_inputs': False, 'full_cuda_graph': True, 'use_inductor_graph_partition': False, 'pass_config': {}, 'max_capture_size': 16, 'local_cache_dir': None} 2025-11-03T17:12:26.5229332Z (EngineCore_DP0 pid=9488) WARNING 11-03 17:12:26 [multiproc_executor.py:744] Reducing Torch parallelism from 24 threads to 1 to avoid unnecessary CPU contention. Set OMP_NUM_THREADS in the external environment to tune this value as needed. 2025-11-03T17:12:26.5233222Z (EngineCore_DP0 pid=9488) INFO 11-03 17:12:26 [shm_broadcast.py:302] vLLM message queue communication handle: Handle(local_reader_ranks=[0, 1], buffer_handle=(2, 16777216, 10, 'psm_7053370f'), local_subscribe_addr='ipc:///tmp/f1086894-303b-4a8a-814d-b5350cb5976d', remote_subscribe_addr=None, remote_addr_ipv6=False) 2025-11-03T17:12:27.0081704Z /opt/conda/envs/py_3.12/lib/python3.12/site-packages/torch/cuda/__init__.py:65: FutureWarning: The pynvml package is deprecated. Please install nvidia-ml-py instead. If you did not install pynvml directly, please report this to the maintainers of the package that installed pynvml for you. 2025-11-03T17:12:27.0082943Z import pynvml # type: ignore[import] 2025-11-03T17:12:27.0084091Z /opt/conda/envs/py_3.12/lib/python3.12/site-packages/torch/cuda/__init__.py:65: FutureWarning: The pynvml package is deprecated. Please install nvidia-ml-py instead. If you did not install pynvml directly, please report this to the maintainers of the package that installed pynvml for you. 2025-11-03T17:12:27.0085240Z import pynvml # type: ignore[import] 2025-11-03T17:12:29.3650667Z INFO 11-03 17:12:29 [__init__.py:224] Automatically detected platform cuda. 2025-11-03T17:12:29.3983093Z INFO 11-03 17:12:29 [__init__.py:224] Automatically detected platform cuda. 2025-11-03T17:12:34.2127612Z INFO 11-03 17:12:34 [shm_broadcast.py:302] vLLM message queue communication handle: Handle(local_reader_ranks=[0], buffer_handle=(1, 10485760, 10, 'psm_3c9f13bd'), local_subscribe_addr='ipc:///tmp/ced7cd99-db0b-4a7e-a800-8b42a06e3cc2', remote_subscribe_addr=None, remote_addr_ipv6=False) 2025-11-03T17:12:34.2520678Z INFO 11-03 17:12:34 [shm_broadcast.py:302] vLLM message queue communication handle: Handle(local_reader_ranks=[0], buffer_handle=(1, 10485760, 10, 'psm_c90bf092'), local_subscribe_addr='ipc:///tmp/55e68d9c-c245-4bb0-a336-a250fa92574b', remote_subscribe_addr=None, remote_addr_ipv6=False) 2025-11-03T17:12:34.4348309Z [Gloo] Rank 0 is connected to 1 peer ranks. Expected number of connected peer ranks is : 1 2025-11-03T17:12:34.4349019Z [Gloo] Rank 1 is connected to 1 peer ranks. Expected number of connected peer ranks is : 1 2025-11-03T17:12:34.4393395Z [Gloo] Rank 1 is connected to 1 peer ranks. Expected number of connected peer ranks is : 1 2025-11-03T17:12:34.4394014Z [Gloo] Rank 0 is connected to 1 peer ranks. Expected number of connected peer ranks is : 1 2025-11-03T17:12:34.4435691Z INFO 11-03 17:12:34 [__init__.py:1146] Found nccl from library libnccl.so.2 2025-11-03T17:12:34.4438107Z INFO 11-03 17:12:34 [pynccl.py:109] vLLM is using nccl==2.27.5 2025-11-03T17:12:34.4438545Z INFO 11-03 17:12:34 [__init__.py:1146] Found nccl from library libnccl.so.2 2025-11-03T17:12:34.4441576Z INFO 11-03 17:12:34 [pynccl.py:109] vLLM is using nccl==2.27.5 2025-11-03T17:12:34.6166199Z WARNING 11-03 17:12:34 [symm_mem.py:60] SymmMemCommunicator: Device capability 8.9 not supported, communicator is not available. 2025-11-03T17:12:34.6191979Z WARNING 11-03 17:12:34 [symm_mem.py:60] SymmMemCommunicator: Device capability 8.9 not supported, communicator is not available. 2025-11-03T17:12:34.6192774Z INFO 11-03 17:12:34 [custom_all_reduce.py:37] Skipping P2P check and trusting the driver's P2P report. 2025-11-03T17:12:34.6193660Z INFO 11-03 17:12:34 [custom_all_reduce.py:37] Skipping P2P check and trusting the driver's P2P report. 2025-11-03T17:12:34.6194697Z WARNING 11-03 17:12:34 [custom_all_reduce.py:162] Custom allreduce is disabled because your platform lacks GPU P2P capability or P2P test failed. To silence this warning, specify disable_custom_all_reduce=True explicitly. 2025-11-03T17:12:34.6198435Z WARNING 11-03 17:12:34 [custom_all_reduce.py:162] Custom allreduce is disabled because your platform lacks GPU P2P capability or P2P test failed. To silence this warning, specify disable_custom_all_reduce=True explicitly. 2025-11-03T17:12:34.6924827Z INFO 11-03 17:12:34 [shm_broadcast.py:302] vLLM message queue communication handle: Handle(local_reader_ranks=[1], buffer_handle=(1, 4194304, 6, 'psm_de1025de'), local_subscribe_addr='ipc:///tmp/20eee359-45c2-44e8-b622-0081a5619d5c', remote_subscribe_addr=None, remote_addr_ipv6=False) 2025-11-03T17:12:34.6951187Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:12:34.6951809Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:12:34.6959497Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:12:34.6967725Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:12:34.7822364Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:12:34.7831333Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:12:34.7844071Z [Gloo] Rank 0 is connected to 1 peer ranks. Expected number of connected peer ranks is : 1 2025-11-03T17:12:34.7844660Z [Gloo] Rank 1 is connected to 1 peer ranks. Expected number of connected peer ranks is : 1 2025-11-03T17:12:34.7847238Z INFO 11-03 17:12:34 [__init__.py:1146] Found nccl from library libnccl.so.2 2025-11-03T17:12:34.7848114Z INFO 11-03 17:12:34 [__init__.py:1146] Found nccl from library libnccl.so.2 2025-11-03T17:12:34.7848718Z INFO 11-03 17:12:34 [pynccl.py:109] vLLM is using nccl==2.27.5 2025-11-03T17:12:34.7849108Z INFO 11-03 17:12:34 [pynccl.py:109] vLLM is using nccl==2.27.5 2025-11-03T17:12:34.8183712Z INFO 11-03 17:12:34 [parallel_state.py:1231] rank 0 in world size 2 is assigned as DP rank 0, PP rank 0, TP rank 0, EP rank 0 2025-11-03T17:12:34.8184483Z INFO 11-03 17:12:34 [parallel_state.py:1231] rank 1 in world size 2 is assigned as DP rank 0, PP rank 0, TP rank 1, EP rank 1 2025-11-03T17:12:35.0491700Z INFO 11-03 17:12:35 [topk_topp_sampler.py:59] Using FlashInfer for top-p & top-k sampling. 2025-11-03T17:12:35.0661241Z INFO 11-03 17:12:35 [topk_topp_sampler.py:59] Using FlashInfer for top-p & top-k sampling. 2025-11-03T17:12:35.0812562Z (Worker_TP0 pid=9607) INFO 11-03 17:12:35 [gpu_model_runner.py:2840] Starting to load model meta-llama/Llama-3.2-1B-Instruct... 2025-11-03T17:12:35.0979483Z (Worker_TP1 pid=9608) INFO 11-03 17:12:35 [gpu_model_runner.py:2840] Starting to load model meta-llama/Llama-3.2-1B-Instruct... 2025-11-03T17:12:35.2885009Z (Worker_TP0 pid=9607) INFO 11-03 17:12:35 [gpu_model_runner.py:2870] Loading model from scratch... 2025-11-03T17:12:35.3234209Z (Worker_TP1 pid=9608) INFO 11-03 17:12:35 [gpu_model_runner.py:2870] Loading model from scratch... 2025-11-03T17:12:35.3774308Z (Worker_TP0 pid=9607) INFO 11-03 17:12:35 [cuda.py:404] Using Flash Attention backend on V1 engine. 2025-11-03T17:12:35.4119086Z (Worker_TP1 pid=9608) INFO 11-03 17:12:35 [cuda.py:404] Using Flash Attention backend on V1 engine. 2025-11-03T17:12:35.8609972Z (Worker_TP0 pid=9607) INFO 11-03 17:12:35 [gpu_model_runner.py:2902] Model loading took 1.1667 GiB and 0.164583 seconds 2025-11-03T17:12:35.9215364Z (Worker_TP1 pid=9608) INFO 11-03 17:12:35 [gpu_model_runner.py:2902] Model loading took 1.1667 GiB and 0.167559 seconds 2025-11-03T17:12:39.7232669Z (Worker_TP0 pid=9607) INFO 11-03 17:12:39 [backends.py:594] Using cache directory: /var/lib/jenkins/.cache/vllm/torch_compile_cache/cdd8585f72/rank_0_0/backbone for vLLM's torch.compile 2025-11-03T17:12:39.7237822Z (Worker_TP0 pid=9607) INFO 11-03 17:12:39 [backends.py:608] Dynamo bytecode transform time: 3.52 s 2025-11-03T17:12:39.7396216Z (Worker_TP1 pid=9608) INFO 11-03 17:12:39 [backends.py:594] Using cache directory: /var/lib/jenkins/.cache/vllm/torch_compile_cache/cdd8585f72/rank_1_0/backbone for vLLM's torch.compile 2025-11-03T17:12:39.7403376Z (Worker_TP1 pid=9608) INFO 11-03 17:12:39 [backends.py:608] Dynamo bytecode transform time: 3.54 s 2025-11-03T17:12:40.8256195Z (Worker_TP0 pid=9607) INFO 11-03 17:12:40 [backends.py:179] Directly load the compiled graph(s) for dynamic shape from the cache, took 0.910 s 2025-11-03T17:12:40.8578787Z (Worker_TP1 pid=9608) INFO 11-03 17:12:40 [backends.py:179] Directly load the compiled graph(s) for dynamic shape from the cache, took 0.925 s 2025-11-03T17:12:41.6252752Z (Worker_TP0 pid=9607) INFO 11-03 17:12:41 [monitor.py:33] torch.compile takes 3.52 s in total 2025-11-03T17:12:41.6253500Z (Worker_TP1 pid=9608) INFO 11-03 17:12:41 [monitor.py:33] torch.compile takes 3.54 s in total 2025-11-03T17:12:42.2645607Z (Worker_TP1 pid=9608) INFO 11-03 17:12:42 [gpu_worker.py:314] Available KV cache memory: 18.54 GiB 2025-11-03T17:12:42.2723552Z (Worker_TP0 pid=9607) INFO 11-03 17:12:42 [gpu_worker.py:314] Available KV cache memory: 18.54 GiB 2025-11-03T17:12:42.5102265Z (EngineCore_DP0 pid=9488) INFO 11-03 17:12:42 [kv_cache_utils.py:1199] GPU KV cache size: 1,214,720 tokens 2025-11-03T17:12:42.5103148Z (EngineCore_DP0 pid=9488) INFO 11-03 17:12:42 [kv_cache_utils.py:1204] Maximum concurrency for 2,048 tokens per request: 593.12x 2025-11-03T17:12:42.5104421Z (EngineCore_DP0 pid=9488) INFO 11-03 17:12:42 [kv_cache_utils.py:1199] GPU KV cache size: 1,214,720 tokens 2025-11-03T17:12:42.5105298Z (EngineCore_DP0 pid=9488) INFO 11-03 17:12:42 [kv_cache_utils.py:1204] Maximum concurrency for 2,048 tokens per request: 593.12x 2025-11-03T17:12:42.7566918Z (Worker_TP0 pid=9607) 2025-11-03T17:12:42.8790972Z Capturing CUDA graphs (mixed prefill-decode, PIECEWISE): 0% 0/5 [00:00, 'use_cudagraph': True, 'cudagraph_num_of_warmups': 1, 'cudagraph_capture_sizes': [16, 8, 4, 2], 'cudagraph_copy_inputs': False, 'full_cuda_graph': True, 'use_inductor_graph_partition': False, 'pass_config': {'enable_noop': True, 'enable_sequence_parallelism': True}, 'max_capture_size': 16, 'local_cache_dir': None} 2025-11-03T17:13:02.4597947Z (EngineCore_DP0 pid=9984) WARNING 11-03 17:13:02 [multiproc_executor.py:744] Reducing Torch parallelism from 24 threads to 1 to avoid unnecessary CPU contention. Set OMP_NUM_THREADS in the external environment to tune this value as needed. 2025-11-03T17:13:02.4603635Z (EngineCore_DP0 pid=9984) INFO 11-03 17:13:02 [shm_broadcast.py:302] vLLM message queue communication handle: Handle(local_reader_ranks=[0, 1], buffer_handle=(2, 16777216, 10, 'psm_b4d4c6e7'), local_subscribe_addr='ipc:///tmp/cbe0bdcb-cb92-475d-9ff1-64bd73910fc2', remote_subscribe_addr=None, remote_addr_ipv6=False) 2025-11-03T17:13:02.9429192Z /opt/conda/envs/py_3.12/lib/python3.12/site-packages/torch/cuda/__init__.py:65: FutureWarning: The pynvml package is deprecated. Please install nvidia-ml-py instead. If you did not install pynvml directly, please report this to the maintainers of the package that installed pynvml for you. 2025-11-03T17:13:02.9431295Z import pynvml # type: ignore[import] 2025-11-03T17:13:02.9685806Z /opt/conda/envs/py_3.12/lib/python3.12/site-packages/torch/cuda/__init__.py:65: FutureWarning: The pynvml package is deprecated. Please install nvidia-ml-py instead. If you did not install pynvml directly, please report this to the maintainers of the package that installed pynvml for you. 2025-11-03T17:13:02.9688081Z import pynvml # type: ignore[import] 2025-11-03T17:13:05.2984650Z INFO 11-03 17:13:05 [__init__.py:224] Automatically detected platform cuda. 2025-11-03T17:13:05.3655318Z INFO 11-03 17:13:05 [__init__.py:224] Automatically detected platform cuda. 2025-11-03T17:13:10.0729524Z INFO 11-03 17:13:10 [shm_broadcast.py:302] vLLM message queue communication handle: Handle(local_reader_ranks=[0], buffer_handle=(1, 10485760, 10, 'psm_481d3baf'), local_subscribe_addr='ipc:///tmp/298c14c3-85e0-47a2-88d2-d2a8fc180abe', remote_subscribe_addr=None, remote_addr_ipv6=False) 2025-11-03T17:13:10.1127488Z INFO 11-03 17:13:10 [shm_broadcast.py:302] vLLM message queue communication handle: Handle(local_reader_ranks=[0], buffer_handle=(1, 10485760, 10, 'psm_fd1ccaef'), local_subscribe_addr='ipc:///tmp/bcdec2cc-f288-413c-b70d-ae80560398b8', remote_subscribe_addr=None, remote_addr_ipv6=False) 2025-11-03T17:13:10.9855333Z [Gloo] Rank 1 is connected to 1 peer ranks. Expected number of connected peer ranks is : 1 2025-11-03T17:13:10.9856021Z [Gloo] Rank 0 is connected to 1 peer ranks. Expected number of connected peer ranks is : 1 2025-11-03T17:13:10.9912908Z [Gloo] Rank 1 is connected to 1 peer ranks. Expected number of connected peer ranks is : 1 2025-11-03T17:13:10.9913517Z [Gloo] Rank 0 is connected to 1 peer ranks. Expected number of connected peer ranks is : 1 2025-11-03T17:13:10.9953338Z INFO 11-03 17:13:10 [__init__.py:1146] Found nccl from library libnccl.so.2 2025-11-03T17:13:10.9955641Z INFO 11-03 17:13:10 [pynccl.py:109] vLLM is using nccl==2.27.5 2025-11-03T17:13:10.9956099Z INFO 11-03 17:13:10 [__init__.py:1146] Found nccl from library libnccl.so.2 2025-11-03T17:13:10.9958826Z INFO 11-03 17:13:10 [pynccl.py:109] vLLM is using nccl==2.27.5 2025-11-03T17:13:11.1667104Z WARNING 11-03 17:13:11 [symm_mem.py:60] SymmMemCommunicator: Device capability 8.9 not supported, communicator is not available. 2025-11-03T17:13:11.1668607Z WARNING 11-03 17:13:11 [symm_mem.py:60] SymmMemCommunicator: Device capability 8.9 not supported, communicator is not available. 2025-11-03T17:13:11.1694140Z INFO 11-03 17:13:11 [custom_all_reduce.py:37] Skipping P2P check and trusting the driver's P2P report. 2025-11-03T17:13:11.1695310Z INFO 11-03 17:13:11 [custom_all_reduce.py:37] Skipping P2P check and trusting the driver's P2P report. 2025-11-03T17:13:11.1697262Z WARNING 11-03 17:13:11 [custom_all_reduce.py:162] Custom allreduce is disabled because your platform lacks GPU P2P capability or P2P test failed. To silence this warning, specify disable_custom_all_reduce=True explicitly. 2025-11-03T17:13:11.1699326Z WARNING 11-03 17:13:11 [custom_all_reduce.py:162] Custom allreduce is disabled because your platform lacks GPU P2P capability or P2P test failed. To silence this warning, specify disable_custom_all_reduce=True explicitly. 2025-11-03T17:13:11.2726903Z INFO 11-03 17:13:11 [shm_broadcast.py:302] vLLM message queue communication handle: Handle(local_reader_ranks=[1], buffer_handle=(1, 4194304, 6, 'psm_04b07460'), local_subscribe_addr='ipc:///tmp/3ff3ad1a-b220-46a3-a249-d106ae81c277', remote_subscribe_addr=None, remote_addr_ipv6=False) 2025-11-03T17:13:11.2749321Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:13:11.2750477Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:13:11.2757464Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:13:11.2764387Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:13:11.3122055Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:13:11.3131230Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:13:11.3141416Z [Gloo] Rank 0 is connected to 1 peer ranks. Expected number of connected peer ranks is : 1 2025-11-03T17:13:11.3142033Z [Gloo] Rank 1 is connected to 1 peer ranks. Expected number of connected peer ranks is : 1 2025-11-03T17:13:11.3144391Z INFO 11-03 17:13:11 [__init__.py:1146] Found nccl from library libnccl.so.2 2025-11-03T17:13:11.3144914Z INFO 11-03 17:13:11 [__init__.py:1146] Found nccl from library libnccl.so.2 2025-11-03T17:13:11.3145462Z INFO 11-03 17:13:11 [pynccl.py:109] vLLM is using nccl==2.27.5 2025-11-03T17:13:11.3145846Z INFO 11-03 17:13:11 [pynccl.py:109] vLLM is using nccl==2.27.5 2025-11-03T17:13:11.3485682Z INFO 11-03 17:13:11 [parallel_state.py:1231] rank 1 in world size 2 is assigned as DP rank 0, PP rank 0, TP rank 1, EP rank 1 2025-11-03T17:13:11.3486466Z INFO 11-03 17:13:11 [parallel_state.py:1231] rank 0 in world size 2 is assigned as DP rank 0, PP rank 0, TP rank 0, EP rank 0 2025-11-03T17:13:11.5788867Z INFO 11-03 17:13:11 [topk_topp_sampler.py:59] Using FlashInfer for top-p & top-k sampling. 2025-11-03T17:13:11.5789486Z INFO 11-03 17:13:11 [topk_topp_sampler.py:59] Using FlashInfer for top-p & top-k sampling. 2025-11-03T17:13:11.6110968Z (Worker_TP1 pid=10104) INFO 11-03 17:13:11 [gpu_model_runner.py:2840] Starting to load model meta-llama/Llama-3.2-1B-Instruct... 2025-11-03T17:13:11.6128757Z (Worker_TP0 pid=10103) INFO 11-03 17:13:11 [gpu_model_runner.py:2840] Starting to load model meta-llama/Llama-3.2-1B-Instruct... 2025-11-03T17:13:11.8130397Z (Worker_TP1 pid=10104) INFO 11-03 17:13:11 [gpu_model_runner.py:2870] Loading model from scratch... 2025-11-03T17:13:11.8161462Z (Worker_TP0 pid=10103) INFO 11-03 17:13:11 [gpu_model_runner.py:2870] Loading model from scratch... 2025-11-03T17:13:11.9033658Z (Worker_TP1 pid=10104) INFO 11-03 17:13:11 [cuda.py:404] Using Flash Attention backend on V1 engine. 2025-11-03T17:13:11.9070896Z (Worker_TP0 pid=10103) INFO 11-03 17:13:11 [cuda.py:404] Using Flash Attention backend on V1 engine. 2025-11-03T17:13:12.3899952Z (Worker_TP1 pid=10104) INFO 11-03 17:13:12 [gpu_model_runner.py:2902] Model loading took 1.1667 GiB and 0.168582 seconds 2025-11-03T17:13:12.4016628Z (Worker_TP0 pid=10103) INFO 11-03 17:13:12 [gpu_model_runner.py:2902] Model loading took 1.1667 GiB and 0.169810 seconds 2025-11-03T17:13:15.8000101Z (Worker_TP1 pid=10104) INFO 11-03 17:13:15 [backends.py:594] Using cache directory: /var/lib/jenkins/.cache/vllm/torch_compile_cache/618b881a70/rank_1_0/backbone for vLLM's torch.compile 2025-11-03T17:13:15.8009160Z (Worker_TP1 pid=10104) INFO 11-03 17:13:15 [backends.py:608] Dynamo bytecode transform time: 3.13 s 2025-11-03T17:13:15.9053459Z (Worker_TP0 pid=10103) INFO 11-03 17:13:15 [backends.py:594] Using cache directory: /var/lib/jenkins/.cache/vllm/torch_compile_cache/618b881a70/rank_0_0/backbone for vLLM's torch.compile 2025-11-03T17:13:15.9063755Z (Worker_TP0 pid=10103) INFO 11-03 17:13:15 [backends.py:608] Dynamo bytecode transform time: 3.23 s 2025-11-03T17:13:16.8380235Z (Worker_TP1 pid=10104) INFO 11-03 17:13:16 [backends.py:179] Directly load the compiled graph(s) for dynamic shape from the cache, took 0.772 s 2025-11-03T17:13:17.0005424Z (Worker_TP0 pid=10103) INFO 11-03 17:13:17 [backends.py:179] Directly load the compiled graph(s) for dynamic shape from the cache, took 0.824 s 2025-11-03T17:13:18.4249052Z (Worker_TP1 pid=10104) INFO 11-03 17:13:18 [gpu_worker.py:314] Available KV cache memory: 18.54 GiB 2025-11-03T17:13:18.4510403Z (Worker_TP0 pid=10103) INFO 11-03 17:13:18 [gpu_worker.py:314] Available KV cache memory: 18.54 GiB 2025-11-03T17:13:18.6904955Z (EngineCore_DP0 pid=9984) INFO 11-03 17:13:18 [kv_cache_utils.py:1199] GPU KV cache size: 1,215,232 tokens 2025-11-03T17:13:18.6905840Z (EngineCore_DP0 pid=9984) INFO 11-03 17:13:18 [kv_cache_utils.py:1204] Maximum concurrency for 2,048 tokens per request: 593.38x 2025-11-03T17:13:18.6906664Z (EngineCore_DP0 pid=9984) INFO 11-03 17:13:18 [kv_cache_utils.py:1199] GPU KV cache size: 1,215,232 tokens 2025-11-03T17:13:18.6907483Z (EngineCore_DP0 pid=9984) INFO 11-03 17:13:18 [kv_cache_utils.py:1204] Maximum concurrency for 2,048 tokens per request: 593.38x 2025-11-03T17:13:18.9377781Z (Worker_TP0 pid=10103) 2025-11-03T17:13:19.9409223Z Capturing CUDA graphs (mixed prefill-decode, PIECEWISE): 0% 0/4 [00:00, 'use_cudagraph': True, 'cudagraph_num_of_warmups': 1, 'cudagraph_capture_sizes': [16, 8, 4, 2, 1], 'cudagraph_copy_inputs': False, 'full_cuda_graph': True, 'use_inductor_graph_partition': False, 'pass_config': {}, 'max_capture_size': 16, 'local_cache_dir': None} 2025-11-03T17:13:42.8831576Z (EngineCore_DP0 pid=10530) WARNING 11-03 17:13:42 [multiproc_executor.py:744] Reducing Torch parallelism from 24 threads to 1 to avoid unnecessary CPU contention. Set OMP_NUM_THREADS in the external environment to tune this value as needed. 2025-11-03T17:13:42.8835401Z (EngineCore_DP0 pid=10530) INFO 11-03 17:13:42 [shm_broadcast.py:302] vLLM message queue communication handle: Handle(local_reader_ranks=[0, 1], buffer_handle=(2, 16777216, 10, 'psm_2671f671'), local_subscribe_addr='ipc:///tmp/60339202-1588-4c69-8d60-2c8a10a68244', remote_subscribe_addr=None, remote_addr_ipv6=False) 2025-11-03T17:13:43.3698702Z /opt/conda/envs/py_3.12/lib/python3.12/site-packages/torch/cuda/__init__.py:65: FutureWarning: The pynvml package is deprecated. Please install nvidia-ml-py instead. If you did not install pynvml directly, please report this to the maintainers of the package that installed pynvml for you. 2025-11-03T17:13:43.3699937Z import pynvml # type: ignore[import] 2025-11-03T17:13:43.3780248Z /opt/conda/envs/py_3.12/lib/python3.12/site-packages/torch/cuda/__init__.py:65: FutureWarning: The pynvml package is deprecated. Please install nvidia-ml-py instead. If you did not install pynvml directly, please report this to the maintainers of the package that installed pynvml for you. 2025-11-03T17:13:43.3781833Z import pynvml # type: ignore[import] 2025-11-03T17:13:45.7299656Z INFO 11-03 17:13:45 [__init__.py:224] Automatically detected platform cuda. 2025-11-03T17:13:45.7531059Z INFO 11-03 17:13:45 [__init__.py:224] Automatically detected platform cuda. 2025-11-03T17:13:50.5130169Z INFO 11-03 17:13:50 [shm_broadcast.py:302] vLLM message queue communication handle: Handle(local_reader_ranks=[0], buffer_handle=(1, 10485760, 10, 'psm_51b10813'), local_subscribe_addr='ipc:///tmp/c52fb092-16e8-491e-85fb-b3add28dbaa3', remote_subscribe_addr=None, remote_addr_ipv6=False) 2025-11-03T17:13:50.5132040Z INFO 11-03 17:13:50 [shm_broadcast.py:302] vLLM message queue communication handle: Handle(local_reader_ranks=[0], buffer_handle=(1, 10485760, 10, 'psm_9d2a4609'), local_subscribe_addr='ipc:///tmp/6272d822-7d4e-483b-935f-db3d665976f7', remote_subscribe_addr=None, remote_addr_ipv6=False) 2025-11-03T17:13:51.3271952Z [Gloo] Rank 0 is connected to 1 peer ranks. Expected number of connected peer ranks is : 1 2025-11-03T17:13:51.3273645Z [Gloo] Rank 1 is connected to 1 peer ranks. Expected number of connected peer ranks is : 1 2025-11-03T17:13:51.3319850Z [Gloo] Rank 0 is connected to 1 peer ranks. Expected number of connected peer ranks is : 1 2025-11-03T17:13:51.3320991Z [Gloo] Rank 1 is connected to 1 peer ranks. Expected number of connected peer ranks is : 1 2025-11-03T17:13:51.3362576Z INFO 11-03 17:13:51 [__init__.py:1146] Found nccl from library libnccl.so.2 2025-11-03T17:13:51.3363499Z INFO 11-03 17:13:51 [__init__.py:1146] Found nccl from library libnccl.so.2 2025-11-03T17:13:51.3364428Z INFO 11-03 17:13:51 [pynccl.py:109] vLLM is using nccl==2.27.5 2025-11-03T17:13:51.3365575Z INFO 11-03 17:13:51 [pynccl.py:109] vLLM is using nccl==2.27.5 2025-11-03T17:13:51.5058189Z WARNING 11-03 17:13:51 [symm_mem.py:60] SymmMemCommunicator: Device capability 8.9 not supported, communicator is not available. 2025-11-03T17:13:51.5059059Z WARNING 11-03 17:13:51 [symm_mem.py:60] SymmMemCommunicator: Device capability 8.9 not supported, communicator is not available. 2025-11-03T17:13:51.5082348Z INFO 11-03 17:13:51 [custom_all_reduce.py:37] Skipping P2P check and trusting the driver's P2P report. 2025-11-03T17:13:51.5083109Z INFO 11-03 17:13:51 [custom_all_reduce.py:37] Skipping P2P check and trusting the driver's P2P report. 2025-11-03T17:13:51.5084227Z WARNING 11-03 17:13:51 [custom_all_reduce.py:162] Custom allreduce is disabled because your platform lacks GPU P2P capability or P2P test failed. To silence this warning, specify disable_custom_all_reduce=True explicitly. 2025-11-03T17:13:51.5085603Z WARNING 11-03 17:13:51 [custom_all_reduce.py:162] Custom allreduce is disabled because your platform lacks GPU P2P capability or P2P test failed. To silence this warning, specify disable_custom_all_reduce=True explicitly. 2025-11-03T17:13:51.5825742Z INFO 11-03 17:13:51 [shm_broadcast.py:302] vLLM message queue communication handle: Handle(local_reader_ranks=[1], buffer_handle=(1, 4194304, 6, 'psm_9bcccf73'), local_subscribe_addr='ipc:///tmp/6d488d75-5707-473b-a260-06b93c32224e', remote_subscribe_addr=None, remote_addr_ipv6=False) 2025-11-03T17:13:51.5849140Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:13:51.5849735Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:13:51.5860410Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:13:51.5868800Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:13:51.6221584Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:13:51.6233765Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:13:51.6247149Z [Gloo] Rank 1 is connected to 1 peer ranks. Expected number of connected peer ranks is : 1 2025-11-03T17:13:51.6248041Z [Gloo] Rank 0 is connected to 1 peer ranks. Expected number of connected peer ranks is : 1 2025-11-03T17:13:51.6249606Z INFO 11-03 17:13:51 [__init__.py:1146] Found nccl from library libnccl.so.2 2025-11-03T17:13:51.6250188Z INFO 11-03 17:13:51 [pynccl.py:109] vLLM is using nccl==2.27.5 2025-11-03T17:13:51.6250596Z INFO 11-03 17:13:51 [__init__.py:1146] Found nccl from library libnccl.so.2 2025-11-03T17:13:51.6251020Z INFO 11-03 17:13:51 [pynccl.py:109] vLLM is using nccl==2.27.5 2025-11-03T17:13:51.6590590Z INFO 11-03 17:13:51 [parallel_state.py:1231] rank 1 in world size 2 is assigned as DP rank 0, PP rank 0, TP rank 1, EP rank 1 2025-11-03T17:13:51.6591363Z INFO 11-03 17:13:51 [parallel_state.py:1231] rank 0 in world size 2 is assigned as DP rank 0, PP rank 0, TP rank 0, EP rank 0 2025-11-03T17:13:51.8909174Z INFO 11-03 17:13:51 [topk_topp_sampler.py:59] Using FlashInfer for top-p & top-k sampling. 2025-11-03T17:13:51.8941069Z INFO 11-03 17:13:51 [topk_topp_sampler.py:59] Using FlashInfer for top-p & top-k sampling. 2025-11-03T17:13:51.9226904Z (Worker_TP0 pid=10649) INFO 11-03 17:13:51 [gpu_model_runner.py:2840] Starting to load model meta-llama/Llama-3.2-1B-Instruct... 2025-11-03T17:13:51.9259924Z (Worker_TP1 pid=10650) INFO 11-03 17:13:51 [gpu_model_runner.py:2840] Starting to load model meta-llama/Llama-3.2-1B-Instruct... 2025-11-03T17:13:52.1258451Z (Worker_TP0 pid=10649) INFO 11-03 17:13:52 [gpu_model_runner.py:2870] Loading model from scratch... 2025-11-03T17:13:52.1333848Z (Worker_TP1 pid=10650) INFO 11-03 17:13:52 [gpu_model_runner.py:2870] Loading model from scratch... 2025-11-03T17:13:52.2167777Z (Worker_TP0 pid=10649) INFO 11-03 17:13:52 [cuda.py:404] Using Flash Attention backend on V1 engine. 2025-11-03T17:13:52.2234347Z (Worker_TP1 pid=10650) INFO 11-03 17:13:52 [cuda.py:404] Using Flash Attention backend on V1 engine. 2025-11-03T17:13:52.7015769Z (Worker_TP0 pid=10649) INFO 11-03 17:13:52 [gpu_model_runner.py:2902] Model loading took 1.1667 GiB and 0.167665 seconds 2025-11-03T17:13:52.7127159Z (Worker_TP1 pid=10650) INFO 11-03 17:13:52 [gpu_model_runner.py:2902] Model loading took 1.1667 GiB and 0.168162 seconds 2025-11-03T17:13:56.4733603Z (Worker_TP0 pid=10649) INFO 11-03 17:13:56 [backends.py:594] Using cache directory: /var/lib/jenkins/.cache/vllm/torch_compile_cache/cdd8585f72/rank_0_0/backbone for vLLM's torch.compile 2025-11-03T17:13:56.4738162Z (Worker_TP0 pid=10649) INFO 11-03 17:13:56 [backends.py:608] Dynamo bytecode transform time: 3.49 s 2025-11-03T17:13:56.6640993Z (Worker_TP1 pid=10650) INFO 11-03 17:13:56 [backends.py:594] Using cache directory: /var/lib/jenkins/.cache/vllm/torch_compile_cache/cdd8585f72/rank_1_0/backbone for vLLM's torch.compile 2025-11-03T17:13:56.6647166Z (Worker_TP1 pid=10650) INFO 11-03 17:13:56 [backends.py:608] Dynamo bytecode transform time: 3.68 s 2025-11-03T17:13:57.5431673Z (Worker_TP0 pid=10649) INFO 11-03 17:13:57 [backends.py:179] Directly load the compiled graph(s) for dynamic shape from the cache, took 0.881 s 2025-11-03T17:13:57.7741036Z (Worker_TP1 pid=10650) INFO 11-03 17:13:57 [backends.py:179] Directly load the compiled graph(s) for dynamic shape from the cache, took 0.916 s 2025-11-03T17:13:58.3038439Z (Worker_TP0 pid=10649) INFO 11-03 17:13:58 [monitor.py:33] torch.compile takes 3.49 s in total 2025-11-03T17:13:58.3039216Z (Worker_TP1 pid=10650) INFO 11-03 17:13:58 [monitor.py:33] torch.compile takes 3.68 s in total 2025-11-03T17:13:58.9486010Z (Worker_TP0 pid=10649) INFO 11-03 17:13:58 [gpu_worker.py:314] Available KV cache memory: 18.54 GiB 2025-11-03T17:13:58.9578750Z (Worker_TP1 pid=10650) INFO 11-03 17:13:58 [gpu_worker.py:314] Available KV cache memory: 18.54 GiB 2025-11-03T17:13:59.1966537Z (EngineCore_DP0 pid=10530) INFO 11-03 17:13:59 [kv_cache_utils.py:1199] GPU KV cache size: 1,214,720 tokens 2025-11-03T17:13:59.1967996Z (EngineCore_DP0 pid=10530) INFO 11-03 17:13:59 [kv_cache_utils.py:1204] Maximum concurrency for 2,048 tokens per request: 593.12x 2025-11-03T17:13:59.1968995Z (EngineCore_DP0 pid=10530) INFO 11-03 17:13:59 [kv_cache_utils.py:1199] GPU KV cache size: 1,214,720 tokens 2025-11-03T17:13:59.1969641Z (EngineCore_DP0 pid=10530) INFO 11-03 17:13:59 [kv_cache_utils.py:1204] Maximum concurrency for 2,048 tokens per request: 593.12x 2025-11-03T17:13:59.4397239Z (Worker_TP0 pid=10649) 2025-11-03T17:13:59.5657300Z Capturing CUDA graphs (mixed prefill-decode, PIECEWISE): 0% 0/5 [00:00, 'use_cudagraph': True, 'cudagraph_num_of_warmups': 1, 'cudagraph_capture_sizes': [16, 8, 4, 2], 'cudagraph_copy_inputs': False, 'full_cuda_graph': True, 'use_inductor_graph_partition': False, 'pass_config': {'enable_noop': True, 'enable_sequence_parallelism': True}, 'max_capture_size': 16, 'local_cache_dir': None} 2025-11-03T17:14:21.1406565Z (EngineCore_DP0 pid=11026) 2025-11-03 17:14:21,140 INFO worker.py:1918 -- Started a local Ray instance. View the dashboard at http://127.0.0.1:8265  2025-11-03T17:14:21.9891315Z (EngineCore_DP0 pid=11026) INFO 11-03 17:14:21 [ray_utils.py:373] No current placement group found. Creating a new placement group. 2025-11-03T17:14:22.1695636Z (EngineCore_DP0 pid=11026) INFO 11-03 17:14:22 [ray_distributed_executor.py:179] use_ray_spmd_worker: True 2025-11-03T17:14:22.7464587Z (EngineCore_DP0 pid=11026) (pid=11911) /opt/conda/envs/py_3.12/lib/python3.12/site-packages/torch/cuda/__init__.py:65: FutureWarning: The pynvml package is deprecated. Please install nvidia-ml-py instead. If you did not install pynvml directly, please report this to the maintainers of the package that installed pynvml for you. 2025-11-03T17:14:22.7466617Z (EngineCore_DP0 pid=11026) (pid=11911) import pynvml # type: ignore[import] 2025-11-03T17:14:24.9920023Z (EngineCore_DP0 pid=11026) (pid=11921) INFO 11-03 17:14:24 [__init__.py:224] Automatically detected platform cuda. 2025-11-03T17:14:26.4706574Z (EngineCore_DP0 pid=11026) INFO 11-03 17:14:26 [ray_env.py:67] RAY_NON_CARRY_OVER_ENV_VARS from config: set() 2025-11-03T17:14:26.4710401Z (EngineCore_DP0 pid=11026) INFO 11-03 17:14:26 [ray_env.py:70] Copying the following environment variables to workers: ['VLLM_USE_RAY_COMPILED_DAG', 'VLLM_USE_RAY_SPMD_WORKER', 'MAX_JOBS', 'VLLM_USE_V1', 'LD_LIBRARY_PATH', 'HUGGING_FACE_HUB_TOKEN', 'HF_TOKEN', 'VLLM_WORKER_MULTIPROC_METHOD'] 2025-11-03T17:14:26.4713823Z (EngineCore_DP0 pid=11026) INFO 11-03 17:14:26 [ray_env.py:75] If certain env vars should NOT be copied, add them to /var/lib/jenkins/.config/vllm/ray_non_carry_over_env_vars.json file 2025-11-03T17:14:29.3782197Z (EngineCore_DP0 pid=11026) (RayWorkerWrapper pid=11911) WARNING 11-03 17:14:29 [worker_base.py:307] Missing `shared_worker_lock` argument from executor. This argument is needed for mm_processor_cache_type='shm'. 2025-11-03T17:14:29.5861112Z (EngineCore_DP0 pid=11026) (RayWorkerWrapper pid=11911) [Gloo] Rank 0 is connected to 1 peer ranks. Expected number of connected peer ranks is : 1 2025-11-03T17:14:29.6906962Z (EngineCore_DP0 pid=11026) (RayWorkerWrapper pid=11911) INFO 11-03 17:14:29 [__init__.py:1146] Found nccl from library libnccl.so.2 2025-11-03T17:14:29.6908251Z (EngineCore_DP0 pid=11026) (RayWorkerWrapper pid=11911) INFO 11-03 17:14:29 [pynccl.py:109] vLLM is using nccl==2.27.5 2025-11-03T17:14:29.8975804Z (EngineCore_DP0 pid=11026) (RayWorkerWrapper pid=11911) WARNING 11-03 17:14:29 [symm_mem.py:60] SymmMemCommunicator: Device capability 8.9 not supported, communicator is not available. 2025-11-03T17:14:29.8977351Z (EngineCore_DP0 pid=11026) (RayWorkerWrapper pid=11911) INFO 11-03 17:14:29 [custom_all_reduce.py:37] Skipping P2P check and trusting the driver's P2P report. 2025-11-03T17:14:29.8978998Z (EngineCore_DP0 pid=11026) (RayWorkerWrapper pid=11911) WARNING 11-03 17:14:29 [custom_all_reduce.py:162] Custom allreduce is disabled because your platform lacks GPU P2P capability or P2P test failed. To silence this warning, specify disable_custom_all_reduce=True explicitly. 2025-11-03T17:14:29.8981345Z (EngineCore_DP0 pid=11026) (RayWorkerWrapper pid=11911) INFO 11-03 17:14:29 [shm_broadcast.py:302] vLLM message queue communication handle: Handle(local_reader_ranks=[1], buffer_handle=(1, 4194304, 6, 'psm_4b5d7346'), local_subscribe_addr='ipc:///tmp/5bc2510a-60bf-44d4-9854-fa2a208c9d0b', remote_subscribe_addr=None, remote_addr_ipv6=False) 2025-11-03T17:14:30.0020923Z (EngineCore_DP0 pid=11026) (RayWorkerWrapper pid=11911) INFO 11-03 17:14:29 [parallel_state.py:1231] rank 0 in world size 2 is assigned as DP rank 0, PP rank 0, TP rank 0, EP rank 0 2025-11-03T17:14:30.0023871Z (EngineCore_DP0 pid=11026) (pid=11911) INFO 11-03 17:14:24 [__init__.py:224] Automatically detected platform cuda. 2025-11-03T17:14:30.3104937Z (EngineCore_DP0 pid=11026) (RayWorkerWrapper pid=11911) INFO 11-03 17:14:30 [topk_topp_sampler.py:59] Using FlashInfer for top-p & top-k sampling. 2025-11-03T17:14:30.3106190Z (EngineCore_DP0 pid=11026) (RayWorkerWrapper pid=11911) INFO 11-03 17:14:30 [gpu_model_runner.py:2840] Starting to load model meta-llama/Llama-3.2-1B-Instruct... 2025-11-03T17:14:30.5169700Z (EngineCore_DP0 pid=11026) (RayWorkerWrapper pid=11911) INFO 11-03 17:14:30 [gpu_model_runner.py:2870] Loading model from scratch... 2025-11-03T17:14:30.6214067Z (EngineCore_DP0 pid=11026) (RayWorkerWrapper pid=11911) INFO 11-03 17:14:30 [cuda.py:404] Using Flash Attention backend on V1 engine. 2025-11-03T17:14:31.1337179Z (EngineCore_DP0 pid=11026) (RayWorkerWrapper pid=11911) INFO 11-03 17:14:31 [gpu_model_runner.py:2902] Model loading took 1.1667 GiB and 0.169135 seconds 2025-11-03T17:14:34.3936959Z (EngineCore_DP0 pid=11026) (RayWorkerWrapper pid=11911) INFO 11-03 17:14:34 [backends.py:594] Using cache directory: /var/lib/jenkins/.cache/vllm/torch_compile_cache/618b881a70/rank_0_0/backbone for vLLM's torch.compile 2025-11-03T17:14:34.3938408Z (EngineCore_DP0 pid=11026) (RayWorkerWrapper pid=11911) INFO 11-03 17:14:34 [backends.py:608] Dynamo bytecode transform time: 2.97 s 2025-11-03T17:14:34.3939670Z (EngineCore_DP0 pid=11026) (RayWorkerWrapper pid=11921) WARNING 11-03 17:14:29 [worker_base.py:307] Missing `shared_worker_lock` argument from executor. This argument is needed for mm_processor_cache_type='shm'. 2025-11-03T17:14:35.4184056Z (EngineCore_DP0 pid=11026) (RayWorkerWrapper pid=11911) INFO 11-03 17:14:35 [backends.py:179] Directly load the compiled graph(s) for dynamic shape from the cache, took 0.822 s 2025-11-03T17:14:35.4186433Z (EngineCore_DP0 pid=11026) (RayWorkerWrapper pid=11921) [Gloo] Rank 1 is connected to 1 peer ranks. Expected number of connected peer ranks is : 1 [repeated 11x across cluster] (Ray deduplicates logs by default. Set RAY_DEDUP_LOGS=0 to disable log deduplication, or see https://docs.ray.io/en/master/ray-observability/user-guides/configure-logging.html#log-deduplication for more options.) 2025-11-03T17:14:35.4188489Z (EngineCore_DP0 pid=11026) (RayWorkerWrapper pid=11921) INFO 11-03 17:14:29 [__init__.py:1146] Found nccl from library libnccl.so.2 [repeated 3x across cluster] 2025-11-03T17:14:35.4189704Z (EngineCore_DP0 pid=11026) (RayWorkerWrapper pid=11921) INFO 11-03 17:14:29 [pynccl.py:109] vLLM is using nccl==2.27.5 [repeated 3x across cluster] 2025-11-03T17:14:35.4190770Z (EngineCore_DP0 pid=11026) (RayWorkerWrapper pid=11921) WARNING 11-03 17:14:29 [symm_mem.py:60] SymmMemCommunicator: Device capability 8.9 not supported, communicator is not available. 2025-11-03T17:14:35.4191769Z (EngineCore_DP0 pid=11026) (RayWorkerWrapper pid=11921) INFO 11-03 17:14:29 [custom_all_reduce.py:37] Skipping P2P check and trusting the driver's P2P report. 2025-11-03T17:14:35.4193052Z (EngineCore_DP0 pid=11026) (RayWorkerWrapper pid=11921) WARNING 11-03 17:14:29 [custom_all_reduce.py:162] Custom allreduce is disabled because your platform lacks GPU P2P capability or P2P test failed. To silence this warning, specify disable_custom_all_reduce=True explicitly. 2025-11-03T17:14:35.4194668Z (EngineCore_DP0 pid=11026) (RayWorkerWrapper pid=11921) INFO 11-03 17:14:29 [parallel_state.py:1231] rank 1 in world size 2 is assigned as DP rank 0, PP rank 0, TP rank 1, EP rank 1 2025-11-03T17:14:35.4195831Z (EngineCore_DP0 pid=11026) (RayWorkerWrapper pid=11921) INFO 11-03 17:14:30 [topk_topp_sampler.py:59] Using FlashInfer for top-p & top-k sampling. 2025-11-03T17:14:35.4196782Z (EngineCore_DP0 pid=11026) (RayWorkerWrapper pid=11921) INFO 11-03 17:14:30 [gpu_model_runner.py:2840] Starting to load model meta-llama/Llama-3.2-1B-Instruct... 2025-11-03T17:14:35.5221538Z (EngineCore_DP0 pid=11026) (RayWorkerWrapper pid=11921) INFO 11-03 17:14:30 [gpu_model_runner.py:2870] Loading model from scratch... 2025-11-03T17:14:36.7485328Z (EngineCore_DP0 pid=11026) (RayWorkerWrapper pid=11911) INFO 11-03 17:14:36 [gpu_worker.py:314] Available KV cache memory: 18.54 GiB 2025-11-03T17:14:36.7486488Z (EngineCore_DP0 pid=11026) (RayWorkerWrapper pid=11921) INFO 11-03 17:14:30 [cuda.py:404] Using Flash Attention backend on V1 engine. 2025-11-03T17:14:36.7488070Z (EngineCore_DP0 pid=11026) (RayWorkerWrapper pid=11921) INFO 11-03 17:14:31 [gpu_model_runner.py:2902] Model loading took 1.1667 GiB and 0.169897 seconds 2025-11-03T17:14:36.9660943Z (EngineCore_DP0 pid=11026) INFO 11-03 17:14:36 [kv_cache_utils.py:1199] GPU KV cache size: 1,215,360 tokens 2025-11-03T17:14:36.9662677Z (EngineCore_DP0 pid=11026) INFO 11-03 17:14:36 [kv_cache_utils.py:1204] Maximum concurrency for 2,048 tokens per request: 593.44x 2025-11-03T17:14:36.9663522Z (EngineCore_DP0 pid=11026) INFO 11-03 17:14:36 [kv_cache_utils.py:1199] GPU KV cache size: 1,215,360 tokens 2025-11-03T17:14:36.9664369Z (EngineCore_DP0 pid=11026) INFO 11-03 17:14:36 [kv_cache_utils.py:1204] Maximum concurrency for 2,048 tokens per request: 593.44x 2025-11-03T17:14:37.2612277Z (EngineCore_DP0 pid=11026) (RayWorkerWrapper pid=11911) 2025-11-03T17:14:37.2612804Z Capturing CUDA graphs (mixed prefill-decode, PIECEWISE): 0%| | 0/4 [00:00, 'use_cudagraph': True, 'cudagraph_num_of_warmups': 1, 'cudagraph_capture_sizes': [16, 8, 4, 2, 1], 'cudagraph_copy_inputs': False, 'full_cuda_graph': True, 'use_inductor_graph_partition': False, 'pass_config': {}, 'max_capture_size': 16, 'local_cache_dir': None} 2025-11-03T17:14:58.3640507Z (EngineCore_DP0 pid=15793) WARNING 11-03 17:14:58 [multiproc_executor.py:744] Reducing Torch parallelism from 24 threads to 1 to avoid unnecessary CPU contention. Set OMP_NUM_THREADS in the external environment to tune this value as needed. 2025-11-03T17:14:58.3643688Z (EngineCore_DP0 pid=15793) INFO 11-03 17:14:58 [shm_broadcast.py:302] vLLM message queue communication handle: Handle(local_reader_ranks=[0, 1], buffer_handle=(2, 16777216, 10, 'psm_477d12b4'), local_subscribe_addr='ipc:///tmp/b1495878-13f0-4aff-be2a-56e385c84242', remote_subscribe_addr=None, remote_addr_ipv6=False) 2025-11-03T17:14:58.8498578Z /opt/conda/envs/py_3.12/lib/python3.12/site-packages/torch/cuda/__init__.py:65: FutureWarning: The pynvml package is deprecated. Please install nvidia-ml-py instead. If you did not install pynvml directly, please report this to the maintainers of the package that installed pynvml for you. 2025-11-03T17:14:58.8499822Z import pynvml # type: ignore[import] 2025-11-03T17:14:58.8500938Z /opt/conda/envs/py_3.12/lib/python3.12/site-packages/torch/cuda/__init__.py:65: FutureWarning: The pynvml package is deprecated. Please install nvidia-ml-py instead. If you did not install pynvml directly, please report this to the maintainers of the package that installed pynvml for you. 2025-11-03T17:14:58.8501863Z import pynvml # type: ignore[import] 2025-11-03T17:15:01.1974525Z INFO 11-03 17:15:01 [__init__.py:224] Automatically detected platform cuda. 2025-11-03T17:15:01.2206036Z INFO 11-03 17:15:01 [__init__.py:224] Automatically detected platform cuda. 2025-11-03T17:15:05.9728973Z INFO 11-03 17:15:05 [shm_broadcast.py:302] vLLM message queue communication handle: Handle(local_reader_ranks=[0], buffer_handle=(1, 10485760, 10, 'psm_0b11d9e5'), local_subscribe_addr='ipc:///tmp/5fdda6d0-f2ad-45ca-8ea9-ed86293052c8', remote_subscribe_addr=None, remote_addr_ipv6=False) 2025-11-03T17:15:05.9730873Z INFO 11-03 17:15:05 [shm_broadcast.py:302] vLLM message queue communication handle: Handle(local_reader_ranks=[0], buffer_handle=(1, 10485760, 10, 'psm_7558e62c'), local_subscribe_addr='ipc:///tmp/fcd0f8c4-8ac8-4af5-901e-d8ba0cf15d32', remote_subscribe_addr=None, remote_addr_ipv6=False) 2025-11-03T17:15:06.1687623Z [Gloo] Rank 1 is connected to 1 peer ranks. Expected number of connected peer ranks is : 1 2025-11-03T17:15:06.1688555Z [Gloo] Rank 0 is connected to 1 peer ranks. Expected number of connected peer ranks is : 1 2025-11-03T17:15:06.1736406Z [Gloo] Rank 1 is connected to 1 peer ranks. Expected number of connected peer ranks is : 1 2025-11-03T17:15:06.1737023Z [Gloo] Rank 0 is connected to 1 peer ranks. Expected number of connected peer ranks is : 1 2025-11-03T17:15:06.1778181Z INFO 11-03 17:15:06 [__init__.py:1146] Found nccl from library libnccl.so.2 2025-11-03T17:15:06.1778672Z INFO 11-03 17:15:06 [__init__.py:1146] Found nccl from library libnccl.so.2 2025-11-03T17:15:06.1779913Z INFO 11-03 17:15:06 [pynccl.py:109] vLLM is using nccl==2.27.5 2025-11-03T17:15:06.1780306Z INFO 11-03 17:15:06 [pynccl.py:109] vLLM is using nccl==2.27.5 2025-11-03T17:15:06.3480100Z WARNING 11-03 17:15:06 [symm_mem.py:60] SymmMemCommunicator: Device capability 8.9 not supported, communicator is not available. 2025-11-03T17:15:06.3480984Z WARNING 11-03 17:15:06 [symm_mem.py:60] SymmMemCommunicator: Device capability 8.9 not supported, communicator is not available. 2025-11-03T17:15:06.3504270Z INFO 11-03 17:15:06 [custom_all_reduce.py:37] Skipping P2P check and trusting the driver's P2P report. 2025-11-03T17:15:06.3504928Z INFO 11-03 17:15:06 [custom_all_reduce.py:37] Skipping P2P check and trusting the driver's P2P report. 2025-11-03T17:15:06.3506384Z WARNING 11-03 17:15:06 [custom_all_reduce.py:162] Custom allreduce is disabled because your platform lacks GPU P2P capability or P2P test failed. To silence this warning, specify disable_custom_all_reduce=True explicitly. 2025-11-03T17:15:06.3507699Z WARNING 11-03 17:15:06 [custom_all_reduce.py:162] Custom allreduce is disabled because your platform lacks GPU P2P capability or P2P test failed. To silence this warning, specify disable_custom_all_reduce=True explicitly. 2025-11-03T17:15:06.5026011Z INFO 11-03 17:15:06 [shm_broadcast.py:302] vLLM message queue communication handle: Handle(local_reader_ranks=[1], buffer_handle=(1, 4194304, 6, 'psm_46948c7f'), local_subscribe_addr='ipc:///tmp/fc28bf4b-c7c5-4b79-9f6d-0dad465e4712', remote_subscribe_addr=None, remote_addr_ipv6=False) 2025-11-03T17:15:06.5052271Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:15:06.5052887Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:15:06.5061247Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:15:06.5070999Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:15:06.6221694Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:15:06.6237213Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:15:06.6249212Z [Gloo] Rank 0 is connected to 1 peer ranks. Expected number of connected peer ranks is : 1 2025-11-03T17:15:06.6249803Z [Gloo] Rank 1 is connected to 1 peer ranks. Expected number of connected peer ranks is : 1 2025-11-03T17:15:06.6252118Z INFO 11-03 17:15:06 [__init__.py:1146] Found nccl from library libnccl.so.2 2025-11-03T17:15:06.6252591Z INFO 11-03 17:15:06 [__init__.py:1146] Found nccl from library libnccl.so.2 2025-11-03T17:15:06.6253009Z INFO 11-03 17:15:06 [pynccl.py:109] vLLM is using nccl==2.27.5 2025-11-03T17:15:06.6253416Z INFO 11-03 17:15:06 [pynccl.py:109] vLLM is using nccl==2.27.5 2025-11-03T17:15:06.6590079Z INFO 11-03 17:15:06 [parallel_state.py:1231] rank 0 in world size 2 is assigned as DP rank 0, PP rank 0, TP rank 0, EP rank 0 2025-11-03T17:15:06.6590837Z INFO 11-03 17:15:06 [parallel_state.py:1231] rank 1 in world size 2 is assigned as DP rank 0, PP rank 0, TP rank 1, EP rank 1 2025-11-03T17:15:06.8966046Z INFO 11-03 17:15:06 [topk_topp_sampler.py:59] Using FlashInfer for top-p & top-k sampling. 2025-11-03T17:15:06.8968165Z INFO 11-03 17:15:06 [topk_topp_sampler.py:59] Using FlashInfer for top-p & top-k sampling. 2025-11-03T17:15:06.9294352Z (Worker_TP1 pid=15913) INFO 11-03 17:15:06 [gpu_model_runner.py:2840] Starting to load model meta-llama/Llama-3.2-1B-Instruct... 2025-11-03T17:15:06.9298540Z (Worker_TP0 pid=15912) INFO 11-03 17:15:06 [gpu_model_runner.py:2840] Starting to load model meta-llama/Llama-3.2-1B-Instruct... 2025-11-03T17:15:07.1385601Z (Worker_TP1 pid=15913) INFO 11-03 17:15:07 [gpu_model_runner.py:2870] Loading model from scratch... 2025-11-03T17:15:07.1434805Z (Worker_TP0 pid=15912) INFO 11-03 17:15:07 [gpu_model_runner.py:2870] Loading model from scratch... 2025-11-03T17:15:07.2287760Z (Worker_TP1 pid=15913) INFO 11-03 17:15:07 [cuda.py:404] Using Flash Attention backend on V1 engine. 2025-11-03T17:15:07.2342370Z (Worker_TP0 pid=15912) INFO 11-03 17:15:07 [cuda.py:404] Using Flash Attention backend on V1 engine. 2025-11-03T17:15:07.7202220Z (Worker_TP0 pid=15912) INFO 11-03 17:15:07 [gpu_model_runner.py:2902] Model loading took 1.1667 GiB and 0.169222 seconds 2025-11-03T17:15:07.7247024Z (Worker_TP1 pid=15913) INFO 11-03 17:15:07 [gpu_model_runner.py:2902] Model loading took 1.1667 GiB and 0.168434 seconds 2025-11-03T17:15:11.3684853Z (Worker_TP1 pid=15913) INFO 11-03 17:15:11 [backends.py:594] Using cache directory: /var/lib/jenkins/.cache/vllm/torch_compile_cache/cdd8585f72/rank_1_0/backbone for vLLM's torch.compile 2025-11-03T17:15:11.3691065Z (Worker_TP1 pid=15913) INFO 11-03 17:15:11 [backends.py:608] Dynamo bytecode transform time: 3.38 s 2025-11-03T17:15:11.4721304Z (Worker_TP0 pid=15912) INFO 11-03 17:15:11 [backends.py:594] Using cache directory: /var/lib/jenkins/.cache/vllm/torch_compile_cache/cdd8585f72/rank_0_0/backbone for vLLM's torch.compile 2025-11-03T17:15:11.4727840Z (Worker_TP0 pid=15912) INFO 11-03 17:15:11 [backends.py:608] Dynamo bytecode transform time: 3.49 s 2025-11-03T17:15:12.4308300Z (Worker_TP1 pid=15913) INFO 11-03 17:15:12 [backends.py:179] Directly load the compiled graph(s) for dynamic shape from the cache, took 0.871 s 2025-11-03T17:15:12.5343196Z (Worker_TP0 pid=15912) INFO 11-03 17:15:12 [backends.py:179] Directly load the compiled graph(s) for dynamic shape from the cache, took 0.874 s 2025-11-03T17:15:13.0697298Z (Worker_TP1 pid=15913) INFO 11-03 17:15:13 [monitor.py:33] torch.compile takes 3.38 s in total 2025-11-03T17:15:13.0698110Z (Worker_TP0 pid=15912) INFO 11-03 17:15:13 [monitor.py:33] torch.compile takes 3.49 s in total 2025-11-03T17:15:13.7072984Z (Worker_TP1 pid=15913) INFO 11-03 17:15:13 [gpu_worker.py:314] Available KV cache memory: 18.54 GiB 2025-11-03T17:15:13.7180239Z (Worker_TP0 pid=15912) INFO 11-03 17:15:13 [gpu_worker.py:314] Available KV cache memory: 18.54 GiB 2025-11-03T17:15:13.9553476Z (EngineCore_DP0 pid=15793) INFO 11-03 17:15:13 [kv_cache_utils.py:1199] GPU KV cache size: 1,214,720 tokens 2025-11-03T17:15:13.9554242Z (EngineCore_DP0 pid=15793) INFO 11-03 17:15:13 [kv_cache_utils.py:1204] Maximum concurrency for 2,048 tokens per request: 593.12x 2025-11-03T17:15:13.9554915Z (EngineCore_DP0 pid=15793) INFO 11-03 17:15:13 [kv_cache_utils.py:1199] GPU KV cache size: 1,214,720 tokens 2025-11-03T17:15:13.9555613Z (EngineCore_DP0 pid=15793) INFO 11-03 17:15:13 [kv_cache_utils.py:1204] Maximum concurrency for 2,048 tokens per request: 593.12x 2025-11-03T17:15:14.2013746Z (Worker_TP0 pid=15912) 2025-11-03T17:15:14.3234866Z Capturing CUDA graphs (mixed prefill-decode, PIECEWISE): 0% 0/5 [00:00, 'use_cudagraph': True, 'cudagraph_num_of_warmups': 1, 'cudagraph_capture_sizes': [16, 8, 4, 2], 'cudagraph_copy_inputs': False, 'full_cuda_graph': True, 'use_inductor_graph_partition': False, 'pass_config': {'enable_noop': True, 'enable_sequence_parallelism': True}, 'max_capture_size': 16, 'local_cache_dir': None} 2025-11-03T17:15:37.0118375Z (EngineCore_DP0 pid=16289) WARNING 11-03 17:15:37 [multiproc_executor.py:744] Reducing Torch parallelism from 24 threads to 1 to avoid unnecessary CPU contention. Set OMP_NUM_THREADS in the external environment to tune this value as needed. 2025-11-03T17:15:37.0124180Z (EngineCore_DP0 pid=16289) INFO 11-03 17:15:37 [shm_broadcast.py:302] vLLM message queue communication handle: Handle(local_reader_ranks=[0, 1, 2, 3], buffer_handle=(4, 16777216, 10, 'psm_511ef549'), local_subscribe_addr='ipc:///tmp/4f6ce7c1-6416-4b7c-82d5-d077ea9fe799', remote_subscribe_addr=None, remote_addr_ipv6=False) 2025-11-03T17:15:37.5222175Z /opt/conda/envs/py_3.12/lib/python3.12/site-packages/torch/cuda/__init__.py:65: FutureWarning: The pynvml package is deprecated. Please install nvidia-ml-py instead. If you did not install pynvml directly, please report this to the maintainers of the package that installed pynvml for you. 2025-11-03T17:15:37.5223453Z import pynvml # type: ignore[import] 2025-11-03T17:15:37.5224852Z /opt/conda/envs/py_3.12/lib/python3.12/site-packages/torch/cuda/__init__.py:65: FutureWarning: The pynvml package is deprecated. Please install nvidia-ml-py instead. If you did not install pynvml directly, please report this to the maintainers of the package that installed pynvml for you. 2025-11-03T17:15:37.5225761Z import pynvml # type: ignore[import] 2025-11-03T17:15:37.5278359Z /opt/conda/envs/py_3.12/lib/python3.12/site-packages/torch/cuda/__init__.py:65: FutureWarning: The pynvml package is deprecated. Please install nvidia-ml-py instead. If you did not install pynvml directly, please report this to the maintainers of the package that installed pynvml for you. 2025-11-03T17:15:37.5279589Z import pynvml # type: ignore[import] 2025-11-03T17:15:37.5873693Z /opt/conda/envs/py_3.12/lib/python3.12/site-packages/torch/cuda/__init__.py:65: FutureWarning: The pynvml package is deprecated. Please install nvidia-ml-py instead. If you did not install pynvml directly, please report this to the maintainers of the package that installed pynvml for you. 2025-11-03T17:15:37.5874956Z import pynvml # type: ignore[import] 2025-11-03T17:15:39.8962110Z INFO 11-03 17:15:39 [__init__.py:224] Automatically detected platform cuda. 2025-11-03T17:15:39.9662158Z INFO 11-03 17:15:39 [__init__.py:224] Automatically detected platform cuda. 2025-11-03T17:15:39.9786342Z INFO 11-03 17:15:39 [__init__.py:224] Automatically detected platform cuda. 2025-11-03T17:15:40.0002591Z INFO 11-03 17:15:39 [__init__.py:224] Automatically detected platform cuda. 2025-11-03T17:15:44.7881791Z INFO 11-03 17:15:44 [shm_broadcast.py:302] vLLM message queue communication handle: Handle(local_reader_ranks=[0], buffer_handle=(1, 10485760, 10, 'psm_08188721'), local_subscribe_addr='ipc:///tmp/aa7e7376-d05e-46b3-b745-de823ec44d3e', remote_subscribe_addr=None, remote_addr_ipv6=False) 2025-11-03T17:15:44.8729169Z INFO 11-03 17:15:44 [shm_broadcast.py:302] vLLM message queue communication handle: Handle(local_reader_ranks=[0], buffer_handle=(1, 10485760, 10, 'psm_2c507e53'), local_subscribe_addr='ipc:///tmp/1f3d81a3-9068-4087-89db-5786202b3227', remote_subscribe_addr=None, remote_addr_ipv6=False) 2025-11-03T17:15:44.9125855Z INFO 11-03 17:15:44 [shm_broadcast.py:302] vLLM message queue communication handle: Handle(local_reader_ranks=[0], buffer_handle=(1, 10485760, 10, 'psm_ad6f456f'), local_subscribe_addr='ipc:///tmp/ca6291ed-9d51-48e9-a4a5-3abc796ea9a0', remote_subscribe_addr=None, remote_addr_ipv6=False) 2025-11-03T17:15:44.9527712Z INFO 11-03 17:15:44 [shm_broadcast.py:302] vLLM message queue communication handle: Handle(local_reader_ranks=[0], buffer_handle=(1, 10485760, 10, 'psm_805dc782'), local_subscribe_addr='ipc:///tmp/d462321a-efd4-4108-a4fa-0daffb8d4c2e', remote_subscribe_addr=None, remote_addr_ipv6=False) 2025-11-03T17:15:45.4921311Z [Gloo] Rank 1 is connected to 3 peer ranks. Expected number of connected peer ranks is : 3 2025-11-03T17:15:45.4922064Z [Gloo] Rank 0 is connected to 3 peer ranks. Expected number of connected peer ranks is : 3 2025-11-03T17:15:45.4922701Z [Gloo] Rank 2 is connected to 3 peer ranks. Expected number of connected peer ranks is : 3 2025-11-03T17:15:45.4923269Z [Gloo] Rank 3 is connected to 3 peer ranks. Expected number of connected peer ranks is : 3 2025-11-03T17:15:45.4977599Z [Gloo] Rank 0 is connected to 1 peer ranks. Expected number of connected peer ranks is : 1 2025-11-03T17:15:45.4978255Z [Gloo] Rank 1 is connected to 1 peer ranks. Expected number of connected peer ranks is : 1 2025-11-03T17:15:45.4979296Z [Gloo] Rank 1 is connected to 1 peer ranks. Expected number of connected peer ranks is : 1 2025-11-03T17:15:45.4979841Z [Gloo] Rank 0 is connected to 1 peer ranks. Expected number of connected peer ranks is : 1 2025-11-03T17:15:45.5022878Z INFO 11-03 17:15:45 [__init__.py:1146] Found nccl from library libnccl.so.2 2025-11-03T17:15:45.5023994Z INFO 11-03 17:15:45 [__init__.py:1146] Found nccl from library libnccl.so.2 2025-11-03T17:15:45.5025839Z INFO 11-03 17:15:45 [pynccl.py:109] vLLM is using nccl==2.27.5 2025-11-03T17:15:45.5026648Z INFO 11-03 17:15:45 [pynccl.py:109] vLLM is using nccl==2.27.5 2025-11-03T17:15:45.5034929Z INFO 11-03 17:15:45 [__init__.py:1146] Found nccl from library libnccl.so.2 2025-11-03T17:15:45.5037959Z INFO 11-03 17:15:45 [__init__.py:1146] Found nccl from library libnccl.so.2 2025-11-03T17:15:45.5038371Z INFO 11-03 17:15:45 [pynccl.py:109] vLLM is using nccl==2.27.5 2025-11-03T17:15:45.5041238Z INFO 11-03 17:15:45 [pynccl.py:109] vLLM is using nccl==2.27.5 2025-11-03T17:15:45.7138989Z WARNING 11-03 17:15:45 [symm_mem.py:60] SymmMemCommunicator: Device capability 8.9 not supported, communicator is not available. 2025-11-03T17:15:45.7139848Z WARNING 11-03 17:15:45 [symm_mem.py:60] SymmMemCommunicator: Device capability 8.9 not supported, communicator is not available. 2025-11-03T17:15:45.7160264Z INFO 11-03 17:15:45 [custom_all_reduce.py:37] Skipping P2P check and trusting the driver's P2P report. 2025-11-03T17:15:45.7160889Z INFO 11-03 17:15:45 [custom_all_reduce.py:37] Skipping P2P check and trusting the driver's P2P report. 2025-11-03T17:15:45.7162345Z WARNING 11-03 17:15:45 [custom_all_reduce.py:162] Custom allreduce is disabled because your platform lacks GPU P2P capability or P2P test failed. To silence this warning, specify disable_custom_all_reduce=True explicitly. 2025-11-03T17:15:45.7163745Z WARNING 11-03 17:15:45 [custom_all_reduce.py:162] Custom allreduce is disabled because your platform lacks GPU P2P capability or P2P test failed. To silence this warning, specify disable_custom_all_reduce=True explicitly. 2025-11-03T17:15:45.7186372Z INFO 11-03 17:15:45 [shm_broadcast.py:302] vLLM message queue communication handle: Handle(local_reader_ranks=[1], buffer_handle=(1, 4194304, 6, 'psm_10cf6c81'), local_subscribe_addr='ipc:///tmp/46c5a3b8-3dcf-4ac4-a023-1084f249df08', remote_subscribe_addr=None, remote_addr_ipv6=False) 2025-11-03T17:15:45.7886766Z WARNING 11-03 17:15:45 [symm_mem.py:60] SymmMemCommunicator: Device capability 8.9 not supported, communicator is not available. 2025-11-03T17:15:45.7887639Z WARNING 11-03 17:15:45 [symm_mem.py:60] SymmMemCommunicator: Device capability 8.9 not supported, communicator is not available. 2025-11-03T17:15:45.7915035Z INFO 11-03 17:15:45 [custom_all_reduce.py:37] Skipping P2P check and trusting the driver's P2P report. 2025-11-03T17:15:45.7916119Z WARNING 11-03 17:15:45 [custom_all_reduce.py:162] Custom allreduce is disabled because your platform lacks GPU P2P capability or P2P test failed. To silence this warning, specify disable_custom_all_reduce=True explicitly. 2025-11-03T17:15:45.7917115Z INFO 11-03 17:15:45 [custom_all_reduce.py:37] Skipping P2P check and trusting the driver's P2P report. 2025-11-03T17:15:45.7918114Z WARNING 11-03 17:15:45 [custom_all_reduce.py:162] Custom allreduce is disabled because your platform lacks GPU P2P capability or P2P test failed. To silence this warning, specify disable_custom_all_reduce=True explicitly. 2025-11-03T17:15:45.7935083Z INFO 11-03 17:15:45 [shm_broadcast.py:302] vLLM message queue communication handle: Handle(local_reader_ranks=[1], buffer_handle=(1, 4194304, 6, 'psm_4d4475a4'), local_subscribe_addr='ipc:///tmp/52738470-1e5d-4666-8e37-c7b6c571ed28', remote_subscribe_addr=None, remote_addr_ipv6=False) 2025-11-03T17:15:45.7960670Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:15:45.8121266Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:15:45.8122652Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:15:45.8140096Z [Gloo] Rank 0 is connected to 1 peer ranks. Expected number of connected peer ranks is : 1 2025-11-03T17:15:45.8140765Z [Gloo] Rank 1 is connected to 1 peer ranks. Expected number of connected peer ranks is : 1 2025-11-03T17:15:45.8143665Z INFO 11-03 17:15:45 [__init__.py:1146] Found nccl from library libnccl.so.2 2025-11-03T17:15:45.8144126Z INFO 11-03 17:15:45 [__init__.py:1146] Found nccl from library libnccl.so.2 2025-11-03T17:15:45.8144823Z INFO 11-03 17:15:45 [pynccl.py:109] vLLM is using nccl==2.27.5 2025-11-03T17:15:45.8145206Z INFO 11-03 17:15:45 [pynccl.py:109] vLLM is using nccl==2.27.5 2025-11-03T17:15:45.8498112Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:15:45.8498733Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:15:45.8721050Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:15:45.8738562Z [Gloo] Rank 0 is connected to 1 peer ranks. Expected number of connected peer ranks is : 1 2025-11-03T17:15:45.8739720Z [Gloo] Rank 1 is connected to 1 peer ranks. Expected number of connected peer ranks is : 1 2025-11-03T17:15:45.8741134Z INFO 11-03 17:15:45 [__init__.py:1146] Found nccl from library libnccl.so.2 2025-11-03T17:15:45.8742038Z INFO 11-03 17:15:45 [pynccl.py:109] vLLM is using nccl==2.27.5 2025-11-03T17:15:45.8742548Z INFO 11-03 17:15:45 [__init__.py:1146] Found nccl from library libnccl.so.2 2025-11-03T17:15:45.8742974Z INFO 11-03 17:15:45 [pynccl.py:109] vLLM is using nccl==2.27.5 2025-11-03T17:15:45.9084194Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:15:45.9084832Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:15:45.9095063Z [Gloo] Rank 0 is connected to 1 peer ranks. Expected number of connected peer ranks is : 1 2025-11-03T17:15:45.9095675Z [Gloo] Rank 1 is connected to 1 peer ranks. Expected number of connected peer ranks is : 1 2025-11-03T17:15:45.9096252Z [Gloo] Rank 1 is connected to 1 peer ranks. Expected number of connected peer ranks is : 1 2025-11-03T17:15:45.9096799Z [Gloo] Rank 0 is connected to 1 peer ranks. Expected number of connected peer ranks is : 1 2025-11-03T17:15:45.9097708Z INFO 11-03 17:15:45 [__init__.py:1146] Found nccl from library libnccl.so.2 2025-11-03T17:15:45.9098151Z INFO 11-03 17:15:45 [pynccl.py:109] vLLM is using nccl==2.27.5 2025-11-03T17:15:45.9098889Z INFO 11-03 17:15:45 [__init__.py:1146] Found nccl from library libnccl.so.2 2025-11-03T17:15:45.9099834Z INFO 11-03 17:15:45 [__init__.py:1146] Found nccl from library libnccl.so.2 2025-11-03T17:15:45.9100325Z INFO 11-03 17:15:45 [pynccl.py:109] vLLM is using nccl==2.27.5 2025-11-03T17:15:45.9100792Z INFO 11-03 17:15:45 [__init__.py:1146] Found nccl from library libnccl.so.2 2025-11-03T17:15:45.9101205Z INFO 11-03 17:15:45 [pynccl.py:109] vLLM is using nccl==2.27.5 2025-11-03T17:15:45.9101550Z INFO 11-03 17:15:45 [pynccl.py:109] vLLM is using nccl==2.27.5 2025-11-03T17:15:45.9507553Z INFO 11-03 17:15:45 [parallel_state.py:1231] rank 1 in world size 4 is assigned as DP rank 0, PP rank 0, TP rank 1, EP rank 1 2025-11-03T17:15:45.9508282Z INFO 11-03 17:15:45 [parallel_state.py:1231] rank 0 in world size 4 is assigned as DP rank 0, PP rank 0, TP rank 0, EP rank 0 2025-11-03T17:15:45.9541015Z INFO 11-03 17:15:45 [parallel_state.py:1231] rank 3 in world size 4 is assigned as DP rank 0, PP rank 1, TP rank 1, EP rank 1 2025-11-03T17:15:45.9541799Z INFO 11-03 17:15:45 [parallel_state.py:1231] rank 2 in world size 4 is assigned as DP rank 0, PP rank 1, TP rank 0, EP rank 0 2025-11-03T17:15:46.2006977Z INFO 11-03 17:15:46 [topk_topp_sampler.py:59] Using FlashInfer for top-p & top-k sampling. 2025-11-03T17:15:46.2105735Z INFO 11-03 17:15:46 [topk_topp_sampler.py:59] Using FlashInfer for top-p & top-k sampling. 2025-11-03T17:15:46.2148191Z INFO 11-03 17:15:46 [topk_topp_sampler.py:59] Using FlashInfer for top-p & top-k sampling. 2025-11-03T17:15:46.2189504Z INFO 11-03 17:15:46 [topk_topp_sampler.py:59] Using FlashInfer for top-p & top-k sampling. 2025-11-03T17:15:46.2310602Z (Worker_PP0_TP1 pid=16409) INFO 11-03 17:15:46 [gpu_model_runner.py:2840] Starting to load model meta-llama/Llama-3.2-1B-Instruct... 2025-11-03T17:15:46.2430233Z (Worker_PP0_TP0 pid=16408) INFO 11-03 17:15:46 [gpu_model_runner.py:2840] Starting to load model meta-llama/Llama-3.2-1B-Instruct... 2025-11-03T17:15:46.2447481Z (Worker_PP1_TP0 pid=16410) INFO 11-03 17:15:46 [gpu_model_runner.py:2840] Starting to load model meta-llama/Llama-3.2-1B-Instruct... 2025-11-03T17:15:46.2498410Z (Worker_PP1_TP1 pid=16411) INFO 11-03 17:15:46 [gpu_model_runner.py:2840] Starting to load model meta-llama/Llama-3.2-1B-Instruct... 2025-11-03T17:15:46.4416995Z (Worker_PP0_TP1 pid=16409) INFO 11-03 17:15:46 [gpu_model_runner.py:2870] Loading model from scratch... 2025-11-03T17:15:46.4571198Z (Worker_PP1_TP0 pid=16410) INFO 11-03 17:15:46 [gpu_model_runner.py:2870] Loading model from scratch... 2025-11-03T17:15:46.4804483Z (Worker_PP0_TP0 pid=16408) INFO 11-03 17:15:46 [gpu_model_runner.py:2870] Loading model from scratch... 2025-11-03T17:15:46.4913294Z (Worker_PP1_TP1 pid=16411) INFO 11-03 17:15:46 [gpu_model_runner.py:2870] Loading model from scratch... 2025-11-03T17:15:46.5316249Z (Worker_PP0_TP1 pid=16409) INFO 11-03 17:15:46 [cuda.py:404] Using Flash Attention backend on V1 engine. 2025-11-03T17:15:46.5480786Z (Worker_PP1_TP0 pid=16410) INFO 11-03 17:15:46 [cuda.py:404] Using Flash Attention backend on V1 engine. 2025-11-03T17:15:46.5716117Z (Worker_PP0_TP0 pid=16408) INFO 11-03 17:15:46 [cuda.py:404] Using Flash Attention backend on V1 engine. 2025-11-03T17:15:46.5810416Z (Worker_PP1_TP1 pid=16411) INFO 11-03 17:15:46 [cuda.py:404] Using Flash Attention backend on V1 engine. 2025-11-03T17:15:47.0174434Z (Worker_PP0_TP1 pid=16409) INFO 11-03 17:15:47 [gpu_model_runner.py:2902] Model loading took 0.7135 GiB and 0.153136 seconds 2025-11-03T17:15:47.0270494Z (Worker_PP1_TP0 pid=16410) INFO 11-03 17:15:47 [gpu_model_runner.py:2902] Model loading took 0.7135 GiB and 0.155832 seconds 2025-11-03T17:15:47.1113873Z (Worker_PP0_TP0 pid=16408) INFO 11-03 17:15:47 [gpu_model_runner.py:2902] Model loading took 0.7135 GiB and 0.156916 seconds 2025-11-03T17:15:47.1210016Z (Worker_PP1_TP1 pid=16411) INFO 11-03 17:15:47 [gpu_model_runner.py:2902] Model loading took 0.7135 GiB and 0.154246 seconds 2025-11-03T17:15:49.3856028Z (Worker_PP1_TP0 pid=16410) INFO 11-03 17:15:49 [backends.py:594] Using cache directory: /var/lib/jenkins/.cache/vllm/torch_compile_cache/ba6452aee0/rank_2_0/backbone for vLLM's torch.compile 2025-11-03T17:15:49.3858649Z (Worker_PP1_TP0 pid=16410) INFO 11-03 17:15:49 [backends.py:608] Dynamo bytecode transform time: 2.06 s 2025-11-03T17:15:49.5221603Z (Worker_PP1_TP1 pid=16411) INFO 11-03 17:15:49 [backends.py:594] Using cache directory: /var/lib/jenkins/.cache/vllm/torch_compile_cache/ba6452aee0/rank_3_0/backbone for vLLM's torch.compile 2025-11-03T17:15:49.5225562Z (Worker_PP1_TP1 pid=16411) INFO 11-03 17:15:49 [backends.py:608] Dynamo bytecode transform time: 2.16 s 2025-11-03T17:15:49.5367800Z (Worker_PP0_TP1 pid=16409) INFO 11-03 17:15:49 [backends.py:594] Using cache directory: /var/lib/jenkins/.cache/vllm/torch_compile_cache/3ff2b0020c/rank_1_0/backbone for vLLM's torch.compile 2025-11-03T17:15:49.5372705Z (Worker_PP0_TP1 pid=16409) INFO 11-03 17:15:49 [backends.py:608] Dynamo bytecode transform time: 2.21 s 2025-11-03T17:15:49.5904355Z (Worker_PP0_TP0 pid=16408) INFO 11-03 17:15:49 [backends.py:594] Using cache directory: /var/lib/jenkins/.cache/vllm/torch_compile_cache/3ff2b0020c/rank_0_0/backbone for vLLM's torch.compile 2025-11-03T17:15:49.5908707Z (Worker_PP0_TP0 pid=16408) INFO 11-03 17:15:49 [backends.py:608] Dynamo bytecode transform time: 2.23 s 2025-11-03T17:15:50.3395554Z (Worker_PP1_TP0 pid=16410) [rank2]:W1103 17:15:50.338000 16410 site-packages/torch/_inductor/utils.py:1665] [0/0] Not enough SMs to use max_autotune_gemm mode 2025-11-03T17:15:50.5506038Z (Worker_PP1_TP1 pid=16411) [rank3]:W1103 17:15:50.549000 16411 site-packages/torch/_inductor/utils.py:1665] [0/0] Not enough SMs to use max_autotune_gemm mode 2025-11-03T17:15:50.6159859Z (Worker_PP0_TP1 pid=16409) [rank1]:W1103 17:15:50.615000 16409 site-packages/torch/_inductor/utils.py:1665] [0/0] Not enough SMs to use max_autotune_gemm mode 2025-11-03T17:15:50.6687398Z (Worker_PP0_TP0 pid=16408) [rank0]:W1103 17:15:50.667000 16408 site-packages/torch/_inductor/utils.py:1665] [0/0] Not enough SMs to use max_autotune_gemm mode 2025-11-03T17:15:51.4331917Z (Worker_PP1_TP0 pid=16410) INFO 11-03 17:15:51 [backends.py:214] Cache the graph for dynamic shape for later use 2025-11-03T17:15:51.6446326Z (Worker_PP1_TP1 pid=16411) INFO 11-03 17:15:51 [backends.py:214] Cache the graph for dynamic shape for later use 2025-11-03T17:15:51.9410132Z (Worker_PP0_TP1 pid=16409) INFO 11-03 17:15:51 [backends.py:214] Cache the graph for dynamic shape for later use 2025-11-03T17:15:52.0135240Z (Worker_PP0_TP0 pid=16408) INFO 11-03 17:15:52 [backends.py:214] Cache the graph for dynamic shape for later use 2025-11-03T17:15:55.0961253Z (Worker_PP0_TP1 pid=16409) INFO 11-03 17:15:55 [backends.py:241] Compiling a graph for dynamic shape takes 5.35 s 2025-11-03T17:15:55.1155985Z (Worker_PP0_TP0 pid=16408) INFO 11-03 17:15:55 [backends.py:241] Compiling a graph for dynamic shape takes 5.32 s 2025-11-03T17:15:55.1942878Z (Worker_PP1_TP0 pid=16410) INFO 11-03 17:15:55 [backends.py:241] Compiling a graph for dynamic shape takes 5.61 s 2025-11-03T17:15:55.5334041Z (Worker_PP1_TP1 pid=16411) INFO 11-03 17:15:55 [backends.py:241] Compiling a graph for dynamic shape takes 5.80 s 2025-11-03T17:15:58.9144184Z (Worker_PP0_TP0 pid=16408) INFO 11-03 17:15:58 [gpu_worker.py:314] Available KV cache memory: 18.97 GiB 2025-11-03T17:15:58.9347999Z (Worker_PP0_TP1 pid=16409) INFO 11-03 17:15:58 [gpu_worker.py:314] Available KV cache memory: 18.97 GiB 2025-11-03T17:15:59.5720768Z (Worker_PP1_TP1 pid=16411) INFO 11-03 17:15:59 [gpu_worker.py:314] Available KV cache memory: 18.94 GiB 2025-11-03T17:15:59.5769879Z (Worker_PP1_TP0 pid=16410) INFO 11-03 17:15:59 [gpu_worker.py:314] Available KV cache memory: 18.94 GiB 2025-11-03T17:15:59.8398446Z (EngineCore_DP0 pid=16289) INFO 11-03 17:15:59 [kv_cache_utils.py:1199] GPU KV cache size: 2,482,944 tokens 2025-11-03T17:15:59.8399412Z (EngineCore_DP0 pid=16289) INFO 11-03 17:15:59 [kv_cache_utils.py:1204] Maximum concurrency for 2,048 tokens per request: 1212.38x 2025-11-03T17:15:59.8400269Z (EngineCore_DP0 pid=16289) INFO 11-03 17:15:59 [kv_cache_utils.py:1199] GPU KV cache size: 2,482,944 tokens 2025-11-03T17:15:59.8401101Z (EngineCore_DP0 pid=16289) INFO 11-03 17:15:59 [kv_cache_utils.py:1204] Maximum concurrency for 2,048 tokens per request: 1212.38x 2025-11-03T17:15:59.8401908Z (EngineCore_DP0 pid=16289) INFO 11-03 17:15:59 [kv_cache_utils.py:1199] GPU KV cache size: 2,482,944 tokens 2025-11-03T17:15:59.8402750Z (EngineCore_DP0 pid=16289) INFO 11-03 17:15:59 [kv_cache_utils.py:1204] Maximum concurrency for 2,048 tokens per request: 1212.38x 2025-11-03T17:15:59.8403451Z (EngineCore_DP0 pid=16289) INFO 11-03 17:15:59 [kv_cache_utils.py:1199] GPU KV cache size: 2,482,944 tokens 2025-11-03T17:15:59.8404115Z (EngineCore_DP0 pid=16289) INFO 11-03 17:15:59 [kv_cache_utils.py:1204] Maximum concurrency for 2,048 tokens per request: 1212.38x 2025-11-03T17:16:00.1579360Z (Worker_PP0_TP0 pid=16408) 2025-11-03T17:16:01.1297772Z Capturing CUDA graphs (mixed prefill-decode, PIECEWISE): 0% 0/4 [00:00, 'use_cudagraph': True, 'cudagraph_num_of_warmups': 1, 'cudagraph_capture_sizes': [16, 8, 4, 2, 1], 'cudagraph_copy_inputs': False, 'full_cuda_graph': True, 'use_inductor_graph_partition': False, 'pass_config': {}, 'max_capture_size': 16, 'local_cache_dir': None} 2025-11-03T17:16:46.0927261Z (EngineCore_DP0 pid=20406) WARNING 11-03 17:16:46 [multiproc_executor.py:744] Reducing Torch parallelism from 24 threads to 1 to avoid unnecessary CPU contention. Set OMP_NUM_THREADS in the external environment to tune this value as needed. 2025-11-03T17:16:46.0930021Z (EngineCore_DP0 pid=20406) INFO 11-03 17:16:46 [shm_broadcast.py:302] vLLM message queue communication handle: Handle(local_reader_ranks=[0, 1], buffer_handle=(2, 16777216, 10, 'psm_bf20a3a6'), local_subscribe_addr='ipc:///tmp/f461877f-3f3e-4929-a6b2-8be8fd3bb46b', remote_subscribe_addr=None, remote_addr_ipv6=False) 2025-11-03T17:16:46.5741161Z /opt/conda/envs/py_3.12/lib/python3.12/site-packages/torch/cuda/__init__.py:65: FutureWarning: The pynvml package is deprecated. Please install nvidia-ml-py instead. If you did not install pynvml directly, please report this to the maintainers of the package that installed pynvml for you. 2025-11-03T17:16:46.5742406Z import pynvml # type: ignore[import] 2025-11-03T17:16:46.5882563Z /opt/conda/envs/py_3.12/lib/python3.12/site-packages/torch/cuda/__init__.py:65: FutureWarning: The pynvml package is deprecated. Please install nvidia-ml-py instead. If you did not install pynvml directly, please report this to the maintainers of the package that installed pynvml for you. 2025-11-03T17:16:46.5883724Z import pynvml # type: ignore[import] 2025-11-03T17:16:48.9106722Z INFO 11-03 17:16:48 [__init__.py:224] Automatically detected platform cuda. 2025-11-03T17:16:48.9362105Z INFO 11-03 17:16:48 [__init__.py:224] Automatically detected platform cuda. 2025-11-03T17:16:53.6231213Z INFO 11-03 17:16:53 [shm_broadcast.py:302] vLLM message queue communication handle: Handle(local_reader_ranks=[0], buffer_handle=(1, 10485760, 10, 'psm_74908666'), local_subscribe_addr='ipc:///tmp/fdd4b77a-e03e-48c2-b006-27e66326f50f', remote_subscribe_addr=None, remote_addr_ipv6=False) 2025-11-03T17:16:53.6232898Z INFO 11-03 17:16:53 [shm_broadcast.py:302] vLLM message queue communication handle: Handle(local_reader_ranks=[0], buffer_handle=(1, 10485760, 10, 'psm_75ec3156'), local_subscribe_addr='ipc:///tmp/87dc59ed-3b5c-4e3f-924f-4e03e22e84fe', remote_subscribe_addr=None, remote_addr_ipv6=False) 2025-11-03T17:16:53.8206269Z [Gloo] Rank 0 is connected to 1 peer ranks. Expected number of connected peer ranks is : 1 2025-11-03T17:16:53.8207483Z [Gloo] Rank 1 is connected to 1 peer ranks. Expected number of connected peer ranks is : 1 2025-11-03T17:16:53.8251444Z [Gloo] Rank 0 is connected to 1 peer ranks. Expected number of connected peer ranks is : 1 2025-11-03T17:16:53.8252668Z [Gloo] Rank 1 is connected to 1 peer ranks. Expected number of connected peer ranks is : 1 2025-11-03T17:16:53.8293232Z INFO 11-03 17:16:53 [__init__.py:1146] Found nccl from library libnccl.so.2 2025-11-03T17:16:53.8294138Z INFO 11-03 17:16:53 [__init__.py:1146] Found nccl from library libnccl.so.2 2025-11-03T17:16:53.8295593Z INFO 11-03 17:16:53 [pynccl.py:109] vLLM is using nccl==2.27.5 2025-11-03T17:16:53.8296311Z INFO 11-03 17:16:53 [pynccl.py:109] vLLM is using nccl==2.27.5 2025-11-03T17:16:53.9983202Z WARNING 11-03 17:16:53 [symm_mem.py:60] SymmMemCommunicator: Device capability 8.9 not supported, communicator is not available. 2025-11-03T17:16:53.9983917Z WARNING 11-03 17:16:53 [symm_mem.py:60] SymmMemCommunicator: Device capability 8.9 not supported, communicator is not available. 2025-11-03T17:16:54.0007968Z INFO 11-03 17:16:54 [custom_all_reduce.py:37] Skipping P2P check and trusting the driver's P2P report. 2025-11-03T17:16:54.0011245Z WARNING 11-03 17:16:54 [custom_all_reduce.py:162] Custom allreduce is disabled because your platform lacks GPU P2P capability or P2P test failed. To silence this warning, specify disable_custom_all_reduce=True explicitly. 2025-11-03T17:16:54.0012255Z INFO 11-03 17:16:54 [custom_all_reduce.py:37] Skipping P2P check and trusting the driver's P2P report. 2025-11-03T17:16:54.0013259Z WARNING 11-03 17:16:54 [custom_all_reduce.py:162] Custom allreduce is disabled because your platform lacks GPU P2P capability or P2P test failed. To silence this warning, specify disable_custom_all_reduce=True explicitly. 2025-11-03T17:16:54.0727399Z INFO 11-03 17:16:54 [shm_broadcast.py:302] vLLM message queue communication handle: Handle(local_reader_ranks=[1], buffer_handle=(1, 4194304, 6, 'psm_4333a6ae'), local_subscribe_addr='ipc:///tmp/5e7f4dea-f535-402a-92b8-07b0b5a9575c', remote_subscribe_addr=None, remote_addr_ipv6=False) 2025-11-03T17:16:54.0758768Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:16:54.0759886Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:16:54.0769550Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:16:54.0775950Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:16:54.1121606Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:16:54.1133172Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:16:54.1146950Z [Gloo] Rank 1 is connected to 1 peer ranks. Expected number of connected peer ranks is : 1 2025-11-03T17:16:54.1148068Z [Gloo] Rank 0 is connected to 1 peer ranks. Expected number of connected peer ranks is : 1 2025-11-03T17:16:54.1149484Z INFO 11-03 17:16:54 [__init__.py:1146] Found nccl from library libnccl.so.2 2025-11-03T17:16:54.1150459Z INFO 11-03 17:16:54 [__init__.py:1146] Found nccl from library libnccl.so.2 2025-11-03T17:16:54.1151405Z INFO 11-03 17:16:54 [pynccl.py:109] vLLM is using nccl==2.27.5 2025-11-03T17:16:54.1152047Z INFO 11-03 17:16:54 [pynccl.py:109] vLLM is using nccl==2.27.5 2025-11-03T17:16:54.1491973Z INFO 11-03 17:16:54 [parallel_state.py:1231] rank 0 in world size 2 is assigned as DP rank 0, PP rank 0, TP rank 0, EP rank 0 2025-11-03T17:16:54.1493392Z INFO 11-03 17:16:54 [parallel_state.py:1231] rank 1 in world size 2 is assigned as DP rank 0, PP rank 0, TP rank 1, EP rank 1 2025-11-03T17:16:54.3693133Z INFO 11-03 17:16:54 [topk_topp_sampler.py:59] Using FlashInfer for top-p & top-k sampling. 2025-11-03T17:16:54.3809927Z INFO 11-03 17:16:54 [topk_topp_sampler.py:59] Using FlashInfer for top-p & top-k sampling. 2025-11-03T17:16:54.4018949Z (Worker_TP1 pid=20526) INFO 11-03 17:16:54 [gpu_model_runner.py:2840] Starting to load model meta-llama/Llama-3.2-1B-Instruct... 2025-11-03T17:16:54.4130703Z (Worker_TP0 pid=20525) INFO 11-03 17:16:54 [gpu_model_runner.py:2840] Starting to load model meta-llama/Llama-3.2-1B-Instruct... 2025-11-03T17:16:54.6020651Z (Worker_TP1 pid=20526) INFO 11-03 17:16:54 [gpu_model_runner.py:2870] Loading model from scratch... 2025-11-03T17:16:54.6218222Z (Worker_TP0 pid=20525) INFO 11-03 17:16:54 [gpu_model_runner.py:2870] Loading model from scratch... 2025-11-03T17:16:54.6936693Z (Worker_TP1 pid=20526) INFO 11-03 17:16:54 [cuda.py:404] Using Flash Attention backend on V1 engine. 2025-11-03T17:16:54.7120436Z (Worker_TP0 pid=20525) INFO 11-03 17:16:54 [cuda.py:404] Using Flash Attention backend on V1 engine. 2025-11-03T17:16:55.1778185Z (Worker_TP1 pid=20526) INFO 11-03 17:16:55 [gpu_model_runner.py:2902] Model loading took 1.1667 GiB and 0.169009 seconds 2025-11-03T17:16:55.1923268Z (Worker_TP0 pid=20525) INFO 11-03 17:16:55 [gpu_model_runner.py:2902] Model loading took 1.1667 GiB and 0.167078 seconds 2025-11-03T17:16:58.8290160Z (Worker_TP1 pid=20526) INFO 11-03 17:16:58 [backends.py:594] Using cache directory: /var/lib/jenkins/.cache/vllm/torch_compile_cache/cdd8585f72/rank_1_0/backbone for vLLM's torch.compile 2025-11-03T17:16:58.8294808Z (Worker_TP1 pid=20526) INFO 11-03 17:16:58 [backends.py:608] Dynamo bytecode transform time: 3.36 s 2025-11-03T17:16:58.9203372Z (Worker_TP0 pid=20525) INFO 11-03 17:16:58 [backends.py:594] Using cache directory: /var/lib/jenkins/.cache/vllm/torch_compile_cache/cdd8585f72/rank_0_0/backbone for vLLM's torch.compile 2025-11-03T17:16:58.9210023Z (Worker_TP0 pid=20525) INFO 11-03 17:16:58 [backends.py:608] Dynamo bytecode transform time: 3.45 s 2025-11-03T17:16:59.9277164Z (Worker_TP1 pid=20526) INFO 11-03 17:16:59 [backends.py:179] Directly load the compiled graph(s) for dynamic shape from the cache, took 0.911 s 2025-11-03T17:17:00.0171801Z (Worker_TP0 pid=20525) INFO 11-03 17:17:00 [backends.py:179] Directly load the compiled graph(s) for dynamic shape from the cache, took 0.907 s 2025-11-03T17:17:00.5514405Z (Worker_TP0 pid=20525) INFO 11-03 17:17:00 [monitor.py:33] torch.compile takes 3.45 s in total 2025-11-03T17:17:00.5515458Z (Worker_TP1 pid=20526) INFO 11-03 17:17:00 [monitor.py:33] torch.compile takes 3.36 s in total 2025-11-03T17:17:01.1985945Z (Worker_TP1 pid=20526) INFO 11-03 17:17:01 [gpu_worker.py:314] Available KV cache memory: 18.54 GiB 2025-11-03T17:17:01.2084189Z (Worker_TP0 pid=20525) INFO 11-03 17:17:01 [gpu_worker.py:314] Available KV cache memory: 18.54 GiB 2025-11-03T17:17:01.4508193Z (EngineCore_DP0 pid=20406) INFO 11-03 17:17:01 [kv_cache_utils.py:1199] GPU KV cache size: 1,214,720 tokens 2025-11-03T17:17:01.4509098Z (EngineCore_DP0 pid=20406) INFO 11-03 17:17:01 [kv_cache_utils.py:1204] Maximum concurrency for 2,048 tokens per request: 593.12x 2025-11-03T17:17:01.4510144Z (EngineCore_DP0 pid=20406) INFO 11-03 17:17:01 [kv_cache_utils.py:1199] GPU KV cache size: 1,214,720 tokens 2025-11-03T17:17:01.4511019Z (EngineCore_DP0 pid=20406) INFO 11-03 17:17:01 [kv_cache_utils.py:1204] Maximum concurrency for 2,048 tokens per request: 593.12x 2025-11-03T17:17:01.7003502Z (Worker_TP0 pid=20525) 2025-11-03T17:17:01.8233689Z Capturing CUDA graphs (mixed prefill-decode, PIECEWISE): 0% 0/5 [00:00, 'use_cudagraph': True, 'cudagraph_num_of_warmups': 1, 'cudagraph_capture_sizes': [16, 8, 4, 2], 'cudagraph_copy_inputs': False, 'full_cuda_graph': True, 'use_inductor_graph_partition': False, 'pass_config': {'enable_noop': True, 'enable_sequence_parallelism': True}, 'max_capture_size': 16, 'local_cache_dir': None} 2025-11-03T17:17:22.9029173Z (EngineCore_DP0 pid=20902) 2025-11-03 17:17:22,902 INFO worker.py:1918 -- Started a local Ray instance. View the dashboard at http://127.0.0.1:8265  2025-11-03T17:17:23.8609246Z (EngineCore_DP0 pid=20902) INFO 11-03 17:17:23 [ray_utils.py:373] No current placement group found. Creating a new placement group. 2025-11-03T17:17:24.0388951Z (EngineCore_DP0 pid=20902) INFO 11-03 17:17:24 [ray_distributed_executor.py:179] use_ray_spmd_worker: True 2025-11-03T17:17:24.6823985Z (EngineCore_DP0 pid=20902) (pid=21789) /opt/conda/envs/py_3.12/lib/python3.12/site-packages/torch/cuda/__init__.py:65: FutureWarning: The pynvml package is deprecated. Please install nvidia-ml-py instead. If you did not install pynvml directly, please report this to the maintainers of the package that installed pynvml for you. 2025-11-03T17:17:24.6826414Z (EngineCore_DP0 pid=20902) (pid=21789) import pynvml # type: ignore[import] 2025-11-03T17:17:27.0341803Z (EngineCore_DP0 pid=20902) (pid=21785) INFO 11-03 17:17:26 [__init__.py:224] Automatically detected platform cuda. 2025-11-03T17:17:28.6028548Z (EngineCore_DP0 pid=20902) INFO 11-03 17:17:28 [ray_env.py:67] RAY_NON_CARRY_OVER_ENV_VARS from config: set() 2025-11-03T17:17:28.6030311Z (EngineCore_DP0 pid=20902) INFO 11-03 17:17:28 [ray_env.py:70] Copying the following environment variables to workers: ['HF_TOKEN', 'HUGGING_FACE_HUB_TOKEN', 'LD_LIBRARY_PATH', 'VLLM_USE_V1', 'VLLM_USE_RAY_COMPILED_DAG', 'VLLM_WORKER_MULTIPROC_METHOD', 'VLLM_USE_RAY_SPMD_WORKER', 'MAX_JOBS'] 2025-11-03T17:17:28.6031633Z (EngineCore_DP0 pid=20902) INFO 11-03 17:17:28 [ray_env.py:75] If certain env vars should NOT be copied, add them to /var/lib/jenkins/.config/vllm/ray_non_carry_over_env_vars.json file 2025-11-03T17:17:31.9342049Z (EngineCore_DP0 pid=20902) (RayWorkerWrapper pid=21789) WARNING 11-03 17:17:31 [worker_base.py:307] Missing `shared_worker_lock` argument from executor. This argument is needed for mm_processor_cache_type='shm'. 2025-11-03T17:17:32.0383891Z (EngineCore_DP0 pid=20902) (pid=21792) INFO 11-03 17:17:27 [__init__.py:224] Automatically detected platform cuda. [repeated 3x across cluster] (Ray deduplicates logs by default. Set RAY_DEDUP_LOGS=0 to disable log deduplication, or see https://docs.ray.io/en/master/ray-observability/user-guides/configure-logging.html#log-deduplication for more options.) 2025-11-03T17:17:32.9599726Z (EngineCore_DP0 pid=20902) (RayWorkerWrapper pid=21789) [Gloo] Rank 0 is connected to 3 peer ranks. Expected number of connected peer ranks is : 3 2025-11-03T17:17:33.0654075Z (EngineCore_DP0 pid=20902) (RayWorkerWrapper pid=21789) INFO 11-03 17:17:33 [__init__.py:1146] Found nccl from library libnccl.so.2 2025-11-03T17:17:33.0655147Z (EngineCore_DP0 pid=20902) (RayWorkerWrapper pid=21789) INFO 11-03 17:17:33 [pynccl.py:109] vLLM is using nccl==2.27.5 2025-11-03T17:17:33.4771621Z (EngineCore_DP0 pid=20902) (RayWorkerWrapper pid=21789) WARNING 11-03 17:17:33 [symm_mem.py:60] SymmMemCommunicator: Device capability 8.9 not supported, communicator is not available. 2025-11-03T17:17:33.4773108Z (EngineCore_DP0 pid=20902) (RayWorkerWrapper pid=21789) INFO 11-03 17:17:33 [custom_all_reduce.py:37] Skipping P2P check and trusting the driver's P2P report. 2025-11-03T17:17:33.4774598Z (EngineCore_DP0 pid=20902) (RayWorkerWrapper pid=21789) WARNING 11-03 17:17:33 [custom_all_reduce.py:162] Custom allreduce is disabled because your platform lacks GPU P2P capability or P2P test failed. To silence this warning, specify disable_custom_all_reduce=True explicitly. 2025-11-03T17:17:33.4776436Z (EngineCore_DP0 pid=20902) (RayWorkerWrapper pid=21789) INFO 11-03 17:17:33 [shm_broadcast.py:302] vLLM message queue communication handle: Handle(local_reader_ranks=[1], buffer_handle=(1, 4194304, 6, 'psm_ab9bfa6c'), local_subscribe_addr='ipc:///tmp/b28ae626-2109-47a0-89f6-e30324f8f087', remote_subscribe_addr=None, remote_addr_ipv6=False) 2025-11-03T17:17:33.5827997Z (EngineCore_DP0 pid=20902) (RayWorkerWrapper pid=21789) INFO 11-03 17:17:33 [parallel_state.py:1231] rank 0 in world size 4 is assigned as DP rank 0, PP rank 0, TP rank 0, EP rank 0 2025-11-03T17:17:33.7898688Z (EngineCore_DP0 pid=20902) (RayWorkerWrapper pid=21789) INFO 11-03 17:17:33 [topk_topp_sampler.py:59] Using FlashInfer for top-p & top-k sampling. 2025-11-03T17:17:33.7910036Z (EngineCore_DP0 pid=20902) (RayWorkerWrapper pid=21791) INFO 11-03 17:17:33 [gpu_model_runner.py:2840] Starting to load model meta-llama/Llama-3.2-1B-Instruct... 2025-11-03T17:17:34.1016972Z (EngineCore_DP0 pid=20902) (RayWorkerWrapper pid=21789) INFO 11-03 17:17:34 [gpu_model_runner.py:2870] Loading model from scratch... 2025-11-03T17:17:34.2073112Z (EngineCore_DP0 pid=20902) (RayWorkerWrapper pid=21789) INFO 11-03 17:17:34 [cuda.py:404] Using Flash Attention backend on V1 engine. 2025-11-03T17:17:34.6188947Z (EngineCore_DP0 pid=20902) (RayWorkerWrapper pid=21792) INFO 11-03 17:17:34 [gpu_model_runner.py:2902] Model loading took 0.7135 GiB and 0.156383 seconds 2025-11-03T17:17:37.1739994Z (EngineCore_DP0 pid=20902) (RayWorkerWrapper pid=21789) INFO 11-03 17:17:37 [backends.py:594] Using cache directory: /var/lib/jenkins/.cache/vllm/torch_compile_cache/3ff2b0020c/rank_0_0/backbone for vLLM's torch.compile 2025-11-03T17:17:37.1741425Z (EngineCore_DP0 pid=20902) (RayWorkerWrapper pid=21789) INFO 11-03 17:17:37 [backends.py:608] Dynamo bytecode transform time: 2.21 s 2025-11-03T17:17:37.1742932Z (EngineCore_DP0 pid=20902) (RayWorkerWrapper pid=21792) WARNING 11-03 17:17:31 [worker_base.py:307] Missing `shared_worker_lock` argument from executor. This argument is needed for mm_processor_cache_type='shm'. [repeated 3x across cluster] 2025-11-03T17:17:37.7887458Z (EngineCore_DP0 pid=20902) (RayWorkerWrapper pid=21791) INFO 11-03 17:17:37 [backends.py:179] Directly load the compiled graph(s) for dynamic shape from the cache, took 0.415 s 2025-11-03T17:17:38.8139063Z (EngineCore_DP0 pid=20902) (RayWorkerWrapper pid=21789) INFO 11-03 17:17:38 [gpu_worker.py:314] Available KV cache memory: 19.02 GiB 2025-11-03T17:17:38.8140116Z (EngineCore_DP0 pid=20902) (RayWorkerWrapper pid=21792) [Gloo] Rank 1 is connected to 1 peer ranks. Expected number of connected peer ranks is : 1 [repeated 23x across cluster] 2025-11-03T17:17:38.8141147Z (EngineCore_DP0 pid=20902) (RayWorkerWrapper pid=21792) INFO 11-03 17:17:33 [__init__.py:1146] Found nccl from library libnccl.so.2 [repeated 11x across cluster] 2025-11-03T17:17:38.8142068Z (EngineCore_DP0 pid=20902) (RayWorkerWrapper pid=21792) INFO 11-03 17:17:33 [pynccl.py:109] vLLM is using nccl==2.27.5 [repeated 11x across cluster] 2025-11-03T17:17:38.8143629Z (EngineCore_DP0 pid=20902) (RayWorkerWrapper pid=21792) WARNING 11-03 17:17:33 [symm_mem.py:60] SymmMemCommunicator: Device capability 8.9 not supported, communicator is not available. [repeated 3x across cluster] 2025-11-03T17:17:38.8145191Z (EngineCore_DP0 pid=20902) (RayWorkerWrapper pid=21792) INFO 11-03 17:17:33 [custom_all_reduce.py:37] Skipping P2P check and trusting the driver's P2P report. [repeated 3x across cluster] 2025-11-03T17:17:38.8146840Z (EngineCore_DP0 pid=20902) (RayWorkerWrapper pid=21792) WARNING 11-03 17:17:33 [custom_all_reduce.py:162] Custom allreduce is disabled because your platform lacks GPU P2P capability or P2P test failed. To silence this warning, specify disable_custom_all_reduce=True explicitly. [repeated 3x across cluster] 2025-11-03T17:17:38.8148775Z (EngineCore_DP0 pid=20902) (RayWorkerWrapper pid=21791) INFO 11-03 17:17:33 [shm_broadcast.py:302] vLLM message queue communication handle: Handle(local_reader_ranks=[1], buffer_handle=(1, 4194304, 6, 'psm_f775cb7e'), local_subscribe_addr='ipc:///tmp/c9a4424e-d742-44be-8296-2a032c657543', remote_subscribe_addr=None, remote_addr_ipv6=False) 2025-11-03T17:17:38.8150282Z (EngineCore_DP0 pid=20902) (RayWorkerWrapper pid=21792) INFO 11-03 17:17:33 [parallel_state.py:1231] rank 3 in world size 4 is assigned as DP rank 0, PP rank 1, TP rank 1, EP rank 1 [repeated 3x across cluster] 2025-11-03T17:17:38.8151385Z (EngineCore_DP0 pid=20902) (RayWorkerWrapper pid=21792) INFO 11-03 17:17:33 [topk_topp_sampler.py:59] Using FlashInfer for top-p & top-k sampling. [repeated 3x across cluster] 2025-11-03T17:17:38.8152469Z (EngineCore_DP0 pid=20902) (RayWorkerWrapper pid=21785) INFO 11-03 17:17:33 [gpu_model_runner.py:2840] Starting to load model meta-llama/Llama-3.2-1B-Instruct... [repeated 3x across cluster] 2025-11-03T17:17:39.0606804Z (EngineCore_DP0 pid=20902) INFO 11-03 17:17:39 [kv_cache_utils.py:1199] GPU KV cache size: 2,489,088 tokens 2025-11-03T17:17:39.0608815Z (EngineCore_DP0 pid=20902) INFO 11-03 17:17:39 [kv_cache_utils.py:1204] Maximum concurrency for 2,048 tokens per request: 1215.38x 2025-11-03T17:17:39.0610463Z (EngineCore_DP0 pid=20902) INFO 11-03 17:17:39 [kv_cache_utils.py:1199] GPU KV cache size: 2,489,088 tokens 2025-11-03T17:17:39.0611326Z (EngineCore_DP0 pid=20902) INFO 11-03 17:17:39 [kv_cache_utils.py:1204] Maximum concurrency for 2,048 tokens per request: 1215.38x 2025-11-03T17:17:39.0611997Z (EngineCore_DP0 pid=20902) INFO 11-03 17:17:39 [kv_cache_utils.py:1199] GPU KV cache size: 2,489,088 tokens 2025-11-03T17:17:39.0612645Z (EngineCore_DP0 pid=20902) INFO 11-03 17:17:39 [kv_cache_utils.py:1204] Maximum concurrency for 2,048 tokens per request: 1215.38x 2025-11-03T17:17:39.0613886Z (EngineCore_DP0 pid=20902) INFO 11-03 17:17:39 [kv_cache_utils.py:1199] GPU KV cache size: 2,489,088 tokens 2025-11-03T17:17:39.0614563Z (EngineCore_DP0 pid=20902) INFO 11-03 17:17:39 [kv_cache_utils.py:1204] Maximum concurrency for 2,048 tokens per request: 1215.38x 2025-11-03T17:17:39.3271023Z (EngineCore_DP0 pid=20902) (RayWorkerWrapper pid=21789) 2025-11-03T17:17:39.3271569Z Capturing CUDA graphs (mixed prefill-decode, PIECEWISE): 0%| | 0/4 [00:00, 'use_cudagraph': True, 'cudagraph_num_of_warmups': 1, 'cudagraph_capture_sizes': [16, 8, 4, 2, 1], 'cudagraph_copy_inputs': False, 'full_cuda_graph': True, 'use_inductor_graph_partition': False, 'pass_config': {}, 'max_capture_size': 16, 'local_cache_dir': None} 2025-11-03T17:18:10.8803363Z (EngineCore_DP0 pid=26486) WARNING 11-03 17:18:10 [multiproc_executor.py:744] Reducing Torch parallelism from 24 threads to 1 to avoid unnecessary CPU contention. Set OMP_NUM_THREADS in the external environment to tune this value as needed. 2025-11-03T17:18:10.8812136Z (EngineCore_DP0 pid=26486) INFO 11-03 17:18:10 [shm_broadcast.py:302] vLLM message queue communication handle: Handle(local_reader_ranks=[0, 1], buffer_handle=(2, 16777216, 10, 'psm_847d4500'), local_subscribe_addr='ipc:///tmp/242cf656-ffbb-49a0-a259-07ba712422ae', remote_subscribe_addr=None, remote_addr_ipv6=False) 2025-11-03T17:18:11.3681468Z /opt/conda/envs/py_3.12/lib/python3.12/site-packages/torch/cuda/__init__.py:65: FutureWarning: The pynvml package is deprecated. Please install nvidia-ml-py instead. If you did not install pynvml directly, please report this to the maintainers of the package that installed pynvml for you. 2025-11-03T17:18:11.3682739Z import pynvml # type: ignore[import] 2025-11-03T17:18:11.3781773Z /opt/conda/envs/py_3.12/lib/python3.12/site-packages/torch/cuda/__init__.py:65: FutureWarning: The pynvml package is deprecated. Please install nvidia-ml-py instead. If you did not install pynvml directly, please report this to the maintainers of the package that installed pynvml for you. 2025-11-03T17:18:11.3782908Z import pynvml # type: ignore[import] 2025-11-03T17:18:13.7128994Z INFO 11-03 17:18:13 [__init__.py:224] Automatically detected platform cuda. 2025-11-03T17:18:13.7379868Z INFO 11-03 17:18:13 [__init__.py:224] Automatically detected platform cuda. 2025-11-03T17:18:18.4330039Z INFO 11-03 17:18:18 [shm_broadcast.py:302] vLLM message queue communication handle: Handle(local_reader_ranks=[0], buffer_handle=(1, 10485760, 10, 'psm_730572ac'), local_subscribe_addr='ipc:///tmp/fde86f9f-aed8-4865-9701-fc3dc7149976', remote_subscribe_addr=None, remote_addr_ipv6=False) 2025-11-03T17:18:18.4333668Z INFO 11-03 17:18:18 [shm_broadcast.py:302] vLLM message queue communication handle: Handle(local_reader_ranks=[0], buffer_handle=(1, 10485760, 10, 'psm_821206bb'), local_subscribe_addr='ipc:///tmp/5658c0d1-710c-4437-92e1-4432cc83b8f6', remote_subscribe_addr=None, remote_addr_ipv6=False) 2025-11-03T17:18:19.3840934Z [Gloo] Rank 1 is connected to 1 peer ranks. Expected number of connected peer ranks is : 1 2025-11-03T17:18:19.3842137Z [Gloo] Rank 0 is connected to 1 peer ranks. Expected number of connected peer ranks is : 1 2025-11-03T17:18:19.3887925Z [Gloo] Rank 0 is connected to 1 peer ranks. Expected number of connected peer ranks is : 1 2025-11-03T17:18:19.3889257Z [Gloo] Rank 1 is connected to 1 peer ranks. Expected number of connected peer ranks is : 1 2025-11-03T17:18:19.3932711Z INFO 11-03 17:18:19 [__init__.py:1146] Found nccl from library libnccl.so.2 2025-11-03T17:18:19.3933610Z INFO 11-03 17:18:19 [pynccl.py:109] vLLM is using nccl==2.27.5 2025-11-03T17:18:19.3936421Z INFO 11-03 17:18:19 [__init__.py:1146] Found nccl from library libnccl.so.2 2025-11-03T17:18:19.3938592Z INFO 11-03 17:18:19 [pynccl.py:109] vLLM is using nccl==2.27.5 2025-11-03T17:18:19.5693067Z WARNING 11-03 17:18:19 [symm_mem.py:60] SymmMemCommunicator: Device capability 8.9 not supported, communicator is not available. 2025-11-03T17:18:19.5693977Z WARNING 11-03 17:18:19 [symm_mem.py:60] SymmMemCommunicator: Device capability 8.9 not supported, communicator is not available. 2025-11-03T17:18:19.5721794Z INFO 11-03 17:18:19 [custom_all_reduce.py:37] Skipping P2P check and trusting the driver's P2P report. 2025-11-03T17:18:19.5722336Z INFO 11-03 17:18:19 [custom_all_reduce.py:37] Skipping P2P check and trusting the driver's P2P report. 2025-11-03T17:18:19.5723569Z WARNING 11-03 17:18:19 [custom_all_reduce.py:162] Custom allreduce is disabled because your platform lacks GPU P2P capability or P2P test failed. To silence this warning, specify disable_custom_all_reduce=True explicitly. 2025-11-03T17:18:19.5724816Z WARNING 11-03 17:18:19 [custom_all_reduce.py:162] Custom allreduce is disabled because your platform lacks GPU P2P capability or P2P test failed. To silence this warning, specify disable_custom_all_reduce=True explicitly. 2025-11-03T17:18:19.6324344Z INFO 11-03 17:18:19 [shm_broadcast.py:302] vLLM message queue communication handle: Handle(local_reader_ranks=[1], buffer_handle=(1, 4194304, 6, 'psm_a36d6020'), local_subscribe_addr='ipc:///tmp/1ba794f9-524a-4c3b-8d35-67c9d505adaa', remote_subscribe_addr=None, remote_addr_ipv6=False) 2025-11-03T17:18:19.6352540Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:18:19.6353289Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:18:19.6361238Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:18:19.6368624Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:18:19.7121667Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:18:19.7129639Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:18:19.7140376Z [Gloo] Rank 0 is connected to 1 peer ranks. Expected number of connected peer ranks is : 1 2025-11-03T17:18:19.7140941Z [Gloo] Rank 1 is connected to 1 peer ranks. Expected number of connected peer ranks is : 1 2025-11-03T17:18:19.7143171Z INFO 11-03 17:18:19 [__init__.py:1146] Found nccl from library libnccl.so.2 2025-11-03T17:18:19.7143752Z INFO 11-03 17:18:19 [__init__.py:1146] Found nccl from library libnccl.so.2 2025-11-03T17:18:19.7144163Z INFO 11-03 17:18:19 [pynccl.py:109] vLLM is using nccl==2.27.5 2025-11-03T17:18:19.7144467Z INFO 11-03 17:18:19 [pynccl.py:109] vLLM is using nccl==2.27.5 2025-11-03T17:18:19.7465487Z INFO 11-03 17:18:19 [parallel_state.py:1231] rank 0 in world size 2 is assigned as DP rank 0, PP rank 0, TP rank 0, EP rank 0 2025-11-03T17:18:19.7466302Z INFO 11-03 17:18:19 [parallel_state.py:1231] rank 1 in world size 2 is assigned as DP rank 0, PP rank 0, TP rank 1, EP rank 1 2025-11-03T17:18:19.9681752Z INFO 11-03 17:18:19 [topk_topp_sampler.py:59] Using FlashInfer for top-p & top-k sampling. 2025-11-03T17:18:19.9713768Z INFO 11-03 17:18:19 [topk_topp_sampler.py:59] Using FlashInfer for top-p & top-k sampling. 2025-11-03T17:18:19.9997975Z (Worker_TP1 pid=26606) INFO 11-03 17:18:19 [gpu_model_runner.py:2840] Starting to load model meta-llama/Llama-3.2-1B-Instruct... 2025-11-03T17:18:20.0034975Z (Worker_TP0 pid=26605) INFO 11-03 17:18:20 [gpu_model_runner.py:2840] Starting to load model meta-llama/Llama-3.2-1B-Instruct... 2025-11-03T17:18:20.2065206Z (Worker_TP1 pid=26606) INFO 11-03 17:18:20 [gpu_model_runner.py:2870] Loading model from scratch... 2025-11-03T17:18:20.2113354Z (Worker_TP0 pid=26605) INFO 11-03 17:18:20 [gpu_model_runner.py:2870] Loading model from scratch... 2025-11-03T17:18:20.2963871Z (Worker_TP1 pid=26606) INFO 11-03 17:18:20 [cuda.py:404] Using Flash Attention backend on V1 engine. 2025-11-03T17:18:20.3020295Z (Worker_TP0 pid=26605) INFO 11-03 17:18:20 [cuda.py:404] Using Flash Attention backend on V1 engine. 2025-11-03T17:18:20.7802238Z (Worker_TP1 pid=26606) INFO 11-03 17:18:20 [gpu_model_runner.py:2902] Model loading took 1.1667 GiB and 0.167707 seconds 2025-11-03T17:18:20.8021250Z (Worker_TP0 pid=26605) INFO 11-03 17:18:20 [gpu_model_runner.py:2902] Model loading took 1.1667 GiB and 0.169064 seconds 2025-11-03T17:18:24.3938011Z (Worker_TP1 pid=26606) INFO 11-03 17:18:24 [backends.py:594] Using cache directory: /var/lib/jenkins/.cache/vllm/torch_compile_cache/cdd8585f72/rank_1_0/backbone for vLLM's torch.compile 2025-11-03T17:18:24.3942711Z (Worker_TP1 pid=26606) INFO 11-03 17:18:24 [backends.py:608] Dynamo bytecode transform time: 3.31 s 2025-11-03T17:18:24.5696497Z (Worker_TP0 pid=26605) INFO 11-03 17:18:24 [backends.py:594] Using cache directory: /var/lib/jenkins/.cache/vllm/torch_compile_cache/cdd8585f72/rank_0_0/backbone for vLLM's torch.compile 2025-11-03T17:18:24.5701821Z (Worker_TP0 pid=26605) INFO 11-03 17:18:24 [backends.py:608] Dynamo bytecode transform time: 3.49 s 2025-11-03T17:18:25.4741575Z (Worker_TP1 pid=26606) INFO 11-03 17:18:25 [backends.py:179] Directly load the compiled graph(s) for dynamic shape from the cache, took 0.892 s 2025-11-03T17:18:25.6505211Z (Worker_TP0 pid=26605) INFO 11-03 17:18:25 [backends.py:179] Directly load the compiled graph(s) for dynamic shape from the cache, took 0.889 s 2025-11-03T17:18:26.1908125Z (Worker_TP1 pid=26606) INFO 11-03 17:18:26 [monitor.py:33] torch.compile takes 3.31 s in total 2025-11-03T17:18:26.1912112Z (Worker_TP0 pid=26605) INFO 11-03 17:18:26 [monitor.py:33] torch.compile takes 3.49 s in total 2025-11-03T17:18:26.8300763Z (Worker_TP1 pid=26606) INFO 11-03 17:18:26 [gpu_worker.py:314] Available KV cache memory: 18.54 GiB 2025-11-03T17:18:26.8371334Z (Worker_TP0 pid=26605) INFO 11-03 17:18:26 [gpu_worker.py:314] Available KV cache memory: 18.54 GiB 2025-11-03T17:18:27.0768855Z (EngineCore_DP0 pid=26486) INFO 11-03 17:18:27 [kv_cache_utils.py:1199] GPU KV cache size: 1,214,720 tokens 2025-11-03T17:18:27.0770190Z (EngineCore_DP0 pid=26486) INFO 11-03 17:18:27 [kv_cache_utils.py:1204] Maximum concurrency for 2,048 tokens per request: 593.12x 2025-11-03T17:18:27.0771482Z (EngineCore_DP0 pid=26486) INFO 11-03 17:18:27 [kv_cache_utils.py:1199] GPU KV cache size: 1,214,720 tokens 2025-11-03T17:18:27.0772797Z (EngineCore_DP0 pid=26486) INFO 11-03 17:18:27 [kv_cache_utils.py:1204] Maximum concurrency for 2,048 tokens per request: 593.12x 2025-11-03T17:18:27.3248490Z (Worker_TP0 pid=26605) 2025-11-03T17:18:27.4466461Z Capturing CUDA graphs (mixed prefill-decode, PIECEWISE): 0% 0/5 [00:00, 'use_cudagraph': True, 'cudagraph_num_of_warmups': 1, 'cudagraph_capture_sizes': [16, 8, 4, 2], 'cudagraph_copy_inputs': False, 'full_cuda_graph': True, 'use_inductor_graph_partition': False, 'pass_config': {'enable_noop': True, 'enable_sequence_parallelism': True}, 'max_capture_size': 16, 'local_cache_dir': None} 2025-11-03T17:18:51.5642425Z (EngineCore_DP0 pid=26982) WARNING 11-03 17:18:51 [multiproc_executor.py:744] Reducing Torch parallelism from 24 threads to 1 to avoid unnecessary CPU contention. Set OMP_NUM_THREADS in the external environment to tune this value as needed. 2025-11-03T17:18:51.5645378Z (EngineCore_DP0 pid=26982) INFO 11-03 17:18:51 [shm_broadcast.py:302] vLLM message queue communication handle: Handle(local_reader_ranks=[0, 1, 2, 3], buffer_handle=(4, 16777216, 10, 'psm_2ad2e487'), local_subscribe_addr='ipc:///tmp/297f788f-8cbc-4c75-a88d-c5ee8bcc2779', remote_subscribe_addr=None, remote_addr_ipv6=False) 2025-11-03T17:18:52.0560315Z /opt/conda/envs/py_3.12/lib/python3.12/site-packages/torch/cuda/__init__.py:65: FutureWarning: The pynvml package is deprecated. Please install nvidia-ml-py instead. If you did not install pynvml directly, please report this to the maintainers of the package that installed pynvml for you. 2025-11-03T17:18:52.0561588Z import pynvml # type: ignore[import] 2025-11-03T17:18:52.0575957Z /opt/conda/envs/py_3.12/lib/python3.12/site-packages/torch/cuda/__init__.py:65: FutureWarning: The pynvml package is deprecated. Please install nvidia-ml-py instead. If you did not install pynvml directly, please report this to the maintainers of the package that installed pynvml for you. 2025-11-03T17:18:52.0577115Z import pynvml # type: ignore[import] 2025-11-03T17:18:52.0747077Z /opt/conda/envs/py_3.12/lib/python3.12/site-packages/torch/cuda/__init__.py:65: FutureWarning: The pynvml package is deprecated. Please install nvidia-ml-py instead. If you did not install pynvml directly, please report this to the maintainers of the package that installed pynvml for you. 2025-11-03T17:18:52.0748216Z import pynvml # type: ignore[import] 2025-11-03T17:18:52.2112094Z /opt/conda/envs/py_3.12/lib/python3.12/site-packages/torch/cuda/__init__.py:65: FutureWarning: The pynvml package is deprecated. Please install nvidia-ml-py instead. If you did not install pynvml directly, please report this to the maintainers of the package that installed pynvml for you. 2025-11-03T17:18:52.2113763Z import pynvml # type: ignore[import] 2025-11-03T17:18:54.4527576Z INFO 11-03 17:18:54 [__init__.py:224] Automatically detected platform cuda. 2025-11-03T17:18:54.4671144Z INFO 11-03 17:18:54 [__init__.py:224] Automatically detected platform cuda. 2025-11-03T17:18:54.4855003Z INFO 11-03 17:18:54 [__init__.py:224] Automatically detected platform cuda. 2025-11-03T17:18:54.6229003Z INFO 11-03 17:18:54 [__init__.py:224] Automatically detected platform cuda. 2025-11-03T17:18:59.3301589Z INFO 11-03 17:18:59 [shm_broadcast.py:302] vLLM message queue communication handle: Handle(local_reader_ranks=[0], buffer_handle=(1, 10485760, 10, 'psm_b9deff83'), local_subscribe_addr='ipc:///tmp/5f52d28c-a50e-4a75-80bc-c978b72ba836', remote_subscribe_addr=None, remote_addr_ipv6=False) 2025-11-03T17:18:59.3930858Z INFO 11-03 17:18:59 [shm_broadcast.py:302] vLLM message queue communication handle: Handle(local_reader_ranks=[0], buffer_handle=(1, 10485760, 10, 'psm_fb4f3162'), local_subscribe_addr='ipc:///tmp/8cb260fc-dc2a-4b5c-8139-594c43c3c725', remote_subscribe_addr=None, remote_addr_ipv6=False) 2025-11-03T17:18:59.3934456Z INFO 11-03 17:18:59 [shm_broadcast.py:302] vLLM message queue communication handle: Handle(local_reader_ranks=[0], buffer_handle=(1, 10485760, 10, 'psm_89f39396'), local_subscribe_addr='ipc:///tmp/7d969dd4-c00d-4586-86a8-162f4d2c5957', remote_subscribe_addr=None, remote_addr_ipv6=False) 2025-11-03T17:18:59.4424965Z INFO 11-03 17:18:59 [shm_broadcast.py:302] vLLM message queue communication handle: Handle(local_reader_ranks=[0], buffer_handle=(1, 10485760, 10, 'psm_6652011f'), local_subscribe_addr='ipc:///tmp/3a341ead-b851-4d06-ae6f-e309e5f57733', remote_subscribe_addr=None, remote_addr_ipv6=False) 2025-11-03T17:19:00.0262701Z [Gloo] Rank 0 is connected to 3 peer ranks. Expected number of connected peer ranks is : 3 2025-11-03T17:19:00.0263350Z [Gloo] Rank 1 is connected to 3 peer ranks. Expected number of connected peer ranks is : 3 2025-11-03T17:19:00.0264510Z [Gloo] Rank 3 is connected to 3 peer ranks. Expected number of connected peer ranks is : 3 2025-11-03T17:19:00.0265070Z [Gloo] Rank 2 is connected to 3 peer ranks. Expected number of connected peer ranks is : 3 2025-11-03T17:19:00.0317746Z [Gloo] Rank 1 is connected to 1 peer ranks. Expected number of connected peer ranks is : 1 2025-11-03T17:19:00.0318384Z [Gloo] Rank 0 is connected to 1 peer ranks. Expected number of connected peer ranks is : 1 2025-11-03T17:19:00.0318950Z [Gloo] Rank 1 is connected to 1 peer ranks. Expected number of connected peer ranks is : 1 2025-11-03T17:19:00.0319492Z [Gloo] Rank 0 is connected to 1 peer ranks. Expected number of connected peer ranks is : 1 2025-11-03T17:19:00.0362142Z INFO 11-03 17:19:00 [__init__.py:1146] Found nccl from library libnccl.so.2 2025-11-03T17:19:00.0363052Z INFO 11-03 17:19:00 [__init__.py:1146] Found nccl from library libnccl.so.2 2025-11-03T17:19:00.0363879Z INFO 11-03 17:19:00 [__init__.py:1146] Found nccl from library libnccl.so.2 2025-11-03T17:19:00.0364635Z INFO 11-03 17:19:00 [pynccl.py:109] vLLM is using nccl==2.27.5 2025-11-03T17:19:00.0365381Z INFO 11-03 17:19:00 [__init__.py:1146] Found nccl from library libnccl.so.2 2025-11-03T17:19:00.0366110Z INFO 11-03 17:19:00 [pynccl.py:109] vLLM is using nccl==2.27.5 2025-11-03T17:19:00.0366799Z INFO 11-03 17:19:00 [pynccl.py:109] vLLM is using nccl==2.27.5 2025-11-03T17:19:00.0367466Z INFO 11-03 17:19:00 [pynccl.py:109] vLLM is using nccl==2.27.5 2025-11-03T17:19:00.2235563Z WARNING 11-03 17:19:00 [symm_mem.py:60] SymmMemCommunicator: Device capability 8.9 not supported, communicator is not available. 2025-11-03T17:19:00.2236438Z WARNING 11-03 17:19:00 [symm_mem.py:60] SymmMemCommunicator: Device capability 8.9 not supported, communicator is not available. 2025-11-03T17:19:00.2257569Z INFO 11-03 17:19:00 [custom_all_reduce.py:37] Skipping P2P check and trusting the driver's P2P report. 2025-11-03T17:19:00.2258590Z INFO 11-03 17:19:00 [custom_all_reduce.py:37] Skipping P2P check and trusting the driver's P2P report. 2025-11-03T17:19:00.2259611Z WARNING 11-03 17:19:00 [custom_all_reduce.py:162] Custom allreduce is disabled because your platform lacks GPU P2P capability or P2P test failed. To silence this warning, specify disable_custom_all_reduce=True explicitly. 2025-11-03T17:19:00.2261005Z WARNING 11-03 17:19:00 [custom_all_reduce.py:162] Custom allreduce is disabled because your platform lacks GPU P2P capability or P2P test failed. To silence this warning, specify disable_custom_all_reduce=True explicitly. 2025-11-03T17:19:00.2278692Z INFO 11-03 17:19:00 [shm_broadcast.py:302] vLLM message queue communication handle: Handle(local_reader_ranks=[1], buffer_handle=(1, 4194304, 6, 'psm_490515cc'), local_subscribe_addr='ipc:///tmp/94c22c03-ffd6-4d31-99a7-c7e54a856a48', remote_subscribe_addr=None, remote_addr_ipv6=False) 2025-11-03T17:19:00.3017047Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:19:00.3017921Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:19:00.3180604Z WARNING 11-03 17:19:00 [symm_mem.py:60] SymmMemCommunicator: Device capability 8.9 not supported, communicator is not available. 2025-11-03T17:19:00.3182224Z WARNING 11-03 17:19:00 [symm_mem.py:60] SymmMemCommunicator: Device capability 8.9 not supported, communicator is not available. 2025-11-03T17:19:00.3202407Z INFO 11-03 17:19:00 [custom_all_reduce.py:37] Skipping P2P check and trusting the driver's P2P report. 2025-11-03T17:19:00.3203051Z INFO 11-03 17:19:00 [custom_all_reduce.py:37] Skipping P2P check and trusting the driver's P2P report. 2025-11-03T17:19:00.3204065Z WARNING 11-03 17:19:00 [custom_all_reduce.py:162] Custom allreduce is disabled because your platform lacks GPU P2P capability or P2P test failed. To silence this warning, specify disable_custom_all_reduce=True explicitly. 2025-11-03T17:19:00.3205460Z WARNING 11-03 17:19:00 [custom_all_reduce.py:162] Custom allreduce is disabled because your platform lacks GPU P2P capability or P2P test failed. To silence this warning, specify disable_custom_all_reduce=True explicitly. 2025-11-03T17:19:00.3227022Z INFO 11-03 17:19:00 [shm_broadcast.py:302] vLLM message queue communication handle: Handle(local_reader_ranks=[1], buffer_handle=(1, 4194304, 6, 'psm_17e2a154'), local_subscribe_addr='ipc:///tmp/ef52f172-374b-4907-bbd3-18c21382cac2', remote_subscribe_addr=None, remote_addr_ipv6=False) 2025-11-03T17:19:00.3263233Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:19:00.3276770Z [Gloo] Rank 0 is connected to 1 peer ranks. Expected number of connected peer ranks is : 1 2025-11-03T17:19:00.3277921Z [Gloo] Rank 1 is connected to 1 peer ranks. Expected number of connected peer ranks is : 1 2025-11-03T17:19:00.3278964Z INFO 11-03 17:19:00 [__init__.py:1146] Found nccl from library libnccl.so.2 2025-11-03T17:19:00.3279788Z INFO 11-03 17:19:00 [pynccl.py:109] vLLM is using nccl==2.27.5 2025-11-03T17:19:00.3280999Z INFO 11-03 17:19:00 [__init__.py:1146] Found nccl from library libnccl.so.2 2025-11-03T17:19:00.3282419Z INFO 11-03 17:19:00 [pynccl.py:109] vLLM is using nccl==2.27.5 2025-11-03T17:19:00.3627741Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:19:00.3628369Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:19:00.4321683Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:19:00.4339683Z [Gloo] Rank 1 is connected to 1 peer ranks. Expected number of connected peer ranks is : 1 2025-11-03T17:19:00.4340826Z [Gloo] Rank 0 is connected to 1 peer ranks. Expected number of connected peer ranks is : 1 2025-11-03T17:19:00.4341788Z INFO 11-03 17:19:00 [__init__.py:1146] Found nccl from library libnccl.so.2 2025-11-03T17:19:00.4343928Z INFO 11-03 17:19:00 [pynccl.py:109] vLLM is using nccl==2.27.5 2025-11-03T17:19:00.4345740Z INFO 11-03 17:19:00 [__init__.py:1146] Found nccl from library libnccl.so.2 2025-11-03T17:19:00.4346679Z INFO 11-03 17:19:00 [pynccl.py:109] vLLM is using nccl==2.27.5 2025-11-03T17:19:00.4683446Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:19:00.4684595Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:19:00.4695257Z [Gloo] Rank 1 is connected to 1 peer ranks. Expected number of connected peer ranks is : 1 2025-11-03T17:19:00.4696355Z [Gloo] Rank 0 is connected to 1 peer ranks. Expected number of connected peer ranks is : 1 2025-11-03T17:19:00.4697885Z [Gloo] Rank 0 is connected to 1 peer ranks. Expected number of connected peer ranks is : 1 2025-11-03T17:19:00.4699276Z [Gloo] Rank 1 is connected to 1 peer ranks. Expected number of connected peer ranks is : 1 2025-11-03T17:19:00.4700293Z INFO 11-03 17:19:00 [__init__.py:1146] Found nccl from library libnccl.so.2 2025-11-03T17:19:00.4700747Z INFO 11-03 17:19:00 [__init__.py:1146] Found nccl from library libnccl.so.2 2025-11-03T17:19:00.4701285Z INFO 11-03 17:19:00 [pynccl.py:109] vLLM is using nccl==2.27.5 2025-11-03T17:19:00.4701612Z INFO 11-03 17:19:00 [pynccl.py:109] vLLM is using nccl==2.27.5 2025-11-03T17:19:00.4701932Z INFO 11-03 17:19:00 [__init__.py:1146] Found nccl from library libnccl.so.2 2025-11-03T17:19:00.4702255Z INFO 11-03 17:19:00 [pynccl.py:109] vLLM is using nccl==2.27.5 2025-11-03T17:19:00.4702565Z INFO 11-03 17:19:00 [__init__.py:1146] Found nccl from library libnccl.so.2 2025-11-03T17:19:00.4702873Z INFO 11-03 17:19:00 [pynccl.py:109] vLLM is using nccl==2.27.5 2025-11-03T17:19:00.5154476Z INFO 11-03 17:19:00 [parallel_state.py:1231] rank 2 in world size 4 is assigned as DP rank 0, PP rank 1, TP rank 0, EP rank 0 2025-11-03T17:19:00.5155212Z INFO 11-03 17:19:00 [parallel_state.py:1231] rank 3 in world size 4 is assigned as DP rank 0, PP rank 1, TP rank 1, EP rank 1 2025-11-03T17:19:00.5162179Z INFO 11-03 17:19:00 [parallel_state.py:1231] rank 1 in world size 4 is assigned as DP rank 0, PP rank 0, TP rank 1, EP rank 1 2025-11-03T17:19:00.5162912Z INFO 11-03 17:19:00 [parallel_state.py:1231] rank 0 in world size 4 is assigned as DP rank 0, PP rank 0, TP rank 0, EP rank 0 2025-11-03T17:19:00.7425058Z INFO 11-03 17:19:00 [topk_topp_sampler.py:59] Using FlashInfer for top-p & top-k sampling. 2025-11-03T17:19:00.7727757Z INFO 11-03 17:19:00 [topk_topp_sampler.py:59] Using FlashInfer for top-p & top-k sampling. 2025-11-03T17:19:00.7738724Z (Worker_PP0_TP1 pid=27102) INFO 11-03 17:19:00 [gpu_model_runner.py:2840] Starting to load model meta-llama/Llama-3.2-1B-Instruct... 2025-11-03T17:19:00.7749758Z INFO 11-03 17:19:00 [topk_topp_sampler.py:59] Using FlashInfer for top-p & top-k sampling. 2025-11-03T17:19:00.7760050Z INFO 11-03 17:19:00 [topk_topp_sampler.py:59] Using FlashInfer for top-p & top-k sampling. 2025-11-03T17:19:00.8061331Z (Worker_PP0_TP0 pid=27101) INFO 11-03 17:19:00 [gpu_model_runner.py:2840] Starting to load model meta-llama/Llama-3.2-1B-Instruct... 2025-11-03T17:19:00.8075041Z (Worker_PP1_TP0 pid=27103) INFO 11-03 17:19:00 [gpu_model_runner.py:2840] Starting to load model meta-llama/Llama-3.2-1B-Instruct... 2025-11-03T17:19:00.8077098Z (Worker_PP1_TP1 pid=27104) INFO 11-03 17:19:00 [gpu_model_runner.py:2840] Starting to load model meta-llama/Llama-3.2-1B-Instruct... 2025-11-03T17:19:01.0187345Z (Worker_PP0_TP1 pid=27102) INFO 11-03 17:19:01 [gpu_model_runner.py:2870] Loading model from scratch... 2025-11-03T17:19:01.0410574Z (Worker_PP1_TP0 pid=27103) INFO 11-03 17:19:01 [gpu_model_runner.py:2870] Loading model from scratch... 2025-11-03T17:19:01.0453318Z (Worker_PP0_TP0 pid=27101) INFO 11-03 17:19:01 [gpu_model_runner.py:2870] Loading model from scratch... 2025-11-03T17:19:01.0684129Z (Worker_PP1_TP1 pid=27104) INFO 11-03 17:19:01 [gpu_model_runner.py:2870] Loading model from scratch... 2025-11-03T17:19:01.1079573Z (Worker_PP0_TP1 pid=27102) INFO 11-03 17:19:01 [cuda.py:404] Using Flash Attention backend on V1 engine. 2025-11-03T17:19:01.1329247Z (Worker_PP1_TP0 pid=27103) INFO 11-03 17:19:01 [cuda.py:404] Using Flash Attention backend on V1 engine. 2025-11-03T17:19:01.1369322Z (Worker_PP0_TP0 pid=27101) INFO 11-03 17:19:01 [cuda.py:404] Using Flash Attention backend on V1 engine. 2025-11-03T17:19:01.1582355Z (Worker_PP1_TP1 pid=27104) INFO 11-03 17:19:01 [cuda.py:404] Using Flash Attention backend on V1 engine. 2025-11-03T17:19:01.6166886Z (Worker_PP1_TP0 pid=27103) INFO 11-03 17:19:01 [gpu_model_runner.py:2902] Model loading took 0.7135 GiB and 0.156940 seconds 2025-11-03T17:19:01.6309144Z (Worker_PP1_TP1 pid=27104) INFO 11-03 17:19:01 [gpu_model_runner.py:2902] Model loading took 0.7135 GiB and 0.154143 seconds 2025-11-03T17:19:01.6492619Z (Worker_PP0_TP1 pid=27102) INFO 11-03 17:19:01 [gpu_model_runner.py:2902] Model loading took 0.7135 GiB and 0.153886 seconds 2025-11-03T17:19:01.6703970Z (Worker_PP0_TP0 pid=27101) INFO 11-03 17:19:01 [gpu_model_runner.py:2902] Model loading took 0.7135 GiB and 0.156567 seconds 2025-11-03T17:19:03.9364925Z (Worker_PP1_TP0 pid=27103) INFO 11-03 17:19:03 [backends.py:594] Using cache directory: /var/lib/jenkins/.cache/vllm/torch_compile_cache/ba6452aee0/rank_2_0/backbone for vLLM's torch.compile 2025-11-03T17:19:03.9370393Z (Worker_PP1_TP0 pid=27103) INFO 11-03 17:19:03 [backends.py:608] Dynamo bytecode transform time: 2.06 s 2025-11-03T17:19:04.0071018Z (Worker_PP1_TP1 pid=27104) INFO 11-03 17:19:04 [backends.py:594] Using cache directory: /var/lib/jenkins/.cache/vllm/torch_compile_cache/ba6452aee0/rank_3_0/backbone for vLLM's torch.compile 2025-11-03T17:19:04.0078632Z (Worker_PP1_TP1 pid=27104) INFO 11-03 17:19:04 [backends.py:608] Dynamo bytecode transform time: 2.13 s 2025-11-03T17:19:04.1130206Z (Worker_PP0_TP1 pid=27102) INFO 11-03 17:19:04 [backends.py:594] Using cache directory: /var/lib/jenkins/.cache/vllm/torch_compile_cache/3ff2b0020c/rank_1_0/backbone for vLLM's torch.compile 2025-11-03T17:19:04.1137084Z (Worker_PP0_TP1 pid=27102) INFO 11-03 17:19:04 [backends.py:608] Dynamo bytecode transform time: 2.21 s 2025-11-03T17:19:04.1442058Z (Worker_PP0_TP0 pid=27101) INFO 11-03 17:19:04 [backends.py:594] Using cache directory: /var/lib/jenkins/.cache/vllm/torch_compile_cache/3ff2b0020c/rank_0_0/backbone for vLLM's torch.compile 2025-11-03T17:19:04.1449403Z (Worker_PP0_TP0 pid=27101) INFO 11-03 17:19:04 [backends.py:608] Dynamo bytecode transform time: 2.23 s 2025-11-03T17:19:04.5423911Z (Worker_PP1_TP0 pid=27103) INFO 11-03 17:19:04 [backends.py:179] Directly load the compiled graph(s) for dynamic shape from the cache, took 0.404 s 2025-11-03T17:19:04.5914314Z (Worker_PP1_TP1 pid=27104) INFO 11-03 17:19:04 [backends.py:179] Directly load the compiled graph(s) for dynamic shape from the cache, took 0.384 s 2025-11-03T17:19:04.7475711Z (Worker_PP0_TP1 pid=27102) INFO 11-03 17:19:04 [backends.py:179] Directly load the compiled graph(s) for dynamic shape from the cache, took 0.418 s 2025-11-03T17:19:04.7896397Z (Worker_PP0_TP0 pid=27101) INFO 11-03 17:19:04 [backends.py:179] Directly load the compiled graph(s) for dynamic shape from the cache, took 0.425 s 2025-11-03T17:19:05.5334210Z (Worker_PP1_TP0 pid=27103) INFO 11-03 17:19:05 [gpu_worker.py:314] Available KV cache memory: 18.99 GiB 2025-11-03T17:19:05.6161912Z (Worker_PP1_TP1 pid=27104) INFO 11-03 17:19:05 [gpu_worker.py:314] Available KV cache memory: 18.99 GiB 2025-11-03T17:19:05.6702260Z (Worker_PP0_TP1 pid=27102) INFO 11-03 17:19:05 [gpu_worker.py:314] Available KV cache memory: 19.02 GiB 2025-11-03T17:19:05.7506103Z (Worker_PP0_TP0 pid=27101) INFO 11-03 17:19:05 [gpu_worker.py:314] Available KV cache memory: 19.02 GiB 2025-11-03T17:19:05.9975586Z (EngineCore_DP0 pid=26982) INFO 11-03 17:19:05 [kv_cache_utils.py:1199] GPU KV cache size: 2,489,088 tokens 2025-11-03T17:19:05.9976514Z (EngineCore_DP0 pid=26982) INFO 11-03 17:19:05 [kv_cache_utils.py:1204] Maximum concurrency for 2,048 tokens per request: 1215.38x 2025-11-03T17:19:05.9977360Z (EngineCore_DP0 pid=26982) INFO 11-03 17:19:05 [kv_cache_utils.py:1199] GPU KV cache size: 2,489,088 tokens 2025-11-03T17:19:05.9978174Z (EngineCore_DP0 pid=26982) INFO 11-03 17:19:05 [kv_cache_utils.py:1204] Maximum concurrency for 2,048 tokens per request: 1215.38x 2025-11-03T17:19:05.9978980Z (EngineCore_DP0 pid=26982) INFO 11-03 17:19:05 [kv_cache_utils.py:1199] GPU KV cache size: 2,489,088 tokens 2025-11-03T17:19:05.9979784Z (EngineCore_DP0 pid=26982) INFO 11-03 17:19:05 [kv_cache_utils.py:1204] Maximum concurrency for 2,048 tokens per request: 1215.38x 2025-11-03T17:19:05.9980611Z (EngineCore_DP0 pid=26982) INFO 11-03 17:19:05 [kv_cache_utils.py:1199] GPU KV cache size: 2,489,088 tokens 2025-11-03T17:19:05.9981703Z (EngineCore_DP0 pid=26982) INFO 11-03 17:19:05 [kv_cache_utils.py:1204] Maximum concurrency for 2,048 tokens per request: 1215.38x 2025-11-03T17:19:06.2707069Z (Worker_PP0_TP0 pid=27101) 2025-11-03T17:19:07.3700261Z Capturing CUDA graphs (mixed prefill-decode, PIECEWISE): 0% 0/4 [00:00, 'use_cudagraph': True, 'cudagraph_num_of_warmups': 1, 'cudagraph_capture_sizes': [16, 8, 4, 2, 1], 'cudagraph_copy_inputs': False, 'full_cuda_graph': True, 'use_inductor_graph_partition': False, 'pass_config': {}, 'max_capture_size': 16, 'local_cache_dir': None} 2025-11-03T17:19:33.4866839Z (EngineCore_DP0 pid=27771) WARNING 11-03 17:19:33 [multiproc_executor.py:744] Reducing Torch parallelism from 24 threads to 1 to avoid unnecessary CPU contention. Set OMP_NUM_THREADS in the external environment to tune this value as needed. 2025-11-03T17:19:33.4869115Z (EngineCore_DP0 pid=27771) INFO 11-03 17:19:33 [shm_broadcast.py:302] vLLM message queue communication handle: Handle(local_reader_ranks=[0, 1], buffer_handle=(2, 16777216, 10, 'psm_60be7742'), local_subscribe_addr='ipc:///tmp/07a5844d-eb50-4aff-97cd-a9cf64fc4b06', remote_subscribe_addr=None, remote_addr_ipv6=False) 2025-11-03T17:19:33.9799956Z /opt/conda/envs/py_3.12/lib/python3.12/site-packages/torch/cuda/__init__.py:65: FutureWarning: The pynvml package is deprecated. Please install nvidia-ml-py instead. If you did not install pynvml directly, please report this to the maintainers of the package that installed pynvml for you. 2025-11-03T17:19:33.9801854Z import pynvml # type: ignore[import] 2025-11-03T17:19:33.9803481Z /opt/conda/envs/py_3.12/lib/python3.12/site-packages/torch/cuda/__init__.py:65: FutureWarning: The pynvml package is deprecated. Please install nvidia-ml-py instead. If you did not install pynvml directly, please report this to the maintainers of the package that installed pynvml for you. 2025-11-03T17:19:33.9805098Z import pynvml # type: ignore[import] 2025-11-03T17:19:36.3394173Z INFO 11-03 17:19:36 [__init__.py:224] Automatically detected platform cuda. 2025-11-03T17:19:36.3436120Z INFO 11-03 17:19:36 [__init__.py:224] Automatically detected platform cuda. 2025-11-03T17:19:41.1329217Z INFO 11-03 17:19:41 [shm_broadcast.py:302] vLLM message queue communication handle: Handle(local_reader_ranks=[0], buffer_handle=(1, 10485760, 10, 'psm_38ace9c5'), local_subscribe_addr='ipc:///tmp/372f9a37-def0-4c32-aa96-8997e896597a', remote_subscribe_addr=None, remote_addr_ipv6=False) 2025-11-03T17:19:41.1724002Z INFO 11-03 17:19:41 [shm_broadcast.py:302] vLLM message queue communication handle: Handle(local_reader_ranks=[0], buffer_handle=(1, 10485760, 10, 'psm_5c931580'), local_subscribe_addr='ipc:///tmp/ff2229ed-eafb-44a7-ab13-cdfed18fbf87', remote_subscribe_addr=None, remote_addr_ipv6=False) 2025-11-03T17:19:41.3483611Z [Gloo] Rank 1 is connected to 1 peer ranks. Expected number of connected peer ranks is : 1 2025-11-03T17:19:41.3484516Z [Gloo] Rank 0 is connected to 1 peer ranks. Expected number of connected peer ranks is : 1 2025-11-03T17:19:41.3528558Z [Gloo] Rank 0 is connected to 1 peer ranks. Expected number of connected peer ranks is : 1 2025-11-03T17:19:41.3529545Z [Gloo] Rank 1 is connected to 1 peer ranks. Expected number of connected peer ranks is : 1 2025-11-03T17:19:41.3571089Z INFO 11-03 17:19:41 [__init__.py:1146] Found nccl from library libnccl.so.2 2025-11-03T17:19:41.3571887Z INFO 11-03 17:19:41 [__init__.py:1146] Found nccl from library libnccl.so.2 2025-11-03T17:19:41.3573063Z INFO 11-03 17:19:41 [pynccl.py:109] vLLM is using nccl==2.27.5 2025-11-03T17:19:41.3574151Z INFO 11-03 17:19:41 [pynccl.py:109] vLLM is using nccl==2.27.5 2025-11-03T17:19:41.5311411Z WARNING 11-03 17:19:41 [symm_mem.py:60] SymmMemCommunicator: Device capability 8.9 not supported, communicator is not available. 2025-11-03T17:19:41.5312559Z WARNING 11-03 17:19:41 [symm_mem.py:60] SymmMemCommunicator: Device capability 8.9 not supported, communicator is not available. 2025-11-03T17:19:41.5334646Z INFO 11-03 17:19:41 [custom_all_reduce.py:37] Skipping P2P check and trusting the driver's P2P report. 2025-11-03T17:19:41.5335702Z INFO 11-03 17:19:41 [custom_all_reduce.py:37] Skipping P2P check and trusting the driver's P2P report. 2025-11-03T17:19:41.5336708Z WARNING 11-03 17:19:41 [custom_all_reduce.py:162] Custom allreduce is disabled because your platform lacks GPU P2P capability or P2P test failed. To silence this warning, specify disable_custom_all_reduce=True explicitly. 2025-11-03T17:19:41.5338139Z WARNING 11-03 17:19:41 [custom_all_reduce.py:162] Custom allreduce is disabled because your platform lacks GPU P2P capability or P2P test failed. To silence this warning, specify disable_custom_all_reduce=True explicitly. 2025-11-03T17:19:41.6124388Z INFO 11-03 17:19:41 [shm_broadcast.py:302] vLLM message queue communication handle: Handle(local_reader_ranks=[1], buffer_handle=(1, 4194304, 6, 'psm_f93a9677'), local_subscribe_addr='ipc:///tmp/80485ba0-3cc9-42c3-8692-96cf22cc820e', remote_subscribe_addr=None, remote_addr_ipv6=False) 2025-11-03T17:19:41.6153516Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:19:41.6154512Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:19:41.6161303Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:19:41.6169323Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:19:41.6922486Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:19:41.6929965Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:19:41.6941430Z [Gloo] Rank 0 is connected to 1 peer ranks. Expected number of connected peer ranks is : 1 2025-11-03T17:19:41.6942412Z [Gloo] Rank 1 is connected to 1 peer ranks. Expected number of connected peer ranks is : 1 2025-11-03T17:19:41.6943857Z INFO 11-03 17:19:41 [__init__.py:1146] Found nccl from library libnccl.so.2 2025-11-03T17:19:41.6944601Z INFO 11-03 17:19:41 [__init__.py:1146] Found nccl from library libnccl.so.2 2025-11-03T17:19:41.6945281Z INFO 11-03 17:19:41 [pynccl.py:109] vLLM is using nccl==2.27.5 2025-11-03T17:19:41.6945684Z INFO 11-03 17:19:41 [pynccl.py:109] vLLM is using nccl==2.27.5 2025-11-03T17:19:41.7276297Z INFO 11-03 17:19:41 [parallel_state.py:1231] rank 0 in world size 2 is assigned as DP rank 0, PP rank 0, TP rank 0, EP rank 0 2025-11-03T17:19:41.7277425Z INFO 11-03 17:19:41 [parallel_state.py:1231] rank 1 in world size 2 is assigned as DP rank 0, PP rank 0, TP rank 1, EP rank 1 2025-11-03T17:19:41.9556671Z INFO 11-03 17:19:41 [topk_topp_sampler.py:59] Using FlashInfer for top-p & top-k sampling. 2025-11-03T17:19:41.9577442Z INFO 11-03 17:19:41 [topk_topp_sampler.py:59] Using FlashInfer for top-p & top-k sampling. 2025-11-03T17:19:41.9874991Z (Worker_TP0 pid=27890) INFO 11-03 17:19:41 [gpu_model_runner.py:2840] Starting to load model meta-llama/Llama-3.2-1B-Instruct... 2025-11-03T17:19:41.9893811Z (Worker_TP1 pid=27891) INFO 11-03 17:19:41 [gpu_model_runner.py:2840] Starting to load model meta-llama/Llama-3.2-1B-Instruct... 2025-11-03T17:19:42.1906593Z (Worker_TP0 pid=27890) INFO 11-03 17:19:42 [gpu_model_runner.py:2870] Loading model from scratch... 2025-11-03T17:19:42.1908112Z (Worker_TP1 pid=27891) INFO 11-03 17:19:42 [gpu_model_runner.py:2870] Loading model from scratch... 2025-11-03T17:19:42.2817677Z (Worker_TP0 pid=27890) INFO 11-03 17:19:42 [cuda.py:404] Using Flash Attention backend on V1 engine. 2025-11-03T17:19:42.2819260Z (Worker_TP1 pid=27891) INFO 11-03 17:19:42 [cuda.py:404] Using Flash Attention backend on V1 engine. 2025-11-03T17:19:42.7646263Z (Worker_TP0 pid=27890) INFO 11-03 17:19:42 [gpu_model_runner.py:2902] Model loading took 1.1667 GiB and 0.167299 seconds 2025-11-03T17:19:42.7662335Z (Worker_TP1 pid=27891) INFO 11-03 17:19:42 [gpu_model_runner.py:2902] Model loading took 1.1667 GiB and 0.169222 seconds 2025-11-03T17:19:46.4998462Z (Worker_TP1 pid=27891) INFO 11-03 17:19:46 [backends.py:594] Using cache directory: /var/lib/jenkins/.cache/vllm/torch_compile_cache/cdd8585f72/rank_1_0/backbone for vLLM's torch.compile 2025-11-03T17:19:46.5002632Z (Worker_TP1 pid=27891) INFO 11-03 17:19:46 [backends.py:608] Dynamo bytecode transform time: 3.48 s 2025-11-03T17:19:46.5720336Z (Worker_TP0 pid=27890) INFO 11-03 17:19:46 [backends.py:594] Using cache directory: /var/lib/jenkins/.cache/vllm/torch_compile_cache/cdd8585f72/rank_0_0/backbone for vLLM's torch.compile 2025-11-03T17:19:46.5726868Z (Worker_TP0 pid=27890) INFO 11-03 17:19:46 [backends.py:608] Dynamo bytecode transform time: 3.55 s 2025-11-03T17:19:47.5896922Z (Worker_TP1 pid=27891) INFO 11-03 17:19:47 [backends.py:179] Directly load the compiled graph(s) for dynamic shape from the cache, took 0.900 s 2025-11-03T17:19:47.6646878Z (Worker_TP0 pid=27890) INFO 11-03 17:19:47 [backends.py:179] Directly load the compiled graph(s) for dynamic shape from the cache, took 0.900 s 2025-11-03T17:19:48.2145223Z (Worker_TP1 pid=27891) INFO 11-03 17:19:48 [monitor.py:33] torch.compile takes 3.48 s in total 2025-11-03T17:19:48.2146201Z (Worker_TP0 pid=27890) INFO 11-03 17:19:48 [monitor.py:33] torch.compile takes 3.55 s in total 2025-11-03T17:19:48.8584114Z (Worker_TP0 pid=27890) INFO 11-03 17:19:48 [gpu_worker.py:314] Available KV cache memory: 18.54 GiB 2025-11-03T17:19:48.8602291Z (Worker_TP1 pid=27891) INFO 11-03 17:19:48 [gpu_worker.py:314] Available KV cache memory: 18.54 GiB 2025-11-03T17:19:49.0974385Z (EngineCore_DP0 pid=27771) INFO 11-03 17:19:49 [kv_cache_utils.py:1199] GPU KV cache size: 1,214,720 tokens 2025-11-03T17:19:49.0975448Z (EngineCore_DP0 pid=27771) INFO 11-03 17:19:49 [kv_cache_utils.py:1204] Maximum concurrency for 2,048 tokens per request: 593.12x 2025-11-03T17:19:49.0976303Z (EngineCore_DP0 pid=27771) INFO 11-03 17:19:49 [kv_cache_utils.py:1199] GPU KV cache size: 1,214,720 tokens 2025-11-03T17:19:49.0977142Z (EngineCore_DP0 pid=27771) INFO 11-03 17:19:49 [kv_cache_utils.py:1204] Maximum concurrency for 2,048 tokens per request: 593.12x 2025-11-03T17:19:49.3416406Z (Worker_TP0 pid=27890) 2025-11-03T17:19:49.4645064Z Capturing CUDA graphs (mixed prefill-decode, PIECEWISE): 0% 0/5 [00:00, 'use_cudagraph': True, 'cudagraph_num_of_warmups': 1, 'cudagraph_capture_sizes': [16, 8, 4, 2], 'cudagraph_copy_inputs': False, 'full_cuda_graph': True, 'use_inductor_graph_partition': False, 'pass_config': {'enable_noop': True, 'enable_sequence_parallelism': True}, 'max_capture_size': 16, 'local_cache_dir': None} 2025-11-03T17:20:10.9836414Z (EngineCore_DP0 pid=28267) 2025-11-03 17:20:10,982 INFO worker.py:1918 -- Started a local Ray instance. View the dashboard at http://127.0.0.1:8265  2025-11-03T17:20:12.0678147Z (EngineCore_DP0 pid=28267) INFO 11-03 17:20:12 [ray_utils.py:373] No current placement group found. Creating a new placement group. 2025-11-03T17:20:12.2432087Z (EngineCore_DP0 pid=28267) INFO 11-03 17:20:12 [ray_distributed_executor.py:179] use_ray_spmd_worker: True 2025-11-03T17:20:12.8168082Z (EngineCore_DP0 pid=28267) (pid=29157) /opt/conda/envs/py_3.12/lib/python3.12/site-packages/torch/cuda/__init__.py:65: FutureWarning: The pynvml package is deprecated. Please install nvidia-ml-py instead. If you did not install pynvml directly, please report this to the maintainers of the package that installed pynvml for you. 2025-11-03T17:20:12.8170578Z (EngineCore_DP0 pid=28267) (pid=29157) import pynvml # type: ignore[import] 2025-11-03T17:20:15.1681299Z (EngineCore_DP0 pid=28267) (pid=29157) INFO 11-03 17:20:15 [__init__.py:224] Automatically detected platform cuda. 2025-11-03T17:20:16.7516509Z (EngineCore_DP0 pid=28267) INFO 11-03 17:20:16 [ray_env.py:67] RAY_NON_CARRY_OVER_ENV_VARS from config: set() 2025-11-03T17:20:16.7518061Z (EngineCore_DP0 pid=28267) INFO 11-03 17:20:16 [ray_env.py:70] Copying the following environment variables to workers: ['VLLM_USE_V1', 'VLLM_WORKER_MULTIPROC_METHOD', 'MAX_JOBS', 'HUGGING_FACE_HUB_TOKEN', 'VLLM_USE_RAY_COMPILED_DAG', 'LD_LIBRARY_PATH', 'HF_TOKEN', 'VLLM_USE_RAY_SPMD_WORKER'] 2025-11-03T17:20:16.7519635Z (EngineCore_DP0 pid=28267) INFO 11-03 17:20:16 [ray_env.py:75] If certain env vars should NOT be copied, add them to /var/lib/jenkins/.config/vllm/ray_non_carry_over_env_vars.json file 2025-11-03T17:20:20.0704732Z (EngineCore_DP0 pid=28267) (RayWorkerWrapper pid=29165) WARNING 11-03 17:20:20 [worker_base.py:307] Missing `shared_worker_lock` argument from executor. This argument is needed for mm_processor_cache_type='shm'. 2025-11-03T17:20:20.1741624Z (EngineCore_DP0 pid=28267) (pid=29156) INFO 11-03 17:20:15 [__init__.py:224] Automatically detected platform cuda. [repeated 3x across cluster] (Ray deduplicates logs by default. Set RAY_DEDUP_LOGS=0 to disable log deduplication, or see https://docs.ray.io/en/master/ray-observability/user-guides/configure-logging.html#log-deduplication for more options.) 2025-11-03T17:20:21.1981232Z (EngineCore_DP0 pid=28267) (RayWorkerWrapper pid=29157) [Gloo] Rank 0 is connected to 3 peer ranks. Expected number of connected peer ranks is : 3 2025-11-03T17:20:21.4060588Z (EngineCore_DP0 pid=28267) (RayWorkerWrapper pid=29157) INFO 11-03 17:20:21 [__init__.py:1146] Found nccl from library libnccl.so.2 2025-11-03T17:20:21.4061602Z (EngineCore_DP0 pid=28267) (RayWorkerWrapper pid=29157) INFO 11-03 17:20:21 [pynccl.py:109] vLLM is using nccl==2.27.5 2025-11-03T17:20:21.7150340Z (EngineCore_DP0 pid=28267) (RayWorkerWrapper pid=29157) WARNING 11-03 17:20:21 [symm_mem.py:60] SymmMemCommunicator: Device capability 8.9 not supported, communicator is not available. 2025-11-03T17:20:21.7151714Z (EngineCore_DP0 pid=28267) (RayWorkerWrapper pid=29157) INFO 11-03 17:20:21 [custom_all_reduce.py:37] Skipping P2P check and trusting the driver's P2P report. 2025-11-03T17:20:21.7153340Z (EngineCore_DP0 pid=28267) (RayWorkerWrapper pid=29157) WARNING 11-03 17:20:21 [custom_all_reduce.py:162] Custom allreduce is disabled because your platform lacks GPU P2P capability or P2P test failed. To silence this warning, specify disable_custom_all_reduce=True explicitly. 2025-11-03T17:20:21.7155897Z (EngineCore_DP0 pid=28267) (RayWorkerWrapper pid=29157) INFO 11-03 17:20:21 [shm_broadcast.py:302] vLLM message queue communication handle: Handle(local_reader_ranks=[1], buffer_handle=(1, 4194304, 6, 'psm_6482b242'), local_subscribe_addr='ipc:///tmp/242a81c5-b64e-4b79-8569-4dac266f6a2d', remote_subscribe_addr=None, remote_addr_ipv6=False) 2025-11-03T17:20:21.8203056Z (EngineCore_DP0 pid=28267) (RayWorkerWrapper pid=29157) INFO 11-03 17:20:21 [parallel_state.py:1231] rank 0 in world size 4 is assigned as DP rank 0, PP rank 0, TP rank 0, EP rank 0 2025-11-03T17:20:22.0274273Z (EngineCore_DP0 pid=28267) (RayWorkerWrapper pid=29157) INFO 11-03 17:20:21 [topk_topp_sampler.py:59] Using FlashInfer for top-p & top-k sampling. 2025-11-03T17:20:22.0275550Z (EngineCore_DP0 pid=28267) (RayWorkerWrapper pid=29157) INFO 11-03 17:20:22 [gpu_model_runner.py:2840] Starting to load model meta-llama/Llama-3.2-1B-Instruct... 2025-11-03T17:20:22.3367328Z (EngineCore_DP0 pid=28267) (RayWorkerWrapper pid=29157) INFO 11-03 17:20:22 [gpu_model_runner.py:2870] Loading model from scratch... 2025-11-03T17:20:22.4417029Z (EngineCore_DP0 pid=28267) (RayWorkerWrapper pid=29157) INFO 11-03 17:20:22 [cuda.py:404] Using Flash Attention backend on V1 engine. 2025-11-03T17:20:22.8531556Z (EngineCore_DP0 pid=28267) (RayWorkerWrapper pid=29157) INFO 11-03 17:20:22 [gpu_model_runner.py:2902] Model loading took 0.7135 GiB and 0.157190 seconds 2025-11-03T17:20:25.4076054Z (EngineCore_DP0 pid=28267) (RayWorkerWrapper pid=29157) INFO 11-03 17:20:25 [backends.py:594] Using cache directory: /var/lib/jenkins/.cache/vllm/torch_compile_cache/3ff2b0020c/rank_0_0/backbone for vLLM's torch.compile 2025-11-03T17:20:25.4077489Z (EngineCore_DP0 pid=28267) (RayWorkerWrapper pid=29157) INFO 11-03 17:20:25 [backends.py:608] Dynamo bytecode transform time: 2.22 s 2025-11-03T17:20:25.4078942Z (EngineCore_DP0 pid=28267) (RayWorkerWrapper pid=29151) WARNING 11-03 17:20:20 [worker_base.py:307] Missing `shared_worker_lock` argument from executor. This argument is needed for mm_processor_cache_type='shm'. [repeated 3x across cluster] 2025-11-03T17:20:26.0240261Z (EngineCore_DP0 pid=28267) (RayWorkerWrapper pid=29157) INFO 11-03 17:20:26 [backends.py:179] Directly load the compiled graph(s) for dynamic shape from the cache, took 0.421 s 2025-11-03T17:20:27.0487021Z (EngineCore_DP0 pid=28267) (RayWorkerWrapper pid=29157) INFO 11-03 17:20:26 [gpu_worker.py:314] Available KV cache memory: 19.02 GiB 2025-11-03T17:20:27.0488326Z (EngineCore_DP0 pid=28267) (RayWorkerWrapper pid=29151) [Gloo] Rank 1 is connected to 1 peer ranks. Expected number of connected peer ranks is : 1 [repeated 23x across cluster] 2025-11-03T17:20:27.0489740Z (EngineCore_DP0 pid=28267) (RayWorkerWrapper pid=29151) INFO 11-03 17:20:21 [__init__.py:1146] Found nccl from library libnccl.so.2 [repeated 11x across cluster] 2025-11-03T17:20:27.0490888Z (EngineCore_DP0 pid=28267) (RayWorkerWrapper pid=29151) INFO 11-03 17:20:21 [pynccl.py:109] vLLM is using nccl==2.27.5 [repeated 11x across cluster] 2025-11-03T17:20:27.0492311Z (EngineCore_DP0 pid=28267) (RayWorkerWrapper pid=29151) WARNING 11-03 17:20:21 [symm_mem.py:60] SymmMemCommunicator: Device capability 8.9 not supported, communicator is not available. [repeated 3x across cluster] 2025-11-03T17:20:27.0493830Z (EngineCore_DP0 pid=28267) (RayWorkerWrapper pid=29151) INFO 11-03 17:20:21 [custom_all_reduce.py:37] Skipping P2P check and trusting the driver's P2P report. [repeated 3x across cluster] 2025-11-03T17:20:27.0495466Z (EngineCore_DP0 pid=28267) (RayWorkerWrapper pid=29151) WARNING 11-03 17:20:21 [custom_all_reduce.py:162] Custom allreduce is disabled because your platform lacks GPU P2P capability or P2P test failed. To silence this warning, specify disable_custom_all_reduce=True explicitly. [repeated 3x across cluster] 2025-11-03T17:20:27.0497755Z (EngineCore_DP0 pid=28267) (RayWorkerWrapper pid=29156) INFO 11-03 17:20:21 [shm_broadcast.py:302] vLLM message queue communication handle: Handle(local_reader_ranks=[1], buffer_handle=(1, 4194304, 6, 'psm_6293f7a7'), local_subscribe_addr='ipc:///tmp/dcf7a043-a72c-489b-91c8-48213d748d0e', remote_subscribe_addr=None, remote_addr_ipv6=False) 2025-11-03T17:20:27.0499283Z (EngineCore_DP0 pid=28267) (RayWorkerWrapper pid=29151) INFO 11-03 17:20:21 [parallel_state.py:1231] rank 3 in world size 4 is assigned as DP rank 0, PP rank 1, TP rank 1, EP rank 1 [repeated 3x across cluster] 2025-11-03T17:20:27.0500365Z (EngineCore_DP0 pid=28267) (RayWorkerWrapper pid=29151) INFO 11-03 17:20:21 [topk_topp_sampler.py:59] Using FlashInfer for top-p & top-k sampling. [repeated 3x across cluster] 2025-11-03T17:20:27.0501459Z (EngineCore_DP0 pid=28267) (RayWorkerWrapper pid=29151) INFO 11-03 17:20:22 [gpu_model_runner.py:2840] Starting to load model meta-llama/Llama-3.2-1B-Instruct... [repeated 3x across cluster] 2025-11-03T17:20:27.3320793Z (EngineCore_DP0 pid=28267) INFO 11-03 17:20:27 [kv_cache_utils.py:1199] GPU KV cache size: 2,489,088 tokens 2025-11-03T17:20:27.3322245Z (EngineCore_DP0 pid=28267) INFO 11-03 17:20:27 [kv_cache_utils.py:1204] Maximum concurrency for 2,048 tokens per request: 1215.38x 2025-11-03T17:20:27.3323502Z (EngineCore_DP0 pid=28267) INFO 11-03 17:20:27 [kv_cache_utils.py:1199] GPU KV cache size: 2,489,088 tokens 2025-11-03T17:20:27.3324819Z (EngineCore_DP0 pid=28267) INFO 11-03 17:20:27 [kv_cache_utils.py:1204] Maximum concurrency for 2,048 tokens per request: 1215.38x 2025-11-03T17:20:27.3325727Z (EngineCore_DP0 pid=28267) INFO 11-03 17:20:27 [kv_cache_utils.py:1199] GPU KV cache size: 2,489,088 tokens 2025-11-03T17:20:27.3326415Z (EngineCore_DP0 pid=28267) INFO 11-03 17:20:27 [kv_cache_utils.py:1204] Maximum concurrency for 2,048 tokens per request: 1215.38x 2025-11-03T17:20:27.3327254Z (EngineCore_DP0 pid=28267) INFO 11-03 17:20:27 [kv_cache_utils.py:1199] GPU KV cache size: 2,489,088 tokens 2025-11-03T17:20:27.3327934Z (EngineCore_DP0 pid=28267) INFO 11-03 17:20:27 [kv_cache_utils.py:1204] Maximum concurrency for 2,048 tokens per request: 1215.38x 2025-11-03T17:20:27.6634792Z (EngineCore_DP0 pid=28267) (RayWorkerWrapper pid=29157) 2025-11-03T17:20:27.6635456Z Capturing CUDA graphs (mixed prefill-decode, PIECEWISE): 0%| | 0/4 [00:00, 'use_cudagraph': True, 'cudagraph_num_of_warmups': 1, 'cudagraph_capture_sizes': [16, 8, 4, 2, 1], 'cudagraph_copy_inputs': False, 'full_cuda_graph': True, 'use_inductor_graph_partition': False, 'pass_config': {}, 'max_capture_size': 16, 'local_cache_dir': None} 2025-11-03T17:20:49.8664420Z (EngineCore_DP0 pid=33289) WARNING 11-03 17:20:49 [multiproc_executor.py:744] Reducing Torch parallelism from 24 threads to 1 to avoid unnecessary CPU contention. Set OMP_NUM_THREADS in the external environment to tune this value as needed. 2025-11-03T17:20:49.8666536Z (EngineCore_DP0 pid=33289) INFO 11-03 17:20:49 [shm_broadcast.py:302] vLLM message queue communication handle: Handle(local_reader_ranks=[0, 1], buffer_handle=(2, 16777216, 10, 'psm_379874fb'), local_subscribe_addr='ipc:///tmp/0075fda2-a928-4cdb-989a-10b5faa0bb06', remote_subscribe_addr=None, remote_addr_ipv6=False) 2025-11-03T17:20:50.3730166Z /opt/conda/envs/py_3.12/lib/python3.12/site-packages/torch/cuda/__init__.py:65: FutureWarning: The pynvml package is deprecated. Please install nvidia-ml-py instead. If you did not install pynvml directly, please report this to the maintainers of the package that installed pynvml for you. 2025-11-03T17:20:50.3731398Z import pynvml # type: ignore[import] 2025-11-03T17:20:50.3845127Z /opt/conda/envs/py_3.12/lib/python3.12/site-packages/torch/cuda/__init__.py:65: FutureWarning: The pynvml package is deprecated. Please install nvidia-ml-py instead. If you did not install pynvml directly, please report this to the maintainers of the package that installed pynvml for you. 2025-11-03T17:20:50.3846689Z import pynvml # type: ignore[import] 2025-11-03T17:20:52.7447987Z INFO 11-03 17:20:52 [__init__.py:224] Automatically detected platform cuda. 2025-11-03T17:20:52.7487929Z INFO 11-03 17:20:52 [__init__.py:224] Automatically detected platform cuda. 2025-11-03T17:20:57.5430526Z INFO 11-03 17:20:57 [shm_broadcast.py:302] vLLM message queue communication handle: Handle(local_reader_ranks=[0], buffer_handle=(1, 10485760, 10, 'psm_b8e6d862'), local_subscribe_addr='ipc:///tmp/f72d43cf-0175-46fb-945d-7a2676d3d57f', remote_subscribe_addr=None, remote_addr_ipv6=False) 2025-11-03T17:20:57.5432326Z INFO 11-03 17:20:57 [shm_broadcast.py:302] vLLM message queue communication handle: Handle(local_reader_ranks=[0], buffer_handle=(1, 10485760, 10, 'psm_da7e4452'), local_subscribe_addr='ipc:///tmp/87824656-6e8a-4b2e-bec9-9ebe795ea38d', remote_subscribe_addr=None, remote_addr_ipv6=False) 2025-11-03T17:20:58.4208693Z [Gloo] Rank 1 is connected to 1 peer ranks. Expected number of connected peer ranks is : 1 2025-11-03T17:20:58.4210226Z [Gloo] Rank 0 is connected to 1 peer ranks. Expected number of connected peer ranks is : 1 2025-11-03T17:20:58.4260707Z [Gloo] Rank 1 is connected to 1 peer ranks. Expected number of connected peer ranks is : 1 2025-11-03T17:20:58.4261881Z [Gloo] Rank 0 is connected to 1 peer ranks. Expected number of connected peer ranks is : 1 2025-11-03T17:20:58.4303977Z INFO 11-03 17:20:58 [__init__.py:1146] Found nccl from library libnccl.so.2 2025-11-03T17:20:58.4305138Z INFO 11-03 17:20:58 [__init__.py:1146] Found nccl from library libnccl.so.2 2025-11-03T17:20:58.4306259Z INFO 11-03 17:20:58 [pynccl.py:109] vLLM is using nccl==2.27.5 2025-11-03T17:20:58.4306954Z INFO 11-03 17:20:58 [pynccl.py:109] vLLM is using nccl==2.27.5 2025-11-03T17:20:58.6041662Z WARNING 11-03 17:20:58 [symm_mem.py:60] SymmMemCommunicator: Device capability 8.9 not supported, communicator is not available. 2025-11-03T17:20:58.6043011Z WARNING 11-03 17:20:58 [symm_mem.py:60] SymmMemCommunicator: Device capability 8.9 not supported, communicator is not available. 2025-11-03T17:20:58.6064856Z INFO 11-03 17:20:58 [custom_all_reduce.py:37] Skipping P2P check and trusting the driver's P2P report. 2025-11-03T17:20:58.6066161Z INFO 11-03 17:20:58 [custom_all_reduce.py:37] Skipping P2P check and trusting the driver's P2P report. 2025-11-03T17:20:58.6068085Z WARNING 11-03 17:20:58 [custom_all_reduce.py:162] Custom allreduce is disabled because your platform lacks GPU P2P capability or P2P test failed. To silence this warning, specify disable_custom_all_reduce=True explicitly. 2025-11-03T17:20:58.6070686Z WARNING 11-03 17:20:58 [custom_all_reduce.py:162] Custom allreduce is disabled because your platform lacks GPU P2P capability or P2P test failed. To silence this warning, specify disable_custom_all_reduce=True explicitly. 2025-11-03T17:20:58.6725003Z INFO 11-03 17:20:58 [shm_broadcast.py:302] vLLM message queue communication handle: Handle(local_reader_ranks=[1], buffer_handle=(1, 4194304, 6, 'psm_9048e111'), local_subscribe_addr='ipc:///tmp/d89bfedc-9a6a-4874-84c8-aab027a073d8', remote_subscribe_addr=None, remote_addr_ipv6=False) 2025-11-03T17:20:58.6752018Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:20:58.6756433Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:20:58.6760692Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:20:58.6768874Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:20:58.7320949Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:20:58.7332280Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:20:58.7347009Z [Gloo] Rank 1 is connected to 1 peer ranks. Expected number of connected peer ranks is : 1 2025-11-03T17:20:58.7348404Z [Gloo] Rank 0 is connected to 1 peer ranks. Expected number of connected peer ranks is : 1 2025-11-03T17:20:58.7350503Z INFO 11-03 17:20:58 [__init__.py:1146] Found nccl from library libnccl.so.2 2025-11-03T17:20:58.7351595Z INFO 11-03 17:20:58 [pynccl.py:109] vLLM is using nccl==2.27.5 2025-11-03T17:20:58.7352443Z INFO 11-03 17:20:58 [__init__.py:1146] Found nccl from library libnccl.so.2 2025-11-03T17:20:58.7353019Z INFO 11-03 17:20:58 [pynccl.py:109] vLLM is using nccl==2.27.5 2025-11-03T17:20:58.7687015Z INFO 11-03 17:20:58 [parallel_state.py:1231] rank 1 in world size 2 is assigned as DP rank 0, PP rank 0, TP rank 1, EP rank 1 2025-11-03T17:20:58.7687772Z INFO 11-03 17:20:58 [parallel_state.py:1231] rank 0 in world size 2 is assigned as DP rank 0, PP rank 0, TP rank 0, EP rank 0 2025-11-03T17:20:59.0089831Z INFO 11-03 17:20:59 [topk_topp_sampler.py:59] Using FlashInfer for top-p & top-k sampling. 2025-11-03T17:20:59.0232951Z INFO 11-03 17:20:59 [topk_topp_sampler.py:59] Using FlashInfer for top-p & top-k sampling. 2025-11-03T17:20:59.0419614Z (Worker_TP0 pid=33408) INFO 11-03 17:20:59 [gpu_model_runner.py:2840] Starting to load model meta-llama/Llama-3.2-1B-Instruct... 2025-11-03T17:20:59.0555292Z (Worker_TP1 pid=33409) INFO 11-03 17:20:59 [gpu_model_runner.py:2840] Starting to load model meta-llama/Llama-3.2-1B-Instruct... 2025-11-03T17:20:59.2520202Z (Worker_TP0 pid=33408) INFO 11-03 17:20:59 [gpu_model_runner.py:2870] Loading model from scratch... 2025-11-03T17:20:59.2671167Z (Worker_TP1 pid=33409) INFO 11-03 17:20:59 [gpu_model_runner.py:2870] Loading model from scratch... 2025-11-03T17:20:59.3425472Z (Worker_TP0 pid=33408) INFO 11-03 17:20:59 [cuda.py:404] Using Flash Attention backend on V1 engine. 2025-11-03T17:20:59.3574107Z (Worker_TP1 pid=33409) INFO 11-03 17:20:59 [cuda.py:404] Using Flash Attention backend on V1 engine. 2025-11-03T17:20:59.8355815Z (Worker_TP0 pid=33408) INFO 11-03 17:20:59 [gpu_model_runner.py:2902] Model loading took 1.1667 GiB and 0.170347 seconds 2025-11-03T17:20:59.8521556Z (Worker_TP1 pid=33409) INFO 11-03 17:20:59 [gpu_model_runner.py:2902] Model loading took 1.1667 GiB and 0.169650 seconds 2025-11-03T17:21:03.5019444Z (Worker_TP0 pid=33408) INFO 11-03 17:21:03 [backends.py:594] Using cache directory: /var/lib/jenkins/.cache/vllm/torch_compile_cache/cdd8585f72/rank_0_0/backbone for vLLM's torch.compile 2025-11-03T17:21:03.5024527Z (Worker_TP0 pid=33408) INFO 11-03 17:21:03 [backends.py:608] Dynamo bytecode transform time: 3.37 s 2025-11-03T17:21:03.6181625Z (Worker_TP1 pid=33409) INFO 11-03 17:21:03 [backends.py:594] Using cache directory: /var/lib/jenkins/.cache/vllm/torch_compile_cache/cdd8585f72/rank_1_0/backbone for vLLM's torch.compile 2025-11-03T17:21:03.6188033Z (Worker_TP1 pid=33409) INFO 11-03 17:21:03 [backends.py:608] Dynamo bytecode transform time: 3.49 s 2025-11-03T17:21:04.6030032Z (Worker_TP0 pid=33408) INFO 11-03 17:21:04 [backends.py:179] Directly load the compiled graph(s) for dynamic shape from the cache, took 0.913 s 2025-11-03T17:21:04.7344773Z (Worker_TP1 pid=33409) INFO 11-03 17:21:04 [backends.py:179] Directly load the compiled graph(s) for dynamic shape from the cache, took 0.920 s 2025-11-03T17:21:05.3979438Z (Worker_TP1 pid=33409) INFO 11-03 17:21:05 [monitor.py:33] torch.compile takes 3.49 s in total 2025-11-03T17:21:05.3981261Z (Worker_TP0 pid=33408) INFO 11-03 17:21:05 [monitor.py:33] torch.compile takes 3.37 s in total 2025-11-03T17:21:06.0447063Z (Worker_TP1 pid=33409) INFO 11-03 17:21:06 [gpu_worker.py:314] Available KV cache memory: 18.54 GiB 2025-11-03T17:21:06.0562652Z (Worker_TP0 pid=33408) INFO 11-03 17:21:06 [gpu_worker.py:314] Available KV cache memory: 18.54 GiB 2025-11-03T17:21:06.2952501Z (EngineCore_DP0 pid=33289) INFO 11-03 17:21:06 [kv_cache_utils.py:1199] GPU KV cache size: 1,214,720 tokens 2025-11-03T17:21:06.2953558Z (EngineCore_DP0 pid=33289) INFO 11-03 17:21:06 [kv_cache_utils.py:1204] Maximum concurrency for 2,048 tokens per request: 593.12x 2025-11-03T17:21:06.2957574Z (EngineCore_DP0 pid=33289) INFO 11-03 17:21:06 [kv_cache_utils.py:1199] GPU KV cache size: 1,214,720 tokens 2025-11-03T17:21:06.2958427Z (EngineCore_DP0 pid=33289) INFO 11-03 17:21:06 [kv_cache_utils.py:1204] Maximum concurrency for 2,048 tokens per request: 593.12x 2025-11-03T17:21:06.5432821Z (Worker_TP0 pid=33408) 2025-11-03T17:21:06.6661657Z Capturing CUDA graphs (mixed prefill-decode, PIECEWISE): 0% 0/5 [00:00, 'use_cudagraph': False, 'cudagraph_num_of_warmups': 0, 'cudagraph_capture_sizes': [], 'cudagraph_copy_inputs': False, 'full_cuda_graph': False, 'use_inductor_graph_partition': False, 'pass_config': {'enable_noop': True, 'enable_sequence_parallelism': True}, 'max_capture_size': 0, 'local_cache_dir': None} 2025-11-03T17:22:01.2961034Z (EngineCore_DP0 pid=33785) WARNING 11-03 17:22:01 [multiproc_executor.py:744] Reducing Torch parallelism from 24 threads to 1 to avoid unnecessary CPU contention. Set OMP_NUM_THREADS in the external environment to tune this value as needed. 2025-11-03T17:22:01.2964452Z (EngineCore_DP0 pid=33785) INFO 11-03 17:22:01 [shm_broadcast.py:302] vLLM message queue communication handle: Handle(local_reader_ranks=[0, 1], buffer_handle=(2, 16777216, 10, 'psm_760baa3c'), local_subscribe_addr='ipc:///tmp/7a98014b-3cbc-4e39-845d-4189027acaae', remote_subscribe_addr=None, remote_addr_ipv6=False) 2025-11-03T17:22:01.7894333Z /opt/conda/envs/py_3.12/lib/python3.12/site-packages/torch/cuda/__init__.py:65: FutureWarning: The pynvml package is deprecated. Please install nvidia-ml-py instead. If you did not install pynvml directly, please report this to the maintainers of the package that installed pynvml for you. 2025-11-03T17:22:01.7895561Z import pynvml # type: ignore[import] 2025-11-03T17:22:01.8078278Z /opt/conda/envs/py_3.12/lib/python3.12/site-packages/torch/cuda/__init__.py:65: FutureWarning: The pynvml package is deprecated. Please install nvidia-ml-py instead. If you did not install pynvml directly, please report this to the maintainers of the package that installed pynvml for you. 2025-11-03T17:22:01.8079444Z import pynvml # type: ignore[import] 2025-11-03T17:22:04.1530876Z INFO 11-03 17:22:04 [__init__.py:224] Automatically detected platform cuda. 2025-11-03T17:22:04.1570220Z INFO 11-03 17:22:04 [__init__.py:224] Automatically detected platform cuda. 2025-11-03T17:22:08.8730942Z INFO 11-03 17:22:08 [shm_broadcast.py:302] vLLM message queue communication handle: Handle(local_reader_ranks=[0], buffer_handle=(1, 10485760, 10, 'psm_94714c96'), local_subscribe_addr='ipc:///tmp/19267841-2196-483c-8309-82ecdb6d37d2', remote_subscribe_addr=None, remote_addr_ipv6=False) 2025-11-03T17:22:08.8927813Z INFO 11-03 17:22:08 [shm_broadcast.py:302] vLLM message queue communication handle: Handle(local_reader_ranks=[0], buffer_handle=(1, 10485760, 10, 'psm_0bbcd7dc'), local_subscribe_addr='ipc:///tmp/3064b481-46e2-4c02-8077-f8d2e9d33b05', remote_subscribe_addr=None, remote_addr_ipv6=False) 2025-11-03T17:22:09.4106868Z [Gloo] Rank 0 is connected to 1 peer ranks. Expected number of connected peer ranks is : 1 2025-11-03T17:22:09.4108061Z [Gloo] Rank 1 is connected to 1 peer ranks. Expected number of connected peer ranks is : 1 2025-11-03T17:22:09.4150632Z [Gloo] Rank 1 is connected to 1 peer ranks. Expected number of connected peer ranks is : 1 2025-11-03T17:22:09.4151240Z [Gloo] Rank 0 is connected to 1 peer ranks. Expected number of connected peer ranks is : 1 2025-11-03T17:22:09.4195800Z INFO 11-03 17:22:09 [__init__.py:1146] Found nccl from library libnccl.so.2 2025-11-03T17:22:09.4196705Z INFO 11-03 17:22:09 [__init__.py:1146] Found nccl from library libnccl.so.2 2025-11-03T17:22:09.4197867Z INFO 11-03 17:22:09 [pynccl.py:109] vLLM is using nccl==2.27.5 2025-11-03T17:22:09.4198616Z INFO 11-03 17:22:09 [pynccl.py:109] vLLM is using nccl==2.27.5 2025-11-03T17:22:09.5947424Z WARNING 11-03 17:22:09 [symm_mem.py:60] SymmMemCommunicator: Device capability 8.9 not supported, communicator is not available. 2025-11-03T17:22:09.5949007Z WARNING 11-03 17:22:09 [symm_mem.py:60] SymmMemCommunicator: Device capability 8.9 not supported, communicator is not available. 2025-11-03T17:22:09.5971034Z INFO 11-03 17:22:09 [custom_all_reduce.py:37] Skipping P2P check and trusting the driver's P2P report. 2025-11-03T17:22:09.5972938Z INFO 11-03 17:22:09 [custom_all_reduce.py:37] Skipping P2P check and trusting the driver's P2P report. 2025-11-03T17:22:09.5974539Z WARNING 11-03 17:22:09 [custom_all_reduce.py:162] Custom allreduce is disabled because your platform lacks GPU P2P capability or P2P test failed. To silence this warning, specify disable_custom_all_reduce=True explicitly. 2025-11-03T17:22:09.5975939Z WARNING 11-03 17:22:09 [custom_all_reduce.py:162] Custom allreduce is disabled because your platform lacks GPU P2P capability or P2P test failed. To silence this warning, specify disable_custom_all_reduce=True explicitly. 2025-11-03T17:22:09.6925584Z INFO 11-03 17:22:09 [shm_broadcast.py:302] vLLM message queue communication handle: Handle(local_reader_ranks=[1], buffer_handle=(1, 4194304, 6, 'psm_a82f40bd'), local_subscribe_addr='ipc:///tmp/4daad9f2-1130-48e7-ac43-f563c966fd49', remote_subscribe_addr=None, remote_addr_ipv6=False) 2025-11-03T17:22:09.6957738Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:22:09.6958257Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:22:09.6968584Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:22:09.6976030Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:22:09.7722069Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:22:09.7728916Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:22:09.7742338Z [Gloo] Rank 0 is connected to 1 peer ranks. Expected number of connected peer ranks is : 1 2025-11-03T17:22:09.7743521Z [Gloo] Rank 1 is connected to 1 peer ranks. Expected number of connected peer ranks is : 1 2025-11-03T17:22:09.7744725Z INFO 11-03 17:22:09 [__init__.py:1146] Found nccl from library libnccl.so.2 2025-11-03T17:22:09.7745435Z INFO 11-03 17:22:09 [__init__.py:1146] Found nccl from library libnccl.so.2 2025-11-03T17:22:09.7745832Z INFO 11-03 17:22:09 [pynccl.py:109] vLLM is using nccl==2.27.5 2025-11-03T17:22:09.7746138Z INFO 11-03 17:22:09 [pynccl.py:109] vLLM is using nccl==2.27.5 2025-11-03T17:22:09.8032021Z INFO 11-03 17:22:09 [parallel_state.py:1231] rank 0 in world size 2 is assigned as DP rank 0, PP rank 0, TP rank 0, EP rank 0 2025-11-03T17:22:09.8032626Z INFO 11-03 17:22:09 [parallel_state.py:1231] rank 1 in world size 2 is assigned as DP rank 0, PP rank 0, TP rank 1, EP rank 1 2025-11-03T17:22:10.0357981Z INFO 11-03 17:22:10 [topk_topp_sampler.py:59] Using FlashInfer for top-p & top-k sampling. 2025-11-03T17:22:10.0506977Z INFO 11-03 17:22:10 [topk_topp_sampler.py:59] Using FlashInfer for top-p & top-k sampling. 2025-11-03T17:22:10.0684928Z (Worker_TP1 pid=33905) INFO 11-03 17:22:10 [gpu_model_runner.py:2840] Starting to load model meta-llama/Llama-3.2-1B-Instruct... 2025-11-03T17:22:10.0830653Z (Worker_TP0 pid=33904) INFO 11-03 17:22:10 [gpu_model_runner.py:2840] Starting to load model meta-llama/Llama-3.2-1B-Instruct... 2025-11-03T17:22:10.2756992Z (Worker_TP1 pid=33905) INFO 11-03 17:22:10 [gpu_model_runner.py:2870] Loading model from scratch... 2025-11-03T17:22:10.2930463Z (Worker_TP0 pid=33904) INFO 11-03 17:22:10 [gpu_model_runner.py:2870] Loading model from scratch... 2025-11-03T17:22:10.3652653Z (Worker_TP1 pid=33905) INFO 11-03 17:22:10 [cuda.py:404] Using Flash Attention backend on V1 engine. 2025-11-03T17:22:10.3831677Z (Worker_TP0 pid=33904) INFO 11-03 17:22:10 [cuda.py:404] Using Flash Attention backend on V1 engine. 2025-11-03T17:22:10.8577298Z (Worker_TP1 pid=33905) INFO 11-03 17:22:10 [gpu_model_runner.py:2902] Model loading took 1.1667 GiB and 0.169116 seconds 2025-11-03T17:22:10.8769967Z (Worker_TP0 pid=33904) INFO 11-03 17:22:10 [gpu_model_runner.py:2902] Model loading took 1.1667 GiB and 0.167324 seconds 2025-11-03T17:22:14.1658338Z (Worker_TP1 pid=33905) INFO 11-03 17:22:14 [backends.py:594] Using cache directory: /var/lib/jenkins/.cache/vllm/torch_compile_cache/618b881a70/rank_1_0/backbone for vLLM's torch.compile 2025-11-03T17:22:14.1666780Z (Worker_TP1 pid=33905) INFO 11-03 17:22:14 [backends.py:608] Dynamo bytecode transform time: 3.03 s 2025-11-03T17:22:14.3134655Z (Worker_TP0 pid=33904) INFO 11-03 17:22:14 [backends.py:594] Using cache directory: /var/lib/jenkins/.cache/vllm/torch_compile_cache/618b881a70/rank_0_0/backbone for vLLM's torch.compile 2025-11-03T17:22:14.3144700Z (Worker_TP0 pid=33904) INFO 11-03 17:22:14 [backends.py:608] Dynamo bytecode transform time: 3.18 s 2025-11-03T17:22:15.2584760Z (Worker_TP1 pid=33905) INFO 11-03 17:22:15 [backends.py:179] Directly load the compiled graph(s) for dynamic shape from the cache, took 0.816 s 2025-11-03T17:22:15.4298510Z (Worker_TP0 pid=33904) INFO 11-03 17:22:15 [backends.py:179] Directly load the compiled graph(s) for dynamic shape from the cache, took 0.835 s 2025-11-03T17:22:16.8809973Z (Worker_TP1 pid=33905) INFO 11-03 17:22:16 [gpu_worker.py:314] Available KV cache memory: 18.54 GiB 2025-11-03T17:22:16.8845740Z (Worker_TP0 pid=33904) INFO 11-03 17:22:16 [gpu_worker.py:314] Available KV cache memory: 18.54 GiB 2025-11-03T17:22:17.1232929Z (EngineCore_DP0 pid=33785) INFO 11-03 17:22:17 [kv_cache_utils.py:1199] GPU KV cache size: 1,215,232 tokens 2025-11-03T17:22:17.1233879Z (EngineCore_DP0 pid=33785) INFO 11-03 17:22:17 [kv_cache_utils.py:1204] Maximum concurrency for 2,048 tokens per request: 593.38x 2025-11-03T17:22:17.1234695Z (EngineCore_DP0 pid=33785) INFO 11-03 17:22:17 [kv_cache_utils.py:1199] GPU KV cache size: 1,215,232 tokens 2025-11-03T17:22:17.1235502Z (EngineCore_DP0 pid=33785) INFO 11-03 17:22:17 [kv_cache_utils.py:1204] Maximum concurrency for 2,048 tokens per request: 593.38x 2025-11-03T17:22:17.1302772Z (Worker_TP1 pid=33905) INFO 11-03 17:22:17 [gpu_worker.py:351] Compile and warming up model for size 8 2025-11-03T17:22:17.1303518Z (Worker_TP0 pid=33904) INFO 11-03 17:22:17 [gpu_worker.py:351] Compile and warming up model for size 8 2025-11-03T17:22:18.1750094Z (Worker_TP1 pid=33905) INFO 11-03 17:22:18 [backends.py:185] Directly load the compiled graph(s) for shape 8 from the cache, took 1.044 s 2025-11-03T17:22:18.1788388Z (Worker_TP0 pid=33904) INFO 11-03 17:22:18 [backends.py:185] Directly load the compiled graph(s) for shape 8 from the cache, took 1.048 s 2025-11-03T17:22:18.1794617Z (Worker_TP1 pid=33905) INFO 11-03 17:22:18 [gpu_worker.py:351] Compile and warming up model for size 4 2025-11-03T17:22:18.1795855Z (Worker_TP0 pid=33904) INFO 11-03 17:22:18 [gpu_worker.py:351] Compile and warming up model for size 4 2025-11-03T17:22:18.9164147Z (Worker_TP1 pid=33905) INFO 11-03 17:22:18 [backends.py:185] Directly load the compiled graph(s) for shape 4 from the cache, took 0.736 s 2025-11-03T17:22:18.9165027Z (Worker_TP1 pid=33905) INFO 11-03 17:22:18 [monitor.py:33] torch.compile takes 3.03 s in total 2025-11-03T17:22:18.9186132Z (Worker_TP0 pid=33904) INFO 11-03 17:22:18 [backends.py:185] Directly load the compiled graph(s) for shape 4 from the cache, took 0.739 s 2025-11-03T17:22:18.9186961Z (Worker_TP0 pid=33904) INFO 11-03 17:22:18 [monitor.py:33] torch.compile takes 3.18 s in total 2025-11-03T17:22:18.9312166Z (EngineCore_DP0 pid=33785) INFO 11-03 17:22:18 [core.py:240] init engine (profile, create kv cache, warmup model) took 8.00 seconds 2025-11-03T17:22:19.4118660Z (EngineCore_DP0 pid=33785) INFO 11-03 17:22:19 [vllm.py:354] Cudagraph is disabled under eager mode 2025-11-03T17:22:19.5965063Z (APIServer pid=33663) INFO 11-03 17:22:19 [loggers.py:172] Engine 000: vllm cache_config_info with initialization after num_gpu_blocks is: 75952 2025-11-03T17:22:19.6085321Z (EngineCore_DP0 pid=33785) INFO 11-03 17:22:19 [gc_utils.py:40] GC Debug Config. enabled:False,top_objects:-1 2025-11-03T17:22:19.6303379Z (APIServer pid=33663) INFO 11-03 17:22:19 [api_server.py:1628] Supported tasks: ['generate'] 2025-11-03T17:22:19.6539266Z (APIServer pid=33663) WARNING 11-03 17:22:19 [model.py:1581] Default sampling parameters have been overridden by the model's Hugging Face generation config recommended from the model creator. If this is not intended, please relaunch vLLM instance with `--generation-config vllm`. 2025-11-03T17:22:19.6542257Z (APIServer pid=33663) INFO 11-03 17:22:19 [serving_responses.py:157] Using default chat sampling params from model: {'temperature': 0.6, 'top_p': 0.9} 2025-11-03T17:22:19.6772220Z (APIServer pid=33663) INFO 11-03 17:22:19 [serving_chat.py:159] Using default chat sampling params from model: {'temperature': 0.6, 'top_p': 0.9} 2025-11-03T17:22:19.7070441Z (APIServer pid=33663) INFO 11-03 17:22:19 [serving_completion.py:67] Using default completion sampling params from model: {'temperature': 0.6, 'top_p': 0.9} 2025-11-03T17:22:19.7073274Z (APIServer pid=33663) INFO 11-03 17:22:19 [api_server.py:1939] Starting vLLM API server 0 on http://0.0.0.0:50361 2025-11-03T17:22:19.7073933Z (APIServer pid=33663) INFO 11-03 17:22:19 [launcher.py:38] Available routes are: 2025-11-03T17:22:19.7074564Z (APIServer pid=33663) INFO 11-03 17:22:19 [launcher.py:46] Route: /openapi.json, Methods: GET, HEAD 2025-11-03T17:22:19.7075237Z (APIServer pid=33663) INFO 11-03 17:22:19 [launcher.py:46] Route: /docs, Methods: GET, HEAD 2025-11-03T17:22:19.7075798Z (APIServer pid=33663) INFO 11-03 17:22:19 [launcher.py:46] Route: /docs/oauth2-redirect, Methods: GET, HEAD 2025-11-03T17:22:19.7076355Z (APIServer pid=33663) INFO 11-03 17:22:19 [launcher.py:46] Route: /redoc, Methods: GET, HEAD 2025-11-03T17:22:19.7076867Z (APIServer pid=33663) INFO 11-03 17:22:19 [launcher.py:46] Route: /health, Methods: GET 2025-11-03T17:22:19.7077392Z (APIServer pid=33663) INFO 11-03 17:22:19 [launcher.py:46] Route: /load, Methods: GET 2025-11-03T17:22:19.7077889Z (APIServer pid=33663) INFO 11-03 17:22:19 [launcher.py:46] Route: /ping, Methods: POST 2025-11-03T17:22:19.7078377Z (APIServer pid=33663) INFO 11-03 17:22:19 [launcher.py:46] Route: /ping, Methods: GET 2025-11-03T17:22:19.7078876Z (APIServer pid=33663) INFO 11-03 17:22:19 [launcher.py:46] Route: /tokenize, Methods: POST 2025-11-03T17:22:19.7079387Z (APIServer pid=33663) INFO 11-03 17:22:19 [launcher.py:46] Route: /detokenize, Methods: POST 2025-11-03T17:22:19.7079923Z (APIServer pid=33663) INFO 11-03 17:22:19 [launcher.py:46] Route: /v1/models, Methods: GET 2025-11-03T17:22:19.7080435Z (APIServer pid=33663) INFO 11-03 17:22:19 [launcher.py:46] Route: /version, Methods: GET 2025-11-03T17:22:19.7080963Z (APIServer pid=33663) INFO 11-03 17:22:19 [launcher.py:46] Route: /v1/responses, Methods: POST 2025-11-03T17:22:19.7081536Z (APIServer pid=33663) INFO 11-03 17:22:19 [launcher.py:46] Route: /v1/responses/{response_id}, Methods: GET 2025-11-03T17:22:19.7082156Z (APIServer pid=33663) INFO 11-03 17:22:19 [launcher.py:46] Route: /v1/responses/{response_id}/cancel, Methods: POST 2025-11-03T17:22:19.7082752Z (APIServer pid=33663) INFO 11-03 17:22:19 [launcher.py:46] Route: /v1/chat/completions, Methods: POST 2025-11-03T17:22:19.7083310Z (APIServer pid=33663) INFO 11-03 17:22:19 [launcher.py:46] Route: /v1/completions, Methods: POST 2025-11-03T17:22:19.7083855Z (APIServer pid=33663) INFO 11-03 17:22:19 [launcher.py:46] Route: /v1/embeddings, Methods: POST 2025-11-03T17:22:19.7084381Z (APIServer pid=33663) INFO 11-03 17:22:19 [launcher.py:46] Route: /pooling, Methods: POST 2025-11-03T17:22:19.7085064Z (APIServer pid=33663) INFO 11-03 17:22:19 [launcher.py:46] Route: /classify, Methods: POST 2025-11-03T17:22:19.7085573Z (APIServer pid=33663) INFO 11-03 17:22:19 [launcher.py:46] Route: /score, Methods: POST 2025-11-03T17:22:19.7086076Z (APIServer pid=33663) INFO 11-03 17:22:19 [launcher.py:46] Route: /v1/score, Methods: POST 2025-11-03T17:22:19.7086641Z (APIServer pid=33663) INFO 11-03 17:22:19 [launcher.py:46] Route: /v1/audio/transcriptions, Methods: POST 2025-11-03T17:22:19.7087225Z (APIServer pid=33663) INFO 11-03 17:22:19 [launcher.py:46] Route: /v1/audio/translations, Methods: POST 2025-11-03T17:22:19.7087937Z (APIServer pid=33663) INFO 11-03 17:22:19 [launcher.py:46] Route: /rerank, Methods: POST 2025-11-03T17:22:19.7088550Z (APIServer pid=33663) INFO 11-03 17:22:19 [launcher.py:46] Route: /v1/rerank, Methods: POST 2025-11-03T17:22:19.7089073Z (APIServer pid=33663) INFO 11-03 17:22:19 [launcher.py:46] Route: /v2/rerank, Methods: POST 2025-11-03T17:22:19.7089619Z (APIServer pid=33663) INFO 11-03 17:22:19 [launcher.py:46] Route: /scale_elastic_ep, Methods: POST 2025-11-03T17:22:19.7090283Z (APIServer pid=33663) INFO 11-03 17:22:19 [launcher.py:46] Route: /is_scaling_elastic_ep, Methods: POST 2025-11-03T17:22:19.7090837Z (APIServer pid=33663) INFO 11-03 17:22:19 [launcher.py:46] Route: /invocations, Methods: POST 2025-11-03T17:22:19.7091346Z (APIServer pid=33663) INFO 11-03 17:22:19 [launcher.py:46] Route: /metrics, Methods: GET 2025-11-03T17:22:19.7226696Z (APIServer pid=33663) INFO: Started server process [33663] 2025-11-03T17:22:19.7227874Z (APIServer pid=33663) INFO: Waiting for application startup. 2025-11-03T17:22:19.9272090Z (APIServer pid=33663) INFO: Application startup complete. 2025-11-03T17:22:20.0534274Z (APIServer pid=33663) INFO: 127.0.0.1:59744 - "GET /health HTTP/1.1" 200 OK 2025-11-03T17:22:20.2963965Z (APIServer pid=33663) INFO: 127.0.0.1:59756 - "GET /v1/models HTTP/1.1" 200 OK 2025-11-03T17:22:20.4026624Z (APIServer pid=33663) INFO: 127.0.0.1:59756 - "POST /v1/completions HTTP/1.1" 200 OK 2025-11-03T17:22:20.4602775Z (APIServer pid=33663) INFO: 127.0.0.1:59756 - "POST /v1/completions HTTP/1.1" 200 OK 2025-11-03T17:22:20.5965092Z (APIServer pid=33663) INFO: 127.0.0.1:59756 - "POST /v1/completions HTTP/1.1" 200 OK 2025-11-03T17:22:20.6658337Z (APIServer pid=33663) INFO: 127.0.0.1:59756 - "POST /v1/completions HTTP/1.1" 200 OK 2025-11-03T17:22:20.7215388Z (APIServer pid=33663) INFO: 127.0.0.1:59756 - "POST /v1/completions HTTP/1.1" 200 OK 2025-11-03T17:22:20.7295755Z (APIServer pid=33663) INFO: 127.0.0.1:59756 - "POST /v1/completions HTTP/1.1" 200 OK 2025-11-03T17:22:20.7888956Z (APIServer pid=33663) INFO 11-03 17:22:20 [launcher.py:110] Shutting down FastAPI HTTP server. 2025-11-03T17:22:20.7892664Z (Worker_TP0 pid=33904) INFO 11-03 17:22:20 [multiproc_executor.py:583] Parent process exited, terminating worker 2025-11-03T17:22:20.7894160Z (Worker_TP0 pid=33904) INFO 11-03 17:22:20 [multiproc_executor.py:624] WorkerProc shutting down. 2025-11-03T17:22:20.7895538Z (Worker_TP1 pid=33905) INFO 11-03 17:22:20 [multiproc_executor.py:583] Parent process exited, terminating worker 2025-11-03T17:22:20.7898064Z (Worker_TP1 pid=33905) INFO 11-03 17:22:20 [multiproc_executor.py:624] WorkerProc shutting down. 2025-11-03T17:22:22.8025294Z (APIServer pid=33663) INFO: Shutting down 2025-11-03T17:22:22.9029880Z (APIServer pid=33663) INFO: Waiting for application shutdown. 2025-11-03T17:22:22.9031813Z (APIServer pid=33663) INFO: Application shutdown complete. 2025-11-03T17:22:23.2478641Z INFO 11-03 17:22:23 [model.py:653] Resolved architecture: LlamaForCausalLM 2025-11-03T17:22:23.2479750Z WARNING 11-03 17:22:23 [model.py:1942] Casting torch.bfloat16 to torch.float16. 2025-11-03T17:22:23.2480609Z INFO 11-03 17:22:23 [model.py:1714] Using max model len 2048 2025-11-03T17:22:23.2482850Z Launching RemoteOpenAIServer with: vllm serve meta-llama/Llama-3.2-1B-Instruct --dtype float16 --max-model-len 2048 --max-num-seqs 8 --enforce-eager --tokenizer-mode auto --tensor-parallel-size 2 --distributed-executor-backend mp --load-format dummy --port 59953 --seed 0 2025-11-03T17:22:23.7658994Z /opt/conda/envs/py_3.12/lib/python3.12/site-packages/torch/cuda/__init__.py:65: FutureWarning: The pynvml package is deprecated. Please install nvidia-ml-py instead. If you did not install pynvml directly, please report this to the maintainers of the package that installed pynvml for you. 2025-11-03T17:22:23.7660738Z import pynvml # type: ignore[import] 2025-11-03T17:22:26.0213600Z INFO 11-03 17:22:26 [__init__.py:224] Automatically detected platform cuda. 2025-11-03T17:22:29.6796168Z (APIServer pid=34130) INFO 11-03 17:22:29 [api_server.py:1870] vLLM API server version 0.11.0rc2.dev389+ge51928192.d20251103 2025-11-03T17:22:29.6825394Z (APIServer pid=34130) INFO 11-03 17:22:29 [utils.py:239] non-default args: {'model_tag': 'meta-llama/Llama-3.2-1B-Instruct', 'port': 59953, 'model': 'meta-llama/Llama-3.2-1B-Instruct', 'dtype': 'float16', 'seed': 0, 'max_model_len': 2048, 'enforce_eager': True, 'load_format': 'dummy', 'distributed_executor_backend': 'mp', 'tensor_parallel_size': 2, 'max_num_seqs': 8} 2025-11-03T17:22:29.8461878Z (APIServer pid=34130) INFO 11-03 17:22:29 [model.py:653] Resolved architecture: LlamaForCausalLM 2025-11-03T17:22:29.8463217Z (APIServer pid=34130) `torch_dtype` is deprecated! Use `dtype` instead! 2025-11-03T17:22:29.8464632Z (APIServer pid=34130) WARNING 11-03 17:22:29 [model.py:1942] Casting torch.bfloat16 to torch.float16. 2025-11-03T17:22:29.8465999Z (APIServer pid=34130) INFO 11-03 17:22:29 [model.py:1714] Using max model len 2048 2025-11-03T17:22:30.0132238Z (APIServer pid=34130) INFO 11-03 17:22:30 [scheduler.py:225] Chunked prefill is enabled with max_num_batched_tokens=2048. 2025-11-03T17:22:30.0136622Z (APIServer pid=34130) INFO 11-03 17:22:30 [vllm.py:354] Cudagraph is disabled under eager mode 2025-11-03T17:22:31.0373442Z /opt/conda/envs/py_3.12/lib/python3.12/site-packages/torch/cuda/__init__.py:65: FutureWarning: The pynvml package is deprecated. Please install nvidia-ml-py instead. If you did not install pynvml directly, please report this to the maintainers of the package that installed pynvml for you. 2025-11-03T17:22:31.0374437Z import pynvml # type: ignore[import] 2025-11-03T17:22:33.3765523Z INFO 11-03 17:22:33 [__init__.py:224] Automatically detected platform cuda. 2025-11-03T17:22:37.0727085Z (EngineCore_DP0 pid=34252) INFO 11-03 17:22:37 [core.py:727] Waiting for init message from front-end. 2025-11-03T17:22:37.0987338Z (EngineCore_DP0 pid=34252) INFO 11-03 17:22:37 [core.py:94] Initializing a V1 LLM engine (v0.11.0rc2.dev389+ge51928192.d20251103) with config: model='meta-llama/Llama-3.2-1B-Instruct', speculative_config=None, tokenizer='meta-llama/Llama-3.2-1B-Instruct', skip_tokenizer_init=False, tokenizer_mode=auto, revision=None, tokenizer_revision=None, trust_remote_code=False, dtype=torch.float16, max_seq_len=2048, download_dir=None, load_format=dummy, tensor_parallel_size=2, pipeline_parallel_size=1, data_parallel_size=1, disable_custom_all_reduce=False, quantization=None, enforce_eager=True, kv_cache_dtype=auto, device_config=cuda, structured_outputs_config=StructuredOutputsConfig(backend='auto', disable_fallback=False, disable_any_whitespace=False, disable_additional_properties=False, reasoning_parser=''), observability_config=ObservabilityConfig(show_hidden_metrics_for_version=None, otlp_traces_endpoint=None, collect_detailed_traces=None), seed=0, served_model_name=meta-llama/Llama-3.2-1B-Instruct, enable_prefix_caching=True, chunked_prefill_enabled=True, pooler_config=None, compilation_config={'level': 0, 'debug_dump_path': None, 'cache_dir': '', 'backend': '', 'custom_ops': [], 'splitting_ops': None, 'use_inductor': True, 'compile_sizes': [], 'inductor_compile_config': {'enable_auto_functionalized_v2': False}, 'inductor_passes': {}, 'cudagraph_mode': , 'use_cudagraph': False, 'cudagraph_num_of_warmups': 0, 'cudagraph_capture_sizes': [], 'cudagraph_copy_inputs': False, 'full_cuda_graph': False, 'use_inductor_graph_partition': False, 'pass_config': {}, 'max_capture_size': 0, 'local_cache_dir': None} 2025-11-03T17:22:37.0992861Z (EngineCore_DP0 pid=34252) WARNING 11-03 17:22:37 [multiproc_executor.py:744] Reducing Torch parallelism from 24 threads to 1 to avoid unnecessary CPU contention. Set OMP_NUM_THREADS in the external environment to tune this value as needed. 2025-11-03T17:22:37.1002630Z (EngineCore_DP0 pid=34252) INFO 11-03 17:22:37 [shm_broadcast.py:302] vLLM message queue communication handle: Handle(local_reader_ranks=[0, 1], buffer_handle=(2, 16777216, 10, 'psm_7bde95d8'), local_subscribe_addr='ipc:///tmp/cbfd1615-1b3e-4fac-8a2a-200bb3e42d08', remote_subscribe_addr=None, remote_addr_ipv6=False) 2025-11-03T17:22:37.5827774Z /opt/conda/envs/py_3.12/lib/python3.12/site-packages/torch/cuda/__init__.py:65: FutureWarning: The pynvml package is deprecated. Please install nvidia-ml-py instead. If you did not install pynvml directly, please report this to the maintainers of the package that installed pynvml for you. 2025-11-03T17:22:37.5829005Z import pynvml # type: ignore[import] 2025-11-03T17:22:37.5981894Z /opt/conda/envs/py_3.12/lib/python3.12/site-packages/torch/cuda/__init__.py:65: FutureWarning: The pynvml package is deprecated. Please install nvidia-ml-py instead. If you did not install pynvml directly, please report this to the maintainers of the package that installed pynvml for you. 2025-11-03T17:22:37.5983120Z import pynvml # type: ignore[import] 2025-11-03T17:22:39.9375507Z INFO 11-03 17:22:39 [__init__.py:224] Automatically detected platform cuda. 2025-11-03T17:22:39.9642099Z INFO 11-03 17:22:39 [__init__.py:224] Automatically detected platform cuda. 2025-11-03T17:22:44.6329591Z INFO 11-03 17:22:44 [shm_broadcast.py:302] vLLM message queue communication handle: Handle(local_reader_ranks=[0], buffer_handle=(1, 10485760, 10, 'psm_3e490bbf'), local_subscribe_addr='ipc:///tmp/aa661645-b749-4290-90bd-936dad100b57', remote_subscribe_addr=None, remote_addr_ipv6=False) 2025-11-03T17:22:44.6724306Z INFO 11-03 17:22:44 [shm_broadcast.py:302] vLLM message queue communication handle: Handle(local_reader_ranks=[0], buffer_handle=(1, 10485760, 10, 'psm_cbe5d18a'), local_subscribe_addr='ipc:///tmp/1d57d7b6-5552-4d72-a52b-e08f6d68328b', remote_subscribe_addr=None, remote_addr_ipv6=False) 2025-11-03T17:22:45.3576680Z [Gloo] Rank 1 is connected to 1 peer ranks. Expected number of connected peer ranks is : 1 2025-11-03T17:22:45.3577308Z [Gloo] Rank 0 is connected to 1 peer ranks. Expected number of connected peer ranks is : 1 2025-11-03T17:22:45.3620710Z [Gloo] Rank 1 is connected to 1 peer ranks. Expected number of connected peer ranks is : 1 2025-11-03T17:22:45.3621839Z [Gloo] Rank 0 is connected to 1 peer ranks. Expected number of connected peer ranks is : 1 2025-11-03T17:22:45.3664328Z INFO 11-03 17:22:45 [__init__.py:1146] Found nccl from library libnccl.so.2 2025-11-03T17:22:45.3665802Z INFO 11-03 17:22:45 [__init__.py:1146] Found nccl from library libnccl.so.2 2025-11-03T17:22:45.3666662Z INFO 11-03 17:22:45 [pynccl.py:109] vLLM is using nccl==2.27.5 2025-11-03T17:22:45.3667395Z INFO 11-03 17:22:45 [pynccl.py:109] vLLM is using nccl==2.27.5 2025-11-03T17:22:45.5373799Z WARNING 11-03 17:22:45 [symm_mem.py:60] SymmMemCommunicator: Device capability 8.9 not supported, communicator is not available. 2025-11-03T17:22:45.5376175Z WARNING 11-03 17:22:45 [symm_mem.py:60] SymmMemCommunicator: Device capability 8.9 not supported, communicator is not available. 2025-11-03T17:22:45.5397030Z INFO 11-03 17:22:45 [custom_all_reduce.py:37] Skipping P2P check and trusting the driver's P2P report. 2025-11-03T17:22:45.5398209Z INFO 11-03 17:22:45 [custom_all_reduce.py:37] Skipping P2P check and trusting the driver's P2P report. 2025-11-03T17:22:45.5400135Z WARNING 11-03 17:22:45 [custom_all_reduce.py:162] Custom allreduce is disabled because your platform lacks GPU P2P capability or P2P test failed. To silence this warning, specify disable_custom_all_reduce=True explicitly. 2025-11-03T17:22:45.5402776Z WARNING 11-03 17:22:45 [custom_all_reduce.py:162] Custom allreduce is disabled because your platform lacks GPU P2P capability or P2P test failed. To silence this warning, specify disable_custom_all_reduce=True explicitly. 2025-11-03T17:22:45.6326135Z INFO 11-03 17:22:45 [shm_broadcast.py:302] vLLM message queue communication handle: Handle(local_reader_ranks=[1], buffer_handle=(1, 4194304, 6, 'psm_c5b63e5b'), local_subscribe_addr='ipc:///tmp/2204cc3f-2507-460c-825d-cadf37755783', remote_subscribe_addr=None, remote_addr_ipv6=False) 2025-11-03T17:22:45.6347056Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:22:45.6347540Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:22:45.6356355Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:22:45.6364223Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:22:45.6923092Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:22:45.6929637Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:22:45.6943866Z [Gloo] Rank 1 is connected to 1 peer ranks. Expected number of connected peer ranks is : 1 2025-11-03T17:22:45.6945739Z [Gloo] Rank 0 is connected to 1 peer ranks. Expected number of connected peer ranks is : 1 2025-11-03T17:22:45.6946360Z INFO 11-03 17:22:45 [__init__.py:1146] Found nccl from library libnccl.so.2 2025-11-03T17:22:45.6946781Z INFO 11-03 17:22:45 [pynccl.py:109] vLLM is using nccl==2.27.5 2025-11-03T17:22:45.6947179Z INFO 11-03 17:22:45 [__init__.py:1146] Found nccl from library libnccl.so.2 2025-11-03T17:22:45.6947569Z INFO 11-03 17:22:45 [pynccl.py:109] vLLM is using nccl==2.27.5 2025-11-03T17:22:45.7279421Z INFO 11-03 17:22:45 [parallel_state.py:1231] rank 0 in world size 2 is assigned as DP rank 0, PP rank 0, TP rank 0, EP rank 0 2025-11-03T17:22:45.7280232Z INFO 11-03 17:22:45 [parallel_state.py:1231] rank 1 in world size 2 is assigned as DP rank 0, PP rank 0, TP rank 1, EP rank 1 2025-11-03T17:22:45.9593189Z INFO 11-03 17:22:45 [topk_topp_sampler.py:59] Using FlashInfer for top-p & top-k sampling. 2025-11-03T17:22:45.9769695Z INFO 11-03 17:22:45 [topk_topp_sampler.py:59] Using FlashInfer for top-p & top-k sampling. 2025-11-03T17:22:45.9908682Z (Worker_TP1 pid=34372) INFO 11-03 17:22:45 [gpu_model_runner.py:2840] Starting to load model meta-llama/Llama-3.2-1B-Instruct... 2025-11-03T17:22:46.0100412Z (Worker_TP0 pid=34371) INFO 11-03 17:22:46 [gpu_model_runner.py:2840] Starting to load model meta-llama/Llama-3.2-1B-Instruct... 2025-11-03T17:22:46.1993400Z (Worker_TP1 pid=34372) INFO 11-03 17:22:46 [gpu_model_runner.py:2870] Loading model from scratch... 2025-11-03T17:22:46.2267789Z (Worker_TP0 pid=34371) INFO 11-03 17:22:46 [gpu_model_runner.py:2870] Loading model from scratch... 2025-11-03T17:22:46.2882816Z (Worker_TP1 pid=34372) INFO 11-03 17:22:46 [cuda.py:404] Using Flash Attention backend on V1 engine. 2025-11-03T17:22:46.3165468Z (Worker_TP0 pid=34371) INFO 11-03 17:22:46 [cuda.py:404] Using Flash Attention backend on V1 engine. 2025-11-03T17:22:46.7235124Z (Worker_TP1 pid=34372) INFO 11-03 17:22:46 [gpu_model_runner.py:2902] Model loading took 1.1667 GiB and 0.118478 seconds 2025-11-03T17:22:46.7479675Z (Worker_TP0 pid=34371) INFO 11-03 17:22:46 [gpu_model_runner.py:2902] Model loading took 1.1667 GiB and 0.118494 seconds 2025-11-03T17:22:49.4553950Z (Worker_TP0 pid=34371) INFO 11-03 17:22:49 [gpu_worker.py:314] Available KV cache memory: 18.53 GiB 2025-11-03T17:22:49.4591579Z (Worker_TP1 pid=34372) INFO 11-03 17:22:49 [gpu_worker.py:314] Available KV cache memory: 18.53 GiB 2025-11-03T17:22:49.6810656Z (EngineCore_DP0 pid=34252) INFO 11-03 17:22:49 [kv_cache_utils.py:1199] GPU KV cache size: 1,214,464 tokens 2025-11-03T17:22:49.6811567Z (EngineCore_DP0 pid=34252) INFO 11-03 17:22:49 [kv_cache_utils.py:1204] Maximum concurrency for 2,048 tokens per request: 593.00x 2025-11-03T17:22:49.6812385Z (EngineCore_DP0 pid=34252) INFO 11-03 17:22:49 [kv_cache_utils.py:1199] GPU KV cache size: 1,214,464 tokens 2025-11-03T17:22:49.6813737Z (EngineCore_DP0 pid=34252) INFO 11-03 17:22:49 [kv_cache_utils.py:1204] Maximum concurrency for 2,048 tokens per request: 593.00x 2025-11-03T17:22:49.9780233Z (EngineCore_DP0 pid=34252) INFO 11-03 17:22:49 [core.py:240] init engine (profile, create kv cache, warmup model) took 3.15 seconds 2025-11-03T17:22:50.4777975Z (EngineCore_DP0 pid=34252) INFO 11-03 17:22:50 [vllm.py:354] Cudagraph is disabled under eager mode 2025-11-03T17:22:50.6624713Z (APIServer pid=34130) INFO 11-03 17:22:50 [loggers.py:172] Engine 000: vllm cache_config_info with initialization after num_gpu_blocks is: 75904 2025-11-03T17:22:50.6779529Z (EngineCore_DP0 pid=34252) INFO 11-03 17:22:50 [gc_utils.py:40] GC Debug Config. enabled:False,top_objects:-1 2025-11-03T17:22:50.6987399Z (APIServer pid=34130) INFO 11-03 17:22:50 [api_server.py:1628] Supported tasks: ['generate'] 2025-11-03T17:22:50.7188556Z (APIServer pid=34130) WARNING 11-03 17:22:50 [model.py:1581] Default sampling parameters have been overridden by the model's Hugging Face generation config recommended from the model creator. If this is not intended, please relaunch vLLM instance with `--generation-config vllm`. 2025-11-03T17:22:50.7190200Z (APIServer pid=34130) INFO 11-03 17:22:50 [serving_responses.py:157] Using default chat sampling params from model: {'temperature': 0.6, 'top_p': 0.9} 2025-11-03T17:22:50.7427569Z (APIServer pid=34130) INFO 11-03 17:22:50 [serving_chat.py:159] Using default chat sampling params from model: {'temperature': 0.6, 'top_p': 0.9} 2025-11-03T17:22:50.7734837Z (APIServer pid=34130) INFO 11-03 17:22:50 [serving_completion.py:67] Using default completion sampling params from model: {'temperature': 0.6, 'top_p': 0.9} 2025-11-03T17:22:50.7735874Z (APIServer pid=34130) INFO 11-03 17:22:50 [api_server.py:1939] Starting vLLM API server 0 on http://0.0.0.0:59953 2025-11-03T17:22:50.7736593Z (APIServer pid=34130) INFO 11-03 17:22:50 [launcher.py:38] Available routes are: 2025-11-03T17:22:50.7737287Z (APIServer pid=34130) INFO 11-03 17:22:50 [launcher.py:46] Route: /openapi.json, Methods: GET, HEAD 2025-11-03T17:22:50.7737986Z (APIServer pid=34130) INFO 11-03 17:22:50 [launcher.py:46] Route: /docs, Methods: GET, HEAD 2025-11-03T17:22:50.7738688Z (APIServer pid=34130) INFO 11-03 17:22:50 [launcher.py:46] Route: /docs/oauth2-redirect, Methods: GET, HEAD 2025-11-03T17:22:50.7739379Z (APIServer pid=34130) INFO 11-03 17:22:50 [launcher.py:46] Route: /redoc, Methods: GET, HEAD 2025-11-03T17:22:50.7739922Z (APIServer pid=34130) INFO 11-03 17:22:50 [launcher.py:46] Route: /health, Methods: GET 2025-11-03T17:22:50.7740425Z (APIServer pid=34130) INFO 11-03 17:22:50 [launcher.py:46] Route: /load, Methods: GET 2025-11-03T17:22:50.7741200Z (APIServer pid=34130) INFO 11-03 17:22:50 [launcher.py:46] Route: /ping, Methods: POST 2025-11-03T17:22:50.7741711Z (APIServer pid=34130) INFO 11-03 17:22:50 [launcher.py:46] Route: /ping, Methods: GET 2025-11-03T17:22:50.7742233Z (APIServer pid=34130) INFO 11-03 17:22:50 [launcher.py:46] Route: /tokenize, Methods: POST 2025-11-03T17:22:50.7742755Z (APIServer pid=34130) INFO 11-03 17:22:50 [launcher.py:46] Route: /detokenize, Methods: POST 2025-11-03T17:22:50.7743261Z (APIServer pid=34130) INFO 11-03 17:22:50 [launcher.py:46] Route: /v1/models, Methods: GET 2025-11-03T17:22:50.7743762Z (APIServer pid=34130) INFO 11-03 17:22:50 [launcher.py:46] Route: /version, Methods: GET 2025-11-03T17:22:50.7744282Z (APIServer pid=34130) INFO 11-03 17:22:50 [launcher.py:46] Route: /v1/responses, Methods: POST 2025-11-03T17:22:50.7744850Z (APIServer pid=34130) INFO 11-03 17:22:50 [launcher.py:46] Route: /v1/responses/{response_id}, Methods: GET 2025-11-03T17:22:50.7745470Z (APIServer pid=34130) INFO 11-03 17:22:50 [launcher.py:46] Route: /v1/responses/{response_id}/cancel, Methods: POST 2025-11-03T17:22:50.7746223Z (APIServer pid=34130) INFO 11-03 17:22:50 [launcher.py:46] Route: /v1/chat/completions, Methods: POST 2025-11-03T17:22:50.7746799Z (APIServer pid=34130) INFO 11-03 17:22:50 [launcher.py:46] Route: /v1/completions, Methods: POST 2025-11-03T17:22:50.7747361Z (APIServer pid=34130) INFO 11-03 17:22:50 [launcher.py:46] Route: /v1/embeddings, Methods: POST 2025-11-03T17:22:50.7747893Z (APIServer pid=34130) INFO 11-03 17:22:50 [launcher.py:46] Route: /pooling, Methods: POST 2025-11-03T17:22:50.7748410Z (APIServer pid=34130) INFO 11-03 17:22:50 [launcher.py:46] Route: /classify, Methods: POST 2025-11-03T17:22:50.7748913Z (APIServer pid=34130) INFO 11-03 17:22:50 [launcher.py:46] Route: /score, Methods: POST 2025-11-03T17:22:50.7749440Z (APIServer pid=34130) INFO 11-03 17:22:50 [launcher.py:46] Route: /v1/score, Methods: POST 2025-11-03T17:22:50.7750017Z (APIServer pid=34130) INFO 11-03 17:22:50 [launcher.py:46] Route: /v1/audio/transcriptions, Methods: POST 2025-11-03T17:22:50.7750605Z (APIServer pid=34130) INFO 11-03 17:22:50 [launcher.py:46] Route: /v1/audio/translations, Methods: POST 2025-11-03T17:22:50.7751143Z (APIServer pid=34130) INFO 11-03 17:22:50 [launcher.py:46] Route: /rerank, Methods: POST 2025-11-03T17:22:50.7751663Z (APIServer pid=34130) INFO 11-03 17:22:50 [launcher.py:46] Route: /v1/rerank, Methods: POST 2025-11-03T17:22:50.7752173Z (APIServer pid=34130) INFO 11-03 17:22:50 [launcher.py:46] Route: /v2/rerank, Methods: POST 2025-11-03T17:22:50.7752706Z (APIServer pid=34130) INFO 11-03 17:22:50 [launcher.py:46] Route: /scale_elastic_ep, Methods: POST 2025-11-03T17:22:50.7753275Z (APIServer pid=34130) INFO 11-03 17:22:50 [launcher.py:46] Route: /is_scaling_elastic_ep, Methods: POST 2025-11-03T17:22:50.7753830Z (APIServer pid=34130) INFO 11-03 17:22:50 [launcher.py:46] Route: /invocations, Methods: POST 2025-11-03T17:22:50.7754346Z (APIServer pid=34130) INFO 11-03 17:22:50 [launcher.py:46] Route: /metrics, Methods: GET 2025-11-03T17:22:50.7892619Z (APIServer pid=34130) INFO: Started server process [34130] 2025-11-03T17:22:50.7893240Z (APIServer pid=34130) INFO: Waiting for application startup. 2025-11-03T17:22:51.0037981Z (APIServer pid=34130) INFO: Application startup complete. 2025-11-03T17:22:51.3213323Z (APIServer pid=34130) INFO: 127.0.0.1:50646 - "GET /health HTTP/1.1" 200 OK 2025-11-03T17:22:51.3304710Z (APIServer pid=34130) INFO: 127.0.0.1:50662 - "GET /v1/models HTTP/1.1" 200 OK 2025-11-03T17:22:51.6852333Z (APIServer pid=34130) INFO: 127.0.0.1:50662 - "POST /v1/completions HTTP/1.1" 200 OK 2025-11-03T17:22:51.7419754Z (APIServer pid=34130) INFO: 127.0.0.1:50662 - "POST /v1/completions HTTP/1.1" 200 OK 2025-11-03T17:22:51.8389298Z (APIServer pid=34130) INFO: 127.0.0.1:50662 - "POST /v1/completions HTTP/1.1" 200 OK 2025-11-03T17:22:51.9245285Z (APIServer pid=34130) INFO: 127.0.0.1:50662 - "POST /v1/completions HTTP/1.1" 200 OK 2025-11-03T17:22:51.9959267Z (APIServer pid=34130) INFO: 127.0.0.1:50662 - "POST /v1/completions HTTP/1.1" 200 OK 2025-11-03T17:22:52.0044514Z (APIServer pid=34130) INFO: 127.0.0.1:50662 - "POST /v1/completions HTTP/1.1" 200 OK 2025-11-03T17:22:52.0703248Z (APIServer pid=34130) INFO 11-03 17:22:52 [launcher.py:110] Shutting down FastAPI HTTP server. 2025-11-03T17:22:52.0706414Z (Worker_TP0 pid=34371) INFO 11-03 17:22:52 [multiproc_executor.py:583] Parent process exited, terminating worker 2025-11-03T17:22:52.0708624Z (Worker_TP0 pid=34371) INFO 11-03 17:22:52 [multiproc_executor.py:624] WorkerProc shutting down. 2025-11-03T17:22:52.0711953Z (Worker_TP1 pid=34372) INFO 11-03 17:22:52 [multiproc_executor.py:583] Parent process exited, terminating worker 2025-11-03T17:22:52.0713932Z (Worker_TP1 pid=34372) INFO 11-03 17:22:52 [multiproc_executor.py:624] WorkerProc shutting down. 2025-11-03T17:22:54.3860554Z (APIServer pid=34130) INFO: Shutting down 2025-11-03T17:22:54.4866981Z (APIServer pid=34130) INFO: Waiting for application shutdown. 2025-11-03T17:22:54.4869316Z (APIServer pid=34130) INFO: Application shutdown complete. 2025-11-03T17:22:55.0984845Z PASSED 2025-11-03T17:22:55.1134122Z distributed/test_sequence_parallel.py::test_tp_sp_generation[meta-llama/Llama-3.2-1B-Instruct-parallel_setup9-ray-auto-test_options9] Fork a new process to run a test 34751 2025-11-03T17:22:55.1148799Z Fork a new process to run a test 0 2025-11-03T17:22:55.7823445Z INFO 11-03 17:22:55 [model.py:653] Resolved architecture: LlamaForCausalLM 2025-11-03T17:22:55.7824368Z `torch_dtype` is deprecated! Use `dtype` instead! 2025-11-03T17:22:55.7825185Z WARNING 11-03 17:22:55 [model.py:1942] Casting torch.bfloat16 to torch.float16. 2025-11-03T17:22:55.7826469Z INFO 11-03 17:22:55 [model.py:1714] Using max model len 2048 2025-11-03T17:22:55.7831254Z Launching RemoteOpenAIServer with: vllm serve meta-llama/Llama-3.2-1B-Instruct --dtype float16 --max-model-len 2048 --max-num-seqs 8 --enforce-eager --tokenizer-mode auto --tensor-parallel-size 2 --pipeline-parallel-size 1 --distributed-executor-backend ray --compilation_config {"level": 3, "custom_ops": ["+rms_norm"], "compile_sizes": [4, 8], "pass_config": {"enable_sequence_parallelism": true, "enable_fusion": false, "enable_noop": true}} --load-format dummy --port 52467 --seed 0 2025-11-03T17:22:56.2944192Z /opt/conda/envs/py_3.12/lib/python3.12/site-packages/torch/cuda/__init__.py:65: FutureWarning: The pynvml package is deprecated. Please install nvidia-ml-py instead. If you did not install pynvml directly, please report this to the maintainers of the package that installed pynvml for you. 2025-11-03T17:22:56.2945424Z import pynvml # type: ignore[import] 2025-11-03T17:22:58.5326224Z INFO 11-03 17:22:58 [__init__.py:224] Automatically detected platform cuda. 2025-11-03T17:23:02.1997748Z (APIServer pid=34800) INFO 11-03 17:23:02 [api_server.py:1870] vLLM API server version 0.11.0rc2.dev389+ge51928192.d20251103 2025-11-03T17:23:02.2027180Z (APIServer pid=34800) INFO 11-03 17:23:02 [utils.py:239] non-default args: {'model_tag': 'meta-llama/Llama-3.2-1B-Instruct', 'port': 52467, 'model': 'meta-llama/Llama-3.2-1B-Instruct', 'dtype': 'float16', 'seed': 0, 'max_model_len': 2048, 'enforce_eager': True, 'load_format': 'dummy', 'distributed_executor_backend': 'ray', 'tensor_parallel_size': 2, 'max_num_seqs': 8, 'compilation_config': {'level': 3, 'debug_dump_path': None, 'cache_dir': '', 'backend': '', 'custom_ops': ['+rms_norm'], 'splitting_ops': None, 'use_inductor': True, 'compile_sizes': [4, 8], 'inductor_compile_config': {'enable_auto_functionalized_v2': False}, 'inductor_passes': {}, 'cudagraph_mode': None, 'use_cudagraph': True, 'cudagraph_num_of_warmups': 0, 'cudagraph_capture_sizes': None, 'cudagraph_copy_inputs': False, 'full_cuda_graph': False, 'use_inductor_graph_partition': False, 'pass_config': {'enable_noop': True, 'enable_sequence_parallelism': True}, 'max_capture_size': None, 'local_cache_dir': None}} 2025-11-03T17:23:02.3616260Z (APIServer pid=34800) INFO 11-03 17:23:02 [model.py:653] Resolved architecture: LlamaForCausalLM 2025-11-03T17:23:02.3617624Z (APIServer pid=34800) `torch_dtype` is deprecated! Use `dtype` instead! 2025-11-03T17:23:02.3618993Z (APIServer pid=34800) WARNING 11-03 17:23:02 [model.py:1942] Casting torch.bfloat16 to torch.float16. 2025-11-03T17:23:02.3620231Z (APIServer pid=34800) INFO 11-03 17:23:02 [model.py:1714] Using max model len 2048 2025-11-03T17:23:02.5283217Z (APIServer pid=34800) INFO 11-03 17:23:02 [scheduler.py:225] Chunked prefill is enabled with max_num_batched_tokens=2048. 2025-11-03T17:23:02.5288777Z (APIServer pid=34800) INFO 11-03 17:23:02 [vllm.py:354] Cudagraph is disabled under eager mode 2025-11-03T17:23:03.5010190Z /opt/conda/envs/py_3.12/lib/python3.12/site-packages/torch/cuda/__init__.py:65: FutureWarning: The pynvml package is deprecated. Please install nvidia-ml-py instead. If you did not install pynvml directly, please report this to the maintainers of the package that installed pynvml for you. 2025-11-03T17:23:03.5011448Z import pynvml # type: ignore[import] 2025-11-03T17:23:05.8697170Z INFO 11-03 17:23:05 [__init__.py:224] Automatically detected platform cuda. 2025-11-03T17:23:09.7566923Z (EngineCore_DP0 pid=34922) INFO 11-03 17:23:09 [core.py:727] Waiting for init message from front-end. 2025-11-03T17:23:09.7849790Z (EngineCore_DP0 pid=34922) INFO 11-03 17:23:09 [core.py:94] Initializing a V1 LLM engine (v0.11.0rc2.dev389+ge51928192.d20251103) with config: model='meta-llama/Llama-3.2-1B-Instruct', speculative_config=None, tokenizer='meta-llama/Llama-3.2-1B-Instruct', skip_tokenizer_init=False, tokenizer_mode=auto, revision=None, tokenizer_revision=None, trust_remote_code=False, dtype=torch.float16, max_seq_len=2048, download_dir=None, load_format=dummy, tensor_parallel_size=2, pipeline_parallel_size=1, data_parallel_size=1, disable_custom_all_reduce=False, quantization=None, enforce_eager=True, kv_cache_dtype=auto, device_config=cuda, structured_outputs_config=StructuredOutputsConfig(backend='auto', disable_fallback=False, disable_any_whitespace=False, disable_additional_properties=False, reasoning_parser=''), observability_config=ObservabilityConfig(show_hidden_metrics_for_version=None, otlp_traces_endpoint=None, collect_detailed_traces=None), seed=0, served_model_name=meta-llama/Llama-3.2-1B-Instruct, enable_prefix_caching=True, chunked_prefill_enabled=True, pooler_config=None, compilation_config={'level': 3, 'debug_dump_path': None, 'cache_dir': '', 'backend': '', 'custom_ops': ['+rms_norm', '+rms_norm'], 'splitting_ops': ['vllm.unified_attention', 'vllm.unified_attention_with_output', 'vllm.unified_mla_attention', 'vllm.unified_mla_attention_with_output', 'vllm.mamba_mixer2', 'vllm.mamba_mixer', 'vllm.short_conv', 'vllm.linear_attention', 'vllm.plamo2_mamba_mixer', 'vllm.gdn_attention', 'vllm.sparse_attn_indexer'], 'use_inductor': True, 'compile_sizes': [8, 4], 'inductor_compile_config': {'enable_auto_functionalized_v2': False}, 'inductor_passes': {}, 'cudagraph_mode': , 'use_cudagraph': False, 'cudagraph_num_of_warmups': 0, 'cudagraph_capture_sizes': [], 'cudagraph_copy_inputs': False, 'full_cuda_graph': False, 'use_inductor_graph_partition': False, 'pass_config': {'enable_noop': True, 'enable_sequence_parallelism': True}, 'max_capture_size': 0, 'local_cache_dir': None} 2025-11-03T17:23:11.6898722Z (EngineCore_DP0 pid=34922) 2025-11-03 17:23:11,689 INFO worker.py:1918 -- Started a local Ray instance. View the dashboard at http://127.0.0.1:8265  2025-11-03T17:23:12.5188148Z (EngineCore_DP0 pid=34922) INFO 11-03 17:23:12 [ray_utils.py:373] No current placement group found. Creating a new placement group. 2025-11-03T17:23:12.6931771Z (EngineCore_DP0 pid=34922) INFO 11-03 17:23:12 [ray_distributed_executor.py:179] use_ray_spmd_worker: True 2025-11-03T17:23:13.5806999Z (EngineCore_DP0 pid=34922) (pid=35813) /opt/conda/envs/py_3.12/lib/python3.12/site-packages/torch/cuda/__init__.py:65: FutureWarning: The pynvml package is deprecated. Please install nvidia-ml-py instead. If you did not install pynvml directly, please report this to the maintainers of the package that installed pynvml for you. 2025-11-03T17:23:13.5808829Z (EngineCore_DP0 pid=34922) (pid=35813) import pynvml # type: ignore[import] 2025-11-03T17:23:15.8307163Z (EngineCore_DP0 pid=34922) (pid=35813) INFO 11-03 17:23:15 [__init__.py:224] Automatically detected platform cuda. 2025-11-03T17:23:17.2983961Z (EngineCore_DP0 pid=34922) INFO 11-03 17:23:17 [ray_env.py:67] RAY_NON_CARRY_OVER_ENV_VARS from config: set() 2025-11-03T17:23:17.2985493Z (EngineCore_DP0 pid=34922) INFO 11-03 17:23:17 [ray_env.py:70] Copying the following environment variables to workers: ['MAX_JOBS', 'HUGGING_FACE_HUB_TOKEN', 'VLLM_USE_RAY_SPMD_WORKER', 'LD_LIBRARY_PATH', 'VLLM_WORKER_MULTIPROC_METHOD', 'VLLM_USE_V1', 'VLLM_USE_RAY_COMPILED_DAG', 'HF_TOKEN'] 2025-11-03T17:23:17.2987073Z (EngineCore_DP0 pid=34922) INFO 11-03 17:23:17 [ray_env.py:75] If certain env vars should NOT be copied, add them to /var/lib/jenkins/.config/vllm/ray_non_carry_over_env_vars.json file 2025-11-03T17:23:20.2177244Z (EngineCore_DP0 pid=34922) (RayWorkerWrapper pid=35813) WARNING 11-03 17:23:20 [worker_base.py:307] Missing `shared_worker_lock` argument from executor. This argument is needed for mm_processor_cache_type='shm'. 2025-11-03T17:23:20.4240190Z (EngineCore_DP0 pid=34922) (RayWorkerWrapper pid=35813) [Gloo] Rank 1 is connected to 1 peer ranks. Expected number of connected peer ranks is : 1 2025-11-03T17:23:20.4241371Z (EngineCore_DP0 pid=34922) (RayWorkerWrapper pid=35813) [Gloo] Rank 1 is connected to 1 peer ranks. Expected number of connected peer ranks is : 1 2025-11-03T17:23:20.4259479Z (EngineCore_DP0 pid=34922) (RayWorkerWrapper pid=35813) INFO 11-03 17:23:20 [__init__.py:1146] Found nccl from library libnccl.so.2 2025-11-03T17:23:20.4260467Z (EngineCore_DP0 pid=34922) (RayWorkerWrapper pid=35813) INFO 11-03 17:23:20 [pynccl.py:109] vLLM is using nccl==2.27.5 2025-11-03T17:23:20.7373666Z (EngineCore_DP0 pid=34922) (RayWorkerWrapper pid=35813) WARNING 11-03 17:23:20 [symm_mem.py:60] SymmMemCommunicator: Device capability 8.9 not supported, communicator is not available. 2025-11-03T17:23:20.7375513Z (EngineCore_DP0 pid=34922) (RayWorkerWrapper pid=35813) INFO 11-03 17:23:20 [custom_all_reduce.py:37] Skipping P2P check and trusting the driver's P2P report. 2025-11-03T17:23:20.7376842Z (EngineCore_DP0 pid=34922) (RayWorkerWrapper pid=35813) WARNING 11-03 17:23:20 [custom_all_reduce.py:162] Custom allreduce is disabled because your platform lacks GPU P2P capability or P2P test failed. To silence this warning, specify disable_custom_all_reduce=True explicitly. 2025-11-03T17:23:20.7378675Z (EngineCore_DP0 pid=34922) (RayWorkerWrapper pid=35812) INFO 11-03 17:23:20 [shm_broadcast.py:302] vLLM message queue communication handle: Handle(local_reader_ranks=[1], buffer_handle=(1, 4194304, 6, 'psm_61b53874'), local_subscribe_addr='ipc:///tmp/523fcdc8-6525-480e-95f4-8773d1ed7aea', remote_subscribe_addr=None, remote_addr_ipv6=False) 2025-11-03T17:23:20.8419107Z (EngineCore_DP0 pid=34922) (RayWorkerWrapper pid=35813) INFO 11-03 17:23:20 [parallel_state.py:1231] rank 1 in world size 2 is assigned as DP rank 0, PP rank 0, TP rank 1, EP rank 1 2025-11-03T17:23:20.8420748Z (EngineCore_DP0 pid=34922) (pid=35812) INFO 11-03 17:23:15 [__init__.py:224] Automatically detected platform cuda. 2025-11-03T17:23:21.1498721Z (EngineCore_DP0 pid=34922) (RayWorkerWrapper pid=35813) INFO 11-03 17:23:21 [topk_topp_sampler.py:59] Using FlashInfer for top-p & top-k sampling. 2025-11-03T17:23:21.1499964Z (EngineCore_DP0 pid=34922) (RayWorkerWrapper pid=35813) INFO 11-03 17:23:21 [gpu_model_runner.py:2840] Starting to load model meta-llama/Llama-3.2-1B-Instruct... 2025-11-03T17:23:21.3564929Z (EngineCore_DP0 pid=34922) (RayWorkerWrapper pid=35813) INFO 11-03 17:23:21 [gpu_model_runner.py:2870] Loading model from scratch... 2025-11-03T17:23:21.4605198Z (EngineCore_DP0 pid=34922) (RayWorkerWrapper pid=35813) INFO 11-03 17:23:21 [cuda.py:404] Using Flash Attention backend on V1 engine. 2025-11-03T17:23:21.9726484Z (EngineCore_DP0 pid=34922) (RayWorkerWrapper pid=35813) INFO 11-03 17:23:21 [gpu_model_runner.py:2902] Model loading took 1.1667 GiB and 0.170007 seconds 2025-11-03T17:23:25.1321595Z (EngineCore_DP0 pid=34922) (RayWorkerWrapper pid=35813) INFO 11-03 17:23:25 [backends.py:594] Using cache directory: /var/lib/jenkins/.cache/vllm/torch_compile_cache/618b881a70/rank_1_0/backbone for vLLM's torch.compile 2025-11-03T17:23:25.1322973Z (EngineCore_DP0 pid=34922) (RayWorkerWrapper pid=35813) INFO 11-03 17:23:25 [backends.py:608] Dynamo bytecode transform time: 2.96 s 2025-11-03T17:23:25.2362059Z (EngineCore_DP0 pid=34922) (RayWorkerWrapper pid=35812) WARNING 11-03 17:23:20 [worker_base.py:307] Missing `shared_worker_lock` argument from executor. This argument is needed for mm_processor_cache_type='shm'. 2025-11-03T17:23:26.1570290Z (EngineCore_DP0 pid=34922) (RayWorkerWrapper pid=35813) INFO 11-03 17:23:26 [backends.py:179] Directly load the compiled graph(s) for dynamic shape from the cache, took 0.733 s 2025-11-03T17:23:26.1573270Z (EngineCore_DP0 pid=34922) (RayWorkerWrapper pid=35812) [Gloo] Rank 0 is connected to 1 peer ranks. Expected number of connected peer ranks is : 1 [repeated 10x across cluster] (Ray deduplicates logs by default. Set RAY_DEDUP_LOGS=0 to disable log deduplication, or see https://docs.ray.io/en/master/ray-observability/user-guides/configure-logging.html#log-deduplication for more options.) 2025-11-03T17:23:26.1575567Z (EngineCore_DP0 pid=34922) (RayWorkerWrapper pid=35812) INFO 11-03 17:23:20 [__init__.py:1146] Found nccl from library libnccl.so.2 [repeated 3x across cluster] 2025-11-03T17:23:26.1576836Z (EngineCore_DP0 pid=34922) (RayWorkerWrapper pid=35812) INFO 11-03 17:23:20 [pynccl.py:109] vLLM is using nccl==2.27.5 [repeated 3x across cluster] 2025-11-03T17:23:26.1578094Z (EngineCore_DP0 pid=34922) (RayWorkerWrapper pid=35812) WARNING 11-03 17:23:20 [symm_mem.py:60] SymmMemCommunicator: Device capability 8.9 not supported, communicator is not available. 2025-11-03T17:23:26.1579350Z (EngineCore_DP0 pid=34922) (RayWorkerWrapper pid=35812) INFO 11-03 17:23:20 [custom_all_reduce.py:37] Skipping P2P check and trusting the driver's P2P report. 2025-11-03T17:23:26.1580697Z (EngineCore_DP0 pid=34922) (RayWorkerWrapper pid=35812) WARNING 11-03 17:23:20 [custom_all_reduce.py:162] Custom allreduce is disabled because your platform lacks GPU P2P capability or P2P test failed. To silence this warning, specify disable_custom_all_reduce=True explicitly. 2025-11-03T17:23:26.1582382Z (EngineCore_DP0 pid=34922) (RayWorkerWrapper pid=35812) INFO 11-03 17:23:20 [parallel_state.py:1231] rank 0 in world size 2 is assigned as DP rank 0, PP rank 0, TP rank 0, EP rank 0 2025-11-03T17:23:26.1583319Z (EngineCore_DP0 pid=34922) (RayWorkerWrapper pid=35812) INFO 11-03 17:23:21 [topk_topp_sampler.py:59] Using FlashInfer for top-p & top-k sampling. 2025-11-03T17:23:26.1584234Z (EngineCore_DP0 pid=34922) (RayWorkerWrapper pid=35812) INFO 11-03 17:23:21 [gpu_model_runner.py:2840] Starting to load model meta-llama/Llama-3.2-1B-Instruct... 2025-11-03T17:23:27.4862191Z (EngineCore_DP0 pid=34922) (RayWorkerWrapper pid=35813) INFO 11-03 17:23:27 [gpu_worker.py:314] Available KV cache memory: 18.54 GiB 2025-11-03T17:23:27.4863055Z (EngineCore_DP0 pid=34922) (RayWorkerWrapper pid=35812) INFO 11-03 17:23:21 [gpu_model_runner.py:2870] Loading model from scratch... 2025-11-03T17:23:27.4864176Z (EngineCore_DP0 pid=34922) (RayWorkerWrapper pid=35812) INFO 11-03 17:23:21 [cuda.py:404] Using Flash Attention backend on V1 engine. 2025-11-03T17:23:27.4865092Z (EngineCore_DP0 pid=34922) (RayWorkerWrapper pid=35812) INFO 11-03 17:23:21 [gpu_model_runner.py:2902] Model loading took 1.1667 GiB and 0.169340 seconds 2025-11-03T17:23:27.6982557Z (EngineCore_DP0 pid=34922) INFO 11-03 17:23:27 [kv_cache_utils.py:1199] GPU KV cache size: 1,215,360 tokens 2025-11-03T17:23:27.6983533Z (EngineCore_DP0 pid=34922) INFO 11-03 17:23:27 [kv_cache_utils.py:1204] Maximum concurrency for 2,048 tokens per request: 593.44x 2025-11-03T17:23:27.6984384Z (EngineCore_DP0 pid=34922) INFO 11-03 17:23:27 [kv_cache_utils.py:1199] GPU KV cache size: 1,215,360 tokens 2025-11-03T17:23:27.6985239Z (EngineCore_DP0 pid=34922) INFO 11-03 17:23:27 [kv_cache_utils.py:1204] Maximum concurrency for 2,048 tokens per request: 593.44x 2025-11-03T17:23:27.7950135Z (EngineCore_DP0 pid=34922) (RayWorkerWrapper pid=35813) INFO 11-03 17:23:27 [gpu_worker.py:351] Compile and warming up model for size 8 2025-11-03T17:23:28.3069762Z (EngineCore_DP0 pid=34922) (RayWorkerWrapper pid=35813) INFO 11-03 17:23:28 [backends.py:185] Directly load the compiled graph(s) for shape 8 from the cache, took 0.522 s 2025-11-03T17:23:29.3010362Z (EngineCore_DP0 pid=34922) INFO 11-03 17:23:29 [core.py:240] init engine (profile, create kv cache, warmup model) took 7.34 seconds 2025-11-03T17:23:29.3298138Z (EngineCore_DP0 pid=34922) (RayWorkerWrapper pid=35813) INFO 11-03 17:23:29 [monitor.py:33] torch.compile takes 2.96 s in total 2025-11-03T17:23:29.8104334Z (EngineCore_DP0 pid=34922) INFO 11-03 17:23:29 [vllm.py:354] Cudagraph is disabled under eager mode 2025-11-03T17:23:30.0021187Z (APIServer pid=34800) INFO 11-03 17:23:30 [loggers.py:172] Engine 000: vllm cache_config_info with initialization after num_gpu_blocks is: 75960 2025-11-03T17:23:30.0628759Z (EngineCore_DP0 pid=34922) INFO 11-03 17:23:30 [gc_utils.py:40] GC Debug Config. enabled:False,top_objects:-1 2025-11-03T17:23:30.0861658Z (APIServer pid=34800) INFO 11-03 17:23:30 [api_server.py:1628] Supported tasks: ['generate'] 2025-11-03T17:23:30.1317534Z (APIServer pid=34800) WARNING 11-03 17:23:30 [model.py:1581] Default sampling parameters have been overridden by the model's Hugging Face generation config recommended from the model creator. If this is not intended, please relaunch vLLM instance with `--generation-config vllm`. 2025-11-03T17:23:30.1320764Z (APIServer pid=34800) INFO 11-03 17:23:30 [serving_responses.py:157] Using default chat sampling params from model: {'temperature': 0.6, 'top_p': 0.9} 2025-11-03T17:23:30.1586414Z (APIServer pid=34800) INFO 11-03 17:23:30 [serving_chat.py:159] Using default chat sampling params from model: {'temperature': 0.6, 'top_p': 0.9} 2025-11-03T17:23:30.1798332Z (APIServer pid=34800) INFO 11-03 17:23:30 [serving_completion.py:67] Using default completion sampling params from model: {'temperature': 0.6, 'top_p': 0.9} 2025-11-03T17:23:30.1801142Z (APIServer pid=34800) INFO 11-03 17:23:30 [api_server.py:1939] Starting vLLM API server 0 on http://0.0.0.0:52467 2025-11-03T17:23:30.1802867Z (APIServer pid=34800) INFO 11-03 17:23:30 [launcher.py:38] Available routes are: 2025-11-03T17:23:30.1804328Z (APIServer pid=34800) INFO 11-03 17:23:30 [launcher.py:46] Route: /openapi.json, Methods: GET, HEAD 2025-11-03T17:23:30.1805020Z (APIServer pid=34800) INFO 11-03 17:23:30 [launcher.py:46] Route: /docs, Methods: GET, HEAD 2025-11-03T17:23:30.1805608Z (APIServer pid=34800) INFO 11-03 17:23:30 [launcher.py:46] Route: /docs/oauth2-redirect, Methods: GET, HEAD 2025-11-03T17:23:30.1806196Z (APIServer pid=34800) INFO 11-03 17:23:30 [launcher.py:46] Route: /redoc, Methods: GET, HEAD 2025-11-03T17:23:30.1808760Z (APIServer pid=34800) INFO 11-03 17:23:30 [launcher.py:46] Route: /health, Methods: GET 2025-11-03T17:23:30.1809379Z (APIServer pid=34800) INFO 11-03 17:23:30 [launcher.py:46] Route: /load, Methods: GET 2025-11-03T17:23:30.1810255Z (APIServer pid=34800) INFO 11-03 17:23:30 [launcher.py:46] Route: /ping, Methods: POST 2025-11-03T17:23:30.1810771Z (APIServer pid=34800) INFO 11-03 17:23:30 [launcher.py:46] Route: /ping, Methods: GET 2025-11-03T17:23:30.1811292Z (APIServer pid=34800) INFO 11-03 17:23:30 [launcher.py:46] Route: /tokenize, Methods: POST 2025-11-03T17:23:30.1811839Z (APIServer pid=34800) INFO 11-03 17:23:30 [launcher.py:46] Route: /detokenize, Methods: POST 2025-11-03T17:23:30.1812360Z (APIServer pid=34800) INFO 11-03 17:23:30 [launcher.py:46] Route: /v1/models, Methods: GET 2025-11-03T17:23:30.1812869Z (APIServer pid=34800) INFO 11-03 17:23:30 [launcher.py:46] Route: /version, Methods: GET 2025-11-03T17:23:30.1813429Z (APIServer pid=34800) INFO 11-03 17:23:30 [launcher.py:46] Route: /v1/responses, Methods: POST 2025-11-03T17:23:30.1814009Z (APIServer pid=34800) INFO 11-03 17:23:30 [launcher.py:46] Route: /v1/responses/{response_id}, Methods: GET 2025-11-03T17:23:30.1814642Z (APIServer pid=34800) INFO 11-03 17:23:30 [launcher.py:46] Route: /v1/responses/{response_id}/cancel, Methods: POST 2025-11-03T17:23:30.1815258Z (APIServer pid=34800) INFO 11-03 17:23:30 [launcher.py:46] Route: /v1/chat/completions, Methods: POST 2025-11-03T17:23:30.1815816Z (APIServer pid=34800) INFO 11-03 17:23:30 [launcher.py:46] Route: /v1/completions, Methods: POST 2025-11-03T17:23:30.1816376Z (APIServer pid=34800) INFO 11-03 17:23:30 [launcher.py:46] Route: /v1/embeddings, Methods: POST 2025-11-03T17:23:30.1816919Z (APIServer pid=34800) INFO 11-03 17:23:30 [launcher.py:46] Route: /pooling, Methods: POST 2025-11-03T17:23:30.1817443Z (APIServer pid=34800) INFO 11-03 17:23:30 [launcher.py:46] Route: /classify, Methods: POST 2025-11-03T17:23:30.1817965Z (APIServer pid=34800) INFO 11-03 17:23:30 [launcher.py:46] Route: /score, Methods: POST 2025-11-03T17:23:30.1818462Z (APIServer pid=34800) INFO 11-03 17:23:30 [launcher.py:46] Route: /v1/score, Methods: POST 2025-11-03T17:23:30.1819019Z (APIServer pid=34800) INFO 11-03 17:23:30 [launcher.py:46] Route: /v1/audio/transcriptions, Methods: POST 2025-11-03T17:23:30.1819609Z (APIServer pid=34800) INFO 11-03 17:23:30 [launcher.py:46] Route: /v1/audio/translations, Methods: POST 2025-11-03T17:23:30.1820157Z (APIServer pid=34800) INFO 11-03 17:23:30 [launcher.py:46] Route: /rerank, Methods: POST 2025-11-03T17:23:30.1820676Z (APIServer pid=34800) INFO 11-03 17:23:30 [launcher.py:46] Route: /v1/rerank, Methods: POST 2025-11-03T17:23:30.1821353Z (APIServer pid=34800) INFO 11-03 17:23:30 [launcher.py:46] Route: /v2/rerank, Methods: POST 2025-11-03T17:23:30.1821890Z (APIServer pid=34800) INFO 11-03 17:23:30 [launcher.py:46] Route: /scale_elastic_ep, Methods: POST 2025-11-03T17:23:30.1822461Z (APIServer pid=34800) INFO 11-03 17:23:30 [launcher.py:46] Route: /is_scaling_elastic_ep, Methods: POST 2025-11-03T17:23:30.1823020Z (APIServer pid=34800) INFO 11-03 17:23:30 [launcher.py:46] Route: /invocations, Methods: POST 2025-11-03T17:23:30.1823548Z (APIServer pid=34800) INFO 11-03 17:23:30 [launcher.py:46] Route: /metrics, Methods: GET 2025-11-03T17:23:30.1954910Z (APIServer pid=34800) INFO: Started server process [34800] 2025-11-03T17:23:30.1955569Z (APIServer pid=34800) INFO: Waiting for application startup. 2025-11-03T17:23:30.4130705Z (APIServer pid=34800) INFO: Application startup complete. 2025-11-03T17:23:30.8770266Z (APIServer pid=34800) INFO: 127.0.0.1:49170 - "GET /health HTTP/1.1" 200 OK 2025-11-03T17:23:31.1253749Z (APIServer pid=34800) INFO: 127.0.0.1:49174 - "GET /v1/models HTTP/1.1" 200 OK 2025-11-03T17:23:31.1550660Z (EngineCore_DP0 pid=34922) INFO 11-03 17:23:31 [ray_distributed_executor.py:569] RAY_CGRAPH_get_timeout is set to 300 2025-11-03T17:23:31.1551634Z (EngineCore_DP0 pid=34922) INFO 11-03 17:23:31 [ray_distributed_executor.py:573] VLLM_USE_RAY_COMPILED_DAG_CHANNEL_TYPE = auto 2025-11-03T17:23:31.1552562Z (EngineCore_DP0 pid=34922) INFO 11-03 17:23:31 [ray_distributed_executor.py:577] VLLM_USE_RAY_COMPILED_DAG_OVERLAP_COMM = False 2025-11-03T17:23:31.1557981Z (EngineCore_DP0 pid=34922) INFO 11-03 17:23:31 [ray_distributed_executor.py:653] Using RayPPCommunicator (which wraps vLLM _PP GroupCoordinator) for Ray Compiled Graph communication. 2025-11-03T17:23:31.3274601Z (APIServer pid=34800) INFO: 127.0.0.1:49174 - "POST /v1/completions HTTP/1.1" 200 OK 2025-11-03T17:23:31.3859226Z (APIServer pid=34800) INFO: 127.0.0.1:49174 - "POST /v1/completions HTTP/1.1" 200 OK 2025-11-03T17:23:31.5246977Z (APIServer pid=34800) INFO: 127.0.0.1:49174 - "POST /v1/completions HTTP/1.1" 200 OK 2025-11-03T17:23:31.6091981Z (APIServer pid=34800) INFO: 127.0.0.1:49174 - "POST /v1/completions HTTP/1.1" 200 OK 2025-11-03T17:23:31.6664850Z (APIServer pid=34800) INFO: 127.0.0.1:49174 - "POST /v1/completions HTTP/1.1" 200 OK 2025-11-03T17:23:31.6743976Z (APIServer pid=34800) INFO: 127.0.0.1:49174 - "POST /v1/completions HTTP/1.1" 200 OK 2025-11-03T17:23:31.7572789Z (APIServer pid=34800) INFO 11-03 17:23:31 [launcher.py:110] Shutting down FastAPI HTTP server. 2025-11-03T17:23:31.7573367Z *** SIGTERM received at time=1762190611 on cpu 26 *** 2025-11-03T17:23:31.7620542Z PC: @ 0x7fb1c8efd117 (unknown) (unknown) 2025-11-03T17:23:31.7621707Z @ 0x7fb1c8eae520 (unknown) (unknown) 2025-11-03T17:23:31.7622179Z [2025-11-03 17:23:31,762 E 34922 34922] logging.cc:501: *** SIGTERM received at time=1762190611 on cpu 26 *** 2025-11-03T17:23:31.7624062Z [2025-11-03 17:23:31,762 E 34922 34922] logging.cc:501: PC: @ 0x7fb1c8efd117 (unknown) (unknown) 2025-11-03T17:23:31.7625484Z [2025-11-03 17:23:31,762 E 34922 34922] logging.cc:501: @ 0x7fb1c8eae520 (unknown) (unknown) 2025-11-03T17:23:31.7627407Z (EngineCore_DP0 pid=34922) INFO 11-03 17:23:31 [ray_distributed_executor.py:127] Shutting down Ray distributed executor. If you see error log from logging.cc regarding SIGTERM received, please ignore because this is the expected termination process in Ray. 2025-11-03T17:23:31.7629049Z (EngineCore_DP0 pid=34922) 2025-11-03 17:23:31,762 INFO compiled_dag_node.py:2171 -- Tearing down compiled DAG 2025-11-03T17:23:31.7629941Z (EngineCore_DP0 pid=34922) 2025-11-03 17:23:31,762 INFO compiled_dag_node.py:2176 -- Cancelling compiled worker on actor: Actor(RayWorkerWrapper, 149b14ff77c11bd7aab9b04301000000) 2025-11-03T17:23:31.7630945Z (EngineCore_DP0 pid=34922) 2025-11-03 17:23:31,762 INFO compiled_dag_node.py:2176 -- Cancelling compiled worker on actor: Actor(RayWorkerWrapper, eadb8c88a40457954184b81c01000000) 2025-11-03T17:23:31.7670242Z (EngineCore_DP0 pid=34922) 2025-11-03 17:23:31,766 INFO compiled_dag_node.py:2198 -- Waiting for worker tasks to exit 2025-11-03T17:23:31.7671716Z (EngineCore_DP0 pid=34922) 2025-11-03 17:23:31,766 INFO compiled_dag_node.py:2201 -- Teardown complete 2025-11-03T17:23:32.2701204Z (EngineCore_DP0 pid=34922) (RayWorkerWrapper pid=35812) INFO 11-03 17:23:25 [backends.py:594] Using cache directory: /var/lib/jenkins/.cache/vllm/torch_compile_cache/618b881a70/rank_0_0/backbone for vLLM's torch.compile 2025-11-03T17:23:32.2705239Z (EngineCore_DP0 pid=34922) (RayWorkerWrapper pid=35812) INFO 11-03 17:23:25 [backends.py:608] Dynamo bytecode transform time: 3.01 s 2025-11-03T17:23:32.2706829Z (EngineCore_DP0 pid=34922) (RayWorkerWrapper pid=35812) INFO 11-03 17:23:26 [backends.py:179] Directly load the compiled graph(s) for dynamic shape from the cache, took 0.756 s 2025-11-03T17:23:32.2708222Z (EngineCore_DP0 pid=34922) (RayWorkerWrapper pid=35812) INFO 11-03 17:23:27 [gpu_worker.py:314] Available KV cache memory: 18.54 GiB 2025-11-03T17:23:32.2710258Z (EngineCore_DP0 pid=34922) (RayWorkerWrapper pid=35812) INFO 11-03 17:23:28 [gpu_worker.py:351] Compile and warming up model for size 4 [repeated 3x across cluster] 2025-11-03T17:23:32.2711881Z (EngineCore_DP0 pid=34922) (RayWorkerWrapper pid=35812) INFO 11-03 17:23:29 [backends.py:185] Directly load the compiled graph(s) for shape 4 from the cache, took 1.053 s [repeated 3x across cluster] 2025-11-03T17:23:32.2712877Z (EngineCore_DP0 pid=34922) (RayWorkerWrapper pid=35812) INFO 11-03 17:23:29 [monitor.py:33] torch.compile takes 3.01 s in total 2025-11-03T17:23:32.2714432Z (EngineCore_DP0 pid=34922) (pid=35812) /opt/conda/envs/py_3.12/lib/python3.12/site-packages/torch/cuda/__init__.py:65: FutureWarning: The pynvml package is deprecated. Please install nvidia-ml-py instead. If you did not install pynvml directly, please report this to the maintainers of the package that installed pynvml for you. 2025-11-03T17:23:32.2715627Z (EngineCore_DP0 pid=34922) (pid=35812) import pynvml # type: ignore[import] 2025-11-03T17:23:34.1749965Z (APIServer pid=34800) INFO: Shutting down 2025-11-03T17:23:34.2754009Z (APIServer pid=34800) INFO: Waiting for application shutdown. 2025-11-03T17:23:34.2755371Z (APIServer pid=34800) INFO: Application shutdown complete. 2025-11-03T17:23:34.6422421Z INFO 11-03 17:23:34 [model.py:653] Resolved architecture: LlamaForCausalLM 2025-11-03T17:23:34.6423507Z WARNING 11-03 17:23:34 [model.py:1942] Casting torch.bfloat16 to torch.float16. 2025-11-03T17:23:34.6424371Z INFO 11-03 17:23:34 [model.py:1714] Using max model len 2048 2025-11-03T17:23:34.6426603Z Launching RemoteOpenAIServer with: vllm serve meta-llama/Llama-3.2-1B-Instruct --dtype float16 --max-model-len 2048 --max-num-seqs 8 --enforce-eager --tokenizer-mode auto --tensor-parallel-size 2 --distributed-executor-backend mp --load-format dummy --port 36651 --seed 0 2025-11-03T17:23:35.1822061Z /opt/conda/envs/py_3.12/lib/python3.12/site-packages/torch/cuda/__init__.py:65: FutureWarning: The pynvml package is deprecated. Please install nvidia-ml-py instead. If you did not install pynvml directly, please report this to the maintainers of the package that installed pynvml for you. 2025-11-03T17:23:35.1825218Z import pynvml # type: ignore[import] 2025-11-03T17:23:37.4403616Z INFO 11-03 17:23:37 [__init__.py:224] Automatically detected platform cuda. 2025-11-03T17:23:41.1504055Z (APIServer pid=39562) INFO 11-03 17:23:41 [api_server.py:1870] vLLM API server version 0.11.0rc2.dev389+ge51928192.d20251103 2025-11-03T17:23:41.1529109Z (APIServer pid=39562) INFO 11-03 17:23:41 [utils.py:239] non-default args: {'model_tag': 'meta-llama/Llama-3.2-1B-Instruct', 'port': 36651, 'model': 'meta-llama/Llama-3.2-1B-Instruct', 'dtype': 'float16', 'seed': 0, 'max_model_len': 2048, 'enforce_eager': True, 'load_format': 'dummy', 'distributed_executor_backend': 'mp', 'tensor_parallel_size': 2, 'max_num_seqs': 8} 2025-11-03T17:23:42.3417207Z (APIServer pid=39562) INFO 11-03 17:23:42 [model.py:653] Resolved architecture: LlamaForCausalLM 2025-11-03T17:23:42.3417945Z (APIServer pid=39562) `torch_dtype` is deprecated! Use `dtype` instead! 2025-11-03T17:23:42.3418948Z (APIServer pid=39562) WARNING 11-03 17:23:42 [model.py:1942] Casting torch.bfloat16 to torch.float16. 2025-11-03T17:23:42.3419649Z (APIServer pid=39562) INFO 11-03 17:23:42 [model.py:1714] Using max model len 2048 2025-11-03T17:23:42.5082554Z (APIServer pid=39562) INFO 11-03 17:23:42 [scheduler.py:225] Chunked prefill is enabled with max_num_batched_tokens=2048. 2025-11-03T17:23:42.5087658Z (APIServer pid=39562) INFO 11-03 17:23:42 [vllm.py:354] Cudagraph is disabled under eager mode 2025-11-03T17:23:43.4774275Z /opt/conda/envs/py_3.12/lib/python3.12/site-packages/torch/cuda/__init__.py:65: FutureWarning: The pynvml package is deprecated. Please install nvidia-ml-py instead. If you did not install pynvml directly, please report this to the maintainers of the package that installed pynvml for you. 2025-11-03T17:23:43.4776607Z import pynvml # type: ignore[import] 2025-11-03T17:23:45.8357845Z INFO 11-03 17:23:45 [__init__.py:224] Automatically detected platform cuda. 2025-11-03T17:23:49.5713149Z (EngineCore_DP0 pid=39684) INFO 11-03 17:23:49 [core.py:727] Waiting for init message from front-end. 2025-11-03T17:23:49.5973381Z (EngineCore_DP0 pid=39684) INFO 11-03 17:23:49 [core.py:94] Initializing a V1 LLM engine (v0.11.0rc2.dev389+ge51928192.d20251103) with config: model='meta-llama/Llama-3.2-1B-Instruct', speculative_config=None, tokenizer='meta-llama/Llama-3.2-1B-Instruct', skip_tokenizer_init=False, tokenizer_mode=auto, revision=None, tokenizer_revision=None, trust_remote_code=False, dtype=torch.float16, max_seq_len=2048, download_dir=None, load_format=dummy, tensor_parallel_size=2, pipeline_parallel_size=1, data_parallel_size=1, disable_custom_all_reduce=False, quantization=None, enforce_eager=True, kv_cache_dtype=auto, device_config=cuda, structured_outputs_config=StructuredOutputsConfig(backend='auto', disable_fallback=False, disable_any_whitespace=False, disable_additional_properties=False, reasoning_parser=''), observability_config=ObservabilityConfig(show_hidden_metrics_for_version=None, otlp_traces_endpoint=None, collect_detailed_traces=None), seed=0, served_model_name=meta-llama/Llama-3.2-1B-Instruct, enable_prefix_caching=True, chunked_prefill_enabled=True, pooler_config=None, compilation_config={'level': 0, 'debug_dump_path': None, 'cache_dir': '', 'backend': '', 'custom_ops': [], 'splitting_ops': None, 'use_inductor': True, 'compile_sizes': [], 'inductor_compile_config': {'enable_auto_functionalized_v2': False}, 'inductor_passes': {}, 'cudagraph_mode': , 'use_cudagraph': False, 'cudagraph_num_of_warmups': 0, 'cudagraph_capture_sizes': [], 'cudagraph_copy_inputs': False, 'full_cuda_graph': False, 'use_inductor_graph_partition': False, 'pass_config': {}, 'max_capture_size': 0, 'local_cache_dir': None} 2025-11-03T17:23:49.5978807Z (EngineCore_DP0 pid=39684) WARNING 11-03 17:23:49 [multiproc_executor.py:744] Reducing Torch parallelism from 24 threads to 1 to avoid unnecessary CPU contention. Set OMP_NUM_THREADS in the external environment to tune this value as needed. 2025-11-03T17:23:49.5982140Z (EngineCore_DP0 pid=39684) INFO 11-03 17:23:49 [shm_broadcast.py:302] vLLM message queue communication handle: Handle(local_reader_ranks=[0, 1], buffer_handle=(2, 16777216, 10, 'psm_0633473b'), local_subscribe_addr='ipc:///tmp/57584e63-c955-4580-a7e9-2a99f3a0f727', remote_subscribe_addr=None, remote_addr_ipv6=False) 2025-11-03T17:23:50.0801673Z /opt/conda/envs/py_3.12/lib/python3.12/site-packages/torch/cuda/__init__.py:65: FutureWarning: The pynvml package is deprecated. Please install nvidia-ml-py instead. If you did not install pynvml directly, please report this to the maintainers of the package that installed pynvml for you. 2025-11-03T17:23:50.0803083Z import pynvml # type: ignore[import] 2025-11-03T17:23:50.0804681Z /opt/conda/envs/py_3.12/lib/python3.12/site-packages/torch/cuda/__init__.py:65: FutureWarning: The pynvml package is deprecated. Please install nvidia-ml-py instead. If you did not install pynvml directly, please report this to the maintainers of the package that installed pynvml for you. 2025-11-03T17:23:50.0805907Z import pynvml # type: ignore[import] 2025-11-03T17:23:52.4361890Z INFO 11-03 17:23:52 [__init__.py:224] Automatically detected platform cuda. 2025-11-03T17:23:52.4500814Z INFO 11-03 17:23:52 [__init__.py:224] Automatically detected platform cuda. 2025-11-03T17:23:57.2330575Z INFO 11-03 17:23:57 [shm_broadcast.py:302] vLLM message queue communication handle: Handle(local_reader_ranks=[0], buffer_handle=(1, 10485760, 10, 'psm_42a93daf'), local_subscribe_addr='ipc:///tmp/60bf55fa-7376-411e-a13e-1863c1375cdf', remote_subscribe_addr=None, remote_addr_ipv6=False) 2025-11-03T17:23:57.2525991Z INFO 11-03 17:23:57 [shm_broadcast.py:302] vLLM message queue communication handle: Handle(local_reader_ranks=[0], buffer_handle=(1, 10485760, 10, 'psm_46785ded'), local_subscribe_addr='ipc:///tmp/c2f45d93-7d80-4ac7-804f-3d8e099d76e9', remote_subscribe_addr=None, remote_addr_ipv6=False) 2025-11-03T17:23:57.4548418Z [Gloo] Rank 1 is connected to 1 peer ranks. Expected number of connected peer ranks is : 1 2025-11-03T17:23:57.4549039Z [Gloo] Rank 0 is connected to 1 peer ranks. Expected number of connected peer ranks is : 1 2025-11-03T17:23:57.4590255Z [Gloo] Rank 1 is connected to 1 peer ranks. Expected number of connected peer ranks is : 1 2025-11-03T17:23:57.4591472Z [Gloo] Rank 0 is connected to 1 peer ranks. Expected number of connected peer ranks is : 1 2025-11-03T17:23:57.4634071Z INFO 11-03 17:23:57 [__init__.py:1146] Found nccl from library libnccl.so.2 2025-11-03T17:23:57.4634964Z INFO 11-03 17:23:57 [__init__.py:1146] Found nccl from library libnccl.so.2 2025-11-03T17:23:57.4636109Z INFO 11-03 17:23:57 [pynccl.py:109] vLLM is using nccl==2.27.5 2025-11-03T17:23:57.4636845Z INFO 11-03 17:23:57 [pynccl.py:109] vLLM is using nccl==2.27.5 2025-11-03T17:23:57.6383226Z WARNING 11-03 17:23:57 [symm_mem.py:60] SymmMemCommunicator: Device capability 8.9 not supported, communicator is not available. 2025-11-03T17:23:57.6384866Z WARNING 11-03 17:23:57 [symm_mem.py:60] SymmMemCommunicator: Device capability 8.9 not supported, communicator is not available. 2025-11-03T17:23:57.6411747Z INFO 11-03 17:23:57 [custom_all_reduce.py:37] Skipping P2P check and trusting the driver's P2P report. 2025-11-03T17:23:57.6412782Z WARNING 11-03 17:23:57 [custom_all_reduce.py:162] Custom allreduce is disabled because your platform lacks GPU P2P capability or P2P test failed. To silence this warning, specify disable_custom_all_reduce=True explicitly. 2025-11-03T17:23:57.6413792Z INFO 11-03 17:23:57 [custom_all_reduce.py:37] Skipping P2P check and trusting the driver's P2P report. 2025-11-03T17:23:57.6414785Z WARNING 11-03 17:23:57 [custom_all_reduce.py:162] Custom allreduce is disabled because your platform lacks GPU P2P capability or P2P test failed. To silence this warning, specify disable_custom_all_reduce=True explicitly. 2025-11-03T17:23:57.7325932Z INFO 11-03 17:23:57 [shm_broadcast.py:302] vLLM message queue communication handle: Handle(local_reader_ranks=[1], buffer_handle=(1, 4194304, 6, 'psm_0e351f27'), local_subscribe_addr='ipc:///tmp/36d547a1-cd25-452c-93c5-0b2939352aff', remote_subscribe_addr=None, remote_addr_ipv6=False) 2025-11-03T17:23:57.7349663Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:23:57.7353279Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:23:57.7358485Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:23:57.7365105Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:23:57.8222997Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:23:57.8231579Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:23:57.8243959Z [Gloo] Rank 0 is connected to 1 peer ranks. Expected number of connected peer ranks is : 1 2025-11-03T17:23:57.8245137Z [Gloo] Rank 1 is connected to 1 peer ranks. Expected number of connected peer ranks is : 1 2025-11-03T17:23:57.8246167Z INFO 11-03 17:23:57 [__init__.py:1146] Found nccl from library libnccl.so.2 2025-11-03T17:23:57.8247016Z INFO 11-03 17:23:57 [__init__.py:1146] Found nccl from library libnccl.so.2 2025-11-03T17:23:57.8247781Z INFO 11-03 17:23:57 [pynccl.py:109] vLLM is using nccl==2.27.5 2025-11-03T17:23:57.8248387Z INFO 11-03 17:23:57 [pynccl.py:109] vLLM is using nccl==2.27.5 2025-11-03T17:23:57.8572000Z INFO 11-03 17:23:57 [parallel_state.py:1231] rank 0 in world size 2 is assigned as DP rank 0, PP rank 0, TP rank 0, EP rank 0 2025-11-03T17:23:57.8572735Z INFO 11-03 17:23:57 [parallel_state.py:1231] rank 1 in world size 2 is assigned as DP rank 0, PP rank 0, TP rank 1, EP rank 1 2025-11-03T17:23:58.0854045Z INFO 11-03 17:23:58 [topk_topp_sampler.py:59] Using FlashInfer for top-p & top-k sampling. 2025-11-03T17:23:58.0950589Z INFO 11-03 17:23:58 [topk_topp_sampler.py:59] Using FlashInfer for top-p & top-k sampling. 2025-11-03T17:23:58.1175491Z (Worker_TP1 pid=39804) INFO 11-03 17:23:58 [gpu_model_runner.py:2840] Starting to load model meta-llama/Llama-3.2-1B-Instruct... 2025-11-03T17:23:58.1267450Z (Worker_TP0 pid=39803) INFO 11-03 17:23:58 [gpu_model_runner.py:2840] Starting to load model meta-llama/Llama-3.2-1B-Instruct... 2025-11-03T17:23:58.3275747Z (Worker_TP1 pid=39804) INFO 11-03 17:23:58 [gpu_model_runner.py:2870] Loading model from scratch... 2025-11-03T17:23:58.3336095Z (Worker_TP0 pid=39803) INFO 11-03 17:23:58 [gpu_model_runner.py:2870] Loading model from scratch... 2025-11-03T17:23:58.4174774Z (Worker_TP1 pid=39804) INFO 11-03 17:23:58 [cuda.py:404] Using Flash Attention backend on V1 engine. 2025-11-03T17:23:58.4231652Z (Worker_TP0 pid=39803) INFO 11-03 17:23:58 [cuda.py:404] Using Flash Attention backend on V1 engine. 2025-11-03T17:23:58.8563555Z (Worker_TP0 pid=39803) INFO 11-03 17:23:58 [gpu_model_runner.py:2902] Model loading took 1.1667 GiB and 0.119255 seconds 2025-11-03T17:23:58.8594897Z (Worker_TP1 pid=39804) INFO 11-03 17:23:58 [gpu_model_runner.py:2902] Model loading took 1.1667 GiB and 0.119457 seconds 2025-11-03T17:24:00.7379390Z (Worker_TP0 pid=39803) INFO 11-03 17:24:00 [gpu_worker.py:314] Available KV cache memory: 18.53 GiB 2025-11-03T17:24:00.7449638Z (Worker_TP1 pid=39804) INFO 11-03 17:24:00 [gpu_worker.py:314] Available KV cache memory: 18.53 GiB 2025-11-03T17:24:00.9527880Z (EngineCore_DP0 pid=39684) INFO 11-03 17:24:00 [kv_cache_utils.py:1199] GPU KV cache size: 1,214,464 tokens 2025-11-03T17:24:00.9528939Z (EngineCore_DP0 pid=39684) INFO 11-03 17:24:00 [kv_cache_utils.py:1204] Maximum concurrency for 2,048 tokens per request: 593.00x 2025-11-03T17:24:00.9530266Z (EngineCore_DP0 pid=39684) INFO 11-03 17:24:00 [kv_cache_utils.py:1199] GPU KV cache size: 1,214,464 tokens 2025-11-03T17:24:00.9531125Z (EngineCore_DP0 pid=39684) INFO 11-03 17:24:00 [kv_cache_utils.py:1204] Maximum concurrency for 2,048 tokens per request: 593.00x 2025-11-03T17:24:01.0752457Z (EngineCore_DP0 pid=39684) INFO 11-03 17:24:01 [core.py:240] init engine (profile, create kv cache, warmup model) took 2.17 seconds 2025-11-03T17:24:01.5851218Z (EngineCore_DP0 pid=39684) INFO 11-03 17:24:01 [vllm.py:354] Cudagraph is disabled under eager mode 2025-11-03T17:24:01.7704136Z (APIServer pid=39562) INFO 11-03 17:24:01 [loggers.py:172] Engine 000: vllm cache_config_info with initialization after num_gpu_blocks is: 75904 2025-11-03T17:24:01.7815535Z (EngineCore_DP0 pid=39684) INFO 11-03 17:24:01 [gc_utils.py:40] GC Debug Config. enabled:False,top_objects:-1 2025-11-03T17:24:01.8117961Z (APIServer pid=39562) INFO 11-03 17:24:01 [api_server.py:1628] Supported tasks: ['generate'] 2025-11-03T17:24:01.8382411Z (APIServer pid=39562) WARNING 11-03 17:24:01 [model.py:1581] Default sampling parameters have been overridden by the model's Hugging Face generation config recommended from the model creator. If this is not intended, please relaunch vLLM instance with `--generation-config vllm`. 2025-11-03T17:24:01.8383980Z (APIServer pid=39562) INFO 11-03 17:24:01 [serving_responses.py:157] Using default chat sampling params from model: {'temperature': 0.6, 'top_p': 0.9} 2025-11-03T17:24:01.8629058Z (APIServer pid=39562) INFO 11-03 17:24:01 [serving_chat.py:159] Using default chat sampling params from model: {'temperature': 0.6, 'top_p': 0.9} 2025-11-03T17:24:01.8871340Z (APIServer pid=39562) INFO 11-03 17:24:01 [serving_completion.py:67] Using default completion sampling params from model: {'temperature': 0.6, 'top_p': 0.9} 2025-11-03T17:24:01.8872363Z (APIServer pid=39562) INFO 11-03 17:24:01 [api_server.py:1939] Starting vLLM API server 0 on http://0.0.0.0:36651 2025-11-03T17:24:01.8873085Z (APIServer pid=39562) INFO 11-03 17:24:01 [launcher.py:38] Available routes are: 2025-11-03T17:24:01.8873758Z (APIServer pid=39562) INFO 11-03 17:24:01 [launcher.py:46] Route: /openapi.json, Methods: HEAD, GET 2025-11-03T17:24:01.8874458Z (APIServer pid=39562) INFO 11-03 17:24:01 [launcher.py:46] Route: /docs, Methods: HEAD, GET 2025-11-03T17:24:01.8875160Z (APIServer pid=39562) INFO 11-03 17:24:01 [launcher.py:46] Route: /docs/oauth2-redirect, Methods: HEAD, GET 2025-11-03T17:24:01.8875860Z (APIServer pid=39562) INFO 11-03 17:24:01 [launcher.py:46] Route: /redoc, Methods: HEAD, GET 2025-11-03T17:24:01.8876493Z (APIServer pid=39562) INFO 11-03 17:24:01 [launcher.py:46] Route: /health, Methods: GET 2025-11-03T17:24:01.8877113Z (APIServer pid=39562) INFO 11-03 17:24:01 [launcher.py:46] Route: /load, Methods: GET 2025-11-03T17:24:01.8877728Z (APIServer pid=39562) INFO 11-03 17:24:01 [launcher.py:46] Route: /ping, Methods: POST 2025-11-03T17:24:01.8878361Z (APIServer pid=39562) INFO 11-03 17:24:01 [launcher.py:46] Route: /ping, Methods: GET 2025-11-03T17:24:01.8878981Z (APIServer pid=39562) INFO 11-03 17:24:01 [launcher.py:46] Route: /tokenize, Methods: POST 2025-11-03T17:24:01.8879640Z (APIServer pid=39562) INFO 11-03 17:24:01 [launcher.py:46] Route: /detokenize, Methods: POST 2025-11-03T17:24:01.8880315Z (APIServer pid=39562) INFO 11-03 17:24:01 [launcher.py:46] Route: /v1/models, Methods: GET 2025-11-03T17:24:01.8881060Z (APIServer pid=39562) INFO 11-03 17:24:01 [launcher.py:46] Route: /version, Methods: GET 2025-11-03T17:24:01.8881806Z (APIServer pid=39562) INFO 11-03 17:24:01 [launcher.py:46] Route: /v1/responses, Methods: POST 2025-11-03T17:24:01.8882737Z (APIServer pid=39562) INFO 11-03 17:24:01 [launcher.py:46] Route: /v1/responses/{response_id}, Methods: GET 2025-11-03T17:24:01.8883382Z (APIServer pid=39562) INFO 11-03 17:24:01 [launcher.py:46] Route: /v1/responses/{response_id}/cancel, Methods: POST 2025-11-03T17:24:01.8883987Z (APIServer pid=39562) INFO 11-03 17:24:01 [launcher.py:46] Route: /v1/chat/completions, Methods: POST 2025-11-03T17:24:01.8884554Z (APIServer pid=39562) INFO 11-03 17:24:01 [launcher.py:46] Route: /v1/completions, Methods: POST 2025-11-03T17:24:01.8885099Z (APIServer pid=39562) INFO 11-03 17:24:01 [launcher.py:46] Route: /v1/embeddings, Methods: POST 2025-11-03T17:24:01.8885622Z (APIServer pid=39562) INFO 11-03 17:24:01 [launcher.py:46] Route: /pooling, Methods: POST 2025-11-03T17:24:01.8886131Z (APIServer pid=39562) INFO 11-03 17:24:01 [launcher.py:46] Route: /classify, Methods: POST 2025-11-03T17:24:01.8886638Z (APIServer pid=39562) INFO 11-03 17:24:01 [launcher.py:46] Route: /score, Methods: POST 2025-11-03T17:24:01.8887290Z (APIServer pid=39562) INFO 11-03 17:24:01 [launcher.py:46] Route: /v1/score, Methods: POST 2025-11-03T17:24:01.8887861Z (APIServer pid=39562) INFO 11-03 17:24:01 [launcher.py:46] Route: /v1/audio/transcriptions, Methods: POST 2025-11-03T17:24:01.8888529Z (APIServer pid=39562) INFO 11-03 17:24:01 [launcher.py:46] Route: /v1/audio/translations, Methods: POST 2025-11-03T17:24:01.8889059Z (APIServer pid=39562) INFO 11-03 17:24:01 [launcher.py:46] Route: /rerank, Methods: POST 2025-11-03T17:24:01.8889574Z (APIServer pid=39562) INFO 11-03 17:24:01 [launcher.py:46] Route: /v1/rerank, Methods: POST 2025-11-03T17:24:01.8890086Z (APIServer pid=39562) INFO 11-03 17:24:01 [launcher.py:46] Route: /v2/rerank, Methods: POST 2025-11-03T17:24:01.8890625Z (APIServer pid=39562) INFO 11-03 17:24:01 [launcher.py:46] Route: /scale_elastic_ep, Methods: POST 2025-11-03T17:24:01.8891204Z (APIServer pid=39562) INFO 11-03 17:24:01 [launcher.py:46] Route: /is_scaling_elastic_ep, Methods: POST 2025-11-03T17:24:01.8891752Z (APIServer pid=39562) INFO 11-03 17:24:01 [launcher.py:46] Route: /invocations, Methods: POST 2025-11-03T17:24:01.8892265Z (APIServer pid=39562) INFO 11-03 17:24:01 [launcher.py:46] Route: /metrics, Methods: GET 2025-11-03T17:24:01.9097153Z (APIServer pid=39562) INFO: Started server process [39562] 2025-11-03T17:24:01.9097789Z (APIServer pid=39562) INFO: Waiting for application startup. 2025-11-03T17:24:02.1187977Z (APIServer pid=39562) INFO: Application startup complete. 2025-11-03T17:24:02.2134239Z (APIServer pid=39562) INFO: 127.0.0.1:38080 - "GET /health HTTP/1.1" 200 OK 2025-11-03T17:24:02.2232585Z (APIServer pid=39562) INFO: 127.0.0.1:38082 - "GET /v1/models HTTP/1.1" 200 OK 2025-11-03T17:24:02.3973182Z (APIServer pid=39562) INFO: 127.0.0.1:38082 - "POST /v1/completions HTTP/1.1" 200 OK 2025-11-03T17:24:02.4534344Z (APIServer pid=39562) INFO: 127.0.0.1:38082 - "POST /v1/completions HTTP/1.1" 200 OK 2025-11-03T17:24:02.5496813Z (APIServer pid=39562) INFO: 127.0.0.1:38082 - "POST /v1/completions HTTP/1.1" 200 OK 2025-11-03T17:24:02.6238319Z (APIServer pid=39562) INFO: 127.0.0.1:38082 - "POST /v1/completions HTTP/1.1" 200 OK 2025-11-03T17:24:02.6837468Z (APIServer pid=39562) INFO: 127.0.0.1:38082 - "POST /v1/completions HTTP/1.1" 200 OK 2025-11-03T17:24:02.6910641Z (APIServer pid=39562) INFO: 127.0.0.1:38082 - "POST /v1/completions HTTP/1.1" 200 OK 2025-11-03T17:24:02.7565834Z (APIServer pid=39562) INFO 11-03 17:24:02 [launcher.py:110] Shutting down FastAPI HTTP server. 2025-11-03T17:24:02.7568353Z (Worker_TP0 pid=39803) INFO 11-03 17:24:02 [multiproc_executor.py:583] Parent process exited, terminating worker 2025-11-03T17:24:02.7569262Z (Worker_TP0 pid=39803) INFO 11-03 17:24:02 [multiproc_executor.py:624] WorkerProc shutting down. 2025-11-03T17:24:02.7574511Z (Worker_TP1 pid=39804) INFO 11-03 17:24:02 [multiproc_executor.py:583] Parent process exited, terminating worker 2025-11-03T17:24:02.7575456Z (Worker_TP1 pid=39804) INFO 11-03 17:24:02 [multiproc_executor.py:624] WorkerProc shutting down. 2025-11-03T17:24:04.5722538Z (APIServer pid=39562) INFO: Shutting down 2025-11-03T17:24:04.6718823Z (APIServer pid=39562) INFO: Waiting for application shutdown. 2025-11-03T17:24:04.6721440Z (APIServer pid=39562) INFO: Application shutdown complete. 2025-11-03T17:24:05.2311354Z PASSED 2025-11-03T17:24:05.2424797Z distributed/test_sequence_parallel.py::test_tp_sp_generation[meta-llama/Llama-3.2-1B-Instruct-parallel_setup10-mp-auto-test_options10] Fork a new process to run a test 40007 2025-11-03T17:24:05.2438461Z Fork a new process to run a test 0 2025-11-03T17:24:05.8637227Z INFO 11-03 17:24:05 [model.py:653] Resolved architecture: LlamaForCausalLM 2025-11-03T17:24:05.8638081Z `torch_dtype` is deprecated! Use `dtype` instead! 2025-11-03T17:24:05.8638921Z WARNING 11-03 17:24:05 [model.py:1942] Casting torch.bfloat16 to torch.float16. 2025-11-03T17:24:05.8639731Z INFO 11-03 17:24:05 [model.py:1714] Using max model len 2048 2025-11-03T17:24:05.8644969Z Launching RemoteOpenAIServer with: vllm serve meta-llama/Llama-3.2-1B-Instruct --dtype float16 --max-model-len 2048 --max-num-seqs 8 --enable-chunked-prefill --enforce-eager --tokenizer-mode auto --tensor-parallel-size 2 --pipeline-parallel-size 1 --distributed-executor-backend mp --compilation_config {"level": 3, "custom_ops": ["+rms_norm"], "compile_sizes": [4, 8], "pass_config": {"enable_sequence_parallelism": true, "enable_fusion": false, "enable_noop": true}} --load-format dummy --port 55149 --seed 0 2025-11-03T17:24:06.3825380Z /opt/conda/envs/py_3.12/lib/python3.12/site-packages/torch/cuda/__init__.py:65: FutureWarning: The pynvml package is deprecated. Please install nvidia-ml-py instead. If you did not install pynvml directly, please report this to the maintainers of the package that installed pynvml for you. 2025-11-03T17:24:06.3826757Z import pynvml # type: ignore[import] 2025-11-03T17:24:08.6485988Z INFO 11-03 17:24:08 [__init__.py:224] Automatically detected platform cuda. 2025-11-03T17:24:12.3481569Z (APIServer pid=40056) INFO 11-03 17:24:12 [api_server.py:1870] vLLM API server version 0.11.0rc2.dev389+ge51928192.d20251103 2025-11-03T17:24:12.3510818Z (APIServer pid=40056) INFO 11-03 17:24:12 [utils.py:239] non-default args: {'model_tag': 'meta-llama/Llama-3.2-1B-Instruct', 'port': 55149, 'model': 'meta-llama/Llama-3.2-1B-Instruct', 'dtype': 'float16', 'seed': 0, 'max_model_len': 2048, 'enforce_eager': True, 'load_format': 'dummy', 'distributed_executor_backend': 'mp', 'tensor_parallel_size': 2, 'max_num_seqs': 8, 'enable_chunked_prefill': True, 'compilation_config': {'level': 3, 'debug_dump_path': None, 'cache_dir': '', 'backend': '', 'custom_ops': ['+rms_norm'], 'splitting_ops': None, 'use_inductor': True, 'compile_sizes': [4, 8], 'inductor_compile_config': {'enable_auto_functionalized_v2': False}, 'inductor_passes': {}, 'cudagraph_mode': None, 'use_cudagraph': True, 'cudagraph_num_of_warmups': 0, 'cudagraph_capture_sizes': None, 'cudagraph_copy_inputs': False, 'full_cuda_graph': False, 'use_inductor_graph_partition': False, 'pass_config': {'enable_noop': True, 'enable_sequence_parallelism': True}, 'max_capture_size': None, 'local_cache_dir': None}} 2025-11-03T17:24:12.5266784Z (APIServer pid=40056) INFO 11-03 17:24:12 [model.py:653] Resolved architecture: LlamaForCausalLM 2025-11-03T17:24:12.5268737Z (APIServer pid=40056) `torch_dtype` is deprecated! Use `dtype` instead! 2025-11-03T17:24:12.5270063Z (APIServer pid=40056) WARNING 11-03 17:24:12 [model.py:1942] Casting torch.bfloat16 to torch.float16. 2025-11-03T17:24:12.5271077Z (APIServer pid=40056) INFO 11-03 17:24:12 [model.py:1714] Using max model len 2048 2025-11-03T17:24:12.6895104Z (APIServer pid=40056) INFO 11-03 17:24:12 [scheduler.py:225] Chunked prefill is enabled with max_num_batched_tokens=2048. 2025-11-03T17:24:12.6901536Z (APIServer pid=40056) INFO 11-03 17:24:12 [vllm.py:354] Cudagraph is disabled under eager mode 2025-11-03T17:24:13.6619180Z /opt/conda/envs/py_3.12/lib/python3.12/site-packages/torch/cuda/__init__.py:65: FutureWarning: The pynvml package is deprecated. Please install nvidia-ml-py instead. If you did not install pynvml directly, please report this to the maintainers of the package that installed pynvml for you. 2025-11-03T17:24:13.6620562Z import pynvml # type: ignore[import] 2025-11-03T17:24:16.0032053Z INFO 11-03 17:24:16 [__init__.py:224] Automatically detected platform cuda. 2025-11-03T17:24:19.6929105Z (EngineCore_DP0 pid=40178) INFO 11-03 17:24:19 [core.py:727] Waiting for init message from front-end. 2025-11-03T17:24:19.7196513Z (EngineCore_DP0 pid=40178) INFO 11-03 17:24:19 [core.py:94] Initializing a V1 LLM engine (v0.11.0rc2.dev389+ge51928192.d20251103) with config: model='meta-llama/Llama-3.2-1B-Instruct', speculative_config=None, tokenizer='meta-llama/Llama-3.2-1B-Instruct', skip_tokenizer_init=False, tokenizer_mode=auto, revision=None, tokenizer_revision=None, trust_remote_code=False, dtype=torch.float16, max_seq_len=2048, download_dir=None, load_format=dummy, tensor_parallel_size=2, pipeline_parallel_size=1, data_parallel_size=1, disable_custom_all_reduce=False, quantization=None, enforce_eager=True, kv_cache_dtype=auto, device_config=cuda, structured_outputs_config=StructuredOutputsConfig(backend='auto', disable_fallback=False, disable_any_whitespace=False, disable_additional_properties=False, reasoning_parser=''), observability_config=ObservabilityConfig(show_hidden_metrics_for_version=None, otlp_traces_endpoint=None, collect_detailed_traces=None), seed=0, served_model_name=meta-llama/Llama-3.2-1B-Instruct, enable_prefix_caching=True, chunked_prefill_enabled=True, pooler_config=None, compilation_config={'level': 3, 'debug_dump_path': None, 'cache_dir': '', 'backend': '', 'custom_ops': ['+rms_norm', '+rms_norm'], 'splitting_ops': ['vllm.unified_attention', 'vllm.unified_attention_with_output', 'vllm.unified_mla_attention', 'vllm.unified_mla_attention_with_output', 'vllm.mamba_mixer2', 'vllm.mamba_mixer', 'vllm.short_conv', 'vllm.linear_attention', 'vllm.plamo2_mamba_mixer', 'vllm.gdn_attention', 'vllm.sparse_attn_indexer'], 'use_inductor': True, 'compile_sizes': [8, 4], 'inductor_compile_config': {'enable_auto_functionalized_v2': False}, 'inductor_passes': {}, 'cudagraph_mode': , 'use_cudagraph': False, 'cudagraph_num_of_warmups': 0, 'cudagraph_capture_sizes': [], 'cudagraph_copy_inputs': False, 'full_cuda_graph': False, 'use_inductor_graph_partition': False, 'pass_config': {'enable_noop': True, 'enable_sequence_parallelism': True}, 'max_capture_size': 0, 'local_cache_dir': None} 2025-11-03T17:24:19.7202568Z (EngineCore_DP0 pid=40178) WARNING 11-03 17:24:19 [multiproc_executor.py:744] Reducing Torch parallelism from 24 threads to 1 to avoid unnecessary CPU contention. Set OMP_NUM_THREADS in the external environment to tune this value as needed. 2025-11-03T17:24:19.7204194Z (EngineCore_DP0 pid=40178) INFO 11-03 17:24:19 [shm_broadcast.py:302] vLLM message queue communication handle: Handle(local_reader_ranks=[0, 1], buffer_handle=(2, 16777216, 10, 'psm_39261b90'), local_subscribe_addr='ipc:///tmp/66bdbf9f-b653-4926-8c5a-3d9e0c00c9ab', remote_subscribe_addr=None, remote_addr_ipv6=False) 2025-11-03T17:24:20.1995163Z /opt/conda/envs/py_3.12/lib/python3.12/site-packages/torch/cuda/__init__.py:65: FutureWarning: The pynvml package is deprecated. Please install nvidia-ml-py instead. If you did not install pynvml directly, please report this to the maintainers of the package that installed pynvml for you. 2025-11-03T17:24:20.1996512Z import pynvml # type: ignore[import] 2025-11-03T17:24:20.2180805Z /opt/conda/envs/py_3.12/lib/python3.12/site-packages/torch/cuda/__init__.py:65: FutureWarning: The pynvml package is deprecated. Please install nvidia-ml-py instead. If you did not install pynvml directly, please report this to the maintainers of the package that installed pynvml for you. 2025-11-03T17:24:20.2181964Z import pynvml # type: ignore[import] 2025-11-03T17:24:22.5810902Z INFO 11-03 17:24:22 [__init__.py:224] Automatically detected platform cuda. 2025-11-03T17:24:22.5837301Z INFO 11-03 17:24:22 [__init__.py:224] Automatically detected platform cuda. 2025-11-03T17:24:27.3629759Z INFO 11-03 17:24:27 [shm_broadcast.py:302] vLLM message queue communication handle: Handle(local_reader_ranks=[0], buffer_handle=(1, 10485760, 10, 'psm_87a5ad79'), local_subscribe_addr='ipc:///tmp/c450ba50-caa0-4b2a-ae9f-c581d245bb6f', remote_subscribe_addr=None, remote_addr_ipv6=False) 2025-11-03T17:24:27.3631305Z INFO 11-03 17:24:27 [shm_broadcast.py:302] vLLM message queue communication handle: Handle(local_reader_ranks=[0], buffer_handle=(1, 10485760, 10, 'psm_ab599e6e'), local_subscribe_addr='ipc:///tmp/87152440-613c-41db-b7df-aa1e7fe1a846', remote_subscribe_addr=None, remote_addr_ipv6=False) 2025-11-03T17:24:27.5692096Z [Gloo] Rank 1 is connected to 1 peer ranks. Expected number of connected peer ranks is : 1 2025-11-03T17:24:27.5693334Z [Gloo] Rank 0 is connected to 1 peer ranks. Expected number of connected peer ranks is : 1 2025-11-03T17:24:27.5739744Z [Gloo] Rank 1 is connected to 1 peer ranks. Expected number of connected peer ranks is : 1 2025-11-03T17:24:27.5740813Z [Gloo] Rank 0 is connected to 1 peer ranks. Expected number of connected peer ranks is : 1 2025-11-03T17:24:27.5784743Z INFO 11-03 17:24:27 [__init__.py:1146] Found nccl from library libnccl.so.2 2025-11-03T17:24:27.5785694Z INFO 11-03 17:24:27 [__init__.py:1146] Found nccl from library libnccl.so.2 2025-11-03T17:24:27.5786997Z INFO 11-03 17:24:27 [pynccl.py:109] vLLM is using nccl==2.27.5 2025-11-03T17:24:27.5787881Z INFO 11-03 17:24:27 [pynccl.py:109] vLLM is using nccl==2.27.5 2025-11-03T17:24:27.7522896Z WARNING 11-03 17:24:27 [symm_mem.py:60] SymmMemCommunicator: Device capability 8.9 not supported, communicator is not available. 2025-11-03T17:24:27.7524180Z WARNING 11-03 17:24:27 [symm_mem.py:60] SymmMemCommunicator: Device capability 8.9 not supported, communicator is not available. 2025-11-03T17:24:27.7546708Z INFO 11-03 17:24:27 [custom_all_reduce.py:37] Skipping P2P check and trusting the driver's P2P report. 2025-11-03T17:24:27.7547327Z INFO 11-03 17:24:27 [custom_all_reduce.py:37] Skipping P2P check and trusting the driver's P2P report. 2025-11-03T17:24:27.7548333Z WARNING 11-03 17:24:27 [custom_all_reduce.py:162] Custom allreduce is disabled because your platform lacks GPU P2P capability or P2P test failed. To silence this warning, specify disable_custom_all_reduce=True explicitly. 2025-11-03T17:24:27.7549701Z WARNING 11-03 17:24:27 [custom_all_reduce.py:162] Custom allreduce is disabled because your platform lacks GPU P2P capability or P2P test failed. To silence this warning, specify disable_custom_all_reduce=True explicitly. 2025-11-03T17:24:27.8624439Z INFO 11-03 17:24:27 [shm_broadcast.py:302] vLLM message queue communication handle: Handle(local_reader_ranks=[1], buffer_handle=(1, 4194304, 6, 'psm_9496ae4c'), local_subscribe_addr='ipc:///tmp/ec541d0b-dfa9-421e-8f63-5d353795990a', remote_subscribe_addr=None, remote_addr_ipv6=False) 2025-11-03T17:24:27.8661400Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:24:27.8661987Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:24:27.8670849Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:24:27.8676767Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:24:27.9121714Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:24:27.9129858Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:24:27.9141922Z [Gloo] Rank 1 is connected to 1 peer ranks. Expected number of connected peer ranks is : 1 2025-11-03T17:24:27.9143049Z [Gloo] Rank 0 is connected to 1 peer ranks. Expected number of connected peer ranks is : 1 2025-11-03T17:24:27.9144042Z INFO 11-03 17:24:27 [__init__.py:1146] Found nccl from library libnccl.so.2 2025-11-03T17:24:27.9144483Z INFO 11-03 17:24:27 [__init__.py:1146] Found nccl from library libnccl.so.2 2025-11-03T17:24:27.9144804Z INFO 11-03 17:24:27 [pynccl.py:109] vLLM is using nccl==2.27.5 2025-11-03T17:24:27.9145144Z INFO 11-03 17:24:27 [pynccl.py:109] vLLM is using nccl==2.27.5 2025-11-03T17:24:27.9484391Z INFO 11-03 17:24:27 [parallel_state.py:1231] rank 0 in world size 2 is assigned as DP rank 0, PP rank 0, TP rank 0, EP rank 0 2025-11-03T17:24:27.9485184Z INFO 11-03 17:24:27 [parallel_state.py:1231] rank 1 in world size 2 is assigned as DP rank 0, PP rank 0, TP rank 1, EP rank 1 2025-11-03T17:24:28.1669829Z INFO 11-03 17:24:28 [topk_topp_sampler.py:59] Using FlashInfer for top-p & top-k sampling. 2025-11-03T17:24:28.1874672Z INFO 11-03 17:24:28 [topk_topp_sampler.py:59] Using FlashInfer for top-p & top-k sampling. 2025-11-03T17:24:28.1987299Z (Worker_TP1 pid=40298) INFO 11-03 17:24:28 [gpu_model_runner.py:2840] Starting to load model meta-llama/Llama-3.2-1B-Instruct... 2025-11-03T17:24:28.2193455Z (Worker_TP0 pid=40297) INFO 11-03 17:24:28 [gpu_model_runner.py:2840] Starting to load model meta-llama/Llama-3.2-1B-Instruct... 2025-11-03T17:24:28.4042517Z (Worker_TP1 pid=40298) INFO 11-03 17:24:28 [gpu_model_runner.py:2870] Loading model from scratch... 2025-11-03T17:24:28.4243093Z (Worker_TP0 pid=40297) INFO 11-03 17:24:28 [gpu_model_runner.py:2870] Loading model from scratch... 2025-11-03T17:24:28.4926966Z (Worker_TP1 pid=40298) INFO 11-03 17:24:28 [cuda.py:404] Using Flash Attention backend on V1 engine. 2025-11-03T17:24:28.5137229Z (Worker_TP0 pid=40297) INFO 11-03 17:24:28 [cuda.py:404] Using Flash Attention backend on V1 engine. 2025-11-03T17:24:28.9757708Z (Worker_TP1 pid=40298) INFO 11-03 17:24:28 [gpu_model_runner.py:2902] Model loading took 1.1667 GiB and 0.167711 seconds 2025-11-03T17:24:28.9964603Z (Worker_TP0 pid=40297) INFO 11-03 17:24:28 [gpu_model_runner.py:2902] Model loading took 1.1667 GiB and 0.166499 seconds 2025-11-03T17:24:32.4714566Z (Worker_TP1 pid=40298) INFO 11-03 17:24:32 [backends.py:594] Using cache directory: /var/lib/jenkins/.cache/vllm/torch_compile_cache/618b881a70/rank_1_0/backbone for vLLM's torch.compile 2025-11-03T17:24:32.4723309Z (Worker_TP1 pid=40298) INFO 11-03 17:24:32 [backends.py:608] Dynamo bytecode transform time: 3.18 s 2025-11-03T17:24:32.4875873Z (Worker_TP0 pid=40297) INFO 11-03 17:24:32 [backends.py:594] Using cache directory: /var/lib/jenkins/.cache/vllm/torch_compile_cache/618b881a70/rank_0_0/backbone for vLLM's torch.compile 2025-11-03T17:24:32.4886777Z (Worker_TP0 pid=40297) INFO 11-03 17:24:32 [backends.py:608] Dynamo bytecode transform time: 3.20 s 2025-11-03T17:24:33.5394189Z (Worker_TP1 pid=40298) INFO 11-03 17:24:33 [backends.py:179] Directly load the compiled graph(s) for dynamic shape from the cache, took 0.794 s 2025-11-03T17:24:33.5675500Z (Worker_TP0 pid=40297) INFO 11-03 17:24:33 [backends.py:179] Directly load the compiled graph(s) for dynamic shape from the cache, took 0.802 s 2025-11-03T17:24:35.0091453Z (Worker_TP1 pid=40298) INFO 11-03 17:24:35 [gpu_worker.py:314] Available KV cache memory: 18.54 GiB 2025-11-03T17:24:35.0105202Z (Worker_TP0 pid=40297) INFO 11-03 17:24:35 [gpu_worker.py:314] Available KV cache memory: 18.54 GiB 2025-11-03T17:24:35.2492020Z (EngineCore_DP0 pid=40178) INFO 11-03 17:24:35 [kv_cache_utils.py:1199] GPU KV cache size: 1,215,232 tokens 2025-11-03T17:24:35.2492899Z (EngineCore_DP0 pid=40178) INFO 11-03 17:24:35 [kv_cache_utils.py:1204] Maximum concurrency for 2,048 tokens per request: 593.38x 2025-11-03T17:24:35.2493755Z (EngineCore_DP0 pid=40178) INFO 11-03 17:24:35 [kv_cache_utils.py:1199] GPU KV cache size: 1,215,232 tokens 2025-11-03T17:24:35.2494579Z (EngineCore_DP0 pid=40178) INFO 11-03 17:24:35 [kv_cache_utils.py:1204] Maximum concurrency for 2,048 tokens per request: 593.38x 2025-11-03T17:24:35.2559152Z (Worker_TP1 pid=40298) INFO 11-03 17:24:35 [gpu_worker.py:351] Compile and warming up model for size 8 2025-11-03T17:24:35.2559923Z (Worker_TP0 pid=40297) INFO 11-03 17:24:35 [gpu_worker.py:351] Compile and warming up model for size 8 2025-11-03T17:24:35.7750765Z (Worker_TP1 pid=40298) INFO 11-03 17:24:35 [backends.py:185] Directly load the compiled graph(s) for shape 8 from the cache, took 0.519 s 2025-11-03T17:24:35.7778273Z (Worker_TP0 pid=40297) INFO 11-03 17:24:35 [backends.py:185] Directly load the compiled graph(s) for shape 8 from the cache, took 0.522 s 2025-11-03T17:24:35.7784570Z (Worker_TP1 pid=40298) INFO 11-03 17:24:35 [gpu_worker.py:351] Compile and warming up model for size 4 2025-11-03T17:24:35.7787197Z (Worker_TP0 pid=40297) INFO 11-03 17:24:35 [gpu_worker.py:351] Compile and warming up model for size 4 2025-11-03T17:24:36.2291192Z (Worker_TP1 pid=40298) INFO 11-03 17:24:36 [backends.py:185] Directly load the compiled graph(s) for shape 4 from the cache, took 0.450 s 2025-11-03T17:24:36.2292052Z (Worker_TP1 pid=40298) INFO 11-03 17:24:36 [monitor.py:33] torch.compile takes 3.18 s in total 2025-11-03T17:24:36.2345587Z (Worker_TP0 pid=40297) INFO 11-03 17:24:36 [backends.py:185] Directly load the compiled graph(s) for shape 4 from the cache, took 0.456 s 2025-11-03T17:24:36.2346409Z (Worker_TP0 pid=40297) INFO 11-03 17:24:36 [monitor.py:33] torch.compile takes 3.20 s in total 2025-11-03T17:24:36.2471832Z (EngineCore_DP0 pid=40178) INFO 11-03 17:24:36 [core.py:240] init engine (profile, create kv cache, warmup model) took 7.15 seconds 2025-11-03T17:24:41.6242361Z (EngineCore_DP0 pid=40178) INFO 11-03 17:24:41 [vllm.py:354] Cudagraph is disabled under eager mode 2025-11-03T17:24:41.8087713Z (APIServer pid=40056) INFO 11-03 17:24:41 [loggers.py:172] Engine 000: vllm cache_config_info with initialization after num_gpu_blocks is: 75952 2025-11-03T17:24:41.8203606Z (EngineCore_DP0 pid=40178) INFO 11-03 17:24:41 [gc_utils.py:40] GC Debug Config. enabled:False,top_objects:-1 2025-11-03T17:24:41.8417956Z (APIServer pid=40056) INFO 11-03 17:24:41 [api_server.py:1628] Supported tasks: ['generate'] 2025-11-03T17:24:41.8705717Z (APIServer pid=40056) WARNING 11-03 17:24:41 [model.py:1581] Default sampling parameters have been overridden by the model's Hugging Face generation config recommended from the model creator. If this is not intended, please relaunch vLLM instance with `--generation-config vllm`. 2025-11-03T17:24:41.8708576Z (APIServer pid=40056) INFO 11-03 17:24:41 [serving_responses.py:157] Using default chat sampling params from model: {'temperature': 0.6, 'top_p': 0.9} 2025-11-03T17:24:41.8924816Z (APIServer pid=40056) INFO 11-03 17:24:41 [serving_chat.py:159] Using default chat sampling params from model: {'temperature': 0.6, 'top_p': 0.9} 2025-11-03T17:24:42.0602168Z (APIServer pid=40056) INFO 11-03 17:24:42 [serving_completion.py:67] Using default completion sampling params from model: {'temperature': 0.6, 'top_p': 0.9} 2025-11-03T17:24:42.0604707Z (APIServer pid=40056) INFO 11-03 17:24:42 [api_server.py:1939] Starting vLLM API server 0 on http://0.0.0.0:55149 2025-11-03T17:24:42.0606115Z (APIServer pid=40056) INFO 11-03 17:24:42 [launcher.py:38] Available routes are: 2025-11-03T17:24:42.0606773Z (APIServer pid=40056) INFO 11-03 17:24:42 [launcher.py:46] Route: /openapi.json, Methods: HEAD, GET 2025-11-03T17:24:42.0607340Z (APIServer pid=40056) INFO 11-03 17:24:42 [launcher.py:46] Route: /docs, Methods: HEAD, GET 2025-11-03T17:24:42.0607910Z (APIServer pid=40056) INFO 11-03 17:24:42 [launcher.py:46] Route: /docs/oauth2-redirect, Methods: HEAD, GET 2025-11-03T17:24:42.0608555Z (APIServer pid=40056) INFO 11-03 17:24:42 [launcher.py:46] Route: /redoc, Methods: HEAD, GET 2025-11-03T17:24:42.0609079Z (APIServer pid=40056) INFO 11-03 17:24:42 [launcher.py:46] Route: /health, Methods: GET 2025-11-03T17:24:42.0609752Z (APIServer pid=40056) INFO 11-03 17:24:42 [launcher.py:46] Route: /load, Methods: GET 2025-11-03T17:24:42.0610439Z (APIServer pid=40056) INFO 11-03 17:24:42 [launcher.py:46] Route: /ping, Methods: POST 2025-11-03T17:24:42.0610940Z (APIServer pid=40056) INFO 11-03 17:24:42 [launcher.py:46] Route: /ping, Methods: GET 2025-11-03T17:24:42.0611435Z (APIServer pid=40056) INFO 11-03 17:24:42 [launcher.py:46] Route: /tokenize, Methods: POST 2025-11-03T17:24:42.0611966Z (APIServer pid=40056) INFO 11-03 17:24:42 [launcher.py:46] Route: /detokenize, Methods: POST 2025-11-03T17:24:42.0612502Z (APIServer pid=40056) INFO 11-03 17:24:42 [launcher.py:46] Route: /v1/models, Methods: GET 2025-11-03T17:24:42.0613005Z (APIServer pid=40056) INFO 11-03 17:24:42 [launcher.py:46] Route: /version, Methods: GET 2025-11-03T17:24:42.0613533Z (APIServer pid=40056) INFO 11-03 17:24:42 [launcher.py:46] Route: /v1/responses, Methods: POST 2025-11-03T17:24:42.0614108Z (APIServer pid=40056) INFO 11-03 17:24:42 [launcher.py:46] Route: /v1/responses/{response_id}, Methods: GET 2025-11-03T17:24:42.0614721Z (APIServer pid=40056) INFO 11-03 17:24:42 [launcher.py:46] Route: /v1/responses/{response_id}/cancel, Methods: POST 2025-11-03T17:24:42.0615321Z (APIServer pid=40056) INFO 11-03 17:24:42 [launcher.py:46] Route: /v1/chat/completions, Methods: POST 2025-11-03T17:24:42.0615876Z (APIServer pid=40056) INFO 11-03 17:24:42 [launcher.py:46] Route: /v1/completions, Methods: POST 2025-11-03T17:24:42.0616419Z (APIServer pid=40056) INFO 11-03 17:24:42 [launcher.py:46] Route: /v1/embeddings, Methods: POST 2025-11-03T17:24:42.0616954Z (APIServer pid=40056) INFO 11-03 17:24:42 [launcher.py:46] Route: /pooling, Methods: POST 2025-11-03T17:24:42.0617466Z (APIServer pid=40056) INFO 11-03 17:24:42 [launcher.py:46] Route: /classify, Methods: POST 2025-11-03T17:24:42.0617977Z (APIServer pid=40056) INFO 11-03 17:24:42 [launcher.py:46] Route: /score, Methods: POST 2025-11-03T17:24:42.0618505Z (APIServer pid=40056) INFO 11-03 17:24:42 [launcher.py:46] Route: /v1/score, Methods: POST 2025-11-03T17:24:42.0619069Z (APIServer pid=40056) INFO 11-03 17:24:42 [launcher.py:46] Route: /v1/audio/transcriptions, Methods: POST 2025-11-03T17:24:42.0619655Z (APIServer pid=40056) INFO 11-03 17:24:42 [launcher.py:46] Route: /v1/audio/translations, Methods: POST 2025-11-03T17:24:42.0620182Z (APIServer pid=40056) INFO 11-03 17:24:42 [launcher.py:46] Route: /rerank, Methods: POST 2025-11-03T17:24:42.0620699Z (APIServer pid=40056) INFO 11-03 17:24:42 [launcher.py:46] Route: /v1/rerank, Methods: POST 2025-11-03T17:24:42.0621227Z (APIServer pid=40056) INFO 11-03 17:24:42 [launcher.py:46] Route: /v2/rerank, Methods: POST 2025-11-03T17:24:42.0621765Z (APIServer pid=40056) INFO 11-03 17:24:42 [launcher.py:46] Route: /scale_elastic_ep, Methods: POST 2025-11-03T17:24:42.0622507Z (APIServer pid=40056) INFO 11-03 17:24:42 [launcher.py:46] Route: /is_scaling_elastic_ep, Methods: POST 2025-11-03T17:24:42.0623057Z (APIServer pid=40056) INFO 11-03 17:24:42 [launcher.py:46] Route: /invocations, Methods: POST 2025-11-03T17:24:42.0623573Z (APIServer pid=40056) INFO 11-03 17:24:42 [launcher.py:46] Route: /metrics, Methods: GET 2025-11-03T17:24:42.0759949Z (APIServer pid=40056) INFO: Started server process [40056] 2025-11-03T17:24:42.0761241Z (APIServer pid=40056) INFO: Waiting for application startup. 2025-11-03T17:24:42.2896368Z (APIServer pid=40056) INFO: Application startup complete. 2025-11-03T17:24:42.4543142Z (APIServer pid=40056) INFO: 127.0.0.1:39488 - "GET /health HTTP/1.1" 200 OK 2025-11-03T17:24:42.7047008Z (APIServer pid=40056) INFO: 127.0.0.1:39492 - "GET /v1/models HTTP/1.1" 200 OK 2025-11-03T17:24:42.8102642Z (APIServer pid=40056) INFO: 127.0.0.1:39492 - "POST /v1/completions HTTP/1.1" 200 OK 2025-11-03T17:24:42.8652961Z (APIServer pid=40056) INFO: 127.0.0.1:39492 - "POST /v1/completions HTTP/1.1" 200 OK 2025-11-03T17:24:42.9992719Z (APIServer pid=40056) INFO: 127.0.0.1:39492 - "POST /v1/completions HTTP/1.1" 200 OK 2025-11-03T17:24:43.0679033Z (APIServer pid=40056) INFO: 127.0.0.1:39492 - "POST /v1/completions HTTP/1.1" 200 OK 2025-11-03T17:24:43.1226939Z (APIServer pid=40056) INFO: 127.0.0.1:39492 - "POST /v1/completions HTTP/1.1" 200 OK 2025-11-03T17:24:43.1303575Z (APIServer pid=40056) INFO: 127.0.0.1:39492 - "POST /v1/completions HTTP/1.1" 200 OK 2025-11-03T17:24:43.1898015Z (APIServer pid=40056) INFO 11-03 17:24:43 [launcher.py:110] Shutting down FastAPI HTTP server. 2025-11-03T17:24:43.1901481Z (Worker_TP0 pid=40297) INFO 11-03 17:24:43 [multiproc_executor.py:583] Parent process exited, terminating worker 2025-11-03T17:24:43.1902275Z (Worker_TP0 pid=40297) INFO 11-03 17:24:43 [multiproc_executor.py:624] WorkerProc shutting down. 2025-11-03T17:24:43.1905381Z (Worker_TP1 pid=40298) INFO 11-03 17:24:43 [multiproc_executor.py:583] Parent process exited, terminating worker 2025-11-03T17:24:43.1906795Z (Worker_TP1 pid=40298) INFO 11-03 17:24:43 [multiproc_executor.py:624] WorkerProc shutting down. 2025-11-03T17:24:45.5034528Z (APIServer pid=40056) INFO: Shutting down 2025-11-03T17:24:45.6028631Z (APIServer pid=40056) INFO: Waiting for application shutdown. 2025-11-03T17:24:45.6030776Z (APIServer pid=40056) INFO: Application shutdown complete. 2025-11-03T17:24:45.9321201Z INFO 11-03 17:24:45 [model.py:653] Resolved architecture: LlamaForCausalLM 2025-11-03T17:24:45.9321662Z WARNING 11-03 17:24:45 [model.py:1942] Casting torch.bfloat16 to torch.float16. 2025-11-03T17:24:45.9322015Z INFO 11-03 17:24:45 [model.py:1714] Using max model len 2048 2025-11-03T17:24:45.9323075Z Launching RemoteOpenAIServer with: vllm serve meta-llama/Llama-3.2-1B-Instruct --dtype float16 --max-model-len 2048 --max-num-seqs 8 --enable-chunked-prefill --enforce-eager --tokenizer-mode auto --tensor-parallel-size 2 --distributed-executor-backend mp --load-format dummy --port 52287 --seed 0 2025-11-03T17:24:46.4496566Z /opt/conda/envs/py_3.12/lib/python3.12/site-packages/torch/cuda/__init__.py:65: FutureWarning: The pynvml package is deprecated. Please install nvidia-ml-py instead. If you did not install pynvml directly, please report this to the maintainers of the package that installed pynvml for you. 2025-11-03T17:24:46.4497805Z import pynvml # type: ignore[import] 2025-11-03T17:24:48.6980612Z INFO 11-03 17:24:48 [__init__.py:224] Automatically detected platform cuda. 2025-11-03T17:24:52.3835124Z (APIServer pid=40503) INFO 11-03 17:24:52 [api_server.py:1870] vLLM API server version 0.11.0rc2.dev389+ge51928192.d20251103 2025-11-03T17:24:52.3859831Z (APIServer pid=40503) INFO 11-03 17:24:52 [utils.py:239] non-default args: {'model_tag': 'meta-llama/Llama-3.2-1B-Instruct', 'port': 52287, 'model': 'meta-llama/Llama-3.2-1B-Instruct', 'dtype': 'float16', 'seed': 0, 'max_model_len': 2048, 'enforce_eager': True, 'load_format': 'dummy', 'distributed_executor_backend': 'mp', 'tensor_parallel_size': 2, 'max_num_seqs': 8, 'enable_chunked_prefill': True} 2025-11-03T17:24:52.8510871Z (APIServer pid=40503) INFO 11-03 17:24:52 [model.py:653] Resolved architecture: LlamaForCausalLM 2025-11-03T17:24:52.8512174Z (APIServer pid=40503) `torch_dtype` is deprecated! Use `dtype` instead! 2025-11-03T17:24:52.8513520Z (APIServer pid=40503) WARNING 11-03 17:24:52 [model.py:1942] Casting torch.bfloat16 to torch.float16. 2025-11-03T17:24:52.8515425Z (APIServer pid=40503) INFO 11-03 17:24:52 [model.py:1714] Using max model len 2048 2025-11-03T17:24:53.0119418Z (APIServer pid=40503) INFO 11-03 17:24:53 [scheduler.py:225] Chunked prefill is enabled with max_num_batched_tokens=2048. 2025-11-03T17:24:53.0125367Z (APIServer pid=40503) INFO 11-03 17:24:53 [vllm.py:354] Cudagraph is disabled under eager mode 2025-11-03T17:24:54.0166999Z /opt/conda/envs/py_3.12/lib/python3.12/site-packages/torch/cuda/__init__.py:65: FutureWarning: The pynvml package is deprecated. Please install nvidia-ml-py instead. If you did not install pynvml directly, please report this to the maintainers of the package that installed pynvml for you. 2025-11-03T17:24:54.0168373Z import pynvml # type: ignore[import] 2025-11-03T17:24:56.3727837Z INFO 11-03 17:24:56 [__init__.py:224] Automatically detected platform cuda. 2025-11-03T17:25:00.0978235Z (EngineCore_DP0 pid=40625) INFO 11-03 17:25:00 [core.py:727] Waiting for init message from front-end. 2025-11-03T17:25:00.1254049Z (EngineCore_DP0 pid=40625) INFO 11-03 17:25:00 [core.py:94] Initializing a V1 LLM engine (v0.11.0rc2.dev389+ge51928192.d20251103) with config: model='meta-llama/Llama-3.2-1B-Instruct', speculative_config=None, tokenizer='meta-llama/Llama-3.2-1B-Instruct', skip_tokenizer_init=False, tokenizer_mode=auto, revision=None, tokenizer_revision=None, trust_remote_code=False, dtype=torch.float16, max_seq_len=2048, download_dir=None, load_format=dummy, tensor_parallel_size=2, pipeline_parallel_size=1, data_parallel_size=1, disable_custom_all_reduce=False, quantization=None, enforce_eager=True, kv_cache_dtype=auto, device_config=cuda, structured_outputs_config=StructuredOutputsConfig(backend='auto', disable_fallback=False, disable_any_whitespace=False, disable_additional_properties=False, reasoning_parser=''), observability_config=ObservabilityConfig(show_hidden_metrics_for_version=None, otlp_traces_endpoint=None, collect_detailed_traces=None), seed=0, served_model_name=meta-llama/Llama-3.2-1B-Instruct, enable_prefix_caching=True, chunked_prefill_enabled=True, pooler_config=None, compilation_config={'level': 0, 'debug_dump_path': None, 'cache_dir': '', 'backend': '', 'custom_ops': [], 'splitting_ops': None, 'use_inductor': True, 'compile_sizes': [], 'inductor_compile_config': {'enable_auto_functionalized_v2': False}, 'inductor_passes': {}, 'cudagraph_mode': , 'use_cudagraph': False, 'cudagraph_num_of_warmups': 0, 'cudagraph_capture_sizes': [], 'cudagraph_copy_inputs': False, 'full_cuda_graph': False, 'use_inductor_graph_partition': False, 'pass_config': {}, 'max_capture_size': 0, 'local_cache_dir': None} 2025-11-03T17:25:00.1259100Z (EngineCore_DP0 pid=40625) WARNING 11-03 17:25:00 [multiproc_executor.py:744] Reducing Torch parallelism from 24 threads to 1 to avoid unnecessary CPU contention. Set OMP_NUM_THREADS in the external environment to tune this value as needed. 2025-11-03T17:25:00.1263730Z (EngineCore_DP0 pid=40625) INFO 11-03 17:25:00 [shm_broadcast.py:302] vLLM message queue communication handle: Handle(local_reader_ranks=[0, 1], buffer_handle=(2, 16777216, 10, 'psm_9a78a771'), local_subscribe_addr='ipc:///tmp/973604e7-8c86-47cf-8a2d-91622d9430cc', remote_subscribe_addr=None, remote_addr_ipv6=False) 2025-11-03T17:25:00.6035618Z /opt/conda/envs/py_3.12/lib/python3.12/site-packages/torch/cuda/__init__.py:65: FutureWarning: The pynvml package is deprecated. Please install nvidia-ml-py instead. If you did not install pynvml directly, please report this to the maintainers of the package that installed pynvml for you. 2025-11-03T17:25:00.6036871Z import pynvml # type: ignore[import] 2025-11-03T17:25:00.6284331Z /opt/conda/envs/py_3.12/lib/python3.12/site-packages/torch/cuda/__init__.py:65: FutureWarning: The pynvml package is deprecated. Please install nvidia-ml-py instead. If you did not install pynvml directly, please report this to the maintainers of the package that installed pynvml for you. 2025-11-03T17:25:00.6285542Z import pynvml # type: ignore[import] 2025-11-03T17:25:02.9499854Z INFO 11-03 17:25:02 [__init__.py:224] Automatically detected platform cuda. 2025-11-03T17:25:03.0088686Z INFO 11-03 17:25:03 [__init__.py:224] Automatically detected platform cuda. 2025-11-03T17:25:07.7330156Z INFO 11-03 17:25:07 [shm_broadcast.py:302] vLLM message queue communication handle: Handle(local_reader_ranks=[0], buffer_handle=(1, 10485760, 10, 'psm_9330c61c'), local_subscribe_addr='ipc:///tmp/805018ad-2f51-495a-856e-da2eaf4a7998', remote_subscribe_addr=None, remote_addr_ipv6=False) 2025-11-03T17:25:07.7333746Z INFO 11-03 17:25:07 [shm_broadcast.py:302] vLLM message queue communication handle: Handle(local_reader_ranks=[0], buffer_handle=(1, 10485760, 10, 'psm_7bdb5444'), local_subscribe_addr='ipc:///tmp/4cc7407e-9309-48d9-b895-6d0abbae19fe', remote_subscribe_addr=None, remote_addr_ipv6=False) 2025-11-03T17:25:08.6092581Z [Gloo] Rank 0 is connected to 1 peer ranks. Expected number of connected peer ranks is : 1 2025-11-03T17:25:08.6093798Z [Gloo] Rank 1 is connected to 1 peer ranks. Expected number of connected peer ranks is : 1 2025-11-03T17:25:08.6140697Z [Gloo] Rank 1 is connected to 1 peer ranks. Expected number of connected peer ranks is : 1 2025-11-03T17:25:08.6141279Z [Gloo] Rank 0 is connected to 1 peer ranks. Expected number of connected peer ranks is : 1 2025-11-03T17:25:08.6182310Z INFO 11-03 17:25:08 [__init__.py:1146] Found nccl from library libnccl.so.2 2025-11-03T17:25:08.6183172Z INFO 11-03 17:25:08 [__init__.py:1146] Found nccl from library libnccl.so.2 2025-11-03T17:25:08.6184008Z INFO 11-03 17:25:08 [pynccl.py:109] vLLM is using nccl==2.27.5 2025-11-03T17:25:08.6184718Z INFO 11-03 17:25:08 [pynccl.py:109] vLLM is using nccl==2.27.5 2025-11-03T17:25:08.7898072Z WARNING 11-03 17:25:08 [symm_mem.py:60] SymmMemCommunicator: Device capability 8.9 not supported, communicator is not available. 2025-11-03T17:25:08.7899736Z WARNING 11-03 17:25:08 [symm_mem.py:60] SymmMemCommunicator: Device capability 8.9 not supported, communicator is not available. 2025-11-03T17:25:08.7922247Z INFO 11-03 17:25:08 [custom_all_reduce.py:37] Skipping P2P check and trusting the driver's P2P report. 2025-11-03T17:25:08.7922872Z INFO 11-03 17:25:08 [custom_all_reduce.py:37] Skipping P2P check and trusting the driver's P2P report. 2025-11-03T17:25:08.7923884Z WARNING 11-03 17:25:08 [custom_all_reduce.py:162] Custom allreduce is disabled because your platform lacks GPU P2P capability or P2P test failed. To silence this warning, specify disable_custom_all_reduce=True explicitly. 2025-11-03T17:25:08.7925258Z WARNING 11-03 17:25:08 [custom_all_reduce.py:162] Custom allreduce is disabled because your platform lacks GPU P2P capability or P2P test failed. To silence this warning, specify disable_custom_all_reduce=True explicitly. 2025-11-03T17:25:08.8425414Z INFO 11-03 17:25:08 [shm_broadcast.py:302] vLLM message queue communication handle: Handle(local_reader_ranks=[1], buffer_handle=(1, 4194304, 6, 'psm_2f603fa6'), local_subscribe_addr='ipc:///tmp/98b54536-5bea-4b1f-92df-f842c67aca78', remote_subscribe_addr=None, remote_addr_ipv6=False) 2025-11-03T17:25:08.8456094Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:25:08.8457263Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:25:08.8463664Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:25:08.8473313Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:25:08.8923043Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:25:08.8932399Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:25:08.8944172Z [Gloo] Rank 1 is connected to 1 peer ranks. Expected number of connected peer ranks is : 1 2025-11-03T17:25:08.8945301Z [Gloo] Rank 0 is connected to 1 peer ranks. Expected number of connected peer ranks is : 1 2025-11-03T17:25:08.8946757Z INFO 11-03 17:25:08 [__init__.py:1146] Found nccl from library libnccl.so.2 2025-11-03T17:25:08.8947656Z INFO 11-03 17:25:08 [__init__.py:1146] Found nccl from library libnccl.so.2 2025-11-03T17:25:08.8948451Z INFO 11-03 17:25:08 [pynccl.py:109] vLLM is using nccl==2.27.5 2025-11-03T17:25:08.8949167Z INFO 11-03 17:25:08 [pynccl.py:109] vLLM is using nccl==2.27.5 2025-11-03T17:25:08.9292064Z INFO 11-03 17:25:08 [parallel_state.py:1231] rank 1 in world size 2 is assigned as DP rank 0, PP rank 0, TP rank 1, EP rank 1 2025-11-03T17:25:08.9292828Z INFO 11-03 17:25:08 [parallel_state.py:1231] rank 0 in world size 2 is assigned as DP rank 0, PP rank 0, TP rank 0, EP rank 0 2025-11-03T17:25:09.1420986Z INFO 11-03 17:25:09 [topk_topp_sampler.py:59] Using FlashInfer for top-p & top-k sampling. 2025-11-03T17:25:09.1446629Z INFO 11-03 17:25:09 [topk_topp_sampler.py:59] Using FlashInfer for top-p & top-k sampling. 2025-11-03T17:25:09.1730719Z (Worker_TP1 pid=40745) INFO 11-03 17:25:09 [gpu_model_runner.py:2840] Starting to load model meta-llama/Llama-3.2-1B-Instruct... 2025-11-03T17:25:09.1767829Z (Worker_TP0 pid=40744) INFO 11-03 17:25:09 [gpu_model_runner.py:2840] Starting to load model meta-llama/Llama-3.2-1B-Instruct... 2025-11-03T17:25:09.3783607Z (Worker_TP1 pid=40745) INFO 11-03 17:25:09 [gpu_model_runner.py:2870] Loading model from scratch... 2025-11-03T17:25:09.3844572Z (Worker_TP0 pid=40744) INFO 11-03 17:25:09 [gpu_model_runner.py:2870] Loading model from scratch... 2025-11-03T17:25:09.4684339Z (Worker_TP1 pid=40745) INFO 11-03 17:25:09 [cuda.py:404] Using Flash Attention backend on V1 engine. 2025-11-03T17:25:09.4744990Z (Worker_TP0 pid=40744) INFO 11-03 17:25:09 [cuda.py:404] Using Flash Attention backend on V1 engine. 2025-11-03T17:25:09.9016570Z (Worker_TP1 pid=40745) INFO 11-03 17:25:09 [gpu_model_runner.py:2902] Model loading took 1.1667 GiB and 0.119871 seconds 2025-11-03T17:25:09.9098960Z (Worker_TP0 pid=40744) INFO 11-03 17:25:09 [gpu_model_runner.py:2902] Model loading took 1.1667 GiB and 0.119504 seconds 2025-11-03T17:25:11.7904975Z (Worker_TP1 pid=40745) INFO 11-03 17:25:11 [gpu_worker.py:314] Available KV cache memory: 18.53 GiB 2025-11-03T17:25:11.7911419Z (Worker_TP0 pid=40744) INFO 11-03 17:25:11 [gpu_worker.py:314] Available KV cache memory: 18.53 GiB 2025-11-03T17:25:11.9995551Z (EngineCore_DP0 pid=40625) INFO 11-03 17:25:11 [kv_cache_utils.py:1199] GPU KV cache size: 1,214,464 tokens 2025-11-03T17:25:11.9996439Z (EngineCore_DP0 pid=40625) INFO 11-03 17:25:11 [kv_cache_utils.py:1204] Maximum concurrency for 2,048 tokens per request: 593.00x 2025-11-03T17:25:11.9997268Z (EngineCore_DP0 pid=40625) INFO 11-03 17:25:11 [kv_cache_utils.py:1199] GPU KV cache size: 1,214,464 tokens 2025-11-03T17:25:11.9998381Z (EngineCore_DP0 pid=40625) INFO 11-03 17:25:11 [kv_cache_utils.py:1204] Maximum concurrency for 2,048 tokens per request: 593.00x 2025-11-03T17:25:12.1204937Z (EngineCore_DP0 pid=40625) INFO 11-03 17:25:12 [core.py:240] init engine (profile, create kv cache, warmup model) took 2.16 seconds 2025-11-03T17:25:12.6051528Z (EngineCore_DP0 pid=40625) INFO 11-03 17:25:12 [vllm.py:354] Cudagraph is disabled under eager mode 2025-11-03T17:25:12.7900134Z (APIServer pid=40503) INFO 11-03 17:25:12 [loggers.py:172] Engine 000: vllm cache_config_info with initialization after num_gpu_blocks is: 75904 2025-11-03T17:25:12.7998366Z (EngineCore_DP0 pid=40625) INFO 11-03 17:25:12 [gc_utils.py:40] GC Debug Config. enabled:False,top_objects:-1 2025-11-03T17:25:12.8208379Z (APIServer pid=40503) INFO 11-03 17:25:12 [api_server.py:1628] Supported tasks: ['generate'] 2025-11-03T17:25:12.8514931Z (APIServer pid=40503) WARNING 11-03 17:25:12 [model.py:1581] Default sampling parameters have been overridden by the model's Hugging Face generation config recommended from the model creator. If this is not intended, please relaunch vLLM instance with `--generation-config vllm`. 2025-11-03T17:25:12.8516496Z (APIServer pid=40503) INFO 11-03 17:25:12 [serving_responses.py:157] Using default chat sampling params from model: {'temperature': 0.6, 'top_p': 0.9} 2025-11-03T17:25:12.8751867Z (APIServer pid=40503) INFO 11-03 17:25:12 [serving_chat.py:159] Using default chat sampling params from model: {'temperature': 0.6, 'top_p': 0.9} 2025-11-03T17:25:12.8981857Z (APIServer pid=40503) INFO 11-03 17:25:12 [serving_completion.py:67] Using default completion sampling params from model: {'temperature': 0.6, 'top_p': 0.9} 2025-11-03T17:25:12.8982864Z (APIServer pid=40503) INFO 11-03 17:25:12 [api_server.py:1939] Starting vLLM API server 0 on http://0.0.0.0:52287 2025-11-03T17:25:12.8983595Z (APIServer pid=40503) INFO 11-03 17:25:12 [launcher.py:38] Available routes are: 2025-11-03T17:25:12.8984283Z (APIServer pid=40503) INFO 11-03 17:25:12 [launcher.py:46] Route: /openapi.json, Methods: HEAD, GET 2025-11-03T17:25:12.8984961Z (APIServer pid=40503) INFO 11-03 17:25:12 [launcher.py:46] Route: /docs, Methods: HEAD, GET 2025-11-03T17:25:12.8985664Z (APIServer pid=40503) INFO 11-03 17:25:12 [launcher.py:46] Route: /docs/oauth2-redirect, Methods: HEAD, GET 2025-11-03T17:25:12.8986395Z (APIServer pid=40503) INFO 11-03 17:25:12 [launcher.py:46] Route: /redoc, Methods: HEAD, GET 2025-11-03T17:25:12.8987077Z (APIServer pid=40503) INFO 11-03 17:25:12 [launcher.py:46] Route: /health, Methods: GET 2025-11-03T17:25:12.8987802Z (APIServer pid=40503) INFO 11-03 17:25:12 [launcher.py:46] Route: /load, Methods: GET 2025-11-03T17:25:12.8988437Z (APIServer pid=40503) INFO 11-03 17:25:12 [launcher.py:46] Route: /ping, Methods: POST 2025-11-03T17:25:12.8989084Z (APIServer pid=40503) INFO 11-03 17:25:12 [launcher.py:46] Route: /ping, Methods: GET 2025-11-03T17:25:12.8989715Z (APIServer pid=40503) INFO 11-03 17:25:12 [launcher.py:46] Route: /tokenize, Methods: POST 2025-11-03T17:25:12.8990369Z (APIServer pid=40503) INFO 11-03 17:25:12 [launcher.py:46] Route: /detokenize, Methods: POST 2025-11-03T17:25:12.8991008Z (APIServer pid=40503) INFO 11-03 17:25:12 [launcher.py:46] Route: /v1/models, Methods: GET 2025-11-03T17:25:12.8991634Z (APIServer pid=40503) INFO 11-03 17:25:12 [launcher.py:46] Route: /version, Methods: GET 2025-11-03T17:25:12.8992294Z (APIServer pid=40503) INFO 11-03 17:25:12 [launcher.py:46] Route: /v1/responses, Methods: POST 2025-11-03T17:25:12.8992858Z (APIServer pid=40503) INFO 11-03 17:25:12 [launcher.py:46] Route: /v1/responses/{response_id}, Methods: GET 2025-11-03T17:25:12.8993772Z (APIServer pid=40503) INFO 11-03 17:25:12 [launcher.py:46] Route: /v1/responses/{response_id}/cancel, Methods: POST 2025-11-03T17:25:12.8994389Z (APIServer pid=40503) INFO 11-03 17:25:12 [launcher.py:46] Route: /v1/chat/completions, Methods: POST 2025-11-03T17:25:12.8994953Z (APIServer pid=40503) INFO 11-03 17:25:12 [launcher.py:46] Route: /v1/completions, Methods: POST 2025-11-03T17:25:12.8995496Z (APIServer pid=40503) INFO 11-03 17:25:12 [launcher.py:46] Route: /v1/embeddings, Methods: POST 2025-11-03T17:25:12.8996019Z (APIServer pid=40503) INFO 11-03 17:25:12 [launcher.py:46] Route: /pooling, Methods: POST 2025-11-03T17:25:12.8996540Z (APIServer pid=40503) INFO 11-03 17:25:12 [launcher.py:46] Route: /classify, Methods: POST 2025-11-03T17:25:12.8997053Z (APIServer pid=40503) INFO 11-03 17:25:12 [launcher.py:46] Route: /score, Methods: POST 2025-11-03T17:25:12.8997561Z (APIServer pid=40503) INFO 11-03 17:25:12 [launcher.py:46] Route: /v1/score, Methods: POST 2025-11-03T17:25:12.8998124Z (APIServer pid=40503) INFO 11-03 17:25:12 [launcher.py:46] Route: /v1/audio/transcriptions, Methods: POST 2025-11-03T17:25:12.8998847Z (APIServer pid=40503) INFO 11-03 17:25:12 [launcher.py:46] Route: /v1/audio/translations, Methods: POST 2025-11-03T17:25:12.8999392Z (APIServer pid=40503) INFO 11-03 17:25:12 [launcher.py:46] Route: /rerank, Methods: POST 2025-11-03T17:25:12.8999910Z (APIServer pid=40503) INFO 11-03 17:25:12 [launcher.py:46] Route: /v1/rerank, Methods: POST 2025-11-03T17:25:12.9000431Z (APIServer pid=40503) INFO 11-03 17:25:12 [launcher.py:46] Route: /v2/rerank, Methods: POST 2025-11-03T17:25:12.9000971Z (APIServer pid=40503) INFO 11-03 17:25:12 [launcher.py:46] Route: /scale_elastic_ep, Methods: POST 2025-11-03T17:25:12.9001541Z (APIServer pid=40503) INFO 11-03 17:25:12 [launcher.py:46] Route: /is_scaling_elastic_ep, Methods: POST 2025-11-03T17:25:12.9002092Z (APIServer pid=40503) INFO 11-03 17:25:12 [launcher.py:46] Route: /invocations, Methods: POST 2025-11-03T17:25:12.9002613Z (APIServer pid=40503) INFO 11-03 17:25:12 [launcher.py:46] Route: /metrics, Methods: GET 2025-11-03T17:25:12.9136815Z (APIServer pid=40503) INFO: Started server process [40503] 2025-11-03T17:25:12.9137450Z (APIServer pid=40503) INFO: Waiting for application startup. 2025-11-03T17:25:13.1308637Z (APIServer pid=40503) INFO: Application startup complete. 2025-11-03T17:25:13.5036096Z (APIServer pid=40503) INFO: 127.0.0.1:43318 - "GET /health HTTP/1.1" 200 OK 2025-11-03T17:25:13.5126072Z (APIServer pid=40503) INFO: 127.0.0.1:43328 - "GET /v1/models HTTP/1.1" 200 OK 2025-11-03T17:25:13.6844368Z (APIServer pid=40503) INFO: 127.0.0.1:43328 - "POST /v1/completions HTTP/1.1" 200 OK 2025-11-03T17:25:13.7406789Z (APIServer pid=40503) INFO: 127.0.0.1:43328 - "POST /v1/completions HTTP/1.1" 200 OK 2025-11-03T17:25:13.8287056Z (APIServer pid=40503) INFO: 127.0.0.1:43328 - "POST /v1/completions HTTP/1.1" 200 OK 2025-11-03T17:25:13.9028194Z (APIServer pid=40503) INFO: 127.0.0.1:43328 - "POST /v1/completions HTTP/1.1" 200 OK 2025-11-03T17:25:13.9739456Z (APIServer pid=40503) INFO: 127.0.0.1:43328 - "POST /v1/completions HTTP/1.1" 200 OK 2025-11-03T17:25:13.9810769Z (APIServer pid=40503) INFO: 127.0.0.1:43328 - "POST /v1/completions HTTP/1.1" 200 OK 2025-11-03T17:25:14.0465930Z (APIServer pid=40503) INFO 11-03 17:25:14 [launcher.py:110] Shutting down FastAPI HTTP server. 2025-11-03T17:25:14.0468668Z (Worker_TP0 pid=40744) INFO 11-03 17:25:14 [multiproc_executor.py:583] Parent process exited, terminating worker 2025-11-03T17:25:14.0471421Z (Worker_TP0 pid=40744) INFO 11-03 17:25:14 [multiproc_executor.py:624] WorkerProc shutting down. 2025-11-03T17:25:14.0472764Z (Worker_TP1 pid=40745) INFO 11-03 17:25:14 [multiproc_executor.py:583] Parent process exited, terminating worker 2025-11-03T17:25:14.0473984Z (Worker_TP1 pid=40745) INFO 11-03 17:25:14 [multiproc_executor.py:624] WorkerProc shutting down. 2025-11-03T17:25:16.1622539Z (APIServer pid=40503) INFO: Shutting down 2025-11-03T17:25:16.2627747Z (APIServer pid=40503) INFO: Waiting for application shutdown. 2025-11-03T17:25:16.2629261Z (APIServer pid=40503) INFO: Application shutdown complete. 2025-11-03T17:25:16.8165935Z PASSED 2025-11-03T17:25:16.8281204Z distributed/test_sequence_parallel.py::test_tp_sp_generation[meta-llama/Llama-3.2-1B-Instruct-parallel_setup11-ray-auto-test_options11] Fork a new process to run a test 40948 2025-11-03T17:25:16.8297763Z Fork a new process to run a test 0 2025-11-03T17:25:17.4868188Z INFO 11-03 17:25:17 [model.py:653] Resolved architecture: LlamaForCausalLM 2025-11-03T17:25:17.4869057Z `torch_dtype` is deprecated! Use `dtype` instead! 2025-11-03T17:25:17.4869880Z WARNING 11-03 17:25:17 [model.py:1942] Casting torch.bfloat16 to torch.float16. 2025-11-03T17:25:17.4870698Z INFO 11-03 17:25:17 [model.py:1714] Using max model len 2048 2025-11-03T17:25:17.4875671Z Launching RemoteOpenAIServer with: vllm serve meta-llama/Llama-3.2-1B-Instruct --dtype float16 --max-model-len 2048 --max-num-seqs 8 --enable-chunked-prefill --enforce-eager --tokenizer-mode auto --tensor-parallel-size 2 --pipeline-parallel-size 1 --distributed-executor-backend ray --compilation_config {"level": 3, "custom_ops": ["+rms_norm"], "compile_sizes": [4, 8], "pass_config": {"enable_sequence_parallelism": true, "enable_fusion": false, "enable_noop": true}} --load-format dummy --port 42779 --seed 0 2025-11-03T17:25:18.0047636Z /opt/conda/envs/py_3.12/lib/python3.12/site-packages/torch/cuda/__init__.py:65: FutureWarning: The pynvml package is deprecated. Please install nvidia-ml-py instead. If you did not install pynvml directly, please report this to the maintainers of the package that installed pynvml for you. 2025-11-03T17:25:18.0048912Z import pynvml # type: ignore[import] 2025-11-03T17:25:20.2492702Z INFO 11-03 17:25:20 [__init__.py:224] Automatically detected platform cuda. 2025-11-03T17:25:23.9344491Z (APIServer pid=40997) INFO 11-03 17:25:23 [api_server.py:1870] vLLM API server version 0.11.0rc2.dev389+ge51928192.d20251103 2025-11-03T17:25:23.9373918Z (APIServer pid=40997) INFO 11-03 17:25:23 [utils.py:239] non-default args: {'model_tag': 'meta-llama/Llama-3.2-1B-Instruct', 'port': 42779, 'model': 'meta-llama/Llama-3.2-1B-Instruct', 'dtype': 'float16', 'seed': 0, 'max_model_len': 2048, 'enforce_eager': True, 'load_format': 'dummy', 'distributed_executor_backend': 'ray', 'tensor_parallel_size': 2, 'max_num_seqs': 8, 'enable_chunked_prefill': True, 'compilation_config': {'level': 3, 'debug_dump_path': None, 'cache_dir': '', 'backend': '', 'custom_ops': ['+rms_norm'], 'splitting_ops': None, 'use_inductor': True, 'compile_sizes': [4, 8], 'inductor_compile_config': {'enable_auto_functionalized_v2': False}, 'inductor_passes': {}, 'cudagraph_mode': None, 'use_cudagraph': True, 'cudagraph_num_of_warmups': 0, 'cudagraph_capture_sizes': None, 'cudagraph_copy_inputs': False, 'full_cuda_graph': False, 'use_inductor_graph_partition': False, 'pass_config': {'enable_noop': True, 'enable_sequence_parallelism': True}, 'max_capture_size': None, 'local_cache_dir': None}} 2025-11-03T17:25:24.1884971Z (APIServer pid=40997) INFO 11-03 17:25:24 [model.py:653] Resolved architecture: LlamaForCausalLM 2025-11-03T17:25:24.1886217Z (APIServer pid=40997) `torch_dtype` is deprecated! Use `dtype` instead! 2025-11-03T17:25:24.1888182Z (APIServer pid=40997) WARNING 11-03 17:25:24 [model.py:1942] Casting torch.bfloat16 to torch.float16. 2025-11-03T17:25:24.1889648Z (APIServer pid=40997) INFO 11-03 17:25:24 [model.py:1714] Using max model len 2048 2025-11-03T17:25:24.3465084Z (APIServer pid=40997) INFO 11-03 17:25:24 [scheduler.py:225] Chunked prefill is enabled with max_num_batched_tokens=2048. 2025-11-03T17:25:24.3470919Z (APIServer pid=40997) INFO 11-03 17:25:24 [vllm.py:354] Cudagraph is disabled under eager mode 2025-11-03T17:25:25.3927977Z /opt/conda/envs/py_3.12/lib/python3.12/site-packages/torch/cuda/__init__.py:65: FutureWarning: The pynvml package is deprecated. Please install nvidia-ml-py instead. If you did not install pynvml directly, please report this to the maintainers of the package that installed pynvml for you. 2025-11-03T17:25:25.3929244Z import pynvml # type: ignore[import] 2025-11-03T17:25:27.7381580Z INFO 11-03 17:25:27 [__init__.py:224] Automatically detected platform cuda. 2025-11-03T17:25:31.5889221Z (EngineCore_DP0 pid=41119) INFO 11-03 17:25:31 [core.py:727] Waiting for init message from front-end. 2025-11-03T17:25:31.6161301Z (EngineCore_DP0 pid=41119) INFO 11-03 17:25:31 [core.py:94] Initializing a V1 LLM engine (v0.11.0rc2.dev389+ge51928192.d20251103) with config: model='meta-llama/Llama-3.2-1B-Instruct', speculative_config=None, tokenizer='meta-llama/Llama-3.2-1B-Instruct', skip_tokenizer_init=False, tokenizer_mode=auto, revision=None, tokenizer_revision=None, trust_remote_code=False, dtype=torch.float16, max_seq_len=2048, download_dir=None, load_format=dummy, tensor_parallel_size=2, pipeline_parallel_size=1, data_parallel_size=1, disable_custom_all_reduce=False, quantization=None, enforce_eager=True, kv_cache_dtype=auto, device_config=cuda, structured_outputs_config=StructuredOutputsConfig(backend='auto', disable_fallback=False, disable_any_whitespace=False, disable_additional_properties=False, reasoning_parser=''), observability_config=ObservabilityConfig(show_hidden_metrics_for_version=None, otlp_traces_endpoint=None, collect_detailed_traces=None), seed=0, served_model_name=meta-llama/Llama-3.2-1B-Instruct, enable_prefix_caching=True, chunked_prefill_enabled=True, pooler_config=None, compilation_config={'level': 3, 'debug_dump_path': None, 'cache_dir': '', 'backend': '', 'custom_ops': ['+rms_norm', '+rms_norm'], 'splitting_ops': ['vllm.unified_attention', 'vllm.unified_attention_with_output', 'vllm.unified_mla_attention', 'vllm.unified_mla_attention_with_output', 'vllm.mamba_mixer2', 'vllm.mamba_mixer', 'vllm.short_conv', 'vllm.linear_attention', 'vllm.plamo2_mamba_mixer', 'vllm.gdn_attention', 'vllm.sparse_attn_indexer'], 'use_inductor': True, 'compile_sizes': [8, 4], 'inductor_compile_config': {'enable_auto_functionalized_v2': False}, 'inductor_passes': {}, 'cudagraph_mode': , 'use_cudagraph': False, 'cudagraph_num_of_warmups': 0, 'cudagraph_capture_sizes': [], 'cudagraph_copy_inputs': False, 'full_cuda_graph': False, 'use_inductor_graph_partition': False, 'pass_config': {'enable_noop': True, 'enable_sequence_parallelism': True}, 'max_capture_size': 0, 'local_cache_dir': None} 2025-11-03T17:25:33.5181116Z (EngineCore_DP0 pid=41119) 2025-11-03 17:25:33,517 INFO worker.py:1918 -- Started a local Ray instance. View the dashboard at http://127.0.0.1:8265  2025-11-03T17:25:34.4970420Z (EngineCore_DP0 pid=41119) INFO 11-03 17:25:34 [ray_utils.py:373] No current placement group found. Creating a new placement group. 2025-11-03T17:25:34.6765726Z (EngineCore_DP0 pid=41119) INFO 11-03 17:25:34 [ray_distributed_executor.py:179] use_ray_spmd_worker: True 2025-11-03T17:25:35.6206162Z (EngineCore_DP0 pid=41119) (pid=42007) /opt/conda/envs/py_3.12/lib/python3.12/site-packages/torch/cuda/__init__.py:65: FutureWarning: The pynvml package is deprecated. Please install nvidia-ml-py instead. If you did not install pynvml directly, please report this to the maintainers of the package that installed pynvml for you. 2025-11-03T17:25:35.6208121Z (EngineCore_DP0 pid=41119) (pid=42007) import pynvml # type: ignore[import] 2025-11-03T17:25:37.9710241Z (EngineCore_DP0 pid=41119) (pid=42007) INFO 11-03 17:25:37 [__init__.py:224] Automatically detected platform cuda. 2025-11-03T17:25:39.3697330Z (EngineCore_DP0 pid=41119) INFO 11-03 17:25:39 [ray_env.py:67] RAY_NON_CARRY_OVER_ENV_VARS from config: set() 2025-11-03T17:25:39.3698895Z (EngineCore_DP0 pid=41119) INFO 11-03 17:25:39 [ray_env.py:70] Copying the following environment variables to workers: ['LD_LIBRARY_PATH', 'MAX_JOBS', 'HUGGING_FACE_HUB_TOKEN', 'VLLM_USE_RAY_SPMD_WORKER', 'VLLM_WORKER_MULTIPROC_METHOD', 'VLLM_USE_RAY_COMPILED_DAG', 'HF_TOKEN', 'VLLM_USE_V1'] 2025-11-03T17:25:39.3700535Z (EngineCore_DP0 pid=41119) INFO 11-03 17:25:39 [ray_env.py:75] If certain env vars should NOT be copied, add them to /var/lib/jenkins/.config/vllm/ray_non_carry_over_env_vars.json file 2025-11-03T17:25:42.3598480Z (EngineCore_DP0 pid=41119) (RayWorkerWrapper pid=42007) WARNING 11-03 17:25:42 [worker_base.py:307] Missing `shared_worker_lock` argument from executor. This argument is needed for mm_processor_cache_type='shm'. 2025-11-03T17:25:42.5665153Z (EngineCore_DP0 pid=41119) (RayWorkerWrapper pid=42007) [Gloo] Rank 0 is connected to 1 peer ranks. Expected number of connected peer ranks is : 1 2025-11-03T17:25:42.5666135Z (EngineCore_DP0 pid=41119) (RayWorkerWrapper pid=42007) [Gloo] Rank 0 is connected to 1 peer ranks. Expected number of connected peer ranks is : 1 2025-11-03T17:25:42.5666970Z (EngineCore_DP0 pid=41119) (RayWorkerWrapper pid=42007) INFO 11-03 17:25:42 [__init__.py:1146] Found nccl from library libnccl.so.2 2025-11-03T17:25:42.5667733Z (EngineCore_DP0 pid=41119) (RayWorkerWrapper pid=42007) INFO 11-03 17:25:42 [pynccl.py:109] vLLM is using nccl==2.27.5 2025-11-03T17:25:42.7742909Z (EngineCore_DP0 pid=41119) (RayWorkerWrapper pid=42007) WARNING 11-03 17:25:42 [symm_mem.py:60] SymmMemCommunicator: Device capability 8.9 not supported, communicator is not available. 2025-11-03T17:25:42.7745521Z (EngineCore_DP0 pid=41119) (RayWorkerWrapper pid=42007) INFO 11-03 17:25:42 [custom_all_reduce.py:37] Skipping P2P check and trusting the driver's P2P report. 2025-11-03T17:25:42.7748717Z (EngineCore_DP0 pid=41119) (RayWorkerWrapper pid=42007) WARNING 11-03 17:25:42 [custom_all_reduce.py:162] Custom allreduce is disabled because your platform lacks GPU P2P capability or P2P test failed. To silence this warning, specify disable_custom_all_reduce=True explicitly. 2025-11-03T17:25:42.7750666Z (EngineCore_DP0 pid=41119) (RayWorkerWrapper pid=42007) INFO 11-03 17:25:42 [shm_broadcast.py:302] vLLM message queue communication handle: Handle(local_reader_ranks=[1], buffer_handle=(1, 4194304, 6, 'psm_5b947663'), local_subscribe_addr='ipc:///tmp/18ca80f5-2e6a-46bb-9264-7f7dc04fcc81', remote_subscribe_addr=None, remote_addr_ipv6=False) 2025-11-03T17:25:42.8788540Z (EngineCore_DP0 pid=41119) (RayWorkerWrapper pid=42007) INFO 11-03 17:25:42 [parallel_state.py:1231] rank 0 in world size 2 is assigned as DP rank 0, PP rank 0, TP rank 0, EP rank 0 2025-11-03T17:25:43.1872749Z (EngineCore_DP0 pid=41119) (RayWorkerWrapper pid=42007) INFO 11-03 17:25:43 [topk_topp_sampler.py:59] Using FlashInfer for top-p & top-k sampling. 2025-11-03T17:25:43.1874004Z (EngineCore_DP0 pid=41119) (RayWorkerWrapper pid=42007) INFO 11-03 17:25:43 [gpu_model_runner.py:2840] Starting to load model meta-llama/Llama-3.2-1B-Instruct... 2025-11-03T17:25:43.1875015Z (EngineCore_DP0 pid=41119) (pid=42012) INFO 11-03 17:25:37 [__init__.py:224] Automatically detected platform cuda. 2025-11-03T17:25:43.3934760Z (EngineCore_DP0 pid=41119) (RayWorkerWrapper pid=42007) INFO 11-03 17:25:43 [gpu_model_runner.py:2870] Loading model from scratch... 2025-11-03T17:25:43.4994241Z (EngineCore_DP0 pid=41119) (RayWorkerWrapper pid=42007) INFO 11-03 17:25:43 [cuda.py:404] Using Flash Attention backend on V1 engine. 2025-11-03T17:25:44.0108731Z (EngineCore_DP0 pid=41119) (RayWorkerWrapper pid=42007) INFO 11-03 17:25:43 [gpu_model_runner.py:2902] Model loading took 1.1667 GiB and 0.168032 seconds 2025-11-03T17:25:47.2765267Z (EngineCore_DP0 pid=41119) (RayWorkerWrapper pid=42007) INFO 11-03 17:25:47 [backends.py:594] Using cache directory: /var/lib/jenkins/.cache/vllm/torch_compile_cache/618b881a70/rank_0_0/backbone for vLLM's torch.compile 2025-11-03T17:25:47.2766938Z (EngineCore_DP0 pid=41119) (RayWorkerWrapper pid=42007) INFO 11-03 17:25:47 [backends.py:608] Dynamo bytecode transform time: 2.98 s 2025-11-03T17:25:48.1986248Z (EngineCore_DP0 pid=41119) (RayWorkerWrapper pid=42007) INFO 11-03 17:25:48 [backends.py:179] Directly load the compiled graph(s) for dynamic shape from the cache, took 0.745 s 2025-11-03T17:25:48.1987838Z (EngineCore_DP0 pid=41119) (RayWorkerWrapper pid=42012) WARNING 11-03 17:25:42 [worker_base.py:307] Missing `shared_worker_lock` argument from executor. This argument is needed for mm_processor_cache_type='shm'. 2025-11-03T17:25:48.1990178Z (EngineCore_DP0 pid=41119) (RayWorkerWrapper pid=42012) [Gloo] Rank 1 is connected to 1 peer ranks. Expected number of connected peer ranks is : 1 [repeated 10x across cluster] (Ray deduplicates logs by default. Set RAY_DEDUP_LOGS=0 to disable log deduplication, or see https://docs.ray.io/en/master/ray-observability/user-guides/configure-logging.html#log-deduplication for more options.) 2025-11-03T17:25:48.1992157Z (EngineCore_DP0 pid=41119) (RayWorkerWrapper pid=42012) INFO 11-03 17:25:42 [__init__.py:1146] Found nccl from library libnccl.so.2 [repeated 3x across cluster] 2025-11-03T17:25:48.1993273Z (EngineCore_DP0 pid=41119) (RayWorkerWrapper pid=42012) INFO 11-03 17:25:42 [pynccl.py:109] vLLM is using nccl==2.27.5 [repeated 3x across cluster] 2025-11-03T17:25:48.1994377Z (EngineCore_DP0 pid=41119) (RayWorkerWrapper pid=42012) WARNING 11-03 17:25:42 [symm_mem.py:60] SymmMemCommunicator: Device capability 8.9 not supported, communicator is not available. 2025-11-03T17:25:48.1995483Z (EngineCore_DP0 pid=41119) (RayWorkerWrapper pid=42012) INFO 11-03 17:25:42 [custom_all_reduce.py:37] Skipping P2P check and trusting the driver's P2P report. 2025-11-03T17:25:48.1996784Z (EngineCore_DP0 pid=41119) (RayWorkerWrapper pid=42012) WARNING 11-03 17:25:42 [custom_all_reduce.py:162] Custom allreduce is disabled because your platform lacks GPU P2P capability or P2P test failed. To silence this warning, specify disable_custom_all_reduce=True explicitly. 2025-11-03T17:25:48.1998094Z (EngineCore_DP0 pid=41119) (RayWorkerWrapper pid=42012) INFO 11-03 17:25:42 [parallel_state.py:1231] rank 1 in world size 2 is assigned as DP rank 0, PP rank 0, TP rank 1, EP rank 1 2025-11-03T17:25:48.1999030Z (EngineCore_DP0 pid=41119) (RayWorkerWrapper pid=42012) INFO 11-03 17:25:43 [topk_topp_sampler.py:59] Using FlashInfer for top-p & top-k sampling. 2025-11-03T17:25:48.1999923Z (EngineCore_DP0 pid=41119) (RayWorkerWrapper pid=42012) INFO 11-03 17:25:43 [gpu_model_runner.py:2840] Starting to load model meta-llama/Llama-3.2-1B-Instruct... 2025-11-03T17:25:49.5289832Z (EngineCore_DP0 pid=41119) (RayWorkerWrapper pid=42007) INFO 11-03 17:25:49 [gpu_worker.py:314] Available KV cache memory: 18.54 GiB 2025-11-03T17:25:49.5291523Z (EngineCore_DP0 pid=41119) (RayWorkerWrapper pid=42012) INFO 11-03 17:25:43 [gpu_model_runner.py:2870] Loading model from scratch... 2025-11-03T17:25:49.5293478Z (EngineCore_DP0 pid=41119) (RayWorkerWrapper pid=42012) INFO 11-03 17:25:43 [cuda.py:404] Using Flash Attention backend on V1 engine. 2025-11-03T17:25:49.5294743Z (EngineCore_DP0 pid=41119) (RayWorkerWrapper pid=42012) INFO 11-03 17:25:43 [gpu_model_runner.py:2902] Model loading took 1.1667 GiB and 0.166659 seconds 2025-11-03T17:25:49.7119735Z (EngineCore_DP0 pid=41119) INFO 11-03 17:25:49 [kv_cache_utils.py:1199] GPU KV cache size: 1,215,360 tokens 2025-11-03T17:25:49.7120729Z (EngineCore_DP0 pid=41119) INFO 11-03 17:25:49 [kv_cache_utils.py:1204] Maximum concurrency for 2,048 tokens per request: 593.44x 2025-11-03T17:25:49.7121742Z (EngineCore_DP0 pid=41119) INFO 11-03 17:25:49 [kv_cache_utils.py:1199] GPU KV cache size: 1,215,360 tokens 2025-11-03T17:25:49.7122607Z (EngineCore_DP0 pid=41119) INFO 11-03 17:25:49 [kv_cache_utils.py:1204] Maximum concurrency for 2,048 tokens per request: 593.44x 2025-11-03T17:25:49.7370652Z (EngineCore_DP0 pid=41119) (RayWorkerWrapper pid=42007) INFO 11-03 17:25:49 [gpu_worker.py:351] Compile and warming up model for size 8 2025-11-03T17:25:50.2539256Z (EngineCore_DP0 pid=41119) (RayWorkerWrapper pid=42007) INFO 11-03 17:25:50 [backends.py:185] Directly load the compiled graph(s) for shape 8 from the cache, took 0.525 s 2025-11-03T17:25:50.9583106Z (EngineCore_DP0 pid=41119) INFO 11-03 17:25:50 [core.py:240] init engine (profile, create kv cache, warmup model) took 6.97 seconds 2025-11-03T17:25:50.9727166Z (EngineCore_DP0 pid=41119) (RayWorkerWrapper pid=42007) INFO 11-03 17:25:50 [monitor.py:33] torch.compile takes 2.98 s in total 2025-11-03T17:25:51.4582884Z (EngineCore_DP0 pid=41119) INFO 11-03 17:25:51 [vllm.py:354] Cudagraph is disabled under eager mode 2025-11-03T17:25:51.6504522Z (APIServer pid=40997) INFO 11-03 17:25:51 [loggers.py:172] Engine 000: vllm cache_config_info with initialization after num_gpu_blocks is: 75960 2025-11-03T17:25:51.7066490Z (EngineCore_DP0 pid=41119) INFO 11-03 17:25:51 [gc_utils.py:40] GC Debug Config. enabled:False,top_objects:-1 2025-11-03T17:25:51.7305768Z (APIServer pid=40997) INFO 11-03 17:25:51 [api_server.py:1628] Supported tasks: ['generate'] 2025-11-03T17:25:51.7603448Z (APIServer pid=40997) WARNING 11-03 17:25:51 [model.py:1581] Default sampling parameters have been overridden by the model's Hugging Face generation config recommended from the model creator. If this is not intended, please relaunch vLLM instance with `--generation-config vllm`. 2025-11-03T17:25:51.7606535Z (APIServer pid=40997) INFO 11-03 17:25:51 [serving_responses.py:157] Using default chat sampling params from model: {'temperature': 0.6, 'top_p': 0.9} 2025-11-03T17:25:51.8292646Z (APIServer pid=40997) INFO 11-03 17:25:51 [serving_chat.py:159] Using default chat sampling params from model: {'temperature': 0.6, 'top_p': 0.9} 2025-11-03T17:25:51.8564257Z (APIServer pid=40997) INFO 11-03 17:25:51 [serving_completion.py:67] Using default completion sampling params from model: {'temperature': 0.6, 'top_p': 0.9} 2025-11-03T17:25:51.8566266Z (APIServer pid=40997) INFO 11-03 17:25:51 [api_server.py:1939] Starting vLLM API server 0 on http://0.0.0.0:42779 2025-11-03T17:25:51.8567259Z (APIServer pid=40997) INFO 11-03 17:25:51 [launcher.py:38] Available routes are: 2025-11-03T17:25:51.8567926Z (APIServer pid=40997) INFO 11-03 17:25:51 [launcher.py:46] Route: /openapi.json, Methods: GET, HEAD 2025-11-03T17:25:51.8568707Z (APIServer pid=40997) INFO 11-03 17:25:51 [launcher.py:46] Route: /docs, Methods: GET, HEAD 2025-11-03T17:25:51.8569439Z (APIServer pid=40997) INFO 11-03 17:25:51 [launcher.py:46] Route: /docs/oauth2-redirect, Methods: GET, HEAD 2025-11-03T17:25:51.8570457Z (APIServer pid=40997) INFO 11-03 17:25:51 [launcher.py:46] Route: /redoc, Methods: GET, HEAD 2025-11-03T17:25:51.8571172Z (APIServer pid=40997) INFO 11-03 17:25:51 [launcher.py:46] Route: /health, Methods: GET 2025-11-03T17:25:51.8571786Z (APIServer pid=40997) INFO 11-03 17:25:51 [launcher.py:46] Route: /load, Methods: GET 2025-11-03T17:25:51.8572396Z (APIServer pid=40997) INFO 11-03 17:25:51 [launcher.py:46] Route: /ping, Methods: POST 2025-11-03T17:25:51.8573005Z (APIServer pid=40997) INFO 11-03 17:25:51 [launcher.py:46] Route: /ping, Methods: GET 2025-11-03T17:25:51.8573625Z (APIServer pid=40997) INFO 11-03 17:25:51 [launcher.py:46] Route: /tokenize, Methods: POST 2025-11-03T17:25:51.8574274Z (APIServer pid=40997) INFO 11-03 17:25:51 [launcher.py:46] Route: /detokenize, Methods: POST 2025-11-03T17:25:51.8574913Z (APIServer pid=40997) INFO 11-03 17:25:51 [launcher.py:46] Route: /v1/models, Methods: GET 2025-11-03T17:25:51.8575534Z (APIServer pid=40997) INFO 11-03 17:25:51 [launcher.py:46] Route: /version, Methods: GET 2025-11-03T17:25:51.8576237Z (APIServer pid=40997) INFO 11-03 17:25:51 [launcher.py:46] Route: /v1/responses, Methods: POST 2025-11-03T17:25:51.8576821Z (APIServer pid=40997) INFO 11-03 17:25:51 [launcher.py:46] Route: /v1/responses/{response_id}, Methods: GET 2025-11-03T17:25:51.8577442Z (APIServer pid=40997) INFO 11-03 17:25:51 [launcher.py:46] Route: /v1/responses/{response_id}/cancel, Methods: POST 2025-11-03T17:25:51.8578051Z (APIServer pid=40997) INFO 11-03 17:25:51 [launcher.py:46] Route: /v1/chat/completions, Methods: POST 2025-11-03T17:25:51.8578625Z (APIServer pid=40997) INFO 11-03 17:25:51 [launcher.py:46] Route: /v1/completions, Methods: POST 2025-11-03T17:25:51.8579187Z (APIServer pid=40997) INFO 11-03 17:25:51 [launcher.py:46] Route: /v1/embeddings, Methods: POST 2025-11-03T17:25:51.8579719Z (APIServer pid=40997) INFO 11-03 17:25:51 [launcher.py:46] Route: /pooling, Methods: POST 2025-11-03T17:25:51.8580243Z (APIServer pid=40997) INFO 11-03 17:25:51 [launcher.py:46] Route: /classify, Methods: POST 2025-11-03T17:25:51.8580749Z (APIServer pid=40997) INFO 11-03 17:25:51 [launcher.py:46] Route: /score, Methods: POST 2025-11-03T17:25:51.8581242Z (APIServer pid=40997) INFO 11-03 17:25:51 [launcher.py:46] Route: /v1/score, Methods: POST 2025-11-03T17:25:51.8581793Z (APIServer pid=40997) INFO 11-03 17:25:51 [launcher.py:46] Route: /v1/audio/transcriptions, Methods: POST 2025-11-03T17:25:51.8582387Z (APIServer pid=40997) INFO 11-03 17:25:51 [launcher.py:46] Route: /v1/audio/translations, Methods: POST 2025-11-03T17:25:51.8582920Z (APIServer pid=40997) INFO 11-03 17:25:51 [launcher.py:46] Route: /rerank, Methods: POST 2025-11-03T17:25:51.8583431Z (APIServer pid=40997) INFO 11-03 17:25:51 [launcher.py:46] Route: /v1/rerank, Methods: POST 2025-11-03T17:25:51.8583940Z (APIServer pid=40997) INFO 11-03 17:25:51 [launcher.py:46] Route: /v2/rerank, Methods: POST 2025-11-03T17:25:51.8584469Z (APIServer pid=40997) INFO 11-03 17:25:51 [launcher.py:46] Route: /scale_elastic_ep, Methods: POST 2025-11-03T17:25:51.8585033Z (APIServer pid=40997) INFO 11-03 17:25:51 [launcher.py:46] Route: /is_scaling_elastic_ep, Methods: POST 2025-11-03T17:25:51.8585578Z (APIServer pid=40997) INFO 11-03 17:25:51 [launcher.py:46] Route: /invocations, Methods: POST 2025-11-03T17:25:51.8586100Z (APIServer pid=40997) INFO 11-03 17:25:51 [launcher.py:46] Route: /metrics, Methods: GET 2025-11-03T17:25:51.8720062Z (APIServer pid=40997) INFO: Started server process [40997] 2025-11-03T17:25:51.8720678Z (APIServer pid=40997) INFO: Waiting for application startup. 2025-11-03T17:25:52.0879942Z (APIServer pid=40997) INFO: Application startup complete. 2025-11-03T17:25:52.5815195Z (APIServer pid=40997) INFO: 127.0.0.1:51350 - "GET /health HTTP/1.1" 200 OK 2025-11-03T17:25:52.8283924Z (APIServer pid=40997) INFO: 127.0.0.1:51358 - "GET /v1/models HTTP/1.1" 200 OK 2025-11-03T17:25:52.8587136Z (EngineCore_DP0 pid=41119) INFO 11-03 17:25:52 [ray_distributed_executor.py:569] RAY_CGRAPH_get_timeout is set to 300 2025-11-03T17:25:52.8588443Z (EngineCore_DP0 pid=41119) INFO 11-03 17:25:52 [ray_distributed_executor.py:573] VLLM_USE_RAY_COMPILED_DAG_CHANNEL_TYPE = auto 2025-11-03T17:25:52.8589192Z (EngineCore_DP0 pid=41119) INFO 11-03 17:25:52 [ray_distributed_executor.py:577] VLLM_USE_RAY_COMPILED_DAG_OVERLAP_COMM = False 2025-11-03T17:25:52.8591609Z (EngineCore_DP0 pid=41119) INFO 11-03 17:25:52 [ray_distributed_executor.py:653] Using RayPPCommunicator (which wraps vLLM _PP GroupCoordinator) for Ray Compiled Graph communication. 2025-11-03T17:25:53.0344089Z (APIServer pid=40997) INFO: 127.0.0.1:51358 - "POST /v1/completions HTTP/1.1" 200 OK 2025-11-03T17:25:53.0938751Z (APIServer pid=40997) INFO: 127.0.0.1:51358 - "POST /v1/completions HTTP/1.1" 200 OK 2025-11-03T17:25:53.2291862Z (APIServer pid=40997) INFO: 127.0.0.1:51358 - "POST /v1/completions HTTP/1.1" 200 OK 2025-11-03T17:25:53.3175443Z (APIServer pid=40997) INFO: 127.0.0.1:51358 - "POST /v1/completions HTTP/1.1" 200 OK 2025-11-03T17:25:53.3891535Z (APIServer pid=40997) INFO: 127.0.0.1:51358 - "POST /v1/completions HTTP/1.1" 200 OK 2025-11-03T17:25:53.3973142Z (APIServer pid=40997) INFO: 127.0.0.1:51358 - "POST /v1/completions HTTP/1.1" 200 OK 2025-11-03T17:25:53.4942636Z (APIServer pid=40997) INFO 11-03 17:25:53 [launcher.py:110] Shutting down FastAPI HTTP server. 2025-11-03T17:25:53.4943127Z *** SIGTERM received at time=1762190753 on cpu 19 *** 2025-11-03T17:25:53.4993866Z PC: @ 0x7f145d57a117 (unknown) (unknown) 2025-11-03T17:25:53.4995808Z @ 0x7f145d52b520 (unknown) (unknown) 2025-11-03T17:25:53.4996289Z [2025-11-03 17:25:53,499 E 41119 41119] logging.cc:501: *** SIGTERM received at time=1762190753 on cpu 19 *** 2025-11-03T17:25:53.4997509Z [2025-11-03 17:25:53,499 E 41119 41119] logging.cc:501: PC: @ 0x7f145d57a117 (unknown) (unknown) 2025-11-03T17:25:53.4999211Z [2025-11-03 17:25:53,499 E 41119 41119] logging.cc:501: @ 0x7f145d52b520 (unknown) (unknown) 2025-11-03T17:25:53.5001754Z (EngineCore_DP0 pid=41119) INFO 11-03 17:25:53 [ray_distributed_executor.py:127] Shutting down Ray distributed executor. If you see error log from logging.cc regarding SIGTERM received, please ignore because this is the expected termination process in Ray. 2025-11-03T17:25:53.5003101Z (EngineCore_DP0 pid=41119) 2025-11-03 17:25:53,499 INFO compiled_dag_node.py:2171 -- Tearing down compiled DAG 2025-11-03T17:25:53.5003992Z (EngineCore_DP0 pid=41119) 2025-11-03 17:25:53,500 INFO compiled_dag_node.py:2176 -- Cancelling compiled worker on actor: Actor(RayWorkerWrapper, 1d26af89f17c4f4272dad28201000000) 2025-11-03T17:25:53.5004990Z (EngineCore_DP0 pid=41119) 2025-11-03 17:25:53,500 INFO compiled_dag_node.py:2176 -- Cancelling compiled worker on actor: Actor(RayWorkerWrapper, 717374ec8600319fc138644501000000) 2025-11-03T17:25:53.5045609Z (EngineCore_DP0 pid=41119) 2025-11-03 17:25:53,504 INFO compiled_dag_node.py:2198 -- Waiting for worker tasks to exit 2025-11-03T17:25:53.5047158Z (EngineCore_DP0 pid=41119) 2025-11-03 17:25:53,504 INFO compiled_dag_node.py:2201 -- Teardown complete 2025-11-03T17:25:54.0073816Z (EngineCore_DP0 pid=41119) (RayWorkerWrapper pid=42012) INFO 11-03 17:25:47 [backends.py:594] Using cache directory: /var/lib/jenkins/.cache/vllm/torch_compile_cache/618b881a70/rank_1_0/backbone for vLLM's torch.compile 2025-11-03T17:25:54.0076444Z (EngineCore_DP0 pid=41119) (RayWorkerWrapper pid=42012) INFO 11-03 17:25:47 [backends.py:608] Dynamo bytecode transform time: 2.99 s 2025-11-03T17:25:54.0078401Z (EngineCore_DP0 pid=41119) (RayWorkerWrapper pid=42012) INFO 11-03 17:25:48 [backends.py:179] Directly load the compiled graph(s) for dynamic shape from the cache, took 0.754 s 2025-11-03T17:25:54.0079703Z (EngineCore_DP0 pid=41119) (RayWorkerWrapper pid=42012) INFO 11-03 17:25:49 [gpu_worker.py:314] Available KV cache memory: 18.54 GiB 2025-11-03T17:25:54.0080582Z (EngineCore_DP0 pid=41119) (RayWorkerWrapper pid=42012) INFO 11-03 17:25:50 [gpu_worker.py:351] Compile and warming up model for size 4 [repeated 3x across cluster] 2025-11-03T17:25:54.0085277Z (EngineCore_DP0 pid=41119) (RayWorkerWrapper pid=42012) INFO 11-03 17:25:50 [backends.py:185] Directly load the compiled graph(s) for shape 4 from the cache, took 0.690 s [repeated 3x across cluster] 2025-11-03T17:25:54.0086812Z (EngineCore_DP0 pid=41119) (RayWorkerWrapper pid=42012) INFO 11-03 17:25:50 [monitor.py:33] torch.compile takes 2.99 s in total 2025-11-03T17:25:54.0093559Z (EngineCore_DP0 pid=41119) (pid=42012) /opt/conda/envs/py_3.12/lib/python3.12/site-packages/torch/cuda/__init__.py:65: FutureWarning: The pynvml package is deprecated. Please install nvidia-ml-py instead. If you did not install pynvml directly, please report this to the maintainers of the package that installed pynvml for you. 2025-11-03T17:25:54.0095061Z (EngineCore_DP0 pid=41119) (pid=42012) import pynvml # type: ignore[import] 2025-11-03T17:25:55.9049269Z (APIServer pid=40997) INFO: Shutting down 2025-11-03T17:25:56.0052063Z (APIServer pid=40997) INFO: Waiting for application shutdown. 2025-11-03T17:25:56.0053310Z (APIServer pid=40997) INFO: Application shutdown complete. 2025-11-03T17:25:56.3639367Z INFO 11-03 17:25:56 [model.py:653] Resolved architecture: LlamaForCausalLM 2025-11-03T17:25:56.3640098Z WARNING 11-03 17:25:56 [model.py:1942] Casting torch.bfloat16 to torch.float16. 2025-11-03T17:25:56.3640628Z INFO 11-03 17:25:56 [model.py:1714] Using max model len 2048 2025-11-03T17:25:56.3641858Z Launching RemoteOpenAIServer with: vllm serve meta-llama/Llama-3.2-1B-Instruct --dtype float16 --max-model-len 2048 --max-num-seqs 8 --enable-chunked-prefill --enforce-eager --tokenizer-mode auto --tensor-parallel-size 2 --distributed-executor-backend mp --load-format dummy --port 47785 --seed 0 2025-11-03T17:25:56.8884716Z /opt/conda/envs/py_3.12/lib/python3.12/site-packages/torch/cuda/__init__.py:65: FutureWarning: The pynvml package is deprecated. Please install nvidia-ml-py instead. If you did not install pynvml directly, please report this to the maintainers of the package that installed pynvml for you. 2025-11-03T17:25:56.8885986Z import pynvml # type: ignore[import] 2025-11-03T17:25:59.1779876Z INFO 11-03 17:25:59 [__init__.py:224] Automatically detected platform cuda. 2025-11-03T17:26:02.9322712Z (APIServer pid=45749) INFO 11-03 17:26:02 [api_server.py:1870] vLLM API server version 0.11.0rc2.dev389+ge51928192.d20251103 2025-11-03T17:26:02.9347509Z (APIServer pid=45749) INFO 11-03 17:26:02 [utils.py:239] non-default args: {'model_tag': 'meta-llama/Llama-3.2-1B-Instruct', 'port': 47785, 'model': 'meta-llama/Llama-3.2-1B-Instruct', 'dtype': 'float16', 'seed': 0, 'max_model_len': 2048, 'enforce_eager': True, 'load_format': 'dummy', 'distributed_executor_backend': 'mp', 'tensor_parallel_size': 2, 'max_num_seqs': 8, 'enable_chunked_prefill': True} 2025-11-03T17:26:03.0911864Z (APIServer pid=45749) INFO 11-03 17:26:03 [model.py:653] Resolved architecture: LlamaForCausalLM 2025-11-03T17:26:03.0912853Z (APIServer pid=45749) `torch_dtype` is deprecated! Use `dtype` instead! 2025-11-03T17:26:03.0913559Z (APIServer pid=45749) WARNING 11-03 17:26:03 [model.py:1942] Casting torch.bfloat16 to torch.float16. 2025-11-03T17:26:03.0914224Z (APIServer pid=45749) INFO 11-03 17:26:03 [model.py:1714] Using max model len 2048 2025-11-03T17:26:03.2563495Z (APIServer pid=45749) INFO 11-03 17:26:03 [scheduler.py:225] Chunked prefill is enabled with max_num_batched_tokens=2048. 2025-11-03T17:26:03.2569166Z (APIServer pid=45749) INFO 11-03 17:26:03 [vllm.py:354] Cudagraph is disabled under eager mode 2025-11-03T17:26:04.2328734Z /opt/conda/envs/py_3.12/lib/python3.12/site-packages/torch/cuda/__init__.py:65: FutureWarning: The pynvml package is deprecated. Please install nvidia-ml-py instead. If you did not install pynvml directly, please report this to the maintainers of the package that installed pynvml for you. 2025-11-03T17:26:04.2329987Z import pynvml # type: ignore[import] 2025-11-03T17:26:06.5876672Z INFO 11-03 17:26:06 [__init__.py:224] Automatically detected platform cuda. 2025-11-03T17:26:10.3053824Z (EngineCore_DP0 pid=45871) INFO 11-03 17:26:10 [core.py:727] Waiting for init message from front-end. 2025-11-03T17:26:10.3316523Z (EngineCore_DP0 pid=45871) INFO 11-03 17:26:10 [core.py:94] Initializing a V1 LLM engine (v0.11.0rc2.dev389+ge51928192.d20251103) with config: model='meta-llama/Llama-3.2-1B-Instruct', speculative_config=None, tokenizer='meta-llama/Llama-3.2-1B-Instruct', skip_tokenizer_init=False, tokenizer_mode=auto, revision=None, tokenizer_revision=None, trust_remote_code=False, dtype=torch.float16, max_seq_len=2048, download_dir=None, load_format=dummy, tensor_parallel_size=2, pipeline_parallel_size=1, data_parallel_size=1, disable_custom_all_reduce=False, quantization=None, enforce_eager=True, kv_cache_dtype=auto, device_config=cuda, structured_outputs_config=StructuredOutputsConfig(backend='auto', disable_fallback=False, disable_any_whitespace=False, disable_additional_properties=False, reasoning_parser=''), observability_config=ObservabilityConfig(show_hidden_metrics_for_version=None, otlp_traces_endpoint=None, collect_detailed_traces=None), seed=0, served_model_name=meta-llama/Llama-3.2-1B-Instruct, enable_prefix_caching=True, chunked_prefill_enabled=True, pooler_config=None, compilation_config={'level': 0, 'debug_dump_path': None, 'cache_dir': '', 'backend': '', 'custom_ops': [], 'splitting_ops': None, 'use_inductor': True, 'compile_sizes': [], 'inductor_compile_config': {'enable_auto_functionalized_v2': False}, 'inductor_passes': {}, 'cudagraph_mode': , 'use_cudagraph': False, 'cudagraph_num_of_warmups': 0, 'cudagraph_capture_sizes': [], 'cudagraph_copy_inputs': False, 'full_cuda_graph': False, 'use_inductor_graph_partition': False, 'pass_config': {}, 'max_capture_size': 0, 'local_cache_dir': None} 2025-11-03T17:26:10.3322848Z (EngineCore_DP0 pid=45871) WARNING 11-03 17:26:10 [multiproc_executor.py:744] Reducing Torch parallelism from 24 threads to 1 to avoid unnecessary CPU contention. Set OMP_NUM_THREADS in the external environment to tune this value as needed. 2025-11-03T17:26:10.3331214Z (EngineCore_DP0 pid=45871) INFO 11-03 17:26:10 [shm_broadcast.py:302] vLLM message queue communication handle: Handle(local_reader_ranks=[0, 1], buffer_handle=(2, 16777216, 10, 'psm_dfb2f1e1'), local_subscribe_addr='ipc:///tmp/503fac68-2883-47c4-bbee-64c7d88c238e', remote_subscribe_addr=None, remote_addr_ipv6=False) 2025-11-03T17:26:10.8176365Z /opt/conda/envs/py_3.12/lib/python3.12/site-packages/torch/cuda/__init__.py:65: FutureWarning: The pynvml package is deprecated. Please install nvidia-ml-py instead. If you did not install pynvml directly, please report this to the maintainers of the package that installed pynvml for you. 2025-11-03T17:26:10.8178591Z import pynvml # type: ignore[import] 2025-11-03T17:26:10.8181128Z /opt/conda/envs/py_3.12/lib/python3.12/site-packages/torch/cuda/__init__.py:65: FutureWarning: The pynvml package is deprecated. Please install nvidia-ml-py instead. If you did not install pynvml directly, please report this to the maintainers of the package that installed pynvml for you. 2025-11-03T17:26:10.8183553Z import pynvml # type: ignore[import] 2025-11-03T17:26:13.1734812Z INFO 11-03 17:26:13 [__init__.py:224] Automatically detected platform cuda. 2025-11-03T17:26:13.2036013Z INFO 11-03 17:26:13 [__init__.py:224] Automatically detected platform cuda. 2025-11-03T17:26:17.9530470Z INFO 11-03 17:26:17 [shm_broadcast.py:302] vLLM message queue communication handle: Handle(local_reader_ranks=[0], buffer_handle=(1, 10485760, 10, 'psm_e8a9d5d6'), local_subscribe_addr='ipc:///tmp/91088734-e7ea-4835-8dc1-d9d4708ab823', remote_subscribe_addr=None, remote_addr_ipv6=False) 2025-11-03T17:26:17.9927379Z INFO 11-03 17:26:17 [shm_broadcast.py:302] vLLM message queue communication handle: Handle(local_reader_ranks=[0], buffer_handle=(1, 10485760, 10, 'psm_5bde6961'), local_subscribe_addr='ipc:///tmp/7dbcc191-e792-4f0e-9312-f0fb945e3566', remote_subscribe_addr=None, remote_addr_ipv6=False) 2025-11-03T17:26:18.1723298Z [Gloo] Rank 0 is connected to 1 peer ranks. Expected number of connected peer ranks is : 1 2025-11-03T17:26:18.1724475Z [Gloo] Rank 1 is connected to 1 peer ranks. Expected number of connected peer ranks is : 1 2025-11-03T17:26:18.1772090Z [Gloo] Rank 1 is connected to 1 peer ranks. Expected number of connected peer ranks is : 1 2025-11-03T17:26:18.1773276Z [Gloo] Rank 0 is connected to 1 peer ranks. Expected number of connected peer ranks is : 1 2025-11-03T17:26:18.1812457Z INFO 11-03 17:26:18 [__init__.py:1146] Found nccl from library libnccl.so.2 2025-11-03T17:26:18.1817619Z INFO 11-03 17:26:18 [__init__.py:1146] Found nccl from library libnccl.so.2 2025-11-03T17:26:18.1818549Z INFO 11-03 17:26:18 [pynccl.py:109] vLLM is using nccl==2.27.5 2025-11-03T17:26:18.1819344Z INFO 11-03 17:26:18 [pynccl.py:109] vLLM is using nccl==2.27.5 2025-11-03T17:26:18.3548995Z WARNING 11-03 17:26:18 [symm_mem.py:60] SymmMemCommunicator: Device capability 8.9 not supported, communicator is not available. 2025-11-03T17:26:18.3550563Z WARNING 11-03 17:26:18 [symm_mem.py:60] SymmMemCommunicator: Device capability 8.9 not supported, communicator is not available. 2025-11-03T17:26:18.3575470Z INFO 11-03 17:26:18 [custom_all_reduce.py:37] Skipping P2P check and trusting the driver's P2P report. 2025-11-03T17:26:18.3576101Z INFO 11-03 17:26:18 [custom_all_reduce.py:37] Skipping P2P check and trusting the driver's P2P report. 2025-11-03T17:26:18.3577104Z WARNING 11-03 17:26:18 [custom_all_reduce.py:162] Custom allreduce is disabled because your platform lacks GPU P2P capability or P2P test failed. To silence this warning, specify disable_custom_all_reduce=True explicitly. 2025-11-03T17:26:18.3578474Z WARNING 11-03 17:26:18 [custom_all_reduce.py:162] Custom allreduce is disabled because your platform lacks GPU P2P capability or P2P test failed. To silence this warning, specify disable_custom_all_reduce=True explicitly. 2025-11-03T17:26:18.4224695Z INFO 11-03 17:26:18 [shm_broadcast.py:302] vLLM message queue communication handle: Handle(local_reader_ranks=[1], buffer_handle=(1, 4194304, 6, 'psm_c4767f88'), local_subscribe_addr='ipc:///tmp/bf96cb98-0fcb-48c7-9856-064c9c55ba48', remote_subscribe_addr=None, remote_addr_ipv6=False) 2025-11-03T17:26:18.4248323Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:26:18.4248883Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:26:18.4255908Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:26:18.4274000Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:26:18.4622370Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:26:18.4633281Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:26:18.4647885Z [Gloo] Rank 1 is connected to 1 peer ranks. Expected number of connected peer ranks is : 1 2025-11-03T17:26:18.4649797Z [Gloo] Rank 0 is connected to 1 peer ranks. Expected number of connected peer ranks is : 1 2025-11-03T17:26:18.4650839Z INFO 11-03 17:26:18 [__init__.py:1146] Found nccl from library libnccl.so.2 2025-11-03T17:26:18.4651281Z INFO 11-03 17:26:18 [pynccl.py:109] vLLM is using nccl==2.27.5 2025-11-03T17:26:18.4651625Z INFO 11-03 17:26:18 [__init__.py:1146] Found nccl from library libnccl.so.2 2025-11-03T17:26:18.4651951Z INFO 11-03 17:26:18 [pynccl.py:109] vLLM is using nccl==2.27.5 2025-11-03T17:26:18.4983654Z INFO 11-03 17:26:18 [parallel_state.py:1231] rank 1 in world size 2 is assigned as DP rank 0, PP rank 0, TP rank 1, EP rank 1 2025-11-03T17:26:18.4984274Z INFO 11-03 17:26:18 [parallel_state.py:1231] rank 0 in world size 2 is assigned as DP rank 0, PP rank 0, TP rank 0, EP rank 0 2025-11-03T17:26:18.7289458Z INFO 11-03 17:26:18 [topk_topp_sampler.py:59] Using FlashInfer for top-p & top-k sampling. 2025-11-03T17:26:18.7450382Z INFO 11-03 17:26:18 [topk_topp_sampler.py:59] Using FlashInfer for top-p & top-k sampling. 2025-11-03T17:26:18.7610738Z (Worker_TP1 pid=45991) INFO 11-03 17:26:18 [gpu_model_runner.py:2840] Starting to load model meta-llama/Llama-3.2-1B-Instruct... 2025-11-03T17:26:18.7762578Z (Worker_TP0 pid=45990) INFO 11-03 17:26:18 [gpu_model_runner.py:2840] Starting to load model meta-llama/Llama-3.2-1B-Instruct... 2025-11-03T17:26:18.9676283Z (Worker_TP1 pid=45991) INFO 11-03 17:26:18 [gpu_model_runner.py:2870] Loading model from scratch... 2025-11-03T17:26:18.9888980Z (Worker_TP0 pid=45990) INFO 11-03 17:26:18 [gpu_model_runner.py:2870] Loading model from scratch... 2025-11-03T17:26:19.0565695Z (Worker_TP1 pid=45991) INFO 11-03 17:26:19 [cuda.py:404] Using Flash Attention backend on V1 engine. 2025-11-03T17:26:19.0786089Z (Worker_TP0 pid=45990) INFO 11-03 17:26:19 [cuda.py:404] Using Flash Attention backend on V1 engine. 2025-11-03T17:26:19.5054040Z (Worker_TP1 pid=45991) INFO 11-03 17:26:19 [gpu_model_runner.py:2902] Model loading took 1.1667 GiB and 0.117768 seconds 2025-11-03T17:26:19.5131682Z (Worker_TP0 pid=45990) INFO 11-03 17:26:19 [gpu_model_runner.py:2902] Model loading took 1.1667 GiB and 0.118075 seconds 2025-11-03T17:26:21.4366313Z (Worker_TP1 pid=45991) INFO 11-03 17:26:21 [gpu_worker.py:314] Available KV cache memory: 18.53 GiB 2025-11-03T17:26:21.4371718Z (Worker_TP0 pid=45990) INFO 11-03 17:26:21 [gpu_worker.py:314] Available KV cache memory: 18.53 GiB 2025-11-03T17:26:21.6468265Z (EngineCore_DP0 pid=45871) INFO 11-03 17:26:21 [kv_cache_utils.py:1199] GPU KV cache size: 1,214,464 tokens 2025-11-03T17:26:21.6469153Z (EngineCore_DP0 pid=45871) INFO 11-03 17:26:21 [kv_cache_utils.py:1204] Maximum concurrency for 2,048 tokens per request: 593.00x 2025-11-03T17:26:21.6470019Z (EngineCore_DP0 pid=45871) INFO 11-03 17:26:21 [kv_cache_utils.py:1199] GPU KV cache size: 1,214,464 tokens 2025-11-03T17:26:21.6470847Z (EngineCore_DP0 pid=45871) INFO 11-03 17:26:21 [kv_cache_utils.py:1204] Maximum concurrency for 2,048 tokens per request: 593.00x 2025-11-03T17:26:21.7686558Z (EngineCore_DP0 pid=45871) INFO 11-03 17:26:21 [core.py:240] init engine (profile, create kv cache, warmup model) took 2.18 seconds 2025-11-03T17:26:22.2630154Z (EngineCore_DP0 pid=45871) INFO 11-03 17:26:22 [vllm.py:354] Cudagraph is disabled under eager mode 2025-11-03T17:26:22.4476732Z (APIServer pid=45749) INFO 11-03 17:26:22 [loggers.py:172] Engine 000: vllm cache_config_info with initialization after num_gpu_blocks is: 75904 2025-11-03T17:26:22.4602997Z (EngineCore_DP0 pid=45871) INFO 11-03 17:26:22 [gc_utils.py:40] GC Debug Config. enabled:False,top_objects:-1 2025-11-03T17:26:22.4821737Z (APIServer pid=45749) INFO 11-03 17:26:22 [api_server.py:1628] Supported tasks: ['generate'] 2025-11-03T17:26:22.5102371Z (APIServer pid=45749) WARNING 11-03 17:26:22 [model.py:1581] Default sampling parameters have been overridden by the model's Hugging Face generation config recommended from the model creator. If this is not intended, please relaunch vLLM instance with `--generation-config vllm`. 2025-11-03T17:26:22.5105060Z (APIServer pid=45749) INFO 11-03 17:26:22 [serving_responses.py:157] Using default chat sampling params from model: {'temperature': 0.6, 'top_p': 0.9} 2025-11-03T17:26:22.5343541Z (APIServer pid=45749) INFO 11-03 17:26:22 [serving_chat.py:159] Using default chat sampling params from model: {'temperature': 0.6, 'top_p': 0.9} 2025-11-03T17:26:22.6194517Z (APIServer pid=45749) INFO 11-03 17:26:22 [serving_completion.py:67] Using default completion sampling params from model: {'temperature': 0.6, 'top_p': 0.9} 2025-11-03T17:26:22.6197715Z (APIServer pid=45749) INFO 11-03 17:26:22 [api_server.py:1939] Starting vLLM API server 0 on http://0.0.0.0:47785 2025-11-03T17:26:22.6199761Z (APIServer pid=45749) INFO 11-03 17:26:22 [launcher.py:38] Available routes are: 2025-11-03T17:26:22.6201713Z (APIServer pid=45749) INFO 11-03 17:26:22 [launcher.py:46] Route: /openapi.json, Methods: GET, HEAD 2025-11-03T17:26:22.6203439Z (APIServer pid=45749) INFO 11-03 17:26:22 [launcher.py:46] Route: /docs, Methods: GET, HEAD 2025-11-03T17:26:22.6204134Z (APIServer pid=45749) INFO 11-03 17:26:22 [launcher.py:46] Route: /docs/oauth2-redirect, Methods: GET, HEAD 2025-11-03T17:26:22.6204700Z (APIServer pid=45749) INFO 11-03 17:26:22 [launcher.py:46] Route: /redoc, Methods: GET, HEAD 2025-11-03T17:26:22.6205223Z (APIServer pid=45749) INFO 11-03 17:26:22 [launcher.py:46] Route: /health, Methods: GET 2025-11-03T17:26:22.6205726Z (APIServer pid=45749) INFO 11-03 17:26:22 [launcher.py:46] Route: /load, Methods: GET 2025-11-03T17:26:22.6206224Z (APIServer pid=45749) INFO 11-03 17:26:22 [launcher.py:46] Route: /ping, Methods: POST 2025-11-03T17:26:22.6206711Z (APIServer pid=45749) INFO 11-03 17:26:22 [launcher.py:46] Route: /ping, Methods: GET 2025-11-03T17:26:22.6207206Z (APIServer pid=45749) INFO 11-03 17:26:22 [launcher.py:46] Route: /tokenize, Methods: POST 2025-11-03T17:26:22.6207731Z (APIServer pid=45749) INFO 11-03 17:26:22 [launcher.py:46] Route: /detokenize, Methods: POST 2025-11-03T17:26:22.6208253Z (APIServer pid=45749) INFO 11-03 17:26:22 [launcher.py:46] Route: /v1/models, Methods: GET 2025-11-03T17:26:22.6208864Z (APIServer pid=45749) INFO 11-03 17:26:22 [launcher.py:46] Route: /version, Methods: GET 2025-11-03T17:26:22.6209412Z (APIServer pid=45749) INFO 11-03 17:26:22 [launcher.py:46] Route: /v1/responses, Methods: POST 2025-11-03T17:26:22.6210295Z (APIServer pid=45749) INFO 11-03 17:26:22 [launcher.py:46] Route: /v1/responses/{response_id}, Methods: GET 2025-11-03T17:26:22.6210922Z (APIServer pid=45749) INFO 11-03 17:26:22 [launcher.py:46] Route: /v1/responses/{response_id}/cancel, Methods: POST 2025-11-03T17:26:22.6211522Z (APIServer pid=45749) INFO 11-03 17:26:22 [launcher.py:46] Route: /v1/chat/completions, Methods: POST 2025-11-03T17:26:22.6212084Z (APIServer pid=45749) INFO 11-03 17:26:22 [launcher.py:46] Route: /v1/completions, Methods: POST 2025-11-03T17:26:22.6212629Z (APIServer pid=45749) INFO 11-03 17:26:22 [launcher.py:46] Route: /v1/embeddings, Methods: POST 2025-11-03T17:26:22.6213155Z (APIServer pid=45749) INFO 11-03 17:26:22 [launcher.py:46] Route: /pooling, Methods: POST 2025-11-03T17:26:22.6213676Z (APIServer pid=45749) INFO 11-03 17:26:22 [launcher.py:46] Route: /classify, Methods: POST 2025-11-03T17:26:22.6214324Z (APIServer pid=45749) INFO 11-03 17:26:22 [launcher.py:46] Route: /score, Methods: POST 2025-11-03T17:26:22.6214833Z (APIServer pid=45749) INFO 11-03 17:26:22 [launcher.py:46] Route: /v1/score, Methods: POST 2025-11-03T17:26:22.6215395Z (APIServer pid=45749) INFO 11-03 17:26:22 [launcher.py:46] Route: /v1/audio/transcriptions, Methods: POST 2025-11-03T17:26:22.6215974Z (APIServer pid=45749) INFO 11-03 17:26:22 [launcher.py:46] Route: /v1/audio/translations, Methods: POST 2025-11-03T17:26:22.6216503Z (APIServer pid=45749) INFO 11-03 17:26:22 [launcher.py:46] Route: /rerank, Methods: POST 2025-11-03T17:26:22.6217011Z (APIServer pid=45749) INFO 11-03 17:26:22 [launcher.py:46] Route: /v1/rerank, Methods: POST 2025-11-03T17:26:22.6217518Z (APIServer pid=45749) INFO 11-03 17:26:22 [launcher.py:46] Route: /v2/rerank, Methods: POST 2025-11-03T17:26:22.6218050Z (APIServer pid=45749) INFO 11-03 17:26:22 [launcher.py:46] Route: /scale_elastic_ep, Methods: POST 2025-11-03T17:26:22.6218731Z (APIServer pid=45749) INFO 11-03 17:26:22 [launcher.py:46] Route: /is_scaling_elastic_ep, Methods: POST 2025-11-03T17:26:22.6219277Z (APIServer pid=45749) INFO 11-03 17:26:22 [launcher.py:46] Route: /invocations, Methods: POST 2025-11-03T17:26:22.6219789Z (APIServer pid=45749) INFO 11-03 17:26:22 [launcher.py:46] Route: /metrics, Methods: GET 2025-11-03T17:26:22.6355229Z (APIServer pid=45749) INFO: Started server process [45749] 2025-11-03T17:26:22.6356445Z (APIServer pid=45749) INFO: Waiting for application startup. 2025-11-03T17:26:22.8392884Z (APIServer pid=45749) INFO: Application startup complete. 2025-11-03T17:26:22.9335807Z (APIServer pid=45749) INFO: 127.0.0.1:47450 - "GET /health HTTP/1.1" 200 OK 2025-11-03T17:26:22.9427237Z (APIServer pid=45749) INFO: 127.0.0.1:47454 - "GET /v1/models HTTP/1.1" 200 OK 2025-11-03T17:26:23.1155299Z (APIServer pid=45749) INFO: 127.0.0.1:47454 - "POST /v1/completions HTTP/1.1" 200 OK 2025-11-03T17:26:23.1708471Z (APIServer pid=45749) INFO: 127.0.0.1:47454 - "POST /v1/completions HTTP/1.1" 200 OK 2025-11-03T17:26:23.2576007Z (APIServer pid=45749) INFO: 127.0.0.1:47454 - "POST /v1/completions HTTP/1.1" 200 OK 2025-11-03T17:26:23.3403359Z (APIServer pid=45749) INFO: 127.0.0.1:47454 - "POST /v1/completions HTTP/1.1" 200 OK 2025-11-03T17:26:23.4001774Z (APIServer pid=45749) INFO: 127.0.0.1:47454 - "POST /v1/completions HTTP/1.1" 200 OK 2025-11-03T17:26:23.4074632Z (APIServer pid=45749) INFO: 127.0.0.1:47454 - "POST /v1/completions HTTP/1.1" 200 OK 2025-11-03T17:26:23.4725842Z (APIServer pid=45749) INFO 11-03 17:26:23 [launcher.py:110] Shutting down FastAPI HTTP server. 2025-11-03T17:26:23.4728720Z (Worker_TP0 pid=45990) INFO 11-03 17:26:23 [multiproc_executor.py:583] Parent process exited, terminating worker 2025-11-03T17:26:23.4731003Z (Worker_TP0 pid=45990) INFO 11-03 17:26:23 [multiproc_executor.py:624] WorkerProc shutting down. 2025-11-03T17:26:23.4732712Z (Worker_TP1 pid=45991) INFO 11-03 17:26:23 [multiproc_executor.py:583] Parent process exited, terminating worker 2025-11-03T17:26:23.4733983Z (Worker_TP1 pid=45991) INFO 11-03 17:26:23 [multiproc_executor.py:624] WorkerProc shutting down. 2025-11-03T17:26:25.6898427Z (APIServer pid=45749) INFO: Shutting down 2025-11-03T17:26:25.7904545Z (APIServer pid=45749) INFO: Waiting for application shutdown. 2025-11-03T17:26:25.7905937Z (APIServer pid=45749) INFO: Application shutdown complete. 2025-11-03T17:26:26.4037129Z PASSED 2025-11-03T17:26:26.4182027Z distributed/test_sequence_parallel.py::test_tp_sp_generation[meta-llama/Llama-3.2-1B-Instruct-parallel_setup12-mp-auto-test_options12] Fork a new process to run a test 46194 2025-11-03T17:26:26.4199603Z Fork a new process to run a test 0 2025-11-03T17:26:27.0616831Z INFO 11-03 17:26:27 [model.py:653] Resolved architecture: LlamaForCausalLM 2025-11-03T17:26:27.0617908Z `torch_dtype` is deprecated! Use `dtype` instead! 2025-11-03T17:26:27.0618853Z WARNING 11-03 17:26:27 [model.py:1942] Casting torch.bfloat16 to torch.float16. 2025-11-03T17:26:27.0619682Z INFO 11-03 17:26:27 [model.py:1714] Using max model len 2048 2025-11-03T17:26:27.0624094Z Launching RemoteOpenAIServer with: vllm serve meta-llama/Llama-3.2-1B-Instruct --dtype float16 --max-model-len 2048 --max-num-seqs 8 --enforce-eager --tokenizer-mode auto --tensor-parallel-size 2 --pipeline-parallel-size 2 --distributed-executor-backend mp --compilation_config {"level": 3, "custom_ops": ["+rms_norm"], "compile_sizes": [4, 8], "pass_config": {"enable_sequence_parallelism": true, "enable_fusion": false, "enable_noop": true}} --load-format dummy --port 52725 --seed 0 2025-11-03T17:26:27.5734867Z /opt/conda/envs/py_3.12/lib/python3.12/site-packages/torch/cuda/__init__.py:65: FutureWarning: The pynvml package is deprecated. Please install nvidia-ml-py instead. If you did not install pynvml directly, please report this to the maintainers of the package that installed pynvml for you. 2025-11-03T17:26:27.5736429Z import pynvml # type: ignore[import] 2025-11-03T17:26:29.8404035Z INFO 11-03 17:26:29 [__init__.py:224] Automatically detected platform cuda. 2025-11-03T17:26:33.5260311Z (APIServer pid=46243) INFO 11-03 17:26:33 [api_server.py:1870] vLLM API server version 0.11.0rc2.dev389+ge51928192.d20251103 2025-11-03T17:26:33.5289551Z (APIServer pid=46243) INFO 11-03 17:26:33 [utils.py:239] non-default args: {'model_tag': 'meta-llama/Llama-3.2-1B-Instruct', 'port': 52725, 'model': 'meta-llama/Llama-3.2-1B-Instruct', 'dtype': 'float16', 'seed': 0, 'max_model_len': 2048, 'enforce_eager': True, 'load_format': 'dummy', 'distributed_executor_backend': 'mp', 'pipeline_parallel_size': 2, 'tensor_parallel_size': 2, 'max_num_seqs': 8, 'compilation_config': {'level': 3, 'debug_dump_path': None, 'cache_dir': '', 'backend': '', 'custom_ops': ['+rms_norm'], 'splitting_ops': None, 'use_inductor': True, 'compile_sizes': [4, 8], 'inductor_compile_config': {'enable_auto_functionalized_v2': False}, 'inductor_passes': {}, 'cudagraph_mode': None, 'use_cudagraph': True, 'cudagraph_num_of_warmups': 0, 'cudagraph_capture_sizes': None, 'cudagraph_copy_inputs': False, 'full_cuda_graph': False, 'use_inductor_graph_partition': False, 'pass_config': {'enable_noop': True, 'enable_sequence_parallelism': True}, 'max_capture_size': None, 'local_cache_dir': None}} 2025-11-03T17:26:33.6968906Z (APIServer pid=46243) INFO 11-03 17:26:33 [model.py:653] Resolved architecture: LlamaForCausalLM 2025-11-03T17:26:33.6969976Z (APIServer pid=46243) `torch_dtype` is deprecated! Use `dtype` instead! 2025-11-03T17:26:33.6971060Z (APIServer pid=46243) WARNING 11-03 17:26:33 [model.py:1942] Casting torch.bfloat16 to torch.float16. 2025-11-03T17:26:33.6972245Z (APIServer pid=46243) INFO 11-03 17:26:33 [model.py:1714] Using max model len 2048 2025-11-03T17:26:33.8848314Z (APIServer pid=46243) INFO 11-03 17:26:33 [scheduler.py:225] Chunked prefill is enabled with max_num_batched_tokens=2048. 2025-11-03T17:26:33.8855152Z (APIServer pid=46243) INFO 11-03 17:26:33 [vllm.py:354] Cudagraph is disabled under eager mode 2025-11-03T17:26:34.8664586Z /opt/conda/envs/py_3.12/lib/python3.12/site-packages/torch/cuda/__init__.py:65: FutureWarning: The pynvml package is deprecated. Please install nvidia-ml-py instead. If you did not install pynvml directly, please report this to the maintainers of the package that installed pynvml for you. 2025-11-03T17:26:34.8666165Z import pynvml # type: ignore[import] 2025-11-03T17:26:37.2135225Z INFO 11-03 17:26:37 [__init__.py:224] Automatically detected platform cuda. 2025-11-03T17:26:40.9197943Z (EngineCore_DP0 pid=46365) INFO 11-03 17:26:40 [core.py:727] Waiting for init message from front-end. 2025-11-03T17:26:40.9473374Z (EngineCore_DP0 pid=46365) INFO 11-03 17:26:40 [core.py:94] Initializing a V1 LLM engine (v0.11.0rc2.dev389+ge51928192.d20251103) with config: model='meta-llama/Llama-3.2-1B-Instruct', speculative_config=None, tokenizer='meta-llama/Llama-3.2-1B-Instruct', skip_tokenizer_init=False, tokenizer_mode=auto, revision=None, tokenizer_revision=None, trust_remote_code=False, dtype=torch.float16, max_seq_len=2048, download_dir=None, load_format=dummy, tensor_parallel_size=2, pipeline_parallel_size=2, data_parallel_size=1, disable_custom_all_reduce=False, quantization=None, enforce_eager=True, kv_cache_dtype=auto, device_config=cuda, structured_outputs_config=StructuredOutputsConfig(backend='auto', disable_fallback=False, disable_any_whitespace=False, disable_additional_properties=False, reasoning_parser=''), observability_config=ObservabilityConfig(show_hidden_metrics_for_version=None, otlp_traces_endpoint=None, collect_detailed_traces=None), seed=0, served_model_name=meta-llama/Llama-3.2-1B-Instruct, enable_prefix_caching=True, chunked_prefill_enabled=True, pooler_config=None, compilation_config={'level': 3, 'debug_dump_path': None, 'cache_dir': '', 'backend': '', 'custom_ops': ['+rms_norm', '+rms_norm'], 'splitting_ops': ['vllm.unified_attention', 'vllm.unified_attention_with_output', 'vllm.unified_mla_attention', 'vllm.unified_mla_attention_with_output', 'vllm.mamba_mixer2', 'vllm.mamba_mixer', 'vllm.short_conv', 'vllm.linear_attention', 'vllm.plamo2_mamba_mixer', 'vllm.gdn_attention', 'vllm.sparse_attn_indexer'], 'use_inductor': True, 'compile_sizes': [8, 4], 'inductor_compile_config': {'enable_auto_functionalized_v2': False}, 'inductor_passes': {}, 'cudagraph_mode': , 'use_cudagraph': False, 'cudagraph_num_of_warmups': 0, 'cudagraph_capture_sizes': [], 'cudagraph_copy_inputs': False, 'full_cuda_graph': False, 'use_inductor_graph_partition': False, 'pass_config': {'enable_noop': True, 'enable_sequence_parallelism': True}, 'max_capture_size': 0, 'local_cache_dir': None} 2025-11-03T17:26:40.9479476Z (EngineCore_DP0 pid=46365) WARNING 11-03 17:26:40 [multiproc_executor.py:744] Reducing Torch parallelism from 24 threads to 1 to avoid unnecessary CPU contention. Set OMP_NUM_THREADS in the external environment to tune this value as needed. 2025-11-03T17:26:40.9481279Z (EngineCore_DP0 pid=46365) INFO 11-03 17:26:40 [shm_broadcast.py:302] vLLM message queue communication handle: Handle(local_reader_ranks=[0, 1, 2, 3], buffer_handle=(4, 16777216, 10, 'psm_a859a127'), local_subscribe_addr='ipc:///tmp/c2eb3d6e-77bd-4be1-a9ab-1397d1dc1fbb', remote_subscribe_addr=None, remote_addr_ipv6=False) 2025-11-03T17:26:41.4342895Z /opt/conda/envs/py_3.12/lib/python3.12/site-packages/torch/cuda/__init__.py:65: FutureWarning: The pynvml package is deprecated. Please install nvidia-ml-py instead. If you did not install pynvml directly, please report this to the maintainers of the package that installed pynvml for you. 2025-11-03T17:26:41.4344129Z import pynvml # type: ignore[import] 2025-11-03T17:26:41.4345253Z /opt/conda/envs/py_3.12/lib/python3.12/site-packages/torch/cuda/__init__.py:65: FutureWarning: The pynvml package is deprecated. Please install nvidia-ml-py instead. If you did not install pynvml directly, please report this to the maintainers of the package that installed pynvml for you. 2025-11-03T17:26:41.4346363Z import pynvml # type: ignore[import] 2025-11-03T17:26:41.4539652Z /opt/conda/envs/py_3.12/lib/python3.12/site-packages/torch/cuda/__init__.py:65: FutureWarning: The pynvml package is deprecated. Please install nvidia-ml-py instead. If you did not install pynvml directly, please report this to the maintainers of the package that installed pynvml for you. 2025-11-03T17:26:41.4541189Z import pynvml # type: ignore[import] 2025-11-03T17:26:41.5796449Z /opt/conda/envs/py_3.12/lib/python3.12/site-packages/torch/cuda/__init__.py:65: FutureWarning: The pynvml package is deprecated. Please install nvidia-ml-py instead. If you did not install pynvml directly, please report this to the maintainers of the package that installed pynvml for you. 2025-11-03T17:26:41.5797691Z import pynvml # type: ignore[import] 2025-11-03T17:26:43.7983811Z INFO 11-03 17:26:43 [__init__.py:224] Automatically detected platform cuda. 2025-11-03T17:26:43.8233681Z INFO 11-03 17:26:43 [__init__.py:224] Automatically detected platform cuda. 2025-11-03T17:26:43.8857759Z INFO 11-03 17:26:43 [__init__.py:224] Automatically detected platform cuda. 2025-11-03T17:26:43.9963291Z INFO 11-03 17:26:43 [__init__.py:224] Automatically detected platform cuda. 2025-11-03T17:26:48.7059195Z INFO 11-03 17:26:48 [shm_broadcast.py:302] vLLM message queue communication handle: Handle(local_reader_ranks=[0], buffer_handle=(1, 10485760, 10, 'psm_46f856b6'), local_subscribe_addr='ipc:///tmp/7fec53f0-0695-4c19-b4b0-8ec03867982f', remote_subscribe_addr=None, remote_addr_ipv6=False) 2025-11-03T17:26:48.7425831Z INFO 11-03 17:26:48 [shm_broadcast.py:302] vLLM message queue communication handle: Handle(local_reader_ranks=[0], buffer_handle=(1, 10485760, 10, 'psm_1ba74fef'), local_subscribe_addr='ipc:///tmp/401fb653-9477-4df7-8716-6f4b5ad866e5', remote_subscribe_addr=None, remote_addr_ipv6=False) 2025-11-03T17:26:48.7823186Z INFO 11-03 17:26:48 [shm_broadcast.py:302] vLLM message queue communication handle: Handle(local_reader_ranks=[0], buffer_handle=(1, 10485760, 10, 'psm_c7bb37fe'), local_subscribe_addr='ipc:///tmp/c146a7d2-a17a-4ca0-8c55-3590fe948a20', remote_subscribe_addr=None, remote_addr_ipv6=False) 2025-11-03T17:26:48.7825142Z INFO 11-03 17:26:48 [shm_broadcast.py:302] vLLM message queue communication handle: Handle(local_reader_ranks=[0], buffer_handle=(1, 10485760, 10, 'psm_6d50f7c6'), local_subscribe_addr='ipc:///tmp/be237345-fe99-4aa2-b514-4b9b28d0764d', remote_subscribe_addr=None, remote_addr_ipv6=False) 2025-11-03T17:26:49.6825605Z [Gloo] Rank 1 is connected to 3 peer ranks. Expected number of connected peer ranks is : 3 2025-11-03T17:26:49.6826819Z [Gloo] Rank 0 is connected to 3 peer ranks. Expected number of connected peer ranks is : 3 2025-11-03T17:26:49.6827914Z [Gloo] Rank 2 is connected to 3 peer ranks. Expected number of connected peer ranks is : 3 2025-11-03T17:26:49.6828958Z [Gloo] Rank 3 is connected to 3 peer ranks. Expected number of connected peer ranks is : 3 2025-11-03T17:26:49.6883285Z [Gloo] Rank 0 is connected to 1 peer ranks. Expected number of connected peer ranks is : 1 2025-11-03T17:26:49.6884334Z [Gloo] Rank 1 is connected to 1 peer ranks. Expected number of connected peer ranks is : 1 2025-11-03T17:26:49.6885342Z [Gloo] Rank 0 is connected to 1 peer ranks. Expected number of connected peer ranks is : 1 2025-11-03T17:26:49.6886341Z [Gloo] Rank 1 is connected to 1 peer ranks. Expected number of connected peer ranks is : 1 2025-11-03T17:26:49.6927644Z INFO 11-03 17:26:49 [__init__.py:1146] Found nccl from library libnccl.so.2 2025-11-03T17:26:49.6928890Z INFO 11-03 17:26:49 [__init__.py:1146] Found nccl from library libnccl.so.2 2025-11-03T17:26:49.6930429Z INFO 11-03 17:26:49 [pynccl.py:109] vLLM is using nccl==2.27.5 2025-11-03T17:26:49.6931160Z INFO 11-03 17:26:49 [pynccl.py:109] vLLM is using nccl==2.27.5 2025-11-03T17:26:49.6937872Z INFO 11-03 17:26:49 [__init__.py:1146] Found nccl from library libnccl.so.2 2025-11-03T17:26:49.6938236Z INFO 11-03 17:26:49 [__init__.py:1146] Found nccl from library libnccl.so.2 2025-11-03T17:26:49.6940677Z INFO 11-03 17:26:49 [pynccl.py:109] vLLM is using nccl==2.27.5 2025-11-03T17:26:49.6940999Z INFO 11-03 17:26:49 [pynccl.py:109] vLLM is using nccl==2.27.5 2025-11-03T17:26:49.8930295Z WARNING 11-03 17:26:49 [symm_mem.py:60] SymmMemCommunicator: Device capability 8.9 not supported, communicator is not available. 2025-11-03T17:26:49.8931386Z WARNING 11-03 17:26:49 [symm_mem.py:60] SymmMemCommunicator: Device capability 8.9 not supported, communicator is not available. 2025-11-03T17:26:49.8952653Z INFO 11-03 17:26:49 [custom_all_reduce.py:37] Skipping P2P check and trusting the driver's P2P report. 2025-11-03T17:26:49.8953277Z INFO 11-03 17:26:49 [custom_all_reduce.py:37] Skipping P2P check and trusting the driver's P2P report. 2025-11-03T17:26:49.8954301Z WARNING 11-03 17:26:49 [custom_all_reduce.py:162] Custom allreduce is disabled because your platform lacks GPU P2P capability or P2P test failed. To silence this warning, specify disable_custom_all_reduce=True explicitly. 2025-11-03T17:26:49.8955672Z WARNING 11-03 17:26:49 [custom_all_reduce.py:162] Custom allreduce is disabled because your platform lacks GPU P2P capability or P2P test failed. To silence this warning, specify disable_custom_all_reduce=True explicitly. 2025-11-03T17:26:49.8972872Z INFO 11-03 17:26:49 [shm_broadcast.py:302] vLLM message queue communication handle: Handle(local_reader_ranks=[1], buffer_handle=(1, 4194304, 6, 'psm_465b4615'), local_subscribe_addr='ipc:///tmp/c26c36bd-d019-49de-a40e-d585faf66bfa', remote_subscribe_addr=None, remote_addr_ipv6=False) 2025-11-03T17:26:49.9816154Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:26:49.9816744Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:26:49.9972380Z WARNING 11-03 17:26:49 [symm_mem.py:60] SymmMemCommunicator: Device capability 8.9 not supported, communicator is not available. 2025-11-03T17:26:49.9973182Z WARNING 11-03 17:26:49 [symm_mem.py:60] SymmMemCommunicator: Device capability 8.9 not supported, communicator is not available. 2025-11-03T17:26:49.9993169Z INFO 11-03 17:26:49 [custom_all_reduce.py:37] Skipping P2P check and trusting the driver's P2P report. 2025-11-03T17:26:49.9993784Z INFO 11-03 17:26:49 [custom_all_reduce.py:37] Skipping P2P check and trusting the driver's P2P report. 2025-11-03T17:26:49.9994810Z WARNING 11-03 17:26:49 [custom_all_reduce.py:162] Custom allreduce is disabled because your platform lacks GPU P2P capability or P2P test failed. To silence this warning, specify disable_custom_all_reduce=True explicitly. 2025-11-03T17:26:49.9996169Z WARNING 11-03 17:26:49 [custom_all_reduce.py:162] Custom allreduce is disabled because your platform lacks GPU P2P capability or P2P test failed. To silence this warning, specify disable_custom_all_reduce=True explicitly. 2025-11-03T17:26:50.0012402Z INFO 11-03 17:26:50 [shm_broadcast.py:302] vLLM message queue communication handle: Handle(local_reader_ranks=[1], buffer_handle=(1, 4194304, 6, 'psm_246fde6d'), local_subscribe_addr='ipc:///tmp/5cdfd75b-2527-4138-8f37-53fab36c7c83', remote_subscribe_addr=None, remote_addr_ipv6=False) 2025-11-03T17:26:50.0035458Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:26:50.0049180Z [Gloo] Rank 0 is connected to 1 peer ranks. Expected number of connected peer ranks is : 1 2025-11-03T17:26:50.0049788Z [Gloo] Rank 1 is connected to 1 peer ranks. Expected number of connected peer ranks is : 1 2025-11-03T17:26:50.0052443Z INFO 11-03 17:26:50 [__init__.py:1146] Found nccl from library libnccl.so.2 2025-11-03T17:26:50.0053163Z INFO 11-03 17:26:50 [pynccl.py:109] vLLM is using nccl==2.27.5 2025-11-03T17:26:50.0054020Z INFO 11-03 17:26:50 [__init__.py:1146] Found nccl from library libnccl.so.2 2025-11-03T17:26:50.0054451Z INFO 11-03 17:26:50 [pynccl.py:109] vLLM is using nccl==2.27.5 2025-11-03T17:26:50.0381964Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:26:50.0383078Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:26:50.1221262Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:26:50.1236128Z [Gloo] Rank 1 is connected to 1 peer ranks. Expected number of connected peer ranks is : 1 2025-11-03T17:26:50.1236945Z [Gloo] Rank 0 is connected to 1 peer ranks. Expected number of connected peer ranks is : 1 2025-11-03T17:26:50.1239350Z INFO 11-03 17:26:50 [__init__.py:1146] Found nccl from library libnccl.so.2 2025-11-03T17:26:50.1239830Z INFO 11-03 17:26:50 [__init__.py:1146] Found nccl from library libnccl.so.2 2025-11-03T17:26:50.1242760Z INFO 11-03 17:26:50 [pynccl.py:109] vLLM is using nccl==2.27.5 2025-11-03T17:26:50.1243172Z INFO 11-03 17:26:50 [pynccl.py:109] vLLM is using nccl==2.27.5 2025-11-03T17:26:50.1579154Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:26:50.1579778Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:26:50.1590876Z [Gloo] Rank 1 is connected to 1 peer ranks. Expected number of connected peer ranks is : 1 2025-11-03T17:26:50.1591355Z [Gloo] Rank 0 is connected to 1 peer ranks. Expected number of connected peer ranks is : 1 2025-11-03T17:26:50.1591825Z [Gloo] Rank 0 is connected to 1 peer ranks. Expected number of connected peer ranks is : 1 2025-11-03T17:26:50.1592568Z [Gloo] Rank 1 is connected to 1 peer ranks. Expected number of connected peer ranks is : 1 2025-11-03T17:26:50.1592986Z INFO 11-03 17:26:50 [__init__.py:1146] Found nccl from library libnccl.so.2 2025-11-03T17:26:50.1593315Z INFO 11-03 17:26:50 [pynccl.py:109] vLLM is using nccl==2.27.5 2025-11-03T17:26:50.1593902Z INFO 11-03 17:26:50 [__init__.py:1146] Found nccl from library libnccl.so.2 2025-11-03T17:26:50.1594250Z INFO 11-03 17:26:50 [pynccl.py:109] vLLM is using nccl==2.27.5 2025-11-03T17:26:50.1594571Z INFO 11-03 17:26:50 [__init__.py:1146] Found nccl from library libnccl.so.2 2025-11-03T17:26:50.1595612Z INFO 11-03 17:26:50 [__init__.py:1146] Found nccl from library libnccl.so.2 2025-11-03T17:26:50.1595921Z INFO 11-03 17:26:50 [pynccl.py:109] vLLM is using nccl==2.27.5 2025-11-03T17:26:50.1596211Z INFO 11-03 17:26:50 [pynccl.py:109] vLLM is using nccl==2.27.5 2025-11-03T17:26:50.2037470Z INFO 11-03 17:26:50 [parallel_state.py:1231] rank 1 in world size 4 is assigned as DP rank 0, PP rank 0, TP rank 1, EP rank 1 2025-11-03T17:26:50.2076215Z INFO 11-03 17:26:50 [parallel_state.py:1231] rank 0 in world size 4 is assigned as DP rank 0, PP rank 0, TP rank 0, EP rank 0 2025-11-03T17:26:50.2076930Z INFO 11-03 17:26:50 [parallel_state.py:1231] rank 3 in world size 4 is assigned as DP rank 0, PP rank 1, TP rank 1, EP rank 1 2025-11-03T17:26:50.2077636Z INFO 11-03 17:26:50 [parallel_state.py:1231] rank 2 in world size 4 is assigned as DP rank 0, PP rank 1, TP rank 0, EP rank 0 2025-11-03T17:26:50.4599184Z INFO 11-03 17:26:50 [topk_topp_sampler.py:59] Using FlashInfer for top-p & top-k sampling. 2025-11-03T17:26:50.4620682Z INFO 11-03 17:26:50 [topk_topp_sampler.py:59] Using FlashInfer for top-p & top-k sampling. 2025-11-03T17:26:50.4621991Z INFO 11-03 17:26:50 [topk_topp_sampler.py:59] Using FlashInfer for top-p & top-k sampling. 2025-11-03T17:26:50.4732280Z INFO 11-03 17:26:50 [topk_topp_sampler.py:59] Using FlashInfer for top-p & top-k sampling. 2025-11-03T17:26:50.4907927Z (Worker_PP0_TP0 pid=46484) INFO 11-03 17:26:50 [gpu_model_runner.py:2840] Starting to load model meta-llama/Llama-3.2-1B-Instruct... 2025-11-03T17:26:50.4933321Z (Worker_PP0_TP1 pid=46485) INFO 11-03 17:26:50 [gpu_model_runner.py:2840] Starting to load model meta-llama/Llama-3.2-1B-Instruct... 2025-11-03T17:26:50.4950108Z (Worker_PP1_TP1 pid=46487) INFO 11-03 17:26:50 [gpu_model_runner.py:2840] Starting to load model meta-llama/Llama-3.2-1B-Instruct... 2025-11-03T17:26:50.5038197Z (Worker_PP1_TP0 pid=46486) INFO 11-03 17:26:50 [gpu_model_runner.py:2840] Starting to load model meta-llama/Llama-3.2-1B-Instruct... 2025-11-03T17:26:50.6963117Z (Worker_PP1_TP1 pid=46487) INFO 11-03 17:26:50 [gpu_model_runner.py:2870] Loading model from scratch... 2025-11-03T17:26:50.7068881Z (Worker_PP0_TP1 pid=46485) INFO 11-03 17:26:50 [gpu_model_runner.py:2870] Loading model from scratch... 2025-11-03T17:26:50.7299503Z (Worker_PP0_TP0 pid=46484) INFO 11-03 17:26:50 [gpu_model_runner.py:2870] Loading model from scratch... 2025-11-03T17:26:50.7416040Z (Worker_PP1_TP0 pid=46486) INFO 11-03 17:26:50 [gpu_model_runner.py:2870] Loading model from scratch... 2025-11-03T17:26:50.7857426Z (Worker_PP1_TP1 pid=46487) INFO 11-03 17:26:50 [cuda.py:404] Using Flash Attention backend on V1 engine. 2025-11-03T17:26:50.7978487Z (Worker_PP0_TP1 pid=46485) INFO 11-03 17:26:50 [cuda.py:404] Using Flash Attention backend on V1 engine. 2025-11-03T17:26:50.8203122Z (Worker_PP0_TP0 pid=46484) INFO 11-03 17:26:50 [cuda.py:404] Using Flash Attention backend on V1 engine. 2025-11-03T17:26:50.8312235Z (Worker_PP1_TP0 pid=46486) INFO 11-03 17:26:50 [cuda.py:404] Using Flash Attention backend on V1 engine. 2025-11-03T17:26:51.2573808Z (Worker_PP1_TP1 pid=46487) INFO 11-03 17:26:51 [gpu_model_runner.py:2902] Model loading took 0.7135 GiB and 0.153338 seconds 2025-11-03T17:26:51.2747744Z (Worker_PP0_TP1 pid=46485) INFO 11-03 17:26:51 [gpu_model_runner.py:2902] Model loading took 0.7135 GiB and 0.154877 seconds 2025-11-03T17:26:51.3591976Z (Worker_PP0_TP0 pid=46484) INFO 11-03 17:26:51 [gpu_model_runner.py:2902] Model loading took 0.7135 GiB and 0.153813 seconds 2025-11-03T17:26:51.3834898Z (Worker_PP1_TP0 pid=46486) INFO 11-03 17:26:51 [gpu_model_runner.py:2902] Model loading took 0.7135 GiB and 0.153766 seconds 2025-11-03T17:26:53.6957998Z (Worker_PP0_TP1 pid=46485) INFO 11-03 17:26:53 [backends.py:594] Using cache directory: /var/lib/jenkins/.cache/vllm/torch_compile_cache/3ff2b0020c/rank_1_0/backbone for vLLM's torch.compile 2025-11-03T17:26:53.6963834Z (Worker_PP0_TP1 pid=46485) INFO 11-03 17:26:53 [backends.py:608] Dynamo bytecode transform time: 2.10 s 2025-11-03T17:26:53.7399498Z (Worker_PP1_TP0 pid=46486) INFO 11-03 17:26:53 [backends.py:594] Using cache directory: /var/lib/jenkins/.cache/vllm/torch_compile_cache/ba6452aee0/rank_2_0/backbone for vLLM's torch.compile 2025-11-03T17:26:53.7408852Z (Worker_PP1_TP0 pid=46486) INFO 11-03 17:26:53 [backends.py:608] Dynamo bytecode transform time: 2.11 s 2025-11-03T17:26:53.7496494Z (Worker_PP1_TP1 pid=46487) INFO 11-03 17:26:53 [backends.py:594] Using cache directory: /var/lib/jenkins/.cache/vllm/torch_compile_cache/ba6452aee0/rank_3_0/backbone for vLLM's torch.compile 2025-11-03T17:26:53.7504074Z (Worker_PP1_TP1 pid=46487) INFO 11-03 17:26:53 [backends.py:608] Dynamo bytecode transform time: 2.16 s 2025-11-03T17:26:53.8091754Z (Worker_PP0_TP0 pid=46484) INFO 11-03 17:26:53 [backends.py:594] Using cache directory: /var/lib/jenkins/.cache/vllm/torch_compile_cache/3ff2b0020c/rank_0_0/backbone for vLLM's torch.compile 2025-11-03T17:26:53.8099817Z (Worker_PP0_TP0 pid=46484) INFO 11-03 17:26:53 [backends.py:608] Dynamo bytecode transform time: 2.19 s 2025-11-03T17:26:54.2903774Z (Worker_PP0_TP1 pid=46485) INFO 11-03 17:26:54 [backends.py:179] Directly load the compiled graph(s) for dynamic shape from the cache, took 0.384 s 2025-11-03T17:26:54.3395354Z (Worker_PP1_TP1 pid=46487) INFO 11-03 17:26:54 [backends.py:179] Directly load the compiled graph(s) for dynamic shape from the cache, took 0.387 s 2025-11-03T17:26:54.3468777Z (Worker_PP1_TP0 pid=46486) INFO 11-03 17:26:54 [backends.py:179] Directly load the compiled graph(s) for dynamic shape from the cache, took 0.400 s 2025-11-03T17:26:54.4226374Z (Worker_PP0_TP0 pid=46484) INFO 11-03 17:26:54 [backends.py:179] Directly load the compiled graph(s) for dynamic shape from the cache, took 0.400 s 2025-11-03T17:26:55.2973768Z (Worker_PP0_TP1 pid=46485) INFO 11-03 17:26:55 [gpu_worker.py:314] Available KV cache memory: 19.02 GiB 2025-11-03T17:26:55.3000247Z (Worker_PP0_TP0 pid=46484) INFO 11-03 17:26:55 [gpu_worker.py:314] Available KV cache memory: 19.02 GiB 2025-11-03T17:26:55.3306558Z (Worker_PP1_TP1 pid=46487) INFO 11-03 17:26:55 [gpu_worker.py:314] Available KV cache memory: 18.99 GiB 2025-11-03T17:26:55.3644956Z (Worker_PP1_TP0 pid=46486) INFO 11-03 17:26:55 [gpu_worker.py:314] Available KV cache memory: 18.99 GiB 2025-11-03T17:26:55.6395121Z (EngineCore_DP0 pid=46365) INFO 11-03 17:26:55 [kv_cache_utils.py:1199] GPU KV cache size: 2,489,088 tokens 2025-11-03T17:26:55.6396353Z (EngineCore_DP0 pid=46365) INFO 11-03 17:26:55 [kv_cache_utils.py:1204] Maximum concurrency for 2,048 tokens per request: 1215.38x 2025-11-03T17:26:55.6397379Z (EngineCore_DP0 pid=46365) INFO 11-03 17:26:55 [kv_cache_utils.py:1199] GPU KV cache size: 2,489,088 tokens 2025-11-03T17:26:55.6398203Z (EngineCore_DP0 pid=46365) INFO 11-03 17:26:55 [kv_cache_utils.py:1204] Maximum concurrency for 2,048 tokens per request: 1215.38x 2025-11-03T17:26:55.6399029Z (EngineCore_DP0 pid=46365) INFO 11-03 17:26:55 [kv_cache_utils.py:1199] GPU KV cache size: 2,489,088 tokens 2025-11-03T17:26:55.6400224Z (EngineCore_DP0 pid=46365) INFO 11-03 17:26:55 [kv_cache_utils.py:1204] Maximum concurrency for 2,048 tokens per request: 1215.38x 2025-11-03T17:26:55.6401052Z (EngineCore_DP0 pid=46365) INFO 11-03 17:26:55 [kv_cache_utils.py:1199] GPU KV cache size: 2,489,088 tokens 2025-11-03T17:26:55.6401816Z (EngineCore_DP0 pid=46365) INFO 11-03 17:26:55 [kv_cache_utils.py:1204] Maximum concurrency for 2,048 tokens per request: 1215.38x 2025-11-03T17:26:55.6469783Z (Worker_PP0_TP0 pid=46484) INFO 11-03 17:26:55 [gpu_worker.py:351] Compile and warming up model for size 8 2025-11-03T17:26:55.6470600Z (Worker_PP1_TP0 pid=46486) INFO 11-03 17:26:55 [gpu_worker.py:351] Compile and warming up model for size 8 2025-11-03T17:26:55.6471317Z (Worker_PP0_TP1 pid=46485) INFO 11-03 17:26:55 [gpu_worker.py:351] Compile and warming up model for size 8 2025-11-03T17:26:55.6471898Z (Worker_PP1_TP1 pid=46487) INFO 11-03 17:26:55 [gpu_worker.py:351] Compile and warming up model for size 8 2025-11-03T17:26:56.4784598Z (Worker_PP1_TP0 pid=46486) INFO 11-03 17:26:56 [backends.py:185] Directly load the compiled graph(s) for shape 8 from the cache, took 0.831 s 2025-11-03T17:26:56.4813672Z (Worker_PP1_TP1 pid=46487) INFO 11-03 17:26:56 [backends.py:185] Directly load the compiled graph(s) for shape 8 from the cache, took 0.834 s 2025-11-03T17:26:56.4819893Z (Worker_PP1_TP0 pid=46486) INFO 11-03 17:26:56 [gpu_worker.py:351] Compile and warming up model for size 4 2025-11-03T17:26:56.4820654Z (Worker_PP1_TP1 pid=46487) INFO 11-03 17:26:56 [gpu_worker.py:351] Compile and warming up model for size 4 2025-11-03T17:26:56.5139242Z (Worker_PP0_TP0 pid=46484) INFO 11-03 17:26:56 [backends.py:185] Directly load the compiled graph(s) for shape 8 from the cache, took 0.867 s 2025-11-03T17:26:56.5144339Z (Worker_PP0_TP0 pid=46484) INFO 11-03 17:26:56 [gpu_worker.py:351] Compile and warming up model for size 4 2025-11-03T17:26:56.5205111Z (Worker_PP0_TP1 pid=46485) INFO 11-03 17:26:56 [backends.py:185] Directly load the compiled graph(s) for shape 8 from the cache, took 0.873 s 2025-11-03T17:26:56.5211488Z (Worker_PP0_TP1 pid=46485) INFO 11-03 17:26:56 [gpu_worker.py:351] Compile and warming up model for size 4 2025-11-03T17:26:56.8642990Z (Worker_PP1_TP1 pid=46487) INFO 11-03 17:26:56 [backends.py:185] Directly load the compiled graph(s) for shape 4 from the cache, took 0.382 s 2025-11-03T17:26:56.8643894Z (Worker_PP1_TP1 pid=46487) INFO 11-03 17:26:56 [monitor.py:33] torch.compile takes 2.16 s in total 2025-11-03T17:26:56.8677768Z (Worker_PP1_TP0 pid=46486) INFO 11-03 17:26:56 [backends.py:185] Directly load the compiled graph(s) for shape 4 from the cache, took 0.385 s 2025-11-03T17:26:56.8678908Z (Worker_PP1_TP0 pid=46486) INFO 11-03 17:26:56 [monitor.py:33] torch.compile takes 2.11 s in total 2025-11-03T17:26:57.0636070Z (Worker_PP0_TP0 pid=46484) INFO 11-03 17:26:57 [backends.py:185] Directly load the compiled graph(s) for shape 4 from the cache, took 0.549 s 2025-11-03T17:26:57.0636963Z (Worker_PP0_TP0 pid=46484) INFO 11-03 17:26:57 [monitor.py:33] torch.compile takes 2.19 s in total 2025-11-03T17:26:57.0697196Z (Worker_PP0_TP1 pid=46485) INFO 11-03 17:26:57 [backends.py:185] Directly load the compiled graph(s) for shape 4 from the cache, took 0.548 s 2025-11-03T17:26:57.0698077Z (Worker_PP0_TP1 pid=46485) INFO 11-03 17:26:57 [monitor.py:33] torch.compile takes 2.10 s in total 2025-11-03T17:26:57.0716449Z (EngineCore_DP0 pid=46365) INFO 11-03 17:26:57 [core.py:240] init engine (profile, create kv cache, warmup model) took 5.69 seconds 2025-11-03T17:26:57.8551796Z (EngineCore_DP0 pid=46365) INFO 11-03 17:26:57 [core.py:178] Batch queue is enabled with size 2 2025-11-03T17:26:57.8580296Z (EngineCore_DP0 pid=46365) INFO 11-03 17:26:57 [vllm.py:354] Cudagraph is disabled under eager mode 2025-11-03T17:26:58.0424126Z (APIServer pid=46243) INFO 11-03 17:26:58 [loggers.py:172] Engine 000: vllm cache_config_info with initialization after num_gpu_blocks is: 155568 2025-11-03T17:26:58.0544265Z (EngineCore_DP0 pid=46365) INFO 11-03 17:26:58 [gc_utils.py:40] GC Debug Config. enabled:False,top_objects:-1 2025-11-03T17:26:58.0764744Z (APIServer pid=46243) INFO 11-03 17:26:58 [api_server.py:1628] Supported tasks: ['generate'] 2025-11-03T17:26:58.1024120Z (APIServer pid=46243) WARNING 11-03 17:26:58 [model.py:1581] Default sampling parameters have been overridden by the model's Hugging Face generation config recommended from the model creator. If this is not intended, please relaunch vLLM instance with `--generation-config vllm`. 2025-11-03T17:26:58.1026519Z (APIServer pid=46243) INFO 11-03 17:26:58 [serving_responses.py:157] Using default chat sampling params from model: {'temperature': 0.6, 'top_p': 0.9} 2025-11-03T17:26:58.1248856Z (APIServer pid=46243) INFO 11-03 17:26:58 [serving_chat.py:159] Using default chat sampling params from model: {'temperature': 0.6, 'top_p': 0.9} 2025-11-03T17:26:58.1543365Z (APIServer pid=46243) INFO 11-03 17:26:58 [serving_completion.py:67] Using default completion sampling params from model: {'temperature': 0.6, 'top_p': 0.9} 2025-11-03T17:26:58.1544892Z (APIServer pid=46243) INFO 11-03 17:26:58 [api_server.py:1939] Starting vLLM API server 0 on http://0.0.0.0:52725 2025-11-03T17:26:58.1545965Z (APIServer pid=46243) INFO 11-03 17:26:58 [launcher.py:38] Available routes are: 2025-11-03T17:26:58.1546981Z (APIServer pid=46243) INFO 11-03 17:26:58 [launcher.py:46] Route: /openapi.json, Methods: HEAD, GET 2025-11-03T17:26:58.1548022Z (APIServer pid=46243) INFO 11-03 17:26:58 [launcher.py:46] Route: /docs, Methods: HEAD, GET 2025-11-03T17:26:58.1549148Z (APIServer pid=46243) INFO 11-03 17:26:58 [launcher.py:46] Route: /docs/oauth2-redirect, Methods: HEAD, GET 2025-11-03T17:26:58.1550232Z (APIServer pid=46243) INFO 11-03 17:26:58 [launcher.py:46] Route: /redoc, Methods: HEAD, GET 2025-11-03T17:26:58.1551222Z (APIServer pid=46243) INFO 11-03 17:26:58 [launcher.py:46] Route: /health, Methods: GET 2025-11-03T17:26:58.1552176Z (APIServer pid=46243) INFO 11-03 17:26:58 [launcher.py:46] Route: /load, Methods: GET 2025-11-03T17:26:58.1553119Z (APIServer pid=46243) INFO 11-03 17:26:58 [launcher.py:46] Route: /ping, Methods: POST 2025-11-03T17:26:58.1554075Z (APIServer pid=46243) INFO 11-03 17:26:58 [launcher.py:46] Route: /ping, Methods: GET 2025-11-03T17:26:58.1555044Z (APIServer pid=46243) INFO 11-03 17:26:58 [launcher.py:46] Route: /tokenize, Methods: POST 2025-11-03T17:26:58.1556265Z (APIServer pid=46243) INFO 11-03 17:26:58 [launcher.py:46] Route: /detokenize, Methods: POST 2025-11-03T17:26:58.1557274Z (APIServer pid=46243) INFO 11-03 17:26:58 [launcher.py:46] Route: /v1/models, Methods: GET 2025-11-03T17:26:58.1558245Z (APIServer pid=46243) INFO 11-03 17:26:58 [launcher.py:46] Route: /version, Methods: GET 2025-11-03T17:26:58.1559260Z (APIServer pid=46243) INFO 11-03 17:26:58 [launcher.py:46] Route: /v1/responses, Methods: POST 2025-11-03T17:26:58.1560355Z (APIServer pid=46243) INFO 11-03 17:26:58 [launcher.py:46] Route: /v1/responses/{response_id}, Methods: GET 2025-11-03T17:26:58.1561536Z (APIServer pid=46243) INFO 11-03 17:26:58 [launcher.py:46] Route: /v1/responses/{response_id}/cancel, Methods: POST 2025-11-03T17:26:58.1562691Z (APIServer pid=46243) INFO 11-03 17:26:58 [launcher.py:46] Route: /v1/chat/completions, Methods: POST 2025-11-03T17:26:58.1563772Z (APIServer pid=46243) INFO 11-03 17:26:58 [launcher.py:46] Route: /v1/completions, Methods: POST 2025-11-03T17:26:58.1565030Z (APIServer pid=46243) INFO 11-03 17:26:58 [launcher.py:46] Route: /v1/embeddings, Methods: POST 2025-11-03T17:26:58.1566066Z (APIServer pid=46243) INFO 11-03 17:26:58 [launcher.py:46] Route: /pooling, Methods: POST 2025-11-03T17:26:58.1567063Z (APIServer pid=46243) INFO 11-03 17:26:58 [launcher.py:46] Route: /classify, Methods: POST 2025-11-03T17:26:58.1568187Z (APIServer pid=46243) INFO 11-03 17:26:58 [launcher.py:46] Route: /score, Methods: POST 2025-11-03T17:26:58.1569286Z (APIServer pid=46243) INFO 11-03 17:26:58 [launcher.py:46] Route: /v1/score, Methods: POST 2025-11-03T17:26:58.1570357Z (APIServer pid=46243) INFO 11-03 17:26:58 [launcher.py:46] Route: /v1/audio/transcriptions, Methods: POST 2025-11-03T17:26:58.1571474Z (APIServer pid=46243) INFO 11-03 17:26:58 [launcher.py:46] Route: /v1/audio/translations, Methods: POST 2025-11-03T17:26:58.1572510Z (APIServer pid=46243) INFO 11-03 17:26:58 [launcher.py:46] Route: /rerank, Methods: POST 2025-11-03T17:26:58.1573500Z (APIServer pid=46243) INFO 11-03 17:26:58 [launcher.py:46] Route: /v1/rerank, Methods: POST 2025-11-03T17:26:58.1574498Z (APIServer pid=46243) INFO 11-03 17:26:58 [launcher.py:46] Route: /v2/rerank, Methods: POST 2025-11-03T17:26:58.1575529Z (APIServer pid=46243) INFO 11-03 17:26:58 [launcher.py:46] Route: /scale_elastic_ep, Methods: POST 2025-11-03T17:26:58.1576614Z (APIServer pid=46243) INFO 11-03 17:26:58 [launcher.py:46] Route: /is_scaling_elastic_ep, Methods: POST 2025-11-03T17:26:58.1577670Z (APIServer pid=46243) INFO 11-03 17:26:58 [launcher.py:46] Route: /invocations, Methods: POST 2025-11-03T17:26:58.1578657Z (APIServer pid=46243) INFO 11-03 17:26:58 [launcher.py:46] Route: /metrics, Methods: GET 2025-11-03T17:26:58.1704012Z (APIServer pid=46243) INFO: Started server process [46243] 2025-11-03T17:26:58.1704959Z (APIServer pid=46243) INFO: Waiting for application startup. 2025-11-03T17:26:58.3801922Z (APIServer pid=46243) INFO: Application startup complete. 2025-11-03T17:26:58.6445146Z (APIServer pid=46243) INFO: 127.0.0.1:57756 - "GET /health HTTP/1.1" 200 OK 2025-11-03T17:26:58.8928763Z (APIServer pid=46243) INFO: 127.0.0.1:57760 - "GET /v1/models HTTP/1.1" 200 OK 2025-11-03T17:26:58.9227650Z (Worker_PP0_TP0 pid=46484) /opt/conda/envs/py_3.12/lib/python3.12/site-packages/vllm/distributed/parallel_state.py:530: UserWarning: The given buffer is not writable, and PyTorch does not support non-writable tensors. This means you can write to the underlying (supposedly non-writable) buffer using the tensor. You may want to copy the buffer to protect its data or make it writable before converting it to a tensor. This type of warning will be suppressed for the rest of this program. (Triggered internally at /var/lib/jenkins/workspace/torch/csrc/utils/tensor_new.cpp:1581.) 2025-11-03T17:26:58.9231803Z (Worker_PP0_TP0 pid=46484) object_tensor = torch.frombuffer(pickle.dumps(obj), dtype=torch.uint8) 2025-11-03T17:26:58.9233338Z [rank0]:[W1103 17:26:58.971535824 ProcessGroupNCCL.cpp:4063] Warning: An unbatched P2P op (send/recv) was called on this ProcessGroup with size 2. In lazy initialization mode, this will result in a new 2-rank NCCL communicator to be created. (function operator()) 2025-11-03T17:26:58.9236792Z (Worker_PP0_TP1 pid=46485) /opt/conda/envs/py_3.12/lib/python3.12/site-packages/vllm/distributed/parallel_state.py:530: UserWarning: The given buffer is not writable, and PyTorch does not support non-writable tensors. This means you can write to the underlying (supposedly non-writable) buffer using the tensor. You may want to copy the buffer to protect its data or make it writable before converting it to a tensor. This type of warning will be suppressed for the rest of this program. (Triggered internally at /var/lib/jenkins/workspace/torch/csrc/utils/tensor_new.cpp:1581.) 2025-11-03T17:26:58.9239196Z (Worker_PP0_TP1 pid=46485) object_tensor = torch.frombuffer(pickle.dumps(obj), dtype=torch.uint8) 2025-11-03T17:26:58.9240324Z [rank2]:[W1103 17:26:58.971692476 ProcessGroupNCCL.cpp:4063] Warning: An unbatched P2P op (send/recv) was called on this ProcessGroup with size 2. In lazy initialization mode, this will result in a new 2-rank NCCL communicator to be created. (function operator()) 2025-11-03T17:26:58.9242115Z [rank1]:[W1103 17:26:58.972044933 ProcessGroupNCCL.cpp:4063] Warning: An unbatched P2P op (send/recv) was called on this ProcessGroup with size 2. In lazy initialization mode, this will result in a new 2-rank NCCL communicator to be created. (function operator()) 2025-11-03T17:26:58.9243662Z [rank3]:[W1103 17:26:58.972164632 ProcessGroupNCCL.cpp:4063] Warning: An unbatched P2P op (send/recv) was called on this ProcessGroup with size 2. In lazy initialization mode, this will result in a new 2-rank NCCL communicator to be created. (function operator()) 2025-11-03T17:26:59.0491190Z (APIServer pid=46243) INFO: 127.0.0.1:57760 - "POST /v1/completions HTTP/1.1" 200 OK 2025-11-03T17:26:59.1147746Z (APIServer pid=46243) INFO: 127.0.0.1:57760 - "POST /v1/completions HTTP/1.1" 200 OK 2025-11-03T17:26:59.2580754Z (APIServer pid=46243) INFO: 127.0.0.1:57760 - "POST /v1/completions HTTP/1.1" 200 OK 2025-11-03T17:26:59.3499156Z (APIServer pid=46243) INFO: 127.0.0.1:57760 - "POST /v1/completions HTTP/1.1" 200 OK 2025-11-03T17:26:59.4132511Z (APIServer pid=46243) INFO: 127.0.0.1:57760 - "POST /v1/completions HTTP/1.1" 200 OK 2025-11-03T17:26:59.4209691Z (APIServer pid=46243) INFO: 127.0.0.1:57760 - "POST /v1/completions HTTP/1.1" 200 OK 2025-11-03T17:26:59.4931982Z (APIServer pid=46243) INFO 11-03 17:26:59 [launcher.py:110] Shutting down FastAPI HTTP server. 2025-11-03T17:26:59.4934736Z (Worker_PP0_TP0 pid=46484) INFO 11-03 17:26:59 [multiproc_executor.py:583] Parent process exited, terminating worker 2025-11-03T17:26:59.4936049Z (Worker_PP0_TP0 pid=46484) INFO 11-03 17:26:59 [multiproc_executor.py:624] WorkerProc shutting down. 2025-11-03T17:26:59.4939789Z (Worker_PP0_TP1 pid=46485) INFO 11-03 17:26:59 [multiproc_executor.py:583] Parent process exited, terminating worker 2025-11-03T17:26:59.4943115Z (Worker_PP0_TP1 pid=46485) INFO 11-03 17:26:59 [multiproc_executor.py:624] WorkerProc shutting down. 2025-11-03T17:26:59.4944128Z (Worker_PP1_TP0 pid=46486) INFO 11-03 17:26:59 [multiproc_executor.py:583] Parent process exited, terminating worker 2025-11-03T17:26:59.4945667Z (Worker_PP1_TP0 pid=46486) INFO 11-03 17:26:59 [multiproc_executor.py:624] WorkerProc shutting down. 2025-11-03T17:26:59.4947009Z (Worker_PP1_TP1 pid=46487) INFO 11-03 17:26:59 [multiproc_executor.py:583] Parent process exited, terminating worker 2025-11-03T17:26:59.4948346Z (Worker_PP1_TP1 pid=46487) INFO 11-03 17:26:59 [multiproc_executor.py:624] WorkerProc shutting down. 2025-11-03T17:27:02.5099682Z (APIServer pid=46243) INFO: Shutting down 2025-11-03T17:27:02.6103705Z (APIServer pid=46243) INFO: Waiting for application shutdown. 2025-11-03T17:27:02.6105919Z (APIServer pid=46243) INFO: Application shutdown complete. 2025-11-03T17:27:02.9902896Z INFO 11-03 17:27:02 [model.py:653] Resolved architecture: LlamaForCausalLM 2025-11-03T17:27:02.9904141Z WARNING 11-03 17:27:02 [model.py:1942] Casting torch.bfloat16 to torch.float16. 2025-11-03T17:27:02.9905007Z INFO 11-03 17:27:02 [model.py:1714] Using max model len 2048 2025-11-03T17:27:02.9907700Z Launching RemoteOpenAIServer with: vllm serve meta-llama/Llama-3.2-1B-Instruct --dtype float16 --max-model-len 2048 --max-num-seqs 8 --enforce-eager --tokenizer-mode auto --tensor-parallel-size 2 --distributed-executor-backend mp --load-format dummy --port 56525 --seed 0 2025-11-03T17:27:03.4967426Z /opt/conda/envs/py_3.12/lib/python3.12/site-packages/torch/cuda/__init__.py:65: FutureWarning: The pynvml package is deprecated. Please install nvidia-ml-py instead. If you did not install pynvml directly, please report this to the maintainers of the package that installed pynvml for you. 2025-11-03T17:27:03.4969339Z import pynvml # type: ignore[import] 2025-11-03T17:27:05.7401269Z INFO 11-03 17:27:05 [__init__.py:224] Automatically detected platform cuda. 2025-11-03T17:27:09.4147296Z (APIServer pid=46860) INFO 11-03 17:27:09 [api_server.py:1870] vLLM API server version 0.11.0rc2.dev389+ge51928192.d20251103 2025-11-03T17:27:09.4170667Z (APIServer pid=46860) INFO 11-03 17:27:09 [utils.py:239] non-default args: {'model_tag': 'meta-llama/Llama-3.2-1B-Instruct', 'port': 56525, 'model': 'meta-llama/Llama-3.2-1B-Instruct', 'dtype': 'float16', 'seed': 0, 'max_model_len': 2048, 'enforce_eager': True, 'load_format': 'dummy', 'distributed_executor_backend': 'mp', 'tensor_parallel_size': 2, 'max_num_seqs': 8} 2025-11-03T17:27:09.7140428Z (APIServer pid=46860) INFO 11-03 17:27:09 [model.py:653] Resolved architecture: LlamaForCausalLM 2025-11-03T17:27:09.7141773Z (APIServer pid=46860) `torch_dtype` is deprecated! Use `dtype` instead! 2025-11-03T17:27:09.7143098Z (APIServer pid=46860) WARNING 11-03 17:27:09 [model.py:1942] Casting torch.bfloat16 to torch.float16. 2025-11-03T17:27:09.7144154Z (APIServer pid=46860) INFO 11-03 17:27:09 [model.py:1714] Using max model len 2048 2025-11-03T17:27:09.8721177Z (APIServer pid=46860) INFO 11-03 17:27:09 [scheduler.py:225] Chunked prefill is enabled with max_num_batched_tokens=2048. 2025-11-03T17:27:09.8725797Z (APIServer pid=46860) INFO 11-03 17:27:09 [vllm.py:354] Cudagraph is disabled under eager mode 2025-11-03T17:27:10.8580714Z /opt/conda/envs/py_3.12/lib/python3.12/site-packages/torch/cuda/__init__.py:65: FutureWarning: The pynvml package is deprecated. Please install nvidia-ml-py instead. If you did not install pynvml directly, please report this to the maintainers of the package that installed pynvml for you. 2025-11-03T17:27:10.8582488Z import pynvml # type: ignore[import] 2025-11-03T17:27:13.1689241Z INFO 11-03 17:27:13 [__init__.py:224] Automatically detected platform cuda. 2025-11-03T17:27:16.8554121Z (EngineCore_DP0 pid=46982) INFO 11-03 17:27:16 [core.py:727] Waiting for init message from front-end. 2025-11-03T17:27:16.8824560Z (EngineCore_DP0 pid=46982) INFO 11-03 17:27:16 [core.py:94] Initializing a V1 LLM engine (v0.11.0rc2.dev389+ge51928192.d20251103) with config: model='meta-llama/Llama-3.2-1B-Instruct', speculative_config=None, tokenizer='meta-llama/Llama-3.2-1B-Instruct', skip_tokenizer_init=False, tokenizer_mode=auto, revision=None, tokenizer_revision=None, trust_remote_code=False, dtype=torch.float16, max_seq_len=2048, download_dir=None, load_format=dummy, tensor_parallel_size=2, pipeline_parallel_size=1, data_parallel_size=1, disable_custom_all_reduce=False, quantization=None, enforce_eager=True, kv_cache_dtype=auto, device_config=cuda, structured_outputs_config=StructuredOutputsConfig(backend='auto', disable_fallback=False, disable_any_whitespace=False, disable_additional_properties=False, reasoning_parser=''), observability_config=ObservabilityConfig(show_hidden_metrics_for_version=None, otlp_traces_endpoint=None, collect_detailed_traces=None), seed=0, served_model_name=meta-llama/Llama-3.2-1B-Instruct, enable_prefix_caching=True, chunked_prefill_enabled=True, pooler_config=None, compilation_config={'level': 0, 'debug_dump_path': None, 'cache_dir': '', 'backend': '', 'custom_ops': [], 'splitting_ops': None, 'use_inductor': True, 'compile_sizes': [], 'inductor_compile_config': {'enable_auto_functionalized_v2': False}, 'inductor_passes': {}, 'cudagraph_mode': , 'use_cudagraph': False, 'cudagraph_num_of_warmups': 0, 'cudagraph_capture_sizes': [], 'cudagraph_copy_inputs': False, 'full_cuda_graph': False, 'use_inductor_graph_partition': False, 'pass_config': {}, 'max_capture_size': 0, 'local_cache_dir': None} 2025-11-03T17:27:16.8829900Z (EngineCore_DP0 pid=46982) WARNING 11-03 17:27:16 [multiproc_executor.py:744] Reducing Torch parallelism from 24 threads to 1 to avoid unnecessary CPU contention. Set OMP_NUM_THREADS in the external environment to tune this value as needed. 2025-11-03T17:27:16.8837295Z (EngineCore_DP0 pid=46982) INFO 11-03 17:27:16 [shm_broadcast.py:302] vLLM message queue communication handle: Handle(local_reader_ranks=[0, 1], buffer_handle=(2, 16777216, 10, 'psm_1b14ed64'), local_subscribe_addr='ipc:///tmp/dc7cf4ee-ff1a-46d8-af68-72854c0480f5', remote_subscribe_addr=None, remote_addr_ipv6=False) 2025-11-03T17:27:17.3681403Z /opt/conda/envs/py_3.12/lib/python3.12/site-packages/torch/cuda/__init__.py:65: FutureWarning: The pynvml package is deprecated. Please install nvidia-ml-py instead. If you did not install pynvml directly, please report this to the maintainers of the package that installed pynvml for you. 2025-11-03T17:27:17.3682665Z import pynvml # type: ignore[import] 2025-11-03T17:27:17.3683820Z /opt/conda/envs/py_3.12/lib/python3.12/site-packages/torch/cuda/__init__.py:65: FutureWarning: The pynvml package is deprecated. Please install nvidia-ml-py instead. If you did not install pynvml directly, please report this to the maintainers of the package that installed pynvml for you. 2025-11-03T17:27:17.3684942Z import pynvml # type: ignore[import] 2025-11-03T17:27:19.7056273Z INFO 11-03 17:27:19 [__init__.py:224] Automatically detected platform cuda. 2025-11-03T17:27:19.7245374Z INFO 11-03 17:27:19 [__init__.py:224] Automatically detected platform cuda. 2025-11-03T17:27:24.4931103Z INFO 11-03 17:27:24 [shm_broadcast.py:302] vLLM message queue communication handle: Handle(local_reader_ranks=[0], buffer_handle=(1, 10485760, 10, 'psm_ff7b0df0'), local_subscribe_addr='ipc:///tmp/cb081d74-3905-4e90-82ef-54c832857b19', remote_subscribe_addr=None, remote_addr_ipv6=False) 2025-11-03T17:27:24.5025005Z INFO 11-03 17:27:24 [shm_broadcast.py:302] vLLM message queue communication handle: Handle(local_reader_ranks=[0], buffer_handle=(1, 10485760, 10, 'psm_6135d387'), local_subscribe_addr='ipc:///tmp/0eee4bbf-94ba-41d5-bc48-9b72c652799a', remote_subscribe_addr=None, remote_addr_ipv6=False) 2025-11-03T17:27:24.6959755Z [Gloo] Rank 1 is connected to 1 peer ranks. Expected number of connected peer ranks is : 1 2025-11-03T17:27:24.6960982Z [Gloo] Rank 0 is connected to 1 peer ranks. Expected number of connected peer ranks is : 1 2025-11-03T17:27:24.7003156Z [Gloo] Rank 1 is connected to 1 peer ranks. Expected number of connected peer ranks is : 1 2025-11-03T17:27:24.7004245Z [Gloo] Rank 0 is connected to 1 peer ranks. Expected number of connected peer ranks is : 1 2025-11-03T17:27:24.7045956Z INFO 11-03 17:27:24 [__init__.py:1146] Found nccl from library libnccl.so.2 2025-11-03T17:27:24.7048220Z INFO 11-03 17:27:24 [__init__.py:1146] Found nccl from library libnccl.so.2 2025-11-03T17:27:24.7049280Z INFO 11-03 17:27:24 [pynccl.py:109] vLLM is using nccl==2.27.5 2025-11-03T17:27:24.7050427Z INFO 11-03 17:27:24 [pynccl.py:109] vLLM is using nccl==2.27.5 2025-11-03T17:27:24.8773392Z WARNING 11-03 17:27:24 [symm_mem.py:60] SymmMemCommunicator: Device capability 8.9 not supported, communicator is not available. 2025-11-03T17:27:24.8774266Z WARNING 11-03 17:27:24 [symm_mem.py:60] SymmMemCommunicator: Device capability 8.9 not supported, communicator is not available. 2025-11-03T17:27:24.8798169Z INFO 11-03 17:27:24 [custom_all_reduce.py:37] Skipping P2P check and trusting the driver's P2P report. 2025-11-03T17:27:24.8798843Z INFO 11-03 17:27:24 [custom_all_reduce.py:37] Skipping P2P check and trusting the driver's P2P report. 2025-11-03T17:27:24.8800168Z WARNING 11-03 17:27:24 [custom_all_reduce.py:162] Custom allreduce is disabled because your platform lacks GPU P2P capability or P2P test failed. To silence this warning, specify disable_custom_all_reduce=True explicitly. 2025-11-03T17:27:24.8801560Z WARNING 11-03 17:27:24 [custom_all_reduce.py:162] Custom allreduce is disabled because your platform lacks GPU P2P capability or P2P test failed. To silence this warning, specify disable_custom_all_reduce=True explicitly. 2025-11-03T17:27:24.9825378Z INFO 11-03 17:27:24 [shm_broadcast.py:302] vLLM message queue communication handle: Handle(local_reader_ranks=[1], buffer_handle=(1, 4194304, 6, 'psm_b8f216d1'), local_subscribe_addr='ipc:///tmp/8938b13d-5035-4406-8620-5dd1a27cc142', remote_subscribe_addr=None, remote_addr_ipv6=False) 2025-11-03T17:27:24.9848639Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:27:24.9854311Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:27:24.9857757Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:27:24.9866545Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:27:25.0622494Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:27:25.0630682Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:27:25.0641582Z [Gloo] Rank 0 is connected to 1 peer ranks. Expected number of connected peer ranks is : 1 2025-11-03T17:27:25.0642181Z [Gloo] Rank 1 is connected to 1 peer ranks. Expected number of connected peer ranks is : 1 2025-11-03T17:27:25.0644111Z INFO 11-03 17:27:25 [__init__.py:1146] Found nccl from library libnccl.so.2 2025-11-03T17:27:25.0644573Z INFO 11-03 17:27:25 [pynccl.py:109] vLLM is using nccl==2.27.5 2025-11-03T17:27:25.0645793Z INFO 11-03 17:27:25 [__init__.py:1146] Found nccl from library libnccl.so.2 2025-11-03T17:27:25.0646425Z INFO 11-03 17:27:25 [pynccl.py:109] vLLM is using nccl==2.27.5 2025-11-03T17:27:25.0987587Z INFO 11-03 17:27:25 [parallel_state.py:1231] rank 1 in world size 2 is assigned as DP rank 0, PP rank 0, TP rank 1, EP rank 1 2025-11-03T17:27:25.0988431Z INFO 11-03 17:27:25 [parallel_state.py:1231] rank 0 in world size 2 is assigned as DP rank 0, PP rank 0, TP rank 0, EP rank 0 2025-11-03T17:27:25.3201573Z INFO 11-03 17:27:25 [topk_topp_sampler.py:59] Using FlashInfer for top-p & top-k sampling. 2025-11-03T17:27:25.3203700Z INFO 11-03 17:27:25 [topk_topp_sampler.py:59] Using FlashInfer for top-p & top-k sampling. 2025-11-03T17:27:25.3528140Z (Worker_TP1 pid=47102) INFO 11-03 17:27:25 [gpu_model_runner.py:2840] Starting to load model meta-llama/Llama-3.2-1B-Instruct... 2025-11-03T17:27:25.3537005Z (Worker_TP0 pid=47101) INFO 11-03 17:27:25 [gpu_model_runner.py:2840] Starting to load model meta-llama/Llama-3.2-1B-Instruct... 2025-11-03T17:27:25.5627977Z (Worker_TP1 pid=47102) INFO 11-03 17:27:25 [gpu_model_runner.py:2870] Loading model from scratch... 2025-11-03T17:27:25.5632261Z (Worker_TP0 pid=47101) INFO 11-03 17:27:25 [gpu_model_runner.py:2870] Loading model from scratch... 2025-11-03T17:27:25.6530936Z (Worker_TP1 pid=47102) INFO 11-03 17:27:25 [cuda.py:404] Using Flash Attention backend on V1 engine. 2025-11-03T17:27:25.6543258Z (Worker_TP0 pid=47101) INFO 11-03 17:27:25 [cuda.py:404] Using Flash Attention backend on V1 engine. 2025-11-03T17:27:26.0848375Z (Worker_TP1 pid=47102) INFO 11-03 17:27:26 [gpu_model_runner.py:2902] Model loading took 1.1667 GiB and 0.119732 seconds 2025-11-03T17:27:26.1000038Z (Worker_TP0 pid=47101) INFO 11-03 17:27:26 [gpu_model_runner.py:2902] Model loading took 1.1667 GiB and 0.120791 seconds 2025-11-03T17:27:28.0029253Z (Worker_TP0 pid=47101) INFO 11-03 17:27:28 [gpu_worker.py:314] Available KV cache memory: 18.53 GiB 2025-11-03T17:27:28.1067851Z (Worker_TP1 pid=47102) INFO 11-03 17:27:28 [gpu_worker.py:314] Available KV cache memory: 18.53 GiB 2025-11-03T17:27:28.3560544Z (EngineCore_DP0 pid=46982) INFO 11-03 17:27:28 [kv_cache_utils.py:1199] GPU KV cache size: 1,214,464 tokens 2025-11-03T17:27:28.3561734Z (EngineCore_DP0 pid=46982) INFO 11-03 17:27:28 [kv_cache_utils.py:1204] Maximum concurrency for 2,048 tokens per request: 593.00x 2025-11-03T17:27:28.3562578Z (EngineCore_DP0 pid=46982) INFO 11-03 17:27:28 [kv_cache_utils.py:1199] GPU KV cache size: 1,214,464 tokens 2025-11-03T17:27:28.3563409Z (EngineCore_DP0 pid=46982) INFO 11-03 17:27:28 [kv_cache_utils.py:1204] Maximum concurrency for 2,048 tokens per request: 593.00x 2025-11-03T17:27:28.4773970Z (EngineCore_DP0 pid=46982) INFO 11-03 17:27:28 [core.py:240] init engine (profile, create kv cache, warmup model) took 2.29 seconds 2025-11-03T17:27:28.9968398Z (EngineCore_DP0 pid=46982) INFO 11-03 17:27:28 [vllm.py:354] Cudagraph is disabled under eager mode 2025-11-03T17:27:29.1813904Z (APIServer pid=46860) INFO 11-03 17:27:29 [loggers.py:172] Engine 000: vllm cache_config_info with initialization after num_gpu_blocks is: 75904 2025-11-03T17:27:29.1984889Z (EngineCore_DP0 pid=46982) INFO 11-03 17:27:29 [gc_utils.py:40] GC Debug Config. enabled:False,top_objects:-1 2025-11-03T17:27:29.2195538Z (APIServer pid=46860) INFO 11-03 17:27:29 [api_server.py:1628] Supported tasks: ['generate'] 2025-11-03T17:27:29.2481165Z (APIServer pid=46860) WARNING 11-03 17:27:29 [model.py:1581] Default sampling parameters have been overridden by the model's Hugging Face generation config recommended from the model creator. If this is not intended, please relaunch vLLM instance with `--generation-config vllm`. 2025-11-03T17:27:29.2484156Z (APIServer pid=46860) INFO 11-03 17:27:29 [serving_responses.py:157] Using default chat sampling params from model: {'temperature': 0.6, 'top_p': 0.9} 2025-11-03T17:27:29.2697355Z (APIServer pid=46860) INFO 11-03 17:27:29 [serving_chat.py:159] Using default chat sampling params from model: {'temperature': 0.6, 'top_p': 0.9} 2025-11-03T17:27:29.2910943Z (APIServer pid=46860) INFO 11-03 17:27:29 [serving_completion.py:67] Using default completion sampling params from model: {'temperature': 0.6, 'top_p': 0.9} 2025-11-03T17:27:29.2913139Z (APIServer pid=46860) INFO 11-03 17:27:29 [api_server.py:1939] Starting vLLM API server 0 on http://0.0.0.0:56525 2025-11-03T17:27:29.2915279Z (APIServer pid=46860) INFO 11-03 17:27:29 [launcher.py:38] Available routes are: 2025-11-03T17:27:29.2916969Z (APIServer pid=46860) INFO 11-03 17:27:29 [launcher.py:46] Route: /openapi.json, Methods: GET, HEAD 2025-11-03T17:27:29.2918795Z (APIServer pid=46860) INFO 11-03 17:27:29 [launcher.py:46] Route: /docs, Methods: GET, HEAD 2025-11-03T17:27:29.2920194Z (APIServer pid=46860) INFO 11-03 17:27:29 [launcher.py:46] Route: /docs/oauth2-redirect, Methods: GET, HEAD 2025-11-03T17:27:29.2921109Z (APIServer pid=46860) INFO 11-03 17:27:29 [launcher.py:46] Route: /redoc, Methods: GET, HEAD 2025-11-03T17:27:29.2921632Z (APIServer pid=46860) INFO 11-03 17:27:29 [launcher.py:46] Route: /health, Methods: GET 2025-11-03T17:27:29.2922127Z (APIServer pid=46860) INFO 11-03 17:27:29 [launcher.py:46] Route: /load, Methods: GET 2025-11-03T17:27:29.2922624Z (APIServer pid=46860) INFO 11-03 17:27:29 [launcher.py:46] Route: /ping, Methods: POST 2025-11-03T17:27:29.2923116Z (APIServer pid=46860) INFO 11-03 17:27:29 [launcher.py:46] Route: /ping, Methods: GET 2025-11-03T17:27:29.2923640Z (APIServer pid=46860) INFO 11-03 17:27:29 [launcher.py:46] Route: /tokenize, Methods: POST 2025-11-03T17:27:29.2924160Z (APIServer pid=46860) INFO 11-03 17:27:29 [launcher.py:46] Route: /detokenize, Methods: POST 2025-11-03T17:27:29.2924863Z (APIServer pid=46860) INFO 11-03 17:27:29 [launcher.py:46] Route: /v1/models, Methods: GET 2025-11-03T17:27:29.2925376Z (APIServer pid=46860) INFO 11-03 17:27:29 [launcher.py:46] Route: /version, Methods: GET 2025-11-03T17:27:29.2925896Z (APIServer pid=46860) INFO 11-03 17:27:29 [launcher.py:46] Route: /v1/responses, Methods: POST 2025-11-03T17:27:29.2926551Z (APIServer pid=46860) INFO 11-03 17:27:29 [launcher.py:46] Route: /v1/responses/{response_id}, Methods: GET 2025-11-03T17:27:29.2927165Z (APIServer pid=46860) INFO 11-03 17:27:29 [launcher.py:46] Route: /v1/responses/{response_id}/cancel, Methods: POST 2025-11-03T17:27:29.2927761Z (APIServer pid=46860) INFO 11-03 17:27:29 [launcher.py:46] Route: /v1/chat/completions, Methods: POST 2025-11-03T17:27:29.2928325Z (APIServer pid=46860) INFO 11-03 17:27:29 [launcher.py:46] Route: /v1/completions, Methods: POST 2025-11-03T17:27:29.2928964Z (APIServer pid=46860) INFO 11-03 17:27:29 [launcher.py:46] Route: /v1/embeddings, Methods: POST 2025-11-03T17:27:29.2929487Z (APIServer pid=46860) INFO 11-03 17:27:29 [launcher.py:46] Route: /pooling, Methods: POST 2025-11-03T17:27:29.2930003Z (APIServer pid=46860) INFO 11-03 17:27:29 [launcher.py:46] Route: /classify, Methods: POST 2025-11-03T17:27:29.2930515Z (APIServer pid=46860) INFO 11-03 17:27:29 [launcher.py:46] Route: /score, Methods: POST 2025-11-03T17:27:29.2931011Z (APIServer pid=46860) INFO 11-03 17:27:29 [launcher.py:46] Route: /v1/score, Methods: POST 2025-11-03T17:27:29.2931568Z (APIServer pid=46860) INFO 11-03 17:27:29 [launcher.py:46] Route: /v1/audio/transcriptions, Methods: POST 2025-11-03T17:27:29.2932151Z (APIServer pid=46860) INFO 11-03 17:27:29 [launcher.py:46] Route: /v1/audio/translations, Methods: POST 2025-11-03T17:27:29.2932697Z (APIServer pid=46860) INFO 11-03 17:27:29 [launcher.py:46] Route: /rerank, Methods: POST 2025-11-03T17:27:29.2933227Z (APIServer pid=46860) INFO 11-03 17:27:29 [launcher.py:46] Route: /v1/rerank, Methods: POST 2025-11-03T17:27:29.2933735Z (APIServer pid=46860) INFO 11-03 17:27:29 [launcher.py:46] Route: /v2/rerank, Methods: POST 2025-11-03T17:27:29.2934276Z (APIServer pid=46860) INFO 11-03 17:27:29 [launcher.py:46] Route: /scale_elastic_ep, Methods: POST 2025-11-03T17:27:29.2934839Z (APIServer pid=46860) INFO 11-03 17:27:29 [launcher.py:46] Route: /is_scaling_elastic_ep, Methods: POST 2025-11-03T17:27:29.2935388Z (APIServer pid=46860) INFO 11-03 17:27:29 [launcher.py:46] Route: /invocations, Methods: POST 2025-11-03T17:27:29.2935901Z (APIServer pid=46860) INFO 11-03 17:27:29 [launcher.py:46] Route: /metrics, Methods: GET 2025-11-03T17:27:29.3069000Z (APIServer pid=46860) INFO: Started server process [46860] 2025-11-03T17:27:29.3070254Z (APIServer pid=46860) INFO: Waiting for application startup. 2025-11-03T17:27:29.5093575Z (APIServer pid=46860) INFO: Application startup complete. 2025-11-03T17:27:29.5602296Z (APIServer pid=46860) INFO: 127.0.0.1:59104 - "GET /health HTTP/1.1" 200 OK 2025-11-03T17:27:29.5694434Z (APIServer pid=46860) INFO: 127.0.0.1:59118 - "GET /v1/models HTTP/1.1" 200 OK 2025-11-03T17:27:29.7435739Z (APIServer pid=46860) INFO: 127.0.0.1:59118 - "POST /v1/completions HTTP/1.1" 200 OK 2025-11-03T17:27:29.7999875Z (APIServer pid=46860) INFO: 127.0.0.1:59118 - "POST /v1/completions HTTP/1.1" 200 OK 2025-11-03T17:27:29.8974459Z (APIServer pid=46860) INFO: 127.0.0.1:59118 - "POST /v1/completions HTTP/1.1" 200 OK 2025-11-03T17:27:29.9822969Z (APIServer pid=46860) INFO: 127.0.0.1:59118 - "POST /v1/completions HTTP/1.1" 200 OK 2025-11-03T17:27:30.0437045Z (APIServer pid=46860) INFO: 127.0.0.1:59118 - "POST /v1/completions HTTP/1.1" 200 OK 2025-11-03T17:27:30.0511062Z (APIServer pid=46860) INFO: 127.0.0.1:59118 - "POST /v1/completions HTTP/1.1" 200 OK 2025-11-03T17:27:30.1174782Z (APIServer pid=46860) INFO 11-03 17:27:30 [launcher.py:110] Shutting down FastAPI HTTP server. 2025-11-03T17:27:30.1178478Z (Worker_TP0 pid=47101) INFO 11-03 17:27:30 [multiproc_executor.py:583] Parent process exited, terminating worker 2025-11-03T17:27:30.1180070Z (Worker_TP0 pid=47101) INFO 11-03 17:27:30 [multiproc_executor.py:624] WorkerProc shutting down. 2025-11-03T17:27:30.1181566Z (Worker_TP1 pid=47102) INFO 11-03 17:27:30 [multiproc_executor.py:583] Parent process exited, terminating worker 2025-11-03T17:27:30.1183328Z (Worker_TP1 pid=47102) INFO 11-03 17:27:30 [multiproc_executor.py:624] WorkerProc shutting down. 2025-11-03T17:27:32.3325091Z (APIServer pid=46860) INFO: Shutting down 2025-11-03T17:27:32.4329840Z (APIServer pid=46860) INFO: Waiting for application shutdown. 2025-11-03T17:27:32.4331577Z (APIServer pid=46860) INFO: Application shutdown complete. 2025-11-03T17:27:33.0478664Z PASSED 2025-11-03T17:27:33.0624267Z distributed/test_sequence_parallel.py::test_tp_sp_generation[meta-llama/Llama-3.2-1B-Instruct-parallel_setup13-ray-auto-test_options13] Fork a new process to run a test 47305 2025-11-03T17:27:33.0638866Z Fork a new process to run a test 0 2025-11-03T17:27:35.3068492Z INFO 11-03 17:27:35 [model.py:653] Resolved architecture: LlamaForCausalLM 2025-11-03T17:27:35.3069009Z `torch_dtype` is deprecated! Use `dtype` instead! 2025-11-03T17:27:35.3069716Z WARNING 11-03 17:27:35 [model.py:1942] Casting torch.bfloat16 to torch.float16. 2025-11-03T17:27:35.3070225Z INFO 11-03 17:27:35 [model.py:1714] Using max model len 2048 2025-11-03T17:27:35.3074364Z Launching RemoteOpenAIServer with: vllm serve meta-llama/Llama-3.2-1B-Instruct --dtype float16 --max-model-len 2048 --max-num-seqs 8 --enforce-eager --tokenizer-mode auto --tensor-parallel-size 2 --pipeline-parallel-size 2 --distributed-executor-backend ray --compilation_config {"level": 3, "custom_ops": ["+rms_norm"], "compile_sizes": [4, 8], "pass_config": {"enable_sequence_parallelism": true, "enable_fusion": false, "enable_noop": true}} --load-format dummy --port 43109 --seed 0 2025-11-03T17:27:35.8242496Z /opt/conda/envs/py_3.12/lib/python3.12/site-packages/torch/cuda/__init__.py:65: FutureWarning: The pynvml package is deprecated. Please install nvidia-ml-py instead. If you did not install pynvml directly, please report this to the maintainers of the package that installed pynvml for you. 2025-11-03T17:27:35.8244005Z import pynvml # type: ignore[import] 2025-11-03T17:27:38.0982538Z INFO 11-03 17:27:38 [__init__.py:224] Automatically detected platform cuda. 2025-11-03T17:27:41.7872964Z (APIServer pid=47354) INFO 11-03 17:27:41 [api_server.py:1870] vLLM API server version 0.11.0rc2.dev389+ge51928192.d20251103 2025-11-03T17:27:41.7899702Z (APIServer pid=47354) INFO 11-03 17:27:41 [utils.py:239] non-default args: {'model_tag': 'meta-llama/Llama-3.2-1B-Instruct', 'port': 43109, 'model': 'meta-llama/Llama-3.2-1B-Instruct', 'dtype': 'float16', 'seed': 0, 'max_model_len': 2048, 'enforce_eager': True, 'load_format': 'dummy', 'distributed_executor_backend': 'ray', 'pipeline_parallel_size': 2, 'tensor_parallel_size': 2, 'max_num_seqs': 8, 'compilation_config': {'level': 3, 'debug_dump_path': None, 'cache_dir': '', 'backend': '', 'custom_ops': ['+rms_norm'], 'splitting_ops': None, 'use_inductor': True, 'compile_sizes': [4, 8], 'inductor_compile_config': {'enable_auto_functionalized_v2': False}, 'inductor_passes': {}, 'cudagraph_mode': None, 'use_cudagraph': True, 'cudagraph_num_of_warmups': 0, 'cudagraph_capture_sizes': None, 'cudagraph_copy_inputs': False, 'full_cuda_graph': False, 'use_inductor_graph_partition': False, 'pass_config': {'enable_noop': True, 'enable_sequence_parallelism': True}, 'max_capture_size': None, 'local_cache_dir': None}} 2025-11-03T17:27:41.9556814Z (APIServer pid=47354) INFO 11-03 17:27:41 [model.py:653] Resolved architecture: LlamaForCausalLM 2025-11-03T17:27:41.9557534Z (APIServer pid=47354) `torch_dtype` is deprecated! Use `dtype` instead! 2025-11-03T17:27:41.9558225Z (APIServer pid=47354) WARNING 11-03 17:27:41 [model.py:1942] Casting torch.bfloat16 to torch.float16. 2025-11-03T17:27:41.9558913Z (APIServer pid=47354) INFO 11-03 17:27:41 [model.py:1714] Using max model len 2048 2025-11-03T17:27:42.1184458Z (APIServer pid=47354) INFO 11-03 17:27:42 [scheduler.py:225] Chunked prefill is enabled with max_num_batched_tokens=2048. 2025-11-03T17:27:42.1189851Z (APIServer pid=47354) INFO 11-03 17:27:42 [vllm.py:354] Cudagraph is disabled under eager mode 2025-11-03T17:27:43.1176487Z /opt/conda/envs/py_3.12/lib/python3.12/site-packages/torch/cuda/__init__.py:65: FutureWarning: The pynvml package is deprecated. Please install nvidia-ml-py instead. If you did not install pynvml directly, please report this to the maintainers of the package that installed pynvml for you. 2025-11-03T17:27:43.1177738Z import pynvml # type: ignore[import] 2025-11-03T17:27:45.4618870Z INFO 11-03 17:27:45 [__init__.py:224] Automatically detected platform cuda. 2025-11-03T17:27:49.3212611Z (EngineCore_DP0 pid=47476) INFO 11-03 17:27:49 [core.py:727] Waiting for init message from front-end. 2025-11-03T17:27:49.3480026Z (EngineCore_DP0 pid=47476) INFO 11-03 17:27:49 [core.py:94] Initializing a V1 LLM engine (v0.11.0rc2.dev389+ge51928192.d20251103) with config: model='meta-llama/Llama-3.2-1B-Instruct', speculative_config=None, tokenizer='meta-llama/Llama-3.2-1B-Instruct', skip_tokenizer_init=False, tokenizer_mode=auto, revision=None, tokenizer_revision=None, trust_remote_code=False, dtype=torch.float16, max_seq_len=2048, download_dir=None, load_format=dummy, tensor_parallel_size=2, pipeline_parallel_size=2, data_parallel_size=1, disable_custom_all_reduce=False, quantization=None, enforce_eager=True, kv_cache_dtype=auto, device_config=cuda, structured_outputs_config=StructuredOutputsConfig(backend='auto', disable_fallback=False, disable_any_whitespace=False, disable_additional_properties=False, reasoning_parser=''), observability_config=ObservabilityConfig(show_hidden_metrics_for_version=None, otlp_traces_endpoint=None, collect_detailed_traces=None), seed=0, served_model_name=meta-llama/Llama-3.2-1B-Instruct, enable_prefix_caching=True, chunked_prefill_enabled=True, pooler_config=None, compilation_config={'level': 3, 'debug_dump_path': None, 'cache_dir': '', 'backend': '', 'custom_ops': ['+rms_norm', '+rms_norm'], 'splitting_ops': ['vllm.unified_attention', 'vllm.unified_attention_with_output', 'vllm.unified_mla_attention', 'vllm.unified_mla_attention_with_output', 'vllm.mamba_mixer2', 'vllm.mamba_mixer', 'vllm.short_conv', 'vllm.linear_attention', 'vllm.plamo2_mamba_mixer', 'vllm.gdn_attention', 'vllm.sparse_attn_indexer'], 'use_inductor': True, 'compile_sizes': [8, 4], 'inductor_compile_config': {'enable_auto_functionalized_v2': False}, 'inductor_passes': {}, 'cudagraph_mode': , 'use_cudagraph': False, 'cudagraph_num_of_warmups': 0, 'cudagraph_capture_sizes': [], 'cudagraph_copy_inputs': False, 'full_cuda_graph': False, 'use_inductor_graph_partition': False, 'pass_config': {'enable_noop': True, 'enable_sequence_parallelism': True}, 'max_capture_size': 0, 'local_cache_dir': None} 2025-11-03T17:27:51.2545068Z (EngineCore_DP0 pid=47476) 2025-11-03 17:27:51,253 INFO worker.py:1918 -- Started a local Ray instance. View the dashboard at http://127.0.0.1:8265  2025-11-03T17:27:52.1134975Z (EngineCore_DP0 pid=47476) INFO 11-03 17:27:52 [ray_utils.py:373] No current placement group found. Creating a new placement group. 2025-11-03T17:27:52.2897898Z (EngineCore_DP0 pid=47476) INFO 11-03 17:27:52 [ray_distributed_executor.py:179] use_ray_spmd_worker: True 2025-11-03T17:27:52.9435590Z (EngineCore_DP0 pid=47476) (pid=48368) /opt/conda/envs/py_3.12/lib/python3.12/site-packages/torch/cuda/__init__.py:65: FutureWarning: The pynvml package is deprecated. Please install nvidia-ml-py instead. If you did not install pynvml directly, please report this to the maintainers of the package that installed pynvml for you. 2025-11-03T17:27:52.9438359Z (EngineCore_DP0 pid=47476) (pid=48368) import pynvml # type: ignore[import] 2025-11-03T17:27:55.2922657Z (EngineCore_DP0 pid=47476) (pid=48368) INFO 11-03 17:27:55 [__init__.py:224] Automatically detected platform cuda. 2025-11-03T17:27:56.8158106Z (EngineCore_DP0 pid=47476) INFO 11-03 17:27:56 [ray_env.py:67] RAY_NON_CARRY_OVER_ENV_VARS from config: set() 2025-11-03T17:27:56.8159603Z (EngineCore_DP0 pid=47476) INFO 11-03 17:27:56 [ray_env.py:70] Copying the following environment variables to workers: ['MAX_JOBS', 'HUGGING_FACE_HUB_TOKEN', 'VLLM_USE_V1', 'VLLM_WORKER_MULTIPROC_METHOD', 'HF_TOKEN', 'LD_LIBRARY_PATH', 'VLLM_USE_RAY_COMPILED_DAG', 'VLLM_USE_RAY_SPMD_WORKER'] 2025-11-03T17:27:56.8161227Z (EngineCore_DP0 pid=47476) INFO 11-03 17:27:56 [ray_env.py:75] If certain env vars should NOT be copied, add them to /var/lib/jenkins/.config/vllm/ray_non_carry_over_env_vars.json file 2025-11-03T17:28:00.1912188Z (EngineCore_DP0 pid=47476) (RayWorkerWrapper pid=48368) WARNING 11-03 17:28:00 [worker_base.py:307] Missing `shared_worker_lock` argument from executor. This argument is needed for mm_processor_cache_type='shm'. 2025-11-03T17:28:00.9090588Z (EngineCore_DP0 pid=47476) (RayWorkerWrapper pid=48368) [Gloo] Rank 0 is connected to 3 peer ranks. Expected number of connected peer ranks is : 3 2025-11-03T17:28:00.9092821Z (EngineCore_DP0 pid=47476) (pid=48366) INFO 11-03 17:27:55 [__init__.py:224] Automatically detected platform cuda. [repeated 3x across cluster] (Ray deduplicates logs by default. Set RAY_DEDUP_LOGS=0 to disable log deduplication, or see https://docs.ray.io/en/master/ray-observability/user-guides/configure-logging.html#log-deduplication for more options.) 2025-11-03T17:28:01.0139889Z (EngineCore_DP0 pid=47476) (RayWorkerWrapper pid=48368) INFO 11-03 17:28:00 [__init__.py:1146] Found nccl from library libnccl.so.2 2025-11-03T17:28:01.0140930Z (EngineCore_DP0 pid=47476) (RayWorkerWrapper pid=48368) INFO 11-03 17:28:00 [pynccl.py:109] vLLM is using nccl==2.27.5 2025-11-03T17:28:01.3240116Z (EngineCore_DP0 pid=47476) (RayWorkerWrapper pid=48368) WARNING 11-03 17:28:01 [symm_mem.py:60] SymmMemCommunicator: Device capability 8.9 not supported, communicator is not available. 2025-11-03T17:28:01.3241456Z (EngineCore_DP0 pid=47476) (RayWorkerWrapper pid=48368) INFO 11-03 17:28:01 [custom_all_reduce.py:37] Skipping P2P check and trusting the driver's P2P report. 2025-11-03T17:28:01.3243072Z (EngineCore_DP0 pid=47476) (RayWorkerWrapper pid=48368) WARNING 11-03 17:28:01 [custom_all_reduce.py:162] Custom allreduce is disabled because your platform lacks GPU P2P capability or P2P test failed. To silence this warning, specify disable_custom_all_reduce=True explicitly. 2025-11-03T17:28:01.3245358Z (EngineCore_DP0 pid=47476) (RayWorkerWrapper pid=48368) INFO 11-03 17:28:01 [shm_broadcast.py:302] vLLM message queue communication handle: Handle(local_reader_ranks=[1], buffer_handle=(1, 4194304, 6, 'psm_dd4db156'), local_subscribe_addr='ipc:///tmp/f44aeee3-8831-437d-9211-a0f325a3d6ba', remote_subscribe_addr=None, remote_addr_ipv6=False) 2025-11-03T17:28:01.4350667Z (EngineCore_DP0 pid=47476) (RayWorkerWrapper pid=48368) INFO 11-03 17:28:01 [parallel_state.py:1231] rank 0 in world size 4 is assigned as DP rank 0, PP rank 0, TP rank 0, EP rank 0 2025-11-03T17:28:01.6430020Z (EngineCore_DP0 pid=47476) (RayWorkerWrapper pid=48368) INFO 11-03 17:28:01 [topk_topp_sampler.py:59] Using FlashInfer for top-p & top-k sampling. 2025-11-03T17:28:01.7484238Z (EngineCore_DP0 pid=47476) (RayWorkerWrapper pid=48368) INFO 11-03 17:28:01 [gpu_model_runner.py:2840] Starting to load model meta-llama/Llama-3.2-1B-Instruct... 2025-11-03T17:28:01.9563990Z (EngineCore_DP0 pid=47476) (RayWorkerWrapper pid=48368) INFO 11-03 17:28:01 [gpu_model_runner.py:2870] Loading model from scratch... 2025-11-03T17:28:02.0618961Z (EngineCore_DP0 pid=47476) (RayWorkerWrapper pid=48368) INFO 11-03 17:28:02 [cuda.py:404] Using Flash Attention backend on V1 engine. 2025-11-03T17:28:02.5759351Z (EngineCore_DP0 pid=47476) (RayWorkerWrapper pid=48363) INFO 11-03 17:28:02 [gpu_model_runner.py:2902] Model loading took 0.7135 GiB and 0.154608 seconds 2025-11-03T17:28:05.0263817Z (EngineCore_DP0 pid=47476) (RayWorkerWrapper pid=48368) INFO 11-03 17:28:05 [backends.py:594] Using cache directory: /var/lib/jenkins/.cache/vllm/torch_compile_cache/3ff2b0020c/rank_0_0/backbone for vLLM's torch.compile 2025-11-03T17:28:05.0265308Z (EngineCore_DP0 pid=47476) (RayWorkerWrapper pid=48368) INFO 11-03 17:28:05 [backends.py:608] Dynamo bytecode transform time: 2.13 s 2025-11-03T17:28:05.6446425Z (EngineCore_DP0 pid=47476) (RayWorkerWrapper pid=48368) INFO 11-03 17:28:05 [backends.py:179] Directly load the compiled graph(s) for dynamic shape from the cache, took 0.364 s 2025-11-03T17:28:05.6448537Z (EngineCore_DP0 pid=47476) (RayWorkerWrapper pid=48366) WARNING 11-03 17:28:00 [worker_base.py:307] Missing `shared_worker_lock` argument from executor. This argument is needed for mm_processor_cache_type='shm'. [repeated 3x across cluster] 2025-11-03T17:28:06.5657887Z (EngineCore_DP0 pid=47476) (RayWorkerWrapper pid=48363) INFO 11-03 17:28:06 [gpu_worker.py:314] Available KV cache memory: 19.02 GiB 2025-11-03T17:28:06.5659222Z (EngineCore_DP0 pid=47476) (RayWorkerWrapper pid=48365) [Gloo] Rank 0 is connected to 1 peer ranks. Expected number of connected peer ranks is : 1 [repeated 23x across cluster] 2025-11-03T17:28:06.5660795Z (EngineCore_DP0 pid=47476) (RayWorkerWrapper pid=48366) INFO 11-03 17:28:01 [__init__.py:1146] Found nccl from library libnccl.so.2 [repeated 11x across cluster] 2025-11-03T17:28:06.5661960Z (EngineCore_DP0 pid=47476) (RayWorkerWrapper pid=48366) INFO 11-03 17:28:01 [pynccl.py:109] vLLM is using nccl==2.27.5 [repeated 11x across cluster] 2025-11-03T17:28:06.5663484Z (EngineCore_DP0 pid=47476) (RayWorkerWrapper pid=48366) WARNING 11-03 17:28:01 [symm_mem.py:60] SymmMemCommunicator: Device capability 8.9 not supported, communicator is not available. [repeated 3x across cluster] 2025-11-03T17:28:06.5664703Z (EngineCore_DP0 pid=47476) (RayWorkerWrapper pid=48366) INFO 11-03 17:28:01 [custom_all_reduce.py:37] Skipping P2P check and trusting the driver's P2P report. [repeated 3x across cluster] 2025-11-03T17:28:06.5666201Z (EngineCore_DP0 pid=47476) (RayWorkerWrapper pid=48366) WARNING 11-03 17:28:01 [custom_all_reduce.py:162] Custom allreduce is disabled because your platform lacks GPU P2P capability or P2P test failed. To silence this warning, specify disable_custom_all_reduce=True explicitly. [repeated 3x across cluster] 2025-11-03T17:28:06.5668253Z (EngineCore_DP0 pid=47476) (RayWorkerWrapper pid=48365) INFO 11-03 17:28:01 [shm_broadcast.py:302] vLLM message queue communication handle: Handle(local_reader_ranks=[1], buffer_handle=(1, 4194304, 6, 'psm_75156942'), local_subscribe_addr='ipc:///tmp/90e5105d-76e2-40f5-b554-3a12cedab7f7', remote_subscribe_addr=None, remote_addr_ipv6=False) 2025-11-03T17:28:06.5669782Z (EngineCore_DP0 pid=47476) (RayWorkerWrapper pid=48366) INFO 11-03 17:28:01 [parallel_state.py:1231] rank 3 in world size 4 is assigned as DP rank 0, PP rank 1, TP rank 1, EP rank 1 [repeated 3x across cluster] 2025-11-03T17:28:06.6702905Z (EngineCore_DP0 pid=47476) (RayWorkerWrapper pid=48366) INFO 11-03 17:28:01 [topk_topp_sampler.py:59] Using FlashInfer for top-p & top-k sampling. [repeated 3x across cluster] 2025-11-03T17:28:06.9221401Z (EngineCore_DP0 pid=47476) INFO 11-03 17:28:06 [kv_cache_utils.py:1199] GPU KV cache size: 2,489,088 tokens 2025-11-03T17:28:06.9222199Z (EngineCore_DP0 pid=47476) INFO 11-03 17:28:06 [kv_cache_utils.py:1204] Maximum concurrency for 2,048 tokens per request: 1215.38x 2025-11-03T17:28:06.9222901Z (EngineCore_DP0 pid=47476) INFO 11-03 17:28:06 [kv_cache_utils.py:1199] GPU KV cache size: 2,489,088 tokens 2025-11-03T17:28:06.9223592Z (EngineCore_DP0 pid=47476) INFO 11-03 17:28:06 [kv_cache_utils.py:1204] Maximum concurrency for 2,048 tokens per request: 1215.38x 2025-11-03T17:28:06.9224265Z (EngineCore_DP0 pid=47476) INFO 11-03 17:28:06 [kv_cache_utils.py:1199] GPU KV cache size: 2,489,088 tokens 2025-11-03T17:28:06.9224911Z (EngineCore_DP0 pid=47476) INFO 11-03 17:28:06 [kv_cache_utils.py:1204] Maximum concurrency for 2,048 tokens per request: 1215.38x 2025-11-03T17:28:06.9225546Z (EngineCore_DP0 pid=47476) INFO 11-03 17:28:06 [kv_cache_utils.py:1199] GPU KV cache size: 2,489,088 tokens 2025-11-03T17:28:06.9226191Z (EngineCore_DP0 pid=47476) INFO 11-03 17:28:06 [kv_cache_utils.py:1204] Maximum concurrency for 2,048 tokens per request: 1215.38x 2025-11-03T17:28:06.9794379Z (EngineCore_DP0 pid=47476) (RayWorkerWrapper pid=48368) INFO 11-03 17:28:06 [gpu_worker.py:351] Compile and warming up model for size 8 2025-11-03T17:28:06.9796264Z (EngineCore_DP0 pid=47476) (RayWorkerWrapper pid=48366) INFO 11-03 17:28:01 [gpu_model_runner.py:2840] Starting to load model meta-llama/Llama-3.2-1B-Instruct... [repeated 3x across cluster] 2025-11-03T17:28:06.9798168Z (EngineCore_DP0 pid=47476) (RayWorkerWrapper pid=48366) INFO 11-03 17:28:01 [gpu_model_runner.py:2870] Loading model from scratch... [repeated 3x across cluster] 2025-11-03T17:28:08.3068759Z (EngineCore_DP0 pid=47476) (RayWorkerWrapper pid=48368) INFO 11-03 17:28:08 [backends.py:185] Directly load the compiled graph(s) for shape 8 from the cache, took 1.334 s 2025-11-03T17:28:08.3070252Z (EngineCore_DP0 pid=47476) (RayWorkerWrapper pid=48366) INFO 11-03 17:28:02 [cuda.py:404] Using Flash Attention backend on V1 engine. [repeated 3x across cluster] 2025-11-03T17:28:08.3071853Z (EngineCore_DP0 pid=47476) (RayWorkerWrapper pid=48366) INFO 11-03 17:28:02 [gpu_model_runner.py:2902] Model loading took 0.7135 GiB and 0.157618 seconds [repeated 3x across cluster] 2025-11-03T17:28:08.9233483Z (EngineCore_DP0 pid=47476) (RayWorkerWrapper pid=48365) INFO 11-03 17:28:08 [monitor.py:33] torch.compile takes 2.12 s in total 2025-11-03T17:28:09.2473050Z (EngineCore_DP0 pid=47476) INFO 11-03 17:28:09 [core.py:240] init engine (profile, create kv cache, warmup model) took 6.62 seconds 2025-11-03T17:28:10.0607644Z (EngineCore_DP0 pid=47476) INFO 11-03 17:28:10 [core.py:178] Batch queue is enabled with size 2 2025-11-03T17:28:10.0639440Z (EngineCore_DP0 pid=47476) INFO 11-03 17:28:10 [vllm.py:354] Cudagraph is disabled under eager mode 2025-11-03T17:28:10.2565242Z (APIServer pid=47354) INFO 11-03 17:28:10 [loggers.py:172] Engine 000: vllm cache_config_info with initialization after num_gpu_blocks is: 155568 2025-11-03T17:28:10.2889138Z (EngineCore_DP0 pid=47476) INFO 11-03 17:28:10 [gc_utils.py:40] GC Debug Config. enabled:False,top_objects:-1 2025-11-03T17:28:10.3139270Z (APIServer pid=47354) INFO 11-03 17:28:10 [api_server.py:1628] Supported tasks: ['generate'] 2025-11-03T17:28:10.3400272Z (APIServer pid=47354) WARNING 11-03 17:28:10 [model.py:1581] Default sampling parameters have been overridden by the model's Hugging Face generation config recommended from the model creator. If this is not intended, please relaunch vLLM instance with `--generation-config vllm`. 2025-11-03T17:28:10.3401606Z (APIServer pid=47354) INFO 11-03 17:28:10 [serving_responses.py:157] Using default chat sampling params from model: {'temperature': 0.6, 'top_p': 0.9} 2025-11-03T17:28:10.3643366Z (APIServer pid=47354) INFO 11-03 17:28:10 [serving_chat.py:159] Using default chat sampling params from model: {'temperature': 0.6, 'top_p': 0.9} 2025-11-03T17:28:10.3894376Z (APIServer pid=47354) INFO 11-03 17:28:10 [serving_completion.py:67] Using default completion sampling params from model: {'temperature': 0.6, 'top_p': 0.9} 2025-11-03T17:28:10.3895450Z (APIServer pid=47354) INFO 11-03 17:28:10 [api_server.py:1939] Starting vLLM API server 0 on http://0.0.0.0:43109 2025-11-03T17:28:10.3896205Z (APIServer pid=47354) INFO 11-03 17:28:10 [launcher.py:38] Available routes are: 2025-11-03T17:28:10.3896903Z (APIServer pid=47354) INFO 11-03 17:28:10 [launcher.py:46] Route: /openapi.json, Methods: HEAD, GET 2025-11-03T17:28:10.3897606Z (APIServer pid=47354) INFO 11-03 17:28:10 [launcher.py:46] Route: /docs, Methods: HEAD, GET 2025-11-03T17:28:10.3898241Z (APIServer pid=47354) INFO 11-03 17:28:10 [launcher.py:46] Route: /docs/oauth2-redirect, Methods: HEAD, GET 2025-11-03T17:28:10.3898802Z (APIServer pid=47354) INFO 11-03 17:28:10 [launcher.py:46] Route: /redoc, Methods: HEAD, GET 2025-11-03T17:28:10.3899321Z (APIServer pid=47354) INFO 11-03 17:28:10 [launcher.py:46] Route: /health, Methods: GET 2025-11-03T17:28:10.3899817Z (APIServer pid=47354) INFO 11-03 17:28:10 [launcher.py:46] Route: /load, Methods: GET 2025-11-03T17:28:10.3900313Z (APIServer pid=47354) INFO 11-03 17:28:10 [launcher.py:46] Route: /ping, Methods: POST 2025-11-03T17:28:10.3900801Z (APIServer pid=47354) INFO 11-03 17:28:10 [launcher.py:46] Route: /ping, Methods: GET 2025-11-03T17:28:10.3901315Z (APIServer pid=47354) INFO 11-03 17:28:10 [launcher.py:46] Route: /tokenize, Methods: POST 2025-11-03T17:28:10.3901836Z (APIServer pid=47354) INFO 11-03 17:28:10 [launcher.py:46] Route: /detokenize, Methods: POST 2025-11-03T17:28:10.3902566Z (APIServer pid=47354) INFO 11-03 17:28:10 [launcher.py:46] Route: /v1/models, Methods: GET 2025-11-03T17:28:10.3903081Z (APIServer pid=47354) INFO 11-03 17:28:10 [launcher.py:46] Route: /version, Methods: GET 2025-11-03T17:28:10.3903597Z (APIServer pid=47354) INFO 11-03 17:28:10 [launcher.py:46] Route: /v1/responses, Methods: POST 2025-11-03T17:28:10.3904173Z (APIServer pid=47354) INFO 11-03 17:28:10 [launcher.py:46] Route: /v1/responses/{response_id}, Methods: GET 2025-11-03T17:28:10.3904792Z (APIServer pid=47354) INFO 11-03 17:28:10 [launcher.py:46] Route: /v1/responses/{response_id}/cancel, Methods: POST 2025-11-03T17:28:10.3905390Z (APIServer pid=47354) INFO 11-03 17:28:10 [launcher.py:46] Route: /v1/chat/completions, Methods: POST 2025-11-03T17:28:10.3905946Z (APIServer pid=47354) INFO 11-03 17:28:10 [launcher.py:46] Route: /v1/completions, Methods: POST 2025-11-03T17:28:10.3906482Z (APIServer pid=47354) INFO 11-03 17:28:10 [launcher.py:46] Route: /v1/embeddings, Methods: POST 2025-11-03T17:28:10.3907008Z (APIServer pid=47354) INFO 11-03 17:28:10 [launcher.py:46] Route: /pooling, Methods: POST 2025-11-03T17:28:10.3907706Z (APIServer pid=47354) INFO 11-03 17:28:10 [launcher.py:46] Route: /classify, Methods: POST 2025-11-03T17:28:10.3908233Z (APIServer pid=47354) INFO 11-03 17:28:10 [launcher.py:46] Route: /score, Methods: POST 2025-11-03T17:28:10.3908819Z (APIServer pid=47354) INFO 11-03 17:28:10 [launcher.py:46] Route: /v1/score, Methods: POST 2025-11-03T17:28:10.3909369Z (APIServer pid=47354) INFO 11-03 17:28:10 [launcher.py:46] Route: /v1/audio/transcriptions, Methods: POST 2025-11-03T17:28:10.3910328Z (APIServer pid=47354) INFO 11-03 17:28:10 [launcher.py:46] Route: /v1/audio/translations, Methods: POST 2025-11-03T17:28:10.3910876Z (APIServer pid=47354) INFO 11-03 17:28:10 [launcher.py:46] Route: /rerank, Methods: POST 2025-11-03T17:28:10.3911394Z (APIServer pid=47354) INFO 11-03 17:28:10 [launcher.py:46] Route: /v1/rerank, Methods: POST 2025-11-03T17:28:10.3911916Z (APIServer pid=47354) INFO 11-03 17:28:10 [launcher.py:46] Route: /v2/rerank, Methods: POST 2025-11-03T17:28:10.3912448Z (APIServer pid=47354) INFO 11-03 17:28:10 [launcher.py:46] Route: /scale_elastic_ep, Methods: POST 2025-11-03T17:28:10.3913024Z (APIServer pid=47354) INFO 11-03 17:28:10 [launcher.py:46] Route: /is_scaling_elastic_ep, Methods: POST 2025-11-03T17:28:10.3929727Z (APIServer pid=47354) INFO 11-03 17:28:10 [launcher.py:46] Route: /invocations, Methods: POST 2025-11-03T17:28:10.3930408Z (APIServer pid=47354) INFO 11-03 17:28:10 [launcher.py:46] Route: /metrics, Methods: GET 2025-11-03T17:28:10.4061692Z (APIServer pid=47354) INFO: Started server process [47354] 2025-11-03T17:28:10.4062344Z (APIServer pid=47354) INFO: Waiting for application startup. 2025-11-03T17:28:10.6223408Z (APIServer pid=47354) INFO: Application startup complete. 2025-11-03T17:28:10.9054273Z (APIServer pid=47354) INFO: 127.0.0.1:52562 - "GET /health HTTP/1.1" 200 OK 2025-11-03T17:28:11.1536382Z (APIServer pid=47354) INFO: 127.0.0.1:52566 - "GET /v1/models HTTP/1.1" 200 OK 2025-11-03T17:28:11.1839656Z (EngineCore_DP0 pid=47476) INFO 11-03 17:28:11 [ray_distributed_executor.py:569] RAY_CGRAPH_get_timeout is set to 300 2025-11-03T17:28:11.1840597Z (EngineCore_DP0 pid=47476) INFO 11-03 17:28:11 [ray_distributed_executor.py:573] VLLM_USE_RAY_COMPILED_DAG_CHANNEL_TYPE = auto 2025-11-03T17:28:11.1841517Z (EngineCore_DP0 pid=47476) INFO 11-03 17:28:11 [ray_distributed_executor.py:577] VLLM_USE_RAY_COMPILED_DAG_OVERLAP_COMM = False 2025-11-03T17:28:11.1848664Z (EngineCore_DP0 pid=47476) INFO 11-03 17:28:11 [ray_distributed_executor.py:653] Using RayPPCommunicator (which wraps vLLM _PP GroupCoordinator) for Ray Compiled Graph communication. 2025-11-03T17:28:11.2129547Z (EngineCore_DP0 pid=47476) 2025-11-03 17:28:11,212 INFO torch_tensor_accelerator_channel.py:807 -- Creating communicator group f8441aa1-741c-4c3a-91c8-0527e5fde496 on actors: [Actor(RayWorkerWrapper, 91d2cd2ec71c12ffdbbe3c8101000000), Actor(RayWorkerWrapper, 2ddfdb0b6924ff3fb1cdc0db01000000), Actor(RayWorkerWrapper, bc37947e50383403bf468f0d01000000), Actor(RayWorkerWrapper, ed858cc410e53b01626bd6ab01000000)] 2025-11-03T17:28:11.2707742Z (EngineCore_DP0 pid=47476) 2025-11-03 17:28:11,270 INFO torch_tensor_accelerator_channel.py:833 -- Communicator group initialized. 2025-11-03T17:28:11.2790977Z (EngineCore_DP0 pid=47476) (RayWorkerWrapper pid=48368) /opt/conda/envs/py_3.12/lib/python3.12/site-packages/vllm/distributed/device_communicators/ray_communicator.py:107: UserWarning: The given buffer is not writable, and PyTorch does not support non-writable tensors. This means you can write to the underlying (supposedly non-writable) buffer using the tensor. You may want to copy the buffer to protect its data or make it writable before converting it to a tensor. This type of warning will be suppressed for the rest of this program. (Triggered internally at /var/lib/jenkins/workspace/torch/csrc/utils/tensor_new.cpp:1581.) 2025-11-03T17:28:11.2793826Z (EngineCore_DP0 pid=47476) (RayWorkerWrapper pid=48368) actor_id_tensor = torch.frombuffer(actor_id_bytes, dtype=torch.uint8).to( 2025-11-03T17:28:11.2795942Z (EngineCore_DP0 pid=47476) (pid=48366) /opt/conda/envs/py_3.12/lib/python3.12/site-packages/torch/cuda/__init__.py:65: FutureWarning: The pynvml package is deprecated. Please install nvidia-ml-py instead. If you did not install pynvml directly, please report this to the maintainers of the package that installed pynvml for you. [repeated 3x across cluster] 2025-11-03T17:28:11.2797437Z (EngineCore_DP0 pid=47476) (pid=48366) import pynvml # type: ignore[import] [repeated 3x across cluster] 2025-11-03T17:28:11.5129804Z (APIServer pid=47354) INFO: 127.0.0.1:52566 - "POST /v1/completions HTTP/1.1" 200 OK 2025-11-03T17:28:11.5803004Z (APIServer pid=47354) INFO: 127.0.0.1:52566 - "POST /v1/completions HTTP/1.1" 200 OK 2025-11-03T17:28:11.7358596Z (APIServer pid=47354) INFO: 127.0.0.1:52566 - "POST /v1/completions HTTP/1.1" 200 OK 2025-11-03T17:28:11.8169270Z (APIServer pid=47354) INFO: 127.0.0.1:52566 - "POST /v1/completions HTTP/1.1" 200 OK 2025-11-03T17:28:11.8804718Z (APIServer pid=47354) INFO: 127.0.0.1:52566 - "POST /v1/completions HTTP/1.1" 200 OK 2025-11-03T17:28:11.8886011Z (APIServer pid=47354) INFO: 127.0.0.1:52566 - "POST /v1/completions HTTP/1.1" 200 OK 2025-11-03T17:28:11.9980811Z (APIServer pid=47354) INFO 11-03 17:28:11 [launcher.py:110] Shutting down FastAPI HTTP server. 2025-11-03T17:28:11.9981404Z *** SIGTERM received at time=1762190891 on cpu 7 *** 2025-11-03T17:28:12.0030675Z PC: @ 0x7fac17751117 (unknown) (unknown) 2025-11-03T17:28:12.0031734Z @ 0x7fac17702520 (unknown) (unknown) 2025-11-03T17:28:12.0032226Z [2025-11-03 17:28:12,003 E 47476 47476] logging.cc:501: *** SIGTERM received at time=1762190891 on cpu 7 *** 2025-11-03T17:28:12.0034344Z [2025-11-03 17:28:12,003 E 47476 47476] logging.cc:501: PC: @ 0x7fac17751117 (unknown) (unknown) 2025-11-03T17:28:12.0036133Z [2025-11-03 17:28:12,003 E 47476 47476] logging.cc:501: @ 0x7fac17702520 (unknown) (unknown) 2025-11-03T17:28:12.0038404Z (EngineCore_DP0 pid=47476) INFO 11-03 17:28:12 [ray_distributed_executor.py:127] Shutting down Ray distributed executor. If you see error log from logging.cc regarding SIGTERM received, please ignore because this is the expected termination process in Ray. 2025-11-03T17:28:12.0039950Z (EngineCore_DP0 pid=47476) 2025-11-03 17:28:12,003 INFO compiled_dag_node.py:2171 -- Tearing down compiled DAG 2025-11-03T17:28:12.0040986Z (EngineCore_DP0 pid=47476) 2025-11-03 17:28:12,003 INFO compiled_dag_node.py:2176 -- Cancelling compiled worker on actor: Actor(RayWorkerWrapper, bc37947e50383403bf468f0d01000000) 2025-11-03T17:28:12.0041994Z (EngineCore_DP0 pid=47476) 2025-11-03 17:28:12,003 INFO compiled_dag_node.py:2176 -- Cancelling compiled worker on actor: Actor(RayWorkerWrapper, 2ddfdb0b6924ff3fb1cdc0db01000000) 2025-11-03T17:28:12.0043261Z (EngineCore_DP0 pid=47476) 2025-11-03 17:28:12,003 INFO compiled_dag_node.py:2176 -- Cancelling compiled worker on actor: Actor(RayWorkerWrapper, ed858cc410e53b01626bd6ab01000000) 2025-11-03T17:28:12.0044256Z (EngineCore_DP0 pid=47476) 2025-11-03 17:28:12,004 INFO compiled_dag_node.py:2176 -- Cancelling compiled worker on actor: Actor(RayWorkerWrapper, 91d2cd2ec71c12ffdbbe3c8101000000) 2025-11-03T17:28:12.0097705Z (EngineCore_DP0 pid=47476) 2025-11-03 17:28:12,009 INFO compiled_dag_node.py:2198 -- Waiting for worker tasks to exit 2025-11-03T17:28:12.0098623Z (EngineCore_DP0 pid=47476) 2025-11-03 17:28:12,009 INFO compiled_dag_node.py:2201 -- Teardown complete 2025-11-03T17:28:12.5133857Z (EngineCore_DP0 pid=47476) (RayWorkerWrapper pid=48365) INFO 11-03 17:28:05 [backends.py:594] Using cache directory: /var/lib/jenkins/.cache/vllm/torch_compile_cache/ba6452aee0/rank_2_0/backbone for vLLM's torch.compile [repeated 3x across cluster] 2025-11-03T17:28:12.5137926Z (EngineCore_DP0 pid=47476) (RayWorkerWrapper pid=48365) INFO 11-03 17:28:05 [backends.py:608] Dynamo bytecode transform time: 2.12 s [repeated 3x across cluster] 2025-11-03T17:28:12.5139159Z (EngineCore_DP0 pid=47476) (RayWorkerWrapper pid=48366) INFO 11-03 17:28:05 [backends.py:179] Directly load the compiled graph(s) for dynamic shape from the cache, took 0.382 s [repeated 3x across cluster] 2025-11-03T17:28:12.5140374Z (EngineCore_DP0 pid=47476) (RayWorkerWrapper pid=48366) INFO 11-03 17:28:06 [gpu_worker.py:314] Available KV cache memory: 18.99 GiB [repeated 3x across cluster] 2025-11-03T17:28:12.5141358Z (EngineCore_DP0 pid=47476) (RayWorkerWrapper pid=48366) INFO 11-03 17:28:08 [gpu_worker.py:351] Compile and warming up model for size 4 [repeated 7x across cluster] 2025-11-03T17:28:12.5142431Z (EngineCore_DP0 pid=47476) (RayWorkerWrapper pid=48363) INFO 11-03 17:28:09 [backends.py:185] Directly load the compiled graph(s) for shape 4 from the cache, took 0.974 s [repeated 7x across cluster] 2025-11-03T17:28:12.5143469Z (EngineCore_DP0 pid=47476) (RayWorkerWrapper pid=48363) INFO 11-03 17:28:09 [monitor.py:33] torch.compile takes 2.13 s in total [repeated 3x across cluster] 2025-11-03T17:28:12.5146262Z (EngineCore_DP0 pid=47476) (RayWorkerWrapper pid=48366) /opt/conda/envs/py_3.12/lib/python3.12/site-packages/vllm/distributed/device_communicators/ray_communicator.py:107: UserWarning: The given buffer is not writable, and PyTorch does not support non-writable tensors. This means you can write to the underlying (supposedly non-writable) buffer using the tensor. You may want to copy the buffer to protect its data or make it writable before converting it to a tensor. This type of warning will be suppressed for the rest of this program. (Triggered internally at /var/lib/jenkins/workspace/torch/csrc/utils/tensor_new.cpp:1581.) [repeated 3x across cluster] 2025-11-03T17:28:12.5148679Z (EngineCore_DP0 pid=47476) (RayWorkerWrapper pid=48366) actor_id_tensor = torch.frombuffer(actor_id_bytes, dtype=torch.uint8).to( [repeated 3x across cluster] 2025-11-03T17:28:14.3411531Z (APIServer pid=47354) INFO: Shutting down 2025-11-03T17:28:14.4416427Z (APIServer pid=47354) INFO: Waiting for application shutdown. 2025-11-03T17:28:14.4417198Z (APIServer pid=47354) INFO: Application shutdown complete. 2025-11-03T17:28:14.8097095Z INFO 11-03 17:28:14 [model.py:653] Resolved architecture: LlamaForCausalLM 2025-11-03T17:28:14.8098219Z WARNING 11-03 17:28:14 [model.py:1942] Casting torch.bfloat16 to torch.float16. 2025-11-03T17:28:14.8099093Z INFO 11-03 17:28:14 [model.py:1714] Using max model len 2048 2025-11-03T17:28:14.8100884Z Launching RemoteOpenAIServer with: vllm serve meta-llama/Llama-3.2-1B-Instruct --dtype float16 --max-model-len 2048 --max-num-seqs 8 --enforce-eager --tokenizer-mode auto --tensor-parallel-size 2 --distributed-executor-backend mp --load-format dummy --port 50411 --seed 0 2025-11-03T17:28:15.3402837Z /opt/conda/envs/py_3.12/lib/python3.12/site-packages/torch/cuda/__init__.py:65: FutureWarning: The pynvml package is deprecated. Please install nvidia-ml-py instead. If you did not install pynvml directly, please report this to the maintainers of the package that installed pynvml for you. 2025-11-03T17:28:15.3405156Z import pynvml # type: ignore[import] 2025-11-03T17:28:17.6113379Z INFO 11-03 17:28:17 [__init__.py:224] Automatically detected platform cuda. 2025-11-03T17:28:21.3444589Z (APIServer pid=52361) INFO 11-03 17:28:21 [api_server.py:1870] vLLM API server version 0.11.0rc2.dev389+ge51928192.d20251103 2025-11-03T17:28:21.3471701Z (APIServer pid=52361) INFO 11-03 17:28:21 [utils.py:239] non-default args: {'model_tag': 'meta-llama/Llama-3.2-1B-Instruct', 'port': 50411, 'model': 'meta-llama/Llama-3.2-1B-Instruct', 'dtype': 'float16', 'seed': 0, 'max_model_len': 2048, 'enforce_eager': True, 'load_format': 'dummy', 'distributed_executor_backend': 'mp', 'tensor_parallel_size': 2, 'max_num_seqs': 8} 2025-11-03T17:28:21.5281332Z (APIServer pid=52361) INFO 11-03 17:28:21 [model.py:653] Resolved architecture: LlamaForCausalLM 2025-11-03T17:28:21.5282035Z (APIServer pid=52361) `torch_dtype` is deprecated! Use `dtype` instead! 2025-11-03T17:28:21.5282741Z (APIServer pid=52361) WARNING 11-03 17:28:21 [model.py:1942] Casting torch.bfloat16 to torch.float16. 2025-11-03T17:28:21.5283431Z (APIServer pid=52361) INFO 11-03 17:28:21 [model.py:1714] Using max model len 2048 2025-11-03T17:28:21.6892457Z (APIServer pid=52361) INFO 11-03 17:28:21 [scheduler.py:225] Chunked prefill is enabled with max_num_batched_tokens=2048. 2025-11-03T17:28:21.6898148Z (APIServer pid=52361) INFO 11-03 17:28:21 [vllm.py:354] Cudagraph is disabled under eager mode 2025-11-03T17:28:22.6730775Z /opt/conda/envs/py_3.12/lib/python3.12/site-packages/torch/cuda/__init__.py:65: FutureWarning: The pynvml package is deprecated. Please install nvidia-ml-py instead. If you did not install pynvml directly, please report this to the maintainers of the package that installed pynvml for you. 2025-11-03T17:28:24.9793236Z import pynvml # type: ignore[import] 2025-11-03T17:28:24.9794078Z INFO 11-03 17:28:24 [__init__.py:224] Automatically detected platform cuda. 2025-11-03T17:28:28.6607890Z (EngineCore_DP0 pid=52483) INFO 11-03 17:28:28 [core.py:727] Waiting for init message from front-end. 2025-11-03T17:28:28.6875486Z (EngineCore_DP0 pid=52483) INFO 11-03 17:28:28 [core.py:94] Initializing a V1 LLM engine (v0.11.0rc2.dev389+ge51928192.d20251103) with config: model='meta-llama/Llama-3.2-1B-Instruct', speculative_config=None, tokenizer='meta-llama/Llama-3.2-1B-Instruct', skip_tokenizer_init=False, tokenizer_mode=auto, revision=None, tokenizer_revision=None, trust_remote_code=False, dtype=torch.float16, max_seq_len=2048, download_dir=None, load_format=dummy, tensor_parallel_size=2, pipeline_parallel_size=1, data_parallel_size=1, disable_custom_all_reduce=False, quantization=None, enforce_eager=True, kv_cache_dtype=auto, device_config=cuda, structured_outputs_config=StructuredOutputsConfig(backend='auto', disable_fallback=False, disable_any_whitespace=False, disable_additional_properties=False, reasoning_parser=''), observability_config=ObservabilityConfig(show_hidden_metrics_for_version=None, otlp_traces_endpoint=None, collect_detailed_traces=None), seed=0, served_model_name=meta-llama/Llama-3.2-1B-Instruct, enable_prefix_caching=True, chunked_prefill_enabled=True, pooler_config=None, compilation_config={'level': 0, 'debug_dump_path': None, 'cache_dir': '', 'backend': '', 'custom_ops': [], 'splitting_ops': None, 'use_inductor': True, 'compile_sizes': [], 'inductor_compile_config': {'enable_auto_functionalized_v2': False}, 'inductor_passes': {}, 'cudagraph_mode': , 'use_cudagraph': False, 'cudagraph_num_of_warmups': 0, 'cudagraph_capture_sizes': [], 'cudagraph_copy_inputs': False, 'full_cuda_graph': False, 'use_inductor_graph_partition': False, 'pass_config': {}, 'max_capture_size': 0, 'local_cache_dir': None} 2025-11-03T17:28:28.6881066Z (EngineCore_DP0 pid=52483) WARNING 11-03 17:28:28 [multiproc_executor.py:744] Reducing Torch parallelism from 24 threads to 1 to avoid unnecessary CPU contention. Set OMP_NUM_THREADS in the external environment to tune this value as needed. 2025-11-03T17:28:28.6887056Z (EngineCore_DP0 pid=52483) INFO 11-03 17:28:28 [shm_broadcast.py:302] vLLM message queue communication handle: Handle(local_reader_ranks=[0, 1], buffer_handle=(2, 16777216, 10, 'psm_7e687d5c'), local_subscribe_addr='ipc:///tmp/7c2aa405-6372-444d-8c66-ecc6637e68c5', remote_subscribe_addr=None, remote_addr_ipv6=False) 2025-11-03T17:28:29.1674620Z /opt/conda/envs/py_3.12/lib/python3.12/site-packages/torch/cuda/__init__.py:65: FutureWarning: The pynvml package is deprecated. Please install nvidia-ml-py instead. If you did not install pynvml directly, please report this to the maintainers of the package that installed pynvml for you. 2025-11-03T17:28:29.1676867Z import pynvml # type: ignore[import] 2025-11-03T17:28:29.1777608Z /opt/conda/envs/py_3.12/lib/python3.12/site-packages/torch/cuda/__init__.py:65: FutureWarning: The pynvml package is deprecated. Please install nvidia-ml-py instead. If you did not install pynvml directly, please report this to the maintainers of the package that installed pynvml for you. 2025-11-03T17:28:29.1779785Z import pynvml # type: ignore[import] 2025-11-03T17:28:31.5097634Z INFO 11-03 17:28:31 [__init__.py:224] Automatically detected platform cuda. 2025-11-03T17:28:31.5451133Z INFO 11-03 17:28:31 [__init__.py:224] Automatically detected platform cuda. 2025-11-03T17:28:36.2130930Z INFO 11-03 17:28:36 [shm_broadcast.py:302] vLLM message queue communication handle: Handle(local_reader_ranks=[0], buffer_handle=(1, 10485760, 10, 'psm_65e33cc0'), local_subscribe_addr='ipc:///tmp/a8bc7418-a85d-4dab-a731-f4f229becc41', remote_subscribe_addr=None, remote_addr_ipv6=False) 2025-11-03T17:28:36.2727580Z INFO 11-03 17:28:36 [shm_broadcast.py:302] vLLM message queue communication handle: Handle(local_reader_ranks=[0], buffer_handle=(1, 10485760, 10, 'psm_a142d93c'), local_subscribe_addr='ipc:///tmp/72ef73b1-1753-46e1-a541-c23255807c78', remote_subscribe_addr=None, remote_addr_ipv6=False) 2025-11-03T17:28:36.4378111Z [Gloo] Rank 1 is connected to 1 peer ranks. Expected number of connected peer ranks is : 1 2025-11-03T17:28:36.4379304Z [Gloo] Rank 0 is connected to 1 peer ranks. Expected number of connected peer ranks is : 1 2025-11-03T17:28:36.4425145Z [Gloo] Rank 1 is connected to 1 peer ranks. Expected number of connected peer ranks is : 1 2025-11-03T17:28:36.4469993Z [Gloo] Rank 0 is connected to 1 peer ranks. Expected number of connected peer ranks is : 1 2025-11-03T17:28:36.4471301Z INFO 11-03 17:28:36 [__init__.py:1146] Found nccl from library libnccl.so.2 2025-11-03T17:28:36.4472486Z INFO 11-03 17:28:36 [__init__.py:1146] Found nccl from library libnccl.so.2 2025-11-03T17:28:36.4473930Z INFO 11-03 17:28:36 [pynccl.py:109] vLLM is using nccl==2.27.5 2025-11-03T17:28:36.4474668Z INFO 11-03 17:28:36 [pynccl.py:109] vLLM is using nccl==2.27.5 2025-11-03T17:28:36.6195831Z WARNING 11-03 17:28:36 [symm_mem.py:60] SymmMemCommunicator: Device capability 8.9 not supported, communicator is not available. 2025-11-03T17:28:36.6197486Z WARNING 11-03 17:28:36 [symm_mem.py:60] SymmMemCommunicator: Device capability 8.9 not supported, communicator is not available. 2025-11-03T17:28:36.6220738Z INFO 11-03 17:28:36 [custom_all_reduce.py:37] Skipping P2P check and trusting the driver's P2P report. 2025-11-03T17:28:36.6222221Z INFO 11-03 17:28:36 [custom_all_reduce.py:37] Skipping P2P check and trusting the driver's P2P report. 2025-11-03T17:28:36.6224227Z WARNING 11-03 17:28:36 [custom_all_reduce.py:162] Custom allreduce is disabled because your platform lacks GPU P2P capability or P2P test failed. To silence this warning, specify disable_custom_all_reduce=True explicitly. 2025-11-03T17:28:36.6226917Z WARNING 11-03 17:28:36 [custom_all_reduce.py:162] Custom allreduce is disabled because your platform lacks GPU P2P capability or P2P test failed. To silence this warning, specify disable_custom_all_reduce=True explicitly. 2025-11-03T17:28:36.7527732Z INFO 11-03 17:28:36 [shm_broadcast.py:302] vLLM message queue communication handle: Handle(local_reader_ranks=[1], buffer_handle=(1, 4194304, 6, 'psm_3ea9bc55'), local_subscribe_addr='ipc:///tmp/71189a1c-27bf-44d0-95ec-d4a8b0cf6584', remote_subscribe_addr=None, remote_addr_ipv6=False) 2025-11-03T17:28:36.7563244Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:28:36.7564461Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:28:36.7571053Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:28:36.7578527Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:28:36.8122482Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:28:36.8131616Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:28:36.8143956Z [Gloo] Rank 1 is connected to 1 peer ranks. Expected number of connected peer ranks is : 1 2025-11-03T17:28:36.8145402Z [Gloo] Rank 0 is connected to 1 peer ranks. Expected number of connected peer ranks is : 1 2025-11-03T17:28:36.8146862Z INFO 11-03 17:28:36 [__init__.py:1146] Found nccl from library libnccl.so.2 2025-11-03T17:28:36.8147948Z INFO 11-03 17:28:36 [pynccl.py:109] vLLM is using nccl==2.27.5 2025-11-03T17:28:36.8148791Z INFO 11-03 17:28:36 [__init__.py:1146] Found nccl from library libnccl.so.2 2025-11-03T17:28:36.8149562Z INFO 11-03 17:28:36 [pynccl.py:109] vLLM is using nccl==2.27.5 2025-11-03T17:28:36.8495933Z INFO 11-03 17:28:36 [parallel_state.py:1231] rank 1 in world size 2 is assigned as DP rank 0, PP rank 0, TP rank 1, EP rank 1 2025-11-03T17:28:36.8496761Z INFO 11-03 17:28:36 [parallel_state.py:1231] rank 0 in world size 2 is assigned as DP rank 0, PP rank 0, TP rank 0, EP rank 0 2025-11-03T17:28:37.0649927Z INFO 11-03 17:28:37 [topk_topp_sampler.py:59] Using FlashInfer for top-p & top-k sampling. 2025-11-03T17:28:37.0666777Z INFO 11-03 17:28:37 [topk_topp_sampler.py:59] Using FlashInfer for top-p & top-k sampling. 2025-11-03T17:28:37.0980632Z (Worker_TP0 pid=52602) INFO 11-03 17:28:37 [gpu_model_runner.py:2840] Starting to load model meta-llama/Llama-3.2-1B-Instruct... 2025-11-03T17:28:37.0985867Z (Worker_TP1 pid=52603) INFO 11-03 17:28:37 [gpu_model_runner.py:2840] Starting to load model meta-llama/Llama-3.2-1B-Instruct... 2025-11-03T17:28:37.3015509Z (Worker_TP1 pid=52603) INFO 11-03 17:28:37 [gpu_model_runner.py:2870] Loading model from scratch... 2025-11-03T17:28:37.3042377Z (Worker_TP0 pid=52602) INFO 11-03 17:28:37 [gpu_model_runner.py:2870] Loading model from scratch... 2025-11-03T17:28:37.3908362Z (Worker_TP1 pid=52603) INFO 11-03 17:28:37 [cuda.py:404] Using Flash Attention backend on V1 engine. 2025-11-03T17:28:37.3944389Z (Worker_TP0 pid=52602) INFO 11-03 17:28:37 [cuda.py:404] Using Flash Attention backend on V1 engine. 2025-11-03T17:28:37.8299647Z (Worker_TP1 pid=52603) INFO 11-03 17:28:37 [gpu_model_runner.py:2902] Model loading took 1.1667 GiB and 0.119226 seconds 2025-11-03T17:28:37.8346107Z (Worker_TP0 pid=52602) INFO 11-03 17:28:37 [gpu_model_runner.py:2902] Model loading took 1.1667 GiB and 0.120265 seconds 2025-11-03T17:28:39.7041108Z (Worker_TP1 pid=52603) INFO 11-03 17:28:39 [gpu_worker.py:314] Available KV cache memory: 18.53 GiB 2025-11-03T17:28:39.7163518Z (Worker_TP0 pid=52602) INFO 11-03 17:28:39 [gpu_worker.py:314] Available KV cache memory: 18.53 GiB 2025-11-03T17:28:39.9210851Z (EngineCore_DP0 pid=52483) INFO 11-03 17:28:39 [kv_cache_utils.py:1199] GPU KV cache size: 1,214,464 tokens 2025-11-03T17:28:39.9211959Z (EngineCore_DP0 pid=52483) INFO 11-03 17:28:39 [kv_cache_utils.py:1204] Maximum concurrency for 2,048 tokens per request: 593.00x 2025-11-03T17:28:39.9213227Z (EngineCore_DP0 pid=52483) INFO 11-03 17:28:39 [kv_cache_utils.py:1199] GPU KV cache size: 1,214,464 tokens 2025-11-03T17:28:39.9214134Z (EngineCore_DP0 pid=52483) INFO 11-03 17:28:39 [kv_cache_utils.py:1204] Maximum concurrency for 2,048 tokens per request: 593.00x 2025-11-03T17:28:40.0423195Z (EngineCore_DP0 pid=52483) INFO 11-03 17:28:40 [core.py:240] init engine (profile, create kv cache, warmup model) took 2.15 seconds 2025-11-03T17:28:41.4923947Z (EngineCore_DP0 pid=52483) INFO 11-03 17:28:41 [vllm.py:354] Cudagraph is disabled under eager mode 2025-11-03T17:28:41.6771420Z (APIServer pid=52361) INFO 11-03 17:28:41 [loggers.py:172] Engine 000: vllm cache_config_info with initialization after num_gpu_blocks is: 75904 2025-11-03T17:28:41.6884603Z (EngineCore_DP0 pid=52483) INFO 11-03 17:28:41 [gc_utils.py:40] GC Debug Config. enabled:False,top_objects:-1 2025-11-03T17:28:41.7099718Z (APIServer pid=52361) INFO 11-03 17:28:41 [api_server.py:1628] Supported tasks: ['generate'] 2025-11-03T17:28:41.7353975Z (APIServer pid=52361) WARNING 11-03 17:28:41 [model.py:1581] Default sampling parameters have been overridden by the model's Hugging Face generation config recommended from the model creator. If this is not intended, please relaunch vLLM instance with `--generation-config vllm`. 2025-11-03T17:28:41.7355575Z (APIServer pid=52361) INFO 11-03 17:28:41 [serving_responses.py:157] Using default chat sampling params from model: {'temperature': 0.6, 'top_p': 0.9} 2025-11-03T17:28:41.7574737Z (APIServer pid=52361) INFO 11-03 17:28:41 [serving_chat.py:159] Using default chat sampling params from model: {'temperature': 0.6, 'top_p': 0.9} 2025-11-03T17:28:41.7795432Z (APIServer pid=52361) INFO 11-03 17:28:41 [serving_completion.py:67] Using default completion sampling params from model: {'temperature': 0.6, 'top_p': 0.9} 2025-11-03T17:28:41.7796651Z (APIServer pid=52361) INFO 11-03 17:28:41 [api_server.py:1939] Starting vLLM API server 0 on http://0.0.0.0:50411 2025-11-03T17:28:41.7797704Z (APIServer pid=52361) INFO 11-03 17:28:41 [launcher.py:38] Available routes are: 2025-11-03T17:28:41.7798518Z (APIServer pid=52361) INFO 11-03 17:28:41 [launcher.py:46] Route: /openapi.json, Methods: HEAD, GET 2025-11-03T17:28:41.7799512Z (APIServer pid=52361) INFO 11-03 17:28:41 [launcher.py:46] Route: /docs, Methods: HEAD, GET 2025-11-03T17:28:41.7800429Z (APIServer pid=52361) INFO 11-03 17:28:41 [launcher.py:46] Route: /docs/oauth2-redirect, Methods: HEAD, GET 2025-11-03T17:28:41.7801398Z (APIServer pid=52361) INFO 11-03 17:28:41 [launcher.py:46] Route: /redoc, Methods: HEAD, GET 2025-11-03T17:28:41.7802124Z (APIServer pid=52361) INFO 11-03 17:28:41 [launcher.py:46] Route: /health, Methods: GET 2025-11-03T17:28:41.7803220Z (APIServer pid=52361) INFO 11-03 17:28:41 [launcher.py:46] Route: /load, Methods: GET 2025-11-03T17:28:41.7803801Z (APIServer pid=52361) INFO 11-03 17:28:41 [launcher.py:46] Route: /ping, Methods: POST 2025-11-03T17:28:41.7804464Z (APIServer pid=52361) INFO 11-03 17:28:41 [launcher.py:46] Route: /ping, Methods: GET 2025-11-03T17:28:41.7805039Z (APIServer pid=52361) INFO 11-03 17:28:41 [launcher.py:46] Route: /tokenize, Methods: POST 2025-11-03T17:28:41.7805741Z (APIServer pid=52361) INFO 11-03 17:28:41 [launcher.py:46] Route: /detokenize, Methods: POST 2025-11-03T17:28:41.7806321Z (APIServer pid=52361) INFO 11-03 17:28:41 [launcher.py:46] Route: /v1/models, Methods: GET 2025-11-03T17:28:41.7806961Z (APIServer pid=52361) INFO 11-03 17:28:41 [launcher.py:46] Route: /version, Methods: GET 2025-11-03T17:28:41.7807494Z (APIServer pid=52361) INFO 11-03 17:28:41 [launcher.py:46] Route: /v1/responses, Methods: POST 2025-11-03T17:28:41.7808245Z (APIServer pid=52361) INFO 11-03 17:28:41 [launcher.py:46] Route: /v1/responses/{response_id}, Methods: GET 2025-11-03T17:28:41.7809021Z (APIServer pid=52361) INFO 11-03 17:28:41 [launcher.py:46] Route: /v1/responses/{response_id}/cancel, Methods: POST 2025-11-03T17:28:41.7809831Z (APIServer pid=52361) INFO 11-03 17:28:41 [launcher.py:46] Route: /v1/chat/completions, Methods: POST 2025-11-03T17:28:41.7810522Z (APIServer pid=52361) INFO 11-03 17:28:41 [launcher.py:46] Route: /v1/completions, Methods: POST 2025-11-03T17:28:41.7811087Z (APIServer pid=52361) INFO 11-03 17:28:41 [launcher.py:46] Route: /v1/embeddings, Methods: POST 2025-11-03T17:28:41.7811611Z (APIServer pid=52361) INFO 11-03 17:28:41 [launcher.py:46] Route: /pooling, Methods: POST 2025-11-03T17:28:41.7812122Z (APIServer pid=52361) INFO 11-03 17:28:41 [launcher.py:46] Route: /classify, Methods: POST 2025-11-03T17:28:41.7812631Z (APIServer pid=52361) INFO 11-03 17:28:41 [launcher.py:46] Route: /score, Methods: POST 2025-11-03T17:28:41.7813134Z (APIServer pid=52361) INFO 11-03 17:28:41 [launcher.py:46] Route: /v1/score, Methods: POST 2025-11-03T17:28:41.7813683Z (APIServer pid=52361) INFO 11-03 17:28:41 [launcher.py:46] Route: /v1/audio/transcriptions, Methods: POST 2025-11-03T17:28:41.7814264Z (APIServer pid=52361) INFO 11-03 17:28:41 [launcher.py:46] Route: /v1/audio/translations, Methods: POST 2025-11-03T17:28:41.7814792Z (APIServer pid=52361) INFO 11-03 17:28:41 [launcher.py:46] Route: /rerank, Methods: POST 2025-11-03T17:28:41.7815299Z (APIServer pid=52361) INFO 11-03 17:28:41 [launcher.py:46] Route: /v1/rerank, Methods: POST 2025-11-03T17:28:41.7815805Z (APIServer pid=52361) INFO 11-03 17:28:41 [launcher.py:46] Route: /v2/rerank, Methods: POST 2025-11-03T17:28:41.7816339Z (APIServer pid=52361) INFO 11-03 17:28:41 [launcher.py:46] Route: /scale_elastic_ep, Methods: POST 2025-11-03T17:28:41.7816901Z (APIServer pid=52361) INFO 11-03 17:28:41 [launcher.py:46] Route: /is_scaling_elastic_ep, Methods: POST 2025-11-03T17:28:41.7817452Z (APIServer pid=52361) INFO 11-03 17:28:41 [launcher.py:46] Route: /invocations, Methods: POST 2025-11-03T17:28:41.7817966Z (APIServer pid=52361) INFO 11-03 17:28:41 [launcher.py:46] Route: /metrics, Methods: GET 2025-11-03T17:28:41.7956878Z (APIServer pid=52361) INFO: Started server process [52361] 2025-11-03T17:28:41.7957562Z (APIServer pid=52361) INFO: Waiting for application startup. 2025-11-03T17:28:42.0048665Z (APIServer pid=52361) INFO: Application startup complete. 2025-11-03T17:28:42.3809270Z (APIServer pid=52361) INFO: 127.0.0.1:37412 - "GET /health HTTP/1.1" 200 OK 2025-11-03T17:28:42.3897844Z (APIServer pid=52361) INFO: 127.0.0.1:37418 - "GET /v1/models HTTP/1.1" 200 OK 2025-11-03T17:28:42.5627796Z (APIServer pid=52361) INFO: 127.0.0.1:37418 - "POST /v1/completions HTTP/1.1" 200 OK 2025-11-03T17:28:42.6197055Z (APIServer pid=52361) INFO: 127.0.0.1:37418 - "POST /v1/completions HTTP/1.1" 200 OK 2025-11-03T17:28:42.7164034Z (APIServer pid=52361) INFO: 127.0.0.1:37418 - "POST /v1/completions HTTP/1.1" 200 OK 2025-11-03T17:28:42.7905856Z (APIServer pid=52361) INFO: 127.0.0.1:37418 - "POST /v1/completions HTTP/1.1" 200 OK 2025-11-03T17:28:42.8611042Z (APIServer pid=52361) INFO: 127.0.0.1:37418 - "POST /v1/completions HTTP/1.1" 200 OK 2025-11-03T17:28:42.8689931Z (APIServer pid=52361) INFO: 127.0.0.1:37418 - "POST /v1/completions HTTP/1.1" 200 OK 2025-11-03T17:28:42.9363150Z (APIServer pid=52361) INFO 11-03 17:28:42 [launcher.py:110] Shutting down FastAPI HTTP server. 2025-11-03T17:28:42.9365703Z (Worker_TP0 pid=52602) INFO 11-03 17:28:42 [multiproc_executor.py:583] Parent process exited, terminating worker 2025-11-03T17:28:42.9366678Z (Worker_TP0 pid=52602) INFO 11-03 17:28:42 [multiproc_executor.py:624] WorkerProc shutting down. 2025-11-03T17:28:42.9369908Z (Worker_TP1 pid=52603) INFO 11-03 17:28:42 [multiproc_executor.py:583] Parent process exited, terminating worker 2025-11-03T17:28:42.9371398Z (Worker_TP1 pid=52603) INFO 11-03 17:28:42 [multiproc_executor.py:624] WorkerProc shutting down. 2025-11-03T17:28:45.0539753Z (APIServer pid=52361) INFO: Shutting down 2025-11-03T17:28:45.1544784Z (APIServer pid=52361) INFO: Waiting for application shutdown. 2025-11-03T17:28:45.1547948Z (APIServer pid=52361) INFO: Application shutdown complete. 2025-11-03T17:28:45.7604483Z PASSED 2025-11-03T17:28:45.7749526Z distributed/test_sequence_parallel.py::test_tp_sp_generation[meta-llama/Llama-3.2-1B-Instruct-parallel_setup14-mp-auto-test_options14] Fork a new process to run a test 52806 2025-11-03T17:28:45.7762602Z Fork a new process to run a test 0 2025-11-03T17:28:46.4058282Z INFO 11-03 17:28:46 [model.py:653] Resolved architecture: LlamaForCausalLM 2025-11-03T17:28:46.4059180Z `torch_dtype` is deprecated! Use `dtype` instead! 2025-11-03T17:28:46.4060001Z WARNING 11-03 17:28:46 [model.py:1942] Casting torch.bfloat16 to torch.float16. 2025-11-03T17:28:46.4060833Z INFO 11-03 17:28:46 [model.py:1714] Using max model len 2048 2025-11-03T17:28:46.4065738Z Launching RemoteOpenAIServer with: vllm serve meta-llama/Llama-3.2-1B-Instruct --dtype float16 --max-model-len 2048 --max-num-seqs 8 --enable-chunked-prefill --enforce-eager --tokenizer-mode auto --tensor-parallel-size 2 --pipeline-parallel-size 2 --distributed-executor-backend mp --compilation_config {"level": 3, "custom_ops": ["+rms_norm"], "compile_sizes": [4, 8], "pass_config": {"enable_sequence_parallelism": true, "enable_fusion": false, "enable_noop": true}} --load-format dummy --port 57037 --seed 0 2025-11-03T17:28:46.9179529Z /opt/conda/envs/py_3.12/lib/python3.12/site-packages/torch/cuda/__init__.py:65: FutureWarning: The pynvml package is deprecated. Please install nvidia-ml-py instead. If you did not install pynvml directly, please report this to the maintainers of the package that installed pynvml for you. 2025-11-03T17:28:46.9180731Z import pynvml # type: ignore[import] 2025-11-03T17:28:49.1630678Z INFO 11-03 17:28:49 [__init__.py:224] Automatically detected platform cuda. 2025-11-03T17:28:52.8305273Z (APIServer pid=52855) INFO 11-03 17:28:52 [api_server.py:1870] vLLM API server version 0.11.0rc2.dev389+ge51928192.d20251103 2025-11-03T17:28:52.8333243Z (APIServer pid=52855) INFO 11-03 17:28:52 [utils.py:239] non-default args: {'model_tag': 'meta-llama/Llama-3.2-1B-Instruct', 'port': 57037, 'model': 'meta-llama/Llama-3.2-1B-Instruct', 'dtype': 'float16', 'seed': 0, 'max_model_len': 2048, 'enforce_eager': True, 'load_format': 'dummy', 'distributed_executor_backend': 'mp', 'pipeline_parallel_size': 2, 'tensor_parallel_size': 2, 'max_num_seqs': 8, 'enable_chunked_prefill': True, 'compilation_config': {'level': 3, 'debug_dump_path': None, 'cache_dir': '', 'backend': '', 'custom_ops': ['+rms_norm'], 'splitting_ops': None, 'use_inductor': True, 'compile_sizes': [4, 8], 'inductor_compile_config': {'enable_auto_functionalized_v2': False}, 'inductor_passes': {}, 'cudagraph_mode': None, 'use_cudagraph': True, 'cudagraph_num_of_warmups': 0, 'cudagraph_capture_sizes': None, 'cudagraph_copy_inputs': False, 'full_cuda_graph': False, 'use_inductor_graph_partition': False, 'pass_config': {'enable_noop': True, 'enable_sequence_parallelism': True}, 'max_capture_size': None, 'local_cache_dir': None}} 2025-11-03T17:28:52.9849171Z (APIServer pid=52855) INFO 11-03 17:28:52 [model.py:653] Resolved architecture: LlamaForCausalLM 2025-11-03T17:28:52.9850196Z (APIServer pid=52855) `torch_dtype` is deprecated! Use `dtype` instead! 2025-11-03T17:28:52.9850908Z (APIServer pid=52855) WARNING 11-03 17:28:52 [model.py:1942] Casting torch.bfloat16 to torch.float16. 2025-11-03T17:28:52.9851732Z (APIServer pid=52855) INFO 11-03 17:28:52 [model.py:1714] Using max model len 2048 2025-11-03T17:28:53.1429154Z (APIServer pid=52855) INFO 11-03 17:28:53 [scheduler.py:225] Chunked prefill is enabled with max_num_batched_tokens=2048. 2025-11-03T17:28:53.1435400Z (APIServer pid=52855) INFO 11-03 17:28:53 [vllm.py:354] Cudagraph is disabled under eager mode 2025-11-03T17:28:54.1250973Z /opt/conda/envs/py_3.12/lib/python3.12/site-packages/torch/cuda/__init__.py:65: FutureWarning: The pynvml package is deprecated. Please install nvidia-ml-py instead. If you did not install pynvml directly, please report this to the maintainers of the package that installed pynvml for you. 2025-11-03T17:28:54.1253333Z import pynvml # type: ignore[import] 2025-11-03T17:28:56.4525466Z INFO 11-03 17:28:56 [__init__.py:224] Automatically detected platform cuda. 2025-11-03T17:29:00.1249162Z (EngineCore_DP0 pid=52977) INFO 11-03 17:29:00 [core.py:727] Waiting for init message from front-end. 2025-11-03T17:29:00.1505316Z (EngineCore_DP0 pid=52977) INFO 11-03 17:29:00 [core.py:94] Initializing a V1 LLM engine (v0.11.0rc2.dev389+ge51928192.d20251103) with config: model='meta-llama/Llama-3.2-1B-Instruct', speculative_config=None, tokenizer='meta-llama/Llama-3.2-1B-Instruct', skip_tokenizer_init=False, tokenizer_mode=auto, revision=None, tokenizer_revision=None, trust_remote_code=False, dtype=torch.float16, max_seq_len=2048, download_dir=None, load_format=dummy, tensor_parallel_size=2, pipeline_parallel_size=2, data_parallel_size=1, disable_custom_all_reduce=False, quantization=None, enforce_eager=True, kv_cache_dtype=auto, device_config=cuda, structured_outputs_config=StructuredOutputsConfig(backend='auto', disable_fallback=False, disable_any_whitespace=False, disable_additional_properties=False, reasoning_parser=''), observability_config=ObservabilityConfig(show_hidden_metrics_for_version=None, otlp_traces_endpoint=None, collect_detailed_traces=None), seed=0, served_model_name=meta-llama/Llama-3.2-1B-Instruct, enable_prefix_caching=True, chunked_prefill_enabled=True, pooler_config=None, compilation_config={'level': 3, 'debug_dump_path': None, 'cache_dir': '', 'backend': '', 'custom_ops': ['+rms_norm', '+rms_norm'], 'splitting_ops': ['vllm.unified_attention', 'vllm.unified_attention_with_output', 'vllm.unified_mla_attention', 'vllm.unified_mla_attention_with_output', 'vllm.mamba_mixer2', 'vllm.mamba_mixer', 'vllm.short_conv', 'vllm.linear_attention', 'vllm.plamo2_mamba_mixer', 'vllm.gdn_attention', 'vllm.sparse_attn_indexer'], 'use_inductor': True, 'compile_sizes': [8, 4], 'inductor_compile_config': {'enable_auto_functionalized_v2': False}, 'inductor_passes': {}, 'cudagraph_mode': , 'use_cudagraph': False, 'cudagraph_num_of_warmups': 0, 'cudagraph_capture_sizes': [], 'cudagraph_copy_inputs': False, 'full_cuda_graph': False, 'use_inductor_graph_partition': False, 'pass_config': {'enable_noop': True, 'enable_sequence_parallelism': True}, 'max_capture_size': 0, 'local_cache_dir': None} 2025-11-03T17:29:00.1512428Z (EngineCore_DP0 pid=52977) WARNING 11-03 17:29:00 [multiproc_executor.py:744] Reducing Torch parallelism from 24 threads to 1 to avoid unnecessary CPU contention. Set OMP_NUM_THREADS in the external environment to tune this value as needed. 2025-11-03T17:29:00.1514668Z (EngineCore_DP0 pid=52977) INFO 11-03 17:29:00 [shm_broadcast.py:302] vLLM message queue communication handle: Handle(local_reader_ranks=[0, 1, 2, 3], buffer_handle=(4, 16777216, 10, 'psm_9812c26c'), local_subscribe_addr='ipc:///tmp/8d17d324-a1b2-4294-9ec9-d6d9283bf864', remote_subscribe_addr=None, remote_addr_ipv6=False) 2025-11-03T17:29:00.6534638Z /opt/conda/envs/py_3.12/lib/python3.12/site-packages/torch/cuda/__init__.py:65: FutureWarning: The pynvml package is deprecated. Please install nvidia-ml-py instead. If you did not install pynvml directly, please report this to the maintainers of the package that installed pynvml for you. 2025-11-03T17:29:00.6536038Z import pynvml # type: ignore[import] 2025-11-03T17:29:00.6537170Z /opt/conda/envs/py_3.12/lib/python3.12/site-packages/torch/cuda/__init__.py:65: FutureWarning: The pynvml package is deprecated. Please install nvidia-ml-py instead. If you did not install pynvml directly, please report this to the maintainers of the package that installed pynvml for you. 2025-11-03T17:29:00.6538290Z import pynvml # type: ignore[import] 2025-11-03T17:29:00.6827503Z /opt/conda/envs/py_3.12/lib/python3.12/site-packages/torch/cuda/__init__.py:65: FutureWarning: The pynvml package is deprecated. Please install nvidia-ml-py instead. If you did not install pynvml directly, please report this to the maintainers of the package that installed pynvml for you. 2025-11-03T17:29:00.6828684Z import pynvml # type: ignore[import] 2025-11-03T17:29:00.8114354Z /opt/conda/envs/py_3.12/lib/python3.12/site-packages/torch/cuda/__init__.py:65: FutureWarning: The pynvml package is deprecated. Please install nvidia-ml-py instead. If you did not install pynvml directly, please report this to the maintainers of the package that installed pynvml for you. 2025-11-03T17:29:00.8117303Z import pynvml # type: ignore[import] 2025-11-03T17:29:03.0629157Z INFO 11-03 17:29:03 [__init__.py:224] Automatically detected platform cuda. 2025-11-03T17:29:03.0656752Z INFO 11-03 17:29:03 [__init__.py:224] Automatically detected platform cuda. 2025-11-03T17:29:03.1443053Z INFO 11-03 17:29:03 [__init__.py:224] Automatically detected platform cuda. 2025-11-03T17:29:03.2202875Z INFO 11-03 17:29:03 [__init__.py:224] Automatically detected platform cuda. 2025-11-03T17:29:07.9870964Z INFO 11-03 17:29:07 [shm_broadcast.py:302] vLLM message queue communication handle: Handle(local_reader_ranks=[0], buffer_handle=(1, 10485760, 10, 'psm_e23da392'), local_subscribe_addr='ipc:///tmp/b96b13a5-aa0f-46a0-a136-9aad08b9a3ad', remote_subscribe_addr=None, remote_addr_ipv6=False) 2025-11-03T17:29:08.0224073Z INFO 11-03 17:29:08 [shm_broadcast.py:302] vLLM message queue communication handle: Handle(local_reader_ranks=[0], buffer_handle=(1, 10485760, 10, 'psm_fee2cc8f'), local_subscribe_addr='ipc:///tmp/bc406a42-1efe-4c47-912b-134ce1ddf602', remote_subscribe_addr=None, remote_addr_ipv6=False) 2025-11-03T17:29:08.1025874Z INFO 11-03 17:29:08 [shm_broadcast.py:302] vLLM message queue communication handle: Handle(local_reader_ranks=[0], buffer_handle=(1, 10485760, 10, 'psm_9a09e65d'), local_subscribe_addr='ipc:///tmp/df3652a4-9007-4557-94c2-b26247b2fac3', remote_subscribe_addr=None, remote_addr_ipv6=False) 2025-11-03T17:29:08.1123471Z INFO 11-03 17:29:08 [shm_broadcast.py:302] vLLM message queue communication handle: Handle(local_reader_ranks=[0], buffer_handle=(1, 10485760, 10, 'psm_154e5f17'), local_subscribe_addr='ipc:///tmp/f8d6db1e-f7f4-4a07-86c3-c0b2a796bb5b', remote_subscribe_addr=None, remote_addr_ipv6=False) 2025-11-03T17:29:08.9300335Z [Gloo] Rank 0 is connected to 3 peer ranks. Expected number of connected peer ranks is : 3 2025-11-03T17:29:08.9301048Z [Gloo] Rank 1 is connected to 3 peer ranks. Expected number of connected peer ranks is : 3 2025-11-03T17:29:08.9302358Z [Gloo] Rank 2 is connected to 3 peer ranks. Expected number of connected peer ranks is : 3 2025-11-03T17:29:08.9303045Z [Gloo] Rank 3 is connected to 3 peer ranks. Expected number of connected peer ranks is : 3 2025-11-03T17:29:08.9348318Z [Gloo] Rank 0 is connected to 1 peer ranks. Expected number of connected peer ranks is : 1 2025-11-03T17:29:08.9348826Z [Gloo] Rank 1 is connected to 1 peer ranks. Expected number of connected peer ranks is : 1 2025-11-03T17:29:08.9350851Z [Gloo] Rank 0 is connected to 1 peer ranks. Expected number of connected peer ranks is : 1 2025-11-03T17:29:08.9351560Z [Gloo] Rank 1 is connected to 1 peer ranks. Expected number of connected peer ranks is : 1 2025-11-03T17:29:08.9392324Z INFO 11-03 17:29:08 [__init__.py:1146] Found nccl from library libnccl.so.2 2025-11-03T17:29:08.9392794Z INFO 11-03 17:29:08 [__init__.py:1146] Found nccl from library libnccl.so.2 2025-11-03T17:29:08.9393360Z INFO 11-03 17:29:08 [__init__.py:1146] Found nccl from library libnccl.so.2 2025-11-03T17:29:08.9393764Z INFO 11-03 17:29:08 [__init__.py:1146] Found nccl from library libnccl.so.2 2025-11-03T17:29:08.9395457Z INFO 11-03 17:29:08 [pynccl.py:109] vLLM is using nccl==2.27.5 2025-11-03T17:29:08.9395844Z INFO 11-03 17:29:08 [pynccl.py:109] vLLM is using nccl==2.27.5 2025-11-03T17:29:08.9396216Z INFO 11-03 17:29:08 [pynccl.py:109] vLLM is using nccl==2.27.5 2025-11-03T17:29:08.9396572Z INFO 11-03 17:29:08 [pynccl.py:109] vLLM is using nccl==2.27.5 2025-11-03T17:29:09.1254328Z WARNING 11-03 17:29:09 [symm_mem.py:60] SymmMemCommunicator: Device capability 8.9 not supported, communicator is not available. 2025-11-03T17:29:09.1256049Z WARNING 11-03 17:29:09 [symm_mem.py:60] SymmMemCommunicator: Device capability 8.9 not supported, communicator is not available. 2025-11-03T17:29:09.1278086Z INFO 11-03 17:29:09 [custom_all_reduce.py:37] Skipping P2P check and trusting the driver's P2P report. 2025-11-03T17:29:09.1279011Z INFO 11-03 17:29:09 [custom_all_reduce.py:37] Skipping P2P check and trusting the driver's P2P report. 2025-11-03T17:29:09.1280015Z WARNING 11-03 17:29:09 [custom_all_reduce.py:162] Custom allreduce is disabled because your platform lacks GPU P2P capability or P2P test failed. To silence this warning, specify disable_custom_all_reduce=True explicitly. 2025-11-03T17:29:09.1281375Z WARNING 11-03 17:29:09 [custom_all_reduce.py:162] Custom allreduce is disabled because your platform lacks GPU P2P capability or P2P test failed. To silence this warning, specify disable_custom_all_reduce=True explicitly. 2025-11-03T17:29:09.1300854Z INFO 11-03 17:29:09 [shm_broadcast.py:302] vLLM message queue communication handle: Handle(local_reader_ranks=[1], buffer_handle=(1, 4194304, 6, 'psm_cc98068d'), local_subscribe_addr='ipc:///tmp/a27a48ac-6bda-467b-af9a-14da2b1a210e', remote_subscribe_addr=None, remote_addr_ipv6=False) 2025-11-03T17:29:09.2317303Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:29:09.2318565Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:29:09.2484640Z WARNING 11-03 17:29:09 [symm_mem.py:60] SymmMemCommunicator: Device capability 8.9 not supported, communicator is not available. 2025-11-03T17:29:09.2486227Z WARNING 11-03 17:29:09 [symm_mem.py:60] SymmMemCommunicator: Device capability 8.9 not supported, communicator is not available. 2025-11-03T17:29:09.2504459Z INFO 11-03 17:29:09 [custom_all_reduce.py:37] Skipping P2P check and trusting the driver's P2P report. 2025-11-03T17:29:09.2505790Z WARNING 11-03 17:29:09 [custom_all_reduce.py:162] Custom allreduce is disabled because your platform lacks GPU P2P capability or P2P test failed. To silence this warning, specify disable_custom_all_reduce=True explicitly. 2025-11-03T17:29:09.2506793Z INFO 11-03 17:29:09 [custom_all_reduce.py:37] Skipping P2P check and trusting the driver's P2P report. 2025-11-03T17:29:09.2507780Z WARNING 11-03 17:29:09 [custom_all_reduce.py:162] Custom allreduce is disabled because your platform lacks GPU P2P capability or P2P test failed. To silence this warning, specify disable_custom_all_reduce=True explicitly. 2025-11-03T17:29:09.2532520Z INFO 11-03 17:29:09 [shm_broadcast.py:302] vLLM message queue communication handle: Handle(local_reader_ranks=[1], buffer_handle=(1, 4194304, 6, 'psm_652d2be5'), local_subscribe_addr='ipc:///tmp/6234638d-75a1-4fc9-bf01-238d916116e0', remote_subscribe_addr=None, remote_addr_ipv6=False) 2025-11-03T17:29:09.2551632Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:29:09.2565619Z [Gloo] Rank 1 is connected to 1 peer ranks. Expected number of connected peer ranks is : 1 2025-11-03T17:29:09.2567152Z [Gloo] Rank 0 is connected to 1 peer ranks. Expected number of connected peer ranks is : 1 2025-11-03T17:29:09.2569288Z INFO 11-03 17:29:09 [__init__.py:1146] Found nccl from library libnccl.so.2 2025-11-03T17:29:09.2569720Z INFO 11-03 17:29:09 [__init__.py:1146] Found nccl from library libnccl.so.2 2025-11-03T17:29:09.2570244Z INFO 11-03 17:29:09 [pynccl.py:109] vLLM is using nccl==2.27.5 2025-11-03T17:29:09.2570640Z INFO 11-03 17:29:09 [pynccl.py:109] vLLM is using nccl==2.27.5 2025-11-03T17:29:09.2917906Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:29:09.2919078Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:29:09.3421709Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:29:09.3436847Z [Gloo] Rank 0 is connected to 1 peer ranks. Expected number of connected peer ranks is : 1 2025-11-03T17:29:09.3437443Z [Gloo] Rank 1 is connected to 1 peer ranks. Expected number of connected peer ranks is : 1 2025-11-03T17:29:09.3441102Z INFO 11-03 17:29:09 [__init__.py:1146] Found nccl from library libnccl.so.2 2025-11-03T17:29:09.3441554Z INFO 11-03 17:29:09 [pynccl.py:109] vLLM is using nccl==2.27.5 2025-11-03T17:29:09.3442727Z INFO 11-03 17:29:09 [__init__.py:1146] Found nccl from library libnccl.so.2 2025-11-03T17:29:09.3443120Z INFO 11-03 17:29:09 [pynccl.py:109] vLLM is using nccl==2.27.5 2025-11-03T17:29:09.3799303Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:29:09.3799913Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:29:09.3812016Z [Gloo] Rank 0 is connected to 1 peer ranks. Expected number of connected peer ranks is : 1 2025-11-03T17:29:09.3813746Z [Gloo] Rank 1 is connected to 1 peer ranks. Expected number of connected peer ranks is : 1 2025-11-03T17:29:09.3815240Z [Gloo] Rank 0 is connected to 1 peer ranks. Expected number of connected peer ranks is : 1 2025-11-03T17:29:09.3816817Z [Gloo] Rank 1 is connected to 1 peer ranks. Expected number of connected peer ranks is : 1 2025-11-03T17:29:09.3817950Z INFO 11-03 17:29:09 [__init__.py:1146] Found nccl from library libnccl.so.2 2025-11-03T17:29:09.3818327Z INFO 11-03 17:29:09 [__init__.py:1146] Found nccl from library libnccl.so.2 2025-11-03T17:29:09.3818769Z INFO 11-03 17:29:09 [pynccl.py:109] vLLM is using nccl==2.27.5 2025-11-03T17:29:09.3819084Z INFO 11-03 17:29:09 [pynccl.py:109] vLLM is using nccl==2.27.5 2025-11-03T17:29:09.3819510Z INFO 11-03 17:29:09 [__init__.py:1146] Found nccl from library libnccl.so.2 2025-11-03T17:29:09.3819849Z INFO 11-03 17:29:09 [__init__.py:1146] Found nccl from library libnccl.so.2 2025-11-03T17:29:09.3820493Z INFO 11-03 17:29:09 [pynccl.py:109] vLLM is using nccl==2.27.5 2025-11-03T17:29:09.3820789Z INFO 11-03 17:29:09 [pynccl.py:109] vLLM is using nccl==2.27.5 2025-11-03T17:29:09.4245671Z INFO 11-03 17:29:09 [parallel_state.py:1231] rank 2 in world size 4 is assigned as DP rank 0, PP rank 1, TP rank 0, EP rank 0 2025-11-03T17:29:09.4246426Z INFO 11-03 17:29:09 [parallel_state.py:1231] rank 3 in world size 4 is assigned as DP rank 0, PP rank 1, TP rank 1, EP rank 1 2025-11-03T17:29:09.4269699Z INFO 11-03 17:29:09 [parallel_state.py:1231] rank 1 in world size 4 is assigned as DP rank 0, PP rank 0, TP rank 1, EP rank 1 2025-11-03T17:29:09.4271198Z INFO 11-03 17:29:09 [parallel_state.py:1231] rank 0 in world size 4 is assigned as DP rank 0, PP rank 0, TP rank 0, EP rank 0 2025-11-03T17:29:09.6722264Z INFO 11-03 17:29:09 [topk_topp_sampler.py:59] Using FlashInfer for top-p & top-k sampling. 2025-11-03T17:29:09.6836870Z INFO 11-03 17:29:09 [topk_topp_sampler.py:59] Using FlashInfer for top-p & top-k sampling. 2025-11-03T17:29:09.6854720Z INFO 11-03 17:29:09 [topk_topp_sampler.py:59] Using FlashInfer for top-p & top-k sampling. 2025-11-03T17:29:09.6862988Z INFO 11-03 17:29:09 [topk_topp_sampler.py:59] Using FlashInfer for top-p & top-k sampling. 2025-11-03T17:29:09.7023875Z (Worker_PP1_TP1 pid=53099) INFO 11-03 17:29:09 [gpu_model_runner.py:2840] Starting to load model meta-llama/Llama-3.2-1B-Instruct... 2025-11-03T17:29:09.7154967Z (Worker_PP0_TP0 pid=53096) INFO 11-03 17:29:09 [gpu_model_runner.py:2840] Starting to load model meta-llama/Llama-3.2-1B-Instruct... 2025-11-03T17:29:09.7163192Z (Worker_PP1_TP0 pid=53098) INFO 11-03 17:29:09 [gpu_model_runner.py:2840] Starting to load model meta-llama/Llama-3.2-1B-Instruct... 2025-11-03T17:29:09.7172277Z (Worker_PP0_TP1 pid=53097) INFO 11-03 17:29:09 [gpu_model_runner.py:2840] Starting to load model meta-llama/Llama-3.2-1B-Instruct... 2025-11-03T17:29:09.9111463Z (Worker_PP1_TP1 pid=53099) INFO 11-03 17:29:09 [gpu_model_runner.py:2870] Loading model from scratch... 2025-11-03T17:29:09.9273077Z (Worker_PP1_TP0 pid=53098) INFO 11-03 17:29:09 [gpu_model_runner.py:2870] Loading model from scratch... 2025-11-03T17:29:09.9516525Z (Worker_PP0_TP0 pid=53096) INFO 11-03 17:29:09 [gpu_model_runner.py:2870] Loading model from scratch... 2025-11-03T17:29:09.9632163Z (Worker_PP0_TP1 pid=53097) INFO 11-03 17:29:09 [gpu_model_runner.py:2870] Loading model from scratch... 2025-11-03T17:29:10.0005042Z (Worker_PP1_TP1 pid=53099) INFO 11-03 17:29:10 [cuda.py:404] Using Flash Attention backend on V1 engine. 2025-11-03T17:29:10.0187801Z (Worker_PP1_TP0 pid=53098) INFO 11-03 17:29:10 [cuda.py:404] Using Flash Attention backend on V1 engine. 2025-11-03T17:29:10.0426492Z (Worker_PP0_TP0 pid=53096) INFO 11-03 17:29:10 [cuda.py:404] Using Flash Attention backend on V1 engine. 2025-11-03T17:29:10.0521036Z (Worker_PP0_TP1 pid=53097) INFO 11-03 17:29:10 [cuda.py:404] Using Flash Attention backend on V1 engine. 2025-11-03T17:29:10.4766451Z (Worker_PP1_TP1 pid=53099) INFO 11-03 17:29:10 [gpu_model_runner.py:2902] Model loading took 0.7135 GiB and 0.152969 seconds 2025-11-03T17:29:10.4937488Z (Worker_PP1_TP0 pid=53098) INFO 11-03 17:29:10 [gpu_model_runner.py:2902] Model loading took 0.7135 GiB and 0.155727 seconds 2025-11-03T17:29:10.5767251Z (Worker_PP0_TP0 pid=53096) INFO 11-03 17:29:10 [gpu_model_runner.py:2902] Model loading took 0.7135 GiB and 0.154906 seconds 2025-11-03T17:29:10.5783076Z (Worker_PP0_TP1 pid=53097) INFO 11-03 17:29:10 [gpu_model_runner.py:2902] Model loading took 0.7135 GiB and 0.152714 seconds 2025-11-03T17:29:12.8618674Z (Worker_PP1_TP1 pid=53099) INFO 11-03 17:29:12 [backends.py:594] Using cache directory: /var/lib/jenkins/.cache/vllm/torch_compile_cache/ba6452aee0/rank_3_0/backbone for vLLM's torch.compile 2025-11-03T17:29:12.8625033Z (Worker_PP1_TP1 pid=53099) INFO 11-03 17:29:12 [backends.py:608] Dynamo bytecode transform time: 2.07 s 2025-11-03T17:29:12.9440990Z (Worker_PP1_TP0 pid=53098) INFO 11-03 17:29:12 [backends.py:594] Using cache directory: /var/lib/jenkins/.cache/vllm/torch_compile_cache/ba6452aee0/rank_2_0/backbone for vLLM's torch.compile 2025-11-03T17:29:12.9447831Z (Worker_PP1_TP0 pid=53098) INFO 11-03 17:29:12 [backends.py:608] Dynamo bytecode transform time: 2.16 s 2025-11-03T17:29:13.0329613Z (Worker_PP0_TP0 pid=53096) INFO 11-03 17:29:13 [backends.py:594] Using cache directory: /var/lib/jenkins/.cache/vllm/torch_compile_cache/3ff2b0020c/rank_0_0/backbone for vLLM's torch.compile 2025-11-03T17:29:13.0331050Z (Worker_PP0_TP1 pid=53097) INFO 11-03 17:29:13 [backends.py:594] Using cache directory: /var/lib/jenkins/.cache/vllm/torch_compile_cache/3ff2b0020c/rank_1_0/backbone for vLLM's torch.compile 2025-11-03T17:29:13.0336903Z (Worker_PP0_TP0 pid=53096) INFO 11-03 17:29:13 [backends.py:608] Dynamo bytecode transform time: 2.22 s 2025-11-03T17:29:13.0337698Z (Worker_PP0_TP1 pid=53097) INFO 11-03 17:29:13 [backends.py:608] Dynamo bytecode transform time: 2.22 s 2025-11-03T17:29:13.4476987Z (Worker_PP1_TP1 pid=53099) INFO 11-03 17:29:13 [backends.py:179] Directly load the compiled graph(s) for dynamic shape from the cache, took 0.386 s 2025-11-03T17:29:13.5664940Z (Worker_PP1_TP0 pid=53098) INFO 11-03 17:29:13 [backends.py:179] Directly load the compiled graph(s) for dynamic shape from the cache, took 0.414 s 2025-11-03T17:29:13.6369360Z (Worker_PP0_TP0 pid=53096) INFO 11-03 17:29:13 [backends.py:179] Directly load the compiled graph(s) for dynamic shape from the cache, took 0.391 s 2025-11-03T17:29:13.6623244Z (Worker_PP0_TP1 pid=53097) INFO 11-03 17:29:13 [backends.py:179] Directly load the compiled graph(s) for dynamic shape from the cache, took 0.408 s 2025-11-03T17:29:14.4930331Z (Worker_PP1_TP0 pid=53098) INFO 11-03 17:29:14 [gpu_worker.py:314] Available KV cache memory: 18.99 GiB 2025-11-03T17:29:14.5493177Z (Worker_PP0_TP1 pid=53097) INFO 11-03 17:29:14 [gpu_worker.py:314] Available KV cache memory: 19.02 GiB 2025-11-03T17:29:14.5619760Z (Worker_PP1_TP1 pid=53099) INFO 11-03 17:29:14 [gpu_worker.py:314] Available KV cache memory: 18.99 GiB 2025-11-03T17:29:14.6014145Z (Worker_PP0_TP0 pid=53096) INFO 11-03 17:29:14 [gpu_worker.py:314] Available KV cache memory: 19.02 GiB 2025-11-03T17:29:14.8641405Z (EngineCore_DP0 pid=52977) INFO 11-03 17:29:14 [kv_cache_utils.py:1199] GPU KV cache size: 2,489,088 tokens 2025-11-03T17:29:14.8642948Z (EngineCore_DP0 pid=52977) INFO 11-03 17:29:14 [kv_cache_utils.py:1204] Maximum concurrency for 2,048 tokens per request: 1215.38x 2025-11-03T17:29:14.8643860Z (EngineCore_DP0 pid=52977) INFO 11-03 17:29:14 [kv_cache_utils.py:1199] GPU KV cache size: 2,489,088 tokens 2025-11-03T17:29:14.8644724Z (EngineCore_DP0 pid=52977) INFO 11-03 17:29:14 [kv_cache_utils.py:1204] Maximum concurrency for 2,048 tokens per request: 1215.38x 2025-11-03T17:29:14.8645564Z (EngineCore_DP0 pid=52977) INFO 11-03 17:29:14 [kv_cache_utils.py:1199] GPU KV cache size: 2,489,088 tokens 2025-11-03T17:29:14.8646385Z (EngineCore_DP0 pid=52977) INFO 11-03 17:29:14 [kv_cache_utils.py:1204] Maximum concurrency for 2,048 tokens per request: 1215.38x 2025-11-03T17:29:14.8647185Z (EngineCore_DP0 pid=52977) INFO 11-03 17:29:14 [kv_cache_utils.py:1199] GPU KV cache size: 2,489,088 tokens 2025-11-03T17:29:14.8647981Z (EngineCore_DP0 pid=52977) INFO 11-03 17:29:14 [kv_cache_utils.py:1204] Maximum concurrency for 2,048 tokens per request: 1215.38x 2025-11-03T17:29:14.8715796Z (Worker_PP0_TP0 pid=53096) INFO 11-03 17:29:14 [gpu_worker.py:351] Compile and warming up model for size 8 2025-11-03T17:29:14.8716556Z (Worker_PP1_TP0 pid=53098) INFO 11-03 17:29:14 [gpu_worker.py:351] Compile and warming up model for size 8 2025-11-03T17:29:14.8717532Z (Worker_PP1_TP1 pid=53099) INFO 11-03 17:29:14 [gpu_worker.py:351] Compile and warming up model for size 8 2025-11-03T17:29:14.8718255Z (Worker_PP0_TP1 pid=53097) INFO 11-03 17:29:14 [gpu_worker.py:351] Compile and warming up model for size 8 2025-11-03T17:29:15.5343126Z (Worker_PP0_TP0 pid=53096) INFO 11-03 17:29:15 [backends.py:185] Directly load the compiled graph(s) for shape 8 from the cache, took 0.662 s 2025-11-03T17:29:15.5347180Z (Worker_PP0_TP0 pid=53096) INFO 11-03 17:29:15 [gpu_worker.py:351] Compile and warming up model for size 4 2025-11-03T17:29:15.5375714Z (Worker_PP0_TP1 pid=53097) INFO 11-03 17:29:15 [backends.py:185] Directly load the compiled graph(s) for shape 8 from the cache, took 0.666 s 2025-11-03T17:29:15.5381441Z (Worker_PP0_TP1 pid=53097) INFO 11-03 17:29:15 [gpu_worker.py:351] Compile and warming up model for size 4 2025-11-03T17:29:15.5391153Z (Worker_PP1_TP0 pid=53098) INFO 11-03 17:29:15 [backends.py:185] Directly load the compiled graph(s) for shape 8 from the cache, took 0.667 s 2025-11-03T17:29:15.5429997Z (Worker_PP1_TP1 pid=53099) INFO 11-03 17:29:15 [backends.py:185] Directly load the compiled graph(s) for shape 8 from the cache, took 0.671 s 2025-11-03T17:29:15.5435379Z (Worker_PP1_TP0 pid=53098) INFO 11-03 17:29:15 [gpu_worker.py:351] Compile and warming up model for size 4 2025-11-03T17:29:15.5436160Z (Worker_PP1_TP1 pid=53099) INFO 11-03 17:29:15 [gpu_worker.py:351] Compile and warming up model for size 4 2025-11-03T17:29:15.8785365Z (Worker_PP0_TP0 pid=53096) INFO 11-03 17:29:15 [backends.py:185] Directly load the compiled graph(s) for shape 4 from the cache, took 0.343 s 2025-11-03T17:29:15.8786292Z (Worker_PP0_TP0 pid=53096) INFO 11-03 17:29:15 [monitor.py:33] torch.compile takes 2.22 s in total 2025-11-03T17:29:15.8817677Z (Worker_PP0_TP1 pid=53097) INFO 11-03 17:29:15 [backends.py:185] Directly load the compiled graph(s) for shape 4 from the cache, took 0.343 s 2025-11-03T17:29:15.8818637Z (Worker_PP0_TP1 pid=53097) INFO 11-03 17:29:15 [monitor.py:33] torch.compile takes 2.22 s in total 2025-11-03T17:29:16.2700796Z (Worker_PP1_TP1 pid=53099) INFO 11-03 17:29:16 [backends.py:185] Directly load the compiled graph(s) for shape 4 from the cache, took 0.726 s 2025-11-03T17:29:16.2701743Z (Worker_PP1_TP1 pid=53099) INFO 11-03 17:29:16 [monitor.py:33] torch.compile takes 2.07 s in total 2025-11-03T17:29:16.2734066Z (Worker_PP1_TP0 pid=53098) INFO 11-03 17:29:16 [backends.py:185] Directly load the compiled graph(s) for shape 4 from the cache, took 0.730 s 2025-11-03T17:29:16.2734934Z (Worker_PP1_TP0 pid=53098) INFO 11-03 17:29:16 [monitor.py:33] torch.compile takes 2.16 s in total 2025-11-03T17:29:16.2808330Z (EngineCore_DP0 pid=52977) INFO 11-03 17:29:16 [core.py:240] init engine (profile, create kv cache, warmup model) took 5.70 seconds 2025-11-03T17:29:17.0049472Z (EngineCore_DP0 pid=52977) INFO 11-03 17:29:17 [core.py:178] Batch queue is enabled with size 2 2025-11-03T17:29:17.0078390Z (EngineCore_DP0 pid=52977) INFO 11-03 17:29:17 [vllm.py:354] Cudagraph is disabled under eager mode 2025-11-03T17:29:17.1343734Z (APIServer pid=52855) INFO 11-03 17:29:17 [loggers.py:172] Engine 000: vllm cache_config_info with initialization after num_gpu_blocks is: 155568 2025-11-03T17:29:17.2059700Z (EngineCore_DP0 pid=52977) INFO 11-03 17:29:17 [gc_utils.py:40] GC Debug Config. enabled:False,top_objects:-1 2025-11-03T17:29:17.2280059Z (APIServer pid=52855) INFO 11-03 17:29:17 [api_server.py:1628] Supported tasks: ['generate'] 2025-11-03T17:29:17.2553610Z (APIServer pid=52855) WARNING 11-03 17:29:17 [model.py:1581] Default sampling parameters have been overridden by the model's Hugging Face generation config recommended from the model creator. If this is not intended, please relaunch vLLM instance with `--generation-config vllm`. 2025-11-03T17:29:17.2555779Z (APIServer pid=52855) INFO 11-03 17:29:17 [serving_responses.py:157] Using default chat sampling params from model: {'temperature': 0.6, 'top_p': 0.9} 2025-11-03T17:29:17.2784164Z (APIServer pid=52855) INFO 11-03 17:29:17 [serving_chat.py:159] Using default chat sampling params from model: {'temperature': 0.6, 'top_p': 0.9} 2025-11-03T17:29:17.3057317Z (APIServer pid=52855) INFO 11-03 17:29:17 [serving_completion.py:67] Using default completion sampling params from model: {'temperature': 0.6, 'top_p': 0.9} 2025-11-03T17:29:17.3058764Z (APIServer pid=52855) INFO 11-03 17:29:17 [api_server.py:1939] Starting vLLM API server 0 on http://0.0.0.0:57037 2025-11-03T17:29:17.3059758Z (APIServer pid=52855) INFO 11-03 17:29:17 [launcher.py:38] Available routes are: 2025-11-03T17:29:17.3060667Z (APIServer pid=52855) INFO 11-03 17:29:17 [launcher.py:46] Route: /openapi.json, Methods: GET, HEAD 2025-11-03T17:29:17.3061788Z (APIServer pid=52855) INFO 11-03 17:29:17 [launcher.py:46] Route: /docs, Methods: GET, HEAD 2025-11-03T17:29:17.3062763Z (APIServer pid=52855) INFO 11-03 17:29:17 [launcher.py:46] Route: /docs/oauth2-redirect, Methods: GET, HEAD 2025-11-03T17:29:17.3063602Z (APIServer pid=52855) INFO 11-03 17:29:17 [launcher.py:46] Route: /redoc, Methods: GET, HEAD 2025-11-03T17:29:17.3064472Z (APIServer pid=52855) INFO 11-03 17:29:17 [launcher.py:46] Route: /health, Methods: GET 2025-11-03T17:29:17.3064985Z (APIServer pid=52855) INFO 11-03 17:29:17 [launcher.py:46] Route: /load, Methods: GET 2025-11-03T17:29:17.3065492Z (APIServer pid=52855) INFO 11-03 17:29:17 [launcher.py:46] Route: /ping, Methods: POST 2025-11-03T17:29:17.3065980Z (APIServer pid=52855) INFO 11-03 17:29:17 [launcher.py:46] Route: /ping, Methods: GET 2025-11-03T17:29:17.3066510Z (APIServer pid=52855) INFO 11-03 17:29:17 [launcher.py:46] Route: /tokenize, Methods: POST 2025-11-03T17:29:17.3067045Z (APIServer pid=52855) INFO 11-03 17:29:17 [launcher.py:46] Route: /detokenize, Methods: POST 2025-11-03T17:29:17.3067573Z (APIServer pid=52855) INFO 11-03 17:29:17 [launcher.py:46] Route: /v1/models, Methods: GET 2025-11-03T17:29:17.3068080Z (APIServer pid=52855) INFO 11-03 17:29:17 [launcher.py:46] Route: /version, Methods: GET 2025-11-03T17:29:17.3068604Z (APIServer pid=52855) INFO 11-03 17:29:17 [launcher.py:46] Route: /v1/responses, Methods: POST 2025-11-03T17:29:17.3069167Z (APIServer pid=52855) INFO 11-03 17:29:17 [launcher.py:46] Route: /v1/responses/{response_id}, Methods: GET 2025-11-03T17:29:17.3069785Z (APIServer pid=52855) INFO 11-03 17:29:17 [launcher.py:46] Route: /v1/responses/{response_id}/cancel, Methods: POST 2025-11-03T17:29:17.3070399Z (APIServer pid=52855) INFO 11-03 17:29:17 [launcher.py:46] Route: /v1/chat/completions, Methods: POST 2025-11-03T17:29:17.3070964Z (APIServer pid=52855) INFO 11-03 17:29:17 [launcher.py:46] Route: /v1/completions, Methods: POST 2025-11-03T17:29:17.3071509Z (APIServer pid=52855) INFO 11-03 17:29:17 [launcher.py:46] Route: /v1/embeddings, Methods: POST 2025-11-03T17:29:17.3072028Z (APIServer pid=52855) INFO 11-03 17:29:17 [launcher.py:46] Route: /pooling, Methods: POST 2025-11-03T17:29:17.3072548Z (APIServer pid=52855) INFO 11-03 17:29:17 [launcher.py:46] Route: /classify, Methods: POST 2025-11-03T17:29:17.3073055Z (APIServer pid=52855) INFO 11-03 17:29:17 [launcher.py:46] Route: /score, Methods: POST 2025-11-03T17:29:17.3073578Z (APIServer pid=52855) INFO 11-03 17:29:17 [launcher.py:46] Route: /v1/score, Methods: POST 2025-11-03T17:29:17.3074136Z (APIServer pid=52855) INFO 11-03 17:29:17 [launcher.py:46] Route: /v1/audio/transcriptions, Methods: POST 2025-11-03T17:29:17.3074812Z (APIServer pid=52855) INFO 11-03 17:29:17 [launcher.py:46] Route: /v1/audio/translations, Methods: POST 2025-11-03T17:29:17.3075348Z (APIServer pid=52855) INFO 11-03 17:29:17 [launcher.py:46] Route: /rerank, Methods: POST 2025-11-03T17:29:17.3075873Z (APIServer pid=52855) INFO 11-03 17:29:17 [launcher.py:46] Route: /v1/rerank, Methods: POST 2025-11-03T17:29:17.3076393Z (APIServer pid=52855) INFO 11-03 17:29:17 [launcher.py:46] Route: /v2/rerank, Methods: POST 2025-11-03T17:29:17.3076928Z (APIServer pid=52855) INFO 11-03 17:29:17 [launcher.py:46] Route: /scale_elastic_ep, Methods: POST 2025-11-03T17:29:17.3077487Z (APIServer pid=52855) INFO 11-03 17:29:17 [launcher.py:46] Route: /is_scaling_elastic_ep, Methods: POST 2025-11-03T17:29:17.3078034Z (APIServer pid=52855) INFO 11-03 17:29:17 [launcher.py:46] Route: /invocations, Methods: POST 2025-11-03T17:29:17.3078549Z (APIServer pid=52855) INFO 11-03 17:29:17 [launcher.py:46] Route: /metrics, Methods: GET 2025-11-03T17:29:17.3217410Z (APIServer pid=52855) INFO: Started server process [52855] 2025-11-03T17:29:17.3218060Z (APIServer pid=52855) INFO: Waiting for application startup. 2025-11-03T17:29:17.5363196Z (APIServer pid=52855) INFO: Application startup complete. 2025-11-03T17:29:17.9883384Z (APIServer pid=52855) INFO: 127.0.0.1:38544 - "GET /health HTTP/1.1" 200 OK 2025-11-03T17:29:18.2334480Z (APIServer pid=52855) INFO: 127.0.0.1:38558 - "GET /v1/models HTTP/1.1" 200 OK 2025-11-03T17:29:18.2637227Z (Worker_PP0_TP0 pid=53096) /opt/conda/envs/py_3.12/lib/python3.12/site-packages/vllm/distributed/parallel_state.py:530: UserWarning: The given buffer is not writable, and PyTorch does not support non-writable tensors. This means you can write to the underlying (supposedly non-writable) buffer using the tensor. You may want to copy the buffer to protect its data or make it writable before converting it to a tensor. This type of warning will be suppressed for the rest of this program. (Triggered internally at /var/lib/jenkins/workspace/torch/csrc/utils/tensor_new.cpp:1581.) 2025-11-03T17:29:18.2640284Z (Worker_PP0_TP0 pid=53096) object_tensor = torch.frombuffer(pickle.dumps(obj), dtype=torch.uint8) 2025-11-03T17:29:18.2643775Z (Worker_PP0_TP1 pid=53097) /opt/conda/envs/py_3.12/lib/python3.12/site-packages/vllm/distributed/parallel_state.py:530: UserWarning: The given buffer is not writable, and PyTorch does not support non-writable tensors. This means you can write to the underlying (supposedly non-writable) buffer using the tensor. You may want to copy the buffer to protect its data or make it writable before converting it to a tensor. This type of warning will be suppressed for the rest of this program. (Triggered internally at /var/lib/jenkins/workspace/torch/csrc/utils/tensor_new.cpp:1581.) 2025-11-03T17:29:18.2645919Z (Worker_PP0_TP1 pid=53097) object_tensor = torch.frombuffer(pickle.dumps(obj), dtype=torch.uint8) 2025-11-03T17:29:18.2646856Z [rank0]:[W1103 17:29:18.312816517 ProcessGroupNCCL.cpp:4063] Warning: An unbatched P2P op (send/recv) was called on this ProcessGroup with size 2. In lazy initialization mode, this will result in a new 2-rank NCCL communicator to be created. (function operator()) 2025-11-03T17:29:18.2648148Z [rank1]:[W1103 17:29:18.313015303 ProcessGroupNCCL.cpp:4063] Warning: An unbatched P2P op (send/recv) was called on this ProcessGroup with size 2. In lazy initialization mode, this will result in a new 2-rank NCCL communicator to be created. (function operator()) 2025-11-03T17:29:18.2649535Z [rank2]:[W1103 17:29:18.313038364 ProcessGroupNCCL.cpp:4063] Warning: An unbatched P2P op (send/recv) was called on this ProcessGroup with size 2. In lazy initialization mode, this will result in a new 2-rank NCCL communicator to be created. (function operator()) 2025-11-03T17:29:18.2650989Z [rank3]:[W1103 17:29:18.313147573 ProcessGroupNCCL.cpp:4063] Warning: An unbatched P2P op (send/recv) was called on this ProcessGroup with size 2. In lazy initialization mode, this will result in a new 2-rank NCCL communicator to be created. (function operator()) 2025-11-03T17:29:18.3894166Z (APIServer pid=52855) INFO: 127.0.0.1:38558 - "POST /v1/completions HTTP/1.1" 200 OK 2025-11-03T17:29:18.4560895Z (APIServer pid=52855) INFO: 127.0.0.1:38558 - "POST /v1/completions HTTP/1.1" 200 OK 2025-11-03T17:29:18.5881691Z (APIServer pid=52855) INFO: 127.0.0.1:38558 - "POST /v1/completions HTTP/1.1" 200 OK 2025-11-03T17:29:18.6797586Z (APIServer pid=52855) INFO: 127.0.0.1:38558 - "POST /v1/completions HTTP/1.1" 200 OK 2025-11-03T17:29:18.7568619Z (APIServer pid=52855) INFO: 127.0.0.1:38558 - "POST /v1/completions HTTP/1.1" 200 OK 2025-11-03T17:29:18.7644768Z (APIServer pid=52855) INFO: 127.0.0.1:38558 - "POST /v1/completions HTTP/1.1" 200 OK 2025-11-03T17:29:18.8376608Z (APIServer pid=52855) INFO 11-03 17:29:18 [launcher.py:110] Shutting down FastAPI HTTP server. 2025-11-03T17:29:18.8379354Z (Worker_PP0_TP0 pid=53096) INFO 11-03 17:29:18 [multiproc_executor.py:583] Parent process exited, terminating worker 2025-11-03T17:29:18.8380350Z (Worker_PP0_TP0 pid=53096) INFO 11-03 17:29:18 [multiproc_executor.py:624] WorkerProc shutting down. 2025-11-03T17:29:18.8383822Z (Worker_PP0_TP1 pid=53097) INFO 11-03 17:29:18 [multiproc_executor.py:583] Parent process exited, terminating worker 2025-11-03T17:29:18.8385707Z (Worker_PP0_TP1 pid=53097) INFO 11-03 17:29:18 [multiproc_executor.py:624] WorkerProc shutting down. 2025-11-03T17:29:18.8388688Z (Worker_PP1_TP0 pid=53098) INFO 11-03 17:29:18 [multiproc_executor.py:583] Parent process exited, terminating worker 2025-11-03T17:29:18.8390008Z (Worker_PP1_TP0 pid=53098) INFO 11-03 17:29:18 [multiproc_executor.py:624] WorkerProc shutting down. 2025-11-03T17:29:18.8391526Z (Worker_PP1_TP1 pid=53099) INFO 11-03 17:29:18 [multiproc_executor.py:583] Parent process exited, terminating worker 2025-11-03T17:29:21.5552277Z (APIServer pid=52855) INFO: Shutting down 2025-11-03T17:29:21.6555784Z (APIServer pid=52855) INFO: Waiting for application shutdown. 2025-11-03T17:29:21.6557067Z (APIServer pid=52855) INFO: Application shutdown complete. 2025-11-03T17:29:22.0220231Z INFO 11-03 17:29:22 [model.py:653] Resolved architecture: LlamaForCausalLM 2025-11-03T17:29:22.0220811Z WARNING 11-03 17:29:22 [model.py:1942] Casting torch.bfloat16 to torch.float16. 2025-11-03T17:29:22.0221285Z INFO 11-03 17:29:22 [model.py:1714] Using max model len 2048 2025-11-03T17:29:22.0222584Z Launching RemoteOpenAIServer with: vllm serve meta-llama/Llama-3.2-1B-Instruct --dtype float16 --max-model-len 2048 --max-num-seqs 8 --enable-chunked-prefill --enforce-eager --tokenizer-mode auto --tensor-parallel-size 2 --distributed-executor-backend mp --load-format dummy --port 37967 --seed 0 2025-11-03T17:29:22.5373289Z /opt/conda/envs/py_3.12/lib/python3.12/site-packages/torch/cuda/__init__.py:65: FutureWarning: The pynvml package is deprecated. Please install nvidia-ml-py instead. If you did not install pynvml directly, please report this to the maintainers of the package that installed pynvml for you. 2025-11-03T17:29:22.5374500Z import pynvml # type: ignore[import] 2025-11-03T17:29:24.7883751Z INFO 11-03 17:29:24 [__init__.py:224] Automatically detected platform cuda. 2025-11-03T17:29:28.4687213Z (APIServer pid=53496) INFO 11-03 17:29:28 [api_server.py:1870] vLLM API server version 0.11.0rc2.dev389+ge51928192.d20251103 2025-11-03T17:29:28.4712258Z (APIServer pid=53496) INFO 11-03 17:29:28 [utils.py:239] non-default args: {'model_tag': 'meta-llama/Llama-3.2-1B-Instruct', 'port': 37967, 'model': 'meta-llama/Llama-3.2-1B-Instruct', 'dtype': 'float16', 'seed': 0, 'max_model_len': 2048, 'enforce_eager': True, 'load_format': 'dummy', 'distributed_executor_backend': 'mp', 'tensor_parallel_size': 2, 'max_num_seqs': 8, 'enable_chunked_prefill': True} 2025-11-03T17:29:28.6692379Z (APIServer pid=53496) INFO 11-03 17:29:28 [model.py:653] Resolved architecture: LlamaForCausalLM 2025-11-03T17:29:28.6693738Z (APIServer pid=53496) `torch_dtype` is deprecated! Use `dtype` instead! 2025-11-03T17:29:28.6695264Z (APIServer pid=53496) WARNING 11-03 17:29:28 [model.py:1942] Casting torch.bfloat16 to torch.float16. 2025-11-03T17:29:28.6695905Z (APIServer pid=53496) INFO 11-03 17:29:28 [model.py:1714] Using max model len 2048 2025-11-03T17:29:28.8396294Z (APIServer pid=53496) INFO 11-03 17:29:28 [scheduler.py:225] Chunked prefill is enabled with max_num_batched_tokens=2048. 2025-11-03T17:29:28.8402879Z (APIServer pid=53496) INFO 11-03 17:29:28 [vllm.py:354] Cudagraph is disabled under eager mode 2025-11-03T17:29:29.8394447Z /opt/conda/envs/py_3.12/lib/python3.12/site-packages/torch/cuda/__init__.py:65: FutureWarning: The pynvml package is deprecated. Please install nvidia-ml-py instead. If you did not install pynvml directly, please report this to the maintainers of the package that installed pynvml for you. 2025-11-03T17:29:29.8396310Z import pynvml # type: ignore[import] 2025-11-03T17:29:32.1832047Z INFO 11-03 17:29:32 [__init__.py:224] Automatically detected platform cuda. 2025-11-03T17:29:35.8916859Z (EngineCore_DP0 pid=53618) INFO 11-03 17:29:35 [core.py:727] Waiting for init message from front-end. 2025-11-03T17:29:35.9188814Z (EngineCore_DP0 pid=53618) INFO 11-03 17:29:35 [core.py:94] Initializing a V1 LLM engine (v0.11.0rc2.dev389+ge51928192.d20251103) with config: model='meta-llama/Llama-3.2-1B-Instruct', speculative_config=None, tokenizer='meta-llama/Llama-3.2-1B-Instruct', skip_tokenizer_init=False, tokenizer_mode=auto, revision=None, tokenizer_revision=None, trust_remote_code=False, dtype=torch.float16, max_seq_len=2048, download_dir=None, load_format=dummy, tensor_parallel_size=2, pipeline_parallel_size=1, data_parallel_size=1, disable_custom_all_reduce=False, quantization=None, enforce_eager=True, kv_cache_dtype=auto, device_config=cuda, structured_outputs_config=StructuredOutputsConfig(backend='auto', disable_fallback=False, disable_any_whitespace=False, disable_additional_properties=False, reasoning_parser=''), observability_config=ObservabilityConfig(show_hidden_metrics_for_version=None, otlp_traces_endpoint=None, collect_detailed_traces=None), seed=0, served_model_name=meta-llama/Llama-3.2-1B-Instruct, enable_prefix_caching=True, chunked_prefill_enabled=True, pooler_config=None, compilation_config={'level': 0, 'debug_dump_path': None, 'cache_dir': '', 'backend': '', 'custom_ops': [], 'splitting_ops': None, 'use_inductor': True, 'compile_sizes': [], 'inductor_compile_config': {'enable_auto_functionalized_v2': False}, 'inductor_passes': {}, 'cudagraph_mode': , 'use_cudagraph': False, 'cudagraph_num_of_warmups': 0, 'cudagraph_capture_sizes': [], 'cudagraph_copy_inputs': False, 'full_cuda_graph': False, 'use_inductor_graph_partition': False, 'pass_config': {}, 'max_capture_size': 0, 'local_cache_dir': None} 2025-11-03T17:29:35.9196072Z (EngineCore_DP0 pid=53618) WARNING 11-03 17:29:35 [multiproc_executor.py:744] Reducing Torch parallelism from 24 threads to 1 to avoid unnecessary CPU contention. Set OMP_NUM_THREADS in the external environment to tune this value as needed. 2025-11-03T17:29:35.9197670Z (EngineCore_DP0 pid=53618) INFO 11-03 17:29:35 [shm_broadcast.py:302] vLLM message queue communication handle: Handle(local_reader_ranks=[0, 1], buffer_handle=(2, 16777216, 10, 'psm_f625952a'), local_subscribe_addr='ipc:///tmp/d3a5eaa2-77d0-418f-b2fd-2adaf3901bc5', remote_subscribe_addr=None, remote_addr_ipv6=False) 2025-11-03T17:29:36.3969652Z /opt/conda/envs/py_3.12/lib/python3.12/site-packages/torch/cuda/__init__.py:65: FutureWarning: The pynvml package is deprecated. Please install nvidia-ml-py instead. If you did not install pynvml directly, please report this to the maintainers of the package that installed pynvml for you. 2025-11-03T17:29:36.3971078Z import pynvml # type: ignore[import] 2025-11-03T17:29:36.4092341Z /opt/conda/envs/py_3.12/lib/python3.12/site-packages/torch/cuda/__init__.py:65: FutureWarning: The pynvml package is deprecated. Please install nvidia-ml-py instead. If you did not install pynvml directly, please report this to the maintainers of the package that installed pynvml for you. 2025-11-03T17:29:36.4093469Z import pynvml # type: ignore[import] 2025-11-03T17:29:38.7417571Z INFO 11-03 17:29:38 [__init__.py:224] Automatically detected platform cuda. 2025-11-03T17:29:38.7604055Z INFO 11-03 17:29:38 [__init__.py:224] Automatically detected platform cuda. 2025-11-03T17:29:43.5029891Z INFO 11-03 17:29:43 [shm_broadcast.py:302] vLLM message queue communication handle: Handle(local_reader_ranks=[0], buffer_handle=(1, 10485760, 10, 'psm_4c055624'), local_subscribe_addr='ipc:///tmp/25debc88-545b-43bb-b48f-3a539667456d', remote_subscribe_addr=None, remote_addr_ipv6=False) 2025-11-03T17:29:43.5425824Z INFO 11-03 17:29:43 [shm_broadcast.py:302] vLLM message queue communication handle: Handle(local_reader_ranks=[0], buffer_handle=(1, 10485760, 10, 'psm_716a71f1'), local_subscribe_addr='ipc:///tmp/5227018f-c4e2-4d04-99dc-c3762070e924', remote_subscribe_addr=None, remote_addr_ipv6=False) 2025-11-03T17:29:43.7158088Z [Gloo] Rank 1 is connected to 1 peer ranks. Expected number of connected peer ranks is : 1 2025-11-03T17:29:43.7158611Z [Gloo] Rank 0 is connected to 1 peer ranks. Expected number of connected peer ranks is : 1 2025-11-03T17:29:43.7200525Z [Gloo] Rank 0 is connected to 1 peer ranks. Expected number of connected peer ranks is : 1 2025-11-03T17:29:43.7201166Z [Gloo] Rank 1 is connected to 1 peer ranks. Expected number of connected peer ranks is : 1 2025-11-03T17:29:43.7243240Z INFO 11-03 17:29:43 [__init__.py:1146] Found nccl from library libnccl.so.2 2025-11-03T17:29:43.7244900Z INFO 11-03 17:29:43 [pynccl.py:109] vLLM is using nccl==2.27.5 2025-11-03T17:29:43.7246117Z INFO 11-03 17:29:43 [__init__.py:1146] Found nccl from library libnccl.so.2 2025-11-03T17:29:43.7248550Z INFO 11-03 17:29:43 [pynccl.py:109] vLLM is using nccl==2.27.5 2025-11-03T17:29:43.8972099Z WARNING 11-03 17:29:43 [symm_mem.py:60] SymmMemCommunicator: Device capability 8.9 not supported, communicator is not available. 2025-11-03T17:29:43.8973758Z WARNING 11-03 17:29:43 [symm_mem.py:60] SymmMemCommunicator: Device capability 8.9 not supported, communicator is not available. 2025-11-03T17:29:43.8996196Z INFO 11-03 17:29:43 [custom_all_reduce.py:37] Skipping P2P check and trusting the driver's P2P report. 2025-11-03T17:29:43.8996808Z INFO 11-03 17:29:43 [custom_all_reduce.py:37] Skipping P2P check and trusting the driver's P2P report. 2025-11-03T17:29:43.8998733Z WARNING 11-03 17:29:43 [custom_all_reduce.py:162] Custom allreduce is disabled because your platform lacks GPU P2P capability or P2P test failed. To silence this warning, specify disable_custom_all_reduce=True explicitly. 2025-11-03T17:29:43.9001813Z WARNING 11-03 17:29:43 [custom_all_reduce.py:162] Custom allreduce is disabled because your platform lacks GPU P2P capability or P2P test failed. To silence this warning, specify disable_custom_all_reduce=True explicitly. 2025-11-03T17:29:43.9825066Z INFO 11-03 17:29:43 [shm_broadcast.py:302] vLLM message queue communication handle: Handle(local_reader_ranks=[1], buffer_handle=(1, 4194304, 6, 'psm_a3a03109'), local_subscribe_addr='ipc:///tmp/5b40e056-2b3a-4386-b3ae-63bfe264800b', remote_subscribe_addr=None, remote_addr_ipv6=False) 2025-11-03T17:29:43.9850111Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:29:43.9851259Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:29:43.9858110Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:29:43.9865961Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:29:44.0620911Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:29:44.0628943Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:29:44.0641615Z [Gloo] Rank 1 is connected to 1 peer ranks. Expected number of connected peer ranks is : 1 2025-11-03T17:29:44.0642826Z [Gloo] Rank 0 is connected to 1 peer ranks. Expected number of connected peer ranks is : 1 2025-11-03T17:29:44.0644007Z INFO 11-03 17:29:44 [__init__.py:1146] Found nccl from library libnccl.so.2 2025-11-03T17:29:44.0644890Z INFO 11-03 17:29:44 [__init__.py:1146] Found nccl from library libnccl.so.2 2025-11-03T17:29:44.0646199Z INFO 11-03 17:29:44 [pynccl.py:109] vLLM is using nccl==2.27.5 2025-11-03T17:29:44.0646990Z INFO 11-03 17:29:44 [pynccl.py:109] vLLM is using nccl==2.27.5 2025-11-03T17:29:44.0977238Z INFO 11-03 17:29:44 [parallel_state.py:1231] rank 1 in world size 2 is assigned as DP rank 0, PP rank 0, TP rank 1, EP rank 1 2025-11-03T17:29:44.0978232Z INFO 11-03 17:29:44 [parallel_state.py:1231] rank 0 in world size 2 is assigned as DP rank 0, PP rank 0, TP rank 0, EP rank 0 2025-11-03T17:29:44.3433236Z INFO 11-03 17:29:44 [topk_topp_sampler.py:59] Using FlashInfer for top-p & top-k sampling. 2025-11-03T17:29:44.3516546Z INFO 11-03 17:29:44 [topk_topp_sampler.py:59] Using FlashInfer for top-p & top-k sampling. 2025-11-03T17:29:44.3762047Z (Worker_TP0 pid=53737) INFO 11-03 17:29:44 [gpu_model_runner.py:2840] Starting to load model meta-llama/Llama-3.2-1B-Instruct... 2025-11-03T17:29:44.3827742Z (Worker_TP1 pid=53738) INFO 11-03 17:29:44 [gpu_model_runner.py:2840] Starting to load model meta-llama/Llama-3.2-1B-Instruct... 2025-11-03T17:29:44.5874391Z (Worker_TP0 pid=53737) INFO 11-03 17:29:44 [gpu_model_runner.py:2870] Loading model from scratch... 2025-11-03T17:29:44.5929989Z (Worker_TP1 pid=53738) INFO 11-03 17:29:44 [gpu_model_runner.py:2870] Loading model from scratch... 2025-11-03T17:29:44.6778002Z (Worker_TP0 pid=53737) INFO 11-03 17:29:44 [cuda.py:404] Using Flash Attention backend on V1 engine. 2025-11-03T17:29:44.6822978Z (Worker_TP1 pid=53738) INFO 11-03 17:29:44 [cuda.py:404] Using Flash Attention backend on V1 engine. 2025-11-03T17:29:45.1187023Z (Worker_TP1 pid=53738) INFO 11-03 17:29:45 [gpu_model_runner.py:2902] Model loading took 1.1667 GiB and 0.118401 seconds 2025-11-03T17:29:45.1259418Z (Worker_TP0 pid=53737) INFO 11-03 17:29:45 [gpu_model_runner.py:2902] Model loading took 1.1667 GiB and 0.120759 seconds 2025-11-03T17:29:47.0274218Z (Worker_TP1 pid=53738) INFO 11-03 17:29:47 [gpu_worker.py:314] Available KV cache memory: 18.53 GiB 2025-11-03T17:29:47.0401053Z (Worker_TP0 pid=53737) INFO 11-03 17:29:47 [gpu_worker.py:314] Available KV cache memory: 18.53 GiB 2025-11-03T17:29:47.2477840Z (EngineCore_DP0 pid=53618) INFO 11-03 17:29:47 [kv_cache_utils.py:1199] GPU KV cache size: 1,214,464 tokens 2025-11-03T17:29:47.2478761Z (EngineCore_DP0 pid=53618) INFO 11-03 17:29:47 [kv_cache_utils.py:1204] Maximum concurrency for 2,048 tokens per request: 593.00x 2025-11-03T17:29:47.2479630Z (EngineCore_DP0 pid=53618) INFO 11-03 17:29:47 [kv_cache_utils.py:1199] GPU KV cache size: 1,214,464 tokens 2025-11-03T17:29:47.2480355Z (EngineCore_DP0 pid=53618) INFO 11-03 17:29:47 [kv_cache_utils.py:1204] Maximum concurrency for 2,048 tokens per request: 593.00x 2025-11-03T17:29:47.3694737Z (EngineCore_DP0 pid=53618) INFO 11-03 17:29:47 [core.py:240] init engine (profile, create kv cache, warmup model) took 2.16 seconds 2025-11-03T17:29:48.0661529Z (EngineCore_DP0 pid=53618) INFO 11-03 17:29:48 [vllm.py:354] Cudagraph is disabled under eager mode 2025-11-03T17:29:48.2508117Z (APIServer pid=53496) INFO 11-03 17:29:48 [loggers.py:172] Engine 000: vllm cache_config_info with initialization after num_gpu_blocks is: 75904 2025-11-03T17:29:48.2656688Z (EngineCore_DP0 pid=53618) INFO 11-03 17:29:48 [gc_utils.py:40] GC Debug Config. enabled:False,top_objects:-1 2025-11-03T17:29:48.2866225Z (APIServer pid=53496) INFO 11-03 17:29:48 [api_server.py:1628] Supported tasks: ['generate'] 2025-11-03T17:29:48.3109347Z (APIServer pid=53496) WARNING 11-03 17:29:48 [model.py:1581] Default sampling parameters have been overridden by the model's Hugging Face generation config recommended from the model creator. If this is not intended, please relaunch vLLM instance with `--generation-config vllm`. 2025-11-03T17:29:48.3112312Z (APIServer pid=53496) INFO 11-03 17:29:48 [serving_responses.py:157] Using default chat sampling params from model: {'temperature': 0.6, 'top_p': 0.9} 2025-11-03T17:29:48.3343608Z (APIServer pid=53496) INFO 11-03 17:29:48 [serving_chat.py:159] Using default chat sampling params from model: {'temperature': 0.6, 'top_p': 0.9} 2025-11-03T17:29:48.3549919Z (APIServer pid=53496) INFO 11-03 17:29:48 [serving_completion.py:67] Using default completion sampling params from model: {'temperature': 0.6, 'top_p': 0.9} 2025-11-03T17:29:48.3552208Z (APIServer pid=53496) INFO 11-03 17:29:48 [api_server.py:1939] Starting vLLM API server 0 on http://0.0.0.0:37967 2025-11-03T17:29:48.3553619Z (APIServer pid=53496) INFO 11-03 17:29:48 [launcher.py:38] Available routes are: 2025-11-03T17:29:48.3554889Z (APIServer pid=53496) INFO 11-03 17:29:48 [launcher.py:46] Route: /openapi.json, Methods: GET, HEAD 2025-11-03T17:29:48.3556194Z (APIServer pid=53496) INFO 11-03 17:29:48 [launcher.py:46] Route: /docs, Methods: GET, HEAD 2025-11-03T17:29:48.3557549Z (APIServer pid=53496) INFO 11-03 17:29:48 [launcher.py:46] Route: /docs/oauth2-redirect, Methods: GET, HEAD 2025-11-03T17:29:48.3558862Z (APIServer pid=53496) INFO 11-03 17:29:48 [launcher.py:46] Route: /redoc, Methods: GET, HEAD 2025-11-03T17:29:48.3560085Z (APIServer pid=53496) INFO 11-03 17:29:48 [launcher.py:46] Route: /health, Methods: GET 2025-11-03T17:29:48.3561282Z (APIServer pid=53496) INFO 11-03 17:29:48 [launcher.py:46] Route: /load, Methods: GET 2025-11-03T17:29:48.3561845Z (APIServer pid=53496) INFO 11-03 17:29:48 [launcher.py:46] Route: /ping, Methods: POST 2025-11-03T17:29:48.3562351Z (APIServer pid=53496) INFO 11-03 17:29:48 [launcher.py:46] Route: /ping, Methods: GET 2025-11-03T17:29:48.3562852Z (APIServer pid=53496) INFO 11-03 17:29:48 [launcher.py:46] Route: /tokenize, Methods: POST 2025-11-03T17:29:48.3563391Z (APIServer pid=53496) INFO 11-03 17:29:48 [launcher.py:46] Route: /detokenize, Methods: POST 2025-11-03T17:29:48.3563920Z (APIServer pid=53496) INFO 11-03 17:29:48 [launcher.py:46] Route: /v1/models, Methods: GET 2025-11-03T17:29:48.3564426Z (APIServer pid=53496) INFO 11-03 17:29:48 [launcher.py:46] Route: /version, Methods: GET 2025-11-03T17:29:48.3564947Z (APIServer pid=53496) INFO 11-03 17:29:48 [launcher.py:46] Route: /v1/responses, Methods: POST 2025-11-03T17:29:48.3565509Z (APIServer pid=53496) INFO 11-03 17:29:48 [launcher.py:46] Route: /v1/responses/{response_id}, Methods: GET 2025-11-03T17:29:48.3566135Z (APIServer pid=53496) INFO 11-03 17:29:48 [launcher.py:46] Route: /v1/responses/{response_id}/cancel, Methods: POST 2025-11-03T17:29:48.3566739Z (APIServer pid=53496) INFO 11-03 17:29:48 [launcher.py:46] Route: /v1/chat/completions, Methods: POST 2025-11-03T17:29:48.3567467Z (APIServer pid=53496) INFO 11-03 17:29:48 [launcher.py:46] Route: /v1/completions, Methods: POST 2025-11-03T17:29:48.3568036Z (APIServer pid=53496) INFO 11-03 17:29:48 [launcher.py:46] Route: /v1/embeddings, Methods: POST 2025-11-03T17:29:48.3568647Z (APIServer pid=53496) INFO 11-03 17:29:48 [launcher.py:46] Route: /pooling, Methods: POST 2025-11-03T17:29:48.3569174Z (APIServer pid=53496) INFO 11-03 17:29:48 [launcher.py:46] Route: /classify, Methods: POST 2025-11-03T17:29:48.3569690Z (APIServer pid=53496) INFO 11-03 17:29:48 [launcher.py:46] Route: /score, Methods: POST 2025-11-03T17:29:48.3570200Z (APIServer pid=53496) INFO 11-03 17:29:48 [launcher.py:46] Route: /v1/score, Methods: POST 2025-11-03T17:29:48.3570781Z (APIServer pid=53496) INFO 11-03 17:29:48 [launcher.py:46] Route: /v1/audio/transcriptions, Methods: POST 2025-11-03T17:29:48.3571368Z (APIServer pid=53496) INFO 11-03 17:29:48 [launcher.py:46] Route: /v1/audio/translations, Methods: POST 2025-11-03T17:29:48.3572050Z (APIServer pid=53496) INFO 11-03 17:29:48 [launcher.py:46] Route: /rerank, Methods: POST 2025-11-03T17:29:48.3572588Z (APIServer pid=53496) INFO 11-03 17:29:48 [launcher.py:46] Route: /v1/rerank, Methods: POST 2025-11-03T17:29:48.3573109Z (APIServer pid=53496) INFO 11-03 17:29:48 [launcher.py:46] Route: /v2/rerank, Methods: POST 2025-11-03T17:29:48.3573714Z (APIServer pid=53496) INFO 11-03 17:29:48 [launcher.py:46] Route: /scale_elastic_ep, Methods: POST 2025-11-03T17:29:48.3574275Z (APIServer pid=53496) INFO 11-03 17:29:48 [launcher.py:46] Route: /is_scaling_elastic_ep, Methods: POST 2025-11-03T17:29:48.3574823Z (APIServer pid=53496) INFO 11-03 17:29:48 [launcher.py:46] Route: /invocations, Methods: POST 2025-11-03T17:29:48.3575340Z (APIServer pid=53496) INFO 11-03 17:29:48 [launcher.py:46] Route: /metrics, Methods: GET 2025-11-03T17:29:48.3710164Z (APIServer pid=53496) INFO: Started server process [53496] 2025-11-03T17:29:48.3711495Z (APIServer pid=53496) INFO: Waiting for application startup. 2025-11-03T17:29:48.5812016Z (APIServer pid=53496) INFO: Application startup complete. 2025-11-03T17:29:48.5894806Z (APIServer pid=53496) INFO: 127.0.0.1:49116 - "GET /health HTTP/1.1" 200 OK 2025-11-03T17:29:48.5988615Z (APIServer pid=53496) INFO: 127.0.0.1:49126 - "GET /v1/models HTTP/1.1" 200 OK 2025-11-03T17:29:48.7734996Z (APIServer pid=53496) INFO: 127.0.0.1:49126 - "POST /v1/completions HTTP/1.1" 200 OK 2025-11-03T17:29:48.8308561Z (APIServer pid=53496) INFO: 127.0.0.1:49126 - "POST /v1/completions HTTP/1.1" 200 OK 2025-11-03T17:29:48.9282915Z (APIServer pid=53496) INFO: 127.0.0.1:49126 - "POST /v1/completions HTTP/1.1" 200 OK 2025-11-03T17:29:49.0041932Z (APIServer pid=53496) INFO: 127.0.0.1:49126 - "POST /v1/completions HTTP/1.1" 200 OK 2025-11-03T17:29:49.0762616Z (APIServer pid=53496) INFO: 127.0.0.1:49126 - "POST /v1/completions HTTP/1.1" 200 OK 2025-11-03T17:29:49.0839715Z (APIServer pid=53496) INFO: 127.0.0.1:49126 - "POST /v1/completions HTTP/1.1" 200 OK 2025-11-03T17:29:49.1511973Z (APIServer pid=53496) INFO 11-03 17:29:49 [launcher.py:110] Shutting down FastAPI HTTP server. 2025-11-03T17:29:49.1515108Z (Worker_TP0 pid=53737) INFO 11-03 17:29:49 [multiproc_executor.py:583] Parent process exited, terminating worker 2025-11-03T17:29:49.1516215Z (Worker_TP0 pid=53737) INFO 11-03 17:29:49 [multiproc_executor.py:624] WorkerProc shutting down. 2025-11-03T17:29:49.1522922Z (Worker_TP1 pid=53738) INFO 11-03 17:29:49 [multiproc_executor.py:583] Parent process exited, terminating worker 2025-11-03T17:29:49.1524234Z (Worker_TP1 pid=53738) INFO 11-03 17:29:49 [multiproc_executor.py:624] WorkerProc shutting down. 2025-11-03T17:29:51.3666166Z (APIServer pid=53496) INFO: Shutting down 2025-11-03T17:29:51.4670886Z (APIServer pid=53496) INFO: Waiting for application shutdown. 2025-11-03T17:29:51.4672205Z (APIServer pid=53496) INFO: Application shutdown complete. 2025-11-03T17:29:52.0500943Z PASSED 2025-11-03T17:29:52.0637365Z distributed/test_sequence_parallel.py::test_tp_sp_generation[meta-llama/Llama-3.2-1B-Instruct-parallel_setup15-ray-auto-test_options15] Fork a new process to run a test 53941 2025-11-03T17:29:52.0655515Z Fork a new process to run a test 0 2025-11-03T17:29:52.7104949Z INFO 11-03 17:29:52 [model.py:653] Resolved architecture: LlamaForCausalLM 2025-11-03T17:29:52.7106024Z `torch_dtype` is deprecated! Use `dtype` instead! 2025-11-03T17:29:52.7107441Z WARNING 11-03 17:29:52 [model.py:1942] Casting torch.bfloat16 to torch.float16. 2025-11-03T17:29:52.7108270Z INFO 11-03 17:29:52 [model.py:1714] Using max model len 2048 2025-11-03T17:29:52.7112255Z Launching RemoteOpenAIServer with: vllm serve meta-llama/Llama-3.2-1B-Instruct --dtype float16 --max-model-len 2048 --max-num-seqs 8 --enable-chunked-prefill --enforce-eager --tokenizer-mode auto --tensor-parallel-size 2 --pipeline-parallel-size 2 --distributed-executor-backend ray --compilation_config {"level": 3, "custom_ops": ["+rms_norm"], "compile_sizes": [4, 8], "pass_config": {"enable_sequence_parallelism": true, "enable_fusion": false, "enable_noop": true}} --load-format dummy --port 42435 --seed 0 2025-11-03T17:29:53.2212635Z /opt/conda/envs/py_3.12/lib/python3.12/site-packages/torch/cuda/__init__.py:65: FutureWarning: The pynvml package is deprecated. Please install nvidia-ml-py instead. If you did not install pynvml directly, please report this to the maintainers of the package that installed pynvml for you. 2025-11-03T17:29:53.2213879Z import pynvml # type: ignore[import] 2025-11-03T17:29:55.4608950Z INFO 11-03 17:29:55 [__init__.py:224] Automatically detected platform cuda. 2025-11-03T17:29:59.1302473Z (APIServer pid=53990) INFO 11-03 17:29:59 [api_server.py:1870] vLLM API server version 0.11.0rc2.dev389+ge51928192.d20251103 2025-11-03T17:29:59.1330810Z (APIServer pid=53990) INFO 11-03 17:29:59 [utils.py:239] non-default args: {'model_tag': 'meta-llama/Llama-3.2-1B-Instruct', 'port': 42435, 'model': 'meta-llama/Llama-3.2-1B-Instruct', 'dtype': 'float16', 'seed': 0, 'max_model_len': 2048, 'enforce_eager': True, 'load_format': 'dummy', 'distributed_executor_backend': 'ray', 'pipeline_parallel_size': 2, 'tensor_parallel_size': 2, 'max_num_seqs': 8, 'enable_chunked_prefill': True, 'compilation_config': {'level': 3, 'debug_dump_path': None, 'cache_dir': '', 'backend': '', 'custom_ops': ['+rms_norm'], 'splitting_ops': None, 'use_inductor': True, 'compile_sizes': [4, 8], 'inductor_compile_config': {'enable_auto_functionalized_v2': False}, 'inductor_passes': {}, 'cudagraph_mode': None, 'use_cudagraph': True, 'cudagraph_num_of_warmups': 0, 'cudagraph_capture_sizes': None, 'cudagraph_copy_inputs': False, 'full_cuda_graph': False, 'use_inductor_graph_partition': False, 'pass_config': {'enable_noop': True, 'enable_sequence_parallelism': True}, 'max_capture_size': None, 'local_cache_dir': None}} 2025-11-03T17:29:59.2802412Z (APIServer pid=53990) INFO 11-03 17:29:59 [model.py:653] Resolved architecture: LlamaForCausalLM 2025-11-03T17:29:59.2803720Z (APIServer pid=53990) `torch_dtype` is deprecated! Use `dtype` instead! 2025-11-03T17:29:59.2805060Z (APIServer pid=53990) WARNING 11-03 17:29:59 [model.py:1942] Casting torch.bfloat16 to torch.float16. 2025-11-03T17:29:59.2806385Z (APIServer pid=53990) INFO 11-03 17:29:59 [model.py:1714] Using max model len 2048 2025-11-03T17:29:59.4434349Z (APIServer pid=53990) INFO 11-03 17:29:59 [scheduler.py:225] Chunked prefill is enabled with max_num_batched_tokens=2048. 2025-11-03T17:29:59.4440225Z (APIServer pid=53990) INFO 11-03 17:29:59 [vllm.py:354] Cudagraph is disabled under eager mode 2025-11-03T17:30:00.4174576Z /opt/conda/envs/py_3.12/lib/python3.12/site-packages/torch/cuda/__init__.py:65: FutureWarning: The pynvml package is deprecated. Please install nvidia-ml-py instead. If you did not install pynvml directly, please report this to the maintainers of the package that installed pynvml for you. 2025-11-03T17:30:00.4176339Z import pynvml # type: ignore[import] 2025-11-03T17:30:02.7455650Z INFO 11-03 17:30:02 [__init__.py:224] Automatically detected platform cuda. 2025-11-03T17:30:06.5775026Z (EngineCore_DP0 pid=54112) INFO 11-03 17:30:06 [core.py:727] Waiting for init message from front-end. 2025-11-03T17:30:06.6034443Z (EngineCore_DP0 pid=54112) INFO 11-03 17:30:06 [core.py:94] Initializing a V1 LLM engine (v0.11.0rc2.dev389+ge51928192.d20251103) with config: model='meta-llama/Llama-3.2-1B-Instruct', speculative_config=None, tokenizer='meta-llama/Llama-3.2-1B-Instruct', skip_tokenizer_init=False, tokenizer_mode=auto, revision=None, tokenizer_revision=None, trust_remote_code=False, dtype=torch.float16, max_seq_len=2048, download_dir=None, load_format=dummy, tensor_parallel_size=2, pipeline_parallel_size=2, data_parallel_size=1, disable_custom_all_reduce=False, quantization=None, enforce_eager=True, kv_cache_dtype=auto, device_config=cuda, structured_outputs_config=StructuredOutputsConfig(backend='auto', disable_fallback=False, disable_any_whitespace=False, disable_additional_properties=False, reasoning_parser=''), observability_config=ObservabilityConfig(show_hidden_metrics_for_version=None, otlp_traces_endpoint=None, collect_detailed_traces=None), seed=0, served_model_name=meta-llama/Llama-3.2-1B-Instruct, enable_prefix_caching=True, chunked_prefill_enabled=True, pooler_config=None, compilation_config={'level': 3, 'debug_dump_path': None, 'cache_dir': '', 'backend': '', 'custom_ops': ['+rms_norm', '+rms_norm'], 'splitting_ops': ['vllm.unified_attention', 'vllm.unified_attention_with_output', 'vllm.unified_mla_attention', 'vllm.unified_mla_attention_with_output', 'vllm.mamba_mixer2', 'vllm.mamba_mixer', 'vllm.short_conv', 'vllm.linear_attention', 'vllm.plamo2_mamba_mixer', 'vllm.gdn_attention', 'vllm.sparse_attn_indexer'], 'use_inductor': True, 'compile_sizes': [8, 4], 'inductor_compile_config': {'enable_auto_functionalized_v2': False}, 'inductor_passes': {}, 'cudagraph_mode': , 'use_cudagraph': False, 'cudagraph_num_of_warmups': 0, 'cudagraph_capture_sizes': [], 'cudagraph_copy_inputs': False, 'full_cuda_graph': False, 'use_inductor_graph_partition': False, 'pass_config': {'enable_noop': True, 'enable_sequence_parallelism': True}, 'max_capture_size': 0, 'local_cache_dir': None} 2025-11-03T17:30:08.5034011Z (EngineCore_DP0 pid=54112) 2025-11-03 17:30:08,502 INFO worker.py:1918 -- Started a local Ray instance. View the dashboard at http://127.0.0.1:8265  2025-11-03T17:30:09.3683289Z (EngineCore_DP0 pid=54112) INFO 11-03 17:30:09 [ray_utils.py:373] No current placement group found. Creating a new placement group. 2025-11-03T17:30:09.5451234Z (EngineCore_DP0 pid=54112) INFO 11-03 17:30:09 [ray_distributed_executor.py:179] use_ray_spmd_worker: True 2025-11-03T17:30:10.2652974Z (EngineCore_DP0 pid=54112) (pid=54996) /opt/conda/envs/py_3.12/lib/python3.12/site-packages/torch/cuda/__init__.py:65: FutureWarning: The pynvml package is deprecated. Please install nvidia-ml-py instead. If you did not install pynvml directly, please report this to the maintainers of the package that installed pynvml for you. 2025-11-03T17:30:10.2654464Z (EngineCore_DP0 pid=54112) (pid=54996) import pynvml # type: ignore[import] 2025-11-03T17:30:12.6144840Z (EngineCore_DP0 pid=54112) (pid=54996) INFO 11-03 17:30:12 [__init__.py:224] Automatically detected platform cuda. 2025-11-03T17:30:14.1448188Z (EngineCore_DP0 pid=54112) INFO 11-03 17:30:14 [ray_env.py:67] RAY_NON_CARRY_OVER_ENV_VARS from config: set() 2025-11-03T17:30:14.1449850Z (EngineCore_DP0 pid=54112) INFO 11-03 17:30:14 [ray_env.py:70] Copying the following environment variables to workers: ['VLLM_WORKER_MULTIPROC_METHOD', 'MAX_JOBS', 'VLLM_USE_V1', 'HF_TOKEN', 'HUGGING_FACE_HUB_TOKEN', 'LD_LIBRARY_PATH', 'VLLM_USE_RAY_SPMD_WORKER', 'VLLM_USE_RAY_COMPILED_DAG'] 2025-11-03T17:30:14.1451467Z (EngineCore_DP0 pid=54112) INFO 11-03 17:30:14 [ray_env.py:75] If certain env vars should NOT be copied, add them to /var/lib/jenkins/.config/vllm/ray_non_carry_over_env_vars.json file 2025-11-03T17:30:17.5131927Z (EngineCore_DP0 pid=54112) (RayWorkerWrapper pid=55003) WARNING 11-03 17:30:17 [worker_base.py:307] Missing `shared_worker_lock` argument from executor. This argument is needed for mm_processor_cache_type='shm'. 2025-11-03T17:30:17.6187643Z (EngineCore_DP0 pid=54112) (pid=55005) INFO 11-03 17:30:12 [__init__.py:224] Automatically detected platform cuda. [repeated 3x across cluster] (Ray deduplicates logs by default. Set RAY_DEDUP_LOGS=0 to disable log deduplication, or see https://docs.ray.io/en/master/ray-observability/user-guides/configure-logging.html#log-deduplication for more options.) 2025-11-03T17:30:18.5402271Z (EngineCore_DP0 pid=54112) (RayWorkerWrapper pid=54996) [Gloo] Rank 1 is connected to 3 peer ranks. Expected number of connected peer ranks is : 3 2025-11-03T17:30:18.6463232Z (EngineCore_DP0 pid=54112) (RayWorkerWrapper pid=54996) INFO 11-03 17:30:18 [__init__.py:1146] Found nccl from library libnccl.so.2 2025-11-03T17:30:18.6464272Z (EngineCore_DP0 pid=54112) (RayWorkerWrapper pid=54996) INFO 11-03 17:30:18 [pynccl.py:109] vLLM is using nccl==2.27.5 2025-11-03T17:30:18.9569508Z (EngineCore_DP0 pid=54112) (RayWorkerWrapper pid=54996) WARNING 11-03 17:30:18 [symm_mem.py:60] SymmMemCommunicator: Device capability 8.9 not supported, communicator is not available. 2025-11-03T17:30:18.9571149Z (EngineCore_DP0 pid=54112) (RayWorkerWrapper pid=54996) INFO 11-03 17:30:18 [custom_all_reduce.py:37] Skipping P2P check and trusting the driver's P2P report. 2025-11-03T17:30:18.9573130Z (EngineCore_DP0 pid=54112) (RayWorkerWrapper pid=54996) WARNING 11-03 17:30:18 [custom_all_reduce.py:162] Custom allreduce is disabled because your platform lacks GPU P2P capability or P2P test failed. To silence this warning, specify disable_custom_all_reduce=True explicitly. 2025-11-03T17:30:18.9575970Z (EngineCore_DP0 pid=54112) (RayWorkerWrapper pid=55003) INFO 11-03 17:30:18 [shm_broadcast.py:302] vLLM message queue communication handle: Handle(local_reader_ranks=[1], buffer_handle=(1, 4194304, 6, 'psm_a8ff16cc'), local_subscribe_addr='ipc:///tmp/fa105a6a-95f7-4c35-a40b-cdcf6ec41054', remote_subscribe_addr=None, remote_addr_ipv6=False) 2025-11-03T17:30:19.0625736Z (EngineCore_DP0 pid=54112) (RayWorkerWrapper pid=54996) INFO 11-03 17:30:19 [parallel_state.py:1231] rank 1 in world size 4 is assigned as DP rank 0, PP rank 0, TP rank 1, EP rank 1 2025-11-03T17:30:19.3719418Z (EngineCore_DP0 pid=54112) (RayWorkerWrapper pid=54996) INFO 11-03 17:30:19 [topk_topp_sampler.py:59] Using FlashInfer for top-p & top-k sampling. 2025-11-03T17:30:19.3720451Z (EngineCore_DP0 pid=54112) (RayWorkerWrapper pid=54996) INFO 11-03 17:30:19 [gpu_model_runner.py:2840] Starting to load model meta-llama/Llama-3.2-1B-Instruct... 2025-11-03T17:30:19.5793210Z (EngineCore_DP0 pid=54112) (RayWorkerWrapper pid=55003) INFO 11-03 17:30:19 [gpu_model_runner.py:2870] Loading model from scratch... 2025-11-03T17:30:19.6863065Z (EngineCore_DP0 pid=54112) (RayWorkerWrapper pid=54996) INFO 11-03 17:30:19 [cuda.py:404] Using Flash Attention backend on V1 engine. 2025-11-03T17:30:20.1992129Z (EngineCore_DP0 pid=54112) (RayWorkerWrapper pid=55003) INFO 11-03 17:30:20 [gpu_model_runner.py:2902] Model loading took 0.7135 GiB and 0.152665 seconds 2025-11-03T17:30:22.6504815Z (EngineCore_DP0 pid=54112) (RayWorkerWrapper pid=54996) INFO 11-03 17:30:22 [backends.py:594] Using cache directory: /var/lib/jenkins/.cache/vllm/torch_compile_cache/3ff2b0020c/rank_1_0/backbone for vLLM's torch.compile 2025-11-03T17:30:22.6506438Z (EngineCore_DP0 pid=54112) (RayWorkerWrapper pid=54996) INFO 11-03 17:30:22 [backends.py:608] Dynamo bytecode transform time: 2.13 s 2025-11-03T17:30:22.6508034Z (EngineCore_DP0 pid=54112) (RayWorkerWrapper pid=54997) WARNING 11-03 17:30:17 [worker_base.py:307] Missing `shared_worker_lock` argument from executor. This argument is needed for mm_processor_cache_type='shm'. [repeated 3x across cluster] 2025-11-03T17:30:23.1628297Z (EngineCore_DP0 pid=54112) (RayWorkerWrapper pid=54997) INFO 11-03 17:30:23 [backends.py:179] Directly load the compiled graph(s) for dynamic shape from the cache, took 0.384 s 2025-11-03T17:30:24.1907378Z (EngineCore_DP0 pid=54112) (RayWorkerWrapper pid=54996) INFO 11-03 17:30:24 [gpu_worker.py:314] Available KV cache memory: 19.02 GiB 2025-11-03T17:30:24.1909625Z (EngineCore_DP0 pid=54112) (RayWorkerWrapper pid=55005) [Gloo] Rank 1 is connected to 1 peer ranks. Expected number of connected peer ranks is : 1 [repeated 23x across cluster] 2025-11-03T17:30:24.1910951Z (EngineCore_DP0 pid=54112) (RayWorkerWrapper pid=55005) INFO 11-03 17:30:18 [__init__.py:1146] Found nccl from library libnccl.so.2 [repeated 11x across cluster] 2025-11-03T17:30:24.1912141Z (EngineCore_DP0 pid=54112) (RayWorkerWrapper pid=55005) INFO 11-03 17:30:18 [pynccl.py:109] vLLM is using nccl==2.27.5 [repeated 11x across cluster] 2025-11-03T17:30:24.1913258Z (EngineCore_DP0 pid=54112) (RayWorkerWrapper pid=55005) WARNING 11-03 17:30:18 [symm_mem.py:60] SymmMemCommunicator: Device capability 8.9 not supported, communicator is not available. [repeated 3x across cluster] 2025-11-03T17:30:24.1914504Z (EngineCore_DP0 pid=54112) (RayWorkerWrapper pid=55005) INFO 11-03 17:30:18 [custom_all_reduce.py:37] Skipping P2P check and trusting the driver's P2P report. [repeated 3x across cluster] 2025-11-03T17:30:24.1916008Z (EngineCore_DP0 pid=54112) (RayWorkerWrapper pid=55005) WARNING 11-03 17:30:18 [custom_all_reduce.py:162] Custom allreduce is disabled because your platform lacks GPU P2P capability or P2P test failed. To silence this warning, specify disable_custom_all_reduce=True explicitly. [repeated 3x across cluster] 2025-11-03T17:30:24.1917926Z (EngineCore_DP0 pid=54112) (RayWorkerWrapper pid=54997) INFO 11-03 17:30:18 [shm_broadcast.py:302] vLLM message queue communication handle: Handle(local_reader_ranks=[1], buffer_handle=(1, 4194304, 6, 'psm_9c8fd141'), local_subscribe_addr='ipc:///tmp/a3479107-5318-406e-8fdb-c0bbd3299882', remote_subscribe_addr=None, remote_addr_ipv6=False) 2025-11-03T17:30:24.1919480Z (EngineCore_DP0 pid=54112) (RayWorkerWrapper pid=55005) INFO 11-03 17:30:19 [parallel_state.py:1231] rank 3 in world size 4 is assigned as DP rank 0, PP rank 1, TP rank 1, EP rank 1 [repeated 3x across cluster] 2025-11-03T17:30:24.5206561Z (EngineCore_DP0 pid=54112) INFO 11-03 17:30:24 [kv_cache_utils.py:1199] GPU KV cache size: 2,489,088 tokens 2025-11-03T17:30:24.5208378Z (EngineCore_DP0 pid=54112) INFO 11-03 17:30:24 [kv_cache_utils.py:1204] Maximum concurrency for 2,048 tokens per request: 1215.38x 2025-11-03T17:30:24.5211318Z (EngineCore_DP0 pid=54112) INFO 11-03 17:30:24 [kv_cache_utils.py:1199] GPU KV cache size: 2,489,088 tokens 2025-11-03T17:30:24.5213215Z (EngineCore_DP0 pid=54112) INFO 11-03 17:30:24 [kv_cache_utils.py:1204] Maximum concurrency for 2,048 tokens per request: 1215.38x 2025-11-03T17:30:24.5213863Z (EngineCore_DP0 pid=54112) INFO 11-03 17:30:24 [kv_cache_utils.py:1199] GPU KV cache size: 2,489,088 tokens 2025-11-03T17:30:24.5214513Z (EngineCore_DP0 pid=54112) INFO 11-03 17:30:24 [kv_cache_utils.py:1204] Maximum concurrency for 2,048 tokens per request: 1215.38x 2025-11-03T17:30:24.5215149Z (EngineCore_DP0 pid=54112) INFO 11-03 17:30:24 [kv_cache_utils.py:1199] GPU KV cache size: 2,489,088 tokens 2025-11-03T17:30:24.5215806Z (EngineCore_DP0 pid=54112) INFO 11-03 17:30:24 [kv_cache_utils.py:1204] Maximum concurrency for 2,048 tokens per request: 1215.38x 2025-11-03T17:30:24.6043000Z (EngineCore_DP0 pid=54112) (RayWorkerWrapper pid=54996) INFO 11-03 17:30:24 [gpu_worker.py:351] Compile and warming up model for size 8 2025-11-03T17:30:24.6044589Z (EngineCore_DP0 pid=54112) (RayWorkerWrapper pid=55005) INFO 11-03 17:30:19 [topk_topp_sampler.py:59] Using FlashInfer for top-p & top-k sampling. [repeated 3x across cluster] 2025-11-03T17:30:24.6046016Z (EngineCore_DP0 pid=54112) (RayWorkerWrapper pid=55005) INFO 11-03 17:30:19 [gpu_model_runner.py:2840] Starting to load model meta-llama/Llama-3.2-1B-Instruct... [repeated 3x across cluster] 2025-11-03T17:30:24.6047431Z (EngineCore_DP0 pid=54112) (RayWorkerWrapper pid=55005) INFO 11-03 17:30:19 [gpu_model_runner.py:2870] Loading model from scratch... [repeated 3x across cluster] 2025-11-03T17:30:25.2184003Z (EngineCore_DP0 pid=54112) (RayWorkerWrapper pid=54996) INFO 11-03 17:30:25 [backends.py:185] Directly load the compiled graph(s) for shape 8 from the cache, took 0.666 s 2025-11-03T17:30:25.2186074Z (EngineCore_DP0 pid=54112) (RayWorkerWrapper pid=55005) INFO 11-03 17:30:19 [cuda.py:404] Using Flash Attention backend on V1 engine. [repeated 3x across cluster] 2025-11-03T17:30:25.2188112Z (EngineCore_DP0 pid=54112) (RayWorkerWrapper pid=55005) INFO 11-03 17:30:20 [gpu_model_runner.py:2902] Model loading took 0.7135 GiB and 0.154302 seconds [repeated 3x across cluster] 2025-11-03T17:30:25.6280938Z (EngineCore_DP0 pid=54112) (RayWorkerWrapper pid=54996) INFO 11-03 17:30:25 [monitor.py:33] torch.compile takes 2.13 s in total 2025-11-03T17:30:25.7543920Z (EngineCore_DP0 pid=54112) INFO 11-03 17:30:25 [core.py:240] init engine (profile, create kv cache, warmup model) took 5.51 seconds 2025-11-03T17:30:26.5683296Z (EngineCore_DP0 pid=54112) INFO 11-03 17:30:26 [core.py:178] Batch queue is enabled with size 2 2025-11-03T17:30:26.5717894Z (EngineCore_DP0 pid=54112) INFO 11-03 17:30:26 [vllm.py:354] Cudagraph is disabled under eager mode 2025-11-03T17:30:26.7468961Z (APIServer pid=53990) INFO 11-03 17:30:26 [loggers.py:172] Engine 000: vllm cache_config_info with initialization after num_gpu_blocks is: 155568 2025-11-03T17:30:26.7957305Z (EngineCore_DP0 pid=54112) INFO 11-03 17:30:26 [gc_utils.py:40] GC Debug Config. enabled:False,top_objects:-1 2025-11-03T17:30:26.8234106Z (APIServer pid=53990) INFO 11-03 17:30:26 [api_server.py:1628] Supported tasks: ['generate'] 2025-11-03T17:30:26.8497467Z (APIServer pid=53990) WARNING 11-03 17:30:26 [model.py:1581] Default sampling parameters have been overridden by the model's Hugging Face generation config recommended from the model creator. If this is not intended, please relaunch vLLM instance with `--generation-config vllm`. 2025-11-03T17:30:26.8498847Z (APIServer pid=53990) INFO 11-03 17:30:26 [serving_responses.py:157] Using default chat sampling params from model: {'temperature': 0.6, 'top_p': 0.9} 2025-11-03T17:30:26.8922104Z (APIServer pid=53990) INFO 11-03 17:30:26 [serving_chat.py:159] Using default chat sampling params from model: {'temperature': 0.6, 'top_p': 0.9} 2025-11-03T17:30:26.9150249Z (APIServer pid=53990) INFO 11-03 17:30:26 [serving_completion.py:67] Using default completion sampling params from model: {'temperature': 0.6, 'top_p': 0.9} 2025-11-03T17:30:26.9153729Z (APIServer pid=53990) INFO 11-03 17:30:26 [api_server.py:1939] Starting vLLM API server 0 on http://0.0.0.0:42435 2025-11-03T17:30:26.9156262Z (APIServer pid=53990) INFO 11-03 17:30:26 [launcher.py:38] Available routes are: 2025-11-03T17:30:26.9157221Z (APIServer pid=53990) INFO 11-03 17:30:26 [launcher.py:46] Route: /openapi.json, Methods: HEAD, GET 2025-11-03T17:30:26.9157826Z (APIServer pid=53990) INFO 11-03 17:30:26 [launcher.py:46] Route: /docs, Methods: HEAD, GET 2025-11-03T17:30:26.9158432Z (APIServer pid=53990) INFO 11-03 17:30:26 [launcher.py:46] Route: /docs/oauth2-redirect, Methods: HEAD, GET 2025-11-03T17:30:26.9159250Z (APIServer pid=53990) INFO 11-03 17:30:26 [launcher.py:46] Route: /redoc, Methods: HEAD, GET 2025-11-03T17:30:26.9159807Z (APIServer pid=53990) INFO 11-03 17:30:26 [launcher.py:46] Route: /health, Methods: GET 2025-11-03T17:30:26.9160402Z (APIServer pid=53990) INFO 11-03 17:30:26 [launcher.py:46] Route: /load, Methods: GET 2025-11-03T17:30:26.9160902Z (APIServer pid=53990) INFO 11-03 17:30:26 [launcher.py:46] Route: /ping, Methods: POST 2025-11-03T17:30:26.9161531Z (APIServer pid=53990) INFO 11-03 17:30:26 [launcher.py:46] Route: /ping, Methods: GET 2025-11-03T17:30:26.9162031Z (APIServer pid=53990) INFO 11-03 17:30:26 [launcher.py:46] Route: /tokenize, Methods: POST 2025-11-03T17:30:26.9162551Z (APIServer pid=53990) INFO 11-03 17:30:26 [launcher.py:46] Route: /detokenize, Methods: POST 2025-11-03T17:30:26.9163099Z (APIServer pid=53990) INFO 11-03 17:30:26 [launcher.py:46] Route: /v1/models, Methods: GET 2025-11-03T17:30:26.9163625Z (APIServer pid=53990) INFO 11-03 17:30:26 [launcher.py:46] Route: /version, Methods: GET 2025-11-03T17:30:26.9164152Z (APIServer pid=53990) INFO 11-03 17:30:26 [launcher.py:46] Route: /v1/responses, Methods: POST 2025-11-03T17:30:26.9164731Z (APIServer pid=53990) INFO 11-03 17:30:26 [launcher.py:46] Route: /v1/responses/{response_id}, Methods: GET 2025-11-03T17:30:26.9165368Z (APIServer pid=53990) INFO 11-03 17:30:26 [launcher.py:46] Route: /v1/responses/{response_id}/cancel, Methods: POST 2025-11-03T17:30:26.9165980Z (APIServer pid=53990) INFO 11-03 17:30:26 [launcher.py:46] Route: /v1/chat/completions, Methods: POST 2025-11-03T17:30:26.9166542Z (APIServer pid=53990) INFO 11-03 17:30:26 [launcher.py:46] Route: /v1/completions, Methods: POST 2025-11-03T17:30:26.9167094Z (APIServer pid=53990) INFO 11-03 17:30:26 [launcher.py:46] Route: /v1/embeddings, Methods: POST 2025-11-03T17:30:26.9167659Z (APIServer pid=53990) INFO 11-03 17:30:26 [launcher.py:46] Route: /pooling, Methods: POST 2025-11-03T17:30:26.9168183Z (APIServer pid=53990) INFO 11-03 17:30:26 [launcher.py:46] Route: /classify, Methods: POST 2025-11-03T17:30:26.9168795Z (APIServer pid=53990) INFO 11-03 17:30:26 [launcher.py:46] Route: /score, Methods: POST 2025-11-03T17:30:26.9169314Z (APIServer pid=53990) INFO 11-03 17:30:26 [launcher.py:46] Route: /v1/score, Methods: POST 2025-11-03T17:30:26.9169879Z (APIServer pid=53990) INFO 11-03 17:30:26 [launcher.py:46] Route: /v1/audio/transcriptions, Methods: POST 2025-11-03T17:30:26.9170459Z (APIServer pid=53990) INFO 11-03 17:30:26 [launcher.py:46] Route: /v1/audio/translations, Methods: POST 2025-11-03T17:30:26.9171103Z (APIServer pid=53990) INFO 11-03 17:30:26 [launcher.py:46] Route: /rerank, Methods: POST 2025-11-03T17:30:26.9171644Z (APIServer pid=53990) INFO 11-03 17:30:26 [launcher.py:46] Route: /v1/rerank, Methods: POST 2025-11-03T17:30:26.9172161Z (APIServer pid=53990) INFO 11-03 17:30:26 [launcher.py:46] Route: /v2/rerank, Methods: POST 2025-11-03T17:30:26.9172698Z (APIServer pid=53990) INFO 11-03 17:30:26 [launcher.py:46] Route: /scale_elastic_ep, Methods: POST 2025-11-03T17:30:26.9173261Z (APIServer pid=53990) INFO 11-03 17:30:26 [launcher.py:46] Route: /is_scaling_elastic_ep, Methods: POST 2025-11-03T17:30:26.9173825Z (APIServer pid=53990) INFO 11-03 17:30:26 [launcher.py:46] Route: /invocations, Methods: POST 2025-11-03T17:30:26.9174343Z (APIServer pid=53990) INFO 11-03 17:30:26 [launcher.py:46] Route: /metrics, Methods: GET 2025-11-03T17:30:26.9309365Z (APIServer pid=53990) INFO: Started server process [53990] 2025-11-03T17:30:26.9311689Z (APIServer pid=53990) INFO: Waiting for application startup. 2025-11-03T17:30:27.1502080Z (APIServer pid=53990) INFO: Application startup complete. 2025-11-03T17:30:27.3007674Z (APIServer pid=53990) INFO: 127.0.0.1:46476 - "GET /health HTTP/1.1" 200 OK 2025-11-03T17:30:27.5483318Z (APIServer pid=53990) INFO: 127.0.0.1:46480 - "GET /v1/models HTTP/1.1" 200 OK 2025-11-03T17:30:27.5785873Z (EngineCore_DP0 pid=54112) INFO 11-03 17:30:27 [ray_distributed_executor.py:569] RAY_CGRAPH_get_timeout is set to 300 2025-11-03T17:30:27.5787117Z (EngineCore_DP0 pid=54112) INFO 11-03 17:30:27 [ray_distributed_executor.py:573] VLLM_USE_RAY_COMPILED_DAG_CHANNEL_TYPE = auto 2025-11-03T17:30:27.5788042Z (EngineCore_DP0 pid=54112) INFO 11-03 17:30:27 [ray_distributed_executor.py:577] VLLM_USE_RAY_COMPILED_DAG_OVERLAP_COMM = False 2025-11-03T17:30:27.5792169Z (EngineCore_DP0 pid=54112) INFO 11-03 17:30:27 [ray_distributed_executor.py:653] Using RayPPCommunicator (which wraps vLLM _PP GroupCoordinator) for Ray Compiled Graph communication. 2025-11-03T17:30:27.6064579Z (EngineCore_DP0 pid=54112) 2025-11-03 17:30:27,605 INFO torch_tensor_accelerator_channel.py:807 -- Creating communicator group 73e06cfa-c32f-489e-b016-8755f2c8f6c6 on actors: [Actor(RayWorkerWrapper, a7b6088c11382a14377b7ab301000000), Actor(RayWorkerWrapper, 7d87b3bd4f7af1bdd8972bfe01000000), Actor(RayWorkerWrapper, cf542c3d5e279dc764c3147501000000), Actor(RayWorkerWrapper, 5dcfb6f0b58336aef7f2839001000000)] 2025-11-03T17:30:27.6614800Z (EngineCore_DP0 pid=54112) 2025-11-03 17:30:27,660 INFO torch_tensor_accelerator_channel.py:833 -- Communicator group initialized. 2025-11-03T17:30:27.6779002Z (EngineCore_DP0 pid=54112) (RayWorkerWrapper pid=54996) /opt/conda/envs/py_3.12/lib/python3.12/site-packages/vllm/distributed/device_communicators/ray_communicator.py:107: UserWarning: The given buffer is not writable, and PyTorch does not support non-writable tensors. This means you can write to the underlying (supposedly non-writable) buffer using the tensor. You may want to copy the buffer to protect its data or make it writable before converting it to a tensor. This type of warning will be suppressed for the rest of this program. (Triggered internally at /var/lib/jenkins/workspace/torch/csrc/utils/tensor_new.cpp:1581.) 2025-11-03T17:30:27.6781798Z (EngineCore_DP0 pid=54112) (RayWorkerWrapper pid=54996) actor_id_tensor = torch.frombuffer(actor_id_bytes, dtype=torch.uint8).to( 2025-11-03T17:30:27.6783786Z (EngineCore_DP0 pid=54112) (pid=55005) /opt/conda/envs/py_3.12/lib/python3.12/site-packages/torch/cuda/__init__.py:65: FutureWarning: The pynvml package is deprecated. Please install nvidia-ml-py instead. If you did not install pynvml directly, please report this to the maintainers of the package that installed pynvml for you. [repeated 3x across cluster] 2025-11-03T17:30:27.6785642Z (EngineCore_DP0 pid=54112) (pid=55005) import pynvml # type: ignore[import] [repeated 3x across cluster] 2025-11-03T17:30:27.8909391Z (APIServer pid=53990) INFO: 127.0.0.1:46480 - "POST /v1/completions HTTP/1.1" 200 OK 2025-11-03T17:30:27.9543449Z (APIServer pid=53990) INFO: 127.0.0.1:46480 - "POST /v1/completions HTTP/1.1" 200 OK 2025-11-03T17:30:28.1059252Z (APIServer pid=53990) INFO: 127.0.0.1:46480 - "POST /v1/completions HTTP/1.1" 200 OK 2025-11-03T17:30:28.1847175Z (APIServer pid=53990) INFO: 127.0.0.1:46480 - "POST /v1/completions HTTP/1.1" 200 OK 2025-11-03T17:30:28.2474708Z (APIServer pid=53990) INFO: 127.0.0.1:46480 - "POST /v1/completions HTTP/1.1" 200 OK 2025-11-03T17:30:28.2551763Z (APIServer pid=53990) INFO: 127.0.0.1:46480 - "POST /v1/completions HTTP/1.1" 200 OK 2025-11-03T17:30:28.3624775Z (APIServer pid=53990) INFO 11-03 17:30:28 [launcher.py:110] Shutting down FastAPI HTTP server. 2025-11-03T17:30:28.3625439Z *** SIGTERM received at time=1762191028 on cpu 18 *** 2025-11-03T17:30:28.3673989Z PC: @ 0x7f0da022e117 (unknown) (unknown) 2025-11-03T17:30:28.3675333Z @ 0x7f0da01df520 (unknown) (unknown) 2025-11-03T17:30:28.3676126Z [2025-11-03 17:30:28,367 E 54112 54112] logging.cc:501: *** SIGTERM received at time=1762191028 on cpu 18 *** 2025-11-03T17:30:28.3677704Z [2025-11-03 17:30:28,367 E 54112 54112] logging.cc:501: PC: @ 0x7f0da022e117 (unknown) (unknown) 2025-11-03T17:30:28.3679358Z [2025-11-03 17:30:28,367 E 54112 54112] logging.cc:501: @ 0x7f0da01df520 (unknown) (unknown) 2025-11-03T17:30:28.3682084Z (EngineCore_DP0 pid=54112) INFO 11-03 17:30:28 [ray_distributed_executor.py:127] Shutting down Ray distributed executor. If you see error log from logging.cc regarding SIGTERM received, please ignore because this is the expected termination process in Ray. 2025-11-03T17:30:28.3683936Z (EngineCore_DP0 pid=54112) 2025-11-03 17:30:28,367 INFO compiled_dag_node.py:2171 -- Tearing down compiled DAG 2025-11-03T17:30:28.3685463Z (EngineCore_DP0 pid=54112) 2025-11-03 17:30:28,368 INFO compiled_dag_node.py:2176 -- Cancelling compiled worker on actor: Actor(RayWorkerWrapper, a7b6088c11382a14377b7ab301000000) 2025-11-03T17:30:28.3686520Z (EngineCore_DP0 pid=54112) 2025-11-03 17:30:28,368 INFO compiled_dag_node.py:2176 -- Cancelling compiled worker on actor: Actor(RayWorkerWrapper, 5dcfb6f0b58336aef7f2839001000000) 2025-11-03T17:30:28.3687522Z (EngineCore_DP0 pid=54112) 2025-11-03 17:30:28,368 INFO compiled_dag_node.py:2176 -- Cancelling compiled worker on actor: Actor(RayWorkerWrapper, cf542c3d5e279dc764c3147501000000) 2025-11-03T17:30:28.3688611Z (EngineCore_DP0 pid=54112) 2025-11-03 17:30:28,368 INFO compiled_dag_node.py:2176 -- Cancelling compiled worker on actor: Actor(RayWorkerWrapper, 7d87b3bd4f7af1bdd8972bfe01000000) 2025-11-03T17:30:28.3739050Z (EngineCore_DP0 pid=54112) 2025-11-03 17:30:28,373 INFO compiled_dag_node.py:2198 -- Waiting for worker tasks to exit 2025-11-03T17:30:28.3741741Z (EngineCore_DP0 pid=54112) 2025-11-03 17:30:28,373 INFO compiled_dag_node.py:2201 -- Teardown complete 2025-11-03T17:30:28.8770419Z (EngineCore_DP0 pid=54112) (RayWorkerWrapper pid=55005) INFO 11-03 17:30:22 [backends.py:594] Using cache directory: /var/lib/jenkins/.cache/vllm/torch_compile_cache/ba6452aee0/rank_3_0/backbone for vLLM's torch.compile [repeated 3x across cluster] 2025-11-03T17:30:28.8772733Z (EngineCore_DP0 pid=54112) (RayWorkerWrapper pid=55005) INFO 11-03 17:30:22 [backends.py:608] Dynamo bytecode transform time: 2.08 s [repeated 3x across cluster] 2025-11-03T17:30:28.8775351Z (EngineCore_DP0 pid=54112) (RayWorkerWrapper pid=55005) INFO 11-03 17:30:23 [backends.py:179] Directly load the compiled graph(s) for dynamic shape from the cache, took 0.387 s [repeated 3x across cluster] 2025-11-03T17:30:28.8776627Z (EngineCore_DP0 pid=54112) (RayWorkerWrapper pid=54997) INFO 11-03 17:30:24 [gpu_worker.py:314] Available KV cache memory: 18.99 GiB [repeated 3x across cluster] 2025-11-03T17:30:28.8777637Z (EngineCore_DP0 pid=54112) (RayWorkerWrapper pid=55005) INFO 11-03 17:30:25 [gpu_worker.py:351] Compile and warming up model for size 4 [repeated 7x across cluster] 2025-11-03T17:30:28.8778731Z (EngineCore_DP0 pid=54112) (RayWorkerWrapper pid=55005) INFO 11-03 17:30:25 [backends.py:185] Directly load the compiled graph(s) for shape 4 from the cache, took 0.130 s [repeated 7x across cluster] 2025-11-03T17:30:28.8780056Z (EngineCore_DP0 pid=54112) (RayWorkerWrapper pid=55005) INFO 11-03 17:30:25 [monitor.py:33] torch.compile takes 2.08 s in total [repeated 3x across cluster] 2025-11-03T17:30:28.8784346Z (EngineCore_DP0 pid=54112) (RayWorkerWrapper pid=55005) /opt/conda/envs/py_3.12/lib/python3.12/site-packages/vllm/distributed/device_communicators/ray_communicator.py:107: UserWarning: The given buffer is not writable, and PyTorch does not support non-writable tensors. This means you can write to the underlying (supposedly non-writable) buffer using the tensor. You may want to copy the buffer to protect its data or make it writable before converting it to a tensor. This type of warning will be suppressed for the rest of this program. (Triggered internally at /var/lib/jenkins/workspace/torch/csrc/utils/tensor_new.cpp:1581.) [repeated 3x across cluster] 2025-11-03T17:30:28.8788995Z (EngineCore_DP0 pid=54112) (RayWorkerWrapper pid=55005) actor_id_tensor = torch.frombuffer(actor_id_bytes, dtype=torch.uint8).to( [repeated 3x across cluster] 2025-11-03T17:30:30.7852160Z (APIServer pid=53990) INFO: Shutting down 2025-11-03T17:30:30.8845177Z (APIServer pid=53990) INFO: Waiting for application shutdown. 2025-11-03T17:30:30.8845830Z (APIServer pid=53990) INFO: Application shutdown complete. 2025-11-03T17:30:31.2275740Z INFO 11-03 17:30:31 [model.py:653] Resolved architecture: LlamaForCausalLM 2025-11-03T17:30:31.2276727Z WARNING 11-03 17:30:31 [model.py:1942] Casting torch.bfloat16 to torch.float16. 2025-11-03T17:30:31.2277591Z INFO 11-03 17:30:31 [model.py:1714] Using max model len 2048 2025-11-03T17:30:31.2279665Z Launching RemoteOpenAIServer with: vllm serve meta-llama/Llama-3.2-1B-Instruct --dtype float16 --max-model-len 2048 --max-num-seqs 8 --enable-chunked-prefill --enforce-eager --tokenizer-mode auto --tensor-parallel-size 2 --distributed-executor-backend mp --load-format dummy --port 43537 --seed 0 2025-11-03T17:30:31.7623726Z /opt/conda/envs/py_3.12/lib/python3.12/site-packages/torch/cuda/__init__.py:65: FutureWarning: The pynvml package is deprecated. Please install nvidia-ml-py instead. If you did not install pynvml directly, please report this to the maintainers of the package that installed pynvml for you. 2025-11-03T17:30:31.7624946Z import pynvml # type: ignore[import] 2025-11-03T17:30:34.0111988Z INFO 11-03 17:30:34 [__init__.py:224] Automatically detected platform cuda. 2025-11-03T17:30:37.7053833Z (APIServer pid=58917) INFO 11-03 17:30:37 [api_server.py:1870] vLLM API server version 0.11.0rc2.dev389+ge51928192.d20251103 2025-11-03T17:30:37.7077916Z (APIServer pid=58917) INFO 11-03 17:30:37 [utils.py:239] non-default args: {'model_tag': 'meta-llama/Llama-3.2-1B-Instruct', 'port': 43537, 'model': 'meta-llama/Llama-3.2-1B-Instruct', 'dtype': 'float16', 'seed': 0, 'max_model_len': 2048, 'enforce_eager': True, 'load_format': 'dummy', 'distributed_executor_backend': 'mp', 'tensor_parallel_size': 2, 'max_num_seqs': 8, 'enable_chunked_prefill': True} 2025-11-03T17:30:37.8512354Z (APIServer pid=58917) INFO 11-03 17:30:37 [model.py:653] Resolved architecture: LlamaForCausalLM 2025-11-03T17:30:37.8513833Z (APIServer pid=58917) `torch_dtype` is deprecated! Use `dtype` instead! 2025-11-03T17:30:37.8515300Z (APIServer pid=58917) WARNING 11-03 17:30:37 [model.py:1942] Casting torch.bfloat16 to torch.float16. 2025-11-03T17:30:37.8516100Z (APIServer pid=58917) INFO 11-03 17:30:37 [model.py:1714] Using max model len 2048 2025-11-03T17:30:38.0173269Z (APIServer pid=58917) INFO 11-03 17:30:38 [scheduler.py:225] Chunked prefill is enabled with max_num_batched_tokens=2048. 2025-11-03T17:30:38.0177389Z (APIServer pid=58917) INFO 11-03 17:30:38 [vllm.py:354] Cudagraph is disabled under eager mode 2025-11-03T17:30:39.0230409Z /opt/conda/envs/py_3.12/lib/python3.12/site-packages/torch/cuda/__init__.py:65: FutureWarning: The pynvml package is deprecated. Please install nvidia-ml-py instead. If you did not install pynvml directly, please report this to the maintainers of the package that installed pynvml for you. 2025-11-03T17:30:39.0231670Z import pynvml # type: ignore[import] 2025-11-03T17:30:41.3709257Z INFO 11-03 17:30:41 [__init__.py:224] Automatically detected platform cuda. 2025-11-03T17:30:45.1072520Z (EngineCore_DP0 pid=59039) INFO 11-03 17:30:45 [core.py:727] Waiting for init message from front-end. 2025-11-03T17:30:45.1326627Z (EngineCore_DP0 pid=59039) INFO 11-03 17:30:45 [core.py:94] Initializing a V1 LLM engine (v0.11.0rc2.dev389+ge51928192.d20251103) with config: model='meta-llama/Llama-3.2-1B-Instruct', speculative_config=None, tokenizer='meta-llama/Llama-3.2-1B-Instruct', skip_tokenizer_init=False, tokenizer_mode=auto, revision=None, tokenizer_revision=None, trust_remote_code=False, dtype=torch.float16, max_seq_len=2048, download_dir=None, load_format=dummy, tensor_parallel_size=2, pipeline_parallel_size=1, data_parallel_size=1, disable_custom_all_reduce=False, quantization=None, enforce_eager=True, kv_cache_dtype=auto, device_config=cuda, structured_outputs_config=StructuredOutputsConfig(backend='auto', disable_fallback=False, disable_any_whitespace=False, disable_additional_properties=False, reasoning_parser=''), observability_config=ObservabilityConfig(show_hidden_metrics_for_version=None, otlp_traces_endpoint=None, collect_detailed_traces=None), seed=0, served_model_name=meta-llama/Llama-3.2-1B-Instruct, enable_prefix_caching=True, chunked_prefill_enabled=True, pooler_config=None, compilation_config={'level': 0, 'debug_dump_path': None, 'cache_dir': '', 'backend': '', 'custom_ops': [], 'splitting_ops': None, 'use_inductor': True, 'compile_sizes': [], 'inductor_compile_config': {'enable_auto_functionalized_v2': False}, 'inductor_passes': {}, 'cudagraph_mode': , 'use_cudagraph': False, 'cudagraph_num_of_warmups': 0, 'cudagraph_capture_sizes': [], 'cudagraph_copy_inputs': False, 'full_cuda_graph': False, 'use_inductor_graph_partition': False, 'pass_config': {}, 'max_capture_size': 0, 'local_cache_dir': None} 2025-11-03T17:30:45.1332713Z (EngineCore_DP0 pid=59039) WARNING 11-03 17:30:45 [multiproc_executor.py:744] Reducing Torch parallelism from 24 threads to 1 to avoid unnecessary CPU contention. Set OMP_NUM_THREADS in the external environment to tune this value as needed. 2025-11-03T17:30:45.1340548Z (EngineCore_DP0 pid=59039) INFO 11-03 17:30:45 [shm_broadcast.py:302] vLLM message queue communication handle: Handle(local_reader_ranks=[0, 1], buffer_handle=(2, 16777216, 10, 'psm_8e177590'), local_subscribe_addr='ipc:///tmp/5e48f7d2-0714-4c3d-bfb4-00285d555739', remote_subscribe_addr=None, remote_addr_ipv6=False) 2025-11-03T17:30:45.6232253Z /opt/conda/envs/py_3.12/lib/python3.12/site-packages/torch/cuda/__init__.py:65: FutureWarning: The pynvml package is deprecated. Please install nvidia-ml-py instead. If you did not install pynvml directly, please report this to the maintainers of the package that installed pynvml for you. 2025-11-03T17:30:45.6233803Z import pynvml # type: ignore[import] 2025-11-03T17:30:45.6278623Z /opt/conda/envs/py_3.12/lib/python3.12/site-packages/torch/cuda/__init__.py:65: FutureWarning: The pynvml package is deprecated. Please install nvidia-ml-py instead. If you did not install pynvml directly, please report this to the maintainers of the package that installed pynvml for you. 2025-11-03T17:30:45.6279776Z import pynvml # type: ignore[import] 2025-11-03T17:30:47.9732662Z INFO 11-03 17:30:47 [__init__.py:224] Automatically detected platform cuda. 2025-11-03T17:30:48.0247550Z INFO 11-03 17:30:48 [__init__.py:224] Automatically detected platform cuda. 2025-11-03T17:30:52.7428473Z INFO 11-03 17:30:52 [shm_broadcast.py:302] vLLM message queue communication handle: Handle(local_reader_ranks=[0], buffer_handle=(1, 10485760, 10, 'psm_22836898'), local_subscribe_addr='ipc:///tmp/437bd34d-5980-4944-9981-18d487f4af92', remote_subscribe_addr=None, remote_addr_ipv6=False) 2025-11-03T17:30:52.7430703Z INFO 11-03 17:30:52 [shm_broadcast.py:302] vLLM message queue communication handle: Handle(local_reader_ranks=[0], buffer_handle=(1, 10485760, 10, 'psm_6451e163'), local_subscribe_addr='ipc:///tmp/c270f13e-c6d2-4992-a361-01594410b7da', remote_subscribe_addr=None, remote_addr_ipv6=False) 2025-11-03T17:30:53.5987033Z [Gloo] Rank 1 is connected to 1 peer ranks. Expected number of connected peer ranks is : 1 2025-11-03T17:30:53.5988163Z [Gloo] Rank 0 is connected to 1 peer ranks. Expected number of connected peer ranks is : 1 2025-11-03T17:30:53.6034267Z [Gloo] Rank 1 is connected to 1 peer ranks. Expected number of connected peer ranks is : 1 2025-11-03T17:30:53.6035367Z [Gloo] Rank 0 is connected to 1 peer ranks. Expected number of connected peer ranks is : 1 2025-11-03T17:30:53.6076731Z INFO 11-03 17:30:53 [__init__.py:1146] Found nccl from library libnccl.so.2 2025-11-03T17:30:53.6077631Z INFO 11-03 17:30:53 [__init__.py:1146] Found nccl from library libnccl.so.2 2025-11-03T17:30:53.6078335Z INFO 11-03 17:30:53 [pynccl.py:109] vLLM is using nccl==2.27.5 2025-11-03T17:30:53.6078708Z INFO 11-03 17:30:53 [pynccl.py:109] vLLM is using nccl==2.27.5 2025-11-03T17:30:53.7817595Z WARNING 11-03 17:30:53 [symm_mem.py:60] SymmMemCommunicator: Device capability 8.9 not supported, communicator is not available. 2025-11-03T17:30:53.7818768Z WARNING 11-03 17:30:53 [symm_mem.py:60] SymmMemCommunicator: Device capability 8.9 not supported, communicator is not available. 2025-11-03T17:30:53.7840802Z INFO 11-03 17:30:53 [custom_all_reduce.py:37] Skipping P2P check and trusting the driver's P2P report. 2025-11-03T17:30:53.7841468Z INFO 11-03 17:30:53 [custom_all_reduce.py:37] Skipping P2P check and trusting the driver's P2P report. 2025-11-03T17:30:53.7842463Z WARNING 11-03 17:30:53 [custom_all_reduce.py:162] Custom allreduce is disabled because your platform lacks GPU P2P capability or P2P test failed. To silence this warning, specify disable_custom_all_reduce=True explicitly. 2025-11-03T17:30:53.7843867Z WARNING 11-03 17:30:53 [custom_all_reduce.py:162] Custom allreduce is disabled because your platform lacks GPU P2P capability or P2P test failed. To silence this warning, specify disable_custom_all_reduce=True explicitly. 2025-11-03T17:30:53.8627474Z INFO 11-03 17:30:53 [shm_broadcast.py:302] vLLM message queue communication handle: Handle(local_reader_ranks=[1], buffer_handle=(1, 4194304, 6, 'psm_0667e6fb'), local_subscribe_addr='ipc:///tmp/6bc5d0e8-ca0b-46f7-a54c-a58e0299fc60', remote_subscribe_addr=None, remote_addr_ipv6=False) 2025-11-03T17:30:53.8654342Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:30:53.8654833Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:30:53.8663050Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:30:53.8671458Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:30:53.9023808Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:30:53.9031193Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:30:53.9041286Z [Gloo] Rank 0 is connected to 1 peer ranks. Expected number of connected peer ranks is : 1 2025-11-03T17:30:53.9041872Z [Gloo] Rank 1 is connected to 1 peer ranks. Expected number of connected peer ranks is : 1 2025-11-03T17:30:53.9044809Z INFO 11-03 17:30:53 [__init__.py:1146] Found nccl from library libnccl.so.2 2025-11-03T17:30:53.9045262Z INFO 11-03 17:30:53 [__init__.py:1146] Found nccl from library libnccl.so.2 2025-11-03T17:30:53.9045658Z INFO 11-03 17:30:53 [pynccl.py:109] vLLM is using nccl==2.27.5 2025-11-03T17:30:53.9046047Z INFO 11-03 17:30:53 [pynccl.py:109] vLLM is using nccl==2.27.5 2025-11-03T17:30:53.9391013Z INFO 11-03 17:30:53 [parallel_state.py:1231] rank 1 in world size 2 is assigned as DP rank 0, PP rank 0, TP rank 1, EP rank 1 2025-11-03T17:30:53.9392047Z INFO 11-03 17:30:53 [parallel_state.py:1231] rank 0 in world size 2 is assigned as DP rank 0, PP rank 0, TP rank 0, EP rank 0 2025-11-03T17:30:54.1608657Z INFO 11-03 17:30:54 [topk_topp_sampler.py:59] Using FlashInfer for top-p & top-k sampling. 2025-11-03T17:30:54.1767159Z INFO 11-03 17:30:54 [topk_topp_sampler.py:59] Using FlashInfer for top-p & top-k sampling. 2025-11-03T17:30:54.1925327Z (Worker_TP1 pid=59159) INFO 11-03 17:30:54 [gpu_model_runner.py:2840] Starting to load model meta-llama/Llama-3.2-1B-Instruct... 2025-11-03T17:30:54.2095270Z (Worker_TP0 pid=59158) INFO 11-03 17:30:54 [gpu_model_runner.py:2840] Starting to load model meta-llama/Llama-3.2-1B-Instruct... 2025-11-03T17:30:54.3977021Z (Worker_TP1 pid=59159) INFO 11-03 17:30:54 [gpu_model_runner.py:2870] Loading model from scratch... 2025-11-03T17:30:54.4122319Z (Worker_TP0 pid=59158) INFO 11-03 17:30:54 [gpu_model_runner.py:2870] Loading model from scratch... 2025-11-03T17:30:54.4864725Z (Worker_TP1 pid=59159) INFO 11-03 17:30:54 [cuda.py:404] Using Flash Attention backend on V1 engine. 2025-11-03T17:30:54.5025037Z (Worker_TP0 pid=59158) INFO 11-03 17:30:54 [cuda.py:404] Using Flash Attention backend on V1 engine. 2025-11-03T17:30:54.9198905Z (Worker_TP1 pid=59159) INFO 11-03 17:30:54 [gpu_model_runner.py:2902] Model loading took 1.1667 GiB and 0.117948 seconds 2025-11-03T17:30:54.9466254Z (Worker_TP0 pid=59158) INFO 11-03 17:30:54 [gpu_model_runner.py:2902] Model loading took 1.1667 GiB and 0.119347 seconds 2025-11-03T17:30:56.8851424Z (Worker_TP0 pid=59158) INFO 11-03 17:30:56 [gpu_worker.py:314] Available KV cache memory: 18.53 GiB 2025-11-03T17:30:56.8929617Z (Worker_TP1 pid=59159) INFO 11-03 17:30:56 [gpu_worker.py:314] Available KV cache memory: 18.53 GiB 2025-11-03T17:30:57.1182255Z (EngineCore_DP0 pid=59039) INFO 11-03 17:30:57 [kv_cache_utils.py:1199] GPU KV cache size: 1,214,464 tokens 2025-11-03T17:30:57.1183174Z (EngineCore_DP0 pid=59039) INFO 11-03 17:30:57 [kv_cache_utils.py:1204] Maximum concurrency for 2,048 tokens per request: 593.00x 2025-11-03T17:30:57.1184033Z (EngineCore_DP0 pid=59039) INFO 11-03 17:30:57 [kv_cache_utils.py:1199] GPU KV cache size: 1,214,464 tokens 2025-11-03T17:30:57.1184894Z (EngineCore_DP0 pid=59039) INFO 11-03 17:30:57 [kv_cache_utils.py:1204] Maximum concurrency for 2,048 tokens per request: 593.00x 2025-11-03T17:30:57.2401032Z (EngineCore_DP0 pid=59039) INFO 11-03 17:30:57 [core.py:240] init engine (profile, create kv cache, warmup model) took 2.23 seconds 2025-11-03T17:30:57.7450147Z (EngineCore_DP0 pid=59039) INFO 11-03 17:30:57 [vllm.py:354] Cudagraph is disabled under eager mode 2025-11-03T17:30:57.9299695Z (APIServer pid=58917) INFO 11-03 17:30:57 [loggers.py:172] Engine 000: vllm cache_config_info with initialization after num_gpu_blocks is: 75904 2025-11-03T17:30:57.9482646Z (EngineCore_DP0 pid=59039) INFO 11-03 17:30:57 [gc_utils.py:40] GC Debug Config. enabled:False,top_objects:-1 2025-11-03T17:30:57.9696894Z (APIServer pid=58917) INFO 11-03 17:30:57 [api_server.py:1628] Supported tasks: ['generate'] 2025-11-03T17:30:58.2026623Z (APIServer pid=58917) WARNING 11-03 17:30:58 [model.py:1581] Default sampling parameters have been overridden by the model's Hugging Face generation config recommended from the model creator. If this is not intended, please relaunch vLLM instance with `--generation-config vllm`. 2025-11-03T17:30:58.2029623Z (APIServer pid=58917) INFO 11-03 17:30:58 [serving_responses.py:157] Using default chat sampling params from model: {'temperature': 0.6, 'top_p': 0.9} 2025-11-03T17:30:58.2277103Z (APIServer pid=58917) INFO 11-03 17:30:58 [serving_chat.py:159] Using default chat sampling params from model: {'temperature': 0.6, 'top_p': 0.9} 2025-11-03T17:30:58.3035500Z (APIServer pid=58917) INFO 11-03 17:30:58 [serving_completion.py:67] Using default completion sampling params from model: {'temperature': 0.6, 'top_p': 0.9} 2025-11-03T17:30:58.3038419Z (APIServer pid=58917) INFO 11-03 17:30:58 [api_server.py:1939] Starting vLLM API server 0 on http://0.0.0.0:43537 2025-11-03T17:30:58.3040443Z (APIServer pid=58917) INFO 11-03 17:30:58 [launcher.py:38] Available routes are: 2025-11-03T17:30:58.3041727Z (APIServer pid=58917) INFO 11-03 17:30:58 [launcher.py:46] Route: /openapi.json, Methods: HEAD, GET 2025-11-03T17:30:58.3042619Z (APIServer pid=58917) INFO 11-03 17:30:58 [launcher.py:46] Route: /docs, Methods: HEAD, GET 2025-11-03T17:30:58.3043189Z (APIServer pid=58917) INFO 11-03 17:30:58 [launcher.py:46] Route: /docs/oauth2-redirect, Methods: HEAD, GET 2025-11-03T17:30:58.3043772Z (APIServer pid=58917) INFO 11-03 17:30:58 [launcher.py:46] Route: /redoc, Methods: HEAD, GET 2025-11-03T17:30:58.3044290Z (APIServer pid=58917) INFO 11-03 17:30:58 [launcher.py:46] Route: /health, Methods: GET 2025-11-03T17:30:58.3044787Z (APIServer pid=58917) INFO 11-03 17:30:58 [launcher.py:46] Route: /load, Methods: GET 2025-11-03T17:30:58.3045276Z (APIServer pid=58917) INFO 11-03 17:30:58 [launcher.py:46] Route: /ping, Methods: POST 2025-11-03T17:30:58.3045787Z (APIServer pid=58917) INFO 11-03 17:30:58 [launcher.py:46] Route: /ping, Methods: GET 2025-11-03T17:30:58.3046298Z (APIServer pid=58917) INFO 11-03 17:30:58 [launcher.py:46] Route: /tokenize, Methods: POST 2025-11-03T17:30:58.3046821Z (APIServer pid=58917) INFO 11-03 17:30:58 [launcher.py:46] Route: /detokenize, Methods: POST 2025-11-03T17:30:58.3047338Z (APIServer pid=58917) INFO 11-03 17:30:58 [launcher.py:46] Route: /v1/models, Methods: GET 2025-11-03T17:30:58.3047835Z (APIServer pid=58917) INFO 11-03 17:30:58 [launcher.py:46] Route: /version, Methods: GET 2025-11-03T17:30:58.3048357Z (APIServer pid=58917) INFO 11-03 17:30:58 [launcher.py:46] Route: /v1/responses, Methods: POST 2025-11-03T17:30:58.3049049Z (APIServer pid=58917) INFO 11-03 17:30:58 [launcher.py:46] Route: /v1/responses/{response_id}, Methods: GET 2025-11-03T17:30:58.3049663Z (APIServer pid=58917) INFO 11-03 17:30:58 [launcher.py:46] Route: /v1/responses/{response_id}/cancel, Methods: POST 2025-11-03T17:30:58.3050451Z (APIServer pid=58917) INFO 11-03 17:30:58 [launcher.py:46] Route: /v1/chat/completions, Methods: POST 2025-11-03T17:30:58.3051012Z (APIServer pid=58917) INFO 11-03 17:30:58 [launcher.py:46] Route: /v1/completions, Methods: POST 2025-11-03T17:30:58.3051555Z (APIServer pid=58917) INFO 11-03 17:30:58 [launcher.py:46] Route: /v1/embeddings, Methods: POST 2025-11-03T17:30:58.3052237Z (APIServer pid=58917) INFO 11-03 17:30:58 [launcher.py:46] Route: /pooling, Methods: POST 2025-11-03T17:30:58.3052770Z (APIServer pid=58917) INFO 11-03 17:30:58 [launcher.py:46] Route: /classify, Methods: POST 2025-11-03T17:30:58.3053298Z (APIServer pid=58917) INFO 11-03 17:30:58 [launcher.py:46] Route: /score, Methods: POST 2025-11-03T17:30:58.3053810Z (APIServer pid=58917) INFO 11-03 17:30:58 [launcher.py:46] Route: /v1/score, Methods: POST 2025-11-03T17:30:58.3054362Z (APIServer pid=58917) INFO 11-03 17:30:58 [launcher.py:46] Route: /v1/audio/transcriptions, Methods: POST 2025-11-03T17:30:58.3054948Z (APIServer pid=58917) INFO 11-03 17:30:58 [launcher.py:46] Route: /v1/audio/translations, Methods: POST 2025-11-03T17:30:58.3055484Z (APIServer pid=58917) INFO 11-03 17:30:58 [launcher.py:46] Route: /rerank, Methods: POST 2025-11-03T17:30:58.3056003Z (APIServer pid=58917) INFO 11-03 17:30:58 [launcher.py:46] Route: /v1/rerank, Methods: POST 2025-11-03T17:30:58.3056520Z (APIServer pid=58917) INFO 11-03 17:30:58 [launcher.py:46] Route: /v2/rerank, Methods: POST 2025-11-03T17:30:58.3057152Z (APIServer pid=58917) INFO 11-03 17:30:58 [launcher.py:46] Route: /scale_elastic_ep, Methods: POST 2025-11-03T17:30:58.3057732Z (APIServer pid=58917) INFO 11-03 17:30:58 [launcher.py:46] Route: /is_scaling_elastic_ep, Methods: POST 2025-11-03T17:30:58.3058343Z (APIServer pid=58917) INFO 11-03 17:30:58 [launcher.py:46] Route: /invocations, Methods: POST 2025-11-03T17:30:58.3058863Z (APIServer pid=58917) INFO 11-03 17:30:58 [launcher.py:46] Route: /metrics, Methods: GET 2025-11-03T17:30:58.3190982Z (APIServer pid=58917) INFO: Started server process [58917] 2025-11-03T17:30:58.3192076Z (APIServer pid=58917) INFO: Waiting for application startup. 2025-11-03T17:30:58.5279892Z (APIServer pid=58917) INFO: Application startup complete. 2025-11-03T17:30:58.7985657Z (APIServer pid=58917) INFO: 127.0.0.1:40156 - "GET /health HTTP/1.1" 200 OK 2025-11-03T17:30:58.8085537Z (APIServer pid=58917) INFO: 127.0.0.1:40168 - "GET /v1/models HTTP/1.1" 200 OK 2025-11-03T17:30:58.9817970Z (APIServer pid=58917) INFO: 127.0.0.1:40168 - "POST /v1/completions HTTP/1.1" 200 OK 2025-11-03T17:30:59.0378894Z (APIServer pid=58917) INFO: 127.0.0.1:40168 - "POST /v1/completions HTTP/1.1" 200 OK 2025-11-03T17:30:59.1339012Z (APIServer pid=58917) INFO: 127.0.0.1:40168 - "POST /v1/completions HTTP/1.1" 200 OK 2025-11-03T17:30:59.2063825Z (APIServer pid=58917) INFO: 127.0.0.1:40168 - "POST /v1/completions HTTP/1.1" 200 OK 2025-11-03T17:30:59.2768926Z (APIServer pid=58917) INFO: 127.0.0.1:40168 - "POST /v1/completions HTTP/1.1" 200 OK 2025-11-03T17:30:59.2840402Z (APIServer pid=58917) INFO: 127.0.0.1:40168 - "POST /v1/completions HTTP/1.1" 200 OK 2025-11-03T17:30:59.3506158Z (APIServer pid=58917) INFO 11-03 17:30:59 [launcher.py:110] Shutting down FastAPI HTTP server. 2025-11-03T17:30:59.3508821Z (Worker_TP0 pid=59158) INFO 11-03 17:30:59 [multiproc_executor.py:583] Parent process exited, terminating worker 2025-11-03T17:30:59.3512054Z (Worker_TP0 pid=59158) INFO 11-03 17:30:59 [multiproc_executor.py:624] WorkerProc shutting down. 2025-11-03T17:30:59.3513177Z (Worker_TP1 pid=59159) INFO 11-03 17:30:59 [multiproc_executor.py:583] Parent process exited, terminating worker 2025-11-03T17:30:59.3513957Z (Worker_TP1 pid=59159) INFO 11-03 17:30:59 [multiproc_executor.py:624] WorkerProc shutting down. 2025-11-03T17:31:01.1646929Z (APIServer pid=58917) INFO: Shutting down 2025-11-03T17:31:01.2653761Z (APIServer pid=58917) INFO: Waiting for application shutdown. 2025-11-03T17:31:01.2655905Z (APIServer pid=58917) INFO: Application shutdown complete. 2025-11-03T17:31:01.8841534Z PASSED 2025-11-03T17:31:01.8989903Z distributed/test_sequence_parallel.py::test_tp_sp_generation[RedHatAI/Meta-Llama-3.1-8B-Instruct-FP8-parallel_setup16-mp-auto-test_options16] Fork a new process to run a test 59362 2025-11-03T17:31:01.9006304Z Fork a new process to run a test 0 2025-11-03T17:31:01.9597192Z 2025-11-03T17:31:01.9604765Z tokenizer_config.json: 0.00B [00:00, ?B/s] 2025-11-03T17:31:01.9605471Z tokenizer_config.json: 55.4kB [00:00, 111MB/s] 2025-11-03T17:31:02.1846427Z 2025-11-03T17:31:02.2221456Z tokenizer.json: 0.00B [00:00, ?B/s] 2025-11-03T17:31:02.2222089Z tokenizer.json: 9.09MB [00:00, 242MB/s] 2025-11-03T17:31:02.4544400Z 2025-11-03T17:31:02.4544851Z special_tokens_map.json: 0% 0.00/325 [00:00, 'use_cudagraph': False, 'cudagraph_num_of_warmups': 0, 'cudagraph_capture_sizes': [], 'cudagraph_copy_inputs': False, 'full_cuda_graph': False, 'use_inductor_graph_partition': False, 'pass_config': {'enable_noop': True, 'enable_sequence_parallelism': True}, 'max_capture_size': 0, 'local_cache_dir': None} 2025-11-03T17:31:17.6788723Z (EngineCore_DP0 pid=59536) WARNING 11-03 17:31:17 [multiproc_executor.py:744] Reducing Torch parallelism from 24 threads to 1 to avoid unnecessary CPU contention. Set OMP_NUM_THREADS in the external environment to tune this value as needed. 2025-11-03T17:31:17.6790474Z (EngineCore_DP0 pid=59536) INFO 11-03 17:31:17 [shm_broadcast.py:302] vLLM message queue communication handle: Handle(local_reader_ranks=[0, 1], buffer_handle=(2, 16777216, 10, 'psm_3742c88a'), local_subscribe_addr='ipc:///tmp/1066807b-c9ec-44e1-9fec-840d333b29b8', remote_subscribe_addr=None, remote_addr_ipv6=False) 2025-11-03T17:31:18.1626876Z /opt/conda/envs/py_3.12/lib/python3.12/site-packages/torch/cuda/__init__.py:65: FutureWarning: The pynvml package is deprecated. Please install nvidia-ml-py instead. If you did not install pynvml directly, please report this to the maintainers of the package that installed pynvml for you. 2025-11-03T17:31:18.1628130Z import pynvml # type: ignore[import] 2025-11-03T17:31:18.1629174Z /opt/conda/envs/py_3.12/lib/python3.12/site-packages/torch/cuda/__init__.py:65: FutureWarning: The pynvml package is deprecated. Please install nvidia-ml-py instead. If you did not install pynvml directly, please report this to the maintainers of the package that installed pynvml for you. 2025-11-03T17:31:18.1630070Z import pynvml # type: ignore[import] 2025-11-03T17:31:20.5295482Z INFO 11-03 17:31:20 [__init__.py:224] Automatically detected platform cuda. 2025-11-03T17:31:20.5482640Z INFO 11-03 17:31:20 [__init__.py:224] Automatically detected platform cuda. 2025-11-03T17:31:25.3728980Z INFO 11-03 17:31:25 [shm_broadcast.py:302] vLLM message queue communication handle: Handle(local_reader_ranks=[0], buffer_handle=(1, 10485760, 10, 'psm_327a9e78'), local_subscribe_addr='ipc:///tmp/eff29e05-95e2-4360-ae57-8c8c956691bf', remote_subscribe_addr=None, remote_addr_ipv6=False) 2025-11-03T17:31:25.3732489Z INFO 11-03 17:31:25 [shm_broadcast.py:302] vLLM message queue communication handle: Handle(local_reader_ranks=[0], buffer_handle=(1, 10485760, 10, 'psm_554c8bfe'), local_subscribe_addr='ipc:///tmp/c831e29c-7e9c-4126-a4a9-90a95560357d', remote_subscribe_addr=None, remote_addr_ipv6=False) 2025-11-03T17:31:25.5664523Z [Gloo] Rank 1 is connected to 1 peer ranks. Expected number of connected peer ranks is : 1 2025-11-03T17:31:25.5665144Z [Gloo] Rank 0 is connected to 1 peer ranks. Expected number of connected peer ranks is : 1 2025-11-03T17:31:25.5714298Z [Gloo] Rank 0 is connected to 1 peer ranks. Expected number of connected peer ranks is : 1 2025-11-03T17:31:25.5714898Z [Gloo] Rank 1 is connected to 1 peer ranks. Expected number of connected peer ranks is : 1 2025-11-03T17:31:25.5756458Z INFO 11-03 17:31:25 [__init__.py:1146] Found nccl from library libnccl.so.2 2025-11-03T17:31:25.5757455Z INFO 11-03 17:31:25 [__init__.py:1146] Found nccl from library libnccl.so.2 2025-11-03T17:31:25.5757878Z INFO 11-03 17:31:25 [pynccl.py:109] vLLM is using nccl==2.27.5 2025-11-03T17:31:25.5760458Z INFO 11-03 17:31:25 [pynccl.py:109] vLLM is using nccl==2.27.5 2025-11-03T17:31:25.7478489Z WARNING 11-03 17:31:25 [symm_mem.py:60] SymmMemCommunicator: Device capability 8.9 not supported, communicator is not available. 2025-11-03T17:31:25.7479362Z WARNING 11-03 17:31:25 [symm_mem.py:60] SymmMemCommunicator: Device capability 8.9 not supported, communicator is not available. 2025-11-03T17:31:25.7505125Z INFO 11-03 17:31:25 [custom_all_reduce.py:37] Skipping P2P check and trusting the driver's P2P report. 2025-11-03T17:31:25.7505757Z INFO 11-03 17:31:25 [custom_all_reduce.py:37] Skipping P2P check and trusting the driver's P2P report. 2025-11-03T17:31:25.7506766Z WARNING 11-03 17:31:25 [custom_all_reduce.py:162] Custom allreduce is disabled because your platform lacks GPU P2P capability or P2P test failed. To silence this warning, specify disable_custom_all_reduce=True explicitly. 2025-11-03T17:31:25.7508147Z WARNING 11-03 17:31:25 [custom_all_reduce.py:162] Custom allreduce is disabled because your platform lacks GPU P2P capability or P2P test failed. To silence this warning, specify disable_custom_all_reduce=True explicitly. 2025-11-03T17:31:25.8524576Z INFO 11-03 17:31:25 [shm_broadcast.py:302] vLLM message queue communication handle: Handle(local_reader_ranks=[1], buffer_handle=(1, 4194304, 6, 'psm_171b5d14'), local_subscribe_addr='ipc:///tmp/0c7f2d0d-5157-458a-adba-a73dc292411e', remote_subscribe_addr=None, remote_addr_ipv6=False) 2025-11-03T17:31:25.8556056Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:31:25.8561852Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:31:25.8564237Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:31:25.8572812Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:31:25.9022325Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:31:25.9031411Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:31:25.9043654Z [Gloo] Rank 1 is connected to 1 peer ranks. Expected number of connected peer ranks is : 1 2025-11-03T17:31:25.9044316Z [Gloo] Rank 0 is connected to 1 peer ranks. Expected number of connected peer ranks is : 1 2025-11-03T17:31:25.9046387Z INFO 11-03 17:31:25 [__init__.py:1146] Found nccl from library libnccl.so.2 2025-11-03T17:31:25.9046848Z INFO 11-03 17:31:25 [pynccl.py:109] vLLM is using nccl==2.27.5 2025-11-03T17:31:25.9047526Z INFO 11-03 17:31:25 [__init__.py:1146] Found nccl from library libnccl.so.2 2025-11-03T17:31:25.9047967Z INFO 11-03 17:31:25 [pynccl.py:109] vLLM is using nccl==2.27.5 2025-11-03T17:31:25.9389634Z INFO 11-03 17:31:25 [parallel_state.py:1231] rank 1 in world size 2 is assigned as DP rank 0, PP rank 0, TP rank 1, EP rank 1 2025-11-03T17:31:25.9392385Z INFO 11-03 17:31:25 [parallel_state.py:1231] rank 0 in world size 2 is assigned as DP rank 0, PP rank 0, TP rank 0, EP rank 0 2025-11-03T17:31:26.1749161Z INFO 11-03 17:31:26 [topk_topp_sampler.py:59] Using FlashInfer for top-p & top-k sampling. 2025-11-03T17:31:26.1765630Z INFO 11-03 17:31:26 [topk_topp_sampler.py:59] Using FlashInfer for top-p & top-k sampling. 2025-11-03T17:31:26.2117728Z (Worker_TP0 pid=59655) INFO 11-03 17:31:26 [gpu_model_runner.py:2840] Starting to load model RedHatAI/Meta-Llama-3.1-8B-Instruct-FP8... 2025-11-03T17:31:26.2144567Z (Worker_TP1 pid=59656) INFO 11-03 17:31:26 [gpu_model_runner.py:2840] Starting to load model RedHatAI/Meta-Llama-3.1-8B-Instruct-FP8... 2025-11-03T17:31:26.4222154Z (Worker_TP1 pid=59656) INFO 11-03 17:31:26 [gpu_model_runner.py:2870] Loading model from scratch... 2025-11-03T17:31:26.4223422Z (Worker_TP0 pid=59655) INFO 11-03 17:31:26 [gpu_model_runner.py:2870] Loading model from scratch... 2025-11-03T17:31:26.5137734Z (Worker_TP1 pid=59656) INFO 11-03 17:31:26 [cuda.py:404] Using Flash Attention backend on V1 engine. 2025-11-03T17:31:26.5142841Z (Worker_TP0 pid=59655) INFO 11-03 17:31:26 [cuda.py:404] Using Flash Attention backend on V1 engine. 2025-11-03T17:31:27.3802887Z (Worker_TP0 pid=59655) INFO 11-03 17:31:27 [gpu_model_runner.py:2902] Model loading took 4.2624 GiB and 0.539903 seconds 2025-11-03T17:31:27.3927805Z (Worker_TP1 pid=59656) INFO 11-03 17:31:27 [gpu_model_runner.py:2902] Model loading took 4.2624 GiB and 0.544008 seconds 2025-11-03T17:31:35.3871931Z (Worker_TP1 pid=59656) INFO 11-03 17:31:35 [backends.py:594] Using cache directory: /var/lib/jenkins/.cache/vllm/torch_compile_cache/c23588ba3d/rank_1_0/backbone for vLLM's torch.compile 2025-11-03T17:31:35.3875459Z (Worker_TP1 pid=59656) INFO 11-03 17:31:35 [backends.py:608] Dynamo bytecode transform time: 7.71 s 2025-11-03T17:31:35.6153485Z (Worker_TP0 pid=59655) INFO 11-03 17:31:35 [backends.py:594] Using cache directory: /var/lib/jenkins/.cache/vllm/torch_compile_cache/c23588ba3d/rank_0_0/backbone for vLLM's torch.compile 2025-11-03T17:31:35.6158208Z (Worker_TP0 pid=59655) INFO 11-03 17:31:35 [backends.py:608] Dynamo bytecode transform time: 7.93 s 2025-11-03T17:31:38.5357089Z (Worker_TP1 pid=59656) INFO 11-03 17:31:38 [backends.py:214] Cache the graph for dynamic shape for later use 2025-11-03T17:31:38.7237372Z (Worker_TP0 pid=59655) INFO 11-03 17:31:38 [backends.py:214] Cache the graph for dynamic shape for later use 2025-11-03T17:32:03.6532208Z (Worker_TP0 pid=59655) INFO 11-03 17:32:03 [backends.py:241] Compiling a graph for dynamic shape takes 27.50 s 2025-11-03T17:32:04.5918038Z (Worker_TP1 pid=59656) INFO 11-03 17:32:04 [backends.py:241] Compiling a graph for dynamic shape takes 28.66 s 2025-11-03T17:32:25.8684633Z (Worker_TP0 pid=59655) INFO 11-03 17:32:25 [gpu_worker.py:314] Available KV cache memory: 15.29 GiB 2025-11-03T17:32:25.9403602Z (Worker_TP1 pid=59656) INFO 11-03 17:32:25 [gpu_worker.py:314] Available KV cache memory: 15.29 GiB 2025-11-03T17:32:26.3016336Z (EngineCore_DP0 pid=59536) INFO 11-03 17:32:26 [kv_cache_utils.py:1199] GPU KV cache size: 250,528 tokens 2025-11-03T17:32:26.3017205Z (EngineCore_DP0 pid=59536) INFO 11-03 17:32:26 [kv_cache_utils.py:1204] Maximum concurrency for 2,048 tokens per request: 122.33x 2025-11-03T17:32:26.3018165Z (EngineCore_DP0 pid=59536) INFO 11-03 17:32:26 [kv_cache_utils.py:1199] GPU KV cache size: 250,528 tokens 2025-11-03T17:32:26.3021034Z (EngineCore_DP0 pid=59536) INFO 11-03 17:32:26 [kv_cache_utils.py:1204] Maximum concurrency for 2,048 tokens per request: 122.33x 2025-11-03T17:32:26.3129787Z (Worker_TP1 pid=59656) INFO 11-03 17:32:26 [gpu_worker.py:351] Compile and warming up model for size 8 2025-11-03T17:32:26.3130682Z (Worker_TP0 pid=59655) INFO 11-03 17:32:26 [gpu_worker.py:351] Compile and warming up model for size 8 2025-11-03T17:32:27.5454102Z (Worker_TP0 pid=59655) INFO 11-03 17:32:27 [backends.py:216] Cache the graph of shape 8 for later use 2025-11-03T17:32:27.5525420Z (Worker_TP1 pid=59656) INFO 11-03 17:32:27 [backends.py:216] Cache the graph of shape 8 for later use 2025-11-03T17:32:58.7067450Z (Worker_TP1 pid=59656) INFO 11-03 17:32:58 [backends.py:243] Compiling a graph for shape 8 takes 32.39 s 2025-11-03T17:32:58.7363567Z (Worker_TP0 pid=59655) INFO 11-03 17:32:58 [backends.py:243] Compiling a graph for shape 8 takes 32.42 s 2025-11-03T17:32:59.0141620Z (Worker_TP0 pid=59655) INFO 11-03 17:32:59 [gpu_worker.py:351] Compile and warming up model for size 4 2025-11-03T17:32:59.0160985Z (Worker_TP1 pid=59656) INFO 11-03 17:32:59 [gpu_worker.py:351] Compile and warming up model for size 4 2025-11-03T17:33:00.2071028Z (Worker_TP0 pid=59655) INFO 11-03 17:33:00 [backends.py:216] Cache the graph of shape 4 for later use 2025-11-03T17:33:00.2421237Z (Worker_TP1 pid=59656) INFO 11-03 17:33:00 [backends.py:216] Cache the graph of shape 4 for later use 2025-11-03T17:33:26.3135359Z (EngineCore_DP0 pid=59536) INFO 11-03 17:33:26 [shm_broadcast.py:482] No available shared memory broadcast block found in 60 seconds. This typically happens when some processes are hanging or doing some time-consuming work (e.g. compilation). 2025-11-03T17:33:40.4280485Z (Worker_TP0 pid=59655) INFO 11-03 17:33:40 [backends.py:243] Compiling a graph for shape 4 takes 41.41 s 2025-11-03T17:33:40.4281671Z (Worker_TP1 pid=59656) INFO 11-03 17:33:40 [backends.py:243] Compiling a graph for shape 4 takes 41.41 s 2025-11-03T17:33:40.4301076Z (Worker_TP0 pid=59655) INFO 11-03 17:33:40 [monitor.py:33] torch.compile takes 109.27 s in total 2025-11-03T17:33:40.4303599Z (Worker_TP1 pid=59656) INFO 11-03 17:33:40 [monitor.py:33] torch.compile takes 110.18 s in total 2025-11-03T17:33:40.4570641Z (EngineCore_DP0 pid=59536) INFO 11-03 17:33:40 [core.py:240] init engine (profile, create kv cache, warmup model) took 132.98 seconds 2025-11-03T17:33:40.9199234Z (EngineCore_DP0 pid=59536) INFO 11-03 17:33:40 [vllm.py:354] Cudagraph is disabled under eager mode 2025-11-03T17:33:41.1048027Z (APIServer pid=59412) INFO 11-03 17:33:41 [loggers.py:172] Engine 000: vllm cache_config_info with initialization after num_gpu_blocks is: 15658 2025-11-03T17:33:41.1144409Z (EngineCore_DP0 pid=59536) INFO 11-03 17:33:41 [gc_utils.py:40] GC Debug Config. enabled:False,top_objects:-1 2025-11-03T17:33:41.1357854Z (APIServer pid=59412) INFO 11-03 17:33:41 [api_server.py:1628] Supported tasks: ['generate'] 2025-11-03T17:33:41.1628355Z (APIServer pid=59412) WARNING 11-03 17:33:41 [model.py:1581] Default sampling parameters have been overridden by the model's Hugging Face generation config recommended from the model creator. If this is not intended, please relaunch vLLM instance with `--generation-config vllm`. 2025-11-03T17:33:41.1631319Z (APIServer pid=59412) INFO 11-03 17:33:41 [serving_responses.py:157] Using default chat sampling params from model: {'temperature': 0.6, 'top_p': 0.9} 2025-11-03T17:33:41.1893361Z (APIServer pid=59412) INFO 11-03 17:33:41 [serving_chat.py:159] Using default chat sampling params from model: {'temperature': 0.6, 'top_p': 0.9} 2025-11-03T17:33:41.2171836Z (APIServer pid=59412) INFO 11-03 17:33:41 [serving_completion.py:67] Using default completion sampling params from model: {'temperature': 0.6, 'top_p': 0.9} 2025-11-03T17:33:41.2174203Z (APIServer pid=59412) INFO 11-03 17:33:41 [api_server.py:1939] Starting vLLM API server 0 on http://0.0.0.0:35219 2025-11-03T17:33:41.2175051Z (APIServer pid=59412) INFO 11-03 17:33:41 [launcher.py:38] Available routes are: 2025-11-03T17:33:41.2175978Z (APIServer pid=59412) INFO 11-03 17:33:41 [launcher.py:46] Route: /openapi.json, Methods: GET, HEAD 2025-11-03T17:33:41.2176745Z (APIServer pid=59412) INFO 11-03 17:33:41 [launcher.py:46] Route: /docs, Methods: GET, HEAD 2025-11-03T17:33:41.2177331Z (APIServer pid=59412) INFO 11-03 17:33:41 [launcher.py:46] Route: /docs/oauth2-redirect, Methods: GET, HEAD 2025-11-03T17:33:41.2177885Z (APIServer pid=59412) INFO 11-03 17:33:41 [launcher.py:46] Route: /redoc, Methods: GET, HEAD 2025-11-03T17:33:41.2178413Z (APIServer pid=59412) INFO 11-03 17:33:41 [launcher.py:46] Route: /health, Methods: GET 2025-11-03T17:33:41.2178908Z (APIServer pid=59412) INFO 11-03 17:33:41 [launcher.py:46] Route: /load, Methods: GET 2025-11-03T17:33:41.2179417Z (APIServer pid=59412) INFO 11-03 17:33:41 [launcher.py:46] Route: /ping, Methods: POST 2025-11-03T17:33:41.2179916Z (APIServer pid=59412) INFO 11-03 17:33:41 [launcher.py:46] Route: /ping, Methods: GET 2025-11-03T17:33:41.2180421Z (APIServer pid=59412) INFO 11-03 17:33:41 [launcher.py:46] Route: /tokenize, Methods: POST 2025-11-03T17:33:41.2180957Z (APIServer pid=59412) INFO 11-03 17:33:41 [launcher.py:46] Route: /detokenize, Methods: POST 2025-11-03T17:33:41.2181487Z (APIServer pid=59412) INFO 11-03 17:33:41 [launcher.py:46] Route: /v1/models, Methods: GET 2025-11-03T17:33:41.2181994Z (APIServer pid=59412) INFO 11-03 17:33:41 [launcher.py:46] Route: /version, Methods: GET 2025-11-03T17:33:41.2182540Z (APIServer pid=59412) INFO 11-03 17:33:41 [launcher.py:46] Route: /v1/responses, Methods: POST 2025-11-03T17:33:41.2183116Z (APIServer pid=59412) INFO 11-03 17:33:41 [launcher.py:46] Route: /v1/responses/{response_id}, Methods: GET 2025-11-03T17:33:41.2183730Z (APIServer pid=59412) INFO 11-03 17:33:41 [launcher.py:46] Route: /v1/responses/{response_id}/cancel, Methods: POST 2025-11-03T17:33:41.2184343Z (APIServer pid=59412) INFO 11-03 17:33:41 [launcher.py:46] Route: /v1/chat/completions, Methods: POST 2025-11-03T17:33:41.2184902Z (APIServer pid=59412) INFO 11-03 17:33:41 [launcher.py:46] Route: /v1/completions, Methods: POST 2025-11-03T17:33:41.2185441Z (APIServer pid=59412) INFO 11-03 17:33:41 [launcher.py:46] Route: /v1/embeddings, Methods: POST 2025-11-03T17:33:41.2185958Z (APIServer pid=59412) INFO 11-03 17:33:41 [launcher.py:46] Route: /pooling, Methods: POST 2025-11-03T17:33:41.2186586Z (APIServer pid=59412) INFO 11-03 17:33:41 [launcher.py:46] Route: /classify, Methods: POST 2025-11-03T17:33:41.2187111Z (APIServer pid=59412) INFO 11-03 17:33:41 [launcher.py:46] Route: /score, Methods: POST 2025-11-03T17:33:41.2187614Z (APIServer pid=59412) INFO 11-03 17:33:41 [launcher.py:46] Route: /v1/score, Methods: POST 2025-11-03T17:33:41.2188165Z (APIServer pid=59412) INFO 11-03 17:33:41 [launcher.py:46] Route: /v1/audio/transcriptions, Methods: POST 2025-11-03T17:33:41.2188740Z (APIServer pid=59412) INFO 11-03 17:33:41 [launcher.py:46] Route: /v1/audio/translations, Methods: POST 2025-11-03T17:33:41.2189269Z (APIServer pid=59412) INFO 11-03 17:33:41 [launcher.py:46] Route: /rerank, Methods: POST 2025-11-03T17:33:41.2189780Z (APIServer pid=59412) INFO 11-03 17:33:41 [launcher.py:46] Route: /v1/rerank, Methods: POST 2025-11-03T17:33:41.2190293Z (APIServer pid=59412) INFO 11-03 17:33:41 [launcher.py:46] Route: /v2/rerank, Methods: POST 2025-11-03T17:33:41.2190954Z (APIServer pid=59412) INFO 11-03 17:33:41 [launcher.py:46] Route: /scale_elastic_ep, Methods: POST 2025-11-03T17:33:41.2191523Z (APIServer pid=59412) INFO 11-03 17:33:41 [launcher.py:46] Route: /is_scaling_elastic_ep, Methods: POST 2025-11-03T17:33:41.2192076Z (APIServer pid=59412) INFO 11-03 17:33:41 [launcher.py:46] Route: /invocations, Methods: POST 2025-11-03T17:33:41.2192635Z (APIServer pid=59412) INFO 11-03 17:33:41 [launcher.py:46] Route: /metrics, Methods: GET 2025-11-03T17:33:41.2337997Z (APIServer pid=59412) INFO: Started server process [59412] 2025-11-03T17:33:41.2339150Z (APIServer pid=59412) INFO: Waiting for application startup. 2025-11-03T17:33:41.4571263Z (APIServer pid=59412) INFO: Application startup complete. 2025-11-03T17:33:41.6650132Z (APIServer pid=59412) INFO: 127.0.0.1:56278 - "GET /health HTTP/1.1" 200 OK 2025-11-03T17:33:41.9033371Z (APIServer pid=59412) INFO: 127.0.0.1:56286 - "GET /v1/models HTTP/1.1" 200 OK 2025-11-03T17:33:42.0662514Z (APIServer pid=59412) INFO: 127.0.0.1:56286 - "POST /v1/completions HTTP/1.1" 200 OK 2025-11-03T17:33:42.1836375Z (APIServer pid=59412) INFO: 127.0.0.1:56286 - "POST /v1/completions HTTP/1.1" 200 OK 2025-11-03T17:33:42.3762473Z (APIServer pid=59412) INFO: 127.0.0.1:56286 - "POST /v1/completions HTTP/1.1" 200 OK 2025-11-03T17:33:42.5004448Z (APIServer pid=59412) INFO: 127.0.0.1:56286 - "POST /v1/completions HTTP/1.1" 200 OK 2025-11-03T17:33:42.6177471Z (APIServer pid=59412) INFO: 127.0.0.1:56286 - "POST /v1/completions HTTP/1.1" 200 OK 2025-11-03T17:33:42.6249775Z (APIServer pid=59412) INFO: 127.0.0.1:56286 - "POST /v1/completions HTTP/1.1" 200 OK 2025-11-03T17:33:42.7459299Z (APIServer pid=59412) INFO 11-03 17:33:42 [launcher.py:110] Shutting down FastAPI HTTP server. 2025-11-03T17:33:42.7462391Z (Worker_TP0 pid=59655) INFO 11-03 17:33:42 [multiproc_executor.py:583] Parent process exited, terminating worker 2025-11-03T17:33:42.7463260Z (Worker_TP0 pid=59655) INFO 11-03 17:33:42 [multiproc_executor.py:624] WorkerProc shutting down. 2025-11-03T17:33:42.7466646Z (Worker_TP1 pid=59656) INFO 11-03 17:33:42 [multiproc_executor.py:583] Parent process exited, terminating worker 2025-11-03T17:33:42.7468141Z (Worker_TP1 pid=59656) INFO 11-03 17:33:42 [multiproc_executor.py:624] WorkerProc shutting down. 2025-11-03T17:33:45.9631754Z (APIServer pid=59412) INFO: Shutting down 2025-11-03T17:33:46.0627819Z (APIServer pid=59412) INFO: Waiting for application shutdown. 2025-11-03T17:33:46.0628515Z (APIServer pid=59412) INFO: Application shutdown complete. 2025-11-03T17:33:46.3966550Z INFO 11-03 17:33:46 [model.py:653] Resolved architecture: LlamaForCausalLM 2025-11-03T17:33:46.3967060Z WARNING 11-03 17:33:46 [model.py:1942] Casting torch.bfloat16 to torch.float16. 2025-11-03T17:33:46.3967495Z INFO 11-03 17:33:46 [model.py:1714] Using max model len 2048 2025-11-03T17:33:46.3973816Z Launching RemoteOpenAIServer with: vllm serve RedHatAI/Meta-Llama-3.1-8B-Instruct-FP8 --dtype float16 --max-model-len 2048 --max-num-seqs 8 --enforce-eager --tokenizer-mode auto --tensor-parallel-size 2 --distributed-executor-backend mp --load-format dummy --port 36115 --seed 0 2025-11-03T17:33:46.9168998Z /opt/conda/envs/py_3.12/lib/python3.12/site-packages/torch/cuda/__init__.py:65: FutureWarning: The pynvml package is deprecated. Please install nvidia-ml-py instead. If you did not install pynvml directly, please report this to the maintainers of the package that installed pynvml for you. 2025-11-03T17:33:46.9170198Z import pynvml # type: ignore[import] 2025-11-03T17:33:49.1619496Z INFO 11-03 17:33:49 [__init__.py:224] Automatically detected platform cuda. 2025-11-03T17:33:52.8347005Z (APIServer pid=63676) INFO 11-03 17:33:52 [api_server.py:1870] vLLM API server version 0.11.0rc2.dev389+ge51928192.d20251103 2025-11-03T17:33:52.8367209Z (APIServer pid=63676) INFO 11-03 17:33:52 [utils.py:239] non-default args: {'model_tag': 'RedHatAI/Meta-Llama-3.1-8B-Instruct-FP8', 'port': 36115, 'model': 'RedHatAI/Meta-Llama-3.1-8B-Instruct-FP8', 'dtype': 'float16', 'seed': 0, 'max_model_len': 2048, 'enforce_eager': True, 'load_format': 'dummy', 'distributed_executor_backend': 'mp', 'tensor_parallel_size': 2, 'max_num_seqs': 8} 2025-11-03T17:33:53.2308017Z (APIServer pid=63676) INFO 11-03 17:33:53 [model.py:653] Resolved architecture: LlamaForCausalLM 2025-11-03T17:33:53.2309336Z (APIServer pid=63676) `torch_dtype` is deprecated! Use `dtype` instead! 2025-11-03T17:33:53.2311065Z (APIServer pid=63676) WARNING 11-03 17:33:53 [model.py:1942] Casting torch.bfloat16 to torch.float16. 2025-11-03T17:33:53.2312419Z (APIServer pid=63676) INFO 11-03 17:33:53 [model.py:1714] Using max model len 2048 2025-11-03T17:33:53.5588767Z (APIServer pid=63676) INFO 11-03 17:33:53 [scheduler.py:225] Chunked prefill is enabled with max_num_batched_tokens=2048. 2025-11-03T17:33:53.5961146Z (APIServer pid=63676) INFO 11-03 17:33:53 [vllm.py:354] Cudagraph is disabled under eager mode 2025-11-03T17:33:54.5986220Z /opt/conda/envs/py_3.12/lib/python3.12/site-packages/torch/cuda/__init__.py:65: FutureWarning: The pynvml package is deprecated. Please install nvidia-ml-py instead. If you did not install pynvml directly, please report this to the maintainers of the package that installed pynvml for you. 2025-11-03T17:33:54.5987353Z import pynvml # type: ignore[import] 2025-11-03T17:33:56.9066706Z INFO 11-03 17:33:56 [__init__.py:224] Automatically detected platform cuda. 2025-11-03T17:34:00.7313031Z (EngineCore_DP0 pid=63799) INFO 11-03 17:34:00 [core.py:727] Waiting for init message from front-end. 2025-11-03T17:34:00.7571058Z (EngineCore_DP0 pid=63799) INFO 11-03 17:34:00 [core.py:94] Initializing a V1 LLM engine (v0.11.0rc2.dev389+ge51928192.d20251103) with config: model='RedHatAI/Meta-Llama-3.1-8B-Instruct-FP8', speculative_config=None, tokenizer='RedHatAI/Meta-Llama-3.1-8B-Instruct-FP8', skip_tokenizer_init=False, tokenizer_mode=auto, revision=None, tokenizer_revision=None, trust_remote_code=False, dtype=torch.float16, max_seq_len=2048, download_dir=None, load_format=dummy, tensor_parallel_size=2, pipeline_parallel_size=1, data_parallel_size=1, disable_custom_all_reduce=False, quantization=compressed-tensors, enforce_eager=True, kv_cache_dtype=auto, device_config=cuda, structured_outputs_config=StructuredOutputsConfig(backend='auto', disable_fallback=False, disable_any_whitespace=False, disable_additional_properties=False, reasoning_parser=''), observability_config=ObservabilityConfig(show_hidden_metrics_for_version=None, otlp_traces_endpoint=None, collect_detailed_traces=None), seed=0, served_model_name=RedHatAI/Meta-Llama-3.1-8B-Instruct-FP8, enable_prefix_caching=True, chunked_prefill_enabled=True, pooler_config=None, compilation_config={'level': 0, 'debug_dump_path': None, 'cache_dir': '', 'backend': '', 'custom_ops': [], 'splitting_ops': None, 'use_inductor': True, 'compile_sizes': [], 'inductor_compile_config': {'enable_auto_functionalized_v2': False}, 'inductor_passes': {}, 'cudagraph_mode': , 'use_cudagraph': False, 'cudagraph_num_of_warmups': 0, 'cudagraph_capture_sizes': [], 'cudagraph_copy_inputs': False, 'full_cuda_graph': False, 'use_inductor_graph_partition': False, 'pass_config': {}, 'max_capture_size': 0, 'local_cache_dir': None} 2025-11-03T17:34:00.7577036Z (EngineCore_DP0 pid=63799) WARNING 11-03 17:34:00 [multiproc_executor.py:744] Reducing Torch parallelism from 24 threads to 1 to avoid unnecessary CPU contention. Set OMP_NUM_THREADS in the external environment to tune this value as needed. 2025-11-03T17:34:00.7578682Z (EngineCore_DP0 pid=63799) INFO 11-03 17:34:00 [shm_broadcast.py:302] vLLM message queue communication handle: Handle(local_reader_ranks=[0, 1], buffer_handle=(2, 16777216, 10, 'psm_16d831a6'), local_subscribe_addr='ipc:///tmp/ac48176f-7fbb-4281-a37d-2e24fbb348e9', remote_subscribe_addr=None, remote_addr_ipv6=False) 2025-11-03T17:34:01.2444135Z /opt/conda/envs/py_3.12/lib/python3.12/site-packages/torch/cuda/__init__.py:65: FutureWarning: The pynvml package is deprecated. Please install nvidia-ml-py instead. If you did not install pynvml directly, please report this to the maintainers of the package that installed pynvml for you. 2025-11-03T17:34:01.2445349Z import pynvml # type: ignore[import] 2025-11-03T17:34:01.2584906Z /opt/conda/envs/py_3.12/lib/python3.12/site-packages/torch/cuda/__init__.py:65: FutureWarning: The pynvml package is deprecated. Please install nvidia-ml-py instead. If you did not install pynvml directly, please report this to the maintainers of the package that installed pynvml for you. 2025-11-03T17:34:01.2586041Z import pynvml # type: ignore[import] 2025-11-03T17:34:03.6036047Z INFO 11-03 17:34:03 [__init__.py:224] Automatically detected platform cuda. 2025-11-03T17:34:03.6414930Z INFO 11-03 17:34:03 [__init__.py:224] Automatically detected platform cuda. 2025-11-03T17:34:08.4125549Z INFO 11-03 17:34:08 [shm_broadcast.py:302] vLLM message queue communication handle: Handle(local_reader_ranks=[0], buffer_handle=(1, 10485760, 10, 'psm_37486732'), local_subscribe_addr='ipc:///tmp/8f095ce5-03db-4245-9d4d-b941ef3f96bd', remote_subscribe_addr=None, remote_addr_ipv6=False) 2025-11-03T17:34:08.4527644Z INFO 11-03 17:34:08 [shm_broadcast.py:302] vLLM message queue communication handle: Handle(local_reader_ranks=[0], buffer_handle=(1, 10485760, 10, 'psm_528eeff8'), local_subscribe_addr='ipc:///tmp/d7bc1fd4-861a-4d7b-887c-80f1bf4ed307', remote_subscribe_addr=None, remote_addr_ipv6=False) 2025-11-03T17:34:08.6286512Z [Gloo] Rank 0 is connected to 1 peer ranks. Expected number of connected peer ranks is : 1 2025-11-03T17:34:08.6287448Z [Gloo] Rank 1 is connected to 1 peer ranks. Expected number of connected peer ranks is : 1 2025-11-03T17:34:08.6328146Z [Gloo] Rank 0 is connected to 1 peer ranks. Expected number of connected peer ranks is : 1 2025-11-03T17:34:08.6328810Z [Gloo] Rank 1 is connected to 1 peer ranks. Expected number of connected peer ranks is : 1 2025-11-03T17:34:08.6370160Z INFO 11-03 17:34:08 [__init__.py:1146] Found nccl from library libnccl.so.2 2025-11-03T17:34:08.6370923Z INFO 11-03 17:34:08 [__init__.py:1146] Found nccl from library libnccl.so.2 2025-11-03T17:34:08.6372533Z INFO 11-03 17:34:08 [pynccl.py:109] vLLM is using nccl==2.27.5 2025-11-03T17:34:08.6373388Z INFO 11-03 17:34:08 [pynccl.py:109] vLLM is using nccl==2.27.5 2025-11-03T17:34:08.8104961Z WARNING 11-03 17:34:08 [symm_mem.py:60] SymmMemCommunicator: Device capability 8.9 not supported, communicator is not available. 2025-11-03T17:34:08.8105823Z WARNING 11-03 17:34:08 [symm_mem.py:60] SymmMemCommunicator: Device capability 8.9 not supported, communicator is not available. 2025-11-03T17:34:08.8132533Z INFO 11-03 17:34:08 [custom_all_reduce.py:37] Skipping P2P check and trusting the driver's P2P report. 2025-11-03T17:34:08.8133160Z INFO 11-03 17:34:08 [custom_all_reduce.py:37] Skipping P2P check and trusting the driver's P2P report. 2025-11-03T17:34:08.8134160Z WARNING 11-03 17:34:08 [custom_all_reduce.py:162] Custom allreduce is disabled because your platform lacks GPU P2P capability or P2P test failed. To silence this warning, specify disable_custom_all_reduce=True explicitly. 2025-11-03T17:34:08.8135515Z WARNING 11-03 17:34:08 [custom_all_reduce.py:162] Custom allreduce is disabled because your platform lacks GPU P2P capability or P2P test failed. To silence this warning, specify disable_custom_all_reduce=True explicitly. 2025-11-03T17:34:08.9024296Z INFO 11-03 17:34:08 [shm_broadcast.py:302] vLLM message queue communication handle: Handle(local_reader_ranks=[1], buffer_handle=(1, 4194304, 6, 'psm_6ec3cdf3'), local_subscribe_addr='ipc:///tmp/bfbcfc05-7216-41eb-b46c-2a88813303ef', remote_subscribe_addr=None, remote_addr_ipv6=False) 2025-11-03T17:34:08.9061682Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:34:08.9063314Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:34:08.9070037Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:34:08.9078115Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:34:08.9622244Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:34:08.9630522Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:34:08.9641965Z [Gloo] Rank 1 is connected to 1 peer ranks. Expected number of connected peer ranks is : 1 2025-11-03T17:34:08.9642522Z [Gloo] Rank 0 is connected to 1 peer ranks. Expected number of connected peer ranks is : 1 2025-11-03T17:34:08.9645577Z INFO 11-03 17:34:08 [__init__.py:1146] Found nccl from library libnccl.so.2 2025-11-03T17:34:08.9645982Z INFO 11-03 17:34:08 [__init__.py:1146] Found nccl from library libnccl.so.2 2025-11-03T17:34:08.9646332Z INFO 11-03 17:34:08 [pynccl.py:109] vLLM is using nccl==2.27.5 2025-11-03T17:34:08.9646712Z INFO 11-03 17:34:08 [pynccl.py:109] vLLM is using nccl==2.27.5 2025-11-03T17:34:08.9990053Z INFO 11-03 17:34:08 [parallel_state.py:1231] rank 0 in world size 2 is assigned as DP rank 0, PP rank 0, TP rank 0, EP rank 0 2025-11-03T17:34:08.9990917Z INFO 11-03 17:34:08 [parallel_state.py:1231] rank 1 in world size 2 is assigned as DP rank 0, PP rank 0, TP rank 1, EP rank 1 2025-11-03T17:34:09.2396609Z INFO 11-03 17:34:09 [topk_topp_sampler.py:59] Using FlashInfer for top-p & top-k sampling. 2025-11-03T17:34:09.2502944Z INFO 11-03 17:34:09 [topk_topp_sampler.py:59] Using FlashInfer for top-p & top-k sampling. 2025-11-03T17:34:09.2775540Z (Worker_TP0 pid=63918) INFO 11-03 17:34:09 [gpu_model_runner.py:2840] Starting to load model RedHatAI/Meta-Llama-3.1-8B-Instruct-FP8... 2025-11-03T17:34:09.2869657Z (Worker_TP1 pid=63919) INFO 11-03 17:34:09 [gpu_model_runner.py:2840] Starting to load model RedHatAI/Meta-Llama-3.1-8B-Instruct-FP8... 2025-11-03T17:34:09.4854987Z (Worker_TP0 pid=63918) INFO 11-03 17:34:09 [gpu_model_runner.py:2870] Loading model from scratch... 2025-11-03T17:34:09.4976633Z (Worker_TP1 pid=63919) INFO 11-03 17:34:09 [gpu_model_runner.py:2870] Loading model from scratch... 2025-11-03T17:34:09.5774927Z (Worker_TP0 pid=63918) INFO 11-03 17:34:09 [cuda.py:404] Using Flash Attention backend on V1 engine. 2025-11-03T17:34:09.5893428Z (Worker_TP1 pid=63919) INFO 11-03 17:34:09 [cuda.py:404] Using Flash Attention backend on V1 engine. 2025-11-03T17:34:10.3923704Z (Worker_TP0 pid=63918) INFO 11-03 17:34:10 [gpu_model_runner.py:2902] Model loading took 4.2624 GiB and 0.489344 seconds 2025-11-03T17:34:10.4015395Z (Worker_TP1 pid=63919) INFO 11-03 17:34:10 [gpu_model_runner.py:2902] Model loading took 4.2624 GiB and 0.489864 seconds 2025-11-03T17:34:12.7853803Z (Worker_TP0 pid=63918) INFO 11-03 17:34:12 [gpu_worker.py:314] Available KV cache memory: 15.32 GiB 2025-11-03T17:34:12.7921050Z (Worker_TP1 pid=63919) INFO 11-03 17:34:12 [gpu_worker.py:314] Available KV cache memory: 15.32 GiB 2025-11-03T17:34:13.0018029Z (EngineCore_DP0 pid=63799) INFO 11-03 17:34:13 [kv_cache_utils.py:1199] GPU KV cache size: 250,976 tokens 2025-11-03T17:34:13.0018978Z (EngineCore_DP0 pid=63799) INFO 11-03 17:34:13 [kv_cache_utils.py:1204] Maximum concurrency for 2,048 tokens per request: 122.55x 2025-11-03T17:34:13.0020181Z (EngineCore_DP0 pid=63799) INFO 11-03 17:34:13 [kv_cache_utils.py:1199] GPU KV cache size: 250,976 tokens 2025-11-03T17:34:13.0020876Z (EngineCore_DP0 pid=63799) INFO 11-03 17:34:13 [kv_cache_utils.py:1204] Maximum concurrency for 2,048 tokens per request: 122.55x 2025-11-03T17:34:13.1348436Z (EngineCore_DP0 pid=63799) INFO 11-03 17:34:13 [core.py:240] init engine (profile, create kv cache, warmup model) took 2.65 seconds 2025-11-03T17:34:13.6176360Z (EngineCore_DP0 pid=63799) INFO 11-03 17:34:13 [vllm.py:354] Cudagraph is disabled under eager mode 2025-11-03T17:34:13.8023453Z (APIServer pid=63676) INFO 11-03 17:34:13 [loggers.py:172] Engine 000: vllm cache_config_info with initialization after num_gpu_blocks is: 15686 2025-11-03T17:34:13.8159053Z (EngineCore_DP0 pid=63799) INFO 11-03 17:34:13 [gc_utils.py:40] GC Debug Config. enabled:False,top_objects:-1 2025-11-03T17:34:13.8374160Z (APIServer pid=63676) INFO 11-03 17:34:13 [api_server.py:1628] Supported tasks: ['generate'] 2025-11-03T17:34:13.8712867Z (APIServer pid=63676) WARNING 11-03 17:34:13 [model.py:1581] Default sampling parameters have been overridden by the model's Hugging Face generation config recommended from the model creator. If this is not intended, please relaunch vLLM instance with `--generation-config vllm`. 2025-11-03T17:34:13.8714407Z (APIServer pid=63676) INFO 11-03 17:34:13 [serving_responses.py:157] Using default chat sampling params from model: {'temperature': 0.6, 'top_p': 0.9} 2025-11-03T17:34:13.9078069Z (APIServer pid=63676) INFO 11-03 17:34:13 [serving_chat.py:159] Using default chat sampling params from model: {'temperature': 0.6, 'top_p': 0.9} 2025-11-03T17:34:13.9347653Z (APIServer pid=63676) INFO 11-03 17:34:13 [serving_completion.py:67] Using default completion sampling params from model: {'temperature': 0.6, 'top_p': 0.9} 2025-11-03T17:34:13.9349033Z (APIServer pid=63676) INFO 11-03 17:34:13 [api_server.py:1939] Starting vLLM API server 0 on http://0.0.0.0:36115 2025-11-03T17:34:13.9350151Z (APIServer pid=63676) INFO 11-03 17:34:13 [launcher.py:38] Available routes are: 2025-11-03T17:34:13.9351312Z (APIServer pid=63676) INFO 11-03 17:34:13 [launcher.py:46] Route: /openapi.json, Methods: GET, HEAD 2025-11-03T17:34:13.9358415Z (APIServer pid=63676) INFO 11-03 17:34:13 [launcher.py:46] Route: /docs, Methods: GET, HEAD 2025-11-03T17:34:13.9359041Z (APIServer pid=63676) INFO 11-03 17:34:13 [launcher.py:46] Route: /docs/oauth2-redirect, Methods: GET, HEAD 2025-11-03T17:34:13.9359636Z (APIServer pid=63676) INFO 11-03 17:34:13 [launcher.py:46] Route: /redoc, Methods: GET, HEAD 2025-11-03T17:34:13.9360156Z (APIServer pid=63676) INFO 11-03 17:34:13 [launcher.py:46] Route: /health, Methods: GET 2025-11-03T17:34:13.9360823Z (APIServer pid=63676) INFO 11-03 17:34:13 [launcher.py:46] Route: /load, Methods: GET 2025-11-03T17:34:13.9361322Z (APIServer pid=63676) INFO 11-03 17:34:13 [launcher.py:46] Route: /ping, Methods: POST 2025-11-03T17:34:13.9361815Z (APIServer pid=63676) INFO 11-03 17:34:13 [launcher.py:46] Route: /ping, Methods: GET 2025-11-03T17:34:13.9362329Z (APIServer pid=63676) INFO 11-03 17:34:13 [launcher.py:46] Route: /tokenize, Methods: POST 2025-11-03T17:34:13.9362860Z (APIServer pid=63676) INFO 11-03 17:34:13 [launcher.py:46] Route: /detokenize, Methods: POST 2025-11-03T17:34:13.9363376Z (APIServer pid=63676) INFO 11-03 17:34:13 [launcher.py:46] Route: /v1/models, Methods: GET 2025-11-03T17:34:13.9363873Z (APIServer pid=63676) INFO 11-03 17:34:13 [launcher.py:46] Route: /version, Methods: GET 2025-11-03T17:34:13.9364398Z (APIServer pid=63676) INFO 11-03 17:34:13 [launcher.py:46] Route: /v1/responses, Methods: POST 2025-11-03T17:34:13.9364974Z (APIServer pid=63676) INFO 11-03 17:34:13 [launcher.py:46] Route: /v1/responses/{response_id}, Methods: GET 2025-11-03T17:34:13.9365730Z (APIServer pid=63676) INFO 11-03 17:34:13 [launcher.py:46] Route: /v1/responses/{response_id}/cancel, Methods: POST 2025-11-03T17:34:13.9366352Z (APIServer pid=63676) INFO 11-03 17:34:13 [launcher.py:46] Route: /v1/chat/completions, Methods: POST 2025-11-03T17:34:13.9366971Z (APIServer pid=63676) INFO 11-03 17:34:13 [launcher.py:46] Route: /v1/completions, Methods: POST 2025-11-03T17:34:13.9367522Z (APIServer pid=63676) INFO 11-03 17:34:13 [launcher.py:46] Route: /v1/embeddings, Methods: POST 2025-11-03T17:34:13.9368051Z (APIServer pid=63676) INFO 11-03 17:34:13 [launcher.py:46] Route: /pooling, Methods: POST 2025-11-03T17:34:13.9368714Z (APIServer pid=63676) INFO 11-03 17:34:13 [launcher.py:46] Route: /classify, Methods: POST 2025-11-03T17:34:13.9369239Z (APIServer pid=63676) INFO 11-03 17:34:13 [launcher.py:46] Route: /score, Methods: POST 2025-11-03T17:34:13.9369746Z (APIServer pid=63676) INFO 11-03 17:34:13 [launcher.py:46] Route: /v1/score, Methods: POST 2025-11-03T17:34:13.9370301Z (APIServer pid=63676) INFO 11-03 17:34:13 [launcher.py:46] Route: /v1/audio/transcriptions, Methods: POST 2025-11-03T17:34:13.9370895Z (APIServer pid=63676) INFO 11-03 17:34:13 [launcher.py:46] Route: /v1/audio/translations, Methods: POST 2025-11-03T17:34:13.9371442Z (APIServer pid=63676) INFO 11-03 17:34:13 [launcher.py:46] Route: /rerank, Methods: POST 2025-11-03T17:34:13.9371959Z (APIServer pid=63676) INFO 11-03 17:34:13 [launcher.py:46] Route: /v1/rerank, Methods: POST 2025-11-03T17:34:13.9372471Z (APIServer pid=63676) INFO 11-03 17:34:13 [launcher.py:46] Route: /v2/rerank, Methods: POST 2025-11-03T17:34:13.9373008Z (APIServer pid=63676) INFO 11-03 17:34:13 [launcher.py:46] Route: /scale_elastic_ep, Methods: POST 2025-11-03T17:34:13.9373580Z (APIServer pid=63676) INFO 11-03 17:34:13 [launcher.py:46] Route: /is_scaling_elastic_ep, Methods: POST 2025-11-03T17:34:13.9374134Z (APIServer pid=63676) INFO 11-03 17:34:13 [launcher.py:46] Route: /invocations, Methods: POST 2025-11-03T17:34:13.9374651Z (APIServer pid=63676) INFO 11-03 17:34:13 [launcher.py:46] Route: /metrics, Methods: GET 2025-11-03T17:34:13.9506525Z (APIServer pid=63676) INFO: Started server process [63676] 2025-11-03T17:34:13.9507132Z (APIServer pid=63676) INFO: Waiting for application startup. 2025-11-03T17:34:14.1649227Z (APIServer pid=63676) INFO: Application startup complete. 2025-11-03T17:34:14.4705988Z (APIServer pid=63676) INFO: 127.0.0.1:39952 - "GET /health HTTP/1.1" 200 OK 2025-11-03T17:34:14.4797687Z (APIServer pid=63676) INFO: 127.0.0.1:39960 - "GET /v1/models HTTP/1.1" 200 OK 2025-11-03T17:34:14.7109198Z (APIServer pid=63676) INFO: 127.0.0.1:39960 - "POST /v1/completions HTTP/1.1" 200 OK 2025-11-03T17:34:14.8249450Z (APIServer pid=63676) INFO: 127.0.0.1:39960 - "POST /v1/completions HTTP/1.1" 200 OK 2025-11-03T17:34:14.9763953Z (APIServer pid=63676) INFO: 127.0.0.1:39960 - "POST /v1/completions HTTP/1.1" 200 OK 2025-11-03T17:34:15.1266190Z (APIServer pid=63676) INFO: 127.0.0.1:39960 - "POST /v1/completions HTTP/1.1" 200 OK 2025-11-03T17:34:15.2666558Z (APIServer pid=63676) INFO: 127.0.0.1:39960 - "POST /v1/completions HTTP/1.1" 200 OK 2025-11-03T17:34:15.2743896Z (APIServer pid=63676) INFO: 127.0.0.1:39960 - "POST /v1/completions HTTP/1.1" 200 OK 2025-11-03T17:34:15.3975874Z (APIServer pid=63676) INFO 11-03 17:34:15 [launcher.py:110] Shutting down FastAPI HTTP server. 2025-11-03T17:34:15.3979158Z (Worker_TP0 pid=63918) INFO 11-03 17:34:15 [multiproc_executor.py:583] Parent process exited, terminating worker 2025-11-03T17:34:15.3979972Z (Worker_TP0 pid=63918) INFO 11-03 17:34:15 [multiproc_executor.py:624] WorkerProc shutting down. 2025-11-03T17:34:15.3983941Z (Worker_TP1 pid=63919) INFO 11-03 17:34:15 [multiproc_executor.py:583] Parent process exited, terminating worker 2025-11-03T17:34:15.3984584Z (Worker_TP1 pid=63919) INFO 11-03 17:34:15 [multiproc_executor.py:624] WorkerProc shutting down. 2025-11-03T17:34:17.7145162Z (APIServer pid=63676) INFO: Shutting down 2025-11-03T17:34:17.8152062Z (APIServer pid=63676) INFO: Waiting for application shutdown. 2025-11-03T17:34:17.8153299Z (APIServer pid=63676) INFO: Application shutdown complete. 2025-11-03T17:34:18.4281499Z PASSED 2025-11-03T17:34:18.4426221Z distributed/test_sequence_parallel.py::test_tp_sp_generation[RedHatAI/Meta-Llama-3.1-8B-Instruct-FP8-parallel_setup17-ray-auto-test_options17] Fork a new process to run a test 64122 2025-11-03T17:34:18.4440126Z Fork a new process to run a test 0 2025-11-03T17:34:19.1019406Z INFO 11-03 17:34:19 [model.py:653] Resolved architecture: LlamaForCausalLM 2025-11-03T17:34:19.1020299Z `torch_dtype` is deprecated! Use `dtype` instead! 2025-11-03T17:34:19.1021143Z WARNING 11-03 17:34:19 [model.py:1942] Casting torch.bfloat16 to torch.float16. 2025-11-03T17:34:19.1021949Z INFO 11-03 17:34:19 [model.py:1714] Using max model len 2048 2025-11-03T17:34:19.2095383Z Launching RemoteOpenAIServer with: vllm serve RedHatAI/Meta-Llama-3.1-8B-Instruct-FP8 --dtype float16 --max-model-len 2048 --max-num-seqs 8 --enforce-eager --tokenizer-mode auto --tensor-parallel-size 2 --pipeline-parallel-size 1 --distributed-executor-backend ray --compilation_config {"level": 3, "custom_ops": ["+rms_norm"], "compile_sizes": [4, 8], "pass_config": {"enable_sequence_parallelism": true, "enable_fusion": false, "enable_noop": true}} --load-format dummy --port 49175 --seed 0 2025-11-03T17:34:19.7221312Z /opt/conda/envs/py_3.12/lib/python3.12/site-packages/torch/cuda/__init__.py:65: FutureWarning: The pynvml package is deprecated. Please install nvidia-ml-py instead. If you did not install pynvml directly, please report this to the maintainers of the package that installed pynvml for you. 2025-11-03T17:34:19.7222586Z import pynvml # type: ignore[import] 2025-11-03T17:34:21.9623406Z INFO 11-03 17:34:21 [__init__.py:224] Automatically detected platform cuda. 2025-11-03T17:34:25.6588433Z (APIServer pid=64171) INFO 11-03 17:34:25 [api_server.py:1870] vLLM API server version 0.11.0rc2.dev389+ge51928192.d20251103 2025-11-03T17:34:25.6615365Z (APIServer pid=64171) INFO 11-03 17:34:25 [utils.py:239] non-default args: {'model_tag': 'RedHatAI/Meta-Llama-3.1-8B-Instruct-FP8', 'port': 49175, 'model': 'RedHatAI/Meta-Llama-3.1-8B-Instruct-FP8', 'dtype': 'float16', 'seed': 0, 'max_model_len': 2048, 'enforce_eager': True, 'load_format': 'dummy', 'distributed_executor_backend': 'ray', 'tensor_parallel_size': 2, 'max_num_seqs': 8, 'compilation_config': {'level': 3, 'debug_dump_path': None, 'cache_dir': '', 'backend': '', 'custom_ops': ['+rms_norm'], 'splitting_ops': None, 'use_inductor': True, 'compile_sizes': [4, 8], 'inductor_compile_config': {'enable_auto_functionalized_v2': False}, 'inductor_passes': {}, 'cudagraph_mode': None, 'use_cudagraph': True, 'cudagraph_num_of_warmups': 0, 'cudagraph_capture_sizes': None, 'cudagraph_copy_inputs': False, 'full_cuda_graph': False, 'use_inductor_graph_partition': False, 'pass_config': {'enable_noop': True, 'enable_sequence_parallelism': True}, 'max_capture_size': None, 'local_cache_dir': None}} 2025-11-03T17:34:25.8603974Z (APIServer pid=64171) INFO 11-03 17:34:25 [model.py:653] Resolved architecture: LlamaForCausalLM 2025-11-03T17:34:25.8605327Z (APIServer pid=64171) `torch_dtype` is deprecated! Use `dtype` instead! 2025-11-03T17:34:25.8607075Z (APIServer pid=64171) WARNING 11-03 17:34:25 [model.py:1942] Casting torch.bfloat16 to torch.float16. 2025-11-03T17:34:25.8608647Z (APIServer pid=64171) INFO 11-03 17:34:25 [model.py:1714] Using max model len 2048 2025-11-03T17:34:26.1913013Z (APIServer pid=64171) INFO 11-03 17:34:26 [scheduler.py:225] Chunked prefill is enabled with max_num_batched_tokens=2048. 2025-11-03T17:34:26.2287705Z (APIServer pid=64171) INFO 11-03 17:34:26 [vllm.py:354] Cudagraph is disabled under eager mode 2025-11-03T17:34:27.2042770Z /opt/conda/envs/py_3.12/lib/python3.12/site-packages/torch/cuda/__init__.py:65: FutureWarning: The pynvml package is deprecated. Please install nvidia-ml-py instead. If you did not install pynvml directly, please report this to the maintainers of the package that installed pynvml for you. 2025-11-03T17:34:27.2044201Z import pynvml # type: ignore[import] 2025-11-03T17:34:29.5325169Z INFO 11-03 17:34:29 [__init__.py:224] Automatically detected platform cuda. 2025-11-03T17:34:33.5004017Z (EngineCore_DP0 pid=64294) INFO 11-03 17:34:33 [core.py:727] Waiting for init message from front-end. 2025-11-03T17:34:33.5271995Z (EngineCore_DP0 pid=64294) INFO 11-03 17:34:33 [core.py:94] Initializing a V1 LLM engine (v0.11.0rc2.dev389+ge51928192.d20251103) with config: model='RedHatAI/Meta-Llama-3.1-8B-Instruct-FP8', speculative_config=None, tokenizer='RedHatAI/Meta-Llama-3.1-8B-Instruct-FP8', skip_tokenizer_init=False, tokenizer_mode=auto, revision=None, tokenizer_revision=None, trust_remote_code=False, dtype=torch.float16, max_seq_len=2048, download_dir=None, load_format=dummy, tensor_parallel_size=2, pipeline_parallel_size=1, data_parallel_size=1, disable_custom_all_reduce=False, quantization=compressed-tensors, enforce_eager=True, kv_cache_dtype=auto, device_config=cuda, structured_outputs_config=StructuredOutputsConfig(backend='auto', disable_fallback=False, disable_any_whitespace=False, disable_additional_properties=False, reasoning_parser=''), observability_config=ObservabilityConfig(show_hidden_metrics_for_version=None, otlp_traces_endpoint=None, collect_detailed_traces=None), seed=0, served_model_name=RedHatAI/Meta-Llama-3.1-8B-Instruct-FP8, enable_prefix_caching=True, chunked_prefill_enabled=True, pooler_config=None, compilation_config={'level': 3, 'debug_dump_path': None, 'cache_dir': '', 'backend': '', 'custom_ops': ['+rms_norm', '+rms_norm'], 'splitting_ops': ['vllm.unified_attention', 'vllm.unified_attention_with_output', 'vllm.unified_mla_attention', 'vllm.unified_mla_attention_with_output', 'vllm.mamba_mixer2', 'vllm.mamba_mixer', 'vllm.short_conv', 'vllm.linear_attention', 'vllm.plamo2_mamba_mixer', 'vllm.gdn_attention', 'vllm.sparse_attn_indexer'], 'use_inductor': True, 'compile_sizes': [8, 4], 'inductor_compile_config': {'enable_auto_functionalized_v2': False}, 'inductor_passes': {}, 'cudagraph_mode': , 'use_cudagraph': False, 'cudagraph_num_of_warmups': 0, 'cudagraph_capture_sizes': [], 'cudagraph_copy_inputs': False, 'full_cuda_graph': False, 'use_inductor_graph_partition': False, 'pass_config': {'enable_noop': True, 'enable_sequence_parallelism': True}, 'max_capture_size': 0, 'local_cache_dir': None} 2025-11-03T17:34:35.4317738Z (EngineCore_DP0 pid=64294) 2025-11-03 17:34:35,431 INFO worker.py:1918 -- Started a local Ray instance. View the dashboard at http://127.0.0.1:8265  2025-11-03T17:34:36.2947984Z (EngineCore_DP0 pid=64294) INFO 11-03 17:34:36 [ray_utils.py:373] No current placement group found. Creating a new placement group. 2025-11-03T17:34:36.4710070Z (EngineCore_DP0 pid=64294) INFO 11-03 17:34:36 [ray_distributed_executor.py:179] use_ray_spmd_worker: True 2025-11-03T17:34:37.0176756Z (EngineCore_DP0 pid=64294) (pid=65179) /opt/conda/envs/py_3.12/lib/python3.12/site-packages/torch/cuda/__init__.py:65: FutureWarning: The pynvml package is deprecated. Please install nvidia-ml-py instead. If you did not install pynvml directly, please report this to the maintainers of the package that installed pynvml for you. 2025-11-03T17:34:37.0178516Z (EngineCore_DP0 pid=64294) (pid=65179) import pynvml # type: ignore[import] 2025-11-03T17:34:39.3665092Z (EngineCore_DP0 pid=64294) (pid=65179) INFO 11-03 17:34:39 [__init__.py:224] Automatically detected platform cuda. 2025-11-03T17:34:41.9877908Z (EngineCore_DP0 pid=64294) INFO 11-03 17:34:41 [ray_env.py:67] RAY_NON_CARRY_OVER_ENV_VARS from config: set() 2025-11-03T17:34:41.9879392Z (EngineCore_DP0 pid=64294) INFO 11-03 17:34:41 [ray_env.py:70] Copying the following environment variables to workers: ['HF_TOKEN', 'HUGGING_FACE_HUB_TOKEN', 'VLLM_WORKER_MULTIPROC_METHOD', 'MAX_JOBS', 'LD_LIBRARY_PATH', 'VLLM_USE_V1', 'VLLM_USE_RAY_COMPILED_DAG', 'VLLM_USE_RAY_SPMD_WORKER'] 2025-11-03T17:34:41.9881236Z (EngineCore_DP0 pid=64294) INFO 11-03 17:34:41 [ray_env.py:75] If certain env vars should NOT be copied, add them to /var/lib/jenkins/.config/vllm/ray_non_carry_over_env_vars.json file 2025-11-03T17:34:43.7475406Z (EngineCore_DP0 pid=64294) (RayWorkerWrapper pid=65179) WARNING 11-03 17:34:43 [worker_base.py:307] Missing `shared_worker_lock` argument from executor. This argument is needed for mm_processor_cache_type='shm'. 2025-11-03T17:34:43.9540463Z (EngineCore_DP0 pid=64294) (RayWorkerWrapper pid=65179) [Gloo] Rank 0 is connected to 1 peer ranks. Expected number of connected peer ranks is : 1 2025-11-03T17:34:43.9542146Z (EngineCore_DP0 pid=64294) (RayWorkerWrapper pid=65179) [Gloo] Rank 0 is connected to 1 peer ranks. Expected number of connected peer ranks is : 1 2025-11-03T17:34:43.9543667Z (EngineCore_DP0 pid=64294) (RayWorkerWrapper pid=65179) INFO 11-03 17:34:43 [__init__.py:1146] Found nccl from library libnccl.so.2 2025-11-03T17:34:43.9545082Z (EngineCore_DP0 pid=64294) (RayWorkerWrapper pid=65179) INFO 11-03 17:34:43 [pynccl.py:109] vLLM is using nccl==2.27.5 2025-11-03T17:34:44.2636376Z (EngineCore_DP0 pid=64294) (RayWorkerWrapper pid=65179) WARNING 11-03 17:34:44 [symm_mem.py:60] SymmMemCommunicator: Device capability 8.9 not supported, communicator is not available. 2025-11-03T17:34:44.2638416Z (EngineCore_DP0 pid=64294) (RayWorkerWrapper pid=65179) INFO 11-03 17:34:44 [custom_all_reduce.py:37] Skipping P2P check and trusting the driver's P2P report. 2025-11-03T17:34:44.2640808Z (EngineCore_DP0 pid=64294) (RayWorkerWrapper pid=65179) WARNING 11-03 17:34:44 [custom_all_reduce.py:162] Custom allreduce is disabled because your platform lacks GPU P2P capability or P2P test failed. To silence this warning, specify disable_custom_all_reduce=True explicitly. 2025-11-03T17:34:44.2644363Z (EngineCore_DP0 pid=64294) (RayWorkerWrapper pid=65179) INFO 11-03 17:34:44 [shm_broadcast.py:302] vLLM message queue communication handle: Handle(local_reader_ranks=[1], buffer_handle=(1, 4194304, 6, 'psm_c3412499'), local_subscribe_addr='ipc:///tmp/26a95c9a-475f-41de-bc65-cda97b7b5ee5', remote_subscribe_addr=None, remote_addr_ipv6=False) 2025-11-03T17:34:44.3684407Z (EngineCore_DP0 pid=64294) (RayWorkerWrapper pid=65179) INFO 11-03 17:34:44 [parallel_state.py:1231] rank 0 in world size 2 is assigned as DP rank 0, PP rank 0, TP rank 0, EP rank 0 2025-11-03T17:34:44.3685663Z (EngineCore_DP0 pid=64294) (pid=65187) INFO 11-03 17:34:39 [__init__.py:224] Automatically detected platform cuda. 2025-11-03T17:34:44.5748523Z (EngineCore_DP0 pid=64294) (RayWorkerWrapper pid=65179) INFO 11-03 17:34:44 [topk_topp_sampler.py:59] Using FlashInfer for top-p & top-k sampling. 2025-11-03T17:34:44.6798430Z (EngineCore_DP0 pid=64294) (RayWorkerWrapper pid=65179) INFO 11-03 17:34:44 [gpu_model_runner.py:2840] Starting to load model RedHatAI/Meta-Llama-3.1-8B-Instruct-FP8... 2025-11-03T17:34:44.8863008Z (EngineCore_DP0 pid=64294) (RayWorkerWrapper pid=65179) INFO 11-03 17:34:44 [gpu_model_runner.py:2870] Loading model from scratch... 2025-11-03T17:34:44.9907882Z (EngineCore_DP0 pid=64294) (RayWorkerWrapper pid=65179) INFO 11-03 17:34:44 [cuda.py:404] Using Flash Attention backend on V1 engine. 2025-11-03T17:34:45.8084008Z (EngineCore_DP0 pid=64294) (RayWorkerWrapper pid=65187) INFO 11-03 17:34:45 [gpu_model_runner.py:2902] Model loading took 4.2624 GiB and 0.538570 seconds 2025-11-03T17:34:53.4481209Z (EngineCore_DP0 pid=64294) (RayWorkerWrapper pid=65179) INFO 11-03 17:34:53 [backends.py:594] Using cache directory: /var/lib/jenkins/.cache/vllm/torch_compile_cache/c23588ba3d/rank_0_0/backbone for vLLM's torch.compile 2025-11-03T17:34:53.4482705Z (EngineCore_DP0 pid=64294) (RayWorkerWrapper pid=65179) INFO 11-03 17:34:53 [backends.py:608] Dynamo bytecode transform time: 7.30 s 2025-11-03T17:34:53.4484035Z (EngineCore_DP0 pid=64294) (RayWorkerWrapper pid=65187) WARNING 11-03 17:34:43 [worker_base.py:307] Missing `shared_worker_lock` argument from executor. This argument is needed for mm_processor_cache_type='shm'. 2025-11-03T17:34:53.4486127Z (EngineCore_DP0 pid=64294) (RayWorkerWrapper pid=65187) [Gloo] Rank 1 is connected to 1 peer ranks. Expected number of connected peer ranks is : 1 [repeated 10x across cluster] (Ray deduplicates logs by default. Set RAY_DEDUP_LOGS=0 to disable log deduplication, or see https://docs.ray.io/en/master/ray-observability/user-guides/configure-logging.html#log-deduplication for more options.) 2025-11-03T17:34:53.4487854Z (EngineCore_DP0 pid=64294) (RayWorkerWrapper pid=65187) INFO 11-03 17:34:44 [__init__.py:1146] Found nccl from library libnccl.so.2 [repeated 3x across cluster] 2025-11-03T17:34:53.4488882Z (EngineCore_DP0 pid=64294) (RayWorkerWrapper pid=65187) INFO 11-03 17:34:44 [pynccl.py:109] vLLM is using nccl==2.27.5 [repeated 3x across cluster] 2025-11-03T17:34:53.4489872Z (EngineCore_DP0 pid=64294) (RayWorkerWrapper pid=65187) WARNING 11-03 17:34:44 [symm_mem.py:60] SymmMemCommunicator: Device capability 8.9 not supported, communicator is not available. 2025-11-03T17:34:53.4490864Z (EngineCore_DP0 pid=64294) (RayWorkerWrapper pid=65187) INFO 11-03 17:34:44 [custom_all_reduce.py:37] Skipping P2P check and trusting the driver's P2P report. 2025-11-03T17:34:53.4492142Z (EngineCore_DP0 pid=64294) (RayWorkerWrapper pid=65187) WARNING 11-03 17:34:44 [custom_all_reduce.py:162] Custom allreduce is disabled because your platform lacks GPU P2P capability or P2P test failed. To silence this warning, specify disable_custom_all_reduce=True explicitly. 2025-11-03T17:34:53.4493695Z (EngineCore_DP0 pid=64294) (RayWorkerWrapper pid=65187) INFO 11-03 17:34:44 [parallel_state.py:1231] rank 1 in world size 2 is assigned as DP rank 0, PP rank 0, TP rank 1, EP rank 1 2025-11-03T17:34:53.4494639Z (EngineCore_DP0 pid=64294) (RayWorkerWrapper pid=65187) INFO 11-03 17:34:44 [topk_topp_sampler.py:59] Using FlashInfer for top-p & top-k sampling. 2025-11-03T17:34:53.4495581Z (EngineCore_DP0 pid=64294) (RayWorkerWrapper pid=65187) INFO 11-03 17:34:44 [gpu_model_runner.py:2840] Starting to load model RedHatAI/Meta-Llama-3.1-8B-Instruct-FP8... 2025-11-03T17:34:53.4496563Z (EngineCore_DP0 pid=64294) (RayWorkerWrapper pid=65187) INFO 11-03 17:34:44 [gpu_model_runner.py:2870] Loading model from scratch... 2025-11-03T17:34:53.4497345Z (EngineCore_DP0 pid=64294) (RayWorkerWrapper pid=65187) INFO 11-03 17:34:44 [cuda.py:404] Using Flash Attention backend on V1 engine. 2025-11-03T17:34:53.4498356Z (EngineCore_DP0 pid=64294) (RayWorkerWrapper pid=65179) INFO 11-03 17:34:45 [gpu_model_runner.py:2902] Model loading took 4.2624 GiB and 0.542158 seconds 2025-11-03T17:34:56.0998163Z (EngineCore_DP0 pid=64294) (RayWorkerWrapper pid=65179) INFO 11-03 17:34:56 [backends.py:179] Directly load the compiled graph(s) for dynamic shape from the cache, took 2.197 s 2025-11-03T17:34:58.7513147Z (EngineCore_DP0 pid=64294) (RayWorkerWrapper pid=65179) INFO 11-03 17:34:58 [gpu_worker.py:314] Available KV cache memory: 15.34 GiB 2025-11-03T17:34:58.7514328Z (EngineCore_DP0 pid=64294) (RayWorkerWrapper pid=65187) INFO 11-03 17:34:53 [backends.py:594] Using cache directory: /var/lib/jenkins/.cache/vllm/torch_compile_cache/c23588ba3d/rank_1_0/backbone for vLLM's torch.compile 2025-11-03T17:34:58.7515410Z (EngineCore_DP0 pid=64294) (RayWorkerWrapper pid=65187) INFO 11-03 17:34:53 [backends.py:608] Dynamo bytecode transform time: 7.35 s 2025-11-03T17:34:58.9761295Z (EngineCore_DP0 pid=64294) INFO 11-03 17:34:58 [kv_cache_utils.py:1199] GPU KV cache size: 251,360 tokens 2025-11-03T17:34:58.9762271Z (EngineCore_DP0 pid=64294) INFO 11-03 17:34:58 [kv_cache_utils.py:1204] Maximum concurrency for 2,048 tokens per request: 122.73x 2025-11-03T17:34:58.9763091Z (EngineCore_DP0 pid=64294) INFO 11-03 17:34:58 [kv_cache_utils.py:1199] GPU KV cache size: 251,360 tokens 2025-11-03T17:34:58.9763919Z (EngineCore_DP0 pid=64294) INFO 11-03 17:34:58 [kv_cache_utils.py:1204] Maximum concurrency for 2,048 tokens per request: 122.73x 2025-11-03T17:34:59.0597871Z (EngineCore_DP0 pid=64294) (RayWorkerWrapper pid=65179) INFO 11-03 17:34:58 [gpu_worker.py:351] Compile and warming up model for size 8 2025-11-03T17:35:10.4666499Z (EngineCore_DP0 pid=64294) (RayWorkerWrapper pid=65179) INFO 11-03 17:35:10 [backends.py:185] Directly load the compiled graph(s) for shape 8 from the cache, took 11.409 s 2025-11-03T17:35:10.4667910Z (EngineCore_DP0 pid=64294) (RayWorkerWrapper pid=65187) INFO 11-03 17:34:56 [backends.py:179] Directly load the compiled graph(s) for dynamic shape from the cache, took 2.237 s 2025-11-03T17:35:10.4669033Z (EngineCore_DP0 pid=64294) (RayWorkerWrapper pid=65187) INFO 11-03 17:34:58 [gpu_worker.py:314] Available KV cache memory: 15.34 GiB 2025-11-03T17:35:10.4670073Z (EngineCore_DP0 pid=64294) (RayWorkerWrapper pid=65179) INFO 11-03 17:35:10 [gpu_worker.py:351] Compile and warming up model for size 4 [repeated 2x across cluster] 2025-11-03T17:35:37.8496039Z (EngineCore_DP0 pid=64294) (RayWorkerWrapper pid=65187) INFO 11-03 17:35:37 [monitor.py:33] torch.compile takes 7.35 s in total 2025-11-03T17:35:37.8497985Z (EngineCore_DP0 pid=64294) (RayWorkerWrapper pid=65187) INFO 11-03 17:35:37 [backends.py:185] Directly load the compiled graph(s) for shape 4 from the cache, took 27.430 s [repeated 2x across cluster] 2025-11-03T17:35:37.8500119Z (EngineCore_DP0 pid=64294) (RayWorkerWrapper pid=65187) INFO 11-03 17:35:10 [gpu_worker.py:351] Compile and warming up model for size 4 2025-11-03T17:35:38.0734827Z (EngineCore_DP0 pid=64294) INFO 11-03 17:35:38 [core.py:240] init engine (profile, create kv cache, warmup model) took 52.26 seconds 2025-11-03T17:35:38.5525778Z (EngineCore_DP0 pid=64294) INFO 11-03 17:35:38 [vllm.py:354] Cudagraph is disabled under eager mode 2025-11-03T17:35:38.7388195Z (APIServer pid=64171) INFO 11-03 17:35:38 [loggers.py:172] Engine 000: vllm cache_config_info with initialization after num_gpu_blocks is: 15710 2025-11-03T17:35:38.8010962Z (EngineCore_DP0 pid=64294) INFO 11-03 17:35:38 [gc_utils.py:40] GC Debug Config. enabled:False,top_objects:-1 2025-11-03T17:35:38.8258924Z (APIServer pid=64171) INFO 11-03 17:35:38 [api_server.py:1628] Supported tasks: ['generate'] 2025-11-03T17:35:38.8621669Z (APIServer pid=64171) WARNING 11-03 17:35:38 [model.py:1581] Default sampling parameters have been overridden by the model's Hugging Face generation config recommended from the model creator. If this is not intended, please relaunch vLLM instance with `--generation-config vllm`. 2025-11-03T17:35:38.8624998Z (APIServer pid=64171) INFO 11-03 17:35:38 [serving_responses.py:157] Using default chat sampling params from model: {'temperature': 0.6, 'top_p': 0.9} 2025-11-03T17:35:38.9015122Z (APIServer pid=64171) INFO 11-03 17:35:38 [serving_chat.py:159] Using default chat sampling params from model: {'temperature': 0.6, 'top_p': 0.9} 2025-11-03T17:35:38.9575730Z (APIServer pid=64171) INFO 11-03 17:35:38 [serving_completion.py:67] Using default completion sampling params from model: {'temperature': 0.6, 'top_p': 0.9} 2025-11-03T17:35:38.9576793Z (APIServer pid=64171) INFO 11-03 17:35:38 [api_server.py:1939] Starting vLLM API server 0 on http://0.0.0.0:49175 2025-11-03T17:35:38.9577541Z (APIServer pid=64171) INFO 11-03 17:35:38 [launcher.py:38] Available routes are: 2025-11-03T17:35:38.9578238Z (APIServer pid=64171) INFO 11-03 17:35:38 [launcher.py:46] Route: /openapi.json, Methods: GET, HEAD 2025-11-03T17:35:38.9578914Z (APIServer pid=64171) INFO 11-03 17:35:38 [launcher.py:46] Route: /docs, Methods: GET, HEAD 2025-11-03T17:35:38.9579644Z (APIServer pid=64171) INFO 11-03 17:35:38 [launcher.py:46] Route: /docs/oauth2-redirect, Methods: GET, HEAD 2025-11-03T17:35:38.9580570Z (APIServer pid=64171) INFO 11-03 17:35:38 [launcher.py:46] Route: /redoc, Methods: GET, HEAD 2025-11-03T17:35:38.9581222Z (APIServer pid=64171) INFO 11-03 17:35:38 [launcher.py:46] Route: /health, Methods: GET 2025-11-03T17:35:38.9581835Z (APIServer pid=64171) INFO 11-03 17:35:38 [launcher.py:46] Route: /load, Methods: GET 2025-11-03T17:35:38.9582452Z (APIServer pid=64171) INFO 11-03 17:35:38 [launcher.py:46] Route: /ping, Methods: POST 2025-11-03T17:35:38.9583061Z (APIServer pid=64171) INFO 11-03 17:35:38 [launcher.py:46] Route: /ping, Methods: GET 2025-11-03T17:35:38.9583694Z (APIServer pid=64171) INFO 11-03 17:35:38 [launcher.py:46] Route: /tokenize, Methods: POST 2025-11-03T17:35:38.9584340Z (APIServer pid=64171) INFO 11-03 17:35:38 [launcher.py:46] Route: /detokenize, Methods: POST 2025-11-03T17:35:38.9584853Z (APIServer pid=64171) INFO 11-03 17:35:38 [launcher.py:46] Route: /v1/models, Methods: GET 2025-11-03T17:35:38.9585348Z (APIServer pid=64171) INFO 11-03 17:35:38 [launcher.py:46] Route: /version, Methods: GET 2025-11-03T17:35:38.9585865Z (APIServer pid=64171) INFO 11-03 17:35:38 [launcher.py:46] Route: /v1/responses, Methods: POST 2025-11-03T17:35:38.9586618Z (APIServer pid=64171) INFO 11-03 17:35:38 [launcher.py:46] Route: /v1/responses/{response_id}, Methods: GET 2025-11-03T17:35:38.9587242Z (APIServer pid=64171) INFO 11-03 17:35:38 [launcher.py:46] Route: /v1/responses/{response_id}/cancel, Methods: POST 2025-11-03T17:35:38.9587846Z (APIServer pid=64171) INFO 11-03 17:35:38 [launcher.py:46] Route: /v1/chat/completions, Methods: POST 2025-11-03T17:35:38.9588404Z (APIServer pid=64171) INFO 11-03 17:35:38 [launcher.py:46] Route: /v1/completions, Methods: POST 2025-11-03T17:35:38.9588942Z (APIServer pid=64171) INFO 11-03 17:35:38 [launcher.py:46] Route: /v1/embeddings, Methods: POST 2025-11-03T17:35:38.9589480Z (APIServer pid=64171) INFO 11-03 17:35:38 [launcher.py:46] Route: /pooling, Methods: POST 2025-11-03T17:35:38.9589997Z (APIServer pid=64171) INFO 11-03 17:35:38 [launcher.py:46] Route: /classify, Methods: POST 2025-11-03T17:35:38.9590500Z (APIServer pid=64171) INFO 11-03 17:35:38 [launcher.py:46] Route: /score, Methods: POST 2025-11-03T17:35:38.9591150Z (APIServer pid=64171) INFO 11-03 17:35:38 [launcher.py:46] Route: /v1/score, Methods: POST 2025-11-03T17:35:38.9591711Z (APIServer pid=64171) INFO 11-03 17:35:38 [launcher.py:46] Route: /v1/audio/transcriptions, Methods: POST 2025-11-03T17:35:38.9592292Z (APIServer pid=64171) INFO 11-03 17:35:38 [launcher.py:46] Route: /v1/audio/translations, Methods: POST 2025-11-03T17:35:38.9593878Z (APIServer pid=64171) INFO 11-03 17:35:38 [launcher.py:46] Route: /rerank, Methods: POST 2025-11-03T17:35:38.9594399Z (APIServer pid=64171) INFO 11-03 17:35:38 [launcher.py:46] Route: /v1/rerank, Methods: POST 2025-11-03T17:35:38.9594912Z (APIServer pid=64171) INFO 11-03 17:35:38 [launcher.py:46] Route: /v2/rerank, Methods: POST 2025-11-03T17:35:38.9595441Z (APIServer pid=64171) INFO 11-03 17:35:38 [launcher.py:46] Route: /scale_elastic_ep, Methods: POST 2025-11-03T17:35:38.9596010Z (APIServer pid=64171) INFO 11-03 17:35:38 [launcher.py:46] Route: /is_scaling_elastic_ep, Methods: POST 2025-11-03T17:35:38.9596566Z (APIServer pid=64171) INFO 11-03 17:35:38 [launcher.py:46] Route: /invocations, Methods: POST 2025-11-03T17:35:38.9597079Z (APIServer pid=64171) INFO 11-03 17:35:38 [launcher.py:46] Route: /metrics, Methods: GET 2025-11-03T17:35:38.9734209Z (APIServer pid=64171) INFO: Started server process [64171] 2025-11-03T17:35:38.9735419Z (APIServer pid=64171) INFO: Waiting for application startup. 2025-11-03T17:35:39.1965065Z (APIServer pid=64171) INFO: Application startup complete. 2025-11-03T17:35:39.4167375Z (APIServer pid=64171) INFO: 127.0.0.1:52042 - "GET /health HTTP/1.1" 200 OK 2025-11-03T17:35:39.6651253Z (APIServer pid=64171) INFO: 127.0.0.1:52046 - "GET /v1/models HTTP/1.1" 200 OK 2025-11-03T17:35:39.6958286Z (EngineCore_DP0 pid=64294) INFO 11-03 17:35:39 [ray_distributed_executor.py:569] RAY_CGRAPH_get_timeout is set to 300 2025-11-03T17:35:39.6959501Z (EngineCore_DP0 pid=64294) INFO 11-03 17:35:39 [ray_distributed_executor.py:573] VLLM_USE_RAY_COMPILED_DAG_CHANNEL_TYPE = auto 2025-11-03T17:35:39.6960775Z (EngineCore_DP0 pid=64294) INFO 11-03 17:35:39 [ray_distributed_executor.py:577] VLLM_USE_RAY_COMPILED_DAG_OVERLAP_COMM = False 2025-11-03T17:35:39.6964471Z (EngineCore_DP0 pid=64294) INFO 11-03 17:35:39 [ray_distributed_executor.py:653] Using RayPPCommunicator (which wraps vLLM _PP GroupCoordinator) for Ray Compiled Graph communication. 2025-11-03T17:35:39.9270832Z (APIServer pid=64171) INFO: 127.0.0.1:52046 - "POST /v1/completions HTTP/1.1" 200 OK 2025-11-03T17:35:40.0465470Z (APIServer pid=64171) INFO: 127.0.0.1:52046 - "POST /v1/completions HTTP/1.1" 200 OK 2025-11-03T17:35:40.2640134Z (APIServer pid=64171) INFO: 127.0.0.1:52046 - "POST /v1/completions HTTP/1.1" 200 OK 2025-11-03T17:35:40.3905752Z (APIServer pid=64171) INFO: 127.0.0.1:52046 - "POST /v1/completions HTTP/1.1" 200 OK 2025-11-03T17:35:40.5328056Z (APIServer pid=64171) INFO: 127.0.0.1:52046 - "POST /v1/completions HTTP/1.1" 200 OK 2025-11-03T17:35:40.5403605Z (APIServer pid=64171) INFO: 127.0.0.1:52046 - "POST /v1/completions HTTP/1.1" 200 OK 2025-11-03T17:35:40.6844831Z (APIServer pid=64171) INFO 11-03 17:35:40 [launcher.py:110] Shutting down FastAPI HTTP server. 2025-11-03T17:35:40.6845561Z *** SIGTERM received at time=1762191340 on cpu 28 *** 2025-11-03T17:35:40.6891370Z PC: @ 0x7f4a31da5117 (unknown) (unknown) 2025-11-03T17:35:40.6892251Z @ 0x7f4a31d56520 (unknown) (unknown) 2025-11-03T17:35:40.6892770Z [2025-11-03 17:35:40,689 E 64294 64294] logging.cc:501: *** SIGTERM received at time=1762191340 on cpu 28 *** 2025-11-03T17:35:40.6895413Z [2025-11-03 17:35:40,689 E 64294 64294] logging.cc:501: PC: @ 0x7f4a31da5117 (unknown) (unknown) 2025-11-03T17:35:40.6896152Z [2025-11-03 17:35:40,689 E 64294 64294] logging.cc:501: @ 0x7f4a31d56520 (unknown) (unknown) 2025-11-03T17:35:40.6898267Z (EngineCore_DP0 pid=64294) INFO 11-03 17:35:40 [ray_distributed_executor.py:127] Shutting down Ray distributed executor. If you see error log from logging.cc regarding SIGTERM received, please ignore because this is the expected termination process in Ray. 2025-11-03T17:35:40.6899351Z (EngineCore_DP0 pid=64294) 2025-11-03 17:35:40,689 INFO compiled_dag_node.py:2171 -- Tearing down compiled DAG 2025-11-03T17:35:40.6900638Z (EngineCore_DP0 pid=64294) 2025-11-03 17:35:40,689 INFO compiled_dag_node.py:2176 -- Cancelling compiled worker on actor: Actor(RayWorkerWrapper, 90b0d7bc6771e69e02239c2001000000) 2025-11-03T17:35:40.6901640Z (EngineCore_DP0 pid=64294) 2025-11-03 17:35:40,689 INFO compiled_dag_node.py:2176 -- Cancelling compiled worker on actor: Actor(RayWorkerWrapper, 57901cbb54bd0c00d0ee9b2301000000) 2025-11-03T17:35:40.6940555Z (EngineCore_DP0 pid=64294) 2025-11-03 17:35:40,693 INFO compiled_dag_node.py:2198 -- Waiting for worker tasks to exit 2025-11-03T17:35:40.6941615Z (EngineCore_DP0 pid=64294) 2025-11-03 17:35:40,693 INFO compiled_dag_node.py:2201 -- Teardown complete 2025-11-03T17:35:41.1973594Z (EngineCore_DP0 pid=64294) (RayWorkerWrapper pid=65179) INFO 11-03 17:35:38 [monitor.py:33] torch.compile takes 7.30 s in total 2025-11-03T17:35:41.1975990Z (EngineCore_DP0 pid=64294) (RayWorkerWrapper pid=65179) INFO 11-03 17:35:38 [backends.py:185] Directly load the compiled graph(s) for shape 4 from the cache, took 27.636 s 2025-11-03T17:35:41.1978152Z (EngineCore_DP0 pid=64294) (pid=65187) /opt/conda/envs/py_3.12/lib/python3.12/site-packages/torch/cuda/__init__.py:65: FutureWarning: The pynvml package is deprecated. Please install nvidia-ml-py instead. If you did not install pynvml directly, please report this to the maintainers of the package that installed pynvml for you. 2025-11-03T17:35:41.1979459Z (EngineCore_DP0 pid=64294) (pid=65187) import pynvml # type: ignore[import] 2025-11-03T17:35:43.0512521Z (APIServer pid=64171) INFO: Shutting down 2025-11-03T17:35:43.1514124Z (APIServer pid=64171) INFO: Waiting for application shutdown. 2025-11-03T17:35:43.1515375Z (APIServer pid=64171) INFO: Application shutdown complete. 2025-11-03T17:35:43.5137958Z INFO 11-03 17:35:43 [model.py:653] Resolved architecture: LlamaForCausalLM 2025-11-03T17:35:43.5138596Z WARNING 11-03 17:35:43 [model.py:1942] Casting torch.bfloat16 to torch.float16. 2025-11-03T17:35:43.5139379Z INFO 11-03 17:35:43 [model.py:1714] Using max model len 2048 2025-11-03T17:35:43.5143044Z Launching RemoteOpenAIServer with: vllm serve RedHatAI/Meta-Llama-3.1-8B-Instruct-FP8 --dtype float16 --max-model-len 2048 --max-num-seqs 8 --enforce-eager --tokenizer-mode auto --tensor-parallel-size 2 --distributed-executor-backend mp --load-format dummy --port 54465 --seed 0 2025-11-03T17:35:44.0425552Z /opt/conda/envs/py_3.12/lib/python3.12/site-packages/torch/cuda/__init__.py:65: FutureWarning: The pynvml package is deprecated. Please install nvidia-ml-py instead. If you did not install pynvml directly, please report this to the maintainers of the package that installed pynvml for you. 2025-11-03T17:35:44.0426791Z import pynvml # type: ignore[import] 2025-11-03T17:35:46.2914779Z INFO 11-03 17:35:46 [__init__.py:224] Automatically detected platform cuda. 2025-11-03T17:35:49.9855784Z (APIServer pid=69395) INFO 11-03 17:35:49 [api_server.py:1870] vLLM API server version 0.11.0rc2.dev389+ge51928192.d20251103 2025-11-03T17:35:49.9881273Z (APIServer pid=69395) INFO 11-03 17:35:49 [utils.py:239] non-default args: {'model_tag': 'RedHatAI/Meta-Llama-3.1-8B-Instruct-FP8', 'port': 54465, 'model': 'RedHatAI/Meta-Llama-3.1-8B-Instruct-FP8', 'dtype': 'float16', 'seed': 0, 'max_model_len': 2048, 'enforce_eager': True, 'load_format': 'dummy', 'distributed_executor_backend': 'mp', 'tensor_parallel_size': 2, 'max_num_seqs': 8} 2025-11-03T17:35:50.1793621Z (APIServer pid=69395) INFO 11-03 17:35:50 [model.py:653] Resolved architecture: LlamaForCausalLM 2025-11-03T17:35:50.1794353Z (APIServer pid=69395) `torch_dtype` is deprecated! Use `dtype` instead! 2025-11-03T17:35:50.1795196Z (APIServer pid=69395) WARNING 11-03 17:35:50 [model.py:1942] Casting torch.bfloat16 to torch.float16. 2025-11-03T17:35:50.1796041Z (APIServer pid=69395) INFO 11-03 17:35:50 [model.py:1714] Using max model len 2048 2025-11-03T17:35:50.5327310Z (APIServer pid=69395) INFO 11-03 17:35:50 [scheduler.py:225] Chunked prefill is enabled with max_num_batched_tokens=2048. 2025-11-03T17:35:50.5700701Z (APIServer pid=69395) INFO 11-03 17:35:50 [vllm.py:354] Cudagraph is disabled under eager mode 2025-11-03T17:35:51.5825742Z /opt/conda/envs/py_3.12/lib/python3.12/site-packages/torch/cuda/__init__.py:65: FutureWarning: The pynvml package is deprecated. Please install nvidia-ml-py instead. If you did not install pynvml directly, please report this to the maintainers of the package that installed pynvml for you. 2025-11-03T17:35:51.5827240Z import pynvml # type: ignore[import] 2025-11-03T17:35:53.9266758Z INFO 11-03 17:35:53 [__init__.py:224] Automatically detected platform cuda. 2025-11-03T17:35:57.7887568Z (EngineCore_DP0 pid=69518) INFO 11-03 17:35:57 [core.py:727] Waiting for init message from front-end. 2025-11-03T17:35:57.8161927Z (EngineCore_DP0 pid=69518) INFO 11-03 17:35:57 [core.py:94] Initializing a V1 LLM engine (v0.11.0rc2.dev389+ge51928192.d20251103) with config: model='RedHatAI/Meta-Llama-3.1-8B-Instruct-FP8', speculative_config=None, tokenizer='RedHatAI/Meta-Llama-3.1-8B-Instruct-FP8', skip_tokenizer_init=False, tokenizer_mode=auto, revision=None, tokenizer_revision=None, trust_remote_code=False, dtype=torch.float16, max_seq_len=2048, download_dir=None, load_format=dummy, tensor_parallel_size=2, pipeline_parallel_size=1, data_parallel_size=1, disable_custom_all_reduce=False, quantization=compressed-tensors, enforce_eager=True, kv_cache_dtype=auto, device_config=cuda, structured_outputs_config=StructuredOutputsConfig(backend='auto', disable_fallback=False, disable_any_whitespace=False, disable_additional_properties=False, reasoning_parser=''), observability_config=ObservabilityConfig(show_hidden_metrics_for_version=None, otlp_traces_endpoint=None, collect_detailed_traces=None), seed=0, served_model_name=RedHatAI/Meta-Llama-3.1-8B-Instruct-FP8, enable_prefix_caching=True, chunked_prefill_enabled=True, pooler_config=None, compilation_config={'level': 0, 'debug_dump_path': None, 'cache_dir': '', 'backend': '', 'custom_ops': [], 'splitting_ops': None, 'use_inductor': True, 'compile_sizes': [], 'inductor_compile_config': {'enable_auto_functionalized_v2': False}, 'inductor_passes': {}, 'cudagraph_mode': , 'use_cudagraph': False, 'cudagraph_num_of_warmups': 0, 'cudagraph_capture_sizes': [], 'cudagraph_copy_inputs': False, 'full_cuda_graph': False, 'use_inductor_graph_partition': False, 'pass_config': {}, 'max_capture_size': 0, 'local_cache_dir': None} 2025-11-03T17:35:57.8167496Z (EngineCore_DP0 pid=69518) WARNING 11-03 17:35:57 [multiproc_executor.py:744] Reducing Torch parallelism from 24 threads to 1 to avoid unnecessary CPU contention. Set OMP_NUM_THREADS in the external environment to tune this value as needed. 2025-11-03T17:35:57.8170166Z (EngineCore_DP0 pid=69518) INFO 11-03 17:35:57 [shm_broadcast.py:302] vLLM message queue communication handle: Handle(local_reader_ranks=[0, 1], buffer_handle=(2, 16777216, 10, 'psm_654c8a68'), local_subscribe_addr='ipc:///tmp/4b04f2b4-014c-49f0-ba06-21bbf9610b46', remote_subscribe_addr=None, remote_addr_ipv6=False) 2025-11-03T17:35:58.2985468Z /opt/conda/envs/py_3.12/lib/python3.12/site-packages/torch/cuda/__init__.py:65: FutureWarning: The pynvml package is deprecated. Please install nvidia-ml-py instead. If you did not install pynvml directly, please report this to the maintainers of the package that installed pynvml for you. 2025-11-03T17:35:58.2987050Z import pynvml # type: ignore[import] 2025-11-03T17:35:58.2988205Z /opt/conda/envs/py_3.12/lib/python3.12/site-packages/torch/cuda/__init__.py:65: FutureWarning: The pynvml package is deprecated. Please install nvidia-ml-py instead. If you did not install pynvml directly, please report this to the maintainers of the package that installed pynvml for you. 2025-11-03T17:35:58.2989309Z import pynvml # type: ignore[import] 2025-11-03T17:36:00.6798857Z INFO 11-03 17:36:00 [__init__.py:224] Automatically detected platform cuda. 2025-11-03T17:36:00.6935167Z INFO 11-03 17:36:00 [__init__.py:224] Automatically detected platform cuda. 2025-11-03T17:36:05.5628671Z INFO 11-03 17:36:05 [shm_broadcast.py:302] vLLM message queue communication handle: Handle(local_reader_ranks=[0], buffer_handle=(1, 10485760, 10, 'psm_bc1efd52'), local_subscribe_addr='ipc:///tmp/c00043db-a524-4053-9408-0e85ac118154', remote_subscribe_addr=None, remote_addr_ipv6=False) 2025-11-03T17:36:05.5630541Z INFO 11-03 17:36:05 [shm_broadcast.py:302] vLLM message queue communication handle: Handle(local_reader_ranks=[0], buffer_handle=(1, 10485760, 10, 'psm_1227d46a'), local_subscribe_addr='ipc:///tmp/6d97c44a-106a-4625-a02b-55f6c99716cc', remote_subscribe_addr=None, remote_addr_ipv6=False) 2025-11-03T17:36:05.7560946Z [Gloo] Rank 0 is connected to 1 peer ranks. Expected number of connected peer ranks is : 1 2025-11-03T17:36:05.7561568Z [Gloo] Rank 1 is connected to 1 peer ranks. Expected number of connected peer ranks is : 1 2025-11-03T17:36:05.7605917Z [Gloo] Rank 0 is connected to 1 peer ranks. Expected number of connected peer ranks is : 1 2025-11-03T17:36:05.7606497Z [Gloo] Rank 1 is connected to 1 peer ranks. Expected number of connected peer ranks is : 1 2025-11-03T17:36:05.7649443Z INFO 11-03 17:36:05 [__init__.py:1146] Found nccl from library libnccl.so.2 2025-11-03T17:36:05.7649901Z INFO 11-03 17:36:05 [__init__.py:1146] Found nccl from library libnccl.so.2 2025-11-03T17:36:05.7651724Z INFO 11-03 17:36:05 [pynccl.py:109] vLLM is using nccl==2.27.5 2025-11-03T17:36:05.7652425Z INFO 11-03 17:36:05 [pynccl.py:109] vLLM is using nccl==2.27.5 2025-11-03T17:36:05.9384864Z WARNING 11-03 17:36:05 [symm_mem.py:60] SymmMemCommunicator: Device capability 8.9 not supported, communicator is not available. 2025-11-03T17:36:05.9385633Z WARNING 11-03 17:36:05 [symm_mem.py:60] SymmMemCommunicator: Device capability 8.9 not supported, communicator is not available. 2025-11-03T17:36:05.9406957Z INFO 11-03 17:36:05 [custom_all_reduce.py:37] Skipping P2P check and trusting the driver's P2P report. 2025-11-03T17:36:05.9407898Z INFO 11-03 17:36:05 [custom_all_reduce.py:37] Skipping P2P check and trusting the driver's P2P report. 2025-11-03T17:36:05.9409274Z WARNING 11-03 17:36:05 [custom_all_reduce.py:162] Custom allreduce is disabled because your platform lacks GPU P2P capability or P2P test failed. To silence this warning, specify disable_custom_all_reduce=True explicitly. 2025-11-03T17:36:05.9410693Z WARNING 11-03 17:36:05 [custom_all_reduce.py:162] Custom allreduce is disabled because your platform lacks GPU P2P capability or P2P test failed. To silence this warning, specify disable_custom_all_reduce=True explicitly. 2025-11-03T17:36:06.0426097Z INFO 11-03 17:36:06 [shm_broadcast.py:302] vLLM message queue communication handle: Handle(local_reader_ranks=[1], buffer_handle=(1, 4194304, 6, 'psm_d1f71d28'), local_subscribe_addr='ipc:///tmp/3239feb9-f702-43f2-bbaa-a22b441bf82a', remote_subscribe_addr=None, remote_addr_ipv6=False) 2025-11-03T17:36:06.0451978Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:36:06.0453158Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:36:06.0461264Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:36:06.0471038Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:36:06.1122867Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:36:06.1132380Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:36:06.1145993Z [Gloo] Rank 0 is connected to 1 peer ranks. Expected number of connected peer ranks is : 1 2025-11-03T17:36:06.1147096Z [Gloo] Rank 1 is connected to 1 peer ranks. Expected number of connected peer ranks is : 1 2025-11-03T17:36:06.1148270Z INFO 11-03 17:36:06 [__init__.py:1146] Found nccl from library libnccl.so.2 2025-11-03T17:36:06.1149147Z INFO 11-03 17:36:06 [__init__.py:1146] Found nccl from library libnccl.so.2 2025-11-03T17:36:06.1149806Z INFO 11-03 17:36:06 [pynccl.py:109] vLLM is using nccl==2.27.5 2025-11-03T17:36:06.1150195Z INFO 11-03 17:36:06 [pynccl.py:109] vLLM is using nccl==2.27.5 2025-11-03T17:36:06.1498786Z INFO 11-03 17:36:06 [parallel_state.py:1231] rank 1 in world size 2 is assigned as DP rank 0, PP rank 0, TP rank 1, EP rank 1 2025-11-03T17:36:06.1499611Z INFO 11-03 17:36:06 [parallel_state.py:1231] rank 0 in world size 2 is assigned as DP rank 0, PP rank 0, TP rank 0, EP rank 0 2025-11-03T17:36:06.3962679Z INFO 11-03 17:36:06 [topk_topp_sampler.py:59] Using FlashInfer for top-p & top-k sampling. 2025-11-03T17:36:06.3963277Z INFO 11-03 17:36:06 [topk_topp_sampler.py:59] Using FlashInfer for top-p & top-k sampling. 2025-11-03T17:36:06.4336197Z (Worker_TP1 pid=69638) INFO 11-03 17:36:06 [gpu_model_runner.py:2840] Starting to load model RedHatAI/Meta-Llama-3.1-8B-Instruct-FP8... 2025-11-03T17:36:06.4342311Z (Worker_TP0 pid=69637) INFO 11-03 17:36:06 [gpu_model_runner.py:2840] Starting to load model RedHatAI/Meta-Llama-3.1-8B-Instruct-FP8... 2025-11-03T17:36:06.6526319Z (Worker_TP0 pid=69637) INFO 11-03 17:36:06 [gpu_model_runner.py:2870] Loading model from scratch... 2025-11-03T17:36:06.6562839Z (Worker_TP1 pid=69638) INFO 11-03 17:36:06 [gpu_model_runner.py:2870] Loading model from scratch... 2025-11-03T17:36:06.7443170Z (Worker_TP0 pid=69637) INFO 11-03 17:36:06 [cuda.py:404] Using Flash Attention backend on V1 engine. 2025-11-03T17:36:06.7470824Z (Worker_TP1 pid=69638) INFO 11-03 17:36:06 [cuda.py:404] Using Flash Attention backend on V1 engine. 2025-11-03T17:36:07.5635786Z (Worker_TP0 pid=69637) INFO 11-03 17:36:07 [gpu_model_runner.py:2902] Model loading took 4.2624 GiB and 0.491577 seconds 2025-11-03T17:36:07.5674574Z (Worker_TP1 pid=69638) INFO 11-03 17:36:07 [gpu_model_runner.py:2902] Model loading took 4.2624 GiB and 0.491954 seconds 2025-11-03T17:36:09.9435950Z (Worker_TP0 pid=69637) INFO 11-03 17:36:09 [gpu_worker.py:314] Available KV cache memory: 15.32 GiB 2025-11-03T17:36:09.9442926Z (Worker_TP1 pid=69638) INFO 11-03 17:36:09 [gpu_worker.py:314] Available KV cache memory: 15.32 GiB 2025-11-03T17:36:10.1541276Z (EngineCore_DP0 pid=69518) INFO 11-03 17:36:10 [kv_cache_utils.py:1199] GPU KV cache size: 250,976 tokens 2025-11-03T17:36:10.1542015Z (EngineCore_DP0 pid=69518) INFO 11-03 17:36:10 [kv_cache_utils.py:1204] Maximum concurrency for 2,048 tokens per request: 122.55x 2025-11-03T17:36:10.1542686Z (EngineCore_DP0 pid=69518) INFO 11-03 17:36:10 [kv_cache_utils.py:1199] GPU KV cache size: 250,976 tokens 2025-11-03T17:36:10.1543346Z (EngineCore_DP0 pid=69518) INFO 11-03 17:36:10 [kv_cache_utils.py:1204] Maximum concurrency for 2,048 tokens per request: 122.55x 2025-11-03T17:36:10.2872854Z (EngineCore_DP0 pid=69518) INFO 11-03 17:36:10 [core.py:240] init engine (profile, create kv cache, warmup model) took 2.67 seconds 2025-11-03T17:36:10.7536016Z (EngineCore_DP0 pid=69518) INFO 11-03 17:36:10 [vllm.py:354] Cudagraph is disabled under eager mode 2025-11-03T17:36:10.9382418Z (APIServer pid=69395) INFO 11-03 17:36:10 [loggers.py:172] Engine 000: vllm cache_config_info with initialization after num_gpu_blocks is: 15686 2025-11-03T17:36:10.9531648Z (EngineCore_DP0 pid=69518) INFO 11-03 17:36:10 [gc_utils.py:40] GC Debug Config. enabled:False,top_objects:-1 2025-11-03T17:36:10.9748523Z (APIServer pid=69395) INFO 11-03 17:36:10 [api_server.py:1628] Supported tasks: ['generate'] 2025-11-03T17:36:11.0226277Z (APIServer pid=69395) WARNING 11-03 17:36:11 [model.py:1581] Default sampling parameters have been overridden by the model's Hugging Face generation config recommended from the model creator. If this is not intended, please relaunch vLLM instance with `--generation-config vllm`. 2025-11-03T17:36:11.0227835Z (APIServer pid=69395) INFO 11-03 17:36:11 [serving_responses.py:157] Using default chat sampling params from model: {'temperature': 0.6, 'top_p': 0.9} 2025-11-03T17:36:11.0584724Z (APIServer pid=69395) INFO 11-03 17:36:11 [serving_chat.py:159] Using default chat sampling params from model: {'temperature': 0.6, 'top_p': 0.9} 2025-11-03T17:36:11.0884244Z (APIServer pid=69395) INFO 11-03 17:36:11 [serving_completion.py:67] Using default completion sampling params from model: {'temperature': 0.6, 'top_p': 0.9} 2025-11-03T17:36:11.0885862Z (APIServer pid=69395) INFO 11-03 17:36:11 [api_server.py:1939] Starting vLLM API server 0 on http://0.0.0.0:54465 2025-11-03T17:36:11.0887021Z (APIServer pid=69395) INFO 11-03 17:36:11 [launcher.py:38] Available routes are: 2025-11-03T17:36:11.0887850Z (APIServer pid=69395) INFO 11-03 17:36:11 [launcher.py:46] Route: /openapi.json, Methods: GET, HEAD 2025-11-03T17:36:11.0888671Z (APIServer pid=69395) INFO 11-03 17:36:11 [launcher.py:46] Route: /docs, Methods: GET, HEAD 2025-11-03T17:36:11.0889383Z (APIServer pid=69395) INFO 11-03 17:36:11 [launcher.py:46] Route: /docs/oauth2-redirect, Methods: GET, HEAD 2025-11-03T17:36:11.0890277Z (APIServer pid=69395) INFO 11-03 17:36:11 [launcher.py:46] Route: /redoc, Methods: GET, HEAD 2025-11-03T17:36:11.0891332Z (APIServer pid=69395) INFO 11-03 17:36:11 [launcher.py:46] Route: /health, Methods: GET 2025-11-03T17:36:11.0892291Z (APIServer pid=69395) INFO 11-03 17:36:11 [launcher.py:46] Route: /load, Methods: GET 2025-11-03T17:36:11.0892817Z (APIServer pid=69395) INFO 11-03 17:36:11 [launcher.py:46] Route: /ping, Methods: POST 2025-11-03T17:36:11.0893308Z (APIServer pid=69395) INFO 11-03 17:36:11 [launcher.py:46] Route: /ping, Methods: GET 2025-11-03T17:36:11.0893963Z (APIServer pid=69395) INFO 11-03 17:36:11 [launcher.py:46] Route: /tokenize, Methods: POST 2025-11-03T17:36:11.0894496Z (APIServer pid=69395) INFO 11-03 17:36:11 [launcher.py:46] Route: /detokenize, Methods: POST 2025-11-03T17:36:11.0895022Z (APIServer pid=69395) INFO 11-03 17:36:11 [launcher.py:46] Route: /v1/models, Methods: GET 2025-11-03T17:36:11.0895516Z (APIServer pid=69395) INFO 11-03 17:36:11 [launcher.py:46] Route: /version, Methods: GET 2025-11-03T17:36:11.0896049Z (APIServer pid=69395) INFO 11-03 17:36:11 [launcher.py:46] Route: /v1/responses, Methods: POST 2025-11-03T17:36:11.0896632Z (APIServer pid=69395) INFO 11-03 17:36:11 [launcher.py:46] Route: /v1/responses/{response_id}, Methods: GET 2025-11-03T17:36:11.0897245Z (APIServer pid=69395) INFO 11-03 17:36:11 [launcher.py:46] Route: /v1/responses/{response_id}/cancel, Methods: POST 2025-11-03T17:36:11.0897845Z (APIServer pid=69395) INFO 11-03 17:36:11 [launcher.py:46] Route: /v1/chat/completions, Methods: POST 2025-11-03T17:36:11.0898514Z (APIServer pid=69395) INFO 11-03 17:36:11 [launcher.py:46] Route: /v1/completions, Methods: POST 2025-11-03T17:36:11.0899078Z (APIServer pid=69395) INFO 11-03 17:36:11 [launcher.py:46] Route: /v1/embeddings, Methods: POST 2025-11-03T17:36:11.0899601Z (APIServer pid=69395) INFO 11-03 17:36:11 [launcher.py:46] Route: /pooling, Methods: POST 2025-11-03T17:36:11.0900174Z (APIServer pid=69395) INFO 11-03 17:36:11 [launcher.py:46] Route: /classify, Methods: POST 2025-11-03T17:36:11.0900679Z (APIServer pid=69395) INFO 11-03 17:36:11 [launcher.py:46] Route: /score, Methods: POST 2025-11-03T17:36:11.0901183Z (APIServer pid=69395) INFO 11-03 17:36:11 [launcher.py:46] Route: /v1/score, Methods: POST 2025-11-03T17:36:11.0901897Z (APIServer pid=69395) INFO 11-03 17:36:11 [launcher.py:46] Route: /v1/audio/transcriptions, Methods: POST 2025-11-03T17:36:11.0902499Z (APIServer pid=69395) INFO 11-03 17:36:11 [launcher.py:46] Route: /v1/audio/translations, Methods: POST 2025-11-03T17:36:11.0903040Z (APIServer pid=69395) INFO 11-03 17:36:11 [launcher.py:46] Route: /rerank, Methods: POST 2025-11-03T17:36:11.0903552Z (APIServer pid=69395) INFO 11-03 17:36:11 [launcher.py:46] Route: /v1/rerank, Methods: POST 2025-11-03T17:36:11.0904067Z (APIServer pid=69395) INFO 11-03 17:36:11 [launcher.py:46] Route: /v2/rerank, Methods: POST 2025-11-03T17:36:11.0904622Z (APIServer pid=69395) INFO 11-03 17:36:11 [launcher.py:46] Route: /scale_elastic_ep, Methods: POST 2025-11-03T17:36:11.0905187Z (APIServer pid=69395) INFO 11-03 17:36:11 [launcher.py:46] Route: /is_scaling_elastic_ep, Methods: POST 2025-11-03T17:36:11.0905732Z (APIServer pid=69395) INFO 11-03 17:36:11 [launcher.py:46] Route: /invocations, Methods: POST 2025-11-03T17:36:11.0906253Z (APIServer pid=69395) INFO 11-03 17:36:11 [launcher.py:46] Route: /metrics, Methods: GET 2025-11-03T17:36:11.1042213Z (APIServer pid=69395) INFO: Started server process [69395] 2025-11-03T17:36:11.1042827Z (APIServer pid=69395) INFO: Waiting for application startup. 2025-11-03T17:36:11.3204630Z (APIServer pid=69395) INFO: Application startup complete. 2025-11-03T17:36:11.5861094Z (APIServer pid=69395) INFO: 127.0.0.1:51106 - "GET /health HTTP/1.1" 200 OK 2025-11-03T17:36:11.5950643Z (APIServer pid=69395) INFO: 127.0.0.1:51112 - "GET /v1/models HTTP/1.1" 200 OK 2025-11-03T17:36:11.8289429Z (APIServer pid=69395) INFO: 127.0.0.1:51112 - "POST /v1/completions HTTP/1.1" 200 OK 2025-11-03T17:36:11.9430033Z (APIServer pid=69395) INFO: 127.0.0.1:51112 - "POST /v1/completions HTTP/1.1" 200 OK 2025-11-03T17:36:12.0888765Z (APIServer pid=69395) INFO: 127.0.0.1:51112 - "POST /v1/completions HTTP/1.1" 200 OK 2025-11-03T17:36:12.2206602Z (APIServer pid=69395) INFO: 127.0.0.1:51112 - "POST /v1/completions HTTP/1.1" 200 OK 2025-11-03T17:36:12.3624211Z (APIServer pid=69395) INFO: 127.0.0.1:51112 - "POST /v1/completions HTTP/1.1" 200 OK 2025-11-03T17:36:12.3696829Z (APIServer pid=69395) INFO: 127.0.0.1:51112 - "POST /v1/completions HTTP/1.1" 200 OK 2025-11-03T17:36:12.5162055Z (APIServer pid=69395) INFO 11-03 17:36:12 [launcher.py:110] Shutting down FastAPI HTTP server. 2025-11-03T17:36:12.5165088Z (Worker_TP0 pid=69637) INFO 11-03 17:36:12 [multiproc_executor.py:583] Parent process exited, terminating worker 2025-11-03T17:36:12.5166947Z (Worker_TP0 pid=69637) INFO 11-03 17:36:12 [multiproc_executor.py:624] WorkerProc shutting down. 2025-11-03T17:36:12.5169217Z (Worker_TP1 pid=69638) INFO 11-03 17:36:12 [multiproc_executor.py:583] Parent process exited, terminating worker 2025-11-03T17:36:12.5170603Z (Worker_TP1 pid=69638) INFO 11-03 17:36:12 [multiproc_executor.py:624] WorkerProc shutting down. 2025-11-03T17:36:14.8357327Z (APIServer pid=69395) INFO: Shutting down 2025-11-03T17:36:14.9362430Z (APIServer pid=69395) INFO: Waiting for application shutdown. 2025-11-03T17:36:14.9363713Z (APIServer pid=69395) INFO: Application shutdown complete. 2025-11-03T17:36:15.5511140Z PASSED 2025-11-03T17:36:15.5653799Z distributed/test_sequence_parallel.py::test_tp_sp_generation[RedHatAI/Meta-Llama-3.1-8B-Instruct-FP8-parallel_setup18-mp-auto-test_options18] Fork a new process to run a test 69841 2025-11-03T17:36:15.5669319Z Fork a new process to run a test 0 2025-11-03T17:36:16.3892176Z INFO 11-03 17:36:16 [model.py:653] Resolved architecture: LlamaForCausalLM 2025-11-03T17:36:16.3893082Z `torch_dtype` is deprecated! Use `dtype` instead! 2025-11-03T17:36:16.3895513Z WARNING 11-03 17:36:16 [model.py:1942] Casting torch.bfloat16 to torch.float16. 2025-11-03T17:36:16.3896364Z INFO 11-03 17:36:16 [model.py:1714] Using max model len 2048 2025-11-03T17:36:16.4972451Z Launching RemoteOpenAIServer with: vllm serve RedHatAI/Meta-Llama-3.1-8B-Instruct-FP8 --dtype float16 --max-model-len 2048 --max-num-seqs 8 --enforce-eager --tokenizer-mode auto --tensor-parallel-size 2 --pipeline-parallel-size 1 --distributed-executor-backend mp --compilation_config {"level": 3, "custom_ops": ["+rms_norm"], "compile_sizes": [4, 8], "pass_config": {"enable_sequence_parallelism": true, "enable_fusion": true, "enable_noop": true}} --load-format dummy --port 55847 --seed 0 2025-11-03T17:36:17.0226278Z /opt/conda/envs/py_3.12/lib/python3.12/site-packages/torch/cuda/__init__.py:65: FutureWarning: The pynvml package is deprecated. Please install nvidia-ml-py instead. If you did not install pynvml directly, please report this to the maintainers of the package that installed pynvml for you. 2025-11-03T17:36:17.0227527Z import pynvml # type: ignore[import] 2025-11-03T17:36:19.2589251Z INFO 11-03 17:36:19 [__init__.py:224] Automatically detected platform cuda. 2025-11-03T17:36:22.9124129Z (APIServer pid=69890) INFO 11-03 17:36:22 [api_server.py:1870] vLLM API server version 0.11.0rc2.dev389+ge51928192.d20251103 2025-11-03T17:36:22.9153240Z (APIServer pid=69890) INFO 11-03 17:36:22 [utils.py:239] non-default args: {'model_tag': 'RedHatAI/Meta-Llama-3.1-8B-Instruct-FP8', 'port': 55847, 'model': 'RedHatAI/Meta-Llama-3.1-8B-Instruct-FP8', 'dtype': 'float16', 'seed': 0, 'max_model_len': 2048, 'enforce_eager': True, 'load_format': 'dummy', 'distributed_executor_backend': 'mp', 'tensor_parallel_size': 2, 'max_num_seqs': 8, 'compilation_config': {'level': 3, 'debug_dump_path': None, 'cache_dir': '', 'backend': '', 'custom_ops': ['+rms_norm'], 'splitting_ops': None, 'use_inductor': True, 'compile_sizes': [4, 8], 'inductor_compile_config': {'enable_auto_functionalized_v2': False}, 'inductor_passes': {}, 'cudagraph_mode': None, 'use_cudagraph': True, 'cudagraph_num_of_warmups': 0, 'cudagraph_capture_sizes': None, 'cudagraph_copy_inputs': False, 'full_cuda_graph': False, 'use_inductor_graph_partition': False, 'pass_config': {'enable_fusion': True, 'enable_noop': True, 'enable_sequence_parallelism': True}, 'max_capture_size': None, 'local_cache_dir': None}} 2025-11-03T17:36:23.0768132Z (APIServer pid=69890) INFO 11-03 17:36:23 [model.py:653] Resolved architecture: LlamaForCausalLM 2025-11-03T17:36:23.0769117Z (APIServer pid=69890) `torch_dtype` is deprecated! Use `dtype` instead! 2025-11-03T17:36:23.0769803Z (APIServer pid=69890) WARNING 11-03 17:36:23 [model.py:1942] Casting torch.bfloat16 to torch.float16. 2025-11-03T17:36:23.0770499Z (APIServer pid=69890) INFO 11-03 17:36:23 [model.py:1714] Using max model len 2048 2025-11-03T17:36:23.4018968Z (APIServer pid=69890) INFO 11-03 17:36:23 [scheduler.py:225] Chunked prefill is enabled with max_num_batched_tokens=2048. 2025-11-03T17:36:23.4389113Z (APIServer pid=69890) INFO 11-03 17:36:23 [vllm.py:354] Cudagraph is disabled under eager mode 2025-11-03T17:36:24.5766920Z /opt/conda/envs/py_3.12/lib/python3.12/site-packages/torch/cuda/__init__.py:65: FutureWarning: The pynvml package is deprecated. Please install nvidia-ml-py instead. If you did not install pynvml directly, please report this to the maintainers of the package that installed pynvml for you. 2025-11-03T17:36:24.5768532Z import pynvml # type: ignore[import] 2025-11-03T17:36:26.9166412Z INFO 11-03 17:36:26 [__init__.py:224] Automatically detected platform cuda. 2025-11-03T17:36:30.7614084Z (EngineCore_DP0 pid=70013) INFO 11-03 17:36:30 [core.py:727] Waiting for init message from front-end. 2025-11-03T17:36:30.7883371Z (EngineCore_DP0 pid=70013) INFO 11-03 17:36:30 [core.py:94] Initializing a V1 LLM engine (v0.11.0rc2.dev389+ge51928192.d20251103) with config: model='RedHatAI/Meta-Llama-3.1-8B-Instruct-FP8', speculative_config=None, tokenizer='RedHatAI/Meta-Llama-3.1-8B-Instruct-FP8', skip_tokenizer_init=False, tokenizer_mode=auto, revision=None, tokenizer_revision=None, trust_remote_code=False, dtype=torch.float16, max_seq_len=2048, download_dir=None, load_format=dummy, tensor_parallel_size=2, pipeline_parallel_size=1, data_parallel_size=1, disable_custom_all_reduce=False, quantization=compressed-tensors, enforce_eager=True, kv_cache_dtype=auto, device_config=cuda, structured_outputs_config=StructuredOutputsConfig(backend='auto', disable_fallback=False, disable_any_whitespace=False, disable_additional_properties=False, reasoning_parser=''), observability_config=ObservabilityConfig(show_hidden_metrics_for_version=None, otlp_traces_endpoint=None, collect_detailed_traces=None), seed=0, served_model_name=RedHatAI/Meta-Llama-3.1-8B-Instruct-FP8, enable_prefix_caching=True, chunked_prefill_enabled=True, pooler_config=None, compilation_config={'level': 3, 'debug_dump_path': None, 'cache_dir': '', 'backend': '', 'custom_ops': ['+rms_norm', '+rms_norm'], 'splitting_ops': ['vllm.unified_attention', 'vllm.unified_attention_with_output', 'vllm.unified_mla_attention', 'vllm.unified_mla_attention_with_output', 'vllm.mamba_mixer2', 'vllm.mamba_mixer', 'vllm.short_conv', 'vllm.linear_attention', 'vllm.plamo2_mamba_mixer', 'vllm.gdn_attention', 'vllm.sparse_attn_indexer'], 'use_inductor': True, 'compile_sizes': [8, 4], 'inductor_compile_config': {'enable_auto_functionalized_v2': False}, 'inductor_passes': {}, 'cudagraph_mode': , 'use_cudagraph': False, 'cudagraph_num_of_warmups': 0, 'cudagraph_capture_sizes': [], 'cudagraph_copy_inputs': False, 'full_cuda_graph': False, 'use_inductor_graph_partition': False, 'pass_config': {'enable_fusion': True, 'enable_noop': True, 'enable_sequence_parallelism': True}, 'max_capture_size': 0, 'local_cache_dir': None} 2025-11-03T17:36:30.7890157Z (EngineCore_DP0 pid=70013) WARNING 11-03 17:36:30 [multiproc_executor.py:744] Reducing Torch parallelism from 24 threads to 1 to avoid unnecessary CPU contention. Set OMP_NUM_THREADS in the external environment to tune this value as needed. 2025-11-03T17:36:30.7891791Z (EngineCore_DP0 pid=70013) INFO 11-03 17:36:30 [shm_broadcast.py:302] vLLM message queue communication handle: Handle(local_reader_ranks=[0, 1], buffer_handle=(2, 16777216, 10, 'psm_15f15313'), local_subscribe_addr='ipc:///tmp/1de0072e-fcfe-4b0f-b760-3a090abddadd', remote_subscribe_addr=None, remote_addr_ipv6=False) 2025-11-03T17:36:31.2711328Z /opt/conda/envs/py_3.12/lib/python3.12/site-packages/torch/cuda/__init__.py:65: FutureWarning: The pynvml package is deprecated. Please install nvidia-ml-py instead. If you did not install pynvml directly, please report this to the maintainers of the package that installed pynvml for you. 2025-11-03T17:36:31.2712573Z import pynvml # type: ignore[import] 2025-11-03T17:36:31.2882956Z /opt/conda/envs/py_3.12/lib/python3.12/site-packages/torch/cuda/__init__.py:65: FutureWarning: The pynvml package is deprecated. Please install nvidia-ml-py instead. If you did not install pynvml directly, please report this to the maintainers of the package that installed pynvml for you. 2025-11-03T17:36:31.2884139Z import pynvml # type: ignore[import] 2025-11-03T17:36:33.6447442Z INFO 11-03 17:36:33 [__init__.py:224] Automatically detected platform cuda. 2025-11-03T17:36:33.6677221Z INFO 11-03 17:36:33 [__init__.py:224] Automatically detected platform cuda. 2025-11-03T17:36:38.5130733Z INFO 11-03 17:36:38 [shm_broadcast.py:302] vLLM message queue communication handle: Handle(local_reader_ranks=[0], buffer_handle=(1, 10485760, 10, 'psm_267e7c7b'), local_subscribe_addr='ipc:///tmp/5e7fb234-f29e-4bfa-9381-fbd9e43bdb6d', remote_subscribe_addr=None, remote_addr_ipv6=False) 2025-11-03T17:36:38.5134368Z INFO 11-03 17:36:38 [shm_broadcast.py:302] vLLM message queue communication handle: Handle(local_reader_ranks=[0], buffer_handle=(1, 10485760, 10, 'psm_36028868'), local_subscribe_addr='ipc:///tmp/d458b70b-6444-4f10-937a-d410752f2fa1', remote_subscribe_addr=None, remote_addr_ipv6=False) 2025-11-03T17:36:38.7086106Z [Gloo] Rank 1 is connected to 1 peer ranks. Expected number of connected peer ranks is : 1 2025-11-03T17:36:38.7087213Z [Gloo] Rank 0 is connected to 1 peer ranks. Expected number of connected peer ranks is : 1 2025-11-03T17:36:38.7130822Z [Gloo] Rank 1 is connected to 1 peer ranks. Expected number of connected peer ranks is : 1 2025-11-03T17:36:38.7131295Z [Gloo] Rank 0 is connected to 1 peer ranks. Expected number of connected peer ranks is : 1 2025-11-03T17:36:38.7175491Z INFO 11-03 17:36:38 [__init__.py:1146] Found nccl from library libnccl.so.2 2025-11-03T17:36:38.7177060Z INFO 11-03 17:36:38 [__init__.py:1146] Found nccl from library libnccl.so.2 2025-11-03T17:36:38.7177805Z INFO 11-03 17:36:38 [pynccl.py:109] vLLM is using nccl==2.27.5 2025-11-03T17:36:38.7178152Z INFO 11-03 17:36:38 [pynccl.py:109] vLLM is using nccl==2.27.5 2025-11-03T17:36:38.8889617Z WARNING 11-03 17:36:38 [symm_mem.py:60] SymmMemCommunicator: Device capability 8.9 not supported, communicator is not available. 2025-11-03T17:36:38.8891211Z WARNING 11-03 17:36:38 [symm_mem.py:60] SymmMemCommunicator: Device capability 8.9 not supported, communicator is not available. 2025-11-03T17:36:38.8924601Z INFO 11-03 17:36:38 [custom_all_reduce.py:37] Skipping P2P check and trusting the driver's P2P report. 2025-11-03T17:36:38.8925362Z INFO 11-03 17:36:38 [custom_all_reduce.py:37] Skipping P2P check and trusting the driver's P2P report. 2025-11-03T17:36:38.8926638Z WARNING 11-03 17:36:38 [custom_all_reduce.py:162] Custom allreduce is disabled because your platform lacks GPU P2P capability or P2P test failed. To silence this warning, specify disable_custom_all_reduce=True explicitly. 2025-11-03T17:36:38.8928059Z WARNING 11-03 17:36:38 [custom_all_reduce.py:162] Custom allreduce is disabled because your platform lacks GPU P2P capability or P2P test failed. To silence this warning, specify disable_custom_all_reduce=True explicitly. 2025-11-03T17:36:38.9726315Z INFO 11-03 17:36:38 [shm_broadcast.py:302] vLLM message queue communication handle: Handle(local_reader_ranks=[1], buffer_handle=(1, 4194304, 6, 'psm_ef5e9f3e'), local_subscribe_addr='ipc:///tmp/2b635c51-2007-4bc9-b2f5-a98844afa030', remote_subscribe_addr=None, remote_addr_ipv6=False) 2025-11-03T17:36:38.9763154Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:36:38.9765474Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:36:38.9771192Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:36:38.9779401Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:36:39.0522303Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:36:39.0530927Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:36:39.0542486Z [Gloo] Rank 0 is connected to 1 peer ranks. Expected number of connected peer ranks is : 1 2025-11-03T17:36:39.0543036Z [Gloo] Rank 1 is connected to 1 peer ranks. Expected number of connected peer ranks is : 1 2025-11-03T17:36:39.0545373Z INFO 11-03 17:36:39 [__init__.py:1146] Found nccl from library libnccl.so.2 2025-11-03T17:36:39.0545943Z INFO 11-03 17:36:39 [__init__.py:1146] Found nccl from library libnccl.so.2 2025-11-03T17:36:39.0546316Z INFO 11-03 17:36:39 [pynccl.py:109] vLLM is using nccl==2.27.5 2025-11-03T17:36:39.0546637Z INFO 11-03 17:36:39 [pynccl.py:109] vLLM is using nccl==2.27.5 2025-11-03T17:36:39.0890684Z INFO 11-03 17:36:39 [parallel_state.py:1231] rank 1 in world size 2 is assigned as DP rank 0, PP rank 0, TP rank 1, EP rank 1 2025-11-03T17:36:39.0891292Z INFO 11-03 17:36:39 [parallel_state.py:1231] rank 0 in world size 2 is assigned as DP rank 0, PP rank 0, TP rank 0, EP rank 0 2025-11-03T17:36:39.3264606Z INFO 11-03 17:36:39 [topk_topp_sampler.py:59] Using FlashInfer for top-p & top-k sampling. 2025-11-03T17:36:39.3321397Z INFO 11-03 17:36:39 [topk_topp_sampler.py:59] Using FlashInfer for top-p & top-k sampling. 2025-11-03T17:36:39.3644360Z (Worker_TP0 pid=70132) INFO 11-03 17:36:39 [gpu_model_runner.py:2840] Starting to load model RedHatAI/Meta-Llama-3.1-8B-Instruct-FP8... 2025-11-03T17:36:39.3698533Z (Worker_TP1 pid=70133) INFO 11-03 17:36:39 [gpu_model_runner.py:2840] Starting to load model RedHatAI/Meta-Llama-3.1-8B-Instruct-FP8... 2025-11-03T17:36:39.5790470Z (Worker_TP0 pid=70132) INFO 11-03 17:36:39 [gpu_model_runner.py:2870] Loading model from scratch... 2025-11-03T17:36:39.5798433Z (Worker_TP1 pid=70133) INFO 11-03 17:36:39 [gpu_model_runner.py:2870] Loading model from scratch... 2025-11-03T17:36:39.6707502Z (Worker_TP0 pid=70132) INFO 11-03 17:36:39 [cuda.py:404] Using Flash Attention backend on V1 engine. 2025-11-03T17:36:39.6712713Z (Worker_TP1 pid=70133) INFO 11-03 17:36:39 [cuda.py:404] Using Flash Attention backend on V1 engine. 2025-11-03T17:36:40.5348522Z (Worker_TP0 pid=70132) INFO 11-03 17:36:40 [gpu_model_runner.py:2902] Model loading took 4.2624 GiB and 0.541624 seconds 2025-11-03T17:36:40.5460663Z (Worker_TP1 pid=70133) INFO 11-03 17:36:40 [gpu_model_runner.py:2902] Model loading took 4.2624 GiB and 0.544466 seconds 2025-11-03T17:36:48.4860003Z (Worker_TP0 pid=70132) INFO 11-03 17:36:48 [backends.py:594] Using cache directory: /var/lib/jenkins/.cache/vllm/torch_compile_cache/9e2073b733/rank_0_0/backbone for vLLM's torch.compile 2025-11-03T17:36:48.4863482Z (Worker_TP0 pid=70132) INFO 11-03 17:36:48 [backends.py:608] Dynamo bytecode transform time: 7.67 s 2025-11-03T17:36:48.6345693Z (Worker_TP1 pid=70133) INFO 11-03 17:36:48 [backends.py:594] Using cache directory: /var/lib/jenkins/.cache/vllm/torch_compile_cache/9e2073b733/rank_1_0/backbone for vLLM's torch.compile 2025-11-03T17:36:48.6350561Z (Worker_TP1 pid=70133) INFO 11-03 17:36:48 [backends.py:608] Dynamo bytecode transform time: 7.82 s 2025-11-03T17:36:51.6991437Z (Worker_TP0 pid=70132) INFO 11-03 17:36:51 [backends.py:214] Cache the graph for dynamic shape for later use 2025-11-03T17:36:51.8953602Z (Worker_TP1 pid=70133) INFO 11-03 17:36:51 [backends.py:214] Cache the graph for dynamic shape for later use 2025-11-03T17:37:17.2627350Z (Worker_TP0 pid=70132) INFO 11-03 17:37:17 [backends.py:241] Compiling a graph for dynamic shape takes 28.15 s 2025-11-03T17:37:18.0381694Z (Worker_TP1 pid=70133) INFO 11-03 17:37:18 [backends.py:241] Compiling a graph for dynamic shape takes 28.77 s 2025-11-03T17:37:39.2882159Z (Worker_TP0 pid=70132) INFO 11-03 17:37:39 [gpu_worker.py:314] Available KV cache memory: 15.29 GiB 2025-11-03T17:37:39.2980469Z (Worker_TP1 pid=70133) INFO 11-03 17:37:39 [gpu_worker.py:314] Available KV cache memory: 15.29 GiB 2025-11-03T17:37:39.6530888Z (EngineCore_DP0 pid=70013) INFO 11-03 17:37:39 [kv_cache_utils.py:1199] GPU KV cache size: 250,528 tokens 2025-11-03T17:37:39.6531800Z (EngineCore_DP0 pid=70013) INFO 11-03 17:37:39 [kv_cache_utils.py:1204] Maximum concurrency for 2,048 tokens per request: 122.33x 2025-11-03T17:37:39.6532878Z (EngineCore_DP0 pid=70013) INFO 11-03 17:37:39 [kv_cache_utils.py:1199] GPU KV cache size: 250,528 tokens 2025-11-03T17:37:39.6533708Z (EngineCore_DP0 pid=70013) INFO 11-03 17:37:39 [kv_cache_utils.py:1204] Maximum concurrency for 2,048 tokens per request: 122.33x 2025-11-03T17:37:39.6650258Z (Worker_TP0 pid=70132) INFO 11-03 17:37:39 [gpu_worker.py:351] Compile and warming up model for size 8 2025-11-03T17:37:39.6650974Z (Worker_TP1 pid=70133) INFO 11-03 17:37:39 [gpu_worker.py:351] Compile and warming up model for size 8 2025-11-03T17:37:40.9079388Z (Worker_TP0 pid=70132) INFO 11-03 17:37:40 [backends.py:216] Cache the graph of shape 8 for later use 2025-11-03T17:37:40.9218023Z (Worker_TP1 pid=70133) INFO 11-03 17:37:40 [backends.py:216] Cache the graph of shape 8 for later use 2025-11-03T17:38:13.4555143Z (Worker_TP0 pid=70132) INFO 11-03 17:38:13 [backends.py:243] Compiling a graph for shape 8 takes 33.79 s 2025-11-03T17:38:13.9486037Z (Worker_TP1 pid=70133) INFO 11-03 17:38:13 [backends.py:243] Compiling a graph for shape 8 takes 34.28 s 2025-11-03T17:38:14.0329596Z (Worker_TP1 pid=70133) INFO 11-03 17:38:14 [gpu_worker.py:351] Compile and warming up model for size 4 2025-11-03T17:38:14.0346908Z (Worker_TP0 pid=70132) INFO 11-03 17:38:14 [gpu_worker.py:351] Compile and warming up model for size 4 2025-11-03T17:38:15.2398182Z (Worker_TP0 pid=70132) INFO 11-03 17:38:15 [backends.py:216] Cache the graph of shape 4 for later use 2025-11-03T17:38:15.2566600Z (Worker_TP1 pid=70133) INFO 11-03 17:38:15 [backends.py:216] Cache the graph of shape 4 for later use 2025-11-03T17:38:39.6656289Z (EngineCore_DP0 pid=70013) INFO 11-03 17:38:39 [shm_broadcast.py:482] No available shared memory broadcast block found in 60 seconds. This typically happens when some processes are hanging or doing some time-consuming work (e.g. compilation). 2025-11-03T17:38:57.4423760Z (Worker_TP1 pid=70133) INFO 11-03 17:38:57 [backends.py:243] Compiling a graph for shape 4 takes 43.41 s 2025-11-03T17:38:57.4442527Z (Worker_TP1 pid=70133) INFO 11-03 17:38:57 [monitor.py:33] torch.compile takes 114.28 s in total 2025-11-03T17:38:57.6428648Z (Worker_TP0 pid=70132) INFO 11-03 17:38:57 [backends.py:243] Compiling a graph for shape 4 takes 43.61 s 2025-11-03T17:38:57.6450614Z (Worker_TP0 pid=70132) INFO 11-03 17:38:57 [monitor.py:33] torch.compile takes 113.22 s in total 2025-11-03T17:38:57.6707352Z (EngineCore_DP0 pid=70013) INFO 11-03 17:38:57 [core.py:240] init engine (profile, create kv cache, warmup model) took 137.06 seconds 2025-11-03T17:38:58.1534039Z (EngineCore_DP0 pid=70013) INFO 11-03 17:38:58 [vllm.py:354] Cudagraph is disabled under eager mode 2025-11-03T17:38:58.3381602Z (APIServer pid=69890) INFO 11-03 17:38:58 [loggers.py:172] Engine 000: vllm cache_config_info with initialization after num_gpu_blocks is: 15658 2025-11-03T17:38:58.3510195Z (EngineCore_DP0 pid=70013) INFO 11-03 17:38:58 [gc_utils.py:40] GC Debug Config. enabled:False,top_objects:-1 2025-11-03T17:38:58.3726142Z (APIServer pid=69890) INFO 11-03 17:38:58 [api_server.py:1628] Supported tasks: ['generate'] 2025-11-03T17:38:58.4307854Z (APIServer pid=69890) WARNING 11-03 17:38:58 [model.py:1581] Default sampling parameters have been overridden by the model's Hugging Face generation config recommended from the model creator. If this is not intended, please relaunch vLLM instance with `--generation-config vllm`. 2025-11-03T17:38:58.4309917Z (APIServer pid=69890) INFO 11-03 17:38:58 [serving_responses.py:157] Using default chat sampling params from model: {'temperature': 0.6, 'top_p': 0.9} 2025-11-03T17:38:58.4847380Z (APIServer pid=69890) INFO 11-03 17:38:58 [serving_chat.py:159] Using default chat sampling params from model: {'temperature': 0.6, 'top_p': 0.9} 2025-11-03T17:38:58.5288626Z (APIServer pid=69890) INFO 11-03 17:38:58 [serving_completion.py:67] Using default completion sampling params from model: {'temperature': 0.6, 'top_p': 0.9} 2025-11-03T17:38:58.5290449Z (APIServer pid=69890) INFO 11-03 17:38:58 [api_server.py:1939] Starting vLLM API server 0 on http://0.0.0.0:55847 2025-11-03T17:38:58.5291656Z (APIServer pid=69890) INFO 11-03 17:38:58 [launcher.py:38] Available routes are: 2025-11-03T17:38:58.5292782Z (APIServer pid=69890) INFO 11-03 17:38:58 [launcher.py:46] Route: /openapi.json, Methods: HEAD, GET 2025-11-03T17:38:58.5293983Z (APIServer pid=69890) INFO 11-03 17:38:58 [launcher.py:46] Route: /docs, Methods: HEAD, GET 2025-11-03T17:38:58.5294735Z (APIServer pid=69890) INFO 11-03 17:38:58 [launcher.py:46] Route: /docs/oauth2-redirect, Methods: HEAD, GET 2025-11-03T17:38:58.5295460Z (APIServer pid=69890) INFO 11-03 17:38:58 [launcher.py:46] Route: /redoc, Methods: HEAD, GET 2025-11-03T17:38:58.5296096Z (APIServer pid=69890) INFO 11-03 17:38:58 [launcher.py:46] Route: /health, Methods: GET 2025-11-03T17:38:58.5296739Z (APIServer pid=69890) INFO 11-03 17:38:58 [launcher.py:46] Route: /load, Methods: GET 2025-11-03T17:38:58.5297306Z (APIServer pid=69890) INFO 11-03 17:38:58 [launcher.py:46] Route: /ping, Methods: POST 2025-11-03T17:38:58.5297802Z (APIServer pid=69890) INFO 11-03 17:38:58 [launcher.py:46] Route: /ping, Methods: GET 2025-11-03T17:38:58.5298316Z (APIServer pid=69890) INFO 11-03 17:38:58 [launcher.py:46] Route: /tokenize, Methods: POST 2025-11-03T17:38:58.5298835Z (APIServer pid=69890) INFO 11-03 17:38:58 [launcher.py:46] Route: /detokenize, Methods: POST 2025-11-03T17:38:58.5299351Z (APIServer pid=69890) INFO 11-03 17:38:58 [launcher.py:46] Route: /v1/models, Methods: GET 2025-11-03T17:38:58.5299857Z (APIServer pid=69890) INFO 11-03 17:38:58 [launcher.py:46] Route: /version, Methods: GET 2025-11-03T17:38:58.5300382Z (APIServer pid=69890) INFO 11-03 17:38:58 [launcher.py:46] Route: /v1/responses, Methods: POST 2025-11-03T17:38:58.5300963Z (APIServer pid=69890) INFO 11-03 17:38:58 [launcher.py:46] Route: /v1/responses/{response_id}, Methods: GET 2025-11-03T17:38:58.5301579Z (APIServer pid=69890) INFO 11-03 17:38:58 [launcher.py:46] Route: /v1/responses/{response_id}/cancel, Methods: POST 2025-11-03T17:38:58.5302369Z (APIServer pid=69890) INFO 11-03 17:38:58 [launcher.py:46] Route: /v1/chat/completions, Methods: POST 2025-11-03T17:38:58.5302935Z (APIServer pid=69890) INFO 11-03 17:38:58 [launcher.py:46] Route: /v1/completions, Methods: POST 2025-11-03T17:38:58.5303480Z (APIServer pid=69890) INFO 11-03 17:38:58 [launcher.py:46] Route: /v1/embeddings, Methods: POST 2025-11-03T17:38:58.5304005Z (APIServer pid=69890) INFO 11-03 17:38:58 [launcher.py:46] Route: /pooling, Methods: POST 2025-11-03T17:38:58.5304518Z (APIServer pid=69890) INFO 11-03 17:38:58 [launcher.py:46] Route: /classify, Methods: POST 2025-11-03T17:38:58.5305024Z (APIServer pid=69890) INFO 11-03 17:38:58 [launcher.py:46] Route: /score, Methods: POST 2025-11-03T17:38:58.5305527Z (APIServer pid=69890) INFO 11-03 17:38:58 [launcher.py:46] Route: /v1/score, Methods: POST 2025-11-03T17:38:58.5306082Z (APIServer pid=69890) INFO 11-03 17:38:58 [launcher.py:46] Route: /v1/audio/transcriptions, Methods: POST 2025-11-03T17:38:58.5306667Z (APIServer pid=69890) INFO 11-03 17:38:58 [launcher.py:46] Route: /v1/audio/translations, Methods: POST 2025-11-03T17:38:58.5307340Z (APIServer pid=69890) INFO 11-03 17:38:58 [launcher.py:46] Route: /rerank, Methods: POST 2025-11-03T17:38:58.5307880Z (APIServer pid=69890) INFO 11-03 17:38:58 [launcher.py:46] Route: /v1/rerank, Methods: POST 2025-11-03T17:38:58.5308454Z (APIServer pid=69890) INFO 11-03 17:38:58 [launcher.py:46] Route: /v2/rerank, Methods: POST 2025-11-03T17:38:58.5308986Z (APIServer pid=69890) INFO 11-03 17:38:58 [launcher.py:46] Route: /scale_elastic_ep, Methods: POST 2025-11-03T17:38:58.5309721Z (APIServer pid=69890) INFO 11-03 17:38:58 [launcher.py:46] Route: /is_scaling_elastic_ep, Methods: POST 2025-11-03T17:38:58.5310277Z (APIServer pid=69890) INFO 11-03 17:38:58 [launcher.py:46] Route: /invocations, Methods: POST 2025-11-03T17:38:58.5310790Z (APIServer pid=69890) INFO 11-03 17:38:58 [launcher.py:46] Route: /metrics, Methods: GET 2025-11-03T17:38:58.5446407Z (APIServer pid=69890) INFO: Started server process [69890] 2025-11-03T17:38:58.5448126Z (APIServer pid=69890) INFO: Waiting for application startup. 2025-11-03T17:38:58.7853956Z (APIServer pid=69890) INFO: Application startup complete. 2025-11-03T17:38:58.8597504Z (APIServer pid=69890) INFO: 127.0.0.1:49364 - "GET /health HTTP/1.1" 200 OK 2025-11-03T17:38:59.1034051Z (APIServer pid=69890) INFO: 127.0.0.1:49372 - "GET /v1/models HTTP/1.1" 200 OK 2025-11-03T17:38:59.2646345Z (APIServer pid=69890) INFO: 127.0.0.1:49372 - "POST /v1/completions HTTP/1.1" 200 OK 2025-11-03T17:38:59.3824831Z (APIServer pid=69890) INFO: 127.0.0.1:49372 - "POST /v1/completions HTTP/1.1" 200 OK 2025-11-03T17:38:59.5737141Z (APIServer pid=69890) INFO: 127.0.0.1:49372 - "POST /v1/completions HTTP/1.1" 200 OK 2025-11-03T17:38:59.7272306Z (APIServer pid=69890) INFO: 127.0.0.1:49372 - "POST /v1/completions HTTP/1.1" 200 OK 2025-11-03T17:38:59.8446328Z (APIServer pid=69890) INFO: 127.0.0.1:49372 - "POST /v1/completions HTTP/1.1" 200 OK 2025-11-03T17:38:59.8523769Z (APIServer pid=69890) INFO: 127.0.0.1:49372 - "POST /v1/completions HTTP/1.1" 200 OK 2025-11-03T17:38:59.9729467Z (APIServer pid=69890) INFO 11-03 17:38:59 [launcher.py:110] Shutting down FastAPI HTTP server. 2025-11-03T17:38:59.9732080Z (Worker_TP0 pid=70132) INFO 11-03 17:38:59 [multiproc_executor.py:583] Parent process exited, terminating worker 2025-11-03T17:38:59.9733768Z (Worker_TP0 pid=70132) INFO 11-03 17:38:59 [multiproc_executor.py:624] WorkerProc shutting down. 2025-11-03T17:38:59.9737529Z (Worker_TP1 pid=70133) INFO 11-03 17:38:59 [multiproc_executor.py:583] Parent process exited, terminating worker 2025-11-03T17:38:59.9739219Z (Worker_TP1 pid=70133) INFO 11-03 17:38:59 [multiproc_executor.py:624] WorkerProc shutting down. 2025-11-03T17:39:02.9898103Z (APIServer pid=69890) INFO: Shutting down 2025-11-03T17:39:03.0902495Z (APIServer pid=69890) INFO: Waiting for application shutdown. 2025-11-03T17:39:03.0904640Z (APIServer pid=69890) INFO: Application shutdown complete. 2025-11-03T17:39:03.5132991Z INFO 11-03 17:39:03 [model.py:653] Resolved architecture: LlamaForCausalLM 2025-11-03T17:39:03.5133612Z WARNING 11-03 17:39:03 [model.py:1942] Casting torch.bfloat16 to torch.float16. 2025-11-03T17:39:03.5134052Z INFO 11-03 17:39:03 [model.py:1714] Using max model len 2048 2025-11-03T17:39:03.5138294Z Launching RemoteOpenAIServer with: vllm serve RedHatAI/Meta-Llama-3.1-8B-Instruct-FP8 --dtype float16 --max-model-len 2048 --max-num-seqs 8 --enforce-eager --tokenizer-mode auto --tensor-parallel-size 2 --distributed-executor-backend mp --load-format dummy --port 42403 --seed 0 2025-11-03T17:39:04.0328451Z /opt/conda/envs/py_3.12/lib/python3.12/site-packages/torch/cuda/__init__.py:65: FutureWarning: The pynvml package is deprecated. Please install nvidia-ml-py instead. If you did not install pynvml directly, please report this to the maintainers of the package that installed pynvml for you. 2025-11-03T17:39:04.0329980Z import pynvml # type: ignore[import] 2025-11-03T17:39:06.3111980Z INFO 11-03 17:39:06 [__init__.py:224] Automatically detected platform cuda. 2025-11-03T17:39:10.0620290Z (APIServer pid=74046) INFO 11-03 17:39:10 [api_server.py:1870] vLLM API server version 0.11.0rc2.dev389+ge51928192.d20251103 2025-11-03T17:39:10.0642498Z (APIServer pid=74046) INFO 11-03 17:39:10 [utils.py:239] non-default args: {'model_tag': 'RedHatAI/Meta-Llama-3.1-8B-Instruct-FP8', 'port': 42403, 'model': 'RedHatAI/Meta-Llama-3.1-8B-Instruct-FP8', 'dtype': 'float16', 'seed': 0, 'max_model_len': 2048, 'enforce_eager': True, 'load_format': 'dummy', 'distributed_executor_backend': 'mp', 'tensor_parallel_size': 2, 'max_num_seqs': 8} 2025-11-03T17:39:10.2229510Z (APIServer pid=74046) INFO 11-03 17:39:10 [model.py:653] Resolved architecture: LlamaForCausalLM 2025-11-03T17:39:10.2230853Z (APIServer pid=74046) `torch_dtype` is deprecated! Use `dtype` instead! 2025-11-03T17:39:10.2232150Z (APIServer pid=74046) WARNING 11-03 17:39:10 [model.py:1942] Casting torch.bfloat16 to torch.float16. 2025-11-03T17:39:10.2233410Z (APIServer pid=74046) INFO 11-03 17:39:10 [model.py:1714] Using max model len 2048 2025-11-03T17:39:10.5615732Z (APIServer pid=74046) INFO 11-03 17:39:10 [scheduler.py:225] Chunked prefill is enabled with max_num_batched_tokens=2048. 2025-11-03T17:39:10.5987845Z (APIServer pid=74046) INFO 11-03 17:39:10 [vllm.py:354] Cudagraph is disabled under eager mode 2025-11-03T17:39:11.5987692Z /opt/conda/envs/py_3.12/lib/python3.12/site-packages/torch/cuda/__init__.py:65: FutureWarning: The pynvml package is deprecated. Please install nvidia-ml-py instead. If you did not install pynvml directly, please report this to the maintainers of the package that installed pynvml for you. 2025-11-03T17:39:11.5988865Z import pynvml # type: ignore[import] 2025-11-03T17:39:13.9495945Z INFO 11-03 17:39:13 [__init__.py:224] Automatically detected platform cuda. 2025-11-03T17:39:17.8191751Z (EngineCore_DP0 pid=74169) INFO 11-03 17:39:17 [core.py:727] Waiting for init message from front-end. 2025-11-03T17:39:17.8456872Z (EngineCore_DP0 pid=74169) INFO 11-03 17:39:17 [core.py:94] Initializing a V1 LLM engine (v0.11.0rc2.dev389+ge51928192.d20251103) with config: model='RedHatAI/Meta-Llama-3.1-8B-Instruct-FP8', speculative_config=None, tokenizer='RedHatAI/Meta-Llama-3.1-8B-Instruct-FP8', skip_tokenizer_init=False, tokenizer_mode=auto, revision=None, tokenizer_revision=None, trust_remote_code=False, dtype=torch.float16, max_seq_len=2048, download_dir=None, load_format=dummy, tensor_parallel_size=2, pipeline_parallel_size=1, data_parallel_size=1, disable_custom_all_reduce=False, quantization=compressed-tensors, enforce_eager=True, kv_cache_dtype=auto, device_config=cuda, structured_outputs_config=StructuredOutputsConfig(backend='auto', disable_fallback=False, disable_any_whitespace=False, disable_additional_properties=False, reasoning_parser=''), observability_config=ObservabilityConfig(show_hidden_metrics_for_version=None, otlp_traces_endpoint=None, collect_detailed_traces=None), seed=0, served_model_name=RedHatAI/Meta-Llama-3.1-8B-Instruct-FP8, enable_prefix_caching=True, chunked_prefill_enabled=True, pooler_config=None, compilation_config={'level': 0, 'debug_dump_path': None, 'cache_dir': '', 'backend': '', 'custom_ops': [], 'splitting_ops': None, 'use_inductor': True, 'compile_sizes': [], 'inductor_compile_config': {'enable_auto_functionalized_v2': False}, 'inductor_passes': {}, 'cudagraph_mode': , 'use_cudagraph': False, 'cudagraph_num_of_warmups': 0, 'cudagraph_capture_sizes': [], 'cudagraph_copy_inputs': False, 'full_cuda_graph': False, 'use_inductor_graph_partition': False, 'pass_config': {}, 'max_capture_size': 0, 'local_cache_dir': None} 2025-11-03T17:39:17.8462904Z (EngineCore_DP0 pid=74169) WARNING 11-03 17:39:17 [multiproc_executor.py:744] Reducing Torch parallelism from 24 threads to 1 to avoid unnecessary CPU contention. Set OMP_NUM_THREADS in the external environment to tune this value as needed. 2025-11-03T17:39:17.8470472Z (EngineCore_DP0 pid=74169) INFO 11-03 17:39:17 [shm_broadcast.py:302] vLLM message queue communication handle: Handle(local_reader_ranks=[0, 1], buffer_handle=(2, 16777216, 10, 'psm_dbc58ee5'), local_subscribe_addr='ipc:///tmp/1c668a66-1c0e-4359-bf70-0e2d598b79c7', remote_subscribe_addr=None, remote_addr_ipv6=False) 2025-11-03T17:39:18.3390623Z /opt/conda/envs/py_3.12/lib/python3.12/site-packages/torch/cuda/__init__.py:65: FutureWarning: The pynvml package is deprecated. Please install nvidia-ml-py instead. If you did not install pynvml directly, please report this to the maintainers of the package that installed pynvml for you. 2025-11-03T17:39:18.3391855Z import pynvml # type: ignore[import] 2025-11-03T17:39:18.3393011Z /opt/conda/envs/py_3.12/lib/python3.12/site-packages/torch/cuda/__init__.py:65: FutureWarning: The pynvml package is deprecated. Please install nvidia-ml-py instead. If you did not install pynvml directly, please report this to the maintainers of the package that installed pynvml for you. 2025-11-03T17:39:18.3394119Z import pynvml # type: ignore[import] 2025-11-03T17:39:20.7019557Z INFO 11-03 17:39:20 [__init__.py:224] Automatically detected platform cuda. 2025-11-03T17:39:20.7244153Z INFO 11-03 17:39:20 [__init__.py:224] Automatically detected platform cuda. 2025-11-03T17:39:25.6030981Z INFO 11-03 17:39:25 [shm_broadcast.py:302] vLLM message queue communication handle: Handle(local_reader_ranks=[0], buffer_handle=(1, 10485760, 10, 'psm_8b090b2d'), local_subscribe_addr='ipc:///tmp/efd2b367-903a-4130-8f37-f7ff3762d52e', remote_subscribe_addr=None, remote_addr_ipv6=False) 2025-11-03T17:39:25.6032832Z INFO 11-03 17:39:25 [shm_broadcast.py:302] vLLM message queue communication handle: Handle(local_reader_ranks=[0], buffer_handle=(1, 10485760, 10, 'psm_f400650f'), local_subscribe_addr='ipc:///tmp/70e82ffb-5588-4272-bda3-ac4d48ca4dab', remote_subscribe_addr=None, remote_addr_ipv6=False) 2025-11-03T17:39:26.3754405Z [Gloo] Rank 1 is connected to 1 peer ranks. Expected number of connected peer ranks is : 1 2025-11-03T17:39:26.3755128Z [Gloo] Rank 0 is connected to 1 peer ranks. Expected number of connected peer ranks is : 1 2025-11-03T17:39:26.3801436Z [Gloo] Rank 1 is connected to 1 peer ranks. Expected number of connected peer ranks is : 1 2025-11-03T17:39:26.3802837Z [Gloo] Rank 0 is connected to 1 peer ranks. Expected number of connected peer ranks is : 1 2025-11-03T17:39:26.3842726Z INFO 11-03 17:39:26 [__init__.py:1146] Found nccl from library libnccl.so.2 2025-11-03T17:39:26.3843490Z INFO 11-03 17:39:26 [__init__.py:1146] Found nccl from library libnccl.so.2 2025-11-03T17:39:26.3844665Z INFO 11-03 17:39:26 [pynccl.py:109] vLLM is using nccl==2.27.5 2025-11-03T17:39:26.3845850Z INFO 11-03 17:39:26 [pynccl.py:109] vLLM is using nccl==2.27.5 2025-11-03T17:39:26.5601002Z WARNING 11-03 17:39:26 [symm_mem.py:60] SymmMemCommunicator: Device capability 8.9 not supported, communicator is not available. 2025-11-03T17:39:26.5601850Z WARNING 11-03 17:39:26 [symm_mem.py:60] SymmMemCommunicator: Device capability 8.9 not supported, communicator is not available. 2025-11-03T17:39:26.5627535Z INFO 11-03 17:39:26 [custom_all_reduce.py:37] Skipping P2P check and trusting the driver's P2P report. 2025-11-03T17:39:26.5628183Z INFO 11-03 17:39:26 [custom_all_reduce.py:37] Skipping P2P check and trusting the driver's P2P report. 2025-11-03T17:39:26.5629504Z WARNING 11-03 17:39:26 [custom_all_reduce.py:162] Custom allreduce is disabled because your platform lacks GPU P2P capability or P2P test failed. To silence this warning, specify disable_custom_all_reduce=True explicitly. 2025-11-03T17:39:26.5630923Z WARNING 11-03 17:39:26 [custom_all_reduce.py:162] Custom allreduce is disabled because your platform lacks GPU P2P capability or P2P test failed. To silence this warning, specify disable_custom_all_reduce=True explicitly. 2025-11-03T17:39:26.6625997Z INFO 11-03 17:39:26 [shm_broadcast.py:302] vLLM message queue communication handle: Handle(local_reader_ranks=[1], buffer_handle=(1, 4194304, 6, 'psm_71ae8694'), local_subscribe_addr='ipc:///tmp/6a20fbf2-2032-4afe-924f-96cd34c7a949', remote_subscribe_addr=None, remote_addr_ipv6=False) 2025-11-03T17:39:26.6654942Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:39:26.6657903Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:39:26.6666468Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:39:26.6674428Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:39:26.7220891Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:39:26.7233760Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:39:26.7243338Z [Gloo] Rank 0 is connected to 1 peer ranks. Expected number of connected peer ranks is : 1 2025-11-03T17:39:26.7243933Z [Gloo] Rank 1 is connected to 1 peer ranks. Expected number of connected peer ranks is : 1 2025-11-03T17:39:26.7245140Z INFO 11-03 17:39:26 [__init__.py:1146] Found nccl from library libnccl.so.2 2025-11-03T17:39:26.7245585Z INFO 11-03 17:39:26 [pynccl.py:109] vLLM is using nccl==2.27.5 2025-11-03T17:39:26.7246884Z INFO 11-03 17:39:26 [__init__.py:1146] Found nccl from library libnccl.so.2 2025-11-03T17:39:26.7247388Z INFO 11-03 17:39:26 [pynccl.py:109] vLLM is using nccl==2.27.5 2025-11-03T17:39:26.7624363Z INFO 11-03 17:39:26 [parallel_state.py:1231] rank 1 in world size 2 is assigned as DP rank 0, PP rank 0, TP rank 1, EP rank 1 2025-11-03T17:39:26.7625842Z INFO 11-03 17:39:26 [parallel_state.py:1231] rank 0 in world size 2 is assigned as DP rank 0, PP rank 0, TP rank 0, EP rank 0 2025-11-03T17:39:26.9898143Z INFO 11-03 17:39:26 [topk_topp_sampler.py:59] Using FlashInfer for top-p & top-k sampling. 2025-11-03T17:39:26.9967099Z INFO 11-03 17:39:26 [topk_topp_sampler.py:59] Using FlashInfer for top-p & top-k sampling. 2025-11-03T17:39:27.0271099Z (Worker_TP0 pid=74288) INFO 11-03 17:39:27 [gpu_model_runner.py:2840] Starting to load model RedHatAI/Meta-Llama-3.1-8B-Instruct-FP8... 2025-11-03T17:39:27.0324769Z (Worker_TP1 pid=74289) INFO 11-03 17:39:27 [gpu_model_runner.py:2840] Starting to load model RedHatAI/Meta-Llama-3.1-8B-Instruct-FP8... 2025-11-03T17:39:27.2405113Z (Worker_TP0 pid=74288) INFO 11-03 17:39:27 [gpu_model_runner.py:2870] Loading model from scratch... 2025-11-03T17:39:27.2475528Z (Worker_TP1 pid=74289) INFO 11-03 17:39:27 [gpu_model_runner.py:2870] Loading model from scratch... 2025-11-03T17:39:27.3324818Z (Worker_TP0 pid=74288) INFO 11-03 17:39:27 [cuda.py:404] Using Flash Attention backend on V1 engine. 2025-11-03T17:39:27.3383300Z (Worker_TP1 pid=74289) INFO 11-03 17:39:27 [cuda.py:404] Using Flash Attention backend on V1 engine. 2025-11-03T17:39:28.1491727Z (Worker_TP1 pid=74289) INFO 11-03 17:39:28 [gpu_model_runner.py:2902] Model loading took 4.2624 GiB and 0.487773 seconds 2025-11-03T17:39:28.1561080Z (Worker_TP0 pid=74288) INFO 11-03 17:39:28 [gpu_model_runner.py:2902] Model loading took 4.2624 GiB and 0.492787 seconds 2025-11-03T17:39:30.5694510Z (Worker_TP1 pid=74289) INFO 11-03 17:39:30 [gpu_worker.py:314] Available KV cache memory: 15.32 GiB 2025-11-03T17:39:30.5893177Z (Worker_TP0 pid=74288) INFO 11-03 17:39:30 [gpu_worker.py:314] Available KV cache memory: 15.32 GiB 2025-11-03T17:39:30.7999616Z (EngineCore_DP0 pid=74169) INFO 11-03 17:39:30 [kv_cache_utils.py:1199] GPU KV cache size: 250,976 tokens 2025-11-03T17:39:30.8000568Z (EngineCore_DP0 pid=74169) INFO 11-03 17:39:30 [kv_cache_utils.py:1204] Maximum concurrency for 2,048 tokens per request: 122.55x 2025-11-03T17:39:30.8001764Z (EngineCore_DP0 pid=74169) INFO 11-03 17:39:30 [kv_cache_utils.py:1199] GPU KV cache size: 250,976 tokens 2025-11-03T17:39:30.8002583Z (EngineCore_DP0 pid=74169) INFO 11-03 17:39:30 [kv_cache_utils.py:1204] Maximum concurrency for 2,048 tokens per request: 122.55x 2025-11-03T17:39:30.9310351Z (EngineCore_DP0 pid=74169) INFO 11-03 17:39:30 [core.py:240] init engine (profile, create kv cache, warmup model) took 2.70 seconds 2025-11-03T17:39:32.0102605Z (EngineCore_DP0 pid=74169) INFO 11-03 17:39:32 [vllm.py:354] Cudagraph is disabled under eager mode 2025-11-03T17:39:32.1949219Z (APIServer pid=74046) INFO 11-03 17:39:32 [loggers.py:172] Engine 000: vllm cache_config_info with initialization after num_gpu_blocks is: 15686 2025-11-03T17:39:32.2042796Z (EngineCore_DP0 pid=74169) INFO 11-03 17:39:32 [gc_utils.py:40] GC Debug Config. enabled:False,top_objects:-1 2025-11-03T17:39:32.2262344Z (APIServer pid=74046) INFO 11-03 17:39:32 [api_server.py:1628] Supported tasks: ['generate'] 2025-11-03T17:39:32.2582456Z (APIServer pid=74046) WARNING 11-03 17:39:32 [model.py:1581] Default sampling parameters have been overridden by the model's Hugging Face generation config recommended from the model creator. If this is not intended, please relaunch vLLM instance with `--generation-config vllm`. 2025-11-03T17:39:32.2584016Z (APIServer pid=74046) INFO 11-03 17:39:32 [serving_responses.py:157] Using default chat sampling params from model: {'temperature': 0.6, 'top_p': 0.9} 2025-11-03T17:39:32.3158805Z (APIServer pid=74046) INFO 11-03 17:39:32 [serving_chat.py:159] Using default chat sampling params from model: {'temperature': 0.6, 'top_p': 0.9} 2025-11-03T17:39:32.3552730Z (APIServer pid=74046) INFO 11-03 17:39:32 [serving_completion.py:67] Using default completion sampling params from model: {'temperature': 0.6, 'top_p': 0.9} 2025-11-03T17:39:32.3554203Z (APIServer pid=74046) INFO 11-03 17:39:32 [api_server.py:1939] Starting vLLM API server 0 on http://0.0.0.0:42403 2025-11-03T17:39:32.3555426Z (APIServer pid=74046) INFO 11-03 17:39:32 [launcher.py:38] Available routes are: 2025-11-03T17:39:32.3556733Z (APIServer pid=74046) INFO 11-03 17:39:32 [launcher.py:46] Route: /openapi.json, Methods: HEAD, GET 2025-11-03T17:39:32.3557966Z (APIServer pid=74046) INFO 11-03 17:39:32 [launcher.py:46] Route: /docs, Methods: HEAD, GET 2025-11-03T17:39:32.3559160Z (APIServer pid=74046) INFO 11-03 17:39:32 [launcher.py:46] Route: /docs/oauth2-redirect, Methods: HEAD, GET 2025-11-03T17:39:32.3559989Z (APIServer pid=74046) INFO 11-03 17:39:32 [launcher.py:46] Route: /redoc, Methods: HEAD, GET 2025-11-03T17:39:32.3560763Z (APIServer pid=74046) INFO 11-03 17:39:32 [launcher.py:46] Route: /health, Methods: GET 2025-11-03T17:39:32.3561375Z (APIServer pid=74046) INFO 11-03 17:39:32 [launcher.py:46] Route: /load, Methods: GET 2025-11-03T17:39:32.3562035Z (APIServer pid=74046) INFO 11-03 17:39:32 [launcher.py:46] Route: /ping, Methods: POST 2025-11-03T17:39:32.3562555Z (APIServer pid=74046) INFO 11-03 17:39:32 [launcher.py:46] Route: /ping, Methods: GET 2025-11-03T17:39:32.3563065Z (APIServer pid=74046) INFO 11-03 17:39:32 [launcher.py:46] Route: /tokenize, Methods: POST 2025-11-03T17:39:32.3563585Z (APIServer pid=74046) INFO 11-03 17:39:32 [launcher.py:46] Route: /detokenize, Methods: POST 2025-11-03T17:39:32.3564276Z (APIServer pid=74046) INFO 11-03 17:39:32 [launcher.py:46] Route: /v1/models, Methods: GET 2025-11-03T17:39:32.3564813Z (APIServer pid=74046) INFO 11-03 17:39:32 [launcher.py:46] Route: /version, Methods: GET 2025-11-03T17:39:32.3565343Z (APIServer pid=74046) INFO 11-03 17:39:32 [launcher.py:46] Route: /v1/responses, Methods: POST 2025-11-03T17:39:32.3565988Z (APIServer pid=74046) INFO 11-03 17:39:32 [launcher.py:46] Route: /v1/responses/{response_id}, Methods: GET 2025-11-03T17:39:32.3566603Z (APIServer pid=74046) INFO 11-03 17:39:32 [launcher.py:46] Route: /v1/responses/{response_id}/cancel, Methods: POST 2025-11-03T17:39:32.3567192Z (APIServer pid=74046) INFO 11-03 17:39:32 [launcher.py:46] Route: /v1/chat/completions, Methods: POST 2025-11-03T17:39:32.3567744Z (APIServer pid=74046) INFO 11-03 17:39:32 [launcher.py:46] Route: /v1/completions, Methods: POST 2025-11-03T17:39:32.3568288Z (APIServer pid=74046) INFO 11-03 17:39:32 [launcher.py:46] Route: /v1/embeddings, Methods: POST 2025-11-03T17:39:32.3568952Z (APIServer pid=74046) INFO 11-03 17:39:32 [launcher.py:46] Route: /pooling, Methods: POST 2025-11-03T17:39:32.3569477Z (APIServer pid=74046) INFO 11-03 17:39:32 [launcher.py:46] Route: /classify, Methods: POST 2025-11-03T17:39:32.3569989Z (APIServer pid=74046) INFO 11-03 17:39:32 [launcher.py:46] Route: /score, Methods: POST 2025-11-03T17:39:32.3570495Z (APIServer pid=74046) INFO 11-03 17:39:32 [launcher.py:46] Route: /v1/score, Methods: POST 2025-11-03T17:39:32.3571043Z (APIServer pid=74046) INFO 11-03 17:39:32 [launcher.py:46] Route: /v1/audio/transcriptions, Methods: POST 2025-11-03T17:39:32.3571623Z (APIServer pid=74046) INFO 11-03 17:39:32 [launcher.py:46] Route: /v1/audio/translations, Methods: POST 2025-11-03T17:39:32.3572162Z (APIServer pid=74046) INFO 11-03 17:39:32 [launcher.py:46] Route: /rerank, Methods: POST 2025-11-03T17:39:32.3572675Z (APIServer pid=74046) INFO 11-03 17:39:32 [launcher.py:46] Route: /v1/rerank, Methods: POST 2025-11-03T17:39:32.3573194Z (APIServer pid=74046) INFO 11-03 17:39:32 [launcher.py:46] Route: /v2/rerank, Methods: POST 2025-11-03T17:39:32.3573727Z (APIServer pid=74046) INFO 11-03 17:39:32 [launcher.py:46] Route: /scale_elastic_ep, Methods: POST 2025-11-03T17:39:32.3574290Z (APIServer pid=74046) INFO 11-03 17:39:32 [launcher.py:46] Route: /is_scaling_elastic_ep, Methods: POST 2025-11-03T17:39:32.3574840Z (APIServer pid=74046) INFO 11-03 17:39:32 [launcher.py:46] Route: /invocations, Methods: POST 2025-11-03T17:39:32.3575358Z (APIServer pid=74046) INFO 11-03 17:39:32 [launcher.py:46] Route: /metrics, Methods: GET 2025-11-03T17:39:32.3715211Z (APIServer pid=74046) INFO: Started server process [74046] 2025-11-03T17:39:32.3715992Z (APIServer pid=74046) INFO: Waiting for application startup. 2025-11-03T17:39:32.5898658Z (APIServer pid=74046) INFO: Application startup complete. 2025-11-03T17:39:33.0909297Z (APIServer pid=74046) INFO: 127.0.0.1:39116 - "GET /health HTTP/1.1" 200 OK 2025-11-03T17:39:33.1054405Z (APIServer pid=74046) INFO: 127.0.0.1:39126 - "GET /v1/models HTTP/1.1" 200 OK 2025-11-03T17:39:33.3393169Z (APIServer pid=74046) INFO: 127.0.0.1:39126 - "POST /v1/completions HTTP/1.1" 200 OK 2025-11-03T17:39:33.4573972Z (APIServer pid=74046) INFO: 127.0.0.1:39126 - "POST /v1/completions HTTP/1.1" 200 OK 2025-11-03T17:39:33.6110437Z (APIServer pid=74046) INFO: 127.0.0.1:39126 - "POST /v1/completions HTTP/1.1" 200 OK 2025-11-03T17:39:33.7663493Z (APIServer pid=74046) INFO: 127.0.0.1:39126 - "POST /v1/completions HTTP/1.1" 200 OK 2025-11-03T17:39:33.9086847Z (APIServer pid=74046) INFO: 127.0.0.1:39126 - "POST /v1/completions HTTP/1.1" 200 OK 2025-11-03T17:39:33.9165974Z (APIServer pid=74046) INFO: 127.0.0.1:39126 - "POST /v1/completions HTTP/1.1" 200 OK 2025-11-03T17:39:34.0653763Z (APIServer pid=74046) INFO 11-03 17:39:34 [launcher.py:110] Shutting down FastAPI HTTP server. 2025-11-03T17:39:34.0658767Z (Worker_TP0 pid=74288) INFO 11-03 17:39:34 [multiproc_executor.py:583] Parent process exited, terminating worker 2025-11-03T17:39:34.0660721Z (Worker_TP0 pid=74288) INFO 11-03 17:39:34 [multiproc_executor.py:624] WorkerProc shutting down. 2025-11-03T17:39:34.0663781Z (Worker_TP1 pid=74289) INFO 11-03 17:39:34 [multiproc_executor.py:583] Parent process exited, terminating worker 2025-11-03T17:39:34.0665461Z (Worker_TP1 pid=74289) INFO 11-03 17:39:34 [multiproc_executor.py:624] WorkerProc shutting down. 2025-11-03T17:39:35.9827123Z (APIServer pid=74046) INFO: Shutting down 2025-11-03T17:39:36.0831533Z (APIServer pid=74046) INFO: Waiting for application shutdown. 2025-11-03T17:39:36.0833542Z (APIServer pid=74046) INFO: Application shutdown complete. 2025-11-03T17:39:36.6910447Z PASSED 2025-11-03T17:39:36.7023421Z distributed/test_sequence_parallel.py::test_tp_sp_generation[RedHatAI/Meta-Llama-3.1-8B-Instruct-FP8-parallel_setup19-ray-auto-test_options19] Fork a new process to run a test 74492 2025-11-03T17:39:36.7040633Z Fork a new process to run a test 0 2025-11-03T17:39:37.3671990Z INFO 11-03 17:39:37 [model.py:653] Resolved architecture: LlamaForCausalLM 2025-11-03T17:39:37.3672401Z `torch_dtype` is deprecated! Use `dtype` instead! 2025-11-03T17:39:37.3672837Z WARNING 11-03 17:39:37 [model.py:1942] Casting torch.bfloat16 to torch.float16. 2025-11-03T17:39:37.3673269Z INFO 11-03 17:39:37 [model.py:1714] Using max model len 2048 2025-11-03T17:39:37.4762489Z Launching RemoteOpenAIServer with: vllm serve RedHatAI/Meta-Llama-3.1-8B-Instruct-FP8 --dtype float16 --max-model-len 2048 --max-num-seqs 8 --enforce-eager --tokenizer-mode auto --tensor-parallel-size 2 --pipeline-parallel-size 1 --distributed-executor-backend ray --compilation_config {"level": 3, "custom_ops": ["+rms_norm"], "compile_sizes": [4, 8], "pass_config": {"enable_sequence_parallelism": true, "enable_fusion": true, "enable_noop": true}} --load-format dummy --port 43523 --seed 0 2025-11-03T17:39:37.9925676Z /opt/conda/envs/py_3.12/lib/python3.12/site-packages/torch/cuda/__init__.py:65: FutureWarning: The pynvml package is deprecated. Please install nvidia-ml-py instead. If you did not install pynvml directly, please report this to the maintainers of the package that installed pynvml for you. 2025-11-03T17:39:37.9927119Z import pynvml # type: ignore[import] 2025-11-03T17:39:40.2511290Z INFO 11-03 17:39:40 [__init__.py:224] Automatically detected platform cuda. 2025-11-03T17:39:43.9684867Z (APIServer pid=74541) INFO 11-03 17:39:43 [api_server.py:1870] vLLM API server version 0.11.0rc2.dev389+ge51928192.d20251103 2025-11-03T17:39:43.9710192Z (APIServer pid=74541) INFO 11-03 17:39:43 [utils.py:239] non-default args: {'model_tag': 'RedHatAI/Meta-Llama-3.1-8B-Instruct-FP8', 'port': 43523, 'model': 'RedHatAI/Meta-Llama-3.1-8B-Instruct-FP8', 'dtype': 'float16', 'seed': 0, 'max_model_len': 2048, 'enforce_eager': True, 'load_format': 'dummy', 'distributed_executor_backend': 'ray', 'tensor_parallel_size': 2, 'max_num_seqs': 8, 'compilation_config': {'level': 3, 'debug_dump_path': None, 'cache_dir': '', 'backend': '', 'custom_ops': ['+rms_norm'], 'splitting_ops': None, 'use_inductor': True, 'compile_sizes': [4, 8], 'inductor_compile_config': {'enable_auto_functionalized_v2': False}, 'inductor_passes': {}, 'cudagraph_mode': None, 'use_cudagraph': True, 'cudagraph_num_of_warmups': 0, 'cudagraph_capture_sizes': None, 'cudagraph_copy_inputs': False, 'full_cuda_graph': False, 'use_inductor_graph_partition': False, 'pass_config': {'enable_fusion': True, 'enable_noop': True, 'enable_sequence_parallelism': True}, 'max_capture_size': None, 'local_cache_dir': None}} 2025-11-03T17:39:44.1228349Z (APIServer pid=74541) INFO 11-03 17:39:44 [model.py:653] Resolved architecture: LlamaForCausalLM 2025-11-03T17:39:44.1229256Z (APIServer pid=74541) `torch_dtype` is deprecated! Use `dtype` instead! 2025-11-03T17:39:44.1229934Z (APIServer pid=74541) WARNING 11-03 17:39:44 [model.py:1942] Casting torch.bfloat16 to torch.float16. 2025-11-03T17:39:44.1230644Z (APIServer pid=74541) INFO 11-03 17:39:44 [model.py:1714] Using max model len 2048 2025-11-03T17:39:44.4601914Z (APIServer pid=74541) INFO 11-03 17:39:44 [scheduler.py:225] Chunked prefill is enabled with max_num_batched_tokens=2048. 2025-11-03T17:39:44.4974217Z (APIServer pid=74541) INFO 11-03 17:39:44 [vllm.py:354] Cudagraph is disabled under eager mode 2025-11-03T17:39:45.4826274Z /opt/conda/envs/py_3.12/lib/python3.12/site-packages/torch/cuda/__init__.py:65: FutureWarning: The pynvml package is deprecated. Please install nvidia-ml-py instead. If you did not install pynvml directly, please report this to the maintainers of the package that installed pynvml for you. 2025-11-03T17:39:45.4827526Z import pynvml # type: ignore[import] 2025-11-03T17:39:47.8320766Z INFO 11-03 17:39:47 [__init__.py:224] Automatically detected platform cuda. 2025-11-03T17:39:51.8160571Z (EngineCore_DP0 pid=74664) INFO 11-03 17:39:51 [core.py:727] Waiting for init message from front-end. 2025-11-03T17:39:51.8419504Z (EngineCore_DP0 pid=74664) INFO 11-03 17:39:51 [core.py:94] Initializing a V1 LLM engine (v0.11.0rc2.dev389+ge51928192.d20251103) with config: model='RedHatAI/Meta-Llama-3.1-8B-Instruct-FP8', speculative_config=None, tokenizer='RedHatAI/Meta-Llama-3.1-8B-Instruct-FP8', skip_tokenizer_init=False, tokenizer_mode=auto, revision=None, tokenizer_revision=None, trust_remote_code=False, dtype=torch.float16, max_seq_len=2048, download_dir=None, load_format=dummy, tensor_parallel_size=2, pipeline_parallel_size=1, data_parallel_size=1, disable_custom_all_reduce=False, quantization=compressed-tensors, enforce_eager=True, kv_cache_dtype=auto, device_config=cuda, structured_outputs_config=StructuredOutputsConfig(backend='auto', disable_fallback=False, disable_any_whitespace=False, disable_additional_properties=False, reasoning_parser=''), observability_config=ObservabilityConfig(show_hidden_metrics_for_version=None, otlp_traces_endpoint=None, collect_detailed_traces=None), seed=0, served_model_name=RedHatAI/Meta-Llama-3.1-8B-Instruct-FP8, enable_prefix_caching=True, chunked_prefill_enabled=True, pooler_config=None, compilation_config={'level': 3, 'debug_dump_path': None, 'cache_dir': '', 'backend': '', 'custom_ops': ['+rms_norm', '+rms_norm'], 'splitting_ops': ['vllm.unified_attention', 'vllm.unified_attention_with_output', 'vllm.unified_mla_attention', 'vllm.unified_mla_attention_with_output', 'vllm.mamba_mixer2', 'vllm.mamba_mixer', 'vllm.short_conv', 'vllm.linear_attention', 'vllm.plamo2_mamba_mixer', 'vllm.gdn_attention', 'vllm.sparse_attn_indexer'], 'use_inductor': True, 'compile_sizes': [8, 4], 'inductor_compile_config': {'enable_auto_functionalized_v2': False}, 'inductor_passes': {}, 'cudagraph_mode': , 'use_cudagraph': False, 'cudagraph_num_of_warmups': 0, 'cudagraph_capture_sizes': [], 'cudagraph_copy_inputs': False, 'full_cuda_graph': False, 'use_inductor_graph_partition': False, 'pass_config': {'enable_fusion': True, 'enable_noop': True, 'enable_sequence_parallelism': True}, 'max_capture_size': 0, 'local_cache_dir': None} 2025-11-03T17:39:53.7511429Z (EngineCore_DP0 pid=74664) 2025-11-03 17:39:53,750 INFO worker.py:1918 -- Started a local Ray instance. View the dashboard at http://127.0.0.1:8265  2025-11-03T17:39:54.5881593Z (EngineCore_DP0 pid=74664) INFO 11-03 17:39:54 [ray_utils.py:373] No current placement group found. Creating a new placement group. 2025-11-03T17:39:54.7626137Z (EngineCore_DP0 pid=74664) INFO 11-03 17:39:54 [ray_distributed_executor.py:179] use_ray_spmd_worker: True 2025-11-03T17:39:55.5396680Z (EngineCore_DP0 pid=74664) (pid=75559) /opt/conda/envs/py_3.12/lib/python3.12/site-packages/torch/cuda/__init__.py:65: FutureWarning: The pynvml package is deprecated. Please install nvidia-ml-py instead. If you did not install pynvml directly, please report this to the maintainers of the package that installed pynvml for you. 2025-11-03T17:39:55.5397958Z (EngineCore_DP0 pid=74664) (pid=75559) import pynvml # type: ignore[import] 2025-11-03T17:39:57.8903168Z (EngineCore_DP0 pid=74664) (pid=75559) INFO 11-03 17:39:57 [__init__.py:224] Automatically detected platform cuda. 2025-11-03T17:40:00.4985480Z (EngineCore_DP0 pid=74664) INFO 11-03 17:40:00 [ray_env.py:67] RAY_NON_CARRY_OVER_ENV_VARS from config: set() 2025-11-03T17:40:00.4987044Z (EngineCore_DP0 pid=74664) INFO 11-03 17:40:00 [ray_env.py:70] Copying the following environment variables to workers: ['HUGGING_FACE_HUB_TOKEN', 'VLLM_WORKER_MULTIPROC_METHOD', 'VLLM_USE_V1', 'LD_LIBRARY_PATH', 'VLLM_USE_RAY_COMPILED_DAG', 'VLLM_USE_RAY_SPMD_WORKER', 'HF_TOKEN', 'MAX_JOBS'] 2025-11-03T17:40:00.4988809Z (EngineCore_DP0 pid=74664) INFO 11-03 17:40:00 [ray_env.py:75] If certain env vars should NOT be copied, add them to /var/lib/jenkins/.config/vllm/ray_non_carry_over_env_vars.json file 2025-11-03T17:40:02.2767889Z (EngineCore_DP0 pid=74664) (RayWorkerWrapper pid=75557) WARNING 11-03 17:40:02 [worker_base.py:307] Missing `shared_worker_lock` argument from executor. This argument is needed for mm_processor_cache_type='shm'. 2025-11-03T17:40:02.5880983Z (EngineCore_DP0 pid=74664) (RayWorkerWrapper pid=75559) [Gloo] Rank 0 is connected to 1 peer ranks. Expected number of connected peer ranks is : 1 2025-11-03T17:40:02.5882185Z (EngineCore_DP0 pid=74664) (RayWorkerWrapper pid=75559) [Gloo] Rank 0 is connected to 1 peer ranks. Expected number of connected peer ranks is : 1 2025-11-03T17:40:02.5883242Z (EngineCore_DP0 pid=74664) (RayWorkerWrapper pid=75559) INFO 11-03 17:40:02 [__init__.py:1146] Found nccl from library libnccl.so.2 2025-11-03T17:40:02.5884163Z (EngineCore_DP0 pid=74664) (RayWorkerWrapper pid=75559) INFO 11-03 17:40:02 [pynccl.py:109] vLLM is using nccl==2.27.5 2025-11-03T17:40:02.7954225Z (EngineCore_DP0 pid=74664) (RayWorkerWrapper pid=75559) WARNING 11-03 17:40:02 [symm_mem.py:60] SymmMemCommunicator: Device capability 8.9 not supported, communicator is not available. 2025-11-03T17:40:02.7956172Z (EngineCore_DP0 pid=74664) (RayWorkerWrapper pid=75559) INFO 11-03 17:40:02 [custom_all_reduce.py:37] Skipping P2P check and trusting the driver's P2P report. 2025-11-03T17:40:02.7957820Z (EngineCore_DP0 pid=74664) (RayWorkerWrapper pid=75559) WARNING 11-03 17:40:02 [custom_all_reduce.py:162] Custom allreduce is disabled because your platform lacks GPU P2P capability or P2P test failed. To silence this warning, specify disable_custom_all_reduce=True explicitly. 2025-11-03T17:40:02.7960225Z (EngineCore_DP0 pid=74664) (RayWorkerWrapper pid=75559) INFO 11-03 17:40:02 [shm_broadcast.py:302] vLLM message queue communication handle: Handle(local_reader_ranks=[1], buffer_handle=(1, 4194304, 6, 'psm_28b31740'), local_subscribe_addr='ipc:///tmp/353a585f-838e-4360-a8ed-9f1e37d98662', remote_subscribe_addr=None, remote_addr_ipv6=False) 2025-11-03T17:40:02.9010688Z (EngineCore_DP0 pid=74664) (RayWorkerWrapper pid=75559) INFO 11-03 17:40:02 [parallel_state.py:1231] rank 0 in world size 2 is assigned as DP rank 0, PP rank 0, TP rank 0, EP rank 0 2025-11-03T17:40:02.9013352Z (EngineCore_DP0 pid=74664) (pid=75557) INFO 11-03 17:39:57 [__init__.py:224] Automatically detected platform cuda. 2025-11-03T17:40:03.1079361Z (EngineCore_DP0 pid=74664) (RayWorkerWrapper pid=75559) INFO 11-03 17:40:03 [topk_topp_sampler.py:59] Using FlashInfer for top-p & top-k sampling. 2025-11-03T17:40:03.2117037Z (EngineCore_DP0 pid=74664) (RayWorkerWrapper pid=75559) INFO 11-03 17:40:03 [gpu_model_runner.py:2840] Starting to load model RedHatAI/Meta-Llama-3.1-8B-Instruct-FP8... 2025-11-03T17:40:03.4174339Z (EngineCore_DP0 pid=74664) (RayWorkerWrapper pid=75559) INFO 11-03 17:40:03 [gpu_model_runner.py:2870] Loading model from scratch... 2025-11-03T17:40:03.5210759Z (EngineCore_DP0 pid=74664) (RayWorkerWrapper pid=75559) INFO 11-03 17:40:03 [cuda.py:404] Using Flash Attention backend on V1 engine. 2025-11-03T17:40:04.4400403Z (EngineCore_DP0 pid=74664) (RayWorkerWrapper pid=75557) INFO 11-03 17:40:04 [gpu_model_runner.py:2902] Model loading took 4.2624 GiB and 0.544245 seconds 2025-11-03T17:40:12.0928004Z (EngineCore_DP0 pid=74664) (RayWorkerWrapper pid=75557) INFO 11-03 17:40:12 [backends.py:594] Using cache directory: /var/lib/jenkins/.cache/vllm/torch_compile_cache/9e2073b733/rank_1_0/backbone for vLLM's torch.compile 2025-11-03T17:40:12.0929571Z (EngineCore_DP0 pid=74664) (RayWorkerWrapper pid=75557) INFO 11-03 17:40:12 [backends.py:608] Dynamo bytecode transform time: 7.32 s 2025-11-03T17:40:12.0930994Z (EngineCore_DP0 pid=74664) (RayWorkerWrapper pid=75559) WARNING 11-03 17:40:02 [worker_base.py:307] Missing `shared_worker_lock` argument from executor. This argument is needed for mm_processor_cache_type='shm'. 2025-11-03T17:40:12.0933317Z (EngineCore_DP0 pid=74664) (RayWorkerWrapper pid=75557) [Gloo] Rank 1 is connected to 1 peer ranks. Expected number of connected peer ranks is : 1 [repeated 10x across cluster] (Ray deduplicates logs by default. Set RAY_DEDUP_LOGS=0 to disable log deduplication, or see https://docs.ray.io/en/master/ray-observability/user-guides/configure-logging.html#log-deduplication for more options.) 2025-11-03T17:40:12.0934895Z (EngineCore_DP0 pid=74664) (RayWorkerWrapper pid=75557) INFO 11-03 17:40:02 [__init__.py:1146] Found nccl from library libnccl.so.2 [repeated 3x across cluster] 2025-11-03T17:40:12.0935812Z (EngineCore_DP0 pid=74664) (RayWorkerWrapper pid=75557) INFO 11-03 17:40:02 [pynccl.py:109] vLLM is using nccl==2.27.5 [repeated 3x across cluster] 2025-11-03T17:40:12.0936802Z (EngineCore_DP0 pid=74664) (RayWorkerWrapper pid=75557) WARNING 11-03 17:40:02 [symm_mem.py:60] SymmMemCommunicator: Device capability 8.9 not supported, communicator is not available. 2025-11-03T17:40:12.0938111Z (EngineCore_DP0 pid=74664) (RayWorkerWrapper pid=75557) INFO 11-03 17:40:02 [custom_all_reduce.py:37] Skipping P2P check and trusting the driver's P2P report. 2025-11-03T17:40:12.0939400Z (EngineCore_DP0 pid=74664) (RayWorkerWrapper pid=75557) WARNING 11-03 17:40:02 [custom_all_reduce.py:162] Custom allreduce is disabled because your platform lacks GPU P2P capability or P2P test failed. To silence this warning, specify disable_custom_all_reduce=True explicitly. 2025-11-03T17:40:12.0940836Z (EngineCore_DP0 pid=74664) (RayWorkerWrapper pid=75557) INFO 11-03 17:40:02 [parallel_state.py:1231] rank 1 in world size 2 is assigned as DP rank 0, PP rank 0, TP rank 1, EP rank 1 2025-11-03T17:40:12.0941764Z (EngineCore_DP0 pid=74664) (RayWorkerWrapper pid=75557) INFO 11-03 17:40:03 [topk_topp_sampler.py:59] Using FlashInfer for top-p & top-k sampling. 2025-11-03T17:40:12.0942878Z (EngineCore_DP0 pid=74664) (RayWorkerWrapper pid=75557) INFO 11-03 17:40:03 [gpu_model_runner.py:2840] Starting to load model RedHatAI/Meta-Llama-3.1-8B-Instruct-FP8... 2025-11-03T17:40:12.0943775Z (EngineCore_DP0 pid=74664) (RayWorkerWrapper pid=75557) INFO 11-03 17:40:03 [gpu_model_runner.py:2870] Loading model from scratch... 2025-11-03T17:40:12.0944552Z (EngineCore_DP0 pid=74664) (RayWorkerWrapper pid=75557) INFO 11-03 17:40:03 [cuda.py:404] Using Flash Attention backend on V1 engine. 2025-11-03T17:40:12.0945462Z (EngineCore_DP0 pid=74664) (RayWorkerWrapper pid=75559) INFO 11-03 17:40:04 [gpu_model_runner.py:2902] Model loading took 4.2624 GiB and 0.540099 seconds 2025-11-03T17:40:14.8484336Z (EngineCore_DP0 pid=74664) (RayWorkerWrapper pid=75557) INFO 11-03 17:40:14 [backends.py:179] Directly load the compiled graph(s) for dynamic shape from the cache, took 2.190 s 2025-11-03T17:40:17.5008309Z (EngineCore_DP0 pid=74664) (RayWorkerWrapper pid=75559) INFO 11-03 17:40:17 [gpu_worker.py:314] Available KV cache memory: 15.34 GiB 2025-11-03T17:40:17.5010123Z (EngineCore_DP0 pid=74664) (RayWorkerWrapper pid=75559) INFO 11-03 17:40:12 [backends.py:594] Using cache directory: /var/lib/jenkins/.cache/vllm/torch_compile_cache/9e2073b733/rank_0_0/backbone for vLLM's torch.compile 2025-11-03T17:40:17.5011549Z (EngineCore_DP0 pid=74664) (RayWorkerWrapper pid=75559) INFO 11-03 17:40:12 [backends.py:608] Dynamo bytecode transform time: 7.42 s 2025-11-03T17:40:17.7508438Z (EngineCore_DP0 pid=74664) INFO 11-03 17:40:17 [kv_cache_utils.py:1199] GPU KV cache size: 251,360 tokens 2025-11-03T17:40:17.7509371Z (EngineCore_DP0 pid=74664) INFO 11-03 17:40:17 [kv_cache_utils.py:1204] Maximum concurrency for 2,048 tokens per request: 122.73x 2025-11-03T17:40:17.7510551Z (EngineCore_DP0 pid=74664) INFO 11-03 17:40:17 [kv_cache_utils.py:1199] GPU KV cache size: 251,360 tokens 2025-11-03T17:40:17.7511391Z (EngineCore_DP0 pid=74664) INFO 11-03 17:40:17 [kv_cache_utils.py:1204] Maximum concurrency for 2,048 tokens per request: 122.73x 2025-11-03T17:40:17.8116197Z (EngineCore_DP0 pid=74664) (RayWorkerWrapper pid=75559) INFO 11-03 17:40:17 [gpu_worker.py:351] Compile and warming up model for size 8 2025-11-03T17:40:35.7258648Z (EngineCore_DP0 pid=74664) (RayWorkerWrapper pid=75559) INFO 11-03 17:40:35 [backends.py:185] Directly load the compiled graph(s) for shape 8 from the cache, took 17.906 s 2025-11-03T17:40:35.7261249Z (EngineCore_DP0 pid=74664) (RayWorkerWrapper pid=75559) INFO 11-03 17:40:14 [backends.py:179] Directly load the compiled graph(s) for dynamic shape from the cache, took 2.203 s 2025-11-03T17:40:35.7262612Z (EngineCore_DP0 pid=74664) (RayWorkerWrapper pid=75557) INFO 11-03 17:40:17 [gpu_worker.py:314] Available KV cache memory: 15.34 GiB 2025-11-03T17:40:35.7263802Z (EngineCore_DP0 pid=74664) (RayWorkerWrapper pid=75557) INFO 11-03 17:40:17 [gpu_worker.py:351] Compile and warming up model for size 8 2025-11-03T17:40:35.9311498Z (EngineCore_DP0 pid=74664) (RayWorkerWrapper pid=75559) INFO 11-03 17:40:35 [gpu_worker.py:351] Compile and warming up model for size 4 2025-11-03T17:40:58.2056796Z (EngineCore_DP0 pid=74664) INFO 11-03 17:40:58 [core.py:240] init engine (profile, create kv cache, warmup model) took 53.75 seconds 2025-11-03T17:40:58.2265156Z (EngineCore_DP0 pid=74664) (RayWorkerWrapper pid=75559) INFO 11-03 17:40:58 [monitor.py:33] torch.compile takes 7.42 s in total 2025-11-03T17:40:58.2266629Z (EngineCore_DP0 pid=74664) (RayWorkerWrapper pid=75559) INFO 11-03 17:40:58 [backends.py:185] Directly load the compiled graph(s) for shape 4 from the cache, took 22.286 s [repeated 2x across cluster] 2025-11-03T17:40:58.2267906Z (EngineCore_DP0 pid=74664) (RayWorkerWrapper pid=75557) INFO 11-03 17:40:35 [gpu_worker.py:351] Compile and warming up model for size 4 2025-11-03T17:40:58.7300373Z (EngineCore_DP0 pid=74664) INFO 11-03 17:40:58 [vllm.py:354] Cudagraph is disabled under eager mode 2025-11-03T17:40:58.9155752Z (APIServer pid=74541) INFO 11-03 17:40:58 [loggers.py:172] Engine 000: vllm cache_config_info with initialization after num_gpu_blocks is: 15710 2025-11-03T17:40:58.9869186Z (EngineCore_DP0 pid=74664) INFO 11-03 17:40:58 [gc_utils.py:40] GC Debug Config. enabled:False,top_objects:-1 2025-11-03T17:40:59.0131616Z (APIServer pid=74541) INFO 11-03 17:40:59 [api_server.py:1628] Supported tasks: ['generate'] 2025-11-03T17:40:59.0585824Z (APIServer pid=74541) WARNING 11-03 17:40:59 [model.py:1581] Default sampling parameters have been overridden by the model's Hugging Face generation config recommended from the model creator. If this is not intended, please relaunch vLLM instance with `--generation-config vllm`. 2025-11-03T17:40:59.0587244Z (APIServer pid=74541) INFO 11-03 17:40:59 [serving_responses.py:157] Using default chat sampling params from model: {'temperature': 0.6, 'top_p': 0.9} 2025-11-03T17:40:59.0897959Z (APIServer pid=74541) INFO 11-03 17:40:59 [serving_chat.py:159] Using default chat sampling params from model: {'temperature': 0.6, 'top_p': 0.9} 2025-11-03T17:40:59.1214034Z (APIServer pid=74541) INFO 11-03 17:40:59 [serving_completion.py:67] Using default completion sampling params from model: {'temperature': 0.6, 'top_p': 0.9} 2025-11-03T17:40:59.1215687Z (APIServer pid=74541) INFO 11-03 17:40:59 [api_server.py:1939] Starting vLLM API server 0 on http://0.0.0.0:43523 2025-11-03T17:40:59.1216813Z (APIServer pid=74541) INFO 11-03 17:40:59 [launcher.py:38] Available routes are: 2025-11-03T17:40:59.1217942Z (APIServer pid=74541) INFO 11-03 17:40:59 [launcher.py:46] Route: /openapi.json, Methods: HEAD, GET 2025-11-03T17:40:59.1219241Z (APIServer pid=74541) INFO 11-03 17:40:59 [launcher.py:46] Route: /docs, Methods: HEAD, GET 2025-11-03T17:40:59.1220254Z (APIServer pid=74541) INFO 11-03 17:40:59 [launcher.py:46] Route: /docs/oauth2-redirect, Methods: HEAD, GET 2025-11-03T17:40:59.1220995Z (APIServer pid=74541) INFO 11-03 17:40:59 [launcher.py:46] Route: /redoc, Methods: HEAD, GET 2025-11-03T17:40:59.1221654Z (APIServer pid=74541) INFO 11-03 17:40:59 [launcher.py:46] Route: /health, Methods: GET 2025-11-03T17:40:59.1222267Z (APIServer pid=74541) INFO 11-03 17:40:59 [launcher.py:46] Route: /load, Methods: GET 2025-11-03T17:40:59.1222783Z (APIServer pid=74541) INFO 11-03 17:40:59 [launcher.py:46] Route: /ping, Methods: POST 2025-11-03T17:40:59.1223282Z (APIServer pid=74541) INFO 11-03 17:40:59 [launcher.py:46] Route: /ping, Methods: GET 2025-11-03T17:40:59.1224075Z (APIServer pid=74541) INFO 11-03 17:40:59 [launcher.py:46] Route: /tokenize, Methods: POST 2025-11-03T17:40:59.1224624Z (APIServer pid=74541) INFO 11-03 17:40:59 [launcher.py:46] Route: /detokenize, Methods: POST 2025-11-03T17:40:59.1225139Z (APIServer pid=74541) INFO 11-03 17:40:59 [launcher.py:46] Route: /v1/models, Methods: GET 2025-11-03T17:40:59.1225644Z (APIServer pid=74541) INFO 11-03 17:40:59 [launcher.py:46] Route: /version, Methods: GET 2025-11-03T17:40:59.1226168Z (APIServer pid=74541) INFO 11-03 17:40:59 [launcher.py:46] Route: /v1/responses, Methods: POST 2025-11-03T17:40:59.1226744Z (APIServer pid=74541) INFO 11-03 17:40:59 [launcher.py:46] Route: /v1/responses/{response_id}, Methods: GET 2025-11-03T17:40:59.1227377Z (APIServer pid=74541) INFO 11-03 17:40:59 [launcher.py:46] Route: /v1/responses/{response_id}/cancel, Methods: POST 2025-11-03T17:40:59.1227994Z (APIServer pid=74541) INFO 11-03 17:40:59 [launcher.py:46] Route: /v1/chat/completions, Methods: POST 2025-11-03T17:40:59.1228857Z (APIServer pid=74541) INFO 11-03 17:40:59 [launcher.py:46] Route: /v1/completions, Methods: POST 2025-11-03T17:40:59.1229450Z (APIServer pid=74541) INFO 11-03 17:40:59 [launcher.py:46] Route: /v1/embeddings, Methods: POST 2025-11-03T17:40:59.1229981Z (APIServer pid=74541) INFO 11-03 17:40:59 [launcher.py:46] Route: /pooling, Methods: POST 2025-11-03T17:40:59.1230602Z (APIServer pid=74541) INFO 11-03 17:40:59 [launcher.py:46] Route: /classify, Methods: POST 2025-11-03T17:40:59.1231116Z (APIServer pid=74541) INFO 11-03 17:40:59 [launcher.py:46] Route: /score, Methods: POST 2025-11-03T17:40:59.1231626Z (APIServer pid=74541) INFO 11-03 17:40:59 [launcher.py:46] Route: /v1/score, Methods: POST 2025-11-03T17:40:59.1232182Z (APIServer pid=74541) INFO 11-03 17:40:59 [launcher.py:46] Route: /v1/audio/transcriptions, Methods: POST 2025-11-03T17:40:59.1232774Z (APIServer pid=74541) INFO 11-03 17:40:59 [launcher.py:46] Route: /v1/audio/translations, Methods: POST 2025-11-03T17:40:59.1233317Z (APIServer pid=74541) INFO 11-03 17:40:59 [launcher.py:46] Route: /rerank, Methods: POST 2025-11-03T17:40:59.1233838Z (APIServer pid=74541) INFO 11-03 17:40:59 [launcher.py:46] Route: /v1/rerank, Methods: POST 2025-11-03T17:40:59.1234355Z (APIServer pid=74541) INFO 11-03 17:40:59 [launcher.py:46] Route: /v2/rerank, Methods: POST 2025-11-03T17:40:59.1234892Z (APIServer pid=74541) INFO 11-03 17:40:59 [launcher.py:46] Route: /scale_elastic_ep, Methods: POST 2025-11-03T17:40:59.1235461Z (APIServer pid=74541) INFO 11-03 17:40:59 [launcher.py:46] Route: /is_scaling_elastic_ep, Methods: POST 2025-11-03T17:40:59.1236011Z (APIServer pid=74541) INFO 11-03 17:40:59 [launcher.py:46] Route: /invocations, Methods: POST 2025-11-03T17:40:59.1236528Z (APIServer pid=74541) INFO 11-03 17:40:59 [launcher.py:46] Route: /metrics, Methods: GET 2025-11-03T17:40:59.1373944Z (APIServer pid=74541) INFO: Started server process [74541] 2025-11-03T17:40:59.1375496Z (APIServer pid=74541) INFO: Waiting for application startup. 2025-11-03T17:40:59.3604997Z (APIServer pid=74541) INFO: Application startup complete. 2025-11-03T17:40:59.6834868Z (APIServer pid=74541) INFO: 127.0.0.1:41724 - "GET /health HTTP/1.1" 200 OK 2025-11-03T17:40:59.9337962Z (APIServer pid=74541) INFO: 127.0.0.1:41740 - "GET /v1/models HTTP/1.1" 200 OK 2025-11-03T17:40:59.9631080Z (EngineCore_DP0 pid=74664) INFO 11-03 17:40:59 [ray_distributed_executor.py:569] RAY_CGRAPH_get_timeout is set to 300 2025-11-03T17:40:59.9632051Z (EngineCore_DP0 pid=74664) INFO 11-03 17:40:59 [ray_distributed_executor.py:573] VLLM_USE_RAY_COMPILED_DAG_CHANNEL_TYPE = auto 2025-11-03T17:40:59.9633218Z (EngineCore_DP0 pid=74664) INFO 11-03 17:40:59 [ray_distributed_executor.py:577] VLLM_USE_RAY_COMPILED_DAG_OVERLAP_COMM = False 2025-11-03T17:40:59.9638753Z (EngineCore_DP0 pid=74664) INFO 11-03 17:40:59 [ray_distributed_executor.py:653] Using RayPPCommunicator (which wraps vLLM _PP GroupCoordinator) for Ray Compiled Graph communication. 2025-11-03T17:41:00.1945282Z (APIServer pid=74541) INFO: 127.0.0.1:41740 - "POST /v1/completions HTTP/1.1" 200 OK 2025-11-03T17:41:00.3147909Z (APIServer pid=74541) INFO: 127.0.0.1:41740 - "POST /v1/completions HTTP/1.1" 200 OK 2025-11-03T17:41:00.5204253Z (APIServer pid=74541) INFO: 127.0.0.1:41740 - "POST /v1/completions HTTP/1.1" 200 OK 2025-11-03T17:41:00.6484820Z (APIServer pid=74541) INFO: 127.0.0.1:41740 - "POST /v1/completions HTTP/1.1" 200 OK 2025-11-03T17:41:00.7682697Z (APIServer pid=74541) INFO: 127.0.0.1:41740 - "POST /v1/completions HTTP/1.1" 200 OK 2025-11-03T17:41:00.7764566Z (APIServer pid=74541) INFO: 127.0.0.1:41740 - "POST /v1/completions HTTP/1.1" 200 OK 2025-11-03T17:41:00.9217824Z (APIServer pid=74541) INFO 11-03 17:41:00 [launcher.py:110] Shutting down FastAPI HTTP server. 2025-11-03T17:41:00.9219168Z *** SIGTERM received at time=1762191660 on cpu 32 *** 2025-11-03T17:41:00.9265657Z PC: @ 0x7f13a30eb117 (unknown) (unknown) 2025-11-03T17:41:00.9266926Z @ 0x7f13a309c520 (unknown) (unknown) 2025-11-03T17:41:00.9268291Z [2025-11-03 17:41:00,926 E 74664 74664] logging.cc:501: *** SIGTERM received at time=1762191660 on cpu 32 *** 2025-11-03T17:41:00.9269842Z [2025-11-03 17:41:00,926 E 74664 74664] logging.cc:501: PC: @ 0x7f13a30eb117 (unknown) (unknown) 2025-11-03T17:41:00.9271380Z [2025-11-03 17:41:00,926 E 74664 74664] logging.cc:501: @ 0x7f13a309c520 (unknown) (unknown) 2025-11-03T17:41:00.9274167Z (EngineCore_DP0 pid=74664) INFO 11-03 17:41:00 [ray_distributed_executor.py:127] Shutting down Ray distributed executor. If you see error log from logging.cc regarding SIGTERM received, please ignore because this is the expected termination process in Ray. 2025-11-03T17:41:00.9275316Z (EngineCore_DP0 pid=74664) 2025-11-03 17:41:00,926 INFO compiled_dag_node.py:2171 -- Tearing down compiled DAG 2025-11-03T17:41:00.9276179Z (EngineCore_DP0 pid=74664) 2025-11-03 17:41:00,927 INFO compiled_dag_node.py:2176 -- Cancelling compiled worker on actor: Actor(RayWorkerWrapper, 80edc752a6737f057710a64f01000000) 2025-11-03T17:41:00.9277169Z (EngineCore_DP0 pid=74664) 2025-11-03 17:41:00,927 INFO compiled_dag_node.py:2176 -- Cancelling compiled worker on actor: Actor(RayWorkerWrapper, ef0dcf631eb1df87aa50e00d01000000) 2025-11-03T17:41:00.9319715Z (EngineCore_DP0 pid=74664) 2025-11-03 17:41:00,931 INFO compiled_dag_node.py:2198 -- Waiting for worker tasks to exit 2025-11-03T17:41:00.9321367Z (EngineCore_DP0 pid=74664) 2025-11-03 17:41:00,931 INFO compiled_dag_node.py:2201 -- Teardown complete 2025-11-03T17:41:01.4344279Z (EngineCore_DP0 pid=74664) (RayWorkerWrapper pid=75557) INFO 11-03 17:40:58 [monitor.py:33] torch.compile takes 7.32 s in total 2025-11-03T17:41:01.4345686Z (EngineCore_DP0 pid=74664) (RayWorkerWrapper pid=75557) INFO 11-03 17:40:58 [backends.py:185] Directly load the compiled graph(s) for shape 4 from the cache, took 22.279 s 2025-11-03T17:41:01.4347644Z (EngineCore_DP0 pid=74664) (pid=75557) /opt/conda/envs/py_3.12/lib/python3.12/site-packages/torch/cuda/__init__.py:65: FutureWarning: The pynvml package is deprecated. Please install nvidia-ml-py instead. If you did not install pynvml directly, please report this to the maintainers of the package that installed pynvml for you. 2025-11-03T17:41:01.4349513Z (EngineCore_DP0 pid=74664) (pid=75557) import pynvml # type: ignore[import] 2025-11-03T17:41:03.4733682Z (APIServer pid=74541) INFO: Shutting down 2025-11-03T17:41:03.5740287Z (APIServer pid=74541) INFO: Waiting for application shutdown. 2025-11-03T17:41:03.5741551Z (APIServer pid=74541) INFO: Application shutdown complete. 2025-11-03T17:41:03.9719984Z INFO 11-03 17:41:03 [model.py:653] Resolved architecture: LlamaForCausalLM 2025-11-03T17:41:03.9720574Z WARNING 11-03 17:41:03 [model.py:1942] Casting torch.bfloat16 to torch.float16. 2025-11-03T17:41:03.9721030Z INFO 11-03 17:41:03 [model.py:1714] Using max model len 2048 2025-11-03T17:41:03.9725917Z Launching RemoteOpenAIServer with: vllm serve RedHatAI/Meta-Llama-3.1-8B-Instruct-FP8 --dtype float16 --max-model-len 2048 --max-num-seqs 8 --enforce-eager --tokenizer-mode auto --tensor-parallel-size 2 --distributed-executor-backend mp --load-format dummy --port 55673 --seed 0 2025-11-03T17:41:04.4923027Z /opt/conda/envs/py_3.12/lib/python3.12/site-packages/torch/cuda/__init__.py:65: FutureWarning: The pynvml package is deprecated. Please install nvidia-ml-py instead. If you did not install pynvml directly, please report this to the maintainers of the package that installed pynvml for you. 2025-11-03T17:41:04.4925494Z import pynvml # type: ignore[import] 2025-11-03T17:41:06.7484477Z INFO 11-03 17:41:06 [__init__.py:224] Automatically detected platform cuda. 2025-11-03T17:41:10.4449644Z (APIServer pid=79741) INFO 11-03 17:41:10 [api_server.py:1870] vLLM API server version 0.11.0rc2.dev389+ge51928192.d20251103 2025-11-03T17:41:10.4472274Z (APIServer pid=79741) INFO 11-03 17:41:10 [utils.py:239] non-default args: {'model_tag': 'RedHatAI/Meta-Llama-3.1-8B-Instruct-FP8', 'port': 55673, 'model': 'RedHatAI/Meta-Llama-3.1-8B-Instruct-FP8', 'dtype': 'float16', 'seed': 0, 'max_model_len': 2048, 'enforce_eager': True, 'load_format': 'dummy', 'distributed_executor_backend': 'mp', 'tensor_parallel_size': 2, 'max_num_seqs': 8} 2025-11-03T17:41:10.9006183Z (APIServer pid=79741) INFO 11-03 17:41:10 [model.py:653] Resolved architecture: LlamaForCausalLM 2025-11-03T17:41:10.9007327Z (APIServer pid=79741) `torch_dtype` is deprecated! Use `dtype` instead! 2025-11-03T17:41:10.9008055Z (APIServer pid=79741) WARNING 11-03 17:41:10 [model.py:1942] Casting torch.bfloat16 to torch.float16. 2025-11-03T17:41:10.9008855Z (APIServer pid=79741) INFO 11-03 17:41:10 [model.py:1714] Using max model len 2048 2025-11-03T17:41:11.2379587Z (APIServer pid=79741) INFO 11-03 17:41:11 [scheduler.py:225] Chunked prefill is enabled with max_num_batched_tokens=2048. 2025-11-03T17:41:11.2754652Z (APIServer pid=79741) INFO 11-03 17:41:11 [vllm.py:354] Cudagraph is disabled under eager mode 2025-11-03T17:41:12.3598604Z /opt/conda/envs/py_3.12/lib/python3.12/site-packages/torch/cuda/__init__.py:65: FutureWarning: The pynvml package is deprecated. Please install nvidia-ml-py instead. If you did not install pynvml directly, please report this to the maintainers of the package that installed pynvml for you. 2025-11-03T17:41:12.3600913Z import pynvml # type: ignore[import] 2025-11-03T17:41:14.7091453Z INFO 11-03 17:41:14 [__init__.py:224] Automatically detected platform cuda. 2025-11-03T17:41:18.5664899Z (EngineCore_DP0 pid=79864) INFO 11-03 17:41:18 [core.py:727] Waiting for init message from front-end. 2025-11-03T17:41:18.5926942Z (EngineCore_DP0 pid=79864) INFO 11-03 17:41:18 [core.py:94] Initializing a V1 LLM engine (v0.11.0rc2.dev389+ge51928192.d20251103) with config: model='RedHatAI/Meta-Llama-3.1-8B-Instruct-FP8', speculative_config=None, tokenizer='RedHatAI/Meta-Llama-3.1-8B-Instruct-FP8', skip_tokenizer_init=False, tokenizer_mode=auto, revision=None, tokenizer_revision=None, trust_remote_code=False, dtype=torch.float16, max_seq_len=2048, download_dir=None, load_format=dummy, tensor_parallel_size=2, pipeline_parallel_size=1, data_parallel_size=1, disable_custom_all_reduce=False, quantization=compressed-tensors, enforce_eager=True, kv_cache_dtype=auto, device_config=cuda, structured_outputs_config=StructuredOutputsConfig(backend='auto', disable_fallback=False, disable_any_whitespace=False, disable_additional_properties=False, reasoning_parser=''), observability_config=ObservabilityConfig(show_hidden_metrics_for_version=None, otlp_traces_endpoint=None, collect_detailed_traces=None), seed=0, served_model_name=RedHatAI/Meta-Llama-3.1-8B-Instruct-FP8, enable_prefix_caching=True, chunked_prefill_enabled=True, pooler_config=None, compilation_config={'level': 0, 'debug_dump_path': None, 'cache_dir': '', 'backend': '', 'custom_ops': [], 'splitting_ops': None, 'use_inductor': True, 'compile_sizes': [], 'inductor_compile_config': {'enable_auto_functionalized_v2': False}, 'inductor_passes': {}, 'cudagraph_mode': , 'use_cudagraph': False, 'cudagraph_num_of_warmups': 0, 'cudagraph_capture_sizes': [], 'cudagraph_copy_inputs': False, 'full_cuda_graph': False, 'use_inductor_graph_partition': False, 'pass_config': {}, 'max_capture_size': 0, 'local_cache_dir': None} 2025-11-03T17:41:18.5932710Z (EngineCore_DP0 pid=79864) WARNING 11-03 17:41:18 [multiproc_executor.py:744] Reducing Torch parallelism from 24 threads to 1 to avoid unnecessary CPU contention. Set OMP_NUM_THREADS in the external environment to tune this value as needed. 2025-11-03T17:41:18.5939524Z (EngineCore_DP0 pid=79864) INFO 11-03 17:41:18 [shm_broadcast.py:302] vLLM message queue communication handle: Handle(local_reader_ranks=[0, 1], buffer_handle=(2, 16777216, 10, 'psm_8134f691'), local_subscribe_addr='ipc:///tmp/e5915c8a-557b-4faf-82ed-fe4bb6cf72a9', remote_subscribe_addr=None, remote_addr_ipv6=False) 2025-11-03T17:41:19.0759562Z /opt/conda/envs/py_3.12/lib/python3.12/site-packages/torch/cuda/__init__.py:65: FutureWarning: The pynvml package is deprecated. Please install nvidia-ml-py instead. If you did not install pynvml directly, please report this to the maintainers of the package that installed pynvml for you. 2025-11-03T17:41:19.0761828Z import pynvml # type: ignore[import] 2025-11-03T17:41:19.0876069Z /opt/conda/envs/py_3.12/lib/python3.12/site-packages/torch/cuda/__init__.py:65: FutureWarning: The pynvml package is deprecated. Please install nvidia-ml-py instead. If you did not install pynvml directly, please report this to the maintainers of the package that installed pynvml for you. 2025-11-03T17:41:19.0878233Z import pynvml # type: ignore[import] 2025-11-03T17:41:21.4246634Z INFO 11-03 17:41:21 [__init__.py:224] Automatically detected platform cuda. 2025-11-03T17:41:21.4543355Z INFO 11-03 17:41:21 [__init__.py:224] Automatically detected platform cuda. 2025-11-03T17:41:26.3028860Z INFO 11-03 17:41:26 [shm_broadcast.py:302] vLLM message queue communication handle: Handle(local_reader_ranks=[0], buffer_handle=(1, 10485760, 10, 'psm_64720cb3'), local_subscribe_addr='ipc:///tmp/32476cf0-1719-4937-8933-cb56e77f0966', remote_subscribe_addr=None, remote_addr_ipv6=False) 2025-11-03T17:41:26.3223683Z INFO 11-03 17:41:26 [shm_broadcast.py:302] vLLM message queue communication handle: Handle(local_reader_ranks=[0], buffer_handle=(1, 10485760, 10, 'psm_c9b317b9'), local_subscribe_addr='ipc:///tmp/b7caf9a7-a659-42a8-a77b-4b83e235c85d', remote_subscribe_addr=None, remote_addr_ipv6=False) 2025-11-03T17:41:27.0798742Z [Gloo] Rank 0 is connected to 1 peer ranks. Expected number of connected peer ranks is : 1 2025-11-03T17:41:27.0799375Z [Gloo] Rank 1 is connected to 1 peer ranks. Expected number of connected peer ranks is : 1 2025-11-03T17:41:27.0842926Z [Gloo] Rank 0 is connected to 1 peer ranks. Expected number of connected peer ranks is : 1 2025-11-03T17:41:27.0843533Z [Gloo] Rank 1 is connected to 1 peer ranks. Expected number of connected peer ranks is : 1 2025-11-03T17:41:27.0886335Z INFO 11-03 17:41:27 [__init__.py:1146] Found nccl from library libnccl.so.2 2025-11-03T17:41:27.0887421Z INFO 11-03 17:41:27 [__init__.py:1146] Found nccl from library libnccl.so.2 2025-11-03T17:41:27.0888790Z INFO 11-03 17:41:27 [pynccl.py:109] vLLM is using nccl==2.27.5 2025-11-03T17:41:27.0890021Z INFO 11-03 17:41:27 [pynccl.py:109] vLLM is using nccl==2.27.5 2025-11-03T17:41:27.2605382Z WARNING 11-03 17:41:27 [symm_mem.py:60] SymmMemCommunicator: Device capability 8.9 not supported, communicator is not available. 2025-11-03T17:41:27.2606265Z WARNING 11-03 17:41:27 [symm_mem.py:60] SymmMemCommunicator: Device capability 8.9 not supported, communicator is not available. 2025-11-03T17:41:27.2633346Z INFO 11-03 17:41:27 [custom_all_reduce.py:37] Skipping P2P check and trusting the driver's P2P report. 2025-11-03T17:41:27.2634363Z WARNING 11-03 17:41:27 [custom_all_reduce.py:162] Custom allreduce is disabled because your platform lacks GPU P2P capability or P2P test failed. To silence this warning, specify disable_custom_all_reduce=True explicitly. 2025-11-03T17:41:27.2635371Z INFO 11-03 17:41:27 [custom_all_reduce.py:37] Skipping P2P check and trusting the driver's P2P report. 2025-11-03T17:41:27.2636773Z WARNING 11-03 17:41:27 [custom_all_reduce.py:162] Custom allreduce is disabled because your platform lacks GPU P2P capability or P2P test failed. To silence this warning, specify disable_custom_all_reduce=True explicitly. 2025-11-03T17:41:27.3425057Z INFO 11-03 17:41:27 [shm_broadcast.py:302] vLLM message queue communication handle: Handle(local_reader_ranks=[1], buffer_handle=(1, 4194304, 6, 'psm_407e8764'), local_subscribe_addr='ipc:///tmp/91bc59b3-dc19-48fb-86fb-5708d5efa839', remote_subscribe_addr=None, remote_addr_ipv6=False) 2025-11-03T17:41:27.3454258Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:41:27.3454837Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:41:27.3462476Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:41:27.3470349Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:41:27.4021862Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:41:27.4029638Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:41:27.4040208Z [Gloo] Rank 1 is connected to 1 peer ranks. Expected number of connected peer ranks is : 1 2025-11-03T17:41:27.4040715Z [Gloo] Rank 0 is connected to 1 peer ranks. Expected number of connected peer ranks is : 1 2025-11-03T17:41:27.4042589Z INFO 11-03 17:41:27 [__init__.py:1146] Found nccl from library libnccl.so.2 2025-11-03T17:41:27.4042949Z INFO 11-03 17:41:27 [pynccl.py:109] vLLM is using nccl==2.27.5 2025-11-03T17:41:27.4043293Z INFO 11-03 17:41:27 [__init__.py:1146] Found nccl from library libnccl.so.2 2025-11-03T17:41:27.4044243Z INFO 11-03 17:41:27 [pynccl.py:109] vLLM is using nccl==2.27.5 2025-11-03T17:41:27.4376348Z INFO 11-03 17:41:27 [parallel_state.py:1231] rank 1 in world size 2 is assigned as DP rank 0, PP rank 0, TP rank 1, EP rank 1 2025-11-03T17:41:27.4377161Z INFO 11-03 17:41:27 [parallel_state.py:1231] rank 0 in world size 2 is assigned as DP rank 0, PP rank 0, TP rank 0, EP rank 0 2025-11-03T17:41:27.6624141Z INFO 11-03 17:41:27 [topk_topp_sampler.py:59] Using FlashInfer for top-p & top-k sampling. 2025-11-03T17:41:27.6785084Z INFO 11-03 17:41:27 [topk_topp_sampler.py:59] Using FlashInfer for top-p & top-k sampling. 2025-11-03T17:41:27.6993798Z (Worker_TP0 pid=79983) INFO 11-03 17:41:27 [gpu_model_runner.py:2840] Starting to load model RedHatAI/Meta-Llama-3.1-8B-Instruct-FP8... 2025-11-03T17:41:27.7156344Z (Worker_TP1 pid=79984) INFO 11-03 17:41:27 [gpu_model_runner.py:2840] Starting to load model RedHatAI/Meta-Llama-3.1-8B-Instruct-FP8... 2025-11-03T17:41:27.9102874Z (Worker_TP0 pid=79983) INFO 11-03 17:41:27 [gpu_model_runner.py:2870] Loading model from scratch... 2025-11-03T17:41:27.9239735Z (Worker_TP1 pid=79984) INFO 11-03 17:41:27 [gpu_model_runner.py:2870] Loading model from scratch... 2025-11-03T17:41:28.0025274Z (Worker_TP0 pid=79983) INFO 11-03 17:41:28 [cuda.py:404] Using Flash Attention backend on V1 engine. 2025-11-03T17:41:28.0150159Z (Worker_TP1 pid=79984) INFO 11-03 17:41:28 [cuda.py:404] Using Flash Attention backend on V1 engine. 2025-11-03T17:41:28.8190659Z (Worker_TP0 pid=79983) INFO 11-03 17:41:28 [gpu_model_runner.py:2902] Model loading took 4.2624 GiB and 0.489298 seconds 2025-11-03T17:41:28.8273554Z (Worker_TP1 pid=79984) INFO 11-03 17:41:28 [gpu_model_runner.py:2902] Model loading took 4.2624 GiB and 0.489438 seconds 2025-11-03T17:41:31.2520272Z (Worker_TP1 pid=79984) INFO 11-03 17:41:31 [gpu_worker.py:314] Available KV cache memory: 15.32 GiB 2025-11-03T17:41:31.2563907Z (Worker_TP0 pid=79983) INFO 11-03 17:41:31 [gpu_worker.py:314] Available KV cache memory: 15.32 GiB 2025-11-03T17:41:31.4701650Z (EngineCore_DP0 pid=79864) INFO 11-03 17:41:31 [kv_cache_utils.py:1199] GPU KV cache size: 250,976 tokens 2025-11-03T17:41:31.4702606Z (EngineCore_DP0 pid=79864) INFO 11-03 17:41:31 [kv_cache_utils.py:1204] Maximum concurrency for 2,048 tokens per request: 122.55x 2025-11-03T17:41:31.4703445Z (EngineCore_DP0 pid=79864) INFO 11-03 17:41:31 [kv_cache_utils.py:1199] GPU KV cache size: 250,976 tokens 2025-11-03T17:41:31.4704408Z (EngineCore_DP0 pid=79864) INFO 11-03 17:41:31 [kv_cache_utils.py:1204] Maximum concurrency for 2,048 tokens per request: 122.55x 2025-11-03T17:41:31.6000769Z (EngineCore_DP0 pid=79864) INFO 11-03 17:41:31 [core.py:240] init engine (profile, create kv cache, warmup model) took 2.69 seconds 2025-11-03T17:41:32.0927113Z (EngineCore_DP0 pid=79864) INFO 11-03 17:41:32 [vllm.py:354] Cudagraph is disabled under eager mode 2025-11-03T17:41:32.2773468Z (APIServer pid=79741) INFO 11-03 17:41:32 [loggers.py:172] Engine 000: vllm cache_config_info with initialization after num_gpu_blocks is: 15686 2025-11-03T17:41:32.2923739Z (EngineCore_DP0 pid=79864) INFO 11-03 17:41:32 [gc_utils.py:40] GC Debug Config. enabled:False,top_objects:-1 2025-11-03T17:41:32.3141371Z (APIServer pid=79741) INFO 11-03 17:41:32 [api_server.py:1628] Supported tasks: ['generate'] 2025-11-03T17:41:32.3519750Z (APIServer pid=79741) WARNING 11-03 17:41:32 [model.py:1581] Default sampling parameters have been overridden by the model's Hugging Face generation config recommended from the model creator. If this is not intended, please relaunch vLLM instance with `--generation-config vllm`. 2025-11-03T17:41:32.3521100Z (APIServer pid=79741) INFO 11-03 17:41:32 [serving_responses.py:157] Using default chat sampling params from model: {'temperature': 0.6, 'top_p': 0.9} 2025-11-03T17:41:32.4926264Z (APIServer pid=79741) INFO 11-03 17:41:32 [serving_chat.py:159] Using default chat sampling params from model: {'temperature': 0.6, 'top_p': 0.9} 2025-11-03T17:41:32.5259581Z (APIServer pid=79741) INFO 11-03 17:41:32 [serving_completion.py:67] Using default completion sampling params from model: {'temperature': 0.6, 'top_p': 0.9} 2025-11-03T17:41:32.5260605Z (APIServer pid=79741) INFO 11-03 17:41:32 [api_server.py:1939] Starting vLLM API server 0 on http://0.0.0.0:55673 2025-11-03T17:41:32.5261317Z (APIServer pid=79741) INFO 11-03 17:41:32 [launcher.py:38] Available routes are: 2025-11-03T17:41:32.5261994Z (APIServer pid=79741) INFO 11-03 17:41:32 [launcher.py:46] Route: /openapi.json, Methods: HEAD, GET 2025-11-03T17:41:32.5262681Z (APIServer pid=79741) INFO 11-03 17:41:32 [launcher.py:46] Route: /docs, Methods: HEAD, GET 2025-11-03T17:41:32.5263382Z (APIServer pid=79741) INFO 11-03 17:41:32 [launcher.py:46] Route: /docs/oauth2-redirect, Methods: HEAD, GET 2025-11-03T17:41:32.5264277Z (APIServer pid=79741) INFO 11-03 17:41:32 [launcher.py:46] Route: /redoc, Methods: HEAD, GET 2025-11-03T17:41:32.5264919Z (APIServer pid=79741) INFO 11-03 17:41:32 [launcher.py:46] Route: /health, Methods: GET 2025-11-03T17:41:32.5265561Z (APIServer pid=79741) INFO 11-03 17:41:32 [launcher.py:46] Route: /load, Methods: GET 2025-11-03T17:41:32.5266173Z (APIServer pid=79741) INFO 11-03 17:41:32 [launcher.py:46] Route: /ping, Methods: POST 2025-11-03T17:41:32.5266790Z (APIServer pid=79741) INFO 11-03 17:41:32 [launcher.py:46] Route: /ping, Methods: GET 2025-11-03T17:41:32.5267423Z (APIServer pid=79741) INFO 11-03 17:41:32 [launcher.py:46] Route: /tokenize, Methods: POST 2025-11-03T17:41:32.5268066Z (APIServer pid=79741) INFO 11-03 17:41:32 [launcher.py:46] Route: /detokenize, Methods: POST 2025-11-03T17:41:32.5268703Z (APIServer pid=79741) INFO 11-03 17:41:32 [launcher.py:46] Route: /v1/models, Methods: GET 2025-11-03T17:41:32.5269333Z (APIServer pid=79741) INFO 11-03 17:41:32 [launcher.py:46] Route: /version, Methods: GET 2025-11-03T17:41:32.5270157Z (APIServer pid=79741) INFO 11-03 17:41:32 [launcher.py:46] Route: /v1/responses, Methods: POST 2025-11-03T17:41:32.5270827Z (APIServer pid=79741) INFO 11-03 17:41:32 [launcher.py:46] Route: /v1/responses/{response_id}, Methods: GET 2025-11-03T17:41:32.5271525Z (APIServer pid=79741) INFO 11-03 17:41:32 [launcher.py:46] Route: /v1/responses/{response_id}/cancel, Methods: POST 2025-11-03T17:41:32.5272143Z (APIServer pid=79741) INFO 11-03 17:41:32 [launcher.py:46] Route: /v1/chat/completions, Methods: POST 2025-11-03T17:41:32.5272705Z (APIServer pid=79741) INFO 11-03 17:41:32 [launcher.py:46] Route: /v1/completions, Methods: POST 2025-11-03T17:41:32.5273250Z (APIServer pid=79741) INFO 11-03 17:41:32 [launcher.py:46] Route: /v1/embeddings, Methods: POST 2025-11-03T17:41:32.5273778Z (APIServer pid=79741) INFO 11-03 17:41:32 [launcher.py:46] Route: /pooling, Methods: POST 2025-11-03T17:41:32.5274302Z (APIServer pid=79741) INFO 11-03 17:41:32 [launcher.py:46] Route: /classify, Methods: POST 2025-11-03T17:41:32.5274809Z (APIServer pid=79741) INFO 11-03 17:41:32 [launcher.py:46] Route: /score, Methods: POST 2025-11-03T17:41:32.5275327Z (APIServer pid=79741) INFO 11-03 17:41:32 [launcher.py:46] Route: /v1/score, Methods: POST 2025-11-03T17:41:32.5275886Z (APIServer pid=79741) INFO 11-03 17:41:32 [launcher.py:46] Route: /v1/audio/transcriptions, Methods: POST 2025-11-03T17:41:32.5276470Z (APIServer pid=79741) INFO 11-03 17:41:32 [launcher.py:46] Route: /v1/audio/translations, Methods: POST 2025-11-03T17:41:32.5276995Z (APIServer pid=79741) INFO 11-03 17:41:32 [launcher.py:46] Route: /rerank, Methods: POST 2025-11-03T17:41:32.5277515Z (APIServer pid=79741) INFO 11-03 17:41:32 [launcher.py:46] Route: /v1/rerank, Methods: POST 2025-11-03T17:41:32.5278032Z (APIServer pid=79741) INFO 11-03 17:41:32 [launcher.py:46] Route: /v2/rerank, Methods: POST 2025-11-03T17:41:32.5278565Z (APIServer pid=79741) INFO 11-03 17:41:32 [launcher.py:46] Route: /scale_elastic_ep, Methods: POST 2025-11-03T17:41:32.5279137Z (APIServer pid=79741) INFO 11-03 17:41:32 [launcher.py:46] Route: /is_scaling_elastic_ep, Methods: POST 2025-11-03T17:41:32.5279684Z (APIServer pid=79741) INFO 11-03 17:41:32 [launcher.py:46] Route: /invocations, Methods: POST 2025-11-03T17:41:32.5280196Z (APIServer pid=79741) INFO 11-03 17:41:32 [launcher.py:46] Route: /metrics, Methods: GET 2025-11-03T17:41:32.5426638Z (APIServer pid=79741) INFO: Started server process [79741] 2025-11-03T17:41:32.5427276Z (APIServer pid=79741) INFO: Waiting for application startup. 2025-11-03T17:41:32.7594141Z (APIServer pid=79741) INFO: Application startup complete. 2025-11-03T17:41:33.0465020Z (APIServer pid=79741) INFO: 127.0.0.1:46970 - "GET /health HTTP/1.1" 200 OK 2025-11-03T17:41:33.0556469Z (APIServer pid=79741) INFO: 127.0.0.1:46982 - "GET /v1/models HTTP/1.1" 200 OK 2025-11-03T17:41:33.2858270Z (APIServer pid=79741) INFO: 127.0.0.1:46982 - "POST /v1/completions HTTP/1.1" 200 OK 2025-11-03T17:41:33.4011868Z (APIServer pid=79741) INFO: 127.0.0.1:46982 - "POST /v1/completions HTTP/1.1" 200 OK 2025-11-03T17:41:33.5542763Z (APIServer pid=79741) INFO: 127.0.0.1:46982 - "POST /v1/completions HTTP/1.1" 200 OK 2025-11-03T17:41:33.7065337Z (APIServer pid=79741) INFO: 127.0.0.1:46982 - "POST /v1/completions HTTP/1.1" 200 OK 2025-11-03T17:41:33.8489677Z (APIServer pid=79741) INFO: 127.0.0.1:46982 - "POST /v1/completions HTTP/1.1" 200 OK 2025-11-03T17:41:33.8565072Z (APIServer pid=79741) INFO: 127.0.0.1:46982 - "POST /v1/completions HTTP/1.1" 200 OK 2025-11-03T17:41:34.0023630Z (APIServer pid=79741) INFO 11-03 17:41:34 [launcher.py:110] Shutting down FastAPI HTTP server. 2025-11-03T17:41:34.0027538Z (Worker_TP0 pid=79983) INFO 11-03 17:41:34 [multiproc_executor.py:583] Parent process exited, terminating worker 2025-11-03T17:41:34.0029180Z (Worker_TP0 pid=79983) INFO 11-03 17:41:34 [multiproc_executor.py:624] WorkerProc shutting down. 2025-11-03T17:41:34.0033610Z (Worker_TP1 pid=79984) INFO 11-03 17:41:34 [multiproc_executor.py:583] Parent process exited, terminating worker 2025-11-03T17:41:34.0034257Z (Worker_TP1 pid=79984) INFO 11-03 17:41:34 [multiproc_executor.py:624] WorkerProc shutting down. 2025-11-03T17:41:36.2245291Z (APIServer pid=79741) INFO: Shutting down 2025-11-03T17:41:36.3251350Z (APIServer pid=79741) INFO: Waiting for application shutdown. 2025-11-03T17:41:36.3253030Z (APIServer pid=79741) INFO: Application shutdown complete. 2025-11-03T17:41:36.9385253Z PASSED 2025-11-03T17:41:36.9385507Z 2025-11-03T17:41:36.9385881Z =============================== warnings summary =============================== 2025-11-03T17:41:36.9386573Z ../../../../../../opt/conda/envs/py_3.12/lib/python3.12/site-packages/torch/cuda/__init__.py:65 2025-11-03T17:41:36.9388028Z /opt/conda/envs/py_3.12/lib/python3.12/site-packages/torch/cuda/__init__.py:65: FutureWarning: The pynvml package is deprecated. Please install nvidia-ml-py instead. If you did not install pynvml directly, please report this to the maintainers of the package that installed pynvml for you. 2025-11-03T17:41:36.9389459Z import pynvml # type: ignore[import] 2025-11-03T17:41:36.9389664Z 2025-11-03T17:41:36.9389764Z :488 2025-11-03T17:41:36.9390428Z :488: DeprecationWarning: builtin type SwigPyPacked has no __module__ attribute 2025-11-03T17:41:36.9390922Z 2025-11-03T17:41:36.9391095Z :488 2025-11-03T17:41:36.9391586Z :488: DeprecationWarning: builtin type SwigPyObject has no __module__ attribute 2025-11-03T17:41:36.9392159Z 2025-11-03T17:41:36.9392485Z ../../../../../../opt/conda/envs/py_3.12/lib/python3.12/site-packages/schemathesis/generation/coverage.py:305 2025-11-03T17:41:36.9394198Z /opt/conda/envs/py_3.12/lib/python3.12/site-packages/schemathesis/generation/coverage.py:305: DeprecationWarning: jsonschema.exceptions.RefResolutionError is deprecated as of version 4.18.0. If you wish to catch potential reference resolution errors, directly catch referencing.exceptions.Unresolvable. 2025-11-03T17:41:36.9395753Z ref_error: type[Exception] = jsonschema.RefResolutionError, 2025-11-03T17:41:36.9395977Z 2025-11-03T17:41:36.9396120Z tests/distributed/test_sequence_parallel.py: 20 warnings 2025-11-03T17:41:36.9396881Z /var/lib/jenkins/workspace/vllm/tests/utils.py:872: DeprecationWarning: This process (pid=1107) is multi-threaded, use of fork() may lead to deadlocks in the child. 2025-11-03T17:41:36.9397572Z pid = os.fork() 2025-11-03T17:41:36.9397682Z 2025-11-03T17:41:36.9397879Z -- Docs: https://docs.pytest.org/en/stable/how-to/capture-warnings.html 2025-11-03T17:41:36.9398546Z ================= 20 passed, 24 warnings in 1943.07s (0:32:23) ================= 2025-11-03T17:41:37.1938762Z sys:1: DeprecationWarning: builtin type swigvarlink has no __module__ attribute 2025-11-03T17:41:38.3026725Z 2025-11-03 17:41:38,302 [INFO] cli.lib.core.vllm.lib: Finish running step: pytest -v -s distributed/test_sequence_parallel.py 2025-11-03T17:41:38.3028037Z 2025-11-03 17:41:38,302 [INFO] cli.lib.core.vllm.lib: Done. All tests passed 2025-11-03T17:41:38.3029004Z 2025-11-03 17:41:38,302 [INFO] cli.lib.core.vllm.vllm_test: Double check installed packages 2025-11-03T17:41:38.3036273Z 2025-11-03 17:41:38,303 [INFO] cli.lib.common.pip_helper: torch already exist with version: 2.10.0a0+git3f6538f 2025-11-03T17:41:38.3038139Z 2025-11-03 17:41:38,303 [INFO] cli.lib.common.pip_helper: xformers already exist with version: 0.0.33+5d4b92a5.d20251103 2025-11-03T17:41:38.3041975Z 2025-11-03 17:41:38,303 [INFO] cli.lib.common.pip_helper: torchvision already exist with version: 0.25.0a0+cfbc5c2 2025-11-03T17:41:38.3046446Z 2025-11-03 17:41:38,304 [INFO] cli.lib.common.pip_helper: torchaudio already exist with version: 2.10.0a0+3b0e7a6 2025-11-03T17:41:38.3053439Z 2025-11-03 17:41:38,304 [INFO] cli.lib.common.pip_helper: vllm already exist with version: 0.11.0rc2.dev389+ge51928192.d20251103 2025-11-03T17:41:38.3054749Z 2025-11-03 17:41:38,305 [INFO] cli.lib.core.vllm.vllm_test: Done. checked installed packages 2025-11-03T17:41:38.3300066Z + sccache_epilogue 2025-11-03T17:41:38.3301722Z + echo '::group::Sccache Compilation Log' 2025-11-03T17:41:38.3303364Z ##[group]Sccache Compilation Log 2025-11-03T17:41:38.3304146Z + echo '=================== sccache compilation log ===================' 2025-11-03T17:41:38.3305118Z =================== sccache compilation log =================== 2025-11-03T17:41:38.3306601Z + python /var/lib/jenkins/workspace/.ci/pytorch/print_sccache_log.py /var/lib/jenkins/sccache_error.log 2025-11-03T17:41:38.3760328Z + echo '=========== If your build fails, please take a look at the log above for possible reasons ===========' 2025-11-03T17:41:38.3761508Z =========== If your build fails, please take a look at the log above for possible reasons =========== 2025-11-03T17:41:38.3762318Z + sccache --show-stats 2025-11-03T17:41:38.3790542Z Compile requests 14 2025-11-03T17:41:38.3791092Z Compile requests executed 0 2025-11-03T17:41:38.3791635Z Cache hits 0 2025-11-03T17:41:38.3792153Z Cache misses 0 2025-11-03T17:41:38.3792655Z Cache hits rate - 2025-11-03T17:41:38.3793174Z Cache timeouts 0 2025-11-03T17:41:38.3793665Z Cache read errors 0 2025-11-03T17:41:38.3794133Z Forced recaches 0 2025-11-03T17:41:38.3794613Z Cache write errors 0 2025-11-03T17:41:38.3795098Z Cache errors 0 2025-11-03T17:41:38.3795577Z Compilations 0 2025-11-03T17:41:38.3796055Z Compilation failures 0 2025-11-03T17:41:38.3796405Z Non-cacheable compilations 0 2025-11-03T17:41:38.3796622Z Non-cacheable calls 0 2025-11-03T17:41:38.3796845Z Non-compilation calls 14 2025-11-03T17:41:38.3797064Z Unsupported compiler calls 0 2025-11-03T17:41:38.3797289Z Average cache write 0.000 s 2025-11-03T17:41:38.3797767Z Average compiler 0.000 s 2025-11-03T17:41:38.3797997Z Average cache read hit 0.000 s 2025-11-03T17:41:38.3798224Z Failed distributed compilations 0 2025-11-03T17:41:38.3798549Z Cache location s3, name: ossci-compiler-cache-circleci-v2, prefix: / 2025-11-03T17:41:38.3798870Z Version (client) 0.10.0 2025-11-03T17:41:38.3799098Z + sccache --stop-server 2025-11-03T17:41:38.3822901Z Stopping sccache server... 2025-11-03T17:41:38.3826071Z Compile requests 14 2025-11-03T17:41:38.3826754Z Compile requests executed 0 2025-11-03T17:41:38.3827286Z Cache hits 0 2025-11-03T17:41:38.3827803Z Cache misses 0 2025-11-03T17:41:38.3828298Z Cache hits rate - 2025-11-03T17:41:38.3828796Z Cache timeouts 0 2025-11-03T17:41:38.3829320Z Cache read errors 0 2025-11-03T17:41:38.3829822Z Forced recaches 0 2025-11-03T17:41:38.3830334Z Cache write errors 0 2025-11-03T17:41:38.3830807Z Cache errors 0 2025-11-03T17:41:38.3831764Z Compilations 0 2025-11-03T17:41:38.3832322Z Compilation failures 0 2025-11-03T17:41:38.3832840Z Non-cacheable compilations 0 2025-11-03T17:41:38.3833376Z Non-cacheable calls 0 2025-11-03T17:41:38.3834021Z Non-compilation calls 14 2025-11-03T17:41:38.3834519Z Unsupported compiler calls 0 2025-11-03T17:41:38.3835030Z Average cache write 0.000 s 2025-11-03T17:41:38.3835590Z Average compiler 0.000 s 2025-11-03T17:41:38.3836167Z Average cache read hit 0.000 s 2025-11-03T17:41:38.3836484Z Failed distributed compilations 0 2025-11-03T17:41:38.3836810Z Cache location s3, name: ossci-compiler-cache-circleci-v2, prefix: / 2025-11-03T17:41:38.3837149Z Version (client) 0.10.0 2025-11-03T17:41:38.3837400Z + echo ::endgroup:: 2025-11-03T17:41:38.3837821Z ##[endgroup] 2025-11-03T17:41:38.3837991Z + cleanup_workspace 2025-11-03T17:41:38.3838354Z + echo 'sudo may print the following warning message that can be ignored. The chown command will still run.' 2025-11-03T17:41:38.3838912Z sudo may print the following warning message that can be ignored. The chown command will still run. 2025-11-03T17:41:38.3839357Z + echo ' sudo: setrlimit(RLIMIT_STACK): Operation not permitted' 2025-11-03T17:41:38.3839695Z sudo: setrlimit(RLIMIT_STACK): Operation not permitted 2025-11-03T17:41:38.3840088Z + echo 'For more details refer to https://github.com/sudo-project/sudo/issues/42' 2025-11-03T17:41:38.3840514Z For more details refer to https://github.com/sudo-project/sudo/issues/42 2025-11-03T17:41:38.3840842Z + sudo chown -R 1000 /var/lib/jenkins/workspace 2025-11-03T17:41:39.4074760Z ##[group]Run pytorch/test-infra/.github/actions/upload-benchmark-results@main 2025-11-03T17:41:39.4075152Z with: 2025-11-03T17:41:39.4075345Z benchmark-results-dir: test/test-reports 2025-11-03T17:41:39.4075577Z dry-run: false 2025-11-03T17:41:39.4075738Z schema-version: v3 2025-11-03T17:41:39.4076103Z github-token: *** 2025-11-03T17:41:39.4076281Z env: 2025-11-03T17:41:39.4076436Z GIT_DEFAULT_BRANCH: main 2025-11-03T17:41:39.4076691Z GPU_FLAG: --gpus all -e NVIDIA_DRIVER_CAPABILITIES=all 2025-11-03T17:41:39.4077090Z DOCKER_CONTAINER_ID: 99e00294b4bd7ffe64017e12af3ca26302f354938a9d519b091e205874b31030 2025-11-03T17:41:39.4077443Z ##[endgroup] 2025-11-03T17:41:39.4113143Z ##[group]Run set -eux 2025-11-03T17:41:39.4113350Z set -eux 2025-11-03T17:41:39.4113507Z  2025-11-03T17:41:39.4113659Z if [[ -n "" ]]; then 2025-11-03T17:41:39.4113853Z  source "" 2025-11-03T17:41:39.4114015Z fi 2025-11-03T17:41:39.4114292Z python3 -mpip install boto3==1.35.33 psutil==7.0.0 pynvml==12.0.0 2025-11-03T17:41:39.4114596Z  2025-11-03T17:41:39.4114870Z DEVICE_NAME="" 2025-11-03T17:41:39.4115058Z DEVICE_TYPE="" 2025-11-03T17:41:39.4115229Z  2025-11-03T17:41:39.4115397Z if command -v nvidia-smi; then 2025-11-03T17:41:39.4115713Z  # NB: I'm using PyTorch here to get the device name, however, it needs to 2025-11-03T17:41:39.4116130Z  # install the correct version of PyTorch manually for now. Any PyTorch 2025-11-03T17:41:39.4116514Z  # version is fine, I just use 2.7.1 to satify PYPIDEP linter 2025-11-03T17:41:39.4116821Z  python3 -mpip install torch==2.7.1 2025-11-03T17:41:39.4117071Z elif command -v rocminfo; then 2025-11-03T17:41:39.4117378Z  # NB: Installing torch on ROCm runner with pip here causes CI to fail 2025-11-03T17:41:39.4117785Z  # with a memoryview is too large error only on MI300 runners. Is pip 2025-11-03T17:41:39.4118177Z  # version on ROCm runner there too old? As a workaround, let's use the 2025-11-03T17:41:39.4118527Z  # GPU device name coming from rocminfo instead 2025-11-03T17:41:39.4118787Z  DEVICE_NAME=rocm 2025-11-03T17:41:39.4119133Z  DEVICE_TYPE=$(rocminfo | grep "Marketing Name" | tail -n1 | awk -F':' '{print $2}' | xargs) 2025-11-03T17:41:39.4119484Z fi 2025-11-03T17:41:39.4119640Z  2025-11-03T17:41:39.4119832Z echo "DEVICE_NAME=$DEVICE_NAME" >> $GITHUB_ENV 2025-11-03T17:41:39.4120200Z echo "DEVICE_TYPE=$DEVICE_TYPE" >> $GITHUB_ENV 2025-11-03T17:41:39.4132863Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2025-11-03T17:41:39.4133144Z env: 2025-11-03T17:41:39.4133313Z GIT_DEFAULT_BRANCH: main 2025-11-03T17:41:39.4133553Z GPU_FLAG: --gpus all -e NVIDIA_DRIVER_CAPABILITIES=all 2025-11-03T17:41:39.4133963Z DOCKER_CONTAINER_ID: 99e00294b4bd7ffe64017e12af3ca26302f354938a9d519b091e205874b31030 2025-11-03T17:41:39.4134322Z ##[endgroup] 2025-11-03T17:41:39.4175597Z + [[ -n '' ]] 2025-11-03T17:41:39.4175939Z + python3 -mpip install boto3==1.35.33 psutil==7.0.0 pynvml==12.0.0 2025-11-03T17:41:39.6364658Z Defaulting to user installation because normal site-packages is not writeable 2025-11-03T17:41:40.6864211Z Collecting boto3==1.35.33 2025-11-03T17:41:40.7034654Z Downloading boto3-1.35.33-py3-none-any.whl (139 kB) 2025-11-03T17:41:41.0184190Z Collecting psutil==7.0.0 2025-11-03T17:41:41.0226151Z Downloading psutil-7.0.0-cp36-abi3-manylinux_2_12_x86_64.manylinux2010_x86_64.manylinux_2_17_x86_64.manylinux2014_x86_64.whl (277 kB) 2025-11-03T17:41:41.0545628Z Collecting pynvml==12.0.0 2025-11-03T17:41:41.0589872Z Downloading pynvml-12.0.0-py3-none-any.whl (26 kB) 2025-11-03T17:41:42.1695726Z Collecting botocore<1.36.0,>=1.35.33 2025-11-03T17:41:42.1735787Z Downloading botocore-1.35.99-py3-none-any.whl (13.3 MB) 2025-11-03T17:41:42.3051631Z Requirement already satisfied: jmespath<2.0.0,>=0.7.1 in /usr/lib/python3.9/site-packages (from boto3==1.35.33) (0.10.0) 2025-11-03T17:41:42.3430488Z Collecting s3transfer<0.11.0,>=0.10.0 2025-11-03T17:41:42.3473412Z Downloading s3transfer-0.10.4-py3-none-any.whl (83 kB) 2025-11-03T17:41:42.3921571Z Collecting nvidia-ml-py<13.0.0a0,>=12.0.0 2025-11-03T17:41:42.3970389Z Downloading nvidia_ml_py-12.575.51-py3-none-any.whl (47 kB) 2025-11-03T17:41:42.4057898Z Requirement already satisfied: urllib3<1.27,>=1.25.4 in /usr/lib/python3.9/site-packages (from botocore<1.36.0,>=1.35.33->boto3==1.35.33) (1.25.10) 2025-11-03T17:41:42.4062775Z Requirement already satisfied: python-dateutil<3.0.0,>=2.1 in /usr/lib/python3.9/site-packages (from botocore<1.36.0,>=1.35.33->boto3==1.35.33) (2.8.1) 2025-11-03T17:41:42.5562613Z Requirement already satisfied: six>=1.5 in /usr/lib/python3.9/site-packages (from python-dateutil<3.0.0,>=2.1->botocore<1.36.0,>=1.35.33->boto3==1.35.33) (1.15.0) 2025-11-03T17:41:42.6685050Z Installing collected packages: botocore, s3transfer, nvidia-ml-py, pynvml, psutil, boto3 2025-11-03T17:41:43.1942767Z Attempting uninstall: nvidia-ml-py 2025-11-03T17:41:43.1945803Z Found existing installation: nvidia-ml-py 11.525.84 2025-11-03T17:41:43.1959272Z Uninstalling nvidia-ml-py-11.525.84: 2025-11-03T17:41:43.2178242Z Successfully uninstalled nvidia-ml-py-11.525.84 2025-11-03T17:41:43.2660618Z Attempting uninstall: psutil 2025-11-03T17:41:43.2661582Z Found existing installation: psutil 5.9.8 2025-11-03T17:41:43.2737853Z Uninstalling psutil-5.9.8: 2025-11-03T17:41:43.2744438Z Successfully uninstalled psutil-5.9.8 2025-11-03T17:41:43.4243159Z Successfully installed boto3-1.35.33 botocore-1.35.99 nvidia-ml-py-12.575.51 psutil-7.0.0 pynvml-12.0.0 s3transfer-0.10.4 2025-11-03T17:41:43.5162692Z + DEVICE_NAME= 2025-11-03T17:41:43.5163167Z + DEVICE_TYPE= 2025-11-03T17:41:43.5163571Z + command -v nvidia-smi 2025-11-03T17:41:43.5164082Z + python3 -mpip install torch==2.7.1 2025-11-03T17:41:43.5164635Z /usr/bin/nvidia-smi 2025-11-03T17:41:43.7330024Z Defaulting to user installation because normal site-packages is not writeable 2025-11-03T17:41:43.9894854Z Collecting torch==2.7.1 2025-11-03T17:41:44.0103221Z Downloading torch-2.7.1-cp39-cp39-manylinux_2_28_x86_64.whl (821.1 MB) 2025-11-03T17:41:56.1176416Z Collecting nvidia-cufile-cu12==1.11.1.6 2025-11-03T17:41:56.1268915Z Downloading nvidia_cufile_cu12-1.11.1.6-py3-none-manylinux2014_x86_64.manylinux_2_17_x86_64.whl (1.1 MB) 2025-11-03T17:41:56.2030404Z Collecting networkx 2025-11-03T17:41:56.2076991Z Downloading networkx-3.2.1-py3-none-any.whl (1.6 MB) 2025-11-03T17:41:56.2309989Z Requirement already satisfied: jinja2 in /usr/lib/python3.9/site-packages (from torch==2.7.1) (2.11.3) 2025-11-03T17:41:56.2605359Z Collecting nvidia-cuda-nvrtc-cu12==12.6.77 2025-11-03T17:41:56.2687550Z Downloading nvidia_cuda_nvrtc_cu12-12.6.77-py3-none-manylinux2014_x86_64.whl (23.7 MB) 2025-11-03T17:41:56.5108402Z Collecting nvidia-cusolver-cu12==11.7.1.2 2025-11-03T17:41:56.5186063Z Downloading nvidia_cusolver_cu12-11.7.1.2-py3-none-manylinux2014_x86_64.manylinux_2_17_x86_64.whl (158.2 MB) 2025-11-03T17:41:58.1082638Z Collecting nvidia-nvjitlink-cu12==12.6.85 2025-11-03T17:41:58.1178783Z Downloading nvidia_nvjitlink_cu12-12.6.85-py3-none-manylinux2010_x86_64.manylinux_2_12_x86_64.whl (19.7 MB) 2025-11-03T17:41:58.3141054Z Collecting nvidia-cudnn-cu12==9.5.1.17 2025-11-03T17:41:58.3217950Z Downloading nvidia_cudnn_cu12-9.5.1.17-py3-none-manylinux_2_28_x86_64.whl (571.0 MB) 2025-11-03T17:42:06.3765637Z Collecting nvidia-cusparselt-cu12==0.6.3 2025-11-03T17:42:06.3852614Z Downloading nvidia_cusparselt_cu12-0.6.3-py3-none-manylinux2014_x86_64.whl (156.8 MB) 2025-11-03T17:42:08.0015572Z Collecting fsspec 2025-11-03T17:42:08.0054406Z Downloading fsspec-2025.10.0-py3-none-any.whl (200 kB) 2025-11-03T17:42:08.0365944Z Collecting nvidia-cublas-cu12==12.6.4.1 2025-11-03T17:42:08.0448950Z Downloading nvidia_cublas_cu12-12.6.4.1-py3-none-manylinux2014_x86_64.manylinux_2_17_x86_64.whl (393.1 MB) 2025-11-03T17:42:13.5394267Z Collecting triton==3.3.1 2025-11-03T17:42:13.5474851Z Downloading triton-3.3.1-cp39-cp39-manylinux_2_27_x86_64.manylinux_2_28_x86_64.whl (155.6 MB) 2025-11-03T17:42:15.1233827Z Collecting nvidia-cuda-runtime-cu12==12.6.77 2025-11-03T17:42:15.1304981Z Downloading nvidia_cuda_runtime_cu12-12.6.77-py3-none-manylinux2014_x86_64.manylinux_2_17_x86_64.whl (897 kB) 2025-11-03T17:42:15.1695697Z Collecting nvidia-cufft-cu12==11.3.0.4 2025-11-03T17:42:15.1774537Z Downloading nvidia_cufft_cu12-11.3.0.4-py3-none-manylinux2014_x86_64.manylinux_2_17_x86_64.whl (200.2 MB) 2025-11-03T17:42:17.4331887Z Collecting sympy>=1.13.3 2025-11-03T17:42:17.4394973Z Downloading sympy-1.14.0-py3-none-any.whl (6.3 MB) 2025-11-03T17:42:17.5218015Z Collecting nvidia-nccl-cu12==2.26.2 2025-11-03T17:42:17.5299774Z Downloading nvidia_nccl_cu12-2.26.2-py3-none-manylinux2014_x86_64.manylinux_2_17_x86_64.whl (201.3 MB) 2025-11-03T17:42:19.9189598Z Collecting filelock 2025-11-03T17:42:19.9232314Z Downloading filelock-3.19.1-py3-none-any.whl (15 kB) 2025-11-03T17:42:19.9284364Z Requirement already satisfied: typing-extensions>=4.10.0 in /home/ec2-user/.local/lib/python3.9/site-packages (from torch==2.7.1) (4.15.0) 2025-11-03T17:42:19.9616900Z Collecting nvidia-cuda-cupti-cu12==12.6.80 2025-11-03T17:42:19.9689530Z Downloading nvidia_cuda_cupti_cu12-12.6.80-py3-none-manylinux2014_x86_64.manylinux_2_17_x86_64.whl (8.9 MB) 2025-11-03T17:42:20.0687474Z Collecting nvidia-cusparse-cu12==12.5.4.2 2025-11-03T17:42:20.0760618Z Downloading nvidia_cusparse_cu12-12.5.4.2-py3-none-manylinux2014_x86_64.manylinux_2_17_x86_64.whl (216.6 MB) 2025-11-03T17:42:22.5294625Z Collecting nvidia-curand-cu12==10.3.7.77 2025-11-03T17:42:22.5381522Z Downloading nvidia_curand_cu12-10.3.7.77-py3-none-manylinux2014_x86_64.manylinux_2_17_x86_64.whl (56.3 MB) 2025-11-03T17:42:23.1157698Z Collecting nvidia-nvtx-cu12==12.6.77 2025-11-03T17:42:23.1209180Z Downloading nvidia_nvtx_cu12-12.6.77-py3-none-manylinux2014_x86_64.manylinux_2_17_x86_64.whl (89 kB) 2025-11-03T17:42:23.1499710Z Requirement already satisfied: setuptools>=40.8.0 in /usr/lib/python3.9/site-packages (from triton==3.3.1->torch==2.7.1) (59.6.0) 2025-11-03T17:42:23.1769672Z Collecting mpmath<1.4,>=1.1.0 2025-11-03T17:42:23.1812222Z Downloading mpmath-1.3.0-py3-none-any.whl (536 kB) 2025-11-03T17:42:23.2595260Z Requirement already satisfied: MarkupSafe>=0.23 in /usr/lib64/python3.9/site-packages (from jinja2->torch==2.7.1) (1.1.1) 2025-11-03T17:42:23.5610473Z Installing collected packages: nvidia-nvjitlink-cu12, nvidia-cusparse-cu12, nvidia-cublas-cu12, mpmath, triton, sympy, nvidia-nvtx-cu12, nvidia-nccl-cu12, nvidia-cusparselt-cu12, nvidia-cusolver-cu12, nvidia-curand-cu12, nvidia-cufile-cu12, nvidia-cufft-cu12, nvidia-cudnn-cu12, nvidia-cuda-runtime-cu12, nvidia-cuda-nvrtc-cu12, nvidia-cuda-cupti-cu12, networkx, fsspec, filelock, torch 2025-11-03T17:42:31.2380276Z WARNING: The scripts proton and proton-viewer are installed in '/home/ec2-user/.local/bin' which is not on PATH. 2025-11-03T17:42:31.2384358Z Consider adding this directory to PATH or, if you prefer to suppress this warning, use --no-warn-script-location. 2025-11-03T17:42:34.7061356Z WARNING: The script isympy is installed in '/home/ec2-user/.local/bin' which is not on PATH. 2025-11-03T17:42:34.7062832Z Consider adding this directory to PATH or, if you prefer to suppress this warning, use --no-warn-script-location. 2025-11-03T17:43:01.4602580Z WARNING: The scripts torchfrtrace and torchrun are installed in '/home/ec2-user/.local/bin' which is not on PATH. 2025-11-03T17:43:01.4604322Z Consider adding this directory to PATH or, if you prefer to suppress this warning, use --no-warn-script-location. 2025-11-03T17:43:01.6923073Z Successfully installed filelock-3.19.1 fsspec-2025.10.0 mpmath-1.3.0 networkx-3.2.1 nvidia-cublas-cu12-12.6.4.1 nvidia-cuda-cupti-cu12-12.6.80 nvidia-cuda-nvrtc-cu12-12.6.77 nvidia-cuda-runtime-cu12-12.6.77 nvidia-cudnn-cu12-9.5.1.17 nvidia-cufft-cu12-11.3.0.4 nvidia-cufile-cu12-1.11.1.6 nvidia-curand-cu12-10.3.7.77 nvidia-cusolver-cu12-11.7.1.2 nvidia-cusparse-cu12-12.5.4.2 nvidia-cusparselt-cu12-0.6.3 nvidia-nccl-cu12-2.26.2 nvidia-nvjitlink-cu12-12.6.85 nvidia-nvtx-cu12-12.6.77 sympy-1.14.0 torch-2.7.1 triton-3.3.1 2025-11-03T17:43:02.2022851Z + echo DEVICE_NAME= 2025-11-03T17:43:02.2023358Z + echo DEVICE_TYPE= 2025-11-03T17:43:02.2048934Z ##[group]Run set -eux 2025-11-03T17:43:02.2049155Z set -eux 2025-11-03T17:43:02.2049381Z  2025-11-03T17:43:02.2049564Z if [[ -z "${GITHUB_TOKEN}" ]]; then 2025-11-03T17:43:02.2049832Z  echo "Missing github-token input" 2025-11-03T17:43:02.2050078Z  exit 1 2025-11-03T17:43:02.2050244Z fi 2025-11-03T17:43:02.2060004Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2025-11-03T17:43:02.2060294Z env: 2025-11-03T17:43:02.2060457Z GIT_DEFAULT_BRANCH: main 2025-11-03T17:43:02.2060710Z GPU_FLAG: --gpus all -e NVIDIA_DRIVER_CAPABILITIES=all 2025-11-03T17:43:02.2061137Z DOCKER_CONTAINER_ID: 99e00294b4bd7ffe64017e12af3ca26302f354938a9d519b091e205874b31030 2025-11-03T17:43:02.2061494Z DEVICE_NAME: 2025-11-03T17:43:02.2061813Z DEVICE_TYPE: 2025-11-03T17:43:02.2062177Z GITHUB_TOKEN: *** 2025-11-03T17:43:02.2062353Z ##[endgroup] 2025-11-03T17:43:02.2096765Z + [[ -z *** ]] 2025-11-03T17:43:02.2253889Z ##[group]Run pytorch/test-infra/.github/actions/get-workflow-job-id@main 2025-11-03T17:43:02.2254220Z with: 2025-11-03T17:43:02.2254503Z github-token: *** 2025-11-03T17:43:02.2254691Z env: 2025-11-03T17:43:02.2254859Z GIT_DEFAULT_BRANCH: main 2025-11-03T17:43:02.2255111Z GPU_FLAG: --gpus all -e NVIDIA_DRIVER_CAPABILITIES=all 2025-11-03T17:43:02.2255519Z DOCKER_CONTAINER_ID: 99e00294b4bd7ffe64017e12af3ca26302f354938a9d519b091e205874b31030 2025-11-03T17:43:02.2255881Z DEVICE_NAME: 2025-11-03T17:43:02.2256048Z DEVICE_TYPE: 2025-11-03T17:43:02.2256203Z ##[endgroup] 2025-11-03T17:43:02.2392105Z ##[group]Run set -eux 2025-11-03T17:43:02.2392299Z set -eux 2025-11-03T17:43:02.2392476Z  2025-11-03T17:43:02.2392820Z python3 "${GITHUB_ACTION_PATH}/../../scripts/get_workflow_job_id.py" "${GITHUB_RUN_ID}" "${RUNNER_NAME}" 2025-11-03T17:43:02.2399958Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2025-11-03T17:43:02.2400242Z env: 2025-11-03T17:43:02.2400404Z GIT_DEFAULT_BRANCH: main 2025-11-03T17:43:02.2400655Z GPU_FLAG: --gpus all -e NVIDIA_DRIVER_CAPABILITIES=all 2025-11-03T17:43:02.2401051Z DOCKER_CONTAINER_ID: 99e00294b4bd7ffe64017e12af3ca26302f354938a9d519b091e205874b31030 2025-11-03T17:43:02.2401538Z DEVICE_NAME: 2025-11-03T17:43:02.2401703Z DEVICE_TYPE: 2025-11-03T17:43:02.2401982Z GITHUB_TOKEN: *** 2025-11-03T17:43:02.2402158Z ##[endgroup] 2025-11-03T17:43:02.2437015Z + python3 /home/ec2-user/actions-runner/_work/_actions/pytorch/test-infra/main/.github/actions/get-workflow-job-id/../../scripts/get_workflow_job_id.py 19040285420 i-05b4d4af410353347 2025-11-03T17:43:02.9868313Z setting job-id=54383051348 2025-11-03T17:43:02.9869438Z setting job-name=ci-vllm-test / test (vllm_distributed_test_28_failure_test, 1, 1, linux.g6.12xlarge.nvidia.gpu) 2025-11-03T17:43:03.0129258Z ##[group]Run set -eux 2025-11-03T17:43:03.0129550Z set -eux 2025-11-03T17:43:03.0129715Z  2025-11-03T17:43:03.0129874Z if [[ -n "" ]]; then 2025-11-03T17:43:03.0130102Z  source "" 2025-11-03T17:43:03.0130356Z fi 2025-11-03T17:43:03.0130510Z  2025-11-03T17:43:03.0130796Z python3 "${GITHUB_ACTION_PATH}/../../scripts/benchmarks/gather_metadata.py" \ 2025-11-03T17:43:03.0131183Z  --schema-version "${SCHEMA_VERSION}" \ 2025-11-03T17:43:03.0131423Z  --repo "${REPO}" \ 2025-11-03T17:43:03.0131640Z  --head-branch "${HEAD_BRANCH}" \ 2025-11-03T17:43:03.0131877Z  --head-sha "${HEAD_SHA}" \ 2025-11-03T17:43:03.0132121Z  --workflow-id "${WORKFLOW_RUN_ID}" \ 2025-11-03T17:43:03.0132584Z  --run-attempt "${RUN_ATTEMPT}" \ 2025-11-03T17:43:03.0132845Z  --job-id "${JOB_ID}" \ 2025-11-03T17:43:03.0133061Z  --job-name "${JOB_NAME}" 2025-11-03T17:43:03.0141844Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2025-11-03T17:43:03.0142125Z env: 2025-11-03T17:43:03.0142280Z GIT_DEFAULT_BRANCH: main 2025-11-03T17:43:03.0142520Z GPU_FLAG: --gpus all -e NVIDIA_DRIVER_CAPABILITIES=all 2025-11-03T17:43:03.0142942Z DOCKER_CONTAINER_ID: 99e00294b4bd7ffe64017e12af3ca26302f354938a9d519b091e205874b31030 2025-11-03T17:43:03.0143288Z DEVICE_NAME: 2025-11-03T17:43:03.0143444Z DEVICE_TYPE: 2025-11-03T17:43:03.0143613Z SCHEMA_VERSION: v3 2025-11-03T17:43:03.0143800Z REPO: pytorch/pytorch 2025-11-03T17:43:03.0143988Z HEAD_BRANCH: refs/heads/main 2025-11-03T17:43:03.0144217Z HEAD_SHA: 3f6538febd727b782e6e13cfd026a309fb14351d 2025-11-03T17:43:03.0144469Z WORKFLOW_RUN_ID: 19040285420 2025-11-03T17:43:03.0144659Z RUN_ATTEMPT: 1 2025-11-03T17:43:03.0144823Z JOB_ID: 54383051348 2025-11-03T17:43:03.0145175Z JOB_NAME: ci-vllm-test / test (vllm_distributed_test_28_failure_test, 1, 1, linux.g6.12xlarge.nvidia.gpu) 2025-11-03T17:43:03.0145678Z ##[endgroup] 2025-11-03T17:43:03.0179545Z + [[ -n '' ]] 2025-11-03T17:43:03.0181367Z + python3 /home/ec2-user/actions-runner/_work/_actions/pytorch/test-infra/main/.github/actions/upload-benchmark-results/../../scripts/benchmarks/gather_metadata.py --schema-version v3 --repo pytorch/pytorch --head-branch refs/heads/main --head-sha 3f6538febd727b782e6e13cfd026a309fb14351d --workflow-id 19040285420 --run-attempt 1 --job-id 54383051348 --job-name 'ci-vllm-test / test (vllm_distributed_test_28_failure_test, 1, 1, linux.g6.12xlarge.nvidia.gpu)' 2025-11-03T17:43:03.1030731Z ##[group]Run set -eux 2025-11-03T17:43:03.1030929Z set -eux 2025-11-03T17:43:03.1031081Z  2025-11-03T17:43:03.1031235Z if [[ -n "" ]]; then 2025-11-03T17:43:03.1031425Z  source "" 2025-11-03T17:43:03.1031585Z fi 2025-11-03T17:43:03.1031743Z  2025-11-03T17:43:03.1032024Z python3 "${GITHUB_ACTION_PATH}/../../scripts/benchmarks/gather_runners_info.py" 2025-11-03T17:43:03.1040031Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2025-11-03T17:43:03.1040301Z env: 2025-11-03T17:43:03.1040455Z GIT_DEFAULT_BRANCH: main 2025-11-03T17:43:03.1040694Z GPU_FLAG: --gpus all -e NVIDIA_DRIVER_CAPABILITIES=all 2025-11-03T17:43:03.1041086Z DOCKER_CONTAINER_ID: 99e00294b4bd7ffe64017e12af3ca26302f354938a9d519b091e205874b31030 2025-11-03T17:43:03.1041430Z DEVICE_NAME: 2025-11-03T17:43:03.1041592Z DEVICE_TYPE: 2025-11-03T17:43:03.1041869Z ##[endgroup] 2025-11-03T17:43:03.1073518Z + [[ -n '' ]] 2025-11-03T17:43:03.1074984Z + python3 /home/ec2-user/actions-runner/_work/_actions/pytorch/test-infra/main/.github/actions/upload-benchmark-results/../../scripts/benchmarks/gather_runners_info.py 2025-11-03T17:43:03.9997332Z /home/ec2-user/.local/lib/python3.9/site-packages/torch/_subclasses/functional_tensor.py:276: UserWarning: Failed to initialize NumPy: No module named 'numpy' (Triggered internally at /pytorch/torch/csrc/utils/tensor_numpy.cpp:81.) 2025-11-03T17:43:03.9999153Z cpu = _conversion_method_template(device=torch.device("cpu")) 2025-11-03T17:43:05.3774363Z ##[group]Run set -eux 2025-11-03T17:43:05.3774621Z set -eux 2025-11-03T17:43:05.3774839Z  2025-11-03T17:43:05.3775063Z # TODO (huydhn): Implement this part 2025-11-03T17:43:05.3775384Z echo "dependencies={}" >> "${GITHUB_OUTPUT}" 2025-11-03T17:43:05.3783968Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2025-11-03T17:43:05.3784297Z env: 2025-11-03T17:43:05.3784506Z GIT_DEFAULT_BRANCH: main 2025-11-03T17:43:05.3784823Z GPU_FLAG: --gpus all -e NVIDIA_DRIVER_CAPABILITIES=all 2025-11-03T17:43:05.3795846Z DOCKER_CONTAINER_ID: 99e00294b4bd7ffe64017e12af3ca26302f354938a9d519b091e205874b31030 2025-11-03T17:43:05.3796295Z DEVICE_NAME: 2025-11-03T17:43:05.3796503Z DEVICE_TYPE: 2025-11-03T17:43:05.3796950Z ##[endgroup] 2025-11-03T17:43:05.3831577Z + echo 'dependencies={}' 2025-11-03T17:43:05.3927614Z ##[group]Run set -eux 2025-11-03T17:43:05.3927860Z set -eux 2025-11-03T17:43:05.3928023Z  2025-11-03T17:43:05.3928182Z if [[ -n "" ]]; then 2025-11-03T17:43:05.3928462Z  source "" 2025-11-03T17:43:05.3928640Z fi 2025-11-03T17:43:05.3928797Z  2025-11-03T17:43:05.3928989Z if [[ ! -d "${BENCHMARK_RESULTS_DIR}" ]]; then 2025-11-03T17:43:05.3929309Z  echo "${BENCHMARK_RESULTS_DIR} does not exist, skipping" 2025-11-03T17:43:05.3929666Z  # We don't want the job to fail if the directory doesn't exist 2025-11-03T17:43:05.3929967Z  exit 0 2025-11-03T17:43:05.3930137Z fi 2025-11-03T17:43:05.3930284Z  2025-11-03T17:43:05.3930458Z if [[ "${DRY_RUN}" == "true" ]]; then 2025-11-03T17:43:05.3930806Z  python3 "${GITHUB_ACTION_PATH}/../../scripts/upload_benchmark_results.py" \ 2025-11-03T17:43:05.3931242Z  --benchmark-results-dir "${BENCHMARK_RESULTS_DIR}" \ 2025-11-03T17:43:05.3931675Z  --metadata "${BENCHMARK_METADATA}" \ 2025-11-03T17:43:05.3931917Z  --runners "${RUNNER_INFO}" \ 2025-11-03T17:43:05.3932156Z  --dependencies "${DEPENDENCIES}" \ 2025-11-03T17:43:05.3932385Z  --dry-run 2025-11-03T17:43:05.3932570Z else 2025-11-03T17:43:05.3932842Z  python3 "${GITHUB_ACTION_PATH}/../../scripts/upload_benchmark_results.py" \ 2025-11-03T17:43:05.3933234Z  --benchmark-results-dir "${BENCHMARK_RESULTS_DIR}" \ 2025-11-03T17:43:05.3933539Z  --metadata "${BENCHMARK_METADATA}" \ 2025-11-03T17:43:05.3933785Z  --runners "${RUNNER_INFO}" \ 2025-11-03T17:43:05.3934027Z  --dependencies "${DEPENDENCIES}" 2025-11-03T17:43:05.3934242Z fi 2025-11-03T17:43:05.3942121Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2025-11-03T17:43:05.3942402Z env: 2025-11-03T17:43:05.3942559Z GIT_DEFAULT_BRANCH: main 2025-11-03T17:43:05.3942796Z GPU_FLAG: --gpus all -e NVIDIA_DRIVER_CAPABILITIES=all 2025-11-03T17:43:05.3943198Z DOCKER_CONTAINER_ID: 99e00294b4bd7ffe64017e12af3ca26302f354938a9d519b091e205874b31030 2025-11-03T17:43:05.3943548Z DEVICE_NAME: 2025-11-03T17:43:05.3943716Z DEVICE_TYPE: 2025-11-03T17:43:05.3943900Z BENCHMARK_RESULTS_DIR: test/test-reports 2025-11-03T17:43:05.3944123Z DRY_RUN: false 2025-11-03T17:43:05.3945081Z BENCHMARK_METADATA: {"timestamp": 1762191783, "schema_version": "v3", "name": "ci-vllm-test / test (vllm_distributed_test_28_failure_test, 1, 1, linux.g6.12xlarge.nvidia.gpu)", "repo": "pytorch/pytorch", "head_branch": "refs/heads/main", "head_sha": "3f6538febd727b782e6e13cfd026a309fb14351d", "workflow_id": 19040285420, "run_attempt": 1, "job_id": 54383051348} 2025-11-03T17:43:05.3946522Z RUNNER_INFO: [{"cpu_info": "x86_64", "cpu_count": 48, "avail_mem_in_gb": 181, "extra_info": {"hostname": "ip-10-0-8-115.ec2.internal"}, "name": "cuda", "type": "NVIDIA L4", "gpu_count": 4, "avail_gpu_mem_in_gb": 88}] 2025-11-03T17:43:05.3947063Z DEPENDENCIES: {} 2025-11-03T17:43:05.3947235Z ##[endgroup] 2025-11-03T17:43:05.3982049Z + [[ -n '' ]] 2025-11-03T17:43:05.3982585Z test/test-reports does not exist, skipping 2025-11-03T17:43:05.3983269Z + [[ ! -d test/test-reports ]] 2025-11-03T17:43:05.3983937Z + echo 'test/test-reports does not exist, skipping' 2025-11-03T17:43:05.3984514Z + exit 0 2025-11-03T17:43:05.4105787Z ##[group]Run cat test/**/*_toprint.log || true 2025-11-03T17:43:05.4106106Z cat test/**/*_toprint.log || true 2025-11-03T17:43:05.4113480Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2025-11-03T17:43:05.4113755Z env: 2025-11-03T17:43:05.4113915Z GIT_DEFAULT_BRANCH: main 2025-11-03T17:43:05.4114154Z GPU_FLAG: --gpus all -e NVIDIA_DRIVER_CAPABILITIES=all 2025-11-03T17:43:05.4114537Z DOCKER_CONTAINER_ID: 99e00294b4bd7ffe64017e12af3ca26302f354938a9d519b091e205874b31030 2025-11-03T17:43:05.4114892Z DEVICE_NAME: 2025-11-03T17:43:05.4115056Z DEVICE_TYPE: 2025-11-03T17:43:05.4115213Z ##[endgroup] 2025-11-03T17:43:05.4218205Z cat: 'test/**/*_toprint.log': No such file or directory 2025-11-03T17:43:05.4465334Z ##[group]Run kill "$MONITOR_SCRIPT_PID" 2025-11-03T17:43:05.4465654Z kill "$MONITOR_SCRIPT_PID" 2025-11-03T17:43:05.4473216Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2025-11-03T17:43:05.4473502Z env: 2025-11-03T17:43:05.4473665Z GIT_DEFAULT_BRANCH: main 2025-11-03T17:43:05.4473912Z GPU_FLAG: --gpus all -e NVIDIA_DRIVER_CAPABILITIES=all 2025-11-03T17:43:05.4474316Z DOCKER_CONTAINER_ID: 99e00294b4bd7ffe64017e12af3ca26302f354938a9d519b091e205874b31030 2025-11-03T17:43:05.4474675Z DEVICE_NAME: 2025-11-03T17:43:05.4474853Z DEVICE_TYPE: 2025-11-03T17:43:05.4475040Z MONITOR_SCRIPT_PID: 67419 2025-11-03T17:43:05.4475235Z ##[endgroup] 2025-11-03T17:43:05.4591914Z Prepare all required actions 2025-11-03T17:43:05.4592288Z Getting action download info 2025-11-03T17:43:05.6383712Z Download action repository 'seemethere/upload-artifact-s3@v5' (SHA:baba72d0712b404f646cebe0730933554ebce96a) 2025-11-03T17:43:06.7735344Z Download action repository 'actions/upload-artifact@v4' (SHA:ea165f8d65b6e75b540449e92b4886f43607fa02) 2025-11-03T17:43:11.3206901Z ##[group]Run ./.github/actions/upload-test-artifacts 2025-11-03T17:43:11.3207160Z with: 2025-11-03T17:43:11.3207509Z file-suffix: test-vllm_distributed_test_28_failure_test-1-1-linux.g6.12xlarge.nvidia.gpu_54383051348 2025-11-03T17:43:11.3207911Z s3-bucket: gha-artifacts 2025-11-03T17:43:11.3208118Z env: 2025-11-03T17:43:11.3208266Z GIT_DEFAULT_BRANCH: main 2025-11-03T17:43:11.3208598Z GPU_FLAG: --gpus all -e NVIDIA_DRIVER_CAPABILITIES=all 2025-11-03T17:43:11.3208993Z DOCKER_CONTAINER_ID: 99e00294b4bd7ffe64017e12af3ca26302f354938a9d519b091e205874b31030 2025-11-03T17:43:11.3209368Z DEVICE_NAME: 2025-11-03T17:43:11.3209834Z DEVICE_TYPE: 2025-11-03T17:43:11.3210060Z ##[endgroup] 2025-11-03T17:43:11.3326379Z ##[group]Run # Remove any previous test jsons if they exist 2025-11-03T17:43:11.3326717Z # Remove any previous test jsons if they exist 2025-11-03T17:43:11.3326995Z rm -f test-jsons-*.zip 2025-11-03T17:43:11.3327306Z zip -r "test-jsons-${FILE_SUFFIX}.zip" test/test-reports -i '*.json' 2025-11-03T17:43:11.3336601Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2025-11-03T17:43:11.3336886Z env: 2025-11-03T17:43:11.3337045Z GIT_DEFAULT_BRANCH: main 2025-11-03T17:43:11.3337293Z GPU_FLAG: --gpus all -e NVIDIA_DRIVER_CAPABILITIES=all 2025-11-03T17:43:11.3337810Z DOCKER_CONTAINER_ID: 99e00294b4bd7ffe64017e12af3ca26302f354938a9d519b091e205874b31030 2025-11-03T17:43:11.3338158Z DEVICE_NAME: 2025-11-03T17:43:11.3338324Z DEVICE_TYPE: 2025-11-03T17:43:11.3338671Z FILE_SUFFIX: test-vllm_distributed_test_28_failure_test-1-1-linux.g6.12xlarge.nvidia.gpu_54383051348 2025-11-03T17:43:11.3339056Z ##[endgroup] 2025-11-03T17:43:11.4028962Z zip warning: name not matched: test/test-reports 2025-11-03T17:43:11.4030986Z zip warning: zip file empty 2025-11-03T17:43:11.4165542Z ##[group]Run # Remove any previous test reports if they exist 2025-11-03T17:43:11.4165886Z # Remove any previous test reports if they exist 2025-11-03T17:43:11.4166158Z rm -f test-reports-*.zip 2025-11-03T17:43:11.4166493Z zip -r "test-reports-${FILE_SUFFIX}.zip" test/test-reports -i '*.xml' -i '*.csv' 2025-11-03T17:43:11.4174336Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2025-11-03T17:43:11.4174611Z env: 2025-11-03T17:43:11.4174776Z GIT_DEFAULT_BRANCH: main 2025-11-03T17:43:11.4175025Z GPU_FLAG: --gpus all -e NVIDIA_DRIVER_CAPABILITIES=all 2025-11-03T17:43:11.4175414Z DOCKER_CONTAINER_ID: 99e00294b4bd7ffe64017e12af3ca26302f354938a9d519b091e205874b31030 2025-11-03T17:43:11.4175760Z DEVICE_NAME: 2025-11-03T17:43:11.4175934Z DEVICE_TYPE: 2025-11-03T17:43:11.4176278Z FILE_SUFFIX: test-vllm_distributed_test_28_failure_test-1-1-linux.g6.12xlarge.nvidia.gpu_54383051348 2025-11-03T17:43:11.4176657Z ##[endgroup] 2025-11-03T17:43:11.4279585Z zip warning: name not matched: test/test-reports 2025-11-03T17:43:11.4280571Z zip warning: zip file empty 2025-11-03T17:43:11.4402989Z ##[group]Run # Remove any previous usage logs if they exist 2025-11-03T17:43:11.4403332Z # Remove any previous usage logs if they exist 2025-11-03T17:43:11.4403603Z rm -f logs-*.zip 2025-11-03T17:43:11.4403933Z zip "logs-${FILE_SUFFIX}.zip" 'usage_log.txt' || true 2025-11-03T17:43:11.4404290Z zip -r "logs-${FILE_SUFFIX}.zip" test/test-reports -i '*.log' || true 2025-11-03T17:43:11.4412561Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2025-11-03T17:43:11.4412871Z env: 2025-11-03T17:43:11.4413037Z GIT_DEFAULT_BRANCH: main 2025-11-03T17:43:11.4413284Z GPU_FLAG: --gpus all -e NVIDIA_DRIVER_CAPABILITIES=all 2025-11-03T17:43:11.4413909Z DOCKER_CONTAINER_ID: 99e00294b4bd7ffe64017e12af3ca26302f354938a9d519b091e205874b31030 2025-11-03T17:43:11.4414284Z DEVICE_NAME: 2025-11-03T17:43:11.4414456Z DEVICE_TYPE: 2025-11-03T17:43:11.4414917Z FILE_SUFFIX: test-vllm_distributed_test_28_failure_test-1-1-linux.g6.12xlarge.nvidia.gpu_54383051348 2025-11-03T17:43:11.4415296Z ##[endgroup] 2025-11-03T17:43:11.4598429Z adding: usage_log.txt (deflated 96%) 2025-11-03T17:43:11.4617763Z zip warning: name not matched: test/test-reports 2025-11-03T17:43:11.4618039Z 2025-11-03T17:43:11.4618446Z zip error: Nothing to do! (logs-test-vllm_distributed_test_28_failure_test-1-1-linux.g6.12xlarge.nvidia.gpu_54383051348.zip) 2025-11-03T17:43:11.4675871Z ##[group]Run # Remove any previous debugging artifacts if they exist 2025-11-03T17:43:11.4676252Z # Remove any previous debugging artifacts if they exist 2025-11-03T17:43:11.4676550Z rm -f debug-*.zip 2025-11-03T17:43:11.4676779Z if [ -d 'test/debug' ]; then 2025-11-03T17:43:11.4677047Z  zip -r "debug-${FILE_SUFFIX}.zip" test/debug 2025-11-03T17:43:11.4677282Z fi 2025-11-03T17:43:11.4684271Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2025-11-03T17:43:11.4684571Z env: 2025-11-03T17:43:11.4684733Z GIT_DEFAULT_BRANCH: main 2025-11-03T17:43:11.4684978Z GPU_FLAG: --gpus all -e NVIDIA_DRIVER_CAPABILITIES=all 2025-11-03T17:43:11.4685364Z DOCKER_CONTAINER_ID: 99e00294b4bd7ffe64017e12af3ca26302f354938a9d519b091e205874b31030 2025-11-03T17:43:11.4685718Z DEVICE_NAME: 2025-11-03T17:43:11.4685886Z DEVICE_TYPE: 2025-11-03T17:43:11.4686232Z FILE_SUFFIX: test-vllm_distributed_test_28_failure_test-1-1-linux.g6.12xlarge.nvidia.gpu_54383051348 2025-11-03T17:43:11.4686735Z ##[endgroup] 2025-11-03T17:43:11.4808577Z ##[group]Run seemethere/upload-artifact-s3@v5 2025-11-03T17:43:11.4808840Z with: 2025-11-03T17:43:11.4809022Z s3-bucket: gha-artifacts 2025-11-03T17:43:11.4809283Z s3-prefix: pytorch/pytorch/19040285420/1/artifact 2025-11-03T17:43:11.4809755Z retention-days: 14 2025-11-03T17:43:11.4809947Z if-no-files-found: warn 2025-11-03T17:43:11.4810165Z path: test-jsons-*.zip 2025-11-03T17:43:11.4810368Z name: artifact 2025-11-03T17:43:11.4810540Z region: us-east-1 2025-11-03T17:43:11.4810747Z env: 2025-11-03T17:43:11.4810913Z GIT_DEFAULT_BRANCH: main 2025-11-03T17:43:11.4811219Z GPU_FLAG: --gpus all -e NVIDIA_DRIVER_CAPABILITIES=all 2025-11-03T17:43:11.4811639Z DOCKER_CONTAINER_ID: 99e00294b4bd7ffe64017e12af3ca26302f354938a9d519b091e205874b31030 2025-11-03T17:43:11.4811985Z DEVICE_NAME: 2025-11-03T17:43:11.4812153Z DEVICE_TYPE: 2025-11-03T17:43:11.4812312Z ##[endgroup] 2025-11-03T17:43:11.9398957Z NOTE: s3-prefix specified, ignoring name parameter 2025-11-03T17:43:11.9400038Z With the provided path, there will be 1 file uploaded 2025-11-03T17:43:11.9400837Z Uploading to s3 prefix: pytorch/pytorch/19040285420/1/artifact 2025-11-03T17:43:11.9467593Z Starting upload of test-jsons-test-vllm_distributed_test_28_failure_test-1-1-linux.g6.12xlarge.nvidia.gpu_54383051348.zip 2025-11-03T17:43:12.0648719Z Finished upload of test-jsons-test-vllm_distributed_test_28_failure_test-1-1-linux.g6.12xlarge.nvidia.gpu_54383051348.zip 2025-11-03T17:43:12.1008061Z ##[group]Run seemethere/upload-artifact-s3@v5 2025-11-03T17:43:12.1008308Z with: 2025-11-03T17:43:12.1008573Z s3-bucket: gha-artifacts 2025-11-03T17:43:12.1008817Z s3-prefix: pytorch/pytorch/19040285420/1/artifact 2025-11-03T17:43:12.1009072Z retention-days: 14 2025-11-03T17:43:12.1009260Z if-no-files-found: error 2025-11-03T17:43:12.1009634Z path: test-reports-*.zip 2025-11-03T17:43:12.1009826Z name: artifact 2025-11-03T17:43:12.1009993Z region: us-east-1 2025-11-03T17:43:12.1010163Z env: 2025-11-03T17:43:12.1010317Z GIT_DEFAULT_BRANCH: main 2025-11-03T17:43:12.1010561Z GPU_FLAG: --gpus all -e NVIDIA_DRIVER_CAPABILITIES=all 2025-11-03T17:43:12.1010960Z DOCKER_CONTAINER_ID: 99e00294b4bd7ffe64017e12af3ca26302f354938a9d519b091e205874b31030 2025-11-03T17:43:12.1012075Z DEVICE_NAME: 2025-11-03T17:43:12.1012279Z DEVICE_TYPE: 2025-11-03T17:43:12.1012445Z ##[endgroup] 2025-11-03T17:43:12.6565882Z NOTE: s3-prefix specified, ignoring name parameter 2025-11-03T17:43:12.6567020Z With the provided path, there will be 1 file uploaded 2025-11-03T17:43:12.6567750Z Uploading to s3 prefix: pytorch/pytorch/19040285420/1/artifact 2025-11-03T17:43:12.6632912Z Starting upload of test-reports-test-vllm_distributed_test_28_failure_test-1-1-linux.g6.12xlarge.nvidia.gpu_54383051348.zip 2025-11-03T17:43:12.7765856Z Finished upload of test-reports-test-vllm_distributed_test_28_failure_test-1-1-linux.g6.12xlarge.nvidia.gpu_54383051348.zip 2025-11-03T17:43:12.8125364Z ##[group]Run seemethere/upload-artifact-s3@v5 2025-11-03T17:43:12.8125609Z with: 2025-11-03T17:43:12.8125778Z s3-bucket: gha-artifacts 2025-11-03T17:43:12.8126016Z s3-prefix: pytorch/pytorch/19040285420/1/artifact 2025-11-03T17:43:12.8126269Z retention-days: 14 2025-11-03T17:43:12.8126462Z if-no-files-found: ignore 2025-11-03T17:43:12.8126658Z path: logs-*.zip 2025-11-03T17:43:12.8126826Z name: artifact 2025-11-03T17:43:12.8126992Z region: us-east-1 2025-11-03T17:43:12.8127148Z env: 2025-11-03T17:43:12.8127308Z GIT_DEFAULT_BRANCH: main 2025-11-03T17:43:12.8127566Z GPU_FLAG: --gpus all -e NVIDIA_DRIVER_CAPABILITIES=all 2025-11-03T17:43:12.8127959Z DOCKER_CONTAINER_ID: 99e00294b4bd7ffe64017e12af3ca26302f354938a9d519b091e205874b31030 2025-11-03T17:43:12.8128298Z DEVICE_NAME: 2025-11-03T17:43:12.8128566Z DEVICE_TYPE: 2025-11-03T17:43:12.8128724Z ##[endgroup] 2025-11-03T17:43:13.1174567Z NOTE: s3-prefix specified, ignoring name parameter 2025-11-03T17:43:13.1174982Z With the provided path, there will be 1 file uploaded 2025-11-03T17:43:13.1175648Z Uploading to s3 prefix: pytorch/pytorch/19040285420/1/artifact 2025-11-03T17:43:13.1242630Z Starting upload of logs-test-vllm_distributed_test_28_failure_test-1-1-linux.g6.12xlarge.nvidia.gpu_54383051348.zip 2025-11-03T17:43:13.2487489Z Finished upload of logs-test-vllm_distributed_test_28_failure_test-1-1-linux.g6.12xlarge.nvidia.gpu_54383051348.zip 2025-11-03T17:43:13.2855108Z ##[group]Run seemethere/upload-artifact-s3@v5 2025-11-03T17:43:13.2855355Z with: 2025-11-03T17:43:13.2855555Z s3-bucket: gha-artifacts 2025-11-03T17:43:13.2855800Z s3-prefix: pytorch/pytorch/19040285420/1/artifact 2025-11-03T17:43:13.2856069Z retention-days: 14 2025-11-03T17:43:13.2856265Z if-no-files-found: ignore 2025-11-03T17:43:13.2856461Z path: debug-*.zip 2025-11-03T17:43:13.2856650Z name: artifact 2025-11-03T17:43:13.2856821Z region: us-east-1 2025-11-03T17:43:13.2856982Z env: 2025-11-03T17:43:13.2857152Z GIT_DEFAULT_BRANCH: main 2025-11-03T17:43:13.2857444Z GPU_FLAG: --gpus all -e NVIDIA_DRIVER_CAPABILITIES=all 2025-11-03T17:43:13.2857921Z DOCKER_CONTAINER_ID: 99e00294b4bd7ffe64017e12af3ca26302f354938a9d519b091e205874b31030 2025-11-03T17:43:13.2858341Z DEVICE_NAME: 2025-11-03T17:43:13.2858531Z DEVICE_TYPE: 2025-11-03T17:43:13.2858738Z ##[endgroup] 2025-11-03T17:43:13.5813290Z No files were found with the provided path: debug-*.zip. No artifacts will be uploaded. 2025-11-03T17:43:13.6287889Z ##[group]Run # shellcheck disable=SC2156 2025-11-03T17:43:13.6288209Z # shellcheck disable=SC2156 2025-11-03T17:43:13.6288753Z find . -iname "core.[1-9]*" -exec docker exec "${DOCKER_CONTAINER_ID}" sh -c "gdb python {} -ex 'bt' -ex 'q'" \; 2025-11-03T17:43:13.6296876Z shell: /usr/bin/bash -e {0} 2025-11-03T17:43:13.6297079Z env: 2025-11-03T17:43:13.6297232Z GIT_DEFAULT_BRANCH: main 2025-11-03T17:43:13.6297481Z GPU_FLAG: --gpus all -e NVIDIA_DRIVER_CAPABILITIES=all 2025-11-03T17:43:13.6297879Z DOCKER_CONTAINER_ID: 99e00294b4bd7ffe64017e12af3ca26302f354938a9d519b091e205874b31030 2025-11-03T17:43:13.6298245Z DEVICE_NAME: 2025-11-03T17:43:13.6298402Z DEVICE_TYPE: 2025-11-03T17:43:13.6298561Z ##[endgroup] 2025-11-03T17:43:14.1374538Z Prepare all required actions 2025-11-03T17:43:14.1374879Z Getting action download info 2025-11-03T17:43:14.3014717Z ##[group]Run ./.github/actions/upload-utilization-stats 2025-11-03T17:43:14.3014986Z with: 2025-11-03T17:43:14.3015142Z job_id: 54383051348 2025-11-03T17:43:14.3015501Z job_name: ci-vllm-test / test (vllm_distributed_test_28_failure_test, 1, 1, linux.g6.12xlarge.nvidia.gpu) 2025-11-03T17:43:14.3015993Z workflow_name: vllm-test 2025-11-03T17:43:14.3016184Z workflow_run_id: 19040285420 2025-11-03T17:43:14.3016382Z workflow_attempt: 1 2025-11-03T17:43:14.3016551Z env: 2025-11-03T17:43:14.3016708Z GIT_DEFAULT_BRANCH: main 2025-11-03T17:43:14.3016943Z GPU_FLAG: --gpus all -e NVIDIA_DRIVER_CAPABILITIES=all 2025-11-03T17:43:14.3017335Z DOCKER_CONTAINER_ID: 99e00294b4bd7ffe64017e12af3ca26302f354938a9d519b091e205874b31030 2025-11-03T17:43:14.3017687Z DEVICE_NAME: 2025-11-03T17:43:14.3017854Z DEVICE_TYPE: 2025-11-03T17:43:14.3018005Z ##[endgroup] 2025-11-03T17:43:14.3738067Z ##[group]Run echo "workflow_id: 19040285420" 2025-11-03T17:43:14.3738335Z echo "workflow_id: 19040285420" 2025-11-03T17:43:14.3738567Z echo "workflow_attempt: 1" 2025-11-03T17:43:14.3738787Z echo "workflow_Name: vllm-test" 2025-11-03T17:43:14.3739004Z echo "job_id: 54383051348" 2025-11-03T17:43:14.3739419Z echo "job_name: ci-vllm-test / test (vllm_distributed_test_28_failure_test, 1, 1, linux.g6.12xlarge.nvidia.gpu)" 2025-11-03T17:43:14.3739855Z echo "artifact_prefix: " 2025-11-03T17:43:14.3740078Z python3 --version 2025-11-03T17:43:14.3747167Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2025-11-03T17:43:14.3747447Z env: 2025-11-03T17:43:14.3747619Z GIT_DEFAULT_BRANCH: main 2025-11-03T17:43:14.3747868Z GPU_FLAG: --gpus all -e NVIDIA_DRIVER_CAPABILITIES=all 2025-11-03T17:43:14.3748367Z DOCKER_CONTAINER_ID: 99e00294b4bd7ffe64017e12af3ca26302f354938a9d519b091e205874b31030 2025-11-03T17:43:14.3748741Z DEVICE_NAME: 2025-11-03T17:43:14.3748907Z DEVICE_TYPE: 2025-11-03T17:43:14.3749061Z ##[endgroup] 2025-11-03T17:43:14.3780708Z workflow_id: 19040285420 2025-11-03T17:43:14.3781259Z workflow_attempt: 1 2025-11-03T17:43:14.3781705Z workflow_Name: vllm-test 2025-11-03T17:43:14.3782146Z job_id: 54383051348 2025-11-03T17:43:14.3782968Z job_name: ci-vllm-test / test (vllm_distributed_test_28_failure_test, 1, 1, linux.g6.12xlarge.nvidia.gpu) 2025-11-03T17:43:14.3783898Z artifact_prefix: 2025-11-03T17:43:14.3800940Z Python 3.9.23 2025-11-03T17:43:14.3950747Z ##[group]Run nick-fields/retry@v3.0.0 2025-11-03T17:43:14.3950981Z with: 2025-11-03T17:43:14.3951157Z shell: bash 2025-11-03T17:43:14.3951325Z timeout_minutes: 5 2025-11-03T17:43:14.3951495Z max_attempts: 5 2025-11-03T17:43:14.3951673Z retry_wait_seconds: 30 2025-11-03T17:43:14.3952087Z command: set -eu python3 -m pip install python-dateutil==2.8.2 boto3==1.35.42 pandas==2.1.3 dataclasses_json==0.6.7 2025-11-03T17:43:14.3952536Z polling_interval_seconds: 1 2025-11-03T17:43:14.3952817Z warning_on_retry: true 2025-11-03T17:43:14.3953002Z continue_on_error: false 2025-11-03T17:43:14.3953230Z env: 2025-11-03T17:43:14.3953394Z GIT_DEFAULT_BRANCH: main 2025-11-03T17:43:14.3953627Z GPU_FLAG: --gpus all -e NVIDIA_DRIVER_CAPABILITIES=all 2025-11-03T17:43:14.3954015Z DOCKER_CONTAINER_ID: 99e00294b4bd7ffe64017e12af3ca26302f354938a9d519b091e205874b31030 2025-11-03T17:43:14.3954373Z DEVICE_NAME: 2025-11-03T17:43:14.3954534Z DEVICE_TYPE: 2025-11-03T17:43:14.3954696Z ##[endgroup] 2025-11-03T17:43:14.7214009Z Defaulting to user installation because normal site-packages is not writeable 2025-11-03T17:43:14.8826198Z Collecting python-dateutil==2.8.2 2025-11-03T17:43:14.9014750Z Downloading python_dateutil-2.8.2-py2.py3-none-any.whl (247 kB) 2025-11-03T17:43:15.9067668Z Collecting boto3==1.35.42 2025-11-03T17:43:15.9128906Z Downloading boto3-1.35.42-py3-none-any.whl (139 kB) 2025-11-03T17:43:16.4981199Z Collecting pandas==2.1.3 2025-11-03T17:43:16.5029591Z Downloading pandas-2.1.3-cp39-cp39-manylinux_2_17_x86_64.manylinux2014_x86_64.whl (12.3 MB) 2025-11-03T17:43:16.8514800Z Requirement already satisfied: dataclasses_json==0.6.7 in /home/ec2-user/.local/lib/python3.9/site-packages (0.6.7) 2025-11-03T17:43:16.8529512Z Requirement already satisfied: six>=1.5 in /usr/lib/python3.9/site-packages (from python-dateutil==2.8.2) (1.15.0) 2025-11-03T17:43:16.8570710Z Requirement already satisfied: jmespath<2.0.0,>=0.7.1 in /usr/lib/python3.9/site-packages (from boto3==1.35.42) (0.10.0) 2025-11-03T17:43:16.8575402Z Requirement already satisfied: s3transfer<0.11.0,>=0.10.0 in /home/ec2-user/.local/lib/python3.9/site-packages (from boto3==1.35.42) (0.10.4) 2025-11-03T17:43:16.8578230Z Requirement already satisfied: botocore<1.36.0,>=1.35.42 in /home/ec2-user/.local/lib/python3.9/site-packages (from boto3==1.35.42) (1.35.99) 2025-11-03T17:43:17.7191931Z Collecting numpy<2,>=1.22.4 2025-11-03T17:43:17.7240562Z Downloading numpy-1.26.4-cp39-cp39-manylinux_2_17_x86_64.manylinux2014_x86_64.whl (18.2 MB) 2025-11-03T17:43:18.1440877Z Requirement already satisfied: pytz>=2020.1 in /usr/lib/python3.9/site-packages (from pandas==2.1.3) (2022.7.1) 2025-11-03T17:43:18.2216966Z Collecting tzdata>=2022.1 2025-11-03T17:43:18.2260149Z Downloading tzdata-2025.2-py2.py3-none-any.whl (347 kB) 2025-11-03T17:43:18.2883520Z Requirement already satisfied: marshmallow<4.0.0,>=3.18.0 in /home/ec2-user/.local/lib/python3.9/site-packages (from dataclasses_json==0.6.7) (3.26.1) 2025-11-03T17:43:18.2886519Z Requirement already satisfied: typing-inspect<1,>=0.4.0 in /home/ec2-user/.local/lib/python3.9/site-packages (from dataclasses_json==0.6.7) (0.9.0) 2025-11-03T17:43:18.2968903Z Requirement already satisfied: urllib3<1.27,>=1.25.4 in /usr/lib/python3.9/site-packages (from botocore<1.36.0,>=1.35.42->boto3==1.35.42) (1.25.10) 2025-11-03T17:43:18.3050508Z Requirement already satisfied: packaging>=17.0 in /home/ec2-user/.local/lib/python3.9/site-packages (from marshmallow<4.0.0,>=3.18.0->dataclasses_json==0.6.7) (25.0) 2025-11-03T17:43:18.3140046Z Requirement already satisfied: mypy-extensions>=0.3.0 in /home/ec2-user/.local/lib/python3.9/site-packages (from typing-inspect<1,>=0.4.0->dataclasses_json==0.6.7) (1.1.0) 2025-11-03T17:43:18.3143372Z Requirement already satisfied: typing-extensions>=3.7.4 in /home/ec2-user/.local/lib/python3.9/site-packages (from typing-inspect<1,>=0.4.0->dataclasses_json==0.6.7) (4.15.0) 2025-11-03T17:43:18.5886047Z Installing collected packages: python-dateutil, tzdata, numpy, pandas, boto3 2025-11-03T17:43:23.8097665Z Attempting uninstall: boto3 2025-11-03T17:43:23.8099045Z Found existing installation: boto3 1.35.33 2025-11-03T17:43:23.8207449Z Uninstalling boto3-1.35.33: 2025-11-03T17:43:23.8223035Z Successfully uninstalled boto3-1.35.33 2025-11-03T17:43:24.0065133Z Successfully installed boto3-1.35.42 numpy-1.26.4 pandas-2.1.3 python-dateutil-2.8.2 tzdata-2025.2 2025-11-03T17:43:24.4752718Z Command completed after 1 attempt(s). 2025-11-03T17:43:24.4996576Z ##[group]Run python3 -m tools.stats.upload_utilization_stats.upload_utilization_stats \ 2025-11-03T17:43:24.5000909Z python3 -m tools.stats.upload_utilization_stats.upload_utilization_stats \ 2025-11-03T17:43:24.5001298Z  --workflow-run-id "19040285420" \ 2025-11-03T17:43:24.5001556Z  --workflow-name "vllm-test" \ 2025-11-03T17:43:24.5001824Z  --workflow-run-attempt "1" \ 2025-11-03T17:43:24.5002065Z  --job-id "54383051348" \ 2025-11-03T17:43:24.5002473Z  --job-name "ci-vllm-test / test (vllm_distributed_test_28_failure_test, 1, 1, linux.g6.12xlarge.nvidia.gpu)" \ 2025-11-03T17:43:24.5002890Z  --local-path "" \ 2025-11-03T17:43:24.5003099Z  --artifact-prefix "" 2025-11-03T17:43:24.5012368Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2025-11-03T17:43:24.5012652Z env: 2025-11-03T17:43:24.5012815Z GIT_DEFAULT_BRANCH: main 2025-11-03T17:43:24.5013231Z GPU_FLAG: --gpus all -e NVIDIA_DRIVER_CAPABILITIES=all 2025-11-03T17:43:24.5013646Z DOCKER_CONTAINER_ID: 99e00294b4bd7ffe64017e12af3ca26302f354938a9d519b091e205874b31030 2025-11-03T17:43:24.5013995Z DEVICE_NAME: 2025-11-03T17:43:24.5014156Z DEVICE_TYPE: 2025-11-03T17:43:24.5014328Z ##[endgroup] 2025-11-03T17:43:27.3609162Z repo: pytorch/pytorch 2025-11-03T17:43:27.3610339Z Search for test log in s3 bucket: ossci-utilization 2025-11-03T17:43:27.3611602Z Downloading logs-test-vllm_distributed_test_28_failure_test-1-1-linux.g6.12xlarge.nvidia.gpu_54383051348.zip 2025-11-03T17:43:27.3612911Z extracting usage_log.txt from zip file logs-test-vllm_distributed_test_28_failure_test-1-1-linux.g6.12xlarge.nvidia.gpu_54383051348.zip 2025-11-03T17:43:27.3613895Z Converted Log Model: UtilizationMetadata: 2025-11-03T17:43:27.3615897Z UtilizationMetadata(level='metadata', workflow_id='19040285420', job_id='54383051348', workflow_name='vllm-test', job_name='ci-vllm-test / test (vllm_distributed_test_28_failure_test, 1, 1, linux.g6.12xlarge.nvidia.gpu)', usage_collect_interval=1.0, data_model_version=1.5, start_at=1762189545, gpu_count=4, cpu_count=48, gpu_type='pynvml', error=None) 2025-11-03T17:43:27.3617985Z [Db Segments] detected pytest cmd: 12, generated segments: 12 2025-11-03T17:43:27.3618572Z [db model] Peek db timeseries 2025-11-03T17:43:27.3618949Z :{ 2025-11-03T17:43:27.3619229Z "created_at": 1762191806, 2025-11-03T17:43:27.3619638Z "type": "utilization", 2025-11-03T17:43:27.3620042Z "tags": [ 2025-11-03T17:43:27.3620381Z "record" 2025-11-03T17:43:27.3620724Z ], 2025-11-03T17:43:27.3621074Z "time_stamp": 1762189545, 2025-11-03T17:43:27.3621438Z "repo": "pytorch/pytorch", 2025-11-03T17:43:27.3621668Z "workflow_id": 19040285420, 2025-11-03T17:43:27.3621885Z "run_attempt": 1, 2025-11-03T17:43:27.3632137Z "job_id": 54383051348, 2025-11-03T17:43:27.3632523Z "workflow_name": "vllm-test", 2025-11-03T17:43:27.3632933Z "job_name": "ci-vllm-test / test (vllm_distributed_test_28_failure_test, 1, 1, linux.g6.12xlarge.nvidia.gpu)", 2025-11-03T17:43:27.3633319Z "json_data": "{}" 2025-11-03T17:43:27.3633505Z } 2025-11-03T17:43:27.3633871Z Writing 1 documents to S3 ossci-utilization/util_metadata/v_1.5/pytorch/pytorch/19040285420/1/54383051348/metadata 2025-11-03T17:43:27.3634521Z Done! Finish writing document to S3 ossci-utilization/util_metadata/v_1.5/pytorch/pytorch/19040285420/1/54383051348/metadata 2025-11-03T17:43:27.3635179Z Writing 448 documents to S3 ossci-utilization/util_timeseries/v_1.5/pytorch/pytorch/19040285420/1/54383051348/time_series 2025-11-03T17:43:27.3635853Z Done! Finish writing document to S3 ossci-utilization/util_timeseries/v_1.5/pytorch/pytorch/19040285420/1/54383051348/time_series 2025-11-03T17:43:27.4689127Z ##[group]Run pytorch/test-infra/.github/actions/teardown-linux@main 2025-11-03T17:43:27.4689515Z with: 2025-11-03T17:43:27.4689665Z env: 2025-11-03T17:43:27.4689829Z GIT_DEFAULT_BRANCH: main 2025-11-03T17:43:27.4690107Z GPU_FLAG: --gpus all -e NVIDIA_DRIVER_CAPABILITIES=all 2025-11-03T17:43:27.4690511Z DOCKER_CONTAINER_ID: 99e00294b4bd7ffe64017e12af3ca26302f354938a9d519b091e205874b31030 2025-11-03T17:43:27.4690854Z DEVICE_NAME: 2025-11-03T17:43:27.4691019Z DEVICE_TYPE: 2025-11-03T17:43:27.4691176Z ##[endgroup] 2025-11-03T17:43:27.4836383Z ##[group]Run set -eou pipefail 2025-11-03T17:43:27.4836642Z set -eou pipefail 2025-11-03T17:43:27.4836843Z  2025-11-03T17:43:27.4837124Z echo "Holding runner for 2 hours until all ssh sessions have logged out" 2025-11-03T17:43:27.4837471Z for _ in $(seq 1440); do 2025-11-03T17:43:27.4837727Z  # Break if no ssh session exists anymore 2025-11-03T17:43:27.4838004Z  if [ "$(who)" = "" ]; then 2025-11-03T17:43:27.4838246Z  break 2025-11-03T17:43:27.4838414Z  fi 2025-11-03T17:43:27.4838588Z  echo "." 2025-11-03T17:43:27.4838766Z  sleep 5 2025-11-03T17:43:27.4838933Z done 2025-11-03T17:43:27.4846808Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2025-11-03T17:43:27.4847094Z env: 2025-11-03T17:43:27.4847260Z GIT_DEFAULT_BRANCH: main 2025-11-03T17:43:27.4847506Z GPU_FLAG: --gpus all -e NVIDIA_DRIVER_CAPABILITIES=all 2025-11-03T17:43:27.4847900Z DOCKER_CONTAINER_ID: 99e00294b4bd7ffe64017e12af3ca26302f354938a9d519b091e205874b31030 2025-11-03T17:43:27.4848457Z DEVICE_NAME: 2025-11-03T17:43:27.4848636Z DEVICE_TYPE: 2025-11-03T17:43:27.4848799Z ##[endgroup] 2025-11-03T17:43:27.4882206Z Holding runner for 2 hours until all ssh sessions have logged out 2025-11-03T17:43:27.6014257Z ##[group]Run # ignore expansion of "docker ps -q" since it could be empty 2025-11-03T17:43:27.6014720Z # ignore expansion of "docker ps -q" since it could be empty 2025-11-03T17:43:27.6015037Z # shellcheck disable=SC2046 2025-11-03T17:43:27.6015305Z docker stop $(docker ps -q) || true 2025-11-03T17:43:27.6015546Z # Prune all of the docker images 2025-11-03T17:43:27.6015788Z docker system prune -af 2025-11-03T17:43:27.6023616Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2025-11-03T17:43:27.6023901Z env: 2025-11-03T17:43:27.6024059Z GIT_DEFAULT_BRANCH: main 2025-11-03T17:43:27.6024306Z GPU_FLAG: --gpus all -e NVIDIA_DRIVER_CAPABILITIES=all 2025-11-03T17:43:27.6024704Z DOCKER_CONTAINER_ID: 99e00294b4bd7ffe64017e12af3ca26302f354938a9d519b091e205874b31030 2025-11-03T17:43:27.6025066Z DEVICE_NAME: 2025-11-03T17:43:27.6025247Z DEVICE_TYPE: 2025-11-03T17:43:27.6025412Z ##[endgroup] 2025-11-03T17:43:39.5342474Z 99e00294b4bd 2025-11-03T17:43:47.7643840Z Deleted Containers: 2025-11-03T17:43:47.7644488Z 99e00294b4bd7ffe64017e12af3ca26302f354938a9d519b091e205874b31030 2025-11-03T17:43:47.7644816Z 2025-11-03T17:43:59.3368614Z Deleted Images: 2025-11-03T17:43:59.3369841Z untagged: 308535385114.dkr.ecr.us-east-1.amazonaws.com/pytorch/ci-image:pytorch-linux-jammy-cuda12.8-cudnn9-py3.12-gcc11-vllm-30147fb242f88ab8f13e14b76ab186d189d8605c 2025-11-03T17:43:59.3370824Z untagged: 308535385114.dkr.ecr.us-east-1.amazonaws.com/pytorch/ci-image@sha256:655b37e47c7ab0fab00339d6162ef718600c6b4e76f1601297e6a07026860230 2025-11-03T17:43:59.3371886Z deleted: sha256:123a0a8d2d8124dd2cca825f1276a192d04ee365eb7de45a4829163bae42bcc5 2025-11-03T17:43:59.3372391Z deleted: sha256:31820faffa36fa10d707f81682d96bdb5de2207e3612a625bc2add174a16f33b 2025-11-03T17:43:59.3372857Z deleted: sha256:1bbaafd08d25d515d9317d42cef8b0d6d389624b4d01c65b2374e72e58a98012 2025-11-03T17:43:59.3373300Z deleted: sha256:3a98336219906852cadb9ffa8b260359b85b0ef2ce63f194b2ad17340bf2c5d6 2025-11-03T17:43:59.3373737Z deleted: sha256:a73360da0fd9b9cc724c9a24f957b8b4bd0936a4f97ee5c2fc41a13d6994b0bb 2025-11-03T17:43:59.3374438Z deleted: sha256:38cc6c78e3939cce4347b386ded8480f5b3ba61185cda08e95c53c55b71d1698 2025-11-03T17:43:59.3374922Z deleted: sha256:77104ac67f1c4a71d44f95e55721e88d0d338ccb804df56f916ea3edd8c57177 2025-11-03T17:43:59.3375382Z deleted: sha256:e81aea95b6a623c6eea401a37492286ab3ee28e5042b808f2c9328ba1ba29d53 2025-11-03T17:43:59.3375842Z deleted: sha256:b0de82ab4f930d742562dc8beffaeb4f9e2e37888069111da820e249172d0c04 2025-11-03T17:43:59.3376299Z deleted: sha256:47d87661555964c91d4fc524b43b15505dcbecdf57b6167adcf9774f1c597d53 2025-11-03T17:43:59.3376742Z deleted: sha256:ce3967b7450d7c2a5833ce475b87987bb427bb87bb87d421aaa98cfe8fe093d3 2025-11-03T17:43:59.3377198Z deleted: sha256:65cde699f23961ef3b15b672e8715d50af40ec180991a3fe50c708d5d955da37 2025-11-03T17:43:59.3377739Z deleted: sha256:bda9c0867d0e7ac8556b4cfc02c04d75bd9e845a6b579e50ddbeac76f48824ae 2025-11-03T17:43:59.3378256Z deleted: sha256:b185a3d739322209684243ec5a50898a86f3c18f0058921dcaff683e1a759f53 2025-11-03T17:43:59.3378769Z deleted: sha256:f3b6f40043f4f16c2f36594521bcd3561ae8d9e84c3a1baac1fdd6be2b1d2503 2025-11-03T17:43:59.3379261Z deleted: sha256:c5d92fcf72e25e113b2514a4e18a1faf280ff97600c44e089c8add4ba4bfed67 2025-11-03T17:43:59.3379704Z deleted: sha256:e1b08cb82b0287ee00554e71dc6e69dd5747e05995a19d5be2427a0a0e2ea298 2025-11-03T17:43:59.3380142Z deleted: sha256:f514e296acfd635f5184645775136c340b4bdab12aa80c62965f8fd66a9adc2d 2025-11-03T17:43:59.3380586Z deleted: sha256:a7b8036e533224113c12edff1c24cbd4e6d1b2bfc15ea49643f262a00309092b 2025-11-03T17:43:59.3381026Z deleted: sha256:f23436ff58fe28485a737310c5367061602a90ab5856989457daa060a372a0ae 2025-11-03T17:43:59.3381468Z deleted: sha256:6e1a14c6f6f6089f5b4be23f93de15f5c6898accb4fdc2c577df25bef047c66c 2025-11-03T17:43:59.3382026Z deleted: sha256:cb7a3bd4bcc1a4223e27af60de87c05873f833c649e227b9afb63f5bb208cbb5 2025-11-03T17:43:59.3382469Z deleted: sha256:45210d5bac2c05cfd830e1097010555216604bafbfb71ea5495a0e4b224b7f9a 2025-11-03T17:43:59.3382924Z deleted: sha256:afdad6a8611fbabce05e6714dfffad4c1675e10d3b1e039749ce204b0f8cd41c 2025-11-03T17:43:59.3383377Z deleted: sha256:a44823411001fe4089d1c72c3d5de14d0b2687232a5d8ced0818589e3325316d 2025-11-03T17:43:59.3383813Z deleted: sha256:f263db78292410b172d222dbe7c398e9d8c87663eb725ca5f3ea998b4b2ed080 2025-11-03T17:43:59.3384256Z deleted: sha256:13ecce5bfe277268a0de6e0d01ae45821caad15c1a608da44f82bbc57cbdc020 2025-11-03T17:43:59.3384706Z deleted: sha256:9ab9deacc335d4c360367e7f061aaf66216e015d9d058a73260f35efe0b73b9d 2025-11-03T17:43:59.3385159Z deleted: sha256:bfed55b6ddec7b775d5fa270eb2be23a41a6069ca17ea5b3dd991f8f67712dd8 2025-11-03T17:43:59.3385608Z deleted: sha256:982e6e5dfbb4728f4a51c83a1896cc16587d550ae6f515f89fa807dc87046bfb 2025-11-03T17:43:59.3386049Z deleted: sha256:df42847c61a0101cc7348e49cd2a500cdbbde2f218e4d6fe58ab752b51a43fac 2025-11-03T17:43:59.3386496Z deleted: sha256:de33546992ae1fce63d51acabeec23ebeeaba447c7155cfee96ac2d115b3640f 2025-11-03T17:43:59.3386936Z deleted: sha256:45ba9418147153ef58cb0e6509c4a9139163afeccd32dc80e8a335b32c050b07 2025-11-03T17:43:59.3387372Z deleted: sha256:4e36f594c176514603ccb1a2c4d5e4900ed546cb9499dcc3f6061aab9519c57f 2025-11-03T17:43:59.3387880Z deleted: sha256:a83e194f5cc8b355dda2be872a19f4dc25ed2314ee7952374b67baf90977c9da 2025-11-03T17:43:59.3388319Z deleted: sha256:445972ee71b4614ff3f1d009fba94556bf17e09b2c99292a908ebf5e9529e1c7 2025-11-03T17:43:59.3388755Z deleted: sha256:d2e38abb0bff4a6968be64628e6a661db235e0530d5925a552991f3fe1ce1a30 2025-11-03T17:43:59.3389196Z deleted: sha256:7275b58f0e992bd8031ef2ea708bd2c3b080ecaaecbbebbb2d438ef4413913c1 2025-11-03T17:43:59.3389638Z deleted: sha256:030aae20394d8c1c0d34388b4856c6a2966f30f4cd8393cd33d1defccd696a55 2025-11-03T17:43:59.3390078Z deleted: sha256:f4d5ea7a039db19152ce6bc9827e57a4de69ed4fb053f8e50de52eac2410daca 2025-11-03T17:43:59.3390531Z deleted: sha256:545e10edde5c7ee4d66227f76f9b63db14d5f1b03e8588f429c3961260a2bf09 2025-11-03T17:43:59.3390971Z deleted: sha256:8eced9c38b8d1316270f6ab88b14de97a711d07c7b668f568bd501dc518e7ced 2025-11-03T17:43:59.3391528Z deleted: sha256:709ec82360eef380a961d01fea671225c186880c6cf3375daca37712840f8549 2025-11-03T17:43:59.3391982Z deleted: sha256:7bb05462f24d343c9c9e7e2d24c62d8ba06495e5e07f040ab95a8eaf29724acf 2025-11-03T17:43:59.3392425Z deleted: sha256:0dfdf4b3e5ed9f1d5f79747ed2d29e2b87752d07331442c42efa1a613c1011f9 2025-11-03T17:43:59.3392864Z deleted: sha256:efb8a361d5f170f02f902aaad54acb9dcd69e2b4da2ae2470934d8dad2d6cb25 2025-11-03T17:43:59.3393317Z deleted: sha256:9e8da81757d04e51ce5c02529d343ff92c351b8435f829866a45df37cd4a65eb 2025-11-03T17:43:59.3393753Z deleted: sha256:6a408c52616752ec8052b86a8896a66ea203cfedebd61abc3e660ed783e6778d 2025-11-03T17:43:59.3394191Z deleted: sha256:7cf69cce58132deb3c49815b6e82debd5687c108df52ebc97770b29c97d0c1b6 2025-11-03T17:43:59.3394639Z deleted: sha256:64540da67f68a44e7e37dd5c80e72066f8394abdd1d890073f301097f6234e04 2025-11-03T17:43:59.3395076Z deleted: sha256:3c3d38edfc6f883c7e76990a02f0e977d3d284835cabe63f596c134a4f0408c1 2025-11-03T17:43:59.3395526Z deleted: sha256:dbfcbb7dcb1e0a93c3b284a39a54ac20951ca01bd43d222ef6c21a9a958ef19d 2025-11-03T17:43:59.3395971Z deleted: sha256:89230198d83d52abcd62d73aa0c0919ae60e4c3b0552805ad87fb0e14c78f0ff 2025-11-03T17:43:59.3396410Z deleted: sha256:529d43aa0a1d6c2a11e0ae029bb4a68c2a3cde415a77d094886d6db98794e336 2025-11-03T17:43:59.3396839Z deleted: sha256:a8066515cbb348517543b217f1dcc01167d7202a7dbea9745f573c02834d65a3 2025-11-03T17:43:59.3397291Z deleted: sha256:a13e9abfebb1053045eead7615d90161552798de57c43eaf337b3e64a1140915 2025-11-03T17:43:59.3397803Z deleted: sha256:19ee032c28086715c5dfc7960eaa38dfb68db347946624410c344101d7dee4fb 2025-11-03T17:43:59.3398330Z deleted: sha256:c3a1419c3b6e2c7a3aa39b66ce90f210b6de4804a6a494babc21bda3b8ddb728 2025-11-03T17:43:59.3398910Z deleted: sha256:f6dd774d7904ac47308018e060dd1a6460f1e6b0b495e4637d6dec0e766bbef9 2025-11-03T17:43:59.3399421Z deleted: sha256:b14fbc43504939ea343b2c5ba35bba56b52e208d1262a383ef185124c12f5738 2025-11-03T17:43:59.3399881Z deleted: sha256:58dac273a49f6d6ffd5ecaf85dcf3874a239a37db02320098059eb5686d33ae8 2025-11-03T17:43:59.3400327Z deleted: sha256:7c4c937e97b6d494fb2e144665b1594212d1b854a880e74265e7f2d70f3c3944 2025-11-03T17:43:59.3400757Z deleted: sha256:bf342185a1f51131a608e59d62ad65da006e6322d6aa191876417eef3c624118 2025-11-03T17:43:59.3401184Z deleted: sha256:121408b55e8d21cee32f17ecb5d380729586f6b7d8198be62423161a316f5909 2025-11-03T17:43:59.3401622Z deleted: sha256:fd8ef0586f9366378d192cf475e6e783c5815185a528b077d533f9ecb4f6881d 2025-11-03T17:43:59.3402059Z deleted: sha256:4b4bbdfa85b1154871d9d703b6578679d6ad436b0d157acc03589b20f7e51174 2025-11-03T17:43:59.3402498Z deleted: sha256:a7ec280e3d580ac545fdddb47716864c39c8d5cbd27c48487fb39f6d52ec5813 2025-11-03T17:43:59.3402943Z deleted: sha256:af6a46762136c8cc3cca20b4f523662bda54d2a17a70900b436f5829c2694ab0 2025-11-03T17:43:59.3403377Z deleted: sha256:ed75bd3dae248978cc1d7f962d6313404c2f2cf4c5f5defa282a77834011b95a 2025-11-03T17:43:59.3403825Z deleted: sha256:baab5a5b13fbe97055c0eee16f014775c75da39ee1bfd50ce919e42fd0740eb3 2025-11-03T17:43:59.3404280Z deleted: sha256:abc859f44381c472c6110fe3ea1d5db4f58416a2e4ac461bd7a636be635d19f1 2025-11-03T17:43:59.3404724Z deleted: sha256:120e4c672fc1d0e7e22a7dcfb14bc009fdc45208d24cc41e4fed2c865d4c61b4 2025-11-03T17:43:59.3405200Z deleted: sha256:3a901973b8d19183073f3925d3581e2124a7e019528f651c18ed984dc8920633 2025-11-03T17:43:59.3405620Z deleted: sha256:13d9cf2669287254c4fc4922600894d75d97fb756c308b92ae4067f685b3b9c4 2025-11-03T17:43:59.3406061Z deleted: sha256:d7cc64ff2a7a5cdf4d2e027c93b11862cb741b68bd813300f0e850d03b5894de 2025-11-03T17:43:59.3406505Z deleted: sha256:ddb9c9f2a33a6ff22294d8f0450ec0da44874318928f90f1f53c9e51bd736c90 2025-11-03T17:43:59.3406943Z deleted: sha256:b615ed2509100fbf1d31da5895e95df9b6d890ef5b8cb725aea547c2310d370d 2025-11-03T17:43:59.3407379Z deleted: sha256:cf4cba224210966dd0f57c902103bbe1142ffc9ae140509ba29c98b14df150d2 2025-11-03T17:43:59.3407806Z deleted: sha256:63cc9363306d393f09857fa7d93605c66c884fdcbbb254c37474b6cf4f293a35 2025-11-03T17:43:59.3408312Z deleted: sha256:767e56ba346ae714b6e6b816baa839051145ed78cfa0e4524a86cc287b0c4b00 2025-11-03T17:43:59.3408814Z untagged: public.ecr.aws/docker/library/python:3.13 2025-11-03T17:43:59.3409311Z untagged: public.ecr.aws/docker/library/python@sha256:12513c633252a28bcfee85839aa384e1af322f11275779c6645076c6cd0cfe52 2025-11-03T17:43:59.3410141Z deleted: sha256:438f5c84513b2fc181c3e3cf09b621e66b769789fab066c9c436a35b8588d29e 2025-11-03T17:43:59.3410589Z deleted: sha256:007b97ee297d1b90751d5850e62f9ebc070d655634bd08aa0a4b83fea3a53a14 2025-11-03T17:43:59.3411019Z deleted: sha256:ee90a75564a713c0ee50c40d6581495f346f765072ae44828bf219e3da9fb55c 2025-11-03T17:43:59.3411448Z deleted: sha256:788fe66fc657a4e83d7248083c07adcd949d1e8a803c2835d66e07bff863f4d8 2025-11-03T17:43:59.3411890Z deleted: sha256:a60dcbc8722b7ecb7798ea22a4dad988c0ce2ccb1efb598aa8deee75026f3e1d 2025-11-03T17:43:59.3412345Z deleted: sha256:f8b7333a4ac07b3cf4c5976461e665c3445b16dacf992d1d3130e2d39ebe2570 2025-11-03T17:43:59.3412804Z deleted: sha256:77c2e291dd8342baaa9f49b9265f1925e1357f6a98a61f736d060eaaf3cd7b3b 2025-11-03T17:43:59.3413240Z deleted: sha256:f2522c6ed78b338a9e272dd5038005d008f74729e036073e837f701f221b99ba 2025-11-03T17:43:59.3413509Z 2025-11-03T17:43:59.3413596Z Total reclaimed space: 43.4GB 2025-11-03T17:43:59.3496101Z Post job cleanup. 2025-11-03T17:43:59.3540793Z Post job cleanup. 2025-11-03T17:43:59.4497520Z [command]/usr/bin/git version 2025-11-03T17:43:59.4546804Z git version 2.50.1 2025-11-03T17:43:59.4587560Z Copying '/home/ec2-user/.gitconfig' to '/home/ec2-user/actions-runner/_work/_temp/13ec42dc-575f-4932-835a-d1514adc9b2d/.gitconfig' 2025-11-03T17:43:59.4597475Z Temporarily overriding HOME='/home/ec2-user/actions-runner/_work/_temp/13ec42dc-575f-4932-835a-d1514adc9b2d' before making global git config changes 2025-11-03T17:43:59.4598522Z Adding repository directory to the temporary git global config as a safe directory 2025-11-03T17:43:59.4604210Z [command]/usr/bin/git config --global --add safe.directory /home/ec2-user/actions-runner/_work/pytorch/pytorch 2025-11-03T17:43:59.4665742Z [command]/usr/bin/git config --local --name-only --get-regexp core\.sshCommand 2025-11-03T17:43:59.4716492Z [command]/usr/bin/git submodule foreach --recursive sh -c "git config --local --name-only --get-regexp 'core\.sshCommand' && git config --local --unset-all 'core.sshCommand' || :" 2025-11-03T17:43:59.5160497Z Entering 'android/libs/fbjni' 2025-11-03T17:43:59.5252502Z Entering 'third_party/FP16' 2025-11-03T17:43:59.5339988Z Entering 'third_party/FXdiv' 2025-11-03T17:43:59.5425040Z Entering 'third_party/NNPACK' 2025-11-03T17:43:59.5514477Z Entering 'third_party/NVTX' 2025-11-03T17:43:59.5605702Z Entering 'third_party/VulkanMemoryAllocator' 2025-11-03T17:43:59.5700117Z Entering 'third_party/XNNPACK' 2025-11-03T17:43:59.5799586Z Entering 'third_party/aiter' 2025-11-03T17:43:59.5892874Z Entering 'third_party/aiter/3rdparty/composable_kernel' 2025-11-03T17:43:59.5992808Z Entering 'third_party/benchmark' 2025-11-03T17:43:59.6074928Z Entering 'third_party/composable_kernel' 2025-11-03T17:43:59.6184186Z Entering 'third_party/cpp-httplib' 2025-11-03T17:43:59.6275888Z Entering 'third_party/cpuinfo' 2025-11-03T17:43:59.6366098Z Entering 'third_party/cudnn_frontend' 2025-11-03T17:43:59.6455430Z Entering 'third_party/cutlass' 2025-11-03T17:43:59.6565263Z Entering 'third_party/fbgemm' 2025-11-03T17:43:59.6654738Z Entering 'third_party/fbgemm/external/asmjit' 2025-11-03T17:43:59.6747055Z Entering 'third_party/fbgemm/external/composable_kernel' 2025-11-03T17:43:59.6843951Z Entering 'third_party/fbgemm/external/cpuinfo' 2025-11-03T17:43:59.6929499Z Entering 'third_party/fbgemm/external/cutlass' 2025-11-03T17:43:59.7023532Z Entering 'third_party/fbgemm/external/googletest' 2025-11-03T17:43:59.7106087Z Entering 'third_party/fbgemm/external/hipify_torch' 2025-11-03T17:43:59.7192044Z Entering 'third_party/fbgemm/external/json' 2025-11-03T17:43:59.7276798Z Entering 'third_party/flash-attention' 2025-11-03T17:43:59.7365567Z Entering 'third_party/flash-attention/csrc/composable_kernel' 2025-11-03T17:43:59.7457027Z Entering 'third_party/flash-attention/csrc/cutlass' 2025-11-03T17:43:59.7557833Z Entering 'third_party/flatbuffers' 2025-11-03T17:43:59.7647190Z Entering 'third_party/fmt' 2025-11-03T17:43:59.7739371Z Entering 'third_party/gemmlowp/gemmlowp' 2025-11-03T17:43:59.7824613Z Entering 'third_party/gloo' 2025-11-03T17:43:59.7915424Z Entering 'third_party/googletest' 2025-11-03T17:43:59.8002142Z Entering 'third_party/ideep' 2025-11-03T17:43:59.8080519Z Entering 'third_party/ideep/mkl-dnn' 2025-11-03T17:43:59.8178203Z Entering 'third_party/ittapi' 2025-11-03T17:43:59.8266629Z Entering 'third_party/kineto' 2025-11-03T17:43:59.8362183Z Entering 'third_party/kineto/libkineto/third_party/dynolog' 2025-11-03T17:43:59.8455222Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/DCGM' 2025-11-03T17:43:59.8537557Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/cpr' 2025-11-03T17:43:59.8624193Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/fmt' 2025-11-03T17:43:59.8708837Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/gflags' 2025-11-03T17:43:59.8792060Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/gflags/doc' 2025-11-03T17:43:59.8885343Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/glog' 2025-11-03T17:43:59.8979999Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/googletest' 2025-11-03T17:43:59.9063580Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/json' 2025-11-03T17:43:59.9148228Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/pfs' 2025-11-03T17:43:59.9235083Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/prometheus-cpp' 2025-11-03T17:43:59.9325418Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/prometheus-cpp/3rdparty/civetweb' 2025-11-03T17:43:59.9413956Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/prometheus-cpp/3rdparty/googletest' 2025-11-03T17:43:59.9518983Z Entering 'third_party/kineto/libkineto/third_party/fmt' 2025-11-03T17:43:59.9604641Z Entering 'third_party/kineto/libkineto/third_party/googletest' 2025-11-03T17:43:59.9704471Z Entering 'third_party/kleidiai' 2025-11-03T17:43:59.9796522Z Entering 'third_party/mimalloc' 2025-11-03T17:43:59.9887870Z Entering 'third_party/nlohmann' 2025-11-03T17:43:59.9973039Z Entering 'third_party/onnx' 2025-11-03T17:44:00.0070266Z Entering 'third_party/onnx/third_party/pybind11' 2025-11-03T17:44:00.0162844Z Entering 'third_party/opentelemetry-cpp' 2025-11-03T17:44:00.0244563Z Entering 'third_party/opentelemetry-cpp/third_party/benchmark' 2025-11-03T17:44:00.0321269Z Entering 'third_party/opentelemetry-cpp/third_party/googletest' 2025-11-03T17:44:00.0403510Z Entering 'third_party/opentelemetry-cpp/third_party/ms-gsl' 2025-11-03T17:44:00.0482156Z Entering 'third_party/opentelemetry-cpp/third_party/nlohmann-json' 2025-11-03T17:44:00.0562765Z Entering 'third_party/opentelemetry-cpp/third_party/opentelemetry-proto' 2025-11-03T17:44:00.0636612Z Entering 'third_party/opentelemetry-cpp/third_party/opentracing-cpp' 2025-11-03T17:44:00.0721045Z Entering 'third_party/opentelemetry-cpp/third_party/prometheus-cpp' 2025-11-03T17:44:00.0796753Z Entering 'third_party/opentelemetry-cpp/third_party/prometheus-cpp/3rdparty/civetweb' 2025-11-03T17:44:00.0882820Z Entering 'third_party/opentelemetry-cpp/third_party/prometheus-cpp/3rdparty/googletest' 2025-11-03T17:44:00.0966378Z Entering 'third_party/opentelemetry-cpp/tools/vcpkg' 2025-11-03T17:44:00.1075826Z Entering 'third_party/pocketfft' 2025-11-03T17:44:00.1156237Z Entering 'third_party/protobuf' 2025-11-03T17:44:00.1245098Z Entering 'third_party/protobuf/third_party/benchmark' 2025-11-03T17:44:00.1337132Z Entering 'third_party/protobuf/third_party/googletest' 2025-11-03T17:44:00.1431142Z Entering 'third_party/psimd' 2025-11-03T17:44:00.1513392Z Entering 'third_party/pthreadpool' 2025-11-03T17:44:00.1594202Z Entering 'third_party/pybind11' 2025-11-03T17:44:00.1687887Z Entering 'third_party/python-peachpy' 2025-11-03T17:44:00.1786243Z Entering 'third_party/sleef' 2025-11-03T17:44:00.1879101Z Entering 'third_party/tensorpipe' 2025-11-03T17:44:00.1964667Z Entering 'third_party/tensorpipe/third_party/googletest' 2025-11-03T17:44:00.2055380Z Entering 'third_party/tensorpipe/third_party/libnop' 2025-11-03T17:44:00.2142402Z Entering 'third_party/tensorpipe/third_party/libuv' 2025-11-03T17:44:00.2225394Z Entering 'third_party/tensorpipe/third_party/pybind11' 2025-11-03T17:44:00.2314395Z Entering 'third_party/tensorpipe/third_party/pybind11/tools/clang' 2025-11-03T17:44:00.2436203Z [command]/usr/bin/git config --local --name-only --get-regexp http\.https\:\/\/github\.com\/\.extraheader 2025-11-03T17:44:00.2466620Z http.https://github.com/.extraheader 2025-11-03T17:44:00.2478695Z [command]/usr/bin/git config --local --unset-all http.https://github.com/.extraheader 2025-11-03T17:44:00.2523124Z [command]/usr/bin/git submodule foreach --recursive sh -c "git config --local --name-only --get-regexp 'http\.https\:\/\/github\.com\/\.extraheader' && git config --local --unset-all 'http.https://github.com/.extraheader' || :" 2025-11-03T17:44:00.2992150Z Entering 'android/libs/fbjni' 2025-11-03T17:44:00.3052683Z http.https://github.com/.extraheader 2025-11-03T17:44:00.3112690Z Entering 'third_party/FP16' 2025-11-03T17:44:00.3165695Z http.https://github.com/.extraheader 2025-11-03T17:44:00.3228736Z Entering 'third_party/FXdiv' 2025-11-03T17:44:00.3291709Z http.https://github.com/.extraheader 2025-11-03T17:44:00.3354315Z Entering 'third_party/NNPACK' 2025-11-03T17:44:00.3404565Z http.https://github.com/.extraheader 2025-11-03T17:44:00.3454282Z Entering 'third_party/NVTX' 2025-11-03T17:44:00.3512662Z http.https://github.com/.extraheader 2025-11-03T17:44:00.3570944Z Entering 'third_party/VulkanMemoryAllocator' 2025-11-03T17:44:00.3631719Z http.https://github.com/.extraheader 2025-11-03T17:44:00.3697587Z Entering 'third_party/XNNPACK' 2025-11-03T17:44:00.3753206Z http.https://github.com/.extraheader 2025-11-03T17:44:00.3826195Z Entering 'third_party/aiter' 2025-11-03T17:44:00.3884095Z http.https://github.com/.extraheader 2025-11-03T17:44:00.3942889Z Entering 'third_party/aiter/3rdparty/composable_kernel' 2025-11-03T17:44:00.3994087Z http.https://github.com/.extraheader 2025-11-03T17:44:00.4063075Z Entering 'third_party/benchmark' 2025-11-03T17:44:00.4115843Z http.https://github.com/.extraheader 2025-11-03T17:44:00.4173733Z Entering 'third_party/composable_kernel' 2025-11-03T17:44:00.4232370Z http.https://github.com/.extraheader 2025-11-03T17:44:00.4299817Z Entering 'third_party/cpp-httplib' 2025-11-03T17:44:00.4355357Z http.https://github.com/.extraheader 2025-11-03T17:44:00.4406356Z Entering 'third_party/cpuinfo' 2025-11-03T17:44:00.4464452Z http.https://github.com/.extraheader 2025-11-03T17:44:00.4527515Z Entering 'third_party/cudnn_frontend' 2025-11-03T17:44:00.4583996Z http.https://github.com/.extraheader 2025-11-03T17:44:00.4646299Z Entering 'third_party/cutlass' 2025-11-03T17:44:00.4701477Z http.https://github.com/.extraheader 2025-11-03T17:44:00.4774387Z Entering 'third_party/fbgemm' 2025-11-03T17:44:00.4834149Z http.https://github.com/.extraheader 2025-11-03T17:44:00.4888344Z Entering 'third_party/fbgemm/external/asmjit' 2025-11-03T17:44:00.4945147Z http.https://github.com/.extraheader 2025-11-03T17:44:00.5006622Z Entering 'third_party/fbgemm/external/composable_kernel' 2025-11-03T17:44:00.5060805Z http.https://github.com/.extraheader 2025-11-03T17:44:00.5120111Z Entering 'third_party/fbgemm/external/cpuinfo' 2025-11-03T17:44:00.5174476Z http.https://github.com/.extraheader 2025-11-03T17:44:00.5230096Z Entering 'third_party/fbgemm/external/cutlass' 2025-11-03T17:44:00.5291410Z http.https://github.com/.extraheader 2025-11-03T17:44:00.5358267Z Entering 'third_party/fbgemm/external/googletest' 2025-11-03T17:44:00.5414737Z http.https://github.com/.extraheader 2025-11-03T17:44:00.5471570Z Entering 'third_party/fbgemm/external/hipify_torch' 2025-11-03T17:44:00.5532111Z http.https://github.com/.extraheader 2025-11-03T17:44:00.5589648Z Entering 'third_party/fbgemm/external/json' 2025-11-03T17:44:00.5638996Z http.https://github.com/.extraheader 2025-11-03T17:44:00.5706285Z Entering 'third_party/flash-attention' 2025-11-03T17:44:00.5764832Z http.https://github.com/.extraheader 2025-11-03T17:44:00.5819369Z Entering 'third_party/flash-attention/csrc/composable_kernel' 2025-11-03T17:44:00.5874918Z http.https://github.com/.extraheader 2025-11-03T17:44:00.5942046Z Entering 'third_party/flash-attention/csrc/cutlass' 2025-11-03T17:44:00.5994938Z http.https://github.com/.extraheader 2025-11-03T17:44:00.6061591Z Entering 'third_party/flatbuffers' 2025-11-03T17:44:00.6115409Z http.https://github.com/.extraheader 2025-11-03T17:44:00.6179014Z Entering 'third_party/fmt' 2025-11-03T17:44:00.6236425Z http.https://github.com/.extraheader 2025-11-03T17:44:00.6295410Z Entering 'third_party/gemmlowp/gemmlowp' 2025-11-03T17:44:00.6354864Z http.https://github.com/.extraheader 2025-11-03T17:44:00.6410068Z Entering 'third_party/gloo' 2025-11-03T17:44:00.6472896Z http.https://github.com/.extraheader 2025-11-03T17:44:00.6529722Z Entering 'third_party/googletest' 2025-11-03T17:44:00.6592115Z http.https://github.com/.extraheader 2025-11-03T17:44:00.6650184Z Entering 'third_party/ideep' 2025-11-03T17:44:00.6712359Z http.https://github.com/.extraheader 2025-11-03T17:44:00.6764942Z Entering 'third_party/ideep/mkl-dnn' 2025-11-03T17:44:00.6823099Z http.https://github.com/.extraheader 2025-11-03T17:44:00.6892991Z Entering 'third_party/ittapi' 2025-11-03T17:44:00.6953255Z http.https://github.com/.extraheader 2025-11-03T17:44:00.7010627Z Entering 'third_party/kineto' 2025-11-03T17:44:00.7070800Z http.https://github.com/.extraheader 2025-11-03T17:44:00.7122631Z Entering 'third_party/kineto/libkineto/third_party/dynolog' 2025-11-03T17:44:00.7175936Z http.https://github.com/.extraheader 2025-11-03T17:44:00.7232876Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/DCGM' 2025-11-03T17:44:00.7293034Z http.https://github.com/.extraheader 2025-11-03T17:44:00.7359204Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/cpr' 2025-11-03T17:44:00.7414997Z http.https://github.com/.extraheader 2025-11-03T17:44:00.7477923Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/fmt' 2025-11-03T17:44:00.7534483Z http.https://github.com/.extraheader 2025-11-03T17:44:00.7594265Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/gflags' 2025-11-03T17:44:00.7644272Z http.https://github.com/.extraheader 2025-11-03T17:44:00.7701059Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/gflags/doc' 2025-11-03T17:44:00.7755344Z http.https://github.com/.extraheader 2025-11-03T17:44:00.7827579Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/glog' 2025-11-03T17:44:00.7884131Z http.https://github.com/.extraheader 2025-11-03T17:44:00.7947541Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/googletest' 2025-11-03T17:44:00.8004119Z http.https://github.com/.extraheader 2025-11-03T17:44:00.8061201Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/json' 2025-11-03T17:44:00.8115311Z http.https://github.com/.extraheader 2025-11-03T17:44:00.8176852Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/pfs' 2025-11-03T17:44:00.8234917Z http.https://github.com/.extraheader 2025-11-03T17:44:00.8292280Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/prometheus-cpp' 2025-11-03T17:44:00.8351634Z http.https://github.com/.extraheader 2025-11-03T17:44:00.8408139Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/prometheus-cpp/3rdparty/civetweb' 2025-11-03T17:44:00.8471978Z http.https://github.com/.extraheader 2025-11-03T17:44:00.8535171Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/prometheus-cpp/3rdparty/googletest' 2025-11-03T17:44:00.8593732Z http.https://github.com/.extraheader 2025-11-03T17:44:00.8659179Z Entering 'third_party/kineto/libkineto/third_party/fmt' 2025-11-03T17:44:00.8714881Z http.https://github.com/.extraheader 2025-11-03T17:44:00.8776673Z Entering 'third_party/kineto/libkineto/third_party/googletest' 2025-11-03T17:44:00.8833819Z http.https://github.com/.extraheader 2025-11-03T17:44:00.8900371Z Entering 'third_party/kleidiai' 2025-11-03T17:44:00.8960648Z http.https://github.com/.extraheader 2025-11-03T17:44:00.9014092Z Entering 'third_party/mimalloc' 2025-11-03T17:44:00.9064776Z http.https://github.com/.extraheader 2025-11-03T17:44:00.9124356Z Entering 'third_party/nlohmann' 2025-11-03T17:44:00.9182870Z http.https://github.com/.extraheader 2025-11-03T17:44:00.9245999Z Entering 'third_party/onnx' 2025-11-03T17:44:00.9301821Z http.https://github.com/.extraheader 2025-11-03T17:44:00.9371945Z Entering 'third_party/onnx/third_party/pybind11' 2025-11-03T17:44:00.9432948Z http.https://github.com/.extraheader 2025-11-03T17:44:00.9504137Z Entering 'third_party/opentelemetry-cpp' 2025-11-03T17:44:00.9573835Z http.https://github.com/.extraheader 2025-11-03T17:44:00.9633893Z Entering 'third_party/opentelemetry-cpp/third_party/benchmark' 2025-11-03T17:44:00.9692598Z http.https://github.com/.extraheader 2025-11-03T17:44:00.9755126Z Entering 'third_party/opentelemetry-cpp/third_party/googletest' 2025-11-03T17:44:00.9804826Z http.https://github.com/.extraheader 2025-11-03T17:44:00.9855906Z Entering 'third_party/opentelemetry-cpp/third_party/ms-gsl' 2025-11-03T17:44:00.9905308Z http.https://github.com/.extraheader 2025-11-03T17:44:00.9959402Z Entering 'third_party/opentelemetry-cpp/third_party/nlohmann-json' 2025-11-03T17:44:01.0015354Z http.https://github.com/.extraheader 2025-11-03T17:44:01.0074894Z Entering 'third_party/opentelemetry-cpp/third_party/opentelemetry-proto' 2025-11-03T17:44:01.0128291Z http.https://github.com/.extraheader 2025-11-03T17:44:01.0187543Z Entering 'third_party/opentelemetry-cpp/third_party/opentracing-cpp' 2025-11-03T17:44:01.0243752Z http.https://github.com/.extraheader 2025-11-03T17:44:01.0301827Z Entering 'third_party/opentelemetry-cpp/third_party/prometheus-cpp' 2025-11-03T17:44:01.0354165Z http.https://github.com/.extraheader 2025-11-03T17:44:01.0416853Z Entering 'third_party/opentelemetry-cpp/third_party/prometheus-cpp/3rdparty/civetweb' 2025-11-03T17:44:01.0470694Z http.https://github.com/.extraheader 2025-11-03T17:44:01.0527036Z Entering 'third_party/opentelemetry-cpp/third_party/prometheus-cpp/3rdparty/googletest' 2025-11-03T17:44:01.0582239Z http.https://github.com/.extraheader 2025-11-03T17:44:01.0653588Z Entering 'third_party/opentelemetry-cpp/tools/vcpkg' 2025-11-03T17:44:01.0712650Z http.https://github.com/.extraheader 2025-11-03T17:44:01.0796877Z Entering 'third_party/pocketfft' 2025-11-03T17:44:01.0854679Z http.https://github.com/.extraheader 2025-11-03T17:44:01.0914628Z Entering 'third_party/protobuf' 2025-11-03T17:44:01.0971661Z http.https://github.com/.extraheader 2025-11-03T17:44:01.1028332Z Entering 'third_party/protobuf/third_party/benchmark' 2025-11-03T17:44:01.1084339Z http.https://github.com/.extraheader 2025-11-03T17:44:01.1148431Z Entering 'third_party/protobuf/third_party/googletest' 2025-11-03T17:44:01.1204966Z http.https://github.com/.extraheader 2025-11-03T17:44:01.1266209Z Entering 'third_party/psimd' 2025-11-03T17:44:01.1323797Z http.https://github.com/.extraheader 2025-11-03T17:44:01.1381622Z Entering 'third_party/pthreadpool' 2025-11-03T17:44:01.1434420Z http.https://github.com/.extraheader 2025-11-03T17:44:01.1498798Z Entering 'third_party/pybind11' 2025-11-03T17:44:01.1555416Z http.https://github.com/.extraheader 2025-11-03T17:44:01.1609787Z Entering 'third_party/python-peachpy' 2025-11-03T17:44:01.1665680Z http.https://github.com/.extraheader 2025-11-03T17:44:01.1725124Z Entering 'third_party/sleef' 2025-11-03T17:44:01.1781691Z http.https://github.com/.extraheader 2025-11-03T17:44:01.1839993Z Entering 'third_party/tensorpipe' 2025-11-03T17:44:01.1902068Z http.https://github.com/.extraheader 2025-11-03T17:44:01.1958843Z Entering 'third_party/tensorpipe/third_party/googletest' 2025-11-03T17:44:01.2015814Z http.https://github.com/.extraheader 2025-11-03T17:44:01.2077137Z Entering 'third_party/tensorpipe/third_party/libnop' 2025-11-03T17:44:01.2134338Z http.https://github.com/.extraheader 2025-11-03T17:44:01.2197764Z Entering 'third_party/tensorpipe/third_party/libuv' 2025-11-03T17:44:01.2254827Z http.https://github.com/.extraheader 2025-11-03T17:44:01.2318302Z Entering 'third_party/tensorpipe/third_party/pybind11' 2025-11-03T17:44:01.2374991Z http.https://github.com/.extraheader 2025-11-03T17:44:01.2434240Z Entering 'third_party/tensorpipe/third_party/pybind11/tools/clang' 2025-11-03T17:44:01.2492469Z http.https://github.com/.extraheader 2025-11-03T17:44:01.2675003Z A job completed hook has been configured by the self-hosted runner administrator 2025-11-03T17:44:01.2701208Z ##[group]Run '/home/ec2-user/runner-scripts/after_job.sh' 2025-11-03T17:44:01.2708141Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2025-11-03T17:44:01.2708433Z ##[endgroup] 2025-11-03T17:44:01.2838824Z [!ALERT!] Swap in detected! [!ALERT!] 2025-11-03T17:44:11.4590209Z [!ALERT!] Swap out detected [!ALERT!] 2025-11-03T17:44:29.1072447Z Cleaning up orphan processes