2025-10-10T01:17:19.2034896Z Current runner version: '2.328.0' 2025-10-10T01:17:19.2041261Z Runner name: 'i-0b2d5251fda19aeb9' 2025-10-10T01:17:19.2042106Z Runner group name: 'default' 2025-10-10T01:17:19.2042895Z Machine name: 'ip-10-0-6-129' 2025-10-10T01:17:19.2045947Z ##[group]GITHUB_TOKEN Permissions 2025-10-10T01:17:19.2048023Z Contents: read 2025-10-10T01:17:19.2048663Z Metadata: read 2025-10-10T01:17:19.2049262Z ##[endgroup] 2025-10-10T01:17:19.2051343Z Secret source: Actions 2025-10-10T01:17:19.2052169Z Prepare workflow directory 2025-10-10T01:17:19.2518927Z Prepare all required actions 2025-10-10T01:17:19.2553357Z Getting action download info 2025-10-10T01:17:19.5988869Z Download action repository 'pytorch/test-infra@main' (SHA:264eed5d70b428e3aa5c1a7c98e4330f866e183f) 2025-10-10T01:17:22.0416412Z Download action repository 'pytorch/pytorch@main' (SHA:7614338b69481d702c9f084ac15d9109c7cd3ef0) 2025-10-10T01:17:44.2526141Z Download action repository 'actions/setup-python@a26af69be951a213d495a4c3e4e4022e16d87065' (SHA:a26af69be951a213d495a4c3e4e4022e16d87065) 2025-10-10T01:17:44.6144985Z Download action repository 'aws-actions/configure-aws-credentials@ececac1a45f3b08a01d2dd070d28d111c5fe6722' (SHA:ececac1a45f3b08a01d2dd070d28d111c5fe6722) 2025-10-10T01:17:44.9154046Z Download action repository 'aws-actions/amazon-ecr-login@062b18b96a7aff071d4dc91bc00c4c1a7945b076' (SHA:062b18b96a7aff071d4dc91bc00c4c1a7945b076) 2025-10-10T01:17:45.0721411Z Download action repository 'seemethere/upload-artifact-s3@baba72d0712b404f646cebe0730933554ebce96a' (SHA:baba72d0712b404f646cebe0730933554ebce96a) 2025-10-10T01:17:45.3492248Z Getting action download info 2025-10-10T01:17:45.4736454Z Download action repository 'actions/checkout@v4' (SHA:08eba0b27e820071cde6df949e0beb9ba4906955) 2025-10-10T01:17:45.7851771Z Getting action download info 2025-10-10T01:17:45.9005198Z Download action repository 'nick-fields/retry@v3.0.0' (SHA:7152eba30c6575329ac0576536151aca5a72780e) 2025-10-10T01:17:46.1406444Z Getting action download info 2025-10-10T01:17:46.2442464Z Download action repository 'nick-fields/retry@3e91a01664abd3c5cd539100d10d33b9c5b68482' (SHA:3e91a01664abd3c5cd539100d10d33b9c5b68482) 2025-10-10T01:17:46.4282857Z Getting action download info 2025-10-10T01:17:46.5756243Z Uses: pytorch/pytorch/.github/workflows/_linux-test.yml@refs/heads/main (344e6365a0068c2d2847fcec0c55dd53291d475e) 2025-10-10T01:17:46.5759997Z ##[group] Inputs 2025-10-10T01:17:46.5760339Z build-environment: linux-jammy-cuda12.8-py3.12-gcc11 2025-10-10T01:17:46.5765533Z test-matrix: {"include": [{"config": "vllm_basic_correctness_test", "shard": 1, "num_shards": 1, "runner": "linux.g6.4xlarge.experimental.nvidia.gpu"}, {"config": "vllm_basic_models_test", "shard": 1, "num_shards": 1, "runner": "linux.g6.4xlarge.experimental.nvidia.gpu"}, {"config": "vllm_entrypoints_test", "shard": 1, "num_shards": 1, "runner": "linux.g6.4xlarge.experimental.nvidia.gpu"}, {"config": "vllm_regression_test", "shard": 1, "num_shards": 1, "runner": "linux.g6.4xlarge.experimental.nvidia.gpu"}, {"config": "vllm_multi_model_processor_test", "shard": 1, "num_shards": 1, "runner": "linux.g6.4xlarge.experimental.nvidia.gpu"}, {"config": "vllm_pytorch_compilation_unit_tests", "shard": 1, "num_shards": 1, "runner": "linux.g6.4xlarge.experimental.nvidia.gpu"}, {"config": "vllm_lora_28_failure_test", "shard": 1, "num_shards": 1, "runner": "linux.g6.4xlarge.experimental.nvidia.gpu"}, {"config": "vllm_multi_model_test_28_failure_test", "shard": 1, "num_shards": 1, "runner": "linux.g6.4xlarge.experimental.nvidia.gpu"}, {"config": "vllm_languagde_model_test_extended_generation_28_failure_test", "shard": 1, "num_shards": 1, "runner": "linux.g6.4xlarge.experimental.nvidia.gpu"}, {"config": "vllm_distributed_test_2_gpu_28_failure_test", "shard": 1, "num_shards": 1, "runner": "linux.g6.4xlarge.experimental.nvidia.gpu"}, {"config": "vllm_lora_test", "shard": 0, "num_shards": 4, "runner": "linux.g6.4xlarge.experimental.nvidia.gpu"}, {"config": "vllm_lora_test", "shard": 1, "num_shards": 4, "runner": "linux.g6.4xlarge.experimental.nvidia.gpu"}, {"config": "vllm_lora_test", "shard": 2, "num_shards": 4, "runner": "linux.g6.4xlarge.experimental.nvidia.gpu"}, {"config": "vllm_lora_test", "shard": 3, "num_shards": 4, "runner": "linux.g6.4xlarge.experimental.nvidia.gpu"}, {"config": "vllm_lora_tp_test_distributed", "shard": 1, "num_shards": 1, "runner": "linux.g6.12xlarge.nvidia.gpu"}, {"config": "vllm_distributed_test_28_failure_test", "shard": 1, "num_shards": 1, "runner": "linux.g6.12xlarge.nvidia.gpu"}]} 2025-10-10T01:17:46.5771044Z docker-image: 308535385114.dkr.ecr.us-east-1.amazonaws.com/pytorch/ci-image:pytorch-linux-jammy-cuda12.8-cudnn9-py3.12-gcc11-vllm-d8be0384e085f551506bd739678109fa0f5ee7ac 2025-10-10T01:17:46.5771678Z sync-tag: 2025-10-10T01:17:46.5772350Z timeout-minutes: 240 2025-10-10T01:17:46.5772567Z use-gha: 2025-10-10T01:17:46.5772735Z dashboard-tag: 2025-10-10T01:17:46.5772924Z s3-bucket: gha-artifacts 2025-10-10T01:17:46.5773129Z aws-role-to-assume: 2025-10-10T01:17:46.5773634Z disable-monitor: false 2025-10-10T01:17:46.5773881Z monitor-log-interval: 5 2025-10-10T01:17:46.5774099Z monitor-data-collect-interval: 1 2025-10-10T01:17:46.5774361Z ##[endgroup] 2025-10-10T01:17:46.5774741Z Complete job name: ci-vllm-test / test (vllm_distributed_test_28_failure_test, 1, 1, linux.g6.12xlarge.nvidia.gpu) 2025-10-10T01:17:46.6657743Z A job started hook has been configured by the self-hosted runner administrator 2025-10-10T01:17:46.6756863Z ##[group]Run '/home/ec2-user/runner-scripts/before_job.sh' 2025-10-10T01:17:46.6766836Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2025-10-10T01:17:46.6767372Z ##[endgroup] 2025-10-10T01:17:47.9966567Z Runner Type: linux.g6.12xlarge.nvidia.gpu 2025-10-10T01:17:47.9967027Z Instance Type: g6.12xlarge 2025-10-10T01:17:47.9967245Z AMI Name: unknown 2025-10-10T01:17:48.0008825Z AMI ID: ami-08982f1c5bf93d976 2025-10-10T01:17:53.0118737Z ##[group]Run pytorch/test-infra/.github/actions/setup-ssh@main 2025-10-10T01:17:53.0119100Z with: 2025-10-10T01:17:53.0119631Z github-secret: *** 2025-10-10T01:17:53.0120188Z instructions: All testing is done inside the container, to start an interactive session run: docker exec -it $(docker container ps --format '{{.ID}}') bash 2025-10-10T01:17:53.0121084Z activate-with-label: false 2025-10-10T01:17:53.0121406Z label: with-ssh 2025-10-10T01:17:53.0121679Z remove-existing-keys: true 2025-10-10T01:17:53.0121981Z fail-silently: true 2025-10-10T01:17:53.0122266Z env: 2025-10-10T01:17:53.0122501Z GIT_DEFAULT_BRANCH: main 2025-10-10T01:17:53.0122798Z ##[endgroup] 2025-10-10T01:17:53.1341345Z Please see https://github.com/pytorch/pytorch/wiki/Debugging-using-with-ssh-for-Github-Actions for more info. 2025-10-10T01:17:53.1342591Z Not on pull request and ciflow reference could not be extracted, skipping adding ssh keys 2025-10-10T01:17:53.1521985Z ##[group]Run pytorch/pytorch/.github/actions/checkout-pytorch@main 2025-10-10T01:17:53.1522350Z with: 2025-10-10T01:17:53.1522513Z no-sudo: true 2025-10-10T01:17:53.1522708Z submodules: recursive 2025-10-10T01:17:53.1522937Z fetch-depth: 0 2025-10-10T01:17:53.1523101Z env: 2025-10-10T01:17:53.1523262Z GIT_DEFAULT_BRANCH: main 2025-10-10T01:17:53.1523457Z ##[endgroup] 2025-10-10T01:17:53.1612485Z ##[group]Run echo "IN_CONTAINER_RUNNER=$(if [ -f /.inarc ] || [ -f /.incontainer ]; then echo true ; else echo false; fi)" >> "$GITHUB_OUTPUT" 2025-10-10T01:17:53.1613242Z echo "IN_CONTAINER_RUNNER=$(if [ -f /.inarc ] || [ -f /.incontainer ]; then echo true ; else echo false; fi)" >> "$GITHUB_OUTPUT" 2025-10-10T01:17:53.1625694Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2025-10-10T01:17:53.1625994Z env: 2025-10-10T01:17:53.1626199Z GIT_DEFAULT_BRANCH: main 2025-10-10T01:17:53.1626438Z ##[endgroup] 2025-10-10T01:17:53.1762535Z ##[group]Run # Use all available CPUs for fetching 2025-10-10T01:17:53.1762903Z # Use all available CPUs for fetching 2025-10-10T01:17:53.1763157Z cd "${GITHUB_WORKSPACE}" 2025-10-10T01:17:53.1763406Z git config --global fetch.parallel 0 2025-10-10T01:17:53.1763899Z git config --global submodule.fetchJobs 0 2025-10-10T01:17:53.1764143Z  2025-10-10T01:17:53.1764475Z # Clean workspace. The default checkout action should also do this, but 2025-10-10T01:17:53.1764836Z # do it here as well just in case 2025-10-10T01:17:53.1765071Z if [[ -d .git ]]; then 2025-10-10T01:17:53.1765305Z  if [ -z "${NO_SUDO}" ]; then 2025-10-10T01:17:53.1765534Z  sudo git clean -ffdx 2025-10-10T01:17:53.1765736Z  else 2025-10-10T01:17:53.1765902Z  git clean -ffdx 2025-10-10T01:17:53.1766094Z  fi 2025-10-10T01:17:53.1766250Z fi 2025-10-10T01:17:53.1774413Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2025-10-10T01:17:53.1774693Z env: 2025-10-10T01:17:53.1774856Z GIT_DEFAULT_BRANCH: main 2025-10-10T01:17:53.1775047Z NO_SUDO: true 2025-10-10T01:17:53.1775211Z ##[endgroup] 2025-10-10T01:17:53.1923478Z ##[group]Run actions/checkout@v4 2025-10-10T01:17:53.1923728Z with: 2025-10-10T01:17:53.1923924Z ref: 344e6365a0068c2d2847fcec0c55dd53291d475e 2025-10-10T01:17:53.1924564Z fetch-depth: 0 2025-10-10T01:17:53.1924850Z submodules: recursive 2025-10-10T01:17:53.1925061Z show-progress: false 2025-10-10T01:17:53.1925281Z repository: pytorch/pytorch 2025-10-10T01:17:53.1925613Z token: *** 2025-10-10T01:17:53.1925785Z ssh-strict: true 2025-10-10T01:17:53.1925963Z ssh-user: git 2025-10-10T01:17:53.1926152Z persist-credentials: true 2025-10-10T01:17:53.1926346Z clean: true 2025-10-10T01:17:53.1926533Z sparse-checkout-cone-mode: true 2025-10-10T01:17:53.1926763Z fetch-tags: false 2025-10-10T01:17:53.1926948Z lfs: false 2025-10-10T01:17:53.1927111Z set-safe-directory: true 2025-10-10T01:17:53.1927325Z env: 2025-10-10T01:17:53.1927485Z GIT_DEFAULT_BRANCH: main 2025-10-10T01:17:53.1927678Z ##[endgroup] 2025-10-10T01:17:53.2911166Z Syncing repository: pytorch/pytorch 2025-10-10T01:17:53.2912354Z ##[group]Getting Git version info 2025-10-10T01:17:53.2912745Z Working directory is '/home/ec2-user/actions-runner/_work/pytorch/pytorch' 2025-10-10T01:17:53.2913263Z [command]/usr/bin/git version 2025-10-10T01:17:53.3088238Z git version 2.50.1 2025-10-10T01:17:53.3129007Z ##[endgroup] 2025-10-10T01:17:53.3139531Z Copying '/home/ec2-user/.gitconfig' to '/home/ec2-user/actions-runner/_work/_temp/67aab3e9-a2fb-49d7-8f47-a959e61111ba/.gitconfig' 2025-10-10T01:17:53.3155641Z Temporarily overriding HOME='/home/ec2-user/actions-runner/_work/_temp/67aab3e9-a2fb-49d7-8f47-a959e61111ba' before making global git config changes 2025-10-10T01:17:53.3157364Z Adding repository directory to the temporary git global config as a safe directory 2025-10-10T01:17:53.3160393Z [command]/usr/bin/git config --global --add safe.directory /home/ec2-user/actions-runner/_work/pytorch/pytorch 2025-10-10T01:17:53.3216731Z Deleting the contents of '/home/ec2-user/actions-runner/_work/pytorch/pytorch' 2025-10-10T01:17:53.3219858Z ##[group]Initializing the repository 2025-10-10T01:17:53.3223786Z [command]/usr/bin/git init /home/ec2-user/actions-runner/_work/pytorch/pytorch 2025-10-10T01:17:53.3303133Z hint: Using 'master' as the name for the initial branch. This default branch name 2025-10-10T01:17:53.3304364Z hint: is subject to change. To configure the initial branch name to use in all 2025-10-10T01:17:53.3305511Z hint: of your new repositories, which will suppress this warning, call: 2025-10-10T01:17:53.3306348Z hint: 2025-10-10T01:17:53.3307025Z hint: git config --global init.defaultBranch 2025-10-10T01:17:53.3307736Z hint: 2025-10-10T01:17:53.3308308Z hint: Names commonly chosen instead of 'master' are 'main', 'trunk' and 2025-10-10T01:17:53.3309298Z hint: 'development'. The just-created branch can be renamed via this command: 2025-10-10T01:17:53.3310052Z hint: 2025-10-10T01:17:53.3310443Z hint: git branch -m 2025-10-10T01:17:53.3310892Z hint: 2025-10-10T01:17:53.3311512Z hint: Disable this message with "git config set advice.defaultBranchName false" 2025-10-10T01:17:53.3313536Z Initialized empty Git repository in /home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/ 2025-10-10T01:17:53.3324459Z [command]/usr/bin/git remote add origin https://github.com/pytorch/pytorch 2025-10-10T01:17:53.3378980Z ##[endgroup] 2025-10-10T01:17:53.3379778Z ##[group]Disabling automatic garbage collection 2025-10-10T01:17:53.3386861Z [command]/usr/bin/git config --local gc.auto 0 2025-10-10T01:17:53.3423640Z ##[endgroup] 2025-10-10T01:17:53.3424680Z ##[group]Setting up auth 2025-10-10T01:17:53.3430155Z [command]/usr/bin/git config --local --name-only --get-regexp core\.sshCommand 2025-10-10T01:17:53.3464083Z [command]/usr/bin/git submodule foreach --recursive sh -c "git config --local --name-only --get-regexp 'core\.sshCommand' && git config --local --unset-all 'core.sshCommand' || :" 2025-10-10T01:17:53.3936065Z [command]/usr/bin/git config --local --name-only --get-regexp http\.https\:\/\/github\.com\/\.extraheader 2025-10-10T01:17:53.3972654Z [command]/usr/bin/git submodule foreach --recursive sh -c "git config --local --name-only --get-regexp 'http\.https\:\/\/github\.com\/\.extraheader' && git config --local --unset-all 'http.https://github.com/.extraheader' || :" 2025-10-10T01:17:53.4400155Z [command]/usr/bin/git config --local http.https://github.com/.extraheader AUTHORIZATION: basic *** 2025-10-10T01:17:53.4465165Z ##[endgroup] 2025-10-10T01:17:53.4466074Z ##[group]Fetching the repository 2025-10-10T01:17:53.4472141Z [command]/usr/bin/git -c protocol.version=2 fetch --prune --no-recurse-submodules origin +refs/heads/*:refs/remotes/origin/* +refs/tags/*:refs/tags/* 2025-10-10T01:18:33.1522008Z From https://github.com/pytorch/pytorch 2025-10-10T01:18:33.1523170Z * [new branch] 2.6.0.dev20241004+ -> origin/2.6.0.dev20241004+ 2025-10-10T01:18:33.1524734Z * [new branch] AaronWang04_addmmfusion_perftest -> origin/AaronWang04_addmmfusion_perftest 2025-10-10T01:18:33.1526042Z * [new branch] BootcampDynamo -> origin/BootcampDynamo 2025-10-10T01:18:33.1527023Z * [new branch] DynamoFixGit -> origin/DynamoFixGit 2025-10-10T01:18:33.1527451Z * [new branch] DynamoVariaT -> origin/DynamoVariaT 2025-10-10T01:18:33.1527890Z * [new branch] HDCharles-2.6.0-release-notes -> origin/HDCharles-2.6.0-release-notes 2025-10-10T01:18:33.1528295Z * [new branch] ISSUE-154849 -> origin/ISSUE-154849 2025-10-10T01:18:33.1530234Z * [new branch] IvanKobzarev/stack/1 -> origin/IvanKobzarev/stack/1 2025-10-10T01:18:33.1531958Z * [new branch] IvanKobzarev/stack/2 -> origin/IvanKobzarev/stack/2 2025-10-10T01:18:33.1534259Z * [new branch] NicoshevSVE128 -> origin/NicoshevSVE128 2025-10-10T01:18:33.1535635Z * [new branch] PR-AOTInductorNoneBug -> origin/PR-AOTInductorNoneBug 2025-10-10T01:18:33.1537615Z * [new branch] PR-AOTInductorNoneBugFix -> origin/PR-AOTInductorNoneBugFix 2025-10-10T01:18:33.1539156Z * [new branch] PR-FixConfigsIssue -> origin/PR-FixConfigsIssue 2025-10-10T01:18:33.1541082Z * [new branch] PR-NoneBugFix-viable -> origin/PR-NoneBugFix-viable 2025-10-10T01:18:33.1542931Z * [new branch] PR-ResetToZero -> origin/PR-ResetToZero 2025-10-10T01:18:33.1544778Z * [new branch] Update-Flash-Packaging -> origin/Update-Flash-Packaging 2025-10-10T01:18:33.1547421Z * [new branch] VLA_exp -> origin/VLA_exp 2025-10-10T01:18:33.1549272Z * [new branch] actually-run-mps-aot-inductor -> origin/actually-run-mps-aot-inductor 2025-10-10T01:18:33.1551037Z * [new branch] add_compile_benchmarking -> origin/add_compile_benchmarking 2025-10-10T01:18:33.1552769Z * [new branch] add_op_tests -> origin/add_op_tests 2025-10-10T01:18:33.1554577Z * [new branch] add_op_to_dashboard -> origin/add_op_to_dashboard 2025-10-10T01:18:33.1556431Z * [new branch] addmm-heuristic -> origin/addmm-heuristic 2025-10-10T01:18:33.1558237Z * [new branch] addremovefunction -> origin/addremovefunction 2025-10-10T01:18:33.1560132Z * [new branch] addvllmtest -> origin/addvllmtest 2025-10-10T01:18:33.1562573Z * [new branch] adi/test -> origin/adi/test 2025-10-10T01:18:33.1564324Z * [new branch] adi/test_bgemm -> origin/adi/test_bgemm 2025-10-10T01:18:33.1566017Z * [new branch] adi/test_fusions -> origin/adi/test_fusions 2025-10-10T01:18:33.1567711Z * [new branch] adi/test_onednn -> origin/adi/test_onednn 2025-10-10T01:18:33.1570108Z * [new branch] adi/test_onednn_v3.9 -> origin/adi/test_onednn_v3.9 2025-10-10T01:18:33.1571041Z * [new branch] adi/test_presve_change -> origin/adi/test_presve_change 2025-10-10T01:18:33.1572802Z * [new branch] adi/test_timm -> origin/adi/test_timm 2025-10-10T01:18:33.1574824Z * [new branch] adi/testpresve_change -> origin/adi/testpresve_change 2025-10-10T01:18:33.1577828Z * [new branch] aditew01/test/vec_bf16 -> origin/aditew01/test/vec_bf16 2025-10-10T01:18:33.1579610Z * [new branch] ah-globalfeedback-hook -> origin/ah-globalfeedback-hook 2025-10-10T01:18:33.1581411Z * [new branch] alt-disable -> origin/alt-disable 2025-10-10T01:18:33.1583917Z * [new branch] angelayi/allow_fake -> origin/angelayi/allow_fake 2025-10-10T01:18:33.1585653Z * [new branch] angelayi/aoti_additional_files -> origin/angelayi/aoti_additional_files 2025-10-10T01:18:33.1587401Z * [new branch] angelayi/benchmark -> origin/angelayi/benchmark 2025-10-10T01:18:33.1589075Z * [new branch] angelayi/benchmark2 -> origin/angelayi/benchmark2 2025-10-10T01:18:33.1590815Z * [new branch] angelayi/benchmark3 -> origin/angelayi/benchmark3 2025-10-10T01:18:33.1592606Z * [new branch] angelayi/change_pytree_serialization -> origin/angelayi/change_pytree_serialization 2025-10-10T01:18:33.1594229Z * [new branch] angelayi/cpp_loader -> origin/angelayi/cpp_loader 2025-10-10T01:18:33.1596292Z * [new branch] angelayi/customop -> origin/angelayi/customop 2025-10-10T01:18:33.1598333Z * [new branch] angelayi/fix_mps -> origin/angelayi/fix_mps 2025-10-10T01:18:33.1600595Z * [new branch] angelayi/lint -> origin/angelayi/lint 2025-10-10T01:18:33.1602544Z * [new branch] angelayi/no_so_weight -> origin/angelayi/no_so_weight 2025-10-10T01:18:33.1604254Z * [new branch] angelayi/opaque_obj_v2 -> origin/angelayi/opaque_obj_v2 2025-10-10T01:18:33.1606014Z * [new branch] angelayi/pattern -> origin/angelayi/pattern 2025-10-10T01:18:33.1607854Z * [new branch] angelayi/pattern_in_out_2 -> origin/angelayi/pattern_in_out_2 2025-10-10T01:18:33.1609544Z * [new branch] angelayi/post_grad -> origin/angelayi/post_grad 2025-10-10T01:18:33.1611584Z * [new branch] angelayi/pytree -> origin/angelayi/pytree 2025-10-10T01:18:33.1613323Z * [new branch] angelayi/scan_layers -> origin/angelayi/scan_layers 2025-10-10T01:18:33.1615172Z * [new branch] angelayi/symint_input -> origin/angelayi/symint_input 2025-10-10T01:18:33.1616988Z * [new branch] angelayi/symm_mem -> origin/angelayi/symm_mem 2025-10-10T01:18:33.1618713Z * [new branch] angelayi/test_cpp -> origin/angelayi/test_cpp 2025-10-10T01:18:33.1620479Z * [new branch] angelayi/torch_size -> origin/angelayi/torch_size 2025-10-10T01:18:33.1622215Z * [new branch] angelayi/wrap_grad -> origin/angelayi/wrap_grad 2025-10-10T01:18:33.1623892Z * [new branch] annotate_1 -> origin/annotate_1 2025-10-10T01:18:33.1625952Z * [new branch] annotation_bw -> origin/annotation_bw 2025-10-10T01:18:33.1627528Z * [new branch] annotation_dynamo -> origin/annotation_dynamo 2025-10-10T01:18:33.1629301Z * [new branch] aot_eager_stack_trace -> origin/aot_eager_stack_trace 2025-10-10T01:18:33.1631069Z * [new branch] aoti-cuda-alloc -> origin/aoti-cuda-alloc 2025-10-10T01:18:33.1632772Z * [new branch] aoti_fqn_name_interface -> origin/aoti_fqn_name_interface 2025-10-10T01:18:33.1634439Z * [new branch] aoti_metal_shimify -> origin/aoti_metal_shimify 2025-10-10T01:18:33.1636246Z * [new branch] aoti_package_weights_binary -> origin/aoti_package_weights_binary 2025-10-10T01:18:33.1638002Z * [new branch] aoti_target_windows -> origin/aoti_target_windows 2025-10-10T01:18:33.1639851Z * [new branch] aoti_weight_sharing -> origin/aoti_weight_sharing 2025-10-10T01:18:33.1641604Z * [new branch] aoti_windows_mingw -> origin/aoti_windows_mingw 2025-10-10T01:18:33.1643338Z * [new branch] aoti_windows_mingw_2 -> origin/aoti_windows_mingw_2 2025-10-10T01:18:33.1646353Z * [new branch] arsh/feat/inductor_check_profiling -> origin/arsh/feat/inductor_check_profiling 2025-10-10T01:18:33.1648092Z * [new branch] async_tp -> origin/async_tp 2025-10-10T01:18:33.1649925Z * [new branch] atalman-inductor-perf-cu124 -> origin/atalman-inductor-perf-cu124 2025-10-10T01:18:33.1651717Z * [new branch] atalman-inductor-perf-cu124.1 -> origin/atalman-inductor-perf-cu124.1 2025-10-10T01:18:33.1653397Z * [new branch] atalman-patch-1 -> origin/atalman-patch-1 2025-10-10T01:18:33.1655167Z * [new branch] atalman-patch-2 -> origin/atalman-patch-2 2025-10-10T01:18:33.1657049Z * [new branch] atalman-patch-3 -> origin/atalman-patch-3 2025-10-10T01:18:33.1658798Z * [new branch] atalman-patch-4 -> origin/atalman-patch-4 2025-10-10T01:18:33.1660601Z * [new branch] atalman-patch-5 -> origin/atalman-patch-5 2025-10-10T01:18:33.1662504Z * [new branch] atalman-patch-6 -> origin/atalman-patch-6 2025-10-10T01:18:33.1664244Z * [new branch] atalman-patch-7 -> origin/atalman-patch-7 2025-10-10T01:18:33.1666270Z * [new branch] atalman_inductor_2.3.0 -> origin/atalman_inductor_2.3.0 2025-10-10T01:18:33.1667925Z * [new branch] atalman_inductor_2.3.1 -> origin/atalman_inductor_2.3.1 2025-10-10T01:18:33.1669645Z * [new branch] atalman_inductor_2.4.0 -> origin/atalman_inductor_2.4.0 2025-10-10T01:18:33.1671317Z * [new branch] atalman_inductor_2.4.x -> origin/atalman_inductor_2.4.x 2025-10-10T01:18:33.1673243Z * [new branch] attention_benchmark -> origin/attention_benchmark 2025-10-10T01:18:33.1675060Z * [new branch] attention_benchmarking_clean -> origin/attention_benchmarking_clean 2025-10-10T01:18:33.1676748Z * [new branch] b200_op_bench -> origin/b200_op_bench 2025-10-10T01:18:33.1679214Z * [new branch] bahuang/annotation -> origin/bahuang/annotation 2025-10-10T01:18:33.1680866Z * [new branch] bahuang/debug_mode -> origin/bahuang/debug_mode 2025-10-10T01:18:33.1682566Z * [new branch] bahuang/debug_mode_default -> origin/bahuang/debug_mode_default 2025-10-10T01:18:33.1684285Z * [new branch] bahuang/debug_mode_fix -> origin/bahuang/debug_mode_fix 2025-10-10T01:18:33.1685978Z * [new branch] bahuang/dt_fix_scalar_add -> origin/bahuang/dt_fix_scalar_add 2025-10-10T01:18:33.1687614Z * [new branch] bahuang/dt_reduce_mean -> origin/bahuang/dt_reduce_mean 2025-10-10T01:18:33.1689335Z * [new branch] bahuang/dtensor_demo -> origin/bahuang/dtensor_demo 2025-10-10T01:18:33.1691281Z * [new branch] bahuang/export_dtensor -> origin/bahuang/export_dtensor 2025-10-10T01:18:33.1693352Z * [new branch] bahuang/fix_debug_mode -> origin/bahuang/fix_debug_mode 2025-10-10T01:18:33.1695243Z * [new branch] bahuang/fix_debug_mode2 -> origin/bahuang/fix_debug_mode2 2025-10-10T01:18:33.1696919Z * [new branch] bahuang/fix_expand -> origin/bahuang/fix_expand 2025-10-10T01:18:33.1698722Z * [new branch] bahuang/noop_redistribute -> origin/bahuang/noop_redistribute 2025-10-10T01:18:33.1700518Z * [new branch] bahuang/reland -> origin/bahuang/reland 2025-10-10T01:18:33.1702373Z * [new branch] bahuang/reland_fake_export -> origin/bahuang/reland_fake_export 2025-10-10T01:18:33.1703947Z * [new branch] bahuang/rename -> origin/bahuang/rename 2025-10-10T01:18:33.1705734Z * [new branch] bahuang/test -> origin/bahuang/test 2025-10-10T01:18:33.1708167Z * [new branch] base/1.5 -> origin/base/1.5 2025-10-10T01:18:33.1710058Z * [new branch] batching_sdpa_efficient_attention -> origin/batching_sdpa_efficient_attention 2025-10-10T01:18:33.1711711Z * [new branch] bc-lint-test-new-config -> origin/bc-lint-test-new-config 2025-10-10T01:18:33.1713493Z * [new branch] benchmark-updates -> origin/benchmark-updates 2025-10-10T01:18:33.1715300Z * [new branch] benchmarking-script -> origin/benchmarking-script 2025-10-10T01:18:33.1717632Z * [new branch] bertmaher/pinbump26 -> origin/bertmaher/pinbump26 2025-10-10T01:18:33.1720125Z * [new branch] bertrand/cutlass -> origin/bertrand/cutlass 2025-10-10T01:18:33.1722504Z * [new branch] bf/cg-custom-wrapper -> origin/bf/cg-custom-wrapper 2025-10-10T01:18:33.1724417Z * [new branch] bf/cg-error-re-record -> origin/bf/cg-error-re-record 2025-10-10T01:18:33.1726484Z * [new branch] bf/cg-partition-custom-op-mutation -> origin/bf/cg-partition-custom-op-mutation 2025-10-10T01:18:33.1727961Z * [new branch] bf/cg-remove-check -> origin/bf/cg-remove-check 2025-10-10T01:18:33.1729645Z * [new branch] bf/cg-warn-dynamic-shapes -> origin/bf/cg-warn-dynamic-shapes 2025-10-10T01:18:33.1732178Z * [new branch] bf/cherry-pick-partition-share-default-device-context -> origin/bf/cherry-pick-partition-share-default-device-context 2025-10-10T01:18:33.1733444Z * [new branch] bf/clean-hf -> origin/bf/clean-hf 2025-10-10T01:18:33.1735365Z * [new branch] bf/clean-timm -> origin/bf/clean-timm 2025-10-10T01:18:33.1737285Z * [new branch] bf/clean-torchbench -> origin/bf/clean-torchbench 2025-10-10T01:18:33.1739110Z * [new branch] bf/clean-torchbench-hf -> origin/bf/clean-torchbench-hf 2025-10-10T01:18:33.1740775Z * [new branch] bf/cudagraph -> origin/bf/cudagraph 2025-10-10T01:18:33.1742644Z * [new branch] bf/cudagraph-disable-input-mutation -> origin/bf/cudagraph-disable-input-mutation 2025-10-10T01:18:33.1744621Z * [new branch] bf/cudagraph-enable-input-mutation-support-benchmark -> origin/bf/cudagraph-enable-input-mutation-support-benchmark 2025-10-10T01:18:33.1746186Z * [new branch] bf/cudagraph-partition -> origin/bf/cudagraph-partition 2025-10-10T01:18:33.1747996Z * [new branch] bf/donated-buffer-bench -> origin/bf/donated-buffer-bench 2025-10-10T01:18:33.1749768Z * [new branch] bf/minor-cg-config-doc -> origin/bf/minor-cg-config-doc 2025-10-10T01:18:33.1751586Z * [new branch] bf/minor-fa-tma-config -> origin/bf/minor-fa-tma-config 2025-10-10T01:18:33.1753256Z * [new branch] bf/pa-non-divisible -> origin/bf/pa-non-divisible 2025-10-10T01:18:33.1754994Z * [new branch] bf/partition-custom-op-alias -> origin/bf/partition-custom-op-alias 2025-10-10T01:18:33.1756842Z * [new branch] bf/partition-default-device-context -> origin/bf/partition-default-device-context 2025-10-10T01:18:33.1758603Z * [new branch] bf/partition-move-cpu -> origin/bf/partition-move-cpu 2025-10-10T01:18:33.1760389Z * [new branch] bf/remove-check-55b0c39d -> origin/bf/remove-check-55b0c39d 2025-10-10T01:18:33.1762101Z * [new branch] bf/rope -> origin/bf/rope 2025-10-10T01:18:33.1763891Z * [new branch] bf16_support -> origin/bf16_support 2025-10-10T01:18:33.1765616Z * [new branch] bf16_support_per_channel -> origin/bf16_support_per_channel 2025-10-10T01:18:33.1767305Z * [new branch] bisect_perf_hf_T5_3acc6eac492 -> origin/bisect_perf_hf_T5_3acc6eac492 2025-10-10T01:18:33.1768977Z * [new branch] bisect_perf_hf_T5_3fcf66f61fb -> origin/bisect_perf_hf_T5_3fcf66f61fb 2025-10-10T01:18:33.1770656Z * [new branch] bisect_perf_hf_T5_4009d154129 -> origin/bisect_perf_hf_T5_4009d154129 2025-10-10T01:18:33.1772363Z * [new branch] bisect_perf_hf_T5_40d0740e73d -> origin/bisect_perf_hf_T5_40d0740e73d 2025-10-10T01:18:33.1774037Z * [new branch] bisect_perf_hf_T5_5268754e -> origin/bisect_perf_hf_T5_5268754e 2025-10-10T01:18:33.1775729Z * [new branch] bisect_perf_hf_T5_7d89a8d385c -> origin/bisect_perf_hf_T5_7d89a8d385c 2025-10-10T01:18:33.1777458Z * [new branch] bisect_perf_hf_T5_b7a25c1ee7c -> origin/bisect_perf_hf_T5_b7a25c1ee7c 2025-10-10T01:18:33.1779100Z * [new branch] bisect_perf_hf_T5_c25b201583f -> origin/bisect_perf_hf_T5_c25b201583f 2025-10-10T01:18:33.1780848Z * [new branch] bisect_perf_hf_T5_c93e57efac0 -> origin/bisect_perf_hf_T5_c93e57efac0 2025-10-10T01:18:33.1782552Z * [new branch] bisect_perf_hf_T5_ca9813ea149 -> origin/bisect_perf_hf_T5_ca9813ea149 2025-10-10T01:18:33.1784214Z * [new branch] bisect_perf_hf_T5_d65f194a -> origin/bisect_perf_hf_T5_d65f194a 2025-10-10T01:18:33.1785859Z * [new branch] bisect_perf_hf_T5_da94ab0b -> origin/bisect_perf_hf_T5_da94ab0b 2025-10-10T01:18:33.1787668Z * [new branch] bisect_perf_hf_T5_da94ab0b_new -> origin/bisect_perf_hf_T5_da94ab0b_new 2025-10-10T01:18:33.1789368Z * [new branch] bisect_perf_hf_T5_db4e8a1d8a8 -> origin/bisect_perf_hf_T5_db4e8a1d8a8 2025-10-10T01:18:33.1791024Z * [new branch] bisect_perf_hf_T5_e0d97e936a2 -> origin/bisect_perf_hf_T5_e0d97e936a2 2025-10-10T01:18:33.1792699Z * [new branch] bisect_perf_hf_T5_f23621ec563 -> origin/bisect_perf_hf_T5_f23621ec563 2025-10-10T01:18:33.1795111Z * [new branch] bowbao/wip_prs -> origin/bowbao/wip_prs 2025-10-10T01:18:33.1797867Z * [new branch] brister/break_scatter_src_is_tensor -> origin/brister/break_scatter_src_is_tensor 2025-10-10T01:18:33.1799729Z * [new branch] brister/fx_cond -> origin/brister/fx_cond 2025-10-10T01:18:33.1801460Z * [new branch] brister/fx_dynamic_input -> origin/brister/fx_dynamic_input 2025-10-10T01:18:33.1803062Z * [new branch] brister/fx_index_put -> origin/brister/fx_index_put 2025-10-10T01:18:33.1804742Z * [new branch] brister/fx_no_python_slow -> origin/brister/fx_no_python_slow 2025-10-10T01:18:33.1806432Z * [new branch] brister/fx_scatter_reduce -> origin/brister/fx_scatter_reduce 2025-10-10T01:18:33.1808440Z * [new branch] brister/fx_unbacked_symbols -> origin/brister/fx_unbacked_symbols 2025-10-10T01:18:33.1810580Z * [new branch] brister/property_type_check -> origin/brister/property_type_check 2025-10-10T01:18:33.1812352Z * [new branch] brister/test_inductor_all_fx -> origin/brister/test_inductor_all_fx 2025-10-10T01:18:33.1814215Z * [new branch] brister/tiled_reduction_no_numel_check -> origin/brister/tiled_reduction_no_numel_check 2025-10-10T01:18:33.1815856Z * [new branch] build-aarch64-wheels -> origin/build-aarch64-wheels 2025-10-10T01:18:33.1817576Z * [new branch] bwd-backup -> origin/bwd-backup 2025-10-10T01:18:33.1819404Z * [new branch] c57382a49 -> origin/c57382a49 2025-10-10T01:18:33.1821086Z * [new branch] ca_0431d47eaa -> origin/ca_0431d47eaa 2025-10-10T01:18:33.1822766Z * [new branch] ca_fix_0431d47eaa -> origin/ca_fix_0431d47eaa 2025-10-10T01:18:33.1825744Z * [new branch] camyll/cherrypick_0098e5636d3afa7c75aef8c447a5c402ea9ed524 -> origin/camyll/cherrypick_0098e5636d3afa7c75aef8c447a5c402ea9ed524 2025-10-10T01:18:33.1827375Z * [new branch] camyll/cherrypick_3016616ccbba3dc9bb6a80eb4a81a846ddf49cc9 -> origin/camyll/cherrypick_3016616ccbba3dc9bb6a80eb4a81a846ddf49cc9 2025-10-10T01:18:33.1829788Z * [new branch] camyll/revert-94bc900da97ad7f3c35b3b819bb53b23c74b581a-for-release-2.8 -> origin/camyll/revert-94bc900da97ad7f3c35b3b819bb53b23c74b581a-for-release-2.8 2025-10-10T01:18:33.1831361Z * [new branch] camyll/revert_5d749ceb92c2c28bcfbdf918b4ab99b1a91fcb50 -> origin/camyll/revert_5d749ceb92c2c28bcfbdf918b4ab99b1a91fcb50 2025-10-10T01:18:33.1833882Z * [new branch] camyllh/cherrypick_5e7be988003a38be49227cfaa9bff6a2ea9e6929_v2 -> origin/camyllh/cherrypick_5e7be988003a38be49227cfaa9bff6a2ea9e6929_v2 2025-10-10T01:18:33.1835354Z * [new branch] camyllh/cherrypick_dda071587f0522a16b237f92cbe27fd13a1a1c11 -> origin/camyllh/cherrypick_dda071587f0522a16b237f92cbe27fd13a1a1c11 2025-10-10T01:18:33.1837845Z * [new branch] camyllh/release2_9_cherrypick/dda071587f0522a16b237f92cbe27fd13a1a1c11 -> origin/camyllh/release2_9_cherrypick/dda071587f0522a16b237f92cbe27fd13a1a1c11 2025-10-10T01:18:33.1839688Z * [new branch] camyllh/test_setup_hooks_push -> origin/camyllh/test_setup_hooks_push 2025-10-10T01:18:33.1841591Z * [new branch] cherry-pick-157453-by-pytorch_bot_bot_ -> origin/cherry-pick-157453-by-pytorch_bot_bot_ 2025-10-10T01:18:33.1843513Z * [new branch] cherry-pick-157513-by-pytorch_bot_bot_ -> origin/cherry-pick-157513-by-pytorch_bot_bot_ 2025-10-10T01:18:33.1844879Z * [new branch] cherry-pick-157695-by-pytorch_bot_bot_ -> origin/cherry-pick-157695-by-pytorch_bot_bot_ 2025-10-10T01:18:33.1846905Z * [new branch] cherry-pick-157732-by-pytorch_bot_bot_ -> origin/cherry-pick-157732-by-pytorch_bot_bot_ 2025-10-10T01:18:33.1848613Z * [new branch] cherry-pick-158537-by-pytorch_bot_bot_ -> origin/cherry-pick-158537-by-pytorch_bot_bot_ 2025-10-10T01:18:33.1850393Z * [new branch] cherry-pick-159969-by-pytorch_bot_bot_ -> origin/cherry-pick-159969-by-pytorch_bot_bot_ 2025-10-10T01:18:33.1852289Z * [new branch] cherry-pick-160586-by-pytorch_bot_bot_ -> origin/cherry-pick-160586-by-pytorch_bot_bot_ 2025-10-10T01:18:33.1854043Z * [new branch] cherry-pick-161299-by-pytorch_bot_bot_ -> origin/cherry-pick-161299-by-pytorch_bot_bot_ 2025-10-10T01:18:33.1855859Z * [new branch] cherry-pick-161394-by-pytorch_bot_bot_ -> origin/cherry-pick-161394-by-pytorch_bot_bot_ 2025-10-10T01:18:33.1857523Z * [new branch] cherry-pick-161430-by-pytorch_bot_bot_ -> origin/cherry-pick-161430-by-pytorch_bot_bot_ 2025-10-10T01:18:33.1859377Z * [new branch] cherry-pick-162168-by-pytorch_bot_bot_ -> origin/cherry-pick-162168-by-pytorch_bot_bot_ 2025-10-10T01:18:33.1861547Z * [new branch] cherry-pick-162194-by-pytorch_bot_bot_ -> origin/cherry-pick-162194-by-pytorch_bot_bot_ 2025-10-10T01:18:33.1863329Z * [new branch] cherry-pick-162240-by-pytorch_bot_bot_ -> origin/cherry-pick-162240-by-pytorch_bot_bot_ 2025-10-10T01:18:33.1865388Z * [new branch] cherry-pick-162295-by-pytorch_bot_bot_ -> origin/cherry-pick-162295-by-pytorch_bot_bot_ 2025-10-10T01:18:33.1867237Z * [new branch] cherry-pick-162323-by-pytorch_bot_bot_ -> origin/cherry-pick-162323-by-pytorch_bot_bot_ 2025-10-10T01:18:33.1868923Z * [new branch] cherry-pick-162425-by-pytorch_bot_bot_ -> origin/cherry-pick-162425-by-pytorch_bot_bot_ 2025-10-10T01:18:33.1870846Z * [new branch] cherry-pick-162530-by-pytorch_bot_bot_ -> origin/cherry-pick-162530-by-pytorch_bot_bot_ 2025-10-10T01:18:33.1872612Z * [new branch] cherry-pick-162555-by-pytorch_bot_bot_ -> origin/cherry-pick-162555-by-pytorch_bot_bot_ 2025-10-10T01:18:33.1874572Z * [new branch] cherry-pick-162566-by-pytorch_bot_bot_ -> origin/cherry-pick-162566-by-pytorch_bot_bot_ 2025-10-10T01:18:33.1876819Z * [new branch] cherry-pick-162587-by-pytorch_bot_bot_ -> origin/cherry-pick-162587-by-pytorch_bot_bot_ 2025-10-10T01:18:33.1878602Z * [new branch] cherry-pick-162622-by-pytorch_bot_bot_ -> origin/cherry-pick-162622-by-pytorch_bot_bot_ 2025-10-10T01:18:33.1880433Z * [new branch] cherry-pick-162657-by-pytorch_bot_bot_ -> origin/cherry-pick-162657-by-pytorch_bot_bot_ 2025-10-10T01:18:33.1882143Z * [new branch] cherry-pick-162680-by-pytorch_bot_bot_ -> origin/cherry-pick-162680-by-pytorch_bot_bot_ 2025-10-10T01:18:33.1884076Z * [new branch] cherry-pick-162693-by-pytorch_bot_bot_ -> origin/cherry-pick-162693-by-pytorch_bot_bot_ 2025-10-10T01:18:33.1885788Z * [new branch] cherry-pick-162744-by-pytorch_bot_bot_ -> origin/cherry-pick-162744-by-pytorch_bot_bot_ 2025-10-10T01:18:33.1887609Z * [new branch] cherry-pick-162764-by-pytorch_bot_bot_ -> origin/cherry-pick-162764-by-pytorch_bot_bot_ 2025-10-10T01:18:33.1889438Z * [new branch] cherry-pick-162865-by-pytorch_bot_bot_ -> origin/cherry-pick-162865-by-pytorch_bot_bot_ 2025-10-10T01:18:33.1891174Z * [new branch] cherry-pick-162866-by-pytorch_bot_bot_ -> origin/cherry-pick-162866-by-pytorch_bot_bot_ 2025-10-10T01:18:33.1892931Z * [new branch] cherry-pick-162877-by-pytorch_bot_bot_ -> origin/cherry-pick-162877-by-pytorch_bot_bot_ 2025-10-10T01:18:33.1894683Z * [new branch] cherry-pick-162950-by-pytorch_bot_bot_ -> origin/cherry-pick-162950-by-pytorch_bot_bot_ 2025-10-10T01:18:33.1896418Z * [new branch] cherry-pick-163008-by-pytorch_bot_bot_ -> origin/cherry-pick-163008-by-pytorch_bot_bot_ 2025-10-10T01:18:33.1898331Z * [new branch] cherry-pick-163111-by-pytorch_bot_bot_ -> origin/cherry-pick-163111-by-pytorch_bot_bot_ 2025-10-10T01:18:33.1900058Z * [new branch] cherry-pick-163112-by-pytorch_bot_bot_ -> origin/cherry-pick-163112-by-pytorch_bot_bot_ 2025-10-10T01:18:33.1901829Z * [new branch] cherry-pick-163152-by-pytorch_bot_bot_ -> origin/cherry-pick-163152-by-pytorch_bot_bot_ 2025-10-10T01:18:33.1903537Z * [new branch] cherry-pick-163171-by-pytorch_bot_bot_ -> origin/cherry-pick-163171-by-pytorch_bot_bot_ 2025-10-10T01:18:33.1905336Z * [new branch] cherry-pick-163194-by-pytorch_bot_bot_ -> origin/cherry-pick-163194-by-pytorch_bot_bot_ 2025-10-10T01:18:33.1907154Z * [new branch] cherry-pick-163227-by-pytorch_bot_bot_ -> origin/cherry-pick-163227-by-pytorch_bot_bot_ 2025-10-10T01:18:33.1908854Z * [new branch] cherry-pick-163269-by-pytorch_bot_bot_ -> origin/cherry-pick-163269-by-pytorch_bot_bot_ 2025-10-10T01:18:33.1911088Z * [new branch] cherry-pick-163298-by-pytorch_bot_bot_ -> origin/cherry-pick-163298-by-pytorch_bot_bot_ 2025-10-10T01:18:33.1912923Z * [new branch] cherry-pick-163315-by-pytorch_bot_bot_ -> origin/cherry-pick-163315-by-pytorch_bot_bot_ 2025-10-10T01:18:33.1914663Z * [new branch] cherry-pick-163339-by-pytorch_bot_bot_ -> origin/cherry-pick-163339-by-pytorch_bot_bot_ 2025-10-10T01:18:33.1916984Z * [new branch] cherry-pick-163341-by-pytorch_bot_bot_ -> origin/cherry-pick-163341-by-pytorch_bot_bot_ 2025-10-10T01:18:33.1918090Z * [new branch] cherry-pick-163370-by-pytorch_bot_bot_ -> origin/cherry-pick-163370-by-pytorch_bot_bot_ 2025-10-10T01:18:33.1919829Z * [new branch] cherry-pick-163383-by-pytorch_bot_bot_ -> origin/cherry-pick-163383-by-pytorch_bot_bot_ 2025-10-10T01:18:33.1921866Z * [new branch] cherry-pick-163426-by-pytorch_bot_bot_ -> origin/cherry-pick-163426-by-pytorch_bot_bot_ 2025-10-10T01:18:33.1924127Z * [new branch] cherry-pick-163549-by-pytorch_bot_bot_ -> origin/cherry-pick-163549-by-pytorch_bot_bot_ 2025-10-10T01:18:33.1925584Z * [new branch] cherry-pick-163571-by-pytorch_bot_bot_ -> origin/cherry-pick-163571-by-pytorch_bot_bot_ 2025-10-10T01:18:33.1927435Z * [new branch] cherry-pick-163578-by-pytorch_bot_bot_ -> origin/cherry-pick-163578-by-pytorch_bot_bot_ 2025-10-10T01:18:33.1929175Z * [new branch] cherry-pick-163581-by-pytorch_bot_bot_ -> origin/cherry-pick-163581-by-pytorch_bot_bot_ 2025-10-10T01:18:33.1930867Z * [new branch] cherry-pick-163585-by-pytorch_bot_bot_ -> origin/cherry-pick-163585-by-pytorch_bot_bot_ 2025-10-10T01:18:33.1932622Z * [new branch] cherry-pick-163587-by-pytorch_bot_bot_ -> origin/cherry-pick-163587-by-pytorch_bot_bot_ 2025-10-10T01:18:33.1934409Z * [new branch] cherry-pick-163598-by-pytorch_bot_bot_ -> origin/cherry-pick-163598-by-pytorch_bot_bot_ 2025-10-10T01:18:33.1936238Z * [new branch] cherry-pick-163661-by-pytorch_bot_bot_ -> origin/cherry-pick-163661-by-pytorch_bot_bot_ 2025-10-10T01:18:33.1938049Z * [new branch] cherry-pick-163677-by-pytorch_bot_bot_ -> origin/cherry-pick-163677-by-pytorch_bot_bot_ 2025-10-10T01:18:33.1939799Z * [new branch] cherry-pick-163682-by-pytorch_bot_bot_ -> origin/cherry-pick-163682-by-pytorch_bot_bot_ 2025-10-10T01:18:33.1941467Z * [new branch] cherry-pick-163712-by-pytorch_bot_bot_ -> origin/cherry-pick-163712-by-pytorch_bot_bot_ 2025-10-10T01:18:33.1943165Z * [new branch] cherry-pick-163719-by-pytorch_bot_bot_ -> origin/cherry-pick-163719-by-pytorch_bot_bot_ 2025-10-10T01:18:33.1945125Z * [new branch] cherry-pick-163768-by-pytorch_bot_bot_ -> origin/cherry-pick-163768-by-pytorch_bot_bot_ 2025-10-10T01:18:33.1946767Z * [new branch] cherry-pick-163776-by-pytorch_bot_bot_ -> origin/cherry-pick-163776-by-pytorch_bot_bot_ 2025-10-10T01:18:33.1948499Z * [new branch] cherry-pick-163797-by-pytorch_bot_bot_ -> origin/cherry-pick-163797-by-pytorch_bot_bot_ 2025-10-10T01:18:33.1950205Z * [new branch] cherry-pick-163837-by-pytorch_bot_bot_ -> origin/cherry-pick-163837-by-pytorch_bot_bot_ 2025-10-10T01:18:33.1951973Z * [new branch] cherry-pick-163886-by-pytorch_bot_bot_ -> origin/cherry-pick-163886-by-pytorch_bot_bot_ 2025-10-10T01:18:33.1953711Z * [new branch] cherry-pick-163903-by-pytorch_bot_bot_ -> origin/cherry-pick-163903-by-pytorch_bot_bot_ 2025-10-10T01:18:33.1955606Z * [new branch] cherry-pick-163956-by-pytorch_bot_bot_ -> origin/cherry-pick-163956-by-pytorch_bot_bot_ 2025-10-10T01:18:33.1957413Z * [new branch] cherry-pick-163988-by-pytorch_bot_bot_ -> origin/cherry-pick-163988-by-pytorch_bot_bot_ 2025-10-10T01:18:33.1959222Z * [new branch] cherry-pick-164093-by-pytorch_bot_bot_ -> origin/cherry-pick-164093-by-pytorch_bot_bot_ 2025-10-10T01:18:33.1961076Z * [new branch] cherry-pick-164108-by-pytorch_bot_bot_ -> origin/cherry-pick-164108-by-pytorch_bot_bot_ 2025-10-10T01:18:33.1962843Z * [new branch] cherry-pick-164138-by-pytorch_bot_bot_ -> origin/cherry-pick-164138-by-pytorch_bot_bot_ 2025-10-10T01:18:33.1964426Z * [new branch] cherry-pick-164190 -> origin/cherry-pick-164190 2025-10-10T01:18:33.1966383Z * [new branch] cherry-pick-164470-by-pytorch_bot_bot_ -> origin/cherry-pick-164470-by-pytorch_bot_bot_ 2025-10-10T01:18:33.1968298Z * [new branch] cherry-pick-164575-by-pytorch_bot_bot_ -> origin/cherry-pick-164575-by-pytorch_bot_bot_ 2025-10-10T01:18:33.1970074Z * [new branch] cherry-pick-164774-by-pytorch_bot_bot_ -> origin/cherry-pick-164774-by-pytorch_bot_bot_ 2025-10-10T01:18:33.1971986Z * [new branch] cherry-pick-164870-by-pytorch_bot_bot_ -> origin/cherry-pick-164870-by-pytorch_bot_bot_ 2025-10-10T01:18:33.1973737Z * [new branch] cherry-pick-164946-by-pytorch_bot_bot_ -> origin/cherry-pick-164946-by-pytorch_bot_bot_ 2025-10-10T01:18:33.1975554Z * [new branch] cherry-pick-165013-by-pytorch_bot_bot_ -> origin/cherry-pick-165013-by-pytorch_bot_bot_ 2025-10-10T01:18:33.1977152Z * [new branch] cherry_pick_graph_custom -> origin/cherry_pick_graph_custom 2025-10-10T01:18:33.1978906Z * [new branch] cherrypick-e88cca0691 -> origin/cherrypick-e88cca0691 2025-10-10T01:18:33.1980724Z * [new branch] chuanqi129-patch-1 -> origin/chuanqi129-patch-1 2025-10-10T01:18:33.1982424Z * [new branch] ck_dlpack -> origin/ck_dlpack 2025-10-10T01:18:33.1984287Z * [new branch] codegen_trace -> origin/codegen_trace 2025-10-10T01:18:33.1986279Z * [new branch] codex-testing -> origin/codex-testing 2025-10-10T01:18:33.1988799Z * [new branch] codex/add-metadata-field-for-file-path -> origin/codex/add-metadata-field-for-file-path 2025-10-10T01:18:33.1990985Z * [new branch] codex/add-test-for-inductor-local-cache-behavior -> origin/codex/add-test-for-inductor-local-cache-behavior 2025-10-10T01:18:33.1993048Z * [new branch] codex/create-test-for-tensor-memory-leak-in-cudagraph -> origin/codex/create-test-for-tensor-memory-leak-in-cudagraph 2025-10-10T01:18:33.1995080Z * [new branch] codex/enhance-cuda.matmul-with-allow_splitk-argument -> origin/codex/enhance-cuda.matmul-with-allow_splitk-argument 2025-10-10T01:18:33.1996643Z * [new branch] codex/fix-issue-121219-in-pytorch -> origin/codex/fix-issue-121219-in-pytorch 2025-10-10T01:18:33.1998617Z * [new branch] codex/refactor-dimension-handling-in-shape.cu -> origin/codex/refactor-dimension-handling-in-shape.cu 2025-10-10T01:18:33.2000520Z * [new branch] codex/refactor-lintrunner-config-to-use-uv-run -> origin/codex/refactor-lintrunner-config-to-use-uv-run 2025-10-10T01:18:33.2002002Z * [new branch] codex/remove-allow-untyped-defs-and-fix-type-errors -> origin/codex/remove-allow-untyped-defs-and-fix-type-errors 2025-10-10T01:18:33.2003962Z * [new branch] codex/remove-allow-untyped-defs-and-fix-type-errors-vx0jek -> origin/codex/remove-allow-untyped-defs-and-fix-type-errors-vx0jek 2025-10-10T01:18:33.2005258Z * [new branch] compile_kernel_include_dir -> origin/compile_kernel_include_dir 2025-10-10T01:18:33.2007123Z * [new branch] context_test -> origin/context_test 2025-10-10T01:18:33.2008841Z * [new branch] conv1d_decomp -> origin/conv1d_decomp 2025-10-10T01:18:33.2010568Z * [new branch] conv_autotune -> origin/conv_autotune 2025-10-10T01:18:33.2012996Z * [new branch] copilot/fix-157446 -> origin/copilot/fix-157446 2025-10-10T01:18:33.2014596Z * [new branch] copilot/fix-163730 -> origin/copilot/fix-163730 2025-10-10T01:18:33.2017009Z * [new branch] cpio/fix_new_ami_tests -> origin/cpio/fix_new_ami_tests 2025-10-10T01:18:33.2018814Z * [new branch] cpp-docs-dependency-upgrade -> origin/cpp-docs-dependency-upgrade 2025-10-10T01:18:33.2020568Z * [new branch] cpp_head -> origin/cpp_head 2025-10-10T01:18:33.2022363Z * [new branch] crcrpar-patch-1 -> origin/crcrpar-patch-1 2025-10-10T01:18:33.2024887Z * [new branch] csl/add_win_shard -> origin/csl/add_win_shard 2025-10-10T01:18:33.2026883Z * [new branch] csl/always_produce_xml -> origin/csl/always_produce_xml 2025-10-10T01:18:33.2028586Z * [new branch] csl/build_test_more_procs -> origin/csl/build_test_more_procs 2025-10-10T01:18:33.2030217Z * [new branch] csl/build_test_more_procs2 -> origin/csl/build_test_more_procs2 2025-10-10T01:18:33.2031914Z * [new branch] csl/fix_internal_graph_executor -> origin/csl/fix_internal_graph_executor 2025-10-10T01:18:33.2033819Z * [new branch] csl/fix_nightly_docs_push -> origin/csl/fix_nightly_docs_push 2025-10-10T01:18:33.2035938Z * [new branch] csl/inductor_h100_nightly -> origin/csl/inductor_h100_nightly 2025-10-10T01:18:33.2037537Z * [new branch] csl/katex -> origin/csl/katex 2025-10-10T01:18:33.2039423Z * [new branch] csl/larger_runner -> origin/csl/larger_runner 2025-10-10T01:18:33.2041162Z * [new branch] csl/lint_no_submodules -> origin/csl/lint_no_submodules 2025-10-10T01:18:33.2042788Z * [new branch] csl/lint_testing -> origin/csl/lint_testing 2025-10-10T01:18:33.2044501Z * [new branch] csl/lintrunner_stuff -> origin/csl/lintrunner_stuff 2025-10-10T01:18:33.2046288Z * [new branch] csl/mps_sharding -> origin/csl/mps_sharding 2025-10-10T01:18:33.2048226Z * [new branch] csl/multistage_docker -> origin/csl/multistage_docker 2025-10-10T01:18:33.2049887Z * [new branch] csl/no_keep_goin_rocm -> origin/csl/no_keep_goin_rocm 2025-10-10T01:18:33.2051642Z * [new branch] csl/reuse_old_whl_fix_metadata -> origin/csl/reuse_old_whl_fix_metadata 2025-10-10T01:18:33.2053265Z * [new branch] csl/revert_open -> origin/csl/revert_open 2025-10-10T01:18:33.2055055Z * [new branch] csl/skip_build -> origin/csl/skip_build 2025-10-10T01:18:33.2056770Z * [new branch] csl/smaller_avx_amx_runenrs -> origin/csl/smaller_avx_amx_runenrs 2025-10-10T01:18:33.2058474Z * [new branch] csl/test_cuda_build_large_runner -> origin/csl/test_cuda_build_large_runner 2025-10-10T01:18:33.2060406Z * [new branch] csl/test_info_status -> origin/csl/test_info_status 2025-10-10T01:18:33.2062012Z * [new branch] csl/test_info_upload_changes -> origin/csl/test_info_upload_changes 2025-10-10T01:18:33.2063694Z * [new branch] csl/test_owners_ao_sparse -> origin/csl/test_owners_ao_sparse 2025-10-10T01:18:33.2065526Z * [new branch] csl/test_owners_autograd_dispatch_nn -> origin/csl/test_owners_autograd_dispatch_nn 2025-10-10T01:18:33.2067056Z * [new branch] csl/test_owners_cuda -> origin/csl/test_owners_cuda 2025-10-10T01:18:33.2068790Z * [new branch] csl/test_owners_distributed -> origin/csl/test_owners_distributed 2025-10-10T01:18:33.2070623Z * [new branch] csl/test_owners_higher_confidence -> origin/csl/test_owners_higher_confidence 2025-10-10T01:18:33.2072255Z * [new branch] csl/testing_better_job_name -> origin/csl/testing_better_job_name 2025-10-10T01:18:33.2073978Z * [new branch] csl/vllm_pin_labeler -> origin/csl/vllm_pin_labeler 2025-10-10T01:18:33.2075664Z * [new branch] csl/win_cpp_tests -> origin/csl/win_cpp_tests 2025-10-10T01:18:33.2077437Z * [new branch] csl/win_sccache -> origin/csl/win_sccache 2025-10-10T01:18:33.2079254Z * [new branch] cu_stream_api -> origin/cu_stream_api 2025-10-10T01:18:33.2081011Z * [new branch] cublasltrelax2 -> origin/cublasltrelax2 2025-10-10T01:18:33.2082859Z * [new branch] cublasnowdeterministic -> origin/cublasnowdeterministic 2025-10-10T01:18:33.2084575Z * [new branch] cublasrelax2 -> origin/cublasrelax2 2025-10-10T01:18:33.2086399Z * [new branch] cuda-include-paths-fix -> origin/cuda-include-paths-fix 2025-10-10T01:18:33.2088173Z * [new branch] custom_lowering_dict -> origin/custom_lowering_dict 2025-10-10T01:18:33.2090514Z * [new branch] d4l3k/delete_hook -> origin/d4l3k/delete_hook 2025-10-10T01:18:33.2092843Z * [new branch] daxia6/2.8o3 -> origin/daxia6/2.8o3 2025-10-10T01:18:33.2094592Z * [new branch] dcp_zoc -> origin/dcp_zoc 2025-10-10T01:18:33.2096287Z * [new branch] debug-guard -> origin/debug-guard 2025-10-10T01:18:33.2098249Z * [new branch] delete-quant-docs -> origin/delete-quant-docs 2025-10-10T01:18:33.2103343Z * [new branch] dependabot/pip/dot-ci/docker/ci_commit_pins/main/transformers-4.57.0 -> origin/dependabot/pip/dot-ci/docker/ci_commit_pins/main/transformers-4.57.0 2025-10-10T01:18:33.2105503Z * [new branch] desertfire/test_cpp_wrapper -> origin/desertfire/test_cpp_wrapper 2025-10-10T01:18:33.2107219Z * [new branch] desertfire/triton-cpu-for-aarch64 -> origin/desertfire/triton-cpu-for-aarch64 2025-10-10T01:18:33.2109985Z * [new branch] dev/dhruva/flex_attn_opt -> origin/dev/dhruva/flex_attn_opt 2025-10-10T01:18:33.2112514Z * [new branch] dev/joona/MPSNDArrayAdd -> origin/dev/joona/MPSNDArrayAdd 2025-10-10T01:18:33.2114464Z * [new branch] dev/joona/Unranked -> origin/dev/joona/Unranked 2025-10-10T01:18:33.2116433Z * [new branch] dev/joona/cat -> origin/dev/joona/cat 2025-10-10T01:18:33.2118169Z * [new branch] dev/joona/embeddingbag -> origin/dev/joona/embeddingbag 2025-10-10T01:18:33.2120261Z * [new branch] dev/joona/getTensorsString -> origin/dev/joona/getTensorsString 2025-10-10T01:18:33.2122161Z * [new branch] dev/joona/maxpool2dwithindices_errmsg -> origin/dev/joona/maxpool2dwithindices_errmsg 2025-10-10T01:18:33.2124159Z * [new branch] dev/joona/mps_linear_macos14 -> origin/dev/joona/mps_linear_macos14 2025-10-10T01:18:33.2126051Z * [new branch] dev/joona/sdpa -> origin/dev/joona/sdpa 2025-10-10T01:18:33.2128398Z * [new branch] dev/joona/topk_newapi -> origin/dev/joona/topk_newapi 2025-10-10T01:18:33.2130589Z * [new branch] dev/joona/type_inf -> origin/dev/joona/type_inf 2025-10-10T01:18:33.2132537Z * [new branch] dev/joona/upsize3d -> origin/dev/joona/upsize3d 2025-10-10T01:18:33.2134520Z * [new branch] disable -> origin/disable 2025-10-10T01:18:33.2136190Z * [new branch] disp_counter -> origin/disp_counter 2025-10-10T01:18:33.2138049Z * [new branch] dtensor-issues -> origin/dtensor-issues 2025-10-10T01:18:33.2139828Z * [new branch] eager_model_benchmarks -> origin/eager_model_benchmarks 2025-10-10T01:18:33.2142169Z * [new branch] embg/test_inductor_ci_128B -> origin/embg/test_inductor_ci_128B 2025-10-10T01:18:33.2143852Z * [new branch] embg/test_inductor_ci_base -> origin/embg/test_inductor_ci_base 2025-10-10T01:18:33.2146214Z * [new branch] embg/test_inductor_ci_control -> origin/embg/test_inductor_ci_control 2025-10-10T01:18:33.2147623Z * [new branch] embg/triton_l2_prefetch_128B -> origin/embg/triton_l2_prefetch_128B 2025-10-10T01:18:33.2149607Z * [new branch] embg/triton_l2_prefetch_256B -> origin/embg/triton_l2_prefetch_256B 2025-10-10T01:18:33.2151416Z * [new branch] enable-keep-going-for-trunk-tags -> origin/enable-keep-going-for-trunk-tags 2025-10-10T01:18:33.2153478Z * [new branch] eqy-patch-3 -> origin/eqy-patch-3 2025-10-10T01:18:33.2155286Z * [new branch] eqy-patch-5 -> origin/eqy-patch-5 2025-10-10T01:18:33.2158003Z * [new branch] exclamaforte/amd-ma -> origin/exclamaforte/amd-ma 2025-10-10T01:18:33.2159902Z * [new branch] exclamaforte/combo-kernels-perf-run -> origin/exclamaforte/combo-kernels-perf-run 2025-10-10T01:18:33.2161478Z * [new branch] exclamaforte/do_bench_refactor -> origin/exclamaforte/do_bench_refactor 2025-10-10T01:18:33.2163159Z * [new branch] exclamaforte/enable-mem-dep-fusion -> origin/exclamaforte/enable-mem-dep-fusion 2025-10-10T01:18:33.2164790Z * [new branch] exclamaforte/fix-exhaustive-autotuning -> origin/exclamaforte/fix-exhaustive-autotuning 2025-10-10T01:18:33.2166826Z * [new branch] exclamaforte/fix-exhuastive-autotuning-reland -> origin/exclamaforte/fix-exhuastive-autotuning-reland 2025-10-10T01:18:33.2168877Z * [new branch] exclamaforte/fix-trace-parsing-fx-svg -> origin/exclamaforte/fix-trace-parsing-fx-svg 2025-10-10T01:18:33.2170562Z * [new branch] exclamaforte/force-pointwise-cat-perf-run -> origin/exclamaforte/force-pointwise-cat-perf-run 2025-10-10T01:18:33.2172241Z * [new branch] exclamaforte/fusion-data -> origin/exclamaforte/fusion-data 2025-10-10T01:18:33.2174059Z * [new branch] exclamaforte/gemm-benchmark-run -> origin/exclamaforte/gemm-benchmark-run 2025-10-10T01:18:33.2175734Z * [new branch] exclamaforte/gemm-export-model -> origin/exclamaforte/gemm-export-model 2025-10-10T01:18:33.2177363Z * [new branch] exclamaforte/gemm-model -> origin/exclamaforte/gemm-model 2025-10-10T01:18:33.2179708Z * [new branch] exclamaforte/gemm-model-all-data-collection -> origin/exclamaforte/gemm-model-all-data-collection 2025-10-10T01:18:33.2181285Z * [new branch] exclamaforte/gemm-to-amd -> origin/exclamaforte/gemm-to-amd 2025-10-10T01:18:33.2183086Z * [new branch] exclamaforte/just-gemm-model -> origin/exclamaforte/just-gemm-model 2025-10-10T01:18:33.2184919Z * [new branch] exclamaforte/just-gemm-model-no-refactor -> origin/exclamaforte/just-gemm-model-no-refactor 2025-10-10T01:18:33.2186655Z * [new branch] exclamaforte/profile-diff-algo -> origin/exclamaforte/profile-diff-algo 2025-10-10T01:18:33.2188542Z * [new branch] exclamaforte/profiler-visualization -> origin/exclamaforte/profiler-visualization 2025-10-10T01:18:33.2190131Z * [new branch] exclamaforte/test_cpp_wrapper_mode -> origin/exclamaforte/test_cpp_wrapper_mode 2025-10-10T01:18:33.2191863Z * [new branch] exclamaforte/update-autotune-configs -> origin/exclamaforte/update-autotune-configs 2025-10-10T01:18:33.2193529Z * [new branch] exclamaforte/update-autotune-configs-2 -> origin/exclamaforte/update-autotune-configs-2 2025-10-10T01:18:33.2195818Z * [new branch] exclamforte/gemm-model-final -> origin/exclamforte/gemm-model-final 2025-10-10T01:18:33.2197574Z * [new branch] exec -> origin/exec 2025-10-10T01:18:33.2199663Z * [new branch] experimental-mosaic -> origin/experimental-mosaic 2025-10-10T01:18:33.2201382Z * [new branch] export-D58091437 -> origin/export-D58091437 2025-10-10T01:18:33.2203286Z * [new branch] export-D61047529 -> origin/export-D61047529 2025-10-10T01:18:33.2205126Z * [new branch] export-D71412006 -> origin/export-D71412006 2025-10-10T01:18:33.2207073Z * [new branch] export-D73042989 -> origin/export-D73042989 2025-10-10T01:18:33.2208825Z * [new branch] export-D76797250 -> origin/export-D76797250 2025-10-10T01:18:33.2210560Z * [new branch] export-D76885271 -> origin/export-D76885271 2025-10-10T01:18:33.2212344Z * [new branch] export-D76885620 -> origin/export-D76885620 2025-10-10T01:18:33.2214076Z * [new branch] export-D76936623 -> origin/export-D76936623 2025-10-10T01:18:33.2215856Z * [new branch] export-D76958268 -> origin/export-D76958268 2025-10-10T01:18:33.2217617Z * [new branch] export-D78375400 -> origin/export-D78375400 2025-10-10T01:18:33.2219424Z * [new branch] export-D78431305 -> origin/export-D78431305 2025-10-10T01:18:33.2221231Z * [new branch] export-D78580107 -> origin/export-D78580107 2025-10-10T01:18:33.2222982Z * [new branch] export-D78822171 -> origin/export-D78822171 2025-10-10T01:18:33.2224975Z * [new branch] export-D78822351 -> origin/export-D78822351 2025-10-10T01:18:33.2226702Z * [new branch] export-D78822507 -> origin/export-D78822507 2025-10-10T01:18:33.2228485Z * [new branch] export-D78826994 -> origin/export-D78826994 2025-10-10T01:18:33.2230270Z * [new branch] export-D78894324 -> origin/export-D78894324 2025-10-10T01:18:33.2232059Z * [new branch] export-D78929245 -> origin/export-D78929245 2025-10-10T01:18:33.2233884Z * [new branch] export-D78934925 -> origin/export-D78934925 2025-10-10T01:18:33.2235623Z * [new branch] export-D78953203 -> origin/export-D78953203 2025-10-10T01:18:33.2237367Z * [new branch] export-D78953229 -> origin/export-D78953229 2025-10-10T01:18:33.2239142Z * [new branch] export-D78957093 -> origin/export-D78957093 2025-10-10T01:18:33.2240945Z * [new branch] export-D78957389 -> origin/export-D78957389 2025-10-10T01:18:33.2242698Z * [new branch] export-D78996107 -> origin/export-D78996107 2025-10-10T01:18:33.2244478Z * [new branch] export-D79026433 -> origin/export-D79026433 2025-10-10T01:18:33.2246341Z * [new branch] export-D79230339 -> origin/export-D79230339 2025-10-10T01:18:33.2248072Z * [new branch] export-D79319835 -> origin/export-D79319835 2025-10-10T01:18:33.2249787Z * [new branch] export-D79328456 -> origin/export-D79328456 2025-10-10T01:18:33.2251726Z * [new branch] export-D79378362 -> origin/export-D79378362 2025-10-10T01:18:33.2253386Z * [new branch] export-D80823877 -> origin/export-D80823877 2025-10-10T01:18:33.2255144Z * [new branch] export-D80948073 -> origin/export-D80948073 2025-10-10T01:18:33.2257416Z * [new branch] export-D80958642 -> origin/export-D80958642 2025-10-10T01:18:33.2259237Z * [new branch] export-D81054193 -> origin/export-D81054193 2025-10-10T01:18:33.2261037Z * [new branch] export-D81204584 -> origin/export-D81204584 2025-10-10T01:18:33.2262764Z * [new branch] export-D81429090 -> origin/export-D81429090 2025-10-10T01:18:33.2264641Z * [new branch] export-D81651226 -> origin/export-D81651226 2025-10-10T01:18:33.2266464Z * [new branch] export-D81698719 -> origin/export-D81698719 2025-10-10T01:18:33.2268231Z * [new branch] export-D82140619 -> origin/export-D82140619 2025-10-10T01:18:33.2270508Z * [new branch] export-D82174075 -> origin/export-D82174075 2025-10-10T01:18:33.2272454Z * [new branch] export-D82232574 -> origin/export-D82232574 2025-10-10T01:18:33.2274368Z * [new branch] export-D82250826 -> origin/export-D82250826 2025-10-10T01:18:33.2276124Z * [new branch] export-D82253817 -> origin/export-D82253817 2025-10-10T01:18:33.2277986Z * [new branch] export-D82380307 -> origin/export-D82380307 2025-10-10T01:18:33.2279957Z * [new branch] export-D82597111 -> origin/export-D82597111 2025-10-10T01:18:33.2281770Z * [new branch] export-D83023706 -> origin/export-D83023706 2025-10-10T01:18:33.2283486Z * [new branch] export-D83195687 -> origin/export-D83195687 2025-10-10T01:18:33.2298989Z * [new branch] export-D83200714 -> origin/export-D83200714 2025-10-10T01:18:33.2299597Z * [new branch] export-D83378477 -> origin/export-D83378477 2025-10-10T01:18:33.2299973Z * [new branch] export-D83390563 -> origin/export-D83390563 2025-10-10T01:18:33.2300351Z * [new branch] export-D83390775 -> origin/export-D83390775 2025-10-10T01:18:33.2300701Z * [new branch] export-D83391942 -> origin/export-D83391942 2025-10-10T01:18:33.2301052Z * [new branch] export-D83395610 -> origin/export-D83395610 2025-10-10T01:18:33.2302104Z * [new branch] export-D83539263 -> origin/export-D83539263 2025-10-10T01:18:33.2302457Z * [new branch] export-D83541846 -> origin/export-D83541846 2025-10-10T01:18:33.2302799Z * [new branch] export-D83591083 -> origin/export-D83591083 2025-10-10T01:18:33.2303149Z * [new branch] export-D83609850 -> origin/export-D83609850 2025-10-10T01:18:33.2303502Z * [new branch] export-D83627170 -> origin/export-D83627170 2025-10-10T01:18:33.2305329Z * [new branch] export-D83714690 -> origin/export-D83714690 2025-10-10T01:18:33.2307024Z * [new branch] export-D83766701 -> origin/export-D83766701 2025-10-10T01:18:33.2308837Z * [new branch] export-D83768878 -> origin/export-D83768878 2025-10-10T01:18:33.2310597Z * [new branch] export-D83769447 -> origin/export-D83769447 2025-10-10T01:18:33.2312467Z * [new branch] export-D84009392 -> origin/export-D84009392 2025-10-10T01:18:33.2314221Z * [new branch] export-D84089824 -> origin/export-D84089824 2025-10-10T01:18:33.2316111Z * [new branch] export-D84098898 -> origin/export-D84098898 2025-10-10T01:18:33.2318143Z * [new branch] export-D84103213 -> origin/export-D84103213 2025-10-10T01:18:33.2319896Z * [new branch] export-D84213020 -> origin/export-D84213020 2025-10-10T01:18:33.2321633Z * [new branch] export-reland -> origin/export-reland 2025-10-10T01:18:33.2323603Z * [new branch] exported-model-train-idempotent -> origin/exported-model-train-idempotent 2025-10-10T01:18:33.2325560Z * [new branch] extend_lift_up_op -> origin/extend_lift_up_op 2025-10-10T01:18:33.2327334Z * [new branch] ezyang-titan-october -> origin/ezyang-titan-october 2025-10-10T01:18:33.2329119Z * [new branch] ezyang-titan-october2 -> origin/ezyang-titan-october2 2025-10-10T01:18:33.2330879Z * [new branch] ezyang-war -> origin/ezyang-war 2025-10-10T01:18:33.2333222Z * [new branch] ezyang/wip-aot-descriptors -> origin/ezyang/wip-aot-descriptors 2025-10-10T01:18:33.2334962Z * [new branch] fa_u8_brgemm -> origin/fa_u8_brgemm 2025-10-10T01:18:33.2336915Z * [new branch] fadeputr-fix-fbgemm_genai-build -> origin/fadeputr-fix-fbgemm_genai-build 2025-10-10T01:18:33.2339221Z * [new branch] fadeputr/sequence_fbgemm -> origin/fadeputr/sequence_fbgemm 2025-10-10T01:18:33.2340966Z * [new branch] fastmath_baseline -> origin/fastmath_baseline 2025-10-10T01:18:33.2343382Z * [new branch] fbcode/warm -> origin/fbcode/warm 2025-10-10T01:18:33.2345248Z * [new branch] fca -> origin/fca 2025-10-10T01:18:33.2347007Z * [new branch] fca2_ca5984c -> origin/fca2_ca5984c 2025-10-10T01:18:33.2348840Z * [new branch] fca5 -> origin/fca5 2025-10-10T01:18:33.2351189Z * [new branch] feature/justknobs-cpp -> origin/feature/justknobs-cpp 2025-10-10T01:18:33.2353391Z * [new branch] ffast_math_baseline -> origin/ffast_math_baseline 2025-10-10T01:18:33.2355127Z * [new branch] ffast_math_target -> origin/ffast_math_target 2025-10-10T01:18:33.2357548Z * [new branch] findhao/base_commit -> origin/findhao/base_commit 2025-10-10T01:18:33.2359236Z * [new branch] findhao/base_commit1 -> origin/findhao/base_commit1 2025-10-10T01:18:33.2360987Z * [new branch] findhao/multistream2 -> origin/findhao/multistream2 2025-10-10T01:18:33.2362542Z * [new branch] findhao/multistream5 -> origin/findhao/multistream5 2025-10-10T01:18:33.2364165Z * [new branch] findhao/multistream6 -> origin/findhao/multistream6 2025-10-10T01:18:33.2365804Z * [new branch] findhao/operatorbench3 -> origin/findhao/operatorbench3 2025-10-10T01:18:33.2367567Z * [new branch] findhao/operatorbench5 -> origin/findhao/operatorbench5 2025-10-10T01:18:33.2369204Z * [new branch] findhao/tritonparse -> origin/findhao/tritonparse 2025-10-10T01:18:33.2371010Z * [new branch] fix-ck-gemm-template-format -> origin/fix-ck-gemm-template-format 2025-10-10T01:18:33.2372746Z * [new branch] fix-config-ignore -> origin/fix-config-ignore 2025-10-10T01:18:33.2374500Z * [new branch] fix-dict-guard -> origin/fix-dict-guard 2025-10-10T01:18:33.2376344Z * [new branch] fix-fqn -> origin/fix-fqn 2025-10-10T01:18:33.2378174Z * [new branch] fix-rlease-feature-template -> origin/fix-rlease-feature-template 2025-10-10T01:18:33.2380413Z * [new branch] fix-upload-vllm-wheel-credential -> origin/fix-upload-vllm-wheel-credential 2025-10-10T01:18:33.2382088Z * [new branch] fix_153389 -> origin/fix_153389 2025-10-10T01:18:33.2383893Z * [new branch] fix_nvrtc_discovery -> origin/fix_nvrtc_discovery 2025-10-10T01:18:33.2385759Z * [new branch] fix_op_benchmark -> origin/fix_op_benchmark 2025-10-10T01:18:33.2387347Z * [new branch] fix_op_runner -> origin/fix_op_runner 2025-10-10T01:18:33.2389065Z * [new branch] fix_ubn_159469 -> origin/fix_ubn_159469 2025-10-10T01:18:33.2390923Z * [new branch] fixes -> origin/fixes 2025-10-10T01:18:33.2392701Z * [new branch] fixes-triage -> origin/fixes-triage 2025-10-10T01:18:33.2394930Z * [new branch] fixflashgit -> origin/fixflashgit 2025-10-10T01:18:33.2396738Z * [new branch] fixflashinfer -> origin/fixflashinfer 2025-10-10T01:18:33.2398548Z * [new branch] flash_decoding_cpu -> origin/flash_decoding_cpu 2025-10-10T01:18:33.2400359Z * [new branch] flex-flash -> origin/flex-flash 2025-10-10T01:18:33.2402277Z * [new branch] flex_attention_functorch_grad -> origin/flex_attention_functorch_grad 2025-10-10T01:18:33.2404128Z * [new branch] flex_flash -> origin/flex_flash 2025-10-10T01:18:33.2406530Z * [new branch] fmassa/fix_memeff_sharding_rule -> origin/fmassa/fix_memeff_sharding_rule 2025-10-10T01:18:33.2408241Z * [new branch] free-stack2 -> origin/free-stack2 2025-10-10T01:18:33.2410706Z * [new branch] fsdp2_trace_rules -> origin/fsdp2_trace_rules 2025-10-10T01:18:33.2412638Z * [new branch] fsdpv2_3d -> origin/fsdpv2_3d 2025-10-10T01:18:33.2414453Z * [new branch] fsdpv2_3d_m1 -> origin/fsdpv2_3d_m1 2025-10-10T01:18:33.2416041Z * [new branch] fused_moving_avg_obs_fake_quant_half_support -> origin/fused_moving_avg_obs_fake_quant_half_support 2025-10-10T01:18:33.2417752Z * [new branch] fx_cpp -> origin/fx_cpp 2025-10-10T01:18:33.2420133Z * [new branch] fy/fix-win -> origin/fy/fix-win 2025-10-10T01:18:33.2423681Z * [new branch] gh/AlnisM/1/base -> origin/gh/AlnisM/1/base 2025-10-10T01:18:33.2426425Z * [new branch] gh/AlnisM/1/head -> origin/gh/AlnisM/1/head 2025-10-10T01:18:33.2429329Z * [new branch] gh/ColinPeppler/80/base -> origin/gh/ColinPeppler/80/base 2025-10-10T01:18:33.2431016Z * [new branch] gh/ColinPeppler/80/head -> origin/gh/ColinPeppler/80/head 2025-10-10T01:18:33.2432718Z * [new branch] gh/ColinPeppler/80/orig -> origin/gh/ColinPeppler/80/orig 2025-10-10T01:18:33.2435427Z * [new branch] gh/ColinPeppler/81/base -> origin/gh/ColinPeppler/81/base 2025-10-10T01:18:33.2437193Z * [new branch] gh/ColinPeppler/81/head -> origin/gh/ColinPeppler/81/head 2025-10-10T01:18:33.2439058Z * [new branch] gh/ColinPeppler/81/orig -> origin/gh/ColinPeppler/81/orig 2025-10-10T01:18:33.2441221Z * [new branch] gh/ColinPeppler/82/base -> origin/gh/ColinPeppler/82/base 2025-10-10T01:18:33.2442969Z * [new branch] gh/ColinPeppler/82/head -> origin/gh/ColinPeppler/82/head 2025-10-10T01:18:33.2444691Z * [new branch] gh/ColinPeppler/82/orig -> origin/gh/ColinPeppler/82/orig 2025-10-10T01:18:33.2447105Z * [new branch] gh/ColinPeppler/83/base -> origin/gh/ColinPeppler/83/base 2025-10-10T01:18:33.2448974Z * [new branch] gh/ColinPeppler/83/head -> origin/gh/ColinPeppler/83/head 2025-10-10T01:18:33.2450692Z * [new branch] gh/ColinPeppler/83/orig -> origin/gh/ColinPeppler/83/orig 2025-10-10T01:18:33.2453202Z * [new branch] gh/ColinPeppler/84/base -> origin/gh/ColinPeppler/84/base 2025-10-10T01:18:33.2454877Z * [new branch] gh/ColinPeppler/84/head -> origin/gh/ColinPeppler/84/head 2025-10-10T01:18:33.2457299Z * [new branch] gh/ColinPeppler/85/base -> origin/gh/ColinPeppler/85/base 2025-10-10T01:18:33.2458851Z * [new branch] gh/ColinPeppler/85/head -> origin/gh/ColinPeppler/85/head 2025-10-10T01:18:33.2460963Z * [new branch] gh/ColinPeppler/86/base -> origin/gh/ColinPeppler/86/base 2025-10-10T01:18:33.2462597Z * [new branch] gh/ColinPeppler/86/head -> origin/gh/ColinPeppler/86/head 2025-10-10T01:18:33.2464746Z * [new branch] gh/ColinPeppler/87/base -> origin/gh/ColinPeppler/87/base 2025-10-10T01:18:33.2466555Z * [new branch] gh/ColinPeppler/87/head -> origin/gh/ColinPeppler/87/head 2025-10-10T01:18:33.2468671Z * [new branch] gh/ColinPeppler/88/base -> origin/gh/ColinPeppler/88/base 2025-10-10T01:18:33.2470352Z * [new branch] gh/ColinPeppler/88/head -> origin/gh/ColinPeppler/88/head 2025-10-10T01:18:33.2472547Z * [new branch] gh/ColinPeppler/89/base -> origin/gh/ColinPeppler/89/base 2025-10-10T01:18:33.2474196Z * [new branch] gh/ColinPeppler/89/head -> origin/gh/ColinPeppler/89/head 2025-10-10T01:18:33.2476331Z * [new branch] gh/ColinPeppler/90/base -> origin/gh/ColinPeppler/90/base 2025-10-10T01:18:33.2478028Z * [new branch] gh/ColinPeppler/90/head -> origin/gh/ColinPeppler/90/head 2025-10-10T01:18:33.2480312Z * [new branch] gh/ColinPeppler/91/base -> origin/gh/ColinPeppler/91/base 2025-10-10T01:18:33.2481987Z * [new branch] gh/ColinPeppler/91/head -> origin/gh/ColinPeppler/91/head 2025-10-10T01:18:33.2484131Z * [new branch] gh/ColinPeppler/92/base -> origin/gh/ColinPeppler/92/base 2025-10-10T01:18:33.2485922Z * [new branch] gh/ColinPeppler/92/head -> origin/gh/ColinPeppler/92/head 2025-10-10T01:18:33.2488117Z * [new branch] gh/ColinPeppler/93/base -> origin/gh/ColinPeppler/93/base 2025-10-10T01:18:33.2489802Z * [new branch] gh/ColinPeppler/93/head -> origin/gh/ColinPeppler/93/head 2025-10-10T01:18:33.2491468Z * [new branch] gh/ColinPeppler/93/orig -> origin/gh/ColinPeppler/93/orig 2025-10-10T01:18:33.2494060Z * [new branch] gh/ColinPeppler/94/base -> origin/gh/ColinPeppler/94/base 2025-10-10T01:18:33.2495964Z * [new branch] gh/ColinPeppler/94/head -> origin/gh/ColinPeppler/94/head 2025-10-10T01:18:33.2497699Z * [new branch] gh/ColinPeppler/94/orig -> origin/gh/ColinPeppler/94/orig 2025-10-10T01:18:33.2500182Z * [new branch] gh/ColinPeppler/95/base -> origin/gh/ColinPeppler/95/base 2025-10-10T01:18:33.2502332Z * [new branch] gh/ColinPeppler/95/head -> origin/gh/ColinPeppler/95/head 2025-10-10T01:18:33.2504034Z * [new branch] gh/ColinPeppler/95/orig -> origin/gh/ColinPeppler/95/orig 2025-10-10T01:18:33.2507506Z * [new branch] gh/EikanWang/67/base -> origin/gh/EikanWang/67/base 2025-10-10T01:18:33.2508551Z * [new branch] gh/EikanWang/67/head -> origin/gh/EikanWang/67/head 2025-10-10T01:18:33.2512145Z * [new branch] gh/Gasoonjia/1/base -> origin/gh/Gasoonjia/1/base 2025-10-10T01:18:33.2513328Z * [new branch] gh/Gasoonjia/1/head -> origin/gh/Gasoonjia/1/head 2025-10-10T01:18:33.2516535Z * [new branch] gh/H-Huang/131/base -> origin/gh/H-Huang/131/base 2025-10-10T01:18:33.2518199Z * [new branch] gh/H-Huang/131/head -> origin/gh/H-Huang/131/head 2025-10-10T01:18:33.2520060Z * [new branch] gh/H-Huang/131/orig -> origin/gh/H-Huang/131/orig 2025-10-10T01:18:33.2522307Z * [new branch] gh/H-Huang/132/base -> origin/gh/H-Huang/132/base 2025-10-10T01:18:33.2524131Z * [new branch] gh/H-Huang/132/head -> origin/gh/H-Huang/132/head 2025-10-10T01:18:33.2527124Z * [new branch] gh/H-Huang/132/orig -> origin/gh/H-Huang/132/orig 2025-10-10T01:18:33.2528656Z * [new branch] gh/H-Huang/180/base -> origin/gh/H-Huang/180/base 2025-10-10T01:18:33.2530343Z * [new branch] gh/H-Huang/180/head -> origin/gh/H-Huang/180/head 2025-10-10T01:18:33.2532027Z * [new branch] gh/H-Huang/180/orig -> origin/gh/H-Huang/180/orig 2025-10-10T01:18:33.2534242Z * [new branch] gh/H-Huang/182/base -> origin/gh/H-Huang/182/base 2025-10-10T01:18:33.2536146Z * [new branch] gh/H-Huang/182/head -> origin/gh/H-Huang/182/head 2025-10-10T01:18:33.2537976Z * [new branch] gh/H-Huang/182/orig -> origin/gh/H-Huang/182/orig 2025-10-10T01:18:33.2540297Z * [new branch] gh/H-Huang/187/base -> origin/gh/H-Huang/187/base 2025-10-10T01:18:33.2541904Z * [new branch] gh/H-Huang/187/head -> origin/gh/H-Huang/187/head 2025-10-10T01:18:33.2543618Z * [new branch] gh/H-Huang/187/orig -> origin/gh/H-Huang/187/orig 2025-10-10T01:18:33.2545876Z * [new branch] gh/H-Huang/207/base -> origin/gh/H-Huang/207/base 2025-10-10T01:18:33.2547516Z * [new branch] gh/H-Huang/207/head -> origin/gh/H-Huang/207/head 2025-10-10T01:18:33.2549242Z * [new branch] gh/H-Huang/207/orig -> origin/gh/H-Huang/207/orig 2025-10-10T01:18:33.2551464Z * [new branch] gh/H-Huang/210/base -> origin/gh/H-Huang/210/base 2025-10-10T01:18:33.2553152Z * [new branch] gh/H-Huang/210/head -> origin/gh/H-Huang/210/head 2025-10-10T01:18:33.2554803Z * [new branch] gh/H-Huang/210/orig -> origin/gh/H-Huang/210/orig 2025-10-10T01:18:33.2557201Z * [new branch] gh/H-Huang/212/base -> origin/gh/H-Huang/212/base 2025-10-10T01:18:33.2558902Z * [new branch] gh/H-Huang/212/head -> origin/gh/H-Huang/212/head 2025-10-10T01:18:33.2560609Z * [new branch] gh/H-Huang/212/orig -> origin/gh/H-Huang/212/orig 2025-10-10T01:18:33.2562903Z * [new branch] gh/H-Huang/214/base -> origin/gh/H-Huang/214/base 2025-10-10T01:18:33.2564532Z * [new branch] gh/H-Huang/214/head -> origin/gh/H-Huang/214/head 2025-10-10T01:18:33.2566243Z * [new branch] gh/H-Huang/214/orig -> origin/gh/H-Huang/214/orig 2025-10-10T01:18:33.2568435Z * [new branch] gh/H-Huang/215/base -> origin/gh/H-Huang/215/base 2025-10-10T01:18:33.2570320Z * [new branch] gh/H-Huang/215/head -> origin/gh/H-Huang/215/head 2025-10-10T01:18:33.2571966Z * [new branch] gh/H-Huang/215/orig -> origin/gh/H-Huang/215/orig 2025-10-10T01:18:33.2574225Z * [new branch] gh/H-Huang/216/base -> origin/gh/H-Huang/216/base 2025-10-10T01:18:33.2575919Z * [new branch] gh/H-Huang/216/head -> origin/gh/H-Huang/216/head 2025-10-10T01:18:33.2577548Z * [new branch] gh/H-Huang/216/orig -> origin/gh/H-Huang/216/orig 2025-10-10T01:18:33.2579821Z * [new branch] gh/H-Huang/217/base -> origin/gh/H-Huang/217/base 2025-10-10T01:18:33.2581455Z * [new branch] gh/H-Huang/217/head -> origin/gh/H-Huang/217/head 2025-10-10T01:18:33.2583122Z * [new branch] gh/H-Huang/217/orig -> origin/gh/H-Huang/217/orig 2025-10-10T01:18:33.2585450Z * [new branch] gh/H-Huang/218/base -> origin/gh/H-Huang/218/base 2025-10-10T01:18:33.2587114Z * [new branch] gh/H-Huang/218/head -> origin/gh/H-Huang/218/head 2025-10-10T01:18:33.2588841Z * [new branch] gh/H-Huang/218/orig -> origin/gh/H-Huang/218/orig 2025-10-10T01:18:33.2591169Z * [new branch] gh/H-Huang/219/base -> origin/gh/H-Huang/219/base 2025-10-10T01:18:33.2592899Z * [new branch] gh/H-Huang/219/head -> origin/gh/H-Huang/219/head 2025-10-10T01:18:33.2594639Z * [new branch] gh/H-Huang/219/orig -> origin/gh/H-Huang/219/orig 2025-10-10T01:18:33.2596805Z * [new branch] gh/H-Huang/220/base -> origin/gh/H-Huang/220/base 2025-10-10T01:18:33.2598468Z * [new branch] gh/H-Huang/220/head -> origin/gh/H-Huang/220/head 2025-10-10T01:18:33.2600235Z * [new branch] gh/H-Huang/220/orig -> origin/gh/H-Huang/220/orig 2025-10-10T01:18:33.2602507Z * [new branch] gh/H-Huang/221/base -> origin/gh/H-Huang/221/base 2025-10-10T01:18:33.2604171Z * [new branch] gh/H-Huang/221/head -> origin/gh/H-Huang/221/head 2025-10-10T01:18:33.2605969Z * [new branch] gh/H-Huang/221/orig -> origin/gh/H-Huang/221/orig 2025-10-10T01:18:33.2608649Z * [new branch] gh/H-Huang/222/base -> origin/gh/H-Huang/222/base 2025-10-10T01:18:33.2610363Z * [new branch] gh/H-Huang/222/head -> origin/gh/H-Huang/222/head 2025-10-10T01:18:33.2612047Z * [new branch] gh/H-Huang/222/orig -> origin/gh/H-Huang/222/orig 2025-10-10T01:18:33.2614406Z * [new branch] gh/H-Huang/223/base -> origin/gh/H-Huang/223/base 2025-10-10T01:18:33.2616113Z * [new branch] gh/H-Huang/223/head -> origin/gh/H-Huang/223/head 2025-10-10T01:18:33.2617745Z * [new branch] gh/H-Huang/223/orig -> origin/gh/H-Huang/223/orig 2025-10-10T01:18:33.2620473Z * [new branch] gh/IvanKobzarev/115/base -> origin/gh/IvanKobzarev/115/base 2025-10-10T01:18:33.2622259Z * [new branch] gh/IvanKobzarev/115/head -> origin/gh/IvanKobzarev/115/head 2025-10-10T01:18:33.2624069Z * [new branch] gh/IvanKobzarev/115/orig -> origin/gh/IvanKobzarev/115/orig 2025-10-10T01:18:33.2627388Z * [new branch] gh/IvanKobzarev/116/base -> origin/gh/IvanKobzarev/116/base 2025-10-10T01:18:33.2628495Z * [new branch] gh/IvanKobzarev/116/head -> origin/gh/IvanKobzarev/116/head 2025-10-10T01:18:33.2630310Z * [new branch] gh/IvanKobzarev/116/orig -> origin/gh/IvanKobzarev/116/orig 2025-10-10T01:18:33.2632603Z * [new branch] gh/IvanKobzarev/118/base -> origin/gh/IvanKobzarev/118/base 2025-10-10T01:18:33.2634251Z * [new branch] gh/IvanKobzarev/118/head -> origin/gh/IvanKobzarev/118/head 2025-10-10T01:18:33.2636085Z * [new branch] gh/IvanKobzarev/118/orig -> origin/gh/IvanKobzarev/118/orig 2025-10-10T01:18:33.2638600Z * [new branch] gh/IvanKobzarev/126/base -> origin/gh/IvanKobzarev/126/base 2025-10-10T01:18:33.2640434Z * [new branch] gh/IvanKobzarev/126/head -> origin/gh/IvanKobzarev/126/head 2025-10-10T01:18:33.2642084Z * [new branch] gh/IvanKobzarev/126/orig -> origin/gh/IvanKobzarev/126/orig 2025-10-10T01:18:33.2644485Z * [new branch] gh/IvanKobzarev/127/base -> origin/gh/IvanKobzarev/127/base 2025-10-10T01:18:33.2646172Z * [new branch] gh/IvanKobzarev/127/head -> origin/gh/IvanKobzarev/127/head 2025-10-10T01:18:33.2647801Z * [new branch] gh/IvanKobzarev/127/orig -> origin/gh/IvanKobzarev/127/orig 2025-10-10T01:18:33.2650160Z * [new branch] gh/IvanKobzarev/128/base -> origin/gh/IvanKobzarev/128/base 2025-10-10T01:18:33.2651807Z * [new branch] gh/IvanKobzarev/128/head -> origin/gh/IvanKobzarev/128/head 2025-10-10T01:18:33.2653547Z * [new branch] gh/IvanKobzarev/128/orig -> origin/gh/IvanKobzarev/128/orig 2025-10-10T01:18:33.2656032Z * [new branch] gh/IvanKobzarev/135/base -> origin/gh/IvanKobzarev/135/base 2025-10-10T01:18:33.2657745Z * [new branch] gh/IvanKobzarev/135/head -> origin/gh/IvanKobzarev/135/head 2025-10-10T01:18:33.2659448Z * [new branch] gh/IvanKobzarev/135/orig -> origin/gh/IvanKobzarev/135/orig 2025-10-10T01:18:33.2661844Z * [new branch] gh/IvanKobzarev/138/base -> origin/gh/IvanKobzarev/138/base 2025-10-10T01:18:33.2663484Z * [new branch] gh/IvanKobzarev/138/head -> origin/gh/IvanKobzarev/138/head 2025-10-10T01:18:33.2665196Z * [new branch] gh/IvanKobzarev/138/orig -> origin/gh/IvanKobzarev/138/orig 2025-10-10T01:18:33.2667589Z * [new branch] gh/IvanKobzarev/141/base -> origin/gh/IvanKobzarev/141/base 2025-10-10T01:18:33.2669208Z * [new branch] gh/IvanKobzarev/141/head -> origin/gh/IvanKobzarev/141/head 2025-10-10T01:18:33.2670948Z * [new branch] gh/IvanKobzarev/141/orig -> origin/gh/IvanKobzarev/141/orig 2025-10-10T01:18:33.2673633Z * [new branch] gh/IvanKobzarev/142/base -> origin/gh/IvanKobzarev/142/base 2025-10-10T01:18:33.2675325Z * [new branch] gh/IvanKobzarev/142/head -> origin/gh/IvanKobzarev/142/head 2025-10-10T01:18:33.2676952Z * [new branch] gh/IvanKobzarev/142/orig -> origin/gh/IvanKobzarev/142/orig 2025-10-10T01:18:33.2679558Z * [new branch] gh/IvanKobzarev/144/base -> origin/gh/IvanKobzarev/144/base 2025-10-10T01:18:33.2681198Z * [new branch] gh/IvanKobzarev/144/head -> origin/gh/IvanKobzarev/144/head 2025-10-10T01:18:33.2682957Z * [new branch] gh/IvanKobzarev/144/orig -> origin/gh/IvanKobzarev/144/orig 2025-10-10T01:18:33.2685262Z * [new branch] gh/IvanKobzarev/145/base -> origin/gh/IvanKobzarev/145/base 2025-10-10T01:18:33.2687053Z * [new branch] gh/IvanKobzarev/145/head -> origin/gh/IvanKobzarev/145/head 2025-10-10T01:18:33.2688761Z * [new branch] gh/IvanKobzarev/145/orig -> origin/gh/IvanKobzarev/145/orig 2025-10-10T01:18:33.2691103Z * [new branch] gh/IvanKobzarev/146/base -> origin/gh/IvanKobzarev/146/base 2025-10-10T01:18:33.2692762Z * [new branch] gh/IvanKobzarev/146/head -> origin/gh/IvanKobzarev/146/head 2025-10-10T01:18:33.2694526Z * [new branch] gh/IvanKobzarev/146/orig -> origin/gh/IvanKobzarev/146/orig 2025-10-10T01:18:33.2696907Z * [new branch] gh/IvanKobzarev/147/base -> origin/gh/IvanKobzarev/147/base 2025-10-10T01:18:33.2698556Z * [new branch] gh/IvanKobzarev/147/head -> origin/gh/IvanKobzarev/147/head 2025-10-10T01:18:33.2700239Z * [new branch] gh/IvanKobzarev/147/orig -> origin/gh/IvanKobzarev/147/orig 2025-10-10T01:18:33.2702697Z * [new branch] gh/IvanKobzarev/148/base -> origin/gh/IvanKobzarev/148/base 2025-10-10T01:18:33.2704397Z * [new branch] gh/IvanKobzarev/148/head -> origin/gh/IvanKobzarev/148/head 2025-10-10T01:18:33.2706590Z * [new branch] gh/IvanKobzarev/149/base -> origin/gh/IvanKobzarev/149/base 2025-10-10T01:18:33.2708304Z * [new branch] gh/IvanKobzarev/149/head -> origin/gh/IvanKobzarev/149/head 2025-10-10T01:18:33.2710914Z * [new branch] gh/IvanKobzarev/150/base -> origin/gh/IvanKobzarev/150/base 2025-10-10T01:18:33.2712775Z * [new branch] gh/IvanKobzarev/150/head -> origin/gh/IvanKobzarev/150/head 2025-10-10T01:18:33.2714489Z * [new branch] gh/IvanKobzarev/150/orig -> origin/gh/IvanKobzarev/150/orig 2025-10-10T01:18:33.2716934Z * [new branch] gh/IvanKobzarev/151/base -> origin/gh/IvanKobzarev/151/base 2025-10-10T01:18:33.2718757Z * [new branch] gh/IvanKobzarev/151/head -> origin/gh/IvanKobzarev/151/head 2025-10-10T01:18:33.2720501Z * [new branch] gh/IvanKobzarev/151/orig -> origin/gh/IvanKobzarev/151/orig 2025-10-10T01:18:33.2722951Z * [new branch] gh/IvanKobzarev/152/base -> origin/gh/IvanKobzarev/152/base 2025-10-10T01:18:33.2724748Z * [new branch] gh/IvanKobzarev/152/head -> origin/gh/IvanKobzarev/152/head 2025-10-10T01:18:33.2726454Z * [new branch] gh/IvanKobzarev/152/orig -> origin/gh/IvanKobzarev/152/orig 2025-10-10T01:18:33.2729001Z * [new branch] gh/IvanKobzarev/153/base -> origin/gh/IvanKobzarev/153/base 2025-10-10T01:18:33.2730579Z * [new branch] gh/IvanKobzarev/153/head -> origin/gh/IvanKobzarev/153/head 2025-10-10T01:18:33.2732250Z * [new branch] gh/IvanKobzarev/153/orig -> origin/gh/IvanKobzarev/153/orig 2025-10-10T01:18:33.2734714Z * [new branch] gh/IvanKobzarev/154/base -> origin/gh/IvanKobzarev/154/base 2025-10-10T01:18:33.2736507Z * [new branch] gh/IvanKobzarev/154/head -> origin/gh/IvanKobzarev/154/head 2025-10-10T01:18:33.2738195Z * [new branch] gh/IvanKobzarev/154/orig -> origin/gh/IvanKobzarev/154/orig 2025-10-10T01:18:33.2740536Z * [new branch] gh/IvanKobzarev/155/base -> origin/gh/IvanKobzarev/155/base 2025-10-10T01:18:33.2742237Z * [new branch] gh/IvanKobzarev/155/head -> origin/gh/IvanKobzarev/155/head 2025-10-10T01:18:33.2744057Z * [new branch] gh/IvanKobzarev/155/orig -> origin/gh/IvanKobzarev/155/orig 2025-10-10T01:18:33.2747065Z * [new branch] gh/IvanKobzarev/156/base -> origin/gh/IvanKobzarev/156/base 2025-10-10T01:18:33.2748386Z * [new branch] gh/IvanKobzarev/156/head -> origin/gh/IvanKobzarev/156/head 2025-10-10T01:18:33.2750088Z * [new branch] gh/IvanKobzarev/156/orig -> origin/gh/IvanKobzarev/156/orig 2025-10-10T01:18:33.2752484Z * [new branch] gh/IvanKobzarev/157/base -> origin/gh/IvanKobzarev/157/base 2025-10-10T01:18:33.2754234Z * [new branch] gh/IvanKobzarev/157/head -> origin/gh/IvanKobzarev/157/head 2025-10-10T01:18:33.2756134Z * [new branch] gh/IvanKobzarev/157/orig -> origin/gh/IvanKobzarev/157/orig 2025-10-10T01:18:33.2758632Z * [new branch] gh/IvanKobzarev/158/base -> origin/gh/IvanKobzarev/158/base 2025-10-10T01:18:33.2760432Z * [new branch] gh/IvanKobzarev/158/head -> origin/gh/IvanKobzarev/158/head 2025-10-10T01:18:33.2762101Z * [new branch] gh/IvanKobzarev/158/orig -> origin/gh/IvanKobzarev/158/orig 2025-10-10T01:18:33.2764485Z * [new branch] gh/IvanKobzarev/159/base -> origin/gh/IvanKobzarev/159/base 2025-10-10T01:18:33.2766163Z * [new branch] gh/IvanKobzarev/159/head -> origin/gh/IvanKobzarev/159/head 2025-10-10T01:18:33.2767866Z * [new branch] gh/IvanKobzarev/159/orig -> origin/gh/IvanKobzarev/159/orig 2025-10-10T01:18:33.2770403Z * [new branch] gh/IvanKobzarev/160/base -> origin/gh/IvanKobzarev/160/base 2025-10-10T01:18:33.2772123Z * [new branch] gh/IvanKobzarev/160/head -> origin/gh/IvanKobzarev/160/head 2025-10-10T01:18:33.2773710Z * [new branch] gh/IvanKobzarev/160/orig -> origin/gh/IvanKobzarev/160/orig 2025-10-10T01:18:33.2776404Z * [new branch] gh/NikhilAPatel/1/base -> origin/gh/NikhilAPatel/1/base 2025-10-10T01:18:33.2778204Z * [new branch] gh/NikhilAPatel/1/head -> origin/gh/NikhilAPatel/1/head 2025-10-10T01:18:33.2780311Z * [new branch] gh/NikhilAPatel/2/base -> origin/gh/NikhilAPatel/2/base 2025-10-10T01:18:33.2781949Z * [new branch] gh/NikhilAPatel/2/head -> origin/gh/NikhilAPatel/2/head 2025-10-10T01:18:33.2784314Z * [new branch] gh/NikhilAPatel/4/base -> origin/gh/NikhilAPatel/4/base 2025-10-10T01:18:33.2786104Z * [new branch] gh/NikhilAPatel/4/head -> origin/gh/NikhilAPatel/4/head 2025-10-10T01:18:33.2788844Z * [new branch] gh/PaliC/1/base -> origin/gh/PaliC/1/base 2025-10-10T01:18:33.2790502Z * [new branch] gh/PaliC/1/head -> origin/gh/PaliC/1/head 2025-10-10T01:18:33.2792213Z * [new branch] gh/PaliC/1/orig -> origin/gh/PaliC/1/orig 2025-10-10T01:18:33.2794475Z * [new branch] gh/PaliC/17/base -> origin/gh/PaliC/17/base 2025-10-10T01:18:33.2796251Z * [new branch] gh/PaliC/17/head -> origin/gh/PaliC/17/head 2025-10-10T01:18:33.2797855Z * [new branch] gh/PaliC/17/orig -> origin/gh/PaliC/17/orig 2025-10-10T01:18:33.2800219Z * [new branch] gh/PaliC/18/base -> origin/gh/PaliC/18/base 2025-10-10T01:18:33.2801915Z * [new branch] gh/PaliC/18/head -> origin/gh/PaliC/18/head 2025-10-10T01:18:33.2803521Z * [new branch] gh/PaliC/18/orig -> origin/gh/PaliC/18/orig 2025-10-10T01:18:33.2805806Z * [new branch] gh/PaliC/2/base -> origin/gh/PaliC/2/base 2025-10-10T01:18:33.2807476Z * [new branch] gh/PaliC/2/head -> origin/gh/PaliC/2/head 2025-10-10T01:18:33.2809209Z * [new branch] gh/PaliC/2/orig -> origin/gh/PaliC/2/orig 2025-10-10T01:18:33.2811483Z * [new branch] gh/PaliC/20/base -> origin/gh/PaliC/20/base 2025-10-10T01:18:33.2813182Z * [new branch] gh/PaliC/20/head -> origin/gh/PaliC/20/head 2025-10-10T01:18:33.2814836Z * [new branch] gh/PaliC/20/orig -> origin/gh/PaliC/20/orig 2025-10-10T01:18:33.2817216Z * [new branch] gh/PaliC/21/base -> origin/gh/PaliC/21/base 2025-10-10T01:18:33.2825817Z * [new branch] gh/PaliC/21/head -> origin/gh/PaliC/21/head 2025-10-10T01:18:33.2826889Z * [new branch] gh/PaliC/21/orig -> origin/gh/PaliC/21/orig 2025-10-10T01:18:33.2827815Z * [new branch] gh/PaliC/22/base -> origin/gh/PaliC/22/base 2025-10-10T01:18:33.2828641Z * [new branch] gh/PaliC/22/head -> origin/gh/PaliC/22/head 2025-10-10T01:18:33.2829467Z * [new branch] gh/PaliC/22/orig -> origin/gh/PaliC/22/orig 2025-10-10T01:18:33.2830279Z * [new branch] gh/PaliC/23/base -> origin/gh/PaliC/23/base 2025-10-10T01:18:33.2831643Z * [new branch] gh/PaliC/23/head -> origin/gh/PaliC/23/head 2025-10-10T01:18:33.2833255Z * [new branch] gh/PaliC/23/orig -> origin/gh/PaliC/23/orig 2025-10-10T01:18:33.2835504Z * [new branch] gh/PaliC/24/base -> origin/gh/PaliC/24/base 2025-10-10T01:18:33.2837569Z * [new branch] gh/PaliC/24/head -> origin/gh/PaliC/24/head 2025-10-10T01:18:33.2839406Z * [new branch] gh/PaliC/24/orig -> origin/gh/PaliC/24/orig 2025-10-10T01:18:33.2841898Z * [new branch] gh/PaliC/25/head -> origin/gh/PaliC/25/head 2025-10-10T01:18:33.2843503Z * [new branch] gh/PaliC/25/next -> origin/gh/PaliC/25/next 2025-10-10T01:18:33.2845162Z * [new branch] gh/PaliC/25/orig -> origin/gh/PaliC/25/orig 2025-10-10T01:18:33.2847379Z * [new branch] gh/PaliC/26/head -> origin/gh/PaliC/26/head 2025-10-10T01:18:33.2848981Z * [new branch] gh/PaliC/26/next -> origin/gh/PaliC/26/next 2025-10-10T01:18:33.2850617Z * [new branch] gh/PaliC/26/orig -> origin/gh/PaliC/26/orig 2025-10-10T01:18:33.2852818Z * [new branch] gh/PaliC/27/head -> origin/gh/PaliC/27/head 2025-10-10T01:18:33.2854349Z * [new branch] gh/PaliC/27/next -> origin/gh/PaliC/27/next 2025-10-10T01:18:33.2856008Z * [new branch] gh/PaliC/27/orig -> origin/gh/PaliC/27/orig 2025-10-10T01:18:33.2858270Z * [new branch] gh/PaliC/28/head -> origin/gh/PaliC/28/head 2025-10-10T01:18:33.2859830Z * [new branch] gh/PaliC/28/next -> origin/gh/PaliC/28/next 2025-10-10T01:18:33.2861513Z * [new branch] gh/PaliC/28/orig -> origin/gh/PaliC/28/orig 2025-10-10T01:18:33.2863813Z * [new branch] gh/PaliC/29/head -> origin/gh/PaliC/29/head 2025-10-10T01:18:33.2865483Z * [new branch] gh/PaliC/29/next -> origin/gh/PaliC/29/next 2025-10-10T01:18:33.2867035Z * [new branch] gh/PaliC/29/orig -> origin/gh/PaliC/29/orig 2025-10-10T01:18:33.2869280Z * [new branch] gh/PaliC/30/head -> origin/gh/PaliC/30/head 2025-10-10T01:18:33.2870886Z * [new branch] gh/PaliC/30/next -> origin/gh/PaliC/30/next 2025-10-10T01:18:33.2872625Z * [new branch] gh/PaliC/30/orig -> origin/gh/PaliC/30/orig 2025-10-10T01:18:33.2874822Z * [new branch] gh/PaliC/31/head -> origin/gh/PaliC/31/head 2025-10-10T01:18:33.2876436Z * [new branch] gh/PaliC/31/next -> origin/gh/PaliC/31/next 2025-10-10T01:18:33.2878132Z * [new branch] gh/PaliC/31/orig -> origin/gh/PaliC/31/orig 2025-10-10T01:18:33.2881055Z * [new branch] gh/PaulZhang12/22/base -> origin/gh/PaulZhang12/22/base 2025-10-10T01:18:33.2882715Z * [new branch] gh/PaulZhang12/22/head -> origin/gh/PaulZhang12/22/head 2025-10-10T01:18:33.2884405Z * [new branch] gh/PaulZhang12/22/orig -> origin/gh/PaulZhang12/22/orig 2025-10-10T01:18:33.2886749Z * [new branch] gh/PaulZhang12/24/base -> origin/gh/PaulZhang12/24/base 2025-10-10T01:18:33.2888391Z * [new branch] gh/PaulZhang12/24/head -> origin/gh/PaulZhang12/24/head 2025-10-10T01:18:33.2890012Z * [new branch] gh/PaulZhang12/24/orig -> origin/gh/PaulZhang12/24/orig 2025-10-10T01:18:33.2892381Z * [new branch] gh/PaulZhang12/25/base -> origin/gh/PaulZhang12/25/base 2025-10-10T01:18:33.2894170Z * [new branch] gh/PaulZhang12/25/head -> origin/gh/PaulZhang12/25/head 2025-10-10T01:18:33.2895819Z * [new branch] gh/PaulZhang12/25/orig -> origin/gh/PaulZhang12/25/orig 2025-10-10T01:18:33.2898220Z * [new branch] gh/PaulZhang12/26/base -> origin/gh/PaulZhang12/26/base 2025-10-10T01:18:33.2899875Z * [new branch] gh/PaulZhang12/26/head -> origin/gh/PaulZhang12/26/head 2025-10-10T01:18:33.2901792Z * [new branch] gh/PaulZhang12/26/orig -> origin/gh/PaulZhang12/26/orig 2025-10-10T01:18:33.2904179Z * [new branch] gh/PaulZhang12/27/base -> origin/gh/PaulZhang12/27/base 2025-10-10T01:18:33.2906469Z * [new branch] gh/PaulZhang12/27/head -> origin/gh/PaulZhang12/27/head 2025-10-10T01:18:33.2907683Z * [new branch] gh/PaulZhang12/27/orig -> origin/gh/PaulZhang12/27/orig 2025-10-10T01:18:33.2909967Z * [new branch] gh/PaulZhang12/28/base -> origin/gh/PaulZhang12/28/base 2025-10-10T01:18:33.2911646Z * [new branch] gh/PaulZhang12/28/head -> origin/gh/PaulZhang12/28/head 2025-10-10T01:18:33.2913331Z * [new branch] gh/PaulZhang12/28/orig -> origin/gh/PaulZhang12/28/orig 2025-10-10T01:18:33.2915882Z * [new branch] gh/PaulZhang12/29/base -> origin/gh/PaulZhang12/29/base 2025-10-10T01:18:33.2917573Z * [new branch] gh/PaulZhang12/29/head -> origin/gh/PaulZhang12/29/head 2025-10-10T01:18:33.2919499Z * [new branch] gh/PaulZhang12/29/orig -> origin/gh/PaulZhang12/29/orig 2025-10-10T01:18:33.2921879Z * [new branch] gh/PaulZhang12/30/base -> origin/gh/PaulZhang12/30/base 2025-10-10T01:18:33.2923592Z * [new branch] gh/PaulZhang12/30/head -> origin/gh/PaulZhang12/30/head 2025-10-10T01:18:33.2925541Z * [new branch] gh/PaulZhang12/30/orig -> origin/gh/PaulZhang12/30/orig 2025-10-10T01:18:33.2927927Z * [new branch] gh/PaulZhang12/31/base -> origin/gh/PaulZhang12/31/base 2025-10-10T01:18:33.2929806Z * [new branch] gh/PaulZhang12/31/head -> origin/gh/PaulZhang12/31/head 2025-10-10T01:18:33.2931314Z * [new branch] gh/PaulZhang12/31/orig -> origin/gh/PaulZhang12/31/orig 2025-10-10T01:18:33.2933845Z * [new branch] gh/PaulZhang12/32/base -> origin/gh/PaulZhang12/32/base 2025-10-10T01:18:33.2935435Z * [new branch] gh/PaulZhang12/32/head -> origin/gh/PaulZhang12/32/head 2025-10-10T01:18:33.2937089Z * [new branch] gh/PaulZhang12/32/orig -> origin/gh/PaulZhang12/32/orig 2025-10-10T01:18:33.2940494Z * [new branch] gh/PaulZhang12/33/base -> origin/gh/PaulZhang12/33/base 2025-10-10T01:18:33.2942728Z * [new branch] gh/PaulZhang12/33/head -> origin/gh/PaulZhang12/33/head 2025-10-10T01:18:33.2944429Z * [new branch] gh/PaulZhang12/33/orig -> origin/gh/PaulZhang12/33/orig 2025-10-10T01:18:33.2946848Z * [new branch] gh/PaulZhang12/34/base -> origin/gh/PaulZhang12/34/base 2025-10-10T01:18:33.2948509Z * [new branch] gh/PaulZhang12/34/head -> origin/gh/PaulZhang12/34/head 2025-10-10T01:18:33.2950267Z * [new branch] gh/PaulZhang12/34/orig -> origin/gh/PaulZhang12/34/orig 2025-10-10T01:18:33.2952425Z * [new branch] gh/PaulZhang12/35/base -> origin/gh/PaulZhang12/35/base 2025-10-10T01:18:33.2954125Z * [new branch] gh/PaulZhang12/35/head -> origin/gh/PaulZhang12/35/head 2025-10-10T01:18:33.2955826Z * [new branch] gh/PaulZhang12/35/orig -> origin/gh/PaulZhang12/35/orig 2025-10-10T01:18:33.2958516Z * [new branch] gh/SamGinzburg/11/base -> origin/gh/SamGinzburg/11/base 2025-10-10T01:18:33.2960253Z * [new branch] gh/SamGinzburg/11/head -> origin/gh/SamGinzburg/11/head 2025-10-10T01:18:33.2963758Z * [new branch] gh/SherlockNoMad/1/base -> origin/gh/SherlockNoMad/1/base 2025-10-10T01:18:33.2965826Z * [new branch] gh/SherlockNoMad/1/head -> origin/gh/SherlockNoMad/1/head 2025-10-10T01:18:33.2967933Z * [new branch] gh/SherlockNoMad/10/base -> origin/gh/SherlockNoMad/10/base 2025-10-10T01:18:33.2969629Z * [new branch] gh/SherlockNoMad/10/head -> origin/gh/SherlockNoMad/10/head 2025-10-10T01:18:33.2971580Z * [new branch] gh/SherlockNoMad/10/orig -> origin/gh/SherlockNoMad/10/orig 2025-10-10T01:18:33.2973655Z * [new branch] gh/SherlockNoMad/11/base -> origin/gh/SherlockNoMad/11/base 2025-10-10T01:18:33.2975429Z * [new branch] gh/SherlockNoMad/11/head -> origin/gh/SherlockNoMad/11/head 2025-10-10T01:18:33.2977052Z * [new branch] gh/SherlockNoMad/11/orig -> origin/gh/SherlockNoMad/11/orig 2025-10-10T01:18:33.2979467Z * [new branch] gh/SherlockNoMad/12/base -> origin/gh/SherlockNoMad/12/base 2025-10-10T01:18:33.2981146Z * [new branch] gh/SherlockNoMad/12/head -> origin/gh/SherlockNoMad/12/head 2025-10-10T01:18:33.2982799Z * [new branch] gh/SherlockNoMad/12/orig -> origin/gh/SherlockNoMad/12/orig 2025-10-10T01:18:33.2984952Z * [new branch] gh/SherlockNoMad/13/base -> origin/gh/SherlockNoMad/13/base 2025-10-10T01:18:33.2986797Z * [new branch] gh/SherlockNoMad/13/head -> origin/gh/SherlockNoMad/13/head 2025-10-10T01:18:33.2988560Z * [new branch] gh/SherlockNoMad/13/orig -> origin/gh/SherlockNoMad/13/orig 2025-10-10T01:18:33.2990723Z * [new branch] gh/SherlockNoMad/2/base -> origin/gh/SherlockNoMad/2/base 2025-10-10T01:18:33.2992306Z * [new branch] gh/SherlockNoMad/2/head -> origin/gh/SherlockNoMad/2/head 2025-10-10T01:18:33.2994412Z * [new branch] gh/SherlockNoMad/3/base -> origin/gh/SherlockNoMad/3/base 2025-10-10T01:18:33.2996048Z * [new branch] gh/SherlockNoMad/3/head -> origin/gh/SherlockNoMad/3/head 2025-10-10T01:18:33.2998205Z * [new branch] gh/SherlockNoMad/4/base -> origin/gh/SherlockNoMad/4/base 2025-10-10T01:18:33.2999967Z * [new branch] gh/SherlockNoMad/4/head -> origin/gh/SherlockNoMad/4/head 2025-10-10T01:18:33.3002184Z * [new branch] gh/SherlockNoMad/5/base -> origin/gh/SherlockNoMad/5/base 2025-10-10T01:18:33.3003723Z * [new branch] gh/SherlockNoMad/5/head -> origin/gh/SherlockNoMad/5/head 2025-10-10T01:18:33.3005853Z * [new branch] gh/SherlockNoMad/6/base -> origin/gh/SherlockNoMad/6/base 2025-10-10T01:18:33.3007527Z * [new branch] gh/SherlockNoMad/6/head -> origin/gh/SherlockNoMad/6/head 2025-10-10T01:18:33.3009254Z * [new branch] gh/SherlockNoMad/6/orig -> origin/gh/SherlockNoMad/6/orig 2025-10-10T01:18:33.3011532Z * [new branch] gh/SherlockNoMad/7/base -> origin/gh/SherlockNoMad/7/base 2025-10-10T01:18:33.3013182Z * [new branch] gh/SherlockNoMad/7/head -> origin/gh/SherlockNoMad/7/head 2025-10-10T01:18:33.3014875Z * [new branch] gh/SherlockNoMad/7/orig -> origin/gh/SherlockNoMad/7/orig 2025-10-10T01:18:33.3017109Z * [new branch] gh/SherlockNoMad/8/base -> origin/gh/SherlockNoMad/8/base 2025-10-10T01:18:33.3018800Z * [new branch] gh/SherlockNoMad/8/head -> origin/gh/SherlockNoMad/8/head 2025-10-10T01:18:33.3020489Z * [new branch] gh/SherlockNoMad/8/orig -> origin/gh/SherlockNoMad/8/orig 2025-10-10T01:18:33.3022674Z * [new branch] gh/SherlockNoMad/9/base -> origin/gh/SherlockNoMad/9/base 2025-10-10T01:18:33.3024581Z * [new branch] gh/SherlockNoMad/9/orig -> origin/gh/SherlockNoMad/9/orig 2025-10-10T01:18:33.3027476Z * [new branch] gh/Sidharth123-cpu/24/base -> origin/gh/Sidharth123-cpu/24/base 2025-10-10T01:18:33.3029666Z * [new branch] gh/Sidharth123-cpu/25/base -> origin/gh/Sidharth123-cpu/25/base 2025-10-10T01:18:33.3031776Z * [new branch] gh/Sidharth123-cpu/26/base -> origin/gh/Sidharth123-cpu/26/base 2025-10-10T01:18:33.3034043Z * [new branch] gh/Sidharth123-cpu/27/base -> origin/gh/Sidharth123-cpu/27/base 2025-10-10T01:18:33.3036874Z * [new branch] gh/StrongerXi/1/base -> origin/gh/StrongerXi/1/base 2025-10-10T01:18:33.3038848Z * [new branch] gh/StrongerXi/1/head -> origin/gh/StrongerXi/1/head 2025-10-10T01:18:33.3041608Z * [new branch] gh/StrongerXi/133/base -> origin/gh/StrongerXi/133/base 2025-10-10T01:18:33.3043275Z * [new branch] gh/StrongerXi/133/head -> origin/gh/StrongerXi/133/head 2025-10-10T01:18:33.3045004Z * [new branch] gh/StrongerXi/133/orig -> origin/gh/StrongerXi/133/orig 2025-10-10T01:18:33.3047305Z * [new branch] gh/StrongerXi/134/base -> origin/gh/StrongerXi/134/base 2025-10-10T01:18:33.3049087Z * [new branch] gh/StrongerXi/134/head -> origin/gh/StrongerXi/134/head 2025-10-10T01:18:33.3050731Z * [new branch] gh/StrongerXi/134/orig -> origin/gh/StrongerXi/134/orig 2025-10-10T01:18:33.3053054Z * [new branch] gh/StrongerXi/136/base -> origin/gh/StrongerXi/136/base 2025-10-10T01:18:33.3054740Z * [new branch] gh/StrongerXi/136/head -> origin/gh/StrongerXi/136/head 2025-10-10T01:18:33.3056423Z * [new branch] gh/StrongerXi/136/orig -> origin/gh/StrongerXi/136/orig 2025-10-10T01:18:33.3058698Z * [new branch] gh/StrongerXi/137/base -> origin/gh/StrongerXi/137/base 2025-10-10T01:18:33.3060330Z * [new branch] gh/StrongerXi/137/head -> origin/gh/StrongerXi/137/head 2025-10-10T01:18:33.3061986Z * [new branch] gh/StrongerXi/137/orig -> origin/gh/StrongerXi/137/orig 2025-10-10T01:18:33.3064180Z * [new branch] gh/StrongerXi/138/base -> origin/gh/StrongerXi/138/base 2025-10-10T01:18:33.3065822Z * [new branch] gh/StrongerXi/138/head -> origin/gh/StrongerXi/138/head 2025-10-10T01:18:33.3067880Z * [new branch] gh/StrongerXi/138/orig -> origin/gh/StrongerXi/138/orig 2025-10-10T01:18:33.3070170Z * [new branch] gh/StrongerXi/71/base -> origin/gh/StrongerXi/71/base 2025-10-10T01:18:33.3072057Z * [new branch] gh/StrongerXi/71/head -> origin/gh/StrongerXi/71/head 2025-10-10T01:18:33.3074128Z * [new branch] gh/StrongerXi/72/base -> origin/gh/StrongerXi/72/base 2025-10-10T01:18:33.3076278Z * [new branch] gh/StrongerXi/72/head -> origin/gh/StrongerXi/72/head 2025-10-10T01:18:33.3078675Z * [new branch] gh/XilunWu/147/base -> origin/gh/XilunWu/147/base 2025-10-10T01:18:33.3080305Z * [new branch] gh/XilunWu/147/head -> origin/gh/XilunWu/147/head 2025-10-10T01:18:33.3081994Z * [new branch] gh/XilunWu/147/orig -> origin/gh/XilunWu/147/orig 2025-10-10T01:18:33.3084151Z * [new branch] gh/XilunWu/148/base -> origin/gh/XilunWu/148/base 2025-10-10T01:18:33.3085794Z * [new branch] gh/XilunWu/148/head -> origin/gh/XilunWu/148/head 2025-10-10T01:18:33.3087203Z * [new branch] gh/XilunWu/148/orig -> origin/gh/XilunWu/148/orig 2025-10-10T01:18:33.3089577Z * [new branch] gh/XilunWu/149/base -> origin/gh/XilunWu/149/base 2025-10-10T01:18:33.3091223Z * [new branch] gh/XilunWu/149/head -> origin/gh/XilunWu/149/head 2025-10-10T01:18:33.3092981Z * [new branch] gh/XilunWu/149/orig -> origin/gh/XilunWu/149/orig 2025-10-10T01:18:33.3095185Z * [new branch] gh/XilunWu/150/base -> origin/gh/XilunWu/150/base 2025-10-10T01:18:33.3096905Z * [new branch] gh/XilunWu/150/head -> origin/gh/XilunWu/150/head 2025-10-10T01:18:33.3098591Z * [new branch] gh/XilunWu/150/orig -> origin/gh/XilunWu/150/orig 2025-10-10T01:18:33.3100975Z * [new branch] gh/XilunWu/151/base -> origin/gh/XilunWu/151/base 2025-10-10T01:18:33.3102705Z * [new branch] gh/XilunWu/151/head -> origin/gh/XilunWu/151/head 2025-10-10T01:18:33.3104425Z * [new branch] gh/XilunWu/151/orig -> origin/gh/XilunWu/151/orig 2025-10-10T01:18:33.3106780Z * [new branch] gh/XilunWu/152/base -> origin/gh/XilunWu/152/base 2025-10-10T01:18:33.3108443Z * [new branch] gh/XilunWu/152/head -> origin/gh/XilunWu/152/head 2025-10-10T01:18:33.3110093Z * [new branch] gh/XilunWu/152/orig -> origin/gh/XilunWu/152/orig 2025-10-10T01:18:33.3112566Z * [new branch] gh/XilunWu/153/base -> origin/gh/XilunWu/153/base 2025-10-10T01:18:33.3114239Z * [new branch] gh/XilunWu/153/head -> origin/gh/XilunWu/153/head 2025-10-10T01:18:33.3116021Z * [new branch] gh/XilunWu/153/orig -> origin/gh/XilunWu/153/orig 2025-10-10T01:18:33.3118472Z * [new branch] gh/XilunWu/160/base -> origin/gh/XilunWu/160/base 2025-10-10T01:18:33.3120169Z * [new branch] gh/XilunWu/160/head -> origin/gh/XilunWu/160/head 2025-10-10T01:18:33.3121817Z * [new branch] gh/XilunWu/160/orig -> origin/gh/XilunWu/160/orig 2025-10-10T01:18:33.3124345Z * [new branch] gh/XilunWu/163/base -> origin/gh/XilunWu/163/base 2025-10-10T01:18:33.3126098Z * [new branch] gh/XilunWu/163/head -> origin/gh/XilunWu/163/head 2025-10-10T01:18:33.3127734Z * [new branch] gh/XilunWu/163/orig -> origin/gh/XilunWu/163/orig 2025-10-10T01:18:33.3130250Z * [new branch] gh/XilunWu/166/base -> origin/gh/XilunWu/166/base 2025-10-10T01:18:33.3132037Z * [new branch] gh/XilunWu/166/head -> origin/gh/XilunWu/166/head 2025-10-10T01:18:33.3133663Z * [new branch] gh/XilunWu/166/orig -> origin/gh/XilunWu/166/orig 2025-10-10T01:18:33.3136030Z * [new branch] gh/XilunWu/168/base -> origin/gh/XilunWu/168/base 2025-10-10T01:18:33.3137619Z * [new branch] gh/XilunWu/168/head -> origin/gh/XilunWu/168/head 2025-10-10T01:18:33.3139623Z * [new branch] gh/XilunWu/168/orig -> origin/gh/XilunWu/168/orig 2025-10-10T01:18:33.3141702Z * [new branch] gh/XilunWu/169/base -> origin/gh/XilunWu/169/base 2025-10-10T01:18:33.3143363Z * [new branch] gh/XilunWu/169/head -> origin/gh/XilunWu/169/head 2025-10-10T01:18:33.3145062Z * [new branch] gh/XilunWu/169/orig -> origin/gh/XilunWu/169/orig 2025-10-10T01:18:33.3147220Z * [new branch] gh/XilunWu/170/base -> origin/gh/XilunWu/170/base 2025-10-10T01:18:33.3148901Z * [new branch] gh/XilunWu/170/head -> origin/gh/XilunWu/170/head 2025-10-10T01:18:33.3150585Z * [new branch] gh/XilunWu/170/orig -> origin/gh/XilunWu/170/orig 2025-10-10T01:18:33.3153008Z * [new branch] gh/XilunWu/171/base -> origin/gh/XilunWu/171/base 2025-10-10T01:18:33.3154712Z * [new branch] gh/XilunWu/171/head -> origin/gh/XilunWu/171/head 2025-10-10T01:18:33.3156446Z * [new branch] gh/XilunWu/171/orig -> origin/gh/XilunWu/171/orig 2025-10-10T01:18:33.3158689Z * [new branch] gh/XilunWu/172/base -> origin/gh/XilunWu/172/base 2025-10-10T01:18:33.3160384Z * [new branch] gh/XilunWu/172/head -> origin/gh/XilunWu/172/head 2025-10-10T01:18:33.3162132Z * [new branch] gh/XilunWu/172/orig -> origin/gh/XilunWu/172/orig 2025-10-10T01:18:33.3164703Z * [new branch] gh/XilunWu/173/base -> origin/gh/XilunWu/173/base 2025-10-10T01:18:33.3166109Z * [new branch] gh/XilunWu/173/head -> origin/gh/XilunWu/173/head 2025-10-10T01:18:33.3167743Z * [new branch] gh/XilunWu/173/orig -> origin/gh/XilunWu/173/orig 2025-10-10T01:18:33.3170555Z * [new branch] gh/XilunWu/174/base -> origin/gh/XilunWu/174/base 2025-10-10T01:18:33.3172581Z * [new branch] gh/XilunWu/174/head -> origin/gh/XilunWu/174/head 2025-10-10T01:18:33.3174201Z * [new branch] gh/XilunWu/174/orig -> origin/gh/XilunWu/174/orig 2025-10-10T01:18:33.3176499Z * [new branch] gh/XilunWu/175/base -> origin/gh/XilunWu/175/base 2025-10-10T01:18:33.3178164Z * [new branch] gh/XilunWu/175/head -> origin/gh/XilunWu/175/head 2025-10-10T01:18:33.3179809Z * [new branch] gh/XilunWu/175/orig -> origin/gh/XilunWu/175/orig 2025-10-10T01:18:33.3182828Z * [new branch] gh/XuehaiPan/14/base -> origin/gh/XuehaiPan/14/base 2025-10-10T01:18:33.3184604Z * [new branch] gh/XuehaiPan/14/head -> origin/gh/XuehaiPan/14/head 2025-10-10T01:18:33.3186349Z * [new branch] gh/XuehaiPan/14/orig -> origin/gh/XuehaiPan/14/orig 2025-10-10T01:18:33.3188643Z * [new branch] gh/XuehaiPan/179/base -> origin/gh/XuehaiPan/179/base 2025-10-10T01:18:33.3190339Z * [new branch] gh/XuehaiPan/179/head -> origin/gh/XuehaiPan/179/head 2025-10-10T01:18:33.3192329Z * [new branch] gh/XuehaiPan/179/orig -> origin/gh/XuehaiPan/179/orig 2025-10-10T01:18:33.3194418Z * [new branch] gh/XuehaiPan/189/base -> origin/gh/XuehaiPan/189/base 2025-10-10T01:18:33.3196095Z * [new branch] gh/XuehaiPan/189/head -> origin/gh/XuehaiPan/189/head 2025-10-10T01:18:33.3197845Z * [new branch] gh/XuehaiPan/189/orig -> origin/gh/XuehaiPan/189/orig 2025-10-10T01:18:33.3200332Z * [new branch] gh/XuehaiPan/249/base -> origin/gh/XuehaiPan/249/base 2025-10-10T01:18:33.3201970Z * [new branch] gh/XuehaiPan/249/head -> origin/gh/XuehaiPan/249/head 2025-10-10T01:18:33.3203708Z * [new branch] gh/XuehaiPan/249/orig -> origin/gh/XuehaiPan/249/orig 2025-10-10T01:18:33.3205981Z * [new branch] gh/XuehaiPan/253/base -> origin/gh/XuehaiPan/253/base 2025-10-10T01:18:33.3207794Z * [new branch] gh/XuehaiPan/253/head -> origin/gh/XuehaiPan/253/head 2025-10-10T01:18:33.3209434Z * [new branch] gh/XuehaiPan/253/orig -> origin/gh/XuehaiPan/253/orig 2025-10-10T01:18:33.3211705Z * [new branch] gh/XuehaiPan/254/base -> origin/gh/XuehaiPan/254/base 2025-10-10T01:18:33.3213429Z * [new branch] gh/XuehaiPan/254/head -> origin/gh/XuehaiPan/254/head 2025-10-10T01:18:33.3215114Z * [new branch] gh/XuehaiPan/254/orig -> origin/gh/XuehaiPan/254/orig 2025-10-10T01:18:33.3217292Z * [new branch] gh/XuehaiPan/255/base -> origin/gh/XuehaiPan/255/base 2025-10-10T01:18:33.3218993Z * [new branch] gh/XuehaiPan/255/head -> origin/gh/XuehaiPan/255/head 2025-10-10T01:18:33.3220667Z * [new branch] gh/XuehaiPan/255/orig -> origin/gh/XuehaiPan/255/orig 2025-10-10T01:18:33.3223013Z * [new branch] gh/XuehaiPan/257/base -> origin/gh/XuehaiPan/257/base 2025-10-10T01:18:33.3224728Z * [new branch] gh/XuehaiPan/257/head -> origin/gh/XuehaiPan/257/head 2025-10-10T01:18:33.3227868Z * [new branch] gh/XuehaiPan/257/orig -> origin/gh/XuehaiPan/257/orig 2025-10-10T01:18:33.3230200Z * [new branch] gh/XuehaiPan/271/base -> origin/gh/XuehaiPan/271/base 2025-10-10T01:18:33.3231887Z * [new branch] gh/XuehaiPan/271/head -> origin/gh/XuehaiPan/271/head 2025-10-10T01:18:33.3233637Z * [new branch] gh/XuehaiPan/271/orig -> origin/gh/XuehaiPan/271/orig 2025-10-10T01:18:33.3235926Z * [new branch] gh/XuehaiPan/290/base -> origin/gh/XuehaiPan/290/base 2025-10-10T01:18:33.3237597Z * [new branch] gh/XuehaiPan/290/head -> origin/gh/XuehaiPan/290/head 2025-10-10T01:18:33.3239548Z * [new branch] gh/XuehaiPan/290/orig -> origin/gh/XuehaiPan/290/orig 2025-10-10T01:18:33.3241736Z * [new branch] gh/XuehaiPan/343/base -> origin/gh/XuehaiPan/343/base 2025-10-10T01:18:33.3243450Z * [new branch] gh/XuehaiPan/343/head -> origin/gh/XuehaiPan/343/head 2025-10-10T01:18:33.3245603Z * [new branch] gh/XuehaiPan/343/orig -> origin/gh/XuehaiPan/343/orig 2025-10-10T01:18:33.3247379Z * [new branch] gh/XuehaiPan/347/base -> origin/gh/XuehaiPan/347/base 2025-10-10T01:18:33.3249022Z * [new branch] gh/XuehaiPan/347/head -> origin/gh/XuehaiPan/347/head 2025-10-10T01:18:33.3250711Z * [new branch] gh/XuehaiPan/347/orig -> origin/gh/XuehaiPan/347/orig 2025-10-10T01:18:33.3252873Z * [new branch] gh/XuehaiPan/348/base -> origin/gh/XuehaiPan/348/base 2025-10-10T01:18:33.3254694Z * [new branch] gh/XuehaiPan/348/head -> origin/gh/XuehaiPan/348/head 2025-10-10T01:18:33.3256455Z * [new branch] gh/XuehaiPan/348/orig -> origin/gh/XuehaiPan/348/orig 2025-10-10T01:18:33.3258728Z * [new branch] gh/XuehaiPan/350/base -> origin/gh/XuehaiPan/350/base 2025-10-10T01:18:33.3260484Z * [new branch] gh/XuehaiPan/350/head -> origin/gh/XuehaiPan/350/head 2025-10-10T01:18:33.3262101Z * [new branch] gh/XuehaiPan/350/orig -> origin/gh/XuehaiPan/350/orig 2025-10-10T01:18:33.3264456Z * [new branch] gh/XuehaiPan/356/base -> origin/gh/XuehaiPan/356/base 2025-10-10T01:18:33.3266186Z * [new branch] gh/XuehaiPan/356/head -> origin/gh/XuehaiPan/356/head 2025-10-10T01:18:33.3267821Z * [new branch] gh/XuehaiPan/356/orig -> origin/gh/XuehaiPan/356/orig 2025-10-10T01:18:33.3270089Z * [new branch] gh/XuehaiPan/357/base -> origin/gh/XuehaiPan/357/base 2025-10-10T01:18:33.3271779Z * [new branch] gh/XuehaiPan/357/head -> origin/gh/XuehaiPan/357/head 2025-10-10T01:18:33.3273414Z * [new branch] gh/XuehaiPan/357/orig -> origin/gh/XuehaiPan/357/orig 2025-10-10T01:18:33.3275878Z * [new branch] gh/XuehaiPan/358/base -> origin/gh/XuehaiPan/358/base 2025-10-10T01:18:33.3277452Z * [new branch] gh/XuehaiPan/358/head -> origin/gh/XuehaiPan/358/head 2025-10-10T01:18:33.3279156Z * [new branch] gh/XuehaiPan/358/orig -> origin/gh/XuehaiPan/358/orig 2025-10-10T01:18:33.3281976Z * [new branch] gh/XuehaiPan/359/base -> origin/gh/XuehaiPan/359/base 2025-10-10T01:18:33.3283595Z * [new branch] gh/XuehaiPan/359/head -> origin/gh/XuehaiPan/359/head 2025-10-10T01:18:33.3285296Z * [new branch] gh/XuehaiPan/359/orig -> origin/gh/XuehaiPan/359/orig 2025-10-10T01:18:33.3287558Z * [new branch] gh/XuehaiPan/360/base -> origin/gh/XuehaiPan/360/base 2025-10-10T01:18:33.3289272Z * [new branch] gh/XuehaiPan/360/head -> origin/gh/XuehaiPan/360/head 2025-10-10T01:18:33.3290922Z * [new branch] gh/XuehaiPan/360/orig -> origin/gh/XuehaiPan/360/orig 2025-10-10T01:18:33.3293305Z * [new branch] gh/XuehaiPan/365/base -> origin/gh/XuehaiPan/365/base 2025-10-10T01:18:33.3294938Z * [new branch] gh/XuehaiPan/365/head -> origin/gh/XuehaiPan/365/head 2025-10-10T01:18:33.3296703Z * [new branch] gh/XuehaiPan/365/orig -> origin/gh/XuehaiPan/365/orig 2025-10-10T01:18:33.3299035Z * [new branch] gh/XuehaiPan/366/base -> origin/gh/XuehaiPan/366/base 2025-10-10T01:18:33.3300809Z * [new branch] gh/XuehaiPan/366/head -> origin/gh/XuehaiPan/366/head 2025-10-10T01:18:33.3303147Z * [new branch] gh/XuehaiPan/370/base -> origin/gh/XuehaiPan/370/base 2025-10-10T01:18:33.3305099Z * [new branch] gh/XuehaiPan/370/head -> origin/gh/XuehaiPan/370/head 2025-10-10T01:18:33.3306758Z * [new branch] gh/XuehaiPan/370/orig -> origin/gh/XuehaiPan/370/orig 2025-10-10T01:18:33.3309106Z * [new branch] gh/XuehaiPan/384/base -> origin/gh/XuehaiPan/384/base 2025-10-10T01:18:33.3310771Z * [new branch] gh/XuehaiPan/384/head -> origin/gh/XuehaiPan/384/head 2025-10-10T01:18:33.3312500Z * [new branch] gh/XuehaiPan/384/orig -> origin/gh/XuehaiPan/384/orig 2025-10-10T01:18:33.3314833Z * [new branch] gh/XuehaiPan/385/base -> origin/gh/XuehaiPan/385/base 2025-10-10T01:18:33.3316498Z * [new branch] gh/XuehaiPan/385/head -> origin/gh/XuehaiPan/385/head 2025-10-10T01:18:33.3318206Z * [new branch] gh/XuehaiPan/385/orig -> origin/gh/XuehaiPan/385/orig 2025-10-10T01:18:33.3320630Z * [new branch] gh/XuehaiPan/386/base -> origin/gh/XuehaiPan/386/base 2025-10-10T01:18:33.3322303Z * [new branch] gh/XuehaiPan/386/head -> origin/gh/XuehaiPan/386/head 2025-10-10T01:18:33.3324299Z * [new branch] gh/XuehaiPan/386/orig -> origin/gh/XuehaiPan/386/orig 2025-10-10T01:18:33.3326650Z * [new branch] gh/XuehaiPan/387/base -> origin/gh/XuehaiPan/387/base 2025-10-10T01:18:33.3328282Z * [new branch] gh/XuehaiPan/387/head -> origin/gh/XuehaiPan/387/head 2025-10-10T01:18:33.3329914Z * [new branch] gh/XuehaiPan/387/orig -> origin/gh/XuehaiPan/387/orig 2025-10-10T01:18:33.3332298Z * [new branch] gh/XuehaiPan/388/base -> origin/gh/XuehaiPan/388/base 2025-10-10T01:18:33.3334002Z * [new branch] gh/XuehaiPan/388/head -> origin/gh/XuehaiPan/388/head 2025-10-10T01:18:33.3335743Z * [new branch] gh/XuehaiPan/388/orig -> origin/gh/XuehaiPan/388/orig 2025-10-10T01:18:33.3338064Z * [new branch] gh/XuehaiPan/389/base -> origin/gh/XuehaiPan/389/base 2025-10-10T01:18:33.3339493Z * [new branch] gh/XuehaiPan/389/head -> origin/gh/XuehaiPan/389/head 2025-10-10T01:18:33.3341536Z * [new branch] gh/XuehaiPan/389/orig -> origin/gh/XuehaiPan/389/orig 2025-10-10T01:18:33.3344190Z * [new branch] gh/ZhiweiYan-96/39/base -> origin/gh/ZhiweiYan-96/39/base 2025-10-10T01:18:33.3345821Z * [new branch] gh/ZhiweiYan-96/39/head -> origin/gh/ZhiweiYan-96/39/head 2025-10-10T01:18:33.3347567Z * [new branch] gh/ZhiweiYan-96/39/orig -> origin/gh/ZhiweiYan-96/39/orig 2025-10-10T01:18:33.3349814Z * [new branch] gh/ZhiweiYan-96/44/base -> origin/gh/ZhiweiYan-96/44/base 2025-10-10T01:18:33.3351497Z * [new branch] gh/ZhiweiYan-96/44/head -> origin/gh/ZhiweiYan-96/44/head 2025-10-10T01:18:33.3353770Z * [new branch] gh/ZhiweiYan-96/45/base -> origin/gh/ZhiweiYan-96/45/base 2025-10-10T01:18:33.3355424Z * [new branch] gh/ZhiweiYan-96/45/head -> origin/gh/ZhiweiYan-96/45/head 2025-10-10T01:18:33.3357739Z * [new branch] gh/ZhiweiYan-96/49/base -> origin/gh/ZhiweiYan-96/49/base 2025-10-10T01:18:33.3359528Z * [new branch] gh/ZhiweiYan-96/49/head -> origin/gh/ZhiweiYan-96/49/head 2025-10-10T01:18:33.3361741Z * [new branch] gh/ZhiweiYan-96/62/base -> origin/gh/ZhiweiYan-96/62/base 2025-10-10T01:18:33.3363420Z * [new branch] gh/ZhiweiYan-96/62/head -> origin/gh/ZhiweiYan-96/62/head 2025-10-10T01:18:33.3365629Z * [new branch] gh/ZhiweiYan-96/64/base -> origin/gh/ZhiweiYan-96/64/base 2025-10-10T01:18:33.3367375Z * [new branch] gh/ZhiweiYan-96/64/head -> origin/gh/ZhiweiYan-96/64/head 2025-10-10T01:18:33.3369062Z * [new branch] gh/ZhiweiYan-96/64/orig -> origin/gh/ZhiweiYan-96/64/orig 2025-10-10T01:18:33.3371593Z * [new branch] gh/ZhiweiYan-96/66/base -> origin/gh/ZhiweiYan-96/66/base 2025-10-10T01:18:33.3373232Z * [new branch] gh/ZhiweiYan-96/66/head -> origin/gh/ZhiweiYan-96/66/head 2025-10-10T01:18:33.3375469Z * [new branch] gh/ZhiweiYan-96/67/base -> origin/gh/ZhiweiYan-96/67/base 2025-10-10T01:18:33.3377196Z * [new branch] gh/ZhiweiYan-96/67/head -> origin/gh/ZhiweiYan-96/67/head 2025-10-10T01:18:33.3379368Z * [new branch] gh/ZhiweiYan-96/68/base -> origin/gh/ZhiweiYan-96/68/base 2025-10-10T01:18:33.3381005Z * [new branch] gh/ZhiweiYan-96/68/head -> origin/gh/ZhiweiYan-96/68/head 2025-10-10T01:18:33.3382716Z * [new branch] gh/ZhiweiYan-96/68/orig -> origin/gh/ZhiweiYan-96/68/orig 2025-10-10T01:18:33.3385434Z * [new branch] gh/aakhundov/1/base -> origin/gh/aakhundov/1/base 2025-10-10T01:18:33.3387215Z * [new branch] gh/aakhundov/1/head -> origin/gh/aakhundov/1/head 2025-10-10T01:18:33.3389424Z * [new branch] gh/aakhundov/2/base -> origin/gh/aakhundov/2/base 2025-10-10T01:18:33.3391026Z * [new branch] gh/aakhundov/2/head -> origin/gh/aakhundov/2/head 2025-10-10T01:18:33.3393387Z * [new branch] gh/aakhundov/3/base -> origin/gh/aakhundov/3/base 2025-10-10T01:18:33.3395063Z * [new branch] gh/aakhundov/3/head -> origin/gh/aakhundov/3/head 2025-10-10T01:18:33.3396841Z * [new branch] gh/aakhundov/3/orig -> origin/gh/aakhundov/3/orig 2025-10-10T01:18:33.3399257Z * [new branch] gh/aditew01/openblas -> origin/gh/aditew01/openblas 2025-10-10T01:18:33.3400897Z * [new branch] gh/aditew01/sbgemm -> origin/gh/aditew01/sbgemm 2025-10-10T01:18:33.3402581Z * [new branch] gh/aditew01/vecbf16 -> origin/gh/aditew01/vecbf16 2025-10-10T01:18:33.3405342Z * [new branch] gh/albanD/1/base -> origin/gh/albanD/1/base 2025-10-10T01:18:33.3406998Z * [new branch] gh/albanD/1/head -> origin/gh/albanD/1/head 2025-10-10T01:18:33.3409016Z * [new branch] gh/albanD/1/orig -> origin/gh/albanD/1/orig 2025-10-10T01:18:33.3411153Z * [new branch] gh/albanD/2/base -> origin/gh/albanD/2/base 2025-10-10T01:18:33.3412745Z * [new branch] gh/albanD/2/head -> origin/gh/albanD/2/head 2025-10-10T01:18:33.3414526Z * [new branch] gh/albanD/2/orig -> origin/gh/albanD/2/orig 2025-10-10T01:18:33.3416878Z * [new branch] gh/albanD/3/base -> origin/gh/albanD/3/base 2025-10-10T01:18:33.3418386Z * [new branch] gh/albanD/3/head -> origin/gh/albanD/3/head 2025-10-10T01:18:33.3420060Z * [new branch] gh/albanD/3/orig -> origin/gh/albanD/3/orig 2025-10-10T01:18:33.3422277Z * [new branch] gh/albanD/4/base -> origin/gh/albanD/4/base 2025-10-10T01:18:33.3424095Z * [new branch] gh/albanD/4/head -> origin/gh/albanD/4/head 2025-10-10T01:18:33.3426505Z * [new branch] gh/albanD/4/orig -> origin/gh/albanD/4/orig 2025-10-10T01:18:33.3428772Z * [new branch] gh/alexbrauckmann/paddedtensor_faketensor_init -> origin/gh/alexbrauckmann/paddedtensor_faketensor_init 2025-10-10T01:18:33.3431204Z * [new branch] gh/alexsamardzic/10/base -> origin/gh/alexsamardzic/10/base 2025-10-10T01:18:33.3432868Z * [new branch] gh/alexsamardzic/10/head -> origin/gh/alexsamardzic/10/head 2025-10-10T01:18:33.3434528Z * [new branch] gh/alexsamardzic/10/orig -> origin/gh/alexsamardzic/10/orig 2025-10-10T01:18:33.3436634Z * [new branch] gh/alexsamardzic/11/base -> origin/gh/alexsamardzic/11/base 2025-10-10T01:18:33.3438795Z * [new branch] gh/alexsamardzic/11/head -> origin/gh/alexsamardzic/11/head 2025-10-10T01:18:33.3440479Z * [new branch] gh/alexsamardzic/11/orig -> origin/gh/alexsamardzic/11/orig 2025-10-10T01:18:33.3442685Z * [new branch] gh/alexsamardzic/12/base -> origin/gh/alexsamardzic/12/base 2025-10-10T01:18:33.3444405Z * [new branch] gh/alexsamardzic/12/head -> origin/gh/alexsamardzic/12/head 2025-10-10T01:18:33.3446091Z * [new branch] gh/alexsamardzic/12/orig -> origin/gh/alexsamardzic/12/orig 2025-10-10T01:18:33.3448817Z * [new branch] gh/amjames/18/base -> origin/gh/amjames/18/base 2025-10-10T01:18:33.3450499Z * [new branch] gh/amjames/18/head -> origin/gh/amjames/18/head 2025-10-10T01:18:33.3452185Z * [new branch] gh/amjames/18/orig -> origin/gh/amjames/18/orig 2025-10-10T01:18:33.3455202Z * [new branch] gh/andrewor14/35/base -> origin/gh/andrewor14/35/base 2025-10-10T01:18:33.3456990Z * [new branch] gh/andrewor14/35/head -> origin/gh/andrewor14/35/head 2025-10-10T01:18:33.3458725Z * [new branch] gh/andrewor14/35/orig -> origin/gh/andrewor14/35/orig 2025-10-10T01:18:33.3461187Z * [new branch] gh/andrewor14/50/base -> origin/gh/andrewor14/50/base 2025-10-10T01:18:33.3462940Z * [new branch] gh/andrewor14/50/head -> origin/gh/andrewor14/50/head 2025-10-10T01:18:33.3464708Z * [new branch] gh/andrewor14/50/orig -> origin/gh/andrewor14/50/orig 2025-10-10T01:18:33.3467485Z * [new branch] gh/andyanwang/28/base -> origin/gh/andyanwang/28/base 2025-10-10T01:18:33.3469274Z * [new branch] gh/andyanwang/28/head -> origin/gh/andyanwang/28/head 2025-10-10T01:18:33.3470911Z * [new branch] gh/andyanwang/28/orig -> origin/gh/andyanwang/28/orig 2025-10-10T01:18:33.3473280Z * [new branch] gh/andyanwang/30/base -> origin/gh/andyanwang/30/base 2025-10-10T01:18:33.3475227Z * [new branch] gh/andyanwang/30/orig -> origin/gh/andyanwang/30/orig 2025-10-10T01:18:33.3477668Z * [new branch] gh/andyanwang/31/base -> origin/gh/andyanwang/31/base 2025-10-10T01:18:33.3479604Z * [new branch] gh/andyanwang/31/orig -> origin/gh/andyanwang/31/orig 2025-10-10T01:18:33.3482386Z * [new branch] gh/andyanwang/32/base -> origin/gh/andyanwang/32/base 2025-10-10T01:18:33.3483996Z * [new branch] gh/andyanwang/32/head -> origin/gh/andyanwang/32/head 2025-10-10T01:18:33.3486348Z * [new branch] gh/andyanwang/32/orig -> origin/gh/andyanwang/32/orig 2025-10-10T01:18:33.3488979Z * [new branch] gh/andyanwang/39/base -> origin/gh/andyanwang/39/base 2025-10-10T01:18:33.3490428Z * [new branch] gh/andyanwang/39/head -> origin/gh/andyanwang/39/head 2025-10-10T01:18:33.3492108Z * [new branch] gh/andyanwang/39/orig -> origin/gh/andyanwang/39/orig 2025-10-10T01:18:33.3495310Z * [new branch] gh/angelayi/107/base -> origin/gh/angelayi/107/base 2025-10-10T01:18:33.3496452Z * [new branch] gh/angelayi/107/head -> origin/gh/angelayi/107/head 2025-10-10T01:18:33.3498895Z * [new branch] gh/angelayi/114/base -> origin/gh/angelayi/114/base 2025-10-10T01:18:33.3501158Z * [new branch] gh/angelayi/114/head -> origin/gh/angelayi/114/head 2025-10-10T01:18:33.3502215Z * [new branch] gh/angelayi/114/orig -> origin/gh/angelayi/114/orig 2025-10-10T01:18:33.3504646Z * [new branch] gh/angelayi/116/base -> origin/gh/angelayi/116/base 2025-10-10T01:18:33.3506290Z * [new branch] gh/angelayi/116/head -> origin/gh/angelayi/116/head 2025-10-10T01:18:33.3508224Z * [new branch] gh/angelayi/116/orig -> origin/gh/angelayi/116/orig 2025-10-10T01:18:33.3510678Z * [new branch] gh/angelayi/117/base -> origin/gh/angelayi/117/base 2025-10-10T01:18:33.3512344Z * [new branch] gh/angelayi/117/head -> origin/gh/angelayi/117/head 2025-10-10T01:18:33.3514018Z * [new branch] gh/angelayi/117/orig -> origin/gh/angelayi/117/orig 2025-10-10T01:18:33.3516579Z * [new branch] gh/angelayi/118/base -> origin/gh/angelayi/118/base 2025-10-10T01:18:33.3519429Z * [new branch] gh/angelayi/118/head -> origin/gh/angelayi/118/head 2025-10-10T01:18:33.3520669Z * [new branch] gh/angelayi/118/orig -> origin/gh/angelayi/118/orig 2025-10-10T01:18:33.3523580Z * [new branch] gh/angelayi/119/base -> origin/gh/angelayi/119/base 2025-10-10T01:18:33.3524845Z * [new branch] gh/angelayi/119/head -> origin/gh/angelayi/119/head 2025-10-10T01:18:33.3526858Z * [new branch] gh/angelayi/119/orig -> origin/gh/angelayi/119/orig 2025-10-10T01:18:33.3529268Z * [new branch] gh/angelayi/120/base -> origin/gh/angelayi/120/base 2025-10-10T01:18:33.3531008Z * [new branch] gh/angelayi/120/head -> origin/gh/angelayi/120/head 2025-10-10T01:18:33.3532635Z * [new branch] gh/angelayi/120/orig -> origin/gh/angelayi/120/orig 2025-10-10T01:18:33.3534975Z * [new branch] gh/angelayi/121/base -> origin/gh/angelayi/121/base 2025-10-10T01:18:33.3537546Z * [new branch] gh/angelayi/121/head -> origin/gh/angelayi/121/head 2025-10-10T01:18:33.3538794Z * [new branch] gh/angelayi/121/orig -> origin/gh/angelayi/121/orig 2025-10-10T01:18:33.3541513Z * [new branch] gh/angelayi/122/base -> origin/gh/angelayi/122/base 2025-10-10T01:18:33.3542820Z * [new branch] gh/angelayi/122/head -> origin/gh/angelayi/122/head 2025-10-10T01:18:33.3544488Z * [new branch] gh/angelayi/122/orig -> origin/gh/angelayi/122/orig 2025-10-10T01:18:33.3546951Z * [new branch] gh/angelayi/123/base -> origin/gh/angelayi/123/base 2025-10-10T01:18:33.3548837Z * [new branch] gh/angelayi/123/head -> origin/gh/angelayi/123/head 2025-10-10T01:18:33.3550454Z * [new branch] gh/angelayi/123/orig -> origin/gh/angelayi/123/orig 2025-10-10T01:18:33.3553010Z * [new branch] gh/angelayi/124/base -> origin/gh/angelayi/124/base 2025-10-10T01:18:33.3554509Z * [new branch] gh/angelayi/124/head -> origin/gh/angelayi/124/head 2025-10-10T01:18:33.3556034Z * [new branch] gh/angelayi/124/orig -> origin/gh/angelayi/124/orig 2025-10-10T01:18:33.3558335Z * [new branch] gh/angelayi/125/base -> origin/gh/angelayi/125/base 2025-10-10T01:18:33.3560188Z * [new branch] gh/angelayi/125/head -> origin/gh/angelayi/125/head 2025-10-10T01:18:33.3561888Z * [new branch] gh/angelayi/125/orig -> origin/gh/angelayi/125/orig 2025-10-10T01:18:33.3564178Z * [new branch] gh/angelayi/126/base -> origin/gh/angelayi/126/base 2025-10-10T01:18:33.3565930Z * [new branch] gh/angelayi/126/head -> origin/gh/angelayi/126/head 2025-10-10T01:18:33.3567660Z * [new branch] gh/angelayi/126/orig -> origin/gh/angelayi/126/orig 2025-10-10T01:18:33.3570278Z * [new branch] gh/angelayi/127/base -> origin/gh/angelayi/127/base 2025-10-10T01:18:33.3571957Z * [new branch] gh/angelayi/127/head -> origin/gh/angelayi/127/head 2025-10-10T01:18:33.3573757Z * [new branch] gh/angelayi/127/orig -> origin/gh/angelayi/127/orig 2025-10-10T01:18:33.3576386Z * [new branch] gh/angelayi/128/base -> origin/gh/angelayi/128/base 2025-10-10T01:18:33.3578395Z * [new branch] gh/angelayi/128/head -> origin/gh/angelayi/128/head 2025-10-10T01:18:33.3580012Z * [new branch] gh/angelayi/128/orig -> origin/gh/angelayi/128/orig 2025-10-10T01:18:33.3582573Z * [new branch] gh/angelayi/129/base -> origin/gh/angelayi/129/base 2025-10-10T01:18:33.3583978Z * [new branch] gh/angelayi/129/head -> origin/gh/angelayi/129/head 2025-10-10T01:18:33.3585803Z * [new branch] gh/angelayi/129/orig -> origin/gh/angelayi/129/orig 2025-10-10T01:18:33.3588589Z * [new branch] gh/angelayi/130/base -> origin/gh/angelayi/130/base 2025-10-10T01:18:33.3590334Z * [new branch] gh/angelayi/130/head -> origin/gh/angelayi/130/head 2025-10-10T01:18:33.3591975Z * [new branch] gh/angelayi/130/orig -> origin/gh/angelayi/130/orig 2025-10-10T01:18:33.3594923Z * [new branch] gh/anijain2305/753/base -> origin/gh/anijain2305/753/base 2025-10-10T01:18:33.3596658Z * [new branch] gh/anijain2305/753/head -> origin/gh/anijain2305/753/head 2025-10-10T01:18:33.3598366Z * [new branch] gh/anijain2305/753/orig -> origin/gh/anijain2305/753/orig 2025-10-10T01:18:33.3600808Z * [new branch] gh/anijain2305/790/base -> origin/gh/anijain2305/790/base 2025-10-10T01:18:33.3602551Z * [new branch] gh/anijain2305/790/head -> origin/gh/anijain2305/790/head 2025-10-10T01:18:33.3604315Z * [new branch] gh/anijain2305/790/orig -> origin/gh/anijain2305/790/orig 2025-10-10T01:18:33.3606873Z * [new branch] gh/anijain2305/792/base -> origin/gh/anijain2305/792/base 2025-10-10T01:18:33.3608839Z * [new branch] gh/anijain2305/792/head -> origin/gh/anijain2305/792/head 2025-10-10T01:18:33.3610344Z * [new branch] gh/anijain2305/792/orig -> origin/gh/anijain2305/792/orig 2025-10-10T01:18:33.3612837Z * [new branch] gh/anijain2305/805/base -> origin/gh/anijain2305/805/base 2025-10-10T01:18:33.3614281Z * [new branch] gh/anijain2305/805/head -> origin/gh/anijain2305/805/head 2025-10-10T01:18:33.3616198Z * [new branch] gh/anijain2305/805/orig -> origin/gh/anijain2305/805/orig 2025-10-10T01:18:33.3618641Z * [new branch] gh/anijain2305/810/base -> origin/gh/anijain2305/810/base 2025-10-10T01:18:33.3620540Z * [new branch] gh/anijain2305/810/head -> origin/gh/anijain2305/810/head 2025-10-10T01:18:33.3622102Z * [new branch] gh/anijain2305/810/orig -> origin/gh/anijain2305/810/orig 2025-10-10T01:18:33.3624447Z * [new branch] gh/anijain2305/812/base -> origin/gh/anijain2305/812/base 2025-10-10T01:18:33.3627553Z * [new branch] gh/anijain2305/812/head -> origin/gh/anijain2305/812/head 2025-10-10T01:18:33.3629202Z * [new branch] gh/anijain2305/812/orig -> origin/gh/anijain2305/812/orig 2025-10-10T01:18:33.3631456Z * [new branch] gh/anijain2305/854/base -> origin/gh/anijain2305/854/base 2025-10-10T01:18:33.3633180Z * [new branch] gh/anijain2305/854/head -> origin/gh/anijain2305/854/head 2025-10-10T01:18:33.3634882Z * [new branch] gh/anijain2305/854/orig -> origin/gh/anijain2305/854/orig 2025-10-10T01:18:33.3637243Z * [new branch] gh/anijain2305/855/base -> origin/gh/anijain2305/855/base 2025-10-10T01:18:33.3639062Z * [new branch] gh/anijain2305/855/head -> origin/gh/anijain2305/855/head 2025-10-10T01:18:33.3641993Z * [new branch] gh/anijain2305/855/orig -> origin/gh/anijain2305/855/orig 2025-10-10T01:18:33.3643818Z * [new branch] gh/anijain2305/864/base -> origin/gh/anijain2305/864/base 2025-10-10T01:18:33.3645082Z * [new branch] gh/anijain2305/864/head -> origin/gh/anijain2305/864/head 2025-10-10T01:18:33.3647869Z * [new branch] gh/anijain2305/864/orig -> origin/gh/anijain2305/864/orig 2025-10-10T01:18:33.3649220Z * [new branch] gh/anijain2305/867/base -> origin/gh/anijain2305/867/base 2025-10-10T01:18:33.3650860Z * [new branch] gh/anijain2305/867/head -> origin/gh/anijain2305/867/head 2025-10-10T01:18:33.3652497Z * [new branch] gh/anijain2305/867/orig -> origin/gh/anijain2305/867/orig 2025-10-10T01:18:33.3655351Z * [new branch] gh/anijain2305/868/base -> origin/gh/anijain2305/868/base 2025-10-10T01:18:33.3656472Z * [new branch] gh/anijain2305/868/head -> origin/gh/anijain2305/868/head 2025-10-10T01:18:33.3658244Z * [new branch] gh/anijain2305/868/orig -> origin/gh/anijain2305/868/orig 2025-10-10T01:18:33.3660575Z * [new branch] gh/anijain2305/869/base -> origin/gh/anijain2305/869/base 2025-10-10T01:18:33.3662293Z * [new branch] gh/anijain2305/869/head -> origin/gh/anijain2305/869/head 2025-10-10T01:18:33.3664029Z * [new branch] gh/anijain2305/869/orig -> origin/gh/anijain2305/869/orig 2025-10-10T01:18:33.3666428Z * [new branch] gh/anijain2305/870/base -> origin/gh/anijain2305/870/base 2025-10-10T01:18:33.3668150Z * [new branch] gh/anijain2305/870/head -> origin/gh/anijain2305/870/head 2025-10-10T01:18:33.3669823Z * [new branch] gh/anijain2305/870/orig -> origin/gh/anijain2305/870/orig 2025-10-10T01:18:33.3672130Z * [new branch] gh/anijain2305/871/base -> origin/gh/anijain2305/871/base 2025-10-10T01:18:33.3673826Z * [new branch] gh/anijain2305/871/head -> origin/gh/anijain2305/871/head 2025-10-10T01:18:33.3675626Z * [new branch] gh/anijain2305/871/orig -> origin/gh/anijain2305/871/orig 2025-10-10T01:18:33.3678666Z * [new branch] gh/anijain2305/872/base -> origin/gh/anijain2305/872/base 2025-10-10T01:18:33.3679796Z * [new branch] gh/anijain2305/872/head -> origin/gh/anijain2305/872/head 2025-10-10T01:18:33.3681446Z * [new branch] gh/anijain2305/872/orig -> origin/gh/anijain2305/872/orig 2025-10-10T01:18:33.3684371Z * [new branch] gh/anijain2305/873/base -> origin/gh/anijain2305/873/base 2025-10-10T01:18:33.3685655Z * [new branch] gh/anijain2305/873/head -> origin/gh/anijain2305/873/head 2025-10-10T01:18:33.3687758Z * [new branch] gh/anijain2305/873/orig -> origin/gh/anijain2305/873/orig 2025-10-10T01:18:33.3690406Z * [new branch] gh/anijain2305/874/base -> origin/gh/anijain2305/874/base 2025-10-10T01:18:33.3692135Z * [new branch] gh/anijain2305/874/head -> origin/gh/anijain2305/874/head 2025-10-10T01:18:33.3693867Z * [new branch] gh/anijain2305/874/orig -> origin/gh/anijain2305/874/orig 2025-10-10T01:18:33.3697051Z * [new branch] gh/anijain2305/875/base -> origin/gh/anijain2305/875/base 2025-10-10T01:18:33.3698318Z * [new branch] gh/anijain2305/875/head -> origin/gh/anijain2305/875/head 2025-10-10T01:18:33.3700005Z * [new branch] gh/anijain2305/875/orig -> origin/gh/anijain2305/875/orig 2025-10-10T01:18:33.3702703Z * [new branch] gh/anijain2305/876/base -> origin/gh/anijain2305/876/base 2025-10-10T01:18:33.3704048Z * [new branch] gh/anijain2305/876/head -> origin/gh/anijain2305/876/head 2025-10-10T01:18:33.3706944Z * [new branch] gh/anijain2305/877/base -> origin/gh/anijain2305/877/base 2025-10-10T01:18:33.3708561Z * [new branch] gh/anijain2305/877/head -> origin/gh/anijain2305/877/head 2025-10-10T01:18:33.3710227Z * [new branch] gh/anijain2305/877/orig -> origin/gh/anijain2305/877/orig 2025-10-10T01:18:33.3712735Z * [new branch] gh/anijain2305/878/base -> origin/gh/anijain2305/878/base 2025-10-10T01:18:33.3714422Z * [new branch] gh/anijain2305/878/head -> origin/gh/anijain2305/878/head 2025-10-10T01:18:33.3716241Z * [new branch] gh/anijain2305/878/orig -> origin/gh/anijain2305/878/orig 2025-10-10T01:18:33.3719564Z * [new branch] gh/anijain2305/879/base -> origin/gh/anijain2305/879/base 2025-10-10T01:18:33.3720768Z * [new branch] gh/anijain2305/879/head -> origin/gh/anijain2305/879/head 2025-10-10T01:18:33.3722444Z * [new branch] gh/anijain2305/879/orig -> origin/gh/anijain2305/879/orig 2025-10-10T01:18:33.3724842Z * [new branch] gh/anijain2305/880/base -> origin/gh/anijain2305/880/base 2025-10-10T01:18:33.3726856Z * [new branch] gh/anijain2305/880/head -> origin/gh/anijain2305/880/head 2025-10-10T01:18:33.3728788Z * [new branch] gh/anijain2305/880/orig -> origin/gh/anijain2305/880/orig 2025-10-10T01:18:33.3731063Z * [new branch] gh/anijain2305/881/base -> origin/gh/anijain2305/881/base 2025-10-10T01:18:33.3732715Z * [new branch] gh/anijain2305/881/head -> origin/gh/anijain2305/881/head 2025-10-10T01:18:33.3734389Z * [new branch] gh/anijain2305/881/orig -> origin/gh/anijain2305/881/orig 2025-10-10T01:18:33.3737038Z * [new branch] gh/anijain2305/882/base -> origin/gh/anijain2305/882/base 2025-10-10T01:18:33.3739036Z * [new branch] gh/anijain2305/882/head -> origin/gh/anijain2305/882/head 2025-10-10T01:18:33.3740575Z * [new branch] gh/anijain2305/882/orig -> origin/gh/anijain2305/882/orig 2025-10-10T01:18:33.3743554Z * [new branch] gh/anijain2305/883/base -> origin/gh/anijain2305/883/base 2025-10-10T01:18:33.3744795Z * [new branch] gh/anijain2305/883/head -> origin/gh/anijain2305/883/head 2025-10-10T01:18:33.3746670Z * [new branch] gh/anijain2305/883/orig -> origin/gh/anijain2305/883/orig 2025-10-10T01:18:33.3749135Z * [new branch] gh/anijain2305/884/base -> origin/gh/anijain2305/884/base 2025-10-10T01:18:33.3750815Z * [new branch] gh/anijain2305/884/head -> origin/gh/anijain2305/884/head 2025-10-10T01:18:33.3752453Z * [new branch] gh/anijain2305/884/orig -> origin/gh/anijain2305/884/orig 2025-10-10T01:18:33.3754790Z * [new branch] gh/anijain2305/885/base -> origin/gh/anijain2305/885/base 2025-10-10T01:18:33.3757696Z * [new branch] gh/anijain2305/885/head -> origin/gh/anijain2305/885/head 2025-10-10T01:18:33.3758693Z * [new branch] gh/anijain2305/885/orig -> origin/gh/anijain2305/885/orig 2025-10-10T01:18:33.3761625Z * [new branch] gh/anijain2305/886/base -> origin/gh/anijain2305/886/base 2025-10-10T01:18:33.3762936Z * [new branch] gh/anijain2305/886/head -> origin/gh/anijain2305/886/head 2025-10-10T01:18:33.3764624Z * [new branch] gh/anijain2305/886/orig -> origin/gh/anijain2305/886/orig 2025-10-10T01:18:33.3767370Z * [new branch] gh/anijain2305/887/base -> origin/gh/anijain2305/887/base 2025-10-10T01:18:33.3769153Z * [new branch] gh/anijain2305/887/head -> origin/gh/anijain2305/887/head 2025-10-10T01:18:33.3770781Z * [new branch] gh/anijain2305/887/orig -> origin/gh/anijain2305/887/orig 2025-10-10T01:18:33.3773236Z * [new branch] gh/anijain2305/888/base -> origin/gh/anijain2305/888/base 2025-10-10T01:18:33.3774951Z * [new branch] gh/anijain2305/888/head -> origin/gh/anijain2305/888/head 2025-10-10T01:18:33.3776565Z * [new branch] gh/anijain2305/888/orig -> origin/gh/anijain2305/888/orig 2025-10-10T01:18:33.3778781Z * [new branch] gh/anijain2305/889/base -> origin/gh/anijain2305/889/base 2025-10-10T01:18:33.3780711Z * [new branch] gh/anijain2305/889/head -> origin/gh/anijain2305/889/head 2025-10-10T01:18:33.3782346Z * [new branch] gh/anijain2305/889/orig -> origin/gh/anijain2305/889/orig 2025-10-10T01:18:33.3785137Z * [new branch] gh/anijain2305/890/base -> origin/gh/anijain2305/890/base 2025-10-10T01:18:33.3786877Z * [new branch] gh/anijain2305/890/head -> origin/gh/anijain2305/890/head 2025-10-10T01:18:33.3788525Z * [new branch] gh/anijain2305/890/orig -> origin/gh/anijain2305/890/orig 2025-10-10T01:18:33.3791632Z * [new branch] gh/anijain2305/891/base -> origin/gh/anijain2305/891/base 2025-10-10T01:18:33.3792837Z * [new branch] gh/anijain2305/891/head -> origin/gh/anijain2305/891/head 2025-10-10T01:18:33.3794499Z * [new branch] gh/anijain2305/891/orig -> origin/gh/anijain2305/891/orig 2025-10-10T01:18:33.3797405Z * [new branch] gh/anijain2305/892/base -> origin/gh/anijain2305/892/base 2025-10-10T01:18:33.3799078Z * [new branch] gh/anijain2305/892/head -> origin/gh/anijain2305/892/head 2025-10-10T01:18:33.3800771Z * [new branch] gh/anijain2305/892/orig -> origin/gh/anijain2305/892/orig 2025-10-10T01:18:33.3803130Z * [new branch] gh/anijain2305/893/base -> origin/gh/anijain2305/893/base 2025-10-10T01:18:33.3804860Z * [new branch] gh/anijain2305/893/head -> origin/gh/anijain2305/893/head 2025-10-10T01:18:33.3806474Z * [new branch] gh/anijain2305/893/orig -> origin/gh/anijain2305/893/orig 2025-10-10T01:18:33.3808761Z * [new branch] gh/anijain2305/894/base -> origin/gh/anijain2305/894/base 2025-10-10T01:18:33.3810440Z * [new branch] gh/anijain2305/894/head -> origin/gh/anijain2305/894/head 2025-10-10T01:18:33.3812074Z * [new branch] gh/anijain2305/894/orig -> origin/gh/anijain2305/894/orig 2025-10-10T01:18:33.3814848Z * [new branch] gh/anijain2305/895/base -> origin/gh/anijain2305/895/base 2025-10-10T01:18:33.3816125Z * [new branch] gh/anijain2305/895/head -> origin/gh/anijain2305/895/head 2025-10-10T01:18:33.3817778Z * [new branch] gh/anijain2305/895/orig -> origin/gh/anijain2305/895/orig 2025-10-10T01:18:33.3820182Z * [new branch] gh/anijain2305/896/base -> origin/gh/anijain2305/896/base 2025-10-10T01:18:33.3821913Z * [new branch] gh/anijain2305/896/head -> origin/gh/anijain2305/896/head 2025-10-10T01:18:33.3823580Z * [new branch] gh/anijain2305/896/orig -> origin/gh/anijain2305/896/orig 2025-10-10T01:18:33.3826942Z * [new branch] gh/anijain2305/897/base -> origin/gh/anijain2305/897/base 2025-10-10T01:18:33.3828449Z * [new branch] gh/anijain2305/897/head -> origin/gh/anijain2305/897/head 2025-10-10T01:18:33.3830089Z * [new branch] gh/anijain2305/897/orig -> origin/gh/anijain2305/897/orig 2025-10-10T01:18:33.3833153Z * [new branch] gh/anijain2305/898/base -> origin/gh/anijain2305/898/base 2025-10-10T01:18:33.3834999Z * [new branch] gh/anijain2305/898/head -> origin/gh/anijain2305/898/head 2025-10-10T01:18:33.3836769Z * [new branch] gh/anijain2305/898/orig -> origin/gh/anijain2305/898/orig 2025-10-10T01:18:33.3839340Z * [new branch] gh/anijain2305/899/base -> origin/gh/anijain2305/899/base 2025-10-10T01:18:33.3841085Z * [new branch] gh/anijain2305/899/head -> origin/gh/anijain2305/899/head 2025-10-10T01:18:33.3842754Z * [new branch] gh/anijain2305/899/orig -> origin/gh/anijain2305/899/orig 2025-10-10T01:18:33.3845135Z * [new branch] gh/anijain2305/900/base -> origin/gh/anijain2305/900/base 2025-10-10T01:18:33.3846774Z * [new branch] gh/anijain2305/900/head -> origin/gh/anijain2305/900/head 2025-10-10T01:18:33.3848650Z * [new branch] gh/anijain2305/900/orig -> origin/gh/anijain2305/900/orig 2025-10-10T01:18:33.3850847Z * [new branch] gh/anijain2305/901/base -> origin/gh/anijain2305/901/base 2025-10-10T01:18:33.3852608Z * [new branch] gh/anijain2305/901/head -> origin/gh/anijain2305/901/head 2025-10-10T01:18:33.3854275Z * [new branch] gh/anijain2305/901/orig -> origin/gh/anijain2305/901/orig 2025-10-10T01:18:33.3857014Z * [new branch] gh/anijain2305/902/base -> origin/gh/anijain2305/902/base 2025-10-10T01:18:33.3859073Z * [new branch] gh/anijain2305/902/head -> origin/gh/anijain2305/902/head 2025-10-10T01:18:33.3860607Z * [new branch] gh/anijain2305/902/orig -> origin/gh/anijain2305/902/orig 2025-10-10T01:18:33.3863314Z * [new branch] gh/anijain2305/903/base -> origin/gh/anijain2305/903/base 2025-10-10T01:18:33.3864627Z * [new branch] gh/anijain2305/903/head -> origin/gh/anijain2305/903/head 2025-10-10T01:18:33.3877145Z * [new branch] gh/anijain2305/903/orig -> origin/gh/anijain2305/903/orig 2025-10-10T01:18:33.3877480Z * [new branch] gh/anijain2305/904/base -> origin/gh/anijain2305/904/base 2025-10-10T01:18:33.3877677Z * [new branch] gh/anijain2305/904/head -> origin/gh/anijain2305/904/head 2025-10-10T01:18:33.3877845Z * [new branch] gh/anijain2305/904/orig -> origin/gh/anijain2305/904/orig 2025-10-10T01:18:33.3878012Z * [new branch] gh/anjali411/216/base -> origin/gh/anjali411/216/base 2025-10-10T01:18:33.3878189Z * [new branch] gh/anjali411/216/head -> origin/gh/anjali411/216/head 2025-10-10T01:18:33.3878558Z * [new branch] gh/anjali411/216/orig -> origin/gh/anjali411/216/orig 2025-10-10T01:18:33.3881762Z * [new branch] gh/ankitageorge/17/base -> origin/gh/ankitageorge/17/base 2025-10-10T01:18:33.3883299Z * [new branch] gh/ankitageorge/17/head -> origin/gh/ankitageorge/17/head 2025-10-10T01:18:33.3884972Z * [new branch] gh/ankitageorge/17/orig -> origin/gh/ankitageorge/17/orig 2025-10-10T01:18:33.3887841Z * [new branch] gh/anshul-si/1/base -> origin/gh/anshul-si/1/base 2025-10-10T01:18:33.3889530Z * [new branch] gh/anshul-si/1/head -> origin/gh/anshul-si/1/head 2025-10-10T01:18:33.3891723Z * [new branch] gh/anshul-si/2/base -> origin/gh/anshul-si/2/base 2025-10-10T01:18:33.3893409Z * [new branch] gh/anshul-si/2/head -> origin/gh/anshul-si/2/head 2025-10-10T01:18:33.3896010Z * [new branch] gh/anshul-si/29/base -> origin/gh/anshul-si/29/base 2025-10-10T01:18:33.3897685Z * [new branch] gh/anshul-si/29/head -> origin/gh/anshul-si/29/head 2025-10-10T01:18:33.3899433Z * [new branch] gh/anshul-si/29/orig -> origin/gh/anshul-si/29/orig 2025-10-10T01:18:33.3901596Z * [new branch] gh/anshul-si/3/base -> origin/gh/anshul-si/3/base 2025-10-10T01:18:33.3903205Z * [new branch] gh/anshul-si/3/head -> origin/gh/anshul-si/3/head 2025-10-10T01:18:33.3905520Z * [new branch] gh/anshul-si/30/base -> origin/gh/anshul-si/30/base 2025-10-10T01:18:33.3907239Z * [new branch] gh/anshul-si/30/head -> origin/gh/anshul-si/30/head 2025-10-10T01:18:33.3908895Z * [new branch] gh/anshul-si/30/orig -> origin/gh/anshul-si/30/orig 2025-10-10T01:18:33.3911055Z * [new branch] gh/anshul-si/31/base -> origin/gh/anshul-si/31/base 2025-10-10T01:18:33.3912777Z * [new branch] gh/anshul-si/31/head -> origin/gh/anshul-si/31/head 2025-10-10T01:18:33.3914495Z * [new branch] gh/anshul-si/31/orig -> origin/gh/anshul-si/31/orig 2025-10-10T01:18:33.3916881Z * [new branch] gh/anshul-si/32/base -> origin/gh/anshul-si/32/base 2025-10-10T01:18:33.3918616Z * [new branch] gh/anshul-si/32/head -> origin/gh/anshul-si/32/head 2025-10-10T01:18:33.3920398Z * [new branch] gh/anshul-si/32/orig -> origin/gh/anshul-si/32/orig 2025-10-10T01:18:33.3922825Z * [new branch] gh/anshul-si/33/base -> origin/gh/anshul-si/33/base 2025-10-10T01:18:33.3924801Z * [new branch] gh/anshul-si/33/head -> origin/gh/anshul-si/33/head 2025-10-10T01:18:33.3926551Z * [new branch] gh/anshul-si/33/orig -> origin/gh/anshul-si/33/orig 2025-10-10T01:18:33.3929106Z * [new branch] gh/anshul-si/34/base -> origin/gh/anshul-si/34/base 2025-10-10T01:18:33.3930727Z * [new branch] gh/anshul-si/34/head -> origin/gh/anshul-si/34/head 2025-10-10T01:18:33.3932452Z * [new branch] gh/anshul-si/34/orig -> origin/gh/anshul-si/34/orig 2025-10-10T01:18:33.3934729Z * [new branch] gh/anshul-si/35/base -> origin/gh/anshul-si/35/base 2025-10-10T01:18:33.3936720Z * [new branch] gh/anshul-si/35/head -> origin/gh/anshul-si/35/head 2025-10-10T01:18:33.3938654Z * [new branch] gh/anshul-si/35/orig -> origin/gh/anshul-si/35/orig 2025-10-10T01:18:33.3941336Z * [new branch] gh/anshul-si/36/base -> origin/gh/anshul-si/36/base 2025-10-10T01:18:33.3942609Z * [new branch] gh/anshul-si/36/head -> origin/gh/anshul-si/36/head 2025-10-10T01:18:33.3944175Z * [new branch] gh/anshul-si/36/orig -> origin/gh/anshul-si/36/orig 2025-10-10T01:18:33.3946931Z * [new branch] gh/anshul-si/37/base -> origin/gh/anshul-si/37/base 2025-10-10T01:18:33.3948473Z * [new branch] gh/anshul-si/37/head -> origin/gh/anshul-si/37/head 2025-10-10T01:18:33.3950176Z * [new branch] gh/anshul-si/37/orig -> origin/gh/anshul-si/37/orig 2025-10-10T01:18:33.3952499Z * [new branch] gh/anshul-si/38/base -> origin/gh/anshul-si/38/base 2025-10-10T01:18:33.3954288Z * [new branch] gh/anshul-si/38/head -> origin/gh/anshul-si/38/head 2025-10-10T01:18:33.3956827Z * [new branch] gh/anshul-si/38/orig -> origin/gh/anshul-si/38/orig 2025-10-10T01:18:33.3958852Z * [new branch] gh/anshul-si/39/base -> origin/gh/anshul-si/39/base 2025-10-10T01:18:33.3960537Z * [new branch] gh/anshul-si/39/head -> origin/gh/anshul-si/39/head 2025-10-10T01:18:33.3962142Z * [new branch] gh/anshul-si/39/orig -> origin/gh/anshul-si/39/orig 2025-10-10T01:18:33.3965134Z * [new branch] gh/anshul-si/4/base -> origin/gh/anshul-si/4/base 2025-10-10T01:18:33.3966266Z * [new branch] gh/anshul-si/4/head -> origin/gh/anshul-si/4/head 2025-10-10T01:18:33.3968771Z * [new branch] gh/anshul-si/40/base -> origin/gh/anshul-si/40/base 2025-10-10T01:18:33.3970508Z * [new branch] gh/anshul-si/40/head -> origin/gh/anshul-si/40/head 2025-10-10T01:18:33.3972171Z * [new branch] gh/anshul-si/40/orig -> origin/gh/anshul-si/40/orig 2025-10-10T01:18:33.3974605Z * [new branch] gh/anshul-si/41/base -> origin/gh/anshul-si/41/base 2025-10-10T01:18:33.3976417Z * [new branch] gh/anshul-si/41/head -> origin/gh/anshul-si/41/head 2025-10-10T01:18:33.3978133Z * [new branch] gh/anshul-si/41/orig -> origin/gh/anshul-si/41/orig 2025-10-10T01:18:33.3980415Z * [new branch] gh/anshul-si/42/base -> origin/gh/anshul-si/42/base 2025-10-10T01:18:33.3982148Z * [new branch] gh/anshul-si/42/head -> origin/gh/anshul-si/42/head 2025-10-10T01:18:33.3983925Z * [new branch] gh/anshul-si/42/orig -> origin/gh/anshul-si/42/orig 2025-10-10T01:18:33.3986269Z * [new branch] gh/anshul-si/43/base -> origin/gh/anshul-si/43/base 2025-10-10T01:18:33.3987980Z * [new branch] gh/anshul-si/43/head -> origin/gh/anshul-si/43/head 2025-10-10T01:18:33.3989604Z * [new branch] gh/anshul-si/43/orig -> origin/gh/anshul-si/43/orig 2025-10-10T01:18:33.3992433Z * [new branch] gh/anshul-si/44/base -> origin/gh/anshul-si/44/base 2025-10-10T01:18:33.3994246Z * [new branch] gh/anshul-si/44/head -> origin/gh/anshul-si/44/head 2025-10-10T01:18:33.3996149Z * [new branch] gh/anshul-si/44/orig -> origin/gh/anshul-si/44/orig 2025-10-10T01:18:33.3999305Z * [new branch] gh/anshul-si/45/base -> origin/gh/anshul-si/45/base 2025-10-10T01:18:33.4000775Z * [new branch] gh/anshul-si/45/head -> origin/gh/anshul-si/45/head 2025-10-10T01:18:33.4002422Z * [new branch] gh/anshul-si/45/orig -> origin/gh/anshul-si/45/orig 2025-10-10T01:18:33.4005211Z * [new branch] gh/anshul-si/46/base -> origin/gh/anshul-si/46/base 2025-10-10T01:18:33.4006730Z * [new branch] gh/anshul-si/46/head -> origin/gh/anshul-si/46/head 2025-10-10T01:18:33.4008454Z * [new branch] gh/anshul-si/46/orig -> origin/gh/anshul-si/46/orig 2025-10-10T01:18:33.4011013Z * [new branch] gh/anshul-si/47/base -> origin/gh/anshul-si/47/base 2025-10-10T01:18:33.4013193Z * [new branch] gh/anshul-si/47/head -> origin/gh/anshul-si/47/head 2025-10-10T01:18:33.4014839Z * [new branch] gh/anshul-si/47/orig -> origin/gh/anshul-si/47/orig 2025-10-10T01:18:33.4017516Z * [new branch] gh/anshul-si/48/base -> origin/gh/anshul-si/48/base 2025-10-10T01:18:33.4019166Z * [new branch] gh/anshul-si/48/head -> origin/gh/anshul-si/48/head 2025-10-10T01:18:33.4020907Z * [new branch] gh/anshul-si/48/orig -> origin/gh/anshul-si/48/orig 2025-10-10T01:18:33.4023278Z * [new branch] gh/anshul-si/49/base -> origin/gh/anshul-si/49/base 2025-10-10T01:18:33.4026694Z * [new branch] gh/anshul-si/49/head -> origin/gh/anshul-si/49/head 2025-10-10T01:18:33.4028327Z * [new branch] gh/anshul-si/49/orig -> origin/gh/anshul-si/49/orig 2025-10-10T01:18:33.4030552Z * [new branch] gh/anshul-si/5/base -> origin/gh/anshul-si/5/base 2025-10-10T01:18:33.4032148Z * [new branch] gh/anshul-si/5/head -> origin/gh/anshul-si/5/head 2025-10-10T01:18:33.4034597Z * [new branch] gh/anshul-si/50/base -> origin/gh/anshul-si/50/base 2025-10-10T01:18:33.4036840Z * [new branch] gh/anshul-si/50/head -> origin/gh/anshul-si/50/head 2025-10-10T01:18:33.4038684Z * [new branch] gh/anshul-si/50/orig -> origin/gh/anshul-si/50/orig 2025-10-10T01:18:33.4041504Z * [new branch] gh/anshul-si/51/base -> origin/gh/anshul-si/51/base 2025-10-10T01:18:33.4042855Z * [new branch] gh/anshul-si/51/head -> origin/gh/anshul-si/51/head 2025-10-10T01:18:33.4044473Z * [new branch] gh/anshul-si/51/orig -> origin/gh/anshul-si/51/orig 2025-10-10T01:18:33.4047068Z * [new branch] gh/anshul-si/52/base -> origin/gh/anshul-si/52/base 2025-10-10T01:18:33.4048747Z * [new branch] gh/anshul-si/52/head -> origin/gh/anshul-si/52/head 2025-10-10T01:18:33.4050416Z * [new branch] gh/anshul-si/52/orig -> origin/gh/anshul-si/52/orig 2025-10-10T01:18:33.4053274Z * [new branch] gh/aorenste/132/base -> origin/gh/aorenste/132/base 2025-10-10T01:18:33.4055090Z * [new branch] gh/aorenste/132/head -> origin/gh/aorenste/132/head 2025-10-10T01:18:33.4057364Z * [new branch] gh/aorenste/133/base -> origin/gh/aorenste/133/base 2025-10-10T01:18:33.4059164Z * [new branch] gh/aorenste/133/head -> origin/gh/aorenste/133/head 2025-10-10T01:18:33.4060808Z * [new branch] gh/aorenste/133/orig -> origin/gh/aorenste/133/orig 2025-10-10T01:18:33.4063026Z * [new branch] gh/aorenste/134/base -> origin/gh/aorenste/134/base 2025-10-10T01:18:33.4064865Z * [new branch] gh/aorenste/134/head -> origin/gh/aorenste/134/head 2025-10-10T01:18:33.4066557Z * [new branch] gh/aorenste/134/orig -> origin/gh/aorenste/134/orig 2025-10-10T01:18:33.4068743Z * [new branch] gh/aorenste/135/base -> origin/gh/aorenste/135/base 2025-10-10T01:18:33.4070394Z * [new branch] gh/aorenste/135/head -> origin/gh/aorenste/135/head 2025-10-10T01:18:33.4072081Z * [new branch] gh/aorenste/135/orig -> origin/gh/aorenste/135/orig 2025-10-10T01:18:33.4074478Z * [new branch] gh/aorenste/136/base -> origin/gh/aorenste/136/base 2025-10-10T01:18:33.4076324Z * [new branch] gh/aorenste/136/head -> origin/gh/aorenste/136/head 2025-10-10T01:18:33.4078002Z * [new branch] gh/aorenste/136/orig -> origin/gh/aorenste/136/orig 2025-10-10T01:18:33.4080519Z * [new branch] gh/aorenste/137/base -> origin/gh/aorenste/137/base 2025-10-10T01:18:33.4082315Z * [new branch] gh/aorenste/137/head -> origin/gh/aorenste/137/head 2025-10-10T01:18:33.4084014Z * [new branch] gh/aorenste/137/orig -> origin/gh/aorenste/137/orig 2025-10-10T01:18:33.4087054Z * [new branch] gh/aorenste/138/base -> origin/gh/aorenste/138/base 2025-10-10T01:18:33.4088478Z * [new branch] gh/aorenste/138/head -> origin/gh/aorenste/138/head 2025-10-10T01:18:33.4090106Z * [new branch] gh/aorenste/138/orig -> origin/gh/aorenste/138/orig 2025-10-10T01:18:33.4092748Z * [new branch] gh/aorenste/139/base -> origin/gh/aorenste/139/base 2025-10-10T01:18:33.4094129Z * [new branch] gh/aorenste/139/head -> origin/gh/aorenste/139/head 2025-10-10T01:18:33.4095962Z * [new branch] gh/aorenste/139/orig -> origin/gh/aorenste/139/orig 2025-10-10T01:18:33.4098943Z * [new branch] gh/avikchaudhuri/1/base -> origin/gh/avikchaudhuri/1/base 2025-10-10T01:18:33.4100806Z * [new branch] gh/avikchaudhuri/1/head -> origin/gh/avikchaudhuri/1/head 2025-10-10T01:18:33.4102932Z * [new branch] gh/avikchaudhuri/2/base -> origin/gh/avikchaudhuri/2/base 2025-10-10T01:18:33.4104581Z * [new branch] gh/avikchaudhuri/2/head -> origin/gh/avikchaudhuri/2/head 2025-10-10T01:18:33.4106334Z * [new branch] gh/avikchaudhuri/2/orig -> origin/gh/avikchaudhuri/2/orig 2025-10-10T01:18:33.4109132Z * [new branch] gh/bdhirsh/650/base -> origin/gh/bdhirsh/650/base 2025-10-10T01:18:33.4110895Z * [new branch] gh/bdhirsh/650/head -> origin/gh/bdhirsh/650/head 2025-10-10T01:18:33.4112565Z * [new branch] gh/bdhirsh/650/orig -> origin/gh/bdhirsh/650/orig 2025-10-10T01:18:33.4114983Z * [new branch] gh/bdhirsh/665/base -> origin/gh/bdhirsh/665/base 2025-10-10T01:18:33.4116639Z * [new branch] gh/bdhirsh/665/head -> origin/gh/bdhirsh/665/head 2025-10-10T01:18:33.4118322Z * [new branch] gh/bdhirsh/665/orig -> origin/gh/bdhirsh/665/orig 2025-10-10T01:18:33.4121008Z * [new branch] gh/bdhirsh/666/base -> origin/gh/bdhirsh/666/base 2025-10-10T01:18:33.4122792Z * [new branch] gh/bdhirsh/666/head -> origin/gh/bdhirsh/666/head 2025-10-10T01:18:33.4124741Z * [new branch] gh/bdhirsh/666/orig -> origin/gh/bdhirsh/666/orig 2025-10-10T01:18:33.4126999Z * [new branch] gh/bdhirsh/668/base -> origin/gh/bdhirsh/668/base 2025-10-10T01:18:33.4128700Z * [new branch] gh/bdhirsh/668/head -> origin/gh/bdhirsh/668/head 2025-10-10T01:18:33.4130336Z * [new branch] gh/bdhirsh/668/orig -> origin/gh/bdhirsh/668/orig 2025-10-10T01:18:33.4132809Z * [new branch] gh/bdhirsh/669/base -> origin/gh/bdhirsh/669/base 2025-10-10T01:18:33.4134432Z * [new branch] gh/bdhirsh/669/head -> origin/gh/bdhirsh/669/head 2025-10-10T01:18:33.4136087Z * [new branch] gh/bdhirsh/669/orig -> origin/gh/bdhirsh/669/orig 2025-10-10T01:18:33.4138535Z * [new branch] gh/bdhirsh/670/base -> origin/gh/bdhirsh/670/base 2025-10-10T01:18:33.4140323Z * [new branch] gh/bdhirsh/670/head -> origin/gh/bdhirsh/670/head 2025-10-10T01:18:33.4142018Z * [new branch] gh/bdhirsh/670/orig -> origin/gh/bdhirsh/670/orig 2025-10-10T01:18:33.4144376Z * [new branch] gh/bdhirsh/671/base -> origin/gh/bdhirsh/671/base 2025-10-10T01:18:33.4146348Z * [new branch] gh/bdhirsh/671/head -> origin/gh/bdhirsh/671/head 2025-10-10T01:18:33.4148513Z * [new branch] gh/bdhirsh/671/orig -> origin/gh/bdhirsh/671/orig 2025-10-10T01:18:33.4150632Z * [new branch] gh/bdhirsh/672/base -> origin/gh/bdhirsh/672/base 2025-10-10T01:18:33.4152255Z * [new branch] gh/bdhirsh/672/head -> origin/gh/bdhirsh/672/head 2025-10-10T01:18:33.4153928Z * [new branch] gh/bdhirsh/672/orig -> origin/gh/bdhirsh/672/orig 2025-10-10T01:18:33.4157106Z * [new branch] gh/benjaminglass1/101/base -> origin/gh/benjaminglass1/101/base 2025-10-10T01:18:33.4158859Z * [new branch] gh/benjaminglass1/101/head -> origin/gh/benjaminglass1/101/head 2025-10-10T01:18:33.4160590Z * [new branch] gh/benjaminglass1/101/orig -> origin/gh/benjaminglass1/101/orig 2025-10-10T01:18:33.4162820Z * [new branch] gh/benjaminglass1/102/base -> origin/gh/benjaminglass1/102/base 2025-10-10T01:18:33.4164593Z * [new branch] gh/benjaminglass1/102/head -> origin/gh/benjaminglass1/102/head 2025-10-10T01:18:33.4166569Z * [new branch] gh/benjaminglass1/102/orig -> origin/gh/benjaminglass1/102/orig 2025-10-10T01:18:33.4169744Z * [new branch] gh/benjaminglass1/106/base -> origin/gh/benjaminglass1/106/base 2025-10-10T01:18:33.4170801Z * [new branch] gh/benjaminglass1/106/head -> origin/gh/benjaminglass1/106/head 2025-10-10T01:18:33.4172365Z * [new branch] gh/benjaminglass1/106/orig -> origin/gh/benjaminglass1/106/orig 2025-10-10T01:18:33.4175133Z * [new branch] gh/benjaminglass1/107/base -> origin/gh/benjaminglass1/107/base 2025-10-10T01:18:33.4176967Z * [new branch] gh/benjaminglass1/107/head -> origin/gh/benjaminglass1/107/head 2025-10-10T01:18:33.4178568Z * [new branch] gh/benjaminglass1/107/orig -> origin/gh/benjaminglass1/107/orig 2025-10-10T01:18:33.4180683Z * [new branch] gh/benjaminglass1/108/base -> origin/gh/benjaminglass1/108/base 2025-10-10T01:18:33.4182364Z * [new branch] gh/benjaminglass1/108/head -> origin/gh/benjaminglass1/108/head 2025-10-10T01:18:33.4184027Z * [new branch] gh/benjaminglass1/108/orig -> origin/gh/benjaminglass1/108/orig 2025-10-10T01:18:33.4186790Z * [new branch] gh/benjaminglass1/79/base -> origin/gh/benjaminglass1/79/base 2025-10-10T01:18:33.4188109Z * [new branch] gh/benjaminglass1/79/head -> origin/gh/benjaminglass1/79/head 2025-10-10T01:18:33.4189927Z * [new branch] gh/benjaminglass1/79/orig -> origin/gh/benjaminglass1/79/orig 2025-10-10T01:18:33.4192627Z * [new branch] gh/benjaminglass1/86/base -> origin/gh/benjaminglass1/86/base 2025-10-10T01:18:33.4193894Z * [new branch] gh/benjaminglass1/86/head -> origin/gh/benjaminglass1/86/head 2025-10-10T01:18:33.4195654Z * [new branch] gh/benjaminglass1/86/orig -> origin/gh/benjaminglass1/86/orig 2025-10-10T01:18:33.4197791Z * [new branch] gh/benjaminglass1/95/base -> origin/gh/benjaminglass1/95/base 2025-10-10T01:18:33.4199698Z * [new branch] gh/benjaminglass1/95/head -> origin/gh/benjaminglass1/95/head 2025-10-10T01:18:33.4201310Z * [new branch] gh/benjaminglass1/95/orig -> origin/gh/benjaminglass1/95/orig 2025-10-10T01:18:33.4203697Z * [new branch] gh/benjaminglass1/97/base -> origin/gh/benjaminglass1/97/base 2025-10-10T01:18:33.4205387Z * [new branch] gh/benjaminglass1/97/head -> origin/gh/benjaminglass1/97/head 2025-10-10T01:18:33.4207022Z * [new branch] gh/benjaminglass1/97/orig -> origin/gh/benjaminglass1/97/orig 2025-10-10T01:18:33.4209799Z * [new branch] gh/bobrenjc93/542/base -> origin/gh/bobrenjc93/542/base 2025-10-10T01:18:33.4211478Z * [new branch] gh/bobrenjc93/542/head -> origin/gh/bobrenjc93/542/head 2025-10-10T01:18:33.4213126Z * [new branch] gh/bobrenjc93/542/orig -> origin/gh/bobrenjc93/542/orig 2025-10-10T01:18:33.4215384Z * [new branch] gh/bobrenjc93/543/base -> origin/gh/bobrenjc93/543/base 2025-10-10T01:18:33.4217185Z * [new branch] gh/bobrenjc93/543/head -> origin/gh/bobrenjc93/543/head 2025-10-10T01:18:33.4218839Z * [new branch] gh/bobrenjc93/543/orig -> origin/gh/bobrenjc93/543/orig 2025-10-10T01:18:33.4220959Z * [new branch] gh/bobrenjc93/545/base -> origin/gh/bobrenjc93/545/base 2025-10-10T01:18:33.4222719Z * [new branch] gh/bobrenjc93/545/head -> origin/gh/bobrenjc93/545/head 2025-10-10T01:18:33.4224831Z * [new branch] gh/bobrenjc93/545/orig -> origin/gh/bobrenjc93/545/orig 2025-10-10T01:18:33.4227510Z * [new branch] gh/bobrenjc93/547/base -> origin/gh/bobrenjc93/547/base 2025-10-10T01:18:33.4229200Z * [new branch] gh/bobrenjc93/547/head -> origin/gh/bobrenjc93/547/head 2025-10-10T01:18:33.4230910Z * [new branch] gh/bobrenjc93/547/orig -> origin/gh/bobrenjc93/547/orig 2025-10-10T01:18:33.4233042Z * [new branch] gh/bobrenjc93/548/base -> origin/gh/bobrenjc93/548/base 2025-10-10T01:18:33.4234669Z * [new branch] gh/bobrenjc93/548/head -> origin/gh/bobrenjc93/548/head 2025-10-10T01:18:33.4236306Z * [new branch] gh/bobrenjc93/548/orig -> origin/gh/bobrenjc93/548/orig 2025-10-10T01:18:33.4238579Z * [new branch] gh/bobrenjc93/553/base -> origin/gh/bobrenjc93/553/base 2025-10-10T01:18:33.4240288Z * [new branch] gh/bobrenjc93/553/head -> origin/gh/bobrenjc93/553/head 2025-10-10T01:18:33.4242101Z * [new branch] gh/bobrenjc93/553/orig -> origin/gh/bobrenjc93/553/orig 2025-10-10T01:18:33.4244107Z * [new branch] gh/bobrenjc93/554/base -> origin/gh/bobrenjc93/554/base 2025-10-10T01:18:33.4245953Z * [new branch] gh/bobrenjc93/554/head -> origin/gh/bobrenjc93/554/head 2025-10-10T01:18:33.4248147Z * [new branch] gh/bobrenjc93/554/orig -> origin/gh/bobrenjc93/554/orig 2025-10-10T01:18:33.4250249Z * [new branch] gh/bobrenjc93/555/base -> origin/gh/bobrenjc93/555/base 2025-10-10T01:18:33.4251851Z * [new branch] gh/bobrenjc93/555/head -> origin/gh/bobrenjc93/555/head 2025-10-10T01:18:33.4253461Z * [new branch] gh/bobrenjc93/555/orig -> origin/gh/bobrenjc93/555/orig 2025-10-10T01:18:33.4256779Z * [new branch] gh/bobrenjc93/557/base -> origin/gh/bobrenjc93/557/base 2025-10-10T01:18:33.4258016Z * [new branch] gh/bobrenjc93/557/head -> origin/gh/bobrenjc93/557/head 2025-10-10T01:18:33.4259675Z * [new branch] gh/bobrenjc93/557/orig -> origin/gh/bobrenjc93/557/orig 2025-10-10T01:18:33.4261946Z * [new branch] gh/bobrenjc93/558/base -> origin/gh/bobrenjc93/558/base 2025-10-10T01:18:33.4263769Z * [new branch] gh/bobrenjc93/558/head -> origin/gh/bobrenjc93/558/head 2025-10-10T01:18:33.4265365Z * [new branch] gh/bobrenjc93/558/orig -> origin/gh/bobrenjc93/558/orig 2025-10-10T01:18:33.4267731Z * [new branch] gh/bobrenjc93/559/base -> origin/gh/bobrenjc93/559/base 2025-10-10T01:18:33.4269341Z * [new branch] gh/bobrenjc93/559/head -> origin/gh/bobrenjc93/559/head 2025-10-10T01:18:33.4271101Z * [new branch] gh/bobrenjc93/559/orig -> origin/gh/bobrenjc93/559/orig 2025-10-10T01:18:33.4273299Z * [new branch] gh/bobrenjc93/560/base -> origin/gh/bobrenjc93/560/base 2025-10-10T01:18:33.4275043Z * [new branch] gh/bobrenjc93/560/head -> origin/gh/bobrenjc93/560/head 2025-10-10T01:18:33.4276699Z * [new branch] gh/bobrenjc93/560/orig -> origin/gh/bobrenjc93/560/orig 2025-10-10T01:18:33.4279255Z * [new branch] gh/bobrenjc93/561/base -> origin/gh/bobrenjc93/561/base 2025-10-10T01:18:33.4280911Z * [new branch] gh/bobrenjc93/561/head -> origin/gh/bobrenjc93/561/head 2025-10-10T01:18:33.4282639Z * [new branch] gh/bobrenjc93/561/orig -> origin/gh/bobrenjc93/561/orig 2025-10-10T01:18:33.4284981Z * [new branch] gh/bobrenjc93/562/base -> origin/gh/bobrenjc93/562/base 2025-10-10T01:18:33.4286605Z * [new branch] gh/bobrenjc93/562/head -> origin/gh/bobrenjc93/562/head 2025-10-10T01:18:33.4288312Z * [new branch] gh/bobrenjc93/562/orig -> origin/gh/bobrenjc93/562/orig 2025-10-10T01:18:33.4290657Z * [new branch] gh/bobrenjc93/563/base -> origin/gh/bobrenjc93/563/base 2025-10-10T01:18:33.4292261Z * [new branch] gh/bobrenjc93/563/head -> origin/gh/bobrenjc93/563/head 2025-10-10T01:18:33.4294049Z * [new branch] gh/bobrenjc93/563/orig -> origin/gh/bobrenjc93/563/orig 2025-10-10T01:18:33.4296756Z * [new branch] gh/bobrenjc93/564/base -> origin/gh/bobrenjc93/564/base 2025-10-10T01:18:33.4298738Z * [new branch] gh/bobrenjc93/564/head -> origin/gh/bobrenjc93/564/head 2025-10-10T01:18:33.4300315Z * [new branch] gh/bobrenjc93/564/orig -> origin/gh/bobrenjc93/564/orig 2025-10-10T01:18:33.4303296Z * [new branch] gh/bobrenjc93/565/base -> origin/gh/bobrenjc93/565/base 2025-10-10T01:18:33.4304596Z * [new branch] gh/bobrenjc93/565/head -> origin/gh/bobrenjc93/565/head 2025-10-10T01:18:33.4306609Z * [new branch] gh/bobrenjc93/565/orig -> origin/gh/bobrenjc93/565/orig 2025-10-10T01:18:33.4308998Z * [new branch] gh/bobrenjc93/566/base -> origin/gh/bobrenjc93/566/base 2025-10-10T01:18:33.4310598Z * [new branch] gh/bobrenjc93/566/head -> origin/gh/bobrenjc93/566/head 2025-10-10T01:18:33.4312240Z * [new branch] gh/bobrenjc93/566/orig -> origin/gh/bobrenjc93/566/orig 2025-10-10T01:18:33.4314433Z * [new branch] gh/bobrenjc93/567/base -> origin/gh/bobrenjc93/567/base 2025-10-10T01:18:33.4316099Z * [new branch] gh/bobrenjc93/567/head -> origin/gh/bobrenjc93/567/head 2025-10-10T01:18:33.4317889Z * [new branch] gh/bobrenjc93/567/orig -> origin/gh/bobrenjc93/567/orig 2025-10-10T01:18:33.4320233Z * [new branch] gh/bobrenjc93/568/base -> origin/gh/bobrenjc93/568/base 2025-10-10T01:18:33.4322038Z * [new branch] gh/bobrenjc93/568/head -> origin/gh/bobrenjc93/568/head 2025-10-10T01:18:33.4323674Z * [new branch] gh/bobrenjc93/568/orig -> origin/gh/bobrenjc93/568/orig 2025-10-10T01:18:33.4327576Z * [new branch] gh/bobrenjc93/569/base -> origin/gh/bobrenjc93/569/base 2025-10-10T01:18:33.4329442Z * [new branch] gh/bobrenjc93/569/head -> origin/gh/bobrenjc93/569/head 2025-10-10T01:18:33.4331164Z * [new branch] gh/bobrenjc93/569/orig -> origin/gh/bobrenjc93/569/orig 2025-10-10T01:18:33.4333491Z * [new branch] gh/bobrenjc93/570/base -> origin/gh/bobrenjc93/570/base 2025-10-10T01:18:33.4335219Z * [new branch] gh/bobrenjc93/570/head -> origin/gh/bobrenjc93/570/head 2025-10-10T01:18:33.4336926Z * [new branch] gh/bobrenjc93/570/orig -> origin/gh/bobrenjc93/570/orig 2025-10-10T01:18:33.4339169Z * [new branch] gh/bobrenjc93/571/base -> origin/gh/bobrenjc93/571/base 2025-10-10T01:18:33.4340872Z * [new branch] gh/bobrenjc93/571/head -> origin/gh/bobrenjc93/571/head 2025-10-10T01:18:33.4342593Z * [new branch] gh/bobrenjc93/571/orig -> origin/gh/bobrenjc93/571/orig 2025-10-10T01:18:33.4344954Z * [new branch] gh/bobrenjc93/572/base -> origin/gh/bobrenjc93/572/base 2025-10-10T01:18:33.4346697Z * [new branch] gh/bobrenjc93/572/head -> origin/gh/bobrenjc93/572/head 2025-10-10T01:18:33.4348453Z * [new branch] gh/bobrenjc93/572/orig -> origin/gh/bobrenjc93/572/orig 2025-10-10T01:18:33.4350918Z * [new branch] gh/bobrenjc93/573/base -> origin/gh/bobrenjc93/573/base 2025-10-10T01:18:33.4352751Z * [new branch] gh/bobrenjc93/573/head -> origin/gh/bobrenjc93/573/head 2025-10-10T01:18:33.4354443Z * [new branch] gh/bobrenjc93/573/orig -> origin/gh/bobrenjc93/573/orig 2025-10-10T01:18:33.4357138Z * [new branch] gh/bobrenjc93/574/base -> origin/gh/bobrenjc93/574/base 2025-10-10T01:18:33.4359149Z * [new branch] gh/bobrenjc93/574/head -> origin/gh/bobrenjc93/574/head 2025-10-10T01:18:33.4360886Z * [new branch] gh/bobrenjc93/574/orig -> origin/gh/bobrenjc93/574/orig 2025-10-10T01:18:33.4363593Z * [new branch] gh/bobrenjc93/575/base -> origin/gh/bobrenjc93/575/base 2025-10-10T01:18:33.4364921Z * [new branch] gh/bobrenjc93/575/head -> origin/gh/bobrenjc93/575/head 2025-10-10T01:18:33.4366846Z * [new branch] gh/bobrenjc93/575/orig -> origin/gh/bobrenjc93/575/orig 2025-10-10T01:18:33.4369705Z * [new branch] gh/bobrenjc93/576/base -> origin/gh/bobrenjc93/576/base 2025-10-10T01:18:33.4371414Z * [new branch] gh/bobrenjc93/576/head -> origin/gh/bobrenjc93/576/head 2025-10-10T01:18:33.4373132Z * [new branch] gh/bobrenjc93/576/orig -> origin/gh/bobrenjc93/576/orig 2025-10-10T01:18:33.4375637Z * [new branch] gh/bobrenjc93/577/base -> origin/gh/bobrenjc93/577/base 2025-10-10T01:18:33.4377546Z * [new branch] gh/bobrenjc93/577/head -> origin/gh/bobrenjc93/577/head 2025-10-10T01:18:33.4379140Z * [new branch] gh/bobrenjc93/577/orig -> origin/gh/bobrenjc93/577/orig 2025-10-10T01:18:33.4381490Z * [new branch] gh/bobrenjc93/578/base -> origin/gh/bobrenjc93/578/base 2025-10-10T01:18:33.4383226Z * [new branch] gh/bobrenjc93/578/head -> origin/gh/bobrenjc93/578/head 2025-10-10T01:18:33.4384930Z * [new branch] gh/bobrenjc93/578/orig -> origin/gh/bobrenjc93/578/orig 2025-10-10T01:18:33.4387323Z * [new branch] gh/bobrenjc93/579/base -> origin/gh/bobrenjc93/579/base 2025-10-10T01:18:33.4389003Z * [new branch] gh/bobrenjc93/579/head -> origin/gh/bobrenjc93/579/head 2025-10-10T01:18:33.4390758Z * [new branch] gh/bobrenjc93/579/orig -> origin/gh/bobrenjc93/579/orig 2025-10-10T01:18:33.4393150Z * [new branch] gh/bobrenjc93/580/base -> origin/gh/bobrenjc93/580/base 2025-10-10T01:18:33.4395078Z * [new branch] gh/bobrenjc93/580/head -> origin/gh/bobrenjc93/580/head 2025-10-10T01:18:33.4396783Z * [new branch] gh/bobrenjc93/580/orig -> origin/gh/bobrenjc93/580/orig 2025-10-10T01:18:33.4399368Z * [new branch] gh/bobrenjc93/581/base -> origin/gh/bobrenjc93/581/base 2025-10-10T01:18:33.4401100Z * [new branch] gh/bobrenjc93/581/head -> origin/gh/bobrenjc93/581/head 2025-10-10T01:18:33.4402758Z * [new branch] gh/bobrenjc93/581/orig -> origin/gh/bobrenjc93/581/orig 2025-10-10T01:18:33.4405163Z * [new branch] gh/bobrenjc93/582/base -> origin/gh/bobrenjc93/582/base 2025-10-10T01:18:33.4406849Z * [new branch] gh/bobrenjc93/582/head -> origin/gh/bobrenjc93/582/head 2025-10-10T01:18:33.4408553Z * [new branch] gh/bobrenjc93/582/orig -> origin/gh/bobrenjc93/582/orig 2025-10-10T01:18:33.4410971Z * [new branch] gh/bobrenjc93/583/base -> origin/gh/bobrenjc93/583/base 2025-10-10T01:18:33.4412729Z * [new branch] gh/bobrenjc93/583/head -> origin/gh/bobrenjc93/583/head 2025-10-10T01:18:33.4414460Z * [new branch] gh/bobrenjc93/583/orig -> origin/gh/bobrenjc93/583/orig 2025-10-10T01:18:33.4416947Z * [new branch] gh/bobrenjc93/584/base -> origin/gh/bobrenjc93/584/base 2025-10-10T01:18:33.4418675Z * [new branch] gh/bobrenjc93/584/head -> origin/gh/bobrenjc93/584/head 2025-10-10T01:18:33.4420395Z * [new branch] gh/bobrenjc93/584/orig -> origin/gh/bobrenjc93/584/orig 2025-10-10T01:18:33.4422775Z * [new branch] gh/bobrenjc93/585/base -> origin/gh/bobrenjc93/585/base 2025-10-10T01:18:33.4424842Z * [new branch] gh/bobrenjc93/585/head -> origin/gh/bobrenjc93/585/head 2025-10-10T01:18:33.4426639Z * [new branch] gh/bobrenjc93/585/orig -> origin/gh/bobrenjc93/585/orig 2025-10-10T01:18:33.4429031Z * [new branch] gh/bobrenjc93/586/base -> origin/gh/bobrenjc93/586/base 2025-10-10T01:18:33.4430870Z * [new branch] gh/bobrenjc93/586/head -> origin/gh/bobrenjc93/586/head 2025-10-10T01:18:33.4432339Z * [new branch] gh/bobrenjc93/586/orig -> origin/gh/bobrenjc93/586/orig 2025-10-10T01:18:33.4434680Z * [new branch] gh/bobrenjc93/587/base -> origin/gh/bobrenjc93/587/base 2025-10-10T01:18:33.4436395Z * [new branch] gh/bobrenjc93/587/head -> origin/gh/bobrenjc93/587/head 2025-10-10T01:18:33.4438115Z * [new branch] gh/bobrenjc93/587/orig -> origin/gh/bobrenjc93/587/orig 2025-10-10T01:18:33.4440612Z * [new branch] gh/bobrenjc93/588/base -> origin/gh/bobrenjc93/588/base 2025-10-10T01:18:33.4442294Z * [new branch] gh/bobrenjc93/588/head -> origin/gh/bobrenjc93/588/head 2025-10-10T01:18:33.4443991Z * [new branch] gh/bobrenjc93/588/orig -> origin/gh/bobrenjc93/588/orig 2025-10-10T01:18:33.4446627Z * [new branch] gh/bobrenjc93/589/base -> origin/gh/bobrenjc93/589/base 2025-10-10T01:18:33.4448215Z * [new branch] gh/bobrenjc93/589/head -> origin/gh/bobrenjc93/589/head 2025-10-10T01:18:33.4449863Z * [new branch] gh/bobrenjc93/589/orig -> origin/gh/bobrenjc93/589/orig 2025-10-10T01:18:33.4452247Z * [new branch] gh/bobrenjc93/590/base -> origin/gh/bobrenjc93/590/base 2025-10-10T01:18:33.4453995Z * [new branch] gh/bobrenjc93/590/head -> origin/gh/bobrenjc93/590/head 2025-10-10T01:18:33.4456158Z * [new branch] gh/bobrenjc93/590/orig -> origin/gh/bobrenjc93/590/orig 2025-10-10T01:18:33.4458167Z * [new branch] gh/bobrenjc93/591/base -> origin/gh/bobrenjc93/591/base 2025-10-10T01:18:33.4459797Z * [new branch] gh/bobrenjc93/591/head -> origin/gh/bobrenjc93/591/head 2025-10-10T01:18:33.4461683Z * [new branch] gh/bobrenjc93/591/orig -> origin/gh/bobrenjc93/591/orig 2025-10-10T01:18:33.4464467Z * [new branch] gh/bobrenjc93/592/base -> origin/gh/bobrenjc93/592/base 2025-10-10T01:18:33.4466040Z * [new branch] gh/bobrenjc93/592/head -> origin/gh/bobrenjc93/592/head 2025-10-10T01:18:33.4467733Z * [new branch] gh/bobrenjc93/592/orig -> origin/gh/bobrenjc93/592/orig 2025-10-10T01:18:33.4470300Z * [new branch] gh/bobrenjc93/593/base -> origin/gh/bobrenjc93/593/base 2025-10-10T01:18:33.4472069Z * [new branch] gh/bobrenjc93/593/head -> origin/gh/bobrenjc93/593/head 2025-10-10T01:18:33.4473749Z * [new branch] gh/bobrenjc93/593/orig -> origin/gh/bobrenjc93/593/orig 2025-10-10T01:18:33.4475973Z * [new branch] gh/bobrenjc93/594/base -> origin/gh/bobrenjc93/594/base 2025-10-10T01:18:33.4477852Z * [new branch] gh/bobrenjc93/594/head -> origin/gh/bobrenjc93/594/head 2025-10-10T01:18:33.4479618Z * [new branch] gh/bobrenjc93/594/orig -> origin/gh/bobrenjc93/594/orig 2025-10-10T01:18:33.4481802Z * [new branch] gh/bobrenjc93/595/base -> origin/gh/bobrenjc93/595/base 2025-10-10T01:18:33.4483488Z * [new branch] gh/bobrenjc93/595/head -> origin/gh/bobrenjc93/595/head 2025-10-10T01:18:33.4485143Z * [new branch] gh/bobrenjc93/595/orig -> origin/gh/bobrenjc93/595/orig 2025-10-10T01:18:33.4487386Z * [new branch] gh/bobrenjc93/596/base -> origin/gh/bobrenjc93/596/base 2025-10-10T01:18:33.4489053Z * [new branch] gh/bobrenjc93/596/head -> origin/gh/bobrenjc93/596/head 2025-10-10T01:18:33.4490746Z * [new branch] gh/bobrenjc93/596/orig -> origin/gh/bobrenjc93/596/orig 2025-10-10T01:18:33.4493099Z * [new branch] gh/bobrenjc93/597/base -> origin/gh/bobrenjc93/597/base 2025-10-10T01:18:33.4494802Z * [new branch] gh/bobrenjc93/597/head -> origin/gh/bobrenjc93/597/head 2025-10-10T01:18:33.4496616Z * [new branch] gh/bobrenjc93/597/orig -> origin/gh/bobrenjc93/597/orig 2025-10-10T01:18:33.4499359Z * [new branch] gh/bobrenjc93/598/base -> origin/gh/bobrenjc93/598/base 2025-10-10T01:18:33.4501109Z * [new branch] gh/bobrenjc93/598/head -> origin/gh/bobrenjc93/598/head 2025-10-10T01:18:33.4502791Z * [new branch] gh/bobrenjc93/598/orig -> origin/gh/bobrenjc93/598/orig 2025-10-10T01:18:33.4505165Z * [new branch] gh/bobrenjc93/599/base -> origin/gh/bobrenjc93/599/base 2025-10-10T01:18:33.4506883Z * [new branch] gh/bobrenjc93/599/head -> origin/gh/bobrenjc93/599/head 2025-10-10T01:18:33.4508637Z * [new branch] gh/bobrenjc93/599/orig -> origin/gh/bobrenjc93/599/orig 2025-10-10T01:18:33.4510855Z * [new branch] gh/bobrenjc93/600/base -> origin/gh/bobrenjc93/600/base 2025-10-10T01:18:33.4513261Z * [new branch] gh/bobrenjc93/600/head -> origin/gh/bobrenjc93/600/head 2025-10-10T01:18:33.4514903Z * [new branch] gh/bobrenjc93/600/orig -> origin/gh/bobrenjc93/600/orig 2025-10-10T01:18:33.4518149Z * [new branch] gh/bobrenjc93/601/base -> origin/gh/bobrenjc93/601/base 2025-10-10T01:18:33.4520549Z * [new branch] gh/bobrenjc93/601/head -> origin/gh/bobrenjc93/601/head 2025-10-10T01:18:33.4522883Z * [new branch] gh/bobrenjc93/601/orig -> origin/gh/bobrenjc93/601/orig 2025-10-10T01:18:33.4526316Z * [new branch] gh/bobrenjc93/602/base -> origin/gh/bobrenjc93/602/base 2025-10-10T01:18:33.4528048Z * [new branch] gh/bobrenjc93/602/head -> origin/gh/bobrenjc93/602/head 2025-10-10T01:18:33.4530104Z * [new branch] gh/bobrenjc93/602/orig -> origin/gh/bobrenjc93/602/orig 2025-10-10T01:18:33.4532624Z * [new branch] gh/bobrenjc93/603/base -> origin/gh/bobrenjc93/603/base 2025-10-10T01:18:33.4534399Z * [new branch] gh/bobrenjc93/603/head -> origin/gh/bobrenjc93/603/head 2025-10-10T01:18:33.4536461Z * [new branch] gh/bobrenjc93/603/orig -> origin/gh/bobrenjc93/603/orig 2025-10-10T01:18:33.4539951Z * [new branch] gh/bobrenjc93/604/base -> origin/gh/bobrenjc93/604/base 2025-10-10T01:18:33.4541284Z * [new branch] gh/bobrenjc93/604/head -> origin/gh/bobrenjc93/604/head 2025-10-10T01:18:33.4542999Z * [new branch] gh/bobrenjc93/604/orig -> origin/gh/bobrenjc93/604/orig 2025-10-10T01:18:33.4545885Z * [new branch] gh/bobrenjc93/605/base -> origin/gh/bobrenjc93/605/base 2025-10-10T01:18:33.4547255Z * [new branch] gh/bobrenjc93/605/head -> origin/gh/bobrenjc93/605/head 2025-10-10T01:18:33.4549107Z * [new branch] gh/bobrenjc93/605/orig -> origin/gh/bobrenjc93/605/orig 2025-10-10T01:18:33.4551628Z * [new branch] gh/bobrenjc93/606/base -> origin/gh/bobrenjc93/606/base 2025-10-10T01:18:33.4553237Z * [new branch] gh/bobrenjc93/606/head -> origin/gh/bobrenjc93/606/head 2025-10-10T01:18:33.4555030Z * [new branch] gh/bobrenjc93/606/orig -> origin/gh/bobrenjc93/606/orig 2025-10-10T01:18:33.4557443Z * [new branch] gh/bobrenjc93/607/base -> origin/gh/bobrenjc93/607/base 2025-10-10T01:18:33.4559191Z * [new branch] gh/bobrenjc93/607/head -> origin/gh/bobrenjc93/607/head 2025-10-10T01:18:33.4560853Z * [new branch] gh/bobrenjc93/607/orig -> origin/gh/bobrenjc93/607/orig 2025-10-10T01:18:33.4563327Z * [new branch] gh/bobrenjc93/608/base -> origin/gh/bobrenjc93/608/base 2025-10-10T01:18:33.4565368Z * [new branch] gh/bobrenjc93/608/head -> origin/gh/bobrenjc93/608/head 2025-10-10T01:18:33.4567110Z * [new branch] gh/bobrenjc93/608/orig -> origin/gh/bobrenjc93/608/orig 2025-10-10T01:18:33.4569588Z * [new branch] gh/bobrenjc93/609/base -> origin/gh/bobrenjc93/609/base 2025-10-10T01:18:33.4571288Z * [new branch] gh/bobrenjc93/609/head -> origin/gh/bobrenjc93/609/head 2025-10-10T01:18:33.4573006Z * [new branch] gh/bobrenjc93/609/orig -> origin/gh/bobrenjc93/609/orig 2025-10-10T01:18:33.4575617Z * [new branch] gh/bobrenjc93/610/base -> origin/gh/bobrenjc93/610/base 2025-10-10T01:18:33.4577323Z * [new branch] gh/bobrenjc93/610/head -> origin/gh/bobrenjc93/610/head 2025-10-10T01:18:33.4579086Z * [new branch] gh/bobrenjc93/610/orig -> origin/gh/bobrenjc93/610/orig 2025-10-10T01:18:33.4581509Z * [new branch] gh/bobrenjc93/611/base -> origin/gh/bobrenjc93/611/base 2025-10-10T01:18:33.4583176Z * [new branch] gh/bobrenjc93/611/head -> origin/gh/bobrenjc93/611/head 2025-10-10T01:18:33.4584843Z * [new branch] gh/bobrenjc93/611/orig -> origin/gh/bobrenjc93/611/orig 2025-10-10T01:18:33.4587401Z * [new branch] gh/bobrenjc93/612/base -> origin/gh/bobrenjc93/612/base 2025-10-10T01:18:33.4589022Z * [new branch] gh/bobrenjc93/612/head -> origin/gh/bobrenjc93/612/head 2025-10-10T01:18:33.4590606Z * [new branch] gh/bobrenjc93/612/orig -> origin/gh/bobrenjc93/612/orig 2025-10-10T01:18:33.4593152Z * [new branch] gh/bobrenjc93/613/base -> origin/gh/bobrenjc93/613/base 2025-10-10T01:18:33.4594857Z * [new branch] gh/bobrenjc93/613/head -> origin/gh/bobrenjc93/613/head 2025-10-10T01:18:33.4596443Z * [new branch] gh/bobrenjc93/613/orig -> origin/gh/bobrenjc93/613/orig 2025-10-10T01:18:33.4598986Z * [new branch] gh/bobrenjc93/614/base -> origin/gh/bobrenjc93/614/base 2025-10-10T01:18:33.4600889Z * [new branch] gh/bobrenjc93/614/head -> origin/gh/bobrenjc93/614/head 2025-10-10T01:18:33.4602915Z * [new branch] gh/bobrenjc93/614/orig -> origin/gh/bobrenjc93/614/orig 2025-10-10T01:18:33.4604955Z * [new branch] gh/bobrenjc93/615/base -> origin/gh/bobrenjc93/615/base 2025-10-10T01:18:33.4606674Z * [new branch] gh/bobrenjc93/615/head -> origin/gh/bobrenjc93/615/head 2025-10-10T01:18:33.4608365Z * [new branch] gh/bobrenjc93/615/orig -> origin/gh/bobrenjc93/615/orig 2025-10-10T01:18:33.4610762Z * [new branch] gh/bobrenjc93/616/base -> origin/gh/bobrenjc93/616/base 2025-10-10T01:18:33.4612455Z * [new branch] gh/bobrenjc93/616/head -> origin/gh/bobrenjc93/616/head 2025-10-10T01:18:33.4614237Z * [new branch] gh/bobrenjc93/616/orig -> origin/gh/bobrenjc93/616/orig 2025-10-10T01:18:33.4616617Z * [new branch] gh/bobrenjc93/617/base -> origin/gh/bobrenjc93/617/base 2025-10-10T01:18:33.4618439Z * [new branch] gh/bobrenjc93/617/head -> origin/gh/bobrenjc93/617/head 2025-10-10T01:18:33.4620056Z * [new branch] gh/bobrenjc93/617/orig -> origin/gh/bobrenjc93/617/orig 2025-10-10T01:18:33.4622363Z * [new branch] gh/bobrenjc93/618/base -> origin/gh/bobrenjc93/618/base 2025-10-10T01:18:33.4624274Z * [new branch] gh/bobrenjc93/618/head -> origin/gh/bobrenjc93/618/head 2025-10-10T01:18:33.4626299Z * [new branch] gh/bobrenjc93/618/orig -> origin/gh/bobrenjc93/618/orig 2025-10-10T01:18:33.4629280Z * [new branch] gh/bobrenjc93/619/base -> origin/gh/bobrenjc93/619/base 2025-10-10T01:18:33.4630614Z * [new branch] gh/bobrenjc93/619/head -> origin/gh/bobrenjc93/619/head 2025-10-10T01:18:33.4632279Z * [new branch] gh/bobrenjc93/619/orig -> origin/gh/bobrenjc93/619/orig 2025-10-10T01:18:33.4635081Z * [new branch] gh/bobrenjc93/620/base -> origin/gh/bobrenjc93/620/base 2025-10-10T01:18:33.4636583Z * [new branch] gh/bobrenjc93/620/head -> origin/gh/bobrenjc93/620/head 2025-10-10T01:18:33.4638300Z * [new branch] gh/bobrenjc93/620/orig -> origin/gh/bobrenjc93/620/orig 2025-10-10T01:18:33.4640794Z * [new branch] gh/bobrenjc93/621/base -> origin/gh/bobrenjc93/621/base 2025-10-10T01:18:33.4642455Z * [new branch] gh/bobrenjc93/621/head -> origin/gh/bobrenjc93/621/head 2025-10-10T01:18:33.4644160Z * [new branch] gh/bobrenjc93/621/orig -> origin/gh/bobrenjc93/621/orig 2025-10-10T01:18:33.4647369Z * [new branch] gh/bobrenjc93/622/base -> origin/gh/bobrenjc93/622/base 2025-10-10T01:18:33.4648838Z * [new branch] gh/bobrenjc93/622/head -> origin/gh/bobrenjc93/622/head 2025-10-10T01:18:33.4650392Z * [new branch] gh/bobrenjc93/622/orig -> origin/gh/bobrenjc93/622/orig 2025-10-10T01:18:33.4653468Z * [new branch] gh/bobrenjc93/623/base -> origin/gh/bobrenjc93/623/base 2025-10-10T01:18:33.4655162Z * [new branch] gh/bobrenjc93/623/head -> origin/gh/bobrenjc93/623/head 2025-10-10T01:18:33.4656773Z * [new branch] gh/bobrenjc93/623/orig -> origin/gh/bobrenjc93/623/orig 2025-10-10T01:18:33.4659136Z * [new branch] gh/bobrenjc93/624/base -> origin/gh/bobrenjc93/624/base 2025-10-10T01:18:33.4660889Z * [new branch] gh/bobrenjc93/624/head -> origin/gh/bobrenjc93/624/head 2025-10-10T01:18:33.4662652Z * [new branch] gh/bobrenjc93/624/orig -> origin/gh/bobrenjc93/624/orig 2025-10-10T01:18:33.4665200Z * [new branch] gh/bobrenjc93/625/base -> origin/gh/bobrenjc93/625/base 2025-10-10T01:18:33.4667680Z * [new branch] gh/bobrenjc93/625/head -> origin/gh/bobrenjc93/625/head 2025-10-10T01:18:33.4669126Z * [new branch] gh/bobrenjc93/625/orig -> origin/gh/bobrenjc93/625/orig 2025-10-10T01:18:33.4671776Z * [new branch] gh/bobrenjc93/626/base -> origin/gh/bobrenjc93/626/base 2025-10-10T01:18:33.4673258Z * [new branch] gh/bobrenjc93/626/head -> origin/gh/bobrenjc93/626/head 2025-10-10T01:18:33.4674860Z * [new branch] gh/bobrenjc93/626/orig -> origin/gh/bobrenjc93/626/orig 2025-10-10T01:18:33.4677569Z * [new branch] gh/bobrenjc93/627/base -> origin/gh/bobrenjc93/627/base 2025-10-10T01:18:33.4679431Z * [new branch] gh/bobrenjc93/627/head -> origin/gh/bobrenjc93/627/head 2025-10-10T01:18:33.4681054Z * [new branch] gh/bobrenjc93/627/orig -> origin/gh/bobrenjc93/627/orig 2025-10-10T01:18:33.4683433Z * [new branch] gh/bobrenjc93/628/base -> origin/gh/bobrenjc93/628/base 2025-10-10T01:18:33.4685194Z * [new branch] gh/bobrenjc93/628/head -> origin/gh/bobrenjc93/628/head 2025-10-10T01:18:33.4686842Z * [new branch] gh/bobrenjc93/628/orig -> origin/gh/bobrenjc93/628/orig 2025-10-10T01:18:33.4689265Z * [new branch] gh/bobrenjc93/629/base -> origin/gh/bobrenjc93/629/base 2025-10-10T01:18:33.4690954Z * [new branch] gh/bobrenjc93/629/head -> origin/gh/bobrenjc93/629/head 2025-10-10T01:18:33.4692620Z * [new branch] gh/bobrenjc93/629/orig -> origin/gh/bobrenjc93/629/orig 2025-10-10T01:18:33.4695000Z * [new branch] gh/bobrenjc93/630/base -> origin/gh/bobrenjc93/630/base 2025-10-10T01:18:33.4697205Z * [new branch] gh/bobrenjc93/630/head -> origin/gh/bobrenjc93/630/head 2025-10-10T01:18:33.4698930Z * [new branch] gh/bobrenjc93/630/orig -> origin/gh/bobrenjc93/630/orig 2025-10-10T01:18:33.4701269Z * [new branch] gh/bobrenjc93/631/base -> origin/gh/bobrenjc93/631/base 2025-10-10T01:18:33.4703002Z * [new branch] gh/bobrenjc93/631/head -> origin/gh/bobrenjc93/631/head 2025-10-10T01:18:33.4704718Z * [new branch] gh/bobrenjc93/631/orig -> origin/gh/bobrenjc93/631/orig 2025-10-10T01:18:33.4707116Z * [new branch] gh/bobrenjc93/632/base -> origin/gh/bobrenjc93/632/base 2025-10-10T01:18:33.4708868Z * [new branch] gh/bobrenjc93/632/head -> origin/gh/bobrenjc93/632/head 2025-10-10T01:18:33.4710532Z * [new branch] gh/bobrenjc93/632/orig -> origin/gh/bobrenjc93/632/orig 2025-10-10T01:18:33.4712988Z * [new branch] gh/bobrenjc93/633/base -> origin/gh/bobrenjc93/633/base 2025-10-10T01:18:33.4714725Z * [new branch] gh/bobrenjc93/633/head -> origin/gh/bobrenjc93/633/head 2025-10-10T01:18:33.4716444Z * [new branch] gh/bobrenjc93/633/orig -> origin/gh/bobrenjc93/633/orig 2025-10-10T01:18:33.4718779Z * [new branch] gh/bobrenjc93/634/base -> origin/gh/bobrenjc93/634/base 2025-10-10T01:18:33.4720587Z * [new branch] gh/bobrenjc93/634/head -> origin/gh/bobrenjc93/634/head 2025-10-10T01:18:33.4722496Z * [new branch] gh/bobrenjc93/634/orig -> origin/gh/bobrenjc93/634/orig 2025-10-10T01:18:33.4724833Z * [new branch] gh/bobrenjc93/635/base -> origin/gh/bobrenjc93/635/base 2025-10-10T01:18:33.4727104Z * [new branch] gh/bobrenjc93/635/head -> origin/gh/bobrenjc93/635/head 2025-10-10T01:18:33.4728850Z * [new branch] gh/bobrenjc93/635/orig -> origin/gh/bobrenjc93/635/orig 2025-10-10T01:18:33.4731173Z * [new branch] gh/bobrenjc93/636/base -> origin/gh/bobrenjc93/636/base 2025-10-10T01:18:33.4732872Z * [new branch] gh/bobrenjc93/636/head -> origin/gh/bobrenjc93/636/head 2025-10-10T01:18:33.4734554Z * [new branch] gh/bobrenjc93/636/orig -> origin/gh/bobrenjc93/636/orig 2025-10-10T01:18:33.4737114Z * [new branch] gh/bobrenjc93/637/base -> origin/gh/bobrenjc93/637/base 2025-10-10T01:18:33.4738757Z * [new branch] gh/bobrenjc93/637/head -> origin/gh/bobrenjc93/637/head 2025-10-10T01:18:33.4740524Z * [new branch] gh/bobrenjc93/637/orig -> origin/gh/bobrenjc93/637/orig 2025-10-10T01:18:33.4742885Z * [new branch] gh/bobrenjc93/638/base -> origin/gh/bobrenjc93/638/base 2025-10-10T01:18:33.4744588Z * [new branch] gh/bobrenjc93/638/head -> origin/gh/bobrenjc93/638/head 2025-10-10T01:18:33.4746371Z * [new branch] gh/bobrenjc93/638/orig -> origin/gh/bobrenjc93/638/orig 2025-10-10T01:18:33.4748721Z * [new branch] gh/bobrenjc93/639/base -> origin/gh/bobrenjc93/639/base 2025-10-10T01:18:33.4750486Z * [new branch] gh/bobrenjc93/639/head -> origin/gh/bobrenjc93/639/head 2025-10-10T01:18:33.4752170Z * [new branch] gh/bobrenjc93/639/orig -> origin/gh/bobrenjc93/639/orig 2025-10-10T01:18:33.4754489Z * [new branch] gh/bobrenjc93/640/base -> origin/gh/bobrenjc93/640/base 2025-10-10T01:18:33.4756205Z * [new branch] gh/bobrenjc93/640/head -> origin/gh/bobrenjc93/640/head 2025-10-10T01:18:33.4757900Z * [new branch] gh/bobrenjc93/640/orig -> origin/gh/bobrenjc93/640/orig 2025-10-10T01:18:33.4760513Z * [new branch] gh/bobrenjc93/641/base -> origin/gh/bobrenjc93/641/base 2025-10-10T01:18:33.4762258Z * [new branch] gh/bobrenjc93/641/head -> origin/gh/bobrenjc93/641/head 2025-10-10T01:18:33.4763941Z * [new branch] gh/bobrenjc93/641/orig -> origin/gh/bobrenjc93/641/orig 2025-10-10T01:18:33.4766669Z * [new branch] gh/bobrenjc93/642/base -> origin/gh/bobrenjc93/642/base 2025-10-10T01:18:33.4768669Z * [new branch] gh/bobrenjc93/642/head -> origin/gh/bobrenjc93/642/head 2025-10-10T01:18:33.4770237Z * [new branch] gh/bobrenjc93/642/orig -> origin/gh/bobrenjc93/642/orig 2025-10-10T01:18:33.4773082Z * [new branch] gh/bobrenjc93/643/base -> origin/gh/bobrenjc93/643/base 2025-10-10T01:18:33.4774403Z * [new branch] gh/bobrenjc93/643/head -> origin/gh/bobrenjc93/643/head 2025-10-10T01:18:33.4776215Z * [new branch] gh/bobrenjc93/643/orig -> origin/gh/bobrenjc93/643/orig 2025-10-10T01:18:33.4778728Z * [new branch] gh/bobrenjc93/644/base -> origin/gh/bobrenjc93/644/base 2025-10-10T01:18:33.4780414Z * [new branch] gh/bobrenjc93/644/head -> origin/gh/bobrenjc93/644/head 2025-10-10T01:18:33.4782150Z * [new branch] gh/bobrenjc93/644/orig -> origin/gh/bobrenjc93/644/orig 2025-10-10T01:18:33.4785124Z * [new branch] gh/bobrenjc93/645/base -> origin/gh/bobrenjc93/645/base 2025-10-10T01:18:33.4786819Z * [new branch] gh/bobrenjc93/645/head -> origin/gh/bobrenjc93/645/head 2025-10-10T01:18:33.4788513Z * [new branch] gh/bobrenjc93/645/orig -> origin/gh/bobrenjc93/645/orig 2025-10-10T01:18:33.4790950Z * [new branch] gh/bobrenjc93/646/base -> origin/gh/bobrenjc93/646/base 2025-10-10T01:18:33.4792861Z * [new branch] gh/bobrenjc93/646/head -> origin/gh/bobrenjc93/646/head 2025-10-10T01:18:33.4794424Z * [new branch] gh/bobrenjc93/646/orig -> origin/gh/bobrenjc93/646/orig 2025-10-10T01:18:33.4796987Z * [new branch] gh/bobrenjc93/647/base -> origin/gh/bobrenjc93/647/base 2025-10-10T01:18:33.4799006Z * [new branch] gh/bobrenjc93/647/head -> origin/gh/bobrenjc93/647/head 2025-10-10T01:18:33.4800578Z * [new branch] gh/bobrenjc93/647/orig -> origin/gh/bobrenjc93/647/orig 2025-10-10T01:18:33.4802838Z * [new branch] gh/bobrenjc93/648/base -> origin/gh/bobrenjc93/648/base 2025-10-10T01:18:33.4804632Z * [new branch] gh/bobrenjc93/648/head -> origin/gh/bobrenjc93/648/head 2025-10-10T01:18:33.4806456Z * [new branch] gh/bobrenjc93/648/orig -> origin/gh/bobrenjc93/648/orig 2025-10-10T01:18:33.4808634Z * [new branch] gh/bobrenjc93/649/base -> origin/gh/bobrenjc93/649/base 2025-10-10T01:18:33.4810371Z * [new branch] gh/bobrenjc93/649/head -> origin/gh/bobrenjc93/649/head 2025-10-10T01:18:33.4812049Z * [new branch] gh/bobrenjc93/649/orig -> origin/gh/bobrenjc93/649/orig 2025-10-10T01:18:33.4814471Z * [new branch] gh/bobrenjc93/650/base -> origin/gh/bobrenjc93/650/base 2025-10-10T01:18:33.4816140Z * [new branch] gh/bobrenjc93/650/head -> origin/gh/bobrenjc93/650/head 2025-10-10T01:18:33.4817811Z * [new branch] gh/bobrenjc93/650/orig -> origin/gh/bobrenjc93/650/orig 2025-10-10T01:18:33.4820726Z * [new branch] gh/briancoutinho/2/base -> origin/gh/briancoutinho/2/base 2025-10-10T01:18:33.4822476Z * [new branch] gh/briancoutinho/2/head -> origin/gh/briancoutinho/2/head 2025-10-10T01:18:33.4825734Z * [new branch] gh/c00w/23/base -> origin/gh/c00w/23/base 2025-10-10T01:18:33.4827425Z * [new branch] gh/c00w/23/head -> origin/gh/c00w/23/head 2025-10-10T01:18:33.4829874Z * [new branch] gh/c00w/53/base -> origin/gh/c00w/53/base 2025-10-10T01:18:33.4831546Z * [new branch] gh/c00w/53/head -> origin/gh/c00w/53/head 2025-10-10T01:18:33.4833203Z * [new branch] gh/c00w/53/orig -> origin/gh/c00w/53/orig 2025-10-10T01:18:33.4835414Z * [new branch] gh/c00w/54/base -> origin/gh/c00w/54/base 2025-10-10T01:18:33.4837168Z * [new branch] gh/c00w/54/head -> origin/gh/c00w/54/head 2025-10-10T01:18:33.4838986Z * [new branch] gh/c00w/54/orig -> origin/gh/c00w/54/orig 2025-10-10T01:18:33.4842164Z * [new branch] gh/c00w/57/base -> origin/gh/c00w/57/base 2025-10-10T01:18:33.4843501Z * [new branch] gh/c00w/57/head -> origin/gh/c00w/57/head 2025-10-10T01:18:33.4845341Z * [new branch] gh/c00w/57/orig -> origin/gh/c00w/57/orig 2025-10-10T01:18:33.4848229Z * [new branch] gh/clee2000/1/base -> origin/gh/clee2000/1/base 2025-10-10T01:18:33.4849993Z * [new branch] gh/clee2000/1/head -> origin/gh/clee2000/1/head 2025-10-10T01:18:33.4851702Z * [new branch] gh/clee2000/1/orig -> origin/gh/clee2000/1/orig 2025-10-10T01:18:33.4854657Z * [new branch] gh/coconutruben/1/base -> origin/gh/coconutruben/1/base 2025-10-10T01:18:33.4857337Z * [new branch] gh/coconutruben/1/head -> origin/gh/coconutruben/1/head 2025-10-10T01:18:33.4859427Z * [new branch] gh/coconutruben/20/base -> origin/gh/coconutruben/20/base 2025-10-10T01:18:33.4861153Z * [new branch] gh/coconutruben/20/head -> origin/gh/coconutruben/20/head 2025-10-10T01:18:33.4862903Z * [new branch] gh/coconutruben/20/orig -> origin/gh/coconutruben/20/orig 2025-10-10T01:18:33.4866324Z * [new branch] gh/coconutruben/22/base -> origin/gh/coconutruben/22/base 2025-10-10T01:18:33.4867068Z * [new branch] gh/coconutruben/22/head -> origin/gh/coconutruben/22/head 2025-10-10T01:18:33.4869132Z * [new branch] gh/coconutruben/22/orig -> origin/gh/coconutruben/22/orig 2025-10-10T01:18:33.4871546Z * [new branch] gh/coconutruben/24/base -> origin/gh/coconutruben/24/base 2025-10-10T01:18:33.4873360Z * [new branch] gh/coconutruben/24/head -> origin/gh/coconutruben/24/head 2025-10-10T01:18:33.4875325Z * [new branch] gh/coconutruben/24/orig -> origin/gh/coconutruben/24/orig 2025-10-10T01:18:33.4878145Z * [new branch] gh/coconutruben/25/base -> origin/gh/coconutruben/25/base 2025-10-10T01:18:33.4880317Z * [new branch] gh/coconutruben/25/head -> origin/gh/coconutruben/25/head 2025-10-10T01:18:33.4882235Z * [new branch] gh/coconutruben/25/orig -> origin/gh/coconutruben/25/orig 2025-10-10T01:18:33.4886180Z * [new branch] gh/coconutruben/36/base -> origin/gh/coconutruben/36/base 2025-10-10T01:18:33.4888375Z * [new branch] gh/coconutruben/36/head -> origin/gh/coconutruben/36/head 2025-10-10T01:18:33.4890645Z * [new branch] gh/coconutruben/36/orig -> origin/gh/coconutruben/36/orig 2025-10-10T01:18:33.4893089Z * [new branch] gh/coconutruben/48/base -> origin/gh/coconutruben/48/base 2025-10-10T01:18:33.4894893Z * [new branch] gh/coconutruben/48/head -> origin/gh/coconutruben/48/head 2025-10-10T01:18:33.4896574Z * [new branch] gh/coconutruben/48/orig -> origin/gh/coconutruben/48/orig 2025-10-10T01:18:33.4899451Z * [new branch] gh/coconutruben/49/base -> origin/gh/coconutruben/49/base 2025-10-10T01:18:33.4901217Z * [new branch] gh/coconutruben/49/head -> origin/gh/coconutruben/49/head 2025-10-10T01:18:33.4902970Z * [new branch] gh/coconutruben/49/orig -> origin/gh/coconutruben/49/orig 2025-10-10T01:18:33.4905515Z * [new branch] gh/coconutruben/50/base -> origin/gh/coconutruben/50/base 2025-10-10T01:18:33.4907304Z * [new branch] gh/coconutruben/50/head -> origin/gh/coconutruben/50/head 2025-10-10T01:18:33.4909073Z * [new branch] gh/coconutruben/50/orig -> origin/gh/coconutruben/50/orig 2025-10-10T01:18:33.4911485Z * [new branch] gh/coconutruben/51/base -> origin/gh/coconutruben/51/base 2025-10-10T01:18:33.4913271Z * [new branch] gh/coconutruben/51/head -> origin/gh/coconutruben/51/head 2025-10-10T01:18:33.4915091Z * [new branch] gh/coconutruben/51/orig -> origin/gh/coconutruben/51/orig 2025-10-10T01:18:33.4917563Z * [new branch] gh/coconutruben/52/base -> origin/gh/coconutruben/52/base 2025-10-10T01:18:33.4919545Z * [new branch] gh/coconutruben/52/head -> origin/gh/coconutruben/52/head 2025-10-10T01:18:33.4921343Z * [new branch] gh/coconutruben/52/orig -> origin/gh/coconutruben/52/orig 2025-10-10T01:18:33.4924325Z * [new branch] gh/coconutruben/53/base -> origin/gh/coconutruben/53/base 2025-10-10T01:18:33.4926600Z * [new branch] gh/coconutruben/53/head -> origin/gh/coconutruben/53/head 2025-10-10T01:18:33.4928639Z * [new branch] gh/coconutruben/53/orig -> origin/gh/coconutruben/53/orig 2025-10-10T01:18:33.4931341Z * [new branch] gh/coconutruben/54/base -> origin/gh/coconutruben/54/base 2025-10-10T01:18:33.4932769Z * [new branch] gh/coconutruben/54/head -> origin/gh/coconutruben/54/head 2025-10-10T01:18:33.4934554Z * [new branch] gh/coconutruben/54/orig -> origin/gh/coconutruben/54/orig 2025-10-10T01:18:33.4937413Z * [new branch] gh/coconutruben/55/base -> origin/gh/coconutruben/55/base 2025-10-10T01:18:33.4939242Z * [new branch] gh/coconutruben/55/head -> origin/gh/coconutruben/55/head 2025-10-10T01:18:33.4940890Z * [new branch] gh/coconutruben/55/orig -> origin/gh/coconutruben/55/orig 2025-10-10T01:18:33.4943296Z * [new branch] gh/coconutruben/56/base -> origin/gh/coconutruben/56/base 2025-10-10T01:18:33.4945060Z * [new branch] gh/coconutruben/56/head -> origin/gh/coconutruben/56/head 2025-10-10T01:18:33.4946846Z * [new branch] gh/coconutruben/56/orig -> origin/gh/coconutruben/56/orig 2025-10-10T01:18:33.4949538Z * [new branch] gh/coconutruben/57/base -> origin/gh/coconutruben/57/base 2025-10-10T01:18:33.4951387Z * [new branch] gh/coconutruben/57/head -> origin/gh/coconutruben/57/head 2025-10-10T01:18:33.4953101Z * [new branch] gh/coconutruben/57/orig -> origin/gh/coconutruben/57/orig 2025-10-10T01:18:33.4955834Z * [new branch] gh/coconutruben/58/base -> origin/gh/coconutruben/58/base 2025-10-10T01:18:33.4957703Z * [new branch] gh/coconutruben/58/head -> origin/gh/coconutruben/58/head 2025-10-10T01:18:33.4959577Z * [new branch] gh/coconutruben/58/orig -> origin/gh/coconutruben/58/orig 2025-10-10T01:18:33.4961966Z * [new branch] gh/coconutruben/59/base -> origin/gh/coconutruben/59/base 2025-10-10T01:18:33.4963603Z * [new branch] gh/coconutruben/59/head -> origin/gh/coconutruben/59/head 2025-10-10T01:18:33.4965758Z * [new branch] gh/coconutruben/59/orig -> origin/gh/coconutruben/59/orig 2025-10-10T01:18:33.4968761Z * [new branch] gh/coconutruben/62/base -> origin/gh/coconutruben/62/base 2025-10-10T01:18:33.4970602Z * [new branch] gh/coconutruben/62/head -> origin/gh/coconutruben/62/head 2025-10-10T01:18:33.4972474Z * [new branch] gh/coconutruben/62/orig -> origin/gh/coconutruben/62/orig 2025-10-10T01:18:33.4974989Z * [new branch] gh/coconutruben/64/base -> origin/gh/coconutruben/64/base 2025-10-10T01:18:33.4976754Z * [new branch] gh/coconutruben/64/head -> origin/gh/coconutruben/64/head 2025-10-10T01:18:33.4978644Z * [new branch] gh/coconutruben/64/orig -> origin/gh/coconutruben/64/orig 2025-10-10T01:18:33.4981187Z * [new branch] gh/coconutruben/65/base -> origin/gh/coconutruben/65/base 2025-10-10T01:18:33.4982914Z * [new branch] gh/coconutruben/65/head -> origin/gh/coconutruben/65/head 2025-10-10T01:18:33.4984653Z * [new branch] gh/coconutruben/65/orig -> origin/gh/coconutruben/65/orig 2025-10-10T01:18:33.4987106Z * [new branch] gh/coconutruben/66/base -> origin/gh/coconutruben/66/base 2025-10-10T01:18:33.4988857Z * [new branch] gh/coconutruben/66/head -> origin/gh/coconutruben/66/head 2025-10-10T01:18:33.4990564Z * [new branch] gh/coconutruben/66/orig -> origin/gh/coconutruben/66/orig 2025-10-10T01:18:33.4993095Z * [new branch] gh/coconutruben/67/base -> origin/gh/coconutruben/67/base 2025-10-10T01:18:33.4994874Z * [new branch] gh/coconutruben/67/head -> origin/gh/coconutruben/67/head 2025-10-10T01:18:33.4996626Z * [new branch] gh/coconutruben/67/orig -> origin/gh/coconutruben/67/orig 2025-10-10T01:18:33.4999234Z * [new branch] gh/coconutruben/68/base -> origin/gh/coconutruben/68/base 2025-10-10T01:18:33.5000989Z * [new branch] gh/coconutruben/68/head -> origin/gh/coconutruben/68/head 2025-10-10T01:18:33.5002669Z * [new branch] gh/coconutruben/68/orig -> origin/gh/coconutruben/68/orig 2025-10-10T01:18:33.5005109Z * [new branch] gh/coconutruben/69/base -> origin/gh/coconutruben/69/base 2025-10-10T01:18:33.5006906Z * [new branch] gh/coconutruben/69/head -> origin/gh/coconutruben/69/head 2025-10-10T01:18:33.5008742Z * [new branch] gh/coconutruben/69/orig -> origin/gh/coconutruben/69/orig 2025-10-10T01:18:33.5011057Z * [new branch] gh/coconutruben/70/base -> origin/gh/coconutruben/70/base 2025-10-10T01:18:33.5012798Z * [new branch] gh/coconutruben/70/head -> origin/gh/coconutruben/70/head 2025-10-10T01:18:33.5015100Z * [new branch] gh/coconutruben/70/orig -> origin/gh/coconutruben/70/orig 2025-10-10T01:18:33.5017340Z * [new branch] gh/coconutruben/71/base -> origin/gh/coconutruben/71/base 2025-10-10T01:18:33.5019101Z * [new branch] gh/coconutruben/71/head -> origin/gh/coconutruben/71/head 2025-10-10T01:18:33.5020958Z * [new branch] gh/coconutruben/71/orig -> origin/gh/coconutruben/71/orig 2025-10-10T01:18:33.5023287Z * [new branch] gh/coconutruben/72/base -> origin/gh/coconutruben/72/base 2025-10-10T01:18:33.5025268Z * [new branch] gh/coconutruben/72/head -> origin/gh/coconutruben/72/head 2025-10-10T01:18:33.5027048Z * [new branch] gh/coconutruben/72/orig -> origin/gh/coconutruben/72/orig 2025-10-10T01:18:33.5029363Z * [new branch] gh/coconutruben/73/base -> origin/gh/coconutruben/73/base 2025-10-10T01:18:33.5031096Z * [new branch] gh/coconutruben/73/head -> origin/gh/coconutruben/73/head 2025-10-10T01:18:33.5032809Z * [new branch] gh/coconutruben/73/orig -> origin/gh/coconutruben/73/orig 2025-10-10T01:18:33.5035329Z * [new branch] gh/coconutruben/74/base -> origin/gh/coconutruben/74/base 2025-10-10T01:18:33.5037100Z * [new branch] gh/coconutruben/74/head -> origin/gh/coconutruben/74/head 2025-10-10T01:18:33.5038926Z * [new branch] gh/coconutruben/74/orig -> origin/gh/coconutruben/74/orig 2025-10-10T01:18:33.5041393Z * [new branch] gh/coconutruben/75/base -> origin/gh/coconutruben/75/base 2025-10-10T01:18:33.5043234Z * [new branch] gh/coconutruben/75/head -> origin/gh/coconutruben/75/head 2025-10-10T01:18:33.5044994Z * [new branch] gh/coconutruben/75/orig -> origin/gh/coconutruben/75/orig 2025-10-10T01:18:33.5047551Z * [new branch] gh/coconutruben/76/base -> origin/gh/coconutruben/76/base 2025-10-10T01:18:33.5049395Z * [new branch] gh/coconutruben/76/head -> origin/gh/coconutruben/76/head 2025-10-10T01:18:33.5051115Z * [new branch] gh/coconutruben/76/orig -> origin/gh/coconutruben/76/orig 2025-10-10T01:18:33.5053710Z * [new branch] gh/coconutruben/77/base -> origin/gh/coconutruben/77/base 2025-10-10T01:18:33.5055589Z * [new branch] gh/coconutruben/77/head -> origin/gh/coconutruben/77/head 2025-10-10T01:18:33.5057569Z * [new branch] gh/coconutruben/77/orig -> origin/gh/coconutruben/77/orig 2025-10-10T01:18:33.5059769Z * [new branch] gh/coconutruben/78/base -> origin/gh/coconutruben/78/base 2025-10-10T01:18:33.5061532Z * [new branch] gh/coconutruben/78/head -> origin/gh/coconutruben/78/head 2025-10-10T01:18:33.5063281Z * [new branch] gh/coconutruben/78/orig -> origin/gh/coconutruben/78/orig 2025-10-10T01:18:33.5066195Z * [new branch] gh/coconutruben/79/base -> origin/gh/coconutruben/79/base 2025-10-10T01:18:33.5067481Z * [new branch] gh/coconutruben/79/head -> origin/gh/coconutruben/79/head 2025-10-10T01:18:33.5069289Z * [new branch] gh/coconutruben/79/orig -> origin/gh/coconutruben/79/orig 2025-10-10T01:18:33.5071647Z * [new branch] gh/coconutruben/80/base -> origin/gh/coconutruben/80/base 2025-10-10T01:18:33.5073397Z * [new branch] gh/coconutruben/80/head -> origin/gh/coconutruben/80/head 2025-10-10T01:18:33.5075167Z * [new branch] gh/coconutruben/80/orig -> origin/gh/coconutruben/80/orig 2025-10-10T01:18:33.5077826Z * [new branch] gh/coconutruben/81/base -> origin/gh/coconutruben/81/base 2025-10-10T01:18:33.5079624Z * [new branch] gh/coconutruben/81/head -> origin/gh/coconutruben/81/head 2025-10-10T01:18:33.5081281Z * [new branch] gh/coconutruben/81/orig -> origin/gh/coconutruben/81/orig 2025-10-10T01:18:33.5083642Z * [new branch] gh/coconutruben/82/base -> origin/gh/coconutruben/82/base 2025-10-10T01:18:33.5085355Z * [new branch] gh/coconutruben/82/head -> origin/gh/coconutruben/82/head 2025-10-10T01:18:33.5087219Z * [new branch] gh/coconutruben/82/orig -> origin/gh/coconutruben/82/orig 2025-10-10T01:18:33.5089515Z * [new branch] gh/coconutruben/83/base -> origin/gh/coconutruben/83/base 2025-10-10T01:18:33.5091299Z * [new branch] gh/coconutruben/83/head -> origin/gh/coconutruben/83/head 2025-10-10T01:18:33.5093011Z * [new branch] gh/coconutruben/83/orig -> origin/gh/coconutruben/83/orig 2025-10-10T01:18:33.5095863Z * [new branch] gh/colinchan15/1/base -> origin/gh/colinchan15/1/base 2025-10-10T01:18:33.5097673Z * [new branch] gh/colinchan15/1/head -> origin/gh/colinchan15/1/head 2025-10-10T01:18:33.5099871Z * [new branch] gh/colinchan15/2/base -> origin/gh/colinchan15/2/base 2025-10-10T01:18:33.5101565Z * [new branch] gh/colinchan15/2/head -> origin/gh/colinchan15/2/head 2025-10-10T01:18:33.5103702Z * [new branch] gh/colinchan15/3/base -> origin/gh/colinchan15/3/base 2025-10-10T01:18:33.5105591Z * [new branch] gh/colinchan15/3/head -> origin/gh/colinchan15/3/head 2025-10-10T01:18:33.5108584Z * [new branch] gh/colinchan15/6/base -> origin/gh/colinchan15/6/base 2025-10-10T01:18:33.5109754Z * [new branch] gh/colinchan15/6/head -> origin/gh/colinchan15/6/head 2025-10-10T01:18:33.5113088Z * [new branch] gh/davidberard98/382/base -> origin/gh/davidberard98/382/base 2025-10-10T01:18:33.5114487Z * [new branch] gh/davidberard98/382/head -> origin/gh/davidberard98/382/head 2025-10-10T01:18:33.5116355Z * [new branch] gh/davidberard98/382/orig -> origin/gh/davidberard98/382/orig 2025-10-10T01:18:33.5118921Z * [new branch] gh/davidberard98/386/base -> origin/gh/davidberard98/386/base 2025-10-10T01:18:33.5120825Z * [new branch] gh/davidberard98/386/head -> origin/gh/davidberard98/386/head 2025-10-10T01:18:33.5122558Z * [new branch] gh/davidberard98/386/orig -> origin/gh/davidberard98/386/orig 2025-10-10T01:18:33.5125283Z * [new branch] gh/davidberard98/391/base -> origin/gh/davidberard98/391/base 2025-10-10T01:18:33.5128950Z * [new branch] gh/davidberard98/391/head -> origin/gh/davidberard98/391/head 2025-10-10T01:18:33.5130051Z * [new branch] gh/davidberard98/391/orig -> origin/gh/davidberard98/391/orig 2025-10-10T01:18:33.5132925Z * [new branch] gh/davidberard98/392/base -> origin/gh/davidberard98/392/base 2025-10-10T01:18:33.5134326Z * [new branch] gh/davidberard98/392/head -> origin/gh/davidberard98/392/head 2025-10-10T01:18:33.5135832Z * [new branch] gh/davidberard98/392/orig -> origin/gh/davidberard98/392/orig 2025-10-10T01:18:33.5138981Z * [new branch] gh/davidberard98/399/base -> origin/gh/davidberard98/399/base 2025-10-10T01:18:33.5140991Z * [new branch] gh/davidberard98/399/head -> origin/gh/davidberard98/399/head 2025-10-10T01:18:33.5142423Z * [new branch] gh/davidberard98/399/orig -> origin/gh/davidberard98/399/orig 2025-10-10T01:18:33.5144818Z * [new branch] gh/davidberard98/401/base -> origin/gh/davidberard98/401/base 2025-10-10T01:18:33.5146529Z * [new branch] gh/davidberard98/401/head -> origin/gh/davidberard98/401/head 2025-10-10T01:18:33.5148854Z * [new branch] gh/davidberard98/401/orig -> origin/gh/davidberard98/401/orig 2025-10-10T01:18:33.5150767Z * [new branch] gh/davidberard98/405/base -> origin/gh/davidberard98/405/base 2025-10-10T01:18:33.5153923Z * [new branch] gh/davidberard98/405/head -> origin/gh/davidberard98/405/head 2025-10-10T01:18:33.5154376Z * [new branch] gh/davidberard98/405/orig -> origin/gh/davidberard98/405/orig 2025-10-10T01:18:33.5156683Z * [new branch] gh/davidberard98/410/base -> origin/gh/davidberard98/410/base 2025-10-10T01:18:33.5158293Z * [new branch] gh/davidberard98/410/head -> origin/gh/davidberard98/410/head 2025-10-10T01:18:33.5160394Z * [new branch] gh/davidberard98/410/orig -> origin/gh/davidberard98/410/orig 2025-10-10T01:18:33.5162760Z * [new branch] gh/davidberard98/411/base -> origin/gh/davidberard98/411/base 2025-10-10T01:18:33.5164483Z * [new branch] gh/davidberard98/411/head -> origin/gh/davidberard98/411/head 2025-10-10T01:18:33.5166301Z * [new branch] gh/davidberard98/411/orig -> origin/gh/davidberard98/411/orig 2025-10-10T01:18:33.5169109Z * [new branch] gh/davidberard98/412/base -> origin/gh/davidberard98/412/base 2025-10-10T01:18:33.5171129Z * [new branch] gh/davidberard98/412/head -> origin/gh/davidberard98/412/head 2025-10-10T01:18:33.5172791Z * [new branch] gh/davidberard98/412/orig -> origin/gh/davidberard98/412/orig 2025-10-10T01:18:33.5175625Z * [new branch] gh/desertfire/594/base -> origin/gh/desertfire/594/base 2025-10-10T01:18:33.5177330Z * [new branch] gh/desertfire/594/head -> origin/gh/desertfire/594/head 2025-10-10T01:18:33.5179038Z * [new branch] gh/desertfire/594/orig -> origin/gh/desertfire/594/orig 2025-10-10T01:18:33.5181348Z * [new branch] gh/desertfire/595/base -> origin/gh/desertfire/595/base 2025-10-10T01:18:33.5183061Z * [new branch] gh/desertfire/595/head -> origin/gh/desertfire/595/head 2025-10-10T01:18:33.5184769Z * [new branch] gh/desertfire/595/orig -> origin/gh/desertfire/595/orig 2025-10-10T01:18:33.5187114Z * [new branch] gh/desertfire/597/base -> origin/gh/desertfire/597/base 2025-10-10T01:18:33.5188856Z * [new branch] gh/desertfire/597/head -> origin/gh/desertfire/597/head 2025-10-10T01:18:33.5190538Z * [new branch] gh/desertfire/597/orig -> origin/gh/desertfire/597/orig 2025-10-10T01:18:33.5192872Z * [new branch] gh/desertfire/598/base -> origin/gh/desertfire/598/base 2025-10-10T01:18:33.5194614Z * [new branch] gh/desertfire/598/head -> origin/gh/desertfire/598/head 2025-10-10T01:18:33.5196343Z * [new branch] gh/desertfire/598/orig -> origin/gh/desertfire/598/orig 2025-10-10T01:18:33.5198535Z * [new branch] gh/desertfire/599/base -> origin/gh/desertfire/599/base 2025-10-10T01:18:33.5200318Z * [new branch] gh/desertfire/599/head -> origin/gh/desertfire/599/head 2025-10-10T01:18:33.5202041Z * [new branch] gh/desertfire/599/orig -> origin/gh/desertfire/599/orig 2025-10-10T01:18:33.5204422Z * [new branch] gh/desertfire/600/base -> origin/gh/desertfire/600/base 2025-10-10T01:18:33.5206123Z * [new branch] gh/desertfire/600/head -> origin/gh/desertfire/600/head 2025-10-10T01:18:33.5207870Z * [new branch] gh/desertfire/600/orig -> origin/gh/desertfire/600/orig 2025-10-10T01:18:33.5210222Z * [new branch] gh/desertfire/601/base -> origin/gh/desertfire/601/base 2025-10-10T01:18:33.5211954Z * [new branch] gh/desertfire/601/head -> origin/gh/desertfire/601/head 2025-10-10T01:18:33.5213614Z * [new branch] gh/desertfire/601/orig -> origin/gh/desertfire/601/orig 2025-10-10T01:18:33.5216421Z * [new branch] gh/dharakk/1/base -> origin/gh/dharakk/1/base 2025-10-10T01:18:33.5218305Z * [new branch] gh/dharakk/1/head -> origin/gh/dharakk/1/head 2025-10-10T01:18:33.5221041Z * [new branch] gh/drisspg/159/base -> origin/gh/drisspg/159/base 2025-10-10T01:18:33.5222724Z * [new branch] gh/drisspg/159/head -> origin/gh/drisspg/159/head 2025-10-10T01:18:33.5225102Z * [new branch] gh/drisspg/159/orig -> origin/gh/drisspg/159/orig 2025-10-10T01:18:33.5227537Z * [new branch] gh/drisspg/166/base -> origin/gh/drisspg/166/base 2025-10-10T01:18:33.5229443Z * [new branch] gh/drisspg/166/head -> origin/gh/drisspg/166/head 2025-10-10T01:18:33.5231114Z * [new branch] gh/drisspg/166/orig -> origin/gh/drisspg/166/orig 2025-10-10T01:18:33.5233666Z * [new branch] gh/drisspg/170/base -> origin/gh/drisspg/170/base 2025-10-10T01:18:33.5235177Z * [new branch] gh/drisspg/170/head -> origin/gh/drisspg/170/head 2025-10-10T01:18:33.5239663Z * [new branch] gh/drisspg/170/orig -> origin/gh/drisspg/170/orig 2025-10-10T01:18:33.5240618Z * [new branch] gh/drisspg/177/base -> origin/gh/drisspg/177/base 2025-10-10T01:18:33.5241458Z * [new branch] gh/drisspg/177/head -> origin/gh/drisspg/177/head 2025-10-10T01:18:33.5242585Z * [new branch] gh/drisspg/177/orig -> origin/gh/drisspg/177/orig 2025-10-10T01:18:33.5245263Z * [new branch] gh/drisspg/178/base -> origin/gh/drisspg/178/base 2025-10-10T01:18:33.5246617Z * [new branch] gh/drisspg/178/head -> origin/gh/drisspg/178/head 2025-10-10T01:18:33.5248229Z * [new branch] gh/drisspg/178/orig -> origin/gh/drisspg/178/orig 2025-10-10T01:18:33.5250568Z * [new branch] gh/drisspg/182/base -> origin/gh/drisspg/182/base 2025-10-10T01:18:33.5252284Z * [new branch] gh/drisspg/182/head -> origin/gh/drisspg/182/head 2025-10-10T01:18:33.5254460Z * [new branch] gh/drisspg/183/base -> origin/gh/drisspg/183/base 2025-10-10T01:18:33.5256179Z * [new branch] gh/drisspg/183/head -> origin/gh/drisspg/183/head 2025-10-10T01:18:33.5258324Z * [new branch] gh/drisspg/184/base -> origin/gh/drisspg/184/base 2025-10-10T01:18:33.5260075Z * [new branch] gh/drisspg/184/head -> origin/gh/drisspg/184/head 2025-10-10T01:18:33.5262373Z * [new branch] gh/drisspg/185/base -> origin/gh/drisspg/185/base 2025-10-10T01:18:33.5264050Z * [new branch] gh/drisspg/185/head -> origin/gh/drisspg/185/head 2025-10-10T01:18:33.5266284Z * [new branch] gh/drisspg/187/base -> origin/gh/drisspg/187/base 2025-10-10T01:18:33.5268112Z * [new branch] gh/drisspg/187/head -> origin/gh/drisspg/187/head 2025-10-10T01:18:33.5269747Z * [new branch] gh/drisspg/187/orig -> origin/gh/drisspg/187/orig 2025-10-10T01:18:33.5272039Z * [new branch] gh/drisspg/188/base -> origin/gh/drisspg/188/base 2025-10-10T01:18:33.5273784Z * [new branch] gh/drisspg/188/head -> origin/gh/drisspg/188/head 2025-10-10T01:18:33.5275466Z * [new branch] gh/drisspg/188/orig -> origin/gh/drisspg/188/orig 2025-10-10T01:18:33.5277740Z * [new branch] gh/drisspg/189/base -> origin/gh/drisspg/189/base 2025-10-10T01:18:33.5280129Z * [new branch] gh/drisspg/189/head -> origin/gh/drisspg/189/head 2025-10-10T01:18:33.5281217Z * [new branch] gh/drisspg/189/orig -> origin/gh/drisspg/189/orig 2025-10-10T01:18:33.5283736Z * [new branch] gh/drisspg/193/base -> origin/gh/drisspg/193/base 2025-10-10T01:18:33.5285588Z * [new branch] gh/drisspg/193/head -> origin/gh/drisspg/193/head 2025-10-10T01:18:33.5287245Z * [new branch] gh/drisspg/193/orig -> origin/gh/drisspg/193/orig 2025-10-10T01:18:33.5289404Z * [new branch] gh/drisspg/194/base -> origin/gh/drisspg/194/base 2025-10-10T01:18:33.5291063Z * [new branch] gh/drisspg/194/head -> origin/gh/drisspg/194/head 2025-10-10T01:18:33.5292768Z * [new branch] gh/drisspg/194/orig -> origin/gh/drisspg/194/orig 2025-10-10T01:18:33.5295414Z * [new branch] gh/drisspg/196/base -> origin/gh/drisspg/196/base 2025-10-10T01:18:33.5297125Z * [new branch] gh/drisspg/196/head -> origin/gh/drisspg/196/head 2025-10-10T01:18:33.5298838Z * [new branch] gh/drisspg/196/orig -> origin/gh/drisspg/196/orig 2025-10-10T01:18:33.5301065Z * [new branch] gh/drisspg/197/base -> origin/gh/drisspg/197/base 2025-10-10T01:18:33.5302746Z * [new branch] gh/drisspg/197/head -> origin/gh/drisspg/197/head 2025-10-10T01:18:33.5304371Z * [new branch] gh/drisspg/197/orig -> origin/gh/drisspg/197/orig 2025-10-10T01:18:33.5307390Z * [new branch] gh/drisspg/198/base -> origin/gh/drisspg/198/base 2025-10-10T01:18:33.5309047Z * [new branch] gh/drisspg/198/head -> origin/gh/drisspg/198/head 2025-10-10T01:18:33.5310718Z * [new branch] gh/drisspg/198/orig -> origin/gh/drisspg/198/orig 2025-10-10T01:18:33.5313084Z * [new branch] gh/drisspg/199/base -> origin/gh/drisspg/199/base 2025-10-10T01:18:33.5314764Z * [new branch] gh/drisspg/199/head -> origin/gh/drisspg/199/head 2025-10-10T01:18:33.5316453Z * [new branch] gh/drisspg/199/orig -> origin/gh/drisspg/199/orig 2025-10-10T01:18:33.5318912Z * [new branch] gh/drisspg/200/base -> origin/gh/drisspg/200/base 2025-10-10T01:18:33.5320657Z * [new branch] gh/drisspg/200/head -> origin/gh/drisspg/200/head 2025-10-10T01:18:33.5322328Z * [new branch] gh/drisspg/200/orig -> origin/gh/drisspg/200/orig 2025-10-10T01:18:33.5324680Z * [new branch] gh/drisspg/201/base -> origin/gh/drisspg/201/base 2025-10-10T01:18:33.5326531Z * [new branch] gh/drisspg/201/head -> origin/gh/drisspg/201/head 2025-10-10T01:18:33.5328277Z * [new branch] gh/drisspg/201/orig -> origin/gh/drisspg/201/orig 2025-10-10T01:18:33.5330637Z * [new branch] gh/drisspg/202/base -> origin/gh/drisspg/202/base 2025-10-10T01:18:33.5332326Z * [new branch] gh/drisspg/202/head -> origin/gh/drisspg/202/head 2025-10-10T01:18:33.5334065Z * [new branch] gh/drisspg/202/orig -> origin/gh/drisspg/202/orig 2025-10-10T01:18:33.5336442Z * [new branch] gh/drisspg/203/base -> origin/gh/drisspg/203/base 2025-10-10T01:18:33.5338138Z * [new branch] gh/drisspg/203/head -> origin/gh/drisspg/203/head 2025-10-10T01:18:33.5339808Z * [new branch] gh/drisspg/203/orig -> origin/gh/drisspg/203/orig 2025-10-10T01:18:33.5342186Z * [new branch] gh/drisspg/204/base -> origin/gh/drisspg/204/base 2025-10-10T01:18:33.5343873Z * [new branch] gh/drisspg/204/head -> origin/gh/drisspg/204/head 2025-10-10T01:18:33.5345649Z * [new branch] gh/drisspg/204/orig -> origin/gh/drisspg/204/orig 2025-10-10T01:18:33.5348061Z * [new branch] gh/drisspg/205/base -> origin/gh/drisspg/205/base 2025-10-10T01:18:33.5349674Z * [new branch] gh/drisspg/205/head -> origin/gh/drisspg/205/head 2025-10-10T01:18:33.5351414Z * [new branch] gh/drisspg/205/orig -> origin/gh/drisspg/205/orig 2025-10-10T01:18:33.5353853Z * [new branch] gh/drisspg/206/base -> origin/gh/drisspg/206/base 2025-10-10T01:18:33.5355812Z * [new branch] gh/drisspg/206/head -> origin/gh/drisspg/206/head 2025-10-10T01:18:33.5357369Z * [new branch] gh/drisspg/206/orig -> origin/gh/drisspg/206/orig 2025-10-10T01:18:33.5359780Z * [new branch] gh/drisspg/207/base -> origin/gh/drisspg/207/base 2025-10-10T01:18:33.5361613Z * [new branch] gh/drisspg/207/head -> origin/gh/drisspg/207/head 2025-10-10T01:18:33.5363266Z * [new branch] gh/drisspg/207/orig -> origin/gh/drisspg/207/orig 2025-10-10T01:18:33.5365654Z * [new branch] gh/drisspg/208/base -> origin/gh/drisspg/208/base 2025-10-10T01:18:33.5367318Z * [new branch] gh/drisspg/208/head -> origin/gh/drisspg/208/head 2025-10-10T01:18:33.5369017Z * [new branch] gh/drisspg/208/orig -> origin/gh/drisspg/208/orig 2025-10-10T01:18:33.5371387Z * [new branch] gh/drisspg/209/base -> origin/gh/drisspg/209/base 2025-10-10T01:18:33.5373110Z * [new branch] gh/drisspg/209/head -> origin/gh/drisspg/209/head 2025-10-10T01:18:33.5374765Z * [new branch] gh/drisspg/209/orig -> origin/gh/drisspg/209/orig 2025-10-10T01:18:33.5377747Z * [new branch] gh/dsjohns2/1/base -> origin/gh/dsjohns2/1/base 2025-10-10T01:18:33.5379647Z * [new branch] gh/dsjohns2/1/head -> origin/gh/dsjohns2/1/head 2025-10-10T01:18:33.5382386Z * [new branch] gh/eellison/808/base -> origin/gh/eellison/808/base 2025-10-10T01:18:33.5384104Z * [new branch] gh/eellison/808/head -> origin/gh/eellison/808/head 2025-10-10T01:18:33.5386590Z * [new branch] gh/eellison/808/orig -> origin/gh/eellison/808/orig 2025-10-10T01:18:33.5388532Z * [new branch] gh/eellison/809/base -> origin/gh/eellison/809/base 2025-10-10T01:18:33.5390150Z * [new branch] gh/eellison/809/head -> origin/gh/eellison/809/head 2025-10-10T01:18:33.5391914Z * [new branch] gh/eellison/809/orig -> origin/gh/eellison/809/orig 2025-10-10T01:18:33.5394274Z * [new branch] gh/eellison/822/base -> origin/gh/eellison/822/base 2025-10-10T01:18:33.5395959Z * [new branch] gh/eellison/822/head -> origin/gh/eellison/822/head 2025-10-10T01:18:33.5397718Z * [new branch] gh/eellison/822/orig -> origin/gh/eellison/822/orig 2025-10-10T01:18:33.5400125Z * [new branch] gh/eellison/823/base -> origin/gh/eellison/823/base 2025-10-10T01:18:33.5401765Z * [new branch] gh/eellison/823/head -> origin/gh/eellison/823/head 2025-10-10T01:18:33.5403450Z * [new branch] gh/eellison/823/orig -> origin/gh/eellison/823/orig 2025-10-10T01:18:33.5405770Z * [new branch] gh/eellison/824/base -> origin/gh/eellison/824/base 2025-10-10T01:18:33.5407448Z * [new branch] gh/eellison/824/head -> origin/gh/eellison/824/head 2025-10-10T01:18:33.5409211Z * [new branch] gh/eellison/824/orig -> origin/gh/eellison/824/orig 2025-10-10T01:18:33.5411623Z * [new branch] gh/eellison/825/base -> origin/gh/eellison/825/base 2025-10-10T01:18:33.5413251Z * [new branch] gh/eellison/825/head -> origin/gh/eellison/825/head 2025-10-10T01:18:33.5414950Z * [new branch] gh/eellison/825/orig -> origin/gh/eellison/825/orig 2025-10-10T01:18:33.5417421Z * [new branch] gh/eellison/826/base -> origin/gh/eellison/826/base 2025-10-10T01:18:33.5419054Z * [new branch] gh/eellison/826/head -> origin/gh/eellison/826/head 2025-10-10T01:18:33.5420834Z * [new branch] gh/eellison/826/orig -> origin/gh/eellison/826/orig 2025-10-10T01:18:33.5423140Z * [new branch] gh/eellison/827/base -> origin/gh/eellison/827/base 2025-10-10T01:18:33.5425135Z * [new branch] gh/eellison/827/head -> origin/gh/eellison/827/head 2025-10-10T01:18:33.5426737Z * [new branch] gh/eellison/827/orig -> origin/gh/eellison/827/orig 2025-10-10T01:18:33.5429143Z * [new branch] gh/eellison/828/base -> origin/gh/eellison/828/base 2025-10-10T01:18:33.5430875Z * [new branch] gh/eellison/828/head -> origin/gh/eellison/828/head 2025-10-10T01:18:33.5432522Z * [new branch] gh/eellison/828/orig -> origin/gh/eellison/828/orig 2025-10-10T01:18:33.5434990Z * [new branch] gh/eellison/829/base -> origin/gh/eellison/829/base 2025-10-10T01:18:33.5436713Z * [new branch] gh/eellison/829/head -> origin/gh/eellison/829/head 2025-10-10T01:18:33.5438539Z * [new branch] gh/eellison/829/orig -> origin/gh/eellison/829/orig 2025-10-10T01:18:33.5440904Z * [new branch] gh/eellison/830/base -> origin/gh/eellison/830/base 2025-10-10T01:18:33.5442637Z * [new branch] gh/eellison/830/head -> origin/gh/eellison/830/head 2025-10-10T01:18:33.5444366Z * [new branch] gh/eellison/830/orig -> origin/gh/eellison/830/orig 2025-10-10T01:18:33.5446643Z * [new branch] gh/eellison/831/base -> origin/gh/eellison/831/base 2025-10-10T01:18:33.5448297Z * [new branch] gh/eellison/831/head -> origin/gh/eellison/831/head 2025-10-10T01:18:33.5449979Z * [new branch] gh/eellison/831/orig -> origin/gh/eellison/831/orig 2025-10-10T01:18:33.5452315Z * [new branch] gh/eellison/832/base -> origin/gh/eellison/832/base 2025-10-10T01:18:33.5454006Z * [new branch] gh/eellison/832/head -> origin/gh/eellison/832/head 2025-10-10T01:18:33.5456254Z * [new branch] gh/eellison/832/orig -> origin/gh/eellison/832/orig 2025-10-10T01:18:33.5458115Z * [new branch] gh/eellison/833/base -> origin/gh/eellison/833/base 2025-10-10T01:18:33.5459790Z * [new branch] gh/eellison/833/head -> origin/gh/eellison/833/head 2025-10-10T01:18:33.5461454Z * [new branch] gh/eellison/833/orig -> origin/gh/eellison/833/orig 2025-10-10T01:18:33.5463752Z * [new branch] gh/eellison/834/base -> origin/gh/eellison/834/base 2025-10-10T01:18:33.5465999Z * [new branch] gh/eellison/834/head -> origin/gh/eellison/834/head 2025-10-10T01:18:33.5467818Z * [new branch] gh/eellison/834/orig -> origin/gh/eellison/834/orig 2025-10-10T01:18:33.5470076Z * [new branch] gh/eellison/835/base -> origin/gh/eellison/835/base 2025-10-10T01:18:33.5471749Z * [new branch] gh/eellison/835/head -> origin/gh/eellison/835/head 2025-10-10T01:18:33.5473408Z * [new branch] gh/eellison/835/orig -> origin/gh/eellison/835/orig 2025-10-10T01:18:33.5476414Z * [new branch] gh/eellison/836/base -> origin/gh/eellison/836/base 2025-10-10T01:18:33.5478206Z * [new branch] gh/eellison/836/head -> origin/gh/eellison/836/head 2025-10-10T01:18:33.5480532Z * [new branch] gh/eellison/836/orig -> origin/gh/eellison/836/orig 2025-10-10T01:18:33.5483089Z * [new branch] gh/eellison/837/base -> origin/gh/eellison/837/base 2025-10-10T01:18:33.5484780Z * [new branch] gh/eellison/837/head -> origin/gh/eellison/837/head 2025-10-10T01:18:33.5486463Z * [new branch] gh/eellison/837/orig -> origin/gh/eellison/837/orig 2025-10-10T01:18:33.5488868Z * [new branch] gh/eellison/838/base -> origin/gh/eellison/838/base 2025-10-10T01:18:33.5490591Z * [new branch] gh/eellison/838/head -> origin/gh/eellison/838/head 2025-10-10T01:18:33.5492363Z * [new branch] gh/eellison/838/orig -> origin/gh/eellison/838/orig 2025-10-10T01:18:33.5494856Z * [new branch] gh/eellison/839/base -> origin/gh/eellison/839/base 2025-10-10T01:18:33.5497719Z * [new branch] gh/eellison/839/head -> origin/gh/eellison/839/head 2025-10-10T01:18:33.5498960Z * [new branch] gh/eellison/839/orig -> origin/gh/eellison/839/orig 2025-10-10T01:18:33.5501255Z * [new branch] gh/eellison/840/base -> origin/gh/eellison/840/base 2025-10-10T01:18:33.5502931Z * [new branch] gh/eellison/840/head -> origin/gh/eellison/840/head 2025-10-10T01:18:33.5504623Z * [new branch] gh/eellison/840/orig -> origin/gh/eellison/840/orig 2025-10-10T01:18:33.5507144Z * [new branch] gh/eellison/841/base -> origin/gh/eellison/841/base 2025-10-10T01:18:33.5509088Z * [new branch] gh/eellison/841/head -> origin/gh/eellison/841/head 2025-10-10T01:18:33.5511195Z * [new branch] gh/eellison/841/orig -> origin/gh/eellison/841/orig 2025-10-10T01:18:33.5513494Z * [new branch] gh/eellison/842/base -> origin/gh/eellison/842/base 2025-10-10T01:18:33.5515656Z * [new branch] gh/eellison/842/head -> origin/gh/eellison/842/head 2025-10-10T01:18:33.5517429Z * [new branch] gh/eellison/842/orig -> origin/gh/eellison/842/orig 2025-10-10T01:18:33.5519983Z * [new branch] gh/eellison/843/base -> origin/gh/eellison/843/base 2025-10-10T01:18:33.5521614Z * [new branch] gh/eellison/843/head -> origin/gh/eellison/843/head 2025-10-10T01:18:33.5523246Z * [new branch] gh/eellison/843/orig -> origin/gh/eellison/843/orig 2025-10-10T01:18:33.5527077Z * [new branch] gh/eellison/844/base -> origin/gh/eellison/844/base 2025-10-10T01:18:33.5528814Z * [new branch] gh/eellison/844/head -> origin/gh/eellison/844/head 2025-10-10T01:18:33.5530528Z * [new branch] gh/eellison/844/orig -> origin/gh/eellison/844/orig 2025-10-10T01:18:33.5532902Z * [new branch] gh/eellison/845/base -> origin/gh/eellison/845/base 2025-10-10T01:18:33.5534590Z * [new branch] gh/eellison/845/head -> origin/gh/eellison/845/head 2025-10-10T01:18:33.5536265Z * [new branch] gh/eellison/845/orig -> origin/gh/eellison/845/orig 2025-10-10T01:18:33.5538575Z * [new branch] gh/eellison/846/base -> origin/gh/eellison/846/base 2025-10-10T01:18:33.5540192Z * [new branch] gh/eellison/846/head -> origin/gh/eellison/846/head 2025-10-10T01:18:33.5541867Z * [new branch] gh/eellison/846/orig -> origin/gh/eellison/846/orig 2025-10-10T01:18:33.5544692Z * [new branch] gh/etaf/147/base -> origin/gh/etaf/147/base 2025-10-10T01:18:33.5546495Z * [new branch] gh/etaf/147/head -> origin/gh/etaf/147/head 2025-10-10T01:18:33.5548949Z * [new branch] gh/etaf/154/base -> origin/gh/etaf/154/base 2025-10-10T01:18:33.5550680Z * [new branch] gh/etaf/154/head -> origin/gh/etaf/154/head 2025-10-10T01:18:33.5552349Z * [new branch] gh/etaf/154/orig -> origin/gh/etaf/154/orig 2025-10-10T01:18:33.5554825Z * [new branch] gh/etaf/156/base -> origin/gh/etaf/156/base 2025-10-10T01:18:33.5556661Z * [new branch] gh/etaf/156/head -> origin/gh/etaf/156/head 2025-10-10T01:18:33.5558347Z * [new branch] gh/etaf/156/orig -> origin/gh/etaf/156/orig 2025-10-10T01:18:33.5560947Z * [new branch] gh/etaf/157/base -> origin/gh/etaf/157/base 2025-10-10T01:18:33.5562622Z * [new branch] gh/etaf/157/head -> origin/gh/etaf/157/head 2025-10-10T01:18:33.5564287Z * [new branch] gh/etaf/157/orig -> origin/gh/etaf/157/orig 2025-10-10T01:18:33.5566811Z * [new branch] gh/etaf/158/base -> origin/gh/etaf/158/base 2025-10-10T01:18:33.5568620Z * [new branch] gh/etaf/158/head -> origin/gh/etaf/158/head 2025-10-10T01:18:33.5569994Z * [new branch] gh/etaf/158/orig -> origin/gh/etaf/158/orig 2025-10-10T01:18:33.5572585Z * [new branch] gh/etaf/159/base -> origin/gh/etaf/159/base 2025-10-10T01:18:33.5574399Z * [new branch] gh/etaf/159/head -> origin/gh/etaf/159/head 2025-10-10T01:18:33.5576048Z * [new branch] gh/etaf/159/orig -> origin/gh/etaf/159/orig 2025-10-10T01:18:33.5578438Z * [new branch] gh/etaf/160/base -> origin/gh/etaf/160/base 2025-10-10T01:18:33.5580117Z * [new branch] gh/etaf/160/head -> origin/gh/etaf/160/head 2025-10-10T01:18:33.5581794Z * [new branch] gh/etaf/160/orig -> origin/gh/etaf/160/orig 2025-10-10T01:18:33.5584147Z * [new branch] gh/etaf/161/base -> origin/gh/etaf/161/base 2025-10-10T01:18:33.5585850Z * [new branch] gh/etaf/161/head -> origin/gh/etaf/161/head 2025-10-10T01:18:33.5587454Z * [new branch] gh/etaf/161/orig -> origin/gh/etaf/161/orig 2025-10-10T01:18:33.5590077Z * [new branch] gh/etaf/162/base -> origin/gh/etaf/162/base 2025-10-10T01:18:33.5592273Z * [new branch] gh/etaf/162/head -> origin/gh/etaf/162/head 2025-10-10T01:18:33.5593952Z * [new branch] gh/etaf/162/orig -> origin/gh/etaf/162/orig 2025-10-10T01:18:33.5597145Z * [new branch] gh/etaf/166/base -> origin/gh/etaf/166/base 2025-10-10T01:18:33.5598580Z * [new branch] gh/etaf/166/head -> origin/gh/etaf/166/head 2025-10-10T01:18:33.5600304Z * [new branch] gh/etaf/166/orig -> origin/gh/etaf/166/orig 2025-10-10T01:18:33.5603131Z * [new branch] gh/etaf/167/base -> origin/gh/etaf/167/base 2025-10-10T01:18:33.5604749Z * [new branch] gh/etaf/167/head -> origin/gh/etaf/167/head 2025-10-10T01:18:33.5606362Z * [new branch] gh/etaf/167/orig -> origin/gh/etaf/167/orig 2025-10-10T01:18:33.5608769Z * [new branch] gh/etaf/168/base -> origin/gh/etaf/168/base 2025-10-10T01:18:33.5610563Z * [new branch] gh/etaf/168/head -> origin/gh/etaf/168/head 2025-10-10T01:18:33.5612232Z * [new branch] gh/etaf/168/orig -> origin/gh/etaf/168/orig 2025-10-10T01:18:33.5614634Z * [new branch] gh/etaf/170/base -> origin/gh/etaf/170/base 2025-10-10T01:18:33.5617156Z * [new branch] gh/etaf/170/head -> origin/gh/etaf/170/head 2025-10-10T01:18:33.5618416Z * [new branch] gh/etaf/170/orig -> origin/gh/etaf/170/orig 2025-10-10T01:18:33.5620719Z * [new branch] gh/etaf/171/base -> origin/gh/etaf/171/base 2025-10-10T01:18:33.5622410Z * [new branch] gh/etaf/171/head -> origin/gh/etaf/171/head 2025-10-10T01:18:33.5624287Z * [new branch] gh/etaf/171/orig -> origin/gh/etaf/171/orig 2025-10-10T01:18:33.5626758Z * [new branch] gh/etaf/172/base -> origin/gh/etaf/172/base 2025-10-10T01:18:33.5628469Z * [new branch] gh/etaf/172/head -> origin/gh/etaf/172/head 2025-10-10T01:18:33.5630314Z * [new branch] gh/etaf/172/orig -> origin/gh/etaf/172/orig 2025-10-10T01:18:33.5632980Z * [new branch] gh/exclamaforte/1/base -> origin/gh/exclamaforte/1/base 2025-10-10T01:18:33.5634598Z * [new branch] gh/exclamaforte/1/head -> origin/gh/exclamaforte/1/head 2025-10-10T01:18:33.5636941Z * [new branch] gh/exclamaforte/2/base -> origin/gh/exclamaforte/2/base 2025-10-10T01:18:33.5638596Z * [new branch] gh/exclamaforte/2/head -> origin/gh/exclamaforte/2/head 2025-10-10T01:18:33.5641060Z * [new branch] gh/exclamaforte/3/base -> origin/gh/exclamaforte/3/base 2025-10-10T01:18:33.5642625Z * [new branch] gh/exclamaforte/3/head -> origin/gh/exclamaforte/3/head 2025-10-10T01:18:33.5644838Z * [new branch] gh/exclamaforte/4/base -> origin/gh/exclamaforte/4/base 2025-10-10T01:18:33.5646495Z * [new branch] gh/exclamaforte/4/head -> origin/gh/exclamaforte/4/head 2025-10-10T01:18:33.5649337Z * [new branch] gh/ezyang/2374/base -> origin/gh/ezyang/2374/base 2025-10-10T01:18:33.5651093Z * [new branch] gh/ezyang/2374/head -> origin/gh/ezyang/2374/head 2025-10-10T01:18:33.5652762Z * [new branch] gh/ezyang/2374/orig -> origin/gh/ezyang/2374/orig 2025-10-10T01:18:33.5655368Z * [new branch] gh/ezyang/2973/base -> origin/gh/ezyang/2973/base 2025-10-10T01:18:33.5656663Z * [new branch] gh/ezyang/2973/head -> origin/gh/ezyang/2973/head 2025-10-10T01:18:33.5658596Z * [new branch] gh/ezyang/2973/orig -> origin/gh/ezyang/2973/orig 2025-10-10T01:18:33.5660773Z * [new branch] gh/ezyang/2974/base -> origin/gh/ezyang/2974/base 2025-10-10T01:18:33.5662480Z * [new branch] gh/ezyang/2974/head -> origin/gh/ezyang/2974/head 2025-10-10T01:18:33.5664221Z * [new branch] gh/ezyang/2974/orig -> origin/gh/ezyang/2974/orig 2025-10-10T01:18:33.5666518Z * [new branch] gh/ezyang/3120/base -> origin/gh/ezyang/3120/base 2025-10-10T01:18:33.5668120Z * [new branch] gh/ezyang/3120/head -> origin/gh/ezyang/3120/head 2025-10-10T01:18:33.5669835Z * [new branch] gh/ezyang/3120/orig -> origin/gh/ezyang/3120/orig 2025-10-10T01:18:33.5672120Z * [new branch] gh/ezyang/3122/base -> origin/gh/ezyang/3122/base 2025-10-10T01:18:33.5673820Z * [new branch] gh/ezyang/3122/head -> origin/gh/ezyang/3122/head 2025-10-10T01:18:33.5675481Z * [new branch] gh/ezyang/3122/orig -> origin/gh/ezyang/3122/orig 2025-10-10T01:18:33.5677781Z * [new branch] gh/ezyang/3127/base -> origin/gh/ezyang/3127/base 2025-10-10T01:18:33.5679678Z * [new branch] gh/ezyang/3127/head -> origin/gh/ezyang/3127/head 2025-10-10T01:18:33.5681313Z * [new branch] gh/ezyang/3127/orig -> origin/gh/ezyang/3127/orig 2025-10-10T01:18:33.5683609Z * [new branch] gh/ezyang/3131/base -> origin/gh/ezyang/3131/base 2025-10-10T01:18:33.5685316Z * [new branch] gh/ezyang/3131/head -> origin/gh/ezyang/3131/head 2025-10-10T01:18:33.5686969Z * [new branch] gh/ezyang/3131/orig -> origin/gh/ezyang/3131/orig 2025-10-10T01:18:33.5689235Z * [new branch] gh/ezyang/3134/base -> origin/gh/ezyang/3134/base 2025-10-10T01:18:33.5690861Z * [new branch] gh/ezyang/3134/head -> origin/gh/ezyang/3134/head 2025-10-10T01:18:33.5692581Z * [new branch] gh/ezyang/3134/orig -> origin/gh/ezyang/3134/orig 2025-10-10T01:18:33.5694857Z * [new branch] gh/ezyang/3135/base -> origin/gh/ezyang/3135/base 2025-10-10T01:18:33.5696990Z * [new branch] gh/ezyang/3135/head -> origin/gh/ezyang/3135/head 2025-10-10T01:18:33.5698669Z * [new branch] gh/ezyang/3135/orig -> origin/gh/ezyang/3135/orig 2025-10-10T01:18:33.5700946Z * [new branch] gh/ezyang/3138/base -> origin/gh/ezyang/3138/base 2025-10-10T01:18:33.5702652Z * [new branch] gh/ezyang/3138/head -> origin/gh/ezyang/3138/head 2025-10-10T01:18:33.5704363Z * [new branch] gh/ezyang/3138/orig -> origin/gh/ezyang/3138/orig 2025-10-10T01:18:33.5706614Z * [new branch] gh/ezyang/3139/base -> origin/gh/ezyang/3139/base 2025-10-10T01:18:33.5708384Z * [new branch] gh/ezyang/3139/head -> origin/gh/ezyang/3139/head 2025-10-10T01:18:33.5709967Z * [new branch] gh/ezyang/3139/orig -> origin/gh/ezyang/3139/orig 2025-10-10T01:18:33.5712202Z * [new branch] gh/ezyang/3140/base -> origin/gh/ezyang/3140/base 2025-10-10T01:18:33.5713818Z * [new branch] gh/ezyang/3140/head -> origin/gh/ezyang/3140/head 2025-10-10T01:18:33.5715527Z * [new branch] gh/ezyang/3140/orig -> origin/gh/ezyang/3140/orig 2025-10-10T01:18:33.5717761Z * [new branch] gh/ezyang/3143/base -> origin/gh/ezyang/3143/base 2025-10-10T01:18:33.5719569Z * [new branch] gh/ezyang/3143/head -> origin/gh/ezyang/3143/head 2025-10-10T01:18:33.5721269Z * [new branch] gh/ezyang/3143/orig -> origin/gh/ezyang/3143/orig 2025-10-10T01:18:33.5723614Z * [new branch] gh/ezyang/3144/base -> origin/gh/ezyang/3144/base 2025-10-10T01:18:33.5725706Z * [new branch] gh/ezyang/3144/head -> origin/gh/ezyang/3144/head 2025-10-10T01:18:33.5727389Z * [new branch] gh/ezyang/3144/orig -> origin/gh/ezyang/3144/orig 2025-10-10T01:18:33.5729673Z * [new branch] gh/ezyang/3145/base -> origin/gh/ezyang/3145/base 2025-10-10T01:18:33.5731390Z * [new branch] gh/ezyang/3145/head -> origin/gh/ezyang/3145/head 2025-10-10T01:18:33.5733053Z * [new branch] gh/ezyang/3145/orig -> origin/gh/ezyang/3145/orig 2025-10-10T01:18:33.5735798Z * [new branch] gh/ezyang/3146/base -> origin/gh/ezyang/3146/base 2025-10-10T01:18:33.5737462Z * [new branch] gh/ezyang/3146/head -> origin/gh/ezyang/3146/head 2025-10-10T01:18:33.5739106Z * [new branch] gh/ezyang/3146/orig -> origin/gh/ezyang/3146/orig 2025-10-10T01:18:33.5741439Z * [new branch] gh/ezyang/3147/base -> origin/gh/ezyang/3147/base 2025-10-10T01:18:33.5743129Z * [new branch] gh/ezyang/3147/head -> origin/gh/ezyang/3147/head 2025-10-10T01:18:33.5744782Z * [new branch] gh/ezyang/3147/orig -> origin/gh/ezyang/3147/orig 2025-10-10T01:18:33.5747100Z * [new branch] gh/ezyang/3148/base -> origin/gh/ezyang/3148/base 2025-10-10T01:18:33.5748872Z * [new branch] gh/ezyang/3148/head -> origin/gh/ezyang/3148/head 2025-10-10T01:18:33.5750508Z * [new branch] gh/ezyang/3148/orig -> origin/gh/ezyang/3148/orig 2025-10-10T01:18:33.5752820Z * [new branch] gh/ezyang/3149/base -> origin/gh/ezyang/3149/base 2025-10-10T01:18:33.5754468Z * [new branch] gh/ezyang/3149/head -> origin/gh/ezyang/3149/head 2025-10-10T01:18:33.5756190Z * [new branch] gh/ezyang/3149/orig -> origin/gh/ezyang/3149/orig 2025-10-10T01:18:33.5758524Z * [new branch] gh/ezyang/3150/base -> origin/gh/ezyang/3150/base 2025-10-10T01:18:33.5760257Z * [new branch] gh/ezyang/3150/head -> origin/gh/ezyang/3150/head 2025-10-10T01:18:33.5761905Z * [new branch] gh/ezyang/3150/orig -> origin/gh/ezyang/3150/orig 2025-10-10T01:18:33.5764306Z * [new branch] gh/ezyang/3151/base -> origin/gh/ezyang/3151/base 2025-10-10T01:18:33.5766035Z * [new branch] gh/ezyang/3151/head -> origin/gh/ezyang/3151/head 2025-10-10T01:18:33.5767758Z * [new branch] gh/ezyang/3151/orig -> origin/gh/ezyang/3151/orig 2025-10-10T01:18:33.5770063Z * [new branch] gh/ezyang/3152/base -> origin/gh/ezyang/3152/base 2025-10-10T01:18:33.5771803Z * [new branch] gh/ezyang/3152/head -> origin/gh/ezyang/3152/head 2025-10-10T01:18:33.5773465Z * [new branch] gh/ezyang/3152/orig -> origin/gh/ezyang/3152/orig 2025-10-10T01:18:33.5775894Z * [new branch] gh/ezyang/3153/base -> origin/gh/ezyang/3153/base 2025-10-10T01:18:33.5777402Z * [new branch] gh/ezyang/3153/head -> origin/gh/ezyang/3153/head 2025-10-10T01:18:33.5779103Z * [new branch] gh/ezyang/3153/orig -> origin/gh/ezyang/3153/orig 2025-10-10T01:18:33.5781353Z * [new branch] gh/ezyang/3154/base -> origin/gh/ezyang/3154/base 2025-10-10T01:18:33.5782964Z * [new branch] gh/ezyang/3154/head -> origin/gh/ezyang/3154/head 2025-10-10T01:18:33.5784661Z * [new branch] gh/ezyang/3154/orig -> origin/gh/ezyang/3154/orig 2025-10-10T01:18:33.5787031Z * [new branch] gh/ezyang/3155/base -> origin/gh/ezyang/3155/base 2025-10-10T01:18:33.5788878Z * [new branch] gh/ezyang/3155/head -> origin/gh/ezyang/3155/head 2025-10-10T01:18:33.5790624Z * [new branch] gh/ezyang/3155/orig -> origin/gh/ezyang/3155/orig 2025-10-10T01:18:33.5792982Z * [new branch] gh/ezyang/3156/base -> origin/gh/ezyang/3156/base 2025-10-10T01:18:33.5794739Z * [new branch] gh/ezyang/3156/head -> origin/gh/ezyang/3156/head 2025-10-10T01:18:33.5796494Z * [new branch] gh/ezyang/3156/orig -> origin/gh/ezyang/3156/orig 2025-10-10T01:18:33.5798932Z * [new branch] gh/ezyang/3157/base -> origin/gh/ezyang/3157/base 2025-10-10T01:18:33.5800619Z * [new branch] gh/ezyang/3157/head -> origin/gh/ezyang/3157/head 2025-10-10T01:18:33.5802295Z * [new branch] gh/ezyang/3157/orig -> origin/gh/ezyang/3157/orig 2025-10-10T01:18:33.5804582Z * [new branch] gh/ezyang/3158/base -> origin/gh/ezyang/3158/base 2025-10-10T01:18:33.5806270Z * [new branch] gh/ezyang/3158/head -> origin/gh/ezyang/3158/head 2025-10-10T01:18:33.5807914Z * [new branch] gh/ezyang/3158/orig -> origin/gh/ezyang/3158/orig 2025-10-10T01:18:33.5810346Z * [new branch] gh/ezyang/3159/base -> origin/gh/ezyang/3159/base 2025-10-10T01:18:33.5811976Z * [new branch] gh/ezyang/3159/head -> origin/gh/ezyang/3159/head 2025-10-10T01:18:33.5813656Z * [new branch] gh/ezyang/3159/orig -> origin/gh/ezyang/3159/orig 2025-10-10T01:18:33.5816014Z * [new branch] gh/ezyang/3160/base -> origin/gh/ezyang/3160/base 2025-10-10T01:18:33.5817764Z * [new branch] gh/ezyang/3160/head -> origin/gh/ezyang/3160/head 2025-10-10T01:18:33.5819436Z * [new branch] gh/ezyang/3160/orig -> origin/gh/ezyang/3160/orig 2025-10-10T01:18:33.5821744Z * [new branch] gh/ezyang/3161/base -> origin/gh/ezyang/3161/base 2025-10-10T01:18:33.5823699Z * [new branch] gh/ezyang/3161/head -> origin/gh/ezyang/3161/head 2025-10-10T01:18:33.5825428Z * [new branch] gh/ezyang/3161/orig -> origin/gh/ezyang/3161/orig 2025-10-10T01:18:33.5827644Z * [new branch] gh/ezyang/3162/base -> origin/gh/ezyang/3162/base 2025-10-10T01:18:33.5829340Z * [new branch] gh/ezyang/3162/head -> origin/gh/ezyang/3162/head 2025-10-10T01:18:33.5831102Z * [new branch] gh/ezyang/3162/orig -> origin/gh/ezyang/3162/orig 2025-10-10T01:18:33.5833394Z * [new branch] gh/ezyang/3163/base -> origin/gh/ezyang/3163/base 2025-10-10T01:18:33.5835054Z * [new branch] gh/ezyang/3163/head -> origin/gh/ezyang/3163/head 2025-10-10T01:18:33.5836791Z * [new branch] gh/ezyang/3163/orig -> origin/gh/ezyang/3163/orig 2025-10-10T01:18:33.5839250Z * [new branch] gh/ezyang/3164/base -> origin/gh/ezyang/3164/base 2025-10-10T01:18:33.5841016Z * [new branch] gh/ezyang/3164/head -> origin/gh/ezyang/3164/head 2025-10-10T01:18:33.5842671Z * [new branch] gh/ezyang/3164/orig -> origin/gh/ezyang/3164/orig 2025-10-10T01:18:33.5845100Z * [new branch] gh/ezyang/3165/base -> origin/gh/ezyang/3165/base 2025-10-10T01:18:33.5846658Z * [new branch] gh/ezyang/3165/head -> origin/gh/ezyang/3165/head 2025-10-10T01:18:33.5848300Z * [new branch] gh/ezyang/3165/orig -> origin/gh/ezyang/3165/orig 2025-10-10T01:18:33.5850663Z * [new branch] gh/ezyang/3166/base -> origin/gh/ezyang/3166/base 2025-10-10T01:18:33.5852325Z * [new branch] gh/ezyang/3166/head -> origin/gh/ezyang/3166/head 2025-10-10T01:18:33.5854206Z * [new branch] gh/ezyang/3166/orig -> origin/gh/ezyang/3166/orig 2025-10-10T01:18:33.5856568Z * [new branch] gh/ezyang/3167/base -> origin/gh/ezyang/3167/base 2025-10-10T01:18:33.5858667Z * [new branch] gh/ezyang/3167/head -> origin/gh/ezyang/3167/head 2025-10-10T01:18:33.5860367Z * [new branch] gh/ezyang/3167/orig -> origin/gh/ezyang/3167/orig 2025-10-10T01:18:33.5863147Z * [new branch] gh/ezyang/3168/base -> origin/gh/ezyang/3168/base 2025-10-10T01:18:33.5864905Z * [new branch] gh/ezyang/3168/head -> origin/gh/ezyang/3168/head 2025-10-10T01:18:33.5866620Z * [new branch] gh/ezyang/3168/orig -> origin/gh/ezyang/3168/orig 2025-10-10T01:18:33.5868910Z * [new branch] gh/ezyang/3169/base -> origin/gh/ezyang/3169/base 2025-10-10T01:18:33.5881290Z * [new branch] gh/ezyang/3169/head -> origin/gh/ezyang/3169/head 2025-10-10T01:18:33.5881817Z * [new branch] gh/ezyang/3169/orig -> origin/gh/ezyang/3169/orig 2025-10-10T01:18:33.5882211Z * [new branch] gh/ezyang/3170/base -> origin/gh/ezyang/3170/base 2025-10-10T01:18:33.5882581Z * [new branch] gh/ezyang/3170/head -> origin/gh/ezyang/3170/head 2025-10-10T01:18:33.5882969Z * [new branch] gh/ezyang/3170/orig -> origin/gh/ezyang/3170/orig 2025-10-10T01:18:33.5883329Z * [new branch] gh/ezyang/3171/base -> origin/gh/ezyang/3171/base 2025-10-10T01:18:33.5883708Z * [new branch] gh/ezyang/3171/head -> origin/gh/ezyang/3171/head 2025-10-10T01:18:33.5884229Z * [new branch] gh/ezyang/3171/orig -> origin/gh/ezyang/3171/orig 2025-10-10T01:18:33.5886787Z * [new branch] gh/ezyang/3172/base -> origin/gh/ezyang/3172/base 2025-10-10T01:18:33.5888453Z * [new branch] gh/ezyang/3172/head -> origin/gh/ezyang/3172/head 2025-10-10T01:18:33.5890132Z * [new branch] gh/ezyang/3172/orig -> origin/gh/ezyang/3172/orig 2025-10-10T01:18:33.5892463Z * [new branch] gh/ezyang/3173/base -> origin/gh/ezyang/3173/base 2025-10-10T01:18:33.5894148Z * [new branch] gh/ezyang/3173/head -> origin/gh/ezyang/3173/head 2025-10-10T01:18:33.5895727Z * [new branch] gh/ezyang/3173/orig -> origin/gh/ezyang/3173/orig 2025-10-10T01:18:33.5898443Z * [new branch] gh/fadara01/1/base -> origin/gh/fadara01/1/base 2025-10-10T01:18:33.5900135Z * [new branch] gh/fadara01/1/head -> origin/gh/fadara01/1/head 2025-10-10T01:18:33.5901844Z * [new branch] gh/fadara01/1/orig -> origin/gh/fadara01/1/orig 2025-10-10T01:18:33.5904708Z * [new branch] gh/fduwjj/175/base -> origin/gh/fduwjj/175/base 2025-10-10T01:18:33.5906553Z * [new branch] gh/fduwjj/175/head -> origin/gh/fduwjj/175/head 2025-10-10T01:18:33.5908215Z * [new branch] gh/fduwjj/175/orig -> origin/gh/fduwjj/175/orig 2025-10-10T01:18:33.5910669Z * [new branch] gh/fduwjj/176/base -> origin/gh/fduwjj/176/base 2025-10-10T01:18:33.5912385Z * [new branch] gh/fduwjj/176/head -> origin/gh/fduwjj/176/head 2025-10-10T01:18:33.5914142Z * [new branch] gh/fduwjj/176/orig -> origin/gh/fduwjj/176/orig 2025-10-10T01:18:33.5916252Z * [new branch] gh/fduwjj/177/base -> origin/gh/fduwjj/177/base 2025-10-10T01:18:33.5917973Z * [new branch] gh/fduwjj/177/head -> origin/gh/fduwjj/177/head 2025-10-10T01:18:33.5919739Z * [new branch] gh/fduwjj/177/orig -> origin/gh/fduwjj/177/orig 2025-10-10T01:18:33.5922124Z * [new branch] gh/fduwjj/182/base -> origin/gh/fduwjj/182/base 2025-10-10T01:18:33.5923797Z * [new branch] gh/fduwjj/182/head -> origin/gh/fduwjj/182/head 2025-10-10T01:18:33.5926011Z * [new branch] gh/fduwjj/182/orig -> origin/gh/fduwjj/182/orig 2025-10-10T01:18:33.5928368Z * [new branch] gh/fduwjj/183/base -> origin/gh/fduwjj/183/base 2025-10-10T01:18:33.5930156Z * [new branch] gh/fduwjj/183/head -> origin/gh/fduwjj/183/head 2025-10-10T01:18:33.5931868Z * [new branch] gh/fduwjj/183/orig -> origin/gh/fduwjj/183/orig 2025-10-10T01:18:33.5934424Z * [new branch] gh/fduwjj/184/base -> origin/gh/fduwjj/184/base 2025-10-10T01:18:33.5936145Z * [new branch] gh/fduwjj/184/head -> origin/gh/fduwjj/184/head 2025-10-10T01:18:33.5937788Z * [new branch] gh/fduwjj/184/orig -> origin/gh/fduwjj/184/orig 2025-10-10T01:18:33.5940048Z * [new branch] gh/fduwjj/185/base -> origin/gh/fduwjj/185/base 2025-10-10T01:18:33.5941762Z * [new branch] gh/fduwjj/185/head -> origin/gh/fduwjj/185/head 2025-10-10T01:18:33.5943430Z * [new branch] gh/fduwjj/185/orig -> origin/gh/fduwjj/185/orig 2025-10-10T01:18:33.5945830Z * [new branch] gh/fduwjj/191/base -> origin/gh/fduwjj/191/base 2025-10-10T01:18:33.5947558Z * [new branch] gh/fduwjj/191/head -> origin/gh/fduwjj/191/head 2025-10-10T01:18:33.5949250Z * [new branch] gh/fduwjj/191/orig -> origin/gh/fduwjj/191/orig 2025-10-10T01:18:33.5951515Z * [new branch] gh/fduwjj/192/base -> origin/gh/fduwjj/192/base 2025-10-10T01:18:33.5953283Z * [new branch] gh/fduwjj/192/head -> origin/gh/fduwjj/192/head 2025-10-10T01:18:33.5954932Z * [new branch] gh/fduwjj/192/orig -> origin/gh/fduwjj/192/orig 2025-10-10T01:18:33.5957457Z * [new branch] gh/fduwjj/193/base -> origin/gh/fduwjj/193/base 2025-10-10T01:18:33.5959125Z * [new branch] gh/fduwjj/193/head -> origin/gh/fduwjj/193/head 2025-10-10T01:18:33.5960750Z * [new branch] gh/fduwjj/193/orig -> origin/gh/fduwjj/193/orig 2025-10-10T01:18:33.5963174Z * [new branch] gh/fduwjj/194/base -> origin/gh/fduwjj/194/base 2025-10-10T01:18:33.5964769Z * [new branch] gh/fduwjj/194/head -> origin/gh/fduwjj/194/head 2025-10-10T01:18:33.5966431Z * [new branch] gh/fduwjj/194/orig -> origin/gh/fduwjj/194/orig 2025-10-10T01:18:33.5968857Z * [new branch] gh/fduwjj/195/base -> origin/gh/fduwjj/195/base 2025-10-10T01:18:33.5970597Z * [new branch] gh/fduwjj/195/head -> origin/gh/fduwjj/195/head 2025-10-10T01:18:33.5972347Z * [new branch] gh/fduwjj/195/orig -> origin/gh/fduwjj/195/orig 2025-10-10T01:18:33.5974479Z * [new branch] gh/fduwjj/196/base -> origin/gh/fduwjj/196/base 2025-10-10T01:18:33.5976724Z * [new branch] gh/fduwjj/196/head -> origin/gh/fduwjj/196/head 2025-10-10T01:18:33.5978373Z * [new branch] gh/fduwjj/196/orig -> origin/gh/fduwjj/196/orig 2025-10-10T01:18:33.5980666Z * [new branch] gh/fduwjj/197/base -> origin/gh/fduwjj/197/base 2025-10-10T01:18:33.5982302Z * [new branch] gh/fduwjj/197/head -> origin/gh/fduwjj/197/head 2025-10-10T01:18:33.5984359Z * [new branch] gh/fduwjj/197/orig -> origin/gh/fduwjj/197/orig 2025-10-10T01:18:33.5986170Z * [new branch] gh/fduwjj/198/base -> origin/gh/fduwjj/198/base 2025-10-10T01:18:33.5987804Z * [new branch] gh/fduwjj/198/head -> origin/gh/fduwjj/198/head 2025-10-10T01:18:33.5989441Z * [new branch] gh/fduwjj/198/orig -> origin/gh/fduwjj/198/orig 2025-10-10T01:18:33.5991584Z * [new branch] gh/fduwjj/199/base -> origin/gh/fduwjj/199/base 2025-10-10T01:18:33.5993207Z * [new branch] gh/fduwjj/199/head -> origin/gh/fduwjj/199/head 2025-10-10T01:18:33.5994888Z * [new branch] gh/fduwjj/199/orig -> origin/gh/fduwjj/199/orig 2025-10-10T01:18:33.5997530Z * [new branch] gh/fduwjj/200/base -> origin/gh/fduwjj/200/base 2025-10-10T01:18:33.5999402Z * [new branch] gh/fduwjj/200/head -> origin/gh/fduwjj/200/head 2025-10-10T01:18:33.6001087Z * [new branch] gh/fduwjj/200/orig -> origin/gh/fduwjj/200/orig 2025-10-10T01:18:33.6003385Z * [new branch] gh/fduwjj/201/base -> origin/gh/fduwjj/201/base 2025-10-10T01:18:33.6005085Z * [new branch] gh/fduwjj/201/head -> origin/gh/fduwjj/201/head 2025-10-10T01:18:33.6006712Z * [new branch] gh/fduwjj/201/orig -> origin/gh/fduwjj/201/orig 2025-10-10T01:18:33.6009127Z * [new branch] gh/fduwjj/202/base -> origin/gh/fduwjj/202/base 2025-10-10T01:18:33.6010749Z * [new branch] gh/fduwjj/202/head -> origin/gh/fduwjj/202/head 2025-10-10T01:18:33.6012396Z * [new branch] gh/fduwjj/202/orig -> origin/gh/fduwjj/202/orig 2025-10-10T01:18:33.6014897Z * [new branch] gh/fduwjj/203/base -> origin/gh/fduwjj/203/base 2025-10-10T01:18:33.6016698Z * [new branch] gh/fduwjj/203/head -> origin/gh/fduwjj/203/head 2025-10-10T01:18:33.6018319Z * [new branch] gh/fduwjj/203/orig -> origin/gh/fduwjj/203/orig 2025-10-10T01:18:33.6020866Z * [new branch] gh/fduwjj/204/base -> origin/gh/fduwjj/204/base 2025-10-10T01:18:33.6022542Z * [new branch] gh/fduwjj/204/head -> origin/gh/fduwjj/204/head 2025-10-10T01:18:33.6024471Z * [new branch] gh/fduwjj/204/orig -> origin/gh/fduwjj/204/orig 2025-10-10T01:18:33.6027094Z * [new branch] gh/fduwjj/205/base -> origin/gh/fduwjj/205/base 2025-10-10T01:18:33.6028755Z * [new branch] gh/fduwjj/205/head -> origin/gh/fduwjj/205/head 2025-10-10T01:18:33.6030399Z * [new branch] gh/fduwjj/205/orig -> origin/gh/fduwjj/205/orig 2025-10-10T01:18:33.6032913Z * [new branch] gh/fduwjj/206/base -> origin/gh/fduwjj/206/base 2025-10-10T01:18:33.6034633Z * [new branch] gh/fduwjj/206/head -> origin/gh/fduwjj/206/head 2025-10-10T01:18:33.6036273Z * [new branch] gh/fduwjj/206/orig -> origin/gh/fduwjj/206/orig 2025-10-10T01:18:33.6038842Z * [new branch] gh/fduwjj/207/base -> origin/gh/fduwjj/207/base 2025-10-10T01:18:33.6040590Z * [new branch] gh/fduwjj/207/head -> origin/gh/fduwjj/207/head 2025-10-10T01:18:33.6042156Z * [new branch] gh/fduwjj/207/orig -> origin/gh/fduwjj/207/orig 2025-10-10T01:18:33.6044601Z * [new branch] gh/fduwjj/208/base -> origin/gh/fduwjj/208/base 2025-10-10T01:18:33.6046262Z * [new branch] gh/fduwjj/208/head -> origin/gh/fduwjj/208/head 2025-10-10T01:18:33.6047988Z * [new branch] gh/fduwjj/208/orig -> origin/gh/fduwjj/208/orig 2025-10-10T01:18:33.6050258Z * [new branch] gh/fduwjj/209/base -> origin/gh/fduwjj/209/base 2025-10-10T01:18:33.6052063Z * [new branch] gh/fduwjj/209/head -> origin/gh/fduwjj/209/head 2025-10-10T01:18:33.6053596Z * [new branch] gh/fduwjj/209/orig -> origin/gh/fduwjj/209/orig 2025-10-10T01:18:33.6055911Z * [new branch] gh/fduwjj/210/base -> origin/gh/fduwjj/210/base 2025-10-10T01:18:33.6057572Z * [new branch] gh/fduwjj/210/head -> origin/gh/fduwjj/210/head 2025-10-10T01:18:33.6059824Z * [new branch] gh/fduwjj/210/orig -> origin/gh/fduwjj/210/orig 2025-10-10T01:18:33.6062182Z * [new branch] gh/fduwjj/211/base -> origin/gh/fduwjj/211/base 2025-10-10T01:18:33.6063922Z * [new branch] gh/fduwjj/211/head -> origin/gh/fduwjj/211/head 2025-10-10T01:18:33.6065579Z * [new branch] gh/fduwjj/211/orig -> origin/gh/fduwjj/211/orig 2025-10-10T01:18:33.6067979Z * [new branch] gh/fduwjj/212/base -> origin/gh/fduwjj/212/base 2025-10-10T01:18:33.6069669Z * [new branch] gh/fduwjj/212/head -> origin/gh/fduwjj/212/head 2025-10-10T01:18:33.6071321Z * [new branch] gh/fduwjj/212/orig -> origin/gh/fduwjj/212/orig 2025-10-10T01:18:33.6073764Z * [new branch] gh/fduwjj/213/base -> origin/gh/fduwjj/213/base 2025-10-10T01:18:33.6075525Z * [new branch] gh/fduwjj/213/head -> origin/gh/fduwjj/213/head 2025-10-10T01:18:33.6077175Z * [new branch] gh/fduwjj/213/orig -> origin/gh/fduwjj/213/orig 2025-10-10T01:18:33.6079776Z * [new branch] gh/fduwjj/214/base -> origin/gh/fduwjj/214/base 2025-10-10T01:18:33.6081422Z * [new branch] gh/fduwjj/214/head -> origin/gh/fduwjj/214/head 2025-10-10T01:18:33.6083070Z * [new branch] gh/fduwjj/214/orig -> origin/gh/fduwjj/214/orig 2025-10-10T01:18:33.6085382Z * [new branch] gh/fduwjj/215/base -> origin/gh/fduwjj/215/base 2025-10-10T01:18:33.6087056Z * [new branch] gh/fduwjj/215/head -> origin/gh/fduwjj/215/head 2025-10-10T01:18:33.6088725Z * [new branch] gh/fduwjj/215/orig -> origin/gh/fduwjj/215/orig 2025-10-10T01:18:33.6091517Z * [new branch] gh/fduwjj/216/base -> origin/gh/fduwjj/216/base 2025-10-10T01:18:33.6093319Z * [new branch] gh/fduwjj/216/head -> origin/gh/fduwjj/216/head 2025-10-10T01:18:33.6094881Z * [new branch] gh/fduwjj/216/orig -> origin/gh/fduwjj/216/orig 2025-10-10T01:18:33.6097423Z * [new branch] gh/fduwjj/217/base -> origin/gh/fduwjj/217/base 2025-10-10T01:18:33.6099053Z * [new branch] gh/fduwjj/217/head -> origin/gh/fduwjj/217/head 2025-10-10T01:18:33.6100691Z * [new branch] gh/fduwjj/217/orig -> origin/gh/fduwjj/217/orig 2025-10-10T01:18:33.6103069Z * [new branch] gh/fduwjj/218/base -> origin/gh/fduwjj/218/base 2025-10-10T01:18:33.6104728Z * [new branch] gh/fduwjj/218/head -> origin/gh/fduwjj/218/head 2025-10-10T01:18:33.6106375Z * [new branch] gh/fduwjj/218/orig -> origin/gh/fduwjj/218/orig 2025-10-10T01:18:33.6108757Z * [new branch] gh/fduwjj/219/base -> origin/gh/fduwjj/219/base 2025-10-10T01:18:33.6110557Z * [new branch] gh/fduwjj/219/head -> origin/gh/fduwjj/219/head 2025-10-10T01:18:33.6112222Z * [new branch] gh/fduwjj/219/orig -> origin/gh/fduwjj/219/orig 2025-10-10T01:18:33.6114628Z * [new branch] gh/fduwjj/220/base -> origin/gh/fduwjj/220/base 2025-10-10T01:18:33.6116347Z * [new branch] gh/fduwjj/220/head -> origin/gh/fduwjj/220/head 2025-10-10T01:18:33.6117989Z * [new branch] gh/fduwjj/220/orig -> origin/gh/fduwjj/220/orig 2025-10-10T01:18:33.6121001Z * [new branch] gh/fduwjj/221/base -> origin/gh/fduwjj/221/base 2025-10-10T01:18:33.6122938Z * [new branch] gh/fduwjj/221/head -> origin/gh/fduwjj/221/head 2025-10-10T01:18:33.6124807Z * [new branch] gh/fduwjj/221/orig -> origin/gh/fduwjj/221/orig 2025-10-10T01:18:33.6127347Z * [new branch] gh/fduwjj/222/base -> origin/gh/fduwjj/222/base 2025-10-10T01:18:33.6128710Z * [new branch] gh/fduwjj/222/head -> origin/gh/fduwjj/222/head 2025-10-10T01:18:33.6130326Z * [new branch] gh/fduwjj/222/orig -> origin/gh/fduwjj/222/orig 2025-10-10T01:18:33.6132715Z * [new branch] gh/fduwjj/223/base -> origin/gh/fduwjj/223/base 2025-10-10T01:18:33.6134401Z * [new branch] gh/fduwjj/223/head -> origin/gh/fduwjj/223/head 2025-10-10T01:18:33.6136003Z * [new branch] gh/fduwjj/223/orig -> origin/gh/fduwjj/223/orig 2025-10-10T01:18:33.6138733Z * [new branch] gh/fegin/313/base -> origin/gh/fegin/313/base 2025-10-10T01:18:33.6140458Z * [new branch] gh/fegin/313/head -> origin/gh/fegin/313/head 2025-10-10T01:18:33.6142210Z * [new branch] gh/fegin/313/orig -> origin/gh/fegin/313/orig 2025-10-10T01:18:33.6144480Z * [new branch] gh/fegin/314/base -> origin/gh/fegin/314/base 2025-10-10T01:18:33.6146140Z * [new branch] gh/fegin/314/head -> origin/gh/fegin/314/head 2025-10-10T01:18:33.6147791Z * [new branch] gh/fegin/314/orig -> origin/gh/fegin/314/orig 2025-10-10T01:18:33.6150029Z * [new branch] gh/fegin/315/base -> origin/gh/fegin/315/base 2025-10-10T01:18:33.6151659Z * [new branch] gh/fegin/315/head -> origin/gh/fegin/315/head 2025-10-10T01:18:33.6153455Z * [new branch] gh/fegin/315/orig -> origin/gh/fegin/315/orig 2025-10-10T01:18:33.6155684Z * [new branch] gh/fegin/316/base -> origin/gh/fegin/316/base 2025-10-10T01:18:33.6157379Z * [new branch] gh/fegin/316/head -> origin/gh/fegin/316/head 2025-10-10T01:18:33.6159124Z * [new branch] gh/fegin/316/orig -> origin/gh/fegin/316/orig 2025-10-10T01:18:33.6161458Z * [new branch] gh/fegin/317/base -> origin/gh/fegin/317/base 2025-10-10T01:18:33.6163065Z * [new branch] gh/fegin/317/head -> origin/gh/fegin/317/head 2025-10-10T01:18:33.6164797Z * [new branch] gh/fegin/317/orig -> origin/gh/fegin/317/orig 2025-10-10T01:18:33.6167065Z * [new branch] gh/fegin/318/base -> origin/gh/fegin/318/base 2025-10-10T01:18:33.6168881Z * [new branch] gh/fegin/318/head -> origin/gh/fegin/318/head 2025-10-10T01:18:33.6170546Z * [new branch] gh/fegin/318/orig -> origin/gh/fegin/318/orig 2025-10-10T01:18:33.6172824Z * [new branch] gh/fegin/319/base -> origin/gh/fegin/319/base 2025-10-10T01:18:33.6174497Z * [new branch] gh/fegin/319/head -> origin/gh/fegin/319/head 2025-10-10T01:18:33.6176246Z * [new branch] gh/fegin/319/orig -> origin/gh/fegin/319/orig 2025-10-10T01:18:33.6178409Z * [new branch] gh/fegin/320/base -> origin/gh/fegin/320/base 2025-10-10T01:18:33.6180086Z * [new branch] gh/fegin/320/head -> origin/gh/fegin/320/head 2025-10-10T01:18:33.6181754Z * [new branch] gh/fegin/320/orig -> origin/gh/fegin/320/orig 2025-10-10T01:18:33.6184004Z * [new branch] gh/fegin/321/base -> origin/gh/fegin/321/base 2025-10-10T01:18:33.6185629Z * [new branch] gh/fegin/321/head -> origin/gh/fegin/321/head 2025-10-10T01:18:33.6187499Z * [new branch] gh/fegin/321/orig -> origin/gh/fegin/321/orig 2025-10-10T01:18:33.6189689Z * [new branch] gh/fegin/322/base -> origin/gh/fegin/322/base 2025-10-10T01:18:33.6191529Z * [new branch] gh/fegin/322/head -> origin/gh/fegin/322/head 2025-10-10T01:18:33.6193054Z * [new branch] gh/fegin/322/orig -> origin/gh/fegin/322/orig 2025-10-10T01:18:33.6195257Z * [new branch] gh/fegin/323/base -> origin/gh/fegin/323/base 2025-10-10T01:18:33.6196947Z * [new branch] gh/fegin/323/head -> origin/gh/fegin/323/head 2025-10-10T01:18:33.6199402Z * [new branch] gh/fegin/324/base -> origin/gh/fegin/324/base 2025-10-10T01:18:33.6201198Z * [new branch] gh/fegin/324/head -> origin/gh/fegin/324/head 2025-10-10T01:18:33.6202888Z * [new branch] gh/fegin/324/orig -> origin/gh/fegin/324/orig 2025-10-10T01:18:33.6205191Z * [new branch] gh/fegin/325/base -> origin/gh/fegin/325/base 2025-10-10T01:18:33.6207363Z * [new branch] gh/fegin/325/head -> origin/gh/fegin/325/head 2025-10-10T01:18:33.6209091Z * [new branch] gh/fegin/325/orig -> origin/gh/fegin/325/orig 2025-10-10T01:18:33.6211479Z * [new branch] gh/fegin/326/base -> origin/gh/fegin/326/base 2025-10-10T01:18:33.6213266Z * [new branch] gh/fegin/326/head -> origin/gh/fegin/326/head 2025-10-10T01:18:33.6214928Z * [new branch] gh/fegin/326/orig -> origin/gh/fegin/326/orig 2025-10-10T01:18:33.6217220Z * [new branch] gh/fegin/327/base -> origin/gh/fegin/327/base 2025-10-10T01:18:33.6218883Z * [new branch] gh/fegin/327/head -> origin/gh/fegin/327/head 2025-10-10T01:18:33.6220543Z * [new branch] gh/fegin/327/orig -> origin/gh/fegin/327/orig 2025-10-10T01:18:33.6223306Z * [new branch] gh/fffrog/133/base -> origin/gh/fffrog/133/base 2025-10-10T01:18:33.6225314Z * [new branch] gh/fffrog/133/head -> origin/gh/fffrog/133/head 2025-10-10T01:18:33.6226995Z * [new branch] gh/fffrog/133/orig -> origin/gh/fffrog/133/orig 2025-10-10T01:18:33.6229202Z * [new branch] gh/fffrog/137/base -> origin/gh/fffrog/137/base 2025-10-10T01:18:33.6230871Z * [new branch] gh/fffrog/137/head -> origin/gh/fffrog/137/head 2025-10-10T01:18:33.6232627Z * [new branch] gh/fffrog/137/orig -> origin/gh/fffrog/137/orig 2025-10-10T01:18:33.6234835Z * [new branch] gh/fffrog/147/base -> origin/gh/fffrog/147/base 2025-10-10T01:18:33.6236492Z * [new branch] gh/fffrog/147/head -> origin/gh/fffrog/147/head 2025-10-10T01:18:33.6238701Z * [new branch] gh/fffrog/147/orig -> origin/gh/fffrog/147/orig 2025-10-10T01:18:33.6241097Z * [new branch] gh/fffrog/149/base -> origin/gh/fffrog/149/base 2025-10-10T01:18:33.6242730Z * [new branch] gh/fffrog/149/head -> origin/gh/fffrog/149/head 2025-10-10T01:18:33.6244419Z * [new branch] gh/fffrog/149/orig -> origin/gh/fffrog/149/orig 2025-10-10T01:18:33.6246740Z * [new branch] gh/fffrog/150/base -> origin/gh/fffrog/150/base 2025-10-10T01:18:33.6248371Z * [new branch] gh/fffrog/150/head -> origin/gh/fffrog/150/head 2025-10-10T01:18:33.6250076Z * [new branch] gh/fffrog/150/orig -> origin/gh/fffrog/150/orig 2025-10-10T01:18:33.6252278Z * [new branch] gh/fffrog/153/base -> origin/gh/fffrog/153/base 2025-10-10T01:18:33.6254047Z * [new branch] gh/fffrog/153/head -> origin/gh/fffrog/153/head 2025-10-10T01:18:33.6256373Z * [new branch] gh/fffrog/153/orig -> origin/gh/fffrog/153/orig 2025-10-10T01:18:33.6258447Z * [new branch] gh/fffrog/154/base -> origin/gh/fffrog/154/base 2025-10-10T01:18:33.6261377Z * [new branch] gh/fffrog/154/head -> origin/gh/fffrog/154/head 2025-10-10T01:18:33.6262465Z * [new branch] gh/fffrog/154/orig -> origin/gh/fffrog/154/orig 2025-10-10T01:18:33.6264582Z * [new branch] gh/fffrog/155/base -> origin/gh/fffrog/155/base 2025-10-10T01:18:33.6266541Z * [new branch] gh/fffrog/155/head -> origin/gh/fffrog/155/head 2025-10-10T01:18:33.6268334Z * [new branch] gh/fffrog/155/orig -> origin/gh/fffrog/155/orig 2025-10-10T01:18:33.6271246Z * [new branch] gh/fffrog/156/base -> origin/gh/fffrog/156/base 2025-10-10T01:18:33.6273155Z * [new branch] gh/fffrog/156/head -> origin/gh/fffrog/156/head 2025-10-10T01:18:33.6274827Z * [new branch] gh/fffrog/156/orig -> origin/gh/fffrog/156/orig 2025-10-10T01:18:33.6277061Z * [new branch] gh/fffrog/157/base -> origin/gh/fffrog/157/base 2025-10-10T01:18:33.6278797Z * [new branch] gh/fffrog/157/head -> origin/gh/fffrog/157/head 2025-10-10T01:18:33.6280573Z * [new branch] gh/fffrog/157/orig -> origin/gh/fffrog/157/orig 2025-10-10T01:18:33.6282905Z * [new branch] gh/fffrog/158/base -> origin/gh/fffrog/158/base 2025-10-10T01:18:33.6284662Z * [new branch] gh/fffrog/158/head -> origin/gh/fffrog/158/head 2025-10-10T01:18:33.6286377Z * [new branch] gh/fffrog/158/orig -> origin/gh/fffrog/158/orig 2025-10-10T01:18:33.6289073Z * [new branch] gh/fffrog/159/base -> origin/gh/fffrog/159/base 2025-10-10T01:18:33.6290749Z * [new branch] gh/fffrog/159/head -> origin/gh/fffrog/159/head 2025-10-10T01:18:33.6292558Z * [new branch] gh/fffrog/159/orig -> origin/gh/fffrog/159/orig 2025-10-10T01:18:33.6294739Z * [new branch] gh/fffrog/160/base -> origin/gh/fffrog/160/base 2025-10-10T01:18:33.6296778Z * [new branch] gh/fffrog/160/head -> origin/gh/fffrog/160/head 2025-10-10T01:18:33.6299662Z * [new branch] gh/fffrog/161/base -> origin/gh/fffrog/161/base 2025-10-10T01:18:33.6300917Z * [new branch] gh/fffrog/161/head -> origin/gh/fffrog/161/head 2025-10-10T01:18:33.6302517Z * [new branch] gh/fffrog/161/orig -> origin/gh/fffrog/161/orig 2025-10-10T01:18:33.6305278Z * [new branch] gh/fffrog/162/base -> origin/gh/fffrog/162/base 2025-10-10T01:18:33.6306802Z * [new branch] gh/fffrog/162/head -> origin/gh/fffrog/162/head 2025-10-10T01:18:33.6308528Z * [new branch] gh/fffrog/162/orig -> origin/gh/fffrog/162/orig 2025-10-10T01:18:33.6310847Z * [new branch] gh/fffrog/163/base -> origin/gh/fffrog/163/base 2025-10-10T01:18:33.6312450Z * [new branch] gh/fffrog/163/head -> origin/gh/fffrog/163/head 2025-10-10T01:18:33.6314122Z * [new branch] gh/fffrog/163/orig -> origin/gh/fffrog/163/orig 2025-10-10T01:18:33.6316817Z * [new branch] gh/fffrog/164/base -> origin/gh/fffrog/164/base 2025-10-10T01:18:33.6318545Z * [new branch] gh/fffrog/164/head -> origin/gh/fffrog/164/head 2025-10-10T01:18:33.6320341Z * [new branch] gh/fffrog/164/orig -> origin/gh/fffrog/164/orig 2025-10-10T01:18:33.6322882Z * [new branch] gh/fffrog/165/base -> origin/gh/fffrog/165/base 2025-10-10T01:18:33.6324695Z * [new branch] gh/fffrog/165/head -> origin/gh/fffrog/165/head 2025-10-10T01:18:33.6327549Z * [new branch] gh/fffrog/165/orig -> origin/gh/fffrog/165/orig 2025-10-10T01:18:33.6329938Z * [new branch] gh/fffrog/166/base -> origin/gh/fffrog/166/base 2025-10-10T01:18:33.6331628Z * [new branch] gh/fffrog/166/head -> origin/gh/fffrog/166/head 2025-10-10T01:18:33.6333458Z * [new branch] gh/fffrog/166/orig -> origin/gh/fffrog/166/orig 2025-10-10T01:18:33.6335634Z * [new branch] gh/fffrog/167/base -> origin/gh/fffrog/167/base 2025-10-10T01:18:33.6337236Z * [new branch] gh/fffrog/167/head -> origin/gh/fffrog/167/head 2025-10-10T01:18:33.6338957Z * [new branch] gh/fffrog/167/orig -> origin/gh/fffrog/167/orig 2025-10-10T01:18:33.6341303Z * [new branch] gh/fffrog/168/base -> origin/gh/fffrog/168/base 2025-10-10T01:18:33.6342958Z * [new branch] gh/fffrog/168/head -> origin/gh/fffrog/168/head 2025-10-10T01:18:33.6344664Z * [new branch] gh/fffrog/168/orig -> origin/gh/fffrog/168/orig 2025-10-10T01:18:33.6346973Z * [new branch] gh/fffrog/169/base -> origin/gh/fffrog/169/base 2025-10-10T01:18:33.6348614Z * [new branch] gh/fffrog/169/head -> origin/gh/fffrog/169/head 2025-10-10T01:18:33.6350290Z * [new branch] gh/fffrog/169/orig -> origin/gh/fffrog/169/orig 2025-10-10T01:18:33.6352632Z * [new branch] gh/fffrog/170/base -> origin/gh/fffrog/170/base 2025-10-10T01:18:33.6354274Z * [new branch] gh/fffrog/170/head -> origin/gh/fffrog/170/head 2025-10-10T01:18:33.6356280Z * [new branch] gh/fffrog/170/orig -> origin/gh/fffrog/170/orig 2025-10-10T01:18:33.6359387Z * [new branch] gh/fffrog/171/base -> origin/gh/fffrog/171/base 2025-10-10T01:18:33.6360724Z * [new branch] gh/fffrog/171/head -> origin/gh/fffrog/171/head 2025-10-10T01:18:33.6362328Z * [new branch] gh/fffrog/171/orig -> origin/gh/fffrog/171/orig 2025-10-10T01:18:33.6365085Z * [new branch] gh/fffrog/172/base -> origin/gh/fffrog/172/base 2025-10-10T01:18:33.6366575Z * [new branch] gh/fffrog/172/head -> origin/gh/fffrog/172/head 2025-10-10T01:18:33.6368323Z * [new branch] gh/fffrog/172/orig -> origin/gh/fffrog/172/orig 2025-10-10T01:18:33.6370601Z * [new branch] gh/fffrog/173/base -> origin/gh/fffrog/173/base 2025-10-10T01:18:33.6372234Z * [new branch] gh/fffrog/173/head -> origin/gh/fffrog/173/head 2025-10-10T01:18:33.6373859Z * [new branch] gh/fffrog/173/orig -> origin/gh/fffrog/173/orig 2025-10-10T01:18:33.6376124Z * [new branch] gh/fffrog/174/base -> origin/gh/fffrog/174/base 2025-10-10T01:18:33.6377884Z * [new branch] gh/fffrog/174/head -> origin/gh/fffrog/174/head 2025-10-10T01:18:33.6379493Z * [new branch] gh/fffrog/174/orig -> origin/gh/fffrog/174/orig 2025-10-10T01:18:33.6381874Z * [new branch] gh/fffrog/175/base -> origin/gh/fffrog/175/base 2025-10-10T01:18:33.6383555Z * [new branch] gh/fffrog/175/head -> origin/gh/fffrog/175/head 2025-10-10T01:18:33.6385188Z * [new branch] gh/fffrog/175/orig -> origin/gh/fffrog/175/orig 2025-10-10T01:18:33.6387589Z * [new branch] gh/fffrog/176/base -> origin/gh/fffrog/176/base 2025-10-10T01:18:33.6389363Z * [new branch] gh/fffrog/176/head -> origin/gh/fffrog/176/head 2025-10-10T01:18:33.6391125Z * [new branch] gh/fffrog/176/orig -> origin/gh/fffrog/176/orig 2025-10-10T01:18:33.6393911Z * [new branch] gh/fxdawnn/1/base -> origin/gh/fxdawnn/1/base 2025-10-10T01:18:33.6395600Z * [new branch] gh/fxdawnn/1/head -> origin/gh/fxdawnn/1/head 2025-10-10T01:18:33.6397337Z * [new branch] gh/fxdawnn/1/orig -> origin/gh/fxdawnn/1/orig 2025-10-10T01:18:33.6399743Z * [new branch] gh/fxdawnn/2/base -> origin/gh/fxdawnn/2/base 2025-10-10T01:18:33.6401415Z * [new branch] gh/fxdawnn/2/head -> origin/gh/fxdawnn/2/head 2025-10-10T01:18:33.6403164Z * [new branch] gh/fxdawnn/2/orig -> origin/gh/fxdawnn/2/orig 2025-10-10T01:18:33.6405266Z * [new branch] gh/fxdawnn/3/base -> origin/gh/fxdawnn/3/base 2025-10-10T01:18:33.6406843Z * [new branch] gh/fxdawnn/3/head -> origin/gh/fxdawnn/3/head 2025-10-10T01:18:33.6408734Z * [new branch] gh/fxdawnn/3/orig -> origin/gh/fxdawnn/3/orig 2025-10-10T01:18:33.6410796Z * [new branch] gh/fxdawnn/4/base -> origin/gh/fxdawnn/4/base 2025-10-10T01:18:33.6412506Z * [new branch] gh/fxdawnn/4/orig -> origin/gh/fxdawnn/4/orig 2025-10-10T01:18:33.6415288Z * [new branch] gh/gmagogsfm/1/base -> origin/gh/gmagogsfm/1/base 2025-10-10T01:18:33.6416969Z * [new branch] gh/gmagogsfm/1/head -> origin/gh/gmagogsfm/1/head 2025-10-10T01:18:33.6418750Z * [new branch] gh/gmagogsfm/1/orig -> origin/gh/gmagogsfm/1/orig 2025-10-10T01:18:33.6421499Z * [new branch] gh/gmagogsfm/2/base -> origin/gh/gmagogsfm/2/base 2025-10-10T01:18:33.6423146Z * [new branch] gh/gmagogsfm/2/head -> origin/gh/gmagogsfm/2/head 2025-10-10T01:18:33.6425115Z * [new branch] gh/gmagogsfm/2/orig -> origin/gh/gmagogsfm/2/orig 2025-10-10T01:18:33.6427246Z * [new branch] gh/gmagogsfm/3/base -> origin/gh/gmagogsfm/3/base 2025-10-10T01:18:33.6428871Z * [new branch] gh/gmagogsfm/3/head -> origin/gh/gmagogsfm/3/head 2025-10-10T01:18:33.6430560Z * [new branch] gh/gmagogsfm/3/orig -> origin/gh/gmagogsfm/3/orig 2025-10-10T01:18:33.6433388Z * [new branch] gh/guangyey/134/base -> origin/gh/guangyey/134/base 2025-10-10T01:18:33.6435022Z * [new branch] gh/guangyey/134/head -> origin/gh/guangyey/134/head 2025-10-10T01:18:33.6436700Z * [new branch] gh/guangyey/134/orig -> origin/gh/guangyey/134/orig 2025-10-10T01:18:33.6439022Z * [new branch] gh/guangyey/135/base -> origin/gh/guangyey/135/base 2025-10-10T01:18:33.6440805Z * [new branch] gh/guangyey/135/head -> origin/gh/guangyey/135/head 2025-10-10T01:18:33.6442443Z * [new branch] gh/guangyey/135/orig -> origin/gh/guangyey/135/orig 2025-10-10T01:18:33.6444725Z * [new branch] gh/guangyey/139/base -> origin/gh/guangyey/139/base 2025-10-10T01:18:33.6446675Z * [new branch] gh/guangyey/139/head -> origin/gh/guangyey/139/head 2025-10-10T01:18:33.6448789Z * [new branch] gh/guangyey/139/orig -> origin/gh/guangyey/139/orig 2025-10-10T01:18:33.6450807Z * [new branch] gh/guangyey/140/base -> origin/gh/guangyey/140/base 2025-10-10T01:18:33.6452414Z * [new branch] gh/guangyey/140/head -> origin/gh/guangyey/140/head 2025-10-10T01:18:33.6454102Z * [new branch] gh/guangyey/140/orig -> origin/gh/guangyey/140/orig 2025-10-10T01:18:33.6457281Z * [new branch] gh/guangyey/142/base -> origin/gh/guangyey/142/base 2025-10-10T01:18:33.6458790Z * [new branch] gh/guangyey/142/head -> origin/gh/guangyey/142/head 2025-10-10T01:18:33.6460565Z * [new branch] gh/guangyey/142/orig -> origin/gh/guangyey/142/orig 2025-10-10T01:18:33.6462790Z * [new branch] gh/guangyey/163/base -> origin/gh/guangyey/163/base 2025-10-10T01:18:33.6464554Z * [new branch] gh/guangyey/163/head -> origin/gh/guangyey/163/head 2025-10-10T01:18:33.6466229Z * [new branch] gh/guangyey/163/orig -> origin/gh/guangyey/163/orig 2025-10-10T01:18:33.6468613Z * [new branch] gh/guangyey/168/base -> origin/gh/guangyey/168/base 2025-10-10T01:18:33.6470204Z * [new branch] gh/guangyey/168/head -> origin/gh/guangyey/168/head 2025-10-10T01:18:33.6472033Z * [new branch] gh/guangyey/168/orig -> origin/gh/guangyey/168/orig 2025-10-10T01:18:33.6474151Z * [new branch] gh/guangyey/169/base -> origin/gh/guangyey/169/base 2025-10-10T01:18:33.6475789Z * [new branch] gh/guangyey/169/head -> origin/gh/guangyey/169/head 2025-10-10T01:18:33.6477425Z * [new branch] gh/guangyey/169/orig -> origin/gh/guangyey/169/orig 2025-10-10T01:18:33.6479810Z * [new branch] gh/guangyey/170/base -> origin/gh/guangyey/170/base 2025-10-10T01:18:33.6481457Z * [new branch] gh/guangyey/170/head -> origin/gh/guangyey/170/head 2025-10-10T01:18:33.6483117Z * [new branch] gh/guangyey/170/orig -> origin/gh/guangyey/170/orig 2025-10-10T01:18:33.6485908Z * [new branch] gh/guangyey/171/base -> origin/gh/guangyey/171/base 2025-10-10T01:18:33.6487671Z * [new branch] gh/guangyey/171/head -> origin/gh/guangyey/171/head 2025-10-10T01:18:33.6489375Z * [new branch] gh/guangyey/171/orig -> origin/gh/guangyey/171/orig 2025-10-10T01:18:33.6491620Z * [new branch] gh/guangyey/176/base -> origin/gh/guangyey/176/base 2025-10-10T01:18:33.6493331Z * [new branch] gh/guangyey/176/head -> origin/gh/guangyey/176/head 2025-10-10T01:18:33.6494983Z * [new branch] gh/guangyey/176/orig -> origin/gh/guangyey/176/orig 2025-10-10T01:18:33.6497208Z * [new branch] gh/guangyey/178/base -> origin/gh/guangyey/178/base 2025-10-10T01:18:33.6498845Z * [new branch] gh/guangyey/178/head -> origin/gh/guangyey/178/head 2025-10-10T01:18:33.6500549Z * [new branch] gh/guangyey/178/orig -> origin/gh/guangyey/178/orig 2025-10-10T01:18:33.6502841Z * [new branch] gh/guangyey/181/base -> origin/gh/guangyey/181/base 2025-10-10T01:18:33.6504558Z * [new branch] gh/guangyey/181/head -> origin/gh/guangyey/181/head 2025-10-10T01:18:33.6506235Z * [new branch] gh/guangyey/181/orig -> origin/gh/guangyey/181/orig 2025-10-10T01:18:33.6508480Z * [new branch] gh/guangyey/182/base -> origin/gh/guangyey/182/base 2025-10-10T01:18:33.6510209Z * [new branch] gh/guangyey/182/head -> origin/gh/guangyey/182/head 2025-10-10T01:18:33.6511871Z * [new branch] gh/guangyey/182/orig -> origin/gh/guangyey/182/orig 2025-10-10T01:18:33.6514035Z * [new branch] gh/guangyey/183/base -> origin/gh/guangyey/183/base 2025-10-10T01:18:33.6515711Z * [new branch] gh/guangyey/183/head -> origin/gh/guangyey/183/head 2025-10-10T01:18:33.6517480Z * [new branch] gh/guangyey/183/orig -> origin/gh/guangyey/183/orig 2025-10-10T01:18:33.6519892Z * [new branch] gh/guangyey/185/base -> origin/gh/guangyey/185/base 2025-10-10T01:18:33.6521838Z * [new branch] gh/guangyey/185/head -> origin/gh/guangyey/185/head 2025-10-10T01:18:33.6523486Z * [new branch] gh/guangyey/185/orig -> origin/gh/guangyey/185/orig 2025-10-10T01:18:33.6527596Z * [new branch] gh/guangyey/186/base -> origin/gh/guangyey/186/base 2025-10-10T01:18:33.6529235Z * [new branch] gh/guangyey/186/head -> origin/gh/guangyey/186/head 2025-10-10T01:18:33.6530861Z * [new branch] gh/guangyey/186/orig -> origin/gh/guangyey/186/orig 2025-10-10T01:18:33.6533232Z * [new branch] gh/guangyey/187/base -> origin/gh/guangyey/187/base 2025-10-10T01:18:33.6534972Z * [new branch] gh/guangyey/187/head -> origin/gh/guangyey/187/head 2025-10-10T01:18:33.6536682Z * [new branch] gh/guangyey/187/orig -> origin/gh/guangyey/187/orig 2025-10-10T01:18:33.6538977Z * [new branch] gh/guangyey/188/base -> origin/gh/guangyey/188/base 2025-10-10T01:18:33.6540825Z * [new branch] gh/guangyey/188/head -> origin/gh/guangyey/188/head 2025-10-10T01:18:33.6542361Z * [new branch] gh/guangyey/188/orig -> origin/gh/guangyey/188/orig 2025-10-10T01:18:33.6545548Z * [new branch] gh/guangyey/190/base -> origin/gh/guangyey/190/base 2025-10-10T01:18:33.6547256Z * [new branch] gh/guangyey/190/head -> origin/gh/guangyey/190/head 2025-10-10T01:18:33.6548928Z * [new branch] gh/guangyey/190/orig -> origin/gh/guangyey/190/orig 2025-10-10T01:18:33.6551209Z * [new branch] gh/guangyey/194/base -> origin/gh/guangyey/194/base 2025-10-10T01:18:33.6552866Z * [new branch] gh/guangyey/194/head -> origin/gh/guangyey/194/head 2025-10-10T01:18:33.6554558Z * [new branch] gh/guangyey/194/orig -> origin/gh/guangyey/194/orig 2025-10-10T01:18:33.6556868Z * [new branch] gh/guangyey/195/base -> origin/gh/guangyey/195/base 2025-10-10T01:18:33.6558752Z * [new branch] gh/guangyey/195/head -> origin/gh/guangyey/195/head 2025-10-10T01:18:33.6560464Z * [new branch] gh/guangyey/195/orig -> origin/gh/guangyey/195/orig 2025-10-10T01:18:33.6562790Z * [new branch] gh/guangyey/201/base -> origin/gh/guangyey/201/base 2025-10-10T01:18:33.6564452Z * [new branch] gh/guangyey/201/head -> origin/gh/guangyey/201/head 2025-10-10T01:18:33.6566304Z * [new branch] gh/guangyey/201/orig -> origin/gh/guangyey/201/orig 2025-10-10T01:18:33.6568698Z * [new branch] gh/guangyey/202/base -> origin/gh/guangyey/202/base 2025-10-10T01:18:33.6570351Z * [new branch] gh/guangyey/202/head -> origin/gh/guangyey/202/head 2025-10-10T01:18:33.6572067Z * [new branch] gh/guangyey/202/orig -> origin/gh/guangyey/202/orig 2025-10-10T01:18:33.6574321Z * [new branch] gh/guangyey/203/base -> origin/gh/guangyey/203/base 2025-10-10T01:18:33.6576001Z * [new branch] gh/guangyey/203/head -> origin/gh/guangyey/203/head 2025-10-10T01:18:33.6578201Z * [new branch] gh/guangyey/203/orig -> origin/gh/guangyey/203/orig 2025-10-10T01:18:33.6580285Z * [new branch] gh/guangyey/205/base -> origin/gh/guangyey/205/base 2025-10-10T01:18:33.6582886Z * [new branch] gh/guangyey/205/head -> origin/gh/guangyey/205/head 2025-10-10T01:18:33.6584234Z * [new branch] gh/guangyey/205/orig -> origin/gh/guangyey/205/orig 2025-10-10T01:18:33.6586979Z * [new branch] gh/guangyey/208/base -> origin/gh/guangyey/208/base 2025-10-10T01:18:33.6588449Z * [new branch] gh/guangyey/208/head -> origin/gh/guangyey/208/head 2025-10-10T01:18:33.6590358Z * [new branch] gh/guangyey/208/orig -> origin/gh/guangyey/208/orig 2025-10-10T01:18:33.6592628Z * [new branch] gh/guangyey/209/base -> origin/gh/guangyey/209/base 2025-10-10T01:18:33.6594679Z * [new branch] gh/guangyey/209/head -> origin/gh/guangyey/209/head 2025-10-10T01:18:33.6596632Z * [new branch] gh/guangyey/209/orig -> origin/gh/guangyey/209/orig 2025-10-10T01:18:33.6599740Z * [new branch] gh/guangyey/210/base -> origin/gh/guangyey/210/base 2025-10-10T01:18:33.6600942Z * [new branch] gh/guangyey/210/head -> origin/gh/guangyey/210/head 2025-10-10T01:18:33.6602534Z * [new branch] gh/guangyey/210/orig -> origin/gh/guangyey/210/orig 2025-10-10T01:18:33.6605845Z * [new branch] gh/guangyey/211/base -> origin/gh/guangyey/211/base 2025-10-10T01:18:33.6607255Z * [new branch] gh/guangyey/211/head -> origin/gh/guangyey/211/head 2025-10-10T01:18:33.6609132Z * [new branch] gh/guangyey/211/orig -> origin/gh/guangyey/211/orig 2025-10-10T01:18:33.6611654Z * [new branch] gh/guangyey/89/base -> origin/gh/guangyey/89/base 2025-10-10T01:18:33.6613169Z * [new branch] gh/guangyey/89/head -> origin/gh/guangyey/89/head 2025-10-10T01:18:33.6614854Z * [new branch] gh/guangyey/89/orig -> origin/gh/guangyey/89/orig 2025-10-10T01:18:33.6617743Z * [new branch] gh/guilhermeleobas/107/base -> origin/gh/guilhermeleobas/107/base 2025-10-10T01:18:33.6619431Z * [new branch] gh/guilhermeleobas/107/head -> origin/gh/guilhermeleobas/107/head 2025-10-10T01:18:33.6621042Z * [new branch] gh/guilhermeleobas/107/orig -> origin/gh/guilhermeleobas/107/orig 2025-10-10T01:18:33.6623229Z * [new branch] gh/guilhermeleobas/108/base -> origin/gh/guilhermeleobas/108/base 2025-10-10T01:18:33.6625198Z * [new branch] gh/guilhermeleobas/108/head -> origin/gh/guilhermeleobas/108/head 2025-10-10T01:18:33.6627009Z * [new branch] gh/guilhermeleobas/108/orig -> origin/gh/guilhermeleobas/108/orig 2025-10-10T01:18:33.6629340Z * [new branch] gh/guilhermeleobas/124/base -> origin/gh/guilhermeleobas/124/base 2025-10-10T01:18:33.6631028Z * [new branch] gh/guilhermeleobas/124/head -> origin/gh/guilhermeleobas/124/head 2025-10-10T01:18:33.6632779Z * [new branch] gh/guilhermeleobas/124/orig -> origin/gh/guilhermeleobas/124/orig 2025-10-10T01:18:33.6635171Z * [new branch] gh/guilhermeleobas/147/base -> origin/gh/guilhermeleobas/147/base 2025-10-10T01:18:33.6636797Z * [new branch] gh/guilhermeleobas/147/head -> origin/gh/guilhermeleobas/147/head 2025-10-10T01:18:33.6638595Z * [new branch] gh/guilhermeleobas/147/orig -> origin/gh/guilhermeleobas/147/orig 2025-10-10T01:18:33.6640880Z * [new branch] gh/guilhermeleobas/150/base -> origin/gh/guilhermeleobas/150/base 2025-10-10T01:18:33.6642532Z * [new branch] gh/guilhermeleobas/150/head -> origin/gh/guilhermeleobas/150/head 2025-10-10T01:18:33.6644168Z * [new branch] gh/guilhermeleobas/150/orig -> origin/gh/guilhermeleobas/150/orig 2025-10-10T01:18:33.6647058Z * [new branch] gh/guilhermeleobas/166/base -> origin/gh/guilhermeleobas/166/base 2025-10-10T01:18:33.6648282Z * [new branch] gh/guilhermeleobas/166/head -> origin/gh/guilhermeleobas/166/head 2025-10-10T01:18:33.6649995Z * [new branch] gh/guilhermeleobas/166/orig -> origin/gh/guilhermeleobas/166/orig 2025-10-10T01:18:33.6652897Z * [new branch] gh/guilhermeleobas/168/base -> origin/gh/guilhermeleobas/168/base 2025-10-10T01:18:33.6654094Z * [new branch] gh/guilhermeleobas/168/head -> origin/gh/guilhermeleobas/168/head 2025-10-10T01:18:33.6655910Z * [new branch] gh/guilhermeleobas/168/orig -> origin/gh/guilhermeleobas/168/orig 2025-10-10T01:18:33.6658457Z * [new branch] gh/guilhermeleobas/169/base -> origin/gh/guilhermeleobas/169/base 2025-10-10T01:18:33.6660123Z * [new branch] gh/guilhermeleobas/169/head -> origin/gh/guilhermeleobas/169/head 2025-10-10T01:18:33.6661753Z * [new branch] gh/guilhermeleobas/169/orig -> origin/gh/guilhermeleobas/169/orig 2025-10-10T01:18:33.6664560Z * [new branch] gh/guilhermeleobas/170/base -> origin/gh/guilhermeleobas/170/base 2025-10-10T01:18:33.6666250Z * [new branch] gh/guilhermeleobas/170/head -> origin/gh/guilhermeleobas/170/head 2025-10-10T01:18:33.6667863Z * [new branch] gh/guilhermeleobas/170/orig -> origin/gh/guilhermeleobas/170/orig 2025-10-10T01:18:33.6670148Z * [new branch] gh/guilhermeleobas/171/base -> origin/gh/guilhermeleobas/171/base 2025-10-10T01:18:33.6671807Z * [new branch] gh/guilhermeleobas/171/head -> origin/gh/guilhermeleobas/171/head 2025-10-10T01:18:33.6673484Z * [new branch] gh/guilhermeleobas/171/orig -> origin/gh/guilhermeleobas/171/orig 2025-10-10T01:18:33.6675942Z * [new branch] gh/guilhermeleobas/173/base -> origin/gh/guilhermeleobas/173/base 2025-10-10T01:18:33.6677477Z * [new branch] gh/guilhermeleobas/173/head -> origin/gh/guilhermeleobas/173/head 2025-10-10T01:18:33.6679173Z * [new branch] gh/guilhermeleobas/173/orig -> origin/gh/guilhermeleobas/173/orig 2025-10-10T01:18:33.6681505Z * [new branch] gh/guilhermeleobas/193/base -> origin/gh/guilhermeleobas/193/base 2025-10-10T01:18:33.6683174Z * [new branch] gh/guilhermeleobas/193/head -> origin/gh/guilhermeleobas/193/head 2025-10-10T01:18:33.6684864Z * [new branch] gh/guilhermeleobas/193/orig -> origin/gh/guilhermeleobas/193/orig 2025-10-10T01:18:33.6687102Z * [new branch] gh/guilhermeleobas/204/base -> origin/gh/guilhermeleobas/204/base 2025-10-10T01:18:33.6688718Z * [new branch] gh/guilhermeleobas/204/head -> origin/gh/guilhermeleobas/204/head 2025-10-10T01:18:33.6690441Z * [new branch] gh/guilhermeleobas/204/orig -> origin/gh/guilhermeleobas/204/orig 2025-10-10T01:18:33.6692909Z * [new branch] gh/guilhermeleobas/211/base -> origin/gh/guilhermeleobas/211/base 2025-10-10T01:18:33.6694591Z * [new branch] gh/guilhermeleobas/211/head -> origin/gh/guilhermeleobas/211/head 2025-10-10T01:18:33.6696246Z * [new branch] gh/guilhermeleobas/211/orig -> origin/gh/guilhermeleobas/211/orig 2025-10-10T01:18:33.6698642Z * [new branch] gh/guilhermeleobas/226/base -> origin/gh/guilhermeleobas/226/base 2025-10-10T01:18:33.6700215Z * [new branch] gh/guilhermeleobas/226/head -> origin/gh/guilhermeleobas/226/head 2025-10-10T01:18:33.6701918Z * [new branch] gh/guilhermeleobas/226/orig -> origin/gh/guilhermeleobas/226/orig 2025-10-10T01:18:33.6704154Z * [new branch] gh/guilhermeleobas/236/base -> origin/gh/guilhermeleobas/236/base 2025-10-10T01:18:33.6705778Z * [new branch] gh/guilhermeleobas/236/head -> origin/gh/guilhermeleobas/236/head 2025-10-10T01:18:33.6707390Z * [new branch] gh/guilhermeleobas/236/orig -> origin/gh/guilhermeleobas/236/orig 2025-10-10T01:18:33.6710150Z * [new branch] gh/guilhermeleobas/237/base -> origin/gh/guilhermeleobas/237/base 2025-10-10T01:18:33.6711814Z * [new branch] gh/guilhermeleobas/237/head -> origin/gh/guilhermeleobas/237/head 2025-10-10T01:18:33.6713466Z * [new branch] gh/guilhermeleobas/237/orig -> origin/gh/guilhermeleobas/237/orig 2025-10-10T01:18:33.6715835Z * [new branch] gh/guilhermeleobas/239/base -> origin/gh/guilhermeleobas/239/base 2025-10-10T01:18:33.6717666Z * [new branch] gh/guilhermeleobas/239/head -> origin/gh/guilhermeleobas/239/head 2025-10-10T01:18:33.6719501Z * [new branch] gh/guilhermeleobas/239/orig -> origin/gh/guilhermeleobas/239/orig 2025-10-10T01:18:33.6721935Z * [new branch] gh/guilhermeleobas/246/base -> origin/gh/guilhermeleobas/246/base 2025-10-10T01:18:33.6723720Z * [new branch] gh/guilhermeleobas/246/head -> origin/gh/guilhermeleobas/246/head 2025-10-10T01:18:33.6725802Z * [new branch] gh/guilhermeleobas/246/orig -> origin/gh/guilhermeleobas/246/orig 2025-10-10T01:18:33.6728963Z * [new branch] gh/guilhermeleobas/247/base -> origin/gh/guilhermeleobas/247/base 2025-10-10T01:18:33.6730154Z * [new branch] gh/guilhermeleobas/247/head -> origin/gh/guilhermeleobas/247/head 2025-10-10T01:18:33.6731801Z * [new branch] gh/guilhermeleobas/247/orig -> origin/gh/guilhermeleobas/247/orig 2025-10-10T01:18:33.6734492Z * [new branch] gh/guilhermeleobas/248/base -> origin/gh/guilhermeleobas/248/base 2025-10-10T01:18:33.6736017Z * [new branch] gh/guilhermeleobas/248/head -> origin/gh/guilhermeleobas/248/head 2025-10-10T01:18:33.6737793Z * [new branch] gh/guilhermeleobas/248/orig -> origin/gh/guilhermeleobas/248/orig 2025-10-10T01:18:33.6740263Z * [new branch] gh/guilhermeleobas/249/base -> origin/gh/guilhermeleobas/249/base 2025-10-10T01:18:33.6741869Z * [new branch] gh/guilhermeleobas/249/head -> origin/gh/guilhermeleobas/249/head 2025-10-10T01:18:33.6743544Z * [new branch] gh/guilhermeleobas/249/orig -> origin/gh/guilhermeleobas/249/orig 2025-10-10T01:18:33.6746013Z * [new branch] gh/guilhermeleobas/250/base -> origin/gh/guilhermeleobas/250/base 2025-10-10T01:18:33.6747795Z * [new branch] gh/guilhermeleobas/250/head -> origin/gh/guilhermeleobas/250/head 2025-10-10T01:18:33.6749529Z * [new branch] gh/guilhermeleobas/250/orig -> origin/gh/guilhermeleobas/250/orig 2025-10-10T01:18:33.6752309Z * [new branch] gh/henrylhtsang/150/base -> origin/gh/henrylhtsang/150/base 2025-10-10T01:18:33.6754028Z * [new branch] gh/henrylhtsang/150/head -> origin/gh/henrylhtsang/150/head 2025-10-10T01:18:33.6756218Z * [new branch] gh/henrylhtsang/150/orig -> origin/gh/henrylhtsang/150/orig 2025-10-10T01:18:33.6758673Z * [new branch] gh/henrylhtsang/151/base -> origin/gh/henrylhtsang/151/base 2025-10-10T01:18:33.6760263Z * [new branch] gh/henrylhtsang/151/head -> origin/gh/henrylhtsang/151/head 2025-10-10T01:18:33.6762039Z * [new branch] gh/henrylhtsang/151/orig -> origin/gh/henrylhtsang/151/orig 2025-10-10T01:18:33.6764746Z * [new branch] gh/henrylhtsang/152/base -> origin/gh/henrylhtsang/152/base 2025-10-10T01:18:33.6766235Z * [new branch] gh/henrylhtsang/152/head -> origin/gh/henrylhtsang/152/head 2025-10-10T01:18:33.6767914Z * [new branch] gh/henrylhtsang/152/orig -> origin/gh/henrylhtsang/152/orig 2025-10-10T01:18:33.6770124Z * [new branch] gh/henrylhtsang/153/base -> origin/gh/henrylhtsang/153/base 2025-10-10T01:18:33.6771808Z * [new branch] gh/henrylhtsang/153/head -> origin/gh/henrylhtsang/153/head 2025-10-10T01:18:33.6773479Z * [new branch] gh/henrylhtsang/153/orig -> origin/gh/henrylhtsang/153/orig 2025-10-10T01:18:33.6776685Z * [new branch] gh/huydhn/1/next -> origin/gh/huydhn/1/next 2025-10-10T01:18:33.6778832Z * [new branch] gh/huydhn/2/next -> origin/gh/huydhn/2/next 2025-10-10T01:18:33.6781105Z * [new branch] gh/huydhn/3/next -> origin/gh/huydhn/3/next 2025-10-10T01:18:33.6783292Z * [new branch] gh/huydhn/4/next -> origin/gh/huydhn/4/next 2025-10-10T01:18:33.6785527Z * [new branch] gh/huydhn/5/next -> origin/gh/huydhn/5/next 2025-10-10T01:18:33.6787795Z * [new branch] gh/huydhn/6/next -> origin/gh/huydhn/6/next 2025-10-10T01:18:33.6790559Z * [new branch] gh/int3/97/base -> origin/gh/int3/97/base 2025-10-10T01:18:33.6792236Z * [new branch] gh/int3/97/head -> origin/gh/int3/97/head 2025-10-10T01:18:33.6795228Z * [new branch] gh/isuruf/101/base -> origin/gh/isuruf/101/base 2025-10-10T01:18:33.6797268Z * [new branch] gh/isuruf/101/head -> origin/gh/isuruf/101/head 2025-10-10T01:18:33.6799981Z * [new branch] gh/isuruf/146/base -> origin/gh/isuruf/146/base 2025-10-10T01:18:33.6801718Z * [new branch] gh/isuruf/146/head -> origin/gh/isuruf/146/head 2025-10-10T01:18:33.6803326Z * [new branch] gh/isuruf/146/orig -> origin/gh/isuruf/146/orig 2025-10-10T01:18:33.6806163Z * [new branch] gh/isuruf/147/base -> origin/gh/isuruf/147/base 2025-10-10T01:18:33.6807461Z * [new branch] gh/isuruf/147/head -> origin/gh/isuruf/147/head 2025-10-10T01:18:33.6809503Z * [new branch] gh/isuruf/147/orig -> origin/gh/isuruf/147/orig 2025-10-10T01:18:33.6811775Z * [new branch] gh/isuruf/148/base -> origin/gh/isuruf/148/base 2025-10-10T01:18:33.6813553Z * [new branch] gh/isuruf/148/head -> origin/gh/isuruf/148/head 2025-10-10T01:18:33.6815215Z * [new branch] gh/isuruf/148/orig -> origin/gh/isuruf/148/orig 2025-10-10T01:18:33.6817364Z * [new branch] gh/isuruf/149/base -> origin/gh/isuruf/149/base 2025-10-10T01:18:33.6818998Z * [new branch] gh/isuruf/149/head -> origin/gh/isuruf/149/head 2025-10-10T01:18:33.6820647Z * [new branch] gh/isuruf/149/orig -> origin/gh/isuruf/149/orig 2025-10-10T01:18:33.6822972Z * [new branch] gh/isuruf/150/base -> origin/gh/isuruf/150/base 2025-10-10T01:18:33.6824799Z * [new branch] gh/isuruf/150/head -> origin/gh/isuruf/150/head 2025-10-10T01:18:33.6826404Z * [new branch] gh/isuruf/150/orig -> origin/gh/isuruf/150/orig 2025-10-10T01:18:33.6828538Z * [new branch] gh/isuruf/151/base -> origin/gh/isuruf/151/base 2025-10-10T01:18:33.6830250Z * [new branch] gh/isuruf/151/head -> origin/gh/isuruf/151/head 2025-10-10T01:18:33.6831984Z * [new branch] gh/isuruf/151/orig -> origin/gh/isuruf/151/orig 2025-10-10T01:18:33.6834270Z * [new branch] gh/isuruf/152/base -> origin/gh/isuruf/152/base 2025-10-10T01:18:33.6835915Z * [new branch] gh/isuruf/152/head -> origin/gh/isuruf/152/head 2025-10-10T01:18:33.6837485Z * [new branch] gh/isuruf/152/orig -> origin/gh/isuruf/152/orig 2025-10-10T01:18:33.6839896Z * [new branch] gh/isuruf/153/base -> origin/gh/isuruf/153/base 2025-10-10T01:18:33.6841516Z * [new branch] gh/isuruf/153/head -> origin/gh/isuruf/153/head 2025-10-10T01:18:33.6843137Z * [new branch] gh/isuruf/153/orig -> origin/gh/isuruf/153/orig 2025-10-10T01:18:33.6845418Z * [new branch] gh/isuruf/154/base -> origin/gh/isuruf/154/base 2025-10-10T01:18:33.6847179Z * [new branch] gh/isuruf/154/head -> origin/gh/isuruf/154/head 2025-10-10T01:18:33.6849277Z * [new branch] gh/isuruf/154/orig -> origin/gh/isuruf/154/orig 2025-10-10T01:18:33.6852009Z * [new branch] gh/isuruf/155/base -> origin/gh/isuruf/155/base 2025-10-10T01:18:33.6853632Z * [new branch] gh/isuruf/155/head -> origin/gh/isuruf/155/head 2025-10-10T01:18:33.6855345Z * [new branch] gh/isuruf/155/orig -> origin/gh/isuruf/155/orig 2025-10-10T01:18:33.6857567Z * [new branch] gh/isuruf/156/base -> origin/gh/isuruf/156/base 2025-10-10T01:18:33.6859207Z * [new branch] gh/isuruf/156/head -> origin/gh/isuruf/156/head 2025-10-10T01:18:33.6860927Z * [new branch] gh/isuruf/156/orig -> origin/gh/isuruf/156/orig 2025-10-10T01:18:33.6863172Z * [new branch] gh/isuruf/157/base -> origin/gh/isuruf/157/base 2025-10-10T01:18:33.6865087Z * [new branch] gh/isuruf/157/head -> origin/gh/isuruf/157/head 2025-10-10T01:18:33.6866650Z * [new branch] gh/isuruf/157/orig -> origin/gh/isuruf/157/orig 2025-10-10T01:18:33.6868930Z * [new branch] gh/isuruf/81/base -> origin/gh/isuruf/81/base 2025-10-10T01:18:33.6870537Z * [new branch] gh/isuruf/81/head -> origin/gh/isuruf/81/head 2025-10-10T01:18:33.6872209Z * [new branch] gh/isuruf/81/orig -> origin/gh/isuruf/81/orig 2025-10-10T01:18:33.6874955Z * [new branch] gh/jamesjwu/171/base -> origin/gh/jamesjwu/171/base 2025-10-10T01:18:33.6876691Z * [new branch] gh/jamesjwu/171/head -> origin/gh/jamesjwu/171/head 2025-10-10T01:18:33.6878384Z * [new branch] gh/jamesjwu/171/orig -> origin/gh/jamesjwu/171/orig 2025-10-10T01:18:33.6880790Z * [new branch] gh/jamesjwu/176/base -> origin/gh/jamesjwu/176/base 2025-10-10T01:18:33.6882603Z * [new branch] gh/jamesjwu/176/head -> origin/gh/jamesjwu/176/head 2025-10-10T01:18:33.6884123Z * [new branch] gh/jamesjwu/176/orig -> origin/gh/jamesjwu/176/orig 2025-10-10T01:18:33.6886887Z * [new branch] gh/jamesjwu/186/base -> origin/gh/jamesjwu/186/base 2025-10-10T01:18:33.6888108Z * [new branch] gh/jamesjwu/186/head -> origin/gh/jamesjwu/186/head 2025-10-10T01:18:33.6889762Z * [new branch] gh/jamesjwu/186/orig -> origin/gh/jamesjwu/186/orig 2025-10-10T01:18:33.6892480Z * [new branch] gh/jamesjwu/187/base -> origin/gh/jamesjwu/187/base 2025-10-10T01:18:33.6893677Z * [new branch] gh/jamesjwu/187/head -> origin/gh/jamesjwu/187/head 2025-10-10T01:18:33.6895752Z * [new branch] gh/jamesjwu/187/orig -> origin/gh/jamesjwu/187/orig 2025-10-10T01:18:33.6898525Z * [new branch] gh/jamesjwu/189/base -> origin/gh/jamesjwu/189/base 2025-10-10T01:18:33.6900295Z * [new branch] gh/jamesjwu/189/head -> origin/gh/jamesjwu/189/head 2025-10-10T01:18:33.6901994Z * [new branch] gh/jamesjwu/189/orig -> origin/gh/jamesjwu/189/orig 2025-10-10T01:18:33.6904278Z * [new branch] gh/jamesjwu/190/base -> origin/gh/jamesjwu/190/base 2025-10-10T01:18:33.6906769Z * [new branch] gh/jamesjwu/190/head -> origin/gh/jamesjwu/190/head 2025-10-10T01:18:33.6907982Z * [new branch] gh/jamesjwu/190/orig -> origin/gh/jamesjwu/190/orig 2025-10-10T01:18:33.6910536Z * [new branch] gh/jamesjwu/191/base -> origin/gh/jamesjwu/191/base 2025-10-10T01:18:33.6911838Z * [new branch] gh/jamesjwu/191/head -> origin/gh/jamesjwu/191/head 2025-10-10T01:18:33.6913584Z * [new branch] gh/jamesjwu/191/orig -> origin/gh/jamesjwu/191/orig 2025-10-10T01:18:33.6916672Z * [new branch] gh/jamesjwu/192/base -> origin/gh/jamesjwu/192/base 2025-10-10T01:18:33.6917994Z * [new branch] gh/jamesjwu/192/head -> origin/gh/jamesjwu/192/head 2025-10-10T01:18:33.6920427Z * [new branch] gh/jamesjwu/193/base -> origin/gh/jamesjwu/193/base 2025-10-10T01:18:33.6922109Z * [new branch] gh/jamesjwu/193/head -> origin/gh/jamesjwu/193/head 2025-10-10T01:18:33.6923905Z * [new branch] gh/jamesjwu/193/orig -> origin/gh/jamesjwu/193/orig 2025-10-10T01:18:33.6927550Z * [new branch] gh/jamesjwu/194/base -> origin/gh/jamesjwu/194/base 2025-10-10T01:18:33.6929246Z * [new branch] gh/jamesjwu/194/head -> origin/gh/jamesjwu/194/head 2025-10-10T01:18:33.6931021Z * [new branch] gh/jamesjwu/194/orig -> origin/gh/jamesjwu/194/orig 2025-10-10T01:18:33.6933195Z * [new branch] gh/jamesjwu/195/base -> origin/gh/jamesjwu/195/base 2025-10-10T01:18:33.6934808Z * [new branch] gh/jamesjwu/195/head -> origin/gh/jamesjwu/195/head 2025-10-10T01:18:33.6936495Z * [new branch] gh/jamesjwu/195/orig -> origin/gh/jamesjwu/195/orig 2025-10-10T01:18:33.6938806Z * [new branch] gh/jamesjwu/196/base -> origin/gh/jamesjwu/196/base 2025-10-10T01:18:33.6940417Z * [new branch] gh/jamesjwu/196/head -> origin/gh/jamesjwu/196/head 2025-10-10T01:18:33.6942184Z * [new branch] gh/jamesjwu/196/orig -> origin/gh/jamesjwu/196/orig 2025-10-10T01:18:33.6944520Z * [new branch] gh/jamesjwu/52/base -> origin/gh/jamesjwu/52/base 2025-10-10T01:18:33.6946551Z * [new branch] gh/jamesjwu/52/head -> origin/gh/jamesjwu/52/head 2025-10-10T01:18:33.6949515Z * [new branch] gh/jamesjwu/53/base -> origin/gh/jamesjwu/53/base 2025-10-10T01:18:33.6950577Z * [new branch] gh/jamesjwu/53/head -> origin/gh/jamesjwu/53/head 2025-10-10T01:18:33.6953009Z * [new branch] gh/jamesjwu/54/base -> origin/gh/jamesjwu/54/base 2025-10-10T01:18:33.6954286Z * [new branch] gh/jamesjwu/54/head -> origin/gh/jamesjwu/54/head 2025-10-10T01:18:33.6956682Z * [new branch] gh/jamesjwu/55/base -> origin/gh/jamesjwu/55/base 2025-10-10T01:18:33.6958845Z * [new branch] gh/jamesjwu/55/head -> origin/gh/jamesjwu/55/head 2025-10-10T01:18:33.6961055Z * [new branch] gh/jamesjwu/56/base -> origin/gh/jamesjwu/56/base 2025-10-10T01:18:33.6962624Z * [new branch] gh/jamesjwu/56/head -> origin/gh/jamesjwu/56/head 2025-10-10T01:18:33.6965329Z * [new branch] gh/jamesjwu/57/base -> origin/gh/jamesjwu/57/base 2025-10-10T01:18:33.6967967Z * [new branch] gh/jamesjwu/57/head -> origin/gh/jamesjwu/57/head 2025-10-10T01:18:33.6969685Z * [new branch] gh/jamesjwu/58/base -> origin/gh/jamesjwu/58/base 2025-10-10T01:18:33.6971274Z * [new branch] gh/jamesjwu/58/head -> origin/gh/jamesjwu/58/head 2025-10-10T01:18:33.6974023Z * [new branch] gh/jamesjwu/59/base -> origin/gh/jamesjwu/59/base 2025-10-10T01:18:33.6975208Z * [new branch] gh/jamesjwu/59/head -> origin/gh/jamesjwu/59/head 2025-10-10T01:18:33.6977642Z * [new branch] gh/jamesjwu/60/base -> origin/gh/jamesjwu/60/base 2025-10-10T01:18:33.6979294Z * [new branch] gh/jamesjwu/60/head -> origin/gh/jamesjwu/60/head 2025-10-10T01:18:33.6981436Z * [new branch] gh/jamesjwu/61/base -> origin/gh/jamesjwu/61/base 2025-10-10T01:18:33.6983154Z * [new branch] gh/jamesjwu/61/head -> origin/gh/jamesjwu/61/head 2025-10-10T01:18:33.6985419Z * [new branch] gh/jamesjwu/62/base -> origin/gh/jamesjwu/62/base 2025-10-10T01:18:33.6986986Z * [new branch] gh/jamesjwu/62/head -> origin/gh/jamesjwu/62/head 2025-10-10T01:18:33.6989158Z * [new branch] gh/jamesjwu/63/base -> origin/gh/jamesjwu/63/base 2025-10-10T01:18:33.6990890Z * [new branch] gh/jamesjwu/63/head -> origin/gh/jamesjwu/63/head 2025-10-10T01:18:33.6993336Z * [new branch] gh/jamesjwu/64/base -> origin/gh/jamesjwu/64/base 2025-10-10T01:18:33.6994993Z * [new branch] gh/jamesjwu/64/head -> origin/gh/jamesjwu/64/head 2025-10-10T01:18:33.6997357Z * [new branch] gh/jamesjwu/65/base -> origin/gh/jamesjwu/65/base 2025-10-10T01:18:33.6999098Z * [new branch] gh/jamesjwu/65/head -> origin/gh/jamesjwu/65/head 2025-10-10T01:18:33.7002175Z * [new branch] gh/janeyx99/165/base -> origin/gh/janeyx99/165/base 2025-10-10T01:18:33.7003701Z * [new branch] gh/janeyx99/165/head -> origin/gh/janeyx99/165/head 2025-10-10T01:18:33.7005437Z * [new branch] gh/janeyx99/165/orig -> origin/gh/janeyx99/165/orig 2025-10-10T01:18:33.7008588Z * [new branch] gh/janeyx99/201/base -> origin/gh/janeyx99/201/base 2025-10-10T01:18:33.7009596Z * [new branch] gh/janeyx99/201/head -> origin/gh/janeyx99/201/head 2025-10-10T01:18:33.7011213Z * [new branch] gh/janeyx99/201/orig -> origin/gh/janeyx99/201/orig 2025-10-10T01:18:33.7014270Z * [new branch] gh/janeyx99/225/base -> origin/gh/janeyx99/225/base 2025-10-10T01:18:33.7015625Z * [new branch] gh/janeyx99/225/head -> origin/gh/janeyx99/225/head 2025-10-10T01:18:33.7017372Z * [new branch] gh/janeyx99/225/orig -> origin/gh/janeyx99/225/orig 2025-10-10T01:18:33.7020102Z * [new branch] gh/janeyx99/299/base -> origin/gh/janeyx99/299/base 2025-10-10T01:18:33.7021528Z * [new branch] gh/janeyx99/299/head -> origin/gh/janeyx99/299/head 2025-10-10T01:18:33.7023365Z * [new branch] gh/janeyx99/299/orig -> origin/gh/janeyx99/299/orig 2025-10-10T01:18:33.7025732Z * [new branch] gh/janeyx99/302/base -> origin/gh/janeyx99/302/base 2025-10-10T01:18:33.7027272Z * [new branch] gh/janeyx99/302/head -> origin/gh/janeyx99/302/head 2025-10-10T01:18:33.7029363Z * [new branch] gh/janeyx99/303/base -> origin/gh/janeyx99/303/base 2025-10-10T01:18:33.7030953Z * [new branch] gh/janeyx99/303/head -> origin/gh/janeyx99/303/head 2025-10-10T01:18:33.7033256Z * [new branch] gh/janeyx99/304/base -> origin/gh/janeyx99/304/base 2025-10-10T01:18:33.7034992Z * [new branch] gh/janeyx99/304/head -> origin/gh/janeyx99/304/head 2025-10-10T01:18:33.7036626Z * [new branch] gh/janeyx99/304/orig -> origin/gh/janeyx99/304/orig 2025-10-10T01:18:33.7038850Z * [new branch] gh/janeyx99/305/base -> origin/gh/janeyx99/305/base 2025-10-10T01:18:33.7040699Z * [new branch] gh/janeyx99/305/head -> origin/gh/janeyx99/305/head 2025-10-10T01:18:33.7042770Z * [new branch] gh/janeyx99/306/base -> origin/gh/janeyx99/306/base 2025-10-10T01:18:33.7044397Z * [new branch] gh/janeyx99/306/head -> origin/gh/janeyx99/306/head 2025-10-10T01:18:33.7046711Z * [new branch] gh/janeyx99/307/base -> origin/gh/janeyx99/307/base 2025-10-10T01:18:33.7048358Z * [new branch] gh/janeyx99/307/head -> origin/gh/janeyx99/307/head 2025-10-10T01:18:33.7049954Z * [new branch] gh/janeyx99/307/orig -> origin/gh/janeyx99/307/orig 2025-10-10T01:18:33.7052109Z * [new branch] gh/janeyx99/308/base -> origin/gh/janeyx99/308/base 2025-10-10T01:18:33.7053825Z * [new branch] gh/janeyx99/308/head -> origin/gh/janeyx99/308/head 2025-10-10T01:18:33.7055503Z * [new branch] gh/janeyx99/308/orig -> origin/gh/janeyx99/308/orig 2025-10-10T01:18:33.7057887Z * [new branch] gh/janeyx99/309/base -> origin/gh/janeyx99/309/base 2025-10-10T01:18:33.7059377Z * [new branch] gh/janeyx99/309/head -> origin/gh/janeyx99/309/head 2025-10-10T01:18:33.7061042Z * [new branch] gh/janeyx99/309/orig -> origin/gh/janeyx99/309/orig 2025-10-10T01:18:33.7063491Z * [new branch] gh/janeyx99/310/base -> origin/gh/janeyx99/310/base 2025-10-10T01:18:33.7065171Z * [new branch] gh/janeyx99/310/head -> origin/gh/janeyx99/310/head 2025-10-10T01:18:33.7066844Z * [new branch] gh/janeyx99/310/orig -> origin/gh/janeyx99/310/orig 2025-10-10T01:18:33.7068974Z * [new branch] gh/janeyx99/311/base -> origin/gh/janeyx99/311/base 2025-10-10T01:18:33.7070694Z * [new branch] gh/janeyx99/311/head -> origin/gh/janeyx99/311/head 2025-10-10T01:18:33.7072463Z * [new branch] gh/janeyx99/311/orig -> origin/gh/janeyx99/311/orig 2025-10-10T01:18:33.7074533Z * [new branch] gh/janeyx99/312/base -> origin/gh/janeyx99/312/base 2025-10-10T01:18:33.7076192Z * [new branch] gh/janeyx99/312/head -> origin/gh/janeyx99/312/head 2025-10-10T01:18:33.7077870Z * [new branch] gh/janeyx99/312/orig -> origin/gh/janeyx99/312/orig 2025-10-10T01:18:33.7080199Z * [new branch] gh/janeyx99/313/base -> origin/gh/janeyx99/313/base 2025-10-10T01:18:33.7081857Z * [new branch] gh/janeyx99/313/head -> origin/gh/janeyx99/313/head 2025-10-10T01:18:33.7083489Z * [new branch] gh/janeyx99/313/orig -> origin/gh/janeyx99/313/orig 2025-10-10T01:18:33.7086498Z * [new branch] gh/janeyx99/314/base -> origin/gh/janeyx99/314/base 2025-10-10T01:18:33.7088068Z * [new branch] gh/janeyx99/314/head -> origin/gh/janeyx99/314/head 2025-10-10T01:18:33.7089869Z * [new branch] gh/janeyx99/314/orig -> origin/gh/janeyx99/314/orig 2025-10-10T01:18:33.7092264Z * [new branch] gh/janeyx99/88/base -> origin/gh/janeyx99/88/base 2025-10-10T01:18:33.7095991Z * [new branch] gh/janeyx99/88/head -> origin/gh/janeyx99/88/head 2025-10-10T01:18:33.7097392Z * [new branch] gh/janeyx99/88/orig -> origin/gh/janeyx99/88/orig 2025-10-10T01:18:33.7099466Z * [new branch] gh/jansel/360/base -> origin/gh/jansel/360/base 2025-10-10T01:18:33.7100250Z * [new branch] gh/jansel/360/head -> origin/gh/jansel/360/head 2025-10-10T01:18:33.7103205Z * [new branch] gh/jansel/451/base -> origin/gh/jansel/451/base 2025-10-10T01:18:33.7104352Z * [new branch] gh/jansel/451/head -> origin/gh/jansel/451/head 2025-10-10T01:18:33.7106224Z * [new branch] gh/jansel/451/orig -> origin/gh/jansel/451/orig 2025-10-10T01:18:33.7108665Z * [new branch] gh/jansel/462/base -> origin/gh/jansel/462/base 2025-10-10T01:18:33.7110273Z * [new branch] gh/jansel/462/head -> origin/gh/jansel/462/head 2025-10-10T01:18:33.7111907Z * [new branch] gh/jansel/462/orig -> origin/gh/jansel/462/orig 2025-10-10T01:18:33.7114181Z * [new branch] gh/jansel/531/base -> origin/gh/jansel/531/base 2025-10-10T01:18:33.7115876Z * [new branch] gh/jansel/531/head -> origin/gh/jansel/531/head 2025-10-10T01:18:33.7117894Z * [new branch] gh/jansel/531/orig -> origin/gh/jansel/531/orig 2025-10-10T01:18:33.7120252Z * [new branch] gh/jansel/532/base -> origin/gh/jansel/532/base 2025-10-10T01:18:33.7121915Z * [new branch] gh/jansel/532/head -> origin/gh/jansel/532/head 2025-10-10T01:18:33.7123599Z * [new branch] gh/jansel/532/orig -> origin/gh/jansel/532/orig 2025-10-10T01:18:33.7126226Z * [new branch] gh/jansel/533/base -> origin/gh/jansel/533/base 2025-10-10T01:18:33.7127749Z * [new branch] gh/jansel/533/head -> origin/gh/jansel/533/head 2025-10-10T01:18:33.7129752Z * [new branch] gh/jansel/533/orig -> origin/gh/jansel/533/orig 2025-10-10T01:18:33.7131954Z * [new branch] gh/jansel/534/base -> origin/gh/jansel/534/base 2025-10-10T01:18:33.7133530Z * [new branch] gh/jansel/534/head -> origin/gh/jansel/534/head 2025-10-10T01:18:33.7135152Z * [new branch] gh/jansel/534/orig -> origin/gh/jansel/534/orig 2025-10-10T01:18:33.7137455Z * [new branch] gh/jansel/535/base -> origin/gh/jansel/535/base 2025-10-10T01:18:33.7139070Z * [new branch] gh/jansel/535/head -> origin/gh/jansel/535/head 2025-10-10T01:18:33.7140892Z * [new branch] gh/jansel/535/orig -> origin/gh/jansel/535/orig 2025-10-10T01:18:33.7142993Z * [new branch] gh/jansel/536/base -> origin/gh/jansel/536/base 2025-10-10T01:18:33.7144888Z * [new branch] gh/jansel/536/head -> origin/gh/jansel/536/head 2025-10-10T01:18:33.7146343Z * [new branch] gh/jansel/536/orig -> origin/gh/jansel/536/orig 2025-10-10T01:18:33.7148676Z * [new branch] gh/jansel/537/base -> origin/gh/jansel/537/base 2025-10-10T01:18:33.7150330Z * [new branch] gh/jansel/537/head -> origin/gh/jansel/537/head 2025-10-10T01:18:33.7151982Z * [new branch] gh/jansel/537/orig -> origin/gh/jansel/537/orig 2025-10-10T01:18:33.7154269Z * [new branch] gh/jansel/538/base -> origin/gh/jansel/538/base 2025-10-10T01:18:33.7155914Z * [new branch] gh/jansel/538/head -> origin/gh/jansel/538/head 2025-10-10T01:18:33.7157656Z * [new branch] gh/jansel/538/orig -> origin/gh/jansel/538/orig 2025-10-10T01:18:33.7159956Z * [new branch] gh/jansel/539/base -> origin/gh/jansel/539/base 2025-10-10T01:18:33.7161660Z * [new branch] gh/jansel/539/head -> origin/gh/jansel/539/head 2025-10-10T01:18:33.7163295Z * [new branch] gh/jansel/539/orig -> origin/gh/jansel/539/orig 2025-10-10T01:18:33.7165571Z * [new branch] gh/jansel/540/base -> origin/gh/jansel/540/base 2025-10-10T01:18:33.7167275Z * [new branch] gh/jansel/540/head -> origin/gh/jansel/540/head 2025-10-10T01:18:33.7168925Z * [new branch] gh/jansel/540/orig -> origin/gh/jansel/540/orig 2025-10-10T01:18:33.7171198Z * [new branch] gh/jansel/541/base -> origin/gh/jansel/541/base 2025-10-10T01:18:33.7172891Z * [new branch] gh/jansel/541/head -> origin/gh/jansel/541/head 2025-10-10T01:18:33.7174571Z * [new branch] gh/jansel/541/orig -> origin/gh/jansel/541/orig 2025-10-10T01:18:33.7176810Z * [new branch] gh/jansel/542/base -> origin/gh/jansel/542/base 2025-10-10T01:18:33.7178454Z * [new branch] gh/jansel/542/head -> origin/gh/jansel/542/head 2025-10-10T01:18:33.7180160Z * [new branch] gh/jansel/542/orig -> origin/gh/jansel/542/orig 2025-10-10T01:18:33.7182341Z * [new branch] gh/jansel/543/base -> origin/gh/jansel/543/base 2025-10-10T01:18:33.7184024Z * [new branch] gh/jansel/543/head -> origin/gh/jansel/543/head 2025-10-10T01:18:33.7186070Z * [new branch] gh/jansel/543/orig -> origin/gh/jansel/543/orig 2025-10-10T01:18:33.7189116Z * [new branch] gh/jansel/544/base -> origin/gh/jansel/544/base 2025-10-10T01:18:33.7190468Z * [new branch] gh/jansel/544/head -> origin/gh/jansel/544/head 2025-10-10T01:18:33.7192055Z * [new branch] gh/jansel/544/orig -> origin/gh/jansel/544/orig 2025-10-10T01:18:33.7195612Z * [new branch] gh/jansel/545/base -> origin/gh/jansel/545/base 2025-10-10T01:18:33.7196720Z * [new branch] gh/jansel/545/head -> origin/gh/jansel/545/head 2025-10-10T01:18:33.7198640Z * [new branch] gh/jansel/545/orig -> origin/gh/jansel/545/orig 2025-10-10T01:18:33.7200942Z * [new branch] gh/jansel/546/base -> origin/gh/jansel/546/base 2025-10-10T01:18:33.7202609Z * [new branch] gh/jansel/546/head -> origin/gh/jansel/546/head 2025-10-10T01:18:33.7204269Z * [new branch] gh/jansel/546/orig -> origin/gh/jansel/546/orig 2025-10-10T01:18:33.7206968Z * [new branch] gh/jansel/547/base -> origin/gh/jansel/547/base 2025-10-10T01:18:33.7208941Z * [new branch] gh/jansel/547/head -> origin/gh/jansel/547/head 2025-10-10T01:18:33.7210518Z * [new branch] gh/jansel/547/orig -> origin/gh/jansel/547/orig 2025-10-10T01:18:33.7213082Z * [new branch] gh/jansel/548/base -> origin/gh/jansel/548/base 2025-10-10T01:18:33.7214495Z * [new branch] gh/jansel/548/head -> origin/gh/jansel/548/head 2025-10-10T01:18:33.7216347Z * [new branch] gh/jansel/548/orig -> origin/gh/jansel/548/orig 2025-10-10T01:18:33.7219267Z * [new branch] gh/jbschlosser/247/base -> origin/gh/jbschlosser/247/base 2025-10-10T01:18:33.7220939Z * [new branch] gh/jbschlosser/247/head -> origin/gh/jbschlosser/247/head 2025-10-10T01:18:33.7222614Z * [new branch] gh/jbschlosser/247/orig -> origin/gh/jbschlosser/247/orig 2025-10-10T01:18:33.7226099Z * [new branch] gh/jbschlosser/250/base -> origin/gh/jbschlosser/250/base 2025-10-10T01:18:33.7227768Z * [new branch] gh/jbschlosser/250/head -> origin/gh/jbschlosser/250/head 2025-10-10T01:18:33.7229678Z * [new branch] gh/jbschlosser/250/orig -> origin/gh/jbschlosser/250/orig 2025-10-10T01:18:33.7231788Z * [new branch] gh/jbschlosser/251/base -> origin/gh/jbschlosser/251/base 2025-10-10T01:18:33.7233363Z * [new branch] gh/jbschlosser/251/head -> origin/gh/jbschlosser/251/head 2025-10-10T01:18:33.7235012Z * [new branch] gh/jbschlosser/251/orig -> origin/gh/jbschlosser/251/orig 2025-10-10T01:18:33.7238183Z * [new branch] gh/jiayisunx/59/base -> origin/gh/jiayisunx/59/base 2025-10-10T01:18:33.7240083Z * [new branch] gh/jiayisunx/59/head -> origin/gh/jiayisunx/59/head 2025-10-10T01:18:33.7241669Z * [new branch] gh/jiayisunx/59/orig -> origin/gh/jiayisunx/59/orig 2025-10-10T01:18:33.7243918Z * [new branch] gh/jiayisunx/61/base -> origin/gh/jiayisunx/61/base 2025-10-10T01:18:33.7246053Z * [new branch] gh/jiayisunx/61/head -> origin/gh/jiayisunx/61/head 2025-10-10T01:18:33.7247266Z * [new branch] gh/jiayisunx/61/orig -> origin/gh/jiayisunx/61/orig 2025-10-10T01:18:33.7249810Z * [new branch] gh/jiayisunx/65/base -> origin/gh/jiayisunx/65/base 2025-10-10T01:18:33.7251219Z * [new branch] gh/jiayisunx/65/head -> origin/gh/jiayisunx/65/head 2025-10-10T01:18:33.7252864Z * [new branch] gh/jiayisunx/65/orig -> origin/gh/jiayisunx/65/orig 2025-10-10T01:18:33.7255444Z * [new branch] gh/jiayisunx/67/base -> origin/gh/jiayisunx/67/base 2025-10-10T01:18:33.7256631Z * [new branch] gh/jiayisunx/67/head -> origin/gh/jiayisunx/67/head 2025-10-10T01:18:33.7258408Z * [new branch] gh/jiayisunx/67/orig -> origin/gh/jiayisunx/67/orig 2025-10-10T01:18:33.7260641Z * [new branch] gh/jiayisunx/68/base -> origin/gh/jiayisunx/68/base 2025-10-10T01:18:33.7262474Z * [new branch] gh/jiayisunx/68/head -> origin/gh/jiayisunx/68/head 2025-10-10T01:18:33.7264100Z * [new branch] gh/jiayisunx/68/orig -> origin/gh/jiayisunx/68/orig 2025-10-10T01:18:33.7266307Z * [new branch] gh/jiayisunx/71/base -> origin/gh/jiayisunx/71/base 2025-10-10T01:18:33.7268064Z * [new branch] gh/jiayisunx/71/head -> origin/gh/jiayisunx/71/head 2025-10-10T01:18:33.7269754Z * [new branch] gh/jiayisunx/71/orig -> origin/gh/jiayisunx/71/orig 2025-10-10T01:18:33.7272011Z * [new branch] gh/jiayisunx/72/base -> origin/gh/jiayisunx/72/base 2025-10-10T01:18:33.7273671Z * [new branch] gh/jiayisunx/72/head -> origin/gh/jiayisunx/72/head 2025-10-10T01:18:33.7275389Z * [new branch] gh/jiayisunx/72/orig -> origin/gh/jiayisunx/72/orig 2025-10-10T01:18:33.7277631Z * [new branch] gh/jiayisunx/77/base -> origin/gh/jiayisunx/77/base 2025-10-10T01:18:33.7279422Z * [new branch] gh/jiayisunx/77/head -> origin/gh/jiayisunx/77/head 2025-10-10T01:18:33.7281038Z * [new branch] gh/jiayisunx/77/orig -> origin/gh/jiayisunx/77/orig 2025-10-10T01:18:33.7283256Z * [new branch] gh/jiayisunx/78/base -> origin/gh/jiayisunx/78/base 2025-10-10T01:18:33.7285470Z * [new branch] gh/jiayisunx/78/head -> origin/gh/jiayisunx/78/head 2025-10-10T01:18:33.7287267Z * [new branch] gh/jiayisunx/78/orig -> origin/gh/jiayisunx/78/orig 2025-10-10T01:18:33.7289530Z * [new branch] gh/jiayisunx/79/base -> origin/gh/jiayisunx/79/base 2025-10-10T01:18:33.7291214Z * [new branch] gh/jiayisunx/79/head -> origin/gh/jiayisunx/79/head 2025-10-10T01:18:33.7292855Z * [new branch] gh/jiayisunx/79/orig -> origin/gh/jiayisunx/79/orig 2025-10-10T01:18:33.7295155Z * [new branch] gh/jiayisunx/80/base -> origin/gh/jiayisunx/80/base 2025-10-10T01:18:33.7297431Z * [new branch] gh/jiayisunx/80/head -> origin/gh/jiayisunx/80/head 2025-10-10T01:18:33.7299094Z * [new branch] gh/jiayisunx/80/orig -> origin/gh/jiayisunx/80/orig 2025-10-10T01:18:33.7301750Z * [new branch] gh/jiayisunx/81/base -> origin/gh/jiayisunx/81/base 2025-10-10T01:18:33.7303398Z * [new branch] gh/jiayisunx/81/head -> origin/gh/jiayisunx/81/head 2025-10-10T01:18:33.7305173Z * [new branch] gh/jiayisunx/81/orig -> origin/gh/jiayisunx/81/orig 2025-10-10T01:18:33.7307371Z * [new branch] gh/jiayisunx/82/base -> origin/gh/jiayisunx/82/base 2025-10-10T01:18:33.7309253Z * [new branch] gh/jiayisunx/82/head -> origin/gh/jiayisunx/82/head 2025-10-10T01:18:33.7311109Z * [new branch] gh/jiayisunx/82/orig -> origin/gh/jiayisunx/82/orig 2025-10-10T01:18:33.7313196Z * [new branch] gh/jiayisunx/83/base -> origin/gh/jiayisunx/83/base 2025-10-10T01:18:33.7315283Z * [new branch] gh/jiayisunx/83/head -> origin/gh/jiayisunx/83/head 2025-10-10T01:18:33.7316995Z * [new branch] gh/jiayisunx/83/orig -> origin/gh/jiayisunx/83/orig 2025-10-10T01:18:33.7319337Z * [new branch] gh/jiayisunx/84/base -> origin/gh/jiayisunx/84/base 2025-10-10T01:18:33.7320965Z * [new branch] gh/jiayisunx/84/head -> origin/gh/jiayisunx/84/head 2025-10-10T01:18:33.7322631Z * [new branch] gh/jiayisunx/84/orig -> origin/gh/jiayisunx/84/orig 2025-10-10T01:18:33.7326366Z * [new branch] gh/jjwu@meta.com/1/base -> origin/gh/jjwu@meta.com/1/base 2025-10-10T01:18:33.7328237Z * [new branch] gh/jjwu@meta.com/1/head -> origin/gh/jjwu@meta.com/1/head 2025-10-10T01:18:33.7330892Z * [new branch] gh/karthickai/3/base -> origin/gh/karthickai/3/base 2025-10-10T01:18:33.7332571Z * [new branch] gh/karthickai/3/head -> origin/gh/karthickai/3/head 2025-10-10T01:18:33.7334203Z * [new branch] gh/karthickai/3/orig -> origin/gh/karthickai/3/orig 2025-10-10T01:18:33.7336784Z * [new branch] gh/karthickai/4/base -> origin/gh/karthickai/4/base 2025-10-10T01:18:33.7338924Z * [new branch] gh/karthickai/4/head -> origin/gh/karthickai/4/head 2025-10-10T01:18:33.7340419Z * [new branch] gh/karthickai/4/orig -> origin/gh/karthickai/4/orig 2025-10-10T01:18:33.7343003Z * [new branch] gh/karthickai/5/base -> origin/gh/karthickai/5/base 2025-10-10T01:18:33.7344376Z * [new branch] gh/karthickai/5/head -> origin/gh/karthickai/5/head 2025-10-10T01:18:33.7346219Z * [new branch] gh/karthickai/5/orig -> origin/gh/karthickai/5/orig 2025-10-10T01:18:33.7348708Z * [new branch] gh/karthickai/6/base -> origin/gh/karthickai/6/base 2025-10-10T01:18:33.7350502Z * [new branch] gh/karthickai/6/head -> origin/gh/karthickai/6/head 2025-10-10T01:18:33.7352177Z * [new branch] gh/karthickai/6/orig -> origin/gh/karthickai/6/orig 2025-10-10T01:18:33.7354989Z * [new branch] gh/kurtamohler/32/base -> origin/gh/kurtamohler/32/base 2025-10-10T01:18:33.7356584Z * [new branch] gh/kurtamohler/32/head -> origin/gh/kurtamohler/32/head 2025-10-10T01:18:33.7358240Z * [new branch] gh/kurtamohler/32/orig -> origin/gh/kurtamohler/32/orig 2025-10-10T01:18:33.7360629Z * [new branch] gh/kurtamohler/33/base -> origin/gh/kurtamohler/33/base 2025-10-10T01:18:33.7362282Z * [new branch] gh/kurtamohler/33/head -> origin/gh/kurtamohler/33/head 2025-10-10T01:18:33.7363946Z * [new branch] gh/kurtamohler/33/orig -> origin/gh/kurtamohler/33/orig 2025-10-10T01:18:33.7366292Z * [new branch] gh/kurtamohler/34/base -> origin/gh/kurtamohler/34/base 2025-10-10T01:18:33.7368133Z * [new branch] gh/kurtamohler/34/head -> origin/gh/kurtamohler/34/head 2025-10-10T01:18:33.7369661Z * [new branch] gh/kurtamohler/34/orig -> origin/gh/kurtamohler/34/orig 2025-10-10T01:18:33.7371861Z * [new branch] gh/kurtamohler/51/base -> origin/gh/kurtamohler/51/base 2025-10-10T01:18:33.7373519Z * [new branch] gh/kurtamohler/51/head -> origin/gh/kurtamohler/51/head 2025-10-10T01:18:33.7375146Z * [new branch] gh/kurtamohler/51/orig -> origin/gh/kurtamohler/51/orig 2025-10-10T01:18:33.7377491Z * [new branch] gh/kurtamohler/52/base -> origin/gh/kurtamohler/52/base 2025-10-10T01:18:33.7379155Z * [new branch] gh/kurtamohler/52/head -> origin/gh/kurtamohler/52/head 2025-10-10T01:18:33.7380840Z * [new branch] gh/kurtamohler/52/orig -> origin/gh/kurtamohler/52/orig 2025-10-10T01:18:33.7383235Z * [new branch] gh/kurtamohler/53/base -> origin/gh/kurtamohler/53/base 2025-10-10T01:18:33.7384873Z * [new branch] gh/kurtamohler/53/head -> origin/gh/kurtamohler/53/head 2025-10-10T01:18:33.7386554Z * [new branch] gh/kurtamohler/53/orig -> origin/gh/kurtamohler/53/orig 2025-10-10T01:18:33.7388892Z * [new branch] gh/kurtamohler/54/base -> origin/gh/kurtamohler/54/base 2025-10-10T01:18:33.7390585Z * [new branch] gh/kurtamohler/54/head -> origin/gh/kurtamohler/54/head 2025-10-10T01:18:33.7392227Z * [new branch] gh/kurtamohler/54/orig -> origin/gh/kurtamohler/54/orig 2025-10-10T01:18:33.7394556Z * [new branch] gh/kurtamohler/55/base -> origin/gh/kurtamohler/55/base 2025-10-10T01:18:33.7396394Z * [new branch] gh/kurtamohler/55/head -> origin/gh/kurtamohler/55/head 2025-10-10T01:18:33.7398045Z * [new branch] gh/kurtamohler/55/orig -> origin/gh/kurtamohler/55/orig 2025-10-10T01:18:33.7401134Z * [new branch] gh/kwen2501/130/base -> origin/gh/kwen2501/130/base 2025-10-10T01:18:33.7402872Z * [new branch] gh/kwen2501/130/head -> origin/gh/kwen2501/130/head 2025-10-10T01:18:33.7404578Z * [new branch] gh/kwen2501/130/orig -> origin/gh/kwen2501/130/orig 2025-10-10T01:18:33.7406901Z * [new branch] gh/kwen2501/15/base -> origin/gh/kwen2501/15/base 2025-10-10T01:18:33.7408520Z * [new branch] gh/kwen2501/15/head -> origin/gh/kwen2501/15/head 2025-10-10T01:18:33.7410768Z * [new branch] gh/kwen2501/170/base -> origin/gh/kwen2501/170/base 2025-10-10T01:18:33.7412402Z * [new branch] gh/kwen2501/170/head -> origin/gh/kwen2501/170/head 2025-10-10T01:18:33.7414724Z * [new branch] gh/kwen2501/187/base -> origin/gh/kwen2501/187/base 2025-10-10T01:18:33.7416718Z * [new branch] gh/kwen2501/187/head -> origin/gh/kwen2501/187/head 2025-10-10T01:18:33.7418413Z * [new branch] gh/kwen2501/187/orig -> origin/gh/kwen2501/187/orig 2025-10-10T01:18:33.7420734Z * [new branch] gh/kwen2501/188/base -> origin/gh/kwen2501/188/base 2025-10-10T01:18:33.7422781Z * [new branch] gh/kwen2501/188/head -> origin/gh/kwen2501/188/head 2025-10-10T01:18:33.7424554Z * [new branch] gh/kwen2501/188/orig -> origin/gh/kwen2501/188/orig 2025-10-10T01:18:33.7426964Z * [new branch] gh/kwen2501/211/base -> origin/gh/kwen2501/211/base 2025-10-10T01:18:33.7428598Z * [new branch] gh/kwen2501/211/head -> origin/gh/kwen2501/211/head 2025-10-10T01:18:33.7430832Z * [new branch] gh/kwen2501/222/base -> origin/gh/kwen2501/222/base 2025-10-10T01:18:33.7432477Z * [new branch] gh/kwen2501/222/head -> origin/gh/kwen2501/222/head 2025-10-10T01:18:33.7434332Z * [new branch] gh/kwen2501/222/orig -> origin/gh/kwen2501/222/orig 2025-10-10T01:18:33.7437306Z * [new branch] gh/kwen2501/224/base -> origin/gh/kwen2501/224/base 2025-10-10T01:18:33.7438640Z * [new branch] gh/kwen2501/224/head -> origin/gh/kwen2501/224/head 2025-10-10T01:18:33.7440274Z * [new branch] gh/kwen2501/224/orig -> origin/gh/kwen2501/224/orig 2025-10-10T01:18:33.7442815Z * [new branch] gh/kwen2501/228/base -> origin/gh/kwen2501/228/base 2025-10-10T01:18:33.7444208Z * [new branch] gh/kwen2501/228/head -> origin/gh/kwen2501/228/head 2025-10-10T01:18:33.7446090Z * [new branch] gh/kwen2501/228/orig -> origin/gh/kwen2501/228/orig 2025-10-10T01:18:33.7448507Z * [new branch] gh/kwen2501/230/base -> origin/gh/kwen2501/230/base 2025-10-10T01:18:33.7450275Z * [new branch] gh/kwen2501/230/head -> origin/gh/kwen2501/230/head 2025-10-10T01:18:33.7451944Z * [new branch] gh/kwen2501/230/orig -> origin/gh/kwen2501/230/orig 2025-10-10T01:18:33.7454124Z * [new branch] gh/kwen2501/231/base -> origin/gh/kwen2501/231/base 2025-10-10T01:18:33.7455762Z * [new branch] gh/kwen2501/231/head -> origin/gh/kwen2501/231/head 2025-10-10T01:18:33.7457511Z * [new branch] gh/kwen2501/231/orig -> origin/gh/kwen2501/231/orig 2025-10-10T01:18:33.7459698Z * [new branch] gh/kwen2501/232/base -> origin/gh/kwen2501/232/base 2025-10-10T01:18:33.7461409Z * [new branch] gh/kwen2501/232/head -> origin/gh/kwen2501/232/head 2025-10-10T01:18:33.7463236Z * [new branch] gh/kwen2501/232/orig -> origin/gh/kwen2501/232/orig 2025-10-10T01:18:33.7465443Z * [new branch] gh/kwen2501/233/base -> origin/gh/kwen2501/233/base 2025-10-10T01:18:33.7467197Z * [new branch] gh/kwen2501/233/head -> origin/gh/kwen2501/233/head 2025-10-10T01:18:33.7469377Z * [new branch] gh/kwen2501/233/orig -> origin/gh/kwen2501/233/orig 2025-10-10T01:18:33.7471697Z * [new branch] gh/kwen2501/234/base -> origin/gh/kwen2501/234/base 2025-10-10T01:18:33.7473423Z * [new branch] gh/kwen2501/234/head -> origin/gh/kwen2501/234/head 2025-10-10T01:18:33.7475080Z * [new branch] gh/kwen2501/234/orig -> origin/gh/kwen2501/234/orig 2025-10-10T01:18:33.7477375Z * [new branch] gh/kwen2501/235/base -> origin/gh/kwen2501/235/base 2025-10-10T01:18:33.7479083Z * [new branch] gh/kwen2501/235/head -> origin/gh/kwen2501/235/head 2025-10-10T01:18:33.7480750Z * [new branch] gh/kwen2501/235/orig -> origin/gh/kwen2501/235/orig 2025-10-10T01:18:33.7482998Z * [new branch] gh/kwen2501/236/base -> origin/gh/kwen2501/236/base 2025-10-10T01:18:33.7484665Z * [new branch] gh/kwen2501/236/head -> origin/gh/kwen2501/236/head 2025-10-10T01:18:33.7486409Z * [new branch] gh/kwen2501/236/orig -> origin/gh/kwen2501/236/orig 2025-10-10T01:18:33.7488642Z * [new branch] gh/kwen2501/237/base -> origin/gh/kwen2501/237/base 2025-10-10T01:18:33.7490435Z * [new branch] gh/kwen2501/237/head -> origin/gh/kwen2501/237/head 2025-10-10T01:18:33.7492107Z * [new branch] gh/kwen2501/237/orig -> origin/gh/kwen2501/237/orig 2025-10-10T01:18:33.7494964Z * [new branch] gh/kwen2501/238/base -> origin/gh/kwen2501/238/base 2025-10-10T01:18:33.7496628Z * [new branch] gh/kwen2501/238/head -> origin/gh/kwen2501/238/head 2025-10-10T01:18:33.7498276Z * [new branch] gh/kwen2501/238/orig -> origin/gh/kwen2501/238/orig 2025-10-10T01:18:33.7500624Z * [new branch] gh/kwen2501/239/base -> origin/gh/kwen2501/239/base 2025-10-10T01:18:33.7502401Z * [new branch] gh/kwen2501/239/head -> origin/gh/kwen2501/239/head 2025-10-10T01:18:33.7503983Z * [new branch] gh/kwen2501/239/orig -> origin/gh/kwen2501/239/orig 2025-10-10T01:18:33.7506552Z * [new branch] gh/kwen2501/240/base -> origin/gh/kwen2501/240/base 2025-10-10T01:18:33.7508529Z * [new branch] gh/kwen2501/240/head -> origin/gh/kwen2501/240/head 2025-10-10T01:18:33.7510045Z * [new branch] gh/kwen2501/240/orig -> origin/gh/kwen2501/240/orig 2025-10-10T01:18:33.7512635Z * [new branch] gh/kwen2501/241/base -> origin/gh/kwen2501/241/base 2025-10-10T01:18:33.7514140Z * [new branch] gh/kwen2501/241/head -> origin/gh/kwen2501/241/head 2025-10-10T01:18:33.7515902Z * [new branch] gh/kwen2501/241/orig -> origin/gh/kwen2501/241/orig 2025-10-10T01:18:33.7518270Z * [new branch] gh/kwen2501/242/base -> origin/gh/kwen2501/242/base 2025-10-10T01:18:33.7520097Z * [new branch] gh/kwen2501/242/head -> origin/gh/kwen2501/242/head 2025-10-10T01:18:33.7521741Z * [new branch] gh/kwen2501/242/orig -> origin/gh/kwen2501/242/orig 2025-10-10T01:18:33.7523884Z * [new branch] gh/kwen2501/243/base -> origin/gh/kwen2501/243/base 2025-10-10T01:18:33.7526134Z * [new branch] gh/kwen2501/243/head -> origin/gh/kwen2501/243/head 2025-10-10T01:18:33.7527576Z * [new branch] gh/kwen2501/243/orig -> origin/gh/kwen2501/243/orig 2025-10-10T01:18:33.7529921Z * [new branch] gh/kwen2501/244/base -> origin/gh/kwen2501/244/base 2025-10-10T01:18:33.7531535Z * [new branch] gh/kwen2501/244/head -> origin/gh/kwen2501/244/head 2025-10-10T01:18:33.7533300Z * [new branch] gh/kwen2501/244/orig -> origin/gh/kwen2501/244/orig 2025-10-10T01:18:33.7535695Z * [new branch] gh/kwen2501/245/base -> origin/gh/kwen2501/245/base 2025-10-10T01:18:33.7537297Z * [new branch] gh/kwen2501/245/head -> origin/gh/kwen2501/245/head 2025-10-10T01:18:33.7538917Z * [new branch] gh/kwen2501/245/orig -> origin/gh/kwen2501/245/orig 2025-10-10T01:18:33.7541099Z * [new branch] gh/kwen2501/246/base -> origin/gh/kwen2501/246/base 2025-10-10T01:18:33.7542755Z * [new branch] gh/kwen2501/246/head -> origin/gh/kwen2501/246/head 2025-10-10T01:18:33.7544427Z * [new branch] gh/kwen2501/246/orig -> origin/gh/kwen2501/246/orig 2025-10-10T01:18:33.7546722Z * [new branch] gh/kwen2501/247/base -> origin/gh/kwen2501/247/base 2025-10-10T01:18:33.7548381Z * [new branch] gh/kwen2501/247/head -> origin/gh/kwen2501/247/head 2025-10-10T01:18:33.7550024Z * [new branch] gh/kwen2501/247/orig -> origin/gh/kwen2501/247/orig 2025-10-10T01:18:33.7552395Z * [new branch] gh/kwen2501/248/base -> origin/gh/kwen2501/248/base 2025-10-10T01:18:33.7554063Z * [new branch] gh/kwen2501/248/head -> origin/gh/kwen2501/248/head 2025-10-10T01:18:33.7556213Z * [new branch] gh/kwen2501/248/orig -> origin/gh/kwen2501/248/orig 2025-10-10T01:18:33.7558026Z * [new branch] gh/kwen2501/249/base -> origin/gh/kwen2501/249/base 2025-10-10T01:18:33.7559919Z * [new branch] gh/kwen2501/249/head -> origin/gh/kwen2501/249/head 2025-10-10T01:18:33.7561565Z * [new branch] gh/kwen2501/249/orig -> origin/gh/kwen2501/249/orig 2025-10-10T01:18:33.7564377Z * [new branch] gh/kwen2501/250/base -> origin/gh/kwen2501/250/base 2025-10-10T01:18:33.7565647Z * [new branch] gh/kwen2501/250/head -> origin/gh/kwen2501/250/head 2025-10-10T01:18:33.7567781Z * [new branch] gh/kwen2501/250/orig -> origin/gh/kwen2501/250/orig 2025-10-10T01:18:33.7570182Z * [new branch] gh/kwen2501/251/base -> origin/gh/kwen2501/251/base 2025-10-10T01:18:33.7571579Z * [new branch] gh/kwen2501/251/head -> origin/gh/kwen2501/251/head 2025-10-10T01:18:33.7573203Z * [new branch] gh/kwen2501/251/orig -> origin/gh/kwen2501/251/orig 2025-10-10T01:18:33.7575426Z * [new branch] gh/kwen2501/252/base -> origin/gh/kwen2501/252/base 2025-10-10T01:18:33.7577030Z * [new branch] gh/kwen2501/252/head -> origin/gh/kwen2501/252/head 2025-10-10T01:18:33.7578685Z * [new branch] gh/kwen2501/252/orig -> origin/gh/kwen2501/252/orig 2025-10-10T01:18:33.7581035Z * [new branch] gh/kwen2501/253/base -> origin/gh/kwen2501/253/base 2025-10-10T01:18:33.7582742Z * [new branch] gh/kwen2501/253/head -> origin/gh/kwen2501/253/head 2025-10-10T01:18:33.7584395Z * [new branch] gh/kwen2501/253/orig -> origin/gh/kwen2501/253/orig 2025-10-10T01:18:33.7587066Z * [new branch] gh/kwen2501/254/base -> origin/gh/kwen2501/254/base 2025-10-10T01:18:33.7589105Z * [new branch] gh/kwen2501/254/head -> origin/gh/kwen2501/254/head 2025-10-10T01:18:33.7590586Z * [new branch] gh/kwen2501/254/orig -> origin/gh/kwen2501/254/orig 2025-10-10T01:18:33.7592938Z * [new branch] gh/kwen2501/255/base -> origin/gh/kwen2501/255/base 2025-10-10T01:18:33.7594593Z * [new branch] gh/kwen2501/255/head -> origin/gh/kwen2501/255/head 2025-10-10T01:18:33.7596651Z * [new branch] gh/kwen2501/255/orig -> origin/gh/kwen2501/255/orig 2025-10-10T01:18:33.7599218Z * [new branch] gh/kwen2501/256/base -> origin/gh/kwen2501/256/base 2025-10-10T01:18:33.7600927Z * [new branch] gh/kwen2501/256/head -> origin/gh/kwen2501/256/head 2025-10-10T01:18:33.7602563Z * [new branch] gh/kwen2501/256/orig -> origin/gh/kwen2501/256/orig 2025-10-10T01:18:33.7604941Z * [new branch] gh/kwen2501/257/base -> origin/gh/kwen2501/257/base 2025-10-10T01:18:33.7607469Z * [new branch] gh/kwen2501/257/head -> origin/gh/kwen2501/257/head 2025-10-10T01:18:33.7608850Z * [new branch] gh/kwen2501/257/orig -> origin/gh/kwen2501/257/orig 2025-10-10T01:18:33.7611661Z * [new branch] gh/kwen2501/258/base -> origin/gh/kwen2501/258/base 2025-10-10T01:18:33.7613327Z * [new branch] gh/kwen2501/258/head -> origin/gh/kwen2501/258/head 2025-10-10T01:18:33.7614919Z * [new branch] gh/kwen2501/258/orig -> origin/gh/kwen2501/258/orig 2025-10-10T01:18:33.7618017Z * [new branch] gh/kwen2501/259/base -> origin/gh/kwen2501/259/base 2025-10-10T01:18:33.7619691Z * [new branch] gh/kwen2501/259/head -> origin/gh/kwen2501/259/head 2025-10-10T01:18:33.7621434Z * [new branch] gh/kwen2501/259/orig -> origin/gh/kwen2501/259/orig 2025-10-10T01:18:33.7623747Z * [new branch] gh/kwen2501/260/base -> origin/gh/kwen2501/260/base 2025-10-10T01:18:33.7625712Z * [new branch] gh/kwen2501/260/head -> origin/gh/kwen2501/260/head 2025-10-10T01:18:33.7627396Z * [new branch] gh/kwen2501/260/orig -> origin/gh/kwen2501/260/orig 2025-10-10T01:18:33.7629599Z * [new branch] gh/kwen2501/261/base -> origin/gh/kwen2501/261/base 2025-10-10T01:18:33.7631335Z * [new branch] gh/kwen2501/261/head -> origin/gh/kwen2501/261/head 2025-10-10T01:18:33.7632952Z * [new branch] gh/kwen2501/261/orig -> origin/gh/kwen2501/261/orig 2025-10-10T01:18:33.7635524Z * [new branch] gh/kwen2501/262/base -> origin/gh/kwen2501/262/base 2025-10-10T01:18:33.7636953Z * [new branch] gh/kwen2501/262/head -> origin/gh/kwen2501/262/head 2025-10-10T01:18:33.7638804Z * [new branch] gh/kwen2501/262/orig -> origin/gh/kwen2501/262/orig 2025-10-10T01:18:33.7641137Z * [new branch] gh/kwen2501/263/base -> origin/gh/kwen2501/263/base 2025-10-10T01:18:33.7642736Z * [new branch] gh/kwen2501/263/head -> origin/gh/kwen2501/263/head 2025-10-10T01:18:33.7644377Z * [new branch] gh/kwen2501/263/orig -> origin/gh/kwen2501/263/orig 2025-10-10T01:18:33.7646701Z * [new branch] gh/kwen2501/264/base -> origin/gh/kwen2501/264/base 2025-10-10T01:18:33.7648423Z * [new branch] gh/kwen2501/264/head -> origin/gh/kwen2501/264/head 2025-10-10T01:18:33.7650041Z * [new branch] gh/kwen2501/264/orig -> origin/gh/kwen2501/264/orig 2025-10-10T01:18:33.7652312Z * [new branch] gh/kwen2501/265/base -> origin/gh/kwen2501/265/base 2025-10-10T01:18:33.7654076Z * [new branch] gh/kwen2501/265/head -> origin/gh/kwen2501/265/head 2025-10-10T01:18:33.7655740Z * [new branch] gh/kwen2501/265/orig -> origin/gh/kwen2501/265/orig 2025-10-10T01:18:33.7658605Z * [new branch] gh/kwen2501/266/base -> origin/gh/kwen2501/266/base 2025-10-10T01:18:33.7659690Z * [new branch] gh/kwen2501/266/head -> origin/gh/kwen2501/266/head 2025-10-10T01:18:33.7661360Z * [new branch] gh/kwen2501/266/orig -> origin/gh/kwen2501/266/orig 2025-10-10T01:18:33.7664268Z * [new branch] gh/kwen2501/267/base -> origin/gh/kwen2501/267/base 2025-10-10T01:18:33.7665603Z * [new branch] gh/kwen2501/267/head -> origin/gh/kwen2501/267/head 2025-10-10T01:18:33.7667315Z * [new branch] gh/kwen2501/267/orig -> origin/gh/kwen2501/267/orig 2025-10-10T01:18:33.7670085Z * [new branch] gh/kwen2501/268/base -> origin/gh/kwen2501/268/base 2025-10-10T01:18:33.7671618Z * [new branch] gh/kwen2501/268/head -> origin/gh/kwen2501/268/head 2025-10-10T01:18:33.7673774Z * [new branch] gh/kwen2501/268/orig -> origin/gh/kwen2501/268/orig 2025-10-10T01:18:33.7676287Z * [new branch] gh/kwen2501/269/base -> origin/gh/kwen2501/269/base 2025-10-10T01:18:33.7677905Z * [new branch] gh/kwen2501/269/head -> origin/gh/kwen2501/269/head 2025-10-10T01:18:33.7679739Z * [new branch] gh/kwen2501/269/orig -> origin/gh/kwen2501/269/orig 2025-10-10T01:18:33.7682059Z * [new branch] gh/kwen2501/270/base -> origin/gh/kwen2501/270/base 2025-10-10T01:18:33.7683622Z * [new branch] gh/kwen2501/270/head -> origin/gh/kwen2501/270/head 2025-10-10T01:18:33.7685498Z * [new branch] gh/kwen2501/270/orig -> origin/gh/kwen2501/270/orig 2025-10-10T01:18:33.7687629Z * [new branch] gh/kwen2501/271/base -> origin/gh/kwen2501/271/base 2025-10-10T01:18:33.7689338Z * [new branch] gh/kwen2501/271/head -> origin/gh/kwen2501/271/head 2025-10-10T01:18:33.7691469Z * [new branch] gh/kwen2501/271/orig -> origin/gh/kwen2501/271/orig 2025-10-10T01:18:33.7693756Z * [new branch] gh/kwen2501/272/base -> origin/gh/kwen2501/272/base 2025-10-10T01:18:33.7695419Z * [new branch] gh/kwen2501/272/head -> origin/gh/kwen2501/272/head 2025-10-10T01:18:33.7697033Z * [new branch] gh/kwen2501/272/orig -> origin/gh/kwen2501/272/orig 2025-10-10T01:18:33.7699329Z * [new branch] gh/kwen2501/273/base -> origin/gh/kwen2501/273/base 2025-10-10T01:18:33.7701042Z * [new branch] gh/kwen2501/273/head -> origin/gh/kwen2501/273/head 2025-10-10T01:18:33.7702736Z * [new branch] gh/kwen2501/273/orig -> origin/gh/kwen2501/273/orig 2025-10-10T01:18:33.7705166Z * [new branch] gh/kwen2501/274/base -> origin/gh/kwen2501/274/base 2025-10-10T01:18:33.7708358Z * [new branch] gh/kwen2501/274/head -> origin/gh/kwen2501/274/head 2025-10-10T01:18:33.7709026Z * [new branch] gh/kwen2501/274/orig -> origin/gh/kwen2501/274/orig 2025-10-10T01:18:33.7712223Z * [new branch] gh/laithsakka/251/base -> origin/gh/laithsakka/251/base 2025-10-10T01:18:33.7713495Z * [new branch] gh/laithsakka/251/head -> origin/gh/laithsakka/251/head 2025-10-10T01:18:33.7715102Z * [new branch] gh/laithsakka/251/orig -> origin/gh/laithsakka/251/orig 2025-10-10T01:18:33.7717929Z * [new branch] gh/laithsakka/262/base -> origin/gh/laithsakka/262/base 2025-10-10T01:18:33.7719695Z * [new branch] gh/laithsakka/262/head -> origin/gh/laithsakka/262/head 2025-10-10T01:18:33.7721423Z * [new branch] gh/laithsakka/262/orig -> origin/gh/laithsakka/262/orig 2025-10-10T01:18:33.7723788Z * [new branch] gh/laithsakka/263/base -> origin/gh/laithsakka/263/base 2025-10-10T01:18:33.7727057Z * [new branch] gh/laithsakka/263/head -> origin/gh/laithsakka/263/head 2025-10-10T01:18:33.7728620Z * [new branch] gh/laithsakka/263/orig -> origin/gh/laithsakka/263/orig 2025-10-10T01:18:33.7730801Z * [new branch] gh/laithsakka/264/base -> origin/gh/laithsakka/264/base 2025-10-10T01:18:33.7732550Z * [new branch] gh/laithsakka/264/head -> origin/gh/laithsakka/264/head 2025-10-10T01:18:33.7734170Z * [new branch] gh/laithsakka/264/orig -> origin/gh/laithsakka/264/orig 2025-10-10T01:18:33.7736805Z * [new branch] gh/laithsakka/268/base -> origin/gh/laithsakka/268/base 2025-10-10T01:18:33.7738795Z * [new branch] gh/laithsakka/268/head -> origin/gh/laithsakka/268/head 2025-10-10T01:18:33.7740283Z * [new branch] gh/laithsakka/268/orig -> origin/gh/laithsakka/268/orig 2025-10-10T01:18:33.7742433Z * [new branch] gh/laithsakka/269/base -> origin/gh/laithsakka/269/base 2025-10-10T01:18:33.7744088Z * [new branch] gh/laithsakka/269/head -> origin/gh/laithsakka/269/head 2025-10-10T01:18:33.7745913Z * [new branch] gh/laithsakka/269/orig -> origin/gh/laithsakka/269/orig 2025-10-10T01:18:33.7748455Z * [new branch] gh/laithsakka/271/base -> origin/gh/laithsakka/271/base 2025-10-10T01:18:33.7750024Z * [new branch] gh/laithsakka/271/head -> origin/gh/laithsakka/271/head 2025-10-10T01:18:33.7751794Z * [new branch] gh/laithsakka/271/orig -> origin/gh/laithsakka/271/orig 2025-10-10T01:18:33.7754002Z * [new branch] gh/laithsakka/272/base -> origin/gh/laithsakka/272/base 2025-10-10T01:18:33.7755578Z * [new branch] gh/laithsakka/272/head -> origin/gh/laithsakka/272/head 2025-10-10T01:18:33.7757299Z * [new branch] gh/laithsakka/272/orig -> origin/gh/laithsakka/272/orig 2025-10-10T01:18:33.7759664Z * [new branch] gh/laithsakka/273/base -> origin/gh/laithsakka/273/base 2025-10-10T01:18:33.7761376Z * [new branch] gh/laithsakka/273/head -> origin/gh/laithsakka/273/head 2025-10-10T01:18:33.7762970Z * [new branch] gh/laithsakka/273/orig -> origin/gh/laithsakka/273/orig 2025-10-10T01:18:33.7765294Z * [new branch] gh/laithsakka/274/base -> origin/gh/laithsakka/274/base 2025-10-10T01:18:33.7766866Z * [new branch] gh/laithsakka/274/head -> origin/gh/laithsakka/274/head 2025-10-10T01:18:33.7768957Z * [new branch] gh/laithsakka/274/orig -> origin/gh/laithsakka/274/orig 2025-10-10T01:18:33.7771290Z * [new branch] gh/laithsakka/275/base -> origin/gh/laithsakka/275/base 2025-10-10T01:18:33.7772967Z * [new branch] gh/laithsakka/275/head -> origin/gh/laithsakka/275/head 2025-10-10T01:18:33.7774617Z * [new branch] gh/laithsakka/275/orig -> origin/gh/laithsakka/275/orig 2025-10-10T01:18:33.7777103Z * [new branch] gh/laithsakka/276/base -> origin/gh/laithsakka/276/base 2025-10-10T01:18:33.7778524Z * [new branch] gh/laithsakka/276/head -> origin/gh/laithsakka/276/head 2025-10-10T01:18:33.7780140Z * [new branch] gh/laithsakka/276/orig -> origin/gh/laithsakka/276/orig 2025-10-10T01:18:33.7783023Z * [new branch] gh/laithsakka/277/base -> origin/gh/laithsakka/277/base 2025-10-10T01:18:33.7784696Z * [new branch] gh/laithsakka/277/head -> origin/gh/laithsakka/277/head 2025-10-10T01:18:33.7786401Z * [new branch] gh/laithsakka/277/orig -> origin/gh/laithsakka/277/orig 2025-10-10T01:18:33.7788629Z * [new branch] gh/laithsakka/278/base -> origin/gh/laithsakka/278/base 2025-10-10T01:18:33.7790223Z * [new branch] gh/laithsakka/278/head -> origin/gh/laithsakka/278/head 2025-10-10T01:18:33.7791887Z * [new branch] gh/laithsakka/278/orig -> origin/gh/laithsakka/278/orig 2025-10-10T01:18:33.7794331Z * [new branch] gh/laithsakka/279/base -> origin/gh/laithsakka/279/base 2025-10-10T01:18:33.7795879Z * [new branch] gh/laithsakka/279/head -> origin/gh/laithsakka/279/head 2025-10-10T01:18:33.7797752Z * [new branch] gh/laithsakka/279/orig -> origin/gh/laithsakka/279/orig 2025-10-10T01:18:33.7800237Z * [new branch] gh/laithsakka/28/base -> origin/gh/laithsakka/28/base 2025-10-10T01:18:33.7802477Z * [new branch] gh/laithsakka/280/base -> origin/gh/laithsakka/280/base 2025-10-10T01:18:33.7805202Z * [new branch] gh/laithsakka/280/head -> origin/gh/laithsakka/280/head 2025-10-10T01:18:33.7807985Z * [new branch] gh/laithsakka/280/orig -> origin/gh/laithsakka/280/orig 2025-10-10T01:18:33.7810085Z * [new branch] gh/laithsakka/281/base -> origin/gh/laithsakka/281/base 2025-10-10T01:18:33.7811701Z * [new branch] gh/laithsakka/281/head -> origin/gh/laithsakka/281/head 2025-10-10T01:18:33.7813713Z * [new branch] gh/laithsakka/281/orig -> origin/gh/laithsakka/281/orig 2025-10-10T01:18:33.7816552Z * [new branch] gh/laithsakka/282/base -> origin/gh/laithsakka/282/base 2025-10-10T01:18:33.7818126Z * [new branch] gh/laithsakka/282/head -> origin/gh/laithsakka/282/head 2025-10-10T01:18:33.7819965Z * [new branch] gh/laithsakka/282/orig -> origin/gh/laithsakka/282/orig 2025-10-10T01:18:33.7822343Z * [new branch] gh/laithsakka/283/base -> origin/gh/laithsakka/283/base 2025-10-10T01:18:33.7824107Z * [new branch] gh/laithsakka/283/head -> origin/gh/laithsakka/283/head 2025-10-10T01:18:33.7826522Z * [new branch] gh/laithsakka/283/orig -> origin/gh/laithsakka/283/orig 2025-10-10T01:18:33.7828393Z * [new branch] gh/laithsakka/284/base -> origin/gh/laithsakka/284/base 2025-10-10T01:18:33.7829987Z * [new branch] gh/laithsakka/284/head -> origin/gh/laithsakka/284/head 2025-10-10T01:18:33.7831682Z * [new branch] gh/laithsakka/284/orig -> origin/gh/laithsakka/284/orig 2025-10-10T01:18:33.7834394Z * [new branch] gh/laithsakka/285/base -> origin/gh/laithsakka/285/base 2025-10-10T01:18:33.7835836Z * [new branch] gh/laithsakka/285/head -> origin/gh/laithsakka/285/head 2025-10-10T01:18:33.7837420Z * [new branch] gh/laithsakka/285/orig -> origin/gh/laithsakka/285/orig 2025-10-10T01:18:33.7840186Z * [new branch] gh/laithsakka/286/base -> origin/gh/laithsakka/286/base 2025-10-10T01:18:33.7841713Z * [new branch] gh/laithsakka/286/head -> origin/gh/laithsakka/286/head 2025-10-10T01:18:33.7843730Z * [new branch] gh/laithsakka/286/orig -> origin/gh/laithsakka/286/orig 2025-10-10T01:18:33.7845960Z * [new branch] gh/laithsakka/287/base -> origin/gh/laithsakka/287/base 2025-10-10T01:18:33.7847565Z * [new branch] gh/laithsakka/287/head -> origin/gh/laithsakka/287/head 2025-10-10T01:18:33.7849203Z * [new branch] gh/laithsakka/287/orig -> origin/gh/laithsakka/287/orig 2025-10-10T01:18:33.7851512Z * [new branch] gh/laithsakka/288/base -> origin/gh/laithsakka/288/base 2025-10-10T01:18:33.7853158Z * [new branch] gh/laithsakka/288/head -> origin/gh/laithsakka/288/head 2025-10-10T01:18:33.7854793Z * [new branch] gh/laithsakka/288/orig -> origin/gh/laithsakka/288/orig 2025-10-10T01:18:33.7857262Z * [new branch] gh/laithsakka/289/base -> origin/gh/laithsakka/289/base 2025-10-10T01:18:33.7859057Z * [new branch] gh/laithsakka/289/head -> origin/gh/laithsakka/289/head 2025-10-10T01:18:33.7860753Z * [new branch] gh/laithsakka/289/orig -> origin/gh/laithsakka/289/orig 2025-10-10T01:18:33.7862956Z * [new branch] gh/laithsakka/29/base -> origin/gh/laithsakka/29/base 2025-10-10T01:18:33.7865223Z * [new branch] gh/laithsakka/290/base -> origin/gh/laithsakka/290/base 2025-10-10T01:18:33.7867210Z * [new branch] gh/laithsakka/290/head -> origin/gh/laithsakka/290/head 2025-10-10T01:18:33.7868943Z * [new branch] gh/laithsakka/290/orig -> origin/gh/laithsakka/290/orig 2025-10-10T01:18:33.7871128Z * [new branch] gh/laithsakka/291/base -> origin/gh/laithsakka/291/base 2025-10-10T01:18:33.7872660Z * [new branch] gh/laithsakka/291/head -> origin/gh/laithsakka/291/head 2025-10-10T01:18:33.7874386Z * [new branch] gh/laithsakka/291/orig -> origin/gh/laithsakka/291/orig 2025-10-10T01:18:33.7877158Z * [new branch] gh/laithsakka/292/base -> origin/gh/laithsakka/292/base 2025-10-10T01:18:33.7879215Z * [new branch] gh/laithsakka/292/head -> origin/gh/laithsakka/292/head 2025-10-10T01:18:33.7881095Z * [new branch] gh/laithsakka/292/orig -> origin/gh/laithsakka/292/orig 2025-10-10T01:18:33.7883697Z * [new branch] gh/laithsakka/293/base -> origin/gh/laithsakka/293/base 2025-10-10T01:18:33.7885159Z * [new branch] gh/laithsakka/293/head -> origin/gh/laithsakka/293/head 2025-10-10T01:18:33.7887060Z * [new branch] gh/laithsakka/293/orig -> origin/gh/laithsakka/293/orig 2025-10-10T01:18:33.7889471Z * [new branch] gh/laithsakka/294/base -> origin/gh/laithsakka/294/base 2025-10-10T01:18:33.7891162Z * [new branch] gh/laithsakka/294/head -> origin/gh/laithsakka/294/head 2025-10-10T01:18:33.7892891Z * [new branch] gh/laithsakka/294/orig -> origin/gh/laithsakka/294/orig 2025-10-10T01:18:33.7895267Z * [new branch] gh/laithsakka/295/base -> origin/gh/laithsakka/295/base 2025-10-10T01:18:33.7896947Z * [new branch] gh/laithsakka/295/head -> origin/gh/laithsakka/295/head 2025-10-10T01:18:33.7898532Z * [new branch] gh/laithsakka/295/orig -> origin/gh/laithsakka/295/orig 2025-10-10T01:18:33.7900909Z * [new branch] gh/laithsakka/296/base -> origin/gh/laithsakka/296/base 2025-10-10T01:18:33.7902590Z * [new branch] gh/laithsakka/296/head -> origin/gh/laithsakka/296/head 2025-10-10T01:18:33.7904156Z * [new branch] gh/laithsakka/296/orig -> origin/gh/laithsakka/296/orig 2025-10-10T01:18:33.7907376Z * [new branch] gh/laithsakka/297/base -> origin/gh/laithsakka/297/base 2025-10-10T01:18:33.7909019Z * [new branch] gh/laithsakka/297/head -> origin/gh/laithsakka/297/head 2025-10-10T01:18:33.7910695Z * [new branch] gh/laithsakka/297/orig -> origin/gh/laithsakka/297/orig 2025-10-10T01:18:33.7913581Z * [new branch] gh/laithsakka/298/base -> origin/gh/laithsakka/298/base 2025-10-10T01:18:33.7915144Z * [new branch] gh/laithsakka/298/head -> origin/gh/laithsakka/298/head 2025-10-10T01:18:33.7916817Z * [new branch] gh/laithsakka/298/orig -> origin/gh/laithsakka/298/orig 2025-10-10T01:18:33.7919297Z * [new branch] gh/laithsakka/299/base -> origin/gh/laithsakka/299/base 2025-10-10T01:18:33.7920924Z * [new branch] gh/laithsakka/299/head -> origin/gh/laithsakka/299/head 2025-10-10T01:18:33.7922590Z * [new branch] gh/laithsakka/299/orig -> origin/gh/laithsakka/299/orig 2025-10-10T01:18:33.7925053Z * [new branch] gh/laithsakka/30/base -> origin/gh/laithsakka/30/base 2025-10-10T01:18:33.7926730Z * [new branch] gh/laithsakka/30/head -> origin/gh/laithsakka/30/head 2025-10-10T01:18:33.7929066Z * [new branch] gh/laithsakka/300/base -> origin/gh/laithsakka/300/base 2025-10-10T01:18:33.7930675Z * [new branch] gh/laithsakka/300/head -> origin/gh/laithsakka/300/head 2025-10-10T01:18:33.7932398Z * [new branch] gh/laithsakka/300/orig -> origin/gh/laithsakka/300/orig 2025-10-10T01:18:33.7934899Z * [new branch] gh/laithsakka/301/base -> origin/gh/laithsakka/301/base 2025-10-10T01:18:33.7936648Z * [new branch] gh/laithsakka/301/head -> origin/gh/laithsakka/301/head 2025-10-10T01:18:33.7938442Z * [new branch] gh/laithsakka/301/orig -> origin/gh/laithsakka/301/orig 2025-10-10T01:18:33.7940786Z * [new branch] gh/laithsakka/302/base -> origin/gh/laithsakka/302/base 2025-10-10T01:18:33.7942424Z * [new branch] gh/laithsakka/302/head -> origin/gh/laithsakka/302/head 2025-10-10T01:18:33.7944025Z * [new branch] gh/laithsakka/302/orig -> origin/gh/laithsakka/302/orig 2025-10-10T01:18:33.7946390Z * [new branch] gh/laithsakka/303/base -> origin/gh/laithsakka/303/base 2025-10-10T01:18:33.7948356Z * [new branch] gh/laithsakka/303/head -> origin/gh/laithsakka/303/head 2025-10-10T01:18:33.7949855Z * [new branch] gh/laithsakka/303/orig -> origin/gh/laithsakka/303/orig 2025-10-10T01:18:33.7951986Z * [new branch] gh/laithsakka/304/base -> origin/gh/laithsakka/304/base 2025-10-10T01:18:33.7953647Z * [new branch] gh/laithsakka/304/head -> origin/gh/laithsakka/304/head 2025-10-10T01:18:33.7955246Z * [new branch] gh/laithsakka/304/orig -> origin/gh/laithsakka/304/orig 2025-10-10T01:18:33.7957818Z * [new branch] gh/laithsakka/305/base -> origin/gh/laithsakka/305/base 2025-10-10T01:18:33.7959682Z * [new branch] gh/laithsakka/305/head -> origin/gh/laithsakka/305/head 2025-10-10T01:18:33.7961305Z * [new branch] gh/laithsakka/305/orig -> origin/gh/laithsakka/305/orig 2025-10-10T01:18:33.7963588Z * [new branch] gh/laithsakka/306/base -> origin/gh/laithsakka/306/base 2025-10-10T01:18:33.7965178Z * [new branch] gh/laithsakka/306/head -> origin/gh/laithsakka/306/head 2025-10-10T01:18:33.7966762Z * [new branch] gh/laithsakka/306/orig -> origin/gh/laithsakka/306/orig 2025-10-10T01:18:33.7968975Z * [new branch] gh/laithsakka/307/base -> origin/gh/laithsakka/307/base 2025-10-10T01:18:33.7970650Z * [new branch] gh/laithsakka/307/head -> origin/gh/laithsakka/307/head 2025-10-10T01:18:33.7972301Z * [new branch] gh/laithsakka/307/orig -> origin/gh/laithsakka/307/orig 2025-10-10T01:18:33.7974507Z * [new branch] gh/laithsakka/308/base -> origin/gh/laithsakka/308/base 2025-10-10T01:18:33.7976227Z * [new branch] gh/laithsakka/308/head -> origin/gh/laithsakka/308/head 2025-10-10T01:18:33.7977847Z * [new branch] gh/laithsakka/308/orig -> origin/gh/laithsakka/308/orig 2025-10-10T01:18:33.7980170Z * [new branch] gh/laithsakka/309/base -> origin/gh/laithsakka/309/base 2025-10-10T01:18:33.7981819Z * [new branch] gh/laithsakka/309/head -> origin/gh/laithsakka/309/head 2025-10-10T01:18:33.7983491Z * [new branch] gh/laithsakka/309/orig -> origin/gh/laithsakka/309/orig 2025-10-10T01:18:33.7986074Z * [new branch] gh/laithsakka/31/base -> origin/gh/laithsakka/31/base 2025-10-10T01:18:33.7987692Z * [new branch] gh/laithsakka/31/head -> origin/gh/laithsakka/31/head 2025-10-10T01:18:33.7990138Z * [new branch] gh/laithsakka/310/base -> origin/gh/laithsakka/310/base 2025-10-10T01:18:33.7991744Z * [new branch] gh/laithsakka/310/head -> origin/gh/laithsakka/310/head 2025-10-10T01:18:33.7993366Z * [new branch] gh/laithsakka/310/orig -> origin/gh/laithsakka/310/orig 2025-10-10T01:18:33.7995739Z * [new branch] gh/laithsakka/311/base -> origin/gh/laithsakka/311/base 2025-10-10T01:18:33.7997388Z * [new branch] gh/laithsakka/311/head -> origin/gh/laithsakka/311/head 2025-10-10T01:18:33.7999355Z * [new branch] gh/laithsakka/311/orig -> origin/gh/laithsakka/311/orig 2025-10-10T01:18:33.8001608Z * [new branch] gh/laithsakka/312/base -> origin/gh/laithsakka/312/base 2025-10-10T01:18:33.8003443Z * [new branch] gh/laithsakka/312/head -> origin/gh/laithsakka/312/head 2025-10-10T01:18:33.8005602Z * [new branch] gh/laithsakka/312/orig -> origin/gh/laithsakka/312/orig 2025-10-10T01:18:33.8007810Z * [new branch] gh/laithsakka/313/base -> origin/gh/laithsakka/313/base 2025-10-10T01:18:33.8009463Z * [new branch] gh/laithsakka/313/head -> origin/gh/laithsakka/313/head 2025-10-10T01:18:33.8011094Z * [new branch] gh/laithsakka/313/orig -> origin/gh/laithsakka/313/orig 2025-10-10T01:18:33.8013341Z * [new branch] gh/laithsakka/32/base -> origin/gh/laithsakka/32/base 2025-10-10T01:18:33.8014899Z * [new branch] gh/laithsakka/32/head -> origin/gh/laithsakka/32/head 2025-10-10T01:18:33.8017829Z * [new branch] gh/liangel/1/base -> origin/gh/liangel/1/base 2025-10-10T01:18:33.8019501Z * [new branch] gh/liangel/1/head -> origin/gh/liangel/1/head 2025-10-10T01:18:33.8021399Z * [new branch] gh/liangel/1/orig -> origin/gh/liangel/1/orig 2025-10-10T01:18:33.8023716Z * [new branch] gh/liangel/2/base -> origin/gh/liangel/2/base 2025-10-10T01:18:33.8025568Z * [new branch] gh/liangel/2/head -> origin/gh/liangel/2/head 2025-10-10T01:18:33.8027353Z * [new branch] gh/liangel/2/orig -> origin/gh/liangel/2/orig 2025-10-10T01:18:33.8029856Z * [new branch] gh/liangel/3/base -> origin/gh/liangel/3/base 2025-10-10T01:18:33.8031359Z * [new branch] gh/liangel/3/head -> origin/gh/liangel/3/head 2025-10-10T01:18:33.8032957Z * [new branch] gh/liangel/3/orig -> origin/gh/liangel/3/orig 2025-10-10T01:18:33.8035213Z * [new branch] gh/liangel/4/base -> origin/gh/liangel/4/base 2025-10-10T01:18:33.8036909Z * [new branch] gh/liangel/4/head -> origin/gh/liangel/4/head 2025-10-10T01:18:33.8038835Z * [new branch] gh/liangel/4/orig -> origin/gh/liangel/4/orig 2025-10-10T01:18:33.8042675Z * [new branch] gh/lucaskabela/1/base -> origin/gh/lucaskabela/1/base 2025-10-10T01:18:33.8044440Z * [new branch] gh/lucaskabela/1/head -> origin/gh/lucaskabela/1/head 2025-10-10T01:18:33.8047147Z * [new branch] gh/lw/3/base -> origin/gh/lw/3/base 2025-10-10T01:18:33.8048893Z * [new branch] gh/lw/3/head -> origin/gh/lw/3/head 2025-10-10T01:18:33.8050681Z * [new branch] gh/lw/3/orig -> origin/gh/lw/3/orig 2025-10-10T01:18:33.8052724Z * [new branch] gh/lw/4/base -> origin/gh/lw/4/base 2025-10-10T01:18:33.8054385Z * [new branch] gh/lw/4/head -> origin/gh/lw/4/head 2025-10-10T01:18:33.8056154Z * [new branch] gh/lw/4/orig -> origin/gh/lw/4/orig 2025-10-10T01:18:33.8058541Z * [new branch] gh/lw/5/base -> origin/gh/lw/5/base 2025-10-10T01:18:33.8060162Z * [new branch] gh/lw/5/head -> origin/gh/lw/5/head 2025-10-10T01:18:33.8061963Z * [new branch] gh/lw/5/orig -> origin/gh/lw/5/orig 2025-10-10T01:18:33.8064494Z * [new branch] gh/lw/6/base -> origin/gh/lw/6/base 2025-10-10T01:18:33.8066144Z * [new branch] gh/lw/6/head -> origin/gh/lw/6/head 2025-10-10T01:18:33.8067952Z * [new branch] gh/lw/6/orig -> origin/gh/lw/6/orig 2025-10-10T01:18:33.8070748Z * [new branch] gh/malfet/14/base -> origin/gh/malfet/14/base 2025-10-10T01:18:33.8072978Z * [new branch] gh/malfet/396/base -> origin/gh/malfet/396/base 2025-10-10T01:18:33.8074926Z * [new branch] gh/malfet/396/head -> origin/gh/malfet/396/head 2025-10-10T01:18:33.8076721Z * [new branch] gh/malfet/396/orig -> origin/gh/malfet/396/orig 2025-10-10T01:18:33.8079141Z * [new branch] gh/malfet/397/base -> origin/gh/malfet/397/base 2025-10-10T01:18:33.8080778Z * [new branch] gh/malfet/397/head -> origin/gh/malfet/397/head 2025-10-10T01:18:33.8082894Z * [new branch] gh/malfet/397/orig -> origin/gh/malfet/397/orig 2025-10-10T01:18:33.8085170Z * [new branch] gh/malfet/398/base -> origin/gh/malfet/398/base 2025-10-10T01:18:33.8086800Z * [new branch] gh/malfet/398/head -> origin/gh/malfet/398/head 2025-10-10T01:18:33.8088524Z * [new branch] gh/malfet/398/orig -> origin/gh/malfet/398/orig 2025-10-10T01:18:33.8090753Z * [new branch] gh/malfet/399/base -> origin/gh/malfet/399/base 2025-10-10T01:18:33.8092461Z * [new branch] gh/malfet/399/head -> origin/gh/malfet/399/head 2025-10-10T01:18:33.8094183Z * [new branch] gh/malfet/399/orig -> origin/gh/malfet/399/orig 2025-10-10T01:18:33.8096407Z * [new branch] gh/malfet/414/base -> origin/gh/malfet/414/base 2025-10-10T01:18:33.8098144Z * [new branch] gh/malfet/414/head -> origin/gh/malfet/414/head 2025-10-10T01:18:33.8099715Z * [new branch] gh/malfet/414/orig -> origin/gh/malfet/414/orig 2025-10-10T01:18:33.8101943Z * [new branch] gh/malfet/417/base -> origin/gh/malfet/417/base 2025-10-10T01:18:33.8103681Z * [new branch] gh/malfet/417/head -> origin/gh/malfet/417/head 2025-10-10T01:18:33.8105299Z * [new branch] gh/malfet/417/orig -> origin/gh/malfet/417/orig 2025-10-10T01:18:33.8107454Z * [new branch] gh/malfet/418/base -> origin/gh/malfet/418/base 2025-10-10T01:18:33.8109131Z * [new branch] gh/malfet/418/head -> origin/gh/malfet/418/head 2025-10-10T01:18:33.8110747Z * [new branch] gh/malfet/418/orig -> origin/gh/malfet/418/orig 2025-10-10T01:18:33.8113133Z * [new branch] gh/malfet/505/base -> origin/gh/malfet/505/base 2025-10-10T01:18:33.8114803Z * [new branch] gh/malfet/505/head -> origin/gh/malfet/505/head 2025-10-10T01:18:33.8116656Z * [new branch] gh/malfet/505/orig -> origin/gh/malfet/505/orig 2025-10-10T01:18:33.8118968Z * [new branch] gh/malfet/506/base -> origin/gh/malfet/506/base 2025-10-10T01:18:33.8120775Z * [new branch] gh/malfet/506/head -> origin/gh/malfet/506/head 2025-10-10T01:18:33.8122302Z * [new branch] gh/malfet/506/orig -> origin/gh/malfet/506/orig 2025-10-10T01:18:33.8126638Z * [new branch] gh/malfet/507/base -> origin/gh/malfet/507/base 2025-10-10T01:18:33.8128228Z * [new branch] gh/malfet/507/head -> origin/gh/malfet/507/head 2025-10-10T01:18:33.8129914Z * [new branch] gh/malfet/507/orig -> origin/gh/malfet/507/orig 2025-10-10T01:18:33.8132150Z * [new branch] gh/malfet/513/base -> origin/gh/malfet/513/base 2025-10-10T01:18:33.8133861Z * [new branch] gh/malfet/513/head -> origin/gh/malfet/513/head 2025-10-10T01:18:33.8135704Z * [new branch] gh/malfet/513/orig -> origin/gh/malfet/513/orig 2025-10-10T01:18:33.8138843Z * [new branch] gh/malfet/516/base -> origin/gh/malfet/516/base 2025-10-10T01:18:33.8140402Z * [new branch] gh/malfet/516/head -> origin/gh/malfet/516/head 2025-10-10T01:18:33.8142080Z * [new branch] gh/malfet/516/orig -> origin/gh/malfet/516/orig 2025-10-10T01:18:33.8145291Z * [new branch] gh/malfet/517/base -> origin/gh/malfet/517/base 2025-10-10T01:18:33.8146851Z * [new branch] gh/malfet/517/head -> origin/gh/malfet/517/head 2025-10-10T01:18:33.8149159Z * [new branch] gh/malfet/518/base -> origin/gh/malfet/518/base 2025-10-10T01:18:33.8150767Z * [new branch] gh/malfet/518/head -> origin/gh/malfet/518/head 2025-10-10T01:18:33.8152480Z * [new branch] gh/malfet/518/orig -> origin/gh/malfet/518/orig 2025-10-10T01:18:33.8154694Z * [new branch] gh/malfet/519/base -> origin/gh/malfet/519/base 2025-10-10T01:18:33.8156368Z * [new branch] gh/malfet/519/head -> origin/gh/malfet/519/head 2025-10-10T01:18:33.8158090Z * [new branch] gh/malfet/519/orig -> origin/gh/malfet/519/orig 2025-10-10T01:18:33.8160358Z * [new branch] gh/malfet/520/base -> origin/gh/malfet/520/base 2025-10-10T01:18:33.8162555Z * [new branch] gh/malfet/520/head -> origin/gh/malfet/520/head 2025-10-10T01:18:33.8164295Z * [new branch] gh/malfet/520/orig -> origin/gh/malfet/520/orig 2025-10-10T01:18:33.8167223Z * [new branch] gh/malfet/521/base -> origin/gh/malfet/521/base 2025-10-10T01:18:33.8168879Z * [new branch] gh/malfet/521/head -> origin/gh/malfet/521/head 2025-10-10T01:18:33.8170507Z * [new branch] gh/malfet/521/orig -> origin/gh/malfet/521/orig 2025-10-10T01:18:33.8172835Z * [new branch] gh/malfet/522/base -> origin/gh/malfet/522/base 2025-10-10T01:18:33.8174507Z * [new branch] gh/malfet/522/head -> origin/gh/malfet/522/head 2025-10-10T01:18:33.8176174Z * [new branch] gh/malfet/522/orig -> origin/gh/malfet/522/orig 2025-10-10T01:18:33.8179439Z * [new branch] gh/malfet/523/base -> origin/gh/malfet/523/base 2025-10-10T01:18:33.8180560Z * [new branch] gh/malfet/523/head -> origin/gh/malfet/523/head 2025-10-10T01:18:33.8182207Z * [new branch] gh/malfet/523/orig -> origin/gh/malfet/523/orig 2025-10-10T01:18:33.8184814Z * [new branch] gh/malfet/524/base -> origin/gh/malfet/524/base 2025-10-10T01:18:33.8186343Z * [new branch] gh/malfet/524/head -> origin/gh/malfet/524/head 2025-10-10T01:18:33.8188144Z * [new branch] gh/malfet/524/orig -> origin/gh/malfet/524/orig 2025-10-10T01:18:33.8190568Z * [new branch] gh/malfet/525/base -> origin/gh/malfet/525/base 2025-10-10T01:18:33.8192201Z * [new branch] gh/malfet/525/head -> origin/gh/malfet/525/head 2025-10-10T01:18:33.8194071Z * [new branch] gh/malfet/525/orig -> origin/gh/malfet/525/orig 2025-10-10T01:18:33.8196295Z * [new branch] gh/malfet/526/base -> origin/gh/malfet/526/base 2025-10-10T01:18:33.8198330Z * [new branch] gh/malfet/526/head -> origin/gh/malfet/526/head 2025-10-10T01:18:33.8200165Z * [new branch] gh/malfet/526/orig -> origin/gh/malfet/526/orig 2025-10-10T01:18:33.8202444Z * [new branch] gh/malfet/527/base -> origin/gh/malfet/527/base 2025-10-10T01:18:33.8204192Z * [new branch] gh/malfet/527/head -> origin/gh/malfet/527/head 2025-10-10T01:18:33.8205840Z * [new branch] gh/malfet/527/orig -> origin/gh/malfet/527/orig 2025-10-10T01:18:33.8208344Z * [new branch] gh/malfet/528/base -> origin/gh/malfet/528/base 2025-10-10T01:18:33.8209895Z * [new branch] gh/malfet/528/head -> origin/gh/malfet/528/head 2025-10-10T01:18:33.8211581Z * [new branch] gh/malfet/528/orig -> origin/gh/malfet/528/orig 2025-10-10T01:18:33.8213930Z * [new branch] gh/malfet/529/base -> origin/gh/malfet/529/base 2025-10-10T01:18:33.8215689Z * [new branch] gh/malfet/529/head -> origin/gh/malfet/529/head 2025-10-10T01:18:33.8217225Z * [new branch] gh/malfet/529/orig -> origin/gh/malfet/529/orig 2025-10-10T01:18:33.8219618Z * [new branch] gh/malfet/530/base -> origin/gh/malfet/530/base 2025-10-10T01:18:33.8221264Z * [new branch] gh/malfet/530/head -> origin/gh/malfet/530/head 2025-10-10T01:18:33.8222978Z * [new branch] gh/malfet/530/orig -> origin/gh/malfet/530/orig 2025-10-10T01:18:33.8225648Z * [new branch] gh/malfet/531/base -> origin/gh/malfet/531/base 2025-10-10T01:18:33.8227348Z * [new branch] gh/malfet/531/head -> origin/gh/malfet/531/head 2025-10-10T01:18:33.8229088Z * [new branch] gh/malfet/531/orig -> origin/gh/malfet/531/orig 2025-10-10T01:18:33.8231433Z * [new branch] gh/malfet/532/base -> origin/gh/malfet/532/base 2025-10-10T01:18:33.8233087Z * [new branch] gh/malfet/532/head -> origin/gh/malfet/532/head 2025-10-10T01:18:33.8235343Z * [new branch] gh/malfet/532/orig -> origin/gh/malfet/532/orig 2025-10-10T01:18:33.8237722Z * [new branch] gh/malfet/533/base -> origin/gh/malfet/533/base 2025-10-10T01:18:33.8239491Z * [new branch] gh/malfet/533/head -> origin/gh/malfet/533/head 2025-10-10T01:18:33.8241173Z * [new branch] gh/malfet/533/orig -> origin/gh/malfet/533/orig 2025-10-10T01:18:33.8243493Z * [new branch] gh/malfet/534/base -> origin/gh/malfet/534/base 2025-10-10T01:18:33.8245168Z * [new branch] gh/malfet/534/head -> origin/gh/malfet/534/head 2025-10-10T01:18:33.8246839Z * [new branch] gh/malfet/534/orig -> origin/gh/malfet/534/orig 2025-10-10T01:18:33.8249100Z * [new branch] gh/malfet/535/base -> origin/gh/malfet/535/base 2025-10-10T01:18:33.8250743Z * [new branch] gh/malfet/535/head -> origin/gh/malfet/535/head 2025-10-10T01:18:33.8252896Z * [new branch] gh/malfet/535/orig -> origin/gh/malfet/535/orig 2025-10-10T01:18:33.8255208Z * [new branch] gh/malfet/536/base -> origin/gh/malfet/536/base 2025-10-10T01:18:33.8256899Z * [new branch] gh/malfet/536/head -> origin/gh/malfet/536/head 2025-10-10T01:18:33.8258595Z * [new branch] gh/malfet/536/orig -> origin/gh/malfet/536/orig 2025-10-10T01:18:33.8260966Z * [new branch] gh/malfet/537/base -> origin/gh/malfet/537/base 2025-10-10T01:18:33.8262628Z * [new branch] gh/malfet/537/head -> origin/gh/malfet/537/head 2025-10-10T01:18:33.8264528Z * [new branch] gh/malfet/537/orig -> origin/gh/malfet/537/orig 2025-10-10T01:18:33.8267059Z * [new branch] gh/malfet/538/base -> origin/gh/malfet/538/base 2025-10-10T01:18:33.8269065Z * [new branch] gh/malfet/538/head -> origin/gh/malfet/538/head 2025-10-10T01:18:33.8270732Z * [new branch] gh/malfet/538/orig -> origin/gh/malfet/538/orig 2025-10-10T01:18:33.8273309Z * [new branch] gh/malfet/539/base -> origin/gh/malfet/539/base 2025-10-10T01:18:33.8274594Z * [new branch] gh/malfet/539/head -> origin/gh/malfet/539/head 2025-10-10T01:18:33.8276503Z * [new branch] gh/malfet/539/orig -> origin/gh/malfet/539/orig 2025-10-10T01:18:33.8278937Z * [new branch] gh/malfet/540/base -> origin/gh/malfet/540/base 2025-10-10T01:18:33.8280655Z * [new branch] gh/malfet/540/head -> origin/gh/malfet/540/head 2025-10-10T01:18:33.8282391Z * [new branch] gh/malfet/540/orig -> origin/gh/malfet/540/orig 2025-10-10T01:18:33.8284782Z * [new branch] gh/malfet/541/base -> origin/gh/malfet/541/base 2025-10-10T01:18:33.8287397Z * [new branch] gh/malfet/541/head -> origin/gh/malfet/541/head 2025-10-10T01:18:33.8288634Z * [new branch] gh/malfet/541/orig -> origin/gh/malfet/541/orig 2025-10-10T01:18:33.8291203Z * [new branch] gh/malfet/542/base -> origin/gh/malfet/542/base 2025-10-10T01:18:33.8292558Z * [new branch] gh/malfet/542/head -> origin/gh/malfet/542/head 2025-10-10T01:18:33.8294289Z * [new branch] gh/malfet/542/orig -> origin/gh/malfet/542/orig 2025-10-10T01:18:33.8297607Z * [new branch] gh/malfet/543/base -> origin/gh/malfet/543/base 2025-10-10T01:18:33.8299105Z * [new branch] gh/malfet/543/head -> origin/gh/malfet/543/head 2025-10-10T01:18:33.8300810Z * [new branch] gh/malfet/543/orig -> origin/gh/malfet/543/orig 2025-10-10T01:18:33.8303049Z * [new branch] gh/malfet/544/base -> origin/gh/malfet/544/base 2025-10-10T01:18:33.8304679Z * [new branch] gh/malfet/544/head -> origin/gh/malfet/544/head 2025-10-10T01:18:33.8306448Z * [new branch] gh/malfet/544/orig -> origin/gh/malfet/544/orig 2025-10-10T01:18:33.8308737Z * [new branch] gh/malfet/545/base -> origin/gh/malfet/545/base 2025-10-10T01:18:33.8325661Z * [new branch] gh/malfet/545/head -> origin/gh/malfet/545/head 2025-10-10T01:18:33.8326430Z * [new branch] gh/malfet/545/orig -> origin/gh/malfet/545/orig 2025-10-10T01:18:33.8326784Z * [new branch] gh/malfet/546/base -> origin/gh/malfet/546/base 2025-10-10T01:18:33.8327147Z * [new branch] gh/malfet/546/head -> origin/gh/malfet/546/head 2025-10-10T01:18:33.8327475Z * [new branch] gh/malfet/546/orig -> origin/gh/malfet/546/orig 2025-10-10T01:18:33.8327813Z * [new branch] gh/malfet/547/base -> origin/gh/malfet/547/base 2025-10-10T01:18:33.8328182Z * [new branch] gh/malfet/547/head -> origin/gh/malfet/547/head 2025-10-10T01:18:33.8328507Z * [new branch] gh/malfet/547/orig -> origin/gh/malfet/547/orig 2025-10-10T01:18:33.8328831Z * [new branch] gh/malfet/548/base -> origin/gh/malfet/548/base 2025-10-10T01:18:33.8329165Z * [new branch] gh/malfet/548/head -> origin/gh/malfet/548/head 2025-10-10T01:18:33.8329739Z * [new branch] gh/malfet/548/orig -> origin/gh/malfet/548/orig 2025-10-10T01:18:33.8332634Z * [new branch] gh/malfet/549/base -> origin/gh/malfet/549/base 2025-10-10T01:18:33.8334142Z * [new branch] gh/malfet/549/head -> origin/gh/malfet/549/head 2025-10-10T01:18:33.8335784Z * [new branch] gh/malfet/549/orig -> origin/gh/malfet/549/orig 2025-10-10T01:18:33.8338494Z * [new branch] gh/malfet/550/base -> origin/gh/malfet/550/base 2025-10-10T01:18:33.8340337Z * [new branch] gh/malfet/550/head -> origin/gh/malfet/550/head 2025-10-10T01:18:33.8341913Z * [new branch] gh/malfet/550/orig -> origin/gh/malfet/550/orig 2025-10-10T01:18:33.8344197Z * [new branch] gh/malfet/551/base -> origin/gh/malfet/551/base 2025-10-10T01:18:33.8346074Z * [new branch] gh/malfet/551/head -> origin/gh/malfet/551/head 2025-10-10T01:18:33.8348095Z * [new branch] gh/malfet/551/orig -> origin/gh/malfet/551/orig 2025-10-10T01:18:33.8350690Z * [new branch] gh/malfet/552/base -> origin/gh/malfet/552/base 2025-10-10T01:18:33.8352002Z * [new branch] gh/malfet/552/head -> origin/gh/malfet/552/head 2025-10-10T01:18:33.8353622Z * [new branch] gh/malfet/552/orig -> origin/gh/malfet/552/orig 2025-10-10T01:18:33.8356771Z * [new branch] gh/malfet/553/base -> origin/gh/malfet/553/base 2025-10-10T01:18:33.8357877Z * [new branch] gh/malfet/553/head -> origin/gh/malfet/553/head 2025-10-10T01:18:33.8359838Z * [new branch] gh/malfet/553/orig -> origin/gh/malfet/553/orig 2025-10-10T01:18:33.8362201Z * [new branch] gh/malfet/64/base -> origin/gh/malfet/64/base 2025-10-10T01:18:33.8363858Z * [new branch] gh/malfet/64/head -> origin/gh/malfet/64/head 2025-10-10T01:18:33.8366671Z * [new branch] gh/manuelcandales/10/base -> origin/gh/manuelcandales/10/base 2025-10-10T01:18:33.8368327Z * [new branch] gh/manuelcandales/10/head -> origin/gh/manuelcandales/10/head 2025-10-10T01:18:33.8370082Z * [new branch] gh/manuelcandales/10/orig -> origin/gh/manuelcandales/10/orig 2025-10-10T01:18:33.8372207Z * [new branch] gh/manuelcandales/11/base -> origin/gh/manuelcandales/11/base 2025-10-10T01:18:33.8373842Z * [new branch] gh/manuelcandales/11/head -> origin/gh/manuelcandales/11/head 2025-10-10T01:18:33.8375460Z * [new branch] gh/manuelcandales/11/orig -> origin/gh/manuelcandales/11/orig 2025-10-10T01:18:33.8377722Z * [new branch] gh/manuelcandales/9/base -> origin/gh/manuelcandales/9/base 2025-10-10T01:18:33.8379397Z * [new branch] gh/manuelcandales/9/head -> origin/gh/manuelcandales/9/head 2025-10-10T01:18:33.8381043Z * [new branch] gh/manuelcandales/9/orig -> origin/gh/manuelcandales/9/orig 2025-10-10T01:18:33.8384383Z * [new branch] gh/markkm/1/base -> origin/gh/markkm/1/base 2025-10-10T01:18:33.8387271Z * [new branch] gh/masnesral/235/base -> origin/gh/masnesral/235/base 2025-10-10T01:18:33.8389014Z * [new branch] gh/masnesral/235/head -> origin/gh/masnesral/235/head 2025-10-10T01:18:33.8390705Z * [new branch] gh/masnesral/235/orig -> origin/gh/masnesral/235/orig 2025-10-10T01:18:33.8393013Z * [new branch] gh/masnesral/236/base -> origin/gh/masnesral/236/base 2025-10-10T01:18:33.8394675Z * [new branch] gh/masnesral/236/head -> origin/gh/masnesral/236/head 2025-10-10T01:18:33.8396426Z * [new branch] gh/masnesral/236/orig -> origin/gh/masnesral/236/orig 2025-10-10T01:18:33.8398946Z * [new branch] gh/masnesral/237/base -> origin/gh/masnesral/237/base 2025-10-10T01:18:33.8400766Z * [new branch] gh/masnesral/237/head -> origin/gh/masnesral/237/head 2025-10-10T01:18:33.8402618Z * [new branch] gh/masnesral/237/orig -> origin/gh/masnesral/237/orig 2025-10-10T01:18:33.8405091Z * [new branch] gh/masnesral/238/base -> origin/gh/masnesral/238/base 2025-10-10T01:18:33.8407224Z * [new branch] gh/masnesral/238/head -> origin/gh/masnesral/238/head 2025-10-10T01:18:33.8409125Z * [new branch] gh/masnesral/238/orig -> origin/gh/masnesral/238/orig 2025-10-10T01:18:33.8412211Z * [new branch] gh/mhorowitz/0/base -> origin/gh/mhorowitz/0/base 2025-10-10T01:18:33.8413377Z * [new branch] gh/mhorowitz/0/head -> origin/gh/mhorowitz/0/head 2025-10-10T01:18:33.8416144Z * [new branch] gh/mhorowitz/1/base -> origin/gh/mhorowitz/1/base 2025-10-10T01:18:33.8417361Z * [new branch] gh/mhorowitz/1/head -> origin/gh/mhorowitz/1/head 2025-10-10T01:18:33.8419794Z * [new branch] gh/mhorowitz/2/base -> origin/gh/mhorowitz/2/base 2025-10-10T01:18:33.8421490Z * [new branch] gh/mhorowitz/2/head -> origin/gh/mhorowitz/2/head 2025-10-10T01:18:33.8423696Z * [new branch] gh/mhorowitz/3/base -> origin/gh/mhorowitz/3/base 2025-10-10T01:18:33.8425617Z * [new branch] gh/mhorowitz/3/head -> origin/gh/mhorowitz/3/head 2025-10-10T01:18:33.8427687Z * [new branch] gh/mhorowitz/4/base -> origin/gh/mhorowitz/4/base 2025-10-10T01:18:33.8429255Z * [new branch] gh/mhorowitz/4/head -> origin/gh/mhorowitz/4/head 2025-10-10T01:18:33.8431392Z * [new branch] gh/mhorowitz/5/base -> origin/gh/mhorowitz/5/base 2025-10-10T01:18:33.8432954Z * [new branch] gh/mhorowitz/5/head -> origin/gh/mhorowitz/5/head 2025-10-10T01:18:33.8435091Z * [new branch] gh/mhorowitz/6/base -> origin/gh/mhorowitz/6/base 2025-10-10T01:18:33.8436735Z * [new branch] gh/mhorowitz/6/head -> origin/gh/mhorowitz/6/head 2025-10-10T01:18:33.8439721Z * [new branch] gh/mikaylagawarecki/234/base -> origin/gh/mikaylagawarecki/234/base 2025-10-10T01:18:33.8441390Z * [new branch] gh/mikaylagawarecki/234/head -> origin/gh/mikaylagawarecki/234/head 2025-10-10T01:18:33.8443970Z * [new branch] gh/mikaylagawarecki/235/base -> origin/gh/mikaylagawarecki/235/base 2025-10-10T01:18:33.8445666Z * [new branch] gh/mikaylagawarecki/235/head -> origin/gh/mikaylagawarecki/235/head 2025-10-10T01:18:33.8447901Z * [new branch] gh/mikaylagawarecki/236/base -> origin/gh/mikaylagawarecki/236/base 2025-10-10T01:18:33.8449555Z * [new branch] gh/mikaylagawarecki/236/head -> origin/gh/mikaylagawarecki/236/head 2025-10-10T01:18:33.8451777Z * [new branch] gh/mikaylagawarecki/237/base -> origin/gh/mikaylagawarecki/237/base 2025-10-10T01:18:33.8453367Z * [new branch] gh/mikaylagawarecki/237/head -> origin/gh/mikaylagawarecki/237/head 2025-10-10T01:18:33.8455576Z * [new branch] gh/mikaylagawarecki/238/base -> origin/gh/mikaylagawarecki/238/base 2025-10-10T01:18:33.8457302Z * [new branch] gh/mikaylagawarecki/238/head -> origin/gh/mikaylagawarecki/238/head 2025-10-10T01:18:33.8459563Z * [new branch] gh/mikaylagawarecki/317/base -> origin/gh/mikaylagawarecki/317/base 2025-10-10T01:18:33.8461420Z * [new branch] gh/mikaylagawarecki/317/head -> origin/gh/mikaylagawarecki/317/head 2025-10-10T01:18:33.8463092Z * [new branch] gh/mikaylagawarecki/317/orig -> origin/gh/mikaylagawarecki/317/orig 2025-10-10T01:18:33.8465397Z * [new branch] gh/mikaylagawarecki/336/base -> origin/gh/mikaylagawarecki/336/base 2025-10-10T01:18:33.8467038Z * [new branch] gh/mikaylagawarecki/336/head -> origin/gh/mikaylagawarecki/336/head 2025-10-10T01:18:33.8468719Z * [new branch] gh/mikaylagawarecki/336/orig -> origin/gh/mikaylagawarecki/336/orig 2025-10-10T01:18:33.8471407Z * [new branch] gh/mikaylagawarecki/337/base -> origin/gh/mikaylagawarecki/337/base 2025-10-10T01:18:33.8473360Z * [new branch] gh/mikaylagawarecki/337/head -> origin/gh/mikaylagawarecki/337/head 2025-10-10T01:18:33.8475081Z * [new branch] gh/mikaylagawarecki/337/orig -> origin/gh/mikaylagawarecki/337/orig 2025-10-10T01:18:33.8477505Z * [new branch] gh/mikaylagawarecki/340/base -> origin/gh/mikaylagawarecki/340/base 2025-10-10T01:18:33.8479434Z * [new branch] gh/mikaylagawarecki/340/head -> origin/gh/mikaylagawarecki/340/head 2025-10-10T01:18:33.8481061Z * [new branch] gh/mikaylagawarecki/340/orig -> origin/gh/mikaylagawarecki/340/orig 2025-10-10T01:18:33.8483522Z * [new branch] gh/mikaylagawarecki/341/base -> origin/gh/mikaylagawarecki/341/base 2025-10-10T01:18:33.8485357Z * [new branch] gh/mikaylagawarecki/341/head -> origin/gh/mikaylagawarecki/341/head 2025-10-10T01:18:33.8487034Z * [new branch] gh/mikaylagawarecki/341/orig -> origin/gh/mikaylagawarecki/341/orig 2025-10-10T01:18:33.8489412Z * [new branch] gh/mikaylagawarecki/342/base -> origin/gh/mikaylagawarecki/342/base 2025-10-10T01:18:33.8491058Z * [new branch] gh/mikaylagawarecki/342/head -> origin/gh/mikaylagawarecki/342/head 2025-10-10T01:18:33.8492816Z * [new branch] gh/mikaylagawarecki/342/orig -> origin/gh/mikaylagawarecki/342/orig 2025-10-10T01:18:33.8495054Z * [new branch] gh/mikaylagawarecki/343/base -> origin/gh/mikaylagawarecki/343/base 2025-10-10T01:18:33.8497129Z * [new branch] gh/mikaylagawarecki/343/head -> origin/gh/mikaylagawarecki/343/head 2025-10-10T01:18:33.8498918Z * [new branch] gh/mikaylagawarecki/343/orig -> origin/gh/mikaylagawarecki/343/orig 2025-10-10T01:18:33.8501192Z * [new branch] gh/mikaylagawarecki/344/base -> origin/gh/mikaylagawarecki/344/base 2025-10-10T01:18:33.8502865Z * [new branch] gh/mikaylagawarecki/344/head -> origin/gh/mikaylagawarecki/344/head 2025-10-10T01:18:33.8504991Z * [new branch] gh/mikaylagawarecki/344/orig -> origin/gh/mikaylagawarecki/344/orig 2025-10-10T01:18:33.8507601Z * [new branch] gh/mikaylagawarecki/345/base -> origin/gh/mikaylagawarecki/345/base 2025-10-10T01:18:33.8509726Z * [new branch] gh/mikaylagawarecki/345/head -> origin/gh/mikaylagawarecki/345/head 2025-10-10T01:18:33.8511262Z * [new branch] gh/mikaylagawarecki/345/orig -> origin/gh/mikaylagawarecki/345/orig 2025-10-10T01:18:33.8514222Z * [new branch] gh/mikaylagawarecki/346/base -> origin/gh/mikaylagawarecki/346/base 2025-10-10T01:18:33.8515463Z * [new branch] gh/mikaylagawarecki/346/head -> origin/gh/mikaylagawarecki/346/head 2025-10-10T01:18:33.8517106Z * [new branch] gh/mikaylagawarecki/346/orig -> origin/gh/mikaylagawarecki/346/orig 2025-10-10T01:18:33.8519680Z * [new branch] gh/mikaylagawarecki/347/base -> origin/gh/mikaylagawarecki/347/base 2025-10-10T01:18:33.8521370Z * [new branch] gh/mikaylagawarecki/347/head -> origin/gh/mikaylagawarecki/347/head 2025-10-10T01:18:33.8523053Z * [new branch] gh/mikaylagawarecki/347/orig -> origin/gh/mikaylagawarecki/347/orig 2025-10-10T01:18:33.8527259Z * [new branch] gh/mikaylagawarecki/348/base -> origin/gh/mikaylagawarecki/348/base 2025-10-10T01:18:33.8528923Z * [new branch] gh/mikaylagawarecki/348/head -> origin/gh/mikaylagawarecki/348/head 2025-10-10T01:18:33.8530549Z * [new branch] gh/mikaylagawarecki/348/orig -> origin/gh/mikaylagawarecki/348/orig 2025-10-10T01:18:33.8533112Z * [new branch] gh/mikaylagawarecki/349/base -> origin/gh/mikaylagawarecki/349/base 2025-10-10T01:18:33.8534737Z * [new branch] gh/mikaylagawarecki/349/head -> origin/gh/mikaylagawarecki/349/head 2025-10-10T01:18:33.8536417Z * [new branch] gh/mikaylagawarecki/349/orig -> origin/gh/mikaylagawarecki/349/orig 2025-10-10T01:18:33.8539002Z * [new branch] gh/mikaylagawarecki/350/base -> origin/gh/mikaylagawarecki/350/base 2025-10-10T01:18:33.8540587Z * [new branch] gh/mikaylagawarecki/350/head -> origin/gh/mikaylagawarecki/350/head 2025-10-10T01:18:33.8542190Z * [new branch] gh/mikaylagawarecki/350/orig -> origin/gh/mikaylagawarecki/350/orig 2025-10-10T01:18:33.8545066Z * [new branch] gh/mlazos/18/base -> origin/gh/mlazos/18/base 2025-10-10T01:18:33.8546740Z * [new branch] gh/mlazos/18/head -> origin/gh/mlazos/18/head 2025-10-10T01:18:33.8548491Z * [new branch] gh/mlazos/18/orig -> origin/gh/mlazos/18/orig 2025-10-10T01:18:33.8550500Z * [new branch] gh/mlazos/19/base -> origin/gh/mlazos/19/base 2025-10-10T01:18:33.8552146Z * [new branch] gh/mlazos/19/head -> origin/gh/mlazos/19/head 2025-10-10T01:18:33.8553793Z * [new branch] gh/mlazos/19/orig -> origin/gh/mlazos/19/orig 2025-10-10T01:18:33.8556162Z * [new branch] gh/mlazos/20/base -> origin/gh/mlazos/20/base 2025-10-10T01:18:33.8557973Z * [new branch] gh/mlazos/20/head -> origin/gh/mlazos/20/head 2025-10-10T01:18:33.8559780Z * [new branch] gh/mlazos/20/orig -> origin/gh/mlazos/20/orig 2025-10-10T01:18:33.8562177Z * [new branch] gh/mlazos/21/base -> origin/gh/mlazos/21/base 2025-10-10T01:18:33.8563847Z * [new branch] gh/mlazos/21/head -> origin/gh/mlazos/21/head 2025-10-10T01:18:33.8565616Z * [new branch] gh/mlazos/21/orig -> origin/gh/mlazos/21/orig 2025-10-10T01:18:33.8568907Z * [new branch] gh/mlazos/22/base -> origin/gh/mlazos/22/base 2025-10-10T01:18:33.8570011Z * [new branch] gh/mlazos/22/head -> origin/gh/mlazos/22/head 2025-10-10T01:18:33.8571610Z * [new branch] gh/mlazos/22/orig -> origin/gh/mlazos/22/orig 2025-10-10T01:18:33.8574192Z * [new branch] gh/mlazos/23/base -> origin/gh/mlazos/23/base 2025-10-10T01:18:33.8575754Z * [new branch] gh/mlazos/23/head -> origin/gh/mlazos/23/head 2025-10-10T01:18:33.8577811Z * [new branch] gh/mlazos/23/orig -> origin/gh/mlazos/23/orig 2025-10-10T01:18:33.8580104Z * [new branch] gh/mlazos/24/base -> origin/gh/mlazos/24/base 2025-10-10T01:18:33.8581782Z * [new branch] gh/mlazos/24/head -> origin/gh/mlazos/24/head 2025-10-10T01:18:33.8583404Z * [new branch] gh/mlazos/24/orig -> origin/gh/mlazos/24/orig 2025-10-10T01:18:33.8585846Z * [new branch] gh/mlazos/25/base -> origin/gh/mlazos/25/base 2025-10-10T01:18:33.8587579Z * [new branch] gh/mlazos/25/head -> origin/gh/mlazos/25/head 2025-10-10T01:18:33.8589277Z * [new branch] gh/mlazos/25/orig -> origin/gh/mlazos/25/orig 2025-10-10T01:18:33.8591577Z * [new branch] gh/mlazos/26/base -> origin/gh/mlazos/26/base 2025-10-10T01:18:33.8593226Z * [new branch] gh/mlazos/26/head -> origin/gh/mlazos/26/head 2025-10-10T01:18:33.8594880Z * [new branch] gh/mlazos/26/orig -> origin/gh/mlazos/26/orig 2025-10-10T01:18:33.8597126Z * [new branch] gh/mlazos/27/base -> origin/gh/mlazos/27/base 2025-10-10T01:18:33.8598915Z * [new branch] gh/mlazos/27/head -> origin/gh/mlazos/27/head 2025-10-10T01:18:33.8600518Z * [new branch] gh/mlazos/27/orig -> origin/gh/mlazos/27/orig 2025-10-10T01:18:33.8602858Z * [new branch] gh/mlazos/28/base -> origin/gh/mlazos/28/base 2025-10-10T01:18:33.8604521Z * [new branch] gh/mlazos/28/head -> origin/gh/mlazos/28/head 2025-10-10T01:18:33.8606247Z * [new branch] gh/mlazos/28/orig -> origin/gh/mlazos/28/orig 2025-10-10T01:18:33.8608611Z * [new branch] gh/mlazos/29/base -> origin/gh/mlazos/29/base 2025-10-10T01:18:33.8610362Z * [new branch] gh/mlazos/29/head -> origin/gh/mlazos/29/head 2025-10-10T01:18:33.8611966Z * [new branch] gh/mlazos/29/orig -> origin/gh/mlazos/29/orig 2025-10-10T01:18:33.8614197Z * [new branch] gh/mlazos/30/base -> origin/gh/mlazos/30/base 2025-10-10T01:18:33.8615902Z * [new branch] gh/mlazos/30/head -> origin/gh/mlazos/30/head 2025-10-10T01:18:33.8617574Z * [new branch] gh/mlazos/30/orig -> origin/gh/mlazos/30/orig 2025-10-10T01:18:33.8619844Z * [new branch] gh/mlazos/31/base -> origin/gh/mlazos/31/base 2025-10-10T01:18:33.8621502Z * [new branch] gh/mlazos/31/head -> origin/gh/mlazos/31/head 2025-10-10T01:18:33.8623228Z * [new branch] gh/mlazos/31/orig -> origin/gh/mlazos/31/orig 2025-10-10T01:18:33.8625839Z * [new branch] gh/mlazos/32/base -> origin/gh/mlazos/32/base 2025-10-10T01:18:33.8627517Z * [new branch] gh/mlazos/32/head -> origin/gh/mlazos/32/head 2025-10-10T01:18:33.8629142Z * [new branch] gh/mlazos/32/orig -> origin/gh/mlazos/32/orig 2025-10-10T01:18:33.8631539Z * [new branch] gh/mlazos/33/base -> origin/gh/mlazos/33/base 2025-10-10T01:18:33.8633162Z * [new branch] gh/mlazos/33/head -> origin/gh/mlazos/33/head 2025-10-10T01:18:33.8634834Z * [new branch] gh/mlazos/33/orig -> origin/gh/mlazos/33/orig 2025-10-10T01:18:33.8637112Z * [new branch] gh/mlazos/34/base -> origin/gh/mlazos/34/base 2025-10-10T01:18:33.8638826Z * [new branch] gh/mlazos/34/head -> origin/gh/mlazos/34/head 2025-10-10T01:18:33.8640527Z * [new branch] gh/mlazos/34/orig -> origin/gh/mlazos/34/orig 2025-10-10T01:18:33.8642659Z * [new branch] gh/mlazos/35/base -> origin/gh/mlazos/35/base 2025-10-10T01:18:33.8644294Z * [new branch] gh/mlazos/35/head -> origin/gh/mlazos/35/head 2025-10-10T01:18:33.8646725Z * [new branch] gh/mlazos/35/orig -> origin/gh/mlazos/35/orig 2025-10-10T01:18:33.8649607Z * [new branch] gh/mlazos/36/base -> origin/gh/mlazos/36/base 2025-10-10T01:18:33.8651263Z * [new branch] gh/mlazos/36/head -> origin/gh/mlazos/36/head 2025-10-10T01:18:33.8652852Z * [new branch] gh/mlazos/36/orig -> origin/gh/mlazos/36/orig 2025-10-10T01:18:33.8655973Z * [new branch] gh/mlazos/37/base -> origin/gh/mlazos/37/base 2025-10-10T01:18:33.8657150Z * [new branch] gh/mlazos/37/head -> origin/gh/mlazos/37/head 2025-10-10T01:18:33.8658989Z * [new branch] gh/mlazos/37/orig -> origin/gh/mlazos/37/orig 2025-10-10T01:18:33.8661730Z * [new branch] gh/mrmiywj/1/base -> origin/gh/mrmiywj/1/base 2025-10-10T01:18:33.8663489Z * [new branch] gh/mrmiywj/1/head -> origin/gh/mrmiywj/1/head 2025-10-10T01:18:33.8666408Z * [new branch] gh/muchulee8/62/base -> origin/gh/muchulee8/62/base 2025-10-10T01:18:33.8668150Z * [new branch] gh/muchulee8/62/head -> origin/gh/muchulee8/62/head 2025-10-10T01:18:33.8669841Z * [new branch] gh/muchulee8/62/orig -> origin/gh/muchulee8/62/orig 2025-10-10T01:18:33.8672303Z * [new branch] gh/muchulee8/64/base -> origin/gh/muchulee8/64/base 2025-10-10T01:18:33.8673992Z * [new branch] gh/muchulee8/64/head -> origin/gh/muchulee8/64/head 2025-10-10T01:18:33.8676063Z * [new branch] gh/muchulee8/64/orig -> origin/gh/muchulee8/64/orig 2025-10-10T01:18:33.8679160Z * [new branch] gh/muchulee8/65/base -> origin/gh/muchulee8/65/base 2025-10-10T01:18:33.8680717Z * [new branch] gh/muchulee8/65/head -> origin/gh/muchulee8/65/head 2025-10-10T01:18:33.8682172Z * [new branch] gh/muchulee8/65/orig -> origin/gh/muchulee8/65/orig 2025-10-10T01:18:33.8684578Z * [new branch] gh/muchulee8/66/base -> origin/gh/muchulee8/66/base 2025-10-10T01:18:33.8686289Z * [new branch] gh/muchulee8/66/head -> origin/gh/muchulee8/66/head 2025-10-10T01:18:33.8688206Z * [new branch] gh/muchulee8/66/orig -> origin/gh/muchulee8/66/orig 2025-10-10T01:18:33.8690589Z * [new branch] gh/muchulee8/67/base -> origin/gh/muchulee8/67/base 2025-10-10T01:18:33.8692214Z * [new branch] gh/muchulee8/67/head -> origin/gh/muchulee8/67/head 2025-10-10T01:18:33.8693874Z * [new branch] gh/muchulee8/67/orig -> origin/gh/muchulee8/67/orig 2025-10-10T01:18:33.8697458Z * [new branch] gh/naveenthangudu/1/base -> origin/gh/naveenthangudu/1/base 2025-10-10T01:18:33.8699493Z * [new branch] gh/naveenthangudu/1/head -> origin/gh/naveenthangudu/1/head 2025-10-10T01:18:33.8701084Z * [new branch] gh/naveenthangudu/1/orig -> origin/gh/naveenthangudu/1/orig 2025-10-10T01:18:33.8703391Z * [new branch] gh/naveenthangudu/2/base -> origin/gh/naveenthangudu/2/base 2025-10-10T01:18:33.8705069Z * [new branch] gh/naveenthangudu/2/head -> origin/gh/naveenthangudu/2/head 2025-10-10T01:18:33.8706929Z * [new branch] gh/naveenthangudu/2/orig -> origin/gh/naveenthangudu/2/orig 2025-10-10T01:18:33.8709414Z * [new branch] gh/naveenthangudu/3/base -> origin/gh/naveenthangudu/3/base 2025-10-10T01:18:33.8711112Z * [new branch] gh/naveenthangudu/3/head -> origin/gh/naveenthangudu/3/head 2025-10-10T01:18:33.8712815Z * [new branch] gh/naveenthangudu/3/orig -> origin/gh/naveenthangudu/3/orig 2025-10-10T01:18:33.8715059Z * [new branch] gh/naveenthangudu/4/base -> origin/gh/naveenthangudu/4/base 2025-10-10T01:18:33.8716746Z * [new branch] gh/naveenthangudu/4/head -> origin/gh/naveenthangudu/4/head 2025-10-10T01:18:33.8718674Z * [new branch] gh/naveenthangudu/4/orig -> origin/gh/naveenthangudu/4/orig 2025-10-10T01:18:33.8721051Z * [new branch] gh/naveenthangudu/5/base -> origin/gh/naveenthangudu/5/base 2025-10-10T01:18:33.8722779Z * [new branch] gh/naveenthangudu/5/head -> origin/gh/naveenthangudu/5/head 2025-10-10T01:18:33.8724726Z * [new branch] gh/naveenthangudu/5/orig -> origin/gh/naveenthangudu/5/orig 2025-10-10T01:18:33.8727157Z * [new branch] gh/naveenthangudu/6/base -> origin/gh/naveenthangudu/6/base 2025-10-10T01:18:33.8728758Z * [new branch] gh/naveenthangudu/6/head -> origin/gh/naveenthangudu/6/head 2025-10-10T01:18:33.8730354Z * [new branch] gh/naveenthangudu/6/orig -> origin/gh/naveenthangudu/6/orig 2025-10-10T01:18:33.8732627Z * [new branch] gh/naveenthangudu/7/base -> origin/gh/naveenthangudu/7/base 2025-10-10T01:18:33.8734342Z * [new branch] gh/naveenthangudu/7/head -> origin/gh/naveenthangudu/7/head 2025-10-10T01:18:33.8736706Z * [new branch] gh/naveenthangudu/7/orig -> origin/gh/naveenthangudu/7/orig 2025-10-10T01:18:33.8738574Z * [new branch] gh/naveenthangudu/8/base -> origin/gh/naveenthangudu/8/base 2025-10-10T01:18:33.8740138Z * [new branch] gh/naveenthangudu/8/head -> origin/gh/naveenthangudu/8/head 2025-10-10T01:18:33.8741975Z * [new branch] gh/naveenthangudu/8/orig -> origin/gh/naveenthangudu/8/orig 2025-10-10T01:18:33.8745607Z * [new branch] gh/nikitaved/1/base -> origin/gh/nikitaved/1/base 2025-10-10T01:18:33.8747111Z * [new branch] gh/nikitaved/1/head -> origin/gh/nikitaved/1/head 2025-10-10T01:18:33.8749058Z * [new branch] gh/nikitaved/1/orig -> origin/gh/nikitaved/1/orig 2025-10-10T01:18:33.8751247Z * [new branch] gh/nikitaved/2/base -> origin/gh/nikitaved/2/base 2025-10-10T01:18:33.8753270Z * [new branch] gh/nikitaved/2/head -> origin/gh/nikitaved/2/head 2025-10-10T01:18:33.8754901Z * [new branch] gh/nikitaved/2/orig -> origin/gh/nikitaved/2/orig 2025-10-10T01:18:33.8757619Z * [new branch] gh/nikitaved/3/base -> origin/gh/nikitaved/3/base 2025-10-10T01:18:33.8759438Z * [new branch] gh/nikitaved/3/head -> origin/gh/nikitaved/3/head 2025-10-10T01:18:33.8761030Z * [new branch] gh/nikitaved/3/orig -> origin/gh/nikitaved/3/orig 2025-10-10T01:18:33.8763819Z * [new branch] gh/oulgen/35/base -> origin/gh/oulgen/35/base 2025-10-10T01:18:33.8765493Z * [new branch] gh/oulgen/35/head -> origin/gh/oulgen/35/head 2025-10-10T01:18:33.8767255Z * [new branch] gh/oulgen/35/orig -> origin/gh/oulgen/35/orig 2025-10-10T01:18:33.8769470Z * [new branch] gh/patvig/mtia-serialization -> origin/gh/patvig/mtia-serialization 2025-10-10T01:18:33.8772354Z * [new branch] gh/pearu/108/base -> origin/gh/pearu/108/base 2025-10-10T01:18:33.8773998Z * [new branch] gh/pearu/108/head -> origin/gh/pearu/108/head 2025-10-10T01:18:33.8775920Z * [new branch] gh/pearu/108/orig -> origin/gh/pearu/108/orig 2025-10-10T01:18:33.8778956Z * [new branch] gh/pearu/109/base -> origin/gh/pearu/109/base 2025-10-10T01:18:33.8780106Z * [new branch] gh/pearu/109/head -> origin/gh/pearu/109/head 2025-10-10T01:18:33.8781918Z * [new branch] gh/pearu/109/orig -> origin/gh/pearu/109/orig 2025-10-10T01:18:33.8784560Z * [new branch] gh/pearu/110/base -> origin/gh/pearu/110/base 2025-10-10T01:18:33.8786146Z * [new branch] gh/pearu/110/head -> origin/gh/pearu/110/head 2025-10-10T01:18:33.8787928Z * [new branch] gh/pearu/110/orig -> origin/gh/pearu/110/orig 2025-10-10T01:18:33.8790188Z * [new branch] gh/pearu/111/base -> origin/gh/pearu/111/base 2025-10-10T01:18:33.8791823Z * [new branch] gh/pearu/111/head -> origin/gh/pearu/111/head 2025-10-10T01:18:33.8793497Z * [new branch] gh/pearu/111/orig -> origin/gh/pearu/111/orig 2025-10-10T01:18:33.8795806Z * [new branch] gh/pearu/112/base -> origin/gh/pearu/112/base 2025-10-10T01:18:33.8797392Z * [new branch] gh/pearu/112/head -> origin/gh/pearu/112/head 2025-10-10T01:18:33.8799181Z * [new branch] gh/pearu/112/orig -> origin/gh/pearu/112/orig 2025-10-10T01:18:33.8801412Z * [new branch] gh/pearu/113/base -> origin/gh/pearu/113/base 2025-10-10T01:18:33.8803580Z * [new branch] gh/pearu/113/head -> origin/gh/pearu/113/head 2025-10-10T01:18:33.8805255Z * [new branch] gh/pearu/113/orig -> origin/gh/pearu/113/orig 2025-10-10T01:18:33.8807620Z * [new branch] gh/pearu/114/base -> origin/gh/pearu/114/base 2025-10-10T01:18:33.8809389Z * [new branch] gh/pearu/114/head -> origin/gh/pearu/114/head 2025-10-10T01:18:33.8811040Z * [new branch] gh/pearu/114/orig -> origin/gh/pearu/114/orig 2025-10-10T01:18:33.8813218Z * [new branch] gh/pearu/115/base -> origin/gh/pearu/115/base 2025-10-10T01:18:33.8815120Z * [new branch] gh/pearu/115/head -> origin/gh/pearu/115/head 2025-10-10T01:18:33.8816813Z * [new branch] gh/pearu/115/orig -> origin/gh/pearu/115/orig 2025-10-10T01:18:33.8818982Z * [new branch] gh/pearu/116/base -> origin/gh/pearu/116/base 2025-10-10T01:18:33.8820760Z * [new branch] gh/pearu/116/head -> origin/gh/pearu/116/head 2025-10-10T01:18:33.8822325Z * [new branch] gh/pearu/116/orig -> origin/gh/pearu/116/orig 2025-10-10T01:18:33.8824597Z * [new branch] gh/pearu/117/base -> origin/gh/pearu/117/base 2025-10-10T01:18:33.8826749Z * [new branch] gh/pearu/117/head -> origin/gh/pearu/117/head 2025-10-10T01:18:33.8828471Z * [new branch] gh/pearu/117/orig -> origin/gh/pearu/117/orig 2025-10-10T01:18:33.8830698Z * [new branch] gh/pearu/118/base -> origin/gh/pearu/118/base 2025-10-10T01:18:33.8832442Z * [new branch] gh/pearu/118/head -> origin/gh/pearu/118/head 2025-10-10T01:18:33.8834576Z * [new branch] gh/pearu/118/orig -> origin/gh/pearu/118/orig 2025-10-10T01:18:33.8836907Z * [new branch] gh/pearu/119/base -> origin/gh/pearu/119/base 2025-10-10T01:18:33.8839085Z * [new branch] gh/pearu/119/head -> origin/gh/pearu/119/head 2025-10-10T01:18:33.8840715Z * [new branch] gh/pearu/119/orig -> origin/gh/pearu/119/orig 2025-10-10T01:18:33.8843670Z * [new branch] gh/pearu/120/base -> origin/gh/pearu/120/base 2025-10-10T01:18:33.8844901Z * [new branch] gh/pearu/120/head -> origin/gh/pearu/120/head 2025-10-10T01:18:33.8846737Z * [new branch] gh/pearu/120/orig -> origin/gh/pearu/120/orig 2025-10-10T01:18:33.8849019Z * [new branch] gh/pearu/121/base -> origin/gh/pearu/121/base 2025-10-10T01:18:33.8850657Z * [new branch] gh/pearu/121/head -> origin/gh/pearu/121/head 2025-10-10T01:18:33.8852360Z * [new branch] gh/pearu/121/orig -> origin/gh/pearu/121/orig 2025-10-10T01:18:33.8854563Z * [new branch] gh/pearu/122/base -> origin/gh/pearu/122/base 2025-10-10T01:18:33.8857123Z * [new branch] gh/pearu/122/head -> origin/gh/pearu/122/head 2025-10-10T01:18:33.8858462Z * [new branch] gh/pearu/122/orig -> origin/gh/pearu/122/orig 2025-10-10T01:18:33.8860929Z * [new branch] gh/pearu/123/base -> origin/gh/pearu/123/base 2025-10-10T01:18:33.8862489Z * [new branch] gh/pearu/123/head -> origin/gh/pearu/123/head 2025-10-10T01:18:33.8864100Z * [new branch] gh/pearu/123/orig -> origin/gh/pearu/123/orig 2025-10-10T01:18:33.8867246Z * [new branch] gh/pearu/124/base -> origin/gh/pearu/124/base 2025-10-10T01:18:33.8868871Z * [new branch] gh/pearu/124/head -> origin/gh/pearu/124/head 2025-10-10T01:18:33.8870520Z * [new branch] gh/pearu/124/orig -> origin/gh/pearu/124/orig 2025-10-10T01:18:33.8872826Z * [new branch] gh/pearu/125/base -> origin/gh/pearu/125/base 2025-10-10T01:18:33.8874422Z * [new branch] gh/pearu/125/head -> origin/gh/pearu/125/head 2025-10-10T01:18:33.8876093Z * [new branch] gh/pearu/125/orig -> origin/gh/pearu/125/orig 2025-10-10T01:18:33.8878471Z * [new branch] gh/pearu/126/base -> origin/gh/pearu/126/base 2025-10-10T01:18:33.8880327Z * [new branch] gh/pearu/126/head -> origin/gh/pearu/126/head 2025-10-10T01:18:33.8882101Z * [new branch] gh/pearu/126/orig -> origin/gh/pearu/126/orig 2025-10-10T01:18:33.8884298Z * [new branch] gh/pearu/127/base -> origin/gh/pearu/127/base 2025-10-10T01:18:33.8886190Z * [new branch] gh/pearu/127/head -> origin/gh/pearu/127/head 2025-10-10T01:18:33.8888180Z * [new branch] gh/pearu/127/orig -> origin/gh/pearu/127/orig 2025-10-10T01:18:33.8891269Z * [new branch] gh/pearu/128/base -> origin/gh/pearu/128/base 2025-10-10T01:18:33.8893317Z * [new branch] gh/pearu/128/head -> origin/gh/pearu/128/head 2025-10-10T01:18:33.8894104Z * [new branch] gh/pearu/128/orig -> origin/gh/pearu/128/orig 2025-10-10T01:18:33.8896182Z * [new branch] gh/pearu/129/base -> origin/gh/pearu/129/base 2025-10-10T01:18:33.8897964Z * [new branch] gh/pearu/129/head -> origin/gh/pearu/129/head 2025-10-10T01:18:33.8899638Z * [new branch] gh/pearu/129/orig -> origin/gh/pearu/129/orig 2025-10-10T01:18:33.8901795Z * [new branch] gh/pearu/130/base -> origin/gh/pearu/130/base 2025-10-10T01:18:33.8903567Z * [new branch] gh/pearu/130/head -> origin/gh/pearu/130/head 2025-10-10T01:18:33.8905266Z * [new branch] gh/pearu/130/orig -> origin/gh/pearu/130/orig 2025-10-10T01:18:33.8907394Z * [new branch] gh/pearu/131/base -> origin/gh/pearu/131/base 2025-10-10T01:18:33.8909082Z * [new branch] gh/pearu/131/head -> origin/gh/pearu/131/head 2025-10-10T01:18:33.8910789Z * [new branch] gh/pearu/131/orig -> origin/gh/pearu/131/orig 2025-10-10T01:18:33.8913043Z * [new branch] gh/pearu/132/base -> origin/gh/pearu/132/base 2025-10-10T01:18:33.8914615Z * [new branch] gh/pearu/132/head -> origin/gh/pearu/132/head 2025-10-10T01:18:33.8916233Z * [new branch] gh/pearu/132/orig -> origin/gh/pearu/132/orig 2025-10-10T01:18:33.8918969Z * [new branch] gh/pearu/133/base -> origin/gh/pearu/133/base 2025-10-10T01:18:33.8920590Z * [new branch] gh/pearu/133/head -> origin/gh/pearu/133/head 2025-10-10T01:18:33.8922197Z * [new branch] gh/pearu/133/orig -> origin/gh/pearu/133/orig 2025-10-10T01:18:33.8924646Z * [new branch] gh/pearu/134/base -> origin/gh/pearu/134/base 2025-10-10T01:18:33.8927923Z * [new branch] gh/pearu/134/head -> origin/gh/pearu/134/head 2025-10-10T01:18:33.8929732Z * [new branch] gh/pearu/134/orig -> origin/gh/pearu/134/orig 2025-10-10T01:18:33.8932251Z * [new branch] gh/pearu/135/base -> origin/gh/pearu/135/base 2025-10-10T01:18:33.8933775Z * [new branch] gh/pearu/135/head -> origin/gh/pearu/135/head 2025-10-10T01:18:33.8935414Z * [new branch] gh/pearu/135/orig -> origin/gh/pearu/135/orig 2025-10-10T01:18:33.8937928Z * [new branch] gh/pearu/136/base -> origin/gh/pearu/136/base 2025-10-10T01:18:33.8939640Z * [new branch] gh/pearu/136/head -> origin/gh/pearu/136/head 2025-10-10T01:18:33.8941268Z * [new branch] gh/pearu/136/orig -> origin/gh/pearu/136/orig 2025-10-10T01:18:33.8943561Z * [new branch] gh/pearu/137/base -> origin/gh/pearu/137/base 2025-10-10T01:18:33.8945256Z * [new branch] gh/pearu/137/head -> origin/gh/pearu/137/head 2025-10-10T01:18:33.8946879Z * [new branch] gh/pearu/137/orig -> origin/gh/pearu/137/orig 2025-10-10T01:18:33.8949405Z * [new branch] gh/pearu/138/base -> origin/gh/pearu/138/base 2025-10-10T01:18:33.8951174Z * [new branch] gh/pearu/138/head -> origin/gh/pearu/138/head 2025-10-10T01:18:33.8952857Z * [new branch] gh/pearu/138/orig -> origin/gh/pearu/138/orig 2025-10-10T01:18:33.8955182Z * [new branch] gh/pearu/139/base -> origin/gh/pearu/139/base 2025-10-10T01:18:33.8956836Z * [new branch] gh/pearu/139/head -> origin/gh/pearu/139/head 2025-10-10T01:18:33.8958504Z * [new branch] gh/pearu/139/orig -> origin/gh/pearu/139/orig 2025-10-10T01:18:33.8961289Z * [new branch] gh/pearu/56/base -> origin/gh/pearu/56/base 2025-10-10T01:18:33.8963246Z * [new branch] gh/pearu/56/head -> origin/gh/pearu/56/head 2025-10-10T01:18:33.8964854Z * [new branch] gh/pearu/56/orig -> origin/gh/pearu/56/orig 2025-10-10T01:18:33.8967272Z * [new branch] gh/pearu/97/base -> origin/gh/pearu/97/base 2025-10-10T01:18:33.8968971Z * [new branch] gh/pearu/97/head -> origin/gh/pearu/97/head 2025-10-10T01:18:33.8970633Z * [new branch] gh/pearu/97/orig -> origin/gh/pearu/97/orig 2025-10-10T01:18:33.8973408Z * [new branch] gh/pianpwk/1/base -> origin/gh/pianpwk/1/base 2025-10-10T01:18:33.8975045Z * [new branch] gh/pianpwk/1/head -> origin/gh/pianpwk/1/head 2025-10-10T01:18:33.8976754Z * [new branch] gh/pianpwk/1/orig -> origin/gh/pianpwk/1/orig 2025-10-10T01:18:33.8978972Z * [new branch] gh/pianpwk/2/base -> origin/gh/pianpwk/2/base 2025-10-10T01:18:33.8980605Z * [new branch] gh/pianpwk/2/head -> origin/gh/pianpwk/2/head 2025-10-10T01:18:33.8982225Z * [new branch] gh/pianpwk/2/orig -> origin/gh/pianpwk/2/orig 2025-10-10T01:18:33.8984526Z * [new branch] gh/pianpwk/3/base -> origin/gh/pianpwk/3/base 2025-10-10T01:18:33.8986178Z * [new branch] gh/pianpwk/3/head -> origin/gh/pianpwk/3/head 2025-10-10T01:18:33.8987777Z * [new branch] gh/pianpwk/3/orig -> origin/gh/pianpwk/3/orig 2025-10-10T01:18:33.8990079Z * [new branch] gh/pianpwk/4/base -> origin/gh/pianpwk/4/base 2025-10-10T01:18:33.8991737Z * [new branch] gh/pianpwk/4/head -> origin/gh/pianpwk/4/head 2025-10-10T01:18:33.8993883Z * [new branch] gh/pianpwk/4/orig -> origin/gh/pianpwk/4/orig 2025-10-10T01:18:33.8996027Z * [new branch] gh/pianpwk/5/base -> origin/gh/pianpwk/5/base 2025-10-10T01:18:33.8997841Z * [new branch] gh/pianpwk/5/head -> origin/gh/pianpwk/5/head 2025-10-10T01:18:33.8999521Z * [new branch] gh/pianpwk/5/orig -> origin/gh/pianpwk/5/orig 2025-10-10T01:18:33.9001789Z * [new branch] gh/pianpwk/6/base -> origin/gh/pianpwk/6/base 2025-10-10T01:18:33.9003385Z * [new branch] gh/pianpwk/6/head -> origin/gh/pianpwk/6/head 2025-10-10T01:18:33.9005021Z * [new branch] gh/pianpwk/6/orig -> origin/gh/pianpwk/6/orig 2025-10-10T01:18:33.9007218Z * [new branch] gh/pianpwk/7/base -> origin/gh/pianpwk/7/base 2025-10-10T01:18:33.9008937Z * [new branch] gh/pianpwk/7/head -> origin/gh/pianpwk/7/head 2025-10-10T01:18:33.9010554Z * [new branch] gh/pianpwk/7/orig -> origin/gh/pianpwk/7/orig 2025-10-10T01:18:33.9012791Z * [new branch] gh/pianpwk/8/base -> origin/gh/pianpwk/8/base 2025-10-10T01:18:33.9014550Z * [new branch] gh/pianpwk/8/head -> origin/gh/pianpwk/8/head 2025-10-10T01:18:33.9016503Z * [new branch] gh/pianpwk/8/orig -> origin/gh/pianpwk/8/orig 2025-10-10T01:18:33.9019716Z * [new branch] gh/raymo/refresh-script -> origin/gh/raymo/refresh-script 2025-10-10T01:18:33.9021839Z * [new branch] gh/rec/141/base -> origin/gh/rec/141/base 2025-10-10T01:18:33.9023438Z * [new branch] gh/rec/141/head -> origin/gh/rec/141/head 2025-10-10T01:18:33.9027107Z * [new branch] gh/rec/153/base -> origin/gh/rec/153/base 2025-10-10T01:18:33.9028362Z * [new branch] gh/rec/153/head -> origin/gh/rec/153/head 2025-10-10T01:18:33.9029969Z * [new branch] gh/rec/153/orig -> origin/gh/rec/153/orig 2025-10-10T01:18:33.9032230Z * [new branch] gh/rec/154/base -> origin/gh/rec/154/base 2025-10-10T01:18:33.9034098Z * [new branch] gh/rec/154/head -> origin/gh/rec/154/head 2025-10-10T01:18:33.9035634Z * [new branch] gh/rec/154/orig -> origin/gh/rec/154/orig 2025-10-10T01:18:33.9037893Z * [new branch] gh/rec/162/base -> origin/gh/rec/162/base 2025-10-10T01:18:33.9039608Z * [new branch] gh/rec/162/head -> origin/gh/rec/162/head 2025-10-10T01:18:33.9041324Z * [new branch] gh/rec/162/orig -> origin/gh/rec/162/orig 2025-10-10T01:18:33.9043501Z * [new branch] gh/rec/164/base -> origin/gh/rec/164/base 2025-10-10T01:18:33.9045149Z * [new branch] gh/rec/164/head -> origin/gh/rec/164/head 2025-10-10T01:18:33.9046782Z * [new branch] gh/rec/164/orig -> origin/gh/rec/164/orig 2025-10-10T01:18:33.9049043Z * [new branch] gh/rec/166/base -> origin/gh/rec/166/base 2025-10-10T01:18:33.9050782Z * [new branch] gh/rec/166/head -> origin/gh/rec/166/head 2025-10-10T01:18:33.9052540Z * [new branch] gh/rec/166/orig -> origin/gh/rec/166/orig 2025-10-10T01:18:33.9055372Z * [new branch] gh/robert-hardwick/1/base -> origin/gh/robert-hardwick/1/base 2025-10-10T01:18:33.9057035Z * [new branch] gh/robert-hardwick/1/head -> origin/gh/robert-hardwick/1/head 2025-10-10T01:18:33.9058680Z * [new branch] gh/robert-hardwick/1/orig -> origin/gh/robert-hardwick/1/orig 2025-10-10T01:18:33.9060967Z * [new branch] gh/robert-hardwick/2/base -> origin/gh/robert-hardwick/2/base 2025-10-10T01:18:33.9062687Z * [new branch] gh/robert-hardwick/2/head -> origin/gh/robert-hardwick/2/head 2025-10-10T01:18:33.9064398Z * [new branch] gh/robert-hardwick/2/orig -> origin/gh/robert-hardwick/2/orig 2025-10-10T01:18:33.9067011Z * [new branch] gh/robert-hardwick/3/base -> origin/gh/robert-hardwick/3/base 2025-10-10T01:18:33.9069061Z * [new branch] gh/robert-hardwick/3/head -> origin/gh/robert-hardwick/3/head 2025-10-10T01:18:33.9070541Z * [new branch] gh/robert-hardwick/3/orig -> origin/gh/robert-hardwick/3/orig 2025-10-10T01:18:33.9072695Z * [new branch] gh/robert-hardwick/4/base -> origin/gh/robert-hardwick/4/base 2025-10-10T01:18:33.9074336Z * [new branch] gh/robert-hardwick/4/head -> origin/gh/robert-hardwick/4/head 2025-10-10T01:18:33.9076180Z * [new branch] gh/robert-hardwick/4/orig -> origin/gh/robert-hardwick/4/orig 2025-10-10T01:18:33.9079091Z * [new branch] gh/rtimpe/1/base -> origin/gh/rtimpe/1/base 2025-10-10T01:18:33.9080847Z * [new branch] gh/rtimpe/1/head -> origin/gh/rtimpe/1/head 2025-10-10T01:18:33.9083197Z * [new branch] gh/rtimpe/11/base -> origin/gh/rtimpe/11/base 2025-10-10T01:18:33.9084933Z * [new branch] gh/rtimpe/11/head -> origin/gh/rtimpe/11/head 2025-10-10T01:18:33.9086935Z * [new branch] gh/rtimpe/11/orig -> origin/gh/rtimpe/11/orig 2025-10-10T01:18:33.9089155Z * [new branch] gh/rtimpe/15/base -> origin/gh/rtimpe/15/base 2025-10-10T01:18:33.9090880Z * [new branch] gh/rtimpe/15/head -> origin/gh/rtimpe/15/head 2025-10-10T01:18:33.9092559Z * [new branch] gh/rtimpe/15/orig -> origin/gh/rtimpe/15/orig 2025-10-10T01:18:33.9094968Z * [new branch] gh/rtimpe/16/base -> origin/gh/rtimpe/16/base 2025-10-10T01:18:33.9096793Z * [new branch] gh/rtimpe/16/head -> origin/gh/rtimpe/16/head 2025-10-10T01:18:33.9099064Z * [new branch] gh/rtimpe/16/orig -> origin/gh/rtimpe/16/orig 2025-10-10T01:18:33.9101365Z * [new branch] gh/rtimpe/17/base -> origin/gh/rtimpe/17/base 2025-10-10T01:18:33.9103229Z * [new branch] gh/rtimpe/17/head -> origin/gh/rtimpe/17/head 2025-10-10T01:18:33.9105158Z * [new branch] gh/rtimpe/17/orig -> origin/gh/rtimpe/17/orig 2025-10-10T01:18:33.9107994Z * [new branch] gh/rtimpe/18/base -> origin/gh/rtimpe/18/base 2025-10-10T01:18:33.9109747Z * [new branch] gh/rtimpe/18/head -> origin/gh/rtimpe/18/head 2025-10-10T01:18:33.9111598Z * [new branch] gh/rtimpe/18/orig -> origin/gh/rtimpe/18/orig 2025-10-10T01:18:33.9113815Z * [new branch] gh/rtimpe/2/base -> origin/gh/rtimpe/2/base 2025-10-10T01:18:33.9115468Z * [new branch] gh/rtimpe/2/head -> origin/gh/rtimpe/2/head 2025-10-10T01:18:33.9117628Z * [new branch] gh/rtimpe/3/base -> origin/gh/rtimpe/3/base 2025-10-10T01:18:33.9119399Z * [new branch] gh/rtimpe/3/head -> origin/gh/rtimpe/3/head 2025-10-10T01:18:33.9121644Z * [new branch] gh/rtimpe/4/base -> origin/gh/rtimpe/4/base 2025-10-10T01:18:33.9123246Z * [new branch] gh/rtimpe/4/head -> origin/gh/rtimpe/4/head 2025-10-10T01:18:33.9126336Z * [new branch] gh/ruisizhang123/1/base -> origin/gh/ruisizhang123/1/base 2025-10-10T01:18:33.9128040Z * [new branch] gh/ruisizhang123/1/head -> origin/gh/ruisizhang123/1/head 2025-10-10T01:18:33.9129611Z * [new branch] gh/ruisizhang123/1/orig -> origin/gh/ruisizhang123/1/orig 2025-10-10T01:18:33.9131921Z * [new branch] gh/ruisizhang123/4/base -> origin/gh/ruisizhang123/4/base 2025-10-10T01:18:33.9133617Z * [new branch] gh/ruisizhang123/4/head -> origin/gh/ruisizhang123/4/head 2025-10-10T01:18:33.9135247Z * [new branch] gh/ruisizhang123/4/orig -> origin/gh/ruisizhang123/4/orig 2025-10-10T01:18:33.9137474Z * [new branch] gh/ruisizhang123/5/base -> origin/gh/ruisizhang123/5/base 2025-10-10T01:18:33.9139177Z * [new branch] gh/ruisizhang123/5/head -> origin/gh/ruisizhang123/5/head 2025-10-10T01:18:33.9140806Z * [new branch] gh/ruisizhang123/5/orig -> origin/gh/ruisizhang123/5/orig 2025-10-10T01:18:33.9143034Z * [new branch] gh/ruisizhang123/6/base -> origin/gh/ruisizhang123/6/base 2025-10-10T01:18:33.9144713Z * [new branch] gh/ruisizhang123/6/head -> origin/gh/ruisizhang123/6/head 2025-10-10T01:18:33.9146389Z * [new branch] gh/ruisizhang123/6/orig -> origin/gh/ruisizhang123/6/orig 2025-10-10T01:18:33.9148624Z * [new branch] gh/ruisizhang123/7/base -> origin/gh/ruisizhang123/7/base 2025-10-10T01:18:33.9150277Z * [new branch] gh/ruisizhang123/7/head -> origin/gh/ruisizhang123/7/head 2025-10-10T01:18:33.9152131Z * [new branch] gh/ruisizhang123/7/orig -> origin/gh/ruisizhang123/7/orig 2025-10-10T01:18:33.9154404Z * [new branch] gh/ruisizhang123/8/base -> origin/gh/ruisizhang123/8/base 2025-10-10T01:18:33.9156151Z * [new branch] gh/ruisizhang123/8/head -> origin/gh/ruisizhang123/8/head 2025-10-10T01:18:33.9157790Z * [new branch] gh/ruisizhang123/8/orig -> origin/gh/ruisizhang123/8/orig 2025-10-10T01:18:33.9160221Z * [new branch] gh/ruisizhang123/9/base -> origin/gh/ruisizhang123/9/base 2025-10-10T01:18:33.9161856Z * [new branch] gh/ruisizhang123/9/head -> origin/gh/ruisizhang123/9/head 2025-10-10T01:18:33.9163538Z * [new branch] gh/ruisizhang123/9/orig -> origin/gh/ruisizhang123/9/orig 2025-10-10T01:18:33.9166290Z * [new branch] gh/sarckk/2/base -> origin/gh/sarckk/2/base 2025-10-10T01:18:33.9168021Z * [new branch] gh/sarckk/2/head -> origin/gh/sarckk/2/head 2025-10-10T01:18:33.9169653Z * [new branch] gh/sarckk/2/orig -> origin/gh/sarckk/2/orig 2025-10-10T01:18:33.9172487Z * [new branch] gh/seemethere/35/base -> origin/gh/seemethere/35/base 2025-10-10T01:18:33.9174166Z * [new branch] gh/seemethere/35/head -> origin/gh/seemethere/35/head 2025-10-10T01:18:33.9176419Z * [new branch] gh/seemethere/35/orig -> origin/gh/seemethere/35/orig 2025-10-10T01:18:33.9178546Z * [new branch] gh/seemethere/37/base -> origin/gh/seemethere/37/base 2025-10-10T01:18:33.9180192Z * [new branch] gh/seemethere/37/head -> origin/gh/seemethere/37/head 2025-10-10T01:18:33.9181853Z * [new branch] gh/seemethere/37/orig -> origin/gh/seemethere/37/orig 2025-10-10T01:18:33.9184465Z * [new branch] gh/seemethere/43/base -> origin/gh/seemethere/43/base 2025-10-10T01:18:33.9185934Z * [new branch] gh/seemethere/43/head -> origin/gh/seemethere/43/head 2025-10-10T01:18:33.9187557Z * [new branch] gh/seemethere/43/orig -> origin/gh/seemethere/43/orig 2025-10-10T01:18:33.9189944Z * [new branch] gh/seemethere/44/base -> origin/gh/seemethere/44/base 2025-10-10T01:18:33.9191586Z * [new branch] gh/seemethere/44/head -> origin/gh/seemethere/44/head 2025-10-10T01:18:33.9193274Z * [new branch] gh/seemethere/44/orig -> origin/gh/seemethere/44/orig 2025-10-10T01:18:33.9195540Z * [new branch] gh/seemethere/48/base -> origin/gh/seemethere/48/base 2025-10-10T01:18:33.9197330Z * [new branch] gh/seemethere/48/head -> origin/gh/seemethere/48/head 2025-10-10T01:18:33.9199088Z * [new branch] gh/seemethere/48/orig -> origin/gh/seemethere/48/orig 2025-10-10T01:18:33.9201423Z * [new branch] gh/seemethere/49/base -> origin/gh/seemethere/49/base 2025-10-10T01:18:33.9203136Z * [new branch] gh/seemethere/49/head -> origin/gh/seemethere/49/head 2025-10-10T01:18:33.9204765Z * [new branch] gh/seemethere/49/orig -> origin/gh/seemethere/49/orig 2025-10-10T01:18:33.9207008Z * [new branch] gh/seemethere/52/base -> origin/gh/seemethere/52/base 2025-10-10T01:18:33.9208685Z * [new branch] gh/seemethere/52/head -> origin/gh/seemethere/52/head 2025-10-10T01:18:33.9210440Z * [new branch] gh/seemethere/52/orig -> origin/gh/seemethere/52/orig 2025-10-10T01:18:33.9212646Z * [new branch] gh/seemethere/53/base -> origin/gh/seemethere/53/base 2025-10-10T01:18:33.9214333Z * [new branch] gh/seemethere/53/head -> origin/gh/seemethere/53/head 2025-10-10T01:18:33.9216012Z * [new branch] gh/seemethere/53/orig -> origin/gh/seemethere/53/orig 2025-10-10T01:18:33.9218430Z * [new branch] gh/seemethere/54/base -> origin/gh/seemethere/54/base 2025-10-10T01:18:33.9220156Z * [new branch] gh/seemethere/54/head -> origin/gh/seemethere/54/head 2025-10-10T01:18:33.9221830Z * [new branch] gh/seemethere/54/orig -> origin/gh/seemethere/54/orig 2025-10-10T01:18:33.9224154Z * [new branch] gh/seemethere/55/base -> origin/gh/seemethere/55/base 2025-10-10T01:18:33.9225925Z * [new branch] gh/seemethere/55/head -> origin/gh/seemethere/55/head 2025-10-10T01:18:33.9227662Z * [new branch] gh/seemethere/55/orig -> origin/gh/seemethere/55/orig 2025-10-10T01:18:33.9229826Z * [new branch] gh/seemethere/59/base -> origin/gh/seemethere/59/base 2025-10-10T01:18:33.9231471Z * [new branch] gh/seemethere/59/head -> origin/gh/seemethere/59/head 2025-10-10T01:18:33.9233122Z * [new branch] gh/seemethere/59/orig -> origin/gh/seemethere/59/orig 2025-10-10T01:18:33.9235364Z * [new branch] gh/seemethere/62/base -> origin/gh/seemethere/62/base 2025-10-10T01:18:33.9237070Z * [new branch] gh/seemethere/62/head -> origin/gh/seemethere/62/head 2025-10-10T01:18:33.9239013Z * [new branch] gh/seemethere/62/orig -> origin/gh/seemethere/62/orig 2025-10-10T01:18:33.9241193Z * [new branch] gh/seemethere/63/base -> origin/gh/seemethere/63/base 2025-10-10T01:18:33.9242863Z * [new branch] gh/seemethere/63/head -> origin/gh/seemethere/63/head 2025-10-10T01:18:33.9244532Z * [new branch] gh/seemethere/63/orig -> origin/gh/seemethere/63/orig 2025-10-10T01:18:33.9246743Z * [new branch] gh/seemethere/64/base -> origin/gh/seemethere/64/base 2025-10-10T01:18:33.9248555Z * [new branch] gh/seemethere/64/head -> origin/gh/seemethere/64/head 2025-10-10T01:18:33.9250120Z * [new branch] gh/seemethere/64/orig -> origin/gh/seemethere/64/orig 2025-10-10T01:18:33.9252355Z * [new branch] gh/seemethere/65/base -> origin/gh/seemethere/65/base 2025-10-10T01:18:33.9254032Z * [new branch] gh/seemethere/65/head -> origin/gh/seemethere/65/head 2025-10-10T01:18:33.9255706Z * [new branch] gh/seemethere/65/orig -> origin/gh/seemethere/65/orig 2025-10-10T01:18:33.9258512Z * [new branch] gh/seemethere/66/base -> origin/gh/seemethere/66/base 2025-10-10T01:18:33.9259663Z * [new branch] gh/seemethere/66/head -> origin/gh/seemethere/66/head 2025-10-10T01:18:33.9261285Z * [new branch] gh/seemethere/66/orig -> origin/gh/seemethere/66/orig 2025-10-10T01:18:33.9263934Z * [new branch] gh/seemethere/67/base -> origin/gh/seemethere/67/base 2025-10-10T01:18:33.9265285Z * [new branch] gh/seemethere/67/head -> origin/gh/seemethere/67/head 2025-10-10T01:18:33.9267103Z * [new branch] gh/seemethere/67/orig -> origin/gh/seemethere/67/orig 2025-10-10T01:18:33.9269584Z * [new branch] gh/seemethere/68/base -> origin/gh/seemethere/68/base 2025-10-10T01:18:33.9271243Z * [new branch] gh/seemethere/68/head -> origin/gh/seemethere/68/head 2025-10-10T01:18:33.9272868Z * [new branch] gh/seemethere/68/orig -> origin/gh/seemethere/68/orig 2025-10-10T01:18:33.9275574Z * [new branch] gh/seemethere/69/base -> origin/gh/seemethere/69/base 2025-10-10T01:18:33.9277330Z * [new branch] gh/seemethere/69/head -> origin/gh/seemethere/69/head 2025-10-10T01:18:33.9279082Z * [new branch] gh/seemethere/69/orig -> origin/gh/seemethere/69/orig 2025-10-10T01:18:33.9281489Z * [new branch] gh/seemethere/70/base -> origin/gh/seemethere/70/base 2025-10-10T01:18:33.9283234Z * [new branch] gh/seemethere/70/head -> origin/gh/seemethere/70/head 2025-10-10T01:18:33.9285061Z * [new branch] gh/seemethere/70/orig -> origin/gh/seemethere/70/orig 2025-10-10T01:18:33.9287269Z * [new branch] gh/seemethere/71/base -> origin/gh/seemethere/71/base 2025-10-10T01:18:33.9289016Z * [new branch] gh/seemethere/71/head -> origin/gh/seemethere/71/head 2025-10-10T01:18:33.9290688Z * [new branch] gh/seemethere/71/orig -> origin/gh/seemethere/71/orig 2025-10-10T01:18:33.9293615Z * [new branch] gh/shunting314/145/base -> origin/gh/shunting314/145/base 2025-10-10T01:18:33.9295364Z * [new branch] gh/shunting314/145/head -> origin/gh/shunting314/145/head 2025-10-10T01:18:33.9300341Z * [new branch] gh/shunting314/145/orig -> origin/gh/shunting314/145/orig 2025-10-10T01:18:33.9302904Z * [new branch] gh/shunting314/176/base -> origin/gh/shunting314/176/base 2025-10-10T01:18:33.9304774Z * [new branch] gh/shunting314/176/head -> origin/gh/shunting314/176/head 2025-10-10T01:18:33.9306441Z * [new branch] gh/shunting314/176/orig -> origin/gh/shunting314/176/orig 2025-10-10T01:18:33.9309012Z * [new branch] gh/shunting314/211/base -> origin/gh/shunting314/211/base 2025-10-10T01:18:33.9310519Z * [new branch] gh/shunting314/211/head -> origin/gh/shunting314/211/head 2025-10-10T01:18:33.9312182Z * [new branch] gh/shunting314/211/orig -> origin/gh/shunting314/211/orig 2025-10-10T01:18:33.9315185Z * [new branch] gh/shunting314/212/base -> origin/gh/shunting314/212/base 2025-10-10T01:18:33.9317119Z * [new branch] gh/shunting314/212/head -> origin/gh/shunting314/212/head 2025-10-10T01:18:33.9319167Z * [new branch] gh/shunting314/212/orig -> origin/gh/shunting314/212/orig 2025-10-10T01:18:33.9322571Z * [new branch] gh/shunting314/213/base -> origin/gh/shunting314/213/base 2025-10-10T01:18:33.9323816Z * [new branch] gh/shunting314/213/head -> origin/gh/shunting314/213/head 2025-10-10T01:18:33.9327429Z * [new branch] gh/shunting314/213/orig -> origin/gh/shunting314/213/orig 2025-10-10T01:18:33.9329630Z * [new branch] gh/shunting314/215/base -> origin/gh/shunting314/215/base 2025-10-10T01:18:33.9331221Z * [new branch] gh/shunting314/215/head -> origin/gh/shunting314/215/head 2025-10-10T01:18:33.9332904Z * [new branch] gh/shunting314/215/orig -> origin/gh/shunting314/215/orig 2025-10-10T01:18:33.9335239Z * [new branch] gh/shunting314/216/base -> origin/gh/shunting314/216/base 2025-10-10T01:18:33.9336973Z * [new branch] gh/shunting314/216/head -> origin/gh/shunting314/216/head 2025-10-10T01:18:33.9338647Z * [new branch] gh/shunting314/216/orig -> origin/gh/shunting314/216/orig 2025-10-10T01:18:33.9341040Z * [new branch] gh/shunting314/217/base -> origin/gh/shunting314/217/base 2025-10-10T01:18:33.9342648Z * [new branch] gh/shunting314/217/head -> origin/gh/shunting314/217/head 2025-10-10T01:18:33.9344318Z * [new branch] gh/shunting314/217/orig -> origin/gh/shunting314/217/orig 2025-10-10T01:18:33.9347640Z * [new branch] gh/shunting314/218/base -> origin/gh/shunting314/218/base 2025-10-10T01:18:33.9348911Z * [new branch] gh/shunting314/218/head -> origin/gh/shunting314/218/head 2025-10-10T01:18:33.9350687Z * [new branch] gh/shunting314/218/orig -> origin/gh/shunting314/218/orig 2025-10-10T01:18:33.9352793Z * [new branch] gh/shunting314/219/base -> origin/gh/shunting314/219/base 2025-10-10T01:18:33.9354514Z * [new branch] gh/shunting314/219/head -> origin/gh/shunting314/219/head 2025-10-10T01:18:33.9356628Z * [new branch] gh/shunting314/219/orig -> origin/gh/shunting314/219/orig 2025-10-10T01:18:33.9359030Z * [new branch] gh/shunting314/223/base -> origin/gh/shunting314/223/base 2025-10-10T01:18:33.9360911Z * [new branch] gh/shunting314/223/head -> origin/gh/shunting314/223/head 2025-10-10T01:18:33.9362450Z * [new branch] gh/shunting314/223/orig -> origin/gh/shunting314/223/orig 2025-10-10T01:18:33.9364727Z * [new branch] gh/shunting314/224/base -> origin/gh/shunting314/224/base 2025-10-10T01:18:33.9366571Z * [new branch] gh/shunting314/224/head -> origin/gh/shunting314/224/head 2025-10-10T01:18:33.9368597Z * [new branch] gh/shunting314/224/orig -> origin/gh/shunting314/224/orig 2025-10-10T01:18:33.9370759Z * [new branch] gh/shunting314/225/base -> origin/gh/shunting314/225/base 2025-10-10T01:18:33.9372384Z * [new branch] gh/shunting314/225/head -> origin/gh/shunting314/225/head 2025-10-10T01:18:33.9374038Z * [new branch] gh/shunting314/225/orig -> origin/gh/shunting314/225/orig 2025-10-10T01:18:33.9377149Z * [new branch] gh/shunting314/226/base -> origin/gh/shunting314/226/base 2025-10-10T01:18:33.9378685Z * [new branch] gh/shunting314/226/head -> origin/gh/shunting314/226/head 2025-10-10T01:18:33.9380137Z * [new branch] gh/shunting314/226/orig -> origin/gh/shunting314/226/orig 2025-10-10T01:18:33.9382445Z * [new branch] gh/shunting314/227/base -> origin/gh/shunting314/227/base 2025-10-10T01:18:33.9384093Z * [new branch] gh/shunting314/227/head -> origin/gh/shunting314/227/head 2025-10-10T01:18:33.9385903Z * [new branch] gh/shunting314/227/orig -> origin/gh/shunting314/227/orig 2025-10-10T01:18:33.9388879Z * [new branch] gh/shunting314/228/base -> origin/gh/shunting314/228/base 2025-10-10T01:18:33.9390062Z * [new branch] gh/shunting314/228/head -> origin/gh/shunting314/228/head 2025-10-10T01:18:33.9391711Z * [new branch] gh/shunting314/228/orig -> origin/gh/shunting314/228/orig 2025-10-10T01:18:33.9395267Z * [new branch] gh/shunting314/229/base -> origin/gh/shunting314/229/base 2025-10-10T01:18:33.9396778Z * [new branch] gh/shunting314/229/head -> origin/gh/shunting314/229/head 2025-10-10T01:18:33.9398527Z * [new branch] gh/shunting314/229/orig -> origin/gh/shunting314/229/orig 2025-10-10T01:18:33.9401071Z * [new branch] gh/shunting314/230/base -> origin/gh/shunting314/230/base 2025-10-10T01:18:33.9402724Z * [new branch] gh/shunting314/230/head -> origin/gh/shunting314/230/head 2025-10-10T01:18:33.9404370Z * [new branch] gh/shunting314/230/orig -> origin/gh/shunting314/230/orig 2025-10-10T01:18:33.9406757Z * [new branch] gh/shunting314/231/base -> origin/gh/shunting314/231/base 2025-10-10T01:18:33.9408478Z * [new branch] gh/shunting314/231/head -> origin/gh/shunting314/231/head 2025-10-10T01:18:33.9410129Z * [new branch] gh/shunting314/231/orig -> origin/gh/shunting314/231/orig 2025-10-10T01:18:33.9412574Z * [new branch] gh/shunting314/232/base -> origin/gh/shunting314/232/base 2025-10-10T01:18:33.9414263Z * [new branch] gh/shunting314/232/head -> origin/gh/shunting314/232/head 2025-10-10T01:18:33.9415843Z * [new branch] gh/shunting314/232/orig -> origin/gh/shunting314/232/orig 2025-10-10T01:18:33.9418157Z * [new branch] gh/shunting314/233/base -> origin/gh/shunting314/233/base 2025-10-10T01:18:33.9419756Z * [new branch] gh/shunting314/233/head -> origin/gh/shunting314/233/head 2025-10-10T01:18:33.9421448Z * [new branch] gh/shunting314/233/orig -> origin/gh/shunting314/233/orig 2025-10-10T01:18:33.9423766Z * [new branch] gh/shunting314/234/base -> origin/gh/shunting314/234/base 2025-10-10T01:18:33.9425854Z * [new branch] gh/shunting314/234/head -> origin/gh/shunting314/234/head 2025-10-10T01:18:33.9427612Z * [new branch] gh/shunting314/234/orig -> origin/gh/shunting314/234/orig 2025-10-10T01:18:33.9430025Z * [new branch] gh/shunting314/235/base -> origin/gh/shunting314/235/base 2025-10-10T01:18:33.9431618Z * [new branch] gh/shunting314/235/head -> origin/gh/shunting314/235/head 2025-10-10T01:18:33.9433222Z * [new branch] gh/shunting314/235/orig -> origin/gh/shunting314/235/orig 2025-10-10T01:18:33.9436026Z * [new branch] gh/silverguo/1/base -> origin/gh/silverguo/1/base 2025-10-10T01:18:33.9437737Z * [new branch] gh/silverguo/1/head -> origin/gh/silverguo/1/head 2025-10-10T01:18:33.9439987Z * [new branch] gh/silverguo/2/base -> origin/gh/silverguo/2/base 2025-10-10T01:18:33.9441576Z * [new branch] gh/silverguo/2/head -> origin/gh/silverguo/2/head 2025-10-10T01:18:33.9443696Z * [new branch] gh/silverguo/3/base -> origin/gh/silverguo/3/base 2025-10-10T01:18:33.9445336Z * [new branch] gh/silverguo/3/head -> origin/gh/silverguo/3/head 2025-10-10T01:18:33.9447606Z * [new branch] gh/silverguo/4/base -> origin/gh/silverguo/4/base 2025-10-10T01:18:33.9449216Z * [new branch] gh/silverguo/4/head -> origin/gh/silverguo/4/head 2025-10-10T01:18:33.9451893Z * [new branch] gh/sinhaanhsul/1/base -> origin/gh/sinhaanhsul/1/base 2025-10-10T01:18:33.9453585Z * [new branch] gh/sinhaanhsul/1/head -> origin/gh/sinhaanhsul/1/head 2025-10-10T01:18:33.9456318Z * [new branch] gh/slayton58/1/base -> origin/gh/slayton58/1/base 2025-10-10T01:18:33.9457965Z * [new branch] gh/slayton58/1/head -> origin/gh/slayton58/1/head 2025-10-10T01:18:33.9459671Z * [new branch] gh/slayton58/1/orig -> origin/gh/slayton58/1/orig 2025-10-10T01:18:33.9461946Z * [new branch] gh/slayton58/10/base -> origin/gh/slayton58/10/base 2025-10-10T01:18:33.9463593Z * [new branch] gh/slayton58/10/head -> origin/gh/slayton58/10/head 2025-10-10T01:18:33.9465679Z * [new branch] gh/slayton58/10/orig -> origin/gh/slayton58/10/orig 2025-10-10T01:18:33.9467908Z * [new branch] gh/slayton58/11/base -> origin/gh/slayton58/11/base 2025-10-10T01:18:33.9469709Z * [new branch] gh/slayton58/11/head -> origin/gh/slayton58/11/head 2025-10-10T01:18:33.9471364Z * [new branch] gh/slayton58/11/orig -> origin/gh/slayton58/11/orig 2025-10-10T01:18:33.9473498Z * [new branch] gh/slayton58/12/base -> origin/gh/slayton58/12/base 2025-10-10T01:18:33.9475230Z * [new branch] gh/slayton58/12/head -> origin/gh/slayton58/12/head 2025-10-10T01:18:33.9477006Z * [new branch] gh/slayton58/12/orig -> origin/gh/slayton58/12/orig 2025-10-10T01:18:33.9479439Z * [new branch] gh/slayton58/13/base -> origin/gh/slayton58/13/base 2025-10-10T01:18:33.9481105Z * [new branch] gh/slayton58/13/head -> origin/gh/slayton58/13/head 2025-10-10T01:18:33.9482816Z * [new branch] gh/slayton58/13/orig -> origin/gh/slayton58/13/orig 2025-10-10T01:18:33.9485039Z * [new branch] gh/slayton58/14/base -> origin/gh/slayton58/14/base 2025-10-10T01:18:33.9486693Z * [new branch] gh/slayton58/14/head -> origin/gh/slayton58/14/head 2025-10-10T01:18:33.9488369Z * [new branch] gh/slayton58/14/orig -> origin/gh/slayton58/14/orig 2025-10-10T01:18:33.9490752Z * [new branch] gh/slayton58/15/base -> origin/gh/slayton58/15/base 2025-10-10T01:18:33.9492420Z * [new branch] gh/slayton58/15/head -> origin/gh/slayton58/15/head 2025-10-10T01:18:33.9494095Z * [new branch] gh/slayton58/15/orig -> origin/gh/slayton58/15/orig 2025-10-10T01:18:33.9496804Z * [new branch] gh/slayton58/16/base -> origin/gh/slayton58/16/base 2025-10-10T01:18:33.9498461Z * [new branch] gh/slayton58/16/head -> origin/gh/slayton58/16/head 2025-10-10T01:18:33.9500246Z * [new branch] gh/slayton58/16/orig -> origin/gh/slayton58/16/orig 2025-10-10T01:18:33.9502621Z * [new branch] gh/slayton58/17/base -> origin/gh/slayton58/17/base 2025-10-10T01:18:33.9504283Z * [new branch] gh/slayton58/17/head -> origin/gh/slayton58/17/head 2025-10-10T01:18:33.9506119Z * [new branch] gh/slayton58/17/orig -> origin/gh/slayton58/17/orig 2025-10-10T01:18:33.9508570Z * [new branch] gh/slayton58/18/base -> origin/gh/slayton58/18/base 2025-10-10T01:18:33.9510626Z * [new branch] gh/slayton58/18/head -> origin/gh/slayton58/18/head 2025-10-10T01:18:33.9512707Z * [new branch] gh/slayton58/19/base -> origin/gh/slayton58/19/base 2025-10-10T01:18:33.9514404Z * [new branch] gh/slayton58/19/head -> origin/gh/slayton58/19/head 2025-10-10T01:18:33.9516485Z * [new branch] gh/slayton58/19/orig -> origin/gh/slayton58/19/orig 2025-10-10T01:18:33.9518587Z * [new branch] gh/slayton58/2/base -> origin/gh/slayton58/2/base 2025-10-10T01:18:33.9520355Z * [new branch] gh/slayton58/2/head -> origin/gh/slayton58/2/head 2025-10-10T01:18:33.9522087Z * [new branch] gh/slayton58/2/orig -> origin/gh/slayton58/2/orig 2025-10-10T01:18:33.9524539Z * [new branch] gh/slayton58/20/base -> origin/gh/slayton58/20/base 2025-10-10T01:18:33.9527887Z * [new branch] gh/slayton58/20/head -> origin/gh/slayton58/20/head 2025-10-10T01:18:33.9529881Z * [new branch] gh/slayton58/20/orig -> origin/gh/slayton58/20/orig 2025-10-10T01:18:33.9532648Z * [new branch] gh/slayton58/21/base -> origin/gh/slayton58/21/base 2025-10-10T01:18:33.9533832Z * [new branch] gh/slayton58/21/head -> origin/gh/slayton58/21/head 2025-10-10T01:18:33.9535565Z * [new branch] gh/slayton58/21/orig -> origin/gh/slayton58/21/orig 2025-10-10T01:18:33.9538107Z * [new branch] gh/slayton58/22/base -> origin/gh/slayton58/22/base 2025-10-10T01:18:33.9539825Z * [new branch] gh/slayton58/22/head -> origin/gh/slayton58/22/head 2025-10-10T01:18:33.9541587Z * [new branch] gh/slayton58/22/orig -> origin/gh/slayton58/22/orig 2025-10-10T01:18:33.9543955Z * [new branch] gh/slayton58/23/base -> origin/gh/slayton58/23/base 2025-10-10T01:18:33.9545639Z * [new branch] gh/slayton58/23/head -> origin/gh/slayton58/23/head 2025-10-10T01:18:33.9547416Z * [new branch] gh/slayton58/23/orig -> origin/gh/slayton58/23/orig 2025-10-10T01:18:33.9549673Z * [new branch] gh/slayton58/24/base -> origin/gh/slayton58/24/base 2025-10-10T01:18:33.9551307Z * [new branch] gh/slayton58/24/head -> origin/gh/slayton58/24/head 2025-10-10T01:18:33.9552993Z * [new branch] gh/slayton58/24/orig -> origin/gh/slayton58/24/orig 2025-10-10T01:18:33.9555299Z * [new branch] gh/slayton58/25/base -> origin/gh/slayton58/25/base 2025-10-10T01:18:33.9556977Z * [new branch] gh/slayton58/25/head -> origin/gh/slayton58/25/head 2025-10-10T01:18:33.9558739Z * [new branch] gh/slayton58/25/orig -> origin/gh/slayton58/25/orig 2025-10-10T01:18:33.9561896Z * [new branch] gh/slayton58/26/base -> origin/gh/slayton58/26/base 2025-10-10T01:18:33.9563553Z * [new branch] gh/slayton58/26/head -> origin/gh/slayton58/26/head 2025-10-10T01:18:33.9565186Z * [new branch] gh/slayton58/26/orig -> origin/gh/slayton58/26/orig 2025-10-10T01:18:33.9567720Z * [new branch] gh/slayton58/3/base -> origin/gh/slayton58/3/base 2025-10-10T01:18:33.9569393Z * [new branch] gh/slayton58/3/head -> origin/gh/slayton58/3/head 2025-10-10T01:18:33.9571146Z * [new branch] gh/slayton58/3/orig -> origin/gh/slayton58/3/orig 2025-10-10T01:18:33.9573485Z * [new branch] gh/slayton58/4/base -> origin/gh/slayton58/4/base 2025-10-10T01:18:33.9575123Z * [new branch] gh/slayton58/4/head -> origin/gh/slayton58/4/head 2025-10-10T01:18:33.9576837Z * [new branch] gh/slayton58/4/orig -> origin/gh/slayton58/4/orig 2025-10-10T01:18:33.9579596Z * [new branch] gh/slayton58/5/base -> origin/gh/slayton58/5/base 2025-10-10T01:18:33.9581318Z * [new branch] gh/slayton58/5/head -> origin/gh/slayton58/5/head 2025-10-10T01:18:33.9583024Z * [new branch] gh/slayton58/5/orig -> origin/gh/slayton58/5/orig 2025-10-10T01:18:33.9585309Z * [new branch] gh/slayton58/6/base -> origin/gh/slayton58/6/base 2025-10-10T01:18:33.9587234Z * [new branch] gh/slayton58/6/head -> origin/gh/slayton58/6/head 2025-10-10T01:18:33.9589455Z * [new branch] gh/slayton58/7/base -> origin/gh/slayton58/7/base 2025-10-10T01:18:33.9591081Z * [new branch] gh/slayton58/7/head -> origin/gh/slayton58/7/head 2025-10-10T01:18:33.9593302Z * [new branch] gh/slayton58/8/base -> origin/gh/slayton58/8/base 2025-10-10T01:18:33.9595097Z * [new branch] gh/slayton58/8/head -> origin/gh/slayton58/8/head 2025-10-10T01:18:33.9596733Z * [new branch] gh/slayton58/8/orig -> origin/gh/slayton58/8/orig 2025-10-10T01:18:33.9599100Z * [new branch] gh/slayton58/9/base -> origin/gh/slayton58/9/base 2025-10-10T01:18:33.9600773Z * [new branch] gh/slayton58/9/head -> origin/gh/slayton58/9/head 2025-10-10T01:18:33.9602499Z * [new branch] gh/slayton58/9/orig -> origin/gh/slayton58/9/orig 2025-10-10T01:18:33.9605498Z * [new branch] gh/soulitzer/269/base -> origin/gh/soulitzer/269/base 2025-10-10T01:18:33.9607143Z * [new branch] gh/soulitzer/269/head -> origin/gh/soulitzer/269/head 2025-10-10T01:18:33.9608934Z * [new branch] gh/soulitzer/269/orig -> origin/gh/soulitzer/269/orig 2025-10-10T01:18:33.9611397Z * [new branch] gh/soulitzer/276/base -> origin/gh/soulitzer/276/base 2025-10-10T01:18:33.9613120Z * [new branch] gh/soulitzer/276/head -> origin/gh/soulitzer/276/head 2025-10-10T01:18:33.9614823Z * [new branch] gh/soulitzer/276/orig -> origin/gh/soulitzer/276/orig 2025-10-10T01:18:33.9617405Z * [new branch] gh/soulitzer/287/base -> origin/gh/soulitzer/287/base 2025-10-10T01:18:33.9619080Z * [new branch] gh/soulitzer/287/head -> origin/gh/soulitzer/287/head 2025-10-10T01:18:33.9620787Z * [new branch] gh/soulitzer/287/orig -> origin/gh/soulitzer/287/orig 2025-10-10T01:18:33.9623114Z * [new branch] gh/soulitzer/296/base -> origin/gh/soulitzer/296/base 2025-10-10T01:18:33.9625206Z * [new branch] gh/soulitzer/296/head -> origin/gh/soulitzer/296/head 2025-10-10T01:18:33.9626978Z * [new branch] gh/soulitzer/296/orig -> origin/gh/soulitzer/296/orig 2025-10-10T01:18:33.9629261Z * [new branch] gh/soulitzer/299/base -> origin/gh/soulitzer/299/base 2025-10-10T01:18:33.9630995Z * [new branch] gh/soulitzer/299/head -> origin/gh/soulitzer/299/head 2025-10-10T01:18:33.9632700Z * [new branch] gh/soulitzer/299/orig -> origin/gh/soulitzer/299/orig 2025-10-10T01:18:33.9635196Z * [new branch] gh/soulitzer/300/base -> origin/gh/soulitzer/300/base 2025-10-10T01:18:33.9637014Z * [new branch] gh/soulitzer/300/head -> origin/gh/soulitzer/300/head 2025-10-10T01:18:33.9638764Z * [new branch] gh/soulitzer/300/orig -> origin/gh/soulitzer/300/orig 2025-10-10T01:18:33.9641218Z * [new branch] gh/soulitzer/301/base -> origin/gh/soulitzer/301/base 2025-10-10T01:18:33.9642936Z * [new branch] gh/soulitzer/301/head -> origin/gh/soulitzer/301/head 2025-10-10T01:18:33.9644673Z * [new branch] gh/soulitzer/301/orig -> origin/gh/soulitzer/301/orig 2025-10-10T01:18:33.9646971Z * [new branch] gh/soulitzer/313/base -> origin/gh/soulitzer/313/base 2025-10-10T01:18:33.9648679Z * [new branch] gh/soulitzer/313/head -> origin/gh/soulitzer/313/head 2025-10-10T01:18:33.9650361Z * [new branch] gh/soulitzer/313/orig -> origin/gh/soulitzer/313/orig 2025-10-10T01:18:33.9652633Z * [new branch] gh/soulitzer/319/base -> origin/gh/soulitzer/319/base 2025-10-10T01:18:33.9654323Z * [new branch] gh/soulitzer/319/head -> origin/gh/soulitzer/319/head 2025-10-10T01:18:33.9656596Z * [new branch] gh/soulitzer/319/orig -> origin/gh/soulitzer/319/orig 2025-10-10T01:18:33.9659532Z * [new branch] gh/soulitzer/320/base -> origin/gh/soulitzer/320/base 2025-10-10T01:18:33.9660627Z * [new branch] gh/soulitzer/320/head -> origin/gh/soulitzer/320/head 2025-10-10T01:18:33.9662294Z * [new branch] gh/soulitzer/320/orig -> origin/gh/soulitzer/320/orig 2025-10-10T01:18:33.9665336Z * [new branch] gh/soulitzer/336/base -> origin/gh/soulitzer/336/base 2025-10-10T01:18:33.9666585Z * [new branch] gh/soulitzer/336/head -> origin/gh/soulitzer/336/head 2025-10-10T01:18:33.9668324Z * [new branch] gh/soulitzer/336/orig -> origin/gh/soulitzer/336/orig 2025-10-10T01:18:33.9671128Z * [new branch] gh/soulitzer/347/base -> origin/gh/soulitzer/347/base 2025-10-10T01:18:33.9672804Z * [new branch] gh/soulitzer/347/head -> origin/gh/soulitzer/347/head 2025-10-10T01:18:33.9674456Z * [new branch] gh/soulitzer/347/orig -> origin/gh/soulitzer/347/orig 2025-10-10T01:18:33.9677788Z * [new branch] gh/soulitzer/349/base -> origin/gh/soulitzer/349/base 2025-10-10T01:18:33.9679223Z * [new branch] gh/soulitzer/349/head -> origin/gh/soulitzer/349/head 2025-10-10T01:18:33.9680909Z * [new branch] gh/soulitzer/349/orig -> origin/gh/soulitzer/349/orig 2025-10-10T01:18:33.9683349Z * [new branch] gh/soulitzer/350/base -> origin/gh/soulitzer/350/base 2025-10-10T01:18:33.9684803Z * [new branch] gh/soulitzer/350/head -> origin/gh/soulitzer/350/head 2025-10-10T01:18:33.9686751Z * [new branch] gh/soulitzer/350/orig -> origin/gh/soulitzer/350/orig 2025-10-10T01:18:33.9689156Z * [new branch] gh/soulitzer/351/base -> origin/gh/soulitzer/351/base 2025-10-10T01:18:33.9690895Z * [new branch] gh/soulitzer/351/head -> origin/gh/soulitzer/351/head 2025-10-10T01:18:33.9692556Z * [new branch] gh/soulitzer/351/orig -> origin/gh/soulitzer/351/orig 2025-10-10T01:18:33.9694823Z * [new branch] gh/soulitzer/353/base -> origin/gh/soulitzer/353/base 2025-10-10T01:18:33.9696678Z * [new branch] gh/soulitzer/353/head -> origin/gh/soulitzer/353/head 2025-10-10T01:18:33.9698367Z * [new branch] gh/soulitzer/353/orig -> origin/gh/soulitzer/353/orig 2025-10-10T01:18:33.9701923Z * [new branch] gh/soulitzer/358/base -> origin/gh/soulitzer/358/base 2025-10-10T01:18:33.9703843Z * [new branch] gh/soulitzer/358/head -> origin/gh/soulitzer/358/head 2025-10-10T01:18:33.9705605Z * [new branch] gh/soulitzer/358/orig -> origin/gh/soulitzer/358/orig 2025-10-10T01:18:33.9708551Z * [new branch] gh/soulitzer/359/base -> origin/gh/soulitzer/359/base 2025-10-10T01:18:33.9710259Z * [new branch] gh/soulitzer/359/head -> origin/gh/soulitzer/359/head 2025-10-10T01:18:33.9711941Z * [new branch] gh/soulitzer/359/orig -> origin/gh/soulitzer/359/orig 2025-10-10T01:18:33.9714368Z * [new branch] gh/soulitzer/372/base -> origin/gh/soulitzer/372/base 2025-10-10T01:18:33.9716038Z * [new branch] gh/soulitzer/372/head -> origin/gh/soulitzer/372/head 2025-10-10T01:18:33.9717716Z * [new branch] gh/soulitzer/372/orig -> origin/gh/soulitzer/372/orig 2025-10-10T01:18:33.9720381Z * [new branch] gh/soulitzer/374/base -> origin/gh/soulitzer/374/base 2025-10-10T01:18:33.9722033Z * [new branch] gh/soulitzer/374/head -> origin/gh/soulitzer/374/head 2025-10-10T01:18:33.9723696Z * [new branch] gh/soulitzer/374/orig -> origin/gh/soulitzer/374/orig 2025-10-10T01:18:33.9726530Z * [new branch] gh/soulitzer/375/base -> origin/gh/soulitzer/375/base 2025-10-10T01:18:33.9728068Z * [new branch] gh/soulitzer/375/head -> origin/gh/soulitzer/375/head 2025-10-10T01:18:33.9729732Z * [new branch] gh/soulitzer/375/orig -> origin/gh/soulitzer/375/orig 2025-10-10T01:18:33.9732126Z * [new branch] gh/soulitzer/380/base -> origin/gh/soulitzer/380/base 2025-10-10T01:18:33.9733842Z * [new branch] gh/soulitzer/380/head -> origin/gh/soulitzer/380/head 2025-10-10T01:18:33.9735659Z * [new branch] gh/soulitzer/380/orig -> origin/gh/soulitzer/380/orig 2025-10-10T01:18:33.9738953Z * [new branch] gh/soulitzer/381/base -> origin/gh/soulitzer/381/base 2025-10-10T01:18:33.9740073Z * [new branch] gh/soulitzer/381/head -> origin/gh/soulitzer/381/head 2025-10-10T01:18:33.9741737Z * [new branch] gh/soulitzer/381/orig -> origin/gh/soulitzer/381/orig 2025-10-10T01:18:33.9744622Z * [new branch] gh/soulitzer/382/base -> origin/gh/soulitzer/382/base 2025-10-10T01:18:33.9746029Z * [new branch] gh/soulitzer/382/head -> origin/gh/soulitzer/382/head 2025-10-10T01:18:33.9747557Z * [new branch] gh/soulitzer/382/orig -> origin/gh/soulitzer/382/orig 2025-10-10T01:18:33.9750203Z * [new branch] gh/soulitzer/383/base -> origin/gh/soulitzer/383/base 2025-10-10T01:18:33.9751912Z * [new branch] gh/soulitzer/383/head -> origin/gh/soulitzer/383/head 2025-10-10T01:18:33.9753527Z * [new branch] gh/soulitzer/383/orig -> origin/gh/soulitzer/383/orig 2025-10-10T01:18:33.9755953Z * [new branch] gh/soulitzer/384/base -> origin/gh/soulitzer/384/base 2025-10-10T01:18:33.9757725Z * [new branch] gh/soulitzer/384/head -> origin/gh/soulitzer/384/head 2025-10-10T01:18:33.9759505Z * [new branch] gh/soulitzer/384/orig -> origin/gh/soulitzer/384/orig 2025-10-10T01:18:33.9762380Z * [new branch] gh/swolchok/728/next -> origin/gh/swolchok/728/next 2025-10-10T01:18:33.9764687Z * [new branch] gh/swolchok/786/base -> origin/gh/swolchok/786/base 2025-10-10T01:18:33.9766423Z * [new branch] gh/swolchok/786/head -> origin/gh/swolchok/786/head 2025-10-10T01:18:33.9768213Z * [new branch] gh/swolchok/786/orig -> origin/gh/swolchok/786/orig 2025-10-10T01:18:33.9770406Z * [new branch] gh/swolchok/787/base -> origin/gh/swolchok/787/base 2025-10-10T01:18:33.9772234Z * [new branch] gh/swolchok/787/head -> origin/gh/swolchok/787/head 2025-10-10T01:18:33.9773917Z * [new branch] gh/swolchok/787/orig -> origin/gh/swolchok/787/orig 2025-10-10T01:18:33.9776557Z * [new branch] gh/swolchok/809/base -> origin/gh/swolchok/809/base 2025-10-10T01:18:33.9778533Z * [new branch] gh/swolchok/809/head -> origin/gh/swolchok/809/head 2025-10-10T01:18:33.9780109Z * [new branch] gh/swolchok/809/orig -> origin/gh/swolchok/809/orig 2025-10-10T01:18:33.9782985Z * [new branch] gh/swolchok/815/base -> origin/gh/swolchok/815/base 2025-10-10T01:18:33.9784229Z * [new branch] gh/swolchok/815/head -> origin/gh/swolchok/815/head 2025-10-10T01:18:33.9785978Z * [new branch] gh/swolchok/815/orig -> origin/gh/swolchok/815/orig 2025-10-10T01:18:33.9789015Z * [new branch] gh/swolchok/819/base -> origin/gh/swolchok/819/base 2025-10-10T01:18:33.9790673Z * [new branch] gh/swolchok/819/head -> origin/gh/swolchok/819/head 2025-10-10T01:18:33.9792404Z * [new branch] gh/swolchok/819/orig -> origin/gh/swolchok/819/orig 2025-10-10T01:18:33.9794795Z * [new branch] gh/swolchok/821/base -> origin/gh/swolchok/821/base 2025-10-10T01:18:33.9797662Z * [new branch] gh/swolchok/821/head -> origin/gh/swolchok/821/head 2025-10-10T01:18:33.9798862Z * [new branch] gh/swolchok/821/orig -> origin/gh/swolchok/821/orig 2025-10-10T01:18:33.9801990Z * [new branch] gh/swolchok/823/base -> origin/gh/swolchok/823/base 2025-10-10T01:18:33.9802984Z * [new branch] gh/swolchok/823/head -> origin/gh/swolchok/823/head 2025-10-10T01:18:33.9804690Z * [new branch] gh/swolchok/823/orig -> origin/gh/swolchok/823/orig 2025-10-10T01:18:33.9807311Z * [new branch] gh/swolchok/824/base -> origin/gh/swolchok/824/base 2025-10-10T01:18:33.9809113Z * [new branch] gh/swolchok/824/head -> origin/gh/swolchok/824/head 2025-10-10T01:18:33.9810732Z * [new branch] gh/swolchok/824/orig -> origin/gh/swolchok/824/orig 2025-10-10T01:18:33.9813060Z * [new branch] gh/swolchok/826/base -> origin/gh/swolchok/826/base 2025-10-10T01:18:33.9814842Z * [new branch] gh/swolchok/826/head -> origin/gh/swolchok/826/head 2025-10-10T01:18:33.9816444Z * [new branch] gh/swolchok/826/orig -> origin/gh/swolchok/826/orig 2025-10-10T01:18:33.9818774Z * [new branch] gh/swolchok/829/base -> origin/gh/swolchok/829/base 2025-10-10T01:18:33.9820374Z * [new branch] gh/swolchok/829/head -> origin/gh/swolchok/829/head 2025-10-10T01:18:33.9822276Z * [new branch] gh/swolchok/829/orig -> origin/gh/swolchok/829/orig 2025-10-10T01:18:33.9824990Z * [new branch] gh/swolchok/830/base -> origin/gh/swolchok/830/base 2025-10-10T01:18:33.9826809Z * [new branch] gh/swolchok/830/head -> origin/gh/swolchok/830/head 2025-10-10T01:18:33.9828563Z * [new branch] gh/swolchok/830/orig -> origin/gh/swolchok/830/orig 2025-10-10T01:18:33.9831040Z * [new branch] gh/swolchok/831/base -> origin/gh/swolchok/831/base 2025-10-10T01:18:33.9832650Z * [new branch] gh/swolchok/831/head -> origin/gh/swolchok/831/head 2025-10-10T01:18:33.9834385Z * [new branch] gh/swolchok/831/orig -> origin/gh/swolchok/831/orig 2025-10-10T01:18:33.9837584Z * [new branch] gh/swolchok/832/base -> origin/gh/swolchok/832/base 2025-10-10T01:18:33.9839643Z * [new branch] gh/swolchok/832/head -> origin/gh/swolchok/832/head 2025-10-10T01:18:33.9841390Z * [new branch] gh/swolchok/832/orig -> origin/gh/swolchok/832/orig 2025-10-10T01:18:33.9843988Z * [new branch] gh/swolchok/833/base -> origin/gh/swolchok/833/base 2025-10-10T01:18:33.9845481Z * [new branch] gh/swolchok/833/head -> origin/gh/swolchok/833/head 2025-10-10T01:18:33.9847173Z * [new branch] gh/swolchok/833/orig -> origin/gh/swolchok/833/orig 2025-10-10T01:18:33.9850270Z * [new branch] gh/swolchok/834/base -> origin/gh/swolchok/834/base 2025-10-10T01:18:33.9851935Z * [new branch] gh/swolchok/834/head -> origin/gh/swolchok/834/head 2025-10-10T01:18:33.9853608Z * [new branch] gh/swolchok/834/orig -> origin/gh/swolchok/834/orig 2025-10-10T01:18:33.9856395Z * [new branch] gh/swolchok/835/base -> origin/gh/swolchok/835/base 2025-10-10T01:18:33.9858050Z * [new branch] gh/swolchok/835/head -> origin/gh/swolchok/835/head 2025-10-10T01:18:33.9859723Z * [new branch] gh/swolchok/835/orig -> origin/gh/swolchok/835/orig 2025-10-10T01:18:33.9862262Z * [new branch] gh/swolchok/836/base -> origin/gh/swolchok/836/base 2025-10-10T01:18:33.9864008Z * [new branch] gh/swolchok/836/head -> origin/gh/swolchok/836/head 2025-10-10T01:18:33.9866030Z * [new branch] gh/swolchok/836/orig -> origin/gh/swolchok/836/orig 2025-10-10T01:18:33.9869312Z * [new branch] gh/swolchok/837/base -> origin/gh/swolchok/837/base 2025-10-10T01:18:33.9870253Z * [new branch] gh/swolchok/837/head -> origin/gh/swolchok/837/head 2025-10-10T01:18:33.9871853Z * [new branch] gh/swolchok/837/orig -> origin/gh/swolchok/837/orig 2025-10-10T01:18:33.9874907Z * [new branch] gh/swolchok/838/base -> origin/gh/swolchok/838/base 2025-10-10T01:18:33.9876269Z * [new branch] gh/swolchok/838/head -> origin/gh/swolchok/838/head 2025-10-10T01:18:33.9878645Z * [new branch] gh/swolchok/838/orig -> origin/gh/swolchok/838/orig 2025-10-10T01:18:33.9881186Z * [new branch] gh/swolchok/839/base -> origin/gh/swolchok/839/base 2025-10-10T01:18:33.9882770Z * [new branch] gh/swolchok/839/head -> origin/gh/swolchok/839/head 2025-10-10T01:18:33.9884406Z * [new branch] gh/swolchok/839/orig -> origin/gh/swolchok/839/orig 2025-10-10T01:18:33.9886882Z * [new branch] gh/swolchok/840/base -> origin/gh/swolchok/840/base 2025-10-10T01:18:33.9889055Z * [new branch] gh/swolchok/840/head -> origin/gh/swolchok/840/head 2025-10-10T01:18:33.9890793Z * [new branch] gh/swolchok/840/orig -> origin/gh/swolchok/840/orig 2025-10-10T01:18:33.9893103Z * [new branch] gh/swolchok/841/base -> origin/gh/swolchok/841/base 2025-10-10T01:18:33.9894798Z * [new branch] gh/swolchok/841/head -> origin/gh/swolchok/841/head 2025-10-10T01:18:33.9896427Z * [new branch] gh/swolchok/841/orig -> origin/gh/swolchok/841/orig 2025-10-10T01:18:33.9898998Z * [new branch] gh/swolchok/842/base -> origin/gh/swolchok/842/base 2025-10-10T01:18:33.9900697Z * [new branch] gh/swolchok/842/head -> origin/gh/swolchok/842/head 2025-10-10T01:18:33.9902862Z * [new branch] gh/swolchok/842/orig -> origin/gh/swolchok/842/orig 2025-10-10T01:18:33.9905341Z * [new branch] gh/swolchok/843/base -> origin/gh/swolchok/843/base 2025-10-10T01:18:33.9907854Z * [new branch] gh/swolchok/843/head -> origin/gh/swolchok/843/head 2025-10-10T01:18:33.9909118Z * [new branch] gh/swolchok/843/orig -> origin/gh/swolchok/843/orig 2025-10-10T01:18:33.9913057Z * [new branch] gh/swolchok/844/base -> origin/gh/swolchok/844/base 2025-10-10T01:18:33.9914187Z * [new branch] gh/swolchok/844/head -> origin/gh/swolchok/844/head 2025-10-10T01:18:33.9916172Z * [new branch] gh/swolchok/844/orig -> origin/gh/swolchok/844/orig 2025-10-10T01:18:33.9918650Z * [new branch] gh/swolchok/845/base -> origin/gh/swolchok/845/base 2025-10-10T01:18:33.9920561Z * [new branch] gh/swolchok/845/head -> origin/gh/swolchok/845/head 2025-10-10T01:18:33.9922259Z * [new branch] gh/swolchok/845/orig -> origin/gh/swolchok/845/orig 2025-10-10T01:18:33.9924992Z * [new branch] gh/swolchok/846/base -> origin/gh/swolchok/846/base 2025-10-10T01:18:33.9927544Z * [new branch] gh/swolchok/846/head -> origin/gh/swolchok/846/head 2025-10-10T01:18:33.9929480Z * [new branch] gh/swolchok/846/orig -> origin/gh/swolchok/846/orig 2025-10-10T01:18:33.9932302Z * [new branch] gh/swolchok/847/base -> origin/gh/swolchok/847/base 2025-10-10T01:18:33.9933454Z * [new branch] gh/swolchok/847/head -> origin/gh/swolchok/847/head 2025-10-10T01:18:33.9935156Z * [new branch] gh/swolchok/847/orig -> origin/gh/swolchok/847/orig 2025-10-10T01:18:33.9937917Z * [new branch] gh/swolchok/848/base -> origin/gh/swolchok/848/base 2025-10-10T01:18:33.9939633Z * [new branch] gh/swolchok/848/head -> origin/gh/swolchok/848/head 2025-10-10T01:18:33.9941458Z * [new branch] gh/swolchok/848/orig -> origin/gh/swolchok/848/orig 2025-10-10T01:18:33.9943836Z * [new branch] gh/swolchok/849/base -> origin/gh/swolchok/849/base 2025-10-10T01:18:33.9945659Z * [new branch] gh/swolchok/849/head -> origin/gh/swolchok/849/head 2025-10-10T01:18:33.9947417Z * [new branch] gh/swolchok/849/orig -> origin/gh/swolchok/849/orig 2025-10-10T01:18:33.9949693Z * [new branch] gh/swolchok/850/base -> origin/gh/swolchok/850/base 2025-10-10T01:18:33.9951420Z * [new branch] gh/swolchok/850/head -> origin/gh/swolchok/850/head 2025-10-10T01:18:33.9953211Z * [new branch] gh/swolchok/850/orig -> origin/gh/swolchok/850/orig 2025-10-10T01:18:33.9955360Z * [new branch] gh/swolchok/851/base -> origin/gh/swolchok/851/base 2025-10-10T01:18:33.9957029Z * [new branch] gh/swolchok/851/head -> origin/gh/swolchok/851/head 2025-10-10T01:18:33.9958826Z * [new branch] gh/swolchok/851/orig -> origin/gh/swolchok/851/orig 2025-10-10T01:18:33.9961372Z * [new branch] gh/swolchok/852/base -> origin/gh/swolchok/852/base 2025-10-10T01:18:33.9962998Z * [new branch] gh/swolchok/852/head -> origin/gh/swolchok/852/head 2025-10-10T01:18:33.9964751Z * [new branch] gh/swolchok/852/orig -> origin/gh/swolchok/852/orig 2025-10-10T01:18:33.9967453Z * [new branch] gh/syed-ahmed/5/base -> origin/gh/syed-ahmed/5/base 2025-10-10T01:18:33.9969216Z * [new branch] gh/syed-ahmed/5/head -> origin/gh/syed-ahmed/5/head 2025-10-10T01:18:33.9970899Z * [new branch] gh/syed-ahmed/5/orig -> origin/gh/syed-ahmed/5/orig 2025-10-10T01:18:33.9973228Z * [new branch] gh/syed-ahmed/6/base -> origin/gh/syed-ahmed/6/base 2025-10-10T01:18:33.9975000Z * [new branch] gh/syed-ahmed/6/head -> origin/gh/syed-ahmed/6/head 2025-10-10T01:18:33.9976697Z * [new branch] gh/syed-ahmed/6/orig -> origin/gh/syed-ahmed/6/orig 2025-10-10T01:18:33.9978865Z * [new branch] gh/syed-ahmed/7/base -> origin/gh/syed-ahmed/7/base 2025-10-10T01:18:33.9980760Z * [new branch] gh/syed-ahmed/7/head -> origin/gh/syed-ahmed/7/head 2025-10-10T01:18:33.9982476Z * [new branch] gh/syed-ahmed/7/orig -> origin/gh/syed-ahmed/7/orig 2025-10-10T01:18:33.9985291Z * [new branch] gh/teja-rao/4/base -> origin/gh/teja-rao/4/base 2025-10-10T01:18:33.9987032Z * [new branch] gh/teja-rao/4/head -> origin/gh/teja-rao/4/head 2025-10-10T01:18:33.9988699Z * [new branch] gh/teja-rao/4/orig -> origin/gh/teja-rao/4/orig 2025-10-10T01:18:33.9991558Z * [new branch] gh/tianyu-l/2/base -> origin/gh/tianyu-l/2/base 2025-10-10T01:18:33.9993338Z * [new branch] gh/tianyu-l/2/head -> origin/gh/tianyu-l/2/head 2025-10-10T01:18:33.9995462Z * [new branch] gh/tianyu-l/2/orig -> origin/gh/tianyu-l/2/orig 2025-10-10T01:18:33.9997737Z * [new branch] gh/tianyu-l/5/base -> origin/gh/tianyu-l/5/base 2025-10-10T01:18:33.9999581Z * [new branch] gh/tianyu-l/5/orig -> origin/gh/tianyu-l/5/orig 2025-10-10T01:18:34.0001868Z * [new branch] gh/tianyu-l/6/base -> origin/gh/tianyu-l/6/base 2025-10-10T01:18:34.0003510Z * [new branch] gh/tianyu-l/6/head -> origin/gh/tianyu-l/6/head 2025-10-10T01:18:34.0005196Z * [new branch] gh/tianyu-l/6/orig -> origin/gh/tianyu-l/6/orig 2025-10-10T01:18:34.0007433Z * [new branch] gh/tianyu-l/7/base -> origin/gh/tianyu-l/7/base 2025-10-10T01:18:34.0009155Z * [new branch] gh/tianyu-l/7/orig -> origin/gh/tianyu-l/7/orig 2025-10-10T01:18:34.0012399Z * [new branch] gh/tugsbayasgalan/10/base -> origin/gh/tugsbayasgalan/10/base 2025-10-10T01:18:34.0014032Z * [new branch] gh/tugsbayasgalan/10/head -> origin/gh/tugsbayasgalan/10/head 2025-10-10T01:18:34.0015834Z * [new branch] gh/tugsbayasgalan/10/orig -> origin/gh/tugsbayasgalan/10/orig 2025-10-10T01:18:34.0018187Z * [new branch] gh/tugsbayasgalan/11/base -> origin/gh/tugsbayasgalan/11/base 2025-10-10T01:18:34.0019815Z * [new branch] gh/tugsbayasgalan/11/head -> origin/gh/tugsbayasgalan/11/head 2025-10-10T01:18:34.0021512Z * [new branch] gh/tugsbayasgalan/11/orig -> origin/gh/tugsbayasgalan/11/orig 2025-10-10T01:18:34.0023896Z * [new branch] gh/tugsbayasgalan/12/base -> origin/gh/tugsbayasgalan/12/base 2025-10-10T01:18:34.0026062Z * [new branch] gh/tugsbayasgalan/12/head -> origin/gh/tugsbayasgalan/12/head 2025-10-10T01:18:34.0028365Z * [new branch] gh/tugsbayasgalan/12/orig -> origin/gh/tugsbayasgalan/12/orig 2025-10-10T01:18:34.0030482Z * [new branch] gh/tugsbayasgalan/13/base -> origin/gh/tugsbayasgalan/13/base 2025-10-10T01:18:34.0032185Z * [new branch] gh/tugsbayasgalan/13/head -> origin/gh/tugsbayasgalan/13/head 2025-10-10T01:18:34.0033892Z * [new branch] gh/tugsbayasgalan/13/orig -> origin/gh/tugsbayasgalan/13/orig 2025-10-10T01:18:34.0037130Z * [new branch] gh/tugsbayasgalan/14/base -> origin/gh/tugsbayasgalan/14/base 2025-10-10T01:18:34.0038791Z * [new branch] gh/tugsbayasgalan/14/head -> origin/gh/tugsbayasgalan/14/head 2025-10-10T01:18:34.0040730Z * [new branch] gh/tugsbayasgalan/14/orig -> origin/gh/tugsbayasgalan/14/orig 2025-10-10T01:18:34.0043236Z * [new branch] gh/tugsbayasgalan/15/base -> origin/gh/tugsbayasgalan/15/base 2025-10-10T01:18:34.0044911Z * [new branch] gh/tugsbayasgalan/15/head -> origin/gh/tugsbayasgalan/15/head 2025-10-10T01:18:34.0046692Z * [new branch] gh/tugsbayasgalan/15/orig -> origin/gh/tugsbayasgalan/15/orig 2025-10-10T01:18:34.0049044Z * [new branch] gh/tugsbayasgalan/16/base -> origin/gh/tugsbayasgalan/16/base 2025-10-10T01:18:34.0050949Z * [new branch] gh/tugsbayasgalan/16/head -> origin/gh/tugsbayasgalan/16/head 2025-10-10T01:18:34.0052603Z * [new branch] gh/tugsbayasgalan/16/orig -> origin/gh/tugsbayasgalan/16/orig 2025-10-10T01:18:34.0054987Z * [new branch] gh/tugsbayasgalan/17/base -> origin/gh/tugsbayasgalan/17/base 2025-10-10T01:18:34.0056692Z * [new branch] gh/tugsbayasgalan/17/head -> origin/gh/tugsbayasgalan/17/head 2025-10-10T01:18:34.0058438Z * [new branch] gh/tugsbayasgalan/17/orig -> origin/gh/tugsbayasgalan/17/orig 2025-10-10T01:18:34.0060713Z * [new branch] gh/tugsbayasgalan/18/base -> origin/gh/tugsbayasgalan/18/base 2025-10-10T01:18:34.0062449Z * [new branch] gh/tugsbayasgalan/18/head -> origin/gh/tugsbayasgalan/18/head 2025-10-10T01:18:34.0064704Z * [new branch] gh/tugsbayasgalan/18/orig -> origin/gh/tugsbayasgalan/18/orig 2025-10-10T01:18:34.0067316Z * [new branch] gh/tugsbayasgalan/19/base -> origin/gh/tugsbayasgalan/19/base 2025-10-10T01:18:34.0069204Z * [new branch] gh/tugsbayasgalan/19/head -> origin/gh/tugsbayasgalan/19/head 2025-10-10T01:18:34.0070818Z * [new branch] gh/tugsbayasgalan/19/orig -> origin/gh/tugsbayasgalan/19/orig 2025-10-10T01:18:34.0072996Z * [new branch] gh/tugsbayasgalan/2/base -> origin/gh/tugsbayasgalan/2/base 2025-10-10T01:18:34.0074617Z * [new branch] gh/tugsbayasgalan/2/head -> origin/gh/tugsbayasgalan/2/head 2025-10-10T01:18:34.0076572Z * [new branch] gh/tugsbayasgalan/2/orig -> origin/gh/tugsbayasgalan/2/orig 2025-10-10T01:18:34.0079067Z * [new branch] gh/tugsbayasgalan/20/base -> origin/gh/tugsbayasgalan/20/base 2025-10-10T01:18:34.0081045Z * [new branch] gh/tugsbayasgalan/20/head -> origin/gh/tugsbayasgalan/20/head 2025-10-10T01:18:34.0082592Z * [new branch] gh/tugsbayasgalan/20/orig -> origin/gh/tugsbayasgalan/20/orig 2025-10-10T01:18:34.0085538Z * [new branch] gh/tugsbayasgalan/21/base -> origin/gh/tugsbayasgalan/21/base 2025-10-10T01:18:34.0087668Z * [new branch] gh/tugsbayasgalan/21/head -> origin/gh/tugsbayasgalan/21/head 2025-10-10T01:18:34.0089381Z * [new branch] gh/tugsbayasgalan/21/orig -> origin/gh/tugsbayasgalan/21/orig 2025-10-10T01:18:34.0091934Z * [new branch] gh/tugsbayasgalan/22/base -> origin/gh/tugsbayasgalan/22/base 2025-10-10T01:18:34.0093569Z * [new branch] gh/tugsbayasgalan/22/head -> origin/gh/tugsbayasgalan/22/head 2025-10-10T01:18:34.0095214Z * [new branch] gh/tugsbayasgalan/22/orig -> origin/gh/tugsbayasgalan/22/orig 2025-10-10T01:18:34.0098225Z * [new branch] gh/tugsbayasgalan/23/base -> origin/gh/tugsbayasgalan/23/base 2025-10-10T01:18:34.0099906Z * [new branch] gh/tugsbayasgalan/23/head -> origin/gh/tugsbayasgalan/23/head 2025-10-10T01:18:34.0101602Z * [new branch] gh/tugsbayasgalan/23/orig -> origin/gh/tugsbayasgalan/23/orig 2025-10-10T01:18:34.0104010Z * [new branch] gh/tugsbayasgalan/24/base -> origin/gh/tugsbayasgalan/24/base 2025-10-10T01:18:34.0105731Z * [new branch] gh/tugsbayasgalan/24/head -> origin/gh/tugsbayasgalan/24/head 2025-10-10T01:18:34.0107337Z * [new branch] gh/tugsbayasgalan/24/orig -> origin/gh/tugsbayasgalan/24/orig 2025-10-10T01:18:34.0109912Z * [new branch] gh/tugsbayasgalan/25/base -> origin/gh/tugsbayasgalan/25/base 2025-10-10T01:18:34.0111595Z * [new branch] gh/tugsbayasgalan/25/head -> origin/gh/tugsbayasgalan/25/head 2025-10-10T01:18:34.0113388Z * [new branch] gh/tugsbayasgalan/25/orig -> origin/gh/tugsbayasgalan/25/orig 2025-10-10T01:18:34.0115736Z * [new branch] gh/tugsbayasgalan/26/base -> origin/gh/tugsbayasgalan/26/base 2025-10-10T01:18:34.0117654Z * [new branch] gh/tugsbayasgalan/26/head -> origin/gh/tugsbayasgalan/26/head 2025-10-10T01:18:34.0119601Z * [new branch] gh/tugsbayasgalan/26/orig -> origin/gh/tugsbayasgalan/26/orig 2025-10-10T01:18:34.0122714Z * [new branch] gh/tugsbayasgalan/27/base -> origin/gh/tugsbayasgalan/27/base 2025-10-10T01:18:34.0124453Z * [new branch] gh/tugsbayasgalan/27/head -> origin/gh/tugsbayasgalan/27/head 2025-10-10T01:18:34.0127370Z * [new branch] gh/tugsbayasgalan/27/orig -> origin/gh/tugsbayasgalan/27/orig 2025-10-10T01:18:34.0129594Z * [new branch] gh/tugsbayasgalan/28/base -> origin/gh/tugsbayasgalan/28/base 2025-10-10T01:18:34.0131234Z * [new branch] gh/tugsbayasgalan/28/head -> origin/gh/tugsbayasgalan/28/head 2025-10-10T01:18:34.0132967Z * [new branch] gh/tugsbayasgalan/28/orig -> origin/gh/tugsbayasgalan/28/orig 2025-10-10T01:18:34.0136050Z * [new branch] gh/tugsbayasgalan/29/base -> origin/gh/tugsbayasgalan/29/base 2025-10-10T01:18:34.0137983Z * [new branch] gh/tugsbayasgalan/29/head -> origin/gh/tugsbayasgalan/29/head 2025-10-10T01:18:34.0139714Z * [new branch] gh/tugsbayasgalan/29/orig -> origin/gh/tugsbayasgalan/29/orig 2025-10-10T01:18:34.0141987Z * [new branch] gh/tugsbayasgalan/3/base -> origin/gh/tugsbayasgalan/3/base 2025-10-10T01:18:34.0143756Z * [new branch] gh/tugsbayasgalan/3/head -> origin/gh/tugsbayasgalan/3/head 2025-10-10T01:18:34.0145467Z * [new branch] gh/tugsbayasgalan/3/orig -> origin/gh/tugsbayasgalan/3/orig 2025-10-10T01:18:34.0147814Z * [new branch] gh/tugsbayasgalan/30/base -> origin/gh/tugsbayasgalan/30/base 2025-10-10T01:18:34.0149626Z * [new branch] gh/tugsbayasgalan/30/head -> origin/gh/tugsbayasgalan/30/head 2025-10-10T01:18:34.0151479Z * [new branch] gh/tugsbayasgalan/30/orig -> origin/gh/tugsbayasgalan/30/orig 2025-10-10T01:18:34.0153878Z * [new branch] gh/tugsbayasgalan/31/base -> origin/gh/tugsbayasgalan/31/base 2025-10-10T01:18:34.0155725Z * [new branch] gh/tugsbayasgalan/31/head -> origin/gh/tugsbayasgalan/31/head 2025-10-10T01:18:34.0157364Z * [new branch] gh/tugsbayasgalan/31/orig -> origin/gh/tugsbayasgalan/31/orig 2025-10-10T01:18:34.0159803Z * [new branch] gh/tugsbayasgalan/32/base -> origin/gh/tugsbayasgalan/32/base 2025-10-10T01:18:34.0161539Z * [new branch] gh/tugsbayasgalan/32/head -> origin/gh/tugsbayasgalan/32/head 2025-10-10T01:18:34.0163297Z * [new branch] gh/tugsbayasgalan/32/orig -> origin/gh/tugsbayasgalan/32/orig 2025-10-10T01:18:34.0165704Z * [new branch] gh/tugsbayasgalan/33/base -> origin/gh/tugsbayasgalan/33/base 2025-10-10T01:18:34.0167527Z * [new branch] gh/tugsbayasgalan/33/head -> origin/gh/tugsbayasgalan/33/head 2025-10-10T01:18:34.0169339Z * [new branch] gh/tugsbayasgalan/33/orig -> origin/gh/tugsbayasgalan/33/orig 2025-10-10T01:18:34.0171810Z * [new branch] gh/tugsbayasgalan/34/base -> origin/gh/tugsbayasgalan/34/base 2025-10-10T01:18:34.0173888Z * [new branch] gh/tugsbayasgalan/34/head -> origin/gh/tugsbayasgalan/34/head 2025-10-10T01:18:34.0175573Z * [new branch] gh/tugsbayasgalan/34/orig -> origin/gh/tugsbayasgalan/34/orig 2025-10-10T01:18:34.0177900Z * [new branch] gh/tugsbayasgalan/35/base -> origin/gh/tugsbayasgalan/35/base 2025-10-10T01:18:34.0179544Z * [new branch] gh/tugsbayasgalan/35/head -> origin/gh/tugsbayasgalan/35/head 2025-10-10T01:18:34.0181289Z * [new branch] gh/tugsbayasgalan/35/orig -> origin/gh/tugsbayasgalan/35/orig 2025-10-10T01:18:34.0183671Z * [new branch] gh/tugsbayasgalan/36/base -> origin/gh/tugsbayasgalan/36/base 2025-10-10T01:18:34.0185348Z * [new branch] gh/tugsbayasgalan/36/head -> origin/gh/tugsbayasgalan/36/head 2025-10-10T01:18:34.0187110Z * [new branch] gh/tugsbayasgalan/36/orig -> origin/gh/tugsbayasgalan/36/orig 2025-10-10T01:18:34.0190111Z * [new branch] gh/tugsbayasgalan/37/base -> origin/gh/tugsbayasgalan/37/base 2025-10-10T01:18:34.0191797Z * [new branch] gh/tugsbayasgalan/37/head -> origin/gh/tugsbayasgalan/37/head 2025-10-10T01:18:34.0193556Z * [new branch] gh/tugsbayasgalan/37/orig -> origin/gh/tugsbayasgalan/37/orig 2025-10-10T01:18:34.0196075Z * [new branch] gh/tugsbayasgalan/38/base -> origin/gh/tugsbayasgalan/38/base 2025-10-10T01:18:34.0197878Z * [new branch] gh/tugsbayasgalan/38/head -> origin/gh/tugsbayasgalan/38/head 2025-10-10T01:18:34.0199705Z * [new branch] gh/tugsbayasgalan/38/orig -> origin/gh/tugsbayasgalan/38/orig 2025-10-10T01:18:34.0202048Z * [new branch] gh/tugsbayasgalan/39/base -> origin/gh/tugsbayasgalan/39/base 2025-10-10T01:18:34.0203814Z * [new branch] gh/tugsbayasgalan/39/head -> origin/gh/tugsbayasgalan/39/head 2025-10-10T01:18:34.0205510Z * [new branch] gh/tugsbayasgalan/39/orig -> origin/gh/tugsbayasgalan/39/orig 2025-10-10T01:18:34.0207916Z * [new branch] gh/tugsbayasgalan/40/base -> origin/gh/tugsbayasgalan/40/base 2025-10-10T01:18:34.0209591Z * [new branch] gh/tugsbayasgalan/40/head -> origin/gh/tugsbayasgalan/40/head 2025-10-10T01:18:34.0211403Z * [new branch] gh/tugsbayasgalan/40/orig -> origin/gh/tugsbayasgalan/40/orig 2025-10-10T01:18:34.0213811Z * [new branch] gh/tugsbayasgalan/41/base -> origin/gh/tugsbayasgalan/41/base 2025-10-10T01:18:34.0215586Z * [new branch] gh/tugsbayasgalan/41/head -> origin/gh/tugsbayasgalan/41/head 2025-10-10T01:18:34.0217496Z * [new branch] gh/tugsbayasgalan/41/orig -> origin/gh/tugsbayasgalan/41/orig 2025-10-10T01:18:34.0219871Z * [new branch] gh/tugsbayasgalan/42/base -> origin/gh/tugsbayasgalan/42/base 2025-10-10T01:18:34.0221794Z * [new branch] gh/tugsbayasgalan/42/head -> origin/gh/tugsbayasgalan/42/head 2025-10-10T01:18:34.0223513Z * [new branch] gh/tugsbayasgalan/42/orig -> origin/gh/tugsbayasgalan/42/orig 2025-10-10T01:18:34.0226271Z * [new branch] gh/tugsbayasgalan/43/base -> origin/gh/tugsbayasgalan/43/base 2025-10-10T01:18:34.0227939Z * [new branch] gh/tugsbayasgalan/43/head -> origin/gh/tugsbayasgalan/43/head 2025-10-10T01:18:34.0229659Z * [new branch] gh/tugsbayasgalan/43/orig -> origin/gh/tugsbayasgalan/43/orig 2025-10-10T01:18:34.0232094Z * [new branch] gh/tugsbayasgalan/44/base -> origin/gh/tugsbayasgalan/44/base 2025-10-10T01:18:34.0233752Z * [new branch] gh/tugsbayasgalan/44/head -> origin/gh/tugsbayasgalan/44/head 2025-10-10T01:18:34.0235527Z * [new branch] gh/tugsbayasgalan/44/orig -> origin/gh/tugsbayasgalan/44/orig 2025-10-10T01:18:34.0237803Z * [new branch] gh/tugsbayasgalan/45/base -> origin/gh/tugsbayasgalan/45/base 2025-10-10T01:18:34.0239788Z * [new branch] gh/tugsbayasgalan/45/head -> origin/gh/tugsbayasgalan/45/head 2025-10-10T01:18:34.0241498Z * [new branch] gh/tugsbayasgalan/45/orig -> origin/gh/tugsbayasgalan/45/orig 2025-10-10T01:18:34.0243975Z * [new branch] gh/tugsbayasgalan/46/base -> origin/gh/tugsbayasgalan/46/base 2025-10-10T01:18:34.0245664Z * [new branch] gh/tugsbayasgalan/46/head -> origin/gh/tugsbayasgalan/46/head 2025-10-10T01:18:34.0247397Z * [new branch] gh/tugsbayasgalan/46/orig -> origin/gh/tugsbayasgalan/46/orig 2025-10-10T01:18:34.0249814Z * [new branch] gh/tugsbayasgalan/47/base -> origin/gh/tugsbayasgalan/47/base 2025-10-10T01:18:34.0251553Z * [new branch] gh/tugsbayasgalan/47/head -> origin/gh/tugsbayasgalan/47/head 2025-10-10T01:18:34.0253244Z * [new branch] gh/tugsbayasgalan/47/orig -> origin/gh/tugsbayasgalan/47/orig 2025-10-10T01:18:34.0255514Z * [new branch] gh/tugsbayasgalan/48/base -> origin/gh/tugsbayasgalan/48/base 2025-10-10T01:18:34.0257430Z * [new branch] gh/tugsbayasgalan/48/head -> origin/gh/tugsbayasgalan/48/head 2025-10-10T01:18:34.0259229Z * [new branch] gh/tugsbayasgalan/48/orig -> origin/gh/tugsbayasgalan/48/orig 2025-10-10T01:18:34.0261488Z * [new branch] gh/tugsbayasgalan/49/base -> origin/gh/tugsbayasgalan/49/base 2025-10-10T01:18:34.0263406Z * [new branch] gh/tugsbayasgalan/49/head -> origin/gh/tugsbayasgalan/49/head 2025-10-10T01:18:34.0265154Z * [new branch] gh/tugsbayasgalan/49/orig -> origin/gh/tugsbayasgalan/49/orig 2025-10-10T01:18:34.0267579Z * [new branch] gh/tugsbayasgalan/50/base -> origin/gh/tugsbayasgalan/50/base 2025-10-10T01:18:34.0269425Z * [new branch] gh/tugsbayasgalan/50/head -> origin/gh/tugsbayasgalan/50/head 2025-10-10T01:18:34.0271164Z * [new branch] gh/tugsbayasgalan/50/orig -> origin/gh/tugsbayasgalan/50/orig 2025-10-10T01:18:34.0273440Z * [new branch] gh/tugsbayasgalan/51/base -> origin/gh/tugsbayasgalan/51/base 2025-10-10T01:18:34.0275170Z * [new branch] gh/tugsbayasgalan/51/head -> origin/gh/tugsbayasgalan/51/head 2025-10-10T01:18:34.0276897Z * [new branch] gh/tugsbayasgalan/51/orig -> origin/gh/tugsbayasgalan/51/orig 2025-10-10T01:18:34.0279200Z * [new branch] gh/tugsbayasgalan/52/base -> origin/gh/tugsbayasgalan/52/base 2025-10-10T01:18:34.0280869Z * [new branch] gh/tugsbayasgalan/52/head -> origin/gh/tugsbayasgalan/52/head 2025-10-10T01:18:34.0282637Z * [new branch] gh/tugsbayasgalan/52/orig -> origin/gh/tugsbayasgalan/52/orig 2025-10-10T01:18:34.0285217Z * [new branch] gh/tugsbayasgalan/53/base -> origin/gh/tugsbayasgalan/53/base 2025-10-10T01:18:34.0286777Z * [new branch] gh/tugsbayasgalan/53/head -> origin/gh/tugsbayasgalan/53/head 2025-10-10T01:18:34.0288433Z * [new branch] gh/tugsbayasgalan/53/orig -> origin/gh/tugsbayasgalan/53/orig 2025-10-10T01:18:34.0290674Z * [new branch] gh/tugsbayasgalan/54/base -> origin/gh/tugsbayasgalan/54/base 2025-10-10T01:18:34.0292498Z * [new branch] gh/tugsbayasgalan/54/head -> origin/gh/tugsbayasgalan/54/head 2025-10-10T01:18:34.0294138Z * [new branch] gh/tugsbayasgalan/54/orig -> origin/gh/tugsbayasgalan/54/orig 2025-10-10T01:18:34.0296465Z * [new branch] gh/tugsbayasgalan/6/base -> origin/gh/tugsbayasgalan/6/base 2025-10-10T01:18:34.0298093Z * [new branch] gh/tugsbayasgalan/6/head -> origin/gh/tugsbayasgalan/6/head 2025-10-10T01:18:34.0299855Z * [new branch] gh/tugsbayasgalan/6/orig -> origin/gh/tugsbayasgalan/6/orig 2025-10-10T01:18:34.0302175Z * [new branch] gh/tugsbayasgalan/7/base -> origin/gh/tugsbayasgalan/7/base 2025-10-10T01:18:34.0303863Z * [new branch] gh/tugsbayasgalan/7/head -> origin/gh/tugsbayasgalan/7/head 2025-10-10T01:18:34.0305662Z * [new branch] gh/tugsbayasgalan/7/orig -> origin/gh/tugsbayasgalan/7/orig 2025-10-10T01:18:34.0308033Z * [new branch] gh/tugsbayasgalan/8/base -> origin/gh/tugsbayasgalan/8/base 2025-10-10T01:18:34.0309652Z * [new branch] gh/tugsbayasgalan/8/head -> origin/gh/tugsbayasgalan/8/head 2025-10-10T01:18:34.0311390Z * [new branch] gh/tugsbayasgalan/8/orig -> origin/gh/tugsbayasgalan/8/orig 2025-10-10T01:18:34.0313710Z * [new branch] gh/tugsbayasgalan/9/base -> origin/gh/tugsbayasgalan/9/base 2025-10-10T01:18:34.0315320Z * [new branch] gh/tugsbayasgalan/9/head -> origin/gh/tugsbayasgalan/9/head 2025-10-10T01:18:34.0317095Z * [new branch] gh/tugsbayasgalan/9/orig -> origin/gh/tugsbayasgalan/9/orig 2025-10-10T01:18:34.0320067Z * [new branch] gh/v0i0/10/base -> origin/gh/v0i0/10/base 2025-10-10T01:18:34.0321712Z * [new branch] gh/v0i0/10/head -> origin/gh/v0i0/10/head 2025-10-10T01:18:34.0323563Z * [new branch] gh/v0i0/10/orig -> origin/gh/v0i0/10/orig 2025-10-10T01:18:34.0327318Z * [new branch] gh/v0i0/11/base -> origin/gh/v0i0/11/base 2025-10-10T01:18:34.0329009Z * [new branch] gh/v0i0/11/head -> origin/gh/v0i0/11/head 2025-10-10T01:18:34.0330705Z * [new branch] gh/v0i0/11/orig -> origin/gh/v0i0/11/orig 2025-10-10T01:18:34.0333134Z * [new branch] gh/v0i0/12/base -> origin/gh/v0i0/12/base 2025-10-10T01:18:34.0334840Z * [new branch] gh/v0i0/12/head -> origin/gh/v0i0/12/head 2025-10-10T01:18:34.0336530Z * [new branch] gh/v0i0/12/orig -> origin/gh/v0i0/12/orig 2025-10-10T01:18:34.0339067Z * [new branch] gh/v0i0/13/base -> origin/gh/v0i0/13/base 2025-10-10T01:18:34.0340713Z * [new branch] gh/v0i0/13/head -> origin/gh/v0i0/13/head 2025-10-10T01:18:34.0342368Z * [new branch] gh/v0i0/13/orig -> origin/gh/v0i0/13/orig 2025-10-10T01:18:34.0344827Z * [new branch] gh/v0i0/7/base -> origin/gh/v0i0/7/base 2025-10-10T01:18:34.0346591Z * [new branch] gh/v0i0/7/head -> origin/gh/v0i0/7/head 2025-10-10T01:18:34.0348311Z * [new branch] gh/v0i0/7/orig -> origin/gh/v0i0/7/orig 2025-10-10T01:18:34.0351022Z * [new branch] gh/v0i0/8/base -> origin/gh/v0i0/8/base 2025-10-10T01:18:34.0352812Z * [new branch] gh/v0i0/8/head -> origin/gh/v0i0/8/head 2025-10-10T01:18:34.0354715Z * [new branch] gh/v0i0/8/orig -> origin/gh/v0i0/8/orig 2025-10-10T01:18:34.0357348Z * [new branch] gh/v0i0/9/base -> origin/gh/v0i0/9/base 2025-10-10T01:18:34.0359495Z * [new branch] gh/v0i0/9/head -> origin/gh/v0i0/9/head 2025-10-10T01:18:34.0361069Z * [new branch] gh/v0i0/9/orig -> origin/gh/v0i0/9/orig 2025-10-10T01:18:34.0364184Z * [new branch] gh/vishal9-team/1/base -> origin/gh/vishal9-team/1/base 2025-10-10T01:18:34.0365791Z * [new branch] gh/vishal9-team/1/head -> origin/gh/vishal9-team/1/head 2025-10-10T01:18:34.0368121Z * [new branch] gh/vishal9-team/2/base -> origin/gh/vishal9-team/2/base 2025-10-10T01:18:34.0369847Z * [new branch] gh/vishal9-team/2/head -> origin/gh/vishal9-team/2/head 2025-10-10T01:18:34.0371562Z * [new branch] gh/vishal9-team/2/orig -> origin/gh/vishal9-team/2/orig 2025-10-10T01:18:34.0374414Z * [new branch] gh/vkuzo/1/next -> origin/gh/vkuzo/1/next 2025-10-10T01:18:34.0377611Z * [new branch] gh/vkuzo/2/next -> origin/gh/vkuzo/2/next 2025-10-10T01:18:34.0379499Z * [new branch] gh/vkuzo/3/next -> origin/gh/vkuzo/3/next 2025-10-10T01:18:34.0381872Z * [new branch] gh/vkuzo/7/base -> origin/gh/vkuzo/7/base 2025-10-10T01:18:34.0383951Z * [new branch] gh/vkuzo/7/head -> origin/gh/vkuzo/7/head 2025-10-10T01:18:34.0385820Z * [new branch] gh/vkuzo/7/orig -> origin/gh/vkuzo/7/orig 2025-10-10T01:18:34.0388895Z * [new branch] gh/wconstab/419/base -> origin/gh/wconstab/419/base 2025-10-10T01:18:34.0390573Z * [new branch] gh/wconstab/419/head -> origin/gh/wconstab/419/head 2025-10-10T01:18:34.0392217Z * [new branch] gh/wconstab/419/orig -> origin/gh/wconstab/419/orig 2025-10-10T01:18:34.0394963Z * [new branch] gh/wconstab/424/base -> origin/gh/wconstab/424/base 2025-10-10T01:18:34.0396951Z * [new branch] gh/wconstab/424/head -> origin/gh/wconstab/424/head 2025-10-10T01:18:34.0398743Z * [new branch] gh/wconstab/424/orig -> origin/gh/wconstab/424/orig 2025-10-10T01:18:34.0401089Z * [new branch] gh/wconstab/435/base -> origin/gh/wconstab/435/base 2025-10-10T01:18:34.0402822Z * [new branch] gh/wconstab/435/head -> origin/gh/wconstab/435/head 2025-10-10T01:18:34.0404554Z * [new branch] gh/wconstab/435/orig -> origin/gh/wconstab/435/orig 2025-10-10T01:18:34.0406854Z * [new branch] gh/wconstab/438/base -> origin/gh/wconstab/438/base 2025-10-10T01:18:34.0408497Z * [new branch] gh/wconstab/438/head -> origin/gh/wconstab/438/head 2025-10-10T01:18:34.0410195Z * [new branch] gh/wconstab/438/orig -> origin/gh/wconstab/438/orig 2025-10-10T01:18:34.0412446Z * [new branch] gh/wconstab/444/base -> origin/gh/wconstab/444/base 2025-10-10T01:18:34.0414177Z * [new branch] gh/wconstab/444/head -> origin/gh/wconstab/444/head 2025-10-10T01:18:34.0415953Z * [new branch] gh/wconstab/444/orig -> origin/gh/wconstab/444/orig 2025-10-10T01:18:34.0418296Z * [new branch] gh/wconstab/447/base -> origin/gh/wconstab/447/base 2025-10-10T01:18:34.0420056Z * [new branch] gh/wconstab/447/head -> origin/gh/wconstab/447/head 2025-10-10T01:18:34.0421751Z * [new branch] gh/wconstab/447/orig -> origin/gh/wconstab/447/orig 2025-10-10T01:18:34.0424694Z * [new branch] gh/weifengpy/30/base -> origin/gh/weifengpy/30/base 2025-10-10T01:18:34.0426464Z * [new branch] gh/weifengpy/30/head -> origin/gh/weifengpy/30/head 2025-10-10T01:18:34.0428315Z * [new branch] gh/weifengpy/30/orig -> origin/gh/weifengpy/30/orig 2025-10-10T01:18:34.0430518Z * [new branch] gh/weifengpy/31/base -> origin/gh/weifengpy/31/base 2025-10-10T01:18:34.0432387Z * [new branch] gh/weifengpy/31/head -> origin/gh/weifengpy/31/head 2025-10-10T01:18:34.0434212Z * [new branch] gh/weifengpy/31/orig -> origin/gh/weifengpy/31/orig 2025-10-10T01:18:34.0436880Z * [new branch] gh/weifengpy/32/base -> origin/gh/weifengpy/32/base 2025-10-10T01:18:34.0438365Z * [new branch] gh/weifengpy/32/head -> origin/gh/weifengpy/32/head 2025-10-10T01:18:34.0440147Z * [new branch] gh/weifengpy/32/orig -> origin/gh/weifengpy/32/orig 2025-10-10T01:18:34.0443757Z * [new branch] gh/weifengpy/33/base -> origin/gh/weifengpy/33/base 2025-10-10T01:18:34.0445115Z * [new branch] gh/weifengpy/33/head -> origin/gh/weifengpy/33/head 2025-10-10T01:18:34.0446962Z * [new branch] gh/weifengpy/33/orig -> origin/gh/weifengpy/33/orig 2025-10-10T01:18:34.0449541Z * [new branch] gh/weifengpy/34/base -> origin/gh/weifengpy/34/base 2025-10-10T01:18:34.0451206Z * [new branch] gh/weifengpy/34/head -> origin/gh/weifengpy/34/head 2025-10-10T01:18:34.0453013Z * [new branch] gh/weifengpy/34/orig -> origin/gh/weifengpy/34/orig 2025-10-10T01:18:34.0455419Z * [new branch] gh/weifengpy/35/base -> origin/gh/weifengpy/35/base 2025-10-10T01:18:34.0457031Z * [new branch] gh/weifengpy/35/head -> origin/gh/weifengpy/35/head 2025-10-10T01:18:34.0458743Z * [new branch] gh/weifengpy/35/orig -> origin/gh/weifengpy/35/orig 2025-10-10T01:18:34.0461098Z * [new branch] gh/weifengpy/36/base -> origin/gh/weifengpy/36/base 2025-10-10T01:18:34.0462786Z * [new branch] gh/weifengpy/36/head -> origin/gh/weifengpy/36/head 2025-10-10T01:18:34.0464631Z * [new branch] gh/weifengpy/36/orig -> origin/gh/weifengpy/36/orig 2025-10-10T01:18:34.0466824Z * [new branch] gh/weifengpy/37/base -> origin/gh/weifengpy/37/base 2025-10-10T01:18:34.0468557Z * [new branch] gh/weifengpy/37/head -> origin/gh/weifengpy/37/head 2025-10-10T01:18:34.0470249Z * [new branch] gh/weifengpy/37/orig -> origin/gh/weifengpy/37/orig 2025-10-10T01:18:34.0473052Z * [new branch] gh/williamwen42/250/base -> origin/gh/williamwen42/250/base 2025-10-10T01:18:34.0474797Z * [new branch] gh/williamwen42/250/head -> origin/gh/williamwen42/250/head 2025-10-10T01:18:34.0476515Z * [new branch] gh/williamwen42/250/orig -> origin/gh/williamwen42/250/orig 2025-10-10T01:18:34.0479024Z * [new branch] gh/williamwen42/278/base -> origin/gh/williamwen42/278/base 2025-10-10T01:18:34.0480720Z * [new branch] gh/williamwen42/278/head -> origin/gh/williamwen42/278/head 2025-10-10T01:18:34.0482816Z * [new branch] gh/williamwen42/278/orig -> origin/gh/williamwen42/278/orig 2025-10-10T01:18:34.0485308Z * [new branch] gh/williamwen42/279/base -> origin/gh/williamwen42/279/base 2025-10-10T01:18:34.0487033Z * [new branch] gh/williamwen42/279/head -> origin/gh/williamwen42/279/head 2025-10-10T01:18:34.0488721Z * [new branch] gh/williamwen42/279/orig -> origin/gh/williamwen42/279/orig 2025-10-10T01:18:34.0491148Z * [new branch] gh/williamwen42/281/base -> origin/gh/williamwen42/281/base 2025-10-10T01:18:34.0492986Z * [new branch] gh/williamwen42/281/head -> origin/gh/williamwen42/281/head 2025-10-10T01:18:34.0494777Z * [new branch] gh/williamwen42/281/orig -> origin/gh/williamwen42/281/orig 2025-10-10T01:18:34.0497079Z * [new branch] gh/williamwen42/282/base -> origin/gh/williamwen42/282/base 2025-10-10T01:18:34.0498911Z * [new branch] gh/williamwen42/282/head -> origin/gh/williamwen42/282/head 2025-10-10T01:18:34.0500557Z * [new branch] gh/williamwen42/282/orig -> origin/gh/williamwen42/282/orig 2025-10-10T01:18:34.0503105Z * [new branch] gh/williamwen42/285/base -> origin/gh/williamwen42/285/base 2025-10-10T01:18:34.0504977Z * [new branch] gh/williamwen42/285/head -> origin/gh/williamwen42/285/head 2025-10-10T01:18:34.0505969Z * [new branch] gh/williamwen42/285/orig -> origin/gh/williamwen42/285/orig 2025-10-10T01:18:34.0508520Z * [new branch] gh/williamwen42/286/base -> origin/gh/williamwen42/286/base 2025-10-10T01:18:34.0510117Z * [new branch] gh/williamwen42/286/head -> origin/gh/williamwen42/286/head 2025-10-10T01:18:34.0511778Z * [new branch] gh/williamwen42/286/orig -> origin/gh/williamwen42/286/orig 2025-10-10T01:18:34.0514203Z * [new branch] gh/williamwen42/287/base -> origin/gh/williamwen42/287/base 2025-10-10T01:18:34.0516510Z * [new branch] gh/williamwen42/287/head -> origin/gh/williamwen42/287/head 2025-10-10T01:18:34.0517746Z * [new branch] gh/williamwen42/287/orig -> origin/gh/williamwen42/287/orig 2025-10-10T01:18:34.0520252Z * [new branch] gh/williamwen42/288/base -> origin/gh/williamwen42/288/base 2025-10-10T01:18:34.0521889Z * [new branch] gh/williamwen42/288/head -> origin/gh/williamwen42/288/head 2025-10-10T01:18:34.0523603Z * [new branch] gh/williamwen42/288/orig -> origin/gh/williamwen42/288/orig 2025-10-10T01:18:34.0526178Z * [new branch] gh/williamwen42/289/base -> origin/gh/williamwen42/289/base 2025-10-10T01:18:34.0527836Z * [new branch] gh/williamwen42/289/head -> origin/gh/williamwen42/289/head 2025-10-10T01:18:34.0529459Z * [new branch] gh/williamwen42/289/orig -> origin/gh/williamwen42/289/orig 2025-10-10T01:18:34.0531995Z * [new branch] gh/williamwen42/290/base -> origin/gh/williamwen42/290/base 2025-10-10T01:18:34.0533600Z * [new branch] gh/williamwen42/290/head -> origin/gh/williamwen42/290/head 2025-10-10T01:18:34.0535283Z * [new branch] gh/williamwen42/290/orig -> origin/gh/williamwen42/290/orig 2025-10-10T01:18:34.0537732Z * [new branch] gh/williamwen42/291/base -> origin/gh/williamwen42/291/base 2025-10-10T01:18:34.0539575Z * [new branch] gh/williamwen42/291/head -> origin/gh/williamwen42/291/head 2025-10-10T01:18:34.0541240Z * [new branch] gh/williamwen42/291/orig -> origin/gh/williamwen42/291/orig 2025-10-10T01:18:34.0543512Z * [new branch] gh/williamwen42/292/base -> origin/gh/williamwen42/292/base 2025-10-10T01:18:34.0545186Z * [new branch] gh/williamwen42/292/head -> origin/gh/williamwen42/292/head 2025-10-10T01:18:34.0546951Z * [new branch] gh/williamwen42/292/orig -> origin/gh/williamwen42/292/orig 2025-10-10T01:18:34.0549183Z * [new branch] gh/williamwen42/293/base -> origin/gh/williamwen42/293/base 2025-10-10T01:18:34.0550804Z * [new branch] gh/williamwen42/293/head -> origin/gh/williamwen42/293/head 2025-10-10T01:18:34.0552453Z * [new branch] gh/williamwen42/293/orig -> origin/gh/williamwen42/293/orig 2025-10-10T01:18:34.0555221Z * [new branch] gh/williamwen42/294/base -> origin/gh/williamwen42/294/base 2025-10-10T01:18:34.0556920Z * [new branch] gh/williamwen42/294/head -> origin/gh/williamwen42/294/head 2025-10-10T01:18:34.0558740Z * [new branch] gh/williamwen42/294/orig -> origin/gh/williamwen42/294/orig 2025-10-10T01:18:34.0561186Z * [new branch] gh/williamwen42/295/base -> origin/gh/williamwen42/295/base 2025-10-10T01:18:34.0563003Z * [new branch] gh/williamwen42/295/head -> origin/gh/williamwen42/295/head 2025-10-10T01:18:34.0564848Z * [new branch] gh/williamwen42/295/orig -> origin/gh/williamwen42/295/orig 2025-10-10T01:18:34.0567563Z * [new branch] gh/williamwen42/296/base -> origin/gh/williamwen42/296/base 2025-10-10T01:18:34.0569263Z * [new branch] gh/williamwen42/296/head -> origin/gh/williamwen42/296/head 2025-10-10T01:18:34.0571043Z * [new branch] gh/williamwen42/296/orig -> origin/gh/williamwen42/296/orig 2025-10-10T01:18:34.0573453Z * [new branch] gh/williamwen42/297/base -> origin/gh/williamwen42/297/base 2025-10-10T01:18:34.0575104Z * [new branch] gh/williamwen42/297/head -> origin/gh/williamwen42/297/head 2025-10-10T01:18:34.0576742Z * [new branch] gh/williamwen42/297/orig -> origin/gh/williamwen42/297/orig 2025-10-10T01:18:34.0579185Z * [new branch] gh/williamwen42/298/base -> origin/gh/williamwen42/298/base 2025-10-10T01:18:34.0580926Z * [new branch] gh/williamwen42/298/head -> origin/gh/williamwen42/298/head 2025-10-10T01:18:34.0582623Z * [new branch] gh/williamwen42/298/orig -> origin/gh/williamwen42/298/orig 2025-10-10T01:18:34.0585220Z * [new branch] gh/williamwen42/299/base -> origin/gh/williamwen42/299/base 2025-10-10T01:18:34.0587338Z * [new branch] gh/williamwen42/299/head -> origin/gh/williamwen42/299/head 2025-10-10T01:18:34.0598930Z * [new branch] gh/williamwen42/299/orig -> origin/gh/williamwen42/299/orig 2025-10-10T01:18:34.0599405Z * [new branch] gh/williamwen42/300/base -> origin/gh/williamwen42/300/base 2025-10-10T01:18:34.0599816Z * [new branch] gh/williamwen42/300/head -> origin/gh/williamwen42/300/head 2025-10-10T01:18:34.0600245Z * [new branch] gh/williamwen42/300/orig -> origin/gh/williamwen42/300/orig 2025-10-10T01:18:34.0600642Z * [new branch] gh/williamwen42/301/base -> origin/gh/williamwen42/301/base 2025-10-10T01:18:34.0601054Z * [new branch] gh/williamwen42/301/head -> origin/gh/williamwen42/301/head 2025-10-10T01:18:34.0601471Z * [new branch] gh/williamwen42/301/orig -> origin/gh/williamwen42/301/orig 2025-10-10T01:18:34.0603681Z * [new branch] gh/williamwen42/302/base -> origin/gh/williamwen42/302/base 2025-10-10T01:18:34.0605349Z * [new branch] gh/williamwen42/302/head -> origin/gh/williamwen42/302/head 2025-10-10T01:18:34.0606988Z * [new branch] gh/williamwen42/302/orig -> origin/gh/williamwen42/302/orig 2025-10-10T01:18:34.0609468Z * [new branch] gh/williamwen42/303/base -> origin/gh/williamwen42/303/base 2025-10-10T01:18:34.0611161Z * [new branch] gh/williamwen42/303/head -> origin/gh/williamwen42/303/head 2025-10-10T01:18:34.0612861Z * [new branch] gh/williamwen42/303/orig -> origin/gh/williamwen42/303/orig 2025-10-10T01:18:34.0615277Z * [new branch] gh/williamwen42/304/base -> origin/gh/williamwen42/304/base 2025-10-10T01:18:34.0617066Z * [new branch] gh/williamwen42/304/head -> origin/gh/williamwen42/304/head 2025-10-10T01:18:34.0618732Z * [new branch] gh/williamwen42/304/orig -> origin/gh/williamwen42/304/orig 2025-10-10T01:18:34.0621114Z * [new branch] gh/williamwen42/305/base -> origin/gh/williamwen42/305/base 2025-10-10T01:18:34.0622856Z * [new branch] gh/williamwen42/305/head -> origin/gh/williamwen42/305/head 2025-10-10T01:18:34.0624803Z * [new branch] gh/williamwen42/305/orig -> origin/gh/williamwen42/305/orig 2025-10-10T01:18:34.0627034Z * [new branch] gh/williamwen42/306/base -> origin/gh/williamwen42/306/base 2025-10-10T01:18:34.0628724Z * [new branch] gh/williamwen42/306/head -> origin/gh/williamwen42/306/head 2025-10-10T01:18:34.0630545Z * [new branch] gh/williamwen42/306/orig -> origin/gh/williamwen42/306/orig 2025-10-10T01:18:34.0632873Z * [new branch] gh/williamwen42/307/base -> origin/gh/williamwen42/307/base 2025-10-10T01:18:34.0634610Z * [new branch] gh/williamwen42/307/head -> origin/gh/williamwen42/307/head 2025-10-10T01:18:34.0636307Z * [new branch] gh/williamwen42/307/orig -> origin/gh/williamwen42/307/orig 2025-10-10T01:18:34.0639186Z * [new branch] gh/xmfan/169/base -> origin/gh/xmfan/169/base 2025-10-10T01:18:34.0640915Z * [new branch] gh/xmfan/169/head -> origin/gh/xmfan/169/head 2025-10-10T01:18:34.0643107Z * [new branch] gh/xmfan/170/base -> origin/gh/xmfan/170/base 2025-10-10T01:18:34.0644743Z * [new branch] gh/xmfan/170/head -> origin/gh/xmfan/170/head 2025-10-10T01:18:34.0647008Z * [new branch] gh/xmfan/244/base -> origin/gh/xmfan/244/base 2025-10-10T01:18:34.0648699Z * [new branch] gh/xmfan/244/head -> origin/gh/xmfan/244/head 2025-10-10T01:18:34.0650382Z * [new branch] gh/xmfan/244/orig -> origin/gh/xmfan/244/orig 2025-10-10T01:18:34.0652728Z * [new branch] gh/xmfan/246/base -> origin/gh/xmfan/246/base 2025-10-10T01:18:34.0654569Z * [new branch] gh/xmfan/246/head -> origin/gh/xmfan/246/head 2025-10-10T01:18:34.0656218Z * [new branch] gh/xmfan/246/orig -> origin/gh/xmfan/246/orig 2025-10-10T01:18:34.0658509Z * [new branch] gh/xmfan/253/base -> origin/gh/xmfan/253/base 2025-10-10T01:18:34.0660206Z * [new branch] gh/xmfan/253/head -> origin/gh/xmfan/253/head 2025-10-10T01:18:34.0661905Z * [new branch] gh/xmfan/253/orig -> origin/gh/xmfan/253/orig 2025-10-10T01:18:34.0664687Z * [new branch] gh/xmfan/260/base -> origin/gh/xmfan/260/base 2025-10-10T01:18:34.0666647Z * [new branch] gh/xmfan/260/head -> origin/gh/xmfan/260/head 2025-10-10T01:18:34.0668382Z * [new branch] gh/xmfan/260/orig -> origin/gh/xmfan/260/orig 2025-10-10T01:18:34.0670633Z * [new branch] gh/xmfan/262/base -> origin/gh/xmfan/262/base 2025-10-10T01:18:34.0672352Z * [new branch] gh/xmfan/262/head -> origin/gh/xmfan/262/head 2025-10-10T01:18:34.0674010Z * [new branch] gh/xmfan/262/orig -> origin/gh/xmfan/262/orig 2025-10-10T01:18:34.0676823Z * [new branch] gh/xmfan/274/base -> origin/gh/xmfan/274/base 2025-10-10T01:18:34.0678132Z * [new branch] gh/xmfan/274/head -> origin/gh/xmfan/274/head 2025-10-10T01:18:34.0679959Z * [new branch] gh/xmfan/274/orig -> origin/gh/xmfan/274/orig 2025-10-10T01:18:34.0682148Z * [new branch] gh/xmfan/277/base -> origin/gh/xmfan/277/base 2025-10-10T01:18:34.0683844Z * [new branch] gh/xmfan/277/head -> origin/gh/xmfan/277/head 2025-10-10T01:18:34.0685503Z * [new branch] gh/xmfan/277/orig -> origin/gh/xmfan/277/orig 2025-10-10T01:18:34.0687870Z * [new branch] gh/xmfan/281/base -> origin/gh/xmfan/281/base 2025-10-10T01:18:34.0689487Z * [new branch] gh/xmfan/281/head -> origin/gh/xmfan/281/head 2025-10-10T01:18:34.0691167Z * [new branch] gh/xmfan/281/orig -> origin/gh/xmfan/281/orig 2025-10-10T01:18:34.0693543Z * [new branch] gh/xmfan/284/base -> origin/gh/xmfan/284/base 2025-10-10T01:18:34.0695298Z * [new branch] gh/xmfan/284/head -> origin/gh/xmfan/284/head 2025-10-10T01:18:34.0696991Z * [new branch] gh/xmfan/284/orig -> origin/gh/xmfan/284/orig 2025-10-10T01:18:34.0699618Z * [new branch] gh/xmfan/285/base -> origin/gh/xmfan/285/base 2025-10-10T01:18:34.0701507Z * [new branch] gh/xmfan/285/head -> origin/gh/xmfan/285/head 2025-10-10T01:18:34.0703095Z * [new branch] gh/xmfan/285/orig -> origin/gh/xmfan/285/orig 2025-10-10T01:18:34.0705260Z * [new branch] gh/xmfan/286/base -> origin/gh/xmfan/286/base 2025-10-10T01:18:34.0706932Z * [new branch] gh/xmfan/286/head -> origin/gh/xmfan/286/head 2025-10-10T01:18:34.0708794Z * [new branch] gh/xmfan/286/orig -> origin/gh/xmfan/286/orig 2025-10-10T01:18:34.0711142Z * [new branch] gh/xmfan/287/base -> origin/gh/xmfan/287/base 2025-10-10T01:18:34.0712775Z * [new branch] gh/xmfan/287/head -> origin/gh/xmfan/287/head 2025-10-10T01:18:34.0714420Z * [new branch] gh/xmfan/287/orig -> origin/gh/xmfan/287/orig 2025-10-10T01:18:34.0716798Z * [new branch] gh/xmfan/288/base -> origin/gh/xmfan/288/base 2025-10-10T01:18:34.0718615Z * [new branch] gh/xmfan/288/head -> origin/gh/xmfan/288/head 2025-10-10T01:18:34.0720457Z * [new branch] gh/xmfan/288/orig -> origin/gh/xmfan/288/orig 2025-10-10T01:18:34.0722744Z * [new branch] gh/xmfan/289/base -> origin/gh/xmfan/289/base 2025-10-10T01:18:34.0724686Z * [new branch] gh/xmfan/289/head -> origin/gh/xmfan/289/head 2025-10-10T01:18:34.0727141Z * [new branch] gh/xmfan/289/orig -> origin/gh/xmfan/289/orig 2025-10-10T01:18:34.0729900Z * [new branch] gh/xmfan/290/base -> origin/gh/xmfan/290/base 2025-10-10T01:18:34.0731594Z * [new branch] gh/xmfan/290/head -> origin/gh/xmfan/290/head 2025-10-10T01:18:34.0733464Z * [new branch] gh/xmfan/290/orig -> origin/gh/xmfan/290/orig 2025-10-10T01:18:34.0735869Z * [new branch] gh/xmfan/291/base -> origin/gh/xmfan/291/base 2025-10-10T01:18:34.0737636Z * [new branch] gh/xmfan/291/head -> origin/gh/xmfan/291/head 2025-10-10T01:18:34.0739313Z * [new branch] gh/xmfan/291/orig -> origin/gh/xmfan/291/orig 2025-10-10T01:18:34.0741811Z * [new branch] gh/xmfan/292/base -> origin/gh/xmfan/292/base 2025-10-10T01:18:34.0743626Z * [new branch] gh/xmfan/292/head -> origin/gh/xmfan/292/head 2025-10-10T01:18:34.0745372Z * [new branch] gh/xmfan/292/orig -> origin/gh/xmfan/292/orig 2025-10-10T01:18:34.0747793Z * [new branch] gh/xmfan/293/base -> origin/gh/xmfan/293/base 2025-10-10T01:18:34.0749519Z * [new branch] gh/xmfan/293/head -> origin/gh/xmfan/293/head 2025-10-10T01:18:34.0751170Z * [new branch] gh/xmfan/293/orig -> origin/gh/xmfan/293/orig 2025-10-10T01:18:34.0753379Z * [new branch] gh/xmfan/294/base -> origin/gh/xmfan/294/base 2025-10-10T01:18:34.0755077Z * [new branch] gh/xmfan/294/head -> origin/gh/xmfan/294/head 2025-10-10T01:18:34.0756773Z * [new branch] gh/xmfan/294/orig -> origin/gh/xmfan/294/orig 2025-10-10T01:18:34.0759475Z * [new branch] gh/xmfan/295/base -> origin/gh/xmfan/295/base 2025-10-10T01:18:34.0761293Z * [new branch] gh/xmfan/295/head -> origin/gh/xmfan/295/head 2025-10-10T01:18:34.0763095Z * [new branch] gh/xmfan/295/orig -> origin/gh/xmfan/295/orig 2025-10-10T01:18:34.0765452Z * [new branch] gh/xmfan/296/base -> origin/gh/xmfan/296/base 2025-10-10T01:18:34.0767264Z * [new branch] gh/xmfan/296/head -> origin/gh/xmfan/296/head 2025-10-10T01:18:34.0768905Z * [new branch] gh/xmfan/296/orig -> origin/gh/xmfan/296/orig 2025-10-10T01:18:34.0771403Z * [new branch] gh/xmfan/297/base -> origin/gh/xmfan/297/base 2025-10-10T01:18:34.0773329Z * [new branch] gh/xmfan/297/head -> origin/gh/xmfan/297/head 2025-10-10T01:18:34.0774907Z * [new branch] gh/xmfan/297/orig -> origin/gh/xmfan/297/orig 2025-10-10T01:18:34.0777341Z * [new branch] gh/xmfan/298/base -> origin/gh/xmfan/298/base 2025-10-10T01:18:34.0779053Z * [new branch] gh/xmfan/298/head -> origin/gh/xmfan/298/head 2025-10-10T01:18:34.0780769Z * [new branch] gh/xmfan/298/orig -> origin/gh/xmfan/298/orig 2025-10-10T01:18:34.0783083Z * [new branch] gh/xmfan/299/base -> origin/gh/xmfan/299/base 2025-10-10T01:18:34.0784806Z * [new branch] gh/xmfan/299/head -> origin/gh/xmfan/299/head 2025-10-10T01:18:34.0786537Z * [new branch] gh/xmfan/299/orig -> origin/gh/xmfan/299/orig 2025-10-10T01:18:34.0789234Z * [new branch] gh/xmfan/300/base -> origin/gh/xmfan/300/base 2025-10-10T01:18:34.0791007Z * [new branch] gh/xmfan/300/head -> origin/gh/xmfan/300/head 2025-10-10T01:18:34.0792648Z * [new branch] gh/xmfan/300/orig -> origin/gh/xmfan/300/orig 2025-10-10T01:18:34.0795067Z * [new branch] gh/xmfan/301/base -> origin/gh/xmfan/301/base 2025-10-10T01:18:34.0796803Z * [new branch] gh/xmfan/301/head -> origin/gh/xmfan/301/head 2025-10-10T01:18:34.0798540Z * [new branch] gh/xmfan/301/orig -> origin/gh/xmfan/301/orig 2025-10-10T01:18:34.0801123Z * [new branch] gh/xmfan/302/base -> origin/gh/xmfan/302/base 2025-10-10T01:18:34.0802800Z * [new branch] gh/xmfan/302/head -> origin/gh/xmfan/302/head 2025-10-10T01:18:34.0804497Z * [new branch] gh/xmfan/302/orig -> origin/gh/xmfan/302/orig 2025-10-10T01:18:34.0806828Z * [new branch] gh/xmfan/303/base -> origin/gh/xmfan/303/base 2025-10-10T01:18:34.0808492Z * [new branch] gh/xmfan/303/head -> origin/gh/xmfan/303/head 2025-10-10T01:18:34.0810180Z * [new branch] gh/xmfan/303/orig -> origin/gh/xmfan/303/orig 2025-10-10T01:18:34.0812417Z * [new branch] gh/xmfan/304/base -> origin/gh/xmfan/304/base 2025-10-10T01:18:34.0814086Z * [new branch] gh/xmfan/304/head -> origin/gh/xmfan/304/head 2025-10-10T01:18:34.0815801Z * [new branch] gh/xmfan/304/orig -> origin/gh/xmfan/304/orig 2025-10-10T01:18:34.0818658Z * [new branch] gh/xuanzhang816/14/base -> origin/gh/xuanzhang816/14/base 2025-10-10T01:18:34.0820406Z * [new branch] gh/xuanzhang816/14/head -> origin/gh/xuanzhang816/14/head 2025-10-10T01:18:34.0822092Z * [new branch] gh/xuanzhang816/14/orig -> origin/gh/xuanzhang816/14/orig 2025-10-10T01:18:34.0824541Z * [new branch] gh/xuanzhang816/22/base -> origin/gh/xuanzhang816/22/base 2025-10-10T01:18:34.0826416Z * [new branch] gh/xuanzhang816/22/head -> origin/gh/xuanzhang816/22/head 2025-10-10T01:18:34.0828002Z * [new branch] gh/xuanzhang816/22/orig -> origin/gh/xuanzhang816/22/orig 2025-10-10T01:18:34.0830265Z * [new branch] gh/xuanzhang816/23/base -> origin/gh/xuanzhang816/23/base 2025-10-10T01:18:34.0832001Z * [new branch] gh/xuanzhang816/23/head -> origin/gh/xuanzhang816/23/head 2025-10-10T01:18:34.0833717Z * [new branch] gh/xuanzhang816/23/orig -> origin/gh/xuanzhang816/23/orig 2025-10-10T01:18:34.0835973Z * [new branch] gh/xuanzhang816/25/base -> origin/gh/xuanzhang816/25/base 2025-10-10T01:18:34.0837716Z * [new branch] gh/xuanzhang816/25/head -> origin/gh/xuanzhang816/25/head 2025-10-10T01:18:34.0839537Z * [new branch] gh/xuanzhang816/25/orig -> origin/gh/xuanzhang816/25/orig 2025-10-10T01:18:34.0842080Z * [new branch] gh/xuanzhang816/26/base -> origin/gh/xuanzhang816/26/base 2025-10-10T01:18:34.0843635Z * [new branch] gh/xuanzhang816/26/head -> origin/gh/xuanzhang816/26/head 2025-10-10T01:18:34.0845302Z * [new branch] gh/xuanzhang816/26/orig -> origin/gh/xuanzhang816/26/orig 2025-10-10T01:18:34.0847551Z * [new branch] gh/xuanzhang816/27/base -> origin/gh/xuanzhang816/27/base 2025-10-10T01:18:34.0849224Z * [new branch] gh/xuanzhang816/27/head -> origin/gh/xuanzhang816/27/head 2025-10-10T01:18:34.0850898Z * [new branch] gh/xuanzhang816/27/orig -> origin/gh/xuanzhang816/27/orig 2025-10-10T01:18:34.0853650Z * [new branch] gh/xuanzhang816/28/base -> origin/gh/xuanzhang816/28/base 2025-10-10T01:18:34.0855376Z * [new branch] gh/xuanzhang816/28/head -> origin/gh/xuanzhang816/28/head 2025-10-10T01:18:34.0857161Z * [new branch] gh/xuanzhang816/28/orig -> origin/gh/xuanzhang816/28/orig 2025-10-10T01:18:34.0859453Z * [new branch] gh/xuanzhang816/29/base -> origin/gh/xuanzhang816/29/base 2025-10-10T01:18:34.0861131Z * [new branch] gh/xuanzhang816/29/head -> origin/gh/xuanzhang816/29/head 2025-10-10T01:18:34.0862848Z * [new branch] gh/xuanzhang816/29/orig -> origin/gh/xuanzhang816/29/orig 2025-10-10T01:18:34.0865208Z * [new branch] gh/xuanzhang816/30/base -> origin/gh/xuanzhang816/30/base 2025-10-10T01:18:34.0867505Z * [new branch] gh/xuanzhang816/30/head -> origin/gh/xuanzhang816/30/head 2025-10-10T01:18:34.0869197Z * [new branch] gh/xuanzhang816/30/orig -> origin/gh/xuanzhang816/30/orig 2025-10-10T01:18:34.0871462Z * [new branch] gh/xuanzhang816/31/base -> origin/gh/xuanzhang816/31/base 2025-10-10T01:18:34.0873135Z * [new branch] gh/xuanzhang816/31/head -> origin/gh/xuanzhang816/31/head 2025-10-10T01:18:34.0874938Z * [new branch] gh/xuanzhang816/31/orig -> origin/gh/xuanzhang816/31/orig 2025-10-10T01:18:34.0877301Z * [new branch] gh/xuanzhang816/32/base -> origin/gh/xuanzhang816/32/base 2025-10-10T01:18:34.0879179Z * [new branch] gh/xuanzhang816/32/head -> origin/gh/xuanzhang816/32/head 2025-10-10T01:18:34.0881380Z * [new branch] gh/xuanzhang816/32/orig -> origin/gh/xuanzhang816/32/orig 2025-10-10T01:18:34.0883679Z * [new branch] gh/xuanzhang816/33/base -> origin/gh/xuanzhang816/33/base 2025-10-10T01:18:34.0885430Z * [new branch] gh/xuanzhang816/33/head -> origin/gh/xuanzhang816/33/head 2025-10-10T01:18:34.0887151Z * [new branch] gh/xuanzhang816/33/orig -> origin/gh/xuanzhang816/33/orig 2025-10-10T01:18:34.0890023Z * [new branch] gh/yanbing-j/11/base -> origin/gh/yanbing-j/11/base 2025-10-10T01:18:34.0891685Z * [new branch] gh/yanbing-j/11/head -> origin/gh/yanbing-j/11/head 2025-10-10T01:18:34.0893436Z * [new branch] gh/yanbing-j/11/orig -> origin/gh/yanbing-j/11/orig 2025-10-10T01:18:34.0895710Z * [new branch] gh/yanbing-j/12/base -> origin/gh/yanbing-j/12/base 2025-10-10T01:18:34.0897419Z * [new branch] gh/yanbing-j/12/head -> origin/gh/yanbing-j/12/head 2025-10-10T01:18:34.0899136Z * [new branch] gh/yanbing-j/12/orig -> origin/gh/yanbing-j/12/orig 2025-10-10T01:18:34.0901448Z * [new branch] gh/yanbing-j/13/base -> origin/gh/yanbing-j/13/base 2025-10-10T01:18:34.0903140Z * [new branch] gh/yanbing-j/13/head -> origin/gh/yanbing-j/13/head 2025-10-10T01:18:34.0904862Z * [new branch] gh/yanbing-j/13/orig -> origin/gh/yanbing-j/13/orig 2025-10-10T01:18:34.0907171Z * [new branch] gh/yanbing-j/14/base -> origin/gh/yanbing-j/14/base 2025-10-10T01:18:34.0908894Z * [new branch] gh/yanbing-j/14/head -> origin/gh/yanbing-j/14/head 2025-10-10T01:18:34.0910788Z * [new branch] gh/yanbing-j/14/orig -> origin/gh/yanbing-j/14/orig 2025-10-10T01:18:34.0912875Z * [new branch] gh/yanbing-j/15/base -> origin/gh/yanbing-j/15/base 2025-10-10T01:18:34.0914590Z * [new branch] gh/yanbing-j/15/head -> origin/gh/yanbing-j/15/head 2025-10-10T01:18:34.0916285Z * [new branch] gh/yanbing-j/15/orig -> origin/gh/yanbing-j/15/orig 2025-10-10T01:18:34.0919045Z * [new branch] gh/yanbing-j/18/base -> origin/gh/yanbing-j/18/base 2025-10-10T01:18:34.0920765Z * [new branch] gh/yanbing-j/18/head -> origin/gh/yanbing-j/18/head 2025-10-10T01:18:34.0922447Z * [new branch] gh/yanbing-j/18/orig -> origin/gh/yanbing-j/18/orig 2025-10-10T01:18:34.0924976Z * [new branch] gh/yanbing-j/19/base -> origin/gh/yanbing-j/19/base 2025-10-10T01:18:34.0926668Z * [new branch] gh/yanbing-j/19/head -> origin/gh/yanbing-j/19/head 2025-10-10T01:18:34.0928316Z * [new branch] gh/yanbing-j/19/orig -> origin/gh/yanbing-j/19/orig 2025-10-10T01:18:34.0930742Z * [new branch] gh/yanbing-j/20/base -> origin/gh/yanbing-j/20/base 2025-10-10T01:18:34.0932584Z * [new branch] gh/yanbing-j/20/head -> origin/gh/yanbing-j/20/head 2025-10-10T01:18:34.0934570Z * [new branch] gh/yanbing-j/20/orig -> origin/gh/yanbing-j/20/orig 2025-10-10T01:18:34.0936761Z * [new branch] gh/yanbing-j/21/base -> origin/gh/yanbing-j/21/base 2025-10-10T01:18:34.0938423Z * [new branch] gh/yanbing-j/21/head -> origin/gh/yanbing-j/21/head 2025-10-10T01:18:34.0940649Z * [new branch] gh/yanbing-j/22/base -> origin/gh/yanbing-j/22/base 2025-10-10T01:18:34.0942457Z * [new branch] gh/yanbing-j/22/head -> origin/gh/yanbing-j/22/head 2025-10-10T01:18:34.0944131Z * [new branch] gh/yanbing-j/22/orig -> origin/gh/yanbing-j/22/orig 2025-10-10T01:18:34.0946925Z * [new branch] gh/yanbing-j/23/base -> origin/gh/yanbing-j/23/base 2025-10-10T01:18:34.0948227Z * [new branch] gh/yanbing-j/23/head -> origin/gh/yanbing-j/23/head 2025-10-10T01:18:34.0949921Z * [new branch] gh/yanbing-j/23/orig -> origin/gh/yanbing-j/23/orig 2025-10-10T01:18:34.0952170Z * [new branch] gh/yanbing-j/24/base -> origin/gh/yanbing-j/24/base 2025-10-10T01:18:34.0953856Z * [new branch] gh/yanbing-j/24/head -> origin/gh/yanbing-j/24/head 2025-10-10T01:18:34.0955547Z * [new branch] gh/yanbing-j/24/orig -> origin/gh/yanbing-j/24/orig 2025-10-10T01:18:34.0957894Z * [new branch] gh/yanbing-j/25/base -> origin/gh/yanbing-j/25/base 2025-10-10T01:18:34.0959742Z * [new branch] gh/yanbing-j/25/head -> origin/gh/yanbing-j/25/head 2025-10-10T01:18:34.0961393Z * [new branch] gh/yanbing-j/25/orig -> origin/gh/yanbing-j/25/orig 2025-10-10T01:18:34.0963655Z * [new branch] gh/yanbing-j/26/base -> origin/gh/yanbing-j/26/base 2025-10-10T01:18:34.0965383Z * [new branch] gh/yanbing-j/26/head -> origin/gh/yanbing-j/26/head 2025-10-10T01:18:34.0967086Z * [new branch] gh/yanbing-j/26/orig -> origin/gh/yanbing-j/26/orig 2025-10-10T01:18:34.0969352Z * [new branch] gh/yanbing-j/36/base -> origin/gh/yanbing-j/36/base 2025-10-10T01:18:34.0971027Z * [new branch] gh/yanbing-j/36/head -> origin/gh/yanbing-j/36/head 2025-10-10T01:18:34.0972748Z * [new branch] gh/yanbing-j/36/orig -> origin/gh/yanbing-j/36/orig 2025-10-10T01:18:34.0975510Z * [new branch] gh/yangw-dev/12/base -> origin/gh/yangw-dev/12/base 2025-10-10T01:18:34.0977216Z * [new branch] gh/yangw-dev/12/head -> origin/gh/yangw-dev/12/head 2025-10-10T01:18:34.0979145Z * [new branch] gh/yangw-dev/12/orig -> origin/gh/yangw-dev/12/orig 2025-10-10T01:18:34.0981362Z * [new branch] gh/yangw-dev/13/base -> origin/gh/yangw-dev/13/base 2025-10-10T01:18:34.0983102Z * [new branch] gh/yangw-dev/13/head -> origin/gh/yangw-dev/13/head 2025-10-10T01:18:34.0984773Z * [new branch] gh/yangw-dev/13/orig -> origin/gh/yangw-dev/13/orig 2025-10-10T01:18:34.0987076Z * [new branch] gh/yangw-dev/14/base -> origin/gh/yangw-dev/14/base 2025-10-10T01:18:34.0988869Z * [new branch] gh/yangw-dev/14/head -> origin/gh/yangw-dev/14/head 2025-10-10T01:18:34.0990998Z * [new branch] gh/yangw-dev/14/orig -> origin/gh/yangw-dev/14/orig 2025-10-10T01:18:34.0993315Z * [new branch] gh/yangw-dev/15/base -> origin/gh/yangw-dev/15/base 2025-10-10T01:18:34.0995031Z * [new branch] gh/yangw-dev/15/head -> origin/gh/yangw-dev/15/head 2025-10-10T01:18:34.0996692Z * [new branch] gh/yangw-dev/15/orig -> origin/gh/yangw-dev/15/orig 2025-10-10T01:18:34.0999061Z * [new branch] gh/yangw-dev/19/base -> origin/gh/yangw-dev/19/base 2025-10-10T01:18:34.1000877Z * [new branch] gh/yangw-dev/19/head -> origin/gh/yangw-dev/19/head 2025-10-10T01:18:34.1002839Z * [new branch] gh/yangw-dev/19/orig -> origin/gh/yangw-dev/19/orig 2025-10-10T01:18:34.1005076Z * [new branch] gh/yangw-dev/26/base -> origin/gh/yangw-dev/26/base 2025-10-10T01:18:34.1006794Z * [new branch] gh/yangw-dev/26/head -> origin/gh/yangw-dev/26/head 2025-10-10T01:18:34.1008496Z * [new branch] gh/yangw-dev/26/orig -> origin/gh/yangw-dev/26/orig 2025-10-10T01:18:34.1010796Z * [new branch] gh/yangw-dev/27/base -> origin/gh/yangw-dev/27/base 2025-10-10T01:18:34.1012452Z * [new branch] gh/yangw-dev/27/head -> origin/gh/yangw-dev/27/head 2025-10-10T01:18:34.1014141Z * [new branch] gh/yangw-dev/27/orig -> origin/gh/yangw-dev/27/orig 2025-10-10T01:18:34.1016953Z * [new branch] gh/ydwu4/262/base -> origin/gh/ydwu4/262/base 2025-10-10T01:18:34.1018742Z * [new branch] gh/ydwu4/262/head -> origin/gh/ydwu4/262/head 2025-10-10T01:18:34.1020397Z * [new branch] gh/ydwu4/262/orig -> origin/gh/ydwu4/262/orig 2025-10-10T01:18:34.1022665Z * [new branch] gh/ydwu4/263/base -> origin/gh/ydwu4/263/base 2025-10-10T01:18:34.1024665Z * [new branch] gh/ydwu4/263/head -> origin/gh/ydwu4/263/head 2025-10-10T01:18:34.1026456Z * [new branch] gh/ydwu4/263/orig -> origin/gh/ydwu4/263/orig 2025-10-10T01:18:34.1028805Z * [new branch] gh/ydwu4/269/base -> origin/gh/ydwu4/269/base 2025-10-10T01:18:34.1030422Z * [new branch] gh/ydwu4/269/head -> origin/gh/ydwu4/269/head 2025-10-10T01:18:34.1032109Z * [new branch] gh/ydwu4/269/orig -> origin/gh/ydwu4/269/orig 2025-10-10T01:18:34.1034483Z * [new branch] gh/ydwu4/270/base -> origin/gh/ydwu4/270/base 2025-10-10T01:18:34.1036186Z * [new branch] gh/ydwu4/270/head -> origin/gh/ydwu4/270/head 2025-10-10T01:18:34.1037915Z * [new branch] gh/ydwu4/270/orig -> origin/gh/ydwu4/270/orig 2025-10-10T01:18:34.1040378Z * [new branch] gh/ydwu4/272/base -> origin/gh/ydwu4/272/base 2025-10-10T01:18:34.1042144Z * [new branch] gh/ydwu4/272/head -> origin/gh/ydwu4/272/head 2025-10-10T01:18:34.1043973Z * [new branch] gh/ydwu4/272/orig -> origin/gh/ydwu4/272/orig 2025-10-10T01:18:34.1046602Z * [new branch] gh/ydwu4/275/base -> origin/gh/ydwu4/275/base 2025-10-10T01:18:34.1048292Z * [new branch] gh/ydwu4/275/head -> origin/gh/ydwu4/275/head 2025-10-10T01:18:34.1049669Z * [new branch] gh/ydwu4/275/orig -> origin/gh/ydwu4/275/orig 2025-10-10T01:18:34.1051822Z * [new branch] gh/ydwu4/276/base -> origin/gh/ydwu4/276/base 2025-10-10T01:18:34.1053545Z * [new branch] gh/ydwu4/276/head -> origin/gh/ydwu4/276/head 2025-10-10T01:18:34.1055141Z * [new branch] gh/ydwu4/276/orig -> origin/gh/ydwu4/276/orig 2025-10-10T01:18:34.1057632Z * [new branch] gh/ydwu4/283/base -> origin/gh/ydwu4/283/base 2025-10-10T01:18:34.1059322Z * [new branch] gh/ydwu4/283/head -> origin/gh/ydwu4/283/head 2025-10-10T01:18:34.1060983Z * [new branch] gh/ydwu4/283/orig -> origin/gh/ydwu4/283/orig 2025-10-10T01:18:34.1063340Z * [new branch] gh/ydwu4/292/base -> origin/gh/ydwu4/292/base 2025-10-10T01:18:34.1064958Z * [new branch] gh/ydwu4/292/head -> origin/gh/ydwu4/292/head 2025-10-10T01:18:34.1066649Z * [new branch] gh/ydwu4/292/orig -> origin/gh/ydwu4/292/orig 2025-10-10T01:18:34.1069141Z * [new branch] gh/ydwu4/294/base -> origin/gh/ydwu4/294/base 2025-10-10T01:18:34.1070917Z * [new branch] gh/ydwu4/294/head -> origin/gh/ydwu4/294/head 2025-10-10T01:18:34.1072581Z * [new branch] gh/ydwu4/294/orig -> origin/gh/ydwu4/294/orig 2025-10-10T01:18:34.1075013Z * [new branch] gh/ydwu4/295/base -> origin/gh/ydwu4/295/base 2025-10-10T01:18:34.1076774Z * [new branch] gh/ydwu4/295/head -> origin/gh/ydwu4/295/head 2025-10-10T01:18:34.1078584Z * [new branch] gh/ydwu4/295/orig -> origin/gh/ydwu4/295/orig 2025-10-10T01:18:34.1080926Z * [new branch] gh/ydwu4/296/base -> origin/gh/ydwu4/296/base 2025-10-10T01:18:34.1082509Z * [new branch] gh/ydwu4/296/head -> origin/gh/ydwu4/296/head 2025-10-10T01:18:34.1084239Z * [new branch] gh/ydwu4/296/orig -> origin/gh/ydwu4/296/orig 2025-10-10T01:18:34.1086689Z * [new branch] gh/ydwu4/306/base -> origin/gh/ydwu4/306/base 2025-10-10T01:18:34.1088444Z * [new branch] gh/ydwu4/306/head -> origin/gh/ydwu4/306/head 2025-10-10T01:18:34.1090164Z * [new branch] gh/ydwu4/306/orig -> origin/gh/ydwu4/306/orig 2025-10-10T01:18:34.1092476Z * [new branch] gh/ydwu4/312/base -> origin/gh/ydwu4/312/base 2025-10-10T01:18:34.1094269Z * [new branch] gh/ydwu4/312/head -> origin/gh/ydwu4/312/head 2025-10-10T01:18:34.1095961Z * [new branch] gh/ydwu4/312/orig -> origin/gh/ydwu4/312/orig 2025-10-10T01:18:34.1098390Z * [new branch] gh/ydwu4/318/base -> origin/gh/ydwu4/318/base 2025-10-10T01:18:34.1100135Z * [new branch] gh/ydwu4/318/head -> origin/gh/ydwu4/318/head 2025-10-10T01:18:34.1101894Z * [new branch] gh/ydwu4/318/orig -> origin/gh/ydwu4/318/orig 2025-10-10T01:18:34.1104188Z * [new branch] gh/ydwu4/319/base -> origin/gh/ydwu4/319/base 2025-10-10T01:18:34.1105855Z * [new branch] gh/ydwu4/319/head -> origin/gh/ydwu4/319/head 2025-10-10T01:18:34.1107564Z * [new branch] gh/ydwu4/319/orig -> origin/gh/ydwu4/319/orig 2025-10-10T01:18:34.1110238Z * [new branch] gh/ydwu4/320/base -> origin/gh/ydwu4/320/base 2025-10-10T01:18:34.1111933Z * [new branch] gh/ydwu4/320/head -> origin/gh/ydwu4/320/head 2025-10-10T01:18:34.1113627Z * [new branch] gh/ydwu4/320/orig -> origin/gh/ydwu4/320/orig 2025-10-10T01:18:34.1116040Z * [new branch] gh/ydwu4/321/base -> origin/gh/ydwu4/321/base 2025-10-10T01:18:34.1117911Z * [new branch] gh/ydwu4/321/head -> origin/gh/ydwu4/321/head 2025-10-10T01:18:34.1119653Z * [new branch] gh/ydwu4/321/orig -> origin/gh/ydwu4/321/orig 2025-10-10T01:18:34.1121958Z * [new branch] gh/ydwu4/322/base -> origin/gh/ydwu4/322/base 2025-10-10T01:18:34.1123621Z * [new branch] gh/ydwu4/322/head -> origin/gh/ydwu4/322/head 2025-10-10T01:18:34.1126645Z * [new branch] gh/ydwu4/322/orig -> origin/gh/ydwu4/322/orig 2025-10-10T01:18:34.1129010Z * [new branch] gh/ydwu4/324/base -> origin/gh/ydwu4/324/base 2025-10-10T01:18:34.1130674Z * [new branch] gh/ydwu4/324/head -> origin/gh/ydwu4/324/head 2025-10-10T01:18:34.1132311Z * [new branch] gh/ydwu4/324/orig -> origin/gh/ydwu4/324/orig 2025-10-10T01:18:34.1134712Z * [new branch] gh/ydwu4/325/base -> origin/gh/ydwu4/325/base 2025-10-10T01:18:34.1137097Z * [new branch] gh/ydwu4/325/head -> origin/gh/ydwu4/325/head 2025-10-10T01:18:34.1138415Z * [new branch] gh/ydwu4/325/orig -> origin/gh/ydwu4/325/orig 2025-10-10T01:18:34.1140679Z * [new branch] gh/ydwu4/326/base -> origin/gh/ydwu4/326/base 2025-10-10T01:18:34.1142417Z * [new branch] gh/ydwu4/326/head -> origin/gh/ydwu4/326/head 2025-10-10T01:18:34.1144102Z * [new branch] gh/ydwu4/326/orig -> origin/gh/ydwu4/326/orig 2025-10-10T01:18:34.1146761Z * [new branch] gh/ydwu4/327/base -> origin/gh/ydwu4/327/base 2025-10-10T01:18:34.1148448Z * [new branch] gh/ydwu4/327/head -> origin/gh/ydwu4/327/head 2025-10-10T01:18:34.1150125Z * [new branch] gh/ydwu4/327/orig -> origin/gh/ydwu4/327/orig 2025-10-10T01:18:34.1152486Z * [new branch] gh/ydwu4/328/base -> origin/gh/ydwu4/328/base 2025-10-10T01:18:34.1154404Z * [new branch] gh/ydwu4/328/head -> origin/gh/ydwu4/328/head 2025-10-10T01:18:34.1155955Z * [new branch] gh/ydwu4/328/orig -> origin/gh/ydwu4/328/orig 2025-10-10T01:18:34.1158927Z * [new branch] gh/ydwu4/329/base -> origin/gh/ydwu4/329/base 2025-10-10T01:18:34.1160233Z * [new branch] gh/ydwu4/329/head -> origin/gh/ydwu4/329/head 2025-10-10T01:18:34.1161781Z * [new branch] gh/ydwu4/329/orig -> origin/gh/ydwu4/329/orig 2025-10-10T01:18:34.1164598Z * [new branch] gh/ydwu4/330/base -> origin/gh/ydwu4/330/base 2025-10-10T01:18:34.1165835Z * [new branch] gh/ydwu4/330/head -> origin/gh/ydwu4/330/head 2025-10-10T01:18:34.1167164Z * [new branch] gh/ydwu4/330/orig -> origin/gh/ydwu4/330/orig 2025-10-10T01:18:34.1169594Z * [new branch] gh/ydwu4/331/base -> origin/gh/ydwu4/331/base 2025-10-10T01:18:34.1171314Z * [new branch] gh/ydwu4/331/head -> origin/gh/ydwu4/331/head 2025-10-10T01:18:34.1172961Z * [new branch] gh/ydwu4/331/orig -> origin/gh/ydwu4/331/orig 2025-10-10T01:18:34.1175182Z * [new branch] gh/ydwu4/332/base -> origin/gh/ydwu4/332/base 2025-10-10T01:18:34.1176892Z * [new branch] gh/ydwu4/332/head -> origin/gh/ydwu4/332/head 2025-10-10T01:18:34.1178559Z * [new branch] gh/ydwu4/332/orig -> origin/gh/ydwu4/332/orig 2025-10-10T01:18:34.1180772Z * [new branch] gh/ydwu4/333/base -> origin/gh/ydwu4/333/base 2025-10-10T01:18:34.1182485Z * [new branch] gh/ydwu4/333/head -> origin/gh/ydwu4/333/head 2025-10-10T01:18:34.1184148Z * [new branch] gh/ydwu4/333/orig -> origin/gh/ydwu4/333/orig 2025-10-10T01:18:34.1186822Z * [new branch] gh/ydwu4/334/base -> origin/gh/ydwu4/334/base 2025-10-10T01:18:34.1188664Z * [new branch] gh/ydwu4/334/head -> origin/gh/ydwu4/334/head 2025-10-10T01:18:34.1190236Z * [new branch] gh/ydwu4/334/orig -> origin/gh/ydwu4/334/orig 2025-10-10T01:18:34.1192498Z * [new branch] gh/ydwu4/335/base -> origin/gh/ydwu4/335/base 2025-10-10T01:18:34.1194170Z * [new branch] gh/ydwu4/335/head -> origin/gh/ydwu4/335/head 2025-10-10T01:18:34.1195830Z * [new branch] gh/ydwu4/335/orig -> origin/gh/ydwu4/335/orig 2025-10-10T01:18:34.1198186Z * [new branch] gh/ydwu4/336/base -> origin/gh/ydwu4/336/base 2025-10-10T01:18:34.1199958Z * [new branch] gh/ydwu4/336/head -> origin/gh/ydwu4/336/head 2025-10-10T01:18:34.1201610Z * [new branch] gh/ydwu4/336/orig -> origin/gh/ydwu4/336/orig 2025-10-10T01:18:34.1203788Z * [new branch] gh/ydwu4/337/base -> origin/gh/ydwu4/337/base 2025-10-10T01:18:34.1205487Z * [new branch] gh/ydwu4/337/head -> origin/gh/ydwu4/337/head 2025-10-10T01:18:34.1207092Z * [new branch] gh/ydwu4/337/orig -> origin/gh/ydwu4/337/orig 2025-10-10T01:18:34.1209837Z * [new branch] gh/yf225/133/base -> origin/gh/yf225/133/base 2025-10-10T01:18:34.1211535Z * [new branch] gh/yf225/133/head -> origin/gh/yf225/133/head 2025-10-10T01:18:34.1213850Z * [new branch] gh/yf225/93/base -> origin/gh/yf225/93/base 2025-10-10T01:18:34.1215483Z * [new branch] gh/yf225/93/head -> origin/gh/yf225/93/head 2025-10-10T01:18:34.1218669Z * [new branch] gh/yifuwang/152/base -> origin/gh/yifuwang/152/base 2025-10-10T01:18:34.1220598Z * [new branch] gh/yifuwang/152/head -> origin/gh/yifuwang/152/head 2025-10-10T01:18:34.1222268Z * [new branch] gh/yifuwang/152/orig -> origin/gh/yifuwang/152/orig 2025-10-10T01:18:34.1224867Z * [new branch] gh/yifuwang/195/base -> origin/gh/yifuwang/195/base 2025-10-10T01:18:34.1227046Z * [new branch] gh/yifuwang/195/head -> origin/gh/yifuwang/195/head 2025-10-10T01:18:34.1229101Z * [new branch] gh/yifuwang/195/orig -> origin/gh/yifuwang/195/orig 2025-10-10T01:18:34.1232074Z * [new branch] gh/yiming0416/1/base -> origin/gh/yiming0416/1/base 2025-10-10T01:18:34.1233386Z * [new branch] gh/yiming0416/1/head -> origin/gh/yiming0416/1/head 2025-10-10T01:18:34.1235659Z * [new branch] gh/yiming0416/2/base -> origin/gh/yiming0416/2/base 2025-10-10T01:18:34.1237282Z * [new branch] gh/yiming0416/2/head -> origin/gh/yiming0416/2/head 2025-10-10T01:18:34.1240852Z * [new branch] gh/ysiraichi/88/base -> origin/gh/ysiraichi/88/base 2025-10-10T01:18:34.1242488Z * [new branch] gh/ysiraichi/88/head -> origin/gh/ysiraichi/88/head 2025-10-10T01:18:34.1244208Z * [new branch] gh/ysiraichi/88/orig -> origin/gh/ysiraichi/88/orig 2025-10-10T01:18:34.1247457Z * [new branch] gh/zhxchen17/25/base -> origin/gh/zhxchen17/25/base 2025-10-10T01:18:34.1248799Z * [new branch] gh/zhxchen17/25/head -> origin/gh/zhxchen17/25/head 2025-10-10T01:18:34.1250429Z * [new branch] gh/zhxchen17/25/orig -> origin/gh/zhxchen17/25/orig 2025-10-10T01:18:34.1253245Z * [new branch] gh/zhxchen17/31/base -> origin/gh/zhxchen17/31/base 2025-10-10T01:18:34.1254539Z * [new branch] gh/zhxchen17/31/head -> origin/gh/zhxchen17/31/head 2025-10-10T01:18:34.1256213Z * [new branch] gh/zhxchen17/31/orig -> origin/gh/zhxchen17/31/orig 2025-10-10T01:18:34.1258556Z * [new branch] gh/zhxchen17/34/base -> origin/gh/zhxchen17/34/base 2025-10-10T01:18:34.1260432Z * [new branch] gh/zhxchen17/34/head -> origin/gh/zhxchen17/34/head 2025-10-10T01:18:34.1262681Z * [new branch] gh/zhxchen17/35/base -> origin/gh/zhxchen17/35/base 2025-10-10T01:18:34.1264072Z * [new branch] gh/zhxchen17/35/head -> origin/gh/zhxchen17/35/head 2025-10-10T01:18:34.1266725Z * [new branch] gh/zklaus/10/base -> origin/gh/zklaus/10/base 2025-10-10T01:18:34.1268725Z * [new branch] gh/zklaus/10/head -> origin/gh/zklaus/10/head 2025-10-10T01:18:34.1270232Z * [new branch] gh/zklaus/10/orig -> origin/gh/zklaus/10/orig 2025-10-10T01:18:34.1272378Z * [new branch] gh/zklaus/11/base -> origin/gh/zklaus/11/base 2025-10-10T01:18:34.1274105Z * [new branch] gh/zklaus/11/head -> origin/gh/zklaus/11/head 2025-10-10T01:18:34.1275857Z * [new branch] gh/zklaus/11/orig -> origin/gh/zklaus/11/orig 2025-10-10T01:18:34.1278901Z * [new branch] gh/zklaus/15/base -> origin/gh/zklaus/15/base 2025-10-10T01:18:34.1280599Z * [new branch] gh/zklaus/15/head -> origin/gh/zklaus/15/head 2025-10-10T01:18:34.1282228Z * [new branch] gh/zklaus/15/orig -> origin/gh/zklaus/15/orig 2025-10-10T01:18:34.1284585Z * [new branch] gh/zklaus/16/base -> origin/gh/zklaus/16/base 2025-10-10T01:18:34.1286290Z * [new branch] gh/zklaus/16/head -> origin/gh/zklaus/16/head 2025-10-10T01:18:34.1287940Z * [new branch] gh/zklaus/16/orig -> origin/gh/zklaus/16/orig 2025-10-10T01:18:34.1290159Z * [new branch] gh/zklaus/17/base -> origin/gh/zklaus/17/base 2025-10-10T01:18:34.1291856Z * [new branch] gh/zklaus/17/head -> origin/gh/zklaus/17/head 2025-10-10T01:18:34.1293715Z * [new branch] gh/zklaus/17/orig -> origin/gh/zklaus/17/orig 2025-10-10T01:18:34.1296003Z * [new branch] gh/zklaus/18/base -> origin/gh/zklaus/18/base 2025-10-10T01:18:34.1297594Z * [new branch] gh/zklaus/18/head -> origin/gh/zklaus/18/head 2025-10-10T01:18:34.1299263Z * [new branch] gh/zklaus/18/orig -> origin/gh/zklaus/18/orig 2025-10-10T01:18:34.1301592Z * [new branch] gh/zklaus/19/base -> origin/gh/zklaus/19/base 2025-10-10T01:18:34.1303220Z * [new branch] gh/zklaus/19/head -> origin/gh/zklaus/19/head 2025-10-10T01:18:34.1304926Z * [new branch] gh/zklaus/19/orig -> origin/gh/zklaus/19/orig 2025-10-10T01:18:34.1307190Z * [new branch] gh/zklaus/7/base -> origin/gh/zklaus/7/base 2025-10-10T01:18:34.1308891Z * [new branch] gh/zklaus/7/head -> origin/gh/zklaus/7/head 2025-10-10T01:18:34.1310566Z * [new branch] gh/zklaus/7/orig -> origin/gh/zklaus/7/orig 2025-10-10T01:18:34.1313339Z * [new branch] gh/zou3519/1177/base -> origin/gh/zou3519/1177/base 2025-10-10T01:18:34.1315038Z * [new branch] gh/zou3519/1177/head -> origin/gh/zou3519/1177/head 2025-10-10T01:18:34.1316700Z * [new branch] gh/zou3519/1177/orig -> origin/gh/zou3519/1177/orig 2025-10-10T01:18:34.1319161Z * [new branch] gh/zou3519/1195/base -> origin/gh/zou3519/1195/base 2025-10-10T01:18:34.1321333Z * [new branch] gh/zou3519/1195/head -> origin/gh/zou3519/1195/head 2025-10-10T01:18:34.1322936Z * [new branch] gh/zou3519/1195/orig -> origin/gh/zou3519/1195/orig 2025-10-10T01:18:34.1325847Z * [new branch] gh/zou3519/1196/base -> origin/gh/zou3519/1196/base 2025-10-10T01:18:34.1327408Z * [new branch] gh/zou3519/1196/head -> origin/gh/zou3519/1196/head 2025-10-10T01:18:34.1328994Z * [new branch] gh/zou3519/1196/orig -> origin/gh/zou3519/1196/orig 2025-10-10T01:18:34.1332030Z * [new branch] gh/zou3519/1197/base -> origin/gh/zou3519/1197/base 2025-10-10T01:18:34.1333470Z * [new branch] gh/zou3519/1197/head -> origin/gh/zou3519/1197/head 2025-10-10T01:18:34.1335132Z * [new branch] gh/zou3519/1197/orig -> origin/gh/zou3519/1197/orig 2025-10-10T01:18:34.1337228Z * [new branch] gh/zou3519/1198/base -> origin/gh/zou3519/1198/base 2025-10-10T01:18:34.1338972Z * [new branch] gh/zou3519/1198/head -> origin/gh/zou3519/1198/head 2025-10-10T01:18:34.1340695Z * [new branch] gh/zou3519/1198/orig -> origin/gh/zou3519/1198/orig 2025-10-10T01:18:34.1342869Z * [new branch] gh/zou3519/1199/base -> origin/gh/zou3519/1199/base 2025-10-10T01:18:34.1344651Z * [new branch] gh/zou3519/1199/head -> origin/gh/zou3519/1199/head 2025-10-10T01:18:34.1346850Z * [new branch] gh/zou3519/1199/orig -> origin/gh/zou3519/1199/orig 2025-10-10T01:18:34.1349225Z * [new branch] gh/zou3519/1200/base -> origin/gh/zou3519/1200/base 2025-10-10T01:18:34.1350922Z * [new branch] gh/zou3519/1200/head -> origin/gh/zou3519/1200/head 2025-10-10T01:18:34.1352705Z * [new branch] gh/zou3519/1200/orig -> origin/gh/zou3519/1200/orig 2025-10-10T01:18:34.1354783Z * [new branch] gh/zou3519/1201/base -> origin/gh/zou3519/1201/base 2025-10-10T01:18:34.1357271Z * [new branch] gh/zou3519/1201/head -> origin/gh/zou3519/1201/head 2025-10-10T01:18:34.1358695Z * [new branch] gh/zou3519/1201/orig -> origin/gh/zou3519/1201/orig 2025-10-10T01:18:34.1362137Z * [new branch] gh/zpcore/1/base -> origin/gh/zpcore/1/base 2025-10-10T01:18:34.1363452Z * [new branch] gh/zpcore/1/head -> origin/gh/zpcore/1/head 2025-10-10T01:18:34.1366355Z * [new branch] gh/zpcore/11/base -> origin/gh/zpcore/11/base 2025-10-10T01:18:34.1367629Z * [new branch] gh/zpcore/11/head -> origin/gh/zpcore/11/head 2025-10-10T01:18:34.1369489Z * [new branch] gh/zpcore/11/orig -> origin/gh/zpcore/11/orig 2025-10-10T01:18:34.1372138Z * [new branch] gh/zpcore/12/base -> origin/gh/zpcore/12/base 2025-10-10T01:18:34.1373785Z * [new branch] gh/zpcore/12/head -> origin/gh/zpcore/12/head 2025-10-10T01:18:34.1375469Z * [new branch] gh/zpcore/12/orig -> origin/gh/zpcore/12/orig 2025-10-10T01:18:34.1377853Z * [new branch] gh/zpcore/13/base -> origin/gh/zpcore/13/base 2025-10-10T01:18:34.1379528Z * [new branch] gh/zpcore/13/head -> origin/gh/zpcore/13/head 2025-10-10T01:18:34.1381166Z * [new branch] gh/zpcore/13/orig -> origin/gh/zpcore/13/orig 2025-10-10T01:18:34.1383466Z * [new branch] gh/zpcore/14/base -> origin/gh/zpcore/14/base 2025-10-10T01:18:34.1385206Z * [new branch] gh/zpcore/14/head -> origin/gh/zpcore/14/head 2025-10-10T01:18:34.1386870Z * [new branch] gh/zpcore/14/orig -> origin/gh/zpcore/14/orig 2025-10-10T01:18:34.1389316Z * [new branch] gh/zpcore/15/base -> origin/gh/zpcore/15/base 2025-10-10T01:18:34.1390961Z * [new branch] gh/zpcore/15/head -> origin/gh/zpcore/15/head 2025-10-10T01:18:34.1392758Z * [new branch] gh/zpcore/15/orig -> origin/gh/zpcore/15/orig 2025-10-10T01:18:34.1395054Z * [new branch] gh/zpcore/16/base -> origin/gh/zpcore/16/base 2025-10-10T01:18:34.1396741Z * [new branch] gh/zpcore/16/head -> origin/gh/zpcore/16/head 2025-10-10T01:18:34.1398490Z * [new branch] gh/zpcore/16/orig -> origin/gh/zpcore/16/orig 2025-10-10T01:18:34.1401035Z * [new branch] gh/zpcore/17/base -> origin/gh/zpcore/17/base 2025-10-10T01:18:34.1402692Z * [new branch] gh/zpcore/17/head -> origin/gh/zpcore/17/head 2025-10-10T01:18:34.1404203Z * [new branch] gh/zpcore/17/orig -> origin/gh/zpcore/17/orig 2025-10-10T01:18:34.1406447Z * [new branch] gh/zpcore/18/base -> origin/gh/zpcore/18/base 2025-10-10T01:18:34.1408208Z * [new branch] gh/zpcore/18/head -> origin/gh/zpcore/18/head 2025-10-10T01:18:34.1409804Z * [new branch] gh/zpcore/18/orig -> origin/gh/zpcore/18/orig 2025-10-10T01:18:34.1412039Z * [new branch] gh/zpcore/19/base -> origin/gh/zpcore/19/base 2025-10-10T01:18:34.1413673Z * [new branch] gh/zpcore/19/head -> origin/gh/zpcore/19/head 2025-10-10T01:18:34.1415311Z * [new branch] gh/zpcore/19/orig -> origin/gh/zpcore/19/orig 2025-10-10T01:18:34.1417635Z * [new branch] gh/zpcore/2/base -> origin/gh/zpcore/2/base 2025-10-10T01:18:34.1419364Z * [new branch] gh/zpcore/2/head -> origin/gh/zpcore/2/head 2025-10-10T01:18:34.1421701Z * [new branch] gh/zpcore/20/base -> origin/gh/zpcore/20/base 2025-10-10T01:18:34.1423345Z * [new branch] gh/zpcore/20/head -> origin/gh/zpcore/20/head 2025-10-10T01:18:34.1425404Z * [new branch] gh/zpcore/20/orig -> origin/gh/zpcore/20/orig 2025-10-10T01:18:34.1428362Z * [new branch] gh/zpcore/21/base -> origin/gh/zpcore/21/base 2025-10-10T01:18:34.1430103Z * [new branch] gh/zpcore/21/head -> origin/gh/zpcore/21/head 2025-10-10T01:18:34.1431816Z * [new branch] gh/zpcore/21/orig -> origin/gh/zpcore/21/orig 2025-10-10T01:18:34.1434385Z * [new branch] gh/zpcore/3/base -> origin/gh/zpcore/3/base 2025-10-10T01:18:34.1436254Z * [new branch] gh/zpcore/3/head -> origin/gh/zpcore/3/head 2025-10-10T01:18:34.1439432Z * [new branch] gh/zpcore/4/base -> origin/gh/zpcore/4/base 2025-10-10T01:18:34.1440596Z * [new branch] gh/zpcore/4/head -> origin/gh/zpcore/4/head 2025-10-10T01:18:34.1442611Z * [new branch] gh/zpcore/5/base -> origin/gh/zpcore/5/base 2025-10-10T01:18:34.1444241Z * [new branch] gh/zpcore/5/head -> origin/gh/zpcore/5/head 2025-10-10T01:18:34.1447331Z * [new branch] gh/zpcore/6/base -> origin/gh/zpcore/6/base 2025-10-10T01:18:34.1448792Z * [new branch] gh/zpcore/6/head -> origin/gh/zpcore/6/head 2025-10-10T01:18:34.1450847Z * [new branch] gh/zpcore/7/base -> origin/gh/zpcore/7/base 2025-10-10T01:18:34.1452437Z * [new branch] gh/zpcore/7/head -> origin/gh/zpcore/7/head 2025-10-10T01:18:34.1454614Z * [new branch] gh/zpcore/8/base -> origin/gh/zpcore/8/base 2025-10-10T01:18:34.1456399Z * [new branch] gh/zpcore/8/head -> origin/gh/zpcore/8/head 2025-10-10T01:18:34.1458313Z * [new branch] google-main -> origin/google-main 2025-10-10T01:18:34.1460056Z * [new branch] greencontext -> origin/greencontext 2025-10-10T01:18:34.1462470Z * [new branch] guangyey/config -> origin/guangyey/config 2025-10-10T01:18:34.1464138Z * [new branch] guangyey/external_stream -> origin/guangyey/external_stream 2025-10-10T01:18:34.1465829Z * [new branch] guangyey/reimport -> origin/guangyey/reimport 2025-10-10T01:18:34.1467914Z * [new branch] guangyey/test_2025 -> origin/guangyey/test_2025 2025-10-10T01:18:34.1470759Z * [new branch] guilhermeleobas/cherry-pick-55d87d9dfd9 -> origin/guilhermeleobas/cherry-pick-55d87d9dfd9 2025-10-10T01:18:34.1472974Z * [new branch] hameerabbasi/gradcheck-allclose -> origin/hameerabbasi/gradcheck-allclose 2025-10-10T01:18:34.1474800Z * [new branch] haozhe/bf16-dynamic-shape -> origin/haozhe/bf16-dynamic-shape 2025-10-10T01:18:34.1476648Z * [new branch] hc_baseline -> origin/hc_baseline 2025-10-10T01:18:34.1478603Z * [new branch] hhh_decomp_mul -> origin/hhh_decomp_mul 2025-10-10T01:18:34.1480565Z * [new branch] hhh_rand -> origin/hhh_rand 2025-10-10T01:18:34.1482775Z * [new branch] hoy/triton-PR3973 -> origin/hoy/triton-PR3973 2025-10-10T01:18:34.1485174Z * [new branch] huba/debug_mode -> origin/huba/debug_mode 2025-10-10T01:18:34.1486771Z * [new branch] huba/dtensor_equal -> origin/huba/dtensor_equal 2025-10-10T01:18:34.1488283Z * [new branch] huba/f1 -> origin/huba/f1 2025-10-10T01:18:34.1490011Z * [new branch] huba/local_tensor -> origin/huba/local_tensor 2025-10-10T01:18:34.1491782Z * [new branch] ideep-update -> origin/ideep-update 2025-10-10T01:18:34.1493628Z * [new branch] increase-asan-build-memory -> origin/increase-asan-build-memory 2025-10-10T01:18:34.1495654Z * [new branch] inductor-perf-increase-timeout -> origin/inductor-perf-increase-timeout 2025-10-10T01:18:34.1497235Z * [new branch] inductordecompfix -> origin/inductordecompfix 2025-10-10T01:18:34.1498990Z * [new branch] inline -> origin/inline 2025-10-10T01:18:34.1500816Z * [new branch] inlining -> origin/inlining 2025-10-10T01:18:34.1502605Z * [new branch] inlining-ezyang -> origin/inlining-ezyang 2025-10-10T01:18:34.1504372Z * [new branch] install-torchao-0.13.0 -> origin/install-torchao-0.13.0 2025-10-10T01:18:34.1506382Z * [new branch] install_free_tensors -> origin/install_free_tensors 2025-10-10T01:18:34.1508428Z * [new branch] int8_sdpa -> origin/int8_sdpa 2025-10-10T01:18:34.1510060Z * [new branch] invoke-subgraph -> origin/invoke-subgraph 2025-10-10T01:18:34.1511940Z * [new branch] issue#58739 -> origin/issue#58739 2025-10-10T01:18:34.1513993Z * [new branch] issue-161010-dynamo-stride-clone -> origin/issue-161010-dynamo-stride-clone 2025-10-10T01:18:34.1516180Z * [new branch] jathu/o3 -> origin/jathu/o3 2025-10-10T01:18:34.1517958Z * [new branch] jathu/sve -> origin/jathu/sve 2025-10-10T01:18:34.1520577Z * [new branch] jcaip/test-cusparselt-version-0.6.2 -> origin/jcaip/test-cusparselt-version-0.6.2 2025-10-10T01:18:34.1522136Z * [new branch] jcaip/update-cusparselt-0.6.2 -> origin/jcaip/update-cusparselt-0.6.2 2025-10-10T01:18:34.1523837Z * [new branch] jeanschmidt-patch-1 -> origin/jeanschmidt-patch-1 2025-10-10T01:18:34.1526015Z * [new branch] jerryzh168-patch-1 -> origin/jerryzh168-patch-1 2025-10-10T01:18:34.1527763Z * [new branch] jithunnair-amd-patch-1 -> origin/jithunnair-amd-patch-1 2025-10-10T01:18:34.1529644Z * [new branch] jithunnair-amd-patch-2 -> origin/jithunnair-amd-patch-2 2025-10-10T01:18:34.1531446Z * [new branch] jithunnair-amd-patch-3 -> origin/jithunnair-amd-patch-3 2025-10-10T01:18:34.1533271Z * [new branch] jithunnair-amd-patch-4 -> origin/jithunnair-amd-patch-4 2025-10-10T01:18:34.1535645Z * [new branch] justinchu/allowlist-api-onnx -> origin/justinchu/allowlist-api-onnx 2025-10-10T01:18:34.1537220Z * [new branch] justinchu/attention-tests -> origin/justinchu/attention-tests 2025-10-10T01:18:34.1539307Z * [new branch] justinchu/native-qdq -> origin/justinchu/native-qdq 2025-10-10T01:18:34.1541863Z * [new branch] justinchuby/typo-error -> origin/justinchuby/typo-error 2025-10-10T01:18:34.1544014Z * [new branch] kainan666/xlf_debug -> origin/kainan666/xlf_debug 2025-10-10T01:18:34.1545754Z * [new branch] kainan_test -> origin/kainan_test 2025-10-10T01:18:34.1548185Z * [new branch] leslie/test_group_gemm_epilogues -> origin/leslie/test_group_gemm_epilogues 2025-10-10T01:18:34.1550403Z * [new branch] lessw2020/fix_cutlass_cache_error -> origin/lessw2020/fix_cutlass_cache_error 2025-10-10T01:18:34.1552580Z * [new branch] liaoxuan/shm_all_reduce -> origin/liaoxuan/shm_all_reduce 2025-10-10T01:18:34.1554218Z * [new branch] liaoxuan/test_fa_disable_softmax -> origin/liaoxuan/test_fa_disable_softmax 2025-10-10T01:18:34.1555791Z * [new branch] liaoxuan/test_int8_sdpa -> origin/liaoxuan/test_int8_sdpa 2025-10-10T01:18:34.1557756Z * [new branch] libtorch_free_so -> origin/libtorch_free_so 2025-10-10T01:18:34.1559938Z * [new branch] lintbuilddocker -> origin/lintbuilddocker 2025-10-10T01:18:34.1562034Z * [new branch] llama4-stable -> origin/llama4-stable 2025-10-10T01:18:34.1563844Z * [new branch] logdetfix -> origin/logdetfix 2025-10-10T01:18:34.1565561Z * [new branch] logsumexp -> origin/logsumexp 2025-10-10T01:18:34.1568561Z * [new branch] lts/release/1.8 -> origin/lts/release/1.8 2025-10-10T01:18:34.1570953Z * [new branch] lucaskabela/#94773 -> origin/lucaskabela/#94773 2025-10-10T01:18:34.1572557Z * [new branch] lucaskabela/cherrypick_163769 -> origin/lucaskabela/cherrypick_163769 2025-10-10T01:18:34.1574138Z * [new branch] lucaskabela/fix_164814 -> origin/lucaskabela/fix_164814 2025-10-10T01:18:34.1575880Z * [new branch] lucaskabela/fix_164823 -> origin/lucaskabela/fix_164823 2025-10-10T01:18:34.1577417Z * [new branch] lucaskabela/fix_164875 -> origin/lucaskabela/fix_164875 2025-10-10T01:18:34.1578978Z * [new branch] lucaskabela/flop_counter -> origin/lucaskabela/flop_counter 2025-10-10T01:18:34.1580588Z * [new branch] lucaskabela/func_under_decomp -> origin/lucaskabela/func_under_decomp 2025-10-10T01:18:34.1582521Z * [new branch] lucaskabela/functional_in_dynamo -> origin/lucaskabela/functional_in_dynamo 2025-10-10T01:18:34.1584653Z * [new branch] lucaskabela/install_params_as_graph_attr -> origin/lucaskabela/install_params_as_graph_attr 2025-10-10T01:18:34.1586387Z * [new branch] lucaskabela/parameters_as_graph_attr -> origin/lucaskabela/parameters_as_graph_attr 2025-10-10T01:18:34.1588178Z * [new branch] lucaskabela/remove_aot_dispatcher_metadata -> origin/lucaskabela/remove_aot_dispatcher_metadata 2025-10-10T01:18:34.1589821Z * [new branch] lucaskabela/rnn_decomp -> origin/lucaskabela/rnn_decomp 2025-10-10T01:18:34.1591544Z * [new branch] lucaskabela/typing_backends -> origin/lucaskabela/typing_backends 2025-10-10T01:18:34.1593288Z * [new branch] main -> origin/main 2025-10-10T01:18:34.1595189Z * [new branch] main-enable-b200-distributed-tests -> origin/main-enable-b200-distributed-tests 2025-10-10T01:18:34.1597006Z * [new branch] main-enable-b200-symm-mem-test -> origin/main-enable-b200-symm-mem-test 2025-10-10T01:18:34.1598934Z * [new branch] malfet-patch-1 -> origin/malfet-patch-1 2025-10-10T01:18:34.1600832Z * [new branch] malfet-patch-14 -> origin/malfet-patch-14 2025-10-10T01:18:34.1602973Z * [new branch] malfet-patch-2 -> origin/malfet-patch-2 2025-10-10T01:18:34.1604753Z * [new branch] malfet-patch-3 -> origin/malfet-patch-3 2025-10-10T01:18:34.1606679Z * [new branch] malfet-patch-4 -> origin/malfet-patch-4 2025-10-10T01:18:34.1608374Z * [new branch] malfet-patch-5 -> origin/malfet-patch-5 2025-10-10T01:18:34.1610250Z * [new branch] malfet-patch-6 -> origin/malfet-patch-6 2025-10-10T01:18:34.1612054Z * [new branch] malfet-patch-7 -> origin/malfet-patch-7 2025-10-10T01:18:34.1613865Z * [new branch] malfet-patch-8 -> origin/malfet-patch-8 2025-10-10T01:18:34.1615780Z * [new branch] malfet-patch-9 -> origin/malfet-patch-9 2025-10-10T01:18:34.1618308Z * [new branch] malfet/be-move-more-settings-to-checkout-pytorch -> origin/malfet/be-move-more-settings-to-checkout-pytorch 2025-10-10T01:18:34.1619830Z * [new branch] malfet/mps-implement-col2im -> origin/malfet/mps-implement-col2im 2025-10-10T01:18:34.1622211Z * [new branch] manuel/aoti_metal_shimify-thread_safe -> origin/manuel/aoti_metal_shimify-thread_safe 2025-10-10T01:18:34.1623832Z * [new branch] manuel/test-ops-common-allow-mps -> origin/manuel/test-ops-common-allow-mps 2025-10-10T01:18:34.1626614Z * [new branch] masnesral/metaconda -> origin/masnesral/metaconda 2025-10-10T01:18:34.1628318Z * [new branch] masnesral/pt2_internal_logging -> origin/masnesral/pt2_internal_logging 2025-10-10T01:18:34.1629979Z * [new branch] metascroy-patch-1 -> origin/metascroy-patch-1 2025-10-10T01:18:34.1631692Z * [new branch] mingw_constant_buffer -> origin/mingw_constant_buffer 2025-10-10T01:18:34.1634022Z * [new branch] mlazos/S429861-debug -> origin/mlazos/S429861-debug 2025-10-10T01:18:34.1635783Z * [new branch] mlazos/aa -> origin/mlazos/aa 2025-10-10T01:18:34.1637447Z * [new branch] mlazos/acts -> origin/mlazos/acts 2025-10-10T01:18:34.1639314Z * [new branch] mlazos/arg-renames -> origin/mlazos/arg-renames 2025-10-10T01:18:34.1640986Z * [new branch] mlazos/backup-test-branch -> origin/mlazos/backup-test-branch 2025-10-10T01:18:34.1642594Z * [new branch] mlazos/bad-cudagraphs -> origin/mlazos/bad-cudagraphs 2025-10-10T01:18:34.1644121Z * [new branch] mlazos/baseline -> origin/mlazos/baseline 2025-10-10T01:18:34.1646358Z * [new branch] mlazos/baseline-graph-breaks -> origin/mlazos/baseline-graph-breaks 2025-10-10T01:18:34.1648296Z * [new branch] mlazos/beta-tensor -> origin/mlazos/beta-tensor 2025-10-10T01:18:34.1650884Z * [new branch] mlazos/buffers -> origin/mlazos/buffers 2025-10-10T01:18:34.1652456Z * [new branch] mlazos/buffers2 -> origin/mlazos/buffers2 2025-10-10T01:18:34.1654168Z * [new branch] mlazos/buffers3 -> origin/mlazos/buffers3 2025-10-10T01:18:34.1656256Z * [new branch] mlazos/ck2 -> origin/mlazos/ck2 2025-10-10T01:18:34.1658251Z * [new branch] mlazos/combokernels -> origin/mlazos/combokernels 2025-10-10T01:18:34.1659843Z * [new branch] mlazos/ctx-cleanup -> origin/mlazos/ctx-cleanup 2025-10-10T01:18:34.1661539Z * [new branch] mlazos/cuda-cmd-log -> origin/mlazos/cuda-cmd-log 2025-10-10T01:18:34.1663356Z * [new branch] mlazos/cudagraph-tests -> origin/mlazos/cudagraph-tests 2025-10-10T01:18:34.1665158Z * [new branch] mlazos/cudagraphs-measurement -> origin/mlazos/cudagraphs-measurement 2025-10-10T01:18:34.1666775Z * [new branch] mlazos/cutlass-test -> origin/mlazos/cutlass-test 2025-10-10T01:18:34.1668703Z * [new branch] mlazos/cutlass-topo-bug -> origin/mlazos/cutlass-topo-bug 2025-10-10T01:18:34.1670577Z * [new branch] mlazos/dataclass-proxy -> origin/mlazos/dataclass-proxy 2025-10-10T01:18:34.1672239Z * [new branch] mlazos/dc-attrs -> origin/mlazos/dc-attrs 2025-10-10T01:18:34.1673889Z * [new branch] mlazos/dc-helion -> origin/mlazos/dc-helion 2025-10-10T01:18:34.1675746Z * [new branch] mlazos/dict-fix -> origin/mlazos/dict-fix 2025-10-10T01:18:34.1677947Z * [new branch] mlazos/disable-tf -> origin/mlazos/disable-tf 2025-10-10T01:18:34.1679660Z * [new branch] mlazos/dupe-fix -> origin/mlazos/dupe-fix 2025-10-10T01:18:34.1681355Z * [new branch] mlazos/dyn-batch -> origin/mlazos/dyn-batch 2025-10-10T01:18:34.1682979Z * [new branch] mlazos/evt -> origin/mlazos/evt 2025-10-10T01:18:34.1684806Z * [new branch] mlazos/extract-examples -> origin/mlazos/extract-examples 2025-10-10T01:18:34.1687015Z * [new branch] mlazos/foreach-op -> origin/mlazos/foreach-op 2025-10-10T01:18:34.1689018Z * [new branch] mlazos/fp8 -> origin/mlazos/fp8 2025-10-10T01:18:34.1690732Z * [new branch] mlazos/fp8-bias -> origin/mlazos/fp8-bias 2025-10-10T01:18:34.1692514Z * [new branch] mlazos/fp8-bias-fusion -> origin/mlazos/fp8-bias-fusion 2025-10-10T01:18:34.1694235Z * [new branch] mlazos/fp8-fixes -> origin/mlazos/fp8-fixes 2025-10-10T01:18:34.1696005Z * [new branch] mlazos/freezing -> origin/mlazos/freezing 2025-10-10T01:18:34.1698561Z * [new branch] mlazos/h-comp -> origin/mlazos/h-comp 2025-10-10T01:18:34.1700265Z * [new branch] mlazos/h-comp2 -> origin/mlazos/h-comp2 2025-10-10T01:18:34.1702018Z * [new branch] mlazos/hash-hop -> origin/mlazos/hash-hop 2025-10-10T01:18:34.1703729Z * [new branch] mlazos/hc -> origin/mlazos/hc 2025-10-10T01:18:34.1705469Z * [new branch] mlazos/hc-cycles -> origin/mlazos/hc-cycles 2025-10-10T01:18:34.1707125Z * [new branch] mlazos/hc-fixes -> origin/mlazos/hc-fixes 2025-10-10T01:18:34.1708851Z * [new branch] mlazos/hc-fixes3 -> origin/mlazos/hc-fixes3 2025-10-10T01:18:34.1710521Z * [new branch] mlazos/hc-fixes4 -> origin/mlazos/hc-fixes4 2025-10-10T01:18:34.1712297Z * [new branch] mlazos/hc-hf -> origin/mlazos/hc-hf 2025-10-10T01:18:34.1713984Z * [new branch] mlazos/hc-mut -> origin/mlazos/hc-mut 2025-10-10T01:18:34.1715777Z * [new branch] mlazos/hc10 -> origin/mlazos/hc10 2025-10-10T01:18:34.1717858Z * [new branch] mlazos/hc11 -> origin/mlazos/hc11 2025-10-10T01:18:34.1719575Z * [new branch] mlazos/hc12 -> origin/mlazos/hc12 2025-10-10T01:18:34.1721208Z * [new branch] mlazos/hc13 -> origin/mlazos/hc13 2025-10-10T01:18:34.1722901Z * [new branch] mlazos/hc14 -> origin/mlazos/hc14 2025-10-10T01:18:34.1724692Z * [new branch] mlazos/hc15 -> origin/mlazos/hc15 2025-10-10T01:18:34.1726988Z * [new branch] mlazos/hc2 -> origin/mlazos/hc2 2025-10-10T01:18:34.1728842Z * [new branch] mlazos/hc4 -> origin/mlazos/hc4 2025-10-10T01:18:34.1730549Z * [new branch] mlazos/hc5 -> origin/mlazos/hc5 2025-10-10T01:18:34.1732230Z * [new branch] mlazos/hc6 -> origin/mlazos/hc6 2025-10-10T01:18:34.1733995Z * [new branch] mlazos/hc7 -> origin/mlazos/hc7 2025-10-10T01:18:34.1735608Z * [new branch] mlazos/hc8 -> origin/mlazos/hc8 2025-10-10T01:18:34.1737888Z * [new branch] mlazos/hc9 -> origin/mlazos/hc9 2025-10-10T01:18:34.1738954Z * [new branch] mlazos/hc_baseline2 -> origin/mlazos/hc_baseline2 2025-10-10T01:18:34.1740898Z * [new branch] mlazos/inductor-streams -> origin/mlazos/inductor-streams 2025-10-10T01:18:34.1742629Z * [new branch] mlazos/lr-composibility -> origin/mlazos/lr-composibility 2025-10-10T01:18:34.1744163Z * [new branch] mlazos/main -> origin/mlazos/main 2025-10-10T01:18:34.1745972Z * [new branch] mlazos/main-test-enablement -> origin/mlazos/main-test-enablement 2025-10-10T01:18:34.1748019Z * [new branch] mlazos/mark-static-update -> origin/mlazos/mark-static-update 2025-10-10T01:18:34.1749744Z * [new branch] mlazos/mcg -> origin/mlazos/mcg 2025-10-10T01:18:34.1751432Z * [new branch] mlazos/mcg2 -> origin/mlazos/mcg2 2025-10-10T01:18:34.1753193Z * [new branch] mlazos/meta-guards -> origin/mlazos/meta-guards 2025-10-10T01:18:34.1755757Z * [new branch] mlazos/mlazos/ck2 -> origin/mlazos/mlazos/ck2 2025-10-10T01:18:34.1756991Z * [new branch] mlazos/mlazos/foreach-map-adam -> origin/mlazos/mlazos/foreach-map-adam 2025-10-10T01:18:34.1758882Z * [new branch] mlazos/mlazos/tf-mode-backup -> origin/mlazos/mlazos/tf-mode-backup 2025-10-10T01:18:34.1760595Z * [new branch] mlazos/mod-fix -> origin/mlazos/mod-fix 2025-10-10T01:18:34.1762371Z * [new branch] mlazos/mode-fix -> origin/mlazos/mode-fix 2025-10-10T01:18:34.1764136Z * [new branch] mlazos/more-tests -> origin/mlazos/more-tests 2025-10-10T01:18:34.1765856Z * [new branch] mlazos/offsets -> origin/mlazos/offsets 2025-10-10T01:18:34.1767655Z * [new branch] mlazos/proxy-ctors -> origin/mlazos/proxy-ctors 2025-10-10T01:18:34.1769348Z * [new branch] mlazos/quant-fix -> origin/mlazos/quant-fix 2025-10-10T01:18:34.1771519Z * [new branch] mlazos/resnet-fix -> origin/mlazos/resnet-fix 2025-10-10T01:18:34.1773243Z * [new branch] mlazos/rm-buf-names -> origin/mlazos/rm-buf-names 2025-10-10T01:18:34.1774926Z * [new branch] mlazos/rm-code -> origin/mlazos/rm-code 2025-10-10T01:18:34.1776625Z * [new branch] mlazos/rm-spam -> origin/mlazos/rm-spam 2025-10-10T01:18:34.1778491Z * [new branch] mlazos/rtp -> origin/mlazos/rtp 2025-10-10T01:18:34.1780230Z * [new branch] mlazos/static-idx-dbg -> origin/mlazos/static-idx-dbg 2025-10-10T01:18:34.1781987Z * [new branch] mlazos/static-inputs-log -> origin/mlazos/static-inputs-log 2025-10-10T01:18:34.1783666Z * [new branch] mlazos/td-fix2 -> origin/mlazos/td-fix2 2025-10-10T01:18:34.1785448Z * [new branch] mlazos/tensor-hasattr2 -> origin/mlazos/tensor-hasattr2 2025-10-10T01:18:34.1787126Z * [new branch] mlazos/test -> origin/mlazos/test 2025-10-10T01:18:34.1788887Z * [new branch] mlazos/tf-mode -> origin/mlazos/tf-mode 2025-10-10T01:18:34.1790632Z * [new branch] mlazos/tf-mode-backup2 -> origin/mlazos/tf-mode-backup2 2025-10-10T01:18:34.1792345Z * [new branch] mlazos/tf-mode-reland -> origin/mlazos/tf-mode-reland 2025-10-10T01:18:34.1794170Z * [new branch] mlazos/tf-mode-reland2 -> origin/mlazos/tf-mode-reland2 2025-10-10T01:18:34.1795894Z * [new branch] mlazos/tf-mode-reland3 -> origin/mlazos/tf-mode-reland3 2025-10-10T01:18:34.1797875Z * [new branch] mlazos/triton-no-epi -> origin/mlazos/triton-no-epi 2025-10-10T01:18:34.1799658Z * [new branch] mlazos/tune-proto -> origin/mlazos/tune-proto 2025-10-10T01:18:34.1801586Z * [new branch] mlazos/tuple-fixes -> origin/mlazos/tuple-fixes 2025-10-10T01:18:34.1803105Z * [new branch] mlazos/tuple-fixes2 -> origin/mlazos/tuple-fixes2 2025-10-10T01:18:34.1804841Z * [new branch] mlazos/tuple-handling -> origin/mlazos/tuple-handling 2025-10-10T01:18:34.1806745Z * [new branch] mlazos/user-streams -> origin/mlazos/user-streams 2025-10-10T01:18:34.1808615Z * [new branch] mlazos/user-streams-backup -> origin/mlazos/user-streams-backup 2025-10-10T01:18:34.1810506Z * [new branch] mlazos/vary-beta -> origin/mlazos/vary-beta 2025-10-10T01:18:34.1812245Z * [new branch] mlazos/vary-beta2 -> origin/mlazos/vary-beta2 2025-10-10T01:18:34.1813940Z * [new branch] mlazos/weird-perf1 -> origin/mlazos/weird-perf1 2025-10-10T01:18:34.1815709Z * [new branch] mm_out_dtype_compile -> origin/mm_out_dtype_compile 2025-10-10T01:18:34.1817491Z * [new branch] module-shim -> origin/module-shim 2025-10-10T01:18:34.1819185Z * [new branch] module-stack -> origin/module-stack 2025-10-10T01:18:34.1820925Z * [new branch] more_ck_Fixes -> origin/more_ck_Fixes 2025-10-10T01:18:34.1822678Z * [new branch] move-theme-out-docker -> origin/move-theme-out-docker 2025-10-10T01:18:34.1824570Z * [new branch] move_aws_steps_inside_setup_rocm -> origin/move_aws_steps_inside_setup_rocm 2025-10-10T01:18:34.1827751Z * [new branch] msaroufim-patch-1 -> origin/msaroufim-patch-1 2025-10-10T01:18:34.1830707Z * [new branch] msaroufim/be1 -> origin/msaroufim/be1 2025-10-10T01:18:34.1832159Z * [new branch] msaroufim/cn_path -> origin/msaroufim/cn_path 2025-10-10T01:18:34.1833661Z * [new branch] msaroufim/cub -> origin/msaroufim/cub 2025-10-10T01:18:34.1835511Z * [new branch] msaroufim/dtensorfusedadam -> origin/msaroufim/dtensorfusedadam 2025-10-10T01:18:34.1837111Z * [new branch] msaroufim/patchx -> origin/msaroufim/patchx 2025-10-10T01:18:34.1838998Z * [new branch] msaroufim/reduce -> origin/msaroufim/reduce 2025-10-10T01:18:34.1841287Z * [new branch] mtia/basic-cmake -> origin/mtia/basic-cmake 2025-10-10T01:18:34.1843618Z * [new branch] mwizak/fix-triton-block-shape -> origin/mwizak/fix-triton-block-shape 2025-10-10T01:18:34.1845272Z * [new branch] my_varlen_backup -> origin/my_varlen_backup 2025-10-10T01:18:34.1847623Z * [new branch] nWEIdia/skip-tests-for-pr-159494 -> origin/nWEIdia/skip-tests-for-pr-159494 2025-10-10T01:18:34.1849380Z * [new branch] nativert_num_outputs -> origin/nativert_num_outputs 2025-10-10T01:18:34.1851152Z * [new branch] new-codegen -> origin/new-codegen 2025-10-10T01:18:34.1852917Z * [new branch] newtest-base -> origin/newtest-base 2025-10-10T01:18:34.1855169Z * [new branch] ngimel/allgather_format -> origin/ngimel/allgather_format 2025-10-10T01:18:34.1856760Z * [new branch] ngimel/cat_perf2 -> origin/ngimel/cat_perf2 2025-10-10T01:18:34.1858383Z * [new branch] ngimel/error_index_list -> origin/ngimel/error_index_list 2025-10-10T01:18:34.1859893Z * [new branch] ngimel/gg_new -> origin/ngimel/gg_new 2025-10-10T01:18:34.1861561Z * [new branch] ngimel/scatter_add_multid -> origin/ngimel/scatter_add_multid 2025-10-10T01:18:34.1863347Z * [new branch] nightly -> origin/nightly 2025-10-10T01:18:34.1865831Z * [new branch] nikitaved/addmm_1_rowcol_lt_path_check -> origin/nikitaved/addmm_1_rowcol_lt_path_check 2025-10-10T01:18:34.1867470Z * [new branch] nikitaved/addmm_epilogue_fusions -> origin/nikitaved/addmm_epilogue_fusions 2025-10-10T01:18:34.1868972Z * [new branch] nikitaved/addmm_epilogue_fusions_scratch -> origin/nikitaved/addmm_epilogue_fusions_scratch 2025-10-10T01:18:34.1870529Z * [new branch] nikitaved/simpler_can_use_32bit_index -> origin/nikitaved/simpler_can_use_32bit_index 2025-10-10T01:18:34.1872023Z * [new branch] nikitaved/test -> origin/nikitaved/test 2025-10-10T01:18:34.1873810Z * [new branch] nmacchioni-patch-10 -> origin/nmacchioni-patch-10 2025-10-10T01:18:34.1875828Z * [new branch] nmacchioni-patch-7 -> origin/nmacchioni-patch-7 2025-10-10T01:18:34.1877662Z * [new branch] nmacchioni-patch-8 -> origin/nmacchioni-patch-8 2025-10-10T01:18:34.1879617Z * [new branch] nmacchioni-patch-9 -> origin/nmacchioni-patch-9 2025-10-10T01:18:34.1881281Z * [new branch] no_distributed_log_spew -> origin/no_distributed_log_spew 2025-10-10T01:18:34.1883157Z * [new branch] nofun-hack -> origin/nofun-hack 2025-10-10T01:18:34.1885388Z * [new branch] nullplay/fuse_matmul -> origin/nullplay/fuse_matmul 2025-10-10T01:18:34.1887282Z * [new branch] nullplay_fuse_matmul -> origin/nullplay_fuse_matmul 2025-10-10T01:18:34.1890013Z * [new branch] orig/release/1.10 -> origin/orig/release/1.10 2025-10-10T01:18:34.1891741Z * [new branch] orig/release/1.11 -> origin/orig/release/1.11 2025-10-10T01:18:34.1893458Z * [new branch] orig/release/1.12 -> origin/orig/release/1.12 2025-10-10T01:18:34.1895351Z * [new branch] orig/release/1.13 -> origin/orig/release/1.13 2025-10-10T01:18:34.1897112Z * [new branch] orig/release/1.6 -> origin/orig/release/1.6 2025-10-10T01:18:34.1898976Z * [new branch] orig/release/1.7 -> origin/orig/release/1.7 2025-10-10T01:18:34.1901171Z * [new branch] orig/release/1.8 -> origin/orig/release/1.8 2025-10-10T01:18:34.1902923Z * [new branch] orig/release/1.9 -> origin/orig/release/1.9 2025-10-10T01:18:34.1904636Z * [new branch] orig/release/2.0 -> origin/orig/release/2.0 2025-10-10T01:18:34.1906271Z * [new branch] orig/release/2.1 -> origin/orig/release/2.1 2025-10-10T01:18:34.1907979Z * [new branch] orig/release/2.2 -> origin/orig/release/2.2 2025-10-10T01:18:34.1909651Z * [new branch] orig/release/2.3 -> origin/orig/release/2.3 2025-10-10T01:18:34.1911313Z * [new branch] orig/release/2.4 -> origin/orig/release/2.4 2025-10-10T01:18:34.1913042Z * [new branch] orig/release/2.5 -> origin/orig/release/2.5 2025-10-10T01:18:34.1914716Z * [new branch] orig/release/2.6 -> origin/orig/release/2.6 2025-10-10T01:18:34.1916659Z * [new branch] orig/release/2.7 -> origin/orig/release/2.7 2025-10-10T01:18:34.1918866Z * [new branch] orig/release/2.8 -> origin/orig/release/2.8 2025-10-10T01:18:34.1920580Z * [new branch] orig/release/2.9 -> origin/orig/release/2.9 2025-10-10T01:18:34.1924382Z * [new branch] origin/gh/fxdawnn/1/base -> origin/origin/gh/fxdawnn/1/base 2025-10-10T01:18:34.1926117Z * [new branch] origin/gh/fxdawnn/1/orig -> origin/origin/gh/fxdawnn/1/orig 2025-10-10T01:18:34.1928906Z * [new branch] origin/gh/zpcore/14/orig -> origin/origin/gh/zpcore/14/orig 2025-10-10T01:18:34.1930508Z * [new branch] padded-tensor -> origin/padded-tensor 2025-10-10T01:18:34.1932306Z * [new branch] pca2 -> origin/pca2 2025-10-10T01:18:34.1934380Z * [new branch] perf_ops -> origin/perf_ops 2025-10-10T01:18:34.1935964Z * [new branch] perf_ops_2_9 -> origin/perf_ops_2_9 2025-10-10T01:18:34.1938041Z * [new branch] perserve_node_meta_decomp -> origin/perserve_node_meta_decomp 2025-10-10T01:18:34.1939741Z * [new branch] pianpwk-patch-1 -> origin/pianpwk-patch-1 2025-10-10T01:18:34.1942638Z * [new branch] pianpwk/__draft_debug_mode -> origin/pianpwk/__draft_debug_mode 2025-10-10T01:18:34.1943877Z * [new branch] pianpwk/_super_draft_debug_mode -> origin/pianpwk/_super_draft_debug_mode 2025-10-10T01:18:34.1945500Z * [new branch] pianpwk/backed_size_oblivious_export -> origin/pianpwk/backed_size_oblivious_export 2025-10-10T01:18:34.1946980Z * [new branch] pianpwk/base_view_shape_key -> origin/pianpwk/base_view_shape_key 2025-10-10T01:18:34.1948585Z * [new branch] pianpwk/bert_dynamic_perf -> origin/pianpwk/bert_dynamic_perf 2025-10-10T01:18:34.1950492Z * [new branch] pianpwk/debug_mode_hacks -> origin/pianpwk/debug_mode_hacks 2025-10-10T01:18:34.1952510Z * [new branch] pianpwk/debug_mode_inductor -> origin/pianpwk/debug_mode_inductor 2025-10-10T01:18:34.1954221Z * [new branch] pianpwk/debug_mode_show_ids -> origin/pianpwk/debug_mode_show_ids 2025-10-10T01:18:34.1956281Z * [new branch] pianpwk/debugmode_compile_tf -> origin/pianpwk/debugmode_compile_tf 2025-10-10T01:18:34.1957696Z * [new branch] pianpwk/debugmode_show_ids -> origin/pianpwk/debugmode_show_ids 2025-10-10T01:18:34.1959721Z * [new branch] pianpwk/dispatch_key_debugging_for_debug -> origin/pianpwk/dispatch_key_debugging_for_debug 2025-10-10T01:18:34.1961376Z * [new branch] pianpwk/draft_debug_mode_tfcompile -> origin/pianpwk/draft_debug_mode_tfcompile 2025-10-10T01:18:34.1962987Z * [new branch] pianpwk/draft_multikernel_nn -> origin/pianpwk/draft_multikernel_nn 2025-10-10T01:18:34.1964796Z * [new branch] pianpwk/draft_multikernel_status_10_5 -> origin/pianpwk/draft_multikernel_status_10_5 2025-10-10T01:18:34.1966452Z * [new branch] pianpwk/dtensor_shape_metadata_guard -> origin/pianpwk/dtensor_shape_metadata_guard 2025-10-10T01:18:34.1968015Z * [new branch] pianpwk/false_numel_refs -> origin/pianpwk/false_numel_refs 2025-10-10T01:18:34.1969652Z * [new branch] pianpwk/maybe_guard_rel -> origin/pianpwk/maybe_guard_rel 2025-10-10T01:18:34.1971309Z * [new branch] pianpwk/multi_kernel_l1 -> origin/pianpwk/multi_kernel_l1 2025-10-10T01:18:34.1973035Z * [new branch] pianpwk/multikernel_hints_draft -> origin/pianpwk/multikernel_hints_draft 2025-10-10T01:18:34.1974862Z * [new branch] pianpwk/no_size_oblivious_slice_scat -> origin/pianpwk/no_size_oblivious_slice_scat 2025-10-10T01:18:34.1976532Z * [new branch] pianpwk/oblivious_reshape_view_better -> origin/pianpwk/oblivious_reshape_view_better 2025-10-10T01:18:34.1978203Z * [new branch] pianpwk/pre_forward_hook -> origin/pianpwk/pre_forward_hook 2025-10-10T01:18:34.1980066Z * [new branch] pianpwk/skip_python_keys_in_guards -> origin/pianpwk/skip_python_keys_in_guards 2025-10-10T01:18:34.1981651Z * [new branch] pianpwk/slice_fresh_symbols -> origin/pianpwk/slice_fresh_symbols 2025-10-10T01:18:34.1983775Z * [new branch] pianpwk/sym_tokens_draft -> origin/pianpwk/sym_tokens_draft 2025-10-10T01:18:34.1985558Z * [new branch] pianpwk/test_pointwise_guard_or_false -> origin/pianpwk/test_pointwise_guard_or_false 2025-10-10T01:18:34.1987244Z * [new branch] pianpwk/totally_draft_sym_wrap -> origin/pianpwk/totally_draft_sym_wrap 2025-10-10T01:18:34.1989109Z * [new branch] pianpwk/triton_benchmark_hints -> origin/pianpwk/triton_benchmark_hints 2025-10-10T01:18:34.1990860Z * [new branch] pianpwk/try_dumb_stuff -> origin/pianpwk/try_dumb_stuff 2025-10-10T01:18:34.1992438Z * [new branch] pianpwk/try_dumb_stuff_2 -> origin/pianpwk/try_dumb_stuff_2 2025-10-10T01:18:34.1994191Z * [new branch] pianpwk/unbacked_channels_last -> origin/pianpwk/unbacked_channels_last 2025-10-10T01:18:34.1995840Z * [new branch] pianpwk/unbacked_should_swap_2 -> origin/pianpwk/unbacked_should_swap_2 2025-10-10T01:18:34.1997581Z * [new branch] pianpwk/user_symints -> origin/pianpwk/user_symints 2025-10-10T01:18:34.1999267Z * [new branch] pianpwk/wan21_reshape -> origin/pianpwk/wan21_reshape 2025-10-10T01:18:34.2001192Z * [new branch] pianpwk/whitelist_optimizer -> origin/pianpwk/whitelist_optimizer 2025-10-10T01:18:34.2003387Z * [new branch] piz/add_wait -> origin/piz/add_wait 2025-10-10T01:18:34.2005080Z * [new branch] piz/fall_back_missing_0716 -> origin/piz/fall_back_missing_0716 2025-10-10T01:18:34.2006723Z * [new branch] pool-separate -> origin/pool-separate 2025-10-10T01:18:34.2008447Z * [new branch] pr-156087 -> origin/pr-156087 2025-10-10T01:18:34.2010887Z * [new branch] pr/131860 -> origin/pr/131860 2025-10-10T01:18:34.2012664Z * [new branch] pre_compile_checks -> origin/pre_compile_checks 2025-10-10T01:18:34.2014471Z * [new branch] predispatch_to -> origin/predispatch_to 2025-10-10T01:18:34.2016525Z * [new branch] prepare-perf-baseline-number-2.8 -> origin/prepare-perf-baseline-number-2.8 2025-10-10T01:18:34.2018604Z * [new branch] prepare-perf-number-2.9 -> origin/prepare-perf-number-2.9 2025-10-10T01:18:34.2020146Z * [new branch] profiler-enabled -> origin/profiler-enabled 2025-10-10T01:18:34.2021807Z * [new branch] provenance_doc_2 -> origin/provenance_doc_2 2025-10-10T01:18:34.2023552Z * [new branch] pt-opt-cuda3 -> origin/pt-opt-cuda3 2025-10-10T01:18:34.2025676Z * [new branch] pyobjectslot -> origin/pyobjectslot 2025-10-10T01:18:34.2027914Z * [new branch] python_compiled_autograd -> origin/python_compiled_autograd 2025-10-10T01:18:34.2031054Z * [new branch] qchip/export-D54134695 -> origin/qchip/export-D54134695 2025-10-10T01:18:34.2032750Z * [new branch] quantile-docs -> origin/quantile-docs 2025-10-10T01:18:34.2034500Z * [new branch] quint-bits -> origin/quint-bits 2025-10-10T01:18:34.2036258Z * [new branch] reland-fx-annotate -> origin/reland-fx-annotate 2025-10-10T01:18:34.2038143Z * [new branch] reland_req_nvsh -> origin/reland_req_nvsh 2025-10-10T01:18:34.2040631Z * [new branch] release/1.10 -> origin/release/1.10 2025-10-10T01:18:34.2042473Z * [new branch] release/1.11 -> origin/release/1.11 2025-10-10T01:18:34.2044204Z * [new branch] release/1.12 -> origin/release/1.12 2025-10-10T01:18:34.2045900Z * [new branch] release/1.13 -> origin/release/1.13 2025-10-10T01:18:34.2047693Z * [new branch] release/1.4 -> origin/release/1.4 2025-10-10T01:18:34.2049054Z * [new branch] release/1.4.1 -> origin/release/1.4.1 2025-10-10T01:18:34.2051060Z * [new branch] release/1.5 -> origin/release/1.5 2025-10-10T01:18:34.2052780Z * [new branch] release/1.6 -> origin/release/1.6 2025-10-10T01:18:34.2054569Z * [new branch] release/1.7 -> origin/release/1.7 2025-10-10T01:18:34.2057113Z * [new branch] release/1.8 -> origin/release/1.8 2025-10-10T01:18:34.2059598Z * [new branch] release/1.9 -> origin/release/1.9 2025-10-10T01:18:34.2060038Z * [new branch] release/2.0 -> origin/release/2.0 2025-10-10T01:18:34.2061919Z * [new branch] release/2.1 -> origin/release/2.1 2025-10-10T01:18:34.2063468Z * [new branch] release/2.2 -> origin/release/2.2 2025-10-10T01:18:34.2065694Z * [new branch] release/2.3 -> origin/release/2.3 2025-10-10T01:18:34.2067799Z * [new branch] release/2.4 -> origin/release/2.4 2025-10-10T01:18:34.2069972Z * [new branch] release/2.5 -> origin/release/2.5 2025-10-10T01:18:34.2071780Z * [new branch] release/2.6 -> origin/release/2.6 2025-10-10T01:18:34.2073627Z * [new branch] release/2.7 -> origin/release/2.7 2025-10-10T01:18:34.2075340Z * [new branch] release/2.8 -> origin/release/2.8 2025-10-10T01:18:34.2077210Z * [new branch] release/2.9 -> origin/release/2.9 2025-10-10T01:18:34.2078976Z * [new branch] release_notes -> origin/release_notes 2025-10-10T01:18:34.2080809Z * [new branch] remove_header_code -> origin/remove_header_code 2025-10-10T01:18:34.2082492Z * [new branch] remove_pyinterpreter -> origin/remove_pyinterpreter 2025-10-10T01:18:34.2084719Z * [new branch] repackage-vllm-nightlies -> origin/repackage-vllm-nightlies 2025-10-10T01:18:34.2087609Z * [new branch] replace-pytorch-labs-20250812-195836 -> origin/replace-pytorch-labs-20250812-195836 2025-10-10T01:18:34.2088817Z * [new branch] replace-pytorch-labs-20250812-200248 -> origin/replace-pytorch-labs-20250812-200248 2025-10-10T01:18:34.2090327Z * [new branch] replace-pytorch-labs-20250812-200324 -> origin/replace-pytorch-labs-20250812-200324 2025-10-10T01:18:34.2092118Z * [new branch] replace-pytorch-labs-20250812-204020 -> origin/replace-pytorch-labs-20250812-204020 2025-10-10T01:18:34.2096702Z * [new branch] revert-131069-gh/krzysztofjordan/1/head -> origin/revert-131069-gh/krzysztofjordan/1/head 2025-10-10T01:18:34.2099313Z * [new branch] revert-131469-gh/andrewor14/51/head -> origin/revert-131469-gh/andrewor14/51/head 2025-10-10T01:18:34.2103054Z * [new branch] revert-156870-gh/skarjala/3/head -> origin/revert-156870-gh/skarjala/3/head 2025-10-10T01:18:34.2105046Z * [new branch] revert-157914-cherry-pick-157503-by-pytorch_bot_bot_ -> origin/revert-157914-cherry-pick-157503-by-pytorch_bot_bot_ 2025-10-10T01:18:34.2107722Z * [new branch] revert-163802-camyll/cherrypick_3016616ccbba3dc9bb6a80eb4a81a846ddf49cc9 -> origin/revert-163802-camyll/cherrypick_3016616ccbba3dc9bb6a80eb4a81a846ddf49cc9 2025-10-10T01:18:34.2109508Z * [new branch] revert_always_build_distributed -> origin/revert_always_build_distributed 2025-10-10T01:18:34.2111095Z * [new branch] rocm-test-yml-update -> origin/rocm-test-yml-update 2025-10-10T01:18:34.2112686Z * [new branch] rocm_op_bench -> origin/rocm_op_bench 2025-10-10T01:18:34.2116082Z * [new branch] ruisi/aot_eager_pass -> origin/ruisi/aot_eager_pass 2025-10-10T01:18:34.2117229Z * [new branch] ruisi/placement_trace -> origin/ruisi/placement_trace 2025-10-10T01:18:34.2120254Z * [new branch] ryanguo99/cleanup-dynamo-expected-failures -> origin/ryanguo99/cleanup-dynamo-expected-failures 2025-10-10T01:18:34.2121498Z * [new branch] ryanguo99/fix-closure-var -> origin/ryanguo99/fix-closure-var 2025-10-10T01:18:34.2123843Z * [new branch] rzou/faketensor_bench -> origin/rzou/faketensor_bench 2025-10-10T01:18:34.2126714Z * [new branch] rzou/njt -> origin/rzou/njt 2025-10-10T01:18:34.2128490Z * [new branch] rzou/pca -> origin/rzou/pca 2025-10-10T01:18:34.2129944Z * [new branch] rzou/realprop -> origin/rzou/realprop 2025-10-10T01:18:34.2131514Z * [new branch] rzou/setup_context -> origin/rzou/setup_context 2025-10-10T01:18:34.2133236Z * [new branch] samplevllm -> origin/samplevllm 2025-10-10T01:18:34.2136116Z * [new branch] sanchitintel/weird_thing_with_test_cpu_select_algorithm -> origin/sanchitintel/weird_thing_with_test_cpu_select_algorithm 2025-10-10T01:18:34.2137904Z * [new branch] sapling-pr-archive-SS-JIA -> origin/sapling-pr-archive-SS-JIA 2025-10-10T01:18:34.2139566Z * [new branch] save -> origin/save 2025-10-10T01:18:34.2141885Z * [new branch] sdym/2.5.1 -> origin/sdym/2.5.1 2025-10-10T01:18:34.2143766Z * [new branch] sekyondaMeta-dynamoconfig-fix -> origin/sekyondaMeta-dynamoconfig-fix 2025-10-10T01:18:34.2145963Z * [new branch] shengf/fx-xform-perf -> origin/shengf/fx-xform-perf 2025-10-10T01:18:34.2147705Z * [new branch] shoumikhin-patch-1 -> origin/shoumikhin-patch-1 2025-10-10T01:18:34.2149428Z * [new branch] shoumikhin-patch-12 -> origin/shoumikhin-patch-12 2025-10-10T01:18:34.2151278Z * [new branch] solve-accuracy-fix -> origin/solve-accuracy-fix 2025-10-10T01:18:34.2153630Z * [new branch] soulitzer/reland-codev-grad-dtype -> origin/soulitzer/reland-codev-grad-dtype 2025-10-10T01:18:34.2155181Z * [new branch] soulitzer/stash-tls-ac -> origin/soulitzer/stash-tls-ac 2025-10-10T01:18:34.2157643Z * [new branch] sqzhang/flight4plus -> origin/sqzhang/flight4plus 2025-10-10T01:18:34.2160174Z * [new branch] sraikund16/test -> origin/sraikund16/test 2025-10-10T01:18:34.2161951Z * [new branch] stablize-compilation-time -> origin/stablize-compilation-time 2025-10-10T01:18:34.2163594Z * [new branch] starterTaskUpdate -> origin/starterTaskUpdate 2025-10-10T01:18:34.2165416Z * [new branch] suo -> origin/suo 2025-10-10T01:18:34.2167275Z * [new branch] support-uv-in-collect_env -> origin/support-uv-in-collect_env 2025-10-10T01:18:34.2168921Z * [new branch] sve-poc -> origin/sve-poc 2025-10-10T01:18:34.2170656Z * [new branch] svekars-patch-1 -> origin/svekars-patch-1 2025-10-10T01:18:34.2172441Z * [new branch] svekars-patch-2 -> origin/svekars-patch-2 2025-10-10T01:18:34.2174258Z * [new branch] svekars-patch-3 -> origin/svekars-patch-3 2025-10-10T01:18:34.2175996Z * [new branch] svekars-patch-4 -> origin/svekars-patch-4 2025-10-10T01:18:34.2177745Z * [new branch] svekars-patch-5 -> origin/svekars-patch-5 2025-10-10T01:18:34.2179488Z * [new branch] switch-bn -> origin/switch-bn 2025-10-10T01:18:34.2181344Z * [new branch] sympy-bottleneck-repro -> origin/sympy-bottleneck-repro 2025-10-10T01:18:34.2183605Z * [new branch] tenpercent/ck_rocm_ci_v3 -> origin/tenpercent/ck_rocm_ci_v3 2025-10-10T01:18:34.2185381Z * [new branch] tensordict_integration -> origin/tensordict_integration 2025-10-10T01:18:34.2187120Z * [new branch] test-move-conda-builds -> origin/test-move-conda-builds 2025-10-10T01:18:34.2188998Z * [new branch] test-myst-markdown-docstring -> origin/test-myst-markdown-docstring 2025-10-10T01:18:34.2190620Z * [new branch] test-old -> origin/test-old 2025-10-10T01:18:34.2192548Z * [new branch] test-vec-migration-internally -> origin/test-vec-migration-internally 2025-10-10T01:18:34.2194843Z * [new branch] test/bmm_heur -> origin/test/bmm_heur 2025-10-10T01:18:34.2196742Z * [new branch] test/inductor -> origin/test/inductor 2025-10-10T01:18:34.2198875Z * [new branch] test_quantization -> origin/test_quantization 2025-10-10T01:18:34.2201098Z * [new branch] tianren/customOp_autotune -> origin/tianren/customOp_autotune 2025-10-10T01:18:34.2202881Z * [new branch] tianren/customOp_autotune_fix -> origin/tianren/customOp_autotune_fix 2025-10-10T01:18:34.2204304Z * [new branch] tianren/customOp_fusion -> origin/tianren/customOp_fusion 2025-10-10T01:18:34.2206100Z * [new branch] tianren/flex_paged_attn_fix_temp -> origin/tianren/flex_paged_attn_fix_temp 2025-10-10T01:18:34.2207409Z * [new branch] tianren/remove_repeate -> origin/tianren/remove_repeate 2025-10-10T01:18:34.2209351Z * [new branch] tianren/test -> origin/tianren/test 2025-10-10T01:18:34.2211034Z * [new branch] tidy_performance_cyy -> origin/tidy_performance_cyy 2025-10-10T01:18:34.2212689Z * [new branch] torchtitan_ep -> origin/torchtitan_ep 2025-10-10T01:18:34.2214529Z * [new branch] trace_fsdp_torchtune_lora -> origin/trace_fsdp_torchtune_lora 2025-10-10T01:18:34.2216390Z * [new branch] traceable_fsdp_unit_tests -> origin/traceable_fsdp_unit_tests 2025-10-10T01:18:34.2218611Z * [new branch] transpose_pack_fusion -> origin/transpose_pack_fusion 2025-10-10T01:18:34.2220153Z * [new branch] tree_loop_vec_base -> origin/tree_loop_vec_base 2025-10-10T01:18:34.2221823Z * [new branch] triton_kernel -> origin/triton_kernel 2025-10-10T01:18:34.2223649Z * [new branch] trunk-tagging-multi-commits -> origin/trunk-tagging-multi-commits 2025-10-10T01:18:34.2225621Z * [new branch] tt_pkg_1908 -> origin/tt_pkg_1908 2025-10-10T01:18:34.2227402Z * [new branch] type_dec -> origin/type_dec 2025-10-10T01:18:34.2229807Z * [new branch] udate-sphinx-dependancies -> origin/udate-sphinx-dependancies 2025-10-10T01:18:34.2231640Z * [new branch] unlift -> origin/unlift 2025-10-10T01:18:34.2234102Z * [new branch] update-audio-commit-hash/17567864209-1799-1 -> origin/update-audio-commit-hash/17567864209-1799-1 2025-10-10T01:18:34.2236109Z * [new branch] update-audio-commit-hash/17599208654-1801-1 -> origin/update-audio-commit-hash/17599208654-1801-1 2025-10-10T01:18:34.2237626Z * [new branch] update-audio-commit-hash/17630256502-1803-1 -> origin/update-audio-commit-hash/17630256502-1803-1 2025-10-10T01:18:34.2239335Z * [new branch] update-audio-commit-hash/17657093113-1804-1 -> origin/update-audio-commit-hash/17657093113-1804-1 2025-10-10T01:18:34.2240942Z * [new branch] update-audio-commit-hash/17688961747-1806-1 -> origin/update-audio-commit-hash/17688961747-1806-1 2025-10-10T01:18:34.2242793Z * [new branch] update-audio-commit-hash/17703952853-1807-1 -> origin/update-audio-commit-hash/17703952853-1807-1 2025-10-10T01:18:34.2244865Z * [new branch] update-audio-commit-hash/18392707270-1874-1 -> origin/update-audio-commit-hash/18392707270-1874-1 2025-10-10T01:18:34.2246760Z * [new branch] update-dynamic-shapes-doc -> origin/update-dynamic-shapes-doc 2025-10-10T01:18:34.2249265Z * [new branch] update-executorch-commit-hash/15694981040-1626-1 -> origin/update-executorch-commit-hash/15694981040-1626-1 2025-10-10T01:18:34.2251364Z * [new branch] update-triton-commit-hash/13663274526-1487-2 -> origin/update-triton-commit-hash/13663274526-1487-2 2025-10-10T01:18:34.2253757Z * [new branch] update-vision-commit-hash/15336342773-1607-1 -> origin/update-vision-commit-hash/15336342773-1607-1 2025-10-10T01:18:34.2254987Z * [new branch] update-vision-commit-hash/18361653903-1869-1 -> origin/update-vision-commit-hash/18361653903-1869-1 2025-10-10T01:18:34.2257273Z * [new branch] update-vllm-commit-hash/17536029887-1798-1 -> origin/update-vllm-commit-hash/17536029887-1798-1 2025-10-10T01:18:34.2258910Z * [new branch] update-vllm-commit-hash/17599208654-1801-1 -> origin/update-vllm-commit-hash/17599208654-1801-1 2025-10-10T01:18:34.2260491Z * [new branch] update-vllm-commit-hash/17657093113-1804-1 -> origin/update-vllm-commit-hash/17657093113-1804-1 2025-10-10T01:18:34.2262143Z * [new branch] update-vllm-commit-hash/17703952853-1807-1 -> origin/update-vllm-commit-hash/17703952853-1807-1 2025-10-10T01:18:34.2263764Z * [new branch] update-vllm-commit-hash/17718740812-1808-1 -> origin/update-vllm-commit-hash/17718740812-1808-1 2025-10-10T01:18:34.2265810Z * [new branch] update-vllm-commit-hash/17782703922-1813-1 -> origin/update-vllm-commit-hash/17782703922-1813-1 2025-10-10T01:18:34.2267846Z * [new branch] update-vllm-commit-hash/17814169036-1822-1 -> origin/update-vllm-commit-hash/17814169036-1822-1 2025-10-10T01:18:34.2270017Z * [new branch] update-vllm-commit-hash/17844794719-1823-1 -> origin/update-vllm-commit-hash/17844794719-1823-1 2025-10-10T01:18:34.2272130Z * [new branch] update-vllm-commit-hash/17872674059-1830-1 -> origin/update-vllm-commit-hash/17872674059-1830-1 2025-10-10T01:18:34.2273932Z * [new branch] update-vllm-commit-hash/17901034819-1833-1 -> origin/update-vllm-commit-hash/17901034819-1833-1 2025-10-10T01:18:34.2275673Z * [new branch] update-vllm-commit-hash/17932176396-1836-1 -> origin/update-vllm-commit-hash/17932176396-1836-1 2025-10-10T01:18:34.2277255Z * [new branch] update-vllm-commit-hash/17962545886-1842-1 -> origin/update-vllm-commit-hash/17962545886-1842-1 2025-10-10T01:18:34.2279011Z * [new branch] update-vllm-commit-hash/17993166855-1844-1 -> origin/update-vllm-commit-hash/17993166855-1844-1 2025-10-10T01:18:34.2280805Z * [new branch] update-vllm-commit-hash/18052321282-1848-1 -> origin/update-vllm-commit-hash/18052321282-1848-1 2025-10-10T01:18:34.2282493Z * [new branch] update-vllm-commit-hash/18066820738-1849-1 -> origin/update-vllm-commit-hash/18066820738-1849-1 2025-10-10T01:18:34.2284173Z * [new branch] update-vllm-commit-hash/18081987460-1850-1 -> origin/update-vllm-commit-hash/18081987460-1850-1 2025-10-10T01:18:34.2285923Z * [new branch] update-vllm-commit-hash/18114584510-1852-1 -> origin/update-vllm-commit-hash/18114584510-1852-1 2025-10-10T01:18:34.2287714Z * [new branch] update-vllm-commit-hash/18147226974-1853-1 -> origin/update-vllm-commit-hash/18147226974-1853-1 2025-10-10T01:18:34.2289445Z * [new branch] update-vllm-commit-hash/18236802781-1857-1 -> origin/update-vllm-commit-hash/18236802781-1857-1 2025-10-10T01:18:34.2291665Z * [new branch] update-xla-commit-hash/17725712604-203-1 -> origin/update-xla-commit-hash/17725712604-203-1 2025-10-10T01:18:34.2293263Z * [new branch] update-xla-commit-hash/17908176340-204-1 -> origin/update-xla-commit-hash/17908176340-204-1 2025-10-10T01:18:34.2294833Z * [new branch] update-xla-commit-hash/18273597034-206-1 -> origin/update-xla-commit-hash/18273597034-206-1 2025-10-10T01:18:34.2296731Z * [new branch] update_docs_torch_multinomial_issue#125388 -> origin/update_docs_torch_multinomial_issue#125388 2025-10-10T01:18:34.2298408Z * [new branch] update_executorch_pin -> origin/update_executorch_pin 2025-10-10T01:18:34.2300229Z * [new branch] update_slow_tests_1722488736 -> origin/update_slow_tests_1722488736 2025-10-10T01:18:34.2301916Z * [new branch] update_slow_tests_1722879173 -> origin/update_slow_tests_1722879173 2025-10-10T01:18:34.2303797Z * [new branch] update_slow_tests_1757922057 -> origin/update_slow_tests_1757922057 2025-10-10T01:18:34.2305400Z * [new branch] update_slow_tests_1758526845 -> origin/update_slow_tests_1758526845 2025-10-10T01:18:34.2307008Z * [new branch] update_slow_tests_1759736444 -> origin/update_slow_tests_1759736444 2025-10-10T01:18:34.2308815Z * [new branch] update_submodule_FBGEMM -> origin/update_submodule_FBGEMM 2025-10-10T01:18:34.2310503Z * [new branch] update_submodule_kineto -> origin/update_submodule_kineto 2025-10-10T01:18:34.2312201Z * [new branch] update_submodule_tensorpipe -> origin/update_submodule_tensorpipe 2025-10-10T01:18:34.2314123Z * [new branch] v0.1.2 -> origin/v0.1.2 2025-10-10T01:18:34.2316115Z * [new branch] v1.0.1 -> origin/v1.0.1 2025-10-10T01:18:34.2317775Z * [new branch] v1.0.3 -> origin/v1.0.3 2025-10-10T01:18:34.2320167Z * [new branch] v1.1.0 -> origin/v1.1.0 2025-10-10T01:18:34.2321953Z * [new branch] v1.2.0 -> origin/v1.2.0 2025-10-10T01:18:34.2323850Z * [new branch] v1.3.0 -> origin/v1.3.0 2025-10-10T01:18:34.2326024Z * [new branch] v1.3.1 -> origin/v1.3.1 2025-10-10T01:18:34.2327823Z * [new branch] validate_fn -> origin/validate_fn 2025-10-10T01:18:34.2329811Z * [new branch] validations_2.6 -> origin/validations_2.6 2025-10-10T01:18:34.2331666Z * [new branch] validations_2.8 -> origin/validations_2.8 2025-10-10T01:18:34.2333410Z * [new branch] varlen-api -> origin/varlen-api 2025-10-10T01:18:34.2335014Z * [new branch] varlen_api -> origin/varlen_api 2025-10-10T01:18:34.2337436Z * [new branch] viable/strict -> origin/viable/strict 2025-10-10T01:18:34.2339855Z * [new branch] vishal9-team/dtensor_parallelism_toy -> origin/vishal9-team/dtensor_parallelism_toy 2025-10-10T01:18:34.2341514Z * [new branch] vllmbuildci -> origin/vllmbuildci 2025-10-10T01:18:34.2343276Z * [new branch] vllmpin -> origin/vllmpin 2025-10-10T01:18:34.2345584Z * [new branch] wdvr/iss_145259 -> origin/wdvr/iss_145259 2025-10-10T01:18:34.2347938Z * [new branch] whc/flight51 -> origin/whc/flight51 2025-10-10T01:18:34.2349603Z * [new branch] whc/flight53 -> origin/whc/flight53 2025-10-10T01:18:34.2351282Z * [new branch] whc/stage2 -> origin/whc/stage2 2025-10-10T01:18:34.2352823Z * [new branch] whc/uneven -> origin/whc/uneven 2025-10-10T01:18:34.2354782Z * [new branch] whc/uneven-merge -> origin/whc/uneven-merge 2025-10-10T01:18:34.2356835Z * [new branch] williamwen42-patch-1 -> origin/williamwen42-patch-1 2025-10-10T01:18:34.2359020Z * [new branch] win_warnings -> origin/win_warnings 2025-10-10T01:18:34.2360788Z * [new branch] windows_libtorch_free -> origin/windows_libtorch_free 2025-10-10T01:18:34.2362603Z * [new branch] windows_mmap -> origin/windows_mmap 2025-10-10T01:18:34.2364345Z * [new branch] xmfan-war -> origin/xmfan-war 2025-10-10T01:18:34.2367342Z * [new branch] xmfan/ca_0516 -> origin/xmfan/ca_0516 2025-10-10T01:18:34.2368609Z * [new branch] xmfan/ca_1051b93192 -> origin/xmfan/ca_1051b93192 2025-10-10T01:18:34.2370270Z * [new branch] xmfan/ca_1a722f62c248391fc4a542e8851a5559aa356ae8 -> origin/xmfan/ca_1a722f62c248391fc4a542e8851a5559aa356ae8 2025-10-10T01:18:34.2371928Z * [new branch] xmfan/ca_5a2be192d1 -> origin/xmfan/ca_5a2be192d1 2025-10-10T01:18:34.2373796Z * [new branch] xmfan/ca_9d59b516e9 -> origin/xmfan/ca_9d59b516e9 2025-10-10T01:18:34.2375381Z * [new branch] xmfan/ca_api -> origin/xmfan/ca_api 2025-10-10T01:18:34.2377121Z * [new branch] xmfan/ca_apr8 -> origin/xmfan/ca_apr8 2025-10-10T01:18:34.2378929Z * [new branch] xmfan/ca_base -> origin/xmfan/ca_base 2025-10-10T01:18:34.2381099Z * [new branch] xmfan/ca_cudagraphs -> origin/xmfan/ca_cudagraphs 2025-10-10T01:18:34.2382745Z * [new branch] xmfan/ca_dynamic -> origin/xmfan/ca_dynamic 2025-10-10T01:18:34.2384498Z * [new branch] xmfan/ca_fix_dyn -> origin/xmfan/ca_fix_dyn 2025-10-10T01:18:34.2386383Z * [new branch] xmfan/ca_fix_lowering -> origin/xmfan/ca_fix_lowering 2025-10-10T01:18:34.2388469Z * [new branch] xmfan/ca_fix_polyfills -> origin/xmfan/ca_fix_polyfills 2025-10-10T01:18:34.2389828Z * [new branch] xmfan/ca_jan3 -> origin/xmfan/ca_jan3 2025-10-10T01:18:34.2391746Z * [new branch] xmfan/ca_jun18 -> origin/xmfan/ca_jun18 2025-10-10T01:18:34.2393327Z * [new branch] xmfan/ca_jun24 -> origin/xmfan/ca_jun24 2025-10-10T01:18:34.2395053Z * [new branch] xmfan/ca_mem_base -> origin/xmfan/ca_mem_base 2025-10-10T01:18:34.2397054Z * [new branch] xmfan/ca_mem_fix -> origin/xmfan/ca_mem_fix 2025-10-10T01:18:34.2398916Z * [new branch] xmfan/ca_move_to_cuda -> origin/xmfan/ca_move_to_cuda 2025-10-10T01:18:34.2400573Z * [new branch] xmfan/ca_nested -> origin/xmfan/ca_nested 2025-10-10T01:18:34.2402245Z * [new branch] xmfan/ca_overhead -> origin/xmfan/ca_overhead 2025-10-10T01:18:34.2404016Z * [new branch] xmfan/ca_overhead_0eba7e5451 -> origin/xmfan/ca_overhead_0eba7e5451 2025-10-10T01:18:34.2405748Z * [new branch] xmfan/ca_scalar -> origin/xmfan/ca_scalar 2025-10-10T01:18:34.2407541Z * [new branch] xmfan/ca_subclass_mem_fix -> origin/xmfan/ca_subclass_mem_fix 2025-10-10T01:18:34.2409208Z * [new branch] xmfan/ca_warm_mem -> origin/xmfan/ca_warm_mem 2025-10-10T01:18:34.2410870Z * [new branch] xmfan/ca_warm_mem_base -> origin/xmfan/ca_warm_mem_base 2025-10-10T01:18:34.2412569Z * [new branch] xmfan/cacu_jun18 -> origin/xmfan/cacu_jun18 2025-10-10T01:18:34.2414208Z * [new branch] xmfan/cacu_jun19 -> origin/xmfan/cacu_jun19 2025-10-10T01:18:34.2415900Z * [new branch] xmfan/cacu_jun4 -> origin/xmfan/cacu_jun4 2025-10-10T01:18:34.2417586Z * [new branch] xmfan/cacu_may27 -> origin/xmfan/cacu_may27 2025-10-10T01:18:34.2419335Z * [new branch] xmfan/disable_duck_shape -> origin/xmfan/disable_duck_shape 2025-10-10T01:18:34.2421069Z * [new branch] xmfan/fca_cpp_node_passthrough -> origin/xmfan/fca_cpp_node_passthrough 2025-10-10T01:18:34.2422806Z * [new branch] xmfan/issue_123374 -> origin/xmfan/issue_123374 2025-10-10T01:18:34.2425589Z * [new branch] xmfan/post_3945954741e2d37023c5d6954f9483008e0892f9 -> origin/xmfan/post_3945954741e2d37023c5d6954f9483008e0892f9 2025-10-10T01:18:34.2428511Z * [new branch] xmfan/pre_3945954741e2d37023c5d6954f9483008e0892f9 -> origin/xmfan/pre_3945954741e2d37023c5d6954f9483008e0892f9 2025-10-10T01:18:34.2429285Z * [new branch] xmfan/single_step -> origin/xmfan/single_step 2025-10-10T01:18:34.2430943Z * [new branch] xmfan/sth_0829 -> origin/xmfan/sth_0829 2025-10-10T01:18:34.2432730Z * [new branch] xmfan/test -> origin/xmfan/test 2025-10-10T01:18:34.2436174Z * [new branch] yguo/debug-0226-constexpr -> origin/yguo/debug-0226-constexpr 2025-10-10T01:18:34.2437092Z * [new branch] yguo/new_latest_changes -> origin/yguo/new_latest_changes 2025-10-10T01:18:34.2438696Z * [new branch] yguo/patch_constexpr_changes -> origin/yguo/patch_constexpr_changes 2025-10-10T01:18:34.2440607Z * [new branch] yihan_quantization -> origin/yihan_quantization 2025-10-10T01:18:34.2442780Z * [new branch] yiming/bootcamp -> origin/yiming/bootcamp 2025-10-10T01:18:34.2444455Z * [new branch] yiming/improve_sharding_error_msg -> origin/yiming/improve_sharding_error_msg 2025-10-10T01:18:34.2445947Z * [new branch] yiming/precompile_benchmark -> origin/yiming/precompile_benchmark 2025-10-10T01:18:34.2447914Z * [new branch] yolo-llama3 -> origin/yolo-llama3 2025-10-10T01:18:34.2450720Z * [new branch] ysiraichi/install-fmtlib-headers-v12 -> origin/ysiraichi/install-fmtlib-headers-v12 2025-10-10T01:18:34.2452361Z * [new branch] zainr/canary-test -> origin/zainr/canary-test 2025-10-10T01:18:34.2454394Z * [new branch] zainr/cleanup-gh-runners -> origin/zainr/cleanup-gh-runners 2025-10-10T01:18:34.2455878Z * [new branch] zainr/pull-migration-c -> origin/zainr/pull-migration-c 2025-10-10T01:18:34.2457147Z * [new branch] zainr/test2 -> origin/zainr/test2 2025-10-10T01:18:34.2458985Z * [new branch] zainr/unstable -> origin/zainr/unstable 2025-10-10T01:18:34.2460811Z * [new branch] zasdfgbnm-patch-3 -> origin/zasdfgbnm-patch-3 2025-10-10T01:18:34.2462513Z * [new branch] zb2p -> origin/zb2p 2025-10-10T01:18:34.2464324Z * [new branch] zeros-and-scatter-part2 -> origin/zeros-and-scatter-part2 2025-10-10T01:18:34.2466768Z * [new branch] zhxchen17/aot_compile_fix_load_guard_manager -> origin/zhxchen17/aot_compile_fix_load_guard_manager 2025-10-10T01:18:34.2469001Z * [new branch] zhxchen17/precompile/source_info -> origin/zhxchen17/precompile/source_info 2025-10-10T01:18:34.2471172Z * [new branch] zhxchen17/scratch/0 -> origin/zhxchen17/scratch/0 2025-10-10T01:18:34.2473551Z * [new branch] zhxhcen17/moodycamel -> origin/zhxhcen17/moodycamel 2025-10-10T01:18:34.2475825Z * [new branch] zxiiro/build-times -> origin/zxiiro/build-times 2025-10-10T01:18:34.2477461Z * [new branch] zxiiro/c7i-docs -> origin/zxiiro/c7i-docs 2025-10-10T01:18:34.2479325Z * [new branch] zxiiro/c7i-linux-4xlarge -> origin/zxiiro/c7i-linux-4xlarge 2025-10-10T01:18:34.2480982Z * [new branch] zxiiro/c7i-linux-build-yaml -> origin/zxiiro/c7i-linux-build-yaml 2025-10-10T01:18:34.2482609Z * [new branch] zxiiro/main -> origin/zxiiro/main 2025-10-10T01:18:34.2484255Z * [new branch] zxiiro/test-multicloud-arc -> origin/zxiiro/test-multicloud-arc 2025-10-10T01:18:34.2485890Z * [new tag] bc2caa7fdf006894eff7af936babde69ab5a40f8-huydhn-debug -> bc2caa7fdf006894eff7af936babde69ab5a40f8-huydhn-debug 2025-10-10T01:18:34.2488096Z * [new tag] ci/binaries/77164 -> ci/binaries/77164 2025-10-10T01:18:34.2489405Z * [new tag] ciflow/b200-symm-mem/163767 -> ciflow/b200-symm-mem/163767 2025-10-10T01:18:34.2490629Z * [new tag] ciflow/b200/163955 -> ciflow/b200/163955 2025-10-10T01:18:34.2492033Z * [new tag] ciflow/binaries/157432 -> ciflow/binaries/157432 2025-10-10T01:18:34.2493086Z * [new tag] ciflow/binaries/158104 -> ciflow/binaries/158104 2025-10-10T01:18:34.2494470Z * [new tag] ciflow/binaries/164769 -> ciflow/binaries/164769 2025-10-10T01:18:34.2496072Z * [new tag] ciflow/binaries/164894 -> ciflow/binaries/164894 2025-10-10T01:18:34.2497189Z * [new tag] ciflow/binaries_libtorch/157432 -> ciflow/binaries_libtorch/157432 2025-10-10T01:18:34.2498715Z * [new tag] ciflow/binaries_wheel/157432 -> ciflow/binaries_wheel/157432 2025-10-10T01:18:34.2499956Z * [new tag] ciflow/binaries_wheel/159104 -> ciflow/binaries_wheel/159104 2025-10-10T01:18:34.2501319Z * [new tag] ciflow/binaries_wheel/164935 -> ciflow/binaries_wheel/164935 2025-10-10T01:18:34.2502676Z * [new tag] ciflow/h100-cutlass-backend/163767 -> ciflow/h100-cutlass-backend/163767 2025-10-10T01:18:34.2503589Z * [new tag] ciflow/h100-cutlass-backend/164747 -> ciflow/h100-cutlass-backend/164747 2025-10-10T01:18:34.2505093Z * [new tag] ciflow/h100-distributed/163767 -> ciflow/h100-distributed/163767 2025-10-10T01:18:34.2506636Z * [new tag] ciflow/h100-symm-mem/151845 -> ciflow/h100-symm-mem/151845 2025-10-10T01:18:34.2507804Z * [new tag] ciflow/h100-symm-mem/157635 -> ciflow/h100-symm-mem/157635 2025-10-10T01:18:34.2508937Z * [new tag] ciflow/h100-symm-mem/163767 -> ciflow/h100-symm-mem/163767 2025-10-10T01:18:34.2509892Z * [new tag] ciflow/h100-symm-mem/164747 -> ciflow/h100-symm-mem/164747 2025-10-10T01:18:34.2511207Z * [new tag] ciflow/h100-symm-mem/164965 -> ciflow/h100-symm-mem/164965 2025-10-10T01:18:34.2512127Z * [new tag] ciflow/h100-symm-mem/165101 -> ciflow/h100-symm-mem/165101 2025-10-10T01:18:34.2513655Z * [new tag] ciflow/h100/163955 -> ciflow/h100/163955 2025-10-10T01:18:34.2514798Z * [new tag] ciflow/h100/164474 -> ciflow/h100/164474 2025-10-10T01:18:34.2516076Z * [new tag] ciflow/h100/164705 -> ciflow/h100/164705 2025-10-10T01:18:34.2517115Z * [new tag] ciflow/h100/164790 -> ciflow/h100/164790 2025-10-10T01:18:34.2518271Z * [new tag] ciflow/h100/164930 -> ciflow/h100/164930 2025-10-10T01:18:34.2519563Z * [new tag] ciflow/h100/165055 -> ciflow/h100/165055 2025-10-10T01:18:34.2521298Z * [new tag] ciflow/inductor-micro-benchmark/164747 -> ciflow/inductor-micro-benchmark/164747 2025-10-10T01:18:34.2522667Z * [new tag] ciflow/inductor-perf-compare/163767 -> ciflow/inductor-perf-compare/163767 2025-10-10T01:18:34.2523645Z * [new tag] ciflow/inductor-perf-compare/164747 -> ciflow/inductor-perf-compare/164747 2025-10-10T01:18:34.2526429Z * [new tag] ciflow/inductor-perf-test-nightly-rocm/151845 -> ciflow/inductor-perf-test-nightly-rocm/151845 2025-10-10T01:18:34.2527454Z * [new tag] ciflow/inductor-perf-test-nightly-rocm/164747 -> ciflow/inductor-perf-test-nightly-rocm/164747 2025-10-10T01:18:34.2529065Z * [new tag] ciflow/inductor-perf-test-nightly-x86-zen/161512 -> ciflow/inductor-perf-test-nightly-x86-zen/161512 2025-10-10T01:18:34.2530007Z * [new tag] ciflow/inductor-perf-test-nightly-x86-zen/162954 -> ciflow/inductor-perf-test-nightly-x86-zen/162954 2025-10-10T01:18:34.2531380Z * [new tag] ciflow/inductor-perf-test-nightly-x86-zen/163767 -> ciflow/inductor-perf-test-nightly-x86-zen/163767 2025-10-10T01:18:34.2532311Z * [new tag] ciflow/inductor-perf-test-nightly-x86-zen/164126 -> ciflow/inductor-perf-test-nightly-x86-zen/164126 2025-10-10T01:18:34.2533625Z * [new tag] ciflow/inductor-perf-test-nightly-x86-zen/164747 -> ciflow/inductor-perf-test-nightly-x86-zen/164747 2025-10-10T01:18:34.2535259Z * [new tag] ciflow/inductor-periodic/0d39ecb2ce8556e85343d8da0c87450192c2fdf8 -> ciflow/inductor-periodic/0d39ecb2ce8556e85343d8da0c87450192c2fdf8 2025-10-10T01:18:34.2536656Z * [new tag] ciflow/inductor-periodic/156592 -> ciflow/inductor-periodic/156592 2025-10-10T01:18:34.2537809Z * [new tag] ciflow/inductor-periodic/164492 -> ciflow/inductor-periodic/164492 2025-10-10T01:18:34.2539488Z * [new tag] ciflow/inductor-periodic/73adac05d13babb75410c3e033fdce57aa16881a -> ciflow/inductor-periodic/73adac05d13babb75410c3e033fdce57aa16881a 2025-10-10T01:18:34.2540565Z * [new tag] ciflow/inductor-rocm/151845 -> ciflow/inductor-rocm/151845 2025-10-10T01:18:34.2541890Z * [new tag] ciflow/inductor-rocm/161280 -> ciflow/inductor-rocm/161280 2025-10-10T01:18:34.2543220Z * [new tag] ciflow/inductor-rocm/162478 -> ciflow/inductor-rocm/162478 2025-10-10T01:18:34.2544359Z * [new tag] ciflow/inductor-rocm/163767 -> ciflow/inductor-rocm/163767 2025-10-10T01:18:34.2545686Z * [new tag] ciflow/inductor-rocm/164618 -> ciflow/inductor-rocm/164618 2025-10-10T01:18:34.2546887Z * [new tag] ciflow/inductor-rocm/164747 -> ciflow/inductor-rocm/164747 2025-10-10T01:18:34.2548017Z * [new tag] ciflow/inductor-rocm/164769 -> ciflow/inductor-rocm/164769 2025-10-10T01:18:34.2549249Z * [new tag] ciflow/inductor-rocm/165080 -> ciflow/inductor-rocm/165080 2025-10-10T01:18:34.2550710Z * [new tag] ciflow/inductor-rocm/165115 -> ciflow/inductor-rocm/165115 2025-10-10T01:18:34.2551975Z * [new tag] ciflow/inductor/137400 -> ciflow/inductor/137400 2025-10-10T01:18:34.2553129Z * [new tag] ciflow/inductor/148180 -> ciflow/inductor/148180 2025-10-10T01:18:34.2554259Z * [new tag] ciflow/inductor/148328 -> ciflow/inductor/148328 2025-10-10T01:18:34.2555392Z * [new tag] ciflow/inductor/148484 -> ciflow/inductor/148484 2025-10-10T01:18:34.2556505Z * [new tag] ciflow/inductor/148492 -> ciflow/inductor/148492 2025-10-10T01:18:34.2557635Z * [new tag] ciflow/inductor/149003 -> ciflow/inductor/149003 2025-10-10T01:18:34.2558837Z * [new tag] ciflow/inductor/151845 -> ciflow/inductor/151845 2025-10-10T01:18:34.2560002Z * [new tag] ciflow/inductor/152624 -> ciflow/inductor/152624 2025-10-10T01:18:34.2561155Z * [new tag] ciflow/inductor/156592 -> ciflow/inductor/156592 2025-10-10T01:18:34.2562275Z * [new tag] ciflow/inductor/157635 -> ciflow/inductor/157635 2025-10-10T01:18:34.2563526Z * [new tag] ciflow/inductor/157743 -> ciflow/inductor/157743 2025-10-10T01:18:34.2565203Z * [new tag] ciflow/inductor/157994 -> ciflow/inductor/157994 2025-10-10T01:18:34.2566825Z * [new tag] ciflow/inductor/158104 -> ciflow/inductor/158104 2025-10-10T01:18:34.2568257Z * [new tag] ciflow/inductor/158872 -> ciflow/inductor/158872 2025-10-10T01:18:34.2569691Z * [new tag] ciflow/inductor/158932 -> ciflow/inductor/158932 2025-10-10T01:18:34.2570904Z * [new tag] ciflow/inductor/159523 -> ciflow/inductor/159523 2025-10-10T01:18:34.2572161Z * [new tag] ciflow/inductor/160266 -> ciflow/inductor/160266 2025-10-10T01:18:34.2573617Z * [new tag] ciflow/inductor/160324 -> ciflow/inductor/160324 2025-10-10T01:18:34.2574954Z * [new tag] ciflow/inductor/160325 -> ciflow/inductor/160325 2025-10-10T01:18:34.2576376Z * [new tag] ciflow/inductor/160326 -> ciflow/inductor/160326 2025-10-10T01:18:34.2577669Z * [new tag] ciflow/inductor/160327 -> ciflow/inductor/160327 2025-10-10T01:18:34.2579096Z * [new tag] ciflow/inductor/160328 -> ciflow/inductor/160328 2025-10-10T01:18:34.2580479Z * [new tag] ciflow/inductor/160329 -> ciflow/inductor/160329 2025-10-10T01:18:34.2582202Z * [new tag] ciflow/inductor/160539 -> ciflow/inductor/160539 2025-10-10T01:18:34.2583548Z * [new tag] ciflow/inductor/160611 -> ciflow/inductor/160611 2025-10-10T01:18:34.2584694Z * [new tag] ciflow/inductor/160843 -> ciflow/inductor/160843 2025-10-10T01:18:34.2585887Z * [new tag] ciflow/inductor/160903 -> ciflow/inductor/160903 2025-10-10T01:18:34.2587189Z * [new tag] ciflow/inductor/161118 -> ciflow/inductor/161118 2025-10-10T01:18:34.2588362Z * [new tag] ciflow/inductor/161158 -> ciflow/inductor/161158 2025-10-10T01:18:34.2589544Z * [new tag] ciflow/inductor/161280 -> ciflow/inductor/161280 2025-10-10T01:18:34.2590765Z * [new tag] ciflow/inductor/161320 -> ciflow/inductor/161320 2025-10-10T01:18:34.2592308Z * [new tag] ciflow/inductor/161485 -> ciflow/inductor/161485 2025-10-10T01:18:34.2593496Z * [new tag] ciflow/inductor/161495 -> ciflow/inductor/161495 2025-10-10T01:18:34.2594674Z * [new tag] ciflow/inductor/161512 -> ciflow/inductor/161512 2025-10-10T01:18:34.2595857Z * [new tag] ciflow/inductor/162031 -> ciflow/inductor/162031 2025-10-10T01:18:34.2597032Z * [new tag] ciflow/inductor/162066 -> ciflow/inductor/162066 2025-10-10T01:18:34.2598235Z * [new tag] ciflow/inductor/162294 -> ciflow/inductor/162294 2025-10-10T01:18:34.2599569Z * [new tag] ciflow/inductor/162340 -> ciflow/inductor/162340 2025-10-10T01:18:34.2600756Z * [new tag] ciflow/inductor/162470 -> ciflow/inductor/162470 2025-10-10T01:18:34.2602153Z * [new tag] ciflow/inductor/162523 -> ciflow/inductor/162523 2025-10-10T01:18:34.2603350Z * [new tag] ciflow/inductor/162542 -> ciflow/inductor/162542 2025-10-10T01:18:34.2604544Z * [new tag] ciflow/inductor/162768 -> ciflow/inductor/162768 2025-10-10T01:18:34.2605756Z * [new tag] ciflow/inductor/162899 -> ciflow/inductor/162899 2025-10-10T01:18:34.2607053Z * [new tag] ciflow/inductor/162900 -> ciflow/inductor/162900 2025-10-10T01:18:34.2608214Z * [new tag] ciflow/inductor/162901 -> ciflow/inductor/162901 2025-10-10T01:18:34.2609410Z * [new tag] ciflow/inductor/162903 -> ciflow/inductor/162903 2025-10-10T01:18:34.2610615Z * [new tag] ciflow/inductor/162905 -> ciflow/inductor/162905 2025-10-10T01:18:34.2611857Z * [new tag] ciflow/inductor/162954 -> ciflow/inductor/162954 2025-10-10T01:18:34.2613030Z * [new tag] ciflow/inductor/162990 -> ciflow/inductor/162990 2025-10-10T01:18:34.2614232Z * [new tag] ciflow/inductor/163027 -> ciflow/inductor/163027 2025-10-10T01:18:34.2615422Z * [new tag] ciflow/inductor/163028 -> ciflow/inductor/163028 2025-10-10T01:18:34.2616600Z * [new tag] ciflow/inductor/163053 -> ciflow/inductor/163053 2025-10-10T01:18:34.2617813Z * [new tag] ciflow/inductor/163185 -> ciflow/inductor/163185 2025-10-10T01:18:34.2619000Z * [new tag] ciflow/inductor/163335 -> ciflow/inductor/163335 2025-10-10T01:18:34.2620207Z * [new tag] ciflow/inductor/163490 -> ciflow/inductor/163490 2025-10-10T01:18:34.2621424Z * [new tag] ciflow/inductor/163503 -> ciflow/inductor/163503 2025-10-10T01:18:34.2622956Z * [new tag] ciflow/inductor/163517 -> ciflow/inductor/163517 2025-10-10T01:18:34.2624307Z * [new tag] ciflow/inductor/163527 -> ciflow/inductor/163527 2025-10-10T01:18:34.2626181Z * [new tag] ciflow/inductor/163533 -> ciflow/inductor/163533 2025-10-10T01:18:34.2627357Z * [new tag] ciflow/inductor/163602 -> ciflow/inductor/163602 2025-10-10T01:18:34.2628714Z * [new tag] ciflow/inductor/163617 -> ciflow/inductor/163617 2025-10-10T01:18:34.2629895Z * [new tag] ciflow/inductor/163667 -> ciflow/inductor/163667 2025-10-10T01:18:34.2631057Z * [new tag] ciflow/inductor/163671 -> ciflow/inductor/163671 2025-10-10T01:18:34.2632370Z * [new tag] ciflow/inductor/163767 -> ciflow/inductor/163767 2025-10-10T01:18:34.2633575Z * [new tag] ciflow/inductor/163772 -> ciflow/inductor/163772 2025-10-10T01:18:34.2649816Z * [new tag] ciflow/inductor/163806 -> ciflow/inductor/163806 2025-10-10T01:18:34.2650204Z * [new tag] ciflow/inductor/163936 -> ciflow/inductor/163936 2025-10-10T01:18:34.2650369Z * [new tag] ciflow/inductor/163976 -> ciflow/inductor/163976 2025-10-10T01:18:34.2650519Z * [new tag] ciflow/inductor/164039 -> ciflow/inductor/164039 2025-10-10T01:18:34.2650662Z * [new tag] ciflow/inductor/164040 -> ciflow/inductor/164040 2025-10-10T01:18:34.2650796Z * [new tag] ciflow/inductor/164130 -> ciflow/inductor/164130 2025-10-10T01:18:34.2650925Z * [new tag] ciflow/inductor/164144 -> ciflow/inductor/164144 2025-10-10T01:18:34.2651249Z * [new tag] ciflow/inductor/164202 -> ciflow/inductor/164202 2025-10-10T01:18:34.2651378Z * [new tag] ciflow/inductor/164212 -> ciflow/inductor/164212 2025-10-10T01:18:34.2651505Z * [new tag] ciflow/inductor/164273 -> ciflow/inductor/164273 2025-10-10T01:18:34.2651651Z * [new tag] ciflow/inductor/164277 -> ciflow/inductor/164277 2025-10-10T01:18:34.2651775Z * [new tag] ciflow/inductor/164291 -> ciflow/inductor/164291 2025-10-10T01:18:34.2651901Z * [new tag] ciflow/inductor/164296 -> ciflow/inductor/164296 2025-10-10T01:18:34.2652034Z * [new tag] ciflow/inductor/164304 -> ciflow/inductor/164304 2025-10-10T01:18:34.2652404Z * [new tag] ciflow/inductor/164318 -> ciflow/inductor/164318 2025-10-10T01:18:34.2653812Z * [new tag] ciflow/inductor/164321 -> ciflow/inductor/164321 2025-10-10T01:18:34.2654932Z * [new tag] ciflow/inductor/164324 -> ciflow/inductor/164324 2025-10-10T01:18:34.2656320Z * [new tag] ciflow/inductor/164341 -> ciflow/inductor/164341 2025-10-10T01:18:34.2657510Z * [new tag] ciflow/inductor/164343 -> ciflow/inductor/164343 2025-10-10T01:18:34.2658627Z * [new tag] ciflow/inductor/164344 -> ciflow/inductor/164344 2025-10-10T01:18:34.2659945Z * [new tag] ciflow/inductor/164359 -> ciflow/inductor/164359 2025-10-10T01:18:34.2660963Z * [new tag] ciflow/inductor/164373 -> ciflow/inductor/164373 2025-10-10T01:18:34.2662534Z * [new tag] ciflow/inductor/164379 -> ciflow/inductor/164379 2025-10-10T01:18:34.2663513Z * [new tag] ciflow/inductor/164384 -> ciflow/inductor/164384 2025-10-10T01:18:34.2664899Z * [new tag] ciflow/inductor/164404 -> ciflow/inductor/164404 2025-10-10T01:18:34.2665902Z * [new tag] ciflow/inductor/164405 -> ciflow/inductor/164405 2025-10-10T01:18:34.2667268Z * [new tag] ciflow/inductor/164414 -> ciflow/inductor/164414 2025-10-10T01:18:34.2668429Z * [new tag] ciflow/inductor/164422 -> ciflow/inductor/164422 2025-10-10T01:18:34.2669648Z * [new tag] ciflow/inductor/164433 -> ciflow/inductor/164433 2025-10-10T01:18:34.2670888Z * [new tag] ciflow/inductor/164474 -> ciflow/inductor/164474 2025-10-10T01:18:34.2672160Z * [new tag] ciflow/inductor/164488 -> ciflow/inductor/164488 2025-10-10T01:18:34.2673356Z * [new tag] ciflow/inductor/164492 -> ciflow/inductor/164492 2025-10-10T01:18:34.2675114Z * [new tag] ciflow/inductor/164497 -> ciflow/inductor/164497 2025-10-10T01:18:34.2675742Z * [new tag] ciflow/inductor/164498 -> ciflow/inductor/164498 2025-10-10T01:18:34.2677076Z * [new tag] ciflow/inductor/164500 -> ciflow/inductor/164500 2025-10-10T01:18:34.2678286Z * [new tag] ciflow/inductor/164507 -> ciflow/inductor/164507 2025-10-10T01:18:34.2679625Z * [new tag] ciflow/inductor/164519 -> ciflow/inductor/164519 2025-10-10T01:18:34.2681056Z * [new tag] ciflow/inductor/164521 -> ciflow/inductor/164521 2025-10-10T01:18:34.2682258Z * [new tag] ciflow/inductor/164522 -> ciflow/inductor/164522 2025-10-10T01:18:34.2683455Z * [new tag] ciflow/inductor/164523 -> ciflow/inductor/164523 2025-10-10T01:18:34.2684679Z * [new tag] ciflow/inductor/164524 -> ciflow/inductor/164524 2025-10-10T01:18:34.2685958Z * [new tag] ciflow/inductor/164525 -> ciflow/inductor/164525 2025-10-10T01:18:34.2687100Z * [new tag] ciflow/inductor/164526 -> ciflow/inductor/164526 2025-10-10T01:18:34.2688512Z * [new tag] ciflow/inductor/164527 -> ciflow/inductor/164527 2025-10-10T01:18:34.2689707Z * [new tag] ciflow/inductor/164533 -> ciflow/inductor/164533 2025-10-10T01:18:34.2691027Z * [new tag] ciflow/inductor/164537 -> ciflow/inductor/164537 2025-10-10T01:18:34.2692162Z * [new tag] ciflow/inductor/164548 -> ciflow/inductor/164548 2025-10-10T01:18:34.2693363Z * [new tag] ciflow/inductor/164557 -> ciflow/inductor/164557 2025-10-10T01:18:34.2694600Z * [new tag] ciflow/inductor/164558 -> ciflow/inductor/164558 2025-10-10T01:18:34.2695957Z * [new tag] ciflow/inductor/164560 -> ciflow/inductor/164560 2025-10-10T01:18:34.2697180Z * [new tag] ciflow/inductor/164565 -> ciflow/inductor/164565 2025-10-10T01:18:34.2698379Z * [new tag] ciflow/inductor/164577 -> ciflow/inductor/164577 2025-10-10T01:18:34.2699713Z * [new tag] ciflow/inductor/164609 -> ciflow/inductor/164609 2025-10-10T01:18:34.2700950Z * [new tag] ciflow/inductor/164610 -> ciflow/inductor/164610 2025-10-10T01:18:34.2702138Z * [new tag] ciflow/inductor/164611 -> ciflow/inductor/164611 2025-10-10T01:18:34.2703359Z * [new tag] ciflow/inductor/164612 -> ciflow/inductor/164612 2025-10-10T01:18:34.2704563Z * [new tag] ciflow/inductor/164613 -> ciflow/inductor/164613 2025-10-10T01:18:34.2706210Z * [new tag] ciflow/inductor/164614 -> ciflow/inductor/164614 2025-10-10T01:18:34.2707598Z * [new tag] ciflow/inductor/164623 -> ciflow/inductor/164623 2025-10-10T01:18:34.2708854Z * [new tag] ciflow/inductor/164626 -> ciflow/inductor/164626 2025-10-10T01:18:34.2710057Z * [new tag] ciflow/inductor/164628 -> ciflow/inductor/164628 2025-10-10T01:18:34.2711294Z * [new tag] ciflow/inductor/164631 -> ciflow/inductor/164631 2025-10-10T01:18:34.2712593Z * [new tag] ciflow/inductor/164632 -> ciflow/inductor/164632 2025-10-10T01:18:34.2713903Z * [new tag] ciflow/inductor/164633 -> ciflow/inductor/164633 2025-10-10T01:18:34.2715099Z * [new tag] ciflow/inductor/164640 -> ciflow/inductor/164640 2025-10-10T01:18:34.2716437Z * [new tag] ciflow/inductor/164641 -> ciflow/inductor/164641 2025-10-10T01:18:34.2717815Z * [new tag] ciflow/inductor/164645 -> ciflow/inductor/164645 2025-10-10T01:18:34.2719097Z * [new tag] ciflow/inductor/164648 -> ciflow/inductor/164648 2025-10-10T01:18:34.2720560Z * [new tag] ciflow/inductor/164653 -> ciflow/inductor/164653 2025-10-10T01:18:34.2721795Z * [new tag] ciflow/inductor/164655 -> ciflow/inductor/164655 2025-10-10T01:18:34.2723079Z * [new tag] ciflow/inductor/164657 -> ciflow/inductor/164657 2025-10-10T01:18:34.2724668Z * [new tag] ciflow/inductor/164659 -> ciflow/inductor/164659 2025-10-10T01:18:34.2725994Z * [new tag] ciflow/inductor/164669 -> ciflow/inductor/164669 2025-10-10T01:18:34.2727360Z * [new tag] ciflow/inductor/164690 -> ciflow/inductor/164690 2025-10-10T01:18:34.2728579Z * [new tag] ciflow/inductor/164691 -> ciflow/inductor/164691 2025-10-10T01:18:34.2730024Z * [new tag] ciflow/inductor/164692 -> ciflow/inductor/164692 2025-10-10T01:18:34.2731252Z * [new tag] ciflow/inductor/164711 -> ciflow/inductor/164711 2025-10-10T01:18:34.2732473Z * [new tag] ciflow/inductor/164714 -> ciflow/inductor/164714 2025-10-10T01:18:34.2733668Z * [new tag] ciflow/inductor/164717 -> ciflow/inductor/164717 2025-10-10T01:18:34.2734904Z * [new tag] ciflow/inductor/164718 -> ciflow/inductor/164718 2025-10-10T01:18:34.2736337Z * [new tag] ciflow/inductor/164723 -> ciflow/inductor/164723 2025-10-10T01:18:34.2737598Z * [new tag] ciflow/inductor/164724 -> ciflow/inductor/164724 2025-10-10T01:18:34.2738807Z * [new tag] ciflow/inductor/164734 -> ciflow/inductor/164734 2025-10-10T01:18:34.2740014Z * [new tag] ciflow/inductor/164740 -> ciflow/inductor/164740 2025-10-10T01:18:34.2741409Z * [new tag] ciflow/inductor/164746 -> ciflow/inductor/164746 2025-10-10T01:18:34.2742660Z * [new tag] ciflow/inductor/164747 -> ciflow/inductor/164747 2025-10-10T01:18:34.2743881Z * [new tag] ciflow/inductor/164776 -> ciflow/inductor/164776 2025-10-10T01:18:34.2745087Z * [new tag] ciflow/inductor/164778 -> ciflow/inductor/164778 2025-10-10T01:18:34.2746305Z * [new tag] ciflow/inductor/164780 -> ciflow/inductor/164780 2025-10-10T01:18:34.2747564Z * [new tag] ciflow/inductor/164794 -> ciflow/inductor/164794 2025-10-10T01:18:34.2748778Z * [new tag] ciflow/inductor/164802 -> ciflow/inductor/164802 2025-10-10T01:18:34.2749991Z * [new tag] ciflow/inductor/164806 -> ciflow/inductor/164806 2025-10-10T01:18:34.2751490Z * [new tag] ciflow/inductor/164808 -> ciflow/inductor/164808 2025-10-10T01:18:34.2752673Z * [new tag] ciflow/inductor/164810 -> ciflow/inductor/164810 2025-10-10T01:18:34.2753961Z * [new tag] ciflow/inductor/164811 -> ciflow/inductor/164811 2025-10-10T01:18:34.2755187Z * [new tag] ciflow/inductor/164812 -> ciflow/inductor/164812 2025-10-10T01:18:34.2756430Z * [new tag] ciflow/inductor/164819 -> ciflow/inductor/164819 2025-10-10T01:18:34.2757607Z * [new tag] ciflow/inductor/164820 -> ciflow/inductor/164820 2025-10-10T01:18:34.2758887Z * [new tag] ciflow/inductor/164821 -> ciflow/inductor/164821 2025-10-10T01:18:34.2760104Z * [new tag] ciflow/inductor/164839 -> ciflow/inductor/164839 2025-10-10T01:18:34.2761362Z * [new tag] ciflow/inductor/164842 -> ciflow/inductor/164842 2025-10-10T01:18:34.2762549Z * [new tag] ciflow/inductor/164847 -> ciflow/inductor/164847 2025-10-10T01:18:34.2763748Z * [new tag] ciflow/inductor/164852 -> ciflow/inductor/164852 2025-10-10T01:18:34.2764987Z * [new tag] ciflow/inductor/164863 -> ciflow/inductor/164863 2025-10-10T01:18:34.2766269Z * [new tag] ciflow/inductor/164865 -> ciflow/inductor/164865 2025-10-10T01:18:34.2767635Z * [new tag] ciflow/inductor/164866 -> ciflow/inductor/164866 2025-10-10T01:18:34.2768772Z * [new tag] ciflow/inductor/164867 -> ciflow/inductor/164867 2025-10-10T01:18:34.2769956Z * [new tag] ciflow/inductor/164869 -> ciflow/inductor/164869 2025-10-10T01:18:34.2771205Z * [new tag] ciflow/inductor/164873 -> ciflow/inductor/164873 2025-10-10T01:18:34.2772422Z * [new tag] ciflow/inductor/164889 -> ciflow/inductor/164889 2025-10-10T01:18:34.2773624Z * [new tag] ciflow/inductor/164897 -> ciflow/inductor/164897 2025-10-10T01:18:34.2774839Z * [new tag] ciflow/inductor/164902 -> ciflow/inductor/164902 2025-10-10T01:18:34.2776169Z * [new tag] ciflow/inductor/164903 -> ciflow/inductor/164903 2025-10-10T01:18:34.2777347Z * [new tag] ciflow/inductor/164906 -> ciflow/inductor/164906 2025-10-10T01:18:34.2778567Z * [new tag] ciflow/inductor/164914 -> ciflow/inductor/164914 2025-10-10T01:18:34.2779762Z * [new tag] ciflow/inductor/164919 -> ciflow/inductor/164919 2025-10-10T01:18:34.2781001Z * [new tag] ciflow/inductor/164933 -> ciflow/inductor/164933 2025-10-10T01:18:34.2782387Z * [new tag] ciflow/inductor/164938 -> ciflow/inductor/164938 2025-10-10T01:18:34.2783567Z * [new tag] ciflow/inductor/164948 -> ciflow/inductor/164948 2025-10-10T01:18:34.2784766Z * [new tag] ciflow/inductor/164956 -> ciflow/inductor/164956 2025-10-10T01:18:34.2786015Z * [new tag] ciflow/inductor/164965 -> ciflow/inductor/164965 2025-10-10T01:18:34.2787598Z * [new tag] ciflow/inductor/164978 -> ciflow/inductor/164978 2025-10-10T01:18:34.2788945Z * [new tag] ciflow/inductor/164979 -> ciflow/inductor/164979 2025-10-10T01:18:34.2790185Z * [new tag] ciflow/inductor/164980 -> ciflow/inductor/164980 2025-10-10T01:18:34.2791556Z * [new tag] ciflow/inductor/164984 -> ciflow/inductor/164984 2025-10-10T01:18:34.2792963Z * [new tag] ciflow/inductor/164989 -> ciflow/inductor/164989 2025-10-10T01:18:34.2794235Z * [new tag] ciflow/inductor/164991 -> ciflow/inductor/164991 2025-10-10T01:18:34.2795486Z * [new tag] ciflow/inductor/164992 -> ciflow/inductor/164992 2025-10-10T01:18:34.2796888Z * [new tag] ciflow/inductor/164994 -> ciflow/inductor/164994 2025-10-10T01:18:34.2798131Z * [new tag] ciflow/inductor/164999 -> ciflow/inductor/164999 2025-10-10T01:18:34.2799524Z * [new tag] ciflow/inductor/165001 -> ciflow/inductor/165001 2025-10-10T01:18:34.2800718Z * [new tag] ciflow/inductor/165005 -> ciflow/inductor/165005 2025-10-10T01:18:34.2801944Z * [new tag] ciflow/inductor/165006 -> ciflow/inductor/165006 2025-10-10T01:18:34.2803165Z * [new tag] ciflow/inductor/165010 -> ciflow/inductor/165010 2025-10-10T01:18:34.2804382Z * [new tag] ciflow/inductor/165012 -> ciflow/inductor/165012 2025-10-10T01:18:34.2805600Z * [new tag] ciflow/inductor/165017 -> ciflow/inductor/165017 2025-10-10T01:18:34.2806967Z * [new tag] ciflow/inductor/165018 -> ciflow/inductor/165018 2025-10-10T01:18:34.2808174Z * [new tag] ciflow/inductor/165024 -> ciflow/inductor/165024 2025-10-10T01:18:34.2809523Z * [new tag] ciflow/inductor/165029 -> ciflow/inductor/165029 2025-10-10T01:18:34.2810891Z * [new tag] ciflow/inductor/165030 -> ciflow/inductor/165030 2025-10-10T01:18:34.2812236Z * [new tag] ciflow/inductor/165031 -> ciflow/inductor/165031 2025-10-10T01:18:34.2813654Z * [new tag] ciflow/inductor/165033 -> ciflow/inductor/165033 2025-10-10T01:18:34.2814996Z * [new tag] ciflow/inductor/165036 -> ciflow/inductor/165036 2025-10-10T01:18:34.2816400Z * [new tag] ciflow/inductor/165037 -> ciflow/inductor/165037 2025-10-10T01:18:34.2817624Z * [new tag] ciflow/inductor/165039 -> ciflow/inductor/165039 2025-10-10T01:18:34.2819024Z * [new tag] ciflow/inductor/165047 -> ciflow/inductor/165047 2025-10-10T01:18:34.2820220Z * [new tag] ciflow/inductor/165059 -> ciflow/inductor/165059 2025-10-10T01:18:34.2821457Z * [new tag] ciflow/inductor/165063 -> ciflow/inductor/165063 2025-10-10T01:18:34.2822685Z * [new tag] ciflow/inductor/165064 -> ciflow/inductor/165064 2025-10-10T01:18:34.2823920Z * [new tag] ciflow/inductor/165066 -> ciflow/inductor/165066 2025-10-10T01:18:34.2825515Z * [new tag] ciflow/inductor/165074 -> ciflow/inductor/165074 2025-10-10T01:18:34.2826710Z * [new tag] ciflow/inductor/165076 -> ciflow/inductor/165076 2025-10-10T01:18:34.2827917Z * [new tag] ciflow/inductor/165091 -> ciflow/inductor/165091 2025-10-10T01:18:34.2829147Z * [new tag] ciflow/inductor/165092 -> ciflow/inductor/165092 2025-10-10T01:18:34.2830360Z * [new tag] ciflow/inductor/165106 -> ciflow/inductor/165106 2025-10-10T01:18:34.2831604Z * [new tag] ciflow/inductor/165107 -> ciflow/inductor/165107 2025-10-10T01:18:34.2832987Z * [new tag] ciflow/inductor/165112 -> ciflow/inductor/165112 2025-10-10T01:18:34.2834204Z * [new tag] ciflow/inductor/165113 -> ciflow/inductor/165113 2025-10-10T01:18:34.2835493Z * [new tag] ciflow/inductor/165114 -> ciflow/inductor/165114 2025-10-10T01:18:34.2837011Z * [new tag] ciflow/inductor/3b9a386 -> ciflow/inductor/3b9a386 2025-10-10T01:18:34.2838412Z * [new tag] ciflow/inductor/3d4b92b -> ciflow/inductor/3d4b92b 2025-10-10T01:18:34.2839885Z * [new tag] ciflow/inductor/d224ac7 -> ciflow/inductor/d224ac7 2025-10-10T01:18:34.2841317Z * [new tag] ciflow/linux-aarch64/157994 -> ciflow/linux-aarch64/157994 2025-10-10T01:18:34.2842634Z * [new tag] ciflow/linux-aarch64/163952 -> ciflow/linux-aarch64/163952 2025-10-10T01:18:34.2843756Z * [new tag] ciflow/linux-aarch64/164965 -> ciflow/linux-aarch64/164965 2025-10-10T01:18:34.2844882Z * [new tag] ciflow/linux-aarch64/165010 -> ciflow/linux-aarch64/165010 2025-10-10T01:18:34.2846309Z * [new tag] ciflow/mps/157553 -> ciflow/mps/157553 2025-10-10T01:18:34.2847460Z * [new tag] ciflow/mps/157554 -> ciflow/mps/157554 2025-10-10T01:18:34.2848572Z * [new tag] ciflow/mps/157635 -> ciflow/mps/157635 2025-10-10T01:18:34.2849706Z * [new tag] ciflow/mps/162340 -> ciflow/mps/162340 2025-10-10T01:18:34.2850828Z * [new tag] ciflow/mps/164416 -> ciflow/mps/164416 2025-10-10T01:18:34.2852192Z * [new tag] ciflow/mps/164571 -> ciflow/mps/164571 2025-10-10T01:18:34.2853336Z * [new tag] ciflow/mps/164965 -> ciflow/mps/164965 2025-10-10T01:18:34.2854775Z * [new tag] ciflow/nightly/158104 -> ciflow/nightly/158104 2025-10-10T01:18:34.2855933Z * [new tag] ciflow/nightly/164747 -> ciflow/nightly/164747 2025-10-10T01:18:34.2857054Z * [new tag] ciflow/nightly/164901 -> ciflow/nightly/164901 2025-10-10T01:18:34.2858449Z * [new tag] ciflow/op-benchmark/157994 -> ciflow/op-benchmark/157994 2025-10-10T01:18:34.2859392Z * [new tag] ciflow/op-benchmark/163767 -> ciflow/op-benchmark/163767 2025-10-10T01:18:34.2861069Z * [new tag] ciflow/op-benchmark/164583 -> ciflow/op-benchmark/164583 2025-10-10T01:18:34.2861808Z * [new tag] ciflow/op-benchmark/164747 -> ciflow/op-benchmark/164747 2025-10-10T01:18:34.2863337Z * [new tag] ciflow/periodic-rocm-mi300/162478 -> ciflow/periodic-rocm-mi300/162478 2025-10-10T01:18:34.2864524Z * [new tag] ciflow/periodic-rocm-mi300/163767 -> ciflow/periodic-rocm-mi300/163767 2025-10-10T01:18:34.2865693Z * [new tag] ciflow/periodic-rocm-mi300/164618 -> ciflow/periodic-rocm-mi300/164618 2025-10-10T01:18:34.2866608Z * [new tag] ciflow/periodic-rocm-mi300/164747 -> ciflow/periodic-rocm-mi300/164747 2025-10-10T01:18:34.2868054Z * [new tag] ciflow/periodic-rocm-mi300/165011 -> ciflow/periodic-rocm-mi300/165011 2025-10-10T01:18:34.2869266Z * [new tag] ciflow/periodic-rocm-mi300/165080 -> ciflow/periodic-rocm-mi300/165080 2025-10-10T01:18:34.2870456Z * [new tag] ciflow/periodic-rocm-mi300/165115 -> ciflow/periodic-rocm-mi300/165115 2025-10-10T01:18:34.2872120Z * [new tag] ciflow/periodic/054a2fd -> ciflow/periodic/054a2fd 2025-10-10T01:18:34.2873509Z * [new tag] ciflow/periodic/0d39ecb2ce8556e85343d8da0c87450192c2fdf8 -> ciflow/periodic/0d39ecb2ce8556e85343d8da0c87450192c2fdf8 2025-10-10T01:18:34.2874752Z * [new tag] ciflow/periodic/0ea59c3c55dab37a6edefcc7002bb1428afd6456 -> ciflow/periodic/0ea59c3c55dab37a6edefcc7002bb1428afd6456 2025-10-10T01:18:34.2875999Z * [new tag] ciflow/periodic/156491 -> ciflow/periodic/156491 2025-10-10T01:18:34.2877178Z * [new tag] ciflow/periodic/162990 -> ciflow/periodic/162990 2025-10-10T01:18:34.2878627Z * [new tag] ciflow/periodic/163667 -> ciflow/periodic/163667 2025-10-10T01:18:34.2879835Z * [new tag] ciflow/periodic/163767 -> ciflow/periodic/163767 2025-10-10T01:18:34.2881325Z * [new tag] ciflow/periodic/164747 -> ciflow/periodic/164747 2025-10-10T01:18:34.2882759Z * [new tag] ciflow/periodic/164769 -> ciflow/periodic/164769 2025-10-10T01:18:34.2883960Z * [new tag] ciflow/periodic/165011 -> ciflow/periodic/165011 2025-10-10T01:18:34.2885642Z * [new tag] ciflow/periodic/2a6cdba6e5f74c2294fecc2d1344537522efbaab -> ciflow/periodic/2a6cdba6e5f74c2294fecc2d1344537522efbaab 2025-10-10T01:18:34.2886873Z * [new tag] ciflow/periodic/2a6d37d -> ciflow/periodic/2a6d37d 2025-10-10T01:18:34.2888217Z * [new tag] ciflow/periodic/317eeb8 -> ciflow/periodic/317eeb8 2025-10-10T01:18:34.2889563Z * [new tag] ciflow/periodic/3c32 -> ciflow/periodic/3c32 2025-10-10T01:18:34.2891012Z * [new tag] ciflow/periodic/3e98831 -> ciflow/periodic/3e98831 2025-10-10T01:18:34.2892603Z * [new tag] ciflow/periodic/4bcc05777e780e834d44a2d06dd5321daec316f0 -> ciflow/periodic/4bcc05777e780e834d44a2d06dd5321daec316f0 2025-10-10T01:18:34.2893694Z * [new tag] ciflow/periodic/73adac05d13babb75410c3e033fdce57aa16881a -> ciflow/periodic/73adac05d13babb75410c3e033fdce57aa16881a 2025-10-10T01:18:34.2895220Z * [new tag] ciflow/periodic/94512-point -> ciflow/periodic/94512-point 2025-10-10T01:18:34.2896821Z * [new tag] ciflow/periodic/ac08556f674259ff5b117964e300124e8a92d45b -> ciflow/periodic/ac08556f674259ff5b117964e300124e8a92d45b 2025-10-10T01:18:34.2898331Z * [new tag] ciflow/periodic/csl/test87519 -> ciflow/periodic/csl/test87519 2025-10-10T01:18:34.2899675Z * [new tag] ciflow/periodic/csltest88275 -> ciflow/periodic/csltest88275 2025-10-10T01:18:34.2901038Z * [new tag] ciflow/periodic/csltest88761 -> ciflow/periodic/csltest88761 2025-10-10T01:18:34.2902480Z * [new tag] ciflow/periodic/release_1.12 -> ciflow/periodic/release_1.12 2025-10-10T01:18:34.2904114Z * [new tag] ciflow/periodic/release_1.12.0 -> ciflow/periodic/release_1.12.0 2025-10-10T01:18:34.2905526Z * [new tag] ciflow/periodic/sha-ec5b83 -> ciflow/periodic/sha-ec5b83 2025-10-10T01:18:34.2906994Z * [new tag] ciflow/quantization-periodic/163767 -> ciflow/quantization-periodic/163767 2025-10-10T01:18:34.2908173Z * [new tag] ciflow/quantization-periodic/164747 -> ciflow/quantization-periodic/164747 2025-10-10T01:18:34.2909572Z * [new tag] ciflow/riscv64/163767 -> ciflow/riscv64/163767 2025-10-10T01:18:34.2910661Z * [new tag] ciflow/riscv64/164747 -> ciflow/riscv64/164747 2025-10-10T01:18:34.2912067Z * [new tag] ciflow/rocm-mi300/161280 -> ciflow/rocm-mi300/161280 2025-10-10T01:18:34.2913201Z * [new tag] ciflow/rocm-mi300/162478 -> ciflow/rocm-mi300/162478 2025-10-10T01:18:34.2914367Z * [new tag] ciflow/rocm-mi300/163767 -> ciflow/rocm-mi300/163767 2025-10-10T01:18:34.2915471Z * [new tag] ciflow/rocm-mi300/163955 -> ciflow/rocm-mi300/163955 2025-10-10T01:18:34.2916563Z * [new tag] ciflow/rocm-mi300/164618 -> ciflow/rocm-mi300/164618 2025-10-10T01:18:34.2917558Z * [new tag] ciflow/rocm-mi300/164747 -> ciflow/rocm-mi300/164747 2025-10-10T01:18:34.2919232Z * [new tag] ciflow/rocm-mi300/164927 -> ciflow/rocm-mi300/164927 2025-10-10T01:18:34.2920352Z * [new tag] ciflow/rocm-mi300/164930 -> ciflow/rocm-mi300/164930 2025-10-10T01:18:34.2921594Z * [new tag] ciflow/rocm-mi300/165026 -> ciflow/rocm-mi300/165026 2025-10-10T01:18:34.2922733Z * [new tag] ciflow/rocm-mi300/165080 -> ciflow/rocm-mi300/165080 2025-10-10T01:18:34.2923870Z * [new tag] ciflow/rocm-mi300/165115 -> ciflow/rocm-mi300/165115 2025-10-10T01:18:34.2925798Z * [new tag] ciflow/rocm-mi355/163767 -> ciflow/rocm-mi355/163767 2025-10-10T01:18:34.2926921Z * [new tag] ciflow/rocm-mi355/164747 -> ciflow/rocm-mi355/164747 2025-10-10T01:18:34.2928372Z * [new tag] ciflow/rocm/148492 -> ciflow/rocm/148492 2025-10-10T01:18:34.2929484Z * [new tag] ciflow/rocm/151845 -> ciflow/rocm/151845 2025-10-10T01:18:34.2930582Z * [new tag] ciflow/rocm/156592 -> ciflow/rocm/156592 2025-10-10T01:18:34.2931722Z * [new tag] ciflow/rocm/161280 -> ciflow/rocm/161280 2025-10-10T01:18:34.2932847Z * [new tag] ciflow/rocm/163767 -> ciflow/rocm/163767 2025-10-10T01:18:34.2933960Z * [new tag] ciflow/rocm/163955 -> ciflow/rocm/163955 2025-10-10T01:18:34.2935254Z * [new tag] ciflow/rocm/163965 -> ciflow/rocm/163965 2025-10-10T01:18:34.2936451Z * [new tag] ciflow/rocm/164656 -> ciflow/rocm/164656 2025-10-10T01:18:34.2937710Z * [new tag] ciflow/rocm/164747 -> ciflow/rocm/164747 2025-10-10T01:18:34.2938821Z * [new tag] ciflow/rocm/164769 -> ciflow/rocm/164769 2025-10-10T01:18:34.2939947Z * [new tag] ciflow/rocm/164927 -> ciflow/rocm/164927 2025-10-10T01:18:34.2941389Z * [new tag] ciflow/rocm/164930 -> ciflow/rocm/164930 2025-10-10T01:18:34.2943369Z * [new tag] ciflow/rocm/165026 -> ciflow/rocm/165026 2025-10-10T01:18:34.2944809Z * [new tag] ciflow/rocm/165103 -> ciflow/rocm/165103 2025-10-10T01:18:34.2946267Z * [new tag] ciflow/s390/164747 -> ciflow/s390/164747 2025-10-10T01:18:34.2947575Z * [new tag] ciflow/s390/164917 -> ciflow/s390/164917 2025-10-10T01:18:34.2949131Z * [new tag] ciflow/slow/01c7106 -> ciflow/slow/01c7106 2025-10-10T01:18:34.2950525Z * [new tag] ciflow/slow/0577043 -> ciflow/slow/0577043 2025-10-10T01:18:34.2952500Z * [new tag] ciflow/slow/0d5b74da0cab798fbfdb9caa53fad816999c8386-sdym -> ciflow/slow/0d5b74da0cab798fbfdb9caa53fad816999c8386-sdym 2025-10-10T01:18:34.2953378Z * [new tag] ciflow/slow/0e81104 -> ciflow/slow/0e81104 2025-10-10T01:18:34.2954722Z * [new tag] ciflow/slow/163767 -> ciflow/slow/163767 2025-10-10T01:18:34.2955845Z * [new tag] ciflow/slow/164747 -> ciflow/slow/164747 2025-10-10T01:18:34.2956942Z * [new tag] ciflow/slow/164769 -> ciflow/slow/164769 2025-10-10T01:18:34.2958254Z * [new tag] ciflow/slow/1732077 -> ciflow/slow/1732077 2025-10-10T01:18:34.2960080Z * [new tag] ciflow/slow/187eb7c -> ciflow/slow/187eb7c 2025-10-10T01:18:34.2961624Z * [new tag] ciflow/slow/1faef89 -> ciflow/slow/1faef89 2025-10-10T01:18:34.2963141Z * [new tag] ciflow/slow/3920ec1 -> ciflow/slow/3920ec1 2025-10-10T01:18:34.2964509Z * [new tag] ciflow/slow/3b7c6b2 -> ciflow/slow/3b7c6b2 2025-10-10T01:18:34.2965949Z * [new tag] ciflow/slow/59a3759 -> ciflow/slow/59a3759 2025-10-10T01:18:34.2967305Z * [new tag] ciflow/slow/70ef0bb -> ciflow/slow/70ef0bb 2025-10-10T01:18:34.2968635Z * [new tag] ciflow/slow/788ff06 -> ciflow/slow/788ff06 2025-10-10T01:18:34.2970585Z * [new tag] ciflow/slow/8751002215790a3a88750faa8f4366933e296693-sdym -> ciflow/slow/8751002215790a3a88750faa8f4366933e296693-sdym 2025-10-10T01:18:34.2971749Z * [new tag] ciflow/slow/9d85864 -> ciflow/slow/9d85864 2025-10-10T01:18:34.2973080Z * [new tag] ciflow/slow/9ffad5b -> ciflow/slow/9ffad5b 2025-10-10T01:18:34.2974431Z * [new tag] ciflow/slow/a206e8b -> ciflow/slow/a206e8b 2025-10-10T01:18:34.2975835Z * [new tag] ciflow/slow/a837609 -> ciflow/slow/a837609 2025-10-10T01:18:34.2977249Z * [new tag] ciflow/slow/af841f3 -> ciflow/slow/af841f3 2025-10-10T01:18:34.2978954Z * [new tag] ciflow/slow/da3aba1e46157c4df504b067477cdf2b3c96b194-sdym -> ciflow/slow/da3aba1e46157c4df504b067477cdf2b3c96b194-sdym 2025-10-10T01:18:34.2980169Z * [new tag] ciflow/torchbench/164747 -> ciflow/torchbench/164747 2025-10-10T01:18:34.2981557Z * [new tag] ciflow/trunk/113258 -> ciflow/trunk/113258 2025-10-10T01:18:34.2982723Z * [new tag] ciflow/trunk/137400 -> ciflow/trunk/137400 2025-10-10T01:18:34.2983916Z * [new tag] ciflow/trunk/148180 -> ciflow/trunk/148180 2025-10-10T01:18:34.2985053Z * [new tag] ciflow/trunk/148328 -> ciflow/trunk/148328 2025-10-10T01:18:34.2986194Z * [new tag] ciflow/trunk/148492 -> ciflow/trunk/148492 2025-10-10T01:18:34.2987323Z * [new tag] ciflow/trunk/149003 -> ciflow/trunk/149003 2025-10-10T01:18:34.2988492Z * [new tag] ciflow/trunk/149536 -> ciflow/trunk/149536 2025-10-10T01:18:34.2989585Z * [new tag] ciflow/trunk/151845 -> ciflow/trunk/151845 2025-10-10T01:18:34.2990715Z * [new tag] ciflow/trunk/152624 -> ciflow/trunk/152624 2025-10-10T01:18:34.2992441Z * [new tag] ciflow/trunk/154279 -> ciflow/trunk/154279 2025-10-10T01:18:34.2993589Z * [new tag] ciflow/trunk/154983 -> ciflow/trunk/154983 2025-10-10T01:18:34.2995148Z * [new tag] ciflow/trunk/156418 -> ciflow/trunk/156418 2025-10-10T01:18:34.2996688Z * [new tag] ciflow/trunk/156592 -> ciflow/trunk/156592 2025-10-10T01:18:34.2998130Z * [new tag] ciflow/trunk/157432 -> ciflow/trunk/157432 2025-10-10T01:18:34.2999594Z * [new tag] ciflow/trunk/157994 -> ciflow/trunk/157994 2025-10-10T01:18:34.3000713Z * [new tag] ciflow/trunk/158104 -> ciflow/trunk/158104 2025-10-10T01:18:34.3001886Z * [new tag] ciflow/trunk/158932 -> ciflow/trunk/158932 2025-10-10T01:18:34.3003091Z * [new tag] ciflow/trunk/159104 -> ciflow/trunk/159104 2025-10-10T01:18:34.3004495Z * [new tag] ciflow/trunk/159936 -> ciflow/trunk/159936 2025-10-10T01:18:34.3005711Z * [new tag] ciflow/trunk/160266 -> ciflow/trunk/160266 2025-10-10T01:18:34.3006888Z * [new tag] ciflow/trunk/160328 -> ciflow/trunk/160328 2025-10-10T01:18:34.3008122Z * [new tag] ciflow/trunk/160329 -> ciflow/trunk/160329 2025-10-10T01:18:34.3009319Z * [new tag] ciflow/trunk/160539 -> ciflow/trunk/160539 2025-10-10T01:18:34.3010744Z * [new tag] ciflow/trunk/160610 -> ciflow/trunk/160610 2025-10-10T01:18:34.3011952Z * [new tag] ciflow/trunk/160843 -> ciflow/trunk/160843 2025-10-10T01:18:34.3013296Z * [new tag] ciflow/trunk/161035 -> ciflow/trunk/161035 2025-10-10T01:18:34.3014505Z * [new tag] ciflow/trunk/161320 -> ciflow/trunk/161320 2025-10-10T01:18:34.3015796Z * [new tag] ciflow/trunk/162031 -> ciflow/trunk/162031 2025-10-10T01:18:34.3017020Z * [new tag] ciflow/trunk/162066 -> ciflow/trunk/162066 2025-10-10T01:18:34.3018349Z * [new tag] ciflow/trunk/162203 -> ciflow/trunk/162203 2025-10-10T01:18:34.3019619Z * [new tag] ciflow/trunk/162340 -> ciflow/trunk/162340 2025-10-10T01:18:34.3020826Z * [new tag] ciflow/trunk/162542 -> ciflow/trunk/162542 2025-10-10T01:18:34.3022030Z * [new tag] ciflow/trunk/162899 -> ciflow/trunk/162899 2025-10-10T01:18:34.3023399Z * [new tag] ciflow/trunk/163034 -> ciflow/trunk/163034 2025-10-10T01:18:34.3025097Z * [new tag] ciflow/trunk/163332 -> ciflow/trunk/163332 2025-10-10T01:18:34.3026411Z * [new tag] ciflow/trunk/163446 -> ciflow/trunk/163446 2025-10-10T01:18:34.3027669Z * [new tag] ciflow/trunk/163490 -> ciflow/trunk/163490 2025-10-10T01:18:34.3028876Z * [new tag] ciflow/trunk/163527 -> ciflow/trunk/163527 2025-10-10T01:18:34.3030090Z * [new tag] ciflow/trunk/163533 -> ciflow/trunk/163533 2025-10-10T01:18:34.3031264Z * [new tag] ciflow/trunk/163671 -> ciflow/trunk/163671 2025-10-10T01:18:34.3032489Z * [new tag] ciflow/trunk/163767 -> ciflow/trunk/163767 2025-10-10T01:18:34.3033694Z * [new tag] ciflow/trunk/163846 -> ciflow/trunk/163846 2025-10-10T01:18:34.3034937Z * [new tag] ciflow/trunk/163899 -> ciflow/trunk/163899 2025-10-10T01:18:34.3036109Z * [new tag] ciflow/trunk/163955 -> ciflow/trunk/163955 2025-10-10T01:18:34.3037302Z * [new tag] ciflow/trunk/163976 -> ciflow/trunk/163976 2025-10-10T01:18:34.3038602Z * [new tag] ciflow/trunk/164040 -> ciflow/trunk/164040 2025-10-10T01:18:34.3039883Z * [new tag] ciflow/trunk/164130 -> ciflow/trunk/164130 2025-10-10T01:18:34.3041061Z * [new tag] ciflow/trunk/164144 -> ciflow/trunk/164144 2025-10-10T01:18:34.3042281Z * [new tag] ciflow/trunk/164202 -> ciflow/trunk/164202 2025-10-10T01:18:34.3043475Z * [new tag] ciflow/trunk/164318 -> ciflow/trunk/164318 2025-10-10T01:18:34.3044679Z * [new tag] ciflow/trunk/164414 -> ciflow/trunk/164414 2025-10-10T01:18:34.3045886Z * [new tag] ciflow/trunk/164416 -> ciflow/trunk/164416 2025-10-10T01:18:34.3047264Z * [new tag] ciflow/trunk/164437 -> ciflow/trunk/164437 2025-10-10T01:18:34.3048312Z * [new tag] ciflow/trunk/164467 -> ciflow/trunk/164467 2025-10-10T01:18:34.3049530Z * [new tag] ciflow/trunk/164500 -> ciflow/trunk/164500 2025-10-10T01:18:34.3050718Z * [new tag] ciflow/trunk/164510 -> ciflow/trunk/164510 2025-10-10T01:18:34.3051948Z * [new tag] ciflow/trunk/164519 -> ciflow/trunk/164519 2025-10-10T01:18:34.3053190Z * [new tag] ciflow/trunk/164542 -> ciflow/trunk/164542 2025-10-10T01:18:34.3054382Z * [new tag] ciflow/trunk/164560 -> ciflow/trunk/164560 2025-10-10T01:18:34.3055946Z * [new tag] ciflow/trunk/164566 -> ciflow/trunk/164566 2025-10-10T01:18:34.3057166Z * [new tag] ciflow/trunk/164623 -> ciflow/trunk/164623 2025-10-10T01:18:34.3058343Z * [new tag] ciflow/trunk/164628 -> ciflow/trunk/164628 2025-10-10T01:18:34.3059527Z * [new tag] ciflow/trunk/164641 -> ciflow/trunk/164641 2025-10-10T01:18:34.3060854Z * [new tag] ciflow/trunk/164643 -> ciflow/trunk/164643 2025-10-10T01:18:34.3062146Z * [new tag] ciflow/trunk/164645 -> ciflow/trunk/164645 2025-10-10T01:18:34.3063300Z * [new tag] ciflow/trunk/164653 -> ciflow/trunk/164653 2025-10-10T01:18:34.3064573Z * [new tag] ciflow/trunk/164655 -> ciflow/trunk/164655 2025-10-10T01:18:34.3065799Z * [new tag] ciflow/trunk/164691 -> ciflow/trunk/164691 2025-10-10T01:18:34.3066999Z * [new tag] ciflow/trunk/164692 -> ciflow/trunk/164692 2025-10-10T01:18:34.3068262Z * [new tag] ciflow/trunk/164705 -> ciflow/trunk/164705 2025-10-10T01:18:34.3069478Z * [new tag] ciflow/trunk/164746 -> ciflow/trunk/164746 2025-10-10T01:18:34.3070684Z * [new tag] ciflow/trunk/164747 -> ciflow/trunk/164747 2025-10-10T01:18:34.3071900Z * [new tag] ciflow/trunk/164790 -> ciflow/trunk/164790 2025-10-10T01:18:34.3073536Z * [new tag] ciflow/trunk/164808 -> ciflow/trunk/164808 2025-10-10T01:18:34.3074751Z * [new tag] ciflow/trunk/164812 -> ciflow/trunk/164812 2025-10-10T01:18:34.3076109Z * [new tag] ciflow/trunk/164836 -> ciflow/trunk/164836 2025-10-10T01:18:34.3077409Z * [new tag] ciflow/trunk/164842 -> ciflow/trunk/164842 2025-10-10T01:18:34.3078658Z * [new tag] ciflow/trunk/164882 -> ciflow/trunk/164882 2025-10-10T01:18:34.3079971Z * [new tag] ciflow/trunk/164889 -> ciflow/trunk/164889 2025-10-10T01:18:34.3081147Z * [new tag] ciflow/trunk/164894 -> ciflow/trunk/164894 2025-10-10T01:18:34.3082383Z * [new tag] ciflow/trunk/164930 -> ciflow/trunk/164930 2025-10-10T01:18:34.3083549Z * [new tag] ciflow/trunk/164953 -> ciflow/trunk/164953 2025-10-10T01:18:34.3084731Z * [new tag] ciflow/trunk/164976 -> ciflow/trunk/164976 2025-10-10T01:18:34.3085959Z * [new tag] ciflow/trunk/164999 -> ciflow/trunk/164999 2025-10-10T01:18:34.3087127Z * [new tag] ciflow/trunk/165000 -> ciflow/trunk/165000 2025-10-10T01:18:34.3088311Z * [new tag] ciflow/trunk/165017 -> ciflow/trunk/165017 2025-10-10T01:18:34.3089524Z * [new tag] ciflow/trunk/165018 -> ciflow/trunk/165018 2025-10-10T01:18:34.3090741Z * [new tag] ciflow/trunk/165024 -> ciflow/trunk/165024 2025-10-10T01:18:34.3092036Z * [new tag] ciflow/trunk/165031 -> ciflow/trunk/165031 2025-10-10T01:18:34.3093390Z * [new tag] ciflow/trunk/165033 -> ciflow/trunk/165033 2025-10-10T01:18:34.3094505Z * [new tag] ciflow/trunk/165047 -> ciflow/trunk/165047 2025-10-10T01:18:34.3095683Z * [new tag] ciflow/trunk/165057 -> ciflow/trunk/165057 2025-10-10T01:18:34.3097017Z * [new tag] ciflow/trunk/165060 -> ciflow/trunk/165060 2025-10-10T01:18:34.3098176Z * [new tag] ciflow/trunk/165065 -> ciflow/trunk/165065 2025-10-10T01:18:34.3099370Z * [new tag] ciflow/trunk/165066 -> ciflow/trunk/165066 2025-10-10T01:18:34.3100878Z * [new tag] ciflow/trunk/165090 -> ciflow/trunk/165090 2025-10-10T01:18:34.3102321Z * [new tag] ciflow/trunk/165094 -> ciflow/trunk/165094 2025-10-10T01:18:34.3103502Z * [new tag] ciflow/trunk/165113 -> ciflow/trunk/165113 2025-10-10T01:18:34.3105238Z * [new tag] ciflow/unstable/123 -> ciflow/unstable/123 2025-10-10T01:18:34.3106640Z * [new tag] ciflow/vllm/164628 -> ciflow/vllm/164628 2025-10-10T01:18:34.3108109Z * [new tag] ciflow/win-arm64/158104 -> ciflow/win-arm64/158104 2025-10-10T01:18:34.3109496Z * [new tag] ciflow/xpu/157994 -> ciflow/xpu/157994 2025-10-10T01:18:34.3110646Z * [new tag] ciflow/xpu/161485 -> ciflow/xpu/161485 2025-10-10T01:18:34.3111902Z * [new tag] ciflow/xpu/162454 -> ciflow/xpu/162454 2025-10-10T01:18:34.3113074Z * [new tag] ciflow/xpu/163332 -> ciflow/xpu/163332 2025-10-10T01:18:34.3114287Z * [new tag] cslpull75 -> cslpull75 2025-10-10T01:18:34.3115479Z * [new tag] cslpull76 -> cslpull76 2025-10-10T01:18:34.3116621Z * [new tag] cslpull77 -> cslpull77 2025-10-10T01:18:34.3117940Z * [new tag] cslpull78 -> cslpull78 2025-10-10T01:18:34.3119529Z * [new tag] cslpull79 -> cslpull79 2025-10-10T01:18:34.3121003Z * [new tag] cslpull80 -> cslpull80 2025-10-10T01:18:34.3122324Z * [new tag] cslpull81 -> cslpull81 2025-10-10T01:18:34.3123693Z * [new tag] cslpull82 -> cslpull82 2025-10-10T01:18:34.3125374Z * [new tag] cslpull83 -> cslpull83 2025-10-10T01:18:34.3126571Z * [new tag] cslpull84 -> cslpull84 2025-10-10T01:18:34.3127824Z * [new tag] cslpull85 -> cslpull85 2025-10-10T01:18:34.3129185Z * [new tag] cslpull86 -> cslpull86 2025-10-10T01:18:34.3130414Z * [new tag] cslpull87 -> cslpull87 2025-10-10T01:18:34.3131793Z * [new tag] cslpull88 -> cslpull88 2025-10-10T01:18:34.3133021Z * [new tag] cslpull89 -> cslpull89 2025-10-10T01:18:34.3134202Z * [new tag] cslpull90 -> cslpull90 2025-10-10T01:18:34.3135740Z * [new tag] cslpull91 -> cslpull91 2025-10-10T01:18:34.3136947Z * [new tag] cslpull92 -> cslpull92 2025-10-10T01:18:34.3138210Z * [new tag] flight_5 -> flight_5 2025-10-10T01:18:34.3139639Z * [new tag] flight_5.1 -> flight_5.1 2025-10-10T01:18:34.3140934Z * [new tag] flight_5.2 -> flight_5.2 2025-10-10T01:18:34.3142394Z * [new tag] flight_5.3 -> flight_5.3 2025-10-10T01:18:34.3143668Z * [new tag] forpull1 -> forpull1 2025-10-10T01:18:34.3145218Z * [new tag] malfet/tag-2ef5611 -> malfet/tag-2ef5611 2025-10-10T01:18:34.3146709Z * [new tag] malfet/tag-317b1a0 -> malfet/tag-317b1a0 2025-10-10T01:18:34.3147792Z * [new tag] malfet/tag-ec6f767 -> malfet/tag-ec6f767 2025-10-10T01:18:34.3149160Z * [new tag] nightly-binary -> nightly-binary 2025-10-10T01:18:34.3150312Z * [new tag] sqzhang_flight4_plus -> sqzhang_flight4_plus 2025-10-10T01:18:34.3151710Z * [new tag] sqzhang_flight_3 -> sqzhang_flight_3 2025-10-10T01:18:34.3153383Z * [new tag] trunk/001e1d263746ae9d121d9c8cf55bc87f777d9dba -> trunk/001e1d263746ae9d121d9c8cf55bc87f777d9dba 2025-10-10T01:18:34.3154739Z * [new tag] trunk/005c3d449e4c655d2eb0d76949a8cd41ce88f979 -> trunk/005c3d449e4c655d2eb0d76949a8cd41ce88f979 2025-10-10T01:18:34.3156283Z * [new tag] trunk/00f0365b959323bab89dc0a5bd5d40589e78edc8 -> trunk/00f0365b959323bab89dc0a5bd5d40589e78edc8 2025-10-10T01:18:34.3157862Z * [new tag] trunk/01f3a43462da594b65a6c9e8b46c132cd360cea9 -> trunk/01f3a43462da594b65a6c9e8b46c132cd360cea9 2025-10-10T01:18:34.3159326Z * [new tag] trunk/0319556a35b01e8857f7bf75df9df3287e1e853a -> trunk/0319556a35b01e8857f7bf75df9df3287e1e853a 2025-10-10T01:18:34.3160791Z * [new tag] trunk/054268c9ebb3291c6fd442e4a1f6602a8ea43ab6 -> trunk/054268c9ebb3291c6fd442e4a1f6602a8ea43ab6 2025-10-10T01:18:34.3162088Z * [new tag] trunk/06d86e58d0309aa2c217256f88d1990a22ec6e4f -> trunk/06d86e58d0309aa2c217256f88d1990a22ec6e4f 2025-10-10T01:18:34.3163398Z * [new tag] trunk/078d475d3bb104823e70ce975c2ee0d4d2fb0952 -> trunk/078d475d3bb104823e70ce975c2ee0d4d2fb0952 2025-10-10T01:18:34.3164766Z * [new tag] trunk/086dec3235d463e751c12ce9eeeb2dfcc873e206 -> trunk/086dec3235d463e751c12ce9eeeb2dfcc873e206 2025-10-10T01:18:34.3166766Z * [new tag] trunk/0a3e4e894cbc0cc93568c5d016f3ad72650cf641 -> trunk/0a3e4e894cbc0cc93568c5d016f3ad72650cf641 2025-10-10T01:18:34.3168168Z * [new tag] trunk/0b01ff4de02035eb21c1bc6bf4b1b627bc1cefaa -> trunk/0b01ff4de02035eb21c1bc6bf4b1b627bc1cefaa 2025-10-10T01:18:34.3169534Z * [new tag] trunk/0b15f7ae059cf4fa3909bbb009d83c0253a6385a -> trunk/0b15f7ae059cf4fa3909bbb009d83c0253a6385a 2025-10-10T01:18:34.3170943Z * [new tag] trunk/0b4f2b46d9e14c1858dd3d0ca9b62e349ae316cf -> trunk/0b4f2b46d9e14c1858dd3d0ca9b62e349ae316cf 2025-10-10T01:18:34.3172261Z * [new tag] trunk/0b85236477fe8a0e32510bcc973b2f34ef981df2 -> trunk/0b85236477fe8a0e32510bcc973b2f34ef981df2 2025-10-10T01:18:34.3173473Z * [new tag] trunk/0d39ecb2ce8556e85343d8da0c87450192c2fdf8 -> trunk/0d39ecb2ce8556e85343d8da0c87450192c2fdf8 2025-10-10T01:18:34.3174838Z * [new tag] trunk/0e5773b7fadef9e29b006af470b771fad55b5206 -> trunk/0e5773b7fadef9e29b006af470b771fad55b5206 2025-10-10T01:18:34.3176201Z * [new tag] trunk/0e9b3a772ab96e998ab85591d5b2a9c1d41bacb0 -> trunk/0e9b3a772ab96e998ab85591d5b2a9c1d41bacb0 2025-10-10T01:18:34.3177588Z * [new tag] trunk/0fbe3f19c7e88ee1720d2e1579e3fd2cafdaabf9 -> trunk/0fbe3f19c7e88ee1720d2e1579e3fd2cafdaabf9 2025-10-10T01:18:34.3179002Z * [new tag] trunk/0fd976b65c6daf3799a501d9202e4f50144446d1 -> trunk/0fd976b65c6daf3799a501d9202e4f50144446d1 2025-10-10T01:18:34.3180353Z * [new tag] trunk/1051c1de5c0c1d34bec94c4a3199ac7b23bb19e1 -> trunk/1051c1de5c0c1d34bec94c4a3199ac7b23bb19e1 2025-10-10T01:18:34.3181728Z * [new tag] trunk/115af42e9d57e89c26777be72822107cd7b39e07 -> trunk/115af42e9d57e89c26777be72822107cd7b39e07 2025-10-10T01:18:34.3183121Z * [new tag] trunk/11f5f656867089dac1fa1e64e34c9966578fbddd -> trunk/11f5f656867089dac1fa1e64e34c9966578fbddd 2025-10-10T01:18:34.3184511Z * [new tag] trunk/12d2ef557f6e127100267c31a31572d8ab5cc788 -> trunk/12d2ef557f6e127100267c31a31572d8ab5cc788 2025-10-10T01:18:34.3185853Z * [new tag] trunk/144378615a5a2b347e39c6376cba7d75f7a82926 -> trunk/144378615a5a2b347e39c6376cba7d75f7a82926 2025-10-10T01:18:34.3187211Z * [new tag] trunk/14791ea947349fb5fa7b7d6230cfd3924c36ba27 -> trunk/14791ea947349fb5fa7b7d6230cfd3924c36ba27 2025-10-10T01:18:34.3188244Z * [new tag] trunk/15800888b697bacd555399b3a0ca2e8d0827528e -> trunk/15800888b697bacd555399b3a0ca2e8d0827528e 2025-10-10T01:18:34.3189851Z * [new tag] trunk/15c8bdcc5e3a6dfd14e5c977438f772031e064ff -> trunk/15c8bdcc5e3a6dfd14e5c977438f772031e064ff 2025-10-10T01:18:34.3191376Z * [new tag] trunk/15d726005ddc5558c934c3edd5f815c2e504e501 -> trunk/15d726005ddc5558c934c3edd5f815c2e504e501 2025-10-10T01:18:34.3192725Z * [new tag] trunk/16f9bef642b07b3090a6e4a04517eff84d41a197 -> trunk/16f9bef642b07b3090a6e4a04517eff84d41a197 2025-10-10T01:18:34.3194187Z * [new tag] trunk/17c7170ca6e2efd5ead2b93bd12e226ff48f0669 -> trunk/17c7170ca6e2efd5ead2b93bd12e226ff48f0669 2025-10-10T01:18:34.3195611Z * [new tag] trunk/184817c7a81d5c01e107a84efeb269b063ddf5d6 -> trunk/184817c7a81d5c01e107a84efeb269b063ddf5d6 2025-10-10T01:18:34.3196983Z * [new tag] trunk/18940820006d2304460008575561e2e8e7fc59fc -> trunk/18940820006d2304460008575561e2e8e7fc59fc 2025-10-10T01:18:34.3198346Z * [new tag] trunk/18e18488e8c90e53cc113b1a5eddd9640ee80292 -> trunk/18e18488e8c90e53cc113b1a5eddd9640ee80292 2025-10-10T01:18:34.3199819Z * [new tag] trunk/1927783aa3ad676db6f4c34fc77ef3825a4e2ed5 -> trunk/1927783aa3ad676db6f4c34fc77ef3825a4e2ed5 2025-10-10T01:18:34.3201191Z * [new tag] trunk/19bf67be3286c0e2babe83af0d1593bae850362a -> trunk/19bf67be3286c0e2babe83af0d1593bae850362a 2025-10-10T01:18:34.3202489Z * [new tag] trunk/1bb68271b7ff1b582845384c6c7f7b1593ae1619 -> trunk/1bb68271b7ff1b582845384c6c7f7b1593ae1619 2025-10-10T01:18:34.3203813Z * [new tag] trunk/1d182dd81c3143697337e35d046fd02951dedb09 -> trunk/1d182dd81c3143697337e35d046fd02951dedb09 2025-10-10T01:18:34.3205138Z * [new tag] trunk/1e42fde45eff81845f269e8185f54a19f6d87c5b -> trunk/1e42fde45eff81845f269e8185f54a19f6d87c5b 2025-10-10T01:18:34.3206642Z * [new tag] trunk/1f73b96668bc6ae4c8e7ef5b630ff5f3c69ae005 -> trunk/1f73b96668bc6ae4c8e7ef5b630ff5f3c69ae005 2025-10-10T01:18:34.3207977Z * [new tag] trunk/1f8ee5da117952b03f0050a178d69f8e7189b0f8 -> trunk/1f8ee5da117952b03f0050a178d69f8e7189b0f8 2025-10-10T01:18:34.3209274Z * [new tag] trunk/1f9614cef8e0272c8e3bd99004d2978a6ecc5195 -> trunk/1f9614cef8e0272c8e3bd99004d2978a6ecc5195 2025-10-10T01:18:34.3210531Z * [new tag] trunk/1fb072ac2a33af93a77888dddbdd228b22a3f9c4 -> trunk/1fb072ac2a33af93a77888dddbdd228b22a3f9c4 2025-10-10T01:18:34.3211799Z * [new tag] trunk/1fc71d1b578badb1b3ba7cc2d5795f4f80463749 -> trunk/1fc71d1b578badb1b3ba7cc2d5795f4f80463749 2025-10-10T01:18:34.3213235Z * [new tag] trunk/20082d713666fa1eade588bebd523d86309bfa25 -> trunk/20082d713666fa1eade588bebd523d86309bfa25 2025-10-10T01:18:34.3214539Z * [new tag] trunk/2164b661219ab0a76aa018e955ba3d8e8f99c083 -> trunk/2164b661219ab0a76aa018e955ba3d8e8f99c083 2025-10-10T01:18:34.3215985Z * [new tag] trunk/228973df7f770505aafc6fc17b99f81ac58bdfe1 -> trunk/228973df7f770505aafc6fc17b99f81ac58bdfe1 2025-10-10T01:18:34.3217353Z * [new tag] trunk/22b1710252525d80d47ba95c762ccdbf577b2dc2 -> trunk/22b1710252525d80d47ba95c762ccdbf577b2dc2 2025-10-10T01:18:34.3218693Z * [new tag] trunk/22e219d9969ff3cee85bc5de32fa49d5a549a148 -> trunk/22e219d9969ff3cee85bc5de32fa49d5a549a148 2025-10-10T01:18:34.3220053Z * [new tag] trunk/235b995ce18de632ab816940319fcd66b46039b8 -> trunk/235b995ce18de632ab816940319fcd66b46039b8 2025-10-10T01:18:34.3221175Z * [new tag] trunk/23ab6a45e5c759fb4714905cb8c84ef74c70aa67 -> trunk/23ab6a45e5c759fb4714905cb8c84ef74c70aa67 2025-10-10T01:18:34.3222626Z * [new tag] trunk/24d69c57cbaa94cc828dbbdf83c889f5f244ae28 -> trunk/24d69c57cbaa94cc828dbbdf83c889f5f244ae28 2025-10-10T01:18:34.3224109Z * [new tag] trunk/263db92563f0ae71bf3e4fc265fbb48e79f9f23f -> trunk/263db92563f0ae71bf3e4fc265fbb48e79f9f23f 2025-10-10T01:18:34.3225665Z * [new tag] trunk/27234792add2ee9bedd84ca02dbf34f8f244bc5c -> trunk/27234792add2ee9bedd84ca02dbf34f8f244bc5c 2025-10-10T01:18:34.3227039Z * [new tag] trunk/27eb36debbe3fa2d43a2f893a5c46a6257a09460 -> trunk/27eb36debbe3fa2d43a2f893a5c46a6257a09460 2025-10-10T01:18:34.3228484Z * [new tag] trunk/2855a045b30dafad7a08d66e242be13770189c19 -> trunk/2855a045b30dafad7a08d66e242be13770189c19 2025-10-10T01:18:34.3229835Z * [new tag] trunk/2883b5ab773daf5861d43ff0b65be49a441ab3f9 -> trunk/2883b5ab773daf5861d43ff0b65be49a441ab3f9 2025-10-10T01:18:34.3231178Z * [new tag] trunk/29824067215f3ed9e4044ca0f31a71e9d95f237d -> trunk/29824067215f3ed9e4044ca0f31a71e9d95f237d 2025-10-10T01:18:34.3232679Z * [new tag] trunk/2a11ce2c787b2339ffb8941b849dd487d25b4121 -> trunk/2a11ce2c787b2339ffb8941b849dd487d25b4121 2025-10-10T01:18:34.3233831Z * [new tag] trunk/2a6cdba6e5f74c2294fecc2d1344537522efbaab -> trunk/2a6cdba6e5f74c2294fecc2d1344537522efbaab 2025-10-10T01:18:34.3236023Z * [new tag] trunk/2a760dc51e04d65845440cc09e7016cfc74f9132 -> trunk/2a760dc51e04d65845440cc09e7016cfc74f9132 2025-10-10T01:18:34.3237263Z * [new tag] trunk/2a7c48675010056f23d62b5c6ecb318782801723 -> trunk/2a7c48675010056f23d62b5c6ecb318782801723 2025-10-10T01:18:34.3238099Z * [new tag] trunk/2b58adc3bdcf9476e1cef49ad965b7d3c7b9ac24 -> trunk/2b58adc3bdcf9476e1cef49ad965b7d3c7b9ac24 2025-10-10T01:18:34.3239691Z * [new tag] trunk/2b9ff9953523a2e916234c9197d946f4cff976c7 -> trunk/2b9ff9953523a2e916234c9197d946f4cff976c7 2025-10-10T01:18:34.3240970Z * [new tag] trunk/2c2e1268b7aae8ed610d12f2d38d39f8d93888a3 -> trunk/2c2e1268b7aae8ed610d12f2d38d39f8d93888a3 2025-10-10T01:18:34.3242312Z * [new tag] trunk/2c5ed6e7c067573b093725cd15d13812d9647562 -> trunk/2c5ed6e7c067573b093725cd15d13812d9647562 2025-10-10T01:18:34.3243722Z * [new tag] trunk/2d50678dcc7ab2da13a9bca6af8f2333e8970344 -> trunk/2d50678dcc7ab2da13a9bca6af8f2333e8970344 2025-10-10T01:18:34.3245105Z * [new tag] trunk/2e027e874232fefe7b1c56ce8aeb26c0e6b97f15 -> trunk/2e027e874232fefe7b1c56ce8aeb26c0e6b97f15 2025-10-10T01:18:34.3246477Z * [new tag] trunk/2e1742dd63c2168fd9649dbba96a95abf1f57cae -> trunk/2e1742dd63c2168fd9649dbba96a95abf1f57cae 2025-10-10T01:18:34.3247830Z * [new tag] trunk/2fe37b5fde392535a3238f975c93dd202cd3e24b -> trunk/2fe37b5fde392535a3238f975c93dd202cd3e24b 2025-10-10T01:18:34.3249178Z * [new tag] trunk/3040a5d294bd30d3938d0043a5d93d6c23264827 -> trunk/3040a5d294bd30d3938d0043a5d93d6c23264827 2025-10-10T01:18:34.3250560Z * [new tag] trunk/321e6026925f6b6e8a36e3a8b7c0295cd7541911 -> trunk/321e6026925f6b6e8a36e3a8b7c0295cd7541911 2025-10-10T01:18:34.3251811Z * [new tag] trunk/322091d8d8542a0cbff524306029bef4d7338747 -> trunk/322091d8d8542a0cbff524306029bef4d7338747 2025-10-10T01:18:34.3253108Z * [new tag] trunk/3288fbf374128610928e27d03615ac0d46a6ce14 -> trunk/3288fbf374128610928e27d03615ac0d46a6ce14 2025-10-10T01:18:34.3254980Z * [new tag] trunk/331191ce4b29b5d7d3bb7f0e7454ca70c06fbd26 -> trunk/331191ce4b29b5d7d3bb7f0e7454ca70c06fbd26 2025-10-10T01:18:34.3256315Z * [new tag] trunk/33b17bc619b044a0050797987efb8890d43319df -> trunk/33b17bc619b044a0050797987efb8890d43319df 2025-10-10T01:18:34.3257736Z * [new tag] trunk/34042a9145fe28033e7edb08f1fcf90ed197f4ac -> trunk/34042a9145fe28033e7edb08f1fcf90ed197f4ac 2025-10-10T01:18:34.3259233Z * [new tag] trunk/344e6365a0068c2d2847fcec0c55dd53291d475e -> trunk/344e6365a0068c2d2847fcec0c55dd53291d475e 2025-10-10T01:18:34.3260482Z * [new tag] trunk/34ac9b61cbfcf17328ccb8b729509829447fdddd -> trunk/34ac9b61cbfcf17328ccb8b729509829447fdddd 2025-10-10T01:18:34.3261816Z * [new tag] trunk/35c4130fd1358c98e12301ffa0f1b2294e0c795f -> trunk/35c4130fd1358c98e12301ffa0f1b2294e0c795f 2025-10-10T01:18:34.3263135Z * [new tag] trunk/35f66b83f89a571d0c0abe16c66a23120b92bdaf -> trunk/35f66b83f89a571d0c0abe16c66a23120b92bdaf 2025-10-10T01:18:34.3264446Z * [new tag] trunk/361c5d362c4ea1950e05116899cfcf753c345ebd -> trunk/361c5d362c4ea1950e05116899cfcf753c345ebd 2025-10-10T01:18:34.3265928Z * [new tag] trunk/37c6087334cce3ad4bc9838ea2ef63aba89f2253 -> trunk/37c6087334cce3ad4bc9838ea2ef63aba89f2253 2025-10-10T01:18:34.3267174Z * [new tag] trunk/3912ba3e940b9354622fa09b2ada677cd10723d8 -> trunk/3912ba3e940b9354622fa09b2ada677cd10723d8 2025-10-10T01:18:34.3268535Z * [new tag] trunk/39189592fd688979e56063430ed5a038d999908f -> trunk/39189592fd688979e56063430ed5a038d999908f 2025-10-10T01:18:34.3269902Z * [new tag] trunk/3924f784ba81f87fe09988d6fc9620b57e4d9f72 -> trunk/3924f784ba81f87fe09988d6fc9620b57e4d9f72 2025-10-10T01:18:34.3271233Z * [new tag] trunk/39b31a6bfde6e046383ae2b06fe0b68df5cdbdd2 -> trunk/39b31a6bfde6e046383ae2b06fe0b68df5cdbdd2 2025-10-10T01:18:34.3272510Z * [new tag] trunk/39c340ec9e2ee3011f1d260f581b5a95f3c99039 -> trunk/39c340ec9e2ee3011f1d260f581b5a95f3c99039 2025-10-10T01:18:34.3273849Z * [new tag] trunk/39d0c06ed0d7bc634d7f1a4e84b69f66d1ea0798 -> trunk/39d0c06ed0d7bc634d7f1a4e84b69f66d1ea0798 2025-10-10T01:18:34.3275206Z * [new tag] trunk/3c0577bd15778c96cecf0e7a5e5958d7fcab64f0 -> trunk/3c0577bd15778c96cecf0e7a5e5958d7fcab64f0 2025-10-10T01:18:34.3276636Z * [new tag] trunk/3c59351c6ea2fc29d346903e28e95c5f4d0ccdbb -> trunk/3c59351c6ea2fc29d346903e28e95c5f4d0ccdbb 2025-10-10T01:18:34.3277987Z * [new tag] trunk/3c5ca685d6f5b6f3971c0cd20a054aa355610419 -> trunk/3c5ca685d6f5b6f3971c0cd20a054aa355610419 2025-10-10T01:18:34.3279496Z * [new tag] trunk/3ca09d65f1bdf83142dc9fe47976227ae4a88e7b -> trunk/3ca09d65f1bdf83142dc9fe47976227ae4a88e7b 2025-10-10T01:18:34.3281064Z * [new tag] trunk/3cc8af2d67f42bf2a933796290446c5ab8978aac -> trunk/3cc8af2d67f42bf2a933796290446c5ab8978aac 2025-10-10T01:18:34.3282439Z * [new tag] trunk/3d1fa40ae1fee18ddf3dca89229e3ae828589e0c -> trunk/3d1fa40ae1fee18ddf3dca89229e3ae828589e0c 2025-10-10T01:18:34.3283788Z * [new tag] trunk/3d9d41c80168bcd3c569345a96682c42a5eba36a -> trunk/3d9d41c80168bcd3c569345a96682c42a5eba36a 2025-10-10T01:18:34.3285080Z * [new tag] trunk/3db21643417a04f6f2707a783ac32a538a98d53d -> trunk/3db21643417a04f6f2707a783ac32a538a98d53d 2025-10-10T01:18:34.3286519Z * [new tag] trunk/3ddf2018d0b7b4def0553dc092d928ef831a19c3 -> trunk/3ddf2018d0b7b4def0553dc092d928ef831a19c3 2025-10-10T01:18:34.3287830Z * [new tag] trunk/3e03deab6f3c268c85c8efd9546e28cdda0fa4cc -> trunk/3e03deab6f3c268c85c8efd9546e28cdda0fa4cc 2025-10-10T01:18:34.3289189Z * [new tag] trunk/3e0826c9d792ae87373dc0ff5d46c260020de29f -> trunk/3e0826c9d792ae87373dc0ff5d46c260020de29f 2025-10-10T01:18:34.3290577Z * [new tag] trunk/409aece3f9436a2f740f1b97f1243f738f6bbbf6 -> trunk/409aece3f9436a2f740f1b97f1243f738f6bbbf6 2025-10-10T01:18:34.3291984Z * [new tag] trunk/40b25578e4ecb7ef1c38201b3ce0014eb57c53eb -> trunk/40b25578e4ecb7ef1c38201b3ce0014eb57c53eb 2025-10-10T01:18:34.3293208Z * [new tag] trunk/412c6d28ec3869ef8ba962b290d755251e7cc3c1 -> trunk/412c6d28ec3869ef8ba962b290d755251e7cc3c1 2025-10-10T01:18:34.3294495Z * [new tag] trunk/415e641572473479fc9d9eaea12762e1a223a9e0 -> trunk/415e641572473479fc9d9eaea12762e1a223a9e0 2025-10-10T01:18:34.3295800Z * [new tag] trunk/41808b2ba9a61ab2f4c7af394c1668d09a4a0331 -> trunk/41808b2ba9a61ab2f4c7af394c1668d09a4a0331 2025-10-10T01:18:34.3297181Z * [new tag] trunk/4308b8a28fa332d23ad6d25a472559b354619131 -> trunk/4308b8a28fa332d23ad6d25a472559b354619131 2025-10-10T01:18:34.3298484Z * [new tag] trunk/43848b71d9af0223eafdd1755bf7444aafe9e993 -> trunk/43848b71d9af0223eafdd1755bf7444aafe9e993 2025-10-10T01:18:34.3299829Z * [new tag] trunk/43fc859625f9c0a794307b3ef30c26ab3fc2bfec -> trunk/43fc859625f9c0a794307b3ef30c26ab3fc2bfec 2025-10-10T01:18:34.3301226Z * [new tag] trunk/4412026949b562f940d4c24162de19d299725b62 -> trunk/4412026949b562f940d4c24162de19d299725b62 2025-10-10T01:18:34.3302628Z * [new tag] trunk/44a5d419935a77b3308f247279a457e6d0b9a292 -> trunk/44a5d419935a77b3308f247279a457e6d0b9a292 2025-10-10T01:18:34.3304045Z * [new tag] trunk/4661200125ba9c87aa7d54a55e585403b5ce5040 -> trunk/4661200125ba9c87aa7d54a55e585403b5ce5040 2025-10-10T01:18:34.3305388Z * [new tag] trunk/4691fe60700ac51a878775fd23a8f7c4548c6757 -> trunk/4691fe60700ac51a878775fd23a8f7c4548c6757 2025-10-10T01:18:34.3306856Z * [new tag] trunk/4725871a815fb880e89135a493c8c94ab9bbfece -> trunk/4725871a815fb880e89135a493c8c94ab9bbfece 2025-10-10T01:18:34.3308202Z * [new tag] trunk/47956196d99166fe9083beb2a52fd2e6c90b2011 -> trunk/47956196d99166fe9083beb2a52fd2e6c90b2011 2025-10-10T01:18:34.3309542Z * [new tag] trunk/483f4e0db91166128ad8922d86dc7222338d4ecc -> trunk/483f4e0db91166128ad8922d86dc7222338d4ecc 2025-10-10T01:18:34.3310920Z * [new tag] trunk/48b54b45d62af7ecafccc5afede04474cb236f1a -> trunk/48b54b45d62af7ecafccc5afede04474cb236f1a 2025-10-10T01:18:34.3312322Z * [new tag] trunk/49f7d8d19d24f616b11ef050535a211245aed649 -> trunk/49f7d8d19d24f616b11ef050535a211245aed649 2025-10-10T01:18:34.3313618Z * [new tag] trunk/4a0df39f814afad087e8b29dd2914a8b54567694 -> trunk/4a0df39f814afad087e8b29dd2914a8b54567694 2025-10-10T01:18:34.3314870Z * [new tag] trunk/4a6abba0d9fb3dc0f29b5efe527e26b2962caec1 -> trunk/4a6abba0d9fb3dc0f29b5efe527e26b2962caec1 2025-10-10T01:18:34.3316302Z * [new tag] trunk/4ab847bbc7ba09f29a4e81494e8a752dcb411117 -> trunk/4ab847bbc7ba09f29a4e81494e8a752dcb411117 2025-10-10T01:18:34.3317561Z * [new tag] trunk/4bcc05777e780e834d44a2d06dd5321daec316f0 -> trunk/4bcc05777e780e834d44a2d06dd5321daec316f0 2025-10-10T01:18:34.3318906Z * [new tag] trunk/4bd1505f849e701a8e54f9d185c23f13e7324498 -> trunk/4bd1505f849e701a8e54f9d185c23f13e7324498 2025-10-10T01:18:34.3320247Z * [new tag] trunk/4c0fec3e4dac35b9e9dec2beacfb5967906a4701 -> trunk/4c0fec3e4dac35b9e9dec2beacfb5967906a4701 2025-10-10T01:18:34.3321563Z * [new tag] trunk/4c3c0ef2f10415f5d5b13f1c842f91bb90ee91d3 -> trunk/4c3c0ef2f10415f5d5b13f1c842f91bb90ee91d3 2025-10-10T01:18:34.3323012Z * [new tag] trunk/4d7f9f3aed68729380730ed46e29ff2052f05b73 -> trunk/4d7f9f3aed68729380730ed46e29ff2052f05b73 2025-10-10T01:18:34.3324499Z * [new tag] trunk/50e077beaaf71798f870552f3849e4a52c784df5 -> trunk/50e077beaaf71798f870552f3849e4a52c784df5 2025-10-10T01:18:34.3327152Z * [new tag] trunk/5103ecc5d8f0cc90e686763652e2d84c22d83ca9 -> trunk/5103ecc5d8f0cc90e686763652e2d84c22d83ca9 2025-10-10T01:18:34.3328779Z * [new tag] trunk/5178d0a480f8f4e21da3757de455c8215b249ec5 -> trunk/5178d0a480f8f4e21da3757de455c8215b249ec5 2025-10-10T01:18:34.3330240Z * [new tag] trunk/5209c8ce0704f34ba4bd2a58c19877fbf6cf0392 -> trunk/5209c8ce0704f34ba4bd2a58c19877fbf6cf0392 2025-10-10T01:18:34.3331809Z * [new tag] trunk/5390324984c43f1214b8abf731ad495ba2df5341 -> trunk/5390324984c43f1214b8abf731ad495ba2df5341 2025-10-10T01:18:34.3333049Z * [new tag] trunk/53f5af8c924aba3c0fab1fabc6baf7d6affcb8a1 -> trunk/53f5af8c924aba3c0fab1fabc6baf7d6affcb8a1 2025-10-10T01:18:34.3334493Z * [new tag] trunk/54ae61c573e91fa2a2c6430435059e2d94ecba2e -> trunk/54ae61c573e91fa2a2c6430435059e2d94ecba2e 2025-10-10T01:18:34.3335881Z * [new tag] trunk/5656d45c8ff03cf20fd7d5098247c2250395af8a -> trunk/5656d45c8ff03cf20fd7d5098247c2250395af8a 2025-10-10T01:18:34.3337256Z * [new tag] trunk/56d66ac0d74f44d7b656757795142b5b9a1802a1 -> trunk/56d66ac0d74f44d7b656757795142b5b9a1802a1 2025-10-10T01:18:34.3338605Z * [new tag] trunk/5743d731c1de495ecf3bb03682a2dcbe207ca895 -> trunk/5743d731c1de495ecf3bb03682a2dcbe207ca895 2025-10-10T01:18:34.3340092Z * [new tag] trunk/5a1fbf45ad727353e367740ecd8825ca7ee857e9 -> trunk/5a1fbf45ad727353e367740ecd8825ca7ee857e9 2025-10-10T01:18:34.3341404Z * [new tag] trunk/5a66ff4915ecfd86f1a68e7862e5a2ad473e5a79 -> trunk/5a66ff4915ecfd86f1a68e7862e5a2ad473e5a79 2025-10-10T01:18:34.3342813Z * [new tag] trunk/5b0b4cda4aa03bee16ee67d9d36012a539df3c50 -> trunk/5b0b4cda4aa03bee16ee67d9d36012a539df3c50 2025-10-10T01:18:34.3344598Z * [new tag] trunk/5b8174bc286725f9326fba6dc0ef17c316486bbd -> trunk/5b8174bc286725f9326fba6dc0ef17c316486bbd 2025-10-10T01:18:34.3345968Z * [new tag] trunk/5ba11df4f871717818b88c4eab514d31286601d1 -> trunk/5ba11df4f871717818b88c4eab514d31286601d1 2025-10-10T01:18:34.3346932Z * [new tag] trunk/5c827a4133da69108338d0363bb7ad7f62803c40 -> trunk/5c827a4133da69108338d0363bb7ad7f62803c40 2025-10-10T01:18:34.3348487Z * [new tag] trunk/5d459dd6099ef94d33db9a6d36bcce9f742f1da1 -> trunk/5d459dd6099ef94d33db9a6d36bcce9f742f1da1 2025-10-10T01:18:34.3349659Z * [new tag] trunk/5d7360bb03355c89c0b956df0ab428f5a7b5c9f8 -> trunk/5d7360bb03355c89c0b956df0ab428f5a7b5c9f8 2025-10-10T01:18:34.3351462Z * [new tag] trunk/5dbae1eae26159058f6199fc68fe73fc0e5bef5f -> trunk/5dbae1eae26159058f6199fc68fe73fc0e5bef5f 2025-10-10T01:18:34.3352673Z * [new tag] trunk/5e47b4dd60ff9efb253286af5a2479d9d800ce6a -> trunk/5e47b4dd60ff9efb253286af5a2479d9d800ce6a 2025-10-10T01:18:34.3354183Z * [new tag] trunk/5ed4270440fd0b62d3aa14692f9e377a0061061e -> trunk/5ed4270440fd0b62d3aa14692f9e377a0061061e 2025-10-10T01:18:34.3355623Z * [new tag] trunk/5f18f240de43fc24481ead4d740dda64f174fa86 -> trunk/5f18f240de43fc24481ead4d740dda64f174fa86 2025-10-10T01:18:34.3357031Z * [new tag] trunk/5f775bdfb766d9a2717ffbb64f2a51e53cddc778 -> trunk/5f775bdfb766d9a2717ffbb64f2a51e53cddc778 2025-10-10T01:18:34.3358209Z * [new tag] trunk/600267ea56cafcf8f9a1150a4379184960a757b2 -> trunk/600267ea56cafcf8f9a1150a4379184960a757b2 2025-10-10T01:18:34.3359805Z * [new tag] trunk/600db525bdb5e76c12f30f271d969d43a7f8efef -> trunk/600db525bdb5e76c12f30f271d969d43a7f8efef 2025-10-10T01:18:34.3361359Z * [new tag] trunk/608792153f42254d2d2b5a87d524807a0c2724f1 -> trunk/608792153f42254d2d2b5a87d524807a0c2724f1 2025-10-10T01:18:34.3362788Z * [new tag] trunk/6389658ec6b1ea58cb1de032266d865eeb8d48e9 -> trunk/6389658ec6b1ea58cb1de032266d865eeb8d48e9 2025-10-10T01:18:34.3364194Z * [new tag] trunk/64108bdbed2f099d527060b4c9fdd5a11cad2afc -> trunk/64108bdbed2f099d527060b4c9fdd5a11cad2afc 2025-10-10T01:18:34.3365503Z * [new tag] trunk/65aa62d50d1c83aa1b46ed4d584f12f509bab1c4 -> trunk/65aa62d50d1c83aa1b46ed4d584f12f509bab1c4 2025-10-10T01:18:34.3366878Z * [new tag] trunk/65f10becdf21f3a0947a735904fcce876ce3c4b0 -> trunk/65f10becdf21f3a0947a735904fcce876ce3c4b0 2025-10-10T01:18:34.3368293Z * [new tag] trunk/660e369a68dd8be60ce4eb67c25191ea66efc303 -> trunk/660e369a68dd8be60ce4eb67c25191ea66efc303 2025-10-10T01:18:34.3369862Z * [new tag] trunk/68350660ee2db8c21c84527929b92de9f0bcc3e2 -> trunk/68350660ee2db8c21c84527929b92de9f0bcc3e2 2025-10-10T01:18:34.3371192Z * [new tag] trunk/6861a270624b44954826688f8dad668eb0154452 -> trunk/6861a270624b44954826688f8dad668eb0154452 2025-10-10T01:18:34.3372603Z * [new tag] trunk/6861fa43e5fee7fedc0213e352fa983edea8aa78 -> trunk/6861fa43e5fee7fedc0213e352fa983edea8aa78 2025-10-10T01:18:34.3373958Z * [new tag] trunk/688efd9741dbd18c176729aec3df7a73825f8463 -> trunk/688efd9741dbd18c176729aec3df7a73825f8463 2025-10-10T01:18:34.3375318Z * [new tag] trunk/6a09f9306cadd003b2e6abc3f6422a2d8607779b -> trunk/6a09f9306cadd003b2e6abc3f6422a2d8607779b 2025-10-10T01:18:34.3377008Z * [new tag] trunk/6a31f42da45c0f1cbdb021b3695f0e6388b8b532 -> trunk/6a31f42da45c0f1cbdb021b3695f0e6388b8b532 2025-10-10T01:18:34.3378455Z * [new tag] trunk/6a7f5c0d21a22959d014c8b06f3efe3408336aaf -> trunk/6a7f5c0d21a22959d014c8b06f3efe3408336aaf 2025-10-10T01:18:34.3379968Z * [new tag] trunk/6b768e1890a179122e91395c5532a382d69b96a0 -> trunk/6b768e1890a179122e91395c5532a382d69b96a0 2025-10-10T01:18:34.3381380Z * [new tag] trunk/6b7970192f5de47d29a4fe085f509389ac0bea7d -> trunk/6b7970192f5de47d29a4fe085f509389ac0bea7d 2025-10-10T01:18:34.3382816Z * [new tag] trunk/6bb021c12553755a4f64df0b60dc34b1efdb992b -> trunk/6bb021c12553755a4f64df0b60dc34b1efdb992b 2025-10-10T01:18:34.3384254Z * [new tag] trunk/6bb586eafd723d4972c729f37c14f27c88168adc -> trunk/6bb586eafd723d4972c729f37c14f27c88168adc 2025-10-10T01:18:34.3385610Z * [new tag] trunk/6c0125dbc0241aef962528651df4f67204a8b526 -> trunk/6c0125dbc0241aef962528651df4f67204a8b526 2025-10-10T01:18:34.3386970Z * [new tag] trunk/6c209bfc5c1e1e59e6a62f94151398d66164bb93 -> trunk/6c209bfc5c1e1e59e6a62f94151398d66164bb93 2025-10-10T01:18:34.3388405Z * [new tag] trunk/6c3c9414eb571b34ff0d932978e4733dbb08dc1d -> trunk/6c3c9414eb571b34ff0d932978e4733dbb08dc1d 2025-10-10T01:18:34.3389802Z * [new tag] trunk/6d27a8e5093ee2a21d44dceeeffcb272e6e0f655 -> trunk/6d27a8e5093ee2a21d44dceeeffcb272e6e0f655 2025-10-10T01:18:34.3391183Z * [new tag] trunk/702f6e703b1d3a942346848b65a9f2a37d12ae18 -> trunk/702f6e703b1d3a942346848b65a9f2a37d12ae18 2025-10-10T01:18:34.3392604Z * [new tag] trunk/7158aa22e8dc97fdc2657cf0d4cde34b277e7d9e -> trunk/7158aa22e8dc97fdc2657cf0d4cde34b277e7d9e 2025-10-10T01:18:34.3394018Z * [new tag] trunk/71aefd5595834dd97f38aa978ee32abbd13ac3d6 -> trunk/71aefd5595834dd97f38aa978ee32abbd13ac3d6 2025-10-10T01:18:34.3395398Z * [new tag] trunk/724463d5a2fba369cd14e89215b84d1b01435df7 -> trunk/724463d5a2fba369cd14e89215b84d1b01435df7 2025-10-10T01:18:34.3396703Z * [new tag] trunk/73adac05d13babb75410c3e033fdce57aa16881a -> trunk/73adac05d13babb75410c3e033fdce57aa16881a 2025-10-10T01:18:34.3398163Z * [new tag] trunk/7457d139c51124e5a31a6173f99f81f0deb52178 -> trunk/7457d139c51124e5a31a6173f99f81f0deb52178 2025-10-10T01:18:34.3399735Z * [new tag] trunk/746fe78ecd52f3e9cfddda41f0ac82dada7bdd0b -> trunk/746fe78ecd52f3e9cfddda41f0ac82dada7bdd0b 2025-10-10T01:18:34.3400796Z * [new tag] trunk/7617b113ad0045cdfe5cf1feb8efb634a41c6ce2 -> trunk/7617b113ad0045cdfe5cf1feb8efb634a41c6ce2 2025-10-10T01:18:34.3402374Z * [new tag] trunk/7a1ead755f2e2abe8be49a7a0fb88b6b13973147 -> trunk/7a1ead755f2e2abe8be49a7a0fb88b6b13973147 2025-10-10T01:18:34.3403742Z * [new tag] trunk/7b15534434aeaf59a4c9189f52b4ebd4a5d58803 -> trunk/7b15534434aeaf59a4c9189f52b4ebd4a5d58803 2025-10-10T01:18:34.3405186Z * [new tag] trunk/7b691546d2949790ffc8f6bd3c674faa6a46ff7c -> trunk/7b691546d2949790ffc8f6bd3c674faa6a46ff7c 2025-10-10T01:18:34.3406726Z * [new tag] trunk/7cfecd76b2141d81c90d722dc5e3262bdf7ea900 -> trunk/7cfecd76b2141d81c90d722dc5e3262bdf7ea900 2025-10-10T01:18:34.3408087Z * [new tag] trunk/7d570129e0cea8dd3de0175baff96723656ab8ab -> trunk/7d570129e0cea8dd3de0175baff96723656ab8ab 2025-10-10T01:18:34.3409480Z * [new tag] trunk/7e7ac2039d5d5f35373c4de6cdf0ccdee3734c7a -> trunk/7e7ac2039d5d5f35373c4de6cdf0ccdee3734c7a 2025-10-10T01:18:34.3410674Z * [new tag] trunk/7eb1eb4313cfa3db1beadc6d9d04ea6b76acc39c -> trunk/7eb1eb4313cfa3db1beadc6d9d04ea6b76acc39c 2025-10-10T01:18:34.3411982Z * [new tag] trunk/801e282f39e9ef4424dfd3ecfd2b550a44595229 -> trunk/801e282f39e9ef4424dfd3ecfd2b550a44595229 2025-10-10T01:18:34.3413364Z * [new tag] trunk/81994b08a078b30e076d408713f78c9bf4e329e7 -> trunk/81994b08a078b30e076d408713f78c9bf4e329e7 2025-10-10T01:18:34.3414660Z * [new tag] trunk/81dbeb06f4b3eb6c56625ec25d377eb7c7c6c573 -> trunk/81dbeb06f4b3eb6c56625ec25d377eb7c7c6c573 2025-10-10T01:18:34.3416071Z * [new tag] trunk/83458197d14921f797565135f0f45031c362338d -> trunk/83458197d14921f797565135f0f45031c362338d 2025-10-10T01:18:34.3417484Z * [new tag] trunk/83d71dfb2fd993a6242372b8123549acaa85ffdb -> trunk/83d71dfb2fd993a6242372b8123549acaa85ffdb 2025-10-10T01:18:34.3418832Z * [new tag] trunk/86474ce996d168b404592cbbdfcc30d6607c8bd4 -> trunk/86474ce996d168b404592cbbdfcc30d6607c8bd4 2025-10-10T01:18:34.3420322Z * [new tag] trunk/86c789849eac1f96d03cf273e7995dbc7d319c26 -> trunk/86c789849eac1f96d03cf273e7995dbc7d319c26 2025-10-10T01:18:34.3421961Z * [new tag] trunk/874efa2d72d83b00894097130f18062ce331a265 -> trunk/874efa2d72d83b00894097130f18062ce331a265 2025-10-10T01:18:34.3423453Z * [new tag] trunk/87c9fbda22c229d4e5512011e050efd6ffea1241 -> trunk/87c9fbda22c229d4e5512011e050efd6ffea1241 2025-10-10T01:18:34.3424846Z * [new tag] trunk/87eccf10e8484c9e59ef81ae7bdee68d3db4f605 -> trunk/87eccf10e8484c9e59ef81ae7bdee68d3db4f605 2025-10-10T01:18:34.3426392Z * [new tag] trunk/8c0bc879b97bc580aaa0777b2d266bdd068cb528 -> trunk/8c0bc879b97bc580aaa0777b2d266bdd068cb528 2025-10-10T01:18:34.3427830Z * [new tag] trunk/8c54101933bb7c6ed3f9c1a65629b7f30376f7e2 -> trunk/8c54101933bb7c6ed3f9c1a65629b7f30376f7e2 2025-10-10T01:18:34.3429246Z * [new tag] trunk/8ca986ee60febce075f9e3ff83726048cebbbf68 -> trunk/8ca986ee60febce075f9e3ff83726048cebbbf68 2025-10-10T01:18:34.3430534Z * [new tag] trunk/8d53d788fefc0370931063d91f0c342556c3cf4c -> trunk/8d53d788fefc0370931063d91f0c342556c3cf4c 2025-10-10T01:18:34.3432604Z * [new tag] trunk/8e1f409b8ccf64b2cf3933ece13587ad57e9d8a9 -> trunk/8e1f409b8ccf64b2cf3933ece13587ad57e9d8a9 2025-10-10T01:18:34.3434126Z * [new tag] trunk/8ec8c14aced9f3e7ff4ab663822bed792d6c34f4 -> trunk/8ec8c14aced9f3e7ff4ab663822bed792d6c34f4 2025-10-10T01:18:34.3435978Z * [new tag] trunk/8f54e27e5decf41222f5d744069eb6572dbf275f -> trunk/8f54e27e5decf41222f5d744069eb6572dbf275f 2025-10-10T01:18:34.3437482Z * [new tag] trunk/8f705d019a64b1ca882e043b3eb98559273a9e59 -> trunk/8f705d019a64b1ca882e043b3eb98559273a9e59 2025-10-10T01:18:34.3439030Z * [new tag] trunk/8f83b3e71cb2af6244971af59bfbb6e2abb55f24 -> trunk/8f83b3e71cb2af6244971af59bfbb6e2abb55f24 2025-10-10T01:18:34.3440481Z * [new tag] trunk/90b4e130d6871bee4e1f15bb8294c1bbbf8f4ba5 -> trunk/90b4e130d6871bee4e1f15bb8294c1bbbf8f4ba5 2025-10-10T01:18:34.3441526Z * [new tag] trunk/90c0825e2deb0a46faf5cc2deb7184f6f8ea7a6d -> trunk/90c0825e2deb0a46faf5cc2deb7184f6f8ea7a6d 2025-10-10T01:18:34.3443179Z * [new tag] trunk/91040f49348646d79c6cd3434c34860d25c2e47a -> trunk/91040f49348646d79c6cd3434c34860d25c2e47a 2025-10-10T01:18:34.3444419Z * [new tag] trunk/91b94842645c1a781ab169b0df718545901ebb01 -> trunk/91b94842645c1a781ab169b0df718545901ebb01 2025-10-10T01:18:34.3445936Z * [new tag] trunk/91c211fb8c8ec3065be2a18dfc399ce849ea83bf -> trunk/91c211fb8c8ec3065be2a18dfc399ce849ea83bf 2025-10-10T01:18:34.3446988Z * [new tag] trunk/91c4db76cbb82dfa46d937b8dce4c942eaf5e226 -> trunk/91c4db76cbb82dfa46d937b8dce4c942eaf5e226 2025-10-10T01:18:34.3448645Z * [new tag] trunk/93e833de0f987f66d8c93b76ffe6aad35b714231 -> trunk/93e833de0f987f66d8c93b76ffe6aad35b714231 2025-10-10T01:18:34.3449963Z * [new tag] trunk/94b1ec8c7c5cc63541325abc923973f2fc2ad094 -> trunk/94b1ec8c7c5cc63541325abc923973f2fc2ad094 2025-10-10T01:18:34.3451366Z * [new tag] trunk/955f21dc2c628e09e0d112b3db1ee928cd1da344 -> trunk/955f21dc2c628e09e0d112b3db1ee928cd1da344 2025-10-10T01:18:34.3452853Z * [new tag] trunk/9580539e2f73d68e89544c713ff460bea3038701 -> trunk/9580539e2f73d68e89544c713ff460bea3038701 2025-10-10T01:18:34.3454328Z * [new tag] trunk/95a053284cd28e8d52bd55049bd45aea47adba0c -> trunk/95a053284cd28e8d52bd55049bd45aea47adba0c 2025-10-10T01:18:34.3455782Z * [new tag] trunk/960c4b9937251da01ea588efff0fc06a34eac35b -> trunk/960c4b9937251da01ea588efff0fc06a34eac35b 2025-10-10T01:18:34.3456916Z * [new tag] trunk/96181d6f7619acf938dc743123326c6b5dd25284 -> trunk/96181d6f7619acf938dc743123326c6b5dd25284 2025-10-10T01:18:34.3458533Z * [new tag] trunk/9697a7ce9ea095e933658cfee13f9bbef272551a -> trunk/9697a7ce9ea095e933658cfee13f9bbef272551a 2025-10-10T01:18:34.3459968Z * [new tag] trunk/96d91da792d4b50930318ecdfb8b5b8190c467cd -> trunk/96d91da792d4b50930318ecdfb8b5b8190c467cd 2025-10-10T01:18:34.3461334Z * [new tag] trunk/97463d4cf3c125557ef23502772b12a67dac4dc7 -> trunk/97463d4cf3c125557ef23502772b12a67dac4dc7 2025-10-10T01:18:34.3462735Z * [new tag] trunk/97ca21106d0179f425fc752ec867fe11669c2834 -> trunk/97ca21106d0179f425fc752ec867fe11669c2834 2025-10-10T01:18:34.3464169Z * [new tag] trunk/98a081a24c22072362dc536afd39a469e28939d4 -> trunk/98a081a24c22072362dc536afd39a469e28939d4 2025-10-10T01:18:34.3465586Z * [new tag] trunk/9944cac6e6a95159744a775a8bef40d89eef0f03 -> trunk/9944cac6e6a95159744a775a8bef40d89eef0f03 2025-10-10T01:18:34.3466934Z * [new tag] trunk/9aa92f246fa5fe5cfda17970d41d167b19a0612a -> trunk/9aa92f246fa5fe5cfda17970d41d167b19a0612a 2025-10-10T01:18:34.3468395Z * [new tag] trunk/9d1ab4f4bb508a72c7f549f0b5219c4601944ba1 -> trunk/9d1ab4f4bb508a72c7f549f0b5219c4601944ba1 2025-10-10T01:18:34.3469804Z * [new tag] trunk/9eb89a4ad5965b97c54e498d71fc765c0059acef -> trunk/9eb89a4ad5965b97c54e498d71fc765c0059acef 2025-10-10T01:18:34.3471150Z * [new tag] trunk/9ec10dc26a81dc618ff435edd4ca4819245ecb0f -> trunk/9ec10dc26a81dc618ff435edd4ca4819245ecb0f 2025-10-10T01:18:34.3472612Z * [new tag] trunk/9ecd092bd98f43d1cd4acc88eed6cbc39e946dbe -> trunk/9ecd092bd98f43d1cd4acc88eed6cbc39e946dbe 2025-10-10T01:18:34.3473773Z * [new tag] trunk/9f5e1beaf3c9248a335d2448103240a463187eb5 -> trunk/9f5e1beaf3c9248a335d2448103240a463187eb5 2025-10-10T01:18:34.3475328Z * [new tag] trunk/9fc2c6446d394dd313ed71e9d1ffc4f7f3916423 -> trunk/9fc2c6446d394dd313ed71e9d1ffc4f7f3916423 2025-10-10T01:18:34.3476788Z * [new tag] trunk/9fff8155c362da777e7ce31b85fb2dc7cfced2d5 -> trunk/9fff8155c362da777e7ce31b85fb2dc7cfced2d5 2025-10-10T01:18:34.3478204Z * [new tag] trunk/a029675f6f0b9cf48eb7943d4be8169c67960a8e -> trunk/a029675f6f0b9cf48eb7943d4be8169c67960a8e 2025-10-10T01:18:34.3479733Z * [new tag] trunk/a11a66ef320938cd0fd72b44b2b572b06937e100 -> trunk/a11a66ef320938cd0fd72b44b2b572b06937e100 2025-10-10T01:18:34.3481136Z * [new tag] trunk/a2f29bcd6388acdc3202d8a90974c50ffb605104 -> trunk/a2f29bcd6388acdc3202d8a90974c50ffb605104 2025-10-10T01:18:34.3482667Z * [new tag] trunk/a34797e031727f6a01a2f13a66db2f7e1fcc05b6 -> trunk/a34797e031727f6a01a2f13a66db2f7e1fcc05b6 2025-10-10T01:18:34.3484010Z * [new tag] trunk/a4110fedcf72eaede76324bb5c21a76589d75849 -> trunk/a4110fedcf72eaede76324bb5c21a76589d75849 2025-10-10T01:18:34.3485298Z * [new tag] trunk/a43c4c3972a611db169dde2aed803b91fe78c081 -> trunk/a43c4c3972a611db169dde2aed803b91fe78c081 2025-10-10T01:18:34.3486659Z * [new tag] trunk/a57a14868dcfd9dabf9bd19b6b11f31967c80c87 -> trunk/a57a14868dcfd9dabf9bd19b6b11f31967c80c87 2025-10-10T01:18:34.3488121Z * [new tag] trunk/a6fa4f9c283971c0fb6f60a89674a1f35370ac79 -> trunk/a6fa4f9c283971c0fb6f60a89674a1f35370ac79 2025-10-10T01:18:34.3489558Z * [new tag] trunk/a753ffa9aff47e005c31d6bcbf5b6a61cc54afed -> trunk/a753ffa9aff47e005c31d6bcbf5b6a61cc54afed 2025-10-10T01:18:34.3490917Z * [new tag] trunk/a7fa1a91e386c7708e4c8747680911b0c3174a66 -> trunk/a7fa1a91e386c7708e4c8747680911b0c3174a66 2025-10-10T01:18:34.3492248Z * [new tag] trunk/a9a9a3438a374f96a308b707a1718036aaec790d -> trunk/a9a9a3438a374f96a308b707a1718036aaec790d 2025-10-10T01:18:34.3493627Z * [new tag] trunk/ab01a0d7d352e7fd07989b8d6bf035bf82aea74e -> trunk/ab01a0d7d352e7fd07989b8d6bf035bf82aea74e 2025-10-10T01:18:34.3495336Z * [new tag] trunk/ab94a0d544503b5c27e889b45e45ef8cf75c8183 -> trunk/ab94a0d544503b5c27e889b45e45ef8cf75c8183 2025-10-10T01:18:34.3496326Z * [new tag] trunk/abadea70f3eb5f2f764fd6448d42dd2c29fa28b3 -> trunk/abadea70f3eb5f2f764fd6448d42dd2c29fa28b3 2025-10-10T01:18:34.3497750Z * [new tag] trunk/ac08556f674259ff5b117964e300124e8a92d45b -> trunk/ac08556f674259ff5b117964e300124e8a92d45b 2025-10-10T01:18:34.3499270Z * [new tag] trunk/ac7b4e7fe4d233dcd7f6343d42b4fa3d64bce548 -> trunk/ac7b4e7fe4d233dcd7f6343d42b4fa3d64bce548 2025-10-10T01:18:34.3500667Z * [new tag] trunk/ac901bf79a2d78539ffec272bf32f4ae47035b23 -> trunk/ac901bf79a2d78539ffec272bf32f4ae47035b23 2025-10-10T01:18:34.3502088Z * [new tag] trunk/ad7b2bebc651c297d869f265deedef726bf17048 -> trunk/ad7b2bebc651c297d869f265deedef726bf17048 2025-10-10T01:18:34.3503466Z * [new tag] trunk/ae25ec569c614c2a2274837079578b71f3201a3b -> trunk/ae25ec569c614c2a2274837079578b71f3201a3b 2025-10-10T01:18:34.3505150Z * [new tag] trunk/aea57b3aa38a3d4a058e0a7eba08d0c6c28ed9c5 -> trunk/aea57b3aa38a3d4a058e0a7eba08d0c6c28ed9c5 2025-10-10T01:18:34.3506658Z * [new tag] trunk/aed5ed1076d3e73e0b6357dafac1002aa6a221e9 -> trunk/aed5ed1076d3e73e0b6357dafac1002aa6a221e9 2025-10-10T01:18:34.3508125Z * [new tag] trunk/aed66248a01d309eb2ac1149b5f51310545b0783 -> trunk/aed66248a01d309eb2ac1149b5f51310545b0783 2025-10-10T01:18:34.3509543Z * [new tag] trunk/af32d16a71681ca05c6d410fb1b9cee091d4577d -> trunk/af32d16a71681ca05c6d410fb1b9cee091d4577d 2025-10-10T01:18:34.3510910Z * [new tag] trunk/af40828bbb785f968eda18dbdc8750ba67f57366 -> trunk/af40828bbb785f968eda18dbdc8750ba67f57366 2025-10-10T01:18:34.3512392Z * [new tag] trunk/af4c29fea8f50ac3bb9e4a0e305da4a2c6b53d29 -> trunk/af4c29fea8f50ac3bb9e4a0e305da4a2c6b53d29 2025-10-10T01:18:34.3513842Z * [new tag] trunk/afee8062d511ad63e0af65ffac0e712d86aae8f1 -> trunk/afee8062d511ad63e0af65ffac0e712d86aae8f1 2025-10-10T01:18:34.3515210Z * [new tag] trunk/afeec56a5aa83dd0258565400551a99777c0023b -> trunk/afeec56a5aa83dd0258565400551a99777c0023b 2025-10-10T01:18:34.3516853Z * [new tag] trunk/b0985144b59db8fb20964829b5e0a9d2f9a3f0d6 -> trunk/b0985144b59db8fb20964829b5e0a9d2f9a3f0d6 2025-10-10T01:18:34.3518515Z * [new tag] trunk/b116c5133024be39a2db67cd0112b490b970b710 -> trunk/b116c5133024be39a2db67cd0112b490b970b710 2025-10-10T01:18:34.3520053Z * [new tag] trunk/b13cd141b3585c2ae89ad7747acd11203a2fb837 -> trunk/b13cd141b3585c2ae89ad7747acd11203a2fb837 2025-10-10T01:18:34.3521470Z * [new tag] trunk/b1ac252f55f4a4d0e5488fb2ac9154154decec87 -> trunk/b1ac252f55f4a4d0e5488fb2ac9154154decec87 2025-10-10T01:18:34.3522828Z * [new tag] trunk/b28b24a9fc7d391c5793a94489a3f2d5381f6ad7 -> trunk/b28b24a9fc7d391c5793a94489a3f2d5381f6ad7 2025-10-10T01:18:34.3524453Z * [new tag] trunk/b2b3947565fd0c27ebd4941152c964eab30370e2 -> trunk/b2b3947565fd0c27ebd4941152c964eab30370e2 2025-10-10T01:18:34.3525691Z * [new tag] trunk/b558c986e8ec693b531ad2817026393c55d72eb6 -> trunk/b558c986e8ec693b531ad2817026393c55d72eb6 2025-10-10T01:18:34.3527452Z * [new tag] trunk/b5e93ffdcf779c703af5c8119636b01f250eafcd -> trunk/b5e93ffdcf779c703af5c8119636b01f250eafcd 2025-10-10T01:18:34.3528850Z * [new tag] trunk/b63bbe16615cc7680836dbb151bd848bce4893d6 -> trunk/b63bbe16615cc7680836dbb151bd848bce4893d6 2025-10-10T01:18:34.3530165Z * [new tag] trunk/b6b7a44dec63495d57946cbfe8f2accb8f876db2 -> trunk/b6b7a44dec63495d57946cbfe8f2accb8f876db2 2025-10-10T01:18:34.3531182Z * [new tag] trunk/b9e73e639e36f3aa628752161711e68878231b30 -> trunk/b9e73e639e36f3aa628752161711e68878231b30 2025-10-10T01:18:34.3532858Z * [new tag] trunk/ba480d6bf78ea446d1268d9b5b3a0dbb490c9c88 -> trunk/ba480d6bf78ea446d1268d9b5b3a0dbb490c9c88 2025-10-10T01:18:34.3534381Z * [new tag] trunk/bac0f289a35f05052740076fc5671271a3d487c2 -> trunk/bac0f289a35f05052740076fc5671271a3d487c2 2025-10-10T01:18:34.3535775Z * [new tag] trunk/bc1690c7e859dee8c47a7f0bbd3c43cc27c6fd2a -> trunk/bc1690c7e859dee8c47a7f0bbd3c43cc27c6fd2a 2025-10-10T01:18:34.3537090Z * [new tag] trunk/bc33b10202fb7c3761bcabc166e02d96807d8739 -> trunk/bc33b10202fb7c3761bcabc166e02d96807d8739 2025-10-10T01:18:34.3538493Z * [new tag] trunk/bcafea5c92ca2ee1b0dc8f6d8b62ecabb6f40228 -> trunk/bcafea5c92ca2ee1b0dc8f6d8b62ecabb6f40228 2025-10-10T01:18:34.3539886Z * [new tag] trunk/bcd96cc6ff798281e66aabef6ce72542fdc97c7a -> trunk/bcd96cc6ff798281e66aabef6ce72542fdc97c7a 2025-10-10T01:18:34.3541264Z * [new tag] trunk/bd3b98a8a5d68ddc84b20a4609b9ea90998bf95b -> trunk/bd3b98a8a5d68ddc84b20a4609b9ea90998bf95b 2025-10-10T01:18:34.3542686Z * [new tag] trunk/bdc0a421d7bcc49db12f7593d2c213a6141da614 -> trunk/bdc0a421d7bcc49db12f7593d2c213a6141da614 2025-10-10T01:18:34.3544095Z * [new tag] trunk/bde18c445dcb1d83e8ea0afae52f9b9bf8171f45 -> trunk/bde18c445dcb1d83e8ea0afae52f9b9bf8171f45 2025-10-10T01:18:34.3545427Z * [new tag] trunk/bf717ce346203fc27e792f4bdcc31e979cd74fa9 -> trunk/bf717ce346203fc27e792f4bdcc31e979cd74fa9 2025-10-10T01:18:34.3547148Z * [new tag] trunk/c0510dc447a1f105cb8758d2721380f7a7c380d1 -> trunk/c0510dc447a1f105cb8758d2721380f7a7c380d1 2025-10-10T01:18:34.3548558Z * [new tag] trunk/c1f40d33c89b361a1edad17aa25cfff1ab4014fd -> trunk/c1f40d33c89b361a1edad17aa25cfff1ab4014fd 2025-10-10T01:18:34.3549911Z * [new tag] trunk/c32118dc3e50505fd285e6e448a90883fce11535 -> trunk/c32118dc3e50505fd285e6e448a90883fce11535 2025-10-10T01:18:34.3551331Z * [new tag] trunk/c45d56dd00546daa7d9044674233dba1ac7b6194 -> trunk/c45d56dd00546daa7d9044674233dba1ac7b6194 2025-10-10T01:18:34.3552723Z * [new tag] trunk/c6329524d8670d5f9295cddcf7ebc3040ed9179e -> trunk/c6329524d8670d5f9295cddcf7ebc3040ed9179e 2025-10-10T01:18:34.3554146Z * [new tag] trunk/c6a6c80a730ff4edaec0d2fc4a5ff9344edaed41 -> trunk/c6a6c80a730ff4edaec0d2fc4a5ff9344edaed41 2025-10-10T01:18:34.3555568Z * [new tag] trunk/c7e30ae4dd9a58ed4f4bcbdc6afc2249cac94f28 -> trunk/c7e30ae4dd9a58ed4f4bcbdc6afc2249cac94f28 2025-10-10T01:18:34.3557008Z * [new tag] trunk/c813617c53e6be91e77f47e9a3f713146d54f340 -> trunk/c813617c53e6be91e77f47e9a3f713146d54f340 2025-10-10T01:18:34.3558883Z * [new tag] trunk/c855f8632e331b51d60d5f1bcc59d3181cb4bc82 -> trunk/c855f8632e331b51d60d5f1bcc59d3181cb4bc82 2025-10-10T01:18:34.3560128Z * [new tag] trunk/c965d6dbb23a8b2338ffebf3f01c6f92ce5847d2 -> trunk/c965d6dbb23a8b2338ffebf3f01c6f92ce5847d2 2025-10-10T01:18:34.3561515Z * [new tag] trunk/cac5e13e1384900c5acc4938c33d6037a61850d5 -> trunk/cac5e13e1384900c5acc4938c33d6037a61850d5 2025-10-10T01:18:34.3562916Z * [new tag] trunk/cc71ab86a6985e85645424b727c766e031047ff6 -> trunk/cc71ab86a6985e85645424b727c766e031047ff6 2025-10-10T01:18:34.3564415Z * [new tag] trunk/cd62a73dcb13102069aa827a6657f62d88cce095 -> trunk/cd62a73dcb13102069aa827a6657f62d88cce095 2025-10-10T01:18:34.3565773Z * [new tag] trunk/cf0a00d4f38775e5a82a166e367f40383c606963 -> trunk/cf0a00d4f38775e5a82a166e367f40383c606963 2025-10-10T01:18:34.3567180Z * [new tag] trunk/cfc5cc17dc4fa6be41b4b31eb6e63d3863479452 -> trunk/cfc5cc17dc4fa6be41b4b31eb6e63d3863479452 2025-10-10T01:18:34.3568578Z * [new tag] trunk/cfd46d13e6e1308add3a9f287b4855ccc3f2e66c -> trunk/cfd46d13e6e1308add3a9f287b4855ccc3f2e66c 2025-10-10T01:18:34.3569919Z * [new tag] trunk/d1a62c80363cf769552453eed187e935f905737d -> trunk/d1a62c80363cf769552453eed187e935f905737d 2025-10-10T01:18:34.3571367Z * [new tag] trunk/d1cbb74fb16406488a174832e1b58b7c242f418d -> trunk/d1cbb74fb16406488a174832e1b58b7c242f418d 2025-10-10T01:18:34.3572783Z * [new tag] trunk/d386325ca9a142419f45b987391f4bb175dd7d0b -> trunk/d386325ca9a142419f45b987391f4bb175dd7d0b 2025-10-10T01:18:34.3574246Z * [new tag] trunk/d40a9bfb8da0dc1ac1e6e56b33a25979112874de -> trunk/d40a9bfb8da0dc1ac1e6e56b33a25979112874de 2025-10-10T01:18:34.3575643Z * [new tag] trunk/d4443840036a00a30afcf066cb23f4525e590809 -> trunk/d4443840036a00a30afcf066cb23f4525e590809 2025-10-10T01:18:34.3577077Z * [new tag] trunk/d4752bc7f6818a3df5356a9de61afe1d3e27ade9 -> trunk/d4752bc7f6818a3df5356a9de61afe1d3e27ade9 2025-10-10T01:18:34.3578469Z * [new tag] trunk/da49a57d3462332b26cb7ee58910b5bc67e5772c -> trunk/da49a57d3462332b26cb7ee58910b5bc67e5772c 2025-10-10T01:18:34.3579902Z * [new tag] trunk/da903b6a8be422529d47649e89c0d50bb95c37ca -> trunk/da903b6a8be422529d47649e89c0d50bb95c37ca 2025-10-10T01:18:34.3581460Z * [new tag] trunk/dca73982c53e9f99f96246b5d9ed9bab83c7423f -> trunk/dca73982c53e9f99f96246b5d9ed9bab83c7423f 2025-10-10T01:18:34.3582878Z * [new tag] trunk/ddf8de28c25944a58e739ba9996b06753e4199cc -> trunk/ddf8de28c25944a58e739ba9996b06753e4199cc 2025-10-10T01:18:34.3584246Z * [new tag] trunk/df640df68a5275684eaae3080a9c97a0c61469c8 -> trunk/df640df68a5275684eaae3080a9c97a0c61469c8 2025-10-10T01:18:34.3585572Z * [new tag] trunk/e09fb44ef177005c4a11c28be24781429d416a3e -> trunk/e09fb44ef177005c4a11c28be24781429d416a3e 2025-10-10T01:18:34.3586948Z * [new tag] trunk/e0cb1848d0fd9fb4467ad8b844c565aea5071838 -> trunk/e0cb1848d0fd9fb4467ad8b844c565aea5071838 2025-10-10T01:18:34.3588371Z * [new tag] trunk/e3ae80fc036da356e3748d134689741583552f09 -> trunk/e3ae80fc036da356e3748d134689741583552f09 2025-10-10T01:18:34.3589793Z * [new tag] trunk/e40fe634b1a7aa33e278b1404ee02dea12277080 -> trunk/e40fe634b1a7aa33e278b1404ee02dea12277080 2025-10-10T01:18:34.3591370Z * [new tag] trunk/e438db254602cf39ba536aed0590b4144c019ee8 -> trunk/e438db254602cf39ba536aed0590b4144c019ee8 2025-10-10T01:18:34.3592764Z * [new tag] trunk/e532f62e0d96e56cb28fa6a0ba6d981896a65d52 -> trunk/e532f62e0d96e56cb28fa6a0ba6d981896a65d52 2025-10-10T01:18:34.3594162Z * [new tag] trunk/e6d4b26776842307475b368db60e27ac1bcede86 -> trunk/e6d4b26776842307475b368db60e27ac1bcede86 2025-10-10T01:18:34.3595628Z * [new tag] trunk/e7ed1a00eb5510d1c7dccd17b5c0ebb54231284f -> trunk/e7ed1a00eb5510d1c7dccd17b5c0ebb54231284f 2025-10-10T01:18:34.3596957Z * [new tag] trunk/e7fd2969303ab931f5e5875eca676018e1acd089 -> trunk/e7fd2969303ab931f5e5875eca676018e1acd089 2025-10-10T01:18:34.3598374Z * [new tag] trunk/e89d12bf5d6b69c153cd000ef278fca59f03226d -> trunk/e89d12bf5d6b69c153cd000ef278fca59f03226d 2025-10-10T01:18:34.3599937Z * [new tag] trunk/e98c4e835b1db22092fc93b49d2cddd7b3537d1f -> trunk/e98c4e835b1db22092fc93b49d2cddd7b3537d1f 2025-10-10T01:18:34.3601339Z * [new tag] trunk/ea42517e454d2e47391646bbb897f5fc51147b9d -> trunk/ea42517e454d2e47391646bbb897f5fc51147b9d 2025-10-10T01:18:34.3602752Z * [new tag] trunk/eaa02655eabd24609744c2251ac40d39d86ebb39 -> trunk/eaa02655eabd24609744c2251ac40d39d86ebb39 2025-10-10T01:18:34.3604193Z * [new tag] trunk/eccf561326147894d57482a5aba7a2290005b257 -> trunk/eccf561326147894d57482a5aba7a2290005b257 2025-10-10T01:18:34.3605635Z * [new tag] trunk/ece5e0f01b68509342f85fa388ca61936dc18b20 -> trunk/ece5e0f01b68509342f85fa388ca61936dc18b20 2025-10-10T01:18:34.3607051Z * [new tag] trunk/ed2d514ad860229f6d364688f9db27dad034cd83 -> trunk/ed2d514ad860229f6d364688f9db27dad034cd83 2025-10-10T01:18:34.3608371Z * [new tag] trunk/ed6156e3ea334b9b8d395e5a9f76fa3ba7408c06 -> trunk/ed6156e3ea334b9b8d395e5a9f76fa3ba7408c06 2025-10-10T01:18:34.3609832Z * [new tag] trunk/ee5389d520844db36374e86c986b9ff8f47ac4bb -> trunk/ee5389d520844db36374e86c986b9ff8f47ac4bb 2025-10-10T01:18:34.3611225Z * [new tag] trunk/ee6a1ecb0a1035f068484c8fcfba44b2efc9e837 -> trunk/ee6a1ecb0a1035f068484c8fcfba44b2efc9e837 2025-10-10T01:18:34.3612630Z * [new tag] trunk/ef50c6e3e3d83bfd67e50930eea9a3a9db084061 -> trunk/ef50c6e3e3d83bfd67e50930eea9a3a9db084061 2025-10-10T01:18:34.3614254Z * [new tag] trunk/ef7e2ca77e3f554ced81eb614f15fb84249d4a7e -> trunk/ef7e2ca77e3f554ced81eb614f15fb84249d4a7e 2025-10-10T01:18:34.3615714Z * [new tag] trunk/f006aee601cb72077f4b1dbc3f7f0f685e57a1a9 -> trunk/f006aee601cb72077f4b1dbc3f7f0f685e57a1a9 2025-10-10T01:18:34.3617081Z * [new tag] trunk/f05e23e1bc1439e19145e43e8ffca0051cda2f33 -> trunk/f05e23e1bc1439e19145e43e8ffca0051cda2f33 2025-10-10T01:18:34.3618948Z * [new tag] trunk/f0c9f3bddbf7ad77d5d3a8803c23bb47bfb71d79 -> trunk/f0c9f3bddbf7ad77d5d3a8803c23bb47bfb71d79 2025-10-10T01:18:34.3620410Z * [new tag] trunk/f11ac803d73b90d7e1f7bde962b9afe6b5967eb7 -> trunk/f11ac803d73b90d7e1f7bde962b9afe6b5967eb7 2025-10-10T01:18:34.3621821Z * [new tag] trunk/f1229b6db946c290cc5bdea05dde69fc01e0bed0 -> trunk/f1229b6db946c290cc5bdea05dde69fc01e0bed0 2025-10-10T01:18:34.3623228Z * [new tag] trunk/f231be25c679adb47ac3e483dc68948e5ad137a4 -> trunk/f231be25c679adb47ac3e483dc68948e5ad137a4 2025-10-10T01:18:34.3624782Z * [new tag] trunk/f33201729416ed17467228e80b04d01d4d02b5f3 -> trunk/f33201729416ed17467228e80b04d01d4d02b5f3 2025-10-10T01:18:34.3626354Z * [new tag] trunk/f37a6523efe1b9bf7f6b5b5d0f36dc461a3fda2a -> trunk/f37a6523efe1b9bf7f6b5b5d0f36dc461a3fda2a 2025-10-10T01:18:34.3627860Z * [new tag] trunk/f39789cdabb6465f21666bd001829e1f7284d754 -> trunk/f39789cdabb6465f21666bd001829e1f7284d754 2025-10-10T01:18:34.3629241Z * [new tag] trunk/f3afbcf3407783e54ec2795b06ae744f645320ba -> trunk/f3afbcf3407783e54ec2795b06ae744f645320ba 2025-10-10T01:18:34.3630674Z * [new tag] trunk/f3e43ff2d73f375487b1b71483bbecb6cdad8920 -> trunk/f3e43ff2d73f375487b1b71483bbecb6cdad8920 2025-10-10T01:18:34.3632068Z * [new tag] trunk/f414aa8e0d17e8eff38a93cebd52436e53f50eba -> trunk/f414aa8e0d17e8eff38a93cebd52436e53f50eba 2025-10-10T01:18:34.3633484Z * [new tag] trunk/f465ea6752c91498de63eb57439a74f4836e568a -> trunk/f465ea6752c91498de63eb57439a74f4836e568a 2025-10-10T01:18:34.3635070Z * [new tag] trunk/f46bb04dcc37a9b394e414569aef8aef69f9bf53 -> trunk/f46bb04dcc37a9b394e414569aef8aef69f9bf53 2025-10-10T01:18:34.3636286Z * [new tag] trunk/f46ddb1e65b595c80f285dc42aa8549970736aae -> trunk/f46ddb1e65b595c80f285dc42aa8549970736aae 2025-10-10T01:18:34.3637805Z * [new tag] trunk/f4cf75688f0fd93466589addfb7d0ec33e46e3bf -> trunk/f4cf75688f0fd93466589addfb7d0ec33e46e3bf 2025-10-10T01:18:34.3639243Z * [new tag] trunk/f505caa71bd2e4d1e708e20a3665b834134e08fc -> trunk/f505caa71bd2e4d1e708e20a3665b834134e08fc 2025-10-10T01:18:34.3640681Z * [new tag] trunk/f5fd18f7e24378bd9eb91404f697f1c81a8187d5 -> trunk/f5fd18f7e24378bd9eb91404f697f1c81a8187d5 2025-10-10T01:18:34.3642082Z * [new tag] trunk/f6de195616432f42a545b98ea41cc816019d1c60 -> trunk/f6de195616432f42a545b98ea41cc816019d1c60 2025-10-10T01:18:34.3643518Z * [new tag] trunk/f6f76767563d4293a0f78551edf4675a5794c570 -> trunk/f6f76767563d4293a0f78551edf4675a5794c570 2025-10-10T01:18:34.3644943Z * [new tag] trunk/f7082e92b3635e89906fae514506152a2ec844a0 -> trunk/f7082e92b3635e89906fae514506152a2ec844a0 2025-10-10T01:18:34.3646346Z * [new tag] trunk/f713abab16cb98c15f486e9822dd261279cce252 -> trunk/f713abab16cb98c15f486e9822dd261279cce252 2025-10-10T01:18:34.3647808Z * [new tag] trunk/f76fdcaaf8b6d5f97c7f63705400ebed8984f869 -> trunk/f76fdcaaf8b6d5f97c7f63705400ebed8984f869 2025-10-10T01:18:34.3649260Z * [new tag] trunk/f79e212733ca89ce3cc99a3072e50351686e5568 -> trunk/f79e212733ca89ce3cc99a3072e50351686e5568 2025-10-10T01:18:34.3650658Z * [new tag] trunk/f7ad6dbad67161333a1473d1e0b478b7475a0ec1 -> trunk/f7ad6dbad67161333a1473d1e0b478b7475a0ec1 2025-10-10T01:18:34.3652040Z * [new tag] trunk/fa5306b4f5bea89d80b9f14926712119aab78161 -> trunk/fa5306b4f5bea89d80b9f14926712119aab78161 2025-10-10T01:18:34.3653566Z * [new tag] trunk/fac6f20ae3a68fa49e19571a1fc4bcdddbf87d80 -> trunk/fac6f20ae3a68fa49e19571a1fc4bcdddbf87d80 2025-10-10T01:18:34.3654990Z * [new tag] trunk/fac85fcfb5ad0e63438d808a2f9ba7ea2dff9ad4 -> trunk/fac85fcfb5ad0e63438d808a2f9ba7ea2dff9ad4 2025-10-10T01:18:34.3656375Z * [new tag] trunk/fd3e15c14f4fc474af610b482382a2c85729f50d -> trunk/fd3e15c14f4fc474af610b482382a2c85729f50d 2025-10-10T01:18:34.3657827Z * [new tag] trunk/fd4bde430a51e5f216295c950d962c6343119821 -> trunk/fd4bde430a51e5f216295c950d962c6343119821 2025-10-10T01:18:34.3659255Z * [new tag] trunk/fdc622b513610b53ddcdc0b40282df9beae369bd -> trunk/fdc622b513610b53ddcdc0b40282df9beae369bd 2025-10-10T01:18:34.3660683Z * [new tag] trunk/fdc8ccc5bc433478c2a114016e193f5665d1e370 -> trunk/fdc8ccc5bc433478c2a114016e193f5665d1e370 2025-10-10T01:18:34.3662046Z * [new tag] trunk/ff5faa744a52561f4c6a138089123fd8d41cab73 -> trunk/ff5faa744a52561f4c6a138089123fd8d41cab73 2025-10-10T01:18:34.3663210Z * [new tag] v0.1.1 -> v0.1.1 2025-10-10T01:18:34.3664520Z * [new tag] v0.1.10 -> v0.1.10 2025-10-10T01:18:34.3665783Z * [new tag] v0.1.11 -> v0.1.11 2025-10-10T01:18:34.3667452Z * [new tag] v0.1.12 -> v0.1.12 2025-10-10T01:18:34.3668683Z * [new tag] v0.1.2 -> v0.1.2 2025-10-10T01:18:34.3669922Z * [new tag] v0.1.3 -> v0.1.3 2025-10-10T01:18:34.3671233Z * [new tag] v0.1.4 -> v0.1.4 2025-10-10T01:18:34.3672520Z * [new tag] v0.1.5 -> v0.1.5 2025-10-10T01:18:34.3673822Z * [new tag] v0.1.6 -> v0.1.6 2025-10-10T01:18:34.3675071Z * [new tag] v0.1.7 -> v0.1.7 2025-10-10T01:18:34.3676457Z * [new tag] v0.1.8 -> v0.1.8 2025-10-10T01:18:34.3677630Z * [new tag] v0.1.9 -> v0.1.9 2025-10-10T01:18:34.3679274Z * [new tag] v0.2.0 -> v0.2.0 2025-10-10T01:18:34.3680361Z * [new tag] v0.3.0 -> v0.3.0 2025-10-10T01:18:34.3681822Z * [new tag] v0.3.1 -> v0.3.1 2025-10-10T01:18:34.3683106Z * [new tag] v0.4.0 -> v0.4.0 2025-10-10T01:18:34.3684432Z * [new tag] v0.4.1 -> v0.4.1 2025-10-10T01:18:34.3685747Z * [new tag] v1.0.0 -> v1.0.0 2025-10-10T01:18:34.3687042Z * [new tag] v1.0.0a0 -> v1.0.0a0 2025-10-10T01:18:34.3688353Z * [new tag] v1.0.1 -> v1.0.1 2025-10-10T01:18:34.3689791Z * [new tag] v1.0rc0 -> v1.0rc0 2025-10-10T01:18:34.3690911Z * [new tag] v1.0rc1 -> v1.0rc1 2025-10-10T01:18:34.3692303Z * [new tag] v1.1.0 -> v1.1.0 2025-10-10T01:18:34.3693653Z * [new tag] v1.1.0a0 -> v1.1.0a0 2025-10-10T01:18:34.3695163Z * [new tag] v1.10.0 -> v1.10.0 2025-10-10T01:18:34.3696562Z * [new tag] v1.10.0-rc1 -> v1.10.0-rc1 2025-10-10T01:18:34.3697892Z * [new tag] v1.10.0-rc2 -> v1.10.0-rc2 2025-10-10T01:18:34.3699021Z * [new tag] v1.10.0-rc3 -> v1.10.0-rc3 2025-10-10T01:18:34.3700371Z * [new tag] v1.10.1 -> v1.10.1 2025-10-10T01:18:34.3701495Z * [new tag] v1.10.1-rc1 -> v1.10.1-rc1 2025-10-10T01:18:34.3702712Z * [new tag] v1.10.2 -> v1.10.2 2025-10-10T01:18:34.3703819Z * [new tag] v1.10.2-rc1 -> v1.10.2-rc1 2025-10-10T01:18:34.3705188Z * [new tag] v1.11.0 -> v1.11.0 2025-10-10T01:18:34.3706610Z * [new tag] v1.11.0-rc1 -> v1.11.0-rc1 2025-10-10T01:18:34.3708087Z * [new tag] v1.11.0-rc2 -> v1.11.0-rc2 2025-10-10T01:18:34.3709449Z * [new tag] v1.11.0-rc3 -> v1.11.0-rc3 2025-10-10T01:18:34.3710836Z * [new tag] v1.11.0-rc4 -> v1.11.0-rc4 2025-10-10T01:18:34.3712220Z * [new tag] v1.11.0-rc5 -> v1.11.0-rc5 2025-10-10T01:18:34.3713366Z * [new tag] v1.11.0-rc6 -> v1.11.0-rc6 2025-10-10T01:18:34.3714496Z * [new tag] v1.11.0-rc7 -> v1.11.0-rc7 2025-10-10T01:18:34.3715789Z * [new tag] v1.12.0 -> v1.12.0 2025-10-10T01:18:34.3717146Z * [new tag] v1.12.0-rc1 -> v1.12.0-rc1 2025-10-10T01:18:34.3718537Z * [new tag] v1.12.0-rc2 -> v1.12.0-rc2 2025-10-10T01:18:34.3720071Z * [new tag] v1.12.0-rc3 -> v1.12.0-rc3 2025-10-10T01:18:34.3721399Z * [new tag] v1.12.0-rc4 -> v1.12.0-rc4 2025-10-10T01:18:34.3722764Z * [new tag] v1.12.0-rc5 -> v1.12.0-rc5 2025-10-10T01:18:34.3724468Z * [new tag] v1.12.0-rc6 -> v1.12.0-rc6 2025-10-10T01:18:34.3726461Z * [new tag] v1.12.0-rc7 -> v1.12.0-rc7 2025-10-10T01:18:34.3727675Z * [new tag] v1.12.0-rc8 -> v1.12.0-rc8 2025-10-10T01:18:34.3728945Z * [new tag] v1.12.1 -> v1.12.1 2025-10-10T01:18:34.3730408Z * [new tag] v1.12.1-rc1 -> v1.12.1-rc1 2025-10-10T01:18:34.3731750Z * [new tag] v1.12.1-rc2 -> v1.12.1-rc2 2025-10-10T01:18:34.3733308Z * [new tag] v1.12.1-rc3 -> v1.12.1-rc3 2025-10-10T01:18:34.3734354Z * [new tag] v1.12.1-rc4 -> v1.12.1-rc4 2025-10-10T01:18:34.3735555Z * [new tag] v1.12.1-rc5 -> v1.12.1-rc5 2025-10-10T01:18:34.3736938Z * [new tag] v1.13.0 -> v1.13.0 2025-10-10T01:18:34.3738283Z * [new tag] v1.13.0-rc1 -> v1.13.0-rc1 2025-10-10T01:18:34.3739598Z * [new tag] v1.13.0-rc2 -> v1.13.0-rc2 2025-10-10T01:18:34.3740860Z * [new tag] v1.13.0-rc3 -> v1.13.0-rc3 2025-10-10T01:18:34.3742488Z * [new tag] v1.13.0-rc4 -> v1.13.0-rc4 2025-10-10T01:18:34.3743605Z * [new tag] v1.13.0-rc5 -> v1.13.0-rc5 2025-10-10T01:18:34.3744725Z * [new tag] v1.13.0-rc6 -> v1.13.0-rc6 2025-10-10T01:18:34.3746091Z * [new tag] v1.13.1 -> v1.13.1 2025-10-10T01:18:34.3747240Z * [new tag] v1.13.1-rc1 -> v1.13.1-rc1 2025-10-10T01:18:34.3748541Z * [new tag] v1.2.0 -> v1.2.0 2025-10-10T01:18:34.3749910Z * [new tag] v1.2.0a0 -> v1.2.0a0 2025-10-10T01:18:34.3751199Z * [new tag] v1.3.0 -> v1.3.0 2025-10-10T01:18:34.3752906Z * [new tag] v1.3.0a0 -> v1.3.0a0 2025-10-10T01:18:34.3754034Z * [new tag] v1.3.1 -> v1.3.1 2025-10-10T01:18:34.3755302Z * [new tag] v1.4.0 -> v1.4.0 2025-10-10T01:18:34.3756665Z * [new tag] v1.4.0a0 -> v1.4.0a0 2025-10-10T01:18:34.3757792Z * [new tag] v1.4.1 -> v1.4.1 2025-10-10T01:18:34.3759325Z * [new tag] v1.5.0 -> v1.5.0 2025-10-10T01:18:34.3760680Z * [new tag] v1.5.0-rc1 -> v1.5.0-rc1 2025-10-10T01:18:34.3762033Z * [new tag] v1.5.0-rc2 -> v1.5.0-rc2 2025-10-10T01:18:34.3763332Z * [new tag] v1.5.0-rc3 -> v1.5.0-rc3 2025-10-10T01:18:34.3764537Z * [new tag] v1.5.0-rc4 -> v1.5.0-rc4 2025-10-10T01:18:34.3765651Z * [new tag] v1.5.0-rc5 -> v1.5.0-rc5 2025-10-10T01:18:34.3767109Z * [new tag] v1.5.1 -> v1.5.1 2025-10-10T01:18:34.3768231Z * [new tag] v1.5.1-rc1 -> v1.5.1-rc1 2025-10-10T01:18:34.3769324Z * [new tag] v1.6.0 -> v1.6.0 2025-10-10T01:18:34.3770704Z * [new tag] v1.6.0-rc1 -> v1.6.0-rc1 2025-10-10T01:18:34.3772041Z * [new tag] v1.6.0-rc2 -> v1.6.0-rc2 2025-10-10T01:18:34.3773378Z * [new tag] v1.6.0-rc3 -> v1.6.0-rc3 2025-10-10T01:18:34.3774697Z * [new tag] v1.6.0-rc4 -> v1.6.0-rc4 2025-10-10T01:18:34.3776016Z * [new tag] v1.6.0-rc5 -> v1.6.0-rc5 2025-10-10T01:18:34.3777253Z * [new tag] v1.6.0-rc6 -> v1.6.0-rc6 2025-10-10T01:18:34.3778384Z * [new tag] v1.6.0-rc7 -> v1.6.0-rc7 2025-10-10T01:18:34.3779749Z * [new tag] v1.7.0 -> v1.7.0 2025-10-10T01:18:34.3781069Z * [new tag] v1.7.0-rc1 -> v1.7.0-rc1 2025-10-10T01:18:34.3782452Z * [new tag] v1.7.0-rc2 -> v1.7.0-rc2 2025-10-10T01:18:34.3783759Z * [new tag] v1.7.0-rc3 -> v1.7.0-rc3 2025-10-10T01:18:34.3784879Z * [new tag] v1.7.0-rc4 -> v1.7.0-rc4 2025-10-10T01:18:34.3786367Z * [new tag] v1.7.1 -> v1.7.1 2025-10-10T01:18:34.3787766Z * [new tag] v1.7.1-rc1 -> v1.7.1-rc1 2025-10-10T01:18:34.3789061Z * [new tag] v1.7.1-rc2 -> v1.7.1-rc2 2025-10-10T01:18:34.3790177Z * [new tag] v1.7.1-rc3 -> v1.7.1-rc3 2025-10-10T01:18:34.3791514Z * [new tag] v1.8.0 -> v1.8.0 2025-10-10T01:18:34.3792633Z * [new tag] v1.8.0-rc1 -> v1.8.0-rc1 2025-10-10T01:18:34.3793944Z * [new tag] v1.8.0-rc2 -> v1.8.0-rc2 2025-10-10T01:18:34.3795272Z * [new tag] v1.8.0-rc3 -> v1.8.0-rc3 2025-10-10T01:18:34.3796510Z * [new tag] v1.8.0-rc4 -> v1.8.0-rc4 2025-10-10T01:18:34.3797519Z * [new tag] v1.8.0-rc5 -> v1.8.0-rc5 2025-10-10T01:18:34.3798790Z * [new tag] v1.8.1 -> v1.8.1 2025-10-10T01:18:34.3800227Z * [new tag] v1.8.1-rc1 -> v1.8.1-rc1 2025-10-10T01:18:34.3801227Z * [new tag] v1.8.1-rc2 -> v1.8.1-rc2 2025-10-10T01:18:34.3802367Z * [new tag] v1.8.1-rc3 -> v1.8.1-rc3 2025-10-10T01:18:34.3804114Z * [new tag] v1.8.2 -> v1.8.2 2025-10-10T01:18:34.3805245Z * [new tag] v1.8.2-rc1 -> v1.8.2-rc1 2025-10-10T01:18:34.3806589Z * [new tag] v1.9.0 -> v1.9.0 2025-10-10T01:18:34.3807874Z * [new tag] v1.9.0-rc1 -> v1.9.0-rc1 2025-10-10T01:18:34.3809322Z * [new tag] v1.9.0-rc2 -> v1.9.0-rc2 2025-10-10T01:18:34.3810644Z * [new tag] v1.9.0-rc3 -> v1.9.0-rc3 2025-10-10T01:18:34.3811783Z * [new tag] v1.9.0-rc4 -> v1.9.0-rc4 2025-10-10T01:18:34.3813169Z * [new tag] v1.9.1 -> v1.9.1 2025-10-10T01:18:34.3814666Z * [new tag] v1.9.1-rc1 -> v1.9.1-rc1 2025-10-10T01:18:34.3815963Z * [new tag] v1.9.1-rc2 -> v1.9.1-rc2 2025-10-10T01:18:34.3817390Z * [new tag] v2.0.0 -> v2.0.0 2025-10-10T01:18:34.3818638Z * [new tag] v2.0.0-rc1 -> v2.0.0-rc1 2025-10-10T01:18:34.3819996Z * [new tag] v2.0.0-rc2 -> v2.0.0-rc2 2025-10-10T01:18:34.3821367Z * [new tag] v2.0.0-rc3 -> v2.0.0-rc3 2025-10-10T01:18:34.3822629Z * [new tag] v2.0.0-rc4 -> v2.0.0-rc4 2025-10-10T01:18:34.3824204Z * [new tag] v2.0.0-rc5 -> v2.0.0-rc5 2025-10-10T01:18:34.3825589Z * [new tag] v2.0.0-rc6 -> v2.0.0-rc6 2025-10-10T01:18:34.3826874Z * [new tag] v2.0.1 -> v2.0.1 2025-10-10T01:18:34.3828247Z * [new tag] v2.0.1-rc1 -> v2.0.1-rc1 2025-10-10T01:18:34.3829364Z * [new tag] v2.0.1-rc2 -> v2.0.1-rc2 2025-10-10T01:18:34.3830627Z * [new tag] v2.0.1-rc3 -> v2.0.1-rc3 2025-10-10T01:18:34.3831634Z * [new tag] v2.0.1-rc4 -> v2.0.1-rc4 2025-10-10T01:18:34.3833446Z * [new tag] v2.1.0 -> v2.1.0 2025-10-10T01:18:34.3834772Z * [new tag] v2.1.0-rc1 -> v2.1.0-rc1 2025-10-10T01:18:34.3836477Z * [new tag] v2.1.0-rc2 -> v2.1.0-rc2 2025-10-10T01:18:34.3837913Z * [new tag] v2.1.0-rc3 -> v2.1.0-rc3 2025-10-10T01:18:34.3839439Z * [new tag] v2.1.0-rc4 -> v2.1.0-rc4 2025-10-10T01:18:34.3840917Z * [new tag] v2.1.0-rc5 -> v2.1.0-rc5 2025-10-10T01:18:34.3841738Z * [new tag] v2.1.0-rc6 -> v2.1.0-rc6 2025-10-10T01:18:34.3843218Z * [new tag] v2.1.1 -> v2.1.1 2025-10-10T01:18:34.3844609Z * [new tag] v2.1.1-rc1 -> v2.1.1-rc1 2025-10-10T01:18:34.3845941Z * [new tag] v2.1.1-rc2 -> v2.1.1-rc2 2025-10-10T01:18:34.3847245Z * [new tag] v2.1.1-rc3 -> v2.1.1-rc3 2025-10-10T01:18:34.3848656Z * [new tag] v2.1.1-rc4 -> v2.1.1-rc4 2025-10-10T01:18:34.3849959Z * [new tag] v2.1.1-rc5 -> v2.1.1-rc5 2025-10-10T01:18:34.3851080Z * [new tag] v2.1.1-rc6 -> v2.1.1-rc6 2025-10-10T01:18:34.3852375Z * [new tag] v2.1.2 -> v2.1.2 2025-10-10T01:18:34.3853769Z * [new tag] v2.1.2-rc1 -> v2.1.2-rc1 2025-10-10T01:18:34.3855161Z * [new tag] v2.1.2-rc2 -> v2.1.2-rc2 2025-10-10T01:18:34.3856288Z * [new tag] v2.1.2-rc3 -> v2.1.2-rc3 2025-10-10T01:18:34.3857692Z * [new tag] v2.2.0 -> v2.2.0 2025-10-10T01:18:34.3859011Z * [new tag] v2.2.0-rc1 -> v2.2.0-rc1 2025-10-10T01:18:34.3860212Z * [new tag] v2.2.0-rc2 -> v2.2.0-rc2 2025-10-10T01:18:34.3861478Z * [new tag] v2.2.0-rc3 -> v2.2.0-rc3 2025-10-10T01:18:34.3862783Z * [new tag] v2.2.0-rc4 -> v2.2.0-rc4 2025-10-10T01:18:34.3864113Z * [new tag] v2.2.0-rc5 -> v2.2.0-rc5 2025-10-10T01:18:34.3865479Z * [new tag] v2.2.0-rc6 -> v2.2.0-rc6 2025-10-10T01:18:34.3866541Z * [new tag] v2.2.0-rc7 -> v2.2.0-rc7 2025-10-10T01:18:34.3867742Z * [new tag] v2.2.0-rc8 -> v2.2.0-rc8 2025-10-10T01:18:34.3869141Z * [new tag] v2.2.1 -> v2.2.1 2025-10-10T01:18:34.3870599Z * [new tag] v2.2.1-rc1 -> v2.2.1-rc1 2025-10-10T01:18:34.3871605Z * [new tag] v2.2.1-rc2 -> v2.2.1-rc2 2025-10-10T01:18:34.3872862Z * [new tag] v2.2.1-rc3 -> v2.2.1-rc3 2025-10-10T01:18:34.3873851Z * [new tag] v2.2.2 -> v2.2.2 2025-10-10T01:18:34.3875476Z * [new tag] v2.2.2-rc1 -> v2.2.2-rc1 2025-10-10T01:18:34.3876640Z * [new tag] v2.2.2-rc2 -> v2.2.2-rc2 2025-10-10T01:18:34.3877732Z * [new tag] v2.2.2-rc3 -> v2.2.2-rc3 2025-10-10T01:18:34.3879148Z * [new tag] v2.3.0 -> v2.3.0 2025-10-10T01:18:34.3880456Z * [new tag] v2.3.0-rc1 -> v2.3.0-rc1 2025-10-10T01:18:34.3881842Z * [new tag] v2.3.0-rc10 -> v2.3.0-rc10 2025-10-10T01:18:34.3883243Z * [new tag] v2.3.0-rc11 -> v2.3.0-rc11 2025-10-10T01:18:34.3884376Z * [new tag] v2.3.0-rc12 -> v2.3.0-rc12 2025-10-10T01:18:34.3885738Z * [new tag] v2.3.0-rc2 -> v2.3.0-rc2 2025-10-10T01:18:34.3887112Z * [new tag] v2.3.0-rc3 -> v2.3.0-rc3 2025-10-10T01:18:34.3888399Z * [new tag] v2.3.0-rc4 -> v2.3.0-rc4 2025-10-10T01:18:34.3889700Z * [new tag] v2.3.0-rc5 -> v2.3.0-rc5 2025-10-10T01:18:34.3890816Z * [new tag] v2.3.0-rc6 -> v2.3.0-rc6 2025-10-10T01:18:34.3892096Z * [new tag] v2.3.0-rc7 -> v2.3.0-rc7 2025-10-10T01:18:34.3893461Z * [new tag] v2.3.0-rc8 -> v2.3.0-rc8 2025-10-10T01:18:34.3894690Z * [new tag] v2.3.0-rc9 -> v2.3.0-rc9 2025-10-10T01:18:34.3895535Z * [new tag] v2.3.1 -> v2.3.1 2025-10-10T01:18:34.3897104Z * [new tag] v2.3.1-rc1 -> v2.3.1-rc1 2025-10-10T01:18:34.3898381Z * [new tag] v2.3.1-rc2 -> v2.3.1-rc2 2025-10-10T01:18:34.3899713Z * [new tag] v2.3.1-rc3 -> v2.3.1-rc3 2025-10-10T01:18:34.3901026Z * [new tag] v2.4.0 -> v2.4.0 2025-10-10T01:18:34.3902318Z * [new tag] v2.4.0-rc1 -> v2.4.0-rc1 2025-10-10T01:18:34.3903640Z * [new tag] v2.4.0-rc2 -> v2.4.0-rc2 2025-10-10T01:18:34.3905099Z * [new tag] v2.4.0-rc3 -> v2.4.0-rc3 2025-10-10T01:18:34.3906355Z * [new tag] v2.4.0-rc4 -> v2.4.0-rc4 2025-10-10T01:18:34.3907786Z * [new tag] v2.4.0-rc5 -> v2.4.0-rc5 2025-10-10T01:18:34.3909134Z * [new tag] v2.4.0-rc6 -> v2.4.0-rc6 2025-10-10T01:18:34.3910507Z * [new tag] v2.4.0-rc7 -> v2.4.0-rc7 2025-10-10T01:18:34.3911850Z * [new tag] v2.4.0-rc8 -> v2.4.0-rc8 2025-10-10T01:18:34.3913167Z * [new tag] v2.4.0-rc9 -> v2.4.0-rc9 2025-10-10T01:18:34.3914161Z * [new tag] v2.4.1 -> v2.4.1 2025-10-10T01:18:34.3915636Z * [new tag] v2.4.1-rc1 -> v2.4.1-rc1 2025-10-10T01:18:34.3916977Z * [new tag] v2.4.1-rc2 -> v2.4.1-rc2 2025-10-10T01:18:34.3918331Z * [new tag] v2.4.1-rc3 -> v2.4.1-rc3 2025-10-10T01:18:34.3920147Z * [new tag] v2.5.0 -> v2.5.0 2025-10-10T01:18:34.3921443Z * [new tag] v2.5.0-rc1 -> v2.5.0-rc1 2025-10-10T01:18:34.3922601Z * [new tag] v2.5.0-rc10 -> v2.5.0-rc10 2025-10-10T01:18:34.3923898Z * [new tag] v2.5.0-rc2 -> v2.5.0-rc2 2025-10-10T01:18:34.3925594Z * [new tag] v2.5.0-rc3 -> v2.5.0-rc3 2025-10-10T01:18:34.3926924Z * [new tag] v2.5.0-rc4 -> v2.5.0-rc4 2025-10-10T01:18:34.3928296Z * [new tag] v2.5.0-rc5 -> v2.5.0-rc5 2025-10-10T01:18:34.3929616Z * [new tag] v2.5.0-rc6 -> v2.5.0-rc6 2025-10-10T01:18:34.3930936Z * [new tag] v2.5.0-rc7 -> v2.5.0-rc7 2025-10-10T01:18:34.3932260Z * [new tag] v2.5.0-rc8 -> v2.5.0-rc8 2025-10-10T01:18:34.3933721Z * [new tag] v2.5.0-rc9 -> v2.5.0-rc9 2025-10-10T01:18:34.3934847Z * [new tag] v2.5.1 -> v2.5.1 2025-10-10T01:18:34.3935846Z * [new tag] v2.5.1-rc1 -> v2.5.1-rc1 2025-10-10T01:18:34.3937107Z * [new tag] v2.6.0 -> v2.6.0 2025-10-10T01:18:34.3938500Z * [new tag] v2.6.0-rc1 -> v2.6.0-rc1 2025-10-10T01:18:34.3939902Z * [new tag] v2.6.0-rc2 -> v2.6.0-rc2 2025-10-10T01:18:34.3941239Z * [new tag] v2.6.0-rc3 -> v2.6.0-rc3 2025-10-10T01:18:34.3942554Z * [new tag] v2.6.0-rc4 -> v2.6.0-rc4 2025-10-10T01:18:34.3944082Z * [new tag] v2.6.0-rc5 -> v2.6.0-rc5 2025-10-10T01:18:34.3945437Z * [new tag] v2.6.0-rc6 -> v2.6.0-rc6 2025-10-10T01:18:34.3946741Z * [new tag] v2.6.0-rc7 -> v2.6.0-rc7 2025-10-10T01:18:34.3948112Z * [new tag] v2.6.0-rc8 -> v2.6.0-rc8 2025-10-10T01:18:34.3949721Z * [new tag] v2.6.0-rc9 -> v2.6.0-rc9 2025-10-10T01:18:34.3951029Z * [new tag] v2.7.0 -> v2.7.0 2025-10-10T01:18:34.3952379Z * [new tag] v2.7.0-rc1 -> v2.7.0-rc1 2025-10-10T01:18:34.3953550Z * [new tag] v2.7.0-rc10 -> v2.7.0-rc10 2025-10-10T01:18:34.3954994Z * [new tag] v2.7.0-rc2 -> v2.7.0-rc2 2025-10-10T01:18:34.3956349Z * [new tag] v2.7.0-rc3 -> v2.7.0-rc3 2025-10-10T01:18:34.3957744Z * [new tag] v2.7.0-rc4 -> v2.7.0-rc4 2025-10-10T01:18:34.3959110Z * [new tag] v2.7.0-rc5 -> v2.7.0-rc5 2025-10-10T01:18:34.3960476Z * [new tag] v2.7.0-rc6 -> v2.7.0-rc6 2025-10-10T01:18:34.3961801Z * [new tag] v2.7.0-rc7 -> v2.7.0-rc7 2025-10-10T01:18:34.3963128Z * [new tag] v2.7.0-rc8 -> v2.7.0-rc8 2025-10-10T01:18:34.3964507Z * [new tag] v2.7.0-rc9 -> v2.7.0-rc9 2025-10-10T01:18:34.3965639Z * [new tag] v2.7.1 -> v2.7.1 2025-10-10T01:18:34.3967047Z * [new tag] v2.7.1-rc1 -> v2.7.1-rc1 2025-10-10T01:18:34.3968423Z * [new tag] v2.7.1-rc2 -> v2.7.1-rc2 2025-10-10T01:18:34.3969796Z * [new tag] v2.7.1-rc3 -> v2.7.1-rc3 2025-10-10T01:18:34.3971199Z * [new tag] v2.7.1-rc4 -> v2.7.1-rc4 2025-10-10T01:18:34.3972519Z * [new tag] v2.7.1-rc5 -> v2.7.1-rc5 2025-10-10T01:18:34.3973691Z * [new tag] v2.8.0 -> v2.8.0 2025-10-10T01:18:34.3975067Z * [new tag] v2.8.0-rc1 -> v2.8.0-rc1 2025-10-10T01:18:34.3976462Z * [new tag] v2.8.0-rc2 -> v2.8.0-rc2 2025-10-10T01:18:34.3977830Z * [new tag] v2.8.0-rc3 -> v2.8.0-rc3 2025-10-10T01:18:34.3979232Z * [new tag] v2.8.0-rc4 -> v2.8.0-rc4 2025-10-10T01:18:34.3980594Z * [new tag] v2.8.0-rc5 -> v2.8.0-rc5 2025-10-10T01:18:34.3982026Z * [new tag] v2.8.0-rc6 -> v2.8.0-rc6 2025-10-10T01:18:34.3983397Z * [new tag] v2.8.0-rc7 -> v2.8.0-rc7 2025-10-10T01:18:34.3984701Z * [new tag] v2.8.0-rc8 -> v2.8.0-rc8 2025-10-10T01:18:34.3986072Z * [new tag] v2.9.0-rc1 -> v2.9.0-rc1 2025-10-10T01:18:34.3987693Z * [new tag] v2.9.0-rc2 -> v2.9.0-rc2 2025-10-10T01:18:34.3989022Z * [new tag] v2.9.0-rc3 -> v2.9.0-rc3 2025-10-10T01:18:34.3990383Z * [new tag] v2.9.0-rc4 -> v2.9.0-rc4 2025-10-10T01:18:34.3991605Z * [new tag] v2.9.0-rc5 -> v2.9.0-rc5 2025-10-10T01:18:34.3993375Z * [new tag] v2.9.0-rc6 -> v2.9.0-rc6 2025-10-10T01:18:34.3994512Z * [new tag] v2.9.0-rc7 -> v2.9.0-rc7 2025-10-10T01:18:34.3996316Z * [new tag] v2.9.0-rc8 -> v2.9.0-rc8 2025-10-10T01:18:34.3997498Z * [new tag] v2.9.0-rc9 -> v2.9.0-rc9 2025-10-10T01:18:34.3999511Z * [new tag] viable/strict/1759343184 -> viable/strict/1759343184 2025-10-10T01:18:34.4000736Z * [new tag] viable/strict/1759346540 -> viable/strict/1759346540 2025-10-10T01:18:34.4001862Z * [new tag] viable/strict/1759348181 -> viable/strict/1759348181 2025-10-10T01:18:34.4003136Z * [new tag] viable/strict/1759350324 -> viable/strict/1759350324 2025-10-10T01:18:34.4004541Z * [new tag] viable/strict/1759351793 -> viable/strict/1759351793 2025-10-10T01:18:34.4006186Z * [new tag] viable/strict/1759353844 -> viable/strict/1759353844 2025-10-10T01:18:34.4007466Z * [new tag] viable/strict/1759355374 -> viable/strict/1759355374 2025-10-10T01:18:34.4008506Z * [new tag] viable/strict/1759357472 -> viable/strict/1759357472 2025-10-10T01:18:34.4009879Z * [new tag] viable/strict/1759361002 -> viable/strict/1759361002 2025-10-10T01:18:34.4011397Z * [new tag] viable/strict/1759362585 -> viable/strict/1759362585 2025-10-10T01:18:34.4012944Z * [new tag] viable/strict/1759365359 -> viable/strict/1759365359 2025-10-10T01:18:34.4014311Z * [new tag] viable/strict/1759370089 -> viable/strict/1759370089 2025-10-10T01:18:34.4015677Z * [new tag] viable/strict/1759377554 -> viable/strict/1759377554 2025-10-10T01:18:34.4016886Z * [new tag] viable/strict/1759379133 -> viable/strict/1759379133 2025-10-10T01:18:34.4018081Z * [new tag] viable/strict/1759389871 -> viable/strict/1759389871 2025-10-10T01:18:34.4019316Z * [new tag] viable/strict/1759393562 -> viable/strict/1759393562 2025-10-10T01:18:34.4020350Z * [new tag] viable/strict/1759395076 -> viable/strict/1759395076 2025-10-10T01:18:34.4021726Z * [new tag] viable/strict/1759398579 -> viable/strict/1759398579 2025-10-10T01:18:34.4022923Z * [new tag] viable/strict/1759404142 -> viable/strict/1759404142 2025-10-10T01:18:34.4024274Z * [new tag] viable/strict/1759405773 -> viable/strict/1759405773 2025-10-10T01:18:34.4025696Z * [new tag] viable/strict/1759408041 -> viable/strict/1759408041 2025-10-10T01:18:34.4026857Z * [new tag] viable/strict/1759411593 -> viable/strict/1759411593 2025-10-10T01:18:34.4028071Z * [new tag] viable/strict/1759427395 -> viable/strict/1759427395 2025-10-10T01:18:34.4029252Z * [new tag] viable/strict/1759434582 -> viable/strict/1759434582 2025-10-10T01:18:34.4030479Z * [new tag] viable/strict/1759436720 -> viable/strict/1759436720 2025-10-10T01:18:34.4031675Z * [new tag] viable/strict/1759440219 -> viable/strict/1759440219 2025-10-10T01:18:34.4032850Z * [new tag] viable/strict/1759441948 -> viable/strict/1759441948 2025-10-10T01:18:34.4034118Z * [new tag] viable/strict/1759443860 -> viable/strict/1759443860 2025-10-10T01:18:34.4035397Z * [new tag] viable/strict/1759445377 -> viable/strict/1759445377 2025-10-10T01:18:34.4036595Z * [new tag] viable/strict/1759447415 -> viable/strict/1759447415 2025-10-10T01:18:34.4037788Z * [new tag] viable/strict/1759451750 -> viable/strict/1759451750 2025-10-10T01:18:34.4039078Z * [new tag] viable/strict/1759453910 -> viable/strict/1759453910 2025-10-10T01:18:34.4040275Z * [new tag] viable/strict/1759456483 -> viable/strict/1759456483 2025-10-10T01:18:34.4041453Z * [new tag] viable/strict/1759459279 -> viable/strict/1759459279 2025-10-10T01:18:34.4042758Z * [new tag] viable/strict/1759460742 -> viable/strict/1759460742 2025-10-10T01:18:34.4043985Z * [new tag] viable/strict/1759462025 -> viable/strict/1759462025 2025-10-10T01:18:34.4045157Z * [new tag] viable/strict/1759469086 -> viable/strict/1759469086 2025-10-10T01:18:34.4046392Z * [new tag] viable/strict/1759470581 -> viable/strict/1759470581 2025-10-10T01:18:34.4047556Z * [new tag] viable/strict/1759472786 -> viable/strict/1759472786 2025-10-10T01:18:34.4048745Z * [new tag] viable/strict/1759476294 -> viable/strict/1759476294 2025-10-10T01:18:34.4050102Z * [new tag] viable/strict/1759479963 -> viable/strict/1759479963 2025-10-10T01:18:34.4051011Z * [new tag] viable/strict/1759492177 -> viable/strict/1759492177 2025-10-10T01:18:34.4052367Z * [new tag] viable/strict/1759519278 -> viable/strict/1759519278 2025-10-10T01:18:34.4053563Z * [new tag] viable/strict/1759524580 -> viable/strict/1759524580 2025-10-10T01:18:34.4054736Z * [new tag] viable/strict/1759528193 -> viable/strict/1759528193 2025-10-10T01:18:34.4055933Z * [new tag] viable/strict/1759533797 -> viable/strict/1759533797 2025-10-10T01:18:34.4057151Z * [new tag] viable/strict/1759542780 -> viable/strict/1759542780 2025-10-10T01:18:34.4058342Z * [new tag] viable/strict/1759549779 -> viable/strict/1759549779 2025-10-10T01:18:34.4059569Z * [new tag] viable/strict/1759555455 -> viable/strict/1759555455 2025-10-10T01:18:34.4060759Z * [new tag] viable/strict/1759559176 -> viable/strict/1759559176 2025-10-10T01:18:34.4061974Z * [new tag] viable/strict/1759560629 -> viable/strict/1759560629 2025-10-10T01:18:34.4063167Z * [new tag] viable/strict/1759569848 -> viable/strict/1759569848 2025-10-10T01:18:34.4064507Z * [new tag] viable/strict/1759571382 -> viable/strict/1759571382 2025-10-10T01:18:34.4065657Z * [new tag] viable/strict/1759573474 -> viable/strict/1759573474 2025-10-10T01:18:34.4066791Z * [new tag] viable/strict/1759618187 -> viable/strict/1759618187 2025-10-10T01:18:34.4068046Z * [new tag] viable/strict/1759626742 -> viable/strict/1759626742 2025-10-10T01:18:34.4069253Z * [new tag] viable/strict/1759632427 -> viable/strict/1759632427 2025-10-10T01:18:34.4070472Z * [new tag] viable/strict/1759634971 -> viable/strict/1759634971 2025-10-10T01:18:34.4071662Z * [new tag] viable/strict/1759661382 -> viable/strict/1759661382 2025-10-10T01:18:34.4072875Z * [new tag] viable/strict/1759663294 -> viable/strict/1759663294 2025-10-10T01:18:34.4074112Z * [new tag] viable/strict/1759708178 -> viable/strict/1759708178 2025-10-10T01:18:34.4075308Z * [new tag] viable/strict/1759715695 -> viable/strict/1759715695 2025-10-10T01:18:34.4076482Z * [new tag] viable/strict/1759728293 -> viable/strict/1759728293 2025-10-10T01:18:34.4077733Z * [new tag] viable/strict/1759735513 -> viable/strict/1759735513 2025-10-10T01:18:34.4079049Z * [new tag] viable/strict/1759739177 -> viable/strict/1759739177 2025-10-10T01:18:34.4080378Z * [new tag] viable/strict/1759758635 -> viable/strict/1759758635 2025-10-10T01:18:34.4081588Z * [new tag] viable/strict/1759765784 -> viable/strict/1759765784 2025-10-10T01:18:34.4082825Z * [new tag] viable/strict/1759767948 -> viable/strict/1759767948 2025-10-10T01:18:34.4083989Z * [new tag] viable/strict/1759771461 -> viable/strict/1759771461 2025-10-10T01:18:34.4085661Z * [new tag] viable/strict/1759776706 -> viable/strict/1759776706 2025-10-10T01:18:34.4086905Z * [new tag] viable/strict/1759782317 -> viable/strict/1759782317 2025-10-10T01:18:34.4088127Z * [new tag] viable/strict/1759783777 -> viable/strict/1759783777 2025-10-10T01:18:34.4089308Z * [new tag] viable/strict/1759785815 -> viable/strict/1759785815 2025-10-10T01:18:34.4090467Z * [new tag] viable/strict/1759789459 -> viable/strict/1759789459 2025-10-10T01:18:34.4091670Z * [new tag] viable/strict/1759790974 -> viable/strict/1759790974 2025-10-10T01:18:34.4092926Z * [new tag] viable/strict/1759794583 -> viable/strict/1759794583 2025-10-10T01:18:34.4094170Z * [new tag] viable/strict/1759797408 -> viable/strict/1759797408 2025-10-10T01:18:34.4095233Z * [new tag] viable/strict/1759799518 -> viable/strict/1759799518 2025-10-10T01:18:34.4096479Z * [new tag] viable/strict/1759804909 -> viable/strict/1759804909 2025-10-10T01:18:34.4097684Z * [new tag] viable/strict/1759807643 -> viable/strict/1759807643 2025-10-10T01:18:34.4098896Z * [new tag] viable/strict/1759809089 -> viable/strict/1759809089 2025-10-10T01:18:34.4100104Z * [new tag] viable/strict/1759811145 -> viable/strict/1759811145 2025-10-10T01:18:34.4101299Z * [new tag] viable/strict/1759812581 -> viable/strict/1759812581 2025-10-10T01:18:34.4102478Z * [new tag] viable/strict/1759814683 -> viable/strict/1759814683 2025-10-10T01:18:34.4103653Z * [new tag] viable/strict/1759821889 -> viable/strict/1759821889 2025-10-10T01:18:34.4104868Z * [new tag] viable/strict/1759823376 -> viable/strict/1759823376 2025-10-10T01:18:34.4106069Z * [new tag] viable/strict/1759827107 -> viable/strict/1759827107 2025-10-10T01:18:34.4107355Z * [new tag] viable/strict/1759830577 -> viable/strict/1759830577 2025-10-10T01:18:34.4108606Z * [new tag] viable/strict/1759832720 -> viable/strict/1759832720 2025-10-10T01:18:34.4109763Z * [new tag] viable/strict/1759842063 -> viable/strict/1759842063 2025-10-10T01:18:34.4110985Z * [new tag] viable/strict/1759847121 -> viable/strict/1759847121 2025-10-10T01:18:34.4112165Z * [new tag] viable/strict/1759850721 -> viable/strict/1759850721 2025-10-10T01:18:34.4113428Z * [new tag] viable/strict/1759857870 -> viable/strict/1759857870 2025-10-10T01:18:34.4114656Z * [new tag] viable/strict/1759863143 -> viable/strict/1759863143 2025-10-10T01:18:34.4115860Z * [new tag] viable/strict/1759875874 -> viable/strict/1759875874 2025-10-10T01:18:34.4117064Z * [new tag] viable/strict/1759877385 -> viable/strict/1759877385 2025-10-10T01:18:34.4118258Z * [new tag] viable/strict/1759883801 -> viable/strict/1759883801 2025-10-10T01:18:34.4119574Z * [new tag] viable/strict/1759885922 -> viable/strict/1759885922 2025-10-10T01:18:34.4120785Z * [new tag] viable/strict/1759888488 -> viable/strict/1759888488 2025-10-10T01:18:34.4122010Z * [new tag] viable/strict/1759895471 -> viable/strict/1759895471 2025-10-10T01:18:34.4123171Z * [new tag] viable/strict/1759904803 -> viable/strict/1759904803 2025-10-10T01:18:34.4124497Z * [new tag] viable/strict/1759908300 -> viable/strict/1759908300 2025-10-10T01:18:34.4126209Z * [new tag] viable/strict/1759915520 -> viable/strict/1759915520 2025-10-10T01:18:34.4127380Z * [new tag] viable/strict/1759916978 -> viable/strict/1759916978 2025-10-10T01:18:34.4128571Z * [new tag] viable/strict/1759930024 -> viable/strict/1759930024 2025-10-10T01:18:34.4129757Z * [new tag] viable/strict/1759948122 -> viable/strict/1759948122 2025-10-10T01:18:34.4130976Z * [new tag] viable/strict/1759952983 -> viable/strict/1759952983 2025-10-10T01:18:34.4132192Z * [new tag] viable/strict/1759955121 -> viable/strict/1759955121 2025-10-10T01:18:34.4133387Z * [new tag] viable/strict/1759962298 -> viable/strict/1759962298 2025-10-10T01:18:34.4134574Z * [new tag] viable/strict/1759965837 -> viable/strict/1759965837 2025-10-10T01:18:34.4135785Z * [new tag] viable/strict/1759970213 -> viable/strict/1759970213 2025-10-10T01:18:34.4137016Z * [new tag] viable/strict/1759974894 -> viable/strict/1759974894 2025-10-10T01:18:34.4138416Z * [new tag] viable/strict/1759977763 -> viable/strict/1759977763 2025-10-10T01:18:34.4139498Z * [new tag] viable/strict/1759979241 -> viable/strict/1759979241 2025-10-10T01:18:34.4140746Z * [new tag] viable/strict/1759985417 -> viable/strict/1759985417 2025-10-10T01:18:34.4141931Z * [new tag] viable/strict/1759987490 -> viable/strict/1759987490 2025-10-10T01:18:34.4143136Z * [new tag] viable/strict/1759996180 -> viable/strict/1759996180 2025-10-10T01:18:34.4144476Z * [new tag] whc_flight_1 -> whc_flight_1 2025-10-10T01:18:34.4145923Z * [new tag] whc_flight_2 -> whc_flight_2 2025-10-10T01:18:34.4147423Z * [new tag] whc_flight_4 -> whc_flight_4 2025-10-10T01:18:34.5253102Z [command]/usr/bin/git rev-parse --verify --quiet 344e6365a0068c2d2847fcec0c55dd53291d475e^{object} 2025-10-10T01:18:34.5289413Z 344e6365a0068c2d2847fcec0c55dd53291d475e 2025-10-10T01:18:34.5294406Z ##[endgroup] 2025-10-10T01:18:34.5295124Z ##[group]Determining the checkout info 2025-10-10T01:18:34.5296010Z ##[endgroup] 2025-10-10T01:18:34.5300960Z [command]/usr/bin/git sparse-checkout disable 2025-10-10T01:18:34.5354665Z [command]/usr/bin/git config --local --unset-all extensions.worktreeConfig 2025-10-10T01:18:34.5392670Z ##[group]Checking out the ref 2025-10-10T01:18:34.5394944Z [command]/usr/bin/git checkout --progress --force 344e6365a0068c2d2847fcec0c55dd53291d475e 2025-10-10T01:18:35.5671525Z Updating files: 73% (14689/19929) 2025-10-10T01:18:35.5941478Z Updating files: 74% (14748/19929) 2025-10-10T01:18:35.6331648Z Updating files: 75% (14947/19929) 2025-10-10T01:18:35.6563946Z Updating files: 76% (15147/19929) 2025-10-10T01:18:35.6716890Z Updating files: 77% (15346/19929) 2025-10-10T01:18:35.6875739Z Updating files: 78% (15545/19929) 2025-10-10T01:18:35.7167400Z Updating files: 79% (15744/19929) 2025-10-10T01:18:35.7441143Z Updating files: 80% (15944/19929) 2025-10-10T01:18:35.7697390Z Updating files: 81% (16143/19929) 2025-10-10T01:18:35.7952231Z Updating files: 82% (16342/19929) 2025-10-10T01:18:35.8133787Z Updating files: 83% (16542/19929) 2025-10-10T01:18:35.8290052Z Updating files: 84% (16741/19929) 2025-10-10T01:18:35.8469273Z Updating files: 85% (16940/19929) 2025-10-10T01:18:35.8643596Z Updating files: 86% (17139/19929) 2025-10-10T01:18:35.8804104Z Updating files: 87% (17339/19929) 2025-10-10T01:18:35.8952628Z Updating files: 88% (17538/19929) 2025-10-10T01:18:35.9107923Z Updating files: 89% (17737/19929) 2025-10-10T01:18:35.9298869Z Updating files: 90% (17937/19929) 2025-10-10T01:18:35.9455661Z Updating files: 91% (18136/19929) 2025-10-10T01:18:35.9618801Z Updating files: 92% (18335/19929) 2025-10-10T01:18:35.9829539Z Updating files: 93% (18534/19929) 2025-10-10T01:18:36.0044835Z Updating files: 94% (18734/19929) 2025-10-10T01:18:36.0241341Z Updating files: 95% (18933/19929) 2025-10-10T01:18:36.0418079Z Updating files: 96% (19132/19929) 2025-10-10T01:18:36.0601907Z Updating files: 97% (19332/19929) 2025-10-10T01:18:36.0877059Z Updating files: 98% (19531/19929) 2025-10-10T01:18:36.1066837Z Updating files: 99% (19730/19929) 2025-10-10T01:18:36.1067379Z Updating files: 100% (19929/19929) 2025-10-10T01:18:36.1067888Z Updating files: 100% (19929/19929), done. 2025-10-10T01:18:36.1301808Z Note: switching to '344e6365a0068c2d2847fcec0c55dd53291d475e'. 2025-10-10T01:18:36.1302381Z 2025-10-10T01:18:36.1302780Z You are in 'detached HEAD' state. You can look around, make experimental 2025-10-10T01:18:36.1303759Z changes and commit them, and you can discard any commits you make in this 2025-10-10T01:18:36.1304723Z state without impacting any branches by switching back to a branch. 2025-10-10T01:18:36.1305282Z 2025-10-10T01:18:36.1305636Z If you want to create a new branch to retain commits you create, you may 2025-10-10T01:18:36.1306494Z do so (now or later) by using -c with the switch command. Example: 2025-10-10T01:18:36.1306973Z 2025-10-10T01:18:36.1307361Z git switch -c 2025-10-10T01:18:36.1307520Z 2025-10-10T01:18:36.1307600Z Or undo this operation with: 2025-10-10T01:18:36.1307736Z 2025-10-10T01:18:36.1307806Z git switch - 2025-10-10T01:18:36.1307929Z 2025-10-10T01:18:36.1308127Z Turn off this advice by setting config variable advice.detachedHead to false 2025-10-10T01:18:36.1308380Z 2025-10-10T01:18:36.1308651Z HEAD is now at 344e6365a00 [inductor][eazy] change how torch.use_deterministic_algorithms affect inductor (#164905) 2025-10-10T01:18:36.1432489Z ##[endgroup] 2025-10-10T01:18:36.1433279Z ##[group]Setting up auth for fetching submodules 2025-10-10T01:18:36.1438315Z [command]/usr/bin/git config --global http.https://github.com/.extraheader AUTHORIZATION: basic *** 2025-10-10T01:18:36.1501277Z [command]/usr/bin/git config --global --unset-all url.https://github.com/.insteadOf 2025-10-10T01:18:36.1541811Z [command]/usr/bin/git config --global --add url.https://github.com/.insteadOf git@github.com: 2025-10-10T01:18:36.1581032Z [command]/usr/bin/git config --global --add url.https://github.com/.insteadOf org-21003710@github.com: 2025-10-10T01:18:36.1620077Z ##[endgroup] 2025-10-10T01:18:36.1620797Z ##[group]Fetching submodules 2025-10-10T01:18:36.1622145Z [command]/usr/bin/git submodule sync --recursive 2025-10-10T01:18:36.2112601Z [command]/usr/bin/git -c protocol.version=2 submodule update --init --force --recursive 2025-10-10T01:18:36.2591793Z Submodule 'android/libs/fbjni' (https://github.com/facebookincubator/fbjni.git) registered for path 'android/libs/fbjni' 2025-10-10T01:18:36.2594856Z Submodule 'third_party/NNPACK_deps/FP16' (https://github.com/Maratyszcza/FP16.git) registered for path 'third_party/FP16' 2025-10-10T01:18:36.2597339Z Submodule 'third_party/NNPACK_deps/FXdiv' (https://github.com/Maratyszcza/FXdiv.git) registered for path 'third_party/FXdiv' 2025-10-10T01:18:36.2600805Z Submodule 'third_party/NNPACK' (https://github.com/Maratyszcza/NNPACK.git) registered for path 'third_party/NNPACK' 2025-10-10T01:18:36.2604546Z Submodule 'third_party/NVTX' (https://github.com/NVIDIA/NVTX.git) registered for path 'third_party/NVTX' 2025-10-10T01:18:36.2608730Z Submodule 'third_party/VulkanMemoryAllocator' (https://github.com/GPUOpen-LibrariesAndSDKs/VulkanMemoryAllocator.git) registered for path 'third_party/VulkanMemoryAllocator' 2025-10-10T01:18:36.2612493Z Submodule 'third_party/XNNPACK' (https://github.com/google/XNNPACK.git) registered for path 'third_party/XNNPACK' 2025-10-10T01:18:36.2617801Z Submodule 'third_party/aiter' (https://github.com/ROCm/aiter.git) registered for path 'third_party/aiter' 2025-10-10T01:18:36.2622063Z Submodule 'third_party/benchmark' (https://github.com/google/benchmark.git) registered for path 'third_party/benchmark' 2025-10-10T01:18:36.2627058Z Submodule 'third_party/composable_kernel' (https://github.com/ROCm/composable_kernel.git) registered for path 'third_party/composable_kernel' 2025-10-10T01:18:36.2641646Z Submodule 'third_party/cpp-httplib' (https://github.com/yhirose/cpp-httplib.git) registered for path 'third_party/cpp-httplib' 2025-10-10T01:18:36.2642515Z Submodule 'third_party/cpuinfo' (https://github.com/pytorch/cpuinfo.git) registered for path 'third_party/cpuinfo' 2025-10-10T01:18:36.2643298Z Submodule 'third_party/cudnn_frontend' (https://github.com/NVIDIA/cudnn-frontend.git) registered for path 'third_party/cudnn_frontend' 2025-10-10T01:18:36.2644028Z Submodule 'third_party/cutlass' (https://github.com/NVIDIA/cutlass.git) registered for path 'third_party/cutlass' 2025-10-10T01:18:36.2648582Z Submodule 'third_party/fbgemm' (https://github.com/pytorch/fbgemm) registered for path 'third_party/fbgemm' 2025-10-10T01:18:36.2654555Z Submodule 'third_party/flash-attention' (https://github.com/Dao-AILab/flash-attention.git) registered for path 'third_party/flash-attention' 2025-10-10T01:18:36.2662073Z Submodule 'third_party/flatbuffers' (https://github.com/google/flatbuffers.git) registered for path 'third_party/flatbuffers' 2025-10-10T01:18:36.2667994Z Submodule 'third_party/fmt' (https://github.com/fmtlib/fmt.git) registered for path 'third_party/fmt' 2025-10-10T01:18:36.2672755Z Submodule 'third_party/gemmlowp/gemmlowp' (https://github.com/google/gemmlowp.git) registered for path 'third_party/gemmlowp/gemmlowp' 2025-10-10T01:18:36.2677328Z Submodule 'third_party/gloo' (https://github.com/pytorch/gloo) registered for path 'third_party/gloo' 2025-10-10T01:18:36.2683064Z Submodule 'third_party/googletest' (https://github.com/google/googletest.git) registered for path 'third_party/googletest' 2025-10-10T01:18:36.2687492Z Submodule 'third_party/ideep' (https://github.com/intel/ideep) registered for path 'third_party/ideep' 2025-10-10T01:18:36.2692470Z Submodule 'third_party/ittapi' (https://github.com/intel/ittapi.git) registered for path 'third_party/ittapi' 2025-10-10T01:18:36.2698431Z Submodule 'third_party/kineto' (https://github.com/pytorch/kineto) registered for path 'third_party/kineto' 2025-10-10T01:18:36.2704005Z Submodule 'third_party/kleidiai' (https://github.com/ARM-software/kleidiai.git) registered for path 'third_party/kleidiai' 2025-10-10T01:18:36.2708832Z Submodule 'third_party/mimalloc' (https://github.com/microsoft/mimalloc.git) registered for path 'third_party/mimalloc' 2025-10-10T01:18:36.2714374Z Submodule 'third_party/nlohmann' (https://github.com/nlohmann/json.git) registered for path 'third_party/nlohmann' 2025-10-10T01:18:36.2720123Z Submodule 'third_party/onnx' (https://github.com/onnx/onnx.git) registered for path 'third_party/onnx' 2025-10-10T01:18:36.2726446Z Submodule 'third_party/opentelemetry-cpp' (https://github.com/open-telemetry/opentelemetry-cpp.git) registered for path 'third_party/opentelemetry-cpp' 2025-10-10T01:18:36.2732001Z Submodule 'third_party/pocketfft' (https://github.com/mreineck/pocketfft) registered for path 'third_party/pocketfft' 2025-10-10T01:18:36.2738930Z Submodule 'third_party/protobuf' (https://github.com/protocolbuffers/protobuf.git) registered for path 'third_party/protobuf' 2025-10-10T01:18:36.2744915Z Submodule 'third_party/NNPACK_deps/psimd' (https://github.com/Maratyszcza/psimd.git) registered for path 'third_party/psimd' 2025-10-10T01:18:36.2750173Z Submodule 'third_party/NNPACK_deps/pthreadpool' (https://github.com/Maratyszcza/pthreadpool.git) registered for path 'third_party/pthreadpool' 2025-10-10T01:18:36.2759691Z Submodule 'third_party/pybind11' (https://github.com/pybind/pybind11.git) registered for path 'third_party/pybind11' 2025-10-10T01:18:36.2766076Z Submodule 'third_party/python-peachpy' (https://github.com/malfet/PeachPy.git) registered for path 'third_party/python-peachpy' 2025-10-10T01:18:36.2772232Z Submodule 'third_party/sleef' (https://github.com/shibatch/sleef) registered for path 'third_party/sleef' 2025-10-10T01:18:36.2777970Z Submodule 'third_party/tensorpipe' (https://github.com/pytorch/tensorpipe.git) registered for path 'third_party/tensorpipe' 2025-10-10T01:18:36.2821157Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/android/libs/fbjni'... 2025-10-10T01:18:36.5355344Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/FXdiv'... 2025-10-10T01:18:36.5356309Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/FP16'... 2025-10-10T01:18:36.5357223Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/pocketfft'... 2025-10-10T01:18:36.5358143Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/psimd'... 2025-10-10T01:18:36.5359185Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/pthreadpool'... 2025-10-10T01:18:36.5384253Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/NNPACK'... 2025-10-10T01:18:36.6089982Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/NVTX'... 2025-10-10T01:18:36.7473597Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/python-peachpy'... 2025-10-10T01:18:36.7474603Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/ideep'... 2025-10-10T01:18:36.7475772Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/gloo'... 2025-10-10T01:18:36.7476745Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/gemmlowp/gemmlowp'... 2025-10-10T01:18:36.7477729Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/benchmark'... 2025-10-10T01:18:36.7566799Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/VulkanMemoryAllocator'... 2025-10-10T01:18:37.9316313Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/ittapi'... 2025-10-10T01:18:37.9317163Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/tensorpipe'... 2025-10-10T01:18:37.9318082Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/flash-attention'... 2025-10-10T01:18:37.9319124Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/kleidiai'... 2025-10-10T01:18:37.9320075Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/cpuinfo'... 2025-10-10T01:18:37.9320964Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/cpp-httplib'... 2025-10-10T01:18:37.9321967Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/sleef'... 2025-10-10T01:18:37.9322505Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/mimalloc'... 2025-10-10T01:18:37.9323019Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/googletest'... 2025-10-10T01:18:37.9323688Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/pybind11'... 2025-10-10T01:18:37.9324477Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/fmt'... 2025-10-10T01:18:37.9325119Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/cudnn_frontend'... 2025-10-10T01:18:37.9326062Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/kineto'... 2025-10-10T01:18:38.0317370Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/XNNPACK'... 2025-10-10T01:18:47.1592543Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/flatbuffers'... 2025-10-10T01:18:47.1593962Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/onnx'... 2025-10-10T01:18:47.1595152Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/fbgemm'... 2025-10-10T01:18:47.1596330Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/cutlass'... 2025-10-10T01:18:47.1597596Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/composable_kernel'... 2025-10-10T01:18:47.1598974Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/aiter'... 2025-10-10T01:18:47.1600393Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/opentelemetry-cpp'... 2025-10-10T01:18:47.1601916Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/nlohmann'... 2025-10-10T01:18:47.1603299Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/protobuf'... 2025-10-10T01:18:47.1828482Z Submodule path 'android/libs/fbjni': checked out '7e1e1fe3858c63c251c637ae41a20de425dde96f' 2025-10-10T01:18:47.2003497Z Submodule path 'third_party/FP16': checked out '4dfe081cf6bcd15db339cf2680b9281b8451eeb3' 2025-10-10T01:18:47.2154682Z Submodule path 'third_party/FXdiv': checked out 'b408327ac2a15ec3e43352421954f5b1967701d1' 2025-10-10T01:18:47.2507220Z Submodule path 'third_party/NNPACK': checked out 'c07e3a0400713d546e0dea2d5466dd22ea389c73' 2025-10-10T01:18:47.3406217Z Submodule path 'third_party/NVTX': checked out '2942f167cc30c5e3a44a2aecd5b0d9c07ff61a07' 2025-10-10T01:18:47.3995732Z Submodule path 'third_party/VulkanMemoryAllocator': checked out '1d8f600fd424278486eade7ed3e877c99f0846b1' 2025-10-10T01:18:48.3269613Z Submodule path 'third_party/XNNPACK': checked out '51a0103656eff6fc9bfd39a4597923c4b542c883' 2025-10-10T01:18:48.5092097Z Submodule path 'third_party/aiter': checked out '01aae101b9e5e94d6c16a9514c9fb8df99c93150' 2025-10-10T01:18:48.5121048Z Submodule '3rdparty/composable_kernel' (https://github.com/ROCm/composable_kernel.git) registered for path 'third_party/aiter/3rdparty/composable_kernel' 2025-10-10T01:18:48.5165205Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/aiter/3rdparty/composable_kernel'... 2025-10-10T01:18:52.1127803Z Submodule path 'third_party/aiter/3rdparty/composable_kernel': checked out 'cffe8fa2a442ac8e80dd236a1a5d24fe3d7e0cbf' 2025-10-10T01:18:52.1465318Z Submodule path 'third_party/benchmark': checked out '299e5928955cc62af9968370293b916f5130916f' 2025-10-10T01:18:52.5740118Z Submodule path 'third_party/composable_kernel': checked out '7fe50dc3da2069d6645d9deb8c017a876472a977' 2025-10-10T01:18:52.6306363Z Submodule path 'third_party/cpp-httplib': checked out '89c932f313c6437c38f2982869beacc89c2f2246' 2025-10-10T01:18:52.7365998Z Submodule path 'third_party/cpuinfo': checked out '5e3d2445e6a84d9599bee2bf78edbb4d80865e1d' 2025-10-10T01:18:52.7893443Z Submodule path 'third_party/cudnn_frontend': checked out 'f937055efc6d414d11f4c6577e3977fe74f35fb6' 2025-10-10T01:18:53.5264791Z Submodule path 'third_party/cutlass': checked out 'f3fde58372d33e9a5650ba7b80fc48b3b49d40c8' 2025-10-10T01:18:53.7109347Z Submodule path 'third_party/fbgemm': checked out '3cefe0564a8c3de514a152d40a2b4770f2ee5be0' 2025-10-10T01:18:53.7139220Z Submodule 'external/asmjit' (https://github.com/asmjit/asmjit.git) registered for path 'third_party/fbgemm/external/asmjit' 2025-10-10T01:18:53.7141596Z Submodule 'external/composable_kernel' (https://github.com/ROCm/composable_kernel.git) registered for path 'third_party/fbgemm/external/composable_kernel' 2025-10-10T01:18:53.7145159Z Submodule 'external/cpuinfo' (https://github.com/pytorch/cpuinfo) registered for path 'third_party/fbgemm/external/cpuinfo' 2025-10-10T01:18:53.7148848Z Submodule 'external/cutlass' (https://github.com/jwfromm/cutlass) registered for path 'third_party/fbgemm/external/cutlass' 2025-10-10T01:18:53.7153646Z Submodule 'external/googletest' (https://github.com/google/googletest) registered for path 'third_party/fbgemm/external/googletest' 2025-10-10T01:18:53.7158195Z Submodule 'external/hipify_torch' (https://github.com/ROCmSoftwarePlatform/hipify_torch.git) registered for path 'third_party/fbgemm/external/hipify_torch' 2025-10-10T01:18:53.7161681Z Submodule 'external/json' (https://github.com/nlohmann/json.git) registered for path 'third_party/fbgemm/external/json' 2025-10-10T01:18:53.7204096Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/fbgemm/external/asmjit'... 2025-10-10T01:18:54.8441197Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/fbgemm/external/hipify_torch'... 2025-10-10T01:18:54.8442384Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/fbgemm/external/cpuinfo'... 2025-10-10T01:18:54.8443545Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/fbgemm/external/googletest'... 2025-10-10T01:18:54.9442472Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/fbgemm/external/composable_kernel'... 2025-10-10T01:18:56.8849514Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/fbgemm/external/cutlass'... 2025-10-10T01:18:56.9850090Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/fbgemm/external/json'... 2025-10-10T01:18:59.3830560Z Submodule path 'third_party/fbgemm/external/asmjit': checked out 'a3199e8857792cd10b7589ff5d58343d2c9008ea' 2025-10-10T01:18:59.8059946Z Submodule path 'third_party/fbgemm/external/composable_kernel': checked out '7fe50dc3da2069d6645d9deb8c017a876472a977' 2025-10-10T01:18:59.9131582Z Submodule path 'third_party/fbgemm/external/cpuinfo': checked out '6543fec09b2f04ac4a666882998b534afc9c1349' 2025-10-10T01:19:00.6319698Z Submodule path 'third_party/fbgemm/external/cutlass': checked out '311f3c8e51dc0eb56310cfc6980bf63d0fbd7917' 2025-10-10T01:19:00.6856724Z Submodule path 'third_party/fbgemm/external/googletest': checked out '52eb8108c5bdec04579160ae17225d66034bd723' 2025-10-10T01:19:00.7036459Z Submodule path 'third_party/fbgemm/external/hipify_torch': checked out '63b6a7b541fa7f08f8475ca7d74054db36ff2691' 2025-10-10T01:19:00.8354607Z Submodule path 'third_party/fbgemm/external/json': checked out '9cca280a4d0ccf0c08f47a99aa71d1b0e52f8d03' 2025-10-10T01:19:00.9250347Z Submodule path 'third_party/flash-attention': checked out '979702c87a8713a8e0a5e9fee122b90d2ef13be5' 2025-10-10T01:19:00.9283634Z Submodule 'csrc/composable_kernel' (https://github.com/ROCm/composable_kernel.git) registered for path 'third_party/flash-attention/csrc/composable_kernel' 2025-10-10T01:19:00.9286272Z Submodule 'csrc/cutlass' (https://github.com/NVIDIA/cutlass.git) registered for path 'third_party/flash-attention/csrc/cutlass' 2025-10-10T01:19:00.9328332Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/flash-attention/csrc/composable_kernel'... 2025-10-10T01:19:04.1117184Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/flash-attention/csrc/cutlass'... 2025-10-10T01:19:04.4280326Z Submodule path 'third_party/flash-attention/csrc/composable_kernel': checked out '888317e698e9803c62bd38568abc9e05d7709f33' 2025-10-10T01:19:05.0834839Z Submodule path 'third_party/flash-attention/csrc/cutlass': checked out 'c506e16788cb08416a4a57e11a9067beeee29420' 2025-10-10T01:19:05.2581283Z Submodule path 'third_party/flatbuffers': checked out 'a2cd1ea3b6d3fee220106b5fed3f7ce8da9eb757' 2025-10-10T01:19:05.2951025Z Submodule path 'third_party/fmt': checked out 'e424e3f2e607da02742f73db84873b8084fc714c' 2025-10-10T01:19:05.3425069Z Submodule path 'third_party/gemmlowp/gemmlowp': checked out '3fb5c176c17c765a3492cd2f0321b0dab712f350' 2025-10-10T01:19:05.3787993Z Submodule path 'third_party/gloo': checked out '54cbae0d3a67fa890b4c3d9ee162b7860315e341' 2025-10-10T01:19:05.4322168Z Submodule path 'third_party/googletest': checked out '52eb8108c5bdec04579160ae17225d66034bd723' 2025-10-10T01:19:05.4516862Z Submodule path 'third_party/ideep': checked out '719d8e6cd7f7a0e01b155657526d693acf97c2b3' 2025-10-10T01:19:05.4543232Z Submodule 'mkl-dnn' (https://github.com/intel/mkl-dnn.git) registered for path 'third_party/ideep/mkl-dnn' 2025-10-10T01:19:05.4581448Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/ideep/mkl-dnn'... 2025-10-10T01:19:18.3158543Z Submodule path 'third_party/ideep/mkl-dnn': checked out '8d263e693366ef8db40acc569cc7d8edf644556d' 2025-10-10T01:19:18.3457162Z Submodule path 'third_party/ittapi': checked out 'dec1d23ca65ab069d225dfe40dea14f455170959' 2025-10-10T01:19:18.4372218Z Submodule path 'third_party/kineto': checked out '001ba8eb519438592f79dbc8e86a349f5f6c6829' 2025-10-10T01:19:18.4398221Z Submodule 'libkineto/third_party/dynolog' (https://github.com/facebookincubator/dynolog.git) registered for path 'third_party/kineto/libkineto/third_party/dynolog' 2025-10-10T01:19:18.4401256Z Submodule 'libkineto/third_party/fmt' (https://github.com/fmtlib/fmt.git) registered for path 'third_party/kineto/libkineto/third_party/fmt' 2025-10-10T01:19:18.4404291Z Submodule 'libkineto/third_party/googletest' (https://github.com/google/googletest.git) registered for path 'third_party/kineto/libkineto/third_party/googletest' 2025-10-10T01:19:18.4444675Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/kineto/libkineto/third_party/dynolog'... 2025-10-10T01:19:19.0901987Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/kineto/libkineto/third_party/fmt'... 2025-10-10T01:19:19.4325147Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/kineto/libkineto/third_party/googletest'... 2025-10-10T01:19:19.5360669Z Submodule path 'third_party/kineto/libkineto/third_party/dynolog': checked out 'd2ffe0a4e3acace628db49974246b66fc3e85fb1' 2025-10-10T01:19:19.5393239Z Submodule 'third_party/DCGM' (https://github.com/NVIDIA/DCGM.git) registered for path 'third_party/kineto/libkineto/third_party/dynolog/third_party/DCGM' 2025-10-10T01:19:19.5395672Z Submodule 'third_party/cpr' (https://github.com/libcpr/cpr.git) registered for path 'third_party/kineto/libkineto/third_party/dynolog/third_party/cpr' 2025-10-10T01:19:19.5399025Z Submodule 'third_party/fmt' (https://github.com/fmtlib/fmt.git) registered for path 'third_party/kineto/libkineto/third_party/dynolog/third_party/fmt' 2025-10-10T01:19:19.5402380Z Submodule 'third_party/gflags' (https://github.com/gflags/gflags.git) registered for path 'third_party/kineto/libkineto/third_party/dynolog/third_party/gflags' 2025-10-10T01:19:19.5406430Z Submodule 'third_party/glog' (https://github.com/google/glog.git) registered for path 'third_party/kineto/libkineto/third_party/dynolog/third_party/glog' 2025-10-10T01:19:19.5410597Z Submodule 'third_party/googletest' (https://github.com/google/googletest.git) registered for path 'third_party/kineto/libkineto/third_party/dynolog/third_party/googletest' 2025-10-10T01:19:19.5415688Z Submodule 'third_party/json' (https://github.com/nlohmann/json.git) registered for path 'third_party/kineto/libkineto/third_party/dynolog/third_party/json' 2025-10-10T01:19:19.5419580Z Submodule 'third_party/pfs' (https://github.com/dtrugman/pfs.git) registered for path 'third_party/kineto/libkineto/third_party/dynolog/third_party/pfs' 2025-10-10T01:19:19.5423439Z Submodule 'third_party/prometheus-cpp' (https://github.com/jupp0r/prometheus-cpp.git) registered for path 'third_party/kineto/libkineto/third_party/dynolog/third_party/prometheus-cpp' 2025-10-10T01:19:19.5463511Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/kineto/libkineto/third_party/dynolog/third_party/DCGM'... 2025-10-10T01:19:21.2446306Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/kineto/libkineto/third_party/dynolog/third_party/pfs'... 2025-10-10T01:19:21.2448403Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/kineto/libkineto/third_party/dynolog/third_party/gflags'... 2025-10-10T01:19:21.2450410Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/kineto/libkineto/third_party/dynolog/third_party/prometheus-cpp'... 2025-10-10T01:19:21.2452365Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/kineto/libkineto/third_party/dynolog/third_party/cpr'... 2025-10-10T01:19:21.2453823Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/kineto/libkineto/third_party/dynolog/third_party/glog'... 2025-10-10T01:19:21.2454999Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/kineto/libkineto/third_party/dynolog/third_party/googletest'... 2025-10-10T01:19:21.2455883Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/kineto/libkineto/third_party/dynolog/third_party/fmt'... 2025-10-10T01:19:21.3446651Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/kineto/libkineto/third_party/dynolog/third_party/json'... 2025-10-10T01:19:25.4218369Z Submodule path 'third_party/kineto/libkineto/third_party/dynolog/third_party/DCGM': checked out 'ffde4e54bc7249a6039a5e6b45b395141e1217f9' 2025-10-10T01:19:25.4486675Z Submodule path 'third_party/kineto/libkineto/third_party/dynolog/third_party/cpr': checked out '871ed52d350214a034f6ef8a3b8f51c5ce1bd400' 2025-10-10T01:19:25.4940340Z Submodule path 'third_party/kineto/libkineto/third_party/dynolog/third_party/fmt': checked out 'cd4af11efc9c622896a3e4cb599fa28668ca3d05' 2025-10-10T01:19:25.5151548Z Submodule path 'third_party/kineto/libkineto/third_party/dynolog/third_party/gflags': checked out 'e171aa2d15ed9eb17054558e0b3a6a413bb01067' 2025-10-10T01:19:25.5179816Z Submodule 'doc' (https://github.com/gflags/gflags.git) registered for path 'third_party/kineto/libkineto/third_party/dynolog/third_party/gflags/doc' 2025-10-10T01:19:25.5220405Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/kineto/libkineto/third_party/dynolog/third_party/gflags/doc'... 2025-10-10T01:19:25.8321706Z Submodule path 'third_party/kineto/libkineto/third_party/dynolog/third_party/gflags/doc': checked out '8411df715cf522606e3b1aca386ddfc0b63d34b4' 2025-10-10T01:19:25.8589193Z Submodule path 'third_party/kineto/libkineto/third_party/dynolog/third_party/glog': checked out 'b33e3bad4c46c8a6345525fd822af355e5ef9446' 2025-10-10T01:19:25.9129467Z Submodule path 'third_party/kineto/libkineto/third_party/dynolog/third_party/googletest': checked out '52eb8108c5bdec04579160ae17225d66034bd723' 2025-10-10T01:19:26.0338913Z Submodule path 'third_party/kineto/libkineto/third_party/dynolog/third_party/json': checked out '4f8fba14066156b73f1189a2b8bd568bde5284c5' 2025-10-10T01:19:26.0572854Z Submodule path 'third_party/kineto/libkineto/third_party/dynolog/third_party/pfs': checked out 'f68a2fa8ea36c783bdd760371411fcb495aa3150' 2025-10-10T01:19:26.0828981Z Submodule path 'third_party/kineto/libkineto/third_party/dynolog/third_party/prometheus-cpp': checked out 'b1234816facfdda29845c46696a02998a4af115a' 2025-10-10T01:19:26.0852270Z Submodule 'civetweb' (https://github.com/civetweb/civetweb.git) registered for path 'third_party/kineto/libkineto/third_party/dynolog/third_party/prometheus-cpp/3rdparty/civetweb' 2025-10-10T01:19:26.0855117Z Submodule 'googletest' (https://github.com/google/googletest.git) registered for path 'third_party/kineto/libkineto/third_party/dynolog/third_party/prometheus-cpp/3rdparty/googletest' 2025-10-10T01:19:26.0897950Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/kineto/libkineto/third_party/dynolog/third_party/prometheus-cpp/3rdparty/civetweb'... 2025-10-10T01:19:28.2895815Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/kineto/libkineto/third_party/dynolog/third_party/prometheus-cpp/3rdparty/googletest'... 2025-10-10T01:19:28.5552584Z Submodule path 'third_party/kineto/libkineto/third_party/dynolog/third_party/prometheus-cpp/3rdparty/civetweb': checked out 'd7ba35bbb649209c66e582d5a0244ba988a15159' 2025-10-10T01:19:28.6111259Z Submodule path 'third_party/kineto/libkineto/third_party/dynolog/third_party/prometheus-cpp/3rdparty/googletest': checked out 'e2239ee6043f73722e7aa812a459f54a28552929' 2025-10-10T01:19:28.6513170Z Submodule path 'third_party/kineto/libkineto/third_party/fmt': checked out '40626af88bd7df9a5fb80be7b25ac85b122d6c21' 2025-10-10T01:19:28.7021556Z Submodule path 'third_party/kineto/libkineto/third_party/googletest': checked out '52eb8108c5bdec04579160ae17225d66034bd723' 2025-10-10T01:19:28.7593335Z Submodule path 'third_party/kleidiai': checked out 'cca02c2f69dd18e1f12647c1c0bdc8cf90e680c7' 2025-10-10T01:19:28.8076242Z Submodule path 'third_party/mimalloc': checked out 'fbd8b99c2b828428947d70fdc046bb55609be93e' 2025-10-10T01:19:28.9379358Z Submodule path 'third_party/nlohmann': checked out '55f93686c01528224f448c19128836e7df245f72' 2025-10-10T01:19:29.5028619Z Submodule path 'third_party/onnx': checked out 'e709452ef2bbc1d113faf678c24e6d3467696e83' 2025-10-10T01:19:29.5077862Z Submodule 'third_party/pybind11' (https://github.com/pybind/pybind11.git) registered for path 'third_party/onnx/third_party/pybind11' 2025-10-10T01:19:29.5119748Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/onnx/third_party/pybind11'... 2025-10-10T01:19:30.5269441Z Submodule path 'third_party/onnx/third_party/pybind11': checked out 'a2e59f0e7065404b44dfe92a28aca47ba1378dc4' 2025-10-10T01:19:30.6223549Z Submodule path 'third_party/opentelemetry-cpp': checked out 'a799f4aed9c94b765dcdaabaeab7d5e7e2310878' 2025-10-10T01:19:30.6251906Z Submodule 'third_party/benchmark' (https://github.com/google/benchmark) registered for path 'third_party/opentelemetry-cpp/third_party/benchmark' 2025-10-10T01:19:30.6255050Z Submodule 'third_party/googletest' (https://github.com/google/googletest) registered for path 'third_party/opentelemetry-cpp/third_party/googletest' 2025-10-10T01:19:30.6258600Z Submodule 'third_party/ms-gsl' (https://github.com/microsoft/GSL) registered for path 'third_party/opentelemetry-cpp/third_party/ms-gsl' 2025-10-10T01:19:30.6262365Z Submodule 'third_party/nlohmann-json' (https://github.com/nlohmann/json) registered for path 'third_party/opentelemetry-cpp/third_party/nlohmann-json' 2025-10-10T01:19:30.6266423Z Submodule 'third_party/opentelemetry-proto' (https://github.com/open-telemetry/opentelemetry-proto) registered for path 'third_party/opentelemetry-cpp/third_party/opentelemetry-proto' 2025-10-10T01:19:30.6270341Z Submodule 'third_party/opentracing-cpp' (https://github.com/opentracing/opentracing-cpp.git) registered for path 'third_party/opentelemetry-cpp/third_party/opentracing-cpp' 2025-10-10T01:19:30.6274423Z Submodule 'third_party/prometheus-cpp' (https://github.com/jupp0r/prometheus-cpp) registered for path 'third_party/opentelemetry-cpp/third_party/prometheus-cpp' 2025-10-10T01:19:30.6279387Z Submodule 'tools/vcpkg' (https://github.com/Microsoft/vcpkg) registered for path 'third_party/opentelemetry-cpp/tools/vcpkg' 2025-10-10T01:19:30.6318564Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/opentelemetry-cpp/third_party/benchmark'... 2025-10-10T01:19:31.0520573Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/opentelemetry-cpp/third_party/opentracing-cpp'... 2025-10-10T01:19:31.0522970Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/opentelemetry-cpp/third_party/opentelemetry-proto'... 2025-10-10T01:19:31.0525946Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/opentelemetry-cpp/third_party/ms-gsl'... 2025-10-10T01:19:31.0526659Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/opentelemetry-cpp/third_party/prometheus-cpp'... 2025-10-10T01:19:31.1520570Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/opentelemetry-cpp/third_party/googletest'... 2025-10-10T01:19:31.6334247Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/opentelemetry-cpp/third_party/nlohmann-json'... 2025-10-10T01:19:38.0712071Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/opentelemetry-cpp/tools/vcpkg'... 2025-10-10T01:19:38.7223625Z Submodule path 'third_party/opentelemetry-cpp/third_party/benchmark': checked out 'd572f4777349d43653b21d6c2fc63020ab326db2' 2025-10-10T01:19:38.7716843Z Submodule path 'third_party/opentelemetry-cpp/third_party/googletest': checked out 'b796f7d44681514f58a683a3a71ff17c94edb0c1' 2025-10-10T01:19:38.7946495Z Submodule path 'third_party/opentelemetry-cpp/third_party/ms-gsl': checked out '6f4529395c5b7c2d661812257cd6780c67e54afa' 2025-10-10T01:19:38.9222079Z Submodule path 'third_party/opentelemetry-cpp/third_party/nlohmann-json': checked out 'bc889afb4c5bf1c0d8ee29ef35eaaf4c8bef8a5d' 2025-10-10T01:19:38.9432576Z Submodule path 'third_party/opentelemetry-cpp/third_party/opentelemetry-proto': checked out '4ca4f0335c63cda7ab31ea7ed70d6553aee14dce' 2025-10-10T01:19:38.9646382Z Submodule path 'third_party/opentelemetry-cpp/third_party/opentracing-cpp': checked out '06b57f48ded1fa3bdd3d4346f6ef29e40e08eaf5' 2025-10-10T01:19:38.9890837Z Submodule path 'third_party/opentelemetry-cpp/third_party/prometheus-cpp': checked out 'c9ffcdda9086ffd9e1283ea7a0276d831f3c8a8d' 2025-10-10T01:19:38.9913214Z Submodule 'civetweb' (https://github.com/civetweb/civetweb.git) registered for path 'third_party/opentelemetry-cpp/third_party/prometheus-cpp/3rdparty/civetweb' 2025-10-10T01:19:38.9915574Z Submodule 'googletest' (https://github.com/google/googletest.git) registered for path 'third_party/opentelemetry-cpp/third_party/prometheus-cpp/3rdparty/googletest' 2025-10-10T01:19:38.9962168Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/opentelemetry-cpp/third_party/prometheus-cpp/3rdparty/civetweb'... 2025-10-10T01:19:40.6593207Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/opentelemetry-cpp/third_party/prometheus-cpp/3rdparty/googletest'... 2025-10-10T01:19:40.9214668Z Submodule path 'third_party/opentelemetry-cpp/third_party/prometheus-cpp/3rdparty/civetweb': checked out 'eefb26f82b233268fc98577d265352720d477ba4' 2025-10-10T01:19:40.9766339Z Submodule path 'third_party/opentelemetry-cpp/third_party/prometheus-cpp/3rdparty/googletest': checked out 'e2239ee6043f73722e7aa812a459f54a28552929' 2025-10-10T01:19:41.6617177Z Submodule path 'third_party/opentelemetry-cpp/tools/vcpkg': checked out '8eb57355a4ffb410a2e94c07b4dca2dffbee8e50' 2025-10-10T01:19:41.6794469Z Submodule path 'third_party/pocketfft': checked out '0fa0ef591e38c2758e3184c6c23e497b9f732ffa' 2025-10-10T01:19:41.9902293Z Submodule path 'third_party/protobuf': checked out 'd1eca4e4b421cd2997495c4b4e65cea6be4e9b8a' 2025-10-10T01:19:41.9934551Z Submodule 'third_party/benchmark' (https://github.com/google/benchmark.git) registered for path 'third_party/protobuf/third_party/benchmark' 2025-10-10T01:19:41.9936646Z Submodule 'third_party/googletest' (https://github.com/google/googletest.git) registered for path 'third_party/protobuf/third_party/googletest' 2025-10-10T01:19:41.9982859Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/protobuf/third_party/benchmark'... 2025-10-10T01:19:42.5348996Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/protobuf/third_party/googletest'... 2025-10-10T01:19:42.9037794Z Submodule path 'third_party/protobuf/third_party/benchmark': checked out '5b7683f49e1e9223cf9927b24f6fd3d6bd82e3f8' 2025-10-10T01:19:42.9830387Z Submodule path 'third_party/protobuf/third_party/googletest': checked out '5ec7f0c4a113e2f18ac2c6cc7df51ad6afc24081' 2025-10-10T01:19:42.9993372Z Submodule path 'third_party/psimd': checked out '072586a71b55b7f8c584153d223e95687148a900' 2025-10-10T01:19:43.0170962Z Submodule path 'third_party/pthreadpool': checked out '4fe0e1e183925bf8cfa6aae24237e724a96479b8' 2025-10-10T01:19:43.0684147Z Submodule path 'third_party/pybind11': checked out 'f5fbe867d2d26e4a0a9177a51f6e568868ad3dc8' 2025-10-10T01:19:43.1041260Z Submodule path 'third_party/python-peachpy': checked out 'f45429b087dd7d5bc78bb40dc7cf06425c252d67' 2025-10-10T01:19:43.1547952Z Submodule path 'third_party/sleef': checked out '5a1d179df9cf652951b59010a2d2075372d67f68' 2025-10-10T01:19:43.1916632Z Submodule path 'third_party/tensorpipe': checked out 'af0118d13e52f5a08841464a768e01a0bf3e3075' 2025-10-10T01:19:43.1947636Z Submodule 'third_party/googletest' (https://github.com/google/googletest.git) registered for path 'third_party/tensorpipe/third_party/googletest' 2025-10-10T01:19:43.1949611Z Submodule 'third_party/libnop' (https://github.com/google/libnop.git) registered for path 'third_party/tensorpipe/third_party/libnop' 2025-10-10T01:19:43.1953796Z Submodule 'third_party/libuv' (https://github.com/libuv/libuv.git) registered for path 'third_party/tensorpipe/third_party/libuv' 2025-10-10T01:19:43.1957698Z Submodule 'third_party/pybind11' (https://github.com/pybind/pybind11.git) registered for path 'third_party/tensorpipe/third_party/pybind11' 2025-10-10T01:19:43.1999920Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/tensorpipe/third_party/googletest'... 2025-10-10T01:19:44.1050650Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/tensorpipe/third_party/libnop'... 2025-10-10T01:19:44.1051572Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/tensorpipe/third_party/pybind11'... 2025-10-10T01:19:44.2051233Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/tensorpipe/third_party/libuv'... 2025-10-10T01:19:44.3297102Z Submodule path 'third_party/tensorpipe/third_party/googletest': checked out 'aee0f9d9b5b87796ee8a0ab26b7587ec30e8858e' 2025-10-10T01:19:44.3536338Z Submodule path 'third_party/tensorpipe/third_party/libnop': checked out '910b55815be16109f04f4180e9adee14fb4ce281' 2025-10-10T01:19:44.4383991Z Submodule path 'third_party/tensorpipe/third_party/libuv': checked out '5152db2cbfeb5582e9c27c5ea1dba2cd9e10759b' 2025-10-10T01:19:44.4759097Z Submodule path 'third_party/tensorpipe/third_party/pybind11': checked out 'a23996fce38ff6ccfbcdc09f1e63f2c4be5ea2ef' 2025-10-10T01:19:44.4786078Z Submodule 'tools/clang' (https://github.com/wjakob/clang-cindex-python3) registered for path 'third_party/tensorpipe/third_party/pybind11/tools/clang' 2025-10-10T01:19:44.4826512Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/tensorpipe/third_party/pybind11/tools/clang'... 2025-10-10T01:19:44.6977664Z Submodule path 'third_party/tensorpipe/third_party/pybind11/tools/clang': checked out '6a00cbc4a9b8e68b71caf7f774b3f9c753ae84d5' 2025-10-10T01:19:44.7054313Z [command]/usr/bin/git submodule foreach --recursive git config --local gc.auto 0 2025-10-10T01:19:44.7528344Z Entering 'android/libs/fbjni' 2025-10-10T01:19:44.7591914Z Entering 'third_party/FP16' 2025-10-10T01:19:44.7659087Z Entering 'third_party/FXdiv' 2025-10-10T01:19:44.7728599Z Entering 'third_party/NNPACK' 2025-10-10T01:19:44.7798288Z Entering 'third_party/NVTX' 2025-10-10T01:19:44.7867960Z Entering 'third_party/VulkanMemoryAllocator' 2025-10-10T01:19:44.7930833Z Entering 'third_party/XNNPACK' 2025-10-10T01:19:44.8014518Z Entering 'third_party/aiter' 2025-10-10T01:19:44.8078384Z Entering 'third_party/aiter/3rdparty/composable_kernel' 2025-10-10T01:19:44.8160805Z Entering 'third_party/benchmark' 2025-10-10T01:19:44.8229295Z Entering 'third_party/composable_kernel' 2025-10-10T01:19:44.8306866Z Entering 'third_party/cpp-httplib' 2025-10-10T01:19:44.8375973Z Entering 'third_party/cpuinfo' 2025-10-10T01:19:44.8445275Z Entering 'third_party/cudnn_frontend' 2025-10-10T01:19:44.8511408Z Entering 'third_party/cutlass' 2025-10-10T01:19:44.8589204Z Entering 'third_party/fbgemm' 2025-10-10T01:19:44.8658828Z Entering 'third_party/fbgemm/external/asmjit' 2025-10-10T01:19:44.8719778Z Entering 'third_party/fbgemm/external/composable_kernel' 2025-10-10T01:19:44.8795768Z Entering 'third_party/fbgemm/external/cpuinfo' 2025-10-10T01:19:44.8865105Z Entering 'third_party/fbgemm/external/cutlass' 2025-10-10T01:19:44.8940019Z Entering 'third_party/fbgemm/external/googletest' 2025-10-10T01:19:44.9010804Z Entering 'third_party/fbgemm/external/hipify_torch' 2025-10-10T01:19:44.9077165Z Entering 'third_party/fbgemm/external/json' 2025-10-10T01:19:44.9158002Z Entering 'third_party/flash-attention' 2025-10-10T01:19:44.9226648Z Entering 'third_party/flash-attention/csrc/composable_kernel' 2025-10-10T01:19:44.9293448Z Entering 'third_party/flash-attention/csrc/cutlass' 2025-10-10T01:19:44.9367134Z Entering 'third_party/flatbuffers' 2025-10-10T01:19:44.9434641Z Entering 'third_party/fmt' 2025-10-10T01:19:44.9498323Z Entering 'third_party/gemmlowp/gemmlowp' 2025-10-10T01:19:44.9557933Z Entering 'third_party/gloo' 2025-10-10T01:19:44.9620950Z Entering 'third_party/googletest' 2025-10-10T01:19:44.9688966Z Entering 'third_party/ideep' 2025-10-10T01:19:44.9755667Z Entering 'third_party/ideep/mkl-dnn' 2025-10-10T01:19:44.9824884Z Entering 'third_party/ittapi' 2025-10-10T01:19:44.9890147Z Entering 'third_party/kineto' 2025-10-10T01:19:44.9954819Z Entering 'third_party/kineto/libkineto/third_party/dynolog' 2025-10-10T01:19:45.0015875Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/DCGM' 2025-10-10T01:19:45.0075535Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/cpr' 2025-10-10T01:19:45.0140094Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/fmt' 2025-10-10T01:19:45.0207502Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/gflags' 2025-10-10T01:19:45.0272318Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/gflags/doc' 2025-10-10T01:19:45.0346010Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/glog' 2025-10-10T01:19:45.0410369Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/googletest' 2025-10-10T01:19:45.0474604Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/json' 2025-10-10T01:19:45.0540797Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/pfs' 2025-10-10T01:19:45.0611053Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/prometheus-cpp' 2025-10-10T01:19:45.0670036Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/prometheus-cpp/3rdparty/civetweb' 2025-10-10T01:19:45.0733220Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/prometheus-cpp/3rdparty/googletest' 2025-10-10T01:19:45.0810197Z Entering 'third_party/kineto/libkineto/third_party/fmt' 2025-10-10T01:19:45.0877446Z Entering 'third_party/kineto/libkineto/third_party/googletest' 2025-10-10T01:19:45.0953689Z Entering 'third_party/kleidiai' 2025-10-10T01:19:45.1017722Z Entering 'third_party/mimalloc' 2025-10-10T01:19:45.1080675Z Entering 'third_party/nlohmann' 2025-10-10T01:19:45.1152483Z Entering 'third_party/onnx' 2025-10-10T01:19:45.1240089Z Entering 'third_party/onnx/third_party/pybind11' 2025-10-10T01:19:45.1315389Z Entering 'third_party/opentelemetry-cpp' 2025-10-10T01:19:45.1380142Z Entering 'third_party/opentelemetry-cpp/third_party/benchmark' 2025-10-10T01:19:45.1449800Z Entering 'third_party/opentelemetry-cpp/third_party/googletest' 2025-10-10T01:19:45.1517174Z Entering 'third_party/opentelemetry-cpp/third_party/ms-gsl' 2025-10-10T01:19:45.1590655Z Entering 'third_party/opentelemetry-cpp/third_party/nlohmann-json' 2025-10-10T01:19:45.1659102Z Entering 'third_party/opentelemetry-cpp/third_party/opentelemetry-proto' 2025-10-10T01:19:45.1725456Z Entering 'third_party/opentelemetry-cpp/third_party/opentracing-cpp' 2025-10-10T01:19:45.1788167Z Entering 'third_party/opentelemetry-cpp/third_party/prometheus-cpp' 2025-10-10T01:19:45.1855076Z Entering 'third_party/opentelemetry-cpp/third_party/prometheus-cpp/3rdparty/civetweb' 2025-10-10T01:19:45.1915576Z Entering 'third_party/opentelemetry-cpp/third_party/prometheus-cpp/3rdparty/googletest' 2025-10-10T01:19:45.1980863Z Entering 'third_party/opentelemetry-cpp/tools/vcpkg' 2025-10-10T01:19:45.2066550Z Entering 'third_party/pocketfft' 2025-10-10T01:19:45.2137762Z Entering 'third_party/protobuf' 2025-10-10T01:19:45.2203454Z Entering 'third_party/protobuf/third_party/benchmark' 2025-10-10T01:19:45.2267187Z Entering 'third_party/protobuf/third_party/googletest' 2025-10-10T01:19:45.2340922Z Entering 'third_party/psimd' 2025-10-10T01:19:45.2409745Z Entering 'third_party/pthreadpool' 2025-10-10T01:19:45.2475212Z Entering 'third_party/pybind11' 2025-10-10T01:19:45.2540139Z Entering 'third_party/python-peachpy' 2025-10-10T01:19:45.2607616Z Entering 'third_party/sleef' 2025-10-10T01:19:45.2671137Z Entering 'third_party/tensorpipe' 2025-10-10T01:19:45.2737969Z Entering 'third_party/tensorpipe/third_party/googletest' 2025-10-10T01:19:45.2798966Z Entering 'third_party/tensorpipe/third_party/libnop' 2025-10-10T01:19:45.2867110Z Entering 'third_party/tensorpipe/third_party/libuv' 2025-10-10T01:19:45.2933801Z Entering 'third_party/tensorpipe/third_party/pybind11' 2025-10-10T01:19:45.2989673Z Entering 'third_party/tensorpipe/third_party/pybind11/tools/clang' 2025-10-10T01:19:45.3083762Z ##[endgroup] 2025-10-10T01:19:45.3084624Z ##[group]Persisting credentials for submodules 2025-10-10T01:19:45.3095573Z [command]/usr/bin/git submodule foreach --recursive sh -c "git config --local --name-only --get-regexp 'url\.https\:\/\/github\.com\/\.insteadOf' && git config --local --unset-all 'url.https://github.com/.insteadOf' || :" 2025-10-10T01:19:45.3569662Z Entering 'android/libs/fbjni' 2025-10-10T01:19:45.3660364Z Entering 'third_party/FP16' 2025-10-10T01:19:45.3748286Z Entering 'third_party/FXdiv' 2025-10-10T01:19:45.3838810Z Entering 'third_party/NNPACK' 2025-10-10T01:19:45.3930449Z Entering 'third_party/NVTX' 2025-10-10T01:19:45.4020505Z Entering 'third_party/VulkanMemoryAllocator' 2025-10-10T01:19:45.4111173Z Entering 'third_party/XNNPACK' 2025-10-10T01:19:45.4221036Z Entering 'third_party/aiter' 2025-10-10T01:19:45.4316808Z Entering 'third_party/aiter/3rdparty/composable_kernel' 2025-10-10T01:19:45.4418522Z Entering 'third_party/benchmark' 2025-10-10T01:19:45.4508793Z Entering 'third_party/composable_kernel' 2025-10-10T01:19:45.4606372Z Entering 'third_party/cpp-httplib' 2025-10-10T01:19:45.4699873Z Entering 'third_party/cpuinfo' 2025-10-10T01:19:45.4788055Z Entering 'third_party/cudnn_frontend' 2025-10-10T01:19:45.4879130Z Entering 'third_party/cutlass' 2025-10-10T01:19:45.4975780Z Entering 'third_party/fbgemm' 2025-10-10T01:19:45.5059196Z Entering 'third_party/fbgemm/external/asmjit' 2025-10-10T01:19:45.5144495Z Entering 'third_party/fbgemm/external/composable_kernel' 2025-10-10T01:19:45.5236110Z Entering 'third_party/fbgemm/external/cpuinfo' 2025-10-10T01:19:45.5315971Z Entering 'third_party/fbgemm/external/cutlass' 2025-10-10T01:19:45.5415907Z Entering 'third_party/fbgemm/external/googletest' 2025-10-10T01:19:45.5494335Z Entering 'third_party/fbgemm/external/hipify_torch' 2025-10-10T01:19:45.5574595Z Entering 'third_party/fbgemm/external/json' 2025-10-10T01:19:45.5665803Z Entering 'third_party/flash-attention' 2025-10-10T01:19:45.5757027Z Entering 'third_party/flash-attention/csrc/composable_kernel' 2025-10-10T01:19:45.5853916Z Entering 'third_party/flash-attention/csrc/cutlass' 2025-10-10T01:19:45.5951639Z Entering 'third_party/flatbuffers' 2025-10-10T01:19:45.6050972Z Entering 'third_party/fmt' 2025-10-10T01:19:45.6143065Z Entering 'third_party/gemmlowp/gemmlowp' 2025-10-10T01:19:45.6232308Z Entering 'third_party/gloo' 2025-10-10T01:19:45.6318190Z Entering 'third_party/googletest' 2025-10-10T01:19:45.6410670Z Entering 'third_party/ideep' 2025-10-10T01:19:45.6501778Z Entering 'third_party/ideep/mkl-dnn' 2025-10-10T01:19:45.6593839Z Entering 'third_party/ittapi' 2025-10-10T01:19:45.6686739Z Entering 'third_party/kineto' 2025-10-10T01:19:45.6774371Z Entering 'third_party/kineto/libkineto/third_party/dynolog' 2025-10-10T01:19:45.6856532Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/DCGM' 2025-10-10T01:19:45.6949766Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/cpr' 2025-10-10T01:19:45.7035018Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/fmt' 2025-10-10T01:19:45.7119008Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/gflags' 2025-10-10T01:19:45.7208231Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/gflags/doc' 2025-10-10T01:19:45.7305923Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/glog' 2025-10-10T01:19:45.7390601Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/googletest' 2025-10-10T01:19:45.7477745Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/json' 2025-10-10T01:19:45.7568185Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/pfs' 2025-10-10T01:19:45.7655794Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/prometheus-cpp' 2025-10-10T01:19:45.7733170Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/prometheus-cpp/3rdparty/civetweb' 2025-10-10T01:19:45.7821818Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/prometheus-cpp/3rdparty/googletest' 2025-10-10T01:19:45.7922916Z Entering 'third_party/kineto/libkineto/third_party/fmt' 2025-10-10T01:19:45.8010307Z Entering 'third_party/kineto/libkineto/third_party/googletest' 2025-10-10T01:19:45.8105094Z Entering 'third_party/kleidiai' 2025-10-10T01:19:45.8189351Z Entering 'third_party/mimalloc' 2025-10-10T01:19:45.8274796Z Entering 'third_party/nlohmann' 2025-10-10T01:19:45.8369546Z Entering 'third_party/onnx' 2025-10-10T01:19:45.8474411Z Entering 'third_party/onnx/third_party/pybind11' 2025-10-10T01:19:45.8567503Z Entering 'third_party/opentelemetry-cpp' 2025-10-10T01:19:45.8658119Z Entering 'third_party/opentelemetry-cpp/third_party/benchmark' 2025-10-10T01:19:45.8745634Z Entering 'third_party/opentelemetry-cpp/third_party/googletest' 2025-10-10T01:19:45.8825168Z Entering 'third_party/opentelemetry-cpp/third_party/ms-gsl' 2025-10-10T01:19:45.8905379Z Entering 'third_party/opentelemetry-cpp/third_party/nlohmann-json' 2025-10-10T01:19:45.8990923Z Entering 'third_party/opentelemetry-cpp/third_party/opentelemetry-proto' 2025-10-10T01:19:45.9080705Z Entering 'third_party/opentelemetry-cpp/third_party/opentracing-cpp' 2025-10-10T01:19:45.9166269Z Entering 'third_party/opentelemetry-cpp/third_party/prometheus-cpp' 2025-10-10T01:19:45.9250812Z Entering 'third_party/opentelemetry-cpp/third_party/prometheus-cpp/3rdparty/civetweb' 2025-10-10T01:19:45.9342571Z Entering 'third_party/opentelemetry-cpp/third_party/prometheus-cpp/3rdparty/googletest' 2025-10-10T01:19:45.9442691Z Entering 'third_party/opentelemetry-cpp/tools/vcpkg' 2025-10-10T01:19:45.9558974Z Entering 'third_party/pocketfft' 2025-10-10T01:19:45.9652727Z Entering 'third_party/protobuf' 2025-10-10T01:19:45.9740131Z Entering 'third_party/protobuf/third_party/benchmark' 2025-10-10T01:19:45.9828208Z Entering 'third_party/protobuf/third_party/googletest' 2025-10-10T01:19:45.9923367Z Entering 'third_party/psimd' 2025-10-10T01:19:46.0019479Z Entering 'third_party/pthreadpool' 2025-10-10T01:19:46.0109160Z Entering 'third_party/pybind11' 2025-10-10T01:19:46.0194491Z Entering 'third_party/python-peachpy' 2025-10-10T01:19:46.0278731Z Entering 'third_party/sleef' 2025-10-10T01:19:46.0368523Z Entering 'third_party/tensorpipe' 2025-10-10T01:19:46.0454959Z Entering 'third_party/tensorpipe/third_party/googletest' 2025-10-10T01:19:46.0537133Z Entering 'third_party/tensorpipe/third_party/libnop' 2025-10-10T01:19:46.0615320Z Entering 'third_party/tensorpipe/third_party/libuv' 2025-10-10T01:19:46.0694984Z Entering 'third_party/tensorpipe/third_party/pybind11' 2025-10-10T01:19:46.0770762Z Entering 'third_party/tensorpipe/third_party/pybind11/tools/clang' 2025-10-10T01:19:46.0898201Z [command]/usr/bin/git submodule foreach --recursive sh -c "git config --local 'http.https://github.com/.extraheader' 'AUTHORIZATION: basic ***' && git config --local --show-origin --name-only --get-regexp remote.origin.url" 2025-10-10T01:19:46.1359526Z Entering 'android/libs/fbjni' 2025-10-10T01:19:46.1445684Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/android/libs/fbjni/config remote.origin.url 2025-10-10T01:19:46.1473033Z Entering 'third_party/FP16' 2025-10-10T01:19:46.1551452Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/NNPACK_deps/FP16/config remote.origin.url 2025-10-10T01:19:46.1578684Z Entering 'third_party/FXdiv' 2025-10-10T01:19:46.1655408Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/NNPACK_deps/FXdiv/config remote.origin.url 2025-10-10T01:19:46.1682041Z Entering 'third_party/NNPACK' 2025-10-10T01:19:46.1756584Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/NNPACK/config remote.origin.url 2025-10-10T01:19:46.1782752Z Entering 'third_party/NVTX' 2025-10-10T01:19:46.1863859Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/NVTX/config remote.origin.url 2025-10-10T01:19:46.1892799Z Entering 'third_party/VulkanMemoryAllocator' 2025-10-10T01:19:46.1967552Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/VulkanMemoryAllocator/config remote.origin.url 2025-10-10T01:19:46.1993933Z Entering 'third_party/XNNPACK' 2025-10-10T01:19:46.2075486Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/XNNPACK/config remote.origin.url 2025-10-10T01:19:46.2112584Z Entering 'third_party/aiter' 2025-10-10T01:19:46.2186628Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/aiter/config remote.origin.url 2025-10-10T01:19:46.2216814Z Entering 'third_party/aiter/3rdparty/composable_kernel' 2025-10-10T01:19:46.2292833Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/aiter/modules/3rdparty/composable_kernel/config remote.origin.url 2025-10-10T01:19:46.2328704Z Entering 'third_party/benchmark' 2025-10-10T01:19:46.2406738Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/benchmark/config remote.origin.url 2025-10-10T01:19:46.2436550Z Entering 'third_party/composable_kernel' 2025-10-10T01:19:46.2525565Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/composable_kernel/config remote.origin.url 2025-10-10T01:19:46.2563294Z Entering 'third_party/cpp-httplib' 2025-10-10T01:19:46.2642603Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/cpp-httplib/config remote.origin.url 2025-10-10T01:19:46.2671681Z Entering 'third_party/cpuinfo' 2025-10-10T01:19:46.2755635Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/cpuinfo/config remote.origin.url 2025-10-10T01:19:46.2788285Z Entering 'third_party/cudnn_frontend' 2025-10-10T01:19:46.2867126Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/cudnn_frontend/config remote.origin.url 2025-10-10T01:19:46.2896743Z Entering 'third_party/cutlass' 2025-10-10T01:19:46.2973727Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/cutlass/config remote.origin.url 2025-10-10T01:19:46.3006529Z Entering 'third_party/fbgemm' 2025-10-10T01:19:46.3087190Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/fbgemm/config remote.origin.url 2025-10-10T01:19:46.3118220Z Entering 'third_party/fbgemm/external/asmjit' 2025-10-10T01:19:46.3195747Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/fbgemm/modules/external/asmjit/config remote.origin.url 2025-10-10T01:19:46.3227106Z Entering 'third_party/fbgemm/external/composable_kernel' 2025-10-10T01:19:46.3307770Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/fbgemm/modules/external/composable_kernel/config remote.origin.url 2025-10-10T01:19:46.3345349Z Entering 'third_party/fbgemm/external/cpuinfo' 2025-10-10T01:19:46.3426533Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/fbgemm/modules/external/cpuinfo/config remote.origin.url 2025-10-10T01:19:46.3452368Z Entering 'third_party/fbgemm/external/cutlass' 2025-10-10T01:19:46.3535963Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/fbgemm/modules/external/cutlass/config remote.origin.url 2025-10-10T01:19:46.3574010Z Entering 'third_party/fbgemm/external/googletest' 2025-10-10T01:19:46.3647523Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/fbgemm/modules/external/googletest/config remote.origin.url 2025-10-10T01:19:46.3677753Z Entering 'third_party/fbgemm/external/hipify_torch' 2025-10-10T01:19:46.3758607Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/fbgemm/modules/external/hipify_torch/config remote.origin.url 2025-10-10T01:19:46.3782979Z Entering 'third_party/fbgemm/external/json' 2025-10-10T01:19:46.3865107Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/fbgemm/modules/external/json/config remote.origin.url 2025-10-10T01:19:46.3895758Z Entering 'third_party/flash-attention' 2025-10-10T01:19:46.3976119Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/flash-attention/config remote.origin.url 2025-10-10T01:19:46.4006223Z Entering 'third_party/flash-attention/csrc/composable_kernel' 2025-10-10T01:19:46.4087471Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/flash-attention/modules/csrc/composable_kernel/config remote.origin.url 2025-10-10T01:19:46.4122687Z Entering 'third_party/flash-attention/csrc/cutlass' 2025-10-10T01:19:46.4195135Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/flash-attention/modules/csrc/cutlass/config remote.origin.url 2025-10-10T01:19:46.4233396Z Entering 'third_party/flatbuffers' 2025-10-10T01:19:46.4310216Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/flatbuffers/config remote.origin.url 2025-10-10T01:19:46.4342667Z Entering 'third_party/fmt' 2025-10-10T01:19:46.4425797Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/fmt/config remote.origin.url 2025-10-10T01:19:46.4457962Z Entering 'third_party/gemmlowp/gemmlowp' 2025-10-10T01:19:46.4538213Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/gemmlowp/gemmlowp/config remote.origin.url 2025-10-10T01:19:46.4569755Z Entering 'third_party/gloo' 2025-10-10T01:19:46.4646829Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/gloo/config remote.origin.url 2025-10-10T01:19:46.4678348Z Entering 'third_party/googletest' 2025-10-10T01:19:46.4757443Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/googletest/config remote.origin.url 2025-10-10T01:19:46.4784234Z Entering 'third_party/ideep' 2025-10-10T01:19:46.4864478Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/ideep/config remote.origin.url 2025-10-10T01:19:46.4893932Z Entering 'third_party/ideep/mkl-dnn' 2025-10-10T01:19:46.4972916Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/ideep/modules/mkl-dnn/config remote.origin.url 2025-10-10T01:19:46.5008262Z Entering 'third_party/ittapi' 2025-10-10T01:19:46.5086942Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/ittapi/config remote.origin.url 2025-10-10T01:19:46.5115010Z Entering 'third_party/kineto' 2025-10-10T01:19:46.5186877Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/kineto/config remote.origin.url 2025-10-10T01:19:46.5216295Z Entering 'third_party/kineto/libkineto/third_party/dynolog' 2025-10-10T01:19:46.5288130Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/kineto/modules/libkineto/third_party/dynolog/config remote.origin.url 2025-10-10T01:19:46.5314411Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/DCGM' 2025-10-10T01:19:46.5397356Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/kineto/modules/libkineto/third_party/dynolog/modules/third_party/DCGM/config remote.origin.url 2025-10-10T01:19:46.5429295Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/cpr' 2025-10-10T01:19:46.5507021Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/kineto/modules/libkineto/third_party/dynolog/modules/third_party/cpr/config remote.origin.url 2025-10-10T01:19:46.5538076Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/fmt' 2025-10-10T01:19:46.5617428Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/kineto/modules/libkineto/third_party/dynolog/modules/third_party/fmt/config remote.origin.url 2025-10-10T01:19:46.5646340Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/gflags' 2025-10-10T01:19:46.5723381Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/kineto/modules/libkineto/third_party/dynolog/modules/third_party/gflags/config remote.origin.url 2025-10-10T01:19:46.5744137Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/gflags/doc' 2025-10-10T01:19:46.5828637Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/kineto/modules/libkineto/third_party/dynolog/modules/third_party/gflags/modules/doc/config remote.origin.url 2025-10-10T01:19:46.5860080Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/glog' 2025-10-10T01:19:46.5939596Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/kineto/modules/libkineto/third_party/dynolog/modules/third_party/glog/config remote.origin.url 2025-10-10T01:19:46.5971299Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/googletest' 2025-10-10T01:19:46.6048984Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/kineto/modules/libkineto/third_party/dynolog/modules/third_party/googletest/config remote.origin.url 2025-10-10T01:19:46.6078097Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/json' 2025-10-10T01:19:46.6159089Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/kineto/modules/libkineto/third_party/dynolog/modules/third_party/json/config remote.origin.url 2025-10-10T01:19:46.6186962Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/pfs' 2025-10-10T01:19:46.6266045Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/kineto/modules/libkineto/third_party/dynolog/modules/third_party/pfs/config remote.origin.url 2025-10-10T01:19:46.6288927Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/prometheus-cpp' 2025-10-10T01:19:46.6365009Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/kineto/modules/libkineto/third_party/dynolog/modules/third_party/prometheus-cpp/config remote.origin.url 2025-10-10T01:19:46.6392994Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/prometheus-cpp/3rdparty/civetweb' 2025-10-10T01:19:46.6474349Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/kineto/modules/libkineto/third_party/dynolog/modules/third_party/prometheus-cpp/modules/civetweb/config remote.origin.url 2025-10-10T01:19:46.6502944Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/prometheus-cpp/3rdparty/googletest' 2025-10-10T01:19:46.6587987Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/kineto/modules/libkineto/third_party/dynolog/modules/third_party/prometheus-cpp/modules/googletest/config remote.origin.url 2025-10-10T01:19:46.6628970Z Entering 'third_party/kineto/libkineto/third_party/fmt' 2025-10-10T01:19:46.6706696Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/kineto/modules/libkineto/third_party/fmt/config remote.origin.url 2025-10-10T01:19:46.6733227Z Entering 'third_party/kineto/libkineto/third_party/googletest' 2025-10-10T01:19:46.6807273Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/kineto/modules/libkineto/third_party/googletest/config remote.origin.url 2025-10-10T01:19:46.6837154Z Entering 'third_party/kleidiai' 2025-10-10T01:19:46.6917325Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/kleidiai/config remote.origin.url 2025-10-10T01:19:46.6950087Z Entering 'third_party/mimalloc' 2025-10-10T01:19:46.7027561Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/mimalloc/config remote.origin.url 2025-10-10T01:19:46.7058645Z Entering 'third_party/nlohmann' 2025-10-10T01:19:46.7136704Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/nlohmann/config remote.origin.url 2025-10-10T01:19:46.7169225Z Entering 'third_party/onnx' 2025-10-10T01:19:46.7246635Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/onnx/config remote.origin.url 2025-10-10T01:19:46.7292946Z Entering 'third_party/onnx/third_party/pybind11' 2025-10-10T01:19:46.7365467Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/onnx/modules/third_party/pybind11/config remote.origin.url 2025-10-10T01:19:46.7405246Z Entering 'third_party/opentelemetry-cpp' 2025-10-10T01:19:46.7488309Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/opentelemetry-cpp/config remote.origin.url 2025-10-10T01:19:46.7515956Z Entering 'third_party/opentelemetry-cpp/third_party/benchmark' 2025-10-10T01:19:46.7596480Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/opentelemetry-cpp/modules/third_party/benchmark/config remote.origin.url 2025-10-10T01:19:46.7627108Z Entering 'third_party/opentelemetry-cpp/third_party/googletest' 2025-10-10T01:19:46.7705715Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/opentelemetry-cpp/modules/third_party/googletest/config remote.origin.url 2025-10-10T01:19:46.7731084Z Entering 'third_party/opentelemetry-cpp/third_party/ms-gsl' 2025-10-10T01:19:46.7808270Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/opentelemetry-cpp/modules/third_party/ms-gsl/config remote.origin.url 2025-10-10T01:19:46.7837189Z Entering 'third_party/opentelemetry-cpp/third_party/nlohmann-json' 2025-10-10T01:19:46.7909327Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/opentelemetry-cpp/modules/third_party/nlohmann-json/config remote.origin.url 2025-10-10T01:19:46.7934902Z Entering 'third_party/opentelemetry-cpp/third_party/opentelemetry-proto' 2025-10-10T01:19:46.8015660Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/opentelemetry-cpp/modules/third_party/opentelemetry-proto/config remote.origin.url 2025-10-10T01:19:46.8039932Z Entering 'third_party/opentelemetry-cpp/third_party/opentracing-cpp' 2025-10-10T01:19:46.8118085Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/opentelemetry-cpp/modules/third_party/opentracing-cpp/config remote.origin.url 2025-10-10T01:19:46.8147243Z Entering 'third_party/opentelemetry-cpp/third_party/prometheus-cpp' 2025-10-10T01:19:46.8226928Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/opentelemetry-cpp/modules/third_party/prometheus-cpp/config remote.origin.url 2025-10-10T01:19:46.8253705Z Entering 'third_party/opentelemetry-cpp/third_party/prometheus-cpp/3rdparty/civetweb' 2025-10-10T01:19:46.8336732Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/opentelemetry-cpp/modules/third_party/prometheus-cpp/modules/civetweb/config remote.origin.url 2025-10-10T01:19:46.8368003Z Entering 'third_party/opentelemetry-cpp/third_party/prometheus-cpp/3rdparty/googletest' 2025-10-10T01:19:46.8443445Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/opentelemetry-cpp/modules/third_party/prometheus-cpp/modules/googletest/config remote.origin.url 2025-10-10T01:19:46.8475506Z Entering 'third_party/opentelemetry-cpp/tools/vcpkg' 2025-10-10T01:19:46.8548741Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/opentelemetry-cpp/modules/tools/vcpkg/config remote.origin.url 2025-10-10T01:19:46.8600650Z Entering 'third_party/pocketfft' 2025-10-10T01:19:46.8681930Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/pocketfft/config remote.origin.url 2025-10-10T01:19:46.8708348Z Entering 'third_party/protobuf' 2025-10-10T01:19:46.8787582Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/protobuf/config remote.origin.url 2025-10-10T01:19:46.8817639Z Entering 'third_party/protobuf/third_party/benchmark' 2025-10-10T01:19:46.8898425Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/protobuf/modules/third_party/benchmark/config remote.origin.url 2025-10-10T01:19:46.8928209Z Entering 'third_party/protobuf/third_party/googletest' 2025-10-10T01:19:46.9006747Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/protobuf/modules/third_party/googletest/config remote.origin.url 2025-10-10T01:19:46.9035381Z Entering 'third_party/psimd' 2025-10-10T01:19:46.9106412Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/NNPACK_deps/psimd/config remote.origin.url 2025-10-10T01:19:46.9138862Z Entering 'third_party/pthreadpool' 2025-10-10T01:19:46.9216121Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/NNPACK_deps/pthreadpool/config remote.origin.url 2025-10-10T01:19:46.9246785Z Entering 'third_party/pybind11' 2025-10-10T01:19:46.9326179Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/pybind11/config remote.origin.url 2025-10-10T01:19:46.9353698Z Entering 'third_party/python-peachpy' 2025-10-10T01:19:46.9428129Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/python-peachpy/config remote.origin.url 2025-10-10T01:19:46.9457908Z Entering 'third_party/sleef' 2025-10-10T01:19:46.9538169Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/sleef/config remote.origin.url 2025-10-10T01:19:46.9568983Z Entering 'third_party/tensorpipe' 2025-10-10T01:19:46.9647818Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/tensorpipe/config remote.origin.url 2025-10-10T01:19:46.9676333Z Entering 'third_party/tensorpipe/third_party/googletest' 2025-10-10T01:19:46.9755560Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/tensorpipe/modules/third_party/googletest/config remote.origin.url 2025-10-10T01:19:46.9786248Z Entering 'third_party/tensorpipe/third_party/libnop' 2025-10-10T01:19:46.9863273Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/tensorpipe/modules/third_party/libnop/config remote.origin.url 2025-10-10T01:19:46.9890748Z Entering 'third_party/tensorpipe/third_party/libuv' 2025-10-10T01:19:46.9967049Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/tensorpipe/modules/third_party/libuv/config remote.origin.url 2025-10-10T01:19:46.9996461Z Entering 'third_party/tensorpipe/third_party/pybind11' 2025-10-10T01:19:47.0069042Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/tensorpipe/modules/third_party/pybind11/config remote.origin.url 2025-10-10T01:19:47.0097021Z Entering 'third_party/tensorpipe/third_party/pybind11/tools/clang' 2025-10-10T01:19:47.0177392Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/tensorpipe/modules/third_party/pybind11/modules/tools/clang/config remote.origin.url 2025-10-10T01:19:47.0949457Z [command]/usr/bin/git submodule foreach --recursive git config --local --add 'url.https://github.com/.insteadOf' 'git@github.com:' 2025-10-10T01:19:47.1419132Z Entering 'android/libs/fbjni' 2025-10-10T01:19:47.1492668Z Entering 'third_party/FP16' 2025-10-10T01:19:47.1556882Z Entering 'third_party/FXdiv' 2025-10-10T01:19:47.1630276Z Entering 'third_party/NNPACK' 2025-10-10T01:19:47.1700455Z Entering 'third_party/NVTX' 2025-10-10T01:19:47.1770468Z Entering 'third_party/VulkanMemoryAllocator' 2025-10-10T01:19:47.1836773Z Entering 'third_party/XNNPACK' 2025-10-10T01:19:47.1922141Z Entering 'third_party/aiter' 2025-10-10T01:19:47.1986812Z Entering 'third_party/aiter/3rdparty/composable_kernel' 2025-10-10T01:19:47.2059216Z Entering 'third_party/benchmark' 2025-10-10T01:19:47.2130000Z Entering 'third_party/composable_kernel' 2025-10-10T01:19:47.2209269Z Entering 'third_party/cpp-httplib' 2025-10-10T01:19:47.2268148Z Entering 'third_party/cpuinfo' 2025-10-10T01:19:47.2329081Z Entering 'third_party/cudnn_frontend' 2025-10-10T01:19:47.2398814Z Entering 'third_party/cutlass' 2025-10-10T01:19:47.2478879Z Entering 'third_party/fbgemm' 2025-10-10T01:19:47.2551843Z Entering 'third_party/fbgemm/external/asmjit' 2025-10-10T01:19:47.2622044Z Entering 'third_party/fbgemm/external/composable_kernel' 2025-10-10T01:19:47.2703032Z Entering 'third_party/fbgemm/external/cpuinfo' 2025-10-10T01:19:47.2771121Z Entering 'third_party/fbgemm/external/cutlass' 2025-10-10T01:19:47.2841187Z Entering 'third_party/fbgemm/external/googletest' 2025-10-10T01:19:47.2911329Z Entering 'third_party/fbgemm/external/hipify_torch' 2025-10-10T01:19:47.2983051Z Entering 'third_party/fbgemm/external/json' 2025-10-10T01:19:47.3054229Z Entering 'third_party/flash-attention' 2025-10-10T01:19:47.3114372Z Entering 'third_party/flash-attention/csrc/composable_kernel' 2025-10-10T01:19:47.3185439Z Entering 'third_party/flash-attention/csrc/cutlass' 2025-10-10T01:19:47.3256313Z Entering 'third_party/flatbuffers' 2025-10-10T01:19:47.3322506Z Entering 'third_party/fmt' 2025-10-10T01:19:47.3386075Z Entering 'third_party/gemmlowp/gemmlowp' 2025-10-10T01:19:47.3459587Z Entering 'third_party/gloo' 2025-10-10T01:19:47.3528898Z Entering 'third_party/googletest' 2025-10-10T01:19:47.3597549Z Entering 'third_party/ideep' 2025-10-10T01:19:47.3667575Z Entering 'third_party/ideep/mkl-dnn' 2025-10-10T01:19:47.3748783Z Entering 'third_party/ittapi' 2025-10-10T01:19:47.3819870Z Entering 'third_party/kineto' 2025-10-10T01:19:47.3888775Z Entering 'third_party/kineto/libkineto/third_party/dynolog' 2025-10-10T01:19:47.3944878Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/DCGM' 2025-10-10T01:19:47.4012848Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/cpr' 2025-10-10T01:19:47.4079726Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/fmt' 2025-10-10T01:19:47.4144145Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/gflags' 2025-10-10T01:19:47.4200873Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/gflags/doc' 2025-10-10T01:19:47.4278601Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/glog' 2025-10-10T01:19:47.4338889Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/googletest' 2025-10-10T01:19:47.4409126Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/json' 2025-10-10T01:19:47.4476960Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/pfs' 2025-10-10T01:19:47.4548237Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/prometheus-cpp' 2025-10-10T01:19:47.4607558Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/prometheus-cpp/3rdparty/civetweb' 2025-10-10T01:19:47.4671084Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/prometheus-cpp/3rdparty/googletest' 2025-10-10T01:19:47.4753918Z Entering 'third_party/kineto/libkineto/third_party/fmt' 2025-10-10T01:19:47.4815664Z Entering 'third_party/kineto/libkineto/third_party/googletest' 2025-10-10T01:19:47.4884639Z Entering 'third_party/kleidiai' 2025-10-10T01:19:47.4943204Z Entering 'third_party/mimalloc' 2025-10-10T01:19:47.5008183Z Entering 'third_party/nlohmann' 2025-10-10T01:19:47.5072474Z Entering 'third_party/onnx' 2025-10-10T01:19:47.5152631Z Entering 'third_party/onnx/third_party/pybind11' 2025-10-10T01:19:47.5224965Z Entering 'third_party/opentelemetry-cpp' 2025-10-10T01:19:47.5295557Z Entering 'third_party/opentelemetry-cpp/third_party/benchmark' 2025-10-10T01:19:47.5358672Z Entering 'third_party/opentelemetry-cpp/third_party/googletest' 2025-10-10T01:19:47.5421681Z Entering 'third_party/opentelemetry-cpp/third_party/ms-gsl' 2025-10-10T01:19:47.5479051Z Entering 'third_party/opentelemetry-cpp/third_party/nlohmann-json' 2025-10-10T01:19:47.5538962Z Entering 'third_party/opentelemetry-cpp/third_party/opentelemetry-proto' 2025-10-10T01:19:47.5607976Z Entering 'third_party/opentelemetry-cpp/third_party/opentracing-cpp' 2025-10-10T01:19:47.5669328Z Entering 'third_party/opentelemetry-cpp/third_party/prometheus-cpp' 2025-10-10T01:19:47.5728948Z Entering 'third_party/opentelemetry-cpp/third_party/prometheus-cpp/3rdparty/civetweb' 2025-10-10T01:19:47.5787407Z Entering 'third_party/opentelemetry-cpp/third_party/prometheus-cpp/3rdparty/googletest' 2025-10-10T01:19:47.5849302Z Entering 'third_party/opentelemetry-cpp/tools/vcpkg' 2025-10-10T01:19:47.5939461Z Entering 'third_party/pocketfft' 2025-10-10T01:19:47.6009981Z Entering 'third_party/protobuf' 2025-10-10T01:19:47.6081856Z Entering 'third_party/protobuf/third_party/benchmark' 2025-10-10T01:19:47.6148808Z Entering 'third_party/protobuf/third_party/googletest' 2025-10-10T01:19:47.6223037Z Entering 'third_party/psimd' 2025-10-10T01:19:47.6287309Z Entering 'third_party/pthreadpool' 2025-10-10T01:19:47.6359248Z Entering 'third_party/pybind11' 2025-10-10T01:19:47.6429099Z Entering 'third_party/python-peachpy' 2025-10-10T01:19:47.6501213Z Entering 'third_party/sleef' 2025-10-10T01:19:47.6578068Z Entering 'third_party/tensorpipe' 2025-10-10T01:19:47.6645074Z Entering 'third_party/tensorpipe/third_party/googletest' 2025-10-10T01:19:47.6709015Z Entering 'third_party/tensorpipe/third_party/libnop' 2025-10-10T01:19:47.6770496Z Entering 'third_party/tensorpipe/third_party/libuv' 2025-10-10T01:19:47.6838119Z Entering 'third_party/tensorpipe/third_party/pybind11' 2025-10-10T01:19:47.6901511Z Entering 'third_party/tensorpipe/third_party/pybind11/tools/clang' 2025-10-10T01:19:47.7007208Z [command]/usr/bin/git submodule foreach --recursive git config --local --add 'url.https://github.com/.insteadOf' 'org-21003710@github.com:' 2025-10-10T01:19:47.7487643Z Entering 'android/libs/fbjni' 2025-10-10T01:19:47.7556162Z Entering 'third_party/FP16' 2025-10-10T01:19:47.7614390Z Entering 'third_party/FXdiv' 2025-10-10T01:19:47.7677301Z Entering 'third_party/NNPACK' 2025-10-10T01:19:47.7747234Z Entering 'third_party/NVTX' 2025-10-10T01:19:47.7817331Z Entering 'third_party/VulkanMemoryAllocator' 2025-10-10T01:19:47.7887390Z Entering 'third_party/XNNPACK' 2025-10-10T01:19:47.7971084Z Entering 'third_party/aiter' 2025-10-10T01:19:47.8038343Z Entering 'third_party/aiter/3rdparty/composable_kernel' 2025-10-10T01:19:47.8125291Z Entering 'third_party/benchmark' 2025-10-10T01:19:47.8191794Z Entering 'third_party/composable_kernel' 2025-10-10T01:19:47.8262437Z Entering 'third_party/cpp-httplib' 2025-10-10T01:19:47.8326543Z Entering 'third_party/cpuinfo' 2025-10-10T01:19:47.8392655Z Entering 'third_party/cudnn_frontend' 2025-10-10T01:19:47.8457993Z Entering 'third_party/cutlass' 2025-10-10T01:19:47.8537718Z Entering 'third_party/fbgemm' 2025-10-10T01:19:47.8608596Z Entering 'third_party/fbgemm/external/asmjit' 2025-10-10T01:19:47.8671009Z Entering 'third_party/fbgemm/external/composable_kernel' 2025-10-10T01:19:47.8742757Z Entering 'third_party/fbgemm/external/cpuinfo' 2025-10-10T01:19:47.8809560Z Entering 'third_party/fbgemm/external/cutlass' 2025-10-10T01:19:47.8887119Z Entering 'third_party/fbgemm/external/googletest' 2025-10-10T01:19:47.8946433Z Entering 'third_party/fbgemm/external/hipify_torch' 2025-10-10T01:19:47.9013652Z Entering 'third_party/fbgemm/external/json' 2025-10-10T01:19:47.9083754Z Entering 'third_party/flash-attention' 2025-10-10T01:19:47.9144798Z Entering 'third_party/flash-attention/csrc/composable_kernel' 2025-10-10T01:19:47.9214260Z Entering 'third_party/flash-attention/csrc/cutlass' 2025-10-10T01:19:47.9290775Z Entering 'third_party/flatbuffers' 2025-10-10T01:19:47.9362823Z Entering 'third_party/fmt' 2025-10-10T01:19:47.9430331Z Entering 'third_party/gemmlowp/gemmlowp' 2025-10-10T01:19:47.9501022Z Entering 'third_party/gloo' 2025-10-10T01:19:47.9571072Z Entering 'third_party/googletest' 2025-10-10T01:19:47.9641437Z Entering 'third_party/ideep' 2025-10-10T01:19:47.9706757Z Entering 'third_party/ideep/mkl-dnn' 2025-10-10T01:19:47.9782319Z Entering 'third_party/ittapi' 2025-10-10T01:19:47.9849199Z Entering 'third_party/kineto' 2025-10-10T01:19:47.9916521Z Entering 'third_party/kineto/libkineto/third_party/dynolog' 2025-10-10T01:19:47.9978584Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/DCGM' 2025-10-10T01:19:48.0050835Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/cpr' 2025-10-10T01:19:48.0118566Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/fmt' 2025-10-10T01:19:48.0176371Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/gflags' 2025-10-10T01:19:48.0234601Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/gflags/doc' 2025-10-10T01:19:48.0307102Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/glog' 2025-10-10T01:19:48.0364245Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/googletest' 2025-10-10T01:19:48.0419579Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/json' 2025-10-10T01:19:48.0488108Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/pfs' 2025-10-10T01:19:48.0547119Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/prometheus-cpp' 2025-10-10T01:19:48.0601499Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/prometheus-cpp/3rdparty/civetweb' 2025-10-10T01:19:48.0673843Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/prometheus-cpp/3rdparty/googletest' 2025-10-10T01:19:48.0749406Z Entering 'third_party/kineto/libkineto/third_party/fmt' 2025-10-10T01:19:48.0817922Z Entering 'third_party/kineto/libkineto/third_party/googletest' 2025-10-10T01:19:48.0882969Z Entering 'third_party/kleidiai' 2025-10-10T01:19:48.0951410Z Entering 'third_party/mimalloc' 2025-10-10T01:19:48.1021008Z Entering 'third_party/nlohmann' 2025-10-10T01:19:48.1093699Z Entering 'third_party/onnx' 2025-10-10T01:19:48.1170553Z Entering 'third_party/onnx/third_party/pybind11' 2025-10-10T01:19:48.1247588Z Entering 'third_party/opentelemetry-cpp' 2025-10-10T01:19:48.1308799Z Entering 'third_party/opentelemetry-cpp/third_party/benchmark' 2025-10-10T01:19:48.1377493Z Entering 'third_party/opentelemetry-cpp/third_party/googletest' 2025-10-10T01:19:48.1433897Z Entering 'third_party/opentelemetry-cpp/third_party/ms-gsl' 2025-10-10T01:19:48.1499773Z Entering 'third_party/opentelemetry-cpp/third_party/nlohmann-json' 2025-10-10T01:19:48.1569959Z Entering 'third_party/opentelemetry-cpp/third_party/opentelemetry-proto' 2025-10-10T01:19:48.1633067Z Entering 'third_party/opentelemetry-cpp/third_party/opentracing-cpp' 2025-10-10T01:19:48.1700064Z Entering 'third_party/opentelemetry-cpp/third_party/prometheus-cpp' 2025-10-10T01:19:48.1765418Z Entering 'third_party/opentelemetry-cpp/third_party/prometheus-cpp/3rdparty/civetweb' 2025-10-10T01:19:48.1825789Z Entering 'third_party/opentelemetry-cpp/third_party/prometheus-cpp/3rdparty/googletest' 2025-10-10T01:19:48.1896839Z Entering 'third_party/opentelemetry-cpp/tools/vcpkg' 2025-10-10T01:19:48.1980929Z Entering 'third_party/pocketfft' 2025-10-10T01:19:48.2044794Z Entering 'third_party/protobuf' 2025-10-10T01:19:48.2112262Z Entering 'third_party/protobuf/third_party/benchmark' 2025-10-10T01:19:48.2178974Z Entering 'third_party/protobuf/third_party/googletest' 2025-10-10T01:19:48.2252879Z Entering 'third_party/psimd' 2025-10-10T01:19:48.2317803Z Entering 'third_party/pthreadpool' 2025-10-10T01:19:48.2385494Z Entering 'third_party/pybind11' 2025-10-10T01:19:48.2457126Z Entering 'third_party/python-peachpy' 2025-10-10T01:19:48.2517514Z Entering 'third_party/sleef' 2025-10-10T01:19:48.2585728Z Entering 'third_party/tensorpipe' 2025-10-10T01:19:48.2657096Z Entering 'third_party/tensorpipe/third_party/googletest' 2025-10-10T01:19:48.2719667Z Entering 'third_party/tensorpipe/third_party/libnop' 2025-10-10T01:19:48.2784133Z Entering 'third_party/tensorpipe/third_party/libuv' 2025-10-10T01:19:48.2849105Z Entering 'third_party/tensorpipe/third_party/pybind11' 2025-10-10T01:19:48.2916091Z Entering 'third_party/tensorpipe/third_party/pybind11/tools/clang' 2025-10-10T01:19:48.3009894Z ##[endgroup] 2025-10-10T01:19:48.3061619Z [command]/usr/bin/git log -1 --format=%H 2025-10-10T01:19:48.3093639Z 344e6365a0068c2d2847fcec0c55dd53291d475e 2025-10-10T01:19:48.3219372Z ##[group]Run cd "${GITHUB_WORKSPACE}" 2025-10-10T01:19:48.3219667Z cd "${GITHUB_WORKSPACE}" 2025-10-10T01:19:48.3219895Z # Clean stale submodule dirs 2025-10-10T01:19:48.3220125Z if [ -z "${NO_SUDO}" ]; then 2025-10-10T01:19:48.3220394Z  sudo git submodule foreach --recursive git clean -ffdx 2025-10-10T01:19:48.3220670Z else 2025-10-10T01:19:48.3220906Z  git submodule foreach --recursive git clean -ffdx 2025-10-10T01:19:48.3221183Z fi 2025-10-10T01:19:48.3233804Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2025-10-10T01:19:48.3234108Z env: 2025-10-10T01:19:48.3234278Z GIT_DEFAULT_BRANCH: main 2025-10-10T01:19:48.3234481Z NO_SUDO: true 2025-10-10T01:19:48.3234646Z ##[endgroup] 2025-10-10T01:19:48.3721987Z Entering 'android/libs/fbjni' 2025-10-10T01:19:48.3778504Z Entering 'third_party/FP16' 2025-10-10T01:19:48.3820625Z Entering 'third_party/FXdiv' 2025-10-10T01:19:48.3877370Z Entering 'third_party/NNPACK' 2025-10-10T01:19:48.3935651Z Entering 'third_party/NVTX' 2025-10-10T01:19:48.3999058Z Entering 'third_party/VulkanMemoryAllocator' 2025-10-10T01:19:48.4057888Z Entering 'third_party/XNNPACK' 2025-10-10T01:19:48.4213541Z Entering 'third_party/aiter' 2025-10-10T01:19:48.4278865Z Entering 'third_party/aiter/3rdparty/composable_kernel' 2025-10-10T01:19:48.4424626Z Entering 'third_party/benchmark' 2025-10-10T01:19:48.4475740Z Entering 'third_party/composable_kernel' 2025-10-10T01:19:48.4635483Z Entering 'third_party/cpp-httplib' 2025-10-10T01:19:48.4701741Z Entering 'third_party/cpuinfo' 2025-10-10T01:19:48.4763300Z Entering 'third_party/cudnn_frontend' 2025-10-10T01:19:48.4818837Z Entering 'third_party/cutlass' 2025-10-10T01:19:48.4948616Z Entering 'third_party/fbgemm' 2025-10-10T01:19:48.5039090Z Entering 'third_party/fbgemm/external/asmjit' 2025-10-10T01:19:48.5088548Z Entering 'third_party/fbgemm/external/composable_kernel' 2025-10-10T01:19:48.5246312Z Entering 'third_party/fbgemm/external/cpuinfo' 2025-10-10T01:19:48.5294806Z Entering 'third_party/fbgemm/external/cutlass' 2025-10-10T01:19:48.5418421Z Entering 'third_party/fbgemm/external/googletest' 2025-10-10T01:19:48.5477378Z Entering 'third_party/fbgemm/external/hipify_torch' 2025-10-10T01:19:48.5527643Z Entering 'third_party/fbgemm/external/json' 2025-10-10T01:19:48.5596362Z Entering 'third_party/flash-attention' 2025-10-10T01:19:48.5666881Z Entering 'third_party/flash-attention/csrc/composable_kernel' 2025-10-10T01:19:48.5795161Z Entering 'third_party/flash-attention/csrc/cutlass' 2025-10-10T01:19:48.5920279Z Entering 'third_party/flatbuffers' 2025-10-10T01:19:48.6029598Z Entering 'third_party/fmt' 2025-10-10T01:19:48.6087190Z Entering 'third_party/gemmlowp/gemmlowp' 2025-10-10T01:19:48.6145705Z Entering 'third_party/gloo' 2025-10-10T01:19:48.6205551Z Entering 'third_party/googletest' 2025-10-10T01:19:48.6254083Z Entering 'third_party/ideep' 2025-10-10T01:19:48.6297732Z Entering 'third_party/ideep/mkl-dnn' 2025-10-10T01:19:48.6416906Z Entering 'third_party/ittapi' 2025-10-10T01:19:48.6477571Z Entering 'third_party/kineto' 2025-10-10T01:19:48.6537343Z Entering 'third_party/kineto/libkineto/third_party/dynolog' 2025-10-10T01:19:48.6598768Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/DCGM' 2025-10-10T01:19:48.6672670Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/cpr' 2025-10-10T01:19:48.6723283Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/fmt' 2025-10-10T01:19:48.6779201Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/gflags' 2025-10-10T01:19:48.6824969Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/gflags/doc' 2025-10-10T01:19:48.6882127Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/glog' 2025-10-10T01:19:48.6926084Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/googletest' 2025-10-10T01:19:48.6980951Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/json' 2025-10-10T01:19:48.7049326Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/pfs' 2025-10-10T01:19:48.7108025Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/prometheus-cpp' 2025-10-10T01:19:48.7166753Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/prometheus-cpp/3rdparty/civetweb' 2025-10-10T01:19:48.7240318Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/prometheus-cpp/3rdparty/googletest' 2025-10-10T01:19:48.7304728Z Entering 'third_party/kineto/libkineto/third_party/fmt' 2025-10-10T01:19:48.7353962Z Entering 'third_party/kineto/libkineto/third_party/googletest' 2025-10-10T01:19:48.7410455Z Entering 'third_party/kleidiai' 2025-10-10T01:19:48.7471417Z Entering 'third_party/mimalloc' 2025-10-10T01:19:48.7527314Z Entering 'third_party/nlohmann' 2025-10-10T01:19:48.7600370Z Entering 'third_party/onnx' 2025-10-10T01:19:48.8025438Z Entering 'third_party/onnx/third_party/pybind11' 2025-10-10T01:19:48.8083117Z Entering 'third_party/opentelemetry-cpp' 2025-10-10T01:19:48.8168657Z Entering 'third_party/opentelemetry-cpp/third_party/benchmark' 2025-10-10T01:19:48.8230669Z Entering 'third_party/opentelemetry-cpp/third_party/googletest' 2025-10-10T01:19:48.8288878Z Entering 'third_party/opentelemetry-cpp/third_party/ms-gsl' 2025-10-10T01:19:48.8336531Z Entering 'third_party/opentelemetry-cpp/third_party/nlohmann-json' 2025-10-10T01:19:48.8397513Z Entering 'third_party/opentelemetry-cpp/third_party/opentelemetry-proto' 2025-10-10T01:19:48.8450092Z Entering 'third_party/opentelemetry-cpp/third_party/opentracing-cpp' 2025-10-10T01:19:48.8505962Z Entering 'third_party/opentelemetry-cpp/third_party/prometheus-cpp' 2025-10-10T01:19:48.8553997Z Entering 'third_party/opentelemetry-cpp/third_party/prometheus-cpp/3rdparty/civetweb' 2025-10-10T01:19:48.8630709Z Entering 'third_party/opentelemetry-cpp/third_party/prometheus-cpp/3rdparty/googletest' 2025-10-10T01:19:48.8687567Z Entering 'third_party/opentelemetry-cpp/tools/vcpkg' 2025-10-10T01:19:48.9030069Z Entering 'third_party/pocketfft' 2025-10-10T01:19:48.9087386Z Entering 'third_party/protobuf' 2025-10-10T01:19:48.9198069Z Entering 'third_party/protobuf/third_party/benchmark' 2025-10-10T01:19:48.9251516Z Entering 'third_party/protobuf/third_party/googletest' 2025-10-10T01:19:48.9312309Z Entering 'third_party/psimd' 2025-10-10T01:19:48.9368835Z Entering 'third_party/pthreadpool' 2025-10-10T01:19:48.9416311Z Entering 'third_party/pybind11' 2025-10-10T01:19:48.9469486Z Entering 'third_party/python-peachpy' 2025-10-10T01:19:48.9522146Z Entering 'third_party/sleef' 2025-10-10T01:19:48.9586897Z Entering 'third_party/tensorpipe' 2025-10-10T01:19:48.9649167Z Entering 'third_party/tensorpipe/third_party/googletest' 2025-10-10T01:19:48.9703994Z Entering 'third_party/tensorpipe/third_party/libnop' 2025-10-10T01:19:48.9757286Z Entering 'third_party/tensorpipe/third_party/libuv' 2025-10-10T01:19:48.9805717Z Entering 'third_party/tensorpipe/third_party/pybind11' 2025-10-10T01:19:48.9848051Z Entering 'third_party/tensorpipe/third_party/pybind11/tools/clang' 2025-10-10T01:19:49.0037913Z Prepare all required actions 2025-10-10T01:19:49.0038363Z Getting action download info 2025-10-10T01:19:49.1495501Z ##[group]Run ./.github/actions/setup-linux 2025-10-10T01:19:49.1495734Z env: 2025-10-10T01:19:49.1495898Z GIT_DEFAULT_BRANCH: main 2025-10-10T01:19:49.1496086Z ##[endgroup] 2025-10-10T01:19:49.1542620Z ##[group]Run set -euo pipefail 2025-10-10T01:19:49.1542868Z set -euo pipefail 2025-10-10T01:19:49.1543082Z function get_ec2_metadata() { 2025-10-10T01:19:49.1543356Z  # Pulled from instance metadata endpoint for EC2 2025-10-10T01:19:49.1543822Z  # see https://docs.aws.amazon.com/AWSEC2/latest/UserGuide/instancedata-data-retrieval.html 2025-10-10T01:19:49.1544226Z  category=$1 2025-10-10T01:19:49.1544488Z  # If it is GCP runner (runner name contains gcp), do not run this 2025-10-10T01:19:49.1544810Z  runner_name_str=i-0b2d5251fda19aeb9 2025-10-10T01:19:49.1545081Z  if [[ -f /.inarc ]]; then 2025-10-10T01:19:49.1545333Z  echo "ARC Runner, no info on ec2 metadata" 2025-10-10T01:19:49.1545606Z  elif [[ $runner_name_str == *"gcp"* ]]; then 2025-10-10T01:19:49.1545940Z  echo "Runner is from Google Cloud Platform, No info on ec2 metadata" 2025-10-10T01:19:49.1546239Z  else 2025-10-10T01:19:49.1546850Z  curl -H "X-aws-ec2-metadata-token: $(curl -s -X PUT "http://169.254.169.254/latest/api/token" -H "X-aws-ec2-metadata-token-ttl-seconds: 30")" -fsSL "http://169.254.169.254/latest/meta-data/${category}" 2025-10-10T01:19:49.1547480Z  fi 2025-10-10T01:19:49.1547628Z } 2025-10-10T01:19:49.1547817Z echo "ami-id: $(get_ec2_metadata ami-id)" 2025-10-10T01:19:49.1548117Z echo "instance-id: $(get_ec2_metadata instance-id)" 2025-10-10T01:19:49.1548458Z echo "instance-type: $(get_ec2_metadata instance-type)" 2025-10-10T01:19:49.1548746Z echo "system info $(uname -a)" 2025-10-10T01:19:49.1557864Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2025-10-10T01:19:49.1558153Z env: 2025-10-10T01:19:49.1558310Z GIT_DEFAULT_BRANCH: main 2025-10-10T01:19:49.1558586Z ##[endgroup] 2025-10-10T01:19:49.1748687Z ami-id: ami-08982f1c5bf93d976 2025-10-10T01:19:49.1886269Z instance-id: i-0b2d5251fda19aeb9 2025-10-10T01:19:49.2039603Z instance-type: g6.12xlarge 2025-10-10T01:19:49.2055789Z system info Linux ip-10-0-6-129.ec2.internal 6.1.150-174.273.amzn2023.x86_64 #1 SMP PREEMPT_DYNAMIC Tue Sep 9 12:21:26 UTC 2025 x86_64 x86_64 x86_64 GNU/Linux 2025-10-10T01:19:49.2096089Z ##[group]Run if [ -f /usr/bin/nvidia-smi ]; then nvidia-smi; fi 2025-10-10T01:19:49.2096451Z if [ -f /usr/bin/nvidia-smi ]; then nvidia-smi; fi 2025-10-10T01:19:49.2103975Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2025-10-10T01:19:49.2104248Z env: 2025-10-10T01:19:49.2104409Z GIT_DEFAULT_BRANCH: main 2025-10-10T01:19:49.2104599Z ##[endgroup] 2025-10-10T01:19:51.5702910Z Fri Oct 10 01:19:51 2025 2025-10-10T01:19:51.5703730Z +-----------------------------------------------------------------------------------------+ 2025-10-10T01:19:51.5704695Z | NVIDIA-SMI 580.82.07 Driver Version: 580.82.07 CUDA Version: 13.0 | 2025-10-10T01:19:51.5705585Z +-----------------------------------------+------------------------+----------------------+ 2025-10-10T01:19:51.5706473Z | GPU Name Persistence-M | Bus-Id Disp.A | Volatile Uncorr. ECC | 2025-10-10T01:19:51.5707407Z | Fan Temp Perf Pwr:Usage/Cap | Memory-Usage | GPU-Util Compute M. | 2025-10-10T01:19:51.5739243Z | | | MIG M. | 2025-10-10T01:19:51.5739611Z |=========================================+========================+======================| 2025-10-10T01:19:51.5987030Z | 0 NVIDIA L4 Off | 00000000:38:00.0 Off | 0 | 2025-10-10T01:19:51.5988456Z | N/A 39C P0 27W / 72W | 0MiB / 23034MiB | 0% Default | 2025-10-10T01:19:51.5989211Z | | | N/A | 2025-10-10T01:19:51.5989915Z +-----------------------------------------+------------------------+----------------------+ 2025-10-10T01:19:51.5990687Z | 1 NVIDIA L4 Off | 00000000:3A:00.0 Off | 0 | 2025-10-10T01:19:51.5991409Z | N/A 37C P0 27W / 72W | 0MiB / 23034MiB | 0% Default | 2025-10-10T01:19:51.5992075Z | | | N/A | 2025-10-10T01:19:51.5992753Z +-----------------------------------------+------------------------+----------------------+ 2025-10-10T01:19:51.5993147Z | 2 NVIDIA L4 Off | 00000000:3C:00.0 Off | 0 | 2025-10-10T01:19:51.5993482Z | N/A 39C P0 27W / 72W | 0MiB / 23034MiB | 0% Default | 2025-10-10T01:19:51.5993766Z | | | N/A | 2025-10-10T01:19:51.5994064Z +-----------------------------------------+------------------------+----------------------+ 2025-10-10T01:19:51.5994386Z | 3 NVIDIA L4 Off | 00000000:3E:00.0 Off | 0 | 2025-10-10T01:19:51.5994692Z | N/A 38C P0 27W / 72W | 0MiB / 23034MiB | 0% Default | 2025-10-10T01:19:51.5994963Z | | | N/A | 2025-10-10T01:19:51.5995253Z +-----------------------------------------+------------------------+----------------------+ 2025-10-10T01:19:51.5995489Z 2025-10-10T01:19:51.5995617Z +-----------------------------------------------------------------------------------------+ 2025-10-10T01:19:51.5995940Z | Processes: | 2025-10-10T01:19:51.5996269Z | GPU GI CI PID Type Process name GPU Memory | 2025-10-10T01:19:51.5996576Z | ID ID Usage | 2025-10-10T01:19:51.5996822Z |=========================================================================================| 2025-10-10T01:19:51.6011311Z | No running processes found | 2025-10-10T01:19:51.6012176Z +-----------------------------------------------------------------------------------------+ 2025-10-10T01:19:52.8895287Z ##[group]Run echo "IN_CONTAINER_RUNNER=$(if [ -f /.inarc ] || [ -f /.incontainer ]; then echo true ; else echo false; fi)" >> "$GITHUB_OUTPUT" 2025-10-10T01:19:52.8896012Z echo "IN_CONTAINER_RUNNER=$(if [ -f /.inarc ] || [ -f /.incontainer ]; then echo true ; else echo false; fi)" >> "$GITHUB_OUTPUT" 2025-10-10T01:19:52.8905464Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2025-10-10T01:19:52.8905764Z env: 2025-10-10T01:19:52.8905925Z GIT_DEFAULT_BRANCH: main 2025-10-10T01:19:52.8906127Z ##[endgroup] 2025-10-10T01:19:52.8994385Z ##[group]Run if systemctl is-active --quiet docker; then 2025-10-10T01:19:52.8994729Z if systemctl is-active --quiet docker; then 2025-10-10T01:19:52.8995010Z  echo "Docker daemon is running..."; 2025-10-10T01:19:52.8995241Z else 2025-10-10T01:19:52.8995501Z  echo "Starting docker daemon..." && sudo systemctl start docker; 2025-10-10T01:19:52.8996029Z fi 2025-10-10T01:19:52.9004743Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2025-10-10T01:19:52.9005031Z env: 2025-10-10T01:19:52.9005192Z GIT_DEFAULT_BRANCH: main 2025-10-10T01:19:52.9005386Z ##[endgroup] 2025-10-10T01:19:52.9107855Z Docker daemon is running... 2025-10-10T01:19:52.9163106Z ##[group]Run nick-fields/retry@v3.0.0 2025-10-10T01:19:52.9163329Z with: 2025-10-10T01:19:52.9163473Z shell: bash 2025-10-10T01:19:52.9163899Z timeout_minutes: 5 2025-10-10T01:19:52.9164086Z max_attempts: 3 2025-10-10T01:19:52.9164256Z retry_wait_seconds: 30 2025-10-10T01:19:52.9165884Z command: AWS_ACCOUNT_ID=$(aws sts get-caller-identity|grep Account|cut -f4 -d\") aws ecr get-login-password --region "$AWS_DEFAULT_REGION" | docker login --username AWS \ --password-stdin "$AWS_ACCOUNT_ID.dkr.ecr.$AWS_DEFAULT_REGION.amazonaws.com" # For LF Runners we need to make sure we also login to Meta's ECR docker registry too. META_AWS_ACCOUNT_ID=308535385114 if [ "$AWS_ACCOUNT_ID" != "$META_AWS_ACCOUNT_ID" ] ; then aws ecr get-login-password --region "$AWS_DEFAULT_REGION" | docker login --username AWS \ --password-stdin "$META_AWS_ACCOUNT_ID.dkr.ecr.$AWS_DEFAULT_REGION.amazonaws.com" fi 2025-10-10T01:19:52.9167518Z polling_interval_seconds: 1 2025-10-10T01:19:52.9167722Z warning_on_retry: true 2025-10-10T01:19:52.9167908Z continue_on_error: false 2025-10-10T01:19:52.9168084Z env: 2025-10-10T01:19:52.9168243Z GIT_DEFAULT_BRANCH: main 2025-10-10T01:19:52.9168432Z AWS_RETRY_MODE: standard 2025-10-10T01:19:52.9168612Z AWS_MAX_ATTEMPTS: 5 2025-10-10T01:19:52.9168790Z AWS_DEFAULT_REGION: us-east-1 2025-10-10T01:19:52.9168986Z ##[endgroup] 2025-10-10T01:19:53.9788694Z WARNING! Your password will be stored unencrypted in /home/ec2-user/.docker/config.json. 2025-10-10T01:19:53.9789311Z Configure a credential helper to remove this warning. See 2025-10-10T01:19:53.9789854Z https://docs.docker.com/engine/reference/commandline/login/#credentials-store 2025-10-10T01:19:53.9790233Z 2025-10-10T01:19:53.9790338Z Login Succeeded 2025-10-10T01:19:53.9933253Z Command completed after 1 attempt(s). 2025-10-10T01:19:54.0005410Z ##[group]Run env | grep '^GITHUB' >> "/tmp/github_env_${GITHUB_RUN_ID}" 2025-10-10T01:19:54.0005810Z env | grep '^GITHUB' >> "/tmp/github_env_${GITHUB_RUN_ID}" 2025-10-10T01:19:54.0006138Z env | grep '^CI' >> "/tmp/github_env_${GITHUB_RUN_ID}" 2025-10-10T01:19:54.0014438Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2025-10-10T01:19:54.0014721Z env: 2025-10-10T01:19:54.0014888Z GIT_DEFAULT_BRANCH: main 2025-10-10T01:19:54.0015088Z ##[endgroup] 2025-10-10T01:19:54.0156260Z ##[group]Run # ignore expansion of "docker ps -q" since it could be empty 2025-10-10T01:19:54.0156678Z # ignore expansion of "docker ps -q" since it could be empty 2025-10-10T01:19:54.0156995Z # shellcheck disable=SC2046 2025-10-10T01:19:54.0157262Z docker stop $(docker ps -q) || true 2025-10-10T01:19:54.0157536Z # Prune all of the docker images 2025-10-10T01:19:54.0157777Z docker system prune -af 2025-10-10T01:19:54.0165701Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2025-10-10T01:19:54.0165995Z env: 2025-10-10T01:19:54.0166162Z GIT_DEFAULT_BRANCH: main 2025-10-10T01:19:54.0166359Z ##[endgroup] 2025-10-10T01:19:54.0500945Z "docker stop" requires at least 1 argument. 2025-10-10T01:19:54.0501497Z See 'docker stop --help'. 2025-10-10T01:19:54.0501795Z 2025-10-10T01:19:54.0502028Z Usage: docker stop [OPTIONS] CONTAINER [CONTAINER...] 2025-10-10T01:19:54.0502396Z 2025-10-10T01:19:54.0502568Z Stop one or more running containers 2025-10-10T01:19:54.0733995Z Total reclaimed space: 0B 2025-10-10T01:19:54.0900928Z ##[group]Run pytorch/test-infra/.github/actions/calculate-docker-image@main 2025-10-10T01:19:54.0901313Z with: 2025-10-10T01:19:54.0901929Z docker-image-name: 308535385114.dkr.ecr.us-east-1.amazonaws.com/pytorch/ci-image:pytorch-linux-jammy-cuda12.8-cudnn9-py3.12-gcc11-vllm-d8be0384e085f551506bd739678109fa0f5ee7ac 2025-10-10T01:19:54.0902798Z use-custom-docker-registry: true 2025-10-10T01:19:54.0903043Z docker-build-dir: .ci/docker 2025-10-10T01:19:54.0903256Z docker-build-script: ./build.sh 2025-10-10T01:19:54.0903470Z working-directory: . 2025-10-10T01:19:54.0903724Z docker-registry: 308535385114.dkr.ecr.us-east-1.amazonaws.com 2025-10-10T01:19:54.0904017Z force-push: false 2025-10-10T01:19:54.0904186Z env: 2025-10-10T01:19:54.0904341Z GIT_DEFAULT_BRANCH: main 2025-10-10T01:19:54.0904528Z ##[endgroup] 2025-10-10T01:19:54.0932857Z ##[group]Run set -ex 2025-10-10T01:19:54.0933164Z set -ex 2025-10-10T01:19:54.0933358Z  2025-10-10T01:19:54.0933727Z # If the docker build directory or the build script doesn't exist, the action will 2025-10-10T01:19:54.0934243Z # gracefully return the docker image name as it is. Pulling docker image in Linux 2025-10-10T01:19:54.0934645Z # job could then download the pre-built image as usual 2025-10-10T01:19:54.0935142Z if [[ -d "${DOCKER_BUILD_DIR}" ]] && [[ -f "${DOCKER_BUILD_DIR}/${DOCKER_BUILD_SCRIPT}" ]] && [[ "${USE_CUSTOM_DOCKER_REGISTRY}" == "true" ]]; then 2025-10-10T01:19:54.0935600Z  echo "skip=false" >> "${GITHUB_OUTPUT}" 2025-10-10T01:19:54.0935836Z else 2025-10-10T01:19:54.0936030Z  echo "skip=true" >> "${GITHUB_OUTPUT}" 2025-10-10T01:19:54.0936353Z  echo "docker-image=${DOCKER_IMAGE_NAME}" >> "${GITHUB_OUTPUT}" 2025-10-10T01:19:54.0936648Z  2025-10-10T01:19:54.0937054Z  echo "Not using custom ECR registry. Either it was not requested or there is no Docker build script in the ${REPO_NAME} repo..." 2025-10-10T01:19:54.0937507Z  exit 0 2025-10-10T01:19:54.0937657Z fi 2025-10-10T01:19:54.0937807Z  2025-10-10T01:19:54.0938055Z if [[ "${DOCKER_IMAGE_NAME}" == *"${DOCKER_REGISTRY}/${REPO_NAME}"* ]]; then 2025-10-10T01:19:54.0938491Z  # The docker image name already includes the ECR prefix and tag, so we can just 2025-10-10T01:19:54.0938881Z  # use it as it is, but first let's extract the tag 2025-10-10T01:19:54.0939217Z  DOCKER_TAG=$(echo "${DOCKER_IMAGE_NAME}" | awk -F '[:,]' '{print $2}') 2025-10-10T01:19:54.0939588Z  echo "docker-tag=${DOCKER_TAG}" >> "${GITHUB_OUTPUT}" 2025-10-10T01:19:54.0939935Z  echo "docker-image=${DOCKER_IMAGE_NAME}" >> "${GITHUB_OUTPUT}" 2025-10-10T01:19:54.0940222Z else 2025-10-10T01:19:54.0940409Z  if [[ "${DOCKER_IMAGE_NAME}" == *:* ]]; then 2025-10-10T01:19:54.0940694Z  CUSTOM_TAG_PREFIX=${DOCKER_IMAGE_NAME#*:} 2025-10-10T01:19:54.0940978Z  DOCKER_IMAGE_NAME=${DOCKER_IMAGE_NAME%%:*} 2025-10-10T01:19:54.0941213Z  fi 2025-10-10T01:19:54.0941531Z  DOCKER_TAG=${CUSTOM_TAG_PREFIX:+${CUSTOM_TAG_PREFIX}-}$(git rev-parse HEAD:"${DOCKER_BUILD_DIR}") 2025-10-10T01:19:54.0941956Z  echo "docker-tag=${DOCKER_TAG}" >> "${GITHUB_OUTPUT}" 2025-10-10T01:19:54.0942405Z  echo "docker-image=${DOCKER_REGISTRY}/${REPO_NAME}/${DOCKER_IMAGE_NAME}:${DOCKER_TAG}" >> "${GITHUB_OUTPUT}" 2025-10-10T01:19:54.0942902Z  echo "custom-tag-prefix=${CUSTOM_TAG_PREFIX}" >> "${GITHUB_OUTPUT}" 2025-10-10T01:19:54.0943204Z fi 2025-10-10T01:19:54.0951410Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2025-10-10T01:19:54.0951683Z env: 2025-10-10T01:19:54.0951840Z GIT_DEFAULT_BRANCH: main 2025-10-10T01:19:54.0952035Z REPO_NAME: pytorch 2025-10-10T01:19:54.0952823Z DOCKER_IMAGE_NAME: 308535385114.dkr.ecr.us-east-1.amazonaws.com/pytorch/ci-image:pytorch-linux-jammy-cuda12.8-cudnn9-py3.12-gcc11-vllm-d8be0384e085f551506bd739678109fa0f5ee7ac 2025-10-10T01:19:54.0953498Z DOCKER_BUILD_DIR: .ci/docker 2025-10-10T01:19:54.0953707Z DOCKER_BUILD_SCRIPT: ./build.sh 2025-10-10T01:19:54.0953988Z DOCKER_REGISTRY: 308535385114.dkr.ecr.us-east-1.amazonaws.com 2025-10-10T01:19:54.0954415Z USE_CUSTOM_DOCKER_REGISTRY: true 2025-10-10T01:19:54.0954628Z CUSTOM_TAG_PREFIX: 2025-10-10T01:19:54.0954801Z ##[endgroup] 2025-10-10T01:19:54.0986034Z + [[ -d .ci/docker ]] 2025-10-10T01:19:54.0986591Z + [[ -f .ci/docker/./build.sh ]] 2025-10-10T01:19:54.0987100Z + [[ true == \t\r\u\e ]] 2025-10-10T01:19:54.0987539Z + echo skip=false 2025-10-10T01:19:54.0989401Z + [[ 308535385114.dkr.ecr.us-east-1.amazonaws.com/pytorch/ci-image:pytorch-linux-jammy-cuda12.8-cudnn9-py3.12-gcc11-vllm-d8be0384e085f551506bd739678109fa0f5ee7ac == *\3\0\8\5\3\5\3\8\5\1\1\4\.\d\k\r\.\e\c\r\.\u\s\-\e\a\s\t\-\1\.\a\m\a\z\o\n\a\w\s\.\c\o\m\/\p\y\t\o\r\c\h* ]] 2025-10-10T01:19:54.0995233Z ++ awk -F '[:,]' '{print $2}' 2025-10-10T01:19:54.0996253Z ++ echo 308535385114.dkr.ecr.us-east-1.amazonaws.com/pytorch/ci-image:pytorch-linux-jammy-cuda12.8-cudnn9-py3.12-gcc11-vllm-d8be0384e085f551506bd739678109fa0f5ee7ac 2025-10-10T01:19:54.1023111Z + DOCKER_TAG=pytorch-linux-jammy-cuda12.8-cudnn9-py3.12-gcc11-vllm-d8be0384e085f551506bd739678109fa0f5ee7ac 2025-10-10T01:19:54.1025769Z + echo docker-tag=pytorch-linux-jammy-cuda12.8-cudnn9-py3.12-gcc11-vllm-d8be0384e085f551506bd739678109fa0f5ee7ac 2025-10-10T01:19:54.1027943Z + echo docker-image=308535385114.dkr.ecr.us-east-1.amazonaws.com/pytorch/ci-image:pytorch-linux-jammy-cuda12.8-cudnn9-py3.12-gcc11-vllm-d8be0384e085f551506bd739678109fa0f5ee7ac 2025-10-10T01:19:54.1081250Z ##[group]Run set +e 2025-10-10T01:19:54.1081528Z set +e 2025-10-10T01:19:54.1081709Z set -x 2025-10-10T01:19:54.1081869Z  2025-10-10T01:19:54.1082019Z login() { 2025-10-10T01:19:54.1082392Z  aws ecr get-login-password --region us-east-1 | docker login -u AWS --password-stdin "$1" 2025-10-10T01:19:54.1082764Z } 2025-10-10T01:19:54.1082912Z  2025-10-10T01:19:54.1083062Z retry () { 2025-10-10T01:19:54.1083263Z  $* || (sleep 1 && $*) || (sleep 2 && $*) 2025-10-10T01:19:54.1083493Z } 2025-10-10T01:19:54.1083635Z  2025-10-10T01:19:54.1083798Z retry login "${DOCKER_REGISTRY}" 2025-10-10T01:19:54.1084012Z  2025-10-10T01:19:54.1084169Z START_TIME=$(date +%s) 2025-10-10T01:19:54.1084377Z # Wait up to 120 minutes 2025-10-10T01:19:54.1084645Z while [[ $(( $(date +%s) - 7200 )) -lt $START_TIME ]]; do 2025-10-10T01:19:54.1085014Z  # Check if image already exists, if it does then skip building it 2025-10-10T01:19:54.1085371Z  if docker manifest inspect "${DOCKER_IMAGE}"; then 2025-10-10T01:19:54.1085629Z  exit 0 2025-10-10T01:19:54.1085798Z  fi 2025-10-10T01:19:54.1085953Z  2025-10-10T01:19:54.1086245Z  # NB: This flag is used by Docker build workflow to push the image to ECR, so we can 2025-10-10T01:19:54.1086727Z  # use this to differentiate between the Docker build and regular build jobs. For the 2025-10-10T01:19:54.1087181Z  # latter, it will wait for the Docker images to become available before continuing 2025-10-10T01:19:54.1087558Z  if [ "${DOCKER_PUSH:-false}" == "true" ]; then 2025-10-10T01:19:54.1087856Z  # It's a Docker build job, let's build the image 2025-10-10T01:19:54.1088108Z  break 2025-10-10T01:19:54.1088270Z  else 2025-10-10T01:19:54.1088518Z  # It's a regular build job, wait for the image to become available 2025-10-10T01:19:54.1088815Z  sleep 300 2025-10-10T01:19:54.1088993Z  fi 2025-10-10T01:19:54.1089142Z done 2025-10-10T01:19:54.1089293Z  2025-10-10T01:19:54.1089730Z # NB: This part requires a full checkout. Otherwise, the merge base will 2025-10-10T01:19:54.1090156Z # be empty. The default action would be to continue rebuild the image 2025-10-10T01:19:54.1090523Z if [[ "$BASE_REVISION" = "$(git rev-parse HEAD)" ]]; then 2025-10-10T01:19:54.1090844Z  # if we're on the base branch then use the parent commit 2025-10-10T01:19:54.1091242Z  MERGE_BASE=$(git rev-parse HEAD~) 2025-10-10T01:19:54.1091472Z else 2025-10-10T01:19:54.1091708Z  # otherwise we're on a PR, so use the most recent base commit 2025-10-10T01:19:54.1092045Z  MERGE_BASE=$(git merge-base HEAD "$BASE_REVISION") 2025-10-10T01:19:54.1092299Z fi 2025-10-10T01:19:54.1092456Z  2025-10-10T01:19:54.1092627Z if [[ -z "${MERGE_BASE}" ]]; then 2025-10-10T01:19:54.1092894Z  echo "rebuild=true" >> "${GITHUB_OUTPUT}" 2025-10-10T01:19:54.1093124Z  2025-10-10T01:19:54.1093457Z  echo "Finding merge base only works with full checkout, please set fetch-depth to 0, continuing ..." 2025-10-10T01:19:54.1093844Z  exit 0 2025-10-10T01:19:54.1094002Z fi 2025-10-10T01:19:54.1094139Z  2025-10-10T01:19:54.1094356Z if ! git rev-parse "${MERGE_BASE}:${DOCKER_BUILD_DIR}"; then 2025-10-10T01:19:54.1094849Z  echo "Directory '${DOCKER_BUILD_DIR}' not found in commit $MERGE_BASE, you should rebase onto a more recent commit" 2025-10-10T01:19:54.1095262Z  exit 1 2025-10-10T01:19:54.1095410Z fi 2025-10-10T01:19:54.1095568Z  2025-10-10T01:19:54.1095824Z PREVIOUS_DOCKER_TAG=$(git rev-parse "${MERGE_BASE}:${DOCKER_BUILD_DIR}") 2025-10-10T01:19:54.1096286Z # If no image exists but the hash is the same as the previous hash then we should error out here 2025-10-10T01:19:54.1096705Z if [[ "${PREVIOUS_DOCKER_TAG}" == "${DOCKER_TAG}" ]]; then 2025-10-10T01:19:54.1097184Z  echo "WARNING: Something has gone wrong and the previous image isn't available for the merge-base of your branch" 2025-10-10T01:19:54.1097730Z  echo " Will re-build docker image to store in local cache, TTS may be longer" 2025-10-10T01:19:54.1098052Z fi 2025-10-10T01:19:54.1098202Z  2025-10-10T01:19:54.1098386Z echo "rebuild=true" >> "${GITHUB_OUTPUT}" 2025-10-10T01:19:54.1105431Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2025-10-10T01:19:54.1105713Z env: 2025-10-10T01:19:54.1105885Z GIT_DEFAULT_BRANCH: main 2025-10-10T01:19:54.1106092Z DOCKER_BUILD_DIR: .ci/docker 2025-10-10T01:19:54.1106344Z BASE_REVISION: 344e6365a0068c2d2847fcec0c55dd53291d475e 2025-10-10T01:19:54.1107035Z DOCKER_IMAGE: 308535385114.dkr.ecr.us-east-1.amazonaws.com/pytorch/ci-image:pytorch-linux-jammy-cuda12.8-cudnn9-py3.12-gcc11-vllm-d8be0384e085f551506bd739678109fa0f5ee7ac 2025-10-10T01:19:54.1107894Z DOCKER_TAG: pytorch-linux-jammy-cuda12.8-cudnn9-py3.12-gcc11-vllm-d8be0384e085f551506bd739678109fa0f5ee7ac 2025-10-10T01:19:54.1108415Z DOCKER_REGISTRY: 308535385114.dkr.ecr.us-east-1.amazonaws.com 2025-10-10T01:19:54.1108697Z DOCKER_PUSH: 2025-10-10T01:19:54.1108858Z ##[endgroup] 2025-10-10T01:19:54.1138412Z + retry login 308535385114.dkr.ecr.us-east-1.amazonaws.com 2025-10-10T01:19:54.1138761Z + login 308535385114.dkr.ecr.us-east-1.amazonaws.com 2025-10-10T01:19:54.1143789Z + docker login -u AWS --password-stdin 308535385114.dkr.ecr.us-east-1.amazonaws.com 2025-10-10T01:19:54.1144322Z + aws ecr get-login-password --region us-east-1 2025-10-10T01:19:54.6051752Z WARNING! Your password will be stored unencrypted in /home/ec2-user/.docker/config.json. 2025-10-10T01:19:54.6053105Z Configure a credential helper to remove this warning. See 2025-10-10T01:19:54.6054334Z https://docs.docker.com/engine/reference/commandline/login/#credentials-store 2025-10-10T01:19:54.6055411Z 2025-10-10T01:19:54.6055776Z Login Succeeded 2025-10-10T01:19:54.6087973Z ++ date +%s 2025-10-10T01:19:54.6103798Z + START_TIME=1760059194 2025-10-10T01:19:54.6108308Z ++ date +%s 2025-10-10T01:19:54.6120109Z + [[ 1760051994 -lt 1760059194 ]] 2025-10-10T01:19:54.6122077Z + docker manifest inspect 308535385114.dkr.ecr.us-east-1.amazonaws.com/pytorch/ci-image:pytorch-linux-jammy-cuda12.8-cudnn9-py3.12-gcc11-vllm-d8be0384e085f551506bd739678109fa0f5ee7ac 2025-10-10T01:19:54.8436865Z { 2025-10-10T01:19:54.8437560Z "schemaVersion": 2, 2025-10-10T01:19:54.8438677Z "mediaType": "application/vnd.docker.distribution.manifest.v2+json", 2025-10-10T01:19:54.8439134Z "config": { 2025-10-10T01:19:54.8439450Z "mediaType": "application/vnd.docker.container.image.v1+json", 2025-10-10T01:19:54.8439826Z "size": 31341, 2025-10-10T01:19:54.8440195Z "digest": "sha256:fb0d053588b3e7d110638a9f678211e76aa4515ed1bad57389e38e2ab079e7b1" 2025-10-10T01:19:54.8440621Z }, 2025-10-10T01:19:54.8440808Z "layers": [ 2025-10-10T01:19:54.8441000Z { 2025-10-10T01:19:54.8441302Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-10-10T01:19:54.8441687Z "size": 30447990, 2025-10-10T01:19:54.8442080Z "digest": "sha256:828c1365039a657352c737a62d13e1932951b5658eb6bd9b9096ea9b73562453" 2025-10-10T01:19:54.8442488Z }, 2025-10-10T01:19:54.8442709Z { 2025-10-10T01:19:54.8443006Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-10-10T01:19:54.8443401Z "size": 1554, 2025-10-10T01:19:54.8443830Z "digest": "sha256:140bb29742fd93f99a9a09900147573105606f902b116e98f74b52701d38c892" 2025-10-10T01:19:54.8444294Z }, 2025-10-10T01:19:54.8444506Z { 2025-10-10T01:19:54.8444858Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-10-10T01:19:54.8445221Z "size": 313649951, 2025-10-10T01:19:54.8445593Z "digest": "sha256:99ee2db1caf9d1c2ddb56a54c2f9c956e2904d061a8f3c4cb990ec84d754fe89" 2025-10-10T01:19:54.8445971Z }, 2025-10-10T01:19:54.8446101Z { 2025-10-10T01:19:54.8446323Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-10-10T01:19:54.8446617Z "size": 791, 2025-10-10T01:19:54.8446908Z "digest": "sha256:f5a2330f87a859f908a0c132821f161aca3279ad9972ba6d60cbf7f776ab7611" 2025-10-10T01:19:54.8447245Z }, 2025-10-10T01:19:54.8447379Z { 2025-10-10T01:19:54.8447603Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-10-10T01:19:54.8447894Z "size": 106, 2025-10-10T01:19:54.8448194Z "digest": "sha256:4643d3cb7e67efd8a0ab8abfd1953313811ea4d28f402ccfd94953bae4ee8df3" 2025-10-10T01:19:54.8448526Z }, 2025-10-10T01:19:54.8448656Z { 2025-10-10T01:19:54.8448886Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-10-10T01:19:54.8449180Z "size": 703, 2025-10-10T01:19:54.8449470Z "digest": "sha256:5c871f60f6d4978a0a9d4b692a820cfa1b29c73e1244d641d570114156db0c50" 2025-10-10T01:19:54.8449792Z }, 2025-10-10T01:19:54.8449922Z { 2025-10-10T01:19:54.8450144Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-10-10T01:19:54.8450428Z "size": 1214, 2025-10-10T01:19:54.8450708Z "digest": "sha256:60a2970be007b6e9760d42161c7b5ac43192a16730a7d13b7cd8f268757a952a" 2025-10-10T01:19:54.8451030Z }, 2025-10-10T01:19:54.8451159Z { 2025-10-10T01:19:54.8451383Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-10-10T01:19:54.8451660Z "size": 485, 2025-10-10T01:19:54.8451938Z "digest": "sha256:43d427ee0b00452f669e2850ca0e4b80fbeb33cb3330ab2cbaab0e4fd27f4a49" 2025-10-10T01:19:54.8452274Z }, 2025-10-10T01:19:54.8452403Z { 2025-10-10T01:19:54.8452624Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-10-10T01:19:54.8452914Z "size": 110343651, 2025-10-10T01:19:54.8453224Z "digest": "sha256:728b6e1eb0128b7579525ddaa9ddbe96c7b04e6cb822fc419575c37c2bb3fbc4" 2025-10-10T01:19:54.8453552Z }, 2025-10-10T01:19:54.8453691Z { 2025-10-10T01:19:54.8453911Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-10-10T01:19:54.8454193Z "size": 5038, 2025-10-10T01:19:54.8454480Z "digest": "sha256:1ffc73eae77b7a84ce198ea2a44254cb1ae80fbcc00e7b26877043256b7d8864" 2025-10-10T01:19:54.8454997Z }, 2025-10-10T01:19:54.8455163Z { 2025-10-10T01:19:54.8455453Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-10-10T01:19:54.8455748Z "size": 1710, 2025-10-10T01:19:54.8456022Z "digest": "sha256:1c5379881a3b631c21f573017e01423112af82e3ac53f16992d23abb4d39aa27" 2025-10-10T01:19:54.8456501Z }, 2025-10-10T01:19:54.8456637Z { 2025-10-10T01:19:54.8456867Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-10-10T01:19:54.8457141Z "size": 724, 2025-10-10T01:19:54.8457419Z "digest": "sha256:3e62aac2efe8642154346954eca23d31a01d7d6558ac7f6c2f6ebdc2a2918ec0" 2025-10-10T01:19:54.8457744Z }, 2025-10-10T01:19:54.8457882Z { 2025-10-10T01:19:54.8458096Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-10-10T01:19:54.8458378Z "size": 543, 2025-10-10T01:19:54.8458661Z "digest": "sha256:3c1f345eea91d1915d7438b5d8f77b9bcec1ced3dfff7d3ed56f622a50944ba4" 2025-10-10T01:19:54.8458987Z }, 2025-10-10T01:19:54.8459117Z { 2025-10-10T01:19:54.8459334Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-10-10T01:19:54.8459621Z "size": 3412426191, 2025-10-10T01:19:54.8459908Z "digest": "sha256:a03ec53614528e4964121dfd061df88348df0558d1f8847d293703fb307e8565" 2025-10-10T01:19:54.8460222Z }, 2025-10-10T01:19:54.8460344Z { 2025-10-10T01:19:54.8460560Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-10-10T01:19:54.8460840Z "size": 32, 2025-10-10T01:19:54.8461115Z "digest": "sha256:4f4fb700ef54461cfa02571ae0db9a0dc1e0cdb5577484a6d75e68dc38e8acc1" 2025-10-10T01:19:54.8461440Z }, 2025-10-10T01:19:54.8461586Z { 2025-10-10T01:19:54.8461805Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-10-10T01:19:54.8462081Z "size": 380, 2025-10-10T01:19:54.8462364Z "digest": "sha256:84681debfb5c07ea57fc358c704ffa8b63dbcaf76db8e2dd7b12131db63760e1" 2025-10-10T01:19:54.8462696Z }, 2025-10-10T01:19:54.8462827Z { 2025-10-10T01:19:54.8463043Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-10-10T01:19:54.8463332Z "size": 236058, 2025-10-10T01:19:54.8463619Z "digest": "sha256:d6a77cb34f2d5411c23ad48da2ff399a5f331e2060be3c8cef0711829c69c696" 2025-10-10T01:19:54.8463965Z }, 2025-10-10T01:19:54.8464103Z { 2025-10-10T01:19:54.8464317Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-10-10T01:19:54.8464601Z "size": 231, 2025-10-10T01:19:54.8464875Z "digest": "sha256:b4e994693c0386d6b5a32dbc8f02222c10e8029b5714c7a9a77c04d0555a3cb5" 2025-10-10T01:19:54.8465187Z }, 2025-10-10T01:19:54.8465311Z { 2025-10-10T01:19:54.8465531Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-10-10T01:19:54.8465816Z "size": 3548009, 2025-10-10T01:19:54.8466110Z "digest": "sha256:c6c1bfa3b4c534eaa64a0e9b78533c1c31e6368ceb0f5a8ae5362e44af9a8ee1" 2025-10-10T01:19:54.8466431Z }, 2025-10-10T01:19:54.8466563Z { 2025-10-10T01:19:54.8466787Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-10-10T01:19:54.8467071Z "size": 1479, 2025-10-10T01:19:54.8467351Z "digest": "sha256:eb61ec0f4c410703c6a1d07f978c61c268b52bfb0e5f5139320f4dab72299fd3" 2025-10-10T01:19:54.8467674Z }, 2025-10-10T01:19:54.8467805Z { 2025-10-10T01:19:54.8468025Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-10-10T01:19:54.8468324Z "size": 482, 2025-10-10T01:19:54.8468603Z "digest": "sha256:31c940b54575a2a3c75778d62e786df5f6ea171157fe1a3301e9d44ecd63be4a" 2025-10-10T01:19:54.8468918Z }, 2025-10-10T01:19:54.8469043Z { 2025-10-10T01:19:54.8469307Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-10-10T01:19:54.8469901Z "size": 199, 2025-10-10T01:19:54.8470204Z "digest": "sha256:21e8a5f2f65c173feee790b44ba3e36faef391dbe80291609660503d27c876b7" 2025-10-10T01:19:54.8470539Z }, 2025-10-10T01:19:54.8470672Z { 2025-10-10T01:19:54.8470944Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-10-10T01:19:54.8471337Z "size": 608, 2025-10-10T01:19:54.8471636Z "digest": "sha256:496894807e5ced9b0cfe44157347c1d3369954a2c03be9c8fc7902aa87e83c2f" 2025-10-10T01:19:54.8471965Z }, 2025-10-10T01:19:54.8472092Z { 2025-10-10T01:19:54.8472326Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-10-10T01:19:54.8472698Z "size": 7871294039, 2025-10-10T01:19:54.8472998Z "digest": "sha256:1ae67e80b1bd70abd4d0cc28530080b079fe9851915401fc4efd12b0d30804e3" 2025-10-10T01:19:54.8473319Z }, 2025-10-10T01:19:54.8473449Z { 2025-10-10T01:19:54.8473671Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-10-10T01:19:54.8473958Z "size": 829, 2025-10-10T01:19:54.8474228Z "digest": "sha256:cf225908c05b333296c57368e0e6e51ae79656d0396429e861119103f6dbe42e" 2025-10-10T01:19:54.8474539Z }, 2025-10-10T01:19:54.8474669Z { 2025-10-10T01:19:54.8474888Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-10-10T01:19:54.8475167Z "size": 33451742, 2025-10-10T01:19:54.8475468Z "digest": "sha256:381895ce55f63b7ce900d9f5a6488bf96bb1844eedb4daca0d2ec85c7b2a9d50" 2025-10-10T01:19:54.8475794Z }, 2025-10-10T01:19:54.8475925Z { 2025-10-10T01:19:54.8476141Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-10-10T01:19:54.8476445Z "size": 104, 2025-10-10T01:19:54.8476713Z "digest": "sha256:9263c0b02502319d82502d91c88966066495c32552fc2f0e3346a06a479029e9" 2025-10-10T01:19:54.8477025Z }, 2025-10-10T01:19:54.8477147Z { 2025-10-10T01:19:54.8477367Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-10-10T01:19:54.8477653Z "size": 1494, 2025-10-10T01:19:54.8477944Z "digest": "sha256:df861c28c3a5b031b5566700eceda3adf67fed23ee9a24b470b2655bf58ebd6e" 2025-10-10T01:19:54.8478265Z }, 2025-10-10T01:19:54.8478459Z { 2025-10-10T01:19:54.8478705Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-10-10T01:19:54.8478990Z "size": 453967971, 2025-10-10T01:19:54.8479283Z "digest": "sha256:72757cf9a787c2d65d34892517a1393a5006e11d113414fa1f648eb76a030030" 2025-10-10T01:19:54.8479604Z }, 2025-10-10T01:19:54.8479735Z { 2025-10-10T01:19:54.8479954Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-10-10T01:19:54.8480230Z "size": 165, 2025-10-10T01:19:54.8480511Z "digest": "sha256:6ceacb336387b265adaf889cb35b12576e703638c547d104aab59905da34e10c" 2025-10-10T01:19:54.8480830Z }, 2025-10-10T01:19:54.8480959Z { 2025-10-10T01:19:54.8481186Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-10-10T01:19:54.8481470Z "size": 346, 2025-10-10T01:19:54.8481748Z "digest": "sha256:f75cb5b656f8d33cc03600ad8f842daa41eee9ab78e963b982cc73e6f5a9d5a4" 2025-10-10T01:19:54.8482074Z }, 2025-10-10T01:19:54.8482198Z { 2025-10-10T01:19:54.8482411Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-10-10T01:19:54.8482690Z "size": 32, 2025-10-10T01:19:54.8483013Z "digest": "sha256:4f4fb700ef54461cfa02571ae0db9a0dc1e0cdb5577484a6d75e68dc38e8acc1" 2025-10-10T01:19:54.8483386Z }, 2025-10-10T01:19:54.8483543Z { 2025-10-10T01:19:54.8483796Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-10-10T01:19:54.8484129Z "size": 106, 2025-10-10T01:19:54.8484443Z "digest": "sha256:34e0c35a989365ff6b8bba7ef0a2454944f42d0552b30e564cd29058ed17e922" 2025-10-10T01:19:54.8484825Z }, 2025-10-10T01:19:54.8484954Z { 2025-10-10T01:19:54.8485170Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-10-10T01:19:54.8485446Z "size": 425, 2025-10-10T01:19:54.8485723Z "digest": "sha256:a3022990e5c36bc6dd359bbd4a75577a8afb5eab5f56d246c8365355f45c896d" 2025-10-10T01:19:54.8486045Z }, 2025-10-10T01:19:54.8486175Z { 2025-10-10T01:19:54.8486385Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-10-10T01:19:54.8486670Z "size": 19309421, 2025-10-10T01:19:54.8486957Z "digest": "sha256:93ef959930dc30c9d3829d76c89a3e12a1accde4f296210e73a923775466f6f3" 2025-10-10T01:19:54.8487416Z }, 2025-10-10T01:19:54.8487546Z { 2025-10-10T01:19:54.8487771Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-10-10T01:19:54.8488056Z "size": 108, 2025-10-10T01:19:54.8488333Z "digest": "sha256:fa921863e6e3c4729e4c156241dce0e0e91f06c58137c5a95e23366980f8dc48" 2025-10-10T01:19:54.8488714Z }, 2025-10-10T01:19:54.8488852Z { 2025-10-10T01:19:54.8489072Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-10-10T01:19:54.8489352Z "size": 639, 2025-10-10T01:19:54.8489631Z "digest": "sha256:bab2c667cc1aba875b52e122d6ff86fa80cadb41c3060b4896423fdd41be3bf7" 2025-10-10T01:19:54.8489965Z }, 2025-10-10T01:19:54.8490092Z { 2025-10-10T01:19:54.8490311Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-10-10T01:19:54.8490611Z "size": 724, 2025-10-10T01:19:54.8490893Z "digest": "sha256:3e62aac2efe8642154346954eca23d31a01d7d6558ac7f6c2f6ebdc2a2918ec0" 2025-10-10T01:19:54.8491221Z }, 2025-10-10T01:19:54.8491359Z { 2025-10-10T01:19:54.8491575Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-10-10T01:19:54.8491858Z "size": 149, 2025-10-10T01:19:54.8492137Z "digest": "sha256:3bc21d940f02abbc9ff39086f9d66c1e483b760b25c00efd69260f32a0f21bc3" 2025-10-10T01:19:54.8492464Z }, 2025-10-10T01:19:54.8492598Z { 2025-10-10T01:19:54.8492815Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-10-10T01:19:54.8493093Z "size": 136, 2025-10-10T01:19:54.8493381Z "digest": "sha256:5a5fb6edd17b04afbe57dfd2bc9e4f92067ed6b9509b8a28a274c8802f7577a8" 2025-10-10T01:19:54.8493703Z }, 2025-10-10T01:19:54.8493834Z { 2025-10-10T01:19:54.8494048Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-10-10T01:19:54.8494327Z "size": 140, 2025-10-10T01:19:54.8494598Z "digest": "sha256:62cb56db3356b895eb1bfc6a9352feafb0ba44e2e4bbbb933ee31466a89dc91f" 2025-10-10T01:19:54.8494926Z }, 2025-10-10T01:19:54.8495054Z { 2025-10-10T01:19:54.8495271Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-10-10T01:19:54.8495544Z "size": 32, 2025-10-10T01:19:54.8495820Z "digest": "sha256:4f4fb700ef54461cfa02571ae0db9a0dc1e0cdb5577484a6d75e68dc38e8acc1" 2025-10-10T01:19:54.8496144Z }, 2025-10-10T01:19:54.8496270Z { 2025-10-10T01:19:54.8496486Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-10-10T01:19:54.8496769Z "size": 223, 2025-10-10T01:19:54.8497042Z "digest": "sha256:a1fff40729fd605a37158c121092ced87d96efdb3dad96d4b27f9e99d9e3a25b" 2025-10-10T01:19:54.8497369Z }, 2025-10-10T01:19:54.8497494Z { 2025-10-10T01:19:54.8497712Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-10-10T01:19:54.8497995Z "size": 312468897, 2025-10-10T01:19:54.8498284Z "digest": "sha256:50e3f0227909c5eb4b1e149f9835f43ca029022668ca5d00f48f3ed1dc1d72ea" 2025-10-10T01:19:54.8498607Z }, 2025-10-10T01:19:54.8498741Z { 2025-10-10T01:19:54.8498964Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-10-10T01:19:54.8499249Z "size": 3275597028, 2025-10-10T01:19:54.8499536Z "digest": "sha256:ec62e2e53434b9077ac0ca94ade12194f7e0e4d1c9413fc114875ede7630178d" 2025-10-10T01:19:54.8499860Z }, 2025-10-10T01:19:54.8499986Z { 2025-10-10T01:19:54.8500203Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-10-10T01:19:54.8500482Z "size": 128, 2025-10-10T01:19:54.8500760Z "digest": "sha256:a47c72a0c062803832824190bdf9729fff33877e9e74ee333ba59eedbf97480d" 2025-10-10T01:19:54.8501089Z }, 2025-10-10T01:19:54.8501224Z { 2025-10-10T01:19:54.8501438Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-10-10T01:19:54.8501723Z "size": 880, 2025-10-10T01:19:54.8502010Z "digest": "sha256:04fa9fbfdbbdeb9977625cdcd35321e246859432e4cdd6980b7c4b23009aaebe" 2025-10-10T01:19:54.8502346Z }, 2025-10-10T01:19:54.8502471Z { 2025-10-10T01:19:54.8502691Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-10-10T01:19:54.8503057Z "size": 724, 2025-10-10T01:19:54.8503336Z "digest": "sha256:3e62aac2efe8642154346954eca23d31a01d7d6558ac7f6c2f6ebdc2a2918ec0" 2025-10-10T01:19:54.8503675Z }, 2025-10-10T01:19:54.8503802Z { 2025-10-10T01:19:54.8504022Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-10-10T01:19:54.8504377Z "size": 141, 2025-10-10T01:19:54.8504657Z "digest": "sha256:df1cf69f3334c7338a0642cd84e318daf1a79280116974c76b9cc082096f5a8f" 2025-10-10T01:19:54.8504967Z }, 2025-10-10T01:19:54.8505097Z { 2025-10-10T01:19:54.8505313Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-10-10T01:19:54.8505589Z "size": 32, 2025-10-10T01:19:54.8505860Z "digest": "sha256:4f4fb700ef54461cfa02571ae0db9a0dc1e0cdb5577484a6d75e68dc38e8acc1" 2025-10-10T01:19:54.8506184Z }, 2025-10-10T01:19:54.8506310Z { 2025-10-10T01:19:54.8506522Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-10-10T01:19:54.8506801Z "size": 160, 2025-10-10T01:19:54.8507079Z "digest": "sha256:6ad7c264e84319c670a7ee23b32108cf557c91b12e8438059b6b7cc4cb3b1e92" 2025-10-10T01:19:54.8507396Z }, 2025-10-10T01:19:54.8507527Z { 2025-10-10T01:19:54.8507736Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-10-10T01:19:54.8508025Z "size": 1012, 2025-10-10T01:19:54.8508320Z "digest": "sha256:a08c722c1447f6cfb770780a5e9e0123d27e8e994f6c57b352fa71bb6c40d6f5" 2025-10-10T01:19:54.8508642Z }, 2025-10-10T01:19:54.8508768Z { 2025-10-10T01:19:54.8508984Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-10-10T01:19:54.8509263Z "size": 724, 2025-10-10T01:19:54.8509435Z + exit 0 2025-10-10T01:19:54.8509706Z "digest": "sha256:3e62aac2efe8642154346954eca23d31a01d7d6558ac7f6c2f6ebdc2a2918ec0" 2025-10-10T01:19:54.8510020Z }, 2025-10-10T01:19:54.8510149Z { 2025-10-10T01:19:54.8510371Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-10-10T01:19:54.8510651Z "size": 135, 2025-10-10T01:19:54.8510930Z "digest": "sha256:38deb86def9c0f4fe784a0b2986dab1a87bbbafdf2464831502c7fddc595d2fc" 2025-10-10T01:19:54.8511272Z }, 2025-10-10T01:19:54.8511405Z { 2025-10-10T01:19:54.8511623Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-10-10T01:19:54.8511899Z "size": 32, 2025-10-10T01:19:54.8512183Z "digest": "sha256:4f4fb700ef54461cfa02571ae0db9a0dc1e0cdb5577484a6d75e68dc38e8acc1" 2025-10-10T01:19:54.8512509Z }, 2025-10-10T01:19:54.8512637Z { 2025-10-10T01:19:54.8512849Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-10-10T01:19:54.8513131Z "size": 158, 2025-10-10T01:19:54.8513408Z "digest": "sha256:8a6cf39f45d150cd87db5867fe2dce8a814326a583db1f951f115e7a17fc6a27" 2025-10-10T01:19:54.8513741Z }, 2025-10-10T01:19:54.8513863Z { 2025-10-10T01:19:54.8514079Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-10-10T01:19:54.8514361Z "size": 1369, 2025-10-10T01:19:54.8514637Z "digest": "sha256:a37d784168133fb9c2f7514a9fe459f85805b1717cc33187d3eb30a9f00eeccf" 2025-10-10T01:19:54.8514946Z }, 2025-10-10T01:19:54.8515075Z { 2025-10-10T01:19:54.8515293Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-10-10T01:19:54.8515572Z "size": 32, 2025-10-10T01:19:54.8515841Z "digest": "sha256:4f4fb700ef54461cfa02571ae0db9a0dc1e0cdb5577484a6d75e68dc38e8acc1" 2025-10-10T01:19:54.8516169Z }, 2025-10-10T01:19:54.8516294Z { 2025-10-10T01:19:54.8516511Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-10-10T01:19:54.8516784Z "size": 136, 2025-10-10T01:19:54.8517048Z "digest": "sha256:001023e5b31126d09bd1d68c21a7fae99118967088d1d825c0db18b0968cf4a6" 2025-10-10T01:19:54.8517357Z }, 2025-10-10T01:19:54.8517483Z { 2025-10-10T01:19:54.8517692Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-10-10T01:19:54.8517975Z "size": 528, 2025-10-10T01:19:54.8518248Z "digest": "sha256:5b84bc540263a168bff65155a8a267e3cc7af2274fcfe84ba2d50776cb1f69a7" 2025-10-10T01:19:54.8518729Z }, 2025-10-10T01:19:54.8518858Z { 2025-10-10T01:19:54.8519076Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-10-10T01:19:54.8519357Z "size": 32, 2025-10-10T01:19:54.8519632Z "digest": "sha256:4f4fb700ef54461cfa02571ae0db9a0dc1e0cdb5577484a6d75e68dc38e8acc1" 2025-10-10T01:19:54.8520036Z }, 2025-10-10T01:19:54.8520164Z { 2025-10-10T01:19:54.8520380Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-10-10T01:19:54.8520658Z "size": 104, 2025-10-10T01:19:54.8520927Z "digest": "sha256:7b03cfd9cf3c75c07836bf90040e0afd487cbd07cd3752e9a39f946e1726bf20" 2025-10-10T01:19:54.8521249Z }, 2025-10-10T01:19:54.8521376Z { 2025-10-10T01:19:54.8521594Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-10-10T01:19:54.8521866Z "size": 430, 2025-10-10T01:19:54.8522147Z "digest": "sha256:3ed8e92a019016b1a29fbc84a0d1fba5fc2495e81c5d84deec3c5a0be97fb1b0" 2025-10-10T01:19:54.8522483Z }, 2025-10-10T01:19:54.8522617Z { 2025-10-10T01:19:54.8522828Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-10-10T01:19:54.8523111Z "size": 32, 2025-10-10T01:19:54.8523399Z "digest": "sha256:4f4fb700ef54461cfa02571ae0db9a0dc1e0cdb5577484a6d75e68dc38e8acc1" 2025-10-10T01:19:54.8523732Z }, 2025-10-10T01:19:54.8523855Z { 2025-10-10T01:19:54.8524368Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-10-10T01:19:54.8524668Z "size": 109, 2025-10-10T01:19:54.8524946Z "digest": "sha256:477967cba78d4f983881e91fdfb858a0f628e7533cf303e7cbe80f4c99deb43f" 2025-10-10T01:19:54.8525260Z }, 2025-10-10T01:19:54.8525399Z { 2025-10-10T01:19:54.8525621Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-10-10T01:19:54.8525901Z "size": 1897, 2025-10-10T01:19:54.8526181Z "digest": "sha256:6fffed0cb29593ef2bfbb6035811935c5b2a254e4467e360e058f40d5944d261" 2025-10-10T01:19:54.8526506Z }, 2025-10-10T01:19:54.8526634Z { 2025-10-10T01:19:54.8526859Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-10-10T01:19:54.8527137Z "size": 244514917, 2025-10-10T01:19:54.8527435Z "digest": "sha256:8aa311eb036b9b7ab88d5f00770ec0154cd7705a370477c53454e48ad49949c7" 2025-10-10T01:19:54.8527752Z }, 2025-10-10T01:19:54.8527880Z { 2025-10-10T01:19:54.8528097Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-10-10T01:19:54.8528375Z "size": 106, 2025-10-10T01:19:54.8528643Z "digest": "sha256:33d9650746499f91b89de8ab6ba0f6808dfa248eb37709a1686c3b57a07d0cc7" 2025-10-10T01:19:54.8528958Z }, 2025-10-10T01:19:54.8529079Z { 2025-10-10T01:19:54.8529294Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-10-10T01:19:54.8529575Z "size": 164, 2025-10-10T01:19:54.8529849Z "digest": "sha256:dbf89dacc6c6b267a7cd9200c81200a16386a20727ebfe38d3b05cfb223022e4" 2025-10-10T01:19:54.8530168Z }, 2025-10-10T01:19:54.8530310Z { 2025-10-10T01:19:54.8530529Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-10-10T01:19:54.8530813Z "size": 7943, 2025-10-10T01:19:54.8531085Z "digest": "sha256:208b50d13dad68e60745da16f1f48474ec502e57b6d3a49c863d432fbfc8eb5b" 2025-10-10T01:19:54.8531403Z }, 2025-10-10T01:19:54.8531530Z { 2025-10-10T01:19:54.8531755Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-10-10T01:19:54.8532040Z "size": 8070, 2025-10-10T01:19:54.8532319Z "digest": "sha256:863fa5fce8b803292f021ae5db81240ea781fdacdb093cd08202d4059b13535e" 2025-10-10T01:19:54.8532642Z }, 2025-10-10T01:19:54.8532772Z { 2025-10-10T01:19:54.8533019Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-10-10T01:19:54.8533354Z "size": 303, 2025-10-10T01:19:54.8533676Z "digest": "sha256:2c57470f9b051ed7b272e9cb78b567b3a9672312e2620dbe5432a70d95bb206b" 2025-10-10T01:19:54.8534050Z }, 2025-10-10T01:19:54.8534194Z { 2025-10-10T01:19:54.8534452Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-10-10T01:19:54.8534957Z "size": 13360730, 2025-10-10T01:19:54.8535320Z "digest": "sha256:4fbac50fbb3bb3f8695474fefc6fe9230fc5c628ab961fc42c5b8ad275f1f2d5" 2025-10-10T01:19:54.8535691Z }, 2025-10-10T01:19:54.8535823Z { 2025-10-10T01:19:54.8536038Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-10-10T01:19:54.8536459Z "size": 108, 2025-10-10T01:19:54.8536736Z "digest": "sha256:6baf3cf73050abcdf92034c01cb77c2d6467688d409f22773e8c21e93ed76a42" 2025-10-10T01:19:54.8537062Z }, 2025-10-10T01:19:54.8537192Z { 2025-10-10T01:19:54.8537410Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-10-10T01:19:54.8537693Z "size": 54145699, 2025-10-10T01:19:54.8537982Z "digest": "sha256:ace2a84685d29b76147a66dfa220c800615a65949511de6b464aff673eeaf062" 2025-10-10T01:19:54.8538301Z }, 2025-10-10T01:19:54.8538428Z { 2025-10-10T01:19:54.8538641Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-10-10T01:19:54.8538926Z "size": 32, 2025-10-10T01:19:54.8539212Z "digest": "sha256:4f4fb700ef54461cfa02571ae0db9a0dc1e0cdb5577484a6d75e68dc38e8acc1" 2025-10-10T01:19:54.8539548Z } 2025-10-10T01:19:54.8539673Z ] 2025-10-10T01:19:54.8539803Z } 2025-10-10T01:19:54.8577004Z ##[group]Run set -eux 2025-10-10T01:19:54.8577212Z set -eux 2025-10-10T01:19:54.8577532Z # It's ok if this steps fails, it would then be an anonymous user like what we used to have 2025-10-10T01:19:54.8578362Z aws secretsmanager get-secret-value --secret-id docker_hub_readonly_token | jq --raw-output '.SecretString' | jq -r .docker_hub_readonly_token | docker login --username pytorchbot --password-stdin || true 2025-10-10T01:19:54.8587007Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2025-10-10T01:19:54.8587303Z env: 2025-10-10T01:19:54.8587463Z GIT_DEFAULT_BRANCH: main 2025-10-10T01:19:54.8587656Z ##[endgroup] 2025-10-10T01:19:54.8625555Z + aws secretsmanager get-secret-value --secret-id docker_hub_readonly_token 2025-10-10T01:19:54.8626501Z + jq --raw-output .SecretString 2025-10-10T01:19:54.8628338Z + docker login --username pytorchbot --password-stdin 2025-10-10T01:19:54.8629527Z + jq -r .docker_hub_readonly_token 2025-10-10T01:19:55.4218364Z WARNING! Your password will be stored unencrypted in /home/ec2-user/.docker/config.json. 2025-10-10T01:19:55.4219537Z Configure a credential helper to remove this warning. See 2025-10-10T01:19:55.4220576Z https://docs.docker.com/engine/reference/commandline/login/#credentials-store 2025-10-10T01:19:55.4221287Z 2025-10-10T01:19:55.4221485Z Login Succeeded 2025-10-10T01:19:55.4320978Z ##[group]Run tag=${ECR_DOCKER_IMAGE##*:} 2025-10-10T01:19:55.4321325Z tag=${ECR_DOCKER_IMAGE##*:} 2025-10-10T01:19:55.4333887Z echo "docker pull ghcr.io/pytorch/ci-image:${tag/:/-}" 2025-10-10T01:19:55.4342081Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2025-10-10T01:19:55.4342369Z env: 2025-10-10T01:19:55.4342533Z GIT_DEFAULT_BRANCH: main 2025-10-10T01:19:55.4343158Z ECR_DOCKER_IMAGE: 308535385114.dkr.ecr.us-east-1.amazonaws.com/pytorch/ci-image:pytorch-linux-jammy-cuda12.8-cudnn9-py3.12-gcc11-vllm-d8be0384e085f551506bd739678109fa0f5ee7ac 2025-10-10T01:19:55.4343803Z ##[endgroup] 2025-10-10T01:19:55.4377235Z docker pull ghcr.io/pytorch/ci-image:pytorch-linux-jammy-cuda12.8-cudnn9-py3.12-gcc11-vllm-d8be0384e085f551506bd739678109fa0f5ee7ac 2025-10-10T01:19:55.4434079Z ##[group]Run pytorch/test-infra/.github/actions/pull-docker-image@main 2025-10-10T01:19:55.4434436Z with: 2025-10-10T01:19:55.4435021Z docker-image: 308535385114.dkr.ecr.us-east-1.amazonaws.com/pytorch/ci-image:pytorch-linux-jammy-cuda12.8-cudnn9-py3.12-gcc11-vllm-d8be0384e085f551506bd739678109fa0f5ee7ac 2025-10-10T01:19:55.4435733Z docker-registry: 308535385114.dkr.ecr.us-east-1.amazonaws.com 2025-10-10T01:19:55.4436003Z env: 2025-10-10T01:19:55.4436156Z GIT_DEFAULT_BRANCH: main 2025-10-10T01:19:55.4436342Z ##[endgroup] 2025-10-10T01:19:55.4458170Z ##[group]Run set -x 2025-10-10T01:19:55.4458364Z set -x 2025-10-10T01:19:55.4458525Z set +e 2025-10-10T01:19:55.4458686Z  2025-10-10T01:19:55.4458829Z login() { 2025-10-10T01:19:55.4459179Z  aws ecr get-login-password --region us-east-1 | docker login -u AWS --password-stdin "$1" 2025-10-10T01:19:55.4459744Z } 2025-10-10T01:19:55.4459897Z  2025-10-10T01:19:55.4460067Z retry () { 2025-10-10T01:19:55.4460273Z  $* || (sleep 1 && $*) || (sleep 2 && $*) 2025-10-10T01:19:55.4460490Z } 2025-10-10T01:19:55.4460639Z  2025-10-10T01:19:55.4460803Z retry login "${DOCKER_REGISTRY}" 2025-10-10T01:19:55.4461019Z  2025-10-10T01:19:55.4461358Z IMAGE_SIZE=$(docker manifest inspect "${DOCKER_IMAGE}" | jq '[.layers[].size, .config.size] | add / 1024 / 1024') 2025-10-10T01:19:55.4461821Z echo "Compressed size of image in MB: ${IMAGE_SIZE}" 2025-10-10T01:19:55.4462083Z  2025-10-10T01:19:55.4462235Z set -e 2025-10-10T01:19:55.4462483Z # ignore output since only exit code is used for conditional 2025-10-10T01:19:55.4462835Z # only pull docker image if it's not available locally 2025-10-10T01:19:55.4463216Z if ! docker inspect --type=image "${DOCKER_IMAGE}" >/dev/null 2>/dev/null; then 2025-10-10T01:19:55.4463584Z  retry docker pull "${DOCKER_IMAGE}" 2025-10-10T01:19:55.4463809Z fi 2025-10-10T01:19:55.4470805Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2025-10-10T01:19:55.4471097Z env: 2025-10-10T01:19:55.4471258Z GIT_DEFAULT_BRANCH: main 2025-10-10T01:19:55.4471891Z DOCKER_IMAGE: 308535385114.dkr.ecr.us-east-1.amazonaws.com/pytorch/ci-image:pytorch-linux-jammy-cuda12.8-cudnn9-py3.12-gcc11-vllm-d8be0384e085f551506bd739678109fa0f5ee7ac 2025-10-10T01:19:55.4472584Z DOCKER_REGISTRY: 308535385114.dkr.ecr.us-east-1.amazonaws.com 2025-10-10T01:19:55.4472864Z ##[endgroup] 2025-10-10T01:19:55.4504457Z + set +e 2025-10-10T01:19:55.4505051Z + retry login 308535385114.dkr.ecr.us-east-1.amazonaws.com 2025-10-10T01:19:55.4505782Z + login 308535385114.dkr.ecr.us-east-1.amazonaws.com 2025-10-10T01:19:55.4509760Z + docker login -u AWS --password-stdin 308535385114.dkr.ecr.us-east-1.amazonaws.com 2025-10-10T01:19:55.4510695Z + aws ecr get-login-password --region us-east-1 2025-10-10T01:19:55.9292031Z WARNING! Your password will be stored unencrypted in /home/ec2-user/.docker/config.json. 2025-10-10T01:19:55.9292889Z Configure a credential helper to remove this warning. See 2025-10-10T01:19:55.9293659Z https://docs.docker.com/engine/reference/commandline/login/#credentials-store 2025-10-10T01:19:55.9294077Z 2025-10-10T01:19:55.9294502Z Login Succeeded 2025-10-10T01:19:55.9328873Z ++ docker manifest inspect 308535385114.dkr.ecr.us-east-1.amazonaws.com/pytorch/ci-image:pytorch-linux-jammy-cuda12.8-cudnn9-py3.12-gcc11-vllm-d8be0384e085f551506bd739678109fa0f5ee7ac 2025-10-10T01:19:55.9329805Z ++ jq '[.layers[].size, .config.size] | add / 1024 / 1024' 2025-10-10T01:19:56.1756724Z + IMAGE_SIZE=15400.734984397888 2025-10-10T01:19:56.1757300Z + echo 'Compressed size of image in MB: 15400.734984397888' 2025-10-10T01:19:56.1757826Z + set -e 2025-10-10T01:19:56.1758198Z Compressed size of image in MB: 15400.734984397888 2025-10-10T01:19:56.1760101Z + docker inspect --type=image 308535385114.dkr.ecr.us-east-1.amazonaws.com/pytorch/ci-image:pytorch-linux-jammy-cuda12.8-cudnn9-py3.12-gcc11-vllm-d8be0384e085f551506bd739678109fa0f5ee7ac 2025-10-10T01:19:56.1923541Z + retry docker pull 308535385114.dkr.ecr.us-east-1.amazonaws.com/pytorch/ci-image:pytorch-linux-jammy-cuda12.8-cudnn9-py3.12-gcc11-vllm-d8be0384e085f551506bd739678109fa0f5ee7ac 2025-10-10T01:19:56.1925472Z + docker pull 308535385114.dkr.ecr.us-east-1.amazonaws.com/pytorch/ci-image:pytorch-linux-jammy-cuda12.8-cudnn9-py3.12-gcc11-vllm-d8be0384e085f551506bd739678109fa0f5ee7ac 2025-10-10T01:19:56.4756749Z pytorch-linux-jammy-cuda12.8-cudnn9-py3.12-gcc11-vllm-d8be0384e085f551506bd739678109fa0f5ee7ac: Pulling from pytorch/ci-image 2025-10-10T01:19:56.4759256Z 828c1365039a: Pulling fs layer 2025-10-10T01:19:56.4759585Z 140bb29742fd: Pulling fs layer 2025-10-10T01:19:56.4759878Z 99ee2db1caf9: Pulling fs layer 2025-10-10T01:19:56.4760122Z f5a2330f87a8: Pulling fs layer 2025-10-10T01:19:56.4760366Z 4643d3cb7e67: Pulling fs layer 2025-10-10T01:19:56.4760942Z 5c871f60f6d4: Pulling fs layer 2025-10-10T01:19:56.4761186Z 60a2970be007: Pulling fs layer 2025-10-10T01:19:56.4761416Z 43d427ee0b00: Pulling fs layer 2025-10-10T01:19:56.4761664Z 728b6e1eb012: Pulling fs layer 2025-10-10T01:19:56.4761920Z 1ffc73eae77b: Pulling fs layer 2025-10-10T01:19:56.4762161Z 1c5379881a3b: Pulling fs layer 2025-10-10T01:19:56.4762383Z 3e62aac2efe8: Pulling fs layer 2025-10-10T01:19:56.4762620Z 3c1f345eea91: Pulling fs layer 2025-10-10T01:19:56.4762852Z a03ec5361452: Pulling fs layer 2025-10-10T01:19:56.4763085Z 4f4fb700ef54: Pulling fs layer 2025-10-10T01:19:56.4763310Z 84681debfb5c: Pulling fs layer 2025-10-10T01:19:56.4763550Z d6a77cb34f2d: Pulling fs layer 2025-10-10T01:19:56.4763786Z b4e994693c03: Pulling fs layer 2025-10-10T01:19:56.4764020Z c6c1bfa3b4c5: Pulling fs layer 2025-10-10T01:19:56.4764265Z eb61ec0f4c41: Pulling fs layer 2025-10-10T01:19:56.4764528Z 31c940b54575: Pulling fs layer 2025-10-10T01:19:56.4764814Z 21e8a5f2f65c: Pulling fs layer 2025-10-10T01:19:56.4765104Z 496894807e5c: Pulling fs layer 2025-10-10T01:19:56.4765373Z 1ae67e80b1bd: Pulling fs layer 2025-10-10T01:19:56.4765649Z cf225908c05b: Pulling fs layer 2025-10-10T01:19:56.4765920Z 381895ce55f6: Pulling fs layer 2025-10-10T01:19:56.4766180Z 9263c0b02502: Pulling fs layer 2025-10-10T01:19:56.4766388Z df861c28c3a5: Pulling fs layer 2025-10-10T01:19:56.4766627Z 72757cf9a787: Pulling fs layer 2025-10-10T01:19:56.4766845Z 6ceacb336387: Pulling fs layer 2025-10-10T01:19:56.4767067Z f75cb5b656f8: Pulling fs layer 2025-10-10T01:19:56.4767296Z 34e0c35a9893: Pulling fs layer 2025-10-10T01:19:56.4767506Z a3022990e5c3: Pulling fs layer 2025-10-10T01:19:56.4767721Z 93ef959930dc: Pulling fs layer 2025-10-10T01:19:56.4767908Z fa921863e6e3: Pulling fs layer 2025-10-10T01:19:56.4768097Z bab2c667cc1a: Pulling fs layer 2025-10-10T01:19:56.4768288Z 5c871f60f6d4: Waiting 2025-10-10T01:19:56.4768477Z 3bc21d940f02: Pulling fs layer 2025-10-10T01:19:56.4768685Z 5a5fb6edd17b: Pulling fs layer 2025-10-10T01:19:56.4768884Z 62cb56db3356: Pulling fs layer 2025-10-10T01:19:56.4769069Z a1fff40729fd: Pulling fs layer 2025-10-10T01:19:56.4769257Z 50e3f0227909: Pulling fs layer 2025-10-10T01:19:56.4769444Z ec62e2e53434: Pulling fs layer 2025-10-10T01:19:56.4769638Z a47c72a0c062: Pulling fs layer 2025-10-10T01:19:56.4769822Z 04fa9fbfdbbd: Pulling fs layer 2025-10-10T01:19:56.4770011Z df1cf69f3334: Pulling fs layer 2025-10-10T01:19:56.4770238Z 1c5379881a3b: Waiting 2025-10-10T01:19:56.4770412Z 6ad7c264e843: Pulling fs layer 2025-10-10T01:19:56.4770592Z c6c1bfa3b4c5: Waiting 2025-10-10T01:19:56.4770762Z a08c722c1447: Pulling fs layer 2025-10-10T01:19:56.4770948Z 3e62aac2efe8: Waiting 2025-10-10T01:19:56.4771117Z 38deb86def9c: Pulling fs layer 2025-10-10T01:19:56.4771291Z f5a2330f87a8: Waiting 2025-10-10T01:19:56.4771451Z 4643d3cb7e67: Waiting 2025-10-10T01:19:56.4771609Z 31c940b54575: Waiting 2025-10-10T01:19:56.4771766Z 728b6e1eb012: Waiting 2025-10-10T01:19:56.4771925Z 8a6cf39f45d1: Pulling fs layer 2025-10-10T01:19:56.4772114Z 60a2970be007: Waiting 2025-10-10T01:19:56.4772434Z d6a77cb34f2d: Waiting 2025-10-10T01:19:56.4772614Z 43d427ee0b00: Waiting 2025-10-10T01:19:56.4772776Z a37d78416813: Pulling fs layer 2025-10-10T01:19:56.4772959Z 9263c0b02502: Waiting 2025-10-10T01:19:56.4773130Z 001023e5b311: Pulling fs layer 2025-10-10T01:19:56.4773315Z df861c28c3a5: Waiting 2025-10-10T01:19:56.4773469Z 1ffc73eae77b: Waiting 2025-10-10T01:19:56.4773638Z 5b84bc540263: Pulling fs layer 2025-10-10T01:19:56.4773831Z 1ae67e80b1bd: Waiting 2025-10-10T01:19:56.4773996Z 496894807e5c: Waiting 2025-10-10T01:19:56.4774175Z 7b03cfd9cf3c: Pulling fs layer 2025-10-10T01:19:56.4774373Z 3ed8e92a0190: Pulling fs layer 2025-10-10T01:19:56.4774564Z 477967cba78d: Pulling fs layer 2025-10-10T01:19:56.4774748Z 93ef959930dc: Waiting 2025-10-10T01:19:56.4774900Z 50e3f0227909: Waiting 2025-10-10T01:19:56.4775057Z 4f4fb700ef54: Waiting 2025-10-10T01:19:56.4775213Z 6ad7c264e843: Waiting 2025-10-10T01:19:56.4775374Z 04fa9fbfdbbd: Waiting 2025-10-10T01:19:56.4775615Z a03ec5361452: Waiting 2025-10-10T01:19:56.4775793Z 6fffed0cb295: Pulling fs layer 2025-10-10T01:19:56.4775991Z 8aa311eb036b: Pulling fs layer 2025-10-10T01:19:56.4776169Z 6ceacb336387: Waiting 2025-10-10T01:19:56.4776329Z 8a6cf39f45d1: Waiting 2025-10-10T01:19:56.4776485Z a47c72a0c062: Waiting 2025-10-10T01:19:56.4776643Z ec62e2e53434: Waiting 2025-10-10T01:19:56.4776807Z bab2c667cc1a: Waiting 2025-10-10T01:19:56.4776973Z 33d965074649: Pulling fs layer 2025-10-10T01:19:56.4777179Z dbf89dacc6c6: Pulling fs layer 2025-10-10T01:19:56.4777369Z 208b50d13dad: Pulling fs layer 2025-10-10T01:19:56.4777568Z 863fa5fce8b8: Pulling fs layer 2025-10-10T01:19:56.4777747Z 62cb56db3356: Waiting 2025-10-10T01:19:56.4777908Z 5a5fb6edd17b: Waiting 2025-10-10T01:19:56.4778076Z 2c57470f9b05: Pulling fs layer 2025-10-10T01:19:56.4778256Z 7b03cfd9cf3c: Waiting 2025-10-10T01:19:56.4778419Z 4fbac50fbb3b: Pulling fs layer 2025-10-10T01:19:56.4778609Z 6baf3cf73050: Pulling fs layer 2025-10-10T01:19:56.4778796Z ace2a84685d2: Pulling fs layer 2025-10-10T01:19:56.4778989Z a3022990e5c3: Waiting 2025-10-10T01:19:56.4779146Z 6fffed0cb295: Waiting 2025-10-10T01:19:56.4779306Z 477967cba78d: Waiting 2025-10-10T01:19:56.4779475Z 8aa311eb036b: Waiting 2025-10-10T01:19:56.4779643Z 4fbac50fbb3b: Waiting 2025-10-10T01:19:56.4779804Z 3ed8e92a0190: Waiting 2025-10-10T01:19:56.4779972Z 33d965074649: Waiting 2025-10-10T01:19:56.4780136Z 208b50d13dad: Waiting 2025-10-10T01:19:56.4780295Z 001023e5b311: Waiting 2025-10-10T01:19:56.4780447Z 3c1f345eea91: Waiting 2025-10-10T01:19:56.4780612Z 863fa5fce8b8: Waiting 2025-10-10T01:19:56.4780777Z ace2a84685d2: Waiting 2025-10-10T01:19:56.4780931Z dbf89dacc6c6: Waiting 2025-10-10T01:19:56.4781095Z a08c722c1447: Waiting 2025-10-10T01:19:56.4781253Z 6baf3cf73050: Waiting 2025-10-10T01:19:56.4781411Z 2c57470f9b05: Waiting 2025-10-10T01:19:56.4781562Z 3bc21d940f02: Waiting 2025-10-10T01:19:56.4781726Z 38deb86def9c: Waiting 2025-10-10T01:19:56.4781896Z a37d78416813: Waiting 2025-10-10T01:19:56.4782061Z f75cb5b656f8: Waiting 2025-10-10T01:19:56.4782221Z df1cf69f3334: Waiting 2025-10-10T01:19:56.4782384Z 5b84bc540263: Waiting 2025-10-10T01:19:56.4782542Z 72757cf9a787: Waiting 2025-10-10T01:19:56.4782701Z fa921863e6e3: Waiting 2025-10-10T01:19:56.4782853Z a1fff40729fd: Waiting 2025-10-10T01:19:56.4783020Z 21e8a5f2f65c: Waiting 2025-10-10T01:19:56.4783174Z b4e994693c03: Waiting 2025-10-10T01:19:56.4783329Z eb61ec0f4c41: Waiting 2025-10-10T01:19:56.4783479Z cf225908c05b: Waiting 2025-10-10T01:19:56.5741433Z 140bb29742fd: Verifying Checksum 2025-10-10T01:19:56.5741750Z 140bb29742fd: Download complete 2025-10-10T01:19:56.6713257Z f5a2330f87a8: Verifying Checksum 2025-10-10T01:19:56.6713560Z f5a2330f87a8: Download complete 2025-10-10T01:19:56.7659039Z 4643d3cb7e67: Verifying Checksum 2025-10-10T01:19:56.7659343Z 4643d3cb7e67: Download complete 2025-10-10T01:19:56.8353151Z 828c1365039a: Verifying Checksum 2025-10-10T01:19:56.8353445Z 828c1365039a: Download complete 2025-10-10T01:19:56.8487996Z 5c871f60f6d4: Verifying Checksum 2025-10-10T01:19:56.8488578Z 5c871f60f6d4: Download complete 2025-10-10T01:19:56.9364154Z 43d427ee0b00: Download complete 2025-10-10T01:19:56.9373368Z 60a2970be007: Verifying Checksum 2025-10-10T01:19:56.9373668Z 60a2970be007: Download complete 2025-10-10T01:19:57.0023079Z 1ffc73eae77b: Verifying Checksum 2025-10-10T01:19:57.0023332Z 1ffc73eae77b: Download complete 2025-10-10T01:19:57.1009890Z 1c5379881a3b: Download complete 2025-10-10T01:19:57.2233962Z 3e62aac2efe8: Verifying Checksum 2025-10-10T01:19:57.2234414Z 3e62aac2efe8: Download complete 2025-10-10T01:19:57.3354808Z 3c1f345eea91: Download complete 2025-10-10T01:19:57.7723012Z 828c1365039a: Pull complete 2025-10-10T01:19:57.7968358Z 140bb29742fd: Pull complete 2025-10-10T01:19:58.1117309Z 728b6e1eb012: Verifying Checksum 2025-10-10T01:19:58.1117792Z 728b6e1eb012: Download complete 2025-10-10T01:19:58.1225509Z 4f4fb700ef54: Verifying Checksum 2025-10-10T01:19:58.1225934Z 4f4fb700ef54: Download complete 2025-10-10T01:19:58.2178102Z 84681debfb5c: Download complete 2025-10-10T01:19:58.3330143Z d6a77cb34f2d: Download complete 2025-10-10T01:19:58.3981931Z b4e994693c03: Verifying Checksum 2025-10-10T01:19:58.3982357Z b4e994693c03: Download complete 2025-10-10T01:19:58.5173912Z c6c1bfa3b4c5: Verifying Checksum 2025-10-10T01:19:58.5174353Z c6c1bfa3b4c5: Download complete 2025-10-10T01:19:58.6024475Z eb61ec0f4c41: Verifying Checksum 2025-10-10T01:19:58.6024930Z eb61ec0f4c41: Download complete 2025-10-10T01:19:58.6971461Z 31c940b54575: Download complete 2025-10-10T01:19:58.8106077Z 21e8a5f2f65c: Verifying Checksum 2025-10-10T01:19:58.8106524Z 21e8a5f2f65c: Download complete 2025-10-10T01:19:58.8866513Z 496894807e5c: Verifying Checksum 2025-10-10T01:19:58.8866929Z 496894807e5c: Download complete 2025-10-10T01:19:59.6825110Z 99ee2db1caf9: Download complete 2025-10-10T01:19:59.7940914Z cf225908c05b: Verifying Checksum 2025-10-10T01:19:59.7941223Z cf225908c05b: Download complete 2025-10-10T01:20:00.1991680Z 381895ce55f6: Verifying Checksum 2025-10-10T01:20:00.1992030Z 381895ce55f6: Download complete 2025-10-10T01:20:00.2732795Z 9263c0b02502: Verifying Checksum 2025-10-10T01:20:00.2733229Z 9263c0b02502: Download complete 2025-10-10T01:20:00.3766773Z df861c28c3a5: Download complete 2025-10-10T01:20:06.8773404Z 72757cf9a787: Verifying Checksum 2025-10-10T01:20:06.8773733Z 72757cf9a787: Download complete 2025-10-10T01:20:06.9791922Z 6ceacb336387: Verifying Checksum 2025-10-10T01:20:06.9792806Z 6ceacb336387: Download complete 2025-10-10T01:20:07.0619585Z f75cb5b656f8: Download complete 2025-10-10T01:20:07.1714173Z 34e0c35a9893: Verifying Checksum 2025-10-10T01:20:07.1715020Z 34e0c35a9893: Download complete 2025-10-10T01:20:07.2513469Z a3022990e5c3: Verifying Checksum 2025-10-10T01:20:07.2514108Z a3022990e5c3: Download complete 2025-10-10T01:20:07.5080068Z 93ef959930dc: Verifying Checksum 2025-10-10T01:20:07.5080943Z 93ef959930dc: Download complete 2025-10-10T01:20:07.5919779Z fa921863e6e3: Verifying Checksum 2025-10-10T01:20:07.5920166Z fa921863e6e3: Download complete 2025-10-10T01:20:07.6882886Z bab2c667cc1a: Verifying Checksum 2025-10-10T01:20:07.6883569Z bab2c667cc1a: Download complete 2025-10-10T01:20:07.7888917Z 3bc21d940f02: Download complete 2025-10-10T01:20:07.8615536Z 5a5fb6edd17b: Download complete 2025-10-10T01:20:07.9282722Z 99ee2db1caf9: Pull complete 2025-10-10T01:20:07.9367425Z 62cb56db3356: Download complete 2025-10-10T01:20:08.0119007Z a1fff40729fd: Download complete 2025-10-10T01:20:08.3447429Z f5a2330f87a8: Pull complete 2025-10-10T01:20:08.6991805Z 4643d3cb7e67: Pull complete 2025-10-10T01:20:09.1942689Z 5c871f60f6d4: Pull complete 2025-10-10T01:20:09.4544676Z 60a2970be007: Pull complete 2025-10-10T01:20:09.5229262Z 43d427ee0b00: Pull complete 2025-10-10T01:20:11.6195621Z 728b6e1eb012: Pull complete 2025-10-10T01:20:11.8610597Z 1ffc73eae77b: Pull complete 2025-10-10T01:20:12.1394286Z 1c5379881a3b: Pull complete 2025-10-10T01:20:12.4196984Z 3e62aac2efe8: Pull complete 2025-10-10T01:20:12.8843023Z 3c1f345eea91: Pull complete 2025-10-10T01:20:13.3369122Z 50e3f0227909: Verifying Checksum 2025-10-10T01:20:13.3370017Z 50e3f0227909: Download complete 2025-10-10T01:20:50.7004330Z a03ec5361452: Verifying Checksum 2025-10-10T01:20:50.7004697Z a03ec5361452: Download complete 2025-10-10T01:20:50.7893325Z a47c72a0c062: Verifying Checksum 2025-10-10T01:20:50.7893685Z a47c72a0c062: Download complete 2025-10-10T01:20:50.8740483Z 04fa9fbfdbbd: Verifying Checksum 2025-10-10T01:20:50.8740978Z 04fa9fbfdbbd: Download complete 2025-10-10T01:20:50.9557082Z df1cf69f3334: Verifying Checksum 2025-10-10T01:20:50.9557693Z df1cf69f3334: Download complete 2025-10-10T01:20:51.0444681Z 6ad7c264e843: Download complete 2025-10-10T01:20:51.1475001Z a08c722c1447: Verifying Checksum 2025-10-10T01:20:51.1475314Z a08c722c1447: Download complete 2025-10-10T01:20:51.2324349Z 38deb86def9c: Verifying Checksum 2025-10-10T01:20:51.2324808Z 38deb86def9c: Download complete 2025-10-10T01:20:51.3256324Z 8a6cf39f45d1: Download complete 2025-10-10T01:20:51.4417848Z a37d78416813: Verifying Checksum 2025-10-10T01:20:51.4419189Z a37d78416813: Download complete 2025-10-10T01:20:51.5278772Z 001023e5b311: Verifying Checksum 2025-10-10T01:20:51.5279067Z 001023e5b311: Download complete 2025-10-10T01:20:51.6235264Z 5b84bc540263: Verifying Checksum 2025-10-10T01:20:51.6235569Z 5b84bc540263: Download complete 2025-10-10T01:20:51.7048752Z 7b03cfd9cf3c: Verifying Checksum 2025-10-10T01:20:51.7049105Z 7b03cfd9cf3c: Download complete 2025-10-10T01:20:51.7886085Z 3ed8e92a0190: Verifying Checksum 2025-10-10T01:20:51.7886371Z 3ed8e92a0190: Download complete 2025-10-10T01:20:51.9249221Z 477967cba78d: Verifying Checksum 2025-10-10T01:20:51.9249544Z 477967cba78d: Download complete 2025-10-10T01:20:52.0003473Z 6fffed0cb295: Verifying Checksum 2025-10-10T01:20:56.5737773Z 6fffed0cb295: Download complete 2025-10-10T01:20:56.5738130Z 8aa311eb036b: Verifying Checksum 2025-10-10T01:20:56.5738403Z 8aa311eb036b: Download complete 2025-10-10T01:20:56.6886243Z 33d965074649: Verifying Checksum 2025-10-10T01:20:56.6886543Z 33d965074649: Download complete 2025-10-10T01:20:56.7732204Z dbf89dacc6c6: Verifying Checksum 2025-10-10T01:20:56.7732563Z dbf89dacc6c6: Download complete 2025-10-10T01:20:56.8742869Z 208b50d13dad: Verifying Checksum 2025-10-10T01:20:56.8743158Z 208b50d13dad: Download complete 2025-10-10T01:20:56.9476818Z 863fa5fce8b8: Verifying Checksum 2025-10-10T01:20:56.9477133Z 863fa5fce8b8: Download complete 2025-10-10T01:20:57.0606965Z 2c57470f9b05: Verifying Checksum 2025-10-10T01:20:57.0607264Z 2c57470f9b05: Download complete 2025-10-10T01:20:57.2345038Z 4fbac50fbb3b: Verifying Checksum 2025-10-10T01:20:57.2345368Z 4fbac50fbb3b: Download complete 2025-10-10T01:20:57.3183499Z 6baf3cf73050: Verifying Checksum 2025-10-10T01:20:57.3183824Z 6baf3cf73050: Download complete 2025-10-10T01:20:58.2401751Z ace2a84685d2: Verifying Checksum 2025-10-10T01:20:58.2402091Z ace2a84685d2: Download complete 2025-10-10T01:21:05.2605546Z ec62e2e53434: Verifying Checksum 2025-10-10T01:21:05.2605919Z ec62e2e53434: Download complete 2025-10-10T01:21:32.8316229Z a03ec5361452: Pull complete 2025-10-10T01:21:33.2262096Z 4f4fb700ef54: Pull complete 2025-10-10T01:21:33.6088816Z 84681debfb5c: Pull complete 2025-10-10T01:21:33.7820485Z d6a77cb34f2d: Pull complete 2025-10-10T01:21:34.0512491Z b4e994693c03: Pull complete 2025-10-10T01:21:34.4431791Z c6c1bfa3b4c5: Pull complete 2025-10-10T01:21:34.7506182Z eb61ec0f4c41: Pull complete 2025-10-10T01:21:35.0599190Z 31c940b54575: Pull complete 2025-10-10T01:21:35.3906907Z 21e8a5f2f65c: Pull complete 2025-10-10T01:21:35.7354166Z 496894807e5c: Pull complete 2025-10-10T01:21:58.3604674Z 1ae67e80b1bd: Verifying Checksum 2025-10-10T01:21:58.3607257Z 1ae67e80b1bd: Download complete 2025-10-10T01:22:51.5629875Z 1ae67e80b1bd: Pull complete 2025-10-10T01:22:52.0071599Z cf225908c05b: Pull complete 2025-10-10T01:22:52.8841314Z 381895ce55f6: Pull complete 2025-10-10T01:22:53.3906866Z 9263c0b02502: Pull complete 2025-10-10T01:22:53.7806157Z df861c28c3a5: Pull complete 2025-10-10T01:23:01.7910459Z 72757cf9a787: Pull complete 2025-10-10T01:23:02.2810573Z 6ceacb336387: Pull complete 2025-10-10T01:23:02.7446500Z f75cb5b656f8: Pull complete 2025-10-10T01:23:03.4552263Z 34e0c35a9893: Pull complete 2025-10-10T01:23:03.8074444Z a3022990e5c3: Pull complete 2025-10-10T01:23:04.4152722Z 93ef959930dc: Pull complete 2025-10-10T01:23:04.8049423Z fa921863e6e3: Pull complete 2025-10-10T01:23:05.1612245Z bab2c667cc1a: Pull complete 2025-10-10T01:23:05.8886391Z 3bc21d940f02: Pull complete 2025-10-10T01:23:06.2891697Z 5a5fb6edd17b: Pull complete 2025-10-10T01:23:06.6444727Z 62cb56db3356: Pull complete 2025-10-10T01:23:07.5293020Z a1fff40729fd: Pull complete 2025-10-10T01:23:08.7573355Z 50e3f0227909: Pull complete 2025-10-10T01:23:56.6637812Z ec62e2e53434: Pull complete 2025-10-10T01:23:57.0480059Z a47c72a0c062: Pull complete 2025-10-10T01:23:57.3991886Z 04fa9fbfdbbd: Pull complete 2025-10-10T01:23:58.1459257Z df1cf69f3334: Pull complete 2025-10-10T01:23:58.9241330Z 6ad7c264e843: Pull complete 2025-10-10T01:23:59.3059546Z a08c722c1447: Pull complete 2025-10-10T01:23:59.9730302Z 38deb86def9c: Pull complete 2025-10-10T01:24:00.3658171Z 8a6cf39f45d1: Pull complete 2025-10-10T01:24:00.5610062Z a37d78416813: Pull complete 2025-10-10T01:24:01.5679359Z 001023e5b311: Pull complete 2025-10-10T01:24:02.0506881Z 5b84bc540263: Pull complete 2025-10-10T01:24:02.8860323Z 7b03cfd9cf3c: Pull complete 2025-10-10T01:24:03.4113079Z 3ed8e92a0190: Pull complete 2025-10-10T01:24:04.3851295Z 477967cba78d: Pull complete 2025-10-10T01:24:04.8374991Z 6fffed0cb295: Pull complete 2025-10-10T01:24:11.5058415Z 8aa311eb036b: Pull complete 2025-10-10T01:24:11.8763700Z 33d965074649: Pull complete 2025-10-10T01:24:12.2524992Z dbf89dacc6c6: Pull complete 2025-10-10T01:24:12.7067081Z 208b50d13dad: Pull complete 2025-10-10T01:24:13.2372772Z 863fa5fce8b8: Pull complete 2025-10-10T01:24:13.7048188Z 2c57470f9b05: Pull complete 2025-10-10T01:24:15.5772425Z 4fbac50fbb3b: Pull complete 2025-10-10T01:24:15.8912930Z 6baf3cf73050: Pull complete 2025-10-10T01:24:17.5463011Z ace2a84685d2: Pull complete 2025-10-10T01:24:18.1981576Z Digest: sha256:dbf8617838c93f0ebd67b8b8bb938f5947e45ccc171ead9c6da61e625c32803f 2025-10-10T01:24:18.2741410Z Status: Downloaded newer image for 308535385114.dkr.ecr.us-east-1.amazonaws.com/pytorch/ci-image:pytorch-linux-jammy-cuda12.8-cudnn9-py3.12-gcc11-vllm-d8be0384e085f551506bd739678109fa0f5ee7ac 2025-10-10T01:24:18.3175465Z 308535385114.dkr.ecr.us-east-1.amazonaws.com/pytorch/ci-image:pytorch-linux-jammy-cuda12.8-cudnn9-py3.12-gcc11-vllm-d8be0384e085f551506bd739678109fa0f5ee7ac 2025-10-10T01:24:18.3237631Z ##[group]Run echo "IN_CONTAINER_RUNNER=$(if [ -f /.inarc ] || [ -f /.incontainer ]; then echo true ; else echo false; fi)" >> "$GITHUB_OUTPUT" 2025-10-10T01:24:18.3238469Z echo "IN_CONTAINER_RUNNER=$(if [ -f /.inarc ] || [ -f /.incontainer ]; then echo true ; else echo false; fi)" >> "$GITHUB_OUTPUT" 2025-10-10T01:24:18.3249799Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2025-10-10T01:24:18.3250091Z env: 2025-10-10T01:24:18.3250267Z GIT_DEFAULT_BRANCH: main 2025-10-10T01:24:18.3250476Z ##[endgroup] 2025-10-10T01:24:18.3544384Z ##[group]Run pytorch/test-infra/.github/actions/setup-nvidia@main 2025-10-10T01:24:18.3544728Z with: 2025-10-10T01:24:18.3544902Z driver-version: 580.82.07 2025-10-10T01:24:18.3545085Z env: 2025-10-10T01:24:18.3545239Z GIT_DEFAULT_BRANCH: main 2025-10-10T01:24:18.3545430Z ##[endgroup] 2025-10-10T01:24:18.3696348Z ##[group]Run nick-fields/retry@3e91a01664abd3c5cd539100d10d33b9c5b68482 2025-10-10T01:24:18.3696652Z with: 2025-10-10T01:24:18.3696811Z timeout_minutes: 10 2025-10-10T01:24:18.3696985Z max_attempts: 3 2025-10-10T01:24:18.3716671Z command: # Is it disgusting to have a full shell script here in this github action? Sure # But is it the best way to make it so that this action relies on nothing else? Absolutely set -eou pipefail DISTRIBUTION=$(. /etc/os-release;echo $ID$VERSION_ID) DRIVER_FN="NVIDIA-Linux-x86_64-${DRIVER_VERSION}.run" install_nvidia_docker2_amzn2() { ( set -x # Needed for yum-config-manager sudo yum install -y yum-utils if [[ "${DISTRIBUTION}" == "amzn2023" ]] ; then YUM_REPO_URL="https://nvidia.github.io/libnvidia-container/stable/rpm/nvidia-container-toolkit.repo" else # Amazon Linux 2 YUM_REPO_URL="https://nvidia.github.io/nvidia-docker/${DISTRIBUTION}/nvidia-docker.repo" fi sudo yum-config-manager --add-repo "${YUM_REPO_URL}" sudo yum install -y \ nvidia-container-toolkit-1.17.8 \ libnvidia-container-tools-1.17.8 \ libnvidia-container1-1.17.8 \ nvidia-container-toolkit-base-1.17.8 sudo systemctl restart docker ) } install_nvidia_docker2_ubuntu20() { ( set -x # Install nvidia-driver package if not installed status="$(dpkg-query -W --showformat='${db:Status-Status}' nvidia-docker2 2>&1)" if [ ! $? = 0 ] || [ ! "$status" = installed ]; then sudo apt-get install -y nvidia-container-toolkit-1.17.8 sudo systemctl restart docker fi ) } pre_install_nvidia_driver_amzn2() { ( # Purge any nvidia driver installed from RHEL repo sudo yum remove -y nvidia-driver-latest-dkms ) } install_nvidia_driver_common() { ( # Try to gather more information about the runner and its existing NVIDIA driver if any echo "Before installing NVIDIA driver" lspci lsmod modinfo nvidia || true HAS_NVIDIA_DRIVER=0 # Check if NVIDIA driver has already been installed if [ -x "$(command -v nvidia-smi)" ]; then set +e # The driver exists, check its version next. Also check only the first GPU if there are more than one of them # so that the same driver version is not print over multiple lines INSTALLED_DRIVER_VERSION=$(nvidia-smi --query-gpu=driver_version --format=csv,noheader --id=0) NVIDIA_SMI_STATUS=$? if [ "$NVIDIA_SMI_STATUS" -ne 0 ] && [ "$NVIDIA_SMI_STATUS" -ne 14 ]; then echo "Failed to get NVIDIA driver version ($INSTALLED_DRIVER_VERSION). Continuing" elif [ "$INSTALLED_DRIVER_VERSION" != "$DRIVER_VERSION" ]; then echo "NVIDIA driver ($INSTALLED_DRIVER_VERSION) has been installed, but we expect to have $DRIVER_VERSION instead. Continuing" # Turn off persistent mode so that the installation script can unload the kernel module sudo killall nvidia-persistenced || true else HAS_NVIDIA_DRIVER=1 echo "NVIDIA driver ($INSTALLED_DRIVER_VERSION) has already been installed. Skipping NVIDIA driver installation" fi set -e fi if [ "$HAS_NVIDIA_DRIVER" -eq 0 ]; then # CAUTION: this may need to be updated in future if [ "${DISTRIBUTION}" != ubuntu20.04 ]; then sudo yum groupinstall -y "Development Tools" # ensure our kernel install is the same as our underlying kernel, # groupinstall "Development Tools" has a habit of mismatching kernel headers sudo yum install -y "kernel-devel-uname-r == $(uname -r)" sudo modprobe backlight fi sudo curl -fsL -o /tmp/nvidia_driver "https://s3.amazonaws.com/ossci-linux/nvidia_driver/$DRIVER_FN" set +e sudo /bin/bash /tmp/nvidia_driver -s --no-drm NVIDIA_INSTALLATION_STATUS=$? RESET_GPU=0 if [ "$NVIDIA_INSTALLATION_STATUS" -ne 0 ]; then sudo cat /var/log/nvidia-installer.log # Fail to install NVIDIA driver, try to reset the GPU RESET_GPU=1 elif [ -x "$(command -v nvidia-smi)" ]; then # Check again if nvidia-smi works even if the driver installation completes successfully INSTALLED_DRIVER_VERSION=$(nvidia-smi --query-gpu=driver_version --format=csv,noheader --id=0) NVIDIA_SMI_STATUS=$? if [ "$NVIDIA_SMI_STATUS" -ne 0 ] && [ "$NVIDIA_SMI_STATUS" -ne 14 ]; then RESET_GPU=1 fi fi if [ "$RESET_GPU" -eq 1 ]; then NVIDIA_DEVICES=$(lspci -D | grep -i NVIDIA | cut -d' ' -f1) # The GPU can get stuck in a failure state if somehow the test crashs the GPU microcode. When this # happens, we'll try to reset all NVIDIA devices https://github.com/pytorch/pytorch/issues/88388 for PCI_ID in $NVIDIA_DEVICES; do DEVICE_ENABLED=$(cat /sys/bus/pci/devices/$PCI_ID/enable) echo "Reseting $PCI_ID (enabled state: $DEVICE_ENABLED)" # This requires sudo permission of course echo "1" | sudo tee /sys/bus/pci/devices/$PCI_ID/reset sleep 1 done fi sudo rm -fv /tmp/nvidia_driver set -e fi ) } post_install_nvidia_driver_common() { ( sudo modprobe nvidia || true echo "After installing NVIDIA driver" lspci lsmod modinfo nvidia || true ( set +e nvidia-smi # NB: Annoyingly, nvidia-smi command returns successfully with return code 0 even in # the case where the driver has already crashed as it still can get the driver version # and some basic information like the bus ID. However, the rest of the information # would be missing (ERR!), for example: # # +-----------------------------------------------------------------------------+ # | NVIDIA-SMI 525.89.02 Driver Version: 525.89.02 CUDA Version: 12.0 | # |-------------------------------+----------------------+----------------------+ # | GPU Name Persistence-M| Bus-Id Disp.A | Volatile Uncorr. ECC | # | Fan Temp Perf Pwr:Usage/Cap| Memory-Usage | GPU-Util Compute M. | # | | | MIG M. | # |===============================+======================+======================| # | 0 ERR! Off | 00000000:00:1E.0 Off | ERR! | # |ERR! ERR! ERR! ERR! / ERR! | 4184MiB / 23028MiB | ERR! Default | # | | | ERR! | # +-------------------------------+----------------------+----------------------+ # # +-----------------------------------------------------------------------------+ # | Processes: | # | GPU GI CI PID Type Process name GPU Memory | # | ID ID Usage | # |=============================================================================| # +-----------------------------------------------------------------------------+ # # This should be reported as a failure instead as it will guarantee to fail when # Docker tries to run with --gpus all # # So, the correct check here is to query one of the missing piece of info like # GPU name, so that the command can fail accordingly nvidia-smi --query-gpu=gpu_name --format=csv,noheader --id=0 NVIDIA_SMI_STATUS=$? # Allowable exit statuses for nvidia-smi, see: https://github.com/NVIDIA/gpu-operator/issues/285 if [ "$NVIDIA_SMI_STATUS" -eq 0 ] || [ "$NVIDIA_SMI_STATUS" -eq 14 ]; then echo "INFO: Ignoring allowed status ${NVIDIA_SMI_STATUS}" else echo "ERROR: nvidia-smi exited with unresolved status ${NVIDIA_SMI_STATUS}" exit ${NVIDIA_SMI_STATUS} fi set -e ) ) } install_nvidia_driver_amzn2() { ( set -x pre_install_nvidia_driver_amzn2 install_nvidia_driver_common post_install_nvidia_driver_common ) } install_nvidia_driver_ubuntu20() { ( set -x install_nvidia_driver_common post_install_nvidia_driver_common ) } echo "== Installing nvidia driver ${DRIVER_FN} ==" case "${DISTRIBUTION}" in amzn*) install_nvidia_driver_amzn2 ;; ubuntu20.04) install_nvidia_driver_ubuntu20 ;; *) echo "ERROR: Unknown distribution ${DISTRIBUTION}" exit 1 ;; esac # Install container toolkit based on distribution echo "== Installing nvidia container toolkit for ${DISTRIBUTION} ==" case "${DISTRIBUTION}" in amzn*) install_nvidia_docker2_amzn2 ;; ubuntu20.04) install_nvidia_docker2_ubuntu20 ;; *) echo "ERROR: Unknown distribution ${DISTRIBUTION}" exit 1 ;; esac echo "GPU_FLAG=--gpus all -e NVIDIA_DRIVER_CAPABILITIES=all" >> "${GITHUB_ENV}" # Fix https://github.com/NVIDIA/nvidia-docker/issues/1648 on runners with # more than one GPUs. This just needs to be run once. The command fails # on subsequent runs and complains that the mode is already on, but that's # ok sudo nvidia-persistenced || true # This should show persistence mode ON nvidia-smi # check if the container-toolkit is correctly installed and CUDA is available inside a container docker run --rm -t --gpus=all public.ecr.aws/docker/library/python:3.13 nvidia-smi 2025-10-10T01:24:18.3736745Z retry_wait_seconds: 10 2025-10-10T01:24:18.3736955Z polling_interval_seconds: 1 2025-10-10T01:24:18.3737160Z warning_on_retry: true 2025-10-10T01:24:18.3737342Z continue_on_error: false 2025-10-10T01:24:18.3737540Z env: 2025-10-10T01:24:18.3737695Z GIT_DEFAULT_BRANCH: main 2025-10-10T01:24:18.3737886Z DRIVER_VERSION: 580.82.07 2025-10-10T01:24:18.3738063Z ##[endgroup] 2025-10-10T01:24:18.5235841Z == Installing nvidia driver NVIDIA-Linux-x86_64-580.82.07.run == 2025-10-10T01:24:18.5239089Z + pre_install_nvidia_driver_amzn2 2025-10-10T01:24:18.5242063Z + sudo yum remove -y nvidia-driver-latest-dkms 2025-10-10T01:24:19.3651759Z No match for argument: nvidia-driver-latest-dkms 2025-10-10T01:24:19.3652498Z No packages marked for removal. 2025-10-10T01:24:19.3706494Z Dependencies resolved. 2025-10-10T01:24:19.3715173Z Nothing to do. 2025-10-10T01:24:19.3715615Z Complete! 2025-10-10T01:24:19.5190139Z + install_nvidia_driver_common 2025-10-10T01:24:19.5197261Z + echo 'Before installing NVIDIA driver' 2025-10-10T01:24:19.5197496Z + lspci 2025-10-10T01:24:19.5199401Z Before installing NVIDIA driver 2025-10-10T01:24:19.6705279Z 00:00.0 Host bridge: Intel Corporation 440FX - 82441FX PMC [Natoma] 2025-10-10T01:24:19.6705819Z 00:01.0 ISA bridge: Intel Corporation 82371SB PIIX3 ISA [Natoma/Triton II] 2025-10-10T01:24:19.6706358Z 00:01.3 Non-VGA unclassified device: Intel Corporation 82371AB/EB/MB PIIX4 ACPI (rev 08) 2025-10-10T01:24:19.6707140Z 00:03.0 VGA compatible controller: Amazon.com, Inc. Device 1111 2025-10-10T01:24:19.6707667Z 00:04.0 Non-Volatile memory controller: Amazon.com, Inc. NVMe EBS Controller 2025-10-10T01:24:19.6708105Z 01:00.0 PCI bridge: Amazon.com, Inc. Device 0200 2025-10-10T01:24:19.6708443Z 02:00.0 PCI bridge: Amazon.com, Inc. Device 0200 2025-10-10T01:24:19.6708761Z 03:00.0 PCI bridge: Amazon.com, Inc. Device 0200 2025-10-10T01:24:19.6709086Z 03:00.1 PCI bridge: Amazon.com, Inc. Device 0200 2025-10-10T01:24:19.6709395Z 03:00.2 PCI bridge: Amazon.com, Inc. Device 0200 2025-10-10T01:24:19.6709700Z 03:00.3 PCI bridge: Amazon.com, Inc. Device 0200 2025-10-10T01:24:19.6709996Z 03:00.4 PCI bridge: Amazon.com, Inc. Device 0200 2025-10-10T01:24:19.6710306Z 03:00.5 PCI bridge: Amazon.com, Inc. Device 0200 2025-10-10T01:24:19.6710609Z 03:00.6 PCI bridge: Amazon.com, Inc. Device 0200 2025-10-10T01:24:19.6710911Z 03:00.7 PCI bridge: Amazon.com, Inc. Device 0200 2025-10-10T01:24:19.6711206Z 03:01.0 PCI bridge: Amazon.com, Inc. Device 0200 2025-10-10T01:24:19.6711517Z 03:01.1 PCI bridge: Amazon.com, Inc. Device 0200 2025-10-10T01:24:19.6711820Z 03:01.2 PCI bridge: Amazon.com, Inc. Device 0200 2025-10-10T01:24:19.6712121Z 03:01.3 PCI bridge: Amazon.com, Inc. Device 0200 2025-10-10T01:24:19.6712413Z 03:01.4 PCI bridge: Amazon.com, Inc. Device 0200 2025-10-10T01:24:19.6712715Z 03:01.5 PCI bridge: Amazon.com, Inc. Device 0200 2025-10-10T01:24:19.6713014Z 03:01.6 PCI bridge: Amazon.com, Inc. Device 0200 2025-10-10T01:24:19.6713314Z 03:01.7 PCI bridge: Amazon.com, Inc. Device 0200 2025-10-10T01:24:19.6713630Z 03:02.0 PCI bridge: Amazon.com, Inc. Device 0200 2025-10-10T01:24:19.6713927Z 03:02.1 PCI bridge: Amazon.com, Inc. Device 0200 2025-10-10T01:24:19.6714308Z 03:02.2 PCI bridge: Amazon.com, Inc. Device 0200 2025-10-10T01:24:19.6714667Z 03:02.3 PCI bridge: Amazon.com, Inc. Device 0200 2025-10-10T01:24:19.6715025Z 03:02.4 PCI bridge: Amazon.com, Inc. Device 0200 2025-10-10T01:24:19.6715387Z 03:02.5 PCI bridge: Amazon.com, Inc. Device 0200 2025-10-10T01:24:19.6715906Z 03:02.6 PCI bridge: Amazon.com, Inc. Device 0200 2025-10-10T01:24:19.6716198Z 03:02.7 PCI bridge: Amazon.com, Inc. Device 0200 2025-10-10T01:24:19.6716484Z 03:03.0 PCI bridge: Amazon.com, Inc. Device 0200 2025-10-10T01:24:19.6716741Z 03:03.1 PCI bridge: Amazon.com, Inc. Device 0200 2025-10-10T01:24:19.6716978Z 03:03.2 PCI bridge: Amazon.com, Inc. Device 0200 2025-10-10T01:24:19.6717225Z 03:03.3 PCI bridge: Amazon.com, Inc. Device 0200 2025-10-10T01:24:19.6717469Z 03:03.4 PCI bridge: Amazon.com, Inc. Device 0200 2025-10-10T01:24:19.6717711Z 03:03.5 PCI bridge: Amazon.com, Inc. Device 0200 2025-10-10T01:24:19.6717950Z 03:03.6 PCI bridge: Amazon.com, Inc. Device 0200 2025-10-10T01:24:19.6718194Z 03:03.7 PCI bridge: Amazon.com, Inc. Device 0200 2025-10-10T01:24:19.6718530Z 24:00.0 PCI bridge: Amazon.com, Inc. Device 0200 2025-10-10T01:24:19.6718781Z 25:00.0 PCI bridge: Amazon.com, Inc. Device 0200 2025-10-10T01:24:19.6719020Z 26:00.0 PCI bridge: Amazon.com, Inc. Device 0200 2025-10-10T01:24:19.6719275Z 26:00.1 PCI bridge: Amazon.com, Inc. Device 0200 2025-10-10T01:24:19.6719518Z 26:00.2 PCI bridge: Amazon.com, Inc. Device 0200 2025-10-10T01:24:19.6719760Z 26:00.3 PCI bridge: Amazon.com, Inc. Device 0200 2025-10-10T01:24:19.6719997Z 26:00.4 PCI bridge: Amazon.com, Inc. Device 0200 2025-10-10T01:24:19.6720243Z 26:00.5 PCI bridge: Amazon.com, Inc. Device 0200 2025-10-10T01:24:19.6720486Z 26:00.6 PCI bridge: Amazon.com, Inc. Device 0200 2025-10-10T01:24:19.6720731Z 26:00.7 PCI bridge: Amazon.com, Inc. Device 0200 2025-10-10T01:24:19.6720969Z 26:01.0 PCI bridge: Amazon.com, Inc. Device 0200 2025-10-10T01:24:19.6721296Z 27:00.0 Ethernet controller: Amazon.com, Inc. Elastic Network Adapter (ENA) 2025-10-10T01:24:19.6721635Z 30:00.0 PCI bridge: Amazon.com, Inc. Device 0200 2025-10-10T01:24:19.6721883Z 31:00.0 PCI bridge: Amazon.com, Inc. Device 0200 2025-10-10T01:24:19.6722131Z 32:00.0 PCI bridge: Amazon.com, Inc. Device 0200 2025-10-10T01:24:19.6722374Z 32:00.1 PCI bridge: Amazon.com, Inc. Device 0200 2025-10-10T01:24:19.6722735Z 32:00.2 PCI bridge: Amazon.com, Inc. Device 0200 2025-10-10T01:24:19.6722986Z 32:00.3 PCI bridge: Amazon.com, Inc. Device 0200 2025-10-10T01:24:19.6723309Z 33:00.0 Non-Volatile memory controller: Amazon.com, Inc. NVMe SSD Controller 2025-10-10T01:24:19.6723721Z 34:00.0 Non-Volatile memory controller: Amazon.com, Inc. NVMe SSD Controller 2025-10-10T01:24:19.6724561Z 35:00.0 Non-Volatile memory controller: Amazon.com, Inc. NVMe SSD Controller 2025-10-10T01:24:19.6724989Z 36:00.0 Non-Volatile memory controller: Amazon.com, Inc. NVMe SSD Controller 2025-10-10T01:24:19.6725313Z 37:00.0 PCI bridge: Amazon.com, Inc. Device 0200 2025-10-10T01:24:19.6725614Z 38:00.0 3D controller: NVIDIA Corporation AD104GL [L4] (rev a1) 2025-10-10T01:24:19.6725916Z 39:00.0 PCI bridge: Amazon.com, Inc. Device 0200 2025-10-10T01:24:19.6726210Z 3a:00.0 3D controller: NVIDIA Corporation AD104GL [L4] (rev a1) 2025-10-10T01:24:19.6726499Z 3b:00.0 PCI bridge: Amazon.com, Inc. Device 0200 2025-10-10T01:24:19.6726791Z 3c:00.0 3D controller: NVIDIA Corporation AD104GL [L4] (rev a1) 2025-10-10T01:24:19.6727063Z 3d:00.0 PCI bridge: Amazon.com, Inc. Device 0200 2025-10-10T01:24:19.6727337Z 3e:00.0 3D controller: NVIDIA Corporation AD104GL [L4] (rev a1) 2025-10-10T01:24:19.6727596Z + lsmod 2025-10-10T01:24:19.6764970Z Module Size Used by 2025-10-10T01:24:19.6765477Z nvidia_uvm 1925120 0 2025-10-10T01:24:19.6765686Z nvidia 14286848 1 nvidia_uvm 2025-10-10T01:24:19.6765910Z drm 602112 1 nvidia 2025-10-10T01:24:19.6766148Z drm_panel_orientation_quirks 32768 1 drm 2025-10-10T01:24:19.6766396Z backlight 24576 1 drm 2025-10-10T01:24:19.6766620Z i2c_core 110592 2 nvidia,drm 2025-10-10T01:24:19.6766839Z xt_conntrack 16384 1 2025-10-10T01:24:19.6767051Z nft_chain_nat 16384 3 2025-10-10T01:24:19.6767243Z xt_MASQUERADE 20480 1 2025-10-10T01:24:19.6767478Z nf_nat 57344 2 nft_chain_nat,xt_MASQUERADE 2025-10-10T01:24:19.6770144Z nf_conntrack_netlink 57344 0 2025-10-10T01:24:19.6770456Z nf_conntrack 184320 4 xt_conntrack,nf_nat,nf_conntrack_netlink,xt_MASQUERADE 2025-10-10T01:24:19.6770794Z nf_defrag_ipv6 24576 1 nf_conntrack 2025-10-10T01:24:19.6771032Z nf_defrag_ipv4 16384 1 nf_conntrack 2025-10-10T01:24:19.6771260Z xfrm_user 57344 1 2025-10-10T01:24:19.6771473Z xfrm_algo 16384 1 xfrm_user 2025-10-10T01:24:19.6771696Z xt_addrtype 16384 2 2025-10-10T01:24:19.6771889Z nft_compat 20480 4 2025-10-10T01:24:19.6772109Z nf_tables 311296 57 nft_compat,nft_chain_nat 2025-10-10T01:24:19.6772424Z nfnetlink 20480 4 nft_compat,nf_conntrack_netlink,nf_tables 2025-10-10T01:24:19.6772709Z br_netfilter 36864 0 2025-10-10T01:24:19.6772914Z bridge 323584 1 br_netfilter 2025-10-10T01:24:19.6773128Z stp 16384 1 bridge 2025-10-10T01:24:19.6773347Z llc 16384 2 bridge,stp 2025-10-10T01:24:19.6773577Z overlay 167936 0 2025-10-10T01:24:19.6773776Z tls 139264 0 2025-10-10T01:24:19.6773952Z nls_ascii 16384 1 2025-10-10T01:24:19.6774134Z nls_cp437 20480 1 2025-10-10T01:24:19.6774311Z vfat 24576 1 2025-10-10T01:24:19.6774494Z fat 86016 1 vfat 2025-10-10T01:24:19.6774687Z ghash_clmulni_intel 16384 0 2025-10-10T01:24:19.6774876Z i8042 45056 0 2025-10-10T01:24:19.6775049Z ena 184320 0 2025-10-10T01:24:19.6775242Z serio 28672 3 i8042 2025-10-10T01:24:19.6775439Z button 24576 0 2025-10-10T01:24:19.6775623Z sunrpc 700416 1 2025-10-10T01:24:19.6775807Z sch_fq_codel 20480 9 2025-10-10T01:24:19.6775988Z dm_mod 188416 0 2025-10-10T01:24:19.6776161Z fuse 184320 1 2025-10-10T01:24:19.6776336Z loop 36864 0 2025-10-10T01:24:19.6776653Z configfs 57344 1 2025-10-10T01:24:19.6776844Z dmi_sysfs 20480 0 2025-10-10T01:24:19.6777022Z crc32_pclmul 16384 0 2025-10-10T01:24:19.6777206Z crc32c_intel 24576 0 2025-10-10T01:24:19.6777387Z efivarfs 24576 1 2025-10-10T01:24:19.6777577Z + modinfo nvidia 2025-10-10T01:24:19.6785778Z filename: /lib/modules/6.1.150-174.273.amzn2023.x86_64/kernel/drivers/video/nvidia.ko 2025-10-10T01:24:19.6786680Z import_ns: DMA_BUF 2025-10-10T01:24:19.6787131Z alias: char-major-195-* 2025-10-10T01:24:19.6787614Z version: 580.82.07 2025-10-10T01:24:19.6788050Z supported: external 2025-10-10T01:24:19.6788482Z license: Dual MIT/GPL 2025-10-10T01:24:19.6789004Z firmware: nvidia/580.82.07/gsp_tu10x.bin 2025-10-10T01:24:19.6789614Z firmware: nvidia/580.82.07/gsp_ga10x.bin 2025-10-10T01:24:19.6790193Z srcversion: BA7240A71DCF7DC6FE88C1D 2025-10-10T01:24:19.6790787Z alias: of:N*T*Cnvidia,tegra264-displayC* 2025-10-10T01:24:19.6791432Z alias: of:N*T*Cnvidia,tegra264-display 2025-10-10T01:24:19.6792075Z alias: of:N*T*Cnvidia,tegra234-displayC* 2025-10-10T01:24:19.6792698Z alias: of:N*T*Cnvidia,tegra234-display 2025-10-10T01:24:19.6793289Z alias: pci:v000010DEd*sv*sd*bc06sc80i00* 2025-10-10T01:24:19.6793866Z alias: pci:v000010DEd*sv*sd*bc03sc02i00* 2025-10-10T01:24:19.6794426Z alias: pci:v000010DEd*sv*sd*bc03sc00i00* 2025-10-10T01:24:19.6794971Z depends: i2c-core,drm 2025-10-10T01:24:19.6795394Z retpoline: Y 2025-10-10T01:24:19.6795582Z name: nvidia 2025-10-10T01:24:19.6795857Z vermagic: 6.1.150-174.273.amzn2023.x86_64 SMP preempt mod_unload modversions 2025-10-10T01:24:19.6796232Z parm: NvSwitchRegDwords:NvSwitch regkey (charp) 2025-10-10T01:24:19.6796570Z parm: NvSwitchBlacklist:NvSwitchBlacklist=uuid[,uuid...] (charp) 2025-10-10T01:24:19.6796874Z parm: NVreg_ResmanDebugLevel:int 2025-10-10T01:24:19.6797222Z parm: NVreg_RmLogonRC:int 2025-10-10T01:24:19.6797449Z parm: NVreg_ModifyDeviceFiles:int 2025-10-10T01:24:19.6797685Z parm: NVreg_DeviceFileUID:int 2025-10-10T01:24:19.6797907Z parm: NVreg_DeviceFileGID:int 2025-10-10T01:24:19.6798122Z parm: NVreg_DeviceFileMode:int 2025-10-10T01:24:19.6798471Z parm: NVreg_InitializeSystemMemoryAllocations:int 2025-10-10T01:24:19.6798765Z parm: NVreg_UsePageAttributeTable:int 2025-10-10T01:24:19.6799012Z parm: NVreg_EnablePCIeGen3:int 2025-10-10T01:24:19.6799232Z parm: NVreg_EnableMSI:int 2025-10-10T01:24:19.6799458Z parm: NVreg_EnableStreamMemOPs:int 2025-10-10T01:24:19.6799726Z parm: NVreg_RestrictProfilingToAdminUsers:int 2025-10-10T01:24:19.6800035Z parm: NVreg_PreserveVideoMemoryAllocations:int 2025-10-10T01:24:19.6800311Z parm: NVreg_EnableS0ixPowerManagement:int 2025-10-10T01:24:19.6800618Z parm: NVreg_S0ixPowerManagementVideoMemoryThreshold:int 2025-10-10T01:24:19.6800936Z parm: NVreg_DynamicPowerManagement:int 2025-10-10T01:24:19.6801245Z parm: NVreg_DynamicPowerManagementVideoMemoryThreshold:int 2025-10-10T01:24:19.6801540Z parm: NVreg_EnableGpuFirmware:int 2025-10-10T01:24:19.6801788Z parm: NVreg_EnableGpuFirmwareLogs:int 2025-10-10T01:24:19.6802061Z parm: NVreg_OpenRmEnableUnsupportedGpus:int 2025-10-10T01:24:19.6802354Z parm: NVreg_EnableUserNUMAManagement:int 2025-10-10T01:24:19.6802602Z parm: NVreg_MemoryPoolSize:int 2025-10-10T01:24:19.6802844Z parm: NVreg_KMallocHeapMaxSize:int 2025-10-10T01:24:19.6803091Z parm: NVreg_VMallocHeapMaxSize:int 2025-10-10T01:24:19.6803329Z parm: NVreg_IgnoreMMIOCheck:int 2025-10-10T01:24:19.6803557Z parm: NVreg_NvLinkDisable:int 2025-10-10T01:24:19.6803814Z parm: NVreg_EnablePCIERelaxedOrderingMode:int 2025-10-10T01:24:19.6804078Z parm: NVreg_RegisterPCIDriver:int 2025-10-10T01:24:19.6804441Z parm: NVreg_RegisterPlatformDeviceDriver:int 2025-10-10T01:24:19.6804714Z parm: NVreg_EnableResizableBar:int 2025-10-10T01:24:19.6804961Z parm: NVreg_EnableDbgBreakpoint:int 2025-10-10T01:24:19.6805214Z parm: NVreg_EnableNonblockingOpen:int 2025-10-10T01:24:19.6805484Z parm: NVreg_CoherentGPUMemoryMode:charp 2025-10-10T01:24:19.6805738Z parm: NVreg_RegistryDwords:charp 2025-10-10T01:24:19.6805991Z parm: NVreg_RegistryDwordsPerDevice:charp 2025-10-10T01:24:19.6806237Z parm: NVreg_RmMsg:charp 2025-10-10T01:24:19.6806454Z parm: NVreg_GpuBlacklist:charp 2025-10-10T01:24:19.6806695Z parm: NVreg_TemporaryFilePath:charp 2025-10-10T01:24:19.6806930Z parm: NVreg_ExcludedGpus:charp 2025-10-10T01:24:19.6807176Z parm: NVreg_DmaRemapPeerMmio:int 2025-10-10T01:24:19.6807424Z parm: NVreg_RmNvlinkBandwidth:charp 2025-10-10T01:24:19.6807694Z parm: NVreg_RmNvlinkBandwidthLinkCount:int 2025-10-10T01:24:19.6807954Z parm: NVreg_ImexChannelCount:int 2025-10-10T01:24:19.6808210Z parm: NVreg_CreateImexChannel0:int 2025-10-10T01:24:19.6808471Z parm: NVreg_GrdmaPciTopoCheckOverride:int 2025-10-10T01:24:19.6808723Z parm: rm_firmware_active:charp 2025-10-10T01:24:19.6808941Z + HAS_NVIDIA_DRIVER=0 2025-10-10T01:24:19.6809124Z ++ command -v nvidia-smi 2025-10-10T01:24:19.6809318Z + '[' -x /usr/bin/nvidia-smi ']' 2025-10-10T01:24:19.6809510Z + set +e 2025-10-10T01:24:19.6809738Z ++ nvidia-smi --query-gpu=driver_version --format=csv,noheader --id=0 2025-10-10T01:24:23.0499523Z + INSTALLED_DRIVER_VERSION=580.82.07 2025-10-10T01:24:23.0500227Z + NVIDIA_SMI_STATUS=0 2025-10-10T01:24:23.0500678Z + '[' 0 -ne 0 ']' 2025-10-10T01:24:23.0501072Z + '[' 580.82.07 '!=' 580.82.07 ']' 2025-10-10T01:24:23.0501542Z + HAS_NVIDIA_DRIVER=1 2025-10-10T01:24:23.0502354Z + echo 'NVIDIA driver (580.82.07) has already been installed. Skipping NVIDIA driver installation' 2025-10-10T01:24:23.0504086Z + set -e 2025-10-10T01:24:23.0504443Z + '[' 1 -eq 0 ']' 2025-10-10T01:24:23.0505150Z NVIDIA driver (580.82.07) has already been installed. Skipping NVIDIA driver installation 2025-10-10T01:24:23.0506010Z + post_install_nvidia_driver_common 2025-10-10T01:24:23.0513134Z + sudo modprobe nvidia 2025-10-10T01:24:23.2034021Z + echo 'After installing NVIDIA driver' 2025-10-10T01:24:23.2034633Z + lspci 2025-10-10T01:24:23.2035040Z After installing NVIDIA driver 2025-10-10T01:24:23.2236861Z 00:00.0 Host bridge: Intel Corporation 440FX - 82441FX PMC [Natoma] 2025-10-10T01:24:23.2237872Z 00:01.0 ISA bridge: Intel Corporation 82371SB PIIX3 ISA [Natoma/Triton II] 2025-10-10T01:24:23.2239058Z 00:01.3 Non-VGA unclassified device: Intel Corporation 82371AB/EB/MB PIIX4 ACPI (rev 08) 2025-10-10T01:24:23.2240060Z 00:03.0 VGA compatible controller: Amazon.com, Inc. Device 1111 2025-10-10T01:24:23.2240990Z 00:04.0 Non-Volatile memory controller: Amazon.com, Inc. NVMe EBS Controller 2025-10-10T01:24:23.2241872Z 01:00.0 PCI bridge: Amazon.com, Inc. Device 0200 2025-10-10T01:24:23.2242519Z 02:00.0 PCI bridge: Amazon.com, Inc. Device 0200 2025-10-10T01:24:23.2243130Z 03:00.0 PCI bridge: Amazon.com, Inc. Device 0200 2025-10-10T01:24:23.2243719Z 03:00.1 PCI bridge: Amazon.com, Inc. Device 0200 2025-10-10T01:24:23.2244298Z 03:00.2 PCI bridge: Amazon.com, Inc. Device 0200 2025-10-10T01:24:23.2244863Z 03:00.3 PCI bridge: Amazon.com, Inc. Device 0200 2025-10-10T01:24:23.2245467Z 03:00.4 PCI bridge: Amazon.com, Inc. Device 0200 2025-10-10T01:24:23.2246144Z 03:00.5 PCI bridge: Amazon.com, Inc. Device 0200 2025-10-10T01:24:23.2246824Z 03:00.6 PCI bridge: Amazon.com, Inc. Device 0200 2025-10-10T01:24:23.2247279Z 03:00.7 PCI bridge: Amazon.com, Inc. Device 0200 2025-10-10T01:24:23.2247568Z 03:01.0 PCI bridge: Amazon.com, Inc. Device 0200 2025-10-10T01:24:23.2247829Z 03:01.1 PCI bridge: Amazon.com, Inc. Device 0200 2025-10-10T01:24:23.2248087Z 03:01.2 PCI bridge: Amazon.com, Inc. Device 0200 2025-10-10T01:24:23.2248623Z 03:01.3 PCI bridge: Amazon.com, Inc. Device 0200 2025-10-10T01:24:23.2248901Z 03:01.4 PCI bridge: Amazon.com, Inc. Device 0200 2025-10-10T01:24:23.2249147Z 03:01.5 PCI bridge: Amazon.com, Inc. Device 0200 2025-10-10T01:24:23.2249389Z 03:01.6 PCI bridge: Amazon.com, Inc. Device 0200 2025-10-10T01:24:23.2249629Z 03:01.7 PCI bridge: Amazon.com, Inc. Device 0200 2025-10-10T01:24:23.2249872Z 03:02.0 PCI bridge: Amazon.com, Inc. Device 0200 2025-10-10T01:24:23.2250126Z 03:02.1 PCI bridge: Amazon.com, Inc. Device 0200 2025-10-10T01:24:23.2250379Z 03:02.2 PCI bridge: Amazon.com, Inc. Device 0200 2025-10-10T01:24:23.2250616Z 03:02.3 PCI bridge: Amazon.com, Inc. Device 0200 2025-10-10T01:24:23.2250856Z 03:02.4 PCI bridge: Amazon.com, Inc. Device 0200 2025-10-10T01:24:23.2251097Z 03:02.5 PCI bridge: Amazon.com, Inc. Device 0200 2025-10-10T01:24:23.2251339Z 03:02.6 PCI bridge: Amazon.com, Inc. Device 0200 2025-10-10T01:24:23.2251585Z 03:02.7 PCI bridge: Amazon.com, Inc. Device 0200 2025-10-10T01:24:23.2251833Z 03:03.0 PCI bridge: Amazon.com, Inc. Device 0200 2025-10-10T01:24:23.2252080Z 03:03.1 PCI bridge: Amazon.com, Inc. Device 0200 2025-10-10T01:24:23.2252321Z 03:03.2 PCI bridge: Amazon.com, Inc. Device 0200 2025-10-10T01:24:23.2252564Z 03:03.3 PCI bridge: Amazon.com, Inc. Device 0200 2025-10-10T01:24:23.2252801Z 03:03.4 PCI bridge: Amazon.com, Inc. Device 0200 2025-10-10T01:24:23.2253043Z 03:03.5 PCI bridge: Amazon.com, Inc. Device 0200 2025-10-10T01:24:23.2253285Z 03:03.6 PCI bridge: Amazon.com, Inc. Device 0200 2025-10-10T01:24:23.2253524Z 03:03.7 PCI bridge: Amazon.com, Inc. Device 0200 2025-10-10T01:24:23.2253775Z 24:00.0 PCI bridge: Amazon.com, Inc. Device 0200 2025-10-10T01:24:23.2254025Z 25:00.0 PCI bridge: Amazon.com, Inc. Device 0200 2025-10-10T01:24:23.2254267Z 26:00.0 PCI bridge: Amazon.com, Inc. Device 0200 2025-10-10T01:24:23.2254509Z 26:00.1 PCI bridge: Amazon.com, Inc. Device 0200 2025-10-10T01:24:23.2254746Z 26:00.2 PCI bridge: Amazon.com, Inc. Device 0200 2025-10-10T01:24:23.2255112Z 26:00.3 PCI bridge: Amazon.com, Inc. Device 0200 2025-10-10T01:24:23.2255356Z 26:00.4 PCI bridge: Amazon.com, Inc. Device 0200 2025-10-10T01:24:23.2255606Z 26:00.5 PCI bridge: Amazon.com, Inc. Device 0200 2025-10-10T01:24:23.2255843Z 26:00.6 PCI bridge: Amazon.com, Inc. Device 0200 2025-10-10T01:24:23.2256088Z 26:00.7 PCI bridge: Amazon.com, Inc. Device 0200 2025-10-10T01:24:23.2256329Z 26:01.0 PCI bridge: Amazon.com, Inc. Device 0200 2025-10-10T01:24:23.2256653Z 27:00.0 Ethernet controller: Amazon.com, Inc. Elastic Network Adapter (ENA) 2025-10-10T01:24:23.2256983Z 30:00.0 PCI bridge: Amazon.com, Inc. Device 0200 2025-10-10T01:24:23.2257262Z 31:00.0 PCI bridge: Amazon.com, Inc. Device 0200 2025-10-10T01:24:23.2257516Z 32:00.0 PCI bridge: Amazon.com, Inc. Device 0200 2025-10-10T01:24:23.2257769Z 32:00.1 PCI bridge: Amazon.com, Inc. Device 0200 2025-10-10T01:24:23.2258010Z 32:00.2 PCI bridge: Amazon.com, Inc. Device 0200 2025-10-10T01:24:23.2258279Z 32:00.3 PCI bridge: Amazon.com, Inc. Device 0200 2025-10-10T01:24:23.2258624Z 33:00.0 Non-Volatile memory controller: Amazon.com, Inc. NVMe SSD Controller 2025-10-10T01:24:23.2259027Z 34:00.0 Non-Volatile memory controller: Amazon.com, Inc. NVMe SSD Controller 2025-10-10T01:24:23.2259413Z 35:00.0 Non-Volatile memory controller: Amazon.com, Inc. NVMe SSD Controller 2025-10-10T01:24:23.2259789Z 36:00.0 Non-Volatile memory controller: Amazon.com, Inc. NVMe SSD Controller 2025-10-10T01:24:23.2260112Z 37:00.0 PCI bridge: Amazon.com, Inc. Device 0200 2025-10-10T01:24:23.2260407Z 38:00.0 3D controller: NVIDIA Corporation AD104GL [L4] (rev a1) 2025-10-10T01:24:23.2260696Z 39:00.0 PCI bridge: Amazon.com, Inc. Device 0200 2025-10-10T01:24:23.2260967Z 3a:00.0 3D controller: NVIDIA Corporation AD104GL [L4] (rev a1) 2025-10-10T01:24:23.2261246Z 3b:00.0 PCI bridge: Amazon.com, Inc. Device 0200 2025-10-10T01:24:23.2261518Z 3c:00.0 3D controller: NVIDIA Corporation AD104GL [L4] (rev a1) 2025-10-10T01:24:23.2261793Z 3d:00.0 PCI bridge: Amazon.com, Inc. Device 0200 2025-10-10T01:24:23.2262156Z 3e:00.0 3D controller: NVIDIA Corporation AD104GL [L4] (rev a1) 2025-10-10T01:24:23.2262423Z + lsmod 2025-10-10T01:24:23.2276332Z Module Size Used by 2025-10-10T01:24:23.2277037Z nvidia_uvm 1925120 0 2025-10-10T01:24:23.2277574Z nvidia 14286848 1 nvidia_uvm 2025-10-10T01:24:23.2278095Z drm 602112 1 nvidia 2025-10-10T01:24:23.2278775Z drm_panel_orientation_quirks 32768 1 drm 2025-10-10T01:24:23.2279334Z backlight 24576 1 drm 2025-10-10T01:24:23.2279838Z i2c_core 110592 2 nvidia,drm 2025-10-10T01:24:23.2280340Z xt_conntrack 16384 1 2025-10-10T01:24:23.2280792Z nft_chain_nat 16384 3 2025-10-10T01:24:23.2281228Z xt_MASQUERADE 20480 1 2025-10-10T01:24:23.2281741Z nf_nat 57344 2 nft_chain_nat,xt_MASQUERADE 2025-10-10T01:24:23.2282324Z nf_conntrack_netlink 57344 0 2025-10-10T01:24:23.2283023Z nf_conntrack 184320 4 xt_conntrack,nf_nat,nf_conntrack_netlink,xt_MASQUERADE 2025-10-10T01:24:23.2283831Z nf_defrag_ipv6 24576 1 nf_conntrack 2025-10-10T01:24:23.2284385Z nf_defrag_ipv4 16384 1 nf_conntrack 2025-10-10T01:24:23.2284886Z xfrm_user 57344 1 2025-10-10T01:24:23.2285330Z xfrm_algo 16384 1 xfrm_user 2025-10-10T01:24:23.2285828Z xt_addrtype 16384 2 2025-10-10T01:24:23.2286272Z nft_compat 20480 4 2025-10-10T01:24:23.2286682Z nf_tables 311296 57 nft_compat,nft_chain_nat 2025-10-10T01:24:23.2286994Z nfnetlink 20480 4 nft_compat,nf_conntrack_netlink,nf_tables 2025-10-10T01:24:23.2287281Z br_netfilter 36864 0 2025-10-10T01:24:23.2287482Z bridge 323584 1 br_netfilter 2025-10-10T01:24:23.2287705Z stp 16384 1 bridge 2025-10-10T01:24:23.2287915Z llc 16384 2 bridge,stp 2025-10-10T01:24:23.2288126Z overlay 167936 0 2025-10-10T01:24:23.2288309Z tls 139264 0 2025-10-10T01:24:23.2288607Z nls_ascii 16384 1 2025-10-10T01:24:23.2288787Z nls_cp437 20480 1 2025-10-10T01:24:23.2288967Z vfat 24576 1 2025-10-10T01:24:23.2289152Z fat 86016 1 vfat 2025-10-10T01:24:23.2289357Z ghash_clmulni_intel 16384 0 2025-10-10T01:24:23.2289538Z i8042 45056 0 2025-10-10T01:24:23.2289719Z ena 184320 0 2025-10-10T01:24:23.2289910Z serio 28672 3 i8042 2025-10-10T01:24:23.2290113Z button 24576 0 2025-10-10T01:24:23.2290292Z sunrpc 700416 1 2025-10-10T01:24:23.2290480Z sch_fq_codel 20480 9 2025-10-10T01:24:23.2290660Z dm_mod 188416 0 2025-10-10T01:24:23.2290838Z fuse 184320 1 2025-10-10T01:24:23.2291009Z loop 36864 0 2025-10-10T01:24:23.2291191Z configfs 57344 1 2025-10-10T01:24:23.2291371Z dmi_sysfs 20480 0 2025-10-10T01:24:23.2291551Z crc32_pclmul 16384 0 2025-10-10T01:24:23.2291736Z crc32c_intel 24576 0 2025-10-10T01:24:23.2291929Z efivarfs 24576 1 2025-10-10T01:24:23.2292123Z + modinfo nvidia 2025-10-10T01:24:23.2301353Z filename: /lib/modules/6.1.150-174.273.amzn2023.x86_64/kernel/drivers/video/nvidia.ko 2025-10-10T01:24:23.2302261Z import_ns: DMA_BUF 2025-10-10T01:24:23.2302723Z alias: char-major-195-* 2025-10-10T01:24:23.2303200Z version: 580.82.07 2025-10-10T01:24:23.2303635Z supported: external 2025-10-10T01:24:23.2304076Z license: Dual MIT/GPL 2025-10-10T01:24:23.2304584Z firmware: nvidia/580.82.07/gsp_tu10x.bin 2025-10-10T01:24:23.2305184Z firmware: nvidia/580.82.07/gsp_ga10x.bin 2025-10-10T01:24:23.2305770Z srcversion: BA7240A71DCF7DC6FE88C1D 2025-10-10T01:24:23.2306397Z alias: of:N*T*Cnvidia,tegra264-displayC* 2025-10-10T01:24:23.2306866Z alias: of:N*T*Cnvidia,tegra264-display 2025-10-10T01:24:23.2307120Z alias: of:N*T*Cnvidia,tegra234-displayC* 2025-10-10T01:24:23.2307481Z alias: of:N*T*Cnvidia,tegra234-display 2025-10-10T01:24:23.2307741Z alias: pci:v000010DEd*sv*sd*bc06sc80i00* 2025-10-10T01:24:23.2308008Z alias: pci:v000010DEd*sv*sd*bc03sc02i00* 2025-10-10T01:24:23.2308244Z alias: pci:v000010DEd*sv*sd*bc03sc00i00* 2025-10-10T01:24:23.2308476Z depends: i2c-core,drm 2025-10-10T01:24:23.2308671Z retpoline: Y 2025-10-10T01:24:23.2308836Z name: nvidia 2025-10-10T01:24:23.2309106Z vermagic: 6.1.150-174.273.amzn2023.x86_64 SMP preempt mod_unload modversions 2025-10-10T01:24:23.2309471Z parm: NvSwitchRegDwords:NvSwitch regkey (charp) 2025-10-10T01:24:23.2309810Z parm: NvSwitchBlacklist:NvSwitchBlacklist=uuid[,uuid...] (charp) 2025-10-10T01:24:23.2310124Z parm: NVreg_ResmanDebugLevel:int 2025-10-10T01:24:23.2310357Z parm: NVreg_RmLogonRC:int 2025-10-10T01:24:23.2310603Z parm: NVreg_ModifyDeviceFiles:int 2025-10-10T01:24:23.2310853Z parm: NVreg_DeviceFileUID:int 2025-10-10T01:24:23.2311079Z parm: NVreg_DeviceFileGID:int 2025-10-10T01:24:23.2311299Z parm: NVreg_DeviceFileMode:int 2025-10-10T01:24:23.2311570Z parm: NVreg_InitializeSystemMemoryAllocations:int 2025-10-10T01:24:23.2311864Z parm: NVreg_UsePageAttributeTable:int 2025-10-10T01:24:23.2312113Z parm: NVreg_EnablePCIeGen3:int 2025-10-10T01:24:23.2312332Z parm: NVreg_EnableMSI:int 2025-10-10T01:24:23.2312563Z parm: NVreg_EnableStreamMemOPs:int 2025-10-10T01:24:23.2312836Z parm: NVreg_RestrictProfilingToAdminUsers:int 2025-10-10T01:24:23.2313141Z parm: NVreg_PreserveVideoMemoryAllocations:int 2025-10-10T01:24:23.2313426Z parm: NVreg_EnableS0ixPowerManagement:int 2025-10-10T01:24:23.2313735Z parm: NVreg_S0ixPowerManagementVideoMemoryThreshold:int 2025-10-10T01:24:23.2314044Z parm: NVreg_DynamicPowerManagement:int 2025-10-10T01:24:23.2314360Z parm: NVreg_DynamicPowerManagementVideoMemoryThreshold:int 2025-10-10T01:24:23.2314750Z parm: NVreg_EnableGpuFirmware:int 2025-10-10T01:24:23.2314996Z parm: NVreg_EnableGpuFirmwareLogs:int 2025-10-10T01:24:23.2315276Z parm: NVreg_OpenRmEnableUnsupportedGpus:int 2025-10-10T01:24:23.2315568Z parm: NVreg_EnableUserNUMAManagement:int 2025-10-10T01:24:23.2315827Z parm: NVreg_MemoryPoolSize:int 2025-10-10T01:24:23.2316069Z parm: NVreg_KMallocHeapMaxSize:int 2025-10-10T01:24:23.2316309Z parm: NVreg_VMallocHeapMaxSize:int 2025-10-10T01:24:23.2316543Z parm: NVreg_IgnoreMMIOCheck:int 2025-10-10T01:24:23.2316773Z parm: NVreg_NvLinkDisable:int 2025-10-10T01:24:23.2317023Z parm: NVreg_EnablePCIERelaxedOrderingMode:int 2025-10-10T01:24:23.2317290Z parm: NVreg_RegisterPCIDriver:int 2025-10-10T01:24:23.2317548Z parm: NVreg_RegisterPlatformDeviceDriver:int 2025-10-10T01:24:23.2317824Z parm: NVreg_EnableResizableBar:int 2025-10-10T01:24:23.2318084Z parm: NVreg_EnableDbgBreakpoint:int 2025-10-10T01:24:23.2318343Z parm: NVreg_EnableNonblockingOpen:int 2025-10-10T01:24:23.2318669Z parm: NVreg_CoherentGPUMemoryMode:charp 2025-10-10T01:24:23.2318923Z parm: NVreg_RegistryDwords:charp 2025-10-10T01:24:23.2319175Z parm: NVreg_RegistryDwordsPerDevice:charp 2025-10-10T01:24:23.2319429Z parm: NVreg_RmMsg:charp 2025-10-10T01:24:23.2319652Z parm: NVreg_GpuBlacklist:charp 2025-10-10T01:24:23.2319895Z parm: NVreg_TemporaryFilePath:charp 2025-10-10T01:24:23.2320134Z parm: NVreg_ExcludedGpus:charp 2025-10-10T01:24:23.2320368Z parm: NVreg_DmaRemapPeerMmio:int 2025-10-10T01:24:23.2320607Z parm: NVreg_RmNvlinkBandwidth:charp 2025-10-10T01:24:23.2320872Z parm: NVreg_RmNvlinkBandwidthLinkCount:int 2025-10-10T01:24:23.2321149Z parm: NVreg_ImexChannelCount:int 2025-10-10T01:24:23.2321388Z parm: NVreg_CreateImexChannel0:int 2025-10-10T01:24:23.2321736Z parm: NVreg_GrdmaPciTopoCheckOverride:int 2025-10-10T01:24:23.2321994Z parm: rm_firmware_active:charp 2025-10-10T01:24:23.2322210Z + set +e 2025-10-10T01:24:23.2322355Z + nvidia-smi 2025-10-10T01:24:25.8910288Z Fri Oct 10 01:24:25 2025 2025-10-10T01:24:25.8911092Z +-----------------------------------------------------------------------------------------+ 2025-10-10T01:24:25.8912050Z | NVIDIA-SMI 580.82.07 Driver Version: 580.82.07 CUDA Version: 13.0 | 2025-10-10T01:24:25.8912922Z +-----------------------------------------+------------------------+----------------------+ 2025-10-10T01:24:25.8913807Z | GPU Name Persistence-M | Bus-Id Disp.A | Volatile Uncorr. ECC | 2025-10-10T01:24:25.8914749Z | Fan Temp Perf Pwr:Usage/Cap | Memory-Usage | GPU-Util Compute M. | 2025-10-10T01:24:25.8915502Z | | | MIG M. | 2025-10-10T01:24:25.8916131Z |=========================================+========================+======================| 2025-10-10T01:24:25.9186423Z | 0 NVIDIA L4 Off | 00000000:38:00.0 Off | 0 | 2025-10-10T01:24:25.9187298Z | N/A 38C P0 27W / 72W | 0MiB / 23034MiB | 0% Default | 2025-10-10T01:24:25.9187995Z | | | N/A | 2025-10-10T01:24:25.9188695Z +-----------------------------------------+------------------------+----------------------+ 2025-10-10T01:24:25.9189452Z | 1 NVIDIA L4 Off | 00000000:3A:00.0 Off | 0 | 2025-10-10T01:24:25.9190200Z | N/A 36C P0 27W / 72W | 0MiB / 23034MiB | 0% Default | 2025-10-10T01:24:25.9190850Z | | | N/A | 2025-10-10T01:24:25.9192197Z +-----------------------------------------+------------------------+----------------------+ 2025-10-10T01:24:25.9192985Z | 2 NVIDIA L4 Off | 00000000:3C:00.0 Off | 0 | 2025-10-10T01:24:25.9193750Z | N/A 38C P0 28W / 72W | 0MiB / 23034MiB | 0% Default | 2025-10-10T01:24:25.9194408Z | | | N/A | 2025-10-10T01:24:25.9195082Z +-----------------------------------------+------------------------+----------------------+ 2025-10-10T01:24:25.9195835Z | 3 NVIDIA L4 Off | 00000000:3E:00.0 Off | 0 | 2025-10-10T01:24:25.9196546Z | N/A 37C P0 28W / 72W | 0MiB / 23034MiB | 0% Default | 2025-10-10T01:24:25.9197175Z | | | N/A | 2025-10-10T01:24:25.9197537Z +-----------------------------------------+------------------------+----------------------+ 2025-10-10T01:24:25.9197774Z 2025-10-10T01:24:25.9197901Z +-----------------------------------------------------------------------------------------+ 2025-10-10T01:24:25.9198216Z | Processes: | 2025-10-10T01:24:25.9198619Z | GPU GI CI PID Type Process name GPU Memory | 2025-10-10T01:24:25.9198914Z | ID ID Usage | 2025-10-10T01:24:25.9199162Z |=========================================================================================| 2025-10-10T01:24:25.9210975Z | No running processes found | 2025-10-10T01:24:25.9211866Z +-----------------------------------------------------------------------------------------+ 2025-10-10T01:24:27.2188900Z + nvidia-smi --query-gpu=gpu_name --format=csv,noheader --id=0 2025-10-10T01:24:29.8279177Z NVIDIA L4 2025-10-10T01:24:30.5772918Z + NVIDIA_SMI_STATUS=0 2025-10-10T01:24:30.5773196Z + '[' 0 -eq 0 ']' 2025-10-10T01:24:30.5773402Z + echo 'INFO: Ignoring allowed status 0' 2025-10-10T01:24:30.5773637Z + set -e 2025-10-10T01:24:30.5773824Z INFO: Ignoring allowed status 0 2025-10-10T01:24:30.5782246Z == Installing nvidia container toolkit for amzn2023 == 2025-10-10T01:24:30.5786649Z + sudo yum install -y yum-utils 2025-10-10T01:24:31.0330345Z Last metadata expiration check: 0:31:27 ago on Fri Oct 10 00:53:04 2025. 2025-10-10T01:24:31.0575736Z Package dnf-utils-4.3.0-13.amzn2023.0.5.noarch is already installed. 2025-10-10T01:24:31.0997637Z Dependencies resolved. 2025-10-10T01:24:31.1239659Z Nothing to do. 2025-10-10T01:24:31.1240091Z Complete! 2025-10-10T01:24:31.2533366Z + [[ amzn2023 == \a\m\z\n\2\0\2\3 ]] 2025-10-10T01:24:31.2547638Z + YUM_REPO_URL=https://nvidia.github.io/libnvidia-container/stable/rpm/nvidia-container-toolkit.repo 2025-10-10T01:24:31.2548355Z + sudo yum-config-manager --add-repo https://nvidia.github.io/libnvidia-container/stable/rpm/nvidia-container-toolkit.repo 2025-10-10T01:24:31.6039660Z Adding repo from: https://nvidia.github.io/libnvidia-container/stable/rpm/nvidia-container-toolkit.repo 2025-10-10T01:24:31.6521541Z + sudo yum install -y nvidia-container-toolkit-1.17.8 libnvidia-container-tools-1.17.8 libnvidia-container1-1.17.8 nvidia-container-toolkit-base-1.17.8 2025-10-10T01:24:32.0950592Z nvidia-container-toolkit 18 kB/s | 833 B 00:00 2025-10-10T01:24:32.1179584Z Package nvidia-container-toolkit-1.17.8-1.x86_64 is already installed. 2025-10-10T01:24:32.1184657Z Package libnvidia-container-tools-1.17.8-1.x86_64 is already installed. 2025-10-10T01:24:32.1187427Z Package libnvidia-container1-1.17.8-1.x86_64 is already installed. 2025-10-10T01:24:32.1193006Z Package nvidia-container-toolkit-base-1.17.8-1.x86_64 is already installed. 2025-10-10T01:24:32.1637620Z Dependencies resolved. 2025-10-10T01:24:32.1865086Z Nothing to do. 2025-10-10T01:24:32.1866135Z Complete! 2025-10-10T01:24:32.4043246Z + sudo systemctl restart docker 2025-10-10T01:26:05.5041741Z Fri Oct 10 01:26:05 2025 2025-10-10T01:26:05.5043082Z +-----------------------------------------------------------------------------------------+ 2025-10-10T01:26:05.5043540Z | NVIDIA-SMI 580.82.07 Driver Version: 580.82.07 CUDA Version: 13.0 | 2025-10-10T01:26:05.5043910Z +-----------------------------------------+------------------------+----------------------+ 2025-10-10T01:26:05.5044290Z | GPU Name Persistence-M | Bus-Id Disp.A | Volatile Uncorr. ECC | 2025-10-10T01:26:05.5044682Z | Fan Temp Perf Pwr:Usage/Cap | Memory-Usage | GPU-Util Compute M. | 2025-10-10T01:26:05.5045025Z | | | MIG M. | 2025-10-10T01:26:05.5045280Z |=========================================+========================+======================| 2025-10-10T01:26:05.5327112Z | 0 NVIDIA L4 On | 00000000:38:00.0 Off | 0 | 2025-10-10T01:26:05.5327577Z | N/A 38C P0 27W / 72W | 0MiB / 23034MiB | 0% Default | 2025-10-10T01:26:05.5327927Z | | | N/A | 2025-10-10T01:26:05.5328297Z +-----------------------------------------+------------------------+----------------------+ 2025-10-10T01:26:05.5328717Z | 1 NVIDIA L4 On | 00000000:3A:00.0 Off | 0 | 2025-10-10T01:26:05.5329104Z | N/A 36C P0 27W / 72W | 0MiB / 23034MiB | 0% Default | 2025-10-10T01:26:05.5329443Z | | | N/A | 2025-10-10T01:26:05.5329794Z +-----------------------------------------+------------------------+----------------------+ 2025-10-10T01:26:05.5330190Z | 2 NVIDIA L4 On | 00000000:3C:00.0 Off | 0 | 2025-10-10T01:26:05.5331052Z | N/A 38C P0 27W / 72W | 0MiB / 23034MiB | 0% Default | 2025-10-10T01:26:05.5331415Z | | | N/A | 2025-10-10T01:26:05.5331704Z +-----------------------------------------+------------------------+----------------------+ 2025-10-10T01:26:05.5332015Z | 3 NVIDIA L4 On | 00000000:3E:00.0 Off | 0 | 2025-10-10T01:26:05.5332316Z | N/A 37C P0 29W / 72W | 0MiB / 23034MiB | 4% Default | 2025-10-10T01:26:05.5332582Z | | | N/A | 2025-10-10T01:26:05.5332863Z +-----------------------------------------+------------------------+----------------------+ 2025-10-10T01:26:05.5333532Z 2025-10-10T01:26:05.5333673Z +-----------------------------------------------------------------------------------------+ 2025-10-10T01:26:05.5334000Z | Processes: | 2025-10-10T01:26:05.5334332Z | GPU GI CI PID Type Process name GPU Memory | 2025-10-10T01:26:05.5334634Z | ID ID Usage | 2025-10-10T01:26:05.5334883Z |=========================================================================================| 2025-10-10T01:26:05.5351774Z | No running processes found | 2025-10-10T01:26:05.5352241Z +-----------------------------------------------------------------------------------------+ 2025-10-10T01:26:06.1261363Z Unable to find image 'public.ecr.aws/docker/library/python:3.13' locally 2025-10-10T01:26:06.5132948Z 3.13: Pulling from docker/library/python 2025-10-10T01:26:06.6061418Z cae3b572364a: Pulling fs layer 2025-10-10T01:26:06.6061821Z bd090f42c4b7: Pulling fs layer 2025-10-10T01:26:06.6062668Z f0c9d6d993ac: Pulling fs layer 2025-10-10T01:26:06.6062967Z a2ade626d67a: Pulling fs layer 2025-10-10T01:26:06.6063246Z 7f924d696c9c: Pulling fs layer 2025-10-10T01:26:06.6063510Z 12e6ee790ad5: Pulling fs layer 2025-10-10T01:26:06.6063736Z 54ea66483f67: Pulling fs layer 2025-10-10T01:26:06.6063963Z a2ade626d67a: Waiting 2025-10-10T01:26:06.6064176Z 7f924d696c9c: Waiting 2025-10-10T01:26:06.6064373Z 12e6ee790ad5: Waiting 2025-10-10T01:26:06.6064568Z 54ea66483f67: Waiting 2025-10-10T01:26:06.8660406Z bd090f42c4b7: Verifying Checksum 2025-10-10T01:26:06.8660796Z bd090f42c4b7: Download complete 2025-10-10T01:26:06.9504030Z cae3b572364a: Verifying Checksum 2025-10-10T01:26:06.9504370Z cae3b572364a: Download complete 2025-10-10T01:26:06.9860815Z f0c9d6d993ac: Verifying Checksum 2025-10-10T01:26:06.9861262Z f0c9d6d993ac: Download complete 2025-10-10T01:26:07.0252111Z 7f924d696c9c: Verifying Checksum 2025-10-10T01:26:07.0252483Z 7f924d696c9c: Download complete 2025-10-10T01:26:07.0957782Z 54ea66483f67: Download complete 2025-10-10T01:26:07.1068404Z 12e6ee790ad5: Verifying Checksum 2025-10-10T01:26:07.1068828Z 12e6ee790ad5: Download complete 2025-10-10T01:26:07.9814398Z a2ade626d67a: Verifying Checksum 2025-10-10T01:26:07.9814704Z a2ade626d67a: Download complete 2025-10-10T01:26:08.4229849Z cae3b572364a: Pull complete 2025-10-10T01:26:09.0289679Z bd090f42c4b7: Pull complete 2025-10-10T01:26:11.0948458Z f0c9d6d993ac: Pull complete 2025-10-10T01:26:16.3452356Z a2ade626d67a: Pull complete 2025-10-10T01:26:16.5911450Z 7f924d696c9c: Pull complete 2025-10-10T01:26:17.2025350Z 12e6ee790ad5: Pull complete 2025-10-10T01:26:17.2155480Z 54ea66483f67: Pull complete 2025-10-10T01:26:17.2219577Z Digest: sha256:4889af0e45f04b7c5dd741421a1280919499d38d3125d714b69fa86b23b1052a 2025-10-10T01:26:17.2238117Z Status: Downloaded newer image for public.ecr.aws/docker/library/python:3.13 2025-10-10T01:26:26.0937972Z Fri Oct 10 01:26:26 2025 2025-10-10T01:26:26.0939381Z +-----------------------------------------------------------------------------------------+ 2025-10-10T01:26:26.0940386Z | NVIDIA-SMI 580.82.07 Driver Version: 580.82.07 CUDA Version: 13.0 | 2025-10-10T01:26:26.0941247Z +-----------------------------------------+------------------------+----------------------+ 2025-10-10T01:26:26.0942112Z | GPU Name Persistence-M | Bus-Id Disp.A | Volatile Uncorr. ECC | 2025-10-10T01:26:26.0943097Z | Fan Temp Perf Pwr:Usage/Cap | Memory-Usage | GPU-Util Compute M. | 2025-10-10T01:26:26.0943874Z | | | MIG M. | 2025-10-10T01:26:26.0944448Z |=========================================+========================+======================| 2025-10-10T01:26:26.1385188Z | 0 NVIDIA L4 On | 00000000:38:00.0 Off | 0 | 2025-10-10T01:26:26.1386432Z | N/A 36C P8 12W / 72W | 0MiB / 23034MiB | 0% Default | 2025-10-10T01:26:26.1387201Z | | | N/A | 2025-10-10T01:26:26.1387896Z +-----------------------------------------+------------------------+----------------------+ 2025-10-10T01:26:26.1388659Z | 1 NVIDIA L4 On | 00000000:3A:00.0 Off | 0 | 2025-10-10T01:26:26.1389170Z | N/A 34C P8 12W / 72W | 0MiB / 23034MiB | 0% Default | 2025-10-10T01:26:26.1389444Z | | | N/A | 2025-10-10T01:26:26.1389736Z +-----------------------------------------+------------------------+----------------------+ 2025-10-10T01:26:26.1390050Z | 2 NVIDIA L4 On | 00000000:3C:00.0 Off | 0 | 2025-10-10T01:26:26.1390363Z | N/A 36C P8 12W / 72W | 0MiB / 23034MiB | 0% Default | 2025-10-10T01:26:26.1390638Z | | | N/A | 2025-10-10T01:26:26.1391158Z +-----------------------------------------+------------------------+----------------------+ 2025-10-10T01:26:26.1391472Z | 3 NVIDIA L4 On | 00000000:3E:00.0 Off | 0 | 2025-10-10T01:26:26.1391800Z | N/A 35C P8 12W / 72W | 0MiB / 23034MiB | 0% Default | 2025-10-10T01:26:26.1392082Z | | | N/A | 2025-10-10T01:26:26.1392378Z +-----------------------------------------+------------------------+----------------------+ 2025-10-10T01:26:26.1405413Z 2025-10-10T01:26:26.1405825Z +-----------------------------------------------------------------------------------------+ 2025-10-10T01:26:26.1406956Z | Processes: | 2025-10-10T01:26:26.1407764Z | GPU GI CI PID Type Process name GPU Memory | 2025-10-10T01:26:26.1408510Z | ID ID Usage | 2025-10-10T01:26:26.1409055Z |=========================================================================================| 2025-10-10T01:26:26.1444428Z | No running processes found | 2025-10-10T01:26:26.1445554Z +-----------------------------------------------------------------------------------------+ 2025-10-10T01:26:28.6061206Z Command completed after 1 attempt(s). 2025-10-10T01:26:28.6153474Z Prepare all required actions 2025-10-10T01:26:28.6178028Z ##[group]Run ./.github/actions/get-workflow-job-id 2025-10-10T01:26:28.6178288Z with: 2025-10-10T01:26:28.6178869Z github-token: *** 2025-10-10T01:26:28.6179069Z env: 2025-10-10T01:26:28.6179226Z GIT_DEFAULT_BRANCH: main 2025-10-10T01:26:28.6179476Z GPU_FLAG: --gpus all -e NVIDIA_DRIVER_CAPABILITIES=all 2025-10-10T01:26:28.6179737Z ##[endgroup] 2025-10-10T01:26:28.6205872Z ##[group]Run set -eux 2025-10-10T01:26:28.6206066Z set -eux 2025-10-10T01:26:28.6206387Z python3 .github/scripts/get_workflow_job_id.py "${GITHUB_RUN_ID}" "${RUNNER_NAME}" 2025-10-10T01:26:28.6219568Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2025-10-10T01:26:28.6219857Z env: 2025-10-10T01:26:28.6220019Z GIT_DEFAULT_BRANCH: main 2025-10-10T01:26:28.6220272Z GPU_FLAG: --gpus all -e NVIDIA_DRIVER_CAPABILITIES=all 2025-10-10T01:26:28.6220673Z GITHUB_TOKEN: *** 2025-10-10T01:26:28.6220853Z ##[endgroup] 2025-10-10T01:26:28.6263397Z + python3 .github/scripts/get_workflow_job_id.py 18392306035 i-0b2d5251fda19aeb9 2025-10-10T01:26:29.3499036Z Setting output job-id=52408504942 2025-10-10T01:26:29.3500873Z Setting output job-name=ci-vllm-test / test (vllm_distributed_test_28_failure_test, 1, 1, linux.g6.12xlarge.nvidia.gpu) 2025-10-10T01:26:29.3604611Z ##[group]Run python3 -m pip install psutil==5.9.8 dataclasses_json==0.6.7 nvidia-ml-py==11.525.84 2025-10-10T01:26:29.3605261Z python3 -m pip install psutil==5.9.8 dataclasses_json==0.6.7 nvidia-ml-py==11.525.84 2025-10-10T01:26:29.3605982Z python3 -m tools.stats.monitor --log-interval "$MONITOR_LOG_INTERVAL" --data-collect-interval "$MONITOR_DATA_COLLECT_INTERVAL" > usage_log.txt 2>&1 & 2025-10-10T01:26:29.3606602Z echo "monitor-script-pid=${!}" >> "${GITHUB_OUTPUT}" 2025-10-10T01:26:29.3614698Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2025-10-10T01:26:29.3614977Z env: 2025-10-10T01:26:29.3615135Z GIT_DEFAULT_BRANCH: main 2025-10-10T01:26:29.3615377Z GPU_FLAG: --gpus all -e NVIDIA_DRIVER_CAPABILITIES=all 2025-10-10T01:26:29.3615627Z JOB_ID: 52408504942 2025-10-10T01:26:29.3615985Z JOB_NAME: ci-vllm-test / test (vllm_distributed_test_28_failure_test, 1, 1, linux.g6.12xlarge.nvidia.gpu) 2025-10-10T01:26:29.3616402Z WORKFLOW_NAME: vllm-test 2025-10-10T01:26:29.3616598Z WORKFLOW_RUN_ID: 18392306035 2025-10-10T01:26:29.3616791Z MONITOR_LOG_INTERVAL: 5 2025-10-10T01:26:29.3617191Z MONITOR_DATA_COLLECT_INTERVAL: 1 2025-10-10T01:26:29.3617402Z ##[endgroup] 2025-10-10T01:26:29.6263843Z Defaulting to user installation because normal site-packages is not writeable 2025-10-10T01:26:29.9556216Z Collecting psutil==5.9.8 2025-10-10T01:26:29.9735528Z Downloading psutil-5.9.8-cp36-abi3-manylinux_2_12_x86_64.manylinux2010_x86_64.manylinux_2_17_x86_64.manylinux2014_x86_64.whl (288 kB) 2025-10-10T01:26:30.0397157Z Collecting dataclasses_json==0.6.7 2025-10-10T01:26:30.0442068Z Downloading dataclasses_json-0.6.7-py3-none-any.whl (28 kB) 2025-10-10T01:26:30.0705113Z Collecting nvidia-ml-py==11.525.84 2025-10-10T01:26:30.0778748Z Downloading nvidia_ml_py-11.525.84-py3-none-any.whl (34 kB) 2025-10-10T01:26:30.1107404Z Collecting typing-inspect<1,>=0.4.0 2025-10-10T01:26:30.1148222Z Downloading typing_inspect-0.9.0-py3-none-any.whl (8.8 kB) 2025-10-10T01:26:30.2133300Z Collecting marshmallow<4.0.0,>=3.18.0 2025-10-10T01:26:30.2187958Z Downloading marshmallow-3.26.1-py3-none-any.whl (50 kB) 2025-10-10T01:26:30.2695767Z Collecting packaging>=17.0 2025-10-10T01:26:30.2738373Z Downloading packaging-25.0-py3-none-any.whl (66 kB) 2025-10-10T01:26:30.3240468Z Collecting typing-extensions>=3.7.4 2025-10-10T01:26:30.3281911Z Downloading typing_extensions-4.15.0-py3-none-any.whl (44 kB) 2025-10-10T01:26:30.3535958Z Collecting mypy-extensions>=0.3.0 2025-10-10T01:26:30.3580291Z Downloading mypy_extensions-1.1.0-py3-none-any.whl (5.0 kB) 2025-10-10T01:26:30.4386448Z Installing collected packages: typing-extensions, packaging, mypy-extensions, typing-inspect, marshmallow, psutil, nvidia-ml-py, dataclasses-json 2025-10-10T01:26:30.6850757Z Successfully installed dataclasses-json-0.6.7 marshmallow-3.26.1 mypy-extensions-1.1.0 nvidia-ml-py-11.525.84 packaging-25.0 psutil-5.9.8 typing-extensions-4.15.0 typing-inspect-0.9.0 2025-10-10T01:26:30.8450663Z Prepare all required actions 2025-10-10T01:26:30.8451019Z Getting action download info 2025-10-10T01:26:30.9977673Z Download action repository 'seemethere/download-artifact-s3@v4' (SHA:1da556a7aa0a088e3153970611f6c432d58e80e6) 2025-10-10T01:26:31.2114085Z Download action repository 'actions/download-artifact@v4' (SHA:d3f86a106a0bac45b974a628896c90dbdf5c8093) 2025-10-10T01:26:31.6501935Z ##[group]Run ./.github/actions/download-build-artifacts 2025-10-10T01:26:31.6502234Z with: 2025-10-10T01:26:31.6502425Z name: linux-jammy-cuda12.8-py3.12-gcc11 2025-10-10T01:26:31.6502707Z s3-bucket: gha-artifacts 2025-10-10T01:26:31.6502888Z env: 2025-10-10T01:26:31.6503032Z GIT_DEFAULT_BRANCH: main 2025-10-10T01:26:31.6503272Z GPU_FLAG: --gpus all -e NVIDIA_DRIVER_CAPABILITIES=all 2025-10-10T01:26:31.6503522Z ##[endgroup] 2025-10-10T01:26:31.6554887Z ##[group]Run seemethere/download-artifact-s3@v4 2025-10-10T01:26:31.6555132Z with: 2025-10-10T01:26:31.6555334Z name: linux-jammy-cuda12.8-py3.12-gcc11 2025-10-10T01:26:31.6555579Z s3-bucket: gha-artifacts 2025-10-10T01:26:31.6555780Z region: us-east-1 2025-10-10T01:26:31.6555955Z env: 2025-10-10T01:26:31.6556104Z GIT_DEFAULT_BRANCH: main 2025-10-10T01:26:31.6556345Z GPU_FLAG: --gpus all -e NVIDIA_DRIVER_CAPABILITIES=all 2025-10-10T01:26:31.6556601Z ##[endgroup] 2025-10-10T01:26:32.0640784Z (node:67428) NOTE: We are formalizing our plans to enter AWS SDK for JavaScript (v2) into maintenance mode in 2023. 2025-10-10T01:26:32.0641736Z 2025-10-10T01:26:32.0642085Z Please migrate your code to use AWS SDK for JavaScript (v3). 2025-10-10T01:26:32.0643018Z For more information, check the migration guide at https://a.co/7PzMCcy 2025-10-10T01:26:32.0643969Z (Use `node --trace-warnings ...` to show where the warning was created) 2025-10-10T01:26:32.3044152Z Found 1 objects with prefix pytorch/pytorch/18392306035/linux-jammy-cuda12.8-py3.12-gcc11/ 2025-10-10T01:26:32.3045614Z Starting download (1/1): /home/ec2-user/actions-runner/_work/pytorch/pytorch/artifacts.zip 2025-10-10T01:26:59.7595873Z Finished download (1/1): /home/ec2-user/actions-runner/_work/pytorch/pytorch/artifacts.zip 2025-10-10T01:26:59.7600599Z Artifact download has finished successfully 2025-10-10T01:26:59.7868328Z ##[group]Run unzip -o artifacts.zip 2025-10-10T01:26:59.7868621Z unzip -o artifacts.zip 2025-10-10T01:26:59.7876553Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2025-10-10T01:26:59.7876965Z env: 2025-10-10T01:26:59.7877198Z GIT_DEFAULT_BRANCH: main 2025-10-10T01:26:59.7877450Z GPU_FLAG: --gpus all -e NVIDIA_DRIVER_CAPABILITIES=all 2025-10-10T01:26:59.7877723Z ##[endgroup] 2025-10-10T01:26:59.8909058Z Archive: artifacts.zip 2025-10-10T01:26:59.8912492Z creating: dist/ 2025-10-10T01:27:02.3559091Z inflating: dist/torch-2.10.0a0+git344e636-cp312-cp312-linux_x86_64.whl 2025-10-10T01:27:02.3559966Z creating: dist/vision/ 2025-10-10T01:27:02.3685741Z inflating: dist/vision/torchvision-0.22.0a0+966da7e-cp312-cp312-linux_x86_64.whl 2025-10-10T01:27:02.3686715Z creating: dist/audio/ 2025-10-10T01:27:02.3762603Z inflating: dist/audio/torchaudio-2.8.0a0+87ff22e-cp312-cp312-linux_x86_64.whl 2025-10-10T01:27:02.3877809Z inflating: dist/.ninja_log 2025-10-10T01:27:02.3878407Z creating: dist/external/ 2025-10-10T01:27:02.3879045Z creating: dist/external/vllm/ 2025-10-10T01:27:02.3879578Z creating: dist/external/vllm/wheels/ 2025-10-10T01:27:02.3880290Z inflating: dist/external/vllm/wheels/build_summary.txt 2025-10-10T01:27:02.3881109Z creating: dist/external/vllm/wheels/flashinfer-python/ 2025-10-10T01:27:03.0028843Z inflating: dist/external/vllm/wheels/flashinfer-python/flashinfer_python-0.2.14.post1-cp39-abi3-linux_x86_64.whl 2025-10-10T01:27:03.0030189Z creating: dist/external/vllm/wheels/vllm/ 2025-10-10T01:27:05.3002905Z inflating: dist/external/vllm/wheels/vllm/vllm-0.11.0rc2.dev157+g0ad9951c4.d20251010-cp38-abi3-linux_x86_64.whl 2025-10-10T01:27:05.3004948Z creating: dist/external/vllm/wheels/xformers/ 2025-10-10T01:27:06.8637208Z inflating: dist/external/vllm/wheels/xformers/xformers-0.0.33+5d4b92a5.d20251010-cp39-abi3-linux_x86_64.whl 2025-10-10T01:27:06.8638650Z creating: build/custom_test_artifacts/ 2025-10-10T01:27:06.8639353Z creating: build/custom_test_artifacts/custom-op-build/ 2025-10-10T01:27:06.8640202Z creating: build/custom_test_artifacts/custom-op-build/CMakeFiles/ 2025-10-10T01:27:06.8641197Z creating: build/custom_test_artifacts/custom-op-build/CMakeFiles/pkgRedirects/ 2025-10-10T01:27:06.8646117Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/CMakeConfigureLog.yaml 2025-10-10T01:27:06.8646617Z creating: build/custom_test_artifacts/custom-op-build/CMakeFiles/3.31.6/ 2025-10-10T01:27:06.8647265Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/3.31.6/CMakeSystem.cmake 2025-10-10T01:27:06.8647778Z creating: build/custom_test_artifacts/custom-op-build/CMakeFiles/3.31.6/CompilerIdC/ 2025-10-10T01:27:06.8648534Z creating: build/custom_test_artifacts/custom-op-build/CMakeFiles/3.31.6/CompilerIdC/tmp/ 2025-10-10T01:27:06.8651336Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/3.31.6/CompilerIdC/CMakeCCompilerId.c 2025-10-10T01:27:06.8652950Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/3.31.6/CompilerIdC/a.out 2025-10-10T01:27:06.8654273Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/3.31.6/CMakeCCompiler.cmake 2025-10-10T01:27:06.8655512Z creating: build/custom_test_artifacts/custom-op-build/CMakeFiles/3.31.6/CompilerIdCXX/ 2025-10-10T01:27:06.8656699Z creating: build/custom_test_artifacts/custom-op-build/CMakeFiles/3.31.6/CompilerIdCXX/tmp/ 2025-10-10T01:27:06.8658094Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/3.31.6/CompilerIdCXX/CMakeCXXCompilerId.cpp 2025-10-10T01:27:06.8659514Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/3.31.6/CompilerIdCXX/a.out 2025-10-10T01:27:06.8660814Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/3.31.6/CMakeCXXCompiler.cmake 2025-10-10T01:27:06.8662200Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/3.31.6/CMakeDetermineCompilerABI_C.bin 2025-10-10T01:27:06.8663530Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/3.31.6/CMakeDetermineCompilerABI_CXX.bin 2025-10-10T01:27:06.8664155Z creating: build/custom_test_artifacts/custom-op-build/CMakeFiles/3.31.6/CompilerIdCUDA/ 2025-10-10T01:27:06.8664695Z creating: build/custom_test_artifacts/custom-op-build/CMakeFiles/3.31.6/CompilerIdCUDA/tmp/ 2025-10-10T01:27:06.8717377Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/3.31.6/CompilerIdCUDA/tmp/CMakeCUDACompilerId.cpp4.ii 2025-10-10T01:27:06.8770044Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/3.31.6/CompilerIdCUDA/tmp/CMakeCUDACompilerId.cudafe1.cpp 2025-10-10T01:27:06.8771930Z extracting: build/custom_test_artifacts/custom-op-build/CMakeFiles/3.31.6/CompilerIdCUDA/tmp/CMakeCUDACompilerId.module_id 2025-10-10T01:27:06.8826907Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/3.31.6/CompilerIdCUDA/tmp/CMakeCUDACompilerId.cpp1.ii 2025-10-10T01:27:06.8828718Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/3.31.6/CompilerIdCUDA/tmp/CMakeCUDACompilerId.cudafe1.c 2025-10-10T01:27:06.8830456Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/3.31.6/CompilerIdCUDA/tmp/CMakeCUDACompilerId.cudafe1.gpu 2025-10-10T01:27:06.8832259Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/3.31.6/CompilerIdCUDA/tmp/CMakeCUDACompilerId.cudafe1.stub.c 2025-10-10T01:27:06.8833403Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/3.31.6/CompilerIdCUDA/tmp/CMakeCUDACompilerId.ptx 2025-10-10T01:27:06.8834129Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/3.31.6/CompilerIdCUDA/tmp/CMakeCUDACompilerId.sm_52.cubin 2025-10-10T01:27:06.8835067Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/3.31.6/CompilerIdCUDA/tmp/CMakeCUDACompilerId.fatbin 2025-10-10T01:27:06.8835824Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/3.31.6/CompilerIdCUDA/tmp/CMakeCUDACompilerId.fatbin.c 2025-10-10T01:27:06.8836522Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/3.31.6/CompilerIdCUDA/tmp/CMakeCUDACompilerId.o 2025-10-10T01:27:06.8837185Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/3.31.6/CompilerIdCUDA/tmp/a_dlink.sm_52.cubin 2025-10-10T01:27:06.8837837Z extracting: build/custom_test_artifacts/custom-op-build/CMakeFiles/3.31.6/CompilerIdCUDA/tmp/a_dlink.reg.c 2025-10-10T01:27:06.8838555Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/3.31.6/CompilerIdCUDA/tmp/a_dlink.fatbin 2025-10-10T01:27:06.8839215Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/3.31.6/CompilerIdCUDA/tmp/a_dlink.fatbin.c 2025-10-10T01:27:06.8839835Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/3.31.6/CompilerIdCUDA/tmp/a_dlink.o 2025-10-10T01:27:06.8841876Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/3.31.6/CompilerIdCUDA/CMakeCUDACompilerId.cu 2025-10-10T01:27:06.8907341Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/3.31.6/CompilerIdCUDA/a.out 2025-10-10T01:27:06.8908796Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/3.31.6/CMakeCUDACompiler.cmake 2025-10-10T01:27:06.8972348Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/3.31.6/CMakeDetermineCompilerABI_CUDA.bin 2025-10-10T01:27:06.8973724Z creating: build/custom_test_artifacts/custom-op-build/CMakeFiles/CMakeScratch/ 2025-10-10T01:27:06.8974782Z creating: build/custom_test_artifacts/custom-op-build/CMakeFiles/CMakeTmp/ 2025-10-10T01:27:06.8975847Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/cmake.check_cache 2025-10-10T01:27:06.8977015Z creating: build/custom_test_artifacts/custom-op-build/CMakeFiles/custom_ops.dir/ 2025-10-10T01:27:06.8978297Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/custom_ops.dir/compiler_depend.ts 2025-10-10T01:27:06.8980345Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/custom_ops.dir/compiler_depend.make 2025-10-10T01:27:06.8981707Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/custom_ops.dir/depend.make 2025-10-10T01:27:06.8982929Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/custom_ops.dir/link.txt 2025-10-10T01:27:06.8983480Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/custom_ops.dir/cmake_clean.cmake 2025-10-10T01:27:06.8984034Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/custom_ops.dir/build.make 2025-10-10T01:27:06.8984590Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/custom_ops.dir/DependInfo.cmake 2025-10-10T01:27:06.8985148Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/custom_ops.dir/flags.make 2025-10-10T01:27:06.8985693Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/custom_ops.dir/progress.make 2025-10-10T01:27:06.9000929Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/custom_ops.dir/op.cpp.o.d 2025-10-10T01:27:06.9174889Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/custom_ops.dir/op.cpp.o 2025-10-10T01:27:06.9176161Z creating: build/custom_test_artifacts/custom-op-build/CMakeFiles/test_custom_ops.dir/ 2025-10-10T01:27:06.9177481Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/test_custom_ops.dir/compiler_depend.ts 2025-10-10T01:27:06.9178968Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/test_custom_ops.dir/compiler_depend.make 2025-10-10T01:27:06.9180368Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/test_custom_ops.dir/depend.make 2025-10-10T01:27:06.9182086Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/test_custom_ops.dir/link.txt 2025-10-10T01:27:06.9183187Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/test_custom_ops.dir/cmake_clean.cmake 2025-10-10T01:27:06.9183777Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/test_custom_ops.dir/build.make 2025-10-10T01:27:06.9184351Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/test_custom_ops.dir/DependInfo.cmake 2025-10-10T01:27:06.9184919Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/test_custom_ops.dir/flags.make 2025-10-10T01:27:06.9185517Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/test_custom_ops.dir/progress.make 2025-10-10T01:27:06.9202026Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/test_custom_ops.dir/test_custom_ops.cpp.o.d 2025-10-10T01:27:06.9271875Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/test_custom_ops.dir/test_custom_ops.cpp.o 2025-10-10T01:27:06.9273419Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/CMakeDirectoryInformation.cmake 2025-10-10T01:27:06.9274800Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/TargetDirectories.txt 2025-10-10T01:27:06.9276013Z extracting: build/custom_test_artifacts/custom-op-build/CMakeFiles/progress.marks 2025-10-10T01:27:06.9277114Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/Makefile2 2025-10-10T01:27:06.9278173Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/Makefile.cmake 2025-10-10T01:27:06.9279351Z inflating: build/custom_test_artifacts/custom-op-build/detect_cuda_version.cc 2025-10-10T01:27:06.9280350Z inflating: build/custom_test_artifacts/custom-op-build/CMakeCache.txt 2025-10-10T01:27:06.9281260Z inflating: build/custom_test_artifacts/custom-op-build/Makefile 2025-10-10T01:27:06.9282256Z inflating: build/custom_test_artifacts/custom-op-build/cmake_install.cmake 2025-10-10T01:27:06.9431728Z inflating: build/custom_test_artifacts/custom-op-build/libcustom_ops.so 2025-10-10T01:27:06.9480291Z inflating: build/custom_test_artifacts/custom-op-build/test_custom_ops 2025-10-10T01:27:06.9481583Z creating: build/custom_test_artifacts/jit-hook-build/ 2025-10-10T01:27:06.9482334Z creating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/ 2025-10-10T01:27:06.9483150Z creating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/pkgRedirects/ 2025-10-10T01:27:06.9488789Z inflating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/CMakeConfigureLog.yaml 2025-10-10T01:27:06.9489726Z creating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/3.31.6/ 2025-10-10T01:27:06.9490651Z inflating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/3.31.6/CMakeSystem.cmake 2025-10-10T01:27:06.9491631Z creating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/3.31.6/CompilerIdC/ 2025-10-10T01:27:06.9492592Z creating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/3.31.6/CompilerIdC/tmp/ 2025-10-10T01:27:06.9493304Z inflating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/3.31.6/CompilerIdC/CMakeCCompilerId.c 2025-10-10T01:27:06.9494347Z inflating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/3.31.6/CompilerIdC/a.out 2025-10-10T01:27:06.9495326Z inflating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/3.31.6/CMakeCCompiler.cmake 2025-10-10T01:27:06.9495832Z creating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/3.31.6/CompilerIdCXX/ 2025-10-10T01:27:06.9496310Z creating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/3.31.6/CompilerIdCXX/tmp/ 2025-10-10T01:27:06.9498958Z inflating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/3.31.6/CompilerIdCXX/CMakeCXXCompilerId.cpp 2025-10-10T01:27:06.9500283Z inflating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/3.31.6/CompilerIdCXX/a.out 2025-10-10T01:27:06.9501884Z inflating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/3.31.6/CMakeCXXCompiler.cmake 2025-10-10T01:27:06.9503634Z inflating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/3.31.6/CMakeDetermineCompilerABI_C.bin 2025-10-10T01:27:06.9505526Z inflating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/3.31.6/CMakeDetermineCompilerABI_CXX.bin 2025-10-10T01:27:06.9506098Z creating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/3.31.6/CompilerIdCUDA/ 2025-10-10T01:27:06.9506614Z creating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/3.31.6/CompilerIdCUDA/tmp/ 2025-10-10T01:27:06.9559202Z inflating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/3.31.6/CompilerIdCUDA/tmp/CMakeCUDACompilerId.cpp4.ii 2025-10-10T01:27:06.9611622Z inflating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/3.31.6/CompilerIdCUDA/tmp/CMakeCUDACompilerId.cudafe1.cpp 2025-10-10T01:27:06.9613388Z extracting: build/custom_test_artifacts/jit-hook-build/CMakeFiles/3.31.6/CompilerIdCUDA/tmp/CMakeCUDACompilerId.module_id 2025-10-10T01:27:06.9668501Z inflating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/3.31.6/CompilerIdCUDA/tmp/CMakeCUDACompilerId.cpp1.ii 2025-10-10T01:27:06.9670218Z inflating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/3.31.6/CompilerIdCUDA/tmp/CMakeCUDACompilerId.cudafe1.c 2025-10-10T01:27:06.9671934Z inflating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/3.31.6/CompilerIdCUDA/tmp/CMakeCUDACompilerId.cudafe1.gpu 2025-10-10T01:27:06.9673251Z inflating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/3.31.6/CompilerIdCUDA/tmp/CMakeCUDACompilerId.cudafe1.stub.c 2025-10-10T01:27:06.9673966Z inflating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/3.31.6/CompilerIdCUDA/tmp/CMakeCUDACompilerId.ptx 2025-10-10T01:27:06.9674660Z inflating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/3.31.6/CompilerIdCUDA/tmp/CMakeCUDACompilerId.sm_52.cubin 2025-10-10T01:27:06.9675364Z inflating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/3.31.6/CompilerIdCUDA/tmp/CMakeCUDACompilerId.fatbin 2025-10-10T01:27:06.9676070Z inflating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/3.31.6/CompilerIdCUDA/tmp/CMakeCUDACompilerId.fatbin.c 2025-10-10T01:27:06.9676925Z inflating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/3.31.6/CompilerIdCUDA/tmp/CMakeCUDACompilerId.o 2025-10-10T01:27:06.9677580Z inflating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/3.31.6/CompilerIdCUDA/tmp/a_dlink.sm_52.cubin 2025-10-10T01:27:06.9678216Z extracting: build/custom_test_artifacts/jit-hook-build/CMakeFiles/3.31.6/CompilerIdCUDA/tmp/a_dlink.reg.c 2025-10-10T01:27:06.9678927Z inflating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/3.31.6/CompilerIdCUDA/tmp/a_dlink.fatbin 2025-10-10T01:27:06.9679658Z inflating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/3.31.6/CompilerIdCUDA/tmp/a_dlink.fatbin.c 2025-10-10T01:27:06.9680709Z inflating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/3.31.6/CompilerIdCUDA/tmp/a_dlink.o 2025-10-10T01:27:06.9683413Z inflating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/3.31.6/CompilerIdCUDA/CMakeCUDACompilerId.cu 2025-10-10T01:27:06.9748386Z inflating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/3.31.6/CompilerIdCUDA/a.out 2025-10-10T01:27:06.9749758Z inflating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/3.31.6/CMakeCUDACompiler.cmake 2025-10-10T01:27:06.9813841Z inflating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/3.31.6/CMakeDetermineCompilerABI_CUDA.bin 2025-10-10T01:27:06.9815183Z creating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/CMakeScratch/ 2025-10-10T01:27:06.9816200Z creating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/CMakeTmp/ 2025-10-10T01:27:06.9817240Z inflating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/cmake.check_cache 2025-10-10T01:27:06.9818344Z creating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/test_jit_hooks.dir/ 2025-10-10T01:27:06.9819991Z inflating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/test_jit_hooks.dir/compiler_depend.ts 2025-10-10T01:27:06.9821485Z inflating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/test_jit_hooks.dir/compiler_depend.make 2025-10-10T01:27:06.9822853Z inflating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/test_jit_hooks.dir/depend.make 2025-10-10T01:27:06.9823431Z inflating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/test_jit_hooks.dir/link.txt 2025-10-10T01:27:06.9824153Z inflating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/test_jit_hooks.dir/cmake_clean.cmake 2025-10-10T01:27:06.9824766Z inflating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/test_jit_hooks.dir/build.make 2025-10-10T01:27:06.9825372Z inflating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/test_jit_hooks.dir/DependInfo.cmake 2025-10-10T01:27:06.9825971Z inflating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/test_jit_hooks.dir/flags.make 2025-10-10T01:27:06.9826546Z inflating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/test_jit_hooks.dir/progress.make 2025-10-10T01:27:06.9842591Z inflating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/test_jit_hooks.dir/test_jit_hooks.cpp.o.d 2025-10-10T01:27:06.9897662Z inflating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/test_jit_hooks.dir/test_jit_hooks.cpp.o 2025-10-10T01:27:06.9899148Z inflating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/CMakeDirectoryInformation.cmake 2025-10-10T01:27:06.9900507Z inflating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/TargetDirectories.txt 2025-10-10T01:27:06.9901716Z extracting: build/custom_test_artifacts/jit-hook-build/CMakeFiles/progress.marks 2025-10-10T01:27:06.9902816Z inflating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/Makefile2 2025-10-10T01:27:06.9903375Z inflating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/Makefile.cmake 2025-10-10T01:27:06.9903843Z inflating: build/custom_test_artifacts/jit-hook-build/detect_cuda_version.cc 2025-10-10T01:27:06.9905411Z inflating: build/custom_test_artifacts/jit-hook-build/CMakeCache.txt 2025-10-10T01:27:06.9906336Z inflating: build/custom_test_artifacts/jit-hook-build/Makefile 2025-10-10T01:27:06.9907210Z inflating: build/custom_test_artifacts/jit-hook-build/cmake_install.cmake 2025-10-10T01:27:06.9942106Z inflating: build/custom_test_artifacts/jit-hook-build/test_jit_hooks 2025-10-10T01:27:06.9943078Z creating: build/custom_test_artifacts/custom-backend-build/ 2025-10-10T01:27:06.9943994Z creating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/ 2025-10-10T01:27:06.9945083Z creating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/pkgRedirects/ 2025-10-10T01:27:06.9950778Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/CMakeConfigureLog.yaml 2025-10-10T01:27:06.9952047Z creating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/3.31.6/ 2025-10-10T01:27:06.9953097Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/3.31.6/CMakeSystem.cmake 2025-10-10T01:27:06.9953648Z creating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/3.31.6/CompilerIdC/ 2025-10-10T01:27:06.9954193Z creating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/3.31.6/CompilerIdC/tmp/ 2025-10-10T01:27:06.9954793Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/3.31.6/CompilerIdC/CMakeCCompilerId.c 2025-10-10T01:27:06.9956013Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/3.31.6/CompilerIdC/a.out 2025-10-10T01:27:06.9956964Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/3.31.6/CMakeCCompiler.cmake 2025-10-10T01:27:06.9957531Z creating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/3.31.6/CompilerIdCXX/ 2025-10-10T01:27:06.9958261Z creating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/3.31.6/CompilerIdCXX/tmp/ 2025-10-10T01:27:06.9960915Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/3.31.6/CompilerIdCXX/CMakeCXXCompilerId.cpp 2025-10-10T01:27:06.9962394Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/3.31.6/CompilerIdCXX/a.out 2025-10-10T01:27:06.9963820Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/3.31.6/CMakeCXXCompiler.cmake 2025-10-10T01:27:06.9965687Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/3.31.6/CMakeDetermineCompilerABI_C.bin 2025-10-10T01:27:06.9967722Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/3.31.6/CMakeDetermineCompilerABI_CXX.bin 2025-10-10T01:27:06.9969229Z creating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/3.31.6/CompilerIdCUDA/ 2025-10-10T01:27:06.9970521Z creating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/3.31.6/CompilerIdCUDA/tmp/ 2025-10-10T01:27:07.0020479Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/3.31.6/CompilerIdCUDA/tmp/CMakeCUDACompilerId.cpp4.ii 2025-10-10T01:27:07.0073290Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/3.31.6/CompilerIdCUDA/tmp/CMakeCUDACompilerId.cudafe1.cpp 2025-10-10T01:27:07.0075208Z extracting: build/custom_test_artifacts/custom-backend-build/CMakeFiles/3.31.6/CompilerIdCUDA/tmp/CMakeCUDACompilerId.module_id 2025-10-10T01:27:07.0130181Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/3.31.6/CompilerIdCUDA/tmp/CMakeCUDACompilerId.cpp1.ii 2025-10-10T01:27:07.0132008Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/3.31.6/CompilerIdCUDA/tmp/CMakeCUDACompilerId.cudafe1.c 2025-10-10T01:27:07.0133671Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/3.31.6/CompilerIdCUDA/tmp/CMakeCUDACompilerId.cudafe1.gpu 2025-10-10T01:27:07.0135281Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/3.31.6/CompilerIdCUDA/tmp/CMakeCUDACompilerId.cudafe1.stub.c 2025-10-10T01:27:07.0136847Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/3.31.6/CompilerIdCUDA/tmp/CMakeCUDACompilerId.ptx 2025-10-10T01:27:07.0138662Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/3.31.6/CompilerIdCUDA/tmp/CMakeCUDACompilerId.sm_52.cubin 2025-10-10T01:27:07.0140173Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/3.31.6/CompilerIdCUDA/tmp/CMakeCUDACompilerId.fatbin 2025-10-10T01:27:07.0141652Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/3.31.6/CompilerIdCUDA/tmp/CMakeCUDACompilerId.fatbin.c 2025-10-10T01:27:07.0142969Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/3.31.6/CompilerIdCUDA/tmp/CMakeCUDACompilerId.o 2025-10-10T01:27:07.0143655Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/3.31.6/CompilerIdCUDA/tmp/a_dlink.sm_52.cubin 2025-10-10T01:27:07.0144331Z extracting: build/custom_test_artifacts/custom-backend-build/CMakeFiles/3.31.6/CompilerIdCUDA/tmp/a_dlink.reg.c 2025-10-10T01:27:07.0144993Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/3.31.6/CompilerIdCUDA/tmp/a_dlink.fatbin 2025-10-10T01:27:07.0145668Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/3.31.6/CompilerIdCUDA/tmp/a_dlink.fatbin.c 2025-10-10T01:27:07.0146311Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/3.31.6/CompilerIdCUDA/tmp/a_dlink.o 2025-10-10T01:27:07.0146984Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/3.31.6/CompilerIdCUDA/CMakeCUDACompilerId.cu 2025-10-10T01:27:07.0209930Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/3.31.6/CompilerIdCUDA/a.out 2025-10-10T01:27:07.0211431Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/3.31.6/CMakeCUDACompiler.cmake 2025-10-10T01:27:07.0276330Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/3.31.6/CMakeDetermineCompilerABI_CUDA.bin 2025-10-10T01:27:07.0277809Z creating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/CMakeScratch/ 2025-10-10T01:27:07.0279096Z creating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/CMakeTmp/ 2025-10-10T01:27:07.0280257Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/cmake.check_cache 2025-10-10T01:27:07.0281474Z creating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/custom_backend.dir/ 2025-10-10T01:27:07.0282841Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/custom_backend.dir/compiler_depend.ts 2025-10-10T01:27:07.0283541Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/custom_backend.dir/compiler_depend.make 2025-10-10T01:27:07.0284172Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/custom_backend.dir/depend.make 2025-10-10T01:27:07.0284779Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/custom_backend.dir/link.txt 2025-10-10T01:27:07.0285391Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/custom_backend.dir/cmake_clean.cmake 2025-10-10T01:27:07.0286017Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/custom_backend.dir/build.make 2025-10-10T01:27:07.0286633Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/custom_backend.dir/DependInfo.cmake 2025-10-10T01:27:07.0287242Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/custom_backend.dir/flags.make 2025-10-10T01:27:07.0287851Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/custom_backend.dir/progress.make 2025-10-10T01:27:07.0289877Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/custom_backend.dir/custom_backend.cpp.o.d 2025-10-10T01:27:07.0394310Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/custom_backend.dir/custom_backend.cpp.o 2025-10-10T01:27:07.0395788Z creating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/test_custom_backend.dir/ 2025-10-10T01:27:07.0397622Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/test_custom_backend.dir/compiler_depend.ts 2025-10-10T01:27:07.0399394Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/test_custom_backend.dir/compiler_depend.make 2025-10-10T01:27:07.0400972Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/test_custom_backend.dir/depend.make 2025-10-10T01:27:07.0402455Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/test_custom_backend.dir/link.txt 2025-10-10T01:27:07.0403396Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/test_custom_backend.dir/cmake_clean.cmake 2025-10-10T01:27:07.0404041Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/test_custom_backend.dir/build.make 2025-10-10T01:27:07.0404693Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/test_custom_backend.dir/DependInfo.cmake 2025-10-10T01:27:07.0405345Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/test_custom_backend.dir/flags.make 2025-10-10T01:27:07.0406004Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/test_custom_backend.dir/progress.make 2025-10-10T01:27:07.0421178Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/test_custom_backend.dir/test_custom_backend.cpp.o.d 2025-10-10T01:27:07.0468736Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/test_custom_backend.dir/test_custom_backend.cpp.o 2025-10-10T01:27:07.0470388Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/CMakeDirectoryInformation.cmake 2025-10-10T01:27:07.0471800Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/TargetDirectories.txt 2025-10-10T01:27:07.0473200Z extracting: build/custom_test_artifacts/custom-backend-build/CMakeFiles/progress.marks 2025-10-10T01:27:07.0473716Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/Makefile2 2025-10-10T01:27:07.0474213Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/Makefile.cmake 2025-10-10T01:27:07.0474689Z inflating: build/custom_test_artifacts/custom-backend-build/detect_cuda_version.cc 2025-10-10T01:27:07.0476610Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeCache.txt 2025-10-10T01:27:07.0477419Z inflating: build/custom_test_artifacts/custom-backend-build/Makefile 2025-10-10T01:27:07.0478318Z inflating: build/custom_test_artifacts/custom-backend-build/cmake_install.cmake 2025-10-10T01:27:07.0566867Z inflating: build/custom_test_artifacts/custom-backend-build/libcustom_backend.so 2025-10-10T01:27:07.0600763Z inflating: build/custom_test_artifacts/custom-backend-build/test_custom_backend 2025-10-10T01:27:07.0601635Z creating: build/lib/ 2025-10-10T01:27:07.0672967Z inflating: build/lib/libprotobuf-lite.a 2025-10-10T01:27:07.1057011Z inflating: build/lib/libprotobuf.a 2025-10-10T01:27:07.1065670Z inflating: build/lib/libpthreadpool.a 2025-10-10T01:27:07.1073005Z inflating: build/lib/libcpuinfo.a 2025-10-10T01:27:07.1080106Z inflating: build/lib/libcpuinfo_internals.a 2025-10-10T01:27:07.1080999Z inflating: build/lib/libclog.a 2025-10-10T01:27:07.1082943Z inflating: build/lib/libnnpack_reference_layers.a 2025-10-10T01:27:07.1513596Z inflating: build/lib/libprotoc.a 2025-10-10T01:27:07.1530633Z inflating: build/lib/libpytorch_qnnpack.a 2025-10-10T01:27:07.1546222Z inflating: build/lib/libnnpack.a 2025-10-10T01:27:07.1707602Z inflating: build/lib/libmicrokernels-prod.a 2025-10-10T01:27:07.2463585Z inflating: build/lib/libmicrokernels-all.a 2025-10-10T01:27:07.2529410Z inflating: build/lib/libbenchmark.a 2025-10-10T01:27:07.2530097Z inflating: build/lib/libbenchmark_main.a 2025-10-10T01:27:07.2591170Z inflating: build/lib/libgtest.a 2025-10-10T01:27:07.2606435Z inflating: build/lib/libgmock.a 2025-10-10T01:27:07.2607505Z inflating: build/lib/libgmock_main.a 2025-10-10T01:27:07.2608091Z inflating: build/lib/libgtest_main.a 2025-10-10T01:27:07.2687013Z inflating: build/lib/libXNNPACK.a 2025-10-10T01:27:07.2687675Z inflating: build/lib/libjitprofiling.a 2025-10-10T01:27:07.2695340Z inflating: build/lib/libittnotify.a 2025-10-10T01:27:07.2752685Z inflating: build/lib/libasmjit.a 2025-10-10T01:27:07.3777205Z inflating: build/lib/libfbgemm.a 2025-10-10T01:27:07.3803686Z inflating: build/lib/libtensorpipe_uv.a 2025-10-10T01:27:07.4282217Z inflating: build/lib/libtensorpipe.a 2025-10-10T01:27:07.4497019Z inflating: build/lib/libtensorpipe_cuda.a 2025-10-10T01:27:07.4613992Z inflating: build/lib/libgloo.a 2025-10-10T01:27:07.4655472Z inflating: build/lib/libonnx_proto.a 2025-10-10T01:27:07.5276781Z inflating: build/lib/libonnx.a 2025-10-10T01:27:07.5659886Z inflating: build/lib/libgloo_cuda.a 2025-10-10T01:27:07.5676781Z inflating: build/lib/libfmt.a 2025-10-10T01:27:08.4513778Z inflating: build/lib/libdnnl.a 2025-10-10T01:27:08.4925422Z inflating: build/lib/libkineto.a 2025-10-10T01:27:08.4926452Z inflating: build/lib/libtorch_global_deps.so 2025-10-10T01:27:08.5026477Z inflating: build/lib/libc10.so 2025-10-10T01:27:08.5027848Z inflating: build/lib/libcaffe2_nvrtc.so 2025-10-10T01:27:08.5079920Z inflating: build/lib/libc10_cuda.so 2025-10-10T01:27:11.1104274Z inflating: build/lib/libtorch_cpu.so 2025-10-10T01:27:11.1785567Z inflating: build/lib/libtorch_nvshmem.so 2025-10-10T01:27:11.1789612Z inflating: build/lib/libshm.so 2025-10-10T01:27:15.1826503Z inflating: build/lib/libtorch_cuda.so 2025-10-10T01:27:15.1828037Z inflating: build/lib/libtorch.so 2025-10-10T01:27:15.1831244Z inflating: build/lib/libc10d_cuda_test.so 2025-10-10T01:27:15.1876157Z inflating: build/lib/libtorch_cuda_linalg.so 2025-10-10T01:27:15.1892040Z inflating: build/lib/libjitbackend_test.so 2025-10-10T01:27:15.1953562Z inflating: build/lib/libtorchbind_test.so 2025-10-10T01:27:15.1974913Z inflating: build/lib/libbackend_with_compiler.so 2025-10-10T01:27:15.1998039Z inflating: build/lib/libaoti_custom_ops.so 2025-10-10T01:27:15.3971350Z inflating: build/lib/libtorch_python.so 2025-10-10T01:27:15.4002633Z inflating: build/lib/libnnapi_backend.so 2025-10-10T01:27:15.4003245Z creating: build/bin/ 2025-10-10T01:27:15.4397690Z inflating: build/bin/protoc-3.13.0.0 2025-10-10T01:27:15.4792018Z inflating: build/bin/protoc 2025-10-10T01:27:15.4840441Z inflating: build/bin/c10_CompileTimeFunctionPointer_test 2025-10-10T01:27:15.4889640Z inflating: build/bin/c10_DeviceGuard_test 2025-10-10T01:27:15.4946288Z inflating: build/bin/c10_DispatchKeySet_test 2025-10-10T01:27:15.4995904Z inflating: build/bin/c10_Device_test 2025-10-10T01:27:15.5050905Z inflating: build/bin/c10_SymInt_test 2025-10-10T01:27:15.5098447Z inflating: build/bin/c10_StreamGuard_test 2025-10-10T01:27:15.5152291Z inflating: build/bin/c10_SizesAndStrides_test 2025-10-10T01:27:15.5203439Z inflating: build/bin/c10_AllocatorConfig_test 2025-10-10T01:27:15.5252060Z inflating: build/bin/c10_Half_test 2025-10-10T01:27:15.5299999Z inflating: build/bin/c10_DeadlockDetection_test 2025-10-10T01:27:15.5347503Z inflating: build/bin/c10_ConstexprCrc_test 2025-10-10T01:27:15.5399608Z inflating: build/bin/c10_InlineDeviceGuard_test 2025-10-10T01:27:15.5450513Z inflating: build/bin/c10_Bitset_test 2025-10-10T01:27:15.5499177Z inflating: build/bin/c10_generic_math_test 2025-10-10T01:27:15.5547484Z inflating: build/bin/c10_Synchronized_test 2025-10-10T01:27:15.5597135Z inflating: build/bin/c10_TypeIndex_test 2025-10-10T01:27:15.5650833Z inflating: build/bin/c10_LeftRight_test 2025-10-10T01:27:15.5721453Z inflating: build/bin/c10_optional_test 2025-10-10T01:27:15.5772689Z inflating: build/bin/c10_NetworkFlow_test 2025-10-10T01:27:15.5827261Z inflating: build/bin/c10_Enumerate_test 2025-10-10T01:27:15.5880236Z inflating: build/bin/c10_complex_test 2025-10-10T01:27:15.5934262Z inflating: build/bin/c10_ThreadLocal_test 2025-10-10T01:27:15.5981986Z inflating: build/bin/c10_error_test 2025-10-10T01:27:15.6031741Z inflating: build/bin/c10_accumulate_test 2025-10-10T01:27:15.6082911Z inflating: build/bin/c10_lazy_test 2025-10-10T01:27:15.6141988Z inflating: build/bin/c10_ordered_preserving_dict_test 2025-10-10T01:27:15.6189838Z inflating: build/bin/c10_Semaphore_test 2025-10-10T01:27:15.6244448Z inflating: build/bin/c10_logging_test 2025-10-10T01:27:15.6292193Z inflating: build/bin/c10_ArrayRef_test 2025-10-10T01:27:15.6345485Z inflating: build/bin/c10_Metaprogramming_test 2025-10-10T01:27:15.6395735Z inflating: build/bin/c10_exception_test 2025-10-10T01:27:15.6446797Z inflating: build/bin/c10_IntrusiveList_test 2025-10-10T01:27:15.6512745Z inflating: build/bin/c10_cow_test 2025-10-10T01:27:15.6560440Z inflating: build/bin/c10_TypeTraits_test 2025-10-10T01:27:15.6609319Z inflating: build/bin/c10_irange_test 2025-10-10T01:27:15.6662797Z inflating: build/bin/c10_bfloat16_test 2025-10-10T01:27:15.6714324Z inflating: build/bin/c10_Scalar_test 2025-10-10T01:27:15.6768641Z inflating: build/bin/c10_string_util_test 2025-10-10T01:27:15.6816976Z inflating: build/bin/c10_flags_test 2025-10-10T01:27:15.6868232Z inflating: build/bin/c10_registry_test 2025-10-10T01:27:15.6917700Z inflating: build/bin/c10_ssize_test 2025-10-10T01:27:15.6971340Z inflating: build/bin/c10_InlineStreamGuard_test 2025-10-10T01:27:15.7018613Z inflating: build/bin/c10_string_view_test 2025-10-10T01:27:15.7067046Z inflating: build/bin/c10_tempfile_test 2025-10-10T01:27:15.7116149Z inflating: build/bin/c10_TypeList_test 2025-10-10T01:27:15.7158983Z inflating: build/bin/c10_intrusive_ptr_benchmark 2025-10-10T01:27:15.7212706Z inflating: build/bin/c10_complex_math_test 2025-10-10T01:27:15.7261315Z inflating: build/bin/c10_bit_cast_test 2025-10-10T01:27:15.7315149Z inflating: build/bin/c10_typeid_test 2025-10-10T01:27:15.7466888Z inflating: build/bin/c10_intrusive_ptr_test 2025-10-10T01:27:15.7608173Z inflating: build/bin/c10_small_vector_test 2025-10-10T01:27:15.7656040Z inflating: build/bin/c10_cuda_CUDATest 2025-10-10T01:27:15.8199752Z inflating: build/bin/vec_test_all_types_AVX2 2025-10-10T01:27:15.8722560Z inflating: build/bin/vec_test_all_types_DEFAULT 2025-10-10T01:27:15.9260834Z inflating: build/bin/vec_test_all_types_AVX512 2025-10-10T01:27:15.9310992Z inflating: build/bin/c10_cuda_CUDAAssertionsTest_from_2_processes 2025-10-10T01:27:15.9362178Z inflating: build/bin/c10_cuda_CUDAAssertionsTest_catches_stream 2025-10-10T01:27:15.9412922Z inflating: build/bin/c10_cuda_CUDAAssertionsTest_multiple_writes_from_same_block 2025-10-10T01:27:15.9463912Z inflating: build/bin/c10_cuda_CUDAAssertionsTest_multiple_writes_from_multiple_blocks 2025-10-10T01:27:15.9514587Z inflating: build/bin/c10_cuda_CUDAAssertionsTest_catches_thread_and_block_and_device 2025-10-10T01:27:15.9565287Z inflating: build/bin/c10_cuda_CUDAAssertionsTest_multiple_writes_from_blocks_and_threads 2025-10-10T01:27:15.9616235Z inflating: build/bin/c10_cuda_CUDAAssertionsTest_1_var_test 2025-10-10T01:27:15.9670881Z inflating: build/bin/TCPStoreTest 2025-10-10T01:27:15.9721997Z inflating: build/bin/FileStoreTest 2025-10-10T01:27:15.9773774Z inflating: build/bin/HashStoreTest 2025-10-10T01:27:15.9824685Z inflating: build/bin/BackoffTest 2025-10-10T01:27:15.9836909Z inflating: build/bin/ProcessGroupMPITest 2025-10-10T01:27:15.9840422Z inflating: build/bin/torch_shm_manager 2025-10-10T01:27:15.9843502Z inflating: build/bin/example_allreduce 2025-10-10T01:27:15.9901683Z inflating: build/bin/ProcessGroupNCCLErrorsTest 2025-10-10T01:27:15.9956099Z inflating: build/bin/ProcessGroupGlooAsyncTest 2025-10-10T01:27:16.0016582Z inflating: build/bin/ProcessGroupNCCLTest 2025-10-10T01:27:16.0065793Z inflating: build/bin/dlconvertor_test 2025-10-10T01:27:16.0120350Z inflating: build/bin/extension_backend_test 2025-10-10T01:27:16.0183206Z inflating: build/bin/ProcessGroupGlooTest 2025-10-10T01:27:16.0253236Z inflating: build/bin/Dict_test 2025-10-10T01:27:16.0306131Z inflating: build/bin/broadcast_test 2025-10-10T01:27:16.0363531Z inflating: build/bin/atest 2025-10-10T01:27:16.0416051Z inflating: build/bin/half_test 2025-10-10T01:27:16.0466629Z inflating: build/bin/Dimname_test 2025-10-10T01:27:16.0521061Z inflating: build/bin/NamedTensor_test 2025-10-10T01:27:16.0583261Z inflating: build/bin/MaybeOwned_test 2025-10-10T01:27:16.0631492Z inflating: build/bin/lazy_tensor_test 2025-10-10T01:27:16.0683888Z inflating: build/bin/test_parallel 2025-10-10T01:27:16.0735163Z inflating: build/bin/cpu_profiling_allocator_test 2025-10-10T01:27:16.0796561Z inflating: build/bin/basic 2025-10-10T01:27:16.0845787Z inflating: build/bin/operators_test 2025-10-10T01:27:16.0932190Z inflating: build/bin/cpu_rng_test 2025-10-10T01:27:16.1043165Z inflating: build/bin/kernel_function_legacy_test 2025-10-10T01:27:16.1093184Z inflating: build/bin/stride_properties_test 2025-10-10T01:27:16.1143045Z inflating: build/bin/packedtensoraccessor_test 2025-10-10T01:27:16.1198371Z inflating: build/bin/cpu_generator_test 2025-10-10T01:27:16.1272562Z inflating: build/bin/tensor_iterator_test 2025-10-10T01:27:16.1326868Z inflating: build/bin/scalar_tensor_test 2025-10-10T01:27:16.1378662Z inflating: build/bin/math_kernel_test 2025-10-10T01:27:16.1434745Z inflating: build/bin/scalar_test 2025-10-10T01:27:16.1482911Z inflating: build/bin/reduce_ops_test 2025-10-10T01:27:16.1581593Z inflating: build/bin/List_test 2025-10-10T01:27:16.1646091Z inflating: build/bin/pow_test 2025-10-10T01:27:16.1693418Z inflating: build/bin/verify_api_visibility 2025-10-10T01:27:16.1741319Z inflating: build/bin/cuda_exchange_device_test 2025-10-10T01:27:16.1807479Z inflating: build/bin/legacy_vmap_test 2025-10-10T01:27:16.1857091Z inflating: build/bin/weakref_test 2025-10-10T01:27:16.1909858Z inflating: build/bin/type_ptr_test 2025-10-10T01:27:16.1960713Z inflating: build/bin/cuda_apply_test 2025-10-10T01:27:16.2010104Z inflating: build/bin/StorageUtils_test 2025-10-10T01:27:16.2073454Z inflating: build/bin/inline_container_test 2025-10-10T01:27:16.2122621Z inflating: build/bin/reportMemoryUsage_test 2025-10-10T01:27:16.2178915Z inflating: build/bin/apply_utils_test 2025-10-10T01:27:16.2236088Z inflating: build/bin/cuda_complex_test 2025-10-10T01:27:16.2286660Z inflating: build/bin/undefined_tensor_test 2025-10-10T01:27:16.2566086Z inflating: build/bin/op_registration_test 2025-10-10T01:27:16.2614790Z inflating: build/bin/operator_name_test 2025-10-10T01:27:16.2667043Z inflating: build/bin/memory_format_test 2025-10-10T01:27:16.2756650Z inflating: build/bin/ivalue_test 2025-10-10T01:27:16.2805833Z inflating: build/bin/CppSignature_test 2025-10-10T01:27:16.2857630Z inflating: build/bin/memory_overlapping_test 2025-10-10T01:27:16.2911143Z inflating: build/bin/native_test 2025-10-10T01:27:16.2966070Z inflating: build/bin/quantized_test 2025-10-10T01:27:16.3029083Z inflating: build/bin/KernelFunction_test 2025-10-10T01:27:16.3076316Z inflating: build/bin/op_allowlist_test 2025-10-10T01:27:16.3144780Z inflating: build/bin/cuda_complex_math_test 2025-10-10T01:27:16.3203441Z inflating: build/bin/cuda_atomic_ops_test 2025-10-10T01:27:16.3261209Z inflating: build/bin/kernel_stackbased_test 2025-10-10T01:27:16.3314227Z inflating: build/bin/backend_fallback_test 2025-10-10T01:27:16.3363586Z inflating: build/bin/cpu_allocator_test 2025-10-10T01:27:16.3417326Z inflating: build/bin/cuda_caching_host_allocator_test 2025-10-10T01:27:16.3473950Z inflating: build/bin/IListRef_test 2025-10-10T01:27:16.3536604Z inflating: build/bin/cuda_cub_test 2025-10-10T01:27:16.3593094Z inflating: build/bin/type_test 2025-10-10T01:27:16.3642946Z inflating: build/bin/xla_tensor_test 2025-10-10T01:27:16.3692780Z inflating: build/bin/wrapdim_test 2025-10-10T01:27:16.3751249Z inflating: build/bin/cuda_stream_test 2025-10-10T01:27:16.3799745Z inflating: build/bin/cuda_half_test 2025-10-10T01:27:16.3879840Z inflating: build/bin/cuda_distributions_test 2025-10-10T01:27:16.3928270Z inflating: build/bin/cuda_allocatorTraceTracker_test 2025-10-10T01:27:16.3977818Z inflating: build/bin/cuda_integer_divider_test 2025-10-10T01:27:16.4026113Z inflating: build/bin/cuda_optional_test 2025-10-10T01:27:16.4076719Z inflating: build/bin/cuda_packedtensoraccessor_test 2025-10-10T01:27:16.4125567Z inflating: build/bin/thread_init_test 2025-10-10T01:27:16.4173426Z inflating: build/bin/cuda_device_test 2025-10-10T01:27:16.4224992Z inflating: build/bin/mobile_memory_cleanup 2025-10-10T01:27:16.4274610Z inflating: build/bin/cuda_dlconvertor_test 2025-10-10T01:27:16.4363157Z inflating: build/bin/make_boxed_from_unboxed_functor_test 2025-10-10T01:27:16.4479183Z inflating: build/bin/kernel_lambda_legacy_test 2025-10-10T01:27:16.4530927Z inflating: build/bin/cuda_allocator_test 2025-10-10T01:27:16.4595366Z inflating: build/bin/test_cpp_rpc 2025-10-10T01:27:16.4683803Z inflating: build/bin/kernel_function_test 2025-10-10T01:27:16.4736671Z inflating: build/bin/test_dist_autograd 2025-10-10T01:27:16.4788025Z inflating: build/bin/cuda_vectorized_test 2025-10-10T01:27:16.4838812Z inflating: build/bin/cuda_reportMemoryUsage_test 2025-10-10T01:27:16.4840783Z inflating: build/bin/parallel_benchmark 2025-10-10T01:27:16.4935639Z inflating: build/bin/kernel_lambda_test 2025-10-10T01:27:16.5245705Z inflating: build/bin/test_lazy 2025-10-10T01:27:16.5294112Z inflating: build/bin/cuda_cudnn_test 2025-10-10T01:27:16.5348683Z inflating: build/bin/cuda_generator_test 2025-10-10T01:27:16.6382523Z inflating: build/bin/test_api 2025-10-10T01:27:16.7359395Z inflating: build/bin/test_jit 2025-10-10T01:27:16.7360045Z creating: .additional_ci_files/ 2025-10-10T01:27:16.7420945Z inflating: .additional_ci_files/test-times.json 2025-10-10T01:27:16.7651082Z inflating: .additional_ci_files/test-class-times.json 2025-10-10T01:27:16.7703539Z ##[group]Run rm artifacts.zip 2025-10-10T01:27:16.7703792Z rm artifacts.zip 2025-10-10T01:27:16.7711721Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2025-10-10T01:27:16.7725575Z env: 2025-10-10T01:27:16.7725760Z GIT_DEFAULT_BRANCH: main 2025-10-10T01:27:16.7726041Z GPU_FLAG: --gpus all -e NVIDIA_DRIVER_CAPABILITIES=all 2025-10-10T01:27:16.7726318Z ##[endgroup] 2025-10-10T01:27:16.9614293Z ##[group]Run df -H 2025-10-10T01:27:16.9614495Z df -H 2025-10-10T01:27:16.9622410Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2025-10-10T01:27:16.9622689Z env: 2025-10-10T01:27:16.9622859Z GIT_DEFAULT_BRANCH: main 2025-10-10T01:27:16.9623111Z GPU_FLAG: --gpus all -e NVIDIA_DRIVER_CAPABILITIES=all 2025-10-10T01:27:16.9623371Z ##[endgroup] 2025-10-10T01:27:16.9686112Z Filesystem Size Used Avail Use% Mounted on 2025-10-10T01:27:16.9686529Z devtmpfs 4.2M 0 4.2M 0% /dev 2025-10-10T01:27:16.9686851Z tmpfs 98G 0 98G 0% /dev/shm 2025-10-10T01:27:16.9687169Z tmpfs 40G 975k 40G 1% /run 2025-10-10T01:27:16.9687465Z /dev/nvme0n1p1 161G 56G 106G 35% / 2025-10-10T01:27:16.9687772Z tmpfs 98G 13k 98G 1% /tmp 2025-10-10T01:27:16.9688076Z /dev/nvme0n1p128 11M 1.4M 9.2M 13% /boot/efi 2025-10-10T01:27:16.9688405Z tmpfs 20G 0 20G 0% /run/user/0 2025-10-10T01:27:16.9722533Z Prepare all required actions 2025-10-10T01:27:16.9723069Z Getting action download info 2025-10-10T01:27:17.1515015Z ##[group]Run ./.github/actions/download-td-artifacts 2025-10-10T01:27:17.1515278Z with: 2025-10-10T01:27:17.1515640Z env: 2025-10-10T01:27:17.1515806Z GIT_DEFAULT_BRANCH: main 2025-10-10T01:27:17.1516046Z GPU_FLAG: --gpus all -e NVIDIA_DRIVER_CAPABILITIES=all 2025-10-10T01:27:17.1516302Z ##[endgroup] 2025-10-10T01:27:17.1550981Z ##[group]Run seemethere/download-artifact-s3@v4 2025-10-10T01:27:17.1551233Z with: 2025-10-10T01:27:17.1551384Z name: td_results 2025-10-10T01:27:17.1551570Z s3-bucket: gha-artifacts 2025-10-10T01:27:17.1551766Z region: us-east-1 2025-10-10T01:27:17.1551935Z env: 2025-10-10T01:27:17.1552082Z GIT_DEFAULT_BRANCH: main 2025-10-10T01:27:17.1552324Z GPU_FLAG: --gpus all -e NVIDIA_DRIVER_CAPABILITIES=all 2025-10-10T01:27:17.1552591Z ##[endgroup] 2025-10-10T01:27:17.5657548Z (node:67457) NOTE: We are formalizing our plans to enter AWS SDK for JavaScript (v2) into maintenance mode in 2023. 2025-10-10T01:27:17.5658436Z 2025-10-10T01:27:17.5658795Z Please migrate your code to use AWS SDK for JavaScript (v3). 2025-10-10T01:27:17.5659726Z For more information, check the migration guide at https://a.co/7PzMCcy 2025-10-10T01:27:17.5660705Z (Use `node --trace-warnings ...` to show where the warning was created) 2025-10-10T01:27:17.6637163Z Found 0 objects with prefix pytorch/pytorch/18392306035/td_results/ 2025-10-10T01:27:17.6642824Z Artifact download has finished successfully 2025-10-10T01:27:17.6903061Z ##[group]Run mkdir -p .additional_ci_files 2025-10-10T01:27:17.6903350Z mkdir -p .additional_ci_files 2025-10-10T01:27:17.6903671Z mv td_results.json .additional_ci_files/td_results.json || true 2025-10-10T01:27:17.6913008Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2025-10-10T01:27:17.6913299Z env: 2025-10-10T01:27:17.6913475Z GIT_DEFAULT_BRANCH: main 2025-10-10T01:27:17.6913750Z GPU_FLAG: --gpus all -e NVIDIA_DRIVER_CAPABILITIES=all 2025-10-10T01:27:17.6914011Z ##[endgroup] 2025-10-10T01:27:17.6985879Z mv: cannot stat 'td_results.json': No such file or directory 2025-10-10T01:27:17.7030270Z ##[group]Run .github/scripts/parse_ref.py 2025-10-10T01:27:17.7030608Z .github/scripts/parse_ref.py 2025-10-10T01:27:17.7037748Z shell: /usr/bin/bash -e {0} 2025-10-10T01:27:17.7037955Z env: 2025-10-10T01:27:17.7038123Z GIT_DEFAULT_BRANCH: main 2025-10-10T01:27:17.7038371Z GPU_FLAG: --gpus all -e NVIDIA_DRIVER_CAPABILITIES=all 2025-10-10T01:27:17.7038766Z ##[endgroup] 2025-10-10T01:27:17.7244034Z Setting output branch=main 2025-10-10T01:27:17.7350903Z Prepare all required actions 2025-10-10T01:27:17.7351233Z Getting action download info 2025-10-10T01:27:17.8514305Z ##[group]Run ./.github/actions/filter-test-configs 2025-10-10T01:27:17.8514587Z with: 2025-10-10T01:27:17.8515007Z github-token: *** 2025-10-10T01:27:17.8519861Z test-matrix: {"include": [{"config": "vllm_basic_correctness_test", "shard": 1, "num_shards": 1, "runner": "linux.g6.4xlarge.experimental.nvidia.gpu"}, {"config": "vllm_basic_models_test", "shard": 1, "num_shards": 1, "runner": "linux.g6.4xlarge.experimental.nvidia.gpu"}, {"config": "vllm_entrypoints_test", "shard": 1, "num_shards": 1, "runner": "linux.g6.4xlarge.experimental.nvidia.gpu"}, {"config": "vllm_regression_test", "shard": 1, "num_shards": 1, "runner": "linux.g6.4xlarge.experimental.nvidia.gpu"}, {"config": "vllm_multi_model_processor_test", "shard": 1, "num_shards": 1, "runner": "linux.g6.4xlarge.experimental.nvidia.gpu"}, {"config": "vllm_pytorch_compilation_unit_tests", "shard": 1, "num_shards": 1, "runner": "linux.g6.4xlarge.experimental.nvidia.gpu"}, {"config": "vllm_lora_28_failure_test", "shard": 1, "num_shards": 1, "runner": "linux.g6.4xlarge.experimental.nvidia.gpu"}, {"config": "vllm_multi_model_test_28_failure_test", "shard": 1, "num_shards": 1, "runner": "linux.g6.4xlarge.experimental.nvidia.gpu"}, {"config": "vllm_languagde_model_test_extended_generation_28_failure_test", "shard": 1, "num_shards": 1, "runner": "linux.g6.4xlarge.experimental.nvidia.gpu"}, {"config": "vllm_distributed_test_2_gpu_28_failure_test", "shard": 1, "num_shards": 1, "runner": "linux.g6.4xlarge.experimental.nvidia.gpu"}, {"config": "vllm_lora_test", "shard": 0, "num_shards": 4, "runner": "linux.g6.4xlarge.experimental.nvidia.gpu"}, {"config": "vllm_lora_test", "shard": 1, "num_shards": 4, "runner": "linux.g6.4xlarge.experimental.nvidia.gpu"}, {"config": "vllm_lora_test", "shard": 2, "num_shards": 4, "runner": "linux.g6.4xlarge.experimental.nvidia.gpu"}, {"config": "vllm_lora_test", "shard": 3, "num_shards": 4, "runner": "linux.g6.4xlarge.experimental.nvidia.gpu"}, {"config": "vllm_lora_tp_test_distributed", "shard": 1, "num_shards": 1, "runner": "linux.g6.12xlarge.nvidia.gpu"}, {"config": "vllm_distributed_test_28_failure_test", "shard": 1, "num_shards": 1, "runner": "linux.g6.12xlarge.nvidia.gpu"}]} 2025-10-10T01:27:17.8525353Z job-name: ci-vllm-test / test (vllm_distributed_test_28_failure_test, 1, 1, linux.g6.12xlarge.nvidia.gpu) 2025-10-10T01:27:17.8525772Z env: 2025-10-10T01:27:17.8525932Z GIT_DEFAULT_BRANCH: main 2025-10-10T01:27:17.8526185Z GPU_FLAG: --gpus all -e NVIDIA_DRIVER_CAPABILITIES=all 2025-10-10T01:27:17.8526454Z ##[endgroup] 2025-10-10T01:27:17.8600484Z ##[group]Run nick-fields/retry@v3.0.0 2025-10-10T01:27:17.8600756Z with: 2025-10-10T01:27:17.8600960Z shell: bash 2025-10-10T01:27:17.8601173Z timeout_minutes: 10 2025-10-10T01:27:17.8601391Z max_attempts: 5 2025-10-10T01:27:17.8601613Z retry_wait_seconds: 30 2025-10-10T01:27:17.8602217Z command: set -eux # PyYAML 6.0 doesn't work with MacOS x86 anymore # This must run on Python-3.7 (AmazonLinux2) so can't use request=3.32.2 python3 -m pip install requests==2.27.1 pyyaml==6.0.2 2025-10-10T01:27:17.8602836Z polling_interval_seconds: 1 2025-10-10T01:27:17.8603076Z warning_on_retry: true 2025-10-10T01:27:17.8603302Z continue_on_error: false 2025-10-10T01:27:17.8603533Z env: 2025-10-10T01:27:17.8603730Z GIT_DEFAULT_BRANCH: main 2025-10-10T01:27:17.8604002Z GPU_FLAG: --gpus all -e NVIDIA_DRIVER_CAPABILITIES=all 2025-10-10T01:27:17.8604442Z GITHUB_TOKEN: *** 2025-10-10T01:27:17.8604655Z ##[endgroup] 2025-10-10T01:27:17.9614229Z + python3 -m pip install requests==2.27.1 pyyaml==6.0.2 2025-10-10T01:27:18.1771348Z Defaulting to user installation because normal site-packages is not writeable 2025-10-10T01:27:18.2916948Z Collecting requests==2.27.1 2025-10-10T01:27:18.3105276Z Downloading requests-2.27.1-py2.py3-none-any.whl (63 kB) 2025-10-10T01:27:18.4752220Z Collecting pyyaml==6.0.2 2025-10-10T01:27:18.4819699Z Downloading PyYAML-6.0.2-cp39-cp39-manylinux_2_17_x86_64.manylinux2014_x86_64.whl (737 kB) 2025-10-10T01:27:18.8202205Z Collecting charset-normalizer~=2.0.0 2025-10-10T01:27:18.8264558Z Downloading charset_normalizer-2.0.12-py3-none-any.whl (39 kB) 2025-10-10T01:27:18.8328100Z Requirement already satisfied: idna<4,>=2.5 in /usr/lib/python3.9/site-packages (from requests==2.27.1) (2.10) 2025-10-10T01:27:18.8831182Z Collecting certifi>=2017.4.17 2025-10-10T01:27:18.8883240Z Downloading certifi-2025.10.5-py3-none-any.whl (163 kB) 2025-10-10T01:27:18.8946060Z Requirement already satisfied: urllib3<1.27,>=1.21.1 in /usr/lib/python3.9/site-packages (from requests==2.27.1) (1.25.10) 2025-10-10T01:27:18.9668316Z Installing collected packages: charset-normalizer, certifi, requests, pyyaml 2025-10-10T01:27:19.0788630Z Successfully installed certifi-2025.10.5 charset-normalizer-2.0.12 pyyaml-6.0.2 requests-2.27.1 2025-10-10T01:27:19.9309782Z Command completed after 1 attempt(s). 2025-10-10T01:27:19.9390265Z ##[group]Run set -x 2025-10-10T01:27:19.9390581Z set -x 2025-10-10T01:27:19.9390748Z  2025-10-10T01:27:19.9391025Z # Use relative path here as this could be checked out anywhere, not necessarily 2025-10-10T01:27:19.9391379Z # in runner workspace 2025-10-10T01:27:19.9391661Z python3 "${GITHUB_ACTION_PATH}/../../scripts/parse_ref.py" 2025-10-10T01:27:19.9400781Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2025-10-10T01:27:19.9401079Z env: 2025-10-10T01:27:19.9401236Z GIT_DEFAULT_BRANCH: main 2025-10-10T01:27:19.9401492Z GPU_FLAG: --gpus all -e NVIDIA_DRIVER_CAPABILITIES=all 2025-10-10T01:27:19.9401936Z ##[endgroup] 2025-10-10T01:27:19.9440372Z + python3 /home/ec2-user/actions-runner/_work/pytorch/pytorch/./.github/actions/filter-test-configs/../../scripts/parse_ref.py 2025-10-10T01:27:19.9611157Z Setting output branch=main 2025-10-10T01:27:19.9672509Z ##[group]Run echo "Workflow: ${GITHUB_WORKFLOW}" 2025-10-10T01:27:19.9672857Z echo "Workflow: ${GITHUB_WORKFLOW}" 2025-10-10T01:27:19.9673111Z echo "Job name: ${JOB_NAME}" 2025-10-10T01:27:19.9673327Z  2025-10-10T01:27:19.9673601Z # Use relative path here as this could be checked out anywhere, not necessarily 2025-10-10T01:27:19.9673954Z # in runner workspace 2025-10-10T01:27:19.9674264Z python3 "${GITHUB_ACTION_PATH}/../../scripts/filter_test_configs.py" \ 2025-10-10T01:27:19.9674625Z  --workflow "${GITHUB_WORKFLOW}" \ 2025-10-10T01:27:19.9674862Z  --job-name "${JOB_NAME}" \ 2025-10-10T01:27:19.9680124Z  --test-matrix "{"include": [{"config": "vllm_basic_correctness_test", "shard": 1, "num_shards": 1, "runner": "linux.g6.4xlarge.experimental.nvidia.gpu"}, {"config": "vllm_basic_models_test", "shard": 1, "num_shards": 1, "runner": "linux.g6.4xlarge.experimental.nvidia.gpu"}, {"config": "vllm_entrypoints_test", "shard": 1, "num_shards": 1, "runner": "linux.g6.4xlarge.experimental.nvidia.gpu"}, {"config": "vllm_regression_test", "shard": 1, "num_shards": 1, "runner": "linux.g6.4xlarge.experimental.nvidia.gpu"}, {"config": "vllm_multi_model_processor_test", "shard": 1, "num_shards": 1, "runner": "linux.g6.4xlarge.experimental.nvidia.gpu"}, {"config": "vllm_pytorch_compilation_unit_tests", "shard": 1, "num_shards": 1, "runner": "linux.g6.4xlarge.experimental.nvidia.gpu"}, {"config": "vllm_lora_28_failure_test", "shard": 1, "num_shards": 1, "runner": "linux.g6.4xlarge.experimental.nvidia.gpu"}, {"config": "vllm_multi_model_test_28_failure_test", "shard": 1, "num_shards": 1, "runner": "linux.g6.4xlarge.experimental.nvidia.gpu"}, {"config": "vllm_languagde_model_test_extended_generation_28_failure_test", "shard": 1, "num_shards": 1, "runner": "linux.g6.4xlarge.experimental.nvidia.gpu"}, {"config": "vllm_distributed_test_2_gpu_28_failure_test", "shard": 1, "num_shards": 1, "runner": "linux.g6.4xlarge.experimental.nvidia.gpu"}, {"config": "vllm_lora_test", "shard": 0, "num_shards": 4, "runner": "linux.g6.4xlarge.experimental.nvidia.gpu"}, {"config": "vllm_lora_test", "shard": 1, "num_shards": 4, "runner": "linux.g6.4xlarge.experimental.nvidia.gpu"}, {"config": "vllm_lora_test", "shard": 2, "num_shards": 4, "runner": "linux.g6.4xlarge.experimental.nvidia.gpu"}, {"config": "vllm_lora_test", "shard": 3, "num_shards": 4, "runner": "linux.g6.4xlarge.experimental.nvidia.gpu"}, {"config": "vllm_lora_tp_test_distributed", "shard": 1, "num_shards": 1, "runner": "linux.g6.12xlarge.nvidia.gpu"}, {"config": "vllm_distributed_test_28_failure_test", "shard": 1, "num_shards": 1, "runner": "linux.g6.12xlarge.nvidia.gpu"}]}" \ 2025-10-10T01:27:19.9685338Z  --selected-test-configs "" \ 2025-10-10T01:27:19.9685586Z  --pr-number "${PR_NUMBER}" \ 2025-10-10T01:27:19.9685807Z  --tag "${TAG}" \ 2025-10-10T01:27:19.9686011Z  --event-name "${EVENT_NAME}" \ 2025-10-10T01:27:19.9686239Z  --schedule "${SCHEDULE}" \ 2025-10-10T01:27:19.9686458Z  --branch "${HEAD_BRANCH}" 2025-10-10T01:27:19.9694448Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2025-10-10T01:27:19.9694728Z env: 2025-10-10T01:27:19.9694886Z GIT_DEFAULT_BRANCH: main 2025-10-10T01:27:19.9695126Z GPU_FLAG: --gpus all -e NVIDIA_DRIVER_CAPABILITIES=all 2025-10-10T01:27:19.9695776Z GITHUB_TOKEN: *** 2025-10-10T01:27:19.9696160Z JOB_NAME: ci-vllm-test / test (vllm_distributed_test_28_failure_test, 1, 1, linux.g6.12xlarge.nvidia.gpu) 2025-10-10T01:27:19.9696566Z PR_NUMBER: 2025-10-10T01:27:19.9696721Z TAG: 2025-10-10T01:27:19.9696872Z EVENT_NAME: push 2025-10-10T01:27:19.9697035Z SCHEDULE: 2025-10-10T01:27:19.9697189Z HEAD_BRANCH: main 2025-10-10T01:27:19.9697542Z ##[endgroup] 2025-10-10T01:27:19.9729899Z Workflow: vllm-test 2025-10-10T01:27:19.9730385Z Job name: ci-vllm-test / test (vllm_distributed_test_28_failure_test, 1, 1, linux.g6.12xlarge.nvidia.gpu) 2025-10-10T01:27:20.1511815Z Setting output keep-going=True 2025-10-10T01:27:20.1512452Z Setting output ci-verbose-test-logs=False 2025-10-10T01:27:20.1513088Z Setting output ci-test-showlocals=False 2025-10-10T01:27:20.1513671Z Setting output ci-no-test-timeout=False 2025-10-10T01:27:20.1514210Z Setting output ci-no-td=False 2025-10-10T01:27:20.1514726Z Setting output ci-td-distributed=False 2025-10-10T01:27:20.1515281Z Setting output is-unstable=False 2025-10-10T01:27:20.1515805Z Setting output reenabled-issues= 2025-10-10T01:27:20.1521787Z Setting output test-matrix={"include": [{"config": "vllm_basic_correctness_test", "shard": 1, "num_shards": 1, "runner": "linux.g6.4xlarge.experimental.nvidia.gpu"}, {"config": "vllm_basic_models_test", "shard": 1, "num_shards": 1, "runner": "linux.g6.4xlarge.experimental.nvidia.gpu"}, {"config": "vllm_entrypoints_test", "shard": 1, "num_shards": 1, "runner": "linux.g6.4xlarge.experimental.nvidia.gpu"}, {"config": "vllm_regression_test", "shard": 1, "num_shards": 1, "runner": "linux.g6.4xlarge.experimental.nvidia.gpu"}, {"config": "vllm_multi_model_processor_test", "shard": 1, "num_shards": 1, "runner": "linux.g6.4xlarge.experimental.nvidia.gpu"}, {"config": "vllm_pytorch_compilation_unit_tests", "shard": 1, "num_shards": 1, "runner": "linux.g6.4xlarge.experimental.nvidia.gpu"}, {"config": "vllm_lora_28_failure_test", "shard": 1, "num_shards": 1, "runner": "linux.g6.4xlarge.experimental.nvidia.gpu"}, {"config": "vllm_multi_model_test_28_failure_test", "shard": 1, "num_shards": 1, "runner": "linux.g6.4xlarge.experimental.nvidia.gpu"}, {"config": "vllm_languagde_model_test_extended_generation_28_failure_test", "shard": 1, "num_shards": 1, "runner": "linux.g6.4xlarge.experimental.nvidia.gpu"}, {"config": "vllm_distributed_test_2_gpu_28_failure_test", "shard": 1, "num_shards": 1, "runner": "linux.g6.4xlarge.experimental.nvidia.gpu"}, {"config": "vllm_lora_test", "shard": 0, "num_shards": 4, "runner": "linux.g6.4xlarge.experimental.nvidia.gpu"}, {"config": "vllm_lora_test", "shard": 1, "num_shards": 4, "runner": "linux.g6.4xlarge.experimental.nvidia.gpu"}, {"config": "vllm_lora_test", "shard": 2, "num_shards": 4, "runner": "linux.g6.4xlarge.experimental.nvidia.gpu"}, {"config": "vllm_lora_test", "shard": 3, "num_shards": 4, "runner": "linux.g6.4xlarge.experimental.nvidia.gpu"}, {"config": "vllm_lora_tp_test_distributed", "shard": 1, "num_shards": 1, "runner": "linux.g6.12xlarge.nvidia.gpu"}, {"config": "vllm_distributed_test_28_failure_test", "shard": 1, "num_shards": 1, "runner": "linux.g6.12xlarge.nvidia.gpu"}]} 2025-10-10T01:27:20.1527058Z Setting output is-test-matrix-empty=False 2025-10-10T01:27:20.1635243Z ##[group]Run echo "Filtered matrix:" 2025-10-10T01:27:20.1635523Z echo "Filtered matrix:" 2025-10-10T01:27:20.1640422Z echo "{"include": [{"config": "vllm_basic_correctness_test", "shard": 1, "num_shards": 1, "runner": "linux.g6.4xlarge.experimental.nvidia.gpu"}, {"config": "vllm_basic_models_test", "shard": 1, "num_shards": 1, "runner": "linux.g6.4xlarge.experimental.nvidia.gpu"}, {"config": "vllm_entrypoints_test", "shard": 1, "num_shards": 1, "runner": "linux.g6.4xlarge.experimental.nvidia.gpu"}, {"config": "vllm_regression_test", "shard": 1, "num_shards": 1, "runner": "linux.g6.4xlarge.experimental.nvidia.gpu"}, {"config": "vllm_multi_model_processor_test", "shard": 1, "num_shards": 1, "runner": "linux.g6.4xlarge.experimental.nvidia.gpu"}, {"config": "vllm_pytorch_compilation_unit_tests", "shard": 1, "num_shards": 1, "runner": "linux.g6.4xlarge.experimental.nvidia.gpu"}, {"config": "vllm_lora_28_failure_test", "shard": 1, "num_shards": 1, "runner": "linux.g6.4xlarge.experimental.nvidia.gpu"}, {"config": "vllm_multi_model_test_28_failure_test", "shard": 1, "num_shards": 1, "runner": "linux.g6.4xlarge.experimental.nvidia.gpu"}, {"config": "vllm_languagde_model_test_extended_generation_28_failure_test", "shard": 1, "num_shards": 1, "runner": "linux.g6.4xlarge.experimental.nvidia.gpu"}, {"config": "vllm_distributed_test_2_gpu_28_failure_test", "shard": 1, "num_shards": 1, "runner": "linux.g6.4xlarge.experimental.nvidia.gpu"}, {"config": "vllm_lora_test", "shard": 0, "num_shards": 4, "runner": "linux.g6.4xlarge.experimental.nvidia.gpu"}, {"config": "vllm_lora_test", "shard": 1, "num_shards": 4, "runner": "linux.g6.4xlarge.experimental.nvidia.gpu"}, {"config": "vllm_lora_test", "shard": 2, "num_shards": 4, "runner": "linux.g6.4xlarge.experimental.nvidia.gpu"}, {"config": "vllm_lora_test", "shard": 3, "num_shards": 4, "runner": "linux.g6.4xlarge.experimental.nvidia.gpu"}, {"config": "vllm_lora_tp_test_distributed", "shard": 1, "num_shards": 1, "runner": "linux.g6.12xlarge.nvidia.gpu"}, {"config": "vllm_distributed_test_28_failure_test", "shard": 1, "num_shards": 1, "runner": "linux.g6.12xlarge.nvidia.gpu"}]}" 2025-10-10T01:27:20.1645551Z  2025-10-10T01:27:20.1645708Z echo 2025-10-10T01:27:20.1645905Z echo "Is the current job unstable? False" 2025-10-10T01:27:20.1646139Z  2025-10-10T01:27:20.1646276Z echo 2025-10-10T01:27:20.1646475Z echo "Is keep-going label set? True" 2025-10-10T01:27:20.1646706Z  2025-10-10T01:27:20.1646855Z echo 2025-10-10T01:27:20.1647016Z echo "Reenabled issues? " 2025-10-10T01:27:20.1654634Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2025-10-10T01:27:20.1654914Z env: 2025-10-10T01:27:20.1655080Z GIT_DEFAULT_BRANCH: main 2025-10-10T01:27:20.1655331Z GPU_FLAG: --gpus all -e NVIDIA_DRIVER_CAPABILITIES=all 2025-10-10T01:27:20.1655589Z ##[endgroup] 2025-10-10T01:27:20.1686323Z Filtered matrix: 2025-10-10T01:27:20.1693159Z {include: [{config: vllm_basic_correctness_test, shard: 1, num_shards: 1, runner: linux.g6.4xlarge.experimental.nvidia.gpu}, {config: vllm_basic_models_test, shard: 1, num_shards: 1, runner: linux.g6.4xlarge.experimental.nvidia.gpu}, {config: vllm_entrypoints_test, shard: 1, num_shards: 1, runner: linux.g6.4xlarge.experimental.nvidia.gpu}, {config: vllm_regression_test, shard: 1, num_shards: 1, runner: linux.g6.4xlarge.experimental.nvidia.gpu}, {config: vllm_multi_model_processor_test, shard: 1, num_shards: 1, runner: linux.g6.4xlarge.experimental.nvidia.gpu}, {config: vllm_pytorch_compilation_unit_tests, shard: 1, num_shards: 1, runner: linux.g6.4xlarge.experimental.nvidia.gpu}, {config: vllm_lora_28_failure_test, shard: 1, num_shards: 1, runner: linux.g6.4xlarge.experimental.nvidia.gpu}, {config: vllm_multi_model_test_28_failure_test, shard: 1, num_shards: 1, runner: linux.g6.4xlarge.experimental.nvidia.gpu}, {config: vllm_languagde_model_test_extended_generation_28_failure_test, shard: 1, num_shards: 1, runner: linux.g6.4xlarge.experimental.nvidia.gpu}, {config: vllm_distributed_test_2_gpu_28_failure_test, shard: 1, num_shards: 1, runner: linux.g6.4xlarge.experimental.nvidia.gpu}, {config: vllm_lora_test, shard: 0, num_shards: 4, runner: linux.g6.4xlarge.experimental.nvidia.gpu}, {config: vllm_lora_test, shard: 1, num_shards: 4, runner: linux.g6.4xlarge.experimental.nvidia.gpu}, {config: vllm_lora_test, shard: 2, num_shards: 4, runner: linux.g6.4xlarge.experimental.nvidia.gpu}, {config: vllm_lora_test, shard: 3, num_shards: 4, runner: linux.g6.4xlarge.experimental.nvidia.gpu}, {config: vllm_lora_tp_test_distributed, shard: 1, num_shards: 1, runner: linux.g6.12xlarge.nvidia.gpu}, {config: vllm_distributed_test_28_failure_test, shard: 1, num_shards: 1, runner: linux.g6.12xlarge.nvidia.gpu}]} 2025-10-10T01:27:20.1698651Z 2025-10-10T01:27:20.1698736Z Is the current job unstable? False 2025-10-10T01:27:20.1698894Z 2025-10-10T01:27:20.1698973Z Is keep-going label set? True 2025-10-10T01:27:20.1699116Z 2025-10-10T01:27:20.1699185Z Reenabled issues? 2025-10-10T01:27:20.1745101Z ##[group]Run echo "timeout=$((JOB_TIMEOUT-30))" >> "${GITHUB_OUTPUT}" 2025-10-10T01:27:20.1745528Z echo "timeout=$((JOB_TIMEOUT-30))" >> "${GITHUB_OUTPUT}" 2025-10-10T01:27:20.1752994Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2025-10-10T01:27:20.1753277Z env: 2025-10-10T01:27:20.1753436Z GIT_DEFAULT_BRANCH: main 2025-10-10T01:27:20.1753683Z GPU_FLAG: --gpus all -e NVIDIA_DRIVER_CAPABILITIES=all 2025-10-10T01:27:20.1753932Z JOB_TIMEOUT: 240 2025-10-10T01:27:20.1754103Z ##[endgroup] 2025-10-10T01:27:20.1823687Z ##[group]Run env | grep '^GITHUB' >> "/tmp/github_env_${GITHUB_RUN_ID}" 2025-10-10T01:27:20.1824455Z env | grep '^GITHUB' >> "/tmp/github_env_${GITHUB_RUN_ID}" 2025-10-10T01:27:20.1824801Z env | grep '^CI' >> "/tmp/github_env_${GITHUB_RUN_ID}" 2025-10-10T01:27:20.1832060Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2025-10-10T01:27:20.1832348Z env: 2025-10-10T01:27:20.1832512Z GIT_DEFAULT_BRANCH: main 2025-10-10T01:27:20.1832766Z GPU_FLAG: --gpus all -e NVIDIA_DRIVER_CAPABILITIES=all 2025-10-10T01:27:20.1833029Z ##[endgroup] 2025-10-10T01:27:20.2005948Z ##[group]Run set -x 2025-10-10T01:27:20.2006245Z set -x 2025-10-10T01:27:20.2006428Z  2025-10-10T01:27:20.2006629Z if [[ $TEST_CONFIG == 'multigpu' ]]; then 2025-10-10T01:27:20.2006937Z  TEST_COMMAND=.ci/pytorch/multigpu-test.sh 2025-10-10T01:27:20.2007242Z elif [[ $BUILD_ENVIRONMENT == *onnx* ]]; then 2025-10-10T01:27:20.2007507Z  TEST_COMMAND=.ci/onnx/test.sh 2025-10-10T01:27:20.2007725Z else 2025-10-10T01:27:20.2007916Z  TEST_COMMAND=.ci/pytorch/test.sh 2025-10-10T01:27:20.2008147Z fi 2025-10-10T01:27:20.2008292Z  2025-10-10T01:27:20.2008475Z # Leaving 1GB for the runner and other things 2025-10-10T01:27:20.2008909Z TOTAL_AVAILABLE_MEMORY_IN_GB=$(awk '/MemTotal/ { printf "%.3f \n", $2/1024/1024 - 1 }' /proc/meminfo) 2025-10-10T01:27:20.2009540Z # https://docs.docker.com/engine/containers/resource_constraints/#--memory-swap-details, the 3GB swap 2025-10-10T01:27:20.2010074Z # comes from https://github.com/pytorch/test-infra/pull/6058 2025-10-10T01:27:20.2010463Z TOTAL_MEMORY_WITH_SWAP=$(("${TOTAL_AVAILABLE_MEMORY_IN_GB%.*}" + 3)) 2025-10-10T01:27:20.2010759Z  2025-10-10T01:27:20.2010955Z if [[ ${BUILD_ENVIRONMENT} == *"s390x"* ]]; then 2025-10-10T01:27:20.2011205Z  SHM_OPTS= 2025-10-10T01:27:20.2011383Z  JENKINS_USER= 2025-10-10T01:27:20.2011652Z  # ensure that docker container cleanly exits in 12 hours 2025-10-10T01:27:20.2011998Z  # if for some reason cleanup action doesn't stop container 2025-10-10T01:27:20.2012281Z  # when job is cancelled 2025-10-10T01:27:20.2012509Z  DOCKER_SHELL_CMD="sleep 12h" 2025-10-10T01:27:20.2012719Z else 2025-10-10T01:27:20.2012905Z  SHM_OPTS="--shm-size=${SHM_SIZE}" 2025-10-10T01:27:20.2013146Z  JENKINS_USER="--user jenkins" 2025-10-10T01:27:20.2013372Z  DOCKER_SHELL_CMD= 2025-10-10T01:27:20.2013558Z fi 2025-10-10T01:27:20.2013718Z  2025-10-10T01:27:20.2013961Z # detached container should get cleaned up by teardown_ec2_linux 2025-10-10T01:27:20.2014340Z # TODO: Stop building test binaries as part of the build phase 2025-10-10T01:27:20.2014765Z # Used for GPU_FLAG, SHM_OPTS, JENKINS_USER and DOCKER_SHELL_CMD since that doesn't play nice 2025-10-10T01:27:20.2015146Z # shellcheck disable=SC2086,SC2090 2025-10-10T01:27:20.2015386Z container_name=$(docker run \ 2025-10-10T01:27:20.2015614Z  ${GPU_FLAG:-} \ 2025-10-10T01:27:20.2015840Z  ${SCCACHE_SERVER_PORT_DOCKER_FLAG:-} \ 2025-10-10T01:27:20.2016084Z  -e BUILD_ENVIRONMENT \ 2025-10-10T01:27:20.2016313Z  -e PR_NUMBER \ 2025-10-10T01:27:20.2016513Z  -e GITHUB_ACTIONS \ 2025-10-10T01:27:20.2016733Z  -e GITHUB_REPOSITORY \ 2025-10-10T01:27:20.2016956Z  -e GITHUB_WORKFLOW \ 2025-10-10T01:27:20.2017342Z  -e GITHUB_JOB \ 2025-10-10T01:27:20.2017537Z  -e GITHUB_RUN_ID \ 2025-10-10T01:27:20.2017743Z  -e GITHUB_RUN_NUMBER \ 2025-10-10T01:27:20.2017947Z  -e GITHUB_RUN_ATTEMPT \ 2025-10-10T01:27:20.2018155Z  -e JOB_ID \ 2025-10-10T01:27:20.2018345Z  -e JOB_NAME \ 2025-10-10T01:27:20.2018529Z  -e BASE_SHA \ 2025-10-10T01:27:20.2018698Z  -e BRANCH \ 2025-10-10T01:27:20.2018877Z  -e SHA1 \ 2025-10-10T01:27:20.2019061Z  -e AWS_DEFAULT_REGION \ 2025-10-10T01:27:20.2019271Z  -e IN_WHEEL_TEST \ 2025-10-10T01:27:20.2019473Z  -e SHARD_NUMBER \ 2025-10-10T01:27:20.2019672Z  -e TEST_CONFIG \ 2025-10-10T01:27:20.2019868Z  -e NUM_TEST_SHARDS \ 2025-10-10T01:27:20.2020074Z  -e REENABLED_ISSUES \ 2025-10-10T01:27:20.2020284Z  -e CONTINUE_THROUGH_ERROR \ 2025-10-10T01:27:20.2020619Z  -e VERBOSE_TEST_LOGS \ 2025-10-10T01:27:20.2020840Z  -e TEST_SHOWLOCALS \ 2025-10-10T01:27:20.2021042Z  -e NO_TEST_TIMEOUT \ 2025-10-10T01:27:20.2021232Z  -e NO_TD \ 2025-10-10T01:27:20.2021414Z  -e TD_DISTRIBUTED \ 2025-10-10T01:27:20.2021629Z  -e PR_LABELS \ 2025-10-10T01:27:20.2021850Z  -e MAX_JOBS="$(nproc --ignore=2)" \ 2025-10-10T01:27:20.2022085Z  -e SCCACHE_BUCKET \ 2025-10-10T01:27:20.2022287Z  -e SCCACHE_REGION \ 2025-10-10T01:27:20.2022483Z  -e XLA_CUDA \ 2025-10-10T01:27:20.2022694Z  -e XLA_CLANG_CACHE_S3_BUCKET_NAME \ 2025-10-10T01:27:20.2022953Z  -e PYTORCH_TEST_CUDA_MEM_LEAK_CHECK \ 2025-10-10T01:27:20.2023220Z  -e PYTORCH_TEST_RERUN_DISABLED_TESTS \ 2025-10-10T01:27:20.2023488Z  -e SKIP_SCCACHE_INITIALIZATION=1 \ 2025-10-10T01:27:20.2023733Z  -e HUGGING_FACE_HUB_TOKEN \ 2025-10-10T01:27:20.2024157Z  -e VLLM_TEST_HUGGING_FACE_TOKEN \ 2025-10-10T01:27:20.2024443Z  -e SCRIBE_GRAPHQL_ACCESS_TOKEN \ 2025-10-10T01:27:20.2024695Z  -e DASHBOARD_TAG \ 2025-10-10T01:27:20.2024910Z  -e ARTIFACTS_FILE_SUFFIX \ 2025-10-10T01:27:20.2025180Z  --memory="${TOTAL_AVAILABLE_MEMORY_IN_GB%.*}g" \ 2025-10-10T01:27:20.2025482Z  --memory-swap="${TOTAL_MEMORY_WITH_SWAP}g" \ 2025-10-10T01:27:20.2025786Z  --env-file="/tmp/github_env_${GITHUB_RUN_ID}" \ 2025-10-10T01:27:20.2026075Z  --security-opt seccomp=unconfined \ 2025-10-10T01:27:20.2026323Z  --cap-add=SYS_PTRACE \ 2025-10-10T01:27:20.2026539Z  --ipc=host \ 2025-10-10T01:27:20.2026723Z  ${SHM_OPTS} \ 2025-10-10T01:27:20.2026919Z  --tty \ 2025-10-10T01:27:20.2027095Z  --detach \ 2025-10-10T01:27:20.2027296Z  --name="${container_name}" \ 2025-10-10T01:27:20.2027517Z  ${JENKINS_USER} \ 2025-10-10T01:27:20.2027775Z  -v "${GITHUB_WORKSPACE}:/var/lib/jenkins/workspace" \ 2025-10-10T01:27:20.2028069Z  -w /var/lib/jenkins/workspace \ 2025-10-10T01:27:20.2028297Z  "${DOCKER_IMAGE}" \ 2025-10-10T01:27:20.2028488Z  ${DOCKER_SHELL_CMD} 2025-10-10T01:27:20.2028671Z ) 2025-10-10T01:27:20.2028905Z echo "DOCKER_CONTAINER_ID=${container_name}" >> "${GITHUB_ENV}" 2025-10-10T01:27:20.2029207Z  2025-10-10T01:27:20.2029392Z if [[ ${BUILD_ENVIRONMENT} == *"s390x"* ]]; then 2025-10-10T01:27:20.2029823Z  docker exec -t "${container_name}" sh -c "python3 -m pip install -r .ci/docker/requirements-ci.txt" 2025-10-10T01:27:20.2030198Z fi 2025-10-10T01:27:20.2030348Z  2025-10-10T01:27:20.2030698Z docker exec -t "${container_name}" sh -c "python3 -m pip install $(echo dist/*.whl)[opt-einsum] && ${TEST_COMMAND}" 2025-10-10T01:27:20.2037836Z shell: /usr/bin/bash -e {0} 2025-10-10T01:27:20.2038040Z env: 2025-10-10T01:27:20.2038204Z GIT_DEFAULT_BRANCH: main 2025-10-10T01:27:20.2038686Z GPU_FLAG: --gpus all -e NVIDIA_DRIVER_CAPABILITIES=all 2025-10-10T01:27:20.2039006Z BUILD_ENVIRONMENT: linux-jammy-cuda12.8-py3.12-gcc11 2025-10-10T01:27:20.2039268Z PR_NUMBER: 2025-10-10T01:27:20.2039446Z GITHUB_REPOSITORY: pytorch/pytorch 2025-10-10T01:27:20.2039679Z GITHUB_WORKFLOW: vllm-test 2025-10-10T01:27:20.2039871Z GITHUB_JOB: test 2025-10-10T01:27:20.2040041Z GITHUB_RUN_ID: 18392306035 2025-10-10T01:27:20.2040228Z GITHUB_RUN_NUMBER: 2144 2025-10-10T01:27:20.2040416Z GITHUB_RUN_ATTEMPT: 1 2025-10-10T01:27:20.2040582Z JOB_ID: 52408504942 2025-10-10T01:27:20.2040933Z JOB_NAME: ci-vllm-test / test (vllm_distributed_test_28_failure_test, 1, 1, linux.g6.12xlarge.nvidia.gpu) 2025-10-10T01:27:20.2041319Z BRANCH: main 2025-10-10T01:27:20.2041505Z SHA1: 344e6365a0068c2d2847fcec0c55dd53291d475e 2025-10-10T01:27:20.2041768Z BASE_SHA: 344e6365a0068c2d2847fcec0c55dd53291d475e 2025-10-10T01:27:20.2042193Z TEST_CONFIG: vllm_distributed_test_28_failure_test 2025-10-10T01:27:20.2042449Z SHARD_NUMBER: 1 2025-10-10T01:27:20.2042618Z NUM_TEST_SHARDS: 1 2025-10-10T01:27:20.2042779Z EXTRA_FLAGS: 2025-10-10T01:27:20.2042938Z OP_BENCHMARK_TESTS: 2025-10-10T01:27:20.2043126Z REENABLED_ISSUES: 2025-10-10T01:27:20.2043305Z CONTINUE_THROUGH_ERROR: True 2025-10-10T01:27:20.2043508Z VERBOSE_TEST_LOGS: False 2025-10-10T01:27:20.2043698Z TEST_SHOWLOCALS: False 2025-10-10T01:27:20.2043896Z NO_TEST_TIMEOUT: False 2025-10-10T01:27:20.2044074Z NO_TD: False 2025-10-10T01:27:20.2044231Z TD_DISTRIBUTED: False 2025-10-10T01:27:20.2044456Z SCCACHE_BUCKET: ossci-compiler-cache-circleci-v2 2025-10-10T01:27:20.2044716Z SCCACHE_REGION: us-east-1 2025-10-10T01:27:20.2044904Z SHM_SIZE: 2g 2025-10-10T01:27:20.2045483Z DOCKER_IMAGE: 308535385114.dkr.ecr.us-east-1.amazonaws.com/pytorch/ci-image:pytorch-linux-jammy-cuda12.8-cudnn9-py3.12-gcc11-vllm-d8be0384e085f551506bd739678109fa0f5ee7ac 2025-10-10T01:27:20.2046107Z XLA_CUDA: 2025-10-10T01:27:20.2046368Z XLA_CLANG_CACHE_S3_BUCKET_NAME: ossci-compiler-clang-cache-circleci-xla 2025-10-10T01:27:20.2046697Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK: 0 2025-10-10T01:27:20.2046929Z PYTORCH_TEST_RERUN_DISABLED_TESTS: 0 2025-10-10T01:27:20.2047139Z DASHBOARD_TAG: 2025-10-10T01:27:20.2047475Z VLLM_TEST_HUGGING_FACE_TOKEN: *** 2025-10-10T01:27:20.2047779Z HUGGING_FACE_HUB_TOKEN: *** 2025-10-10T01:27:20.2048072Z SCRIBE_GRAPHQL_ACCESS_TOKEN: *** 2025-10-10T01:27:20.2048497Z ARTIFACTS_FILE_SUFFIX: test-vllm_distributed_test_28_failure_test-1-1-linux.g6.12xlarge.nvidia.gpu_52408504942 2025-10-10T01:27:20.2048907Z ##[endgroup] 2025-10-10T01:27:20.2079299Z + [[ vllm_distributed_test_28_failure_test == \m\u\l\t\i\g\p\u ]] 2025-10-10T01:27:20.2080122Z + [[ linux-jammy-cuda12.8-py3.12-gcc11 == *onnx* ]] 2025-10-10T01:27:20.2080744Z + TEST_COMMAND=.ci/pytorch/test.sh 2025-10-10T01:27:20.2085473Z ++ awk '/MemTotal/ { printf "%.3f \n", $2/1024/1024 - 1 }' /proc/meminfo 2025-10-10T01:27:20.2110711Z + TOTAL_AVAILABLE_MEMORY_IN_GB='180.760 ' 2025-10-10T01:27:20.2111375Z + TOTAL_MEMORY_WITH_SWAP=183 2025-10-10T01:27:20.2111974Z + [[ linux-jammy-cuda12.8-py3.12-gcc11 == *\s\3\9\0\x* ]] 2025-10-10T01:27:20.2112627Z + SHM_OPTS=--shm-size=2g 2025-10-10T01:27:20.2113086Z + JENKINS_USER='--user jenkins' 2025-10-10T01:27:20.2113547Z + DOCKER_SHELL_CMD= 2025-10-10T01:27:20.2123366Z +++ nproc --ignore=2 2025-10-10T01:27:20.2154174Z ++ docker run --gpus all -e NVIDIA_DRIVER_CAPABILITIES=all -e BUILD_ENVIRONMENT -e PR_NUMBER -e GITHUB_ACTIONS -e GITHUB_REPOSITORY -e GITHUB_WORKFLOW -e GITHUB_JOB -e GITHUB_RUN_ID -e GITHUB_RUN_NUMBER -e GITHUB_RUN_ATTEMPT -e JOB_ID -e JOB_NAME -e BASE_SHA -e BRANCH -e SHA1 -e AWS_DEFAULT_REGION -e IN_WHEEL_TEST -e SHARD_NUMBER -e TEST_CONFIG -e NUM_TEST_SHARDS -e REENABLED_ISSUES -e CONTINUE_THROUGH_ERROR -e VERBOSE_TEST_LOGS -e TEST_SHOWLOCALS -e NO_TEST_TIMEOUT -e NO_TD -e TD_DISTRIBUTED -e PR_LABELS -e MAX_JOBS=46 -e SCCACHE_BUCKET -e SCCACHE_REGION -e XLA_CUDA -e XLA_CLANG_CACHE_S3_BUCKET_NAME -e PYTORCH_TEST_CUDA_MEM_LEAK_CHECK -e PYTORCH_TEST_RERUN_DISABLED_TESTS -e SKIP_SCCACHE_INITIALIZATION=1 -e HUGGING_FACE_HUB_TOKEN -e VLLM_TEST_HUGGING_FACE_TOKEN -e SCRIBE_GRAPHQL_ACCESS_TOKEN -e DASHBOARD_TAG -e ARTIFACTS_FILE_SUFFIX --memory=180g --memory-swap=183g --env-file=/tmp/github_env_18392306035 --security-opt seccomp=unconfined --cap-add=SYS_PTRACE --ipc=host --shm-size=2g --tty --detach --name= --user jenkins -v /home/ec2-user/actions-runner/_work/pytorch/pytorch:/var/lib/jenkins/workspace -w /var/lib/jenkins/workspace 308535385114.dkr.ecr.us-east-1.amazonaws.com/pytorch/ci-image:pytorch-linux-jammy-cuda12.8-cudnn9-py3.12-gcc11-vllm-d8be0384e085f551506bd739678109fa0f5ee7ac 2025-10-10T01:27:41.3746955Z + container_name=92bc90bdf047b31a0d3a4858f15294a82b80fed2ff3cd44e76b438de5ef8fa55 2025-10-10T01:27:41.3750211Z + echo DOCKER_CONTAINER_ID=92bc90bdf047b31a0d3a4858f15294a82b80fed2ff3cd44e76b438de5ef8fa55 2025-10-10T01:27:41.3751112Z + [[ linux-jammy-cuda12.8-py3.12-gcc11 == *\s\3\9\0\x* ]] 2025-10-10T01:27:41.3754672Z ++ echo dist/torch-2.10.0a0+git344e636-cp312-cp312-linux_x86_64.whl 2025-10-10T01:27:41.3757539Z + docker exec -t 92bc90bdf047b31a0d3a4858f15294a82b80fed2ff3cd44e76b438de5ef8fa55 sh -c 'python3 -m pip install dist/torch-2.10.0a0+git344e636-cp312-cp312-linux_x86_64.whl[opt-einsum] && .ci/pytorch/test.sh' 2025-10-10T01:27:41.8513942Z Processing ./dist/torch-2.10.0a0+git344e636-cp312-cp312-linux_x86_64.whl (from torch==2.10.0a0+git344e636) 2025-10-10T01:27:42.2429696Z Requirement already satisfied: filelock in /opt/conda/envs/py_3.12/lib/python3.12/site-packages (from torch==2.10.0a0+git344e636->torch==2.10.0a0+git344e636) (3.18.0) 2025-10-10T01:27:42.2432184Z Requirement already satisfied: typing-extensions>=4.10.0 in /opt/conda/envs/py_3.12/lib/python3.12/site-packages (from torch==2.10.0a0+git344e636->torch==2.10.0a0+git344e636) (4.12.2) 2025-10-10T01:27:42.2435916Z Requirement already satisfied: setuptools in /opt/conda/envs/py_3.12/lib/python3.12/site-packages (from torch==2.10.0a0+git344e636->torch==2.10.0a0+git344e636) (78.1.1) 2025-10-10T01:27:42.2437483Z Requirement already satisfied: sympy>=1.13.3 in /opt/conda/envs/py_3.12/lib/python3.12/site-packages (from torch==2.10.0a0+git344e636->torch==2.10.0a0+git344e636) (1.13.3) 2025-10-10T01:27:42.2440668Z Requirement already satisfied: networkx>=2.5.1 in /opt/conda/envs/py_3.12/lib/python3.12/site-packages (from torch==2.10.0a0+git344e636->torch==2.10.0a0+git344e636) (2.8.8) 2025-10-10T01:27:42.2443096Z Requirement already satisfied: jinja2 in /opt/conda/envs/py_3.12/lib/python3.12/site-packages (from torch==2.10.0a0+git344e636->torch==2.10.0a0+git344e636) (3.1.6) 2025-10-10T01:27:42.2446596Z Requirement already satisfied: fsspec>=0.8.5 in /opt/conda/envs/py_3.12/lib/python3.12/site-packages (from torch==2.10.0a0+git344e636->torch==2.10.0a0+git344e636) (2025.9.0) 2025-10-10T01:27:42.2458513Z Requirement already satisfied: opt-einsum>=3.3 in /opt/conda/envs/py_3.12/lib/python3.12/site-packages (from torch==2.10.0a0+git344e636->torch==2.10.0a0+git344e636) (3.3.0) 2025-10-10T01:27:42.2556248Z Requirement already satisfied: numpy>=1.7 in /opt/conda/envs/py_3.12/lib/python3.12/site-packages (from opt-einsum>=3.3->torch==2.10.0a0+git344e636->torch==2.10.0a0+git344e636) (1.26.2) 2025-10-10T01:27:42.2593795Z Requirement already satisfied: mpmath<1.4,>=1.1.0 in /opt/conda/envs/py_3.12/lib/python3.12/site-packages (from sympy>=1.13.3->torch==2.10.0a0+git344e636->torch==2.10.0a0+git344e636) (1.3.0) 2025-10-10T01:27:42.2635555Z Requirement already satisfied: MarkupSafe>=2.0 in /opt/conda/envs/py_3.12/lib/python3.12/site-packages (from jinja2->torch==2.10.0a0+git344e636->torch==2.10.0a0+git344e636) (3.0.3) 2025-10-10T01:27:42.4582309Z Installing collected packages: torch 2025-10-10T01:27:54.7296095Z Successfully installed torch-2.10.0a0+git344e636 2025-10-10T01:27:54.8075147Z + export TERM=vt100 2025-10-10T01:27:54.8075407Z + TERM=vt100 2025-10-10T01:27:54.8080859Z ++ dirname .ci/pytorch/test.sh 2025-10-10T01:27:54.8092489Z + source .ci/pytorch/common.sh 2025-10-10T01:27:54.8098952Z +++ dirname .ci/pytorch/common.sh 2025-10-10T01:27:54.8111092Z ++ source .ci/pytorch/common_utils.sh 2025-10-10T01:27:54.8114142Z +++ declare -f -t trap_add 2025-10-10T01:27:54.8118205Z ++ set -ex -o pipefail 2025-10-10T01:27:54.8119017Z ++ [[ linux-jammy-cuda12.8-py3.12-gcc11 == *rocm* ]] 2025-10-10T01:27:54.8119640Z ++ BUILD_TEST_LIBTORCH=0 2025-10-10T01:27:54.8124587Z ++ dirname .ci/pytorch/test.sh 2025-10-10T01:27:54.8136863Z + source .ci/pytorch/common-build.sh 2025-10-10T01:27:54.8139113Z ++ [[ linux-jammy-cuda12.8-py3.12-gcc11 != *win-* ]] 2025-10-10T01:27:54.8146566Z ++++ dirname .ci/pytorch/common-build.sh 2025-10-10T01:27:54.8164664Z +++ cd .ci/pytorch 2025-10-10T01:27:54.8165096Z +++ pwd -P 2025-10-10T01:27:54.8168661Z ++ script_dir=/var/lib/jenkins/workspace/.ci/pytorch 2025-10-10T01:27:54.8169538Z ++ [[ linux-jammy-cuda12.8-py3.12-gcc11 == *-pch* ]] 2025-10-10T01:27:54.8170089Z ++ which sccache 2025-10-10T01:27:54.8278865Z ++ [[ -z ossci-compiler-cache-circleci-v2 ]] 2025-10-10T01:27:54.8279524Z ++ sccache --stop-server 2025-10-10T01:27:54.8315760Z ++ true 2025-10-10T01:27:54.8316229Z ++ rm -f /var/lib/jenkins/sccache_error.log 2025-10-10T01:27:54.8326963Z ++ trap_add sccache_epilogue EXIT 2025-10-10T01:27:54.8327284Z ++ trap_add_cmd=sccache_epilogue 2025-10-10T01:27:54.8327535Z ++ shift 2025-10-10T01:27:54.8327744Z ++ for trap_add_name in "$@" 2025-10-10T01:27:54.8336982Z ++++ trap -p EXIT 2025-10-10T01:27:54.8340887Z +++ eval 'extract_trap_cmd ' 2025-10-10T01:27:54.8341169Z ++++ extract_trap_cmd 2025-10-10T01:27:54.8341396Z ++++ printf '%s\n' '' 2025-10-10T01:27:54.8342112Z +++ printf '%s\n' sccache_epilogue 2025-10-10T01:27:54.8346544Z ++ trap -- ' 2025-10-10T01:27:54.8346779Z sccache_epilogue' EXIT 2025-10-10T01:27:54.8347556Z ++ [[ -n 1 ]] 2025-10-10T01:27:54.8348022Z ++ echo 'Skipping sccache server initialization, setting environment variables' 2025-10-10T01:27:54.8349062Z Skipping sccache server initialization, setting environment variables 2025-10-10T01:27:54.8349599Z ++ export SCCACHE_IDLE_TIMEOUT=0 2025-10-10T01:27:54.8349935Z ++ SCCACHE_IDLE_TIMEOUT=0 2025-10-10T01:27:54.8350305Z ++ export SCCACHE_ERROR_LOG=/var/lib/jenkins/sccache_error.log 2025-10-10T01:27:54.8350756Z ++ SCCACHE_ERROR_LOG=/var/lib/jenkins/sccache_error.log 2025-10-10T01:27:54.8358450Z ++ export RUST_LOG=sccache::server=error 2025-10-10T01:27:54.8358873Z ++ RUST_LOG=sccache::server=error 2025-10-10T01:27:54.8359114Z ++ sccache --zero-stats 2025-10-10T01:27:55.3169723Z Statistics zeroed. 2025-10-10T01:27:55.3180393Z ++ which ccache 2025-10-10T01:27:55.3324408Z + [[ linux-jammy-cuda12.8-py3.12-gcc11 != *rocm* ]] 2025-10-10T01:27:55.3325290Z + [[ linux-jammy-cuda12.8-py3.12-gcc11 != *s390x* ]] 2025-10-10T01:27:55.3325954Z + [[ -d /var/lib/jenkins/workspace ]] 2025-10-10T01:27:55.3330233Z ++ stat -c %u /var/lib/jenkins/workspace 2025-10-10T01:27:55.3347752Z + WORKSPACE_ORIGINAL_OWNER_ID=1000 2025-10-10T01:27:55.3348102Z + trap_add cleanup_workspace EXIT 2025-10-10T01:27:55.3348400Z + trap_add_cmd=cleanup_workspace 2025-10-10T01:27:55.3348660Z + shift 2025-10-10T01:27:55.3348860Z + for trap_add_name in "$@" 2025-10-10T01:27:55.3360273Z +++ trap -p EXIT 2025-10-10T01:27:55.3364161Z ++ eval 'extract_trap_cmd trap -- '\'' 2025-10-10T01:27:55.3364552Z sccache_epilogue'\'' EXIT' 2025-10-10T01:27:55.3364784Z +++ extract_trap_cmd trap -- ' 2025-10-10T01:27:55.3364993Z sccache_epilogue' EXIT 2025-10-10T01:27:55.3365181Z +++ printf '%s\n' ' 2025-10-10T01:27:55.3365366Z sccache_epilogue' 2025-10-10T01:27:55.3365557Z ++ printf '%s\n' cleanup_workspace 2025-10-10T01:27:55.3368540Z + trap -- ' 2025-10-10T01:27:55.3368725Z sccache_epilogue 2025-10-10T01:27:55.3368996Z cleanup_workspace' EXIT 2025-10-10T01:27:55.3369234Z + sudo chown -R jenkins /var/lib/jenkins/workspace 2025-10-10T01:27:56.3407170Z + git config --global --add safe.directory /var/lib/jenkins/workspace 2025-10-10T01:27:56.3438819Z + [[ linux-jammy-cuda12.8-py3.12-gcc11 == *cuda* ]] 2025-10-10T01:27:56.3444482Z ++ python -c 'import os;import numba.cuda; print(os.path.dirname(numba.cuda.__file__))' 2025-10-10T01:27:56.8403797Z + NUMBA_CUDA_DIR=/opt/conda/envs/py_3.12/lib/python3.12/site-packages/numba/cuda 2025-10-10T01:27:56.8404381Z + '[' -n /opt/conda/envs/py_3.12/lib/python3.12/site-packages/numba/cuda ']' 2025-10-10T01:27:56.8413509Z +++ realpath .ci/pytorch/test.sh 2025-10-10T01:27:56.8430474Z ++ dirname /var/lib/jenkins/workspace/.ci/pytorch/test.sh 2025-10-10T01:27:56.8709681Z + NUMBA_PATCH=/var/lib/jenkins/workspace/.ci/pytorch/numba-cuda-13.patch 2025-10-10T01:27:56.8710963Z + pushd /opt/conda/envs/py_3.12/lib/python3.12/site-packages/numba/cuda 2025-10-10T01:27:56.8711999Z /opt/conda/envs/py_3.12/lib/python3.12/site-packages/numba/cuda ~/workspace 2025-10-10T01:27:56.8712481Z + patch -p4 2025-10-10T01:27:56.8730171Z patching file cudadrv/driver.py 2025-10-10T01:27:56.8736764Z + popd 2025-10-10T01:27:56.8737467Z ~/workspace 2025-10-10T01:27:56.8737667Z + echo 'Environment variables:' 2025-10-10T01:27:56.8738150Z Environment variables: 2025-10-10T01:27:56.8738353Z + env 2025-10-10T01:27:56.8751376Z GITHUB_WORKSPACE=/home/ec2-user/actions-runner/_work/pytorch/pytorch 2025-10-10T01:27:56.8752016Z CONTINUE_THROUGH_ERROR=True 2025-10-10T01:27:56.8752430Z BUILD_ENVIRONMENT=linux-jammy-cuda12.8-py3.12-gcc11 2025-10-10T01:27:56.8753199Z VLLM_TEST_HUGGING_FACE_TOKEN=*** 2025-10-10T01:27:56.8753479Z HOSTNAME=92bc90bdf047 2025-10-10T01:27:56.8754211Z GITHUB_PATH=/home/ec2-user/actions-runner/_work/_temp/_runner_file_commands/add_path_9c160c0e-9d52-49be-a0c1-01555b45e00a 2025-10-10T01:27:56.8754785Z GITHUB_ACTION=__run_2 2025-10-10T01:27:56.8755028Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=0 2025-10-10T01:27:56.8755293Z GITHUB_RUN_NUMBER=2144 2025-10-10T01:27:56.8755564Z TEST_CONFIG=vllm_distributed_test_28_failure_test 2025-10-10T01:27:56.8755900Z GITHUB_REPOSITORY_OWNER_ID=21003710 2025-10-10T01:27:56.8756197Z TORCH_NVCC_FLAGS=-Xfatbin -compress-all 2025-10-10T01:27:56.8756506Z SCCACHE_IDLE_TIMEOUT=0 2025-10-10T01:27:56.8756866Z SCRIBE_GRAPHQL_ACCESS_TOKEN=*** 2025-10-10T01:27:56.8757168Z GITHUB_TRIGGERING_ACTOR=pytorchmergebot 2025-10-10T01:27:56.8757458Z GITHUB_REF_TYPE=branch 2025-10-10T01:27:56.8757725Z BASE_SHA=344e6365a0068c2d2847fcec0c55dd53291d475e 2025-10-10T01:27:56.8758017Z XLA_CUDA= 2025-10-10T01:27:56.8758218Z NCCL_LIB_DIR=/usr/local/cuda/lib64/ 2025-10-10T01:27:56.8758693Z HUGGING_FACE_HUB_TOKEN=*** 2025-10-10T01:27:56.8759170Z *** 2025-10-10T01:27:56.8759332Z GITHUB_REPOSITORY_ID=65600975 2025-10-10T01:27:56.8759534Z GITHUB_ACTIONS=true 2025-10-10T01:27:56.8759720Z NVIDIA_DRIVER_CAPABILITIES=all 2025-10-10T01:27:56.8760019Z SCCACHE_ERROR_LOG=/var/lib/jenkins/sccache_error.log 2025-10-10T01:27:56.8760381Z SHA1=344e6365a0068c2d2847fcec0c55dd53291d475e 2025-10-10T01:27:56.8760646Z GITHUB_SHA=344e6365a0068c2d2847fcec0c55dd53291d475e 2025-10-10T01:27:56.8761120Z GITHUB_WORKFLOW_REF=pytorch/pytorch/.github/workflows/vllm.yml@refs/heads/main 2025-10-10T01:27:56.8761453Z UCC_HOME=/usr 2025-10-10T01:27:56.8761631Z VERBOSE_TEST_LOGS=False 2025-10-10T01:27:56.8761828Z GITHUB_REF=refs/heads/main 2025-10-10T01:27:56.8762013Z SHARD_NUMBER=1 2025-10-10T01:27:56.8762181Z GITHUB_REF_PROTECTED=true 2025-10-10T01:27:56.8762373Z HOME=/var/lib/jenkins 2025-10-10T01:27:56.8762581Z GITHUB_API_URL=https://api.github.com 2025-10-10T01:27:56.8762821Z PYTORCH_TEST_RERUN_DISABLED_TESTS=0 2025-10-10T01:27:56.8763068Z UCX_COMMIT=7836b165abdbe468a2f607e7254011c07d788152 2025-10-10T01:27:56.8763310Z USE_SYSTEM_NCCL=1 2025-10-10T01:27:56.8763476Z NUM_TEST_SHARDS=1 2025-10-10T01:27:56.8763629Z UCX_HOME=/usr 2025-10-10T01:27:56.8764039Z GITHUB_STATE=/home/ec2-user/actions-runner/_work/_temp/_runner_file_commands/save_state_9c160c0e-9d52-49be-a0c1-01555b45e00a 2025-10-10T01:27:56.8764668Z JOB_NAME=ci-vllm-test / test (vllm_distributed_test_28_failure_test, 1, 1, linux.g6.12xlarge.nvidia.gpu) 2025-10-10T01:27:56.8765291Z GITHUB_ENV=/home/ec2-user/actions-runner/_work/_temp/_runner_file_commands/set_env_9c160c0e-9d52-49be-a0c1-01555b45e00a 2025-10-10T01:27:56.8765871Z GITHUB_EVENT_PATH=/home/ec2-user/actions-runner/_work/_temp/_github_workflow/event.json 2025-10-10T01:27:56.8766912Z GITHUB_EVENT_NAME=push 2025-10-10T01:27:56.8767089Z DASHBOARD_TAG= 2025-10-10T01:27:56.8767254Z GITHUB_RUN_ID=18392306035 2025-10-10T01:27:56.8767438Z INSTALLED_OPENBLAS= 2025-10-10T01:27:56.8767867Z GITHUB_STEP_SUMMARY=/home/ec2-user/actions-runner/_work/_temp/_runner_file_commands/step_summary_9c160c0e-9d52-49be-a0c1-01555b45e00a 2025-10-10T01:27:56.8768353Z GITHUB_ACTOR=pytorchmergebot 2025-10-10T01:27:56.8768543Z PR_NUMBER= 2025-10-10T01:27:56.8768692Z DESIRED_CUDA=12.8.1 2025-10-10T01:27:56.8768856Z GITHUB_RUN_ATTEMPT=1 2025-10-10T01:27:56.8769034Z ANACONDA_PYTHON_VERSION=3.12 2025-10-10T01:27:56.8769272Z GITHUB_GRAPHQL_URL=https://api.github.com/graphql 2025-10-10T01:27:56.8769511Z TERM=vt100 2025-10-10T01:27:56.8769656Z INSTALLED_VISION=yes 2025-10-10T01:27:56.8769824Z BRANCH=main 2025-10-10T01:27:56.8769984Z SCCACHE_REGION=us-east-1 2025-10-10T01:27:56.8770182Z OPENSSL_ROOT_DIR=/opt/openssl 2025-10-10T01:27:56.8770477Z CUDA_PATH=/usr/local/cuda 2025-10-10T01:27:56.8770864Z GITHUB_ACTION_PATH=/home/ec2-user/actions-runner/_work/pytorch/pytorch/./.github/actions/setup-linux 2025-10-10T01:27:56.8771274Z GITHUB_SERVER_URL=https://github.com 2025-10-10T01:27:56.8771530Z UCC_COMMIT=430e241bf5d38cbc73fc7a6b89155397232e3f96 2025-10-10T01:27:56.8771763Z REENABLED_ISSUES= 2025-10-10T01:27:56.8771921Z DOCS= 2025-10-10T01:27:56.8772057Z SHLVL=1 2025-10-10T01:27:56.8772197Z MAX_JOBS=46 2025-10-10T01:27:56.8772343Z GITHUB_ACTOR_ID=97764156 2025-10-10T01:27:56.8772587Z GITHUB_WORKFLOW_SHA=344e6365a0068c2d2847fcec0c55dd53291d475e 2025-10-10T01:27:56.8772852Z GITHUB_REF_NAME=main 2025-10-10T01:27:56.8773124Z XLA_CLANG_CACHE_S3_BUCKET_NAME=ossci-compiler-clang-cache-circleci-xla 2025-10-10T01:27:56.8773420Z GITHUB_JOB=test 2025-10-10T01:27:56.8773582Z NO_TEST_TIMEOUT=False 2025-10-10T01:27:56.8773756Z TD_DISTRIBUTED=False 2025-10-10T01:27:56.8773964Z GITHUB_REPOSITORY=pytorch/pytorch 2025-10-10T01:27:56.8774177Z GITHUB_RETENTION_DAYS=90 2025-10-10T01:27:56.8774368Z OPENSSL_DIR=/opt/openssl 2025-10-10T01:27:56.8774556Z GITHUB_ACTION_REPOSITORY= 2025-10-10T01:27:56.8775103Z PATH=/opt/cache/bin:/usr/local/nvidia/bin:/usr/local/cuda/bin:/opt/conda/envs/py_3.12/bin:/opt/conda/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin 2025-10-10T01:27:56.8775648Z GITHUB_BASE_REF= 2025-10-10T01:27:56.8775935Z INSTALLED_ACL= 2025-10-10T01:27:56.8776315Z ARTIFACTS_FILE_SUFFIX=test-vllm_distributed_test_28_failure_test-1-1-linux.g6.12xlarge.nvidia.gpu_52408504942 2025-10-10T01:27:56.8776729Z CI=true 2025-10-10T01:27:56.8776884Z GITHUB_REPOSITORY_OWNER=pytorch 2025-10-10T01:27:56.8777125Z RUST_LOG=sccache::server=error 2025-10-10T01:27:56.8777320Z JOB_ID=52408504942 2025-10-10T01:27:56.8777481Z GITHUB_HEAD_REF= 2025-10-10T01:27:56.8777634Z GITHUB_ACTION_REF= 2025-10-10T01:27:56.8777841Z SCCACHE_BUCKET=ossci-compiler-cache-circleci-v2 2025-10-10T01:27:56.8778100Z TEST_SHOWLOCALS=False 2025-10-10T01:27:56.8778289Z GITHUB_WORKFLOW=vllm-test 2025-10-10T01:27:56.8778485Z DEBIAN_FRONTEND=noninteractive 2025-10-10T01:27:56.8778931Z GITHUB_OUTPUT=/home/ec2-user/actions-runner/_work/_temp/_runner_file_commands/set_output_9c160c0e-9d52-49be-a0c1-01555b45e00a 2025-10-10T01:27:56.8779380Z NO_TD=False 2025-10-10T01:27:56.8779544Z SKIP_SCCACHE_INITIALIZATION=1 2025-10-10T01:27:56.8779752Z NCCL_INCLUDE_DIR=/usr/local/cuda/include/ 2025-10-10T01:27:56.8779984Z _=/usr/bin/env 2025-10-10T01:27:56.8780241Z OLDPWD=/opt/conda/envs/py_3.12/lib/python3.12/site-packages/numba/cuda 2025-10-10T01:27:56.8780618Z ++ python -c 'import site; print(site.getsitepackages()[0])' 2025-10-10T01:27:56.8889039Z + TORCH_INSTALL_DIR=/opt/conda/envs/py_3.12/lib/python3.12/site-packages/torch 2025-10-10T01:27:56.8890355Z + TORCH_BIN_DIR=/opt/conda/envs/py_3.12/lib/python3.12/site-packages/torch/bin 2025-10-10T01:27:56.8890910Z + TORCH_LIB_DIR=/opt/conda/envs/py_3.12/lib/python3.12/site-packages/torch/lib 2025-10-10T01:27:56.8891564Z + TORCH_TEST_DIR=/opt/conda/envs/py_3.12/lib/python3.12/site-packages/torch/test 2025-10-10T01:27:56.8892321Z + BUILD_DIR=build 2025-10-10T01:27:56.8892549Z + BUILD_RENAMED_DIR=build_renamed 2025-10-10T01:27:56.8892823Z + BUILD_BIN_DIR=build/bin 2025-10-10T01:27:56.8893048Z + SHARD_NUMBER=1 2025-10-10T01:27:56.8893249Z + NUM_TEST_SHARDS=1 2025-10-10T01:27:56.8893481Z + export TORCH_SERIALIZATION_DEBUG=1 2025-10-10T01:27:56.8893760Z + TORCH_SERIALIZATION_DEBUG=1 2025-10-10T01:27:56.8894004Z + export VALGRIND=ON 2025-10-10T01:27:56.8894216Z + VALGRIND=ON 2025-10-10T01:27:56.8894463Z + [[ linux-jammy-cuda12.8-py3.12-gcc11 == *clang9* ]] 2025-10-10T01:27:56.8894825Z + [[ linux-jammy-cuda12.8-py3.12-gcc11 == *xpu* ]] 2025-10-10T01:27:56.8895122Z + detect_cuda_arch 2025-10-10T01:27:56.8895384Z + [[ linux-jammy-cuda12.8-py3.12-gcc11 == *cuda* ]] 2025-10-10T01:27:56.8895698Z + command -v nvidia-smi 2025-10-10T01:27:56.8895926Z /usr/bin/nvidia-smi 2025-10-10T01:27:56.8902217Z ++ nvidia-smi --query-gpu=compute_cap --format=csv 2025-10-10T01:27:56.8904154Z ++ tail -n 1 2025-10-10T01:27:56.9419108Z + TORCH_CUDA_ARCH_LIST=8.9 2025-10-10T01:27:56.9419611Z + export TORCH_CUDA_ARCH_LIST 2025-10-10T01:27:56.9420021Z + [[ linux-jammy-cuda12.8-py3.12-gcc11 == *s390x* ]] 2025-10-10T01:27:56.9420418Z + [[ 0 == \1 ]] 2025-10-10T01:27:56.9420664Z + [[ True == \1 ]] 2025-10-10T01:27:56.9420958Z + [[ linux-jammy-cuda12.8-py3.12-gcc11 != *bazel* ]] 2025-10-10T01:27:56.9426584Z ++ realpath build/custom_test_artifacts 2025-10-10T01:27:56.9569905Z + CUSTOM_TEST_ARTIFACT_BUILD_DIR=/var/lib/jenkins/workspace/build/custom_test_artifacts 2025-10-10T01:27:56.9571626Z + [[ -n '' ]] 2025-10-10T01:27:56.9572472Z + echo 'Environment variables' 2025-10-10T01:27:56.9573263Z Environment variables 2025-10-10T01:27:56.9573712Z + env 2025-10-10T01:27:56.9694295Z GITHUB_WORKSPACE=/home/ec2-user/actions-runner/_work/pytorch/pytorch 2025-10-10T01:27:56.9695146Z CONTINUE_THROUGH_ERROR=True 2025-10-10T01:27:56.9695721Z BUILD_ENVIRONMENT=linux-jammy-cuda12.8-py3.12-gcc11 2025-10-10T01:27:56.9696744Z VLLM_TEST_HUGGING_FACE_TOKEN=*** 2025-10-10T01:27:56.9697274Z HOSTNAME=92bc90bdf047 2025-10-10T01:27:56.9698261Z GITHUB_PATH=/home/ec2-user/actions-runner/_work/_temp/_runner_file_commands/add_path_9c160c0e-9d52-49be-a0c1-01555b45e00a 2025-10-10T01:27:56.9699307Z GITHUB_ACTION=__run_2 2025-10-10T01:27:56.9699741Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=0 2025-10-10T01:27:56.9700250Z GITHUB_RUN_NUMBER=2144 2025-10-10T01:27:56.9700612Z TEST_CONFIG=vllm_distributed_test_28_failure_test 2025-10-10T01:27:56.9700877Z GITHUB_REPOSITORY_OWNER_ID=21003710 2025-10-10T01:27:56.9701110Z TORCH_NVCC_FLAGS=-Xfatbin -compress-all 2025-10-10T01:27:56.9701376Z SCCACHE_IDLE_TIMEOUT=0 2025-10-10T01:27:56.9701652Z SCRIBE_GRAPHQL_ACCESS_TOKEN=*** 2025-10-10T01:27:56.9701882Z GITHUB_TRIGGERING_ACTOR=pytorchmergebot 2025-10-10T01:27:56.9702110Z GITHUB_REF_TYPE=branch 2025-10-10T01:27:56.9702287Z TORCH_CUDA_ARCH_LIST=8.9 2025-10-10T01:27:56.9702500Z BASE_SHA=344e6365a0068c2d2847fcec0c55dd53291d475e 2025-10-10T01:27:56.9702736Z XLA_CUDA= 2025-10-10T01:27:56.9702901Z NCCL_LIB_DIR=/usr/local/cuda/lib64/ 2025-10-10T01:27:56.9703300Z HUGGING_FACE_HUB_TOKEN=*** 2025-10-10T01:27:56.9703543Z *** 2025-10-10T01:27:56.9703702Z GITHUB_REPOSITORY_ID=65600975 2025-10-10T01:27:56.9703906Z GITHUB_ACTIONS=true 2025-10-10T01:27:56.9704090Z NVIDIA_DRIVER_CAPABILITIES=all 2025-10-10T01:27:56.9704340Z SCCACHE_ERROR_LOG=/var/lib/jenkins/sccache_error.log 2025-10-10T01:27:56.9704618Z SHA1=344e6365a0068c2d2847fcec0c55dd53291d475e 2025-10-10T01:27:56.9704877Z GITHUB_SHA=344e6365a0068c2d2847fcec0c55dd53291d475e 2025-10-10T01:27:56.9705239Z GITHUB_WORKFLOW_REF=pytorch/pytorch/.github/workflows/vllm.yml@refs/heads/main 2025-10-10T01:27:56.9705556Z UCC_HOME=/usr 2025-10-10T01:27:56.9705726Z TORCH_SERIALIZATION_DEBUG=1 2025-10-10T01:27:56.9705922Z VERBOSE_TEST_LOGS=False 2025-10-10T01:27:56.9706106Z GITHUB_REF=refs/heads/main 2025-10-10T01:27:56.9706280Z SHARD_NUMBER=1 2025-10-10T01:27:56.9706445Z GITHUB_REF_PROTECTED=true 2025-10-10T01:27:56.9706629Z HOME=/var/lib/jenkins 2025-10-10T01:27:56.9706834Z GITHUB_API_URL=https://api.github.com 2025-10-10T01:27:56.9707315Z PYTORCH_TEST_RERUN_DISABLED_TESTS=0 2025-10-10T01:27:56.9707572Z UCX_COMMIT=7836b165abdbe468a2f607e7254011c07d788152 2025-10-10T01:27:56.9707815Z USE_SYSTEM_NCCL=1 2025-10-10T01:27:56.9707984Z NUM_TEST_SHARDS=1 2025-10-10T01:27:56.9708143Z UCX_HOME=/usr 2025-10-10T01:27:56.9708556Z GITHUB_STATE=/home/ec2-user/actions-runner/_work/_temp/_runner_file_commands/save_state_9c160c0e-9d52-49be-a0c1-01555b45e00a 2025-10-10T01:27:56.9709195Z JOB_NAME=ci-vllm-test / test (vllm_distributed_test_28_failure_test, 1, 1, linux.g6.12xlarge.nvidia.gpu) 2025-10-10T01:27:56.9709796Z GITHUB_ENV=/home/ec2-user/actions-runner/_work/_temp/_runner_file_commands/set_env_9c160c0e-9d52-49be-a0c1-01555b45e00a 2025-10-10T01:27:56.9710365Z GITHUB_EVENT_PATH=/home/ec2-user/actions-runner/_work/_temp/_github_workflow/event.json 2025-10-10T01:27:56.9710710Z GITHUB_EVENT_NAME=push 2025-10-10T01:27:56.9710884Z DASHBOARD_TAG= 2025-10-10T01:27:56.9711201Z GITHUB_RUN_ID=18392306035 2025-10-10T01:27:56.9711408Z INSTALLED_OPENBLAS= 2025-10-10T01:27:56.9711843Z GITHUB_STEP_SUMMARY=/home/ec2-user/actions-runner/_work/_temp/_runner_file_commands/step_summary_9c160c0e-9d52-49be-a0c1-01555b45e00a 2025-10-10T01:27:56.9712330Z GITHUB_ACTOR=pytorchmergebot 2025-10-10T01:27:56.9712531Z PR_NUMBER= 2025-10-10T01:27:56.9712693Z DESIRED_CUDA=12.8.1 2025-10-10T01:27:56.9712864Z GITHUB_RUN_ATTEMPT=1 2025-10-10T01:27:56.9713036Z VALGRIND=ON 2025-10-10T01:27:56.9713202Z ANACONDA_PYTHON_VERSION=3.12 2025-10-10T01:27:56.9713441Z GITHUB_GRAPHQL_URL=https://api.github.com/graphql 2025-10-10T01:27:56.9713693Z TERM=vt100 2025-10-10T01:27:56.9713859Z INSTALLED_VISION=yes 2025-10-10T01:27:56.9714022Z BRANCH=main 2025-10-10T01:27:56.9714178Z SCCACHE_REGION=us-east-1 2025-10-10T01:27:56.9714373Z OPENSSL_ROOT_DIR=/opt/openssl 2025-10-10T01:27:56.9714571Z CUDA_PATH=/usr/local/cuda 2025-10-10T01:27:56.9714948Z GITHUB_ACTION_PATH=/home/ec2-user/actions-runner/_work/pytorch/pytorch/./.github/actions/setup-linux 2025-10-10T01:27:56.9715376Z GITHUB_SERVER_URL=https://github.com 2025-10-10T01:27:56.9715633Z UCC_COMMIT=430e241bf5d38cbc73fc7a6b89155397232e3f96 2025-10-10T01:27:56.9715880Z REENABLED_ISSUES= 2025-10-10T01:27:56.9716036Z DOCS= 2025-10-10T01:27:56.9716167Z SHLVL=1 2025-10-10T01:27:56.9716304Z MAX_JOBS=46 2025-10-10T01:27:56.9716457Z GITHUB_ACTOR_ID=97764156 2025-10-10T01:27:56.9716698Z GITHUB_WORKFLOW_SHA=344e6365a0068c2d2847fcec0c55dd53291d475e 2025-10-10T01:27:56.9716968Z GITHUB_REF_NAME=main 2025-10-10T01:27:56.9717244Z XLA_CLANG_CACHE_S3_BUCKET_NAME=ossci-compiler-clang-cache-circleci-xla 2025-10-10T01:27:56.9717542Z GITHUB_JOB=test 2025-10-10T01:27:56.9717698Z NO_TEST_TIMEOUT=False 2025-10-10T01:27:56.9717873Z TD_DISTRIBUTED=False 2025-10-10T01:27:56.9718061Z GITHUB_REPOSITORY=pytorch/pytorch 2025-10-10T01:27:56.9718275Z GITHUB_RETENTION_DAYS=90 2025-10-10T01:27:56.9718463Z OPENSSL_DIR=/opt/openssl 2025-10-10T01:27:56.9718856Z GITHUB_ACTION_REPOSITORY= 2025-10-10T01:27:56.9719525Z PATH=/opt/cache/bin:/usr/local/nvidia/bin:/usr/local/cuda/bin:/opt/conda/envs/py_3.12/bin:/opt/conda/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin 2025-10-10T01:27:56.9720095Z GITHUB_BASE_REF= 2025-10-10T01:27:56.9720262Z INSTALLED_ACL= 2025-10-10T01:27:56.9720631Z ARTIFACTS_FILE_SUFFIX=test-vllm_distributed_test_28_failure_test-1-1-linux.g6.12xlarge.nvidia.gpu_52408504942 2025-10-10T01:27:56.9721038Z CI=true 2025-10-10T01:27:56.9721200Z GITHUB_REPOSITORY_OWNER=pytorch 2025-10-10T01:27:56.9721442Z RUST_LOG=sccache::server=error 2025-10-10T01:27:56.9721632Z JOB_ID=52408504942 2025-10-10T01:27:56.9721796Z GITHUB_HEAD_REF= 2025-10-10T01:27:56.9721956Z GITHUB_ACTION_REF= 2025-10-10T01:27:56.9722165Z SCCACHE_BUCKET=ossci-compiler-cache-circleci-v2 2025-10-10T01:27:56.9722411Z TEST_SHOWLOCALS=False 2025-10-10T01:27:56.9722594Z GITHUB_WORKFLOW=vllm-test 2025-10-10T01:27:56.9722790Z DEBIAN_FRONTEND=noninteractive 2025-10-10T01:27:56.9723241Z GITHUB_OUTPUT=/home/ec2-user/actions-runner/_work/_temp/_runner_file_commands/set_output_9c160c0e-9d52-49be-a0c1-01555b45e00a 2025-10-10T01:27:56.9723808Z NO_TD=False 2025-10-10T01:27:56.9724186Z SKIP_SCCACHE_INITIALIZATION=1 2025-10-10T01:27:56.9724422Z NCCL_INCLUDE_DIR=/usr/local/cuda/include/ 2025-10-10T01:27:56.9724755Z OLDPWD=/opt/conda/envs/py_3.12/lib/python3.12/site-packages/numba/cuda 2025-10-10T01:27:56.9725068Z _=/usr/bin/env 2025-10-10T01:27:56.9725244Z + echo 'Testing pytorch' 2025-10-10T01:27:56.9725433Z Testing pytorch 2025-10-10T01:27:56.9725604Z + export LANG=C.UTF-8 2025-10-10T01:27:56.9725770Z + LANG=C.UTF-8 2025-10-10T01:27:56.9725939Z + PR_NUMBER= 2025-10-10T01:27:56.9726161Z + [[ vllm_distributed_test_28_failure_test == \d\e\f\a\u\l\t ]] 2025-10-10T01:27:56.9726510Z + [[ vllm_distributed_test_28_failure_test == \d\i\s\t\r\i\b\u\t\e\d ]] 2025-10-10T01:27:56.9726831Z + [[ vllm_distributed_test_28_failure_test == \s\l\o\w ]] 2025-10-10T01:27:56.9727141Z + [[ linux-jammy-cuda12.8-py3.12-gcc11 == *slow-gradcheck* ]] 2025-10-10T01:27:56.9727594Z + [[ linux-jammy-cuda12.8-py3.12-gcc11 == *cuda* ]] 2025-10-10T01:27:56.9727896Z + export PYTORCH_TESTING_DEVICE_ONLY_FOR=cuda 2025-10-10T01:27:56.9728139Z + PYTORCH_TESTING_DEVICE_ONLY_FOR=cuda 2025-10-10T01:27:56.9728401Z + [[ vllm_distributed_test_28_failure_test == *crossref* ]] 2025-10-10T01:27:56.9728692Z + [[ linux-jammy-cuda12.8-py3.12-gcc11 == *rocm* ]] 2025-10-10T01:27:56.9728963Z + [[ linux-jammy-cuda12.8-py3.12-gcc11 == *xpu* ]] 2025-10-10T01:27:56.9729246Z + [[ linux-jammy-cuda12.8-py3.12-gcc11 != *-bazel-* ]] 2025-10-10T01:27:56.9729497Z + pip_install ninja==1.10.2 2025-10-10T01:27:56.9729768Z + pip_install_pkg='python3 -m pip install --progress-bar off' 2025-10-10T01:27:56.9730090Z + python3 -m pip install --progress-bar off ninja==1.10.2 2025-10-10T01:27:57.4788595Z Collecting ninja==1.10.2 2025-10-10T01:27:57.5097452Z Downloading ninja-1.10.2-py2.py3-none-manylinux_2_5_x86_64.manylinux1_x86_64.whl.metadata (5.0 kB) 2025-10-10T01:27:57.5539509Z Downloading ninja-1.10.2-py2.py3-none-manylinux_2_5_x86_64.manylinux1_x86_64.whl (108 kB) 2025-10-10T01:27:57.7526088Z Installing collected packages: ninja 2025-10-10T01:27:57.7526633Z Attempting uninstall: ninja 2025-10-10T01:27:57.7553018Z Found existing installation: ninja 1.11.1.4 2025-10-10T01:27:57.7575403Z Uninstalling ninja-1.11.1.4: 2025-10-10T01:27:57.7704032Z Successfully uninstalled ninja-1.11.1.4 2025-10-10T01:27:57.8389270Z Successfully installed ninja-1.10.2 2025-10-10T01:27:57.8850884Z + export PATH=/var/lib/jenkins/.local/bin:/opt/cache/bin:/usr/local/nvidia/bin:/usr/local/cuda/bin:/opt/conda/envs/py_3.12/bin:/opt/conda/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin 2025-10-10T01:27:57.8852742Z + PATH=/var/lib/jenkins/.local/bin:/opt/cache/bin:/usr/local/nvidia/bin:/usr/local/cuda/bin:/opt/conda/envs/py_3.12/bin:/opt/conda/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin 2025-10-10T01:27:57.8853607Z + [[ linux-jammy-cuda12.8-py3.12-gcc11 == *aarch64* ]] 2025-10-10T01:27:57.8853997Z + [[ linux-jammy-cuda12.8-py3.12-gcc11 == *asan* ]] 2025-10-10T01:27:57.8854374Z + [[ linux-jammy-cuda12.8-py3.12-gcc11 == *-debug* ]] 2025-10-10T01:27:57.8854725Z + [[ linux-jammy-cuda12.8-py3.12-gcc11 != *-bazel-* ]] 2025-10-10T01:27:57.8855225Z + echo 'We are not in debug mode: linux-jammy-cuda12.8-py3.12-gcc11. Expect the assertion to pass' 2025-10-10T01:27:57.8855881Z We are not in debug mode: linux-jammy-cuda12.8-py3.12-gcc11. Expect the assertion to pass 2025-10-10T01:27:57.8857232Z + cd test 2025-10-10T01:27:57.8857778Z + python -c 'import torch; torch._C._crash_if_debug_asserts_fail(424242)' 2025-10-10T01:27:59.3647549Z + [[ vllm_distributed_test_28_failure_test == \n\o\g\p\u\_\N\O\_\A\V\X\2 ]] 2025-10-10T01:27:59.3649201Z + [[ vllm_distributed_test_28_failure_test == \n\o\g\p\u\_\A\V\X\5\1\2 ]] 2025-10-10T01:27:59.3650349Z + [[ vllm_distributed_test_28_failure_test == \l\e\g\a\c\y\_\n\v\i\d\i\a\_\d\r\i\v\e\r ]] 2025-10-10T01:27:59.3652408Z + DYNAMO_BENCHMARK_FLAGS=() 2025-10-10T01:27:59.3652936Z + [[ vllm_distributed_test_28_failure_test == *pr_time_benchmarks* ]] 2025-10-10T01:27:59.3653955Z + [[ vllm_distributed_test_28_failure_test == *dynamo_eager* ]] 2025-10-10T01:27:59.3654367Z + [[ vllm_distributed_test_28_failure_test == *aot_eager* ]] 2025-10-10T01:27:59.3654676Z + [[ vllm_distributed_test_28_failure_test == *aot_inductor* ]] 2025-10-10T01:27:59.3655032Z + [[ vllm_distributed_test_28_failure_test == *max_autotune_inductor* ]] 2025-10-10T01:27:59.3655393Z + [[ vllm_distributed_test_28_failure_test == *inductor* ]] 2025-10-10T01:27:59.3655705Z + [[ vllm_distributed_test_28_failure_test == *dynamic* ]] 2025-10-10T01:27:59.3655996Z + [[ vllm_distributed_test_28_failure_test == *cpu* ]] 2025-10-10T01:27:59.3656276Z + DYNAMO_BENCHMARK_FLAGS+=(--device cuda) 2025-10-10T01:27:59.3841685Z + [[ linux-jammy-cuda12.8-py3.12-gcc11 == *libtorch* ]] 2025-10-10T01:27:59.3842582Z + [[ linux-jammy-cuda12.8-py3.12-gcc11 == *-bazel-* ]] 2025-10-10T01:27:59.3846240Z + cd test 2025-10-10T01:27:59.3848260Z + python -c 'import torch; print(torch.__config__.show())' 2025-10-10T01:28:01.7335798Z PyTorch built with: 2025-10-10T01:28:01.7336365Z - GCC 11.4 2025-10-10T01:28:01.7336776Z - C++ Version: 201703 2025-10-10T01:28:01.7337763Z - Intel(R) oneAPI Math Kernel Library Version 2024.2-Product Build 20240605 for Intel(R) 64 architecture applications 2025-10-10T01:28:01.7339051Z - Intel(R) MKL-DNN v3.7.1 (Git Hash 8d263e693366ef8db40acc569cc7d8edf644556d) 2025-10-10T01:28:01.7339835Z - OpenMP 201511 (a.k.a. OpenMP 4.5) 2025-10-10T01:28:01.7340420Z - LAPACK is enabled (usually provided by MKL) 2025-10-10T01:28:01.7340974Z - NNPACK is enabled 2025-10-10T01:28:01.7341419Z - CPU capability usage: AVX2 2025-10-10T01:28:01.7341696Z - CUDA Runtime 12.8 2025-10-10T01:28:01.7342149Z - NVCC architecture flags: -gencode;arch=compute_80,code=sm_80;-gencode;arch=compute_89,code=sm_89;-gencode;arch=compute_90,code=sm_90 2025-10-10T01:28:01.7342606Z - CuDNN 90.8 2025-10-10T01:28:01.7345964Z - Build settings: BLAS_INFO=mkl, BUILD_TYPE=Release, COMMIT_SHA=344e6365a0068c2d2847fcec0c55dd53291d475e, CUDA_VERSION=12.8, CUDNN_VERSION=9.8.0, CXX_COMPILER=/opt/cache/bin/c++, CXX_FLAGS= -fvisibility-inlines-hidden -DUSE_PTHREADPOOL -DNDEBUG -DUSE_KINETO -DLIBKINETO_NOROCTRACER -DLIBKINETO_NOXPUPTI=ON -DUSE_FBGEMM -DUSE_PYTORCH_QNNPACK -DUSE_XNNPACK -DSYMBOLICATE_MOBILE_DEBUG_HANDLE -O2 -fPIC -DC10_NODEPRECATED -Wall -Wextra -Werror=return-type -Werror=non-virtual-dtor -Werror=range-loop-construct -Werror=bool-operation -Wnarrowing -Wno-missing-field-initializers -Wno-unknown-pragmas -Wno-unused-parameter -Wno-strict-overflow -Wno-strict-aliasing -Wno-stringop-overflow -Wsuggest-override -Wno-psabi -Wno-error=old-style-cast -faligned-new -Werror -Wno-maybe-uninitialized -fno-math-errno -fno-trapping-math -Werror=format -Wno-stringop-overflow, FORCE_FALLBACK_CUDA_MPI=1, LAPACK_INFO=mkl, PERF_WITH_AVX=1, PERF_WITH_AVX2=1, TORCH_VERSION=2.10.0, USE_CUDA=ON, USE_CUDNN=ON, USE_CUSPARSELT=ON, USE_GFLAGS=OFF, USE_GLOG=OFF, USE_GLOO=ON, USE_MKL=ON, USE_MKLDNN=ON, USE_MPI=ON, USE_NCCL=ON, USE_NNPACK=ON, USE_OPENMP=ON, USE_ROCM=OFF, USE_ROCM_KERNEL_ASSERT=OFF, USE_XCCL=OFF, USE_XPU=OFF, 2025-10-10T01:28:01.7349413Z 2025-10-10T01:28:02.0665971Z + cd test 2025-10-10T01:28:02.0666316Z + python -c 'import torch; print(torch.__config__.parallel_info())' 2025-10-10T01:28:03.3168657Z ATen/Parallel: 2025-10-10T01:28:03.3168969Z at::get_num_threads() : 24 2025-10-10T01:28:03.3169256Z at::get_num_interop_threads() : 24 2025-10-10T01:28:03.3169545Z OpenMP 201511 (a.k.a. OpenMP 4.5) 2025-10-10T01:28:03.3169808Z omp_get_max_threads() : 24 2025-10-10T01:28:03.3170306Z Intel(R) oneAPI Math Kernel Library Version 2024.2-Product Build 20240605 for Intel(R) 64 architecture applications 2025-10-10T01:28:03.3170834Z mkl_get_max_threads() : 24 2025-10-10T01:28:03.3171249Z Intel(R) MKL-DNN v3.7.1 (Git Hash 8d263e693366ef8db40acc569cc7d8edf644556d) 2025-10-10T01:28:03.3171721Z std::thread::hardware_concurrency() : 48 2025-10-10T01:28:03.3172053Z Environment variables: 2025-10-10T01:28:03.3172343Z OMP_NUM_THREADS : [not set] 2025-10-10T01:28:03.3173046Z MKL_NUM_THREADS : [not set] 2025-10-10T01:28:03.3173341Z ATen parallel backend: OpenMP 2025-10-10T01:28:03.3173533Z 2025-10-10T01:28:03.5536052Z + [[ vllm_distributed_test_28_failure_test == *numpy_2* ]] 2025-10-10T01:28:03.5536495Z + [[ linux-jammy-cuda12.8-py3.12-gcc11 == *aarch64* ]] 2025-10-10T01:28:03.5536906Z + [[ vllm_distributed_test_28_failure_test == *backward* ]] 2025-10-10T01:28:03.5537284Z + [[ vllm_distributed_test_28_failure_test == *xla* ]] 2025-10-10T01:28:03.5537640Z + [[ vllm_distributed_test_28_failure_test == *vllm* ]] 2025-10-10T01:28:03.5537997Z + echo 'vLLM CI uses TORCH_CUDA_ARCH_LIST: 8.9' 2025-10-10T01:28:03.5538314Z vLLM CI uses TORCH_CUDA_ARCH_LIST: 8.9 2025-10-10T01:28:03.5541847Z + cd .ci/lumen_cli 2025-10-10T01:28:03.5543645Z + python -m pip install -e . 2025-10-10T01:28:03.8449509Z Obtaining file:///var/lib/jenkins/workspace/.ci/lumen_cli 2025-10-10T01:28:04.5740292Z Installing build dependencies ... [?25l- \ done 2025-10-10T01:28:04.7876829Z [?25h Checking if build backend supports build_editable ... [?25ldone 2025-10-10T01:28:05.0581732Z [?25h Getting requirements to build editable ... [?25l- done 2025-10-10T01:28:05.2227255Z [?25h Preparing editable metadata (pyproject.toml) ... [?25l- done 2025-10-10T01:28:05.2291827Z [?25hRequirement already satisfied: pyyaml==6.0.2 in /opt/conda/envs/py_3.12/lib/python3.12/site-packages (from lumen-ci==0.1.0) (6.0.2) 2025-10-10T01:28:05.2720925Z Collecting GitPython==3.1.45 (from lumen-ci==0.1.0) 2025-10-10T01:28:05.3039959Z Downloading gitpython-3.1.45-py3-none-any.whl.metadata (13 kB) 2025-10-10T01:28:05.3304061Z Collecting docker==7.1.0 (from lumen-ci==0.1.0) 2025-10-10T01:28:05.3341504Z Downloading docker-7.1.0-py3-none-any.whl.metadata (3.8 kB) 2025-10-10T01:28:05.3421295Z Requirement already satisfied: pytest==7.3.2 in /opt/conda/envs/py_3.12/lib/python3.12/site-packages (from lumen-ci==0.1.0) (7.3.2) 2025-10-10T01:28:05.5149464Z Collecting uv==0.8.6 (from lumen-ci==0.1.0) 2025-10-10T01:28:05.5201560Z Downloading uv-0.8.6-py3-none-manylinux_2_17_x86_64.manylinux2014_x86_64.whl.metadata (11 kB) 2025-10-10T01:28:05.5290677Z Requirement already satisfied: requests>=2.26.0 in /opt/conda/envs/py_3.12/lib/python3.12/site-packages (from docker==7.1.0->lumen-ci==0.1.0) (2.32.5) 2025-10-10T01:28:05.5293707Z Requirement already satisfied: urllib3>=1.26.0 in /opt/conda/envs/py_3.12/lib/python3.12/site-packages (from docker==7.1.0->lumen-ci==0.1.0) (2.5.0) 2025-10-10T01:28:05.5678946Z Collecting gitdb<5,>=4.0.1 (from GitPython==3.1.45->lumen-ci==0.1.0) 2025-10-10T01:28:05.5740212Z Downloading gitdb-4.0.12-py3-none-any.whl.metadata (1.2 kB) 2025-10-10T01:28:05.5826213Z Requirement already satisfied: iniconfig in /opt/conda/envs/py_3.12/lib/python3.12/site-packages (from pytest==7.3.2->lumen-ci==0.1.0) (2.1.0) 2025-10-10T01:28:05.5828924Z Requirement already satisfied: packaging in /opt/conda/envs/py_3.12/lib/python3.12/site-packages (from pytest==7.3.2->lumen-ci==0.1.0) (23.1) 2025-10-10T01:28:05.5833848Z Requirement already satisfied: pluggy<2.0,>=0.12 in /opt/conda/envs/py_3.12/lib/python3.12/site-packages (from pytest==7.3.2->lumen-ci==0.1.0) (1.6.0) 2025-10-10T01:28:05.5984534Z Collecting smmap<6,>=3.0.1 (from gitdb<5,>=4.0.1->GitPython==3.1.45->lumen-ci==0.1.0) 2025-10-10T01:28:05.6021851Z Downloading smmap-5.0.2-py3-none-any.whl.metadata (4.3 kB) 2025-10-10T01:28:05.6116350Z Requirement already satisfied: charset_normalizer<4,>=2 in /opt/conda/envs/py_3.12/lib/python3.12/site-packages (from requests>=2.26.0->docker==7.1.0->lumen-ci==0.1.0) (3.4.3) 2025-10-10T01:28:05.6120188Z Requirement already satisfied: idna<4,>=2.5 in /opt/conda/envs/py_3.12/lib/python3.12/site-packages (from requests>=2.26.0->docker==7.1.0->lumen-ci==0.1.0) (3.10) 2025-10-10T01:28:05.6125357Z Requirement already satisfied: certifi>=2017.4.17 in /opt/conda/envs/py_3.12/lib/python3.12/site-packages (from requests>=2.26.0->docker==7.1.0->lumen-ci==0.1.0) (2025.10.5) 2025-10-10T01:28:05.6216408Z Downloading docker-7.1.0-py3-none-any.whl (147 kB) 2025-10-10T01:28:05.6366450Z Downloading gitpython-3.1.45-py3-none-any.whl (208 kB) 2025-10-10T01:28:05.6541073Z Downloading uv-0.8.6-py3-none-manylinux_2_17_x86_64.manylinux2014_x86_64.whl (19.3 MB) 2025-10-10T01:28:05.8583395Z [?25l ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 0.0/19.3 MB ? eta -:--:-- 2025-10-10T01:28:05.9283889Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╸━━━━━━━ 15.7/19.3 MB 79.4 MB/s eta 0:00:01 2025-10-10T01:28:05.9285232Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 19.3/19.3 MB 70.2 MB/s 0:00:00 2025-10-10T01:28:05.9632031Z [?25hDownloading gitdb-4.0.12-py3-none-any.whl (62 kB) 2025-10-10T01:28:05.9750342Z Downloading smmap-5.0.2-py3-none-any.whl (24 kB) 2025-10-10T01:28:05.9978789Z Building wheels for collected packages: lumen-ci 2025-10-10T01:28:06.1825837Z Building editable for lumen-ci (pyproject.toml) ... [?25l- done 2025-10-10T01:28:06.1834047Z [?25h Created wheel for lumen-ci: filename=lumen_ci-0.1.0-0.editable-py3-none-any.whl size=2709 sha256=1cafb68be91db27e45c00e212d9a0b41c7fe7e4c336b1ab0a093182fbd11ec32 2025-10-10T01:28:06.1835203Z Stored in directory: /tmp/pip-ephem-wheel-cache-_3akuzgb/wheels/6d/29/6c/82214c85262c5bdb1d615291039f4ec9f974034103e7a303fa 2025-10-10T01:28:06.1849895Z Successfully built lumen-ci 2025-10-10T01:28:06.3442134Z Installing collected packages: uv, smmap, gitdb, docker, GitPython, lumen-ci 2025-10-10T01:28:06.5129352Z [?25l 2025-10-10T01:28:06.6813388Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 0/6 [uv] 2025-10-10T01:28:06.7723391Z  ━━━━━━━━━━━━━━━━━━━━╺━━━━━━━━━━━━━━━━━━━ 3/6 [docker] 2025-10-10T01:28:06.7724134Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 6/6 [lumen-ci] 2025-10-10T01:28:06.7724483Z [?25h 2025-10-10T01:28:06.7815564Z Successfully installed GitPython-3.1.45 docker-7.1.0 gitdb-4.0.12 lumen-ci-0.1.0 smmap-5.0.2 uv-0.8.6 2025-10-10T01:28:06.8712245Z + python -m cli.run test external vllm --test-plan vllm_distributed_test_28_failure_test --shard-id 1 --num-shards 1 2025-10-10T01:28:07.0733868Z 2025-10-10 01:28:07,072 [INFO] cli.lib.core.vllm.vllm_test: Display VllmTestParameters VllmTestParameters(torch_whls_path=PosixPath('/var/lib/jenkins/workspace/dist'), vllm_whls_path=PosixPath('/var/lib/jenkins/workspace/dist/external/vllm/wheels'), torch_cuda_arch_list='8.9', cleaning_script=PosixPath('/var/lib/jenkins/workspace/.github/ci_configs/vllm/use_existing_torch.py')) 2025-10-10T01:28:07.0736973Z 2025-10-10 01:28:07,072 [INFO] cli.lib.common.git_helper: Cloning vllm to vllm 2025-10-10T01:28:07.3750864Z 2025-10-10 01:28:07,374 [INFO] cli.lib.common.git_helper: Progress: 10% - remote: Counting objects: 10% (7/64) 2025-10-10T01:28:07.3752245Z 2025-10-10 01:28:07,374 [INFO] cli.lib.common.git_helper: Progress: 15% - remote: Counting objects: 15% (10/64) 2025-10-10T01:28:07.3753485Z 2025-10-10 01:28:07,374 [INFO] cli.lib.common.git_helper: Progress: 20% - remote: Counting objects: 20% (13/64) 2025-10-10T01:28:07.3754669Z 2025-10-10 01:28:07,374 [INFO] cli.lib.common.git_helper: Progress: 25% - remote: Counting objects: 25% (16/64) 2025-10-10T01:28:07.3755245Z 2025-10-10 01:28:07,375 [INFO] cli.lib.common.git_helper: Progress: 35% - remote: Counting objects: 35% (23/64) 2025-10-10T01:28:07.3755753Z 2025-10-10 01:28:07,375 [INFO] cli.lib.common.git_helper: Progress: 40% - remote: Counting objects: 40% (26/64) 2025-10-10T01:28:07.3756254Z 2025-10-10 01:28:07,375 [INFO] cli.lib.common.git_helper: Progress: 45% - remote: Counting objects: 45% (29/64) 2025-10-10T01:28:07.3756743Z 2025-10-10 01:28:07,375 [INFO] cli.lib.common.git_helper: Progress: 50% - remote: Counting objects: 50% (32/64) 2025-10-10T01:28:07.3757256Z 2025-10-10 01:28:07,375 [INFO] cli.lib.common.git_helper: Progress: 60% - remote: Counting objects: 60% (39/64) 2025-10-10T01:28:07.3757764Z 2025-10-10 01:28:07,375 [INFO] cli.lib.common.git_helper: Progress: 65% - remote: Counting objects: 65% (42/64) 2025-10-10T01:28:07.3758468Z 2025-10-10 01:28:07,375 [INFO] cli.lib.common.git_helper: Progress: 70% - remote: Counting objects: 70% (45/64) 2025-10-10T01:28:07.3759040Z 2025-10-10 01:28:07,375 [INFO] cli.lib.common.git_helper: Progress: 75% - remote: Counting objects: 75% (48/64) 2025-10-10T01:28:07.3759532Z 2025-10-10 01:28:07,375 [INFO] cli.lib.common.git_helper: Progress: 85% - remote: Counting objects: 85% (55/64) 2025-10-10T01:28:07.3760034Z 2025-10-10 01:28:07,375 [INFO] cli.lib.common.git_helper: Progress: 90% - remote: Counting objects: 90% (58/64) 2025-10-10T01:28:07.3760559Z 2025-10-10 01:28:07,375 [INFO] cli.lib.common.git_helper: Progress: 95% - remote: Counting objects: 95% (61/64) 2025-10-10T01:28:07.3761076Z 2025-10-10 01:28:07,375 [INFO] cli.lib.common.git_helper: Progress: 100% - remote: Counting objects: 100% (64/64) 2025-10-10T01:28:07.3761763Z 2025-10-10 01:28:07,375 [INFO] cli.lib.common.git_helper: Progress: 5% - remote: Compressing objects: 5% (3/54) 2025-10-10T01:28:07.3872334Z 2025-10-10 01:28:07,386 [INFO] cli.lib.common.git_helper: Progress: 20% - remote: Compressing objects: 20% (11/54) 2025-10-10T01:28:07.3898937Z 2025-10-10 01:28:07,389 [INFO] cli.lib.common.git_helper: Progress: 25% - remote: Compressing objects: 25% (14/54) 2025-10-10T01:28:07.3919508Z 2025-10-10 01:28:07,391 [INFO] cli.lib.common.git_helper: Progress: 35% - remote: Compressing objects: 35% (19/54) 2025-10-10T01:28:07.3920815Z 2025-10-10 01:28:07,391 [INFO] cli.lib.common.git_helper: Progress: 40% - remote: Compressing objects: 40% (22/54) 2025-10-10T01:28:07.3922049Z 2025-10-10 01:28:07,391 [INFO] cli.lib.common.git_helper: Progress: 50% - remote: Compressing objects: 50% (27/54) 2025-10-10T01:28:07.3923353Z 2025-10-10 01:28:07,391 [INFO] cli.lib.common.git_helper: Progress: 55% - remote: Compressing objects: 55% (30/54) 2025-10-10T01:28:07.3925288Z 2025-10-10 01:28:07,391 [INFO] cli.lib.common.git_helper: Progress: 70% - remote: Compressing objects: 70% (38/54) 2025-10-10T01:28:07.3927447Z 2025-10-10 01:28:07,392 [INFO] cli.lib.common.git_helper: Progress: 75% - remote: Compressing objects: 75% (41/54) 2025-10-10T01:28:07.3928705Z 2025-10-10 01:28:07,392 [INFO] cli.lib.common.git_helper: Progress: 85% - remote: Compressing objects: 85% (46/54) 2025-10-10T01:28:07.3929335Z 2025-10-10 01:28:07,392 [INFO] cli.lib.common.git_helper: Progress: 90% - remote: Compressing objects: 90% (49/54) 2025-10-10T01:28:07.3929871Z 2025-10-10 01:28:07,392 [INFO] cli.lib.common.git_helper: Progress: 100% - remote: Compressing objects: 100% (54/54) 2025-10-10T01:28:07.4157955Z 2025-10-10 01:28:07,415 [INFO] cli.lib.common.git_helper: Progress: 0% - Receiving objects: 0% (1/126974) 2025-10-10T01:28:07.5072031Z 2025-10-10 01:28:07,506 [INFO] cli.lib.common.git_helper: Progress: 5% - Receiving objects: 5% (6349/126974) 2025-10-10T01:28:07.6566844Z 2025-10-10 01:28:07,656 [INFO] cli.lib.common.git_helper: Progress: 10% - Receiving objects: 10% (12698/126974) 2025-10-10T01:28:07.9071875Z 2025-10-10 01:28:07,906 [INFO] cli.lib.common.git_helper: Progress: 15% - Receiving objects: 15% (19047/126974) 2025-10-10T01:28:08.0489120Z 2025-10-10 01:28:08,048 [INFO] cli.lib.common.git_helper: Progress: 20% - Receiving objects: 20% (25395/126974), 27.63 MiB | 55.24 MiB/s 2025-10-10T01:28:08.1863960Z 2025-10-10 01:28:08,185 [INFO] cli.lib.common.git_helper: Progress: 25% - Receiving objects: 25% (31744/126974), 27.63 MiB | 55.24 MiB/s 2025-10-10T01:28:08.2892137Z 2025-10-10 01:28:08,288 [INFO] cli.lib.common.git_helper: Progress: 30% - Receiving objects: 30% (38093/126974), 27.63 MiB | 55.24 MiB/s 2025-10-10T01:28:08.3842822Z 2025-10-10 01:28:08,383 [INFO] cli.lib.common.git_helper: Progress: 35% - Receiving objects: 35% (44441/126974), 27.63 MiB | 55.24 MiB/s 2025-10-10T01:28:08.4820126Z 2025-10-10 01:28:08,481 [INFO] cli.lib.common.git_helper: Progress: 40% - Receiving objects: 40% (50790/126974), 56.26 MiB | 56.25 MiB/s 2025-10-10T01:28:08.5166297Z 2025-10-10 01:28:08,516 [INFO] cli.lib.common.git_helper: Progress: 45% - Receiving objects: 45% (57139/126974), 56.26 MiB | 56.25 MiB/s 2025-10-10T01:28:08.5530588Z 2025-10-10 01:28:08,552 [INFO] cli.lib.common.git_helper: Progress: 50% - Receiving objects: 50% (63487/126974), 56.26 MiB | 56.25 MiB/s 2025-10-10T01:28:08.5916645Z 2025-10-10 01:28:08,591 [INFO] cli.lib.common.git_helper: Progress: 55% - Receiving objects: 55% (69836/126974), 56.26 MiB | 56.25 MiB/s 2025-10-10T01:28:08.6530617Z 2025-10-10 01:28:08,652 [INFO] cli.lib.common.git_helper: Progress: 60% - Receiving objects: 60% (76185/126974), 56.26 MiB | 56.25 MiB/s 2025-10-10T01:28:08.7232428Z 2025-10-10 01:28:08,722 [INFO] cli.lib.common.git_helper: Progress: 65% - Receiving objects: 65% (82534/126974), 56.26 MiB | 56.25 MiB/s 2025-10-10T01:28:08.7861663Z 2025-10-10 01:28:08,785 [INFO] cli.lib.common.git_helper: Progress: 70% - Receiving objects: 70% (88882/126974), 56.26 MiB | 56.25 MiB/s 2025-10-10T01:28:08.8316634Z 2025-10-10 01:28:08,831 [INFO] cli.lib.common.git_helper: Progress: 75% - Receiving objects: 75% (95231/126974), 56.26 MiB | 56.25 MiB/s 2025-10-10T01:28:08.9225969Z 2025-10-10 01:28:08,922 [INFO] cli.lib.common.git_helper: Progress: 80% - Receiving objects: 80% (101580/126974), 86.46 MiB | 57.63 MiB/s 2025-10-10T01:28:08.9654594Z 2025-10-10 01:28:08,964 [INFO] cli.lib.common.git_helper: Progress: 85% - Receiving objects: 85% (107928/126974), 86.46 MiB | 57.63 MiB/s 2025-10-10T01:28:08.9983436Z 2025-10-10 01:28:08,997 [INFO] cli.lib.common.git_helper: Progress: 90% - Receiving objects: 90% (114277/126974), 86.46 MiB | 57.63 MiB/s 2025-10-10T01:28:09.0447206Z 2025-10-10 01:28:09,044 [INFO] cli.lib.common.git_helper: Progress: 95% - Receiving objects: 95% (120626/126974), 86.46 MiB | 57.63 MiB/s 2025-10-10T01:28:09.0768845Z 2025-10-10 01:28:09,076 [INFO] cli.lib.common.git_helper: Progress: 100% - Receiving objects: 100% (126974/126974), 86.46 MiB | 57.63 MiB/s 2025-10-10T01:28:09.0965645Z 2025-10-10 01:28:09,096 [INFO] cli.lib.common.git_helper: Resolving deltas: 0% (0/100681) 2025-10-10T01:28:09.1117865Z 2025-10-10 01:28:09,111 [INFO] cli.lib.common.git_helper: Progress: 5% - Resolving deltas: 5% (5035/100681) 2025-10-10T01:28:09.1317102Z 2025-10-10 01:28:09,131 [INFO] cli.lib.common.git_helper: Progress: 10% - Resolving deltas: 10% (10069/100681) 2025-10-10T01:28:09.1513795Z 2025-10-10 01:28:09,151 [INFO] cli.lib.common.git_helper: Progress: 15% - Resolving deltas: 15% (15103/100681) 2025-10-10T01:28:09.1724633Z 2025-10-10 01:28:09,172 [INFO] cli.lib.common.git_helper: Progress: 20% - Resolving deltas: 20% (20137/100681) 2025-10-10T01:28:09.1982937Z 2025-10-10 01:28:09,198 [INFO] cli.lib.common.git_helper: Progress: 25% - Resolving deltas: 25% (25171/100681) 2025-10-10T01:28:09.2192616Z 2025-10-10 01:28:09,218 [INFO] cli.lib.common.git_helper: Progress: 30% - Resolving deltas: 30% (30205/100681) 2025-10-10T01:28:09.2439041Z 2025-10-10 01:28:09,243 [INFO] cli.lib.common.git_helper: Progress: 35% - Resolving deltas: 35% (35239/100681) 2025-10-10T01:28:09.2684438Z 2025-10-10 01:28:09,267 [INFO] cli.lib.common.git_helper: Progress: 40% - Resolving deltas: 40% (40273/100681) 2025-10-10T01:28:09.3024973Z 2025-10-10 01:28:09,302 [INFO] cli.lib.common.git_helper: Progress: 45% - Resolving deltas: 45% (45307/100681) 2025-10-10T01:28:09.3402900Z 2025-10-10 01:28:09,339 [INFO] cli.lib.common.git_helper: Progress: 50% - Resolving deltas: 50% (50341/100681) 2025-10-10T01:28:09.3700304Z 2025-10-10 01:28:09,369 [INFO] cli.lib.common.git_helper: Progress: 55% - Resolving deltas: 55% (55375/100681) 2025-10-10T01:28:09.3990725Z 2025-10-10 01:28:09,398 [INFO] cli.lib.common.git_helper: Progress: 60% - Resolving deltas: 60% (60409/100681) 2025-10-10T01:28:09.4367221Z 2025-10-10 01:28:09,436 [INFO] cli.lib.common.git_helper: Progress: 65% - Resolving deltas: 65% (65443/100681) 2025-10-10T01:28:09.4687754Z 2025-10-10 01:28:09,468 [INFO] cli.lib.common.git_helper: Progress: 70% - Resolving deltas: 70% (70478/100681) 2025-10-10T01:28:09.4920872Z 2025-10-10 01:28:09,491 [INFO] cli.lib.common.git_helper: Progress: 75% - Resolving deltas: 75% (75511/100681) 2025-10-10T01:28:09.5212229Z 2025-10-10 01:28:09,520 [INFO] cli.lib.common.git_helper: Progress: 80% - Resolving deltas: 80% (80545/100681) 2025-10-10T01:28:09.5404796Z 2025-10-10 01:28:09,540 [INFO] cli.lib.common.git_helper: Progress: 85% - Resolving deltas: 85% (85579/100681) 2025-10-10T01:28:09.5578942Z 2025-10-10 01:28:09,557 [INFO] cli.lib.common.git_helper: Progress: 90% - Resolving deltas: 90% (90613/100681) 2025-10-10T01:28:09.5878288Z 2025-10-10 01:28:09,587 [INFO] cli.lib.common.git_helper: Progress: 95% - Resolving deltas: 95% (95647/100681) 2025-10-10T01:28:09.6048700Z 2025-10-10 01:28:09,604 [INFO] cli.lib.common.git_helper: Progress: 100% - Resolving deltas: 100% (100681/100681) 2025-10-10T01:28:10.2543303Z 2025-10-10 01:28:10,253 [INFO] cli.lib.common.git_helper: Checking out pinned vllm commit 0ad9951c416d33c5da4f7a504fb162cbe62386f5 2025-10-10T01:28:10.5454916Z 2025-10-10 01:28:10,544 [INFO] cli.lib.common.git_helper: Successfully cloned vllm 2025-10-10T01:28:10.5848537Z 2025-10-10 01:28:10,584 [INFO] cli.lib.core.vllm.vllm_test: Running vllm test with inputs: VllmTestParameters(torch_whls_path=PosixPath('/var/lib/jenkins/workspace/dist'), vllm_whls_path=PosixPath('/var/lib/jenkins/workspace/dist/external/vllm/wheels'), torch_cuda_arch_list='8.9', cleaning_script=PosixPath('/var/lib/jenkins/workspace/.github/ci_configs/vllm/use_existing_torch.py')) 2025-10-10T01:28:10.5865809Z 2025-10-10 01:28:10,586 [INFO] cli.lib.common.pip_helper: torch already exist with version: 2.10.0a0+git344e636 2025-10-10T01:28:10.5871458Z 2025-10-10 01:28:10,586 [INFO] cli.lib.common.pip_helper: Installing /var/lib/jenkins/workspace/dist/vision/torchvision-0.22.0a0+966da7e-cp312-cp312-linux_x86_64.whl... 2025-10-10T01:28:10.5874279Z 2025-10-10 01:28:10,586 [INFO] cli.lib.common.pip_helper: pip installing packages: /opt/conda/envs/py_3.12/bin/python -m pip install /var/lib/jenkins/workspace/dist/vision/torchvision-0.22.0a0+966da7e-cp312-cp312-linux_x86_64.whl 2025-10-10T01:28:10.5875858Z 2025-10-10 01:28:10,586 [INFO] cli.lib.common.utils: [cmd] /opt/conda/envs/py_3.12/bin/python -m pip install /var/lib/jenkins/workspace/dist/vision/torchvision-0.22.0a0+966da7e-cp312-cp312-linux_x86_64.whl 2025-10-10T01:28:10.8812516Z Processing /var/lib/jenkins/workspace/dist/vision/torchvision-0.22.0a0+966da7e-cp312-cp312-linux_x86_64.whl 2025-10-10T01:28:10.8888431Z Requirement already satisfied: numpy in /opt/conda/envs/py_3.12/lib/python3.12/site-packages (from torchvision==0.22.0a0+966da7e) (1.26.2) 2025-10-10T01:28:10.8891052Z Requirement already satisfied: torch in /opt/conda/envs/py_3.12/lib/python3.12/site-packages (from torchvision==0.22.0a0+966da7e) (2.10.0a0+git344e636) 2025-10-10T01:28:10.8894341Z Requirement already satisfied: pillow!=8.3.*,>=5.3.0 in /opt/conda/envs/py_3.12/lib/python3.12/site-packages (from torchvision==0.22.0a0+966da7e) (11.0.0) 2025-10-10T01:28:10.8944819Z Requirement already satisfied: filelock in /opt/conda/envs/py_3.12/lib/python3.12/site-packages (from torch->torchvision==0.22.0a0+966da7e) (3.18.0) 2025-10-10T01:28:10.8948622Z Requirement already satisfied: typing-extensions>=4.10.0 in /opt/conda/envs/py_3.12/lib/python3.12/site-packages (from torch->torchvision==0.22.0a0+966da7e) (4.12.2) 2025-10-10T01:28:10.8953380Z Requirement already satisfied: setuptools in /opt/conda/envs/py_3.12/lib/python3.12/site-packages (from torch->torchvision==0.22.0a0+966da7e) (78.1.1) 2025-10-10T01:28:10.8955927Z Requirement already satisfied: sympy>=1.13.3 in /opt/conda/envs/py_3.12/lib/python3.12/site-packages (from torch->torchvision==0.22.0a0+966da7e) (1.13.3) 2025-10-10T01:28:10.8958353Z Requirement already satisfied: networkx>=2.5.1 in /opt/conda/envs/py_3.12/lib/python3.12/site-packages (from torch->torchvision==0.22.0a0+966da7e) (2.8.8) 2025-10-10T01:28:10.8965917Z Requirement already satisfied: jinja2 in /opt/conda/envs/py_3.12/lib/python3.12/site-packages (from torch->torchvision==0.22.0a0+966da7e) (3.1.6) 2025-10-10T01:28:10.8971539Z Requirement already satisfied: fsspec>=0.8.5 in /opt/conda/envs/py_3.12/lib/python3.12/site-packages (from torch->torchvision==0.22.0a0+966da7e) (2025.9.0) 2025-10-10T01:28:10.9070959Z Requirement already satisfied: mpmath<1.4,>=1.1.0 in /opt/conda/envs/py_3.12/lib/python3.12/site-packages (from sympy>=1.13.3->torch->torchvision==0.22.0a0+966da7e) (1.3.0) 2025-10-10T01:28:10.9109798Z Requirement already satisfied: MarkupSafe>=2.0 in /opt/conda/envs/py_3.12/lib/python3.12/site-packages (from jinja2->torch->torchvision==0.22.0a0+966da7e) (3.0.3) 2025-10-10T01:28:11.0715798Z Installing collected packages: torchvision 2025-10-10T01:28:11.4922934Z Successfully installed torchvision-0.22.0a0+966da7e 2025-10-10T01:28:11.5362723Z 2025-10-10 01:28:11,535 [INFO] cli.lib.common.pip_helper: Installing /var/lib/jenkins/workspace/dist/audio/torchaudio-2.8.0a0+87ff22e-cp312-cp312-linux_x86_64.whl... 2025-10-10T01:28:11.5365550Z 2025-10-10 01:28:11,535 [INFO] cli.lib.common.pip_helper: pip installing packages: /opt/conda/envs/py_3.12/bin/python -m pip install /var/lib/jenkins/workspace/dist/audio/torchaudio-2.8.0a0+87ff22e-cp312-cp312-linux_x86_64.whl 2025-10-10T01:28:11.5367078Z 2025-10-10 01:28:11,535 [INFO] cli.lib.common.utils: [cmd] /opt/conda/envs/py_3.12/bin/python -m pip install /var/lib/jenkins/workspace/dist/audio/torchaudio-2.8.0a0+87ff22e-cp312-cp312-linux_x86_64.whl 2025-10-10T01:28:11.8300137Z Processing /var/lib/jenkins/workspace/dist/audio/torchaudio-2.8.0a0+87ff22e-cp312-cp312-linux_x86_64.whl 2025-10-10T01:28:11.8360736Z Requirement already satisfied: torch in /opt/conda/envs/py_3.12/lib/python3.12/site-packages (from torchaudio==2.8.0a0+87ff22e) (2.10.0a0+git344e636) 2025-10-10T01:28:11.8368734Z Requirement already satisfied: filelock in /opt/conda/envs/py_3.12/lib/python3.12/site-packages (from torch->torchaudio==2.8.0a0+87ff22e) (3.18.0) 2025-10-10T01:28:11.8372795Z Requirement already satisfied: typing-extensions>=4.10.0 in /opt/conda/envs/py_3.12/lib/python3.12/site-packages (from torch->torchaudio==2.8.0a0+87ff22e) (4.12.2) 2025-10-10T01:28:11.8378147Z Requirement already satisfied: setuptools in /opt/conda/envs/py_3.12/lib/python3.12/site-packages (from torch->torchaudio==2.8.0a0+87ff22e) (78.1.1) 2025-10-10T01:28:11.8381792Z Requirement already satisfied: sympy>=1.13.3 in /opt/conda/envs/py_3.12/lib/python3.12/site-packages (from torch->torchaudio==2.8.0a0+87ff22e) (1.13.3) 2025-10-10T01:28:11.8385316Z Requirement already satisfied: networkx>=2.5.1 in /opt/conda/envs/py_3.12/lib/python3.12/site-packages (from torch->torchaudio==2.8.0a0+87ff22e) (2.8.8) 2025-10-10T01:28:11.8387879Z Requirement already satisfied: jinja2 in /opt/conda/envs/py_3.12/lib/python3.12/site-packages (from torch->torchaudio==2.8.0a0+87ff22e) (3.1.6) 2025-10-10T01:28:11.8391351Z Requirement already satisfied: fsspec>=0.8.5 in /opt/conda/envs/py_3.12/lib/python3.12/site-packages (from torch->torchaudio==2.8.0a0+87ff22e) (2025.9.0) 2025-10-10T01:28:11.8498796Z Requirement already satisfied: mpmath<1.4,>=1.1.0 in /opt/conda/envs/py_3.12/lib/python3.12/site-packages (from sympy>=1.13.3->torch->torchaudio==2.8.0a0+87ff22e) (1.3.0) 2025-10-10T01:28:11.8539044Z Requirement already satisfied: MarkupSafe>=2.0 in /opt/conda/envs/py_3.12/lib/python3.12/site-packages (from jinja2->torch->torchaudio==2.8.0a0+87ff22e) (3.0.3) 2025-10-10T01:28:12.0173078Z Installing collected packages: torchaudio 2025-10-10T01:28:12.1666016Z Successfully installed torchaudio-2.8.0a0+87ff22e 2025-10-10T01:28:12.2191892Z 2025-10-10 01:28:12,218 [INFO] cli.lib.core.vllm.vllm_test: Done. Installed torch and other torch-related wheels 2025-10-10T01:28:12.2193152Z 2025-10-10 01:28:12,218 [INFO] cli.lib.core.vllm.vllm_test: Installing vllm wheels 2025-10-10T01:28:12.2196738Z 2025-10-10 01:28:12,219 [INFO] cli.lib.common.pip_helper: Installing /var/lib/jenkins/workspace/dist/external/vllm/wheels/xformers/xformers-0.0.33+5d4b92a5.d20251010-cp39-abi3-linux_x86_64.whl... 2025-10-10T01:28:12.2198144Z 2025-10-10 01:28:12,219 [INFO] cli.lib.common.pip_helper: pip installing packages: /opt/conda/envs/py_3.12/bin/python -m pip install /var/lib/jenkins/workspace/dist/external/vllm/wheels/xformers/xformers-0.0.33+5d4b92a5.d20251010-cp39-abi3-linux_x86_64.whl 2025-10-10T01:28:12.2200242Z 2025-10-10 01:28:12,219 [INFO] cli.lib.common.utils: [cmd] /opt/conda/envs/py_3.12/bin/python -m pip install /var/lib/jenkins/workspace/dist/external/vllm/wheels/xformers/xformers-0.0.33+5d4b92a5.d20251010-cp39-abi3-linux_x86_64.whl 2025-10-10T01:28:12.5174784Z Processing /var/lib/jenkins/workspace/dist/external/vllm/wheels/xformers/xformers-0.0.33+5d4b92a5.d20251010-cp39-abi3-linux_x86_64.whl 2025-10-10T01:28:12.7068296Z Requirement already satisfied: torch>=2.8 in /opt/conda/envs/py_3.12/lib/python3.12/site-packages (from xformers==0.0.33+5d4b92a5.d20251010) (2.10.0a0+git344e636) 2025-10-10T01:28:12.7071460Z Requirement already satisfied: numpy in /opt/conda/envs/py_3.12/lib/python3.12/site-packages (from xformers==0.0.33+5d4b92a5.d20251010) (1.26.2) 2025-10-10T01:28:12.7080453Z Requirement already satisfied: filelock in /opt/conda/envs/py_3.12/lib/python3.12/site-packages (from torch>=2.8->xformers==0.0.33+5d4b92a5.d20251010) (3.18.0) 2025-10-10T01:28:12.7084837Z Requirement already satisfied: typing-extensions>=4.10.0 in /opt/conda/envs/py_3.12/lib/python3.12/site-packages (from torch>=2.8->xformers==0.0.33+5d4b92a5.d20251010) (4.12.2) 2025-10-10T01:28:12.7092306Z Requirement already satisfied: setuptools in /opt/conda/envs/py_3.12/lib/python3.12/site-packages (from torch>=2.8->xformers==0.0.33+5d4b92a5.d20251010) (78.1.1) 2025-10-10T01:28:12.7096141Z Requirement already satisfied: sympy>=1.13.3 in /opt/conda/envs/py_3.12/lib/python3.12/site-packages (from torch>=2.8->xformers==0.0.33+5d4b92a5.d20251010) (1.13.3) 2025-10-10T01:28:12.7100439Z Requirement already satisfied: networkx>=2.5.1 in /opt/conda/envs/py_3.12/lib/python3.12/site-packages (from torch>=2.8->xformers==0.0.33+5d4b92a5.d20251010) (2.8.8) 2025-10-10T01:28:12.7103356Z Requirement already satisfied: jinja2 in /opt/conda/envs/py_3.12/lib/python3.12/site-packages (from torch>=2.8->xformers==0.0.33+5d4b92a5.d20251010) (3.1.6) 2025-10-10T01:28:12.7107765Z Requirement already satisfied: fsspec>=0.8.5 in /opt/conda/envs/py_3.12/lib/python3.12/site-packages (from torch>=2.8->xformers==0.0.33+5d4b92a5.d20251010) (2025.9.0) 2025-10-10T01:28:12.7217762Z Requirement already satisfied: mpmath<1.4,>=1.1.0 in /opt/conda/envs/py_3.12/lib/python3.12/site-packages (from sympy>=1.13.3->torch>=2.8->xformers==0.0.33+5d4b92a5.d20251010) (1.3.0) 2025-10-10T01:28:12.7258551Z Requirement already satisfied: MarkupSafe>=2.0 in /opt/conda/envs/py_3.12/lib/python3.12/site-packages (from jinja2->torch>=2.8->xformers==0.0.33+5d4b92a5.d20251010) (3.0.3) 2025-10-10T01:28:12.8986161Z Installing collected packages: xformers 2025-10-10T01:28:16.5780143Z Successfully installed xformers-0.0.33+5d4b92a5.d20251010 2025-10-10T01:28:16.6273257Z 2025-10-10 01:28:16,626 [INFO] cli.lib.common.pip_helper: Installing /var/lib/jenkins/workspace/dist/external/vllm/wheels/vllm/vllm-0.11.0rc2.dev157+g0ad9951c4.d20251010-cp38-abi3-linux_x86_64.whl... 2025-10-10T01:28:16.6276814Z 2025-10-10 01:28:16,626 [INFO] cli.lib.common.pip_helper: pip installing packages: /opt/conda/envs/py_3.12/bin/python -m pip install /var/lib/jenkins/workspace/dist/external/vllm/wheels/vllm/vllm-0.11.0rc2.dev157+g0ad9951c4.d20251010-cp38-abi3-linux_x86_64.whl 2025-10-10T01:28:16.6278823Z 2025-10-10 01:28:16,626 [INFO] cli.lib.common.utils: [cmd] /opt/conda/envs/py_3.12/bin/python -m pip install /var/lib/jenkins/workspace/dist/external/vllm/wheels/vllm/vllm-0.11.0rc2.dev157+g0ad9951c4.d20251010-cp38-abi3-linux_x86_64.whl 2025-10-10T01:28:16.9257196Z Processing /var/lib/jenkins/workspace/dist/external/vllm/wheels/vllm/vllm-0.11.0rc2.dev157+g0ad9951c4.d20251010-cp38-abi3-linux_x86_64.whl 2025-10-10T01:28:17.4823010Z Collecting regex (from vllm==0.11.0rc2.dev157+g0ad9951c4.d20251010) 2025-10-10T01:28:17.5023341Z Downloading regex-2025.9.18-cp312-cp312-manylinux2014_x86_64.manylinux_2_17_x86_64.manylinux_2_28_x86_64.whl.metadata (40 kB) 2025-10-10T01:28:17.5106117Z Requirement already satisfied: cachetools in /opt/conda/envs/py_3.12/lib/python3.12/site-packages (from vllm==0.11.0rc2.dev157+g0ad9951c4.d20251010) (6.2.0) 2025-10-10T01:28:17.5107828Z Requirement already satisfied: psutil in /opt/conda/envs/py_3.12/lib/python3.12/site-packages (from vllm==0.11.0rc2.dev157+g0ad9951c4.d20251010) (7.1.0) 2025-10-10T01:28:17.5474650Z Collecting sentencepiece (from vllm==0.11.0rc2.dev157+g0ad9951c4.d20251010) 2025-10-10T01:28:17.5515519Z Downloading sentencepiece-0.2.1-cp312-cp312-manylinux_2_27_x86_64.manylinux_2_28_x86_64.whl.metadata (10 kB) 2025-10-10T01:28:17.5570639Z Requirement already satisfied: numpy in /opt/conda/envs/py_3.12/lib/python3.12/site-packages (from vllm==0.11.0rc2.dev157+g0ad9951c4.d20251010) (1.26.2) 2025-10-10T01:28:17.5574134Z Requirement already satisfied: requests>=2.26.0 in /opt/conda/envs/py_3.12/lib/python3.12/site-packages (from vllm==0.11.0rc2.dev157+g0ad9951c4.d20251010) (2.32.5) 2025-10-10T01:28:17.5576463Z Requirement already satisfied: tqdm in /opt/conda/envs/py_3.12/lib/python3.12/site-packages (from vllm==0.11.0rc2.dev157+g0ad9951c4.d20251010) (4.67.1) 2025-10-10T01:28:17.6012044Z Collecting blake3 (from vllm==0.11.0rc2.dev157+g0ad9951c4.d20251010) 2025-10-10T01:28:17.6077262Z Downloading blake3-1.0.7-cp312-cp312-manylinux_2_17_x86_64.manylinux2014_x86_64.whl.metadata (217 bytes) 2025-10-10T01:28:17.6197225Z Collecting py-cpuinfo (from vllm==0.11.0rc2.dev157+g0ad9951c4.d20251010) 2025-10-10T01:28:17.6265042Z Downloading py_cpuinfo-9.0.0-py3-none-any.whl.metadata (794 bytes) 2025-10-10T01:28:17.6630844Z Collecting transformers>=4.55.2 (from vllm==0.11.0rc2.dev157+g0ad9951c4.d20251010) 2025-10-10T01:28:17.6670891Z Downloading transformers-4.57.0-py3-none-any.whl.metadata (41 kB) 2025-10-10T01:28:17.8620348Z Collecting tokenizers>=0.21.1 (from vllm==0.11.0rc2.dev157+g0ad9951c4.d20251010) 2025-10-10T01:28:17.8667846Z Downloading tokenizers-0.22.1-cp39-abi3-manylinux_2_17_x86_64.manylinux2014_x86_64.whl.metadata (6.8 kB) 2025-10-10T01:28:17.8723543Z Requirement already satisfied: protobuf in /opt/conda/envs/py_3.12/lib/python3.12/site-packages (from vllm==0.11.0rc2.dev157+g0ad9951c4.d20251010) (5.29.5) 2025-10-10T01:28:17.9045910Z Collecting fastapi>=0.115.0 (from fastapi[standard]>=0.115.0->vllm==0.11.0rc2.dev157+g0ad9951c4.d20251010) 2025-10-10T01:28:17.9098381Z Downloading fastapi-0.118.2-py3-none-any.whl.metadata (28 kB) 2025-10-10T01:28:17.9181510Z Requirement already satisfied: aiohttp in /opt/conda/envs/py_3.12/lib/python3.12/site-packages (from vllm==0.11.0rc2.dev157+g0ad9951c4.d20251010) (3.13.0) 2025-10-10T01:28:17.9771134Z Collecting openai>=1.99.1 (from vllm==0.11.0rc2.dev157+g0ad9951c4.d20251010) 2025-10-10T01:28:17.9815551Z Downloading openai-2.3.0-py3-none-any.whl.metadata (29 kB) 2025-10-10T01:28:18.0702255Z Collecting pydantic>=2.11.7 (from vllm==0.11.0rc2.dev157+g0ad9951c4.d20251010) 2025-10-10T01:28:18.0746995Z Downloading pydantic-2.12.0-py3-none-any.whl.metadata (83 kB) 2025-10-10T01:28:18.0937185Z Collecting prometheus_client>=0.18.0 (from vllm==0.11.0rc2.dev157+g0ad9951c4.d20251010) 2025-10-10T01:28:18.0991447Z Downloading prometheus_client-0.23.1-py3-none-any.whl.metadata (1.9 kB) 2025-10-10T01:28:18.1030361Z Requirement already satisfied: pillow in /opt/conda/envs/py_3.12/lib/python3.12/site-packages (from vllm==0.11.0rc2.dev157+g0ad9951c4.d20251010) (11.0.0) 2025-10-10T01:28:18.1163256Z Collecting prometheus-fastapi-instrumentator>=7.0.0 (from vllm==0.11.0rc2.dev157+g0ad9951c4.d20251010) 2025-10-10T01:28:18.1226737Z Downloading prometheus_fastapi_instrumentator-7.1.0-py3-none-any.whl.metadata (13 kB) 2025-10-10T01:28:18.1550319Z Collecting tiktoken>=0.6.0 (from vllm==0.11.0rc2.dev157+g0ad9951c4.d20251010) 2025-10-10T01:28:18.1611783Z Downloading tiktoken-0.12.0-cp312-cp312-manylinux_2_28_x86_64.whl.metadata (6.7 kB) 2025-10-10T01:28:18.1812301Z Collecting lm-format-enforcer==0.11.3 (from vllm==0.11.0rc2.dev157+g0ad9951c4.d20251010) 2025-10-10T01:28:18.1874483Z Downloading lm_format_enforcer-0.11.3-py3-none-any.whl.metadata (17 kB) 2025-10-10T01:28:18.2273693Z Collecting llguidance<0.8.0,>=0.7.11 (from vllm==0.11.0rc2.dev157+g0ad9951c4.d20251010) 2025-10-10T01:28:18.2327180Z Downloading llguidance-0.7.30-cp39-abi3-manylinux_2_17_x86_64.manylinux2014_x86_64.whl.metadata (10 kB) 2025-10-10T01:28:18.2590843Z Collecting outlines_core==0.2.11 (from vllm==0.11.0rc2.dev157+g0ad9951c4.d20251010) 2025-10-10T01:28:18.2646690Z Downloading outlines_core-0.2.11-cp312-cp312-manylinux_2_17_x86_64.manylinux2014_x86_64.whl.metadata (5.8 kB) 2025-10-10T01:28:18.2871924Z Collecting diskcache==5.6.3 (from vllm==0.11.0rc2.dev157+g0ad9951c4.d20251010) 2025-10-10T01:28:18.2925154Z Downloading diskcache-5.6.3-py3-none-any.whl.metadata (20 kB) 2025-10-10T01:28:18.3054509Z Collecting lark==1.2.2 (from vllm==0.11.0rc2.dev157+g0ad9951c4.d20251010) 2025-10-10T01:28:18.3107320Z Downloading lark-1.2.2-py3-none-any.whl.metadata (1.8 kB) 2025-10-10T01:28:18.3455041Z Collecting xgrammar==0.1.25 (from vllm==0.11.0rc2.dev157+g0ad9951c4.d20251010) 2025-10-10T01:28:18.3514670Z Downloading xgrammar-0.1.25-cp312-cp312-manylinux_2_17_x86_64.manylinux2014_x86_64.whl.metadata (6.5 kB) 2025-10-10T01:28:18.3563148Z Requirement already satisfied: typing_extensions>=4.10 in /opt/conda/envs/py_3.12/lib/python3.12/site-packages (from vllm==0.11.0rc2.dev157+g0ad9951c4.d20251010) (4.12.2) 2025-10-10T01:28:18.3565673Z Requirement already satisfied: filelock>=3.16.1 in /opt/conda/envs/py_3.12/lib/python3.12/site-packages (from vllm==0.11.0rc2.dev157+g0ad9951c4.d20251010) (3.18.0) 2025-10-10T01:28:18.3671615Z Collecting partial-json-parser (from vllm==0.11.0rc2.dev157+g0ad9951c4.d20251010) 2025-10-10T01:28:18.3735839Z Downloading partial_json_parser-0.2.1.1.post6-py3-none-any.whl.metadata (6.1 kB) 2025-10-10T01:28:18.5569693Z Collecting pyzmq>=25.0.0 (from vllm==0.11.0rc2.dev157+g0ad9951c4.d20251010) 2025-10-10T01:28:18.5629440Z Downloading pyzmq-27.1.0-cp312-abi3-manylinux_2_26_x86_64.manylinux_2_28_x86_64.whl.metadata (6.0 kB) 2025-10-10T01:28:18.6070385Z Collecting msgspec (from vllm==0.11.0rc2.dev157+g0ad9951c4.d20251010) 2025-10-10T01:28:18.6217874Z Downloading msgspec-0.19.0-cp312-cp312-manylinux_2_17_x86_64.manylinux2014_x86_64.whl.metadata (6.9 kB) 2025-10-10T01:28:18.6382957Z Collecting gguf>=0.13.0 (from vllm==0.11.0rc2.dev157+g0ad9951c4.d20251010) 2025-10-10T01:28:18.6442587Z Downloading gguf-0.17.1-py3-none-any.whl.metadata (4.3 kB) 2025-10-10T01:28:18.6596347Z Collecting mistral_common>=1.8.2 (from mistral_common[audio,image]>=1.8.2->vllm==0.11.0rc2.dev157+g0ad9951c4.d20251010) 2025-10-10T01:28:18.6655629Z Downloading mistral_common-1.8.5-py3-none-any.whl.metadata (5.1 kB) 2025-10-10T01:28:18.7439490Z Collecting opencv-python-headless>=4.11.0 (from vllm==0.11.0rc2.dev157+g0ad9951c4.d20251010) 2025-10-10T01:28:18.7498650Z Downloading opencv_python_headless-4.12.0.88-cp37-abi3-manylinux2014_x86_64.manylinux_2_17_x86_64.whl.metadata (19 kB) 2025-10-10T01:28:18.7559763Z Requirement already satisfied: pyyaml in /opt/conda/envs/py_3.12/lib/python3.12/site-packages (from vllm==0.11.0rc2.dev157+g0ad9951c4.d20251010) (6.0.2) 2025-10-10T01:28:18.7565063Z Requirement already satisfied: six>=1.16.0 in /opt/conda/envs/py_3.12/lib/python3.12/site-packages (from vllm==0.11.0rc2.dev157+g0ad9951c4.d20251010) (1.17.0) 2025-10-10T01:28:18.7568745Z Requirement already satisfied: setuptools<80,>=77.0.3 in /opt/conda/envs/py_3.12/lib/python3.12/site-packages (from vllm==0.11.0rc2.dev157+g0ad9951c4.d20251010) (78.1.1) 2025-10-10T01:28:18.7667688Z Collecting einops (from vllm==0.11.0rc2.dev157+g0ad9951c4.d20251010) 2025-10-10T01:28:18.7732956Z Downloading einops-0.8.1-py3-none-any.whl.metadata (13 kB) 2025-10-10T01:28:18.7947648Z Collecting compressed-tensors==0.11.0 (from vllm==0.11.0rc2.dev157+g0ad9951c4.d20251010) 2025-10-10T01:28:18.8006381Z Downloading compressed_tensors-0.11.0-py3-none-any.whl.metadata (7.0 kB) 2025-10-10T01:28:18.8146811Z Collecting depyf==0.19.0 (from vllm==0.11.0rc2.dev157+g0ad9951c4.d20251010) 2025-10-10T01:28:18.8204304Z Downloading depyf-0.19.0-py3-none-any.whl.metadata (7.3 kB) 2025-10-10T01:28:18.8387686Z Collecting cloudpickle (from vllm==0.11.0rc2.dev157+g0ad9951c4.d20251010) 2025-10-10T01:28:18.8422276Z Downloading cloudpickle-3.1.1-py3-none-any.whl.metadata (7.1 kB) 2025-10-10T01:28:18.9045383Z Collecting watchfiles (from vllm==0.11.0rc2.dev157+g0ad9951c4.d20251010) 2025-10-10T01:28:18.9114336Z Downloading watchfiles-1.1.0-cp312-cp312-manylinux_2_17_x86_64.manylinux2014_x86_64.whl.metadata (4.9 kB) 2025-10-10T01:28:18.9275578Z Collecting python-json-logger (from vllm==0.11.0rc2.dev157+g0ad9951c4.d20251010) 2025-10-10T01:28:18.9321008Z Downloading python_json_logger-4.0.0-py3-none-any.whl.metadata (4.0 kB) 2025-10-10T01:28:18.9372573Z Requirement already satisfied: scipy in /opt/conda/envs/py_3.12/lib/python3.12/site-packages (from vllm==0.11.0rc2.dev157+g0ad9951c4.d20251010) (1.14.1) 2025-10-10T01:28:18.9375845Z Requirement already satisfied: ninja in /opt/conda/envs/py_3.12/lib/python3.12/site-packages (from vllm==0.11.0rc2.dev157+g0ad9951c4.d20251010) (1.10.2) 2025-10-10T01:28:19.0045099Z Collecting pybase64 (from vllm==0.11.0rc2.dev157+g0ad9951c4.d20251010) 2025-10-10T01:28:19.0105551Z Downloading pybase64-1.4.2-cp312-cp312-manylinux1_x86_64.manylinux2014_x86_64.manylinux_2_17_x86_64.manylinux_2_5_x86_64.whl.metadata (8.7 kB) 2025-10-10T01:28:19.0457427Z Collecting cbor2 (from vllm==0.11.0rc2.dev157+g0ad9951c4.d20251010) 2025-10-10T01:28:19.0527138Z Downloading cbor2-5.7.0-cp312-cp312-manylinux2014_x86_64.manylinux_2_17_x86_64.manylinux_2_28_x86_64.whl.metadata (5.4 kB) 2025-10-10T01:28:19.0948313Z Collecting setproctitle (from vllm==0.11.0rc2.dev157+g0ad9951c4.d20251010) 2025-10-10T01:28:19.1053763Z Downloading setproctitle-1.3.7-cp312-cp312-manylinux1_x86_64.manylinux_2_28_x86_64.manylinux_2_5_x86_64.whl.metadata (10 kB) 2025-10-10T01:28:19.1430410Z Collecting openai-harmony>=0.0.3 (from vllm==0.11.0rc2.dev157+g0ad9951c4.d20251010) 2025-10-10T01:28:19.1501525Z Downloading openai_harmony-0.0.4-cp38-abi3-manylinux_2_17_x86_64.manylinux2014_x86_64.whl.metadata (8.0 kB) 2025-10-10T01:28:19.2224857Z Collecting numba==0.61.2 (from vllm==0.11.0rc2.dev157+g0ad9951c4.d20251010) 2025-10-10T01:28:19.2284557Z Downloading numba-0.61.2-cp312-cp312-manylinux2014_x86_64.manylinux_2_17_x86_64.whl.metadata (2.8 kB) 2025-10-10T01:28:19.3180254Z Collecting ray>=2.48.0 (from ray[cgraph]>=2.48.0->vllm==0.11.0rc2.dev157+g0ad9951c4.d20251010) 2025-10-10T01:28:19.3261747Z Downloading ray-2.49.2-cp312-cp312-manylinux2014_x86_64.whl.metadata (21 kB) 2025-10-10T01:28:19.3760579Z Requirement already satisfied: torch>=1.7.0 in /opt/conda/envs/py_3.12/lib/python3.12/site-packages (from compressed-tensors==0.11.0->vllm==0.11.0rc2.dev157+g0ad9951c4.d20251010) (2.10.0a0+git344e636) 2025-10-10T01:28:19.4115181Z Collecting frozendict (from compressed-tensors==0.11.0->vllm==0.11.0rc2.dev157+g0ad9951c4.d20251010) 2025-10-10T01:28:19.4187399Z Downloading frozendict-2.4.6-py312-none-any.whl.metadata (23 kB) 2025-10-10T01:28:19.4331872Z Collecting astor (from depyf==0.19.0->vllm==0.11.0rc2.dev157+g0ad9951c4.d20251010) 2025-10-10T01:28:19.4390415Z Downloading astor-0.8.1-py2.py3-none-any.whl.metadata (4.2 kB) 2025-10-10T01:28:19.4444871Z Requirement already satisfied: dill in /opt/conda/envs/py_3.12/lib/python3.12/site-packages (from depyf==0.19.0->vllm==0.11.0rc2.dev157+g0ad9951c4.d20251010) (0.3.7) 2025-10-10T01:28:19.4542152Z Collecting interegular>=0.3.2 (from lm-format-enforcer==0.11.3->vllm==0.11.0rc2.dev157+g0ad9951c4.d20251010) 2025-10-10T01:28:19.4600640Z Downloading interegular-0.3.3-py37-none-any.whl.metadata (3.0 kB) 2025-10-10T01:28:19.4646094Z Requirement already satisfied: packaging in /opt/conda/envs/py_3.12/lib/python3.12/site-packages (from lm-format-enforcer==0.11.3->vllm==0.11.0rc2.dev157+g0ad9951c4.d20251010) (23.1) 2025-10-10T01:28:19.5550376Z Collecting llvmlite<0.45,>=0.44.0dev0 (from numba==0.61.2->vllm==0.11.0rc2.dev157+g0ad9951c4.d20251010) 2025-10-10T01:28:19.5608838Z Downloading llvmlite-0.44.0-cp312-cp312-manylinux_2_17_x86_64.manylinux2014_x86_64.whl.metadata (5.0 kB) 2025-10-10T01:28:19.5752412Z Requirement already satisfied: triton in /opt/conda/envs/py_3.12/lib/python3.12/site-packages (from xgrammar==0.1.25->vllm==0.11.0rc2.dev157+g0ad9951c4.d20251010) (3.5.0+git7416ffcb) 2025-10-10T01:28:19.6076231Z Collecting starlette<0.49.0,>=0.40.0 (from fastapi>=0.115.0->fastapi[standard]>=0.115.0->vllm==0.11.0rc2.dev157+g0ad9951c4.d20251010) 2025-10-10T01:28:19.6117162Z Downloading starlette-0.48.0-py3-none-any.whl.metadata (6.3 kB) 2025-10-10T01:28:19.6543580Z Collecting annotated-types>=0.6.0 (from pydantic>=2.11.7->vllm==0.11.0rc2.dev157+g0ad9951c4.d20251010) 2025-10-10T01:28:19.6590306Z Downloading annotated_types-0.7.0-py3-none-any.whl.metadata (15 kB) 2025-10-10T01:28:20.2113471Z Collecting pydantic-core==2.41.1 (from pydantic>=2.11.7->vllm==0.11.0rc2.dev157+g0ad9951c4.d20251010) 2025-10-10T01:28:20.2165959Z Downloading pydantic_core-2.41.1-cp312-cp312-manylinux_2_17_x86_64.manylinux2014_x86_64.whl.metadata (7.3 kB) 2025-10-10T01:28:20.2344148Z Collecting typing_extensions>=4.10 (from vllm==0.11.0rc2.dev157+g0ad9951c4.d20251010) 2025-10-10T01:28:20.2385737Z Downloading typing_extensions-4.15.0-py3-none-any.whl.metadata (3.3 kB) 2025-10-10T01:28:20.2516221Z Collecting typing-inspection>=0.4.2 (from pydantic>=2.11.7->vllm==0.11.0rc2.dev157+g0ad9951c4.d20251010) 2025-10-10T01:28:20.2558443Z Downloading typing_inspection-0.4.2-py3-none-any.whl.metadata (2.6 kB) 2025-10-10T01:28:20.2779611Z Collecting anyio<5,>=3.6.2 (from starlette<0.49.0,>=0.40.0->fastapi>=0.115.0->fastapi[standard]>=0.115.0->vllm==0.11.0rc2.dev157+g0ad9951c4.d20251010) 2025-10-10T01:28:20.2819344Z Downloading anyio-4.11.0-py3-none-any.whl.metadata (4.1 kB) 2025-10-10T01:28:20.2877383Z Requirement already satisfied: idna>=2.8 in /opt/conda/envs/py_3.12/lib/python3.12/site-packages (from anyio<5,>=3.6.2->starlette<0.49.0,>=0.40.0->fastapi>=0.115.0->fastapi[standard]>=0.115.0->vllm==0.11.0rc2.dev157+g0ad9951c4.d20251010) (3.10) 2025-10-10T01:28:20.2965579Z Collecting sniffio>=1.1 (from anyio<5,>=3.6.2->starlette<0.49.0,>=0.40.0->fastapi>=0.115.0->fastapi[standard]>=0.115.0->vllm==0.11.0rc2.dev157+g0ad9951c4.d20251010) 2025-10-10T01:28:20.3011116Z Downloading sniffio-1.3.1-py3-none-any.whl.metadata (3.9 kB) 2025-10-10T01:28:20.3181980Z Collecting fastapi-cli>=0.0.8 (from fastapi-cli[standard]>=0.0.8; extra == "standard"->fastapi[standard]>=0.115.0->vllm==0.11.0rc2.dev157+g0ad9951c4.d20251010) 2025-10-10T01:28:20.3237130Z Downloading fastapi_cli-0.0.13-py3-none-any.whl.metadata (6.3 kB) 2025-10-10T01:28:20.3454284Z Collecting httpx<1.0.0,>=0.23.0 (from fastapi[standard]>=0.115.0->vllm==0.11.0rc2.dev157+g0ad9951c4.d20251010) 2025-10-10T01:28:20.3560255Z Downloading httpx-0.28.1-py3-none-any.whl.metadata (7.1 kB) 2025-10-10T01:28:20.3614640Z Requirement already satisfied: jinja2>=3.1.5 in /opt/conda/envs/py_3.12/lib/python3.12/site-packages (from fastapi[standard]>=0.115.0->vllm==0.11.0rc2.dev157+g0ad9951c4.d20251010) (3.1.6) 2025-10-10T01:28:20.3726415Z Collecting python-multipart>=0.0.18 (from fastapi[standard]>=0.115.0->vllm==0.11.0rc2.dev157+g0ad9951c4.d20251010) 2025-10-10T01:28:20.3771006Z Downloading python_multipart-0.0.20-py3-none-any.whl.metadata (1.8 kB) 2025-10-10T01:28:20.3922039Z Collecting email-validator>=2.0.0 (from fastapi[standard]>=0.115.0->vllm==0.11.0rc2.dev157+g0ad9951c4.d20251010) 2025-10-10T01:28:20.3964469Z Downloading email_validator-2.3.0-py3-none-any.whl.metadata (26 kB) 2025-10-10T01:28:20.4307641Z Collecting uvicorn>=0.12.0 (from uvicorn[standard]>=0.12.0; extra == "standard"->fastapi[standard]>=0.115.0->vllm==0.11.0rc2.dev157+g0ad9951c4.d20251010) 2025-10-10T01:28:20.4344858Z Downloading uvicorn-0.37.0-py3-none-any.whl.metadata (6.6 kB) 2025-10-10T01:28:20.4475390Z Requirement already satisfied: certifi in /opt/conda/envs/py_3.12/lib/python3.12/site-packages (from httpx<1.0.0,>=0.23.0->fastapi[standard]>=0.115.0->vllm==0.11.0rc2.dev157+g0ad9951c4.d20251010) (2025.10.5) 2025-10-10T01:28:20.4694082Z Collecting httpcore==1.* (from httpx<1.0.0,>=0.23.0->fastapi[standard]>=0.115.0->vllm==0.11.0rc2.dev157+g0ad9951c4.d20251010) 2025-10-10T01:28:20.4733190Z Downloading httpcore-1.0.9-py3-none-any.whl.metadata (21 kB) 2025-10-10T01:28:20.4944612Z Collecting h11>=0.16 (from httpcore==1.*->httpx<1.0.0,>=0.23.0->fastapi[standard]>=0.115.0->vllm==0.11.0rc2.dev157+g0ad9951c4.d20251010) 2025-10-10T01:28:20.4990672Z Downloading h11-0.16.0-py3-none-any.whl.metadata (8.3 kB) 2025-10-10T01:28:20.5208307Z Collecting dnspython>=2.0.0 (from email-validator>=2.0.0->fastapi[standard]>=0.115.0->vllm==0.11.0rc2.dev157+g0ad9951c4.d20251010) 2025-10-10T01:28:20.5251494Z Downloading dnspython-2.8.0-py3-none-any.whl.metadata (5.7 kB) 2025-10-10T01:28:20.5544685Z Collecting typer>=0.15.1 (from fastapi-cli>=0.0.8->fastapi-cli[standard]>=0.0.8; extra == "standard"->fastapi[standard]>=0.115.0->vllm==0.11.0rc2.dev157+g0ad9951c4.d20251010) 2025-10-10T01:28:20.5582488Z Downloading typer-0.19.2-py3-none-any.whl.metadata (16 kB) 2025-10-10T01:28:20.6255321Z Collecting rich-toolkit>=0.14.8 (from fastapi-cli>=0.0.8->fastapi-cli[standard]>=0.0.8; extra == "standard"->fastapi[standard]>=0.115.0->vllm==0.11.0rc2.dev157+g0ad9951c4.d20251010) 2025-10-10T01:28:20.6296156Z Downloading rich_toolkit-0.15.1-py3-none-any.whl.metadata (1.0 kB) 2025-10-10T01:28:20.7264215Z Collecting fastapi-cloud-cli>=0.1.1 (from fastapi-cli[standard]>=0.0.8; extra == "standard"->fastapi[standard]>=0.115.0->vllm==0.11.0rc2.dev157+g0ad9951c4.d20251010) 2025-10-10T01:28:20.7312851Z Downloading fastapi_cloud_cli-0.3.1-py3-none-any.whl.metadata (3.2 kB) 2025-10-10T01:28:20.9499045Z Collecting rignore>=0.5.1 (from fastapi-cloud-cli>=0.1.1->fastapi-cli[standard]>=0.0.8; extra == "standard"->fastapi[standard]>=0.115.0->vllm==0.11.0rc2.dev157+g0ad9951c4.d20251010) 2025-10-10T01:28:20.9571892Z Downloading rignore-0.7.0-cp312-cp312-manylinux_2_17_x86_64.manylinux2014_x86_64.whl.metadata (3.8 kB) 2025-10-10T01:28:21.0363119Z Collecting sentry-sdk>=2.20.0 (from fastapi-cloud-cli>=0.1.1->fastapi-cli[standard]>=0.0.8; extra == "standard"->fastapi[standard]>=0.115.0->vllm==0.11.0rc2.dev157+g0ad9951c4.d20251010) 2025-10-10T01:28:21.0455196Z Downloading sentry_sdk-2.41.0-py2.py3-none-any.whl.metadata (10 kB) 2025-10-10T01:28:21.0696446Z Requirement already satisfied: MarkupSafe>=2.0 in /opt/conda/envs/py_3.12/lib/python3.12/site-packages (from jinja2>=3.1.5->fastapi[standard]>=0.115.0->vllm==0.11.0rc2.dev157+g0ad9951c4.d20251010) (3.0.3) 2025-10-10T01:28:21.1183105Z Collecting jsonschema>=4.21.1 (from mistral_common>=1.8.2->mistral_common[audio,image]>=1.8.2->vllm==0.11.0rc2.dev157+g0ad9951c4.d20251010) 2025-10-10T01:28:21.1228211Z Downloading jsonschema-4.25.1-py3-none-any.whl.metadata (7.6 kB) 2025-10-10T01:28:21.1491867Z Collecting pydantic-extra-types>=2.10.5 (from pydantic-extra-types[pycountry]>=2.10.5->mistral_common>=1.8.2->mistral_common[audio,image]>=1.8.2->vllm==0.11.0rc2.dev157+g0ad9951c4.d20251010) 2025-10-10T01:28:21.1555528Z Downloading pydantic_extra_types-2.10.6-py3-none-any.whl.metadata (4.0 kB) 2025-10-10T01:28:21.1661455Z Requirement already satisfied: attrs>=22.2.0 in /opt/conda/envs/py_3.12/lib/python3.12/site-packages (from jsonschema>=4.21.1->mistral_common>=1.8.2->mistral_common[audio,image]>=1.8.2->vllm==0.11.0rc2.dev157+g0ad9951c4.d20251010) (25.4.0) 2025-10-10T01:28:21.1782477Z Collecting jsonschema-specifications>=2023.03.6 (from jsonschema>=4.21.1->mistral_common>=1.8.2->mistral_common[audio,image]>=1.8.2->vllm==0.11.0rc2.dev157+g0ad9951c4.d20251010) 2025-10-10T01:28:21.1828386Z Downloading jsonschema_specifications-2025.9.1-py3-none-any.whl.metadata (2.9 kB) 2025-10-10T01:28:21.2091920Z Collecting referencing>=0.28.4 (from jsonschema>=4.21.1->mistral_common>=1.8.2->mistral_common[audio,image]>=1.8.2->vllm==0.11.0rc2.dev157+g0ad9951c4.d20251010) 2025-10-10T01:28:21.2131631Z Downloading referencing-0.36.2-py3-none-any.whl.metadata (2.8 kB) 2025-10-10T01:28:21.4165349Z Collecting rpds-py>=0.7.1 (from jsonschema>=4.21.1->mistral_common>=1.8.2->mistral_common[audio,image]>=1.8.2->vllm==0.11.0rc2.dev157+g0ad9951c4.d20251010) 2025-10-10T01:28:21.4212215Z Downloading rpds_py-0.27.1-cp312-cp312-manylinux_2_17_x86_64.manylinux2014_x86_64.whl.metadata (4.2 kB) 2025-10-10T01:28:21.4433204Z Requirement already satisfied: distro<2,>=1.7.0 in /opt/conda/envs/py_3.12/lib/python3.12/site-packages (from openai>=1.99.1->vllm==0.11.0rc2.dev157+g0ad9951c4.d20251010) (1.9.0) 2025-10-10T01:28:21.4929451Z Collecting jiter<1,>=0.10.0 (from openai>=1.99.1->vllm==0.11.0rc2.dev157+g0ad9951c4.d20251010) 2025-10-10T01:28:21.4973107Z Downloading jiter-0.11.0-cp312-cp312-manylinux_2_17_x86_64.manylinux2014_x86_64.whl.metadata (5.2 kB) 2025-10-10T01:28:21.7486896Z Collecting numpy (from vllm==0.11.0rc2.dev157+g0ad9951c4.d20251010) 2025-10-10T01:28:21.8237974Z Downloading numpy-2.2.6-cp312-cp312-manylinux_2_17_x86_64.manylinux2014_x86_64.whl.metadata (62 kB) 2025-10-10T01:28:21.9034900Z Collecting pycountry>=23 (from pydantic-extra-types[pycountry]>=2.10.5->mistral_common>=1.8.2->mistral_common[audio,image]>=1.8.2->vllm==0.11.0rc2.dev157+g0ad9951c4.d20251010) 2025-10-10T01:28:21.9115087Z Downloading pycountry-24.6.1-py3-none-any.whl.metadata (12 kB) 2025-10-10T01:28:21.9270409Z Requirement already satisfied: click>=7.0 in /opt/conda/envs/py_3.12/lib/python3.12/site-packages (from ray>=2.48.0->ray[cgraph]>=2.48.0->vllm==0.11.0rc2.dev157+g0ad9951c4.d20251010) (8.3.0) 2025-10-10T01:28:21.9275513Z Requirement already satisfied: msgpack<2.0.0,>=1.0.0 in /opt/conda/envs/py_3.12/lib/python3.12/site-packages (from ray>=2.48.0->ray[cgraph]>=2.48.0->vllm==0.11.0rc2.dev157+g0ad9951c4.d20251010) (1.1.2) 2025-10-10T01:28:21.9534882Z Collecting cupy-cuda12x (from ray[cgraph]>=2.48.0->vllm==0.11.0rc2.dev157+g0ad9951c4.d20251010) 2025-10-10T01:28:22.0194605Z Downloading cupy_cuda12x-13.6.0-cp312-cp312-manylinux2014_x86_64.whl.metadata (2.4 kB) 2025-10-10T01:28:22.0414467Z Requirement already satisfied: charset_normalizer<4,>=2 in /opt/conda/envs/py_3.12/lib/python3.12/site-packages (from requests>=2.26.0->vllm==0.11.0rc2.dev157+g0ad9951c4.d20251010) (3.4.3) 2025-10-10T01:28:22.0420205Z Requirement already satisfied: urllib3<3,>=1.21.1 in /opt/conda/envs/py_3.12/lib/python3.12/site-packages (from requests>=2.26.0->vllm==0.11.0rc2.dev157+g0ad9951c4.d20251010) (2.5.0) 2025-10-10T01:28:22.0791357Z Collecting rich>=13.7.1 (from rich-toolkit>=0.14.8->fastapi-cli>=0.0.8->fastapi-cli[standard]>=0.0.8; extra == "standard"->fastapi[standard]>=0.115.0->vllm==0.11.0rc2.dev157+g0ad9951c4.d20251010) 2025-10-10T01:28:22.0947688Z Downloading rich-14.2.0-py3-none-any.whl.metadata (18 kB) 2025-10-10T01:28:22.1230579Z Collecting markdown-it-py>=2.2.0 (from rich>=13.7.1->rich-toolkit>=0.14.8->fastapi-cli>=0.0.8->fastapi-cli[standard]>=0.0.8; extra == "standard"->fastapi[standard]>=0.115.0->vllm==0.11.0rc2.dev157+g0ad9951c4.d20251010) 2025-10-10T01:28:22.1276094Z Downloading markdown_it_py-4.0.0-py3-none-any.whl.metadata (7.3 kB) 2025-10-10T01:28:22.1381205Z Requirement already satisfied: pygments<3.0.0,>=2.13.0 in /opt/conda/envs/py_3.12/lib/python3.12/site-packages (from rich>=13.7.1->rich-toolkit>=0.14.8->fastapi-cli>=0.0.8->fastapi-cli[standard]>=0.0.8; extra == "standard"->fastapi[standard]>=0.115.0->vllm==0.11.0rc2.dev157+g0ad9951c4.d20251010) (2.15.0) 2025-10-10T01:28:22.1523606Z Collecting mdurl~=0.1 (from markdown-it-py>=2.2.0->rich>=13.7.1->rich-toolkit>=0.14.8->fastapi-cli>=0.0.8->fastapi-cli[standard]>=0.0.8; extra == "standard"->fastapi[standard]>=0.115.0->vllm==0.11.0rc2.dev157+g0ad9951c4.d20251010) 2025-10-10T01:28:22.1567746Z Downloading mdurl-0.1.2-py3-none-any.whl.metadata (1.6 kB) 2025-10-10T01:28:22.2200419Z Collecting huggingface-hub<2.0,>=0.16.4 (from tokenizers>=0.21.1->vllm==0.11.0rc2.dev157+g0ad9951c4.d20251010) 2025-10-10T01:28:22.2254563Z Downloading huggingface_hub-0.35.3-py3-none-any.whl.metadata (14 kB) 2025-10-10T01:28:22.2483819Z Requirement already satisfied: fsspec>=2023.5.0 in /opt/conda/envs/py_3.12/lib/python3.12/site-packages (from huggingface-hub<2.0,>=0.16.4->tokenizers>=0.21.1->vllm==0.11.0rc2.dev157+g0ad9951c4.d20251010) (2025.9.0) 2025-10-10T01:28:22.2832789Z Collecting hf-xet<2.0.0,>=1.1.3 (from huggingface-hub<2.0,>=0.16.4->tokenizers>=0.21.1->vllm==0.11.0rc2.dev157+g0ad9951c4.d20251010) 2025-10-10T01:28:22.2870540Z Downloading hf_xet-1.1.10-cp37-abi3-manylinux_2_17_x86_64.manylinux2014_x86_64.whl.metadata (4.7 kB) 2025-10-10T01:28:22.3077632Z Requirement already satisfied: sympy>=1.13.3 in /opt/conda/envs/py_3.12/lib/python3.12/site-packages (from torch>=1.7.0->compressed-tensors==0.11.0->vllm==0.11.0rc2.dev157+g0ad9951c4.d20251010) (1.13.3) 2025-10-10T01:28:22.3081644Z Requirement already satisfied: networkx>=2.5.1 in /opt/conda/envs/py_3.12/lib/python3.12/site-packages (from torch>=1.7.0->compressed-tensors==0.11.0->vllm==0.11.0rc2.dev157+g0ad9951c4.d20251010) (2.8.8) 2025-10-10T01:28:22.3135756Z Requirement already satisfied: mpmath<1.4,>=1.1.0 in /opt/conda/envs/py_3.12/lib/python3.12/site-packages (from sympy>=1.13.3->torch>=1.7.0->compressed-tensors==0.11.0->vllm==0.11.0rc2.dev157+g0ad9951c4.d20251010) (1.3.0) 2025-10-10T01:28:22.4312847Z Collecting safetensors>=0.4.3 (from transformers>=4.55.2->vllm==0.11.0rc2.dev157+g0ad9951c4.d20251010) 2025-10-10T01:28:22.4362897Z Downloading safetensors-0.6.2-cp38-abi3-manylinux_2_17_x86_64.manylinux2014_x86_64.whl.metadata (4.1 kB) 2025-10-10T01:28:22.4693674Z Collecting shellingham>=1.3.0 (from typer>=0.15.1->fastapi-cli>=0.0.8->fastapi-cli[standard]>=0.0.8; extra == "standard"->fastapi[standard]>=0.115.0->vllm==0.11.0rc2.dev157+g0ad9951c4.d20251010) 2025-10-10T01:28:22.4740696Z Downloading shellingham-1.5.4-py2.py3-none-any.whl.metadata (3.5 kB) 2025-10-10T01:28:22.5253030Z Collecting httptools>=0.6.3 (from uvicorn[standard]>=0.12.0; extra == "standard"->fastapi[standard]>=0.115.0->vllm==0.11.0rc2.dev157+g0ad9951c4.d20251010) 2025-10-10T01:28:22.5296555Z Downloading httptools-0.6.4-cp312-cp312-manylinux_2_5_x86_64.manylinux1_x86_64.manylinux_2_17_x86_64.manylinux2014_x86_64.whl.metadata (3.6 kB) 2025-10-10T01:28:22.5520416Z Collecting python-dotenv>=0.13 (from uvicorn[standard]>=0.12.0; extra == "standard"->fastapi[standard]>=0.115.0->vllm==0.11.0rc2.dev157+g0ad9951c4.d20251010) 2025-10-10T01:28:22.5564641Z Downloading python_dotenv-1.1.1-py3-none-any.whl.metadata (24 kB) 2025-10-10T01:28:22.5955904Z Collecting uvloop>=0.15.1 (from uvicorn[standard]>=0.12.0; extra == "standard"->fastapi[standard]>=0.115.0->vllm==0.11.0rc2.dev157+g0ad9951c4.d20251010) 2025-10-10T01:28:22.5996336Z Downloading uvloop-0.21.0-cp312-cp312-manylinux_2_17_x86_64.manylinux2014_x86_64.whl.metadata (4.9 kB) 2025-10-10T01:28:22.6805478Z Collecting websockets>=10.4 (from uvicorn[standard]>=0.12.0; extra == "standard"->fastapi[standard]>=0.115.0->vllm==0.11.0rc2.dev157+g0ad9951c4.d20251010) 2025-10-10T01:28:22.6848641Z Downloading websockets-15.0.1-cp312-cp312-manylinux_2_5_x86_64.manylinux1_x86_64.manylinux_2_17_x86_64.manylinux2014_x86_64.whl.metadata (6.8 kB) 2025-10-10T01:28:22.7086206Z Requirement already satisfied: aiohappyeyeballs>=2.5.0 in /opt/conda/envs/py_3.12/lib/python3.12/site-packages (from aiohttp->vllm==0.11.0rc2.dev157+g0ad9951c4.d20251010) (2.6.1) 2025-10-10T01:28:22.7089859Z Requirement already satisfied: aiosignal>=1.4.0 in /opt/conda/envs/py_3.12/lib/python3.12/site-packages (from aiohttp->vllm==0.11.0rc2.dev157+g0ad9951c4.d20251010) (1.4.0) 2025-10-10T01:28:22.7096448Z Requirement already satisfied: frozenlist>=1.1.1 in /opt/conda/envs/py_3.12/lib/python3.12/site-packages (from aiohttp->vllm==0.11.0rc2.dev157+g0ad9951c4.d20251010) (1.8.0) 2025-10-10T01:28:22.7100838Z Requirement already satisfied: multidict<7.0,>=4.5 in /opt/conda/envs/py_3.12/lib/python3.12/site-packages (from aiohttp->vllm==0.11.0rc2.dev157+g0ad9951c4.d20251010) (6.7.0) 2025-10-10T01:28:22.7104509Z Requirement already satisfied: propcache>=0.2.0 in /opt/conda/envs/py_3.12/lib/python3.12/site-packages (from aiohttp->vllm==0.11.0rc2.dev157+g0ad9951c4.d20251010) (0.3.2) 2025-10-10T01:28:22.7108544Z Requirement already satisfied: yarl<2.0,>=1.17.0 in /opt/conda/envs/py_3.12/lib/python3.12/site-packages (from aiohttp->vllm==0.11.0rc2.dev157+g0ad9951c4.d20251010) (1.22.0) 2025-10-10T01:28:22.7686871Z Collecting fastrlock>=0.5 (from cupy-cuda12x->ray[cgraph]>=2.48.0->vllm==0.11.0rc2.dev157+g0ad9951c4.d20251010) 2025-10-10T01:28:22.7741383Z Downloading fastrlock-0.8.3-cp312-cp312-manylinux_2_5_x86_64.manylinux1_x86_64.manylinux_2_28_x86_64.whl.metadata (7.7 kB) 2025-10-10T01:28:22.7943877Z Requirement already satisfied: soundfile>=0.12.1 in /opt/conda/envs/py_3.12/lib/python3.12/site-packages (from mistral_common>=1.8.2->mistral_common[audio,image]>=1.8.2->vllm==0.11.0rc2.dev157+g0ad9951c4.d20251010) (0.13.1) 2025-10-10T01:28:22.7963540Z Requirement already satisfied: cffi>=1.0 in /opt/conda/envs/py_3.12/lib/python3.12/site-packages (from soundfile>=0.12.1->mistral_common>=1.8.2->mistral_common[audio,image]>=1.8.2->vllm==0.11.0rc2.dev157+g0ad9951c4.d20251010) (2.0.0) 2025-10-10T01:28:22.7983818Z Requirement already satisfied: pycparser in /opt/conda/envs/py_3.12/lib/python3.12/site-packages (from cffi>=1.0->soundfile>=0.12.1->mistral_common>=1.8.2->mistral_common[audio,image]>=1.8.2->vllm==0.11.0rc2.dev157+g0ad9951c4.d20251010) (2.23) 2025-10-10T01:28:22.8010127Z Requirement already satisfied: soxr>=0.5.0 in /opt/conda/envs/py_3.12/lib/python3.12/site-packages (from mistral_common>=1.8.2->mistral_common[audio,image]>=1.8.2->vllm==0.11.0rc2.dev157+g0ad9951c4.d20251010) (1.0.0) 2025-10-10T01:28:22.8535566Z Downloading compressed_tensors-0.11.0-py3-none-any.whl (179 kB) 2025-10-10T01:28:22.8671336Z Downloading depyf-0.19.0-py3-none-any.whl (39 kB) 2025-10-10T01:28:22.8774830Z Downloading diskcache-5.6.3-py3-none-any.whl (45 kB) 2025-10-10T01:28:22.8937156Z Downloading lark-1.2.2-py3-none-any.whl (111 kB) 2025-10-10T01:28:22.9065074Z Downloading lm_format_enforcer-0.11.3-py3-none-any.whl (45 kB) 2025-10-10T01:28:22.9157376Z Downloading numba-0.61.2-cp312-cp312-manylinux2014_x86_64.manylinux_2_17_x86_64.whl (3.9 MB) 2025-10-10T01:28:22.9468874Z [?25l ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 0.0/3.9 MB ? eta -:--:-- 2025-10-10T01:28:22.9472254Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 3.9/3.9 MB 131.2 MB/s 0:00:00 2025-10-10T01:28:22.9545574Z [?25hDownloading outlines_core-0.2.11-cp312-cp312-manylinux_2_17_x86_64.manylinux2014_x86_64.whl (2.3 MB) 2025-10-10T01:28:22.9955928Z [?25l ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 0.0/2.3 MB ? eta -:--:-- 2025-10-10T01:28:22.9956546Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 2.3/2.3 MB 53.6 MB/s 0:00:00 2025-10-10T01:28:23.0040888Z [?25hDownloading xgrammar-0.1.25-cp312-cp312-manylinux_2_17_x86_64.manylinux2014_x86_64.whl (8.7 MB) 2025-10-10T01:28:23.2069361Z [?25l ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 0.0/8.7 MB ? eta -:--:-- 2025-10-10T01:28:23.3108285Z  ━━━━━━━╺━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 1.6/8.7 MB 6.6 MB/s eta 0:00:02 2025-10-10T01:28:23.3109084Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 8.7/8.7 MB 27.8 MB/s 0:00:00 2025-10-10T01:28:23.3237089Z [?25hDownloading llguidance-0.7.30-cp39-abi3-manylinux_2_17_x86_64.manylinux2014_x86_64.whl (15.0 MB) 2025-10-10T01:28:23.4952156Z [?25l ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 0.0/15.0 MB ? eta -:--:-- 2025-10-10T01:28:23.4952812Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 15.0/15.0 MB 87.6 MB/s 0:00:00 2025-10-10T01:28:23.5005815Z [?25hDownloading llvmlite-0.44.0-cp312-cp312-manylinux_2_17_x86_64.manylinux2014_x86_64.whl (42.4 MB) 2025-10-10T01:28:23.7038967Z [?25l ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 0.0/42.4 MB ? eta -:--:-- 2025-10-10T01:28:23.9055861Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━╸━━━━━━━━━━━━━ 28.3/42.4 MB 144.8 MB/s eta 0:00:01 2025-10-10T01:28:23.9070990Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╸ 42.2/42.4 MB 128.3 MB/s eta 0:00:01 2025-10-10T01:28:23.9071671Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 42.4/42.4 MB 105.3 MB/s 0:00:00 2025-10-10T01:28:23.9118617Z [?25hDownloading fastapi-0.118.2-py3-none-any.whl (97 kB) 2025-10-10T01:28:23.9199821Z Downloading pydantic-2.12.0-py3-none-any.whl (459 kB) 2025-10-10T01:28:23.9297183Z Downloading pydantic_core-2.41.1-cp312-cp312-manylinux_2_17_x86_64.manylinux2014_x86_64.whl (2.1 MB) 2025-10-10T01:28:23.9567173Z [?25l ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 0.0/2.1 MB ? eta -:--:-- 2025-10-10T01:28:23.9567813Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 2.1/2.1 MB 75.0 MB/s 0:00:00 2025-10-10T01:28:23.9605043Z [?25hDownloading starlette-0.48.0-py3-none-any.whl (73 kB) 2025-10-10T01:28:23.9682037Z Downloading anyio-4.11.0-py3-none-any.whl (109 kB) 2025-10-10T01:28:23.9754263Z Downloading annotated_types-0.7.0-py3-none-any.whl (13 kB) 2025-10-10T01:28:23.9818862Z Downloading httpx-0.28.1-py3-none-any.whl (73 kB) 2025-10-10T01:28:23.9882274Z Downloading httpcore-1.0.9-py3-none-any.whl (78 kB) 2025-10-10T01:28:23.9956044Z Downloading email_validator-2.3.0-py3-none-any.whl (35 kB) 2025-10-10T01:28:24.0018388Z Downloading dnspython-2.8.0-py3-none-any.whl (331 kB) 2025-10-10T01:28:24.0113367Z Downloading fastapi_cli-0.0.13-py3-none-any.whl (11 kB) 2025-10-10T01:28:24.0214708Z Downloading fastapi_cloud_cli-0.3.1-py3-none-any.whl (19 kB) 2025-10-10T01:28:24.0308406Z Downloading gguf-0.17.1-py3-none-any.whl (96 kB) 2025-10-10T01:28:24.0380003Z Downloading h11-0.16.0-py3-none-any.whl (37 kB) 2025-10-10T01:28:24.0489524Z Downloading interegular-0.3.3-py37-none-any.whl (23 kB) 2025-10-10T01:28:24.0626609Z Downloading mistral_common-1.8.5-py3-none-any.whl (6.5 MB) 2025-10-10T01:28:24.1343980Z [?25l ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 0.0/6.5 MB ? eta -:--:-- 2025-10-10T01:28:24.1344623Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 6.5/6.5 MB 91.3 MB/s 0:00:00 2025-10-10T01:28:24.1407516Z [?25hDownloading jsonschema-4.25.1-py3-none-any.whl (90 kB) 2025-10-10T01:28:24.1494449Z Downloading jsonschema_specifications-2025.9.1-py3-none-any.whl (18 kB) 2025-10-10T01:28:24.1606165Z Downloading openai-2.3.0-py3-none-any.whl (999 kB) 2025-10-10T01:28:24.1696780Z [?25l ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 0.0/999.8 kB ? eta -:--:-- 2025-10-10T01:28:24.1697424Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 999.8/999.8 kB 115.5 MB/s 0:00:00 2025-10-10T01:28:24.1738494Z [?25hDownloading jiter-0.11.0-cp312-cp312-manylinux_2_17_x86_64.manylinux2014_x86_64.whl (347 kB) 2025-10-10T01:28:24.1835883Z Downloading typing_extensions-4.15.0-py3-none-any.whl (44 kB) 2025-10-10T01:28:24.2017537Z Downloading openai_harmony-0.0.4-cp38-abi3-manylinux_2_17_x86_64.manylinux2014_x86_64.whl (3.0 MB) 2025-10-10T01:28:24.2447380Z [?25l ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 0.0/3.0 MB ? eta -:--:-- 2025-10-10T01:28:24.2448052Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 3.0/3.0 MB 68.9 MB/s 0:00:00 2025-10-10T01:28:24.2532460Z [?25hDownloading opencv_python_headless-4.12.0.88-cp37-abi3-manylinux2014_x86_64.manylinux_2_17_x86_64.whl (54.0 MB) 2025-10-10T01:28:24.4557090Z [?25l ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 0.0/54.0 MB ? eta -:--:-- 2025-10-10T01:28:24.6576768Z  ━━━━━━━━━━━━━━━━━━╸━━━━━━━━━━━━━━━━━━━━━ 25.4/54.0 MB 126.4 MB/s eta 0:00:01 2025-10-10T01:28:24.7882658Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╸━━━━ 48.2/54.0 MB 121.3 MB/s eta 0:00:01 2025-10-10T01:28:24.7883458Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 54.0/54.0 MB 101.1 MB/s 0:00:00 2025-10-10T01:28:24.7974476Z [?25hDownloading numpy-2.2.6-cp312-cp312-manylinux_2_17_x86_64.manylinux2014_x86_64.whl (16.5 MB) 2025-10-10T01:28:24.9273016Z [?25l ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 0.0/16.5 MB ? eta -:--:-- 2025-10-10T01:28:24.9273693Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 16.5/16.5 MB 128.4 MB/s 0:00:00 2025-10-10T01:28:24.9324755Z [?25hDownloading prometheus_client-0.23.1-py3-none-any.whl (61 kB) 2025-10-10T01:28:24.9406066Z Downloading prometheus_fastapi_instrumentator-7.1.0-py3-none-any.whl (19 kB) 2025-10-10T01:28:24.9498104Z Downloading pydantic_extra_types-2.10.6-py3-none-any.whl (40 kB) 2025-10-10T01:28:24.9573993Z Downloading pycountry-24.6.1-py3-none-any.whl (6.3 MB) 2025-10-10T01:28:25.0152951Z [?25l ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 0.0/6.3 MB ? eta -:--:-- 2025-10-10T01:28:25.0153604Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 6.3/6.3 MB 110.5 MB/s 0:00:00 2025-10-10T01:28:25.0202124Z [?25hDownloading python_multipart-0.0.20-py3-none-any.whl (24 kB) 2025-10-10T01:28:25.0294263Z Downloading pyzmq-27.1.0-cp312-abi3-manylinux_2_26_x86_64.manylinux_2_28_x86_64.whl (840 kB) 2025-10-10T01:28:25.0436454Z [?25l ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 0.0/841.0 kB ? eta -:--:-- 2025-10-10T01:28:25.0437110Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 841.0/841.0 kB 52.8 MB/s 0:00:00 2025-10-10T01:28:25.0482317Z [?25hDownloading ray-2.49.2-cp312-cp312-manylinux2014_x86_64.whl (70.1 MB) 2025-10-10T01:28:25.2514981Z [?25l ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 0.0/70.1 MB ? eta -:--:-- 2025-10-10T01:28:25.4530475Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━╸━━━━━━━━━━━ 50.3/70.1 MB 256.7 MB/s eta 0:00:01 2025-10-10T01:28:25.5621106Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╸ 70.0/70.1 MB 178.8 MB/s eta 0:00:01 2025-10-10T01:28:25.5622396Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 70.1/70.1 MB 136.8 MB/s 0:00:00 2025-10-10T01:28:25.5667572Z [?25hDownloading referencing-0.36.2-py3-none-any.whl (26 kB) 2025-10-10T01:28:25.5750210Z Downloading rich_toolkit-0.15.1-py3-none-any.whl (29 kB) 2025-10-10T01:28:25.5838365Z Downloading rich-14.2.0-py3-none-any.whl (243 kB) 2025-10-10T01:28:25.5939628Z Downloading markdown_it_py-4.0.0-py3-none-any.whl (87 kB) 2025-10-10T01:28:25.6029842Z Downloading mdurl-0.1.2-py3-none-any.whl (10.0 kB) 2025-10-10T01:28:25.6258202Z Downloading rignore-0.7.0-cp312-cp312-manylinux_2_17_x86_64.manylinux2014_x86_64.whl (951 kB) 2025-10-10T01:28:25.6367626Z [?25l ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 0.0/951.1 kB ? eta -:--:-- 2025-10-10T01:28:25.6368584Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 951.1/951.1 kB 84.1 MB/s 0:00:00 2025-10-10T01:28:25.6418219Z [?25hDownloading rpds_py-0.27.1-cp312-cp312-manylinux_2_17_x86_64.manylinux2014_x86_64.whl (386 kB) 2025-10-10T01:28:25.6545300Z Downloading sentry_sdk-2.41.0-py2.py3-none-any.whl (370 kB) 2025-10-10T01:28:25.6639472Z Downloading sniffio-1.3.1-py3-none-any.whl (10 kB) 2025-10-10T01:28:25.6738997Z Downloading tiktoken-0.12.0-cp312-cp312-manylinux_2_28_x86_64.whl (1.2 MB) 2025-10-10T01:28:25.6883315Z [?25l ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 0.0/1.2 MB ? eta -:--:-- 2025-10-10T01:28:25.6884572Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 1.2/1.2 MB 77.0 MB/s 0:00:00 2025-10-10T01:28:25.6933980Z [?25hDownloading regex-2025.9.18-cp312-cp312-manylinux2014_x86_64.manylinux_2_17_x86_64.manylinux_2_28_x86_64.whl (802 kB) 2025-10-10T01:28:25.7105008Z [?25l ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 0.0/802.0 kB ? eta -:--:-- 2025-10-10T01:28:25.7106263Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 802.0/802.0 kB 60.0 MB/s 0:00:00 2025-10-10T01:28:25.7148593Z [?25hDownloading tokenizers-0.22.1-cp39-abi3-manylinux_2_17_x86_64.manylinux2014_x86_64.whl (3.3 MB) 2025-10-10T01:28:25.7349321Z [?25l ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 0.0/3.3 MB ? eta -:--:-- 2025-10-10T01:28:25.7350265Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 3.3/3.3 MB 172.6 MB/s 0:00:00 2025-10-10T01:28:25.7407051Z [?25hDownloading huggingface_hub-0.35.3-py3-none-any.whl (564 kB) 2025-10-10T01:28:25.7489761Z [?25l ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 0.0/564.3 kB ? eta -:--:-- 2025-10-10T01:28:25.7491564Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 564.3/564.3 kB 55.1 MB/s 0:00:00 2025-10-10T01:28:25.7539588Z [?25hDownloading hf_xet-1.1.10-cp37-abi3-manylinux_2_17_x86_64.manylinux2014_x86_64.whl (3.2 MB) 2025-10-10T01:28:25.7727907Z [?25l ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 0.0/3.2 MB ? eta -:--:-- 2025-10-10T01:28:25.7729168Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 3.2/3.2 MB 184.1 MB/s 0:00:00 2025-10-10T01:28:25.7773724Z [?25hDownloading transformers-4.57.0-py3-none-any.whl (12.0 MB) 2025-10-10T01:28:25.8525197Z [?25l ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 0.0/12.0 MB ? eta -:--:-- 2025-10-10T01:28:25.8526491Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 12.0/12.0 MB 162.6 MB/s 0:00:00 2025-10-10T01:28:25.8568849Z [?25hDownloading safetensors-0.6.2-cp38-abi3-manylinux_2_17_x86_64.manylinux2014_x86_64.whl (485 kB) 2025-10-10T01:28:25.8675076Z Downloading typer-0.19.2-py3-none-any.whl (46 kB) 2025-10-10T01:28:25.8770310Z Downloading shellingham-1.5.4-py2.py3-none-any.whl (9.8 kB) 2025-10-10T01:28:25.8849019Z Downloading typing_inspection-0.4.2-py3-none-any.whl (14 kB) 2025-10-10T01:28:25.8916352Z Downloading uvicorn-0.37.0-py3-none-any.whl (67 kB) 2025-10-10T01:28:25.9040525Z Downloading httptools-0.6.4-cp312-cp312-manylinux_2_5_x86_64.manylinux1_x86_64.manylinux_2_17_x86_64.manylinux2014_x86_64.whl (510 kB) 2025-10-10T01:28:25.9147101Z Downloading python_dotenv-1.1.1-py3-none-any.whl (20 kB) 2025-10-10T01:28:25.9242113Z Downloading uvloop-0.21.0-cp312-cp312-manylinux_2_17_x86_64.manylinux2014_x86_64.whl (4.7 MB) 2025-10-10T01:28:26.0028317Z [?25l ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 0.0/4.7 MB ? eta -:--:-- 2025-10-10T01:28:26.0028999Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 4.7/4.7 MB 59.2 MB/s 0:00:00 2025-10-10T01:28:26.0084744Z [?25hDownloading watchfiles-1.1.0-cp312-cp312-manylinux_2_17_x86_64.manylinux2014_x86_64.whl (452 kB) 2025-10-10T01:28:26.0191836Z Downloading websockets-15.0.1-cp312-cp312-manylinux_2_5_x86_64.manylinux1_x86_64.manylinux_2_17_x86_64.manylinux2014_x86_64.whl (182 kB) 2025-10-10T01:28:26.0324682Z Downloading astor-0.8.1-py2.py3-none-any.whl (27 kB) 2025-10-10T01:28:26.0417648Z Downloading blake3-1.0.7-cp312-cp312-manylinux_2_17_x86_64.manylinux2014_x86_64.whl (387 kB) 2025-10-10T01:28:26.0532259Z Downloading cbor2-5.7.0-cp312-cp312-manylinux2014_x86_64.manylinux_2_17_x86_64.manylinux_2_28_x86_64.whl (284 kB) 2025-10-10T01:28:26.0636955Z Downloading cloudpickle-3.1.1-py3-none-any.whl (20 kB) 2025-10-10T01:28:26.0837475Z Downloading cupy_cuda12x-13.6.0-cp312-cp312-manylinux2014_x86_64.whl (112.9 MB) 2025-10-10T01:28:26.2871451Z [?25l ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 0.0/112.9 MB ? eta -:--:-- 2025-10-10T01:28:26.4884148Z  ━━━━━━━━╸━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 25.2/112.9 MB 128.0 MB/s eta 0:00:01 2025-10-10T01:28:26.6902182Z  ━━━━━━━━━━━━━━━━━╺━━━━━━━━━━━━━━━━━━━━━━ 48.8/112.9 MB 120.9 MB/s eta 0:00:01 2025-10-10T01:28:26.8917902Z  ━━━━━━━━━━━━━━━━━━━━━━━━╺━━━━━━━━━━━━━━━ 68.9/112.9 MB 114.2 MB/s eta 0:00:01 2025-10-10T01:28:27.0932347Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╸━━━━━━━━━ 87.0/112.9 MB 108.1 MB/s eta 0:00:01 2025-10-10T01:28:27.2550695Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╸ 112.7/112.9 MB 113.9 MB/s eta 0:00:01 2025-10-10T01:28:27.2551435Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 112.9/112.9 MB 96.4 MB/s 0:00:01 2025-10-10T01:28:27.2618283Z [?25hDownloading fastrlock-0.8.3-cp312-cp312-manylinux_2_5_x86_64.manylinux1_x86_64.manylinux_2_28_x86_64.whl (53 kB) 2025-10-10T01:28:27.2747250Z Downloading einops-0.8.1-py3-none-any.whl (64 kB) 2025-10-10T01:28:27.2828340Z Downloading frozendict-2.4.6-py312-none-any.whl (16 kB) 2025-10-10T01:28:27.2928538Z Downloading msgspec-0.19.0-cp312-cp312-manylinux_2_17_x86_64.manylinux2014_x86_64.whl (213 kB) 2025-10-10T01:28:27.3022108Z Downloading partial_json_parser-0.2.1.1.post6-py3-none-any.whl (10 kB) 2025-10-10T01:28:27.3098818Z Downloading py_cpuinfo-9.0.0-py3-none-any.whl (22 kB) 2025-10-10T01:28:27.3188390Z Downloading pybase64-1.4.2-cp312-cp312-manylinux1_x86_64.manylinux2014_x86_64.manylinux_2_17_x86_64.manylinux_2_5_x86_64.whl (71 kB) 2025-10-10T01:28:27.3278065Z Downloading python_json_logger-4.0.0-py3-none-any.whl (15 kB) 2025-10-10T01:28:27.3431281Z Downloading sentencepiece-0.2.1-cp312-cp312-manylinux_2_27_x86_64.manylinux_2_28_x86_64.whl (1.4 MB) 2025-10-10T01:28:27.3539365Z [?25l ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 0.0/1.4 MB ? eta -:--:-- 2025-10-10T01:28:27.3540012Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 1.4/1.4 MB 138.8 MB/s 0:00:00 2025-10-10T01:28:27.3639969Z [?25hDownloading setproctitle-1.3.7-cp312-cp312-manylinux1_x86_64.manylinux_2_28_x86_64.manylinux_2_5_x86_64.whl (32 kB) 2025-10-10T01:28:27.9876075Z Installing collected packages: py-cpuinfo, fastrlock, websockets, uvloop, typing_extensions, sniffio, shellingham, setproctitle, sentry-sdk, sentencepiece, safetensors, rpds-py, rignore, regex, pyzmq, python-multipart, python-json-logger, python-dotenv, pycountry, pybase64, prometheus_client, partial-json-parser, outlines_core, numpy, msgspec, mdurl, llvmlite, llguidance, lark, jiter, interegular, httptools, hf-xet, h11, frozendict, einops, dnspython, diskcache, cloudpickle, cbor2, blake3, astor, annotated-types, uvicorn, typing-inspection, tiktoken, referencing, pydantic-core, opencv-python-headless, numba, markdown-it-py, huggingface-hub, httpcore, gguf, email-validator, depyf, cupy-cuda12x, anyio, watchfiles, tokenizers, starlette, rich, pydantic, jsonschema-specifications, httpx, typer, transformers, rich-toolkit, pydantic-extra-types, prometheus-fastapi-instrumentator, openai-harmony, openai, lm-format-enforcer, jsonschema, fastapi, xgrammar, ray, fastapi-cloud-cli, fastapi-cli, compressed-tensors, mistral_common, vllm 2025-10-10T01:28:28.1557016Z [?25l 2025-10-10T01:28:28.1686047Z  ━╺━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━  3/82 [uvloop] 2025-10-10T01:28:28.1686542Z  Attempting uninstall: typing_extensions 2025-10-10T01:28:28.1742789Z ━╺━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━  3/82 [uvloop] 2025-10-10T01:28:28.1743300Z  Found existing installation: typing_extensions 4.12.2 2025-10-10T01:28:28.1763470Z ━╺━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━  3/82 [uvloop] 2025-10-10T01:28:28.1764177Z  Uninstalling typing_extensions-4.12.2: 2025-10-10T01:28:28.3238310Z ━╺━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━  3/82 [uvloop] 2025-10-10T01:28:28.4917662Z  ━╸━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━  4/82 [typing_extensions] 2025-10-10T01:28:28.6499312Z  ━╸━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━  4/82 [typing_extensions] 2025-10-10T01:28:28.6499863Z  Successfully uninstalled typing_extensions-4.12.2 2025-10-10T01:28:28.6688957Z ━╸━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━  4/82 [typing_extensions] 2025-10-10T01:28:28.8377568Z  ━╸━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━  4/82 [typing_extensions] 2025-10-10T01:28:29.0229163Z  ━━━╸━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━  8/82 [sentry-sdk] 2025-10-10T01:28:29.1908329Z  ━━━━━━╺━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 13/82 [regex] 2025-10-10T01:28:29.3587982Z  ━━━━━━━━╺━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 17/82 [python-dotenv] 2025-10-10T01:28:29.4788850Z  ━━━━━━━━╸━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 18/82 [pycountry] 2025-10-10T01:28:29.4789332Z  Attempting uninstall: numpy 2025-10-10T01:28:29.4874533Z ━━━━━━━━╸━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 18/82 [pycountry] 2025-10-10T01:28:29.4875013Z  Found existing installation: numpy 1.26.2 2025-10-10T01:28:29.5267196Z ━━━━━━━━╸━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 18/82 [pycountry] 2025-10-10T01:28:29.5452675Z  ━━━━━━━━━━━╺━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 23/82 [numpy] 2025-10-10T01:28:29.5453133Z  Uninstalling numpy-1.26.2: 2025-10-10T01:28:29.6945646Z ━━━━━━━━━━━╺━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 23/82 [numpy] 2025-10-10T01:28:29.8627221Z  ━━━━━━━━━━━╺━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 23/82 [numpy] 2025-10-10T01:28:29.8961593Z  ━━━━━━━━━━━╺━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 23/82 [numpy] 2025-10-10T01:28:29.8962070Z  Successfully uninstalled numpy-1.26.2 2025-10-10T01:28:30.0307346Z ━━━━━━━━━━━╺━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 23/82 [numpy] 2025-10-10T01:28:30.1984919Z  ━━━━━━━━━━━╺━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 23/82 [numpy] 2025-10-10T01:28:30.3678310Z  ━━━━━━━━━━━╺━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 23/82 [numpy] 2025-10-10T01:28:30.5427658Z  ━━━━━━━━━━━╺━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 23/82 [numpy] 2025-10-10T01:28:30.7107004Z  ━━━━━━━━━━━╺━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 23/82 [numpy] 2025-10-10T01:28:30.8788395Z  ━━━━━━━━━━━╺━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 23/82 [numpy] 2025-10-10T01:28:31.0487802Z  ━━━━━━━━━━━╺━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 23/82 [numpy] 2025-10-10T01:28:31.2178321Z  ━━━━━━━━━━━╺━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 23/82 [numpy] 2025-10-10T01:28:31.3858660Z  ━━━━━━━━━━━╺━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 23/82 [numpy] 2025-10-10T01:28:31.4514072Z  ━━━━━━━━━━━╺━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 23/82 [numpy] 2025-10-10T01:28:31.4514534Z  Attempting uninstall: llvmlite 2025-10-10T01:28:31.4555727Z ━━━━━━━━━━━╺━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 23/82 [numpy] 2025-10-10T01:28:31.4556224Z  Found existing installation: llvmlite 0.43.0 2025-10-10T01:28:31.4607493Z ━━━━━━━━━━━╺━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 23/82 [numpy] 2025-10-10T01:28:31.4607969Z  Uninstalling llvmlite-0.43.0: 2025-10-10T01:28:31.5462811Z ━━━━━━━━━━━╺━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 23/82 [numpy] 2025-10-10T01:28:31.5463337Z  Successfully uninstalled llvmlite-0.43.0 2025-10-10T01:28:31.5537840Z ━━━━━━━━━━━╺━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 23/82 [numpy] 2025-10-10T01:28:31.7215828Z  ━━━━━━━━━━━━╸━━━━━━━━━━━━━━━━━━━━━━━━━━━ 26/82 [llvmlite] 2025-10-10T01:28:31.8895726Z  ━━━━━━━━━━━━╸━━━━━━━━━━━━━━━━━━━━━━━━━━━ 26/82 [llvmlite] 2025-10-10T01:28:32.0577569Z  ━━━━━━━━━━━━╸━━━━━━━━━━━━━━━━━━━━━━━━━━━ 26/82 [llvmlite] 2025-10-10T01:28:32.2257686Z  ━━━━━━━━━━━━╸━━━━━━━━━━━━━━━━━━━━━━━━━━━ 26/82 [llvmlite] 2025-10-10T01:28:32.3222322Z  ━━━━━━━━━━━━━╺━━━━━━━━━━━━━━━━━━━━━━━━━━ 27/82 [llguidance] 2025-10-10T01:28:32.3223127Z  Attempting uninstall: lark 2025-10-10T01:28:32.3263192Z ━━━━━━━━━━━━━╺━━━━━━━━━━━━━━━━━━━━━━━━━━ 27/82 [llguidance] 2025-10-10T01:28:32.3263696Z  Found existing installation: lark 0.12.0 2025-10-10T01:28:32.3321820Z ━━━━━━━━━━━━━╺━━━━━━━━━━━━━━━━━━━━━━━━━━ 27/82 [llguidance] 2025-10-10T01:28:32.3322262Z  Uninstalling lark-0.12.0: 2025-10-10T01:28:32.3568011Z ━━━━━━━━━━━━━╺━━━━━━━━━━━━━━━━━━━━━━━━━━ 27/82 [llguidance] 2025-10-10T01:28:32.3568459Z  Successfully uninstalled lark-0.12.0 2025-10-10T01:28:32.3950208Z ━━━━━━━━━━━━━╺━━━━━━━━━━━━━━━━━━━━━━━━━━ 27/82 [llguidance] 2025-10-10T01:28:32.5628645Z  ━━━━━━━━━━━━━╸━━━━━━━━━━━━━━━━━━━━━━━━━━ 28/82 [lark] 2025-10-10T01:28:32.7318073Z  ━━━━━━━━━━━━━━━━━╺━━━━━━━━━━━━━━━━━━━━━━ 35/82 [einops] 2025-10-10T01:28:32.9007679Z  ━━━━━━━━━━━━━━━━━╸━━━━━━━━━━━━━━━━━━━━━━ 36/82 [dnspython] 2025-10-10T01:28:33.0686873Z  ━━━━━━━━━━━━━━━━━━━━╸━━━━━━━━━━━━━━━━━━━ 43/82 [uvicorn] 2025-10-10T01:28:33.2366140Z  ━━━━━━━━━━━━━━━━━━━━━━━╺━━━━━━━━━━━━━━━━ 48/82 [opencv-python-headless] 2025-10-10T01:28:33.4046341Z  ━━━━━━━━━━━━━━━━━━━━━━━╺━━━━━━━━━━━━━━━━ 48/82 [opencv-python-headless] 2025-10-10T01:28:33.5726020Z  ━━━━━━━━━━━━━━━━━━━━━━━╺━━━━━━━━━━━━━━━━ 48/82 [opencv-python-headless] 2025-10-10T01:28:33.6190146Z  ━━━━━━━━━━━━━━━━━━━━━━━╺━━━━━━━━━━━━━━━━ 48/82 [opencv-python-headless] 2025-10-10T01:28:33.6190625Z  Attempting uninstall: numba 2025-10-10T01:28:33.6235493Z ━━━━━━━━━━━━━━━━━━━━━━━╺━━━━━━━━━━━━━━━━ 48/82 [opencv-python-headless] 2025-10-10T01:28:33.6236293Z  Found existing installation: numba 0.60.0 2025-10-10T01:28:33.6972874Z ━━━━━━━━━━━━━━━━━━━━━━━╺━━━━━━━━━━━━━━━━ 48/82 [opencv-python-headless] 2025-10-10T01:28:33.6973340Z  Uninstalling numba-0.60.0: 2025-10-10T01:28:33.7407349Z ━━━━━━━━━━━━━━━━━━━━━━━╺━━━━━━━━━━━━━━━━ 48/82 [opencv-python-headless] 2025-10-10T01:28:33.9087655Z  ━━━━━━━━━━━━━━━━━━━━━━━╸━━━━━━━━━━━━━━━━ 49/82 [numba] 2025-10-10T01:28:34.0766835Z  ━━━━━━━━━━━━━━━━━━━━━━━╸━━━━━━━━━━━━━━━━ 49/82 [numba] 2025-10-10T01:28:34.0868554Z  ━━━━━━━━━━━━━━━━━━━━━━━╸━━━━━━━━━━━━━━━━ 49/82 [numba] 2025-10-10T01:28:34.0869029Z  Successfully uninstalled numba-0.60.0 2025-10-10T01:28:34.2448447Z ━━━━━━━━━━━━━━━━━━━━━━━╸━━━━━━━━━━━━━━━━ 49/82 [numba] 2025-10-10T01:28:34.4230913Z  ━━━━━━━━━━━━━━━━━━━━━━━╸━━━━━━━━━━━━━━━━ 49/82 [numba] 2025-10-10T01:28:34.5908177Z  ━━━━━━━━━━━━━━━━━━━━━━━╸━━━━━━━━━━━━━━━━ 49/82 [numba] 2025-10-10T01:28:34.7601354Z  ━━━━━━━━━━━━━━━━━━━━━━━╸━━━━━━━━━━━━━━━━ 49/82 [numba] 2025-10-10T01:28:34.9278568Z  ━━━━━━━━━━━━━━━━━━━━━━━╸━━━━━━━━━━━━━━━━ 49/82 [numba] 2025-10-10T01:28:35.1046447Z  ━━━━━━━━━━━━━━━━━━━━━━━╸━━━━━━━━━━━━━━━━ 49/82 [numba] 2025-10-10T01:28:35.2738671Z  ━━━━━━━━━━━━━━━━━━━━━━━╸━━━━━━━━━━━━━━━━ 49/82 [numba] 2025-10-10T01:28:35.4417481Z  ━━━━━━━━━━━━━━━━━━━━━━━╸━━━━━━━━━━━━━━━━ 49/82 [numba] 2025-10-10T01:28:35.6436718Z  ━━━━━━━━━━━━━━━━━━━━━━━╸━━━━━━━━━━━━━━━━ 49/82 [numba] 2025-10-10T01:28:35.8121098Z  ━━━━━━━━━━━━━━━━━━━━━━━╸━━━━━━━━━━━━━━━━ 49/82 [numba] 2025-10-10T01:28:35.9799598Z  ━━━━━━━━━━━━━━━━━━━━━━━╸━━━━━━━━━━━━━━━━ 49/82 [numba] 2025-10-10T01:28:36.1480743Z  ━━━━━━━━━━━━━━━━━━━━━━━╸━━━━━━━━━━━━━━━━ 49/82 [numba] 2025-10-10T01:28:36.3171865Z  ━━━━━━━━━━━━━━━━━━━━━━━━╸━━━━━━━━━━━━━━━ 51/82 [huggingface-hub] 2025-10-10T01:28:36.4850514Z  ━━━━━━━━━━━━━━━━━━━━━━━━━╺━━━━━━━━━━━━━━ 52/82 [httpcore] 2025-10-10T01:28:36.6530902Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━╺━━━━━━━━━━━━ 56/82 [cupy-cuda12x] 2025-10-10T01:28:36.8206502Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━╺━━━━━━━━━━━━ 56/82 [cupy-cuda12x] 2025-10-10T01:28:36.9883064Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━╺━━━━━━━━━━━━ 56/82 [cupy-cuda12x] 2025-10-10T01:28:37.1558407Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━╺━━━━━━━━━━━━ 56/82 [cupy-cuda12x] 2025-10-10T01:28:37.3234687Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━╺━━━━━━━━━━━━ 56/82 [cupy-cuda12x] 2025-10-10T01:28:37.4910367Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━╺━━━━━━━━━━━━ 56/82 [cupy-cuda12x] 2025-10-10T01:28:37.6585695Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━╺━━━━━━━━━━━━ 56/82 [cupy-cuda12x] 2025-10-10T01:28:37.8280275Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━╺━━━━━━━━━━━━ 56/82 [cupy-cuda12x] 2025-10-10T01:28:37.9997983Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━╺━━━━━━━━━━━━ 56/82 [cupy-cuda12x] 2025-10-10T01:28:38.1682578Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━╺━━━━━━━━━━━━ 56/82 [cupy-cuda12x] 2025-10-10T01:28:38.3361961Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━╺━━━━━━━━━━━━ 56/82 [cupy-cuda12x] 2025-10-10T01:28:38.5040999Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━╸━━━━━━━━━━━ 59/82 [tokenizers] 2025-10-10T01:28:38.6720758Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╸━━━━━━━━━━ 61/82 [rich] 2025-10-10T01:28:38.8402569Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╺━━━━━━━━━ 62/82 [pydantic] 2025-10-10T01:28:39.0097594Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╺━━━━━━━━━ 62/82 [pydantic] 2025-10-10T01:28:39.1772597Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╺━━━━━━━ 66/82 [transformers] 2025-10-10T01:28:39.3448217Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╺━━━━━━━ 66/82 [transformers] 2025-10-10T01:28:39.5130490Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╺━━━━━━━ 66/82 [transformers] 2025-10-10T01:28:39.6802900Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╺━━━━━━━ 66/82 [transformers] 2025-10-10T01:28:39.8635699Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╺━━━━━━━ 66/82 [transformers] 2025-10-10T01:28:40.0319751Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╺━━━━━━━ 66/82 [transformers] 2025-10-10T01:28:40.2031029Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╺━━━━━━━ 66/82 [transformers] 2025-10-10T01:28:40.3716756Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╺━━━━━━━ 66/82 [transformers] 2025-10-10T01:28:40.5407996Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╺━━━━━━━ 66/82 [transformers] 2025-10-10T01:28:40.7083963Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╺━━━━━━━ 66/82 [transformers] 2025-10-10T01:28:40.8763396Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╺━━━━━━━ 66/82 [transformers] 2025-10-10T01:28:41.0470415Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╺━━━━━━━ 66/82 [transformers] 2025-10-10T01:28:41.2166017Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╺━━━━━━━ 66/82 [transformers] 2025-10-10T01:28:41.3880352Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╺━━━━━━━ 66/82 [transformers] 2025-10-10T01:28:41.5569977Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╺━━━━━━━ 66/82 [transformers] 2025-10-10T01:28:41.7265623Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╺━━━━━━━ 66/82 [transformers] 2025-10-10T01:28:41.8940469Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╺━━━━━━━ 66/82 [transformers] 2025-10-10T01:28:42.0630246Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╺━━━━━━━ 66/82 [transformers] 2025-10-10T01:28:42.2306290Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╺━━━━━━━ 66/82 [transformers] 2025-10-10T01:28:42.4039076Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╺━━━━━━━ 66/82 [transformers] 2025-10-10T01:28:42.5720915Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╺━━━━━━━ 66/82 [transformers] 2025-10-10T01:28:42.7435377Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╺━━━━━━━ 66/82 [transformers] 2025-10-10T01:28:42.9110506Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╺━━━━━━━ 66/82 [transformers] 2025-10-10T01:28:43.0786449Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╺━━━━━━━ 66/82 [transformers] 2025-10-10T01:28:43.2461360Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╺━━━━━━━ 66/82 [transformers] 2025-10-10T01:28:43.4143057Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╺━━━━━━━ 66/82 [transformers] 2025-10-10T01:28:43.5853285Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╺━━━━━━━ 66/82 [transformers] 2025-10-10T01:28:43.7550625Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╺━━━━━━━ 66/82 [transformers] 2025-10-10T01:28:43.9225207Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╺━━━━━━━ 66/82 [transformers] 2025-10-10T01:28:44.0924840Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╺━━━━━━━ 66/82 [transformers] 2025-10-10T01:28:44.2602829Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╺━━━━━━━ 66/82 [transformers] 2025-10-10T01:28:44.4282185Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╺━━━━━━━ 66/82 [transformers] 2025-10-10T01:28:44.5960109Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╺━━━━━━━ 66/82 [transformers] 2025-10-10T01:28:44.7635935Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╺━━━━━ 70/82 [openai-harmony] 2025-10-10T01:28:44.9323427Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╸━━━━━ 71/82 [openai] 2025-10-10T01:28:45.0998304Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╸━━━━━ 71/82 [openai] 2025-10-10T01:28:45.2671563Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╸━━━━━ 71/82 [openai] 2025-10-10T01:28:45.4348589Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╸━━━━━ 71/82 [openai] 2025-10-10T01:28:45.6023750Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╺━━━ 74/82 [fastapi] 2025-10-10T01:28:45.7698925Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╸━━━ 75/82 [xgrammar] 2025-10-10T01:28:45.9372924Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╺━━ 76/82 [ray] 2025-10-10T01:28:46.1047712Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╺━━ 76/82 [ray] 2025-10-10T01:28:46.2722364Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╺━━ 76/82 [ray] 2025-10-10T01:28:46.4396909Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╺━━ 76/82 [ray] 2025-10-10T01:28:46.6113138Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╺━━ 76/82 [ray] 2025-10-10T01:28:46.7788399Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╺━━ 76/82 [ray] 2025-10-10T01:28:46.9592883Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╺━━ 76/82 [ray] 2025-10-10T01:28:47.1269213Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╺━━ 76/82 [ray] 2025-10-10T01:28:47.2948008Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╺━━ 76/82 [ray] 2025-10-10T01:28:47.4621950Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╺━━ 76/82 [ray] 2025-10-10T01:28:47.6299461Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╺━━ 76/82 [ray] 2025-10-10T01:28:47.8088032Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╺━━ 76/82 [ray] 2025-10-10T01:28:47.9763747Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╺━━ 76/82 [ray] 2025-10-10T01:28:48.1443902Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╺━━ 76/82 [ray] 2025-10-10T01:28:48.3122051Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╺━━ 76/82 [ray] 2025-10-10T01:28:48.4837767Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╺━━ 76/82 [ray] 2025-10-10T01:28:48.6513699Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╺━━ 76/82 [ray] 2025-10-10T01:28:48.8211965Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╺━━ 76/82 [ray] 2025-10-10T01:28:48.9902446Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╺━━ 76/82 [ray] 2025-10-10T01:28:49.1578186Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╺━━ 76/82 [ray] 2025-10-10T01:28:49.3261375Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╺━━ 76/82 [ray] 2025-10-10T01:28:49.4941930Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╺━━ 76/82 [ray] 2025-10-10T01:28:49.6620340Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╺━━ 76/82 [ray] 2025-10-10T01:28:49.8295034Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╺━ 78/82 [fastapi-cli] 2025-10-10T01:28:49.9970218Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╸━ 79/82 [compressed-tensors] 2025-10-10T01:28:50.1649082Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╺ 80/82 [mistral_common] 2025-10-10T01:28:50.3325125Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╺ 80/82 [mistral_common] 2025-10-10T01:28:50.4999298Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╸ 81/82 [vllm] 2025-10-10T01:28:50.6673946Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╸ 81/82 [vllm] 2025-10-10T01:28:50.8349390Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╸ 81/82 [vllm] 2025-10-10T01:28:51.0024902Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╸ 81/82 [vllm] 2025-10-10T01:28:51.1700442Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╸ 81/82 [vllm] 2025-10-10T01:28:51.3375738Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╸ 81/82 [vllm] 2025-10-10T01:28:51.5050892Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╸ 81/82 [vllm] 2025-10-10T01:28:51.6724863Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╸ 81/82 [vllm] 2025-10-10T01:28:51.8402023Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╸ 81/82 [vllm] 2025-10-10T01:28:52.0075494Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╸ 81/82 [vllm] 2025-10-10T01:28:52.1749582Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╸ 81/82 [vllm] 2025-10-10T01:28:52.3424363Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╸ 81/82 [vllm] 2025-10-10T01:28:52.5098024Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╸ 81/82 [vllm] 2025-10-10T01:28:52.6772908Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╸ 81/82 [vllm] 2025-10-10T01:28:52.8446794Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╸ 81/82 [vllm] 2025-10-10T01:28:53.0121575Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╸ 81/82 [vllm] 2025-10-10T01:28:53.1799435Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╸ 81/82 [vllm] 2025-10-10T01:28:53.3473430Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╸ 81/82 [vllm] 2025-10-10T01:28:53.5154656Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╸ 81/82 [vllm] 2025-10-10T01:28:53.6824233Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╸ 81/82 [vllm] 2025-10-10T01:28:53.8498085Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╸ 81/82 [vllm] 2025-10-10T01:28:54.0171513Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╸ 81/82 [vllm] 2025-10-10T01:28:54.1845485Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╸ 81/82 [vllm] 2025-10-10T01:28:54.3520623Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╸ 81/82 [vllm] 2025-10-10T01:28:54.5193722Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╸ 81/82 [vllm] 2025-10-10T01:28:54.6868126Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╸ 81/82 [vllm] 2025-10-10T01:28:54.8612036Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╸ 81/82 [vllm] 2025-10-10T01:28:55.0312909Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╸ 81/82 [vllm] 2025-10-10T01:28:55.1992081Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╸ 81/82 [vllm] 2025-10-10T01:28:55.3734659Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╸ 81/82 [vllm] 2025-10-10T01:28:55.5414454Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╸ 81/82 [vllm] 2025-10-10T01:28:55.7111712Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╸ 81/82 [vllm] 2025-10-10T01:28:55.8794297Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╸ 81/82 [vllm] 2025-10-10T01:28:56.0471157Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╸ 81/82 [vllm] 2025-10-10T01:28:56.2149206Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╸ 81/82 [vllm] 2025-10-10T01:28:56.3824881Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╸ 81/82 [vllm] 2025-10-10T01:28:56.4502001Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╸ 81/82 [vllm] 2025-10-10T01:28:56.4506551Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 82/82 [vllm] 2025-10-10T01:28:56.4507296Z [?25h 2025-10-10T01:28:56.4651668Z Successfully installed annotated-types-0.7.0 anyio-4.11.0 astor-0.8.1 blake3-1.0.7 cbor2-5.7.0 cloudpickle-3.1.1 compressed-tensors-0.11.0 cupy-cuda12x-13.6.0 depyf-0.19.0 diskcache-5.6.3 dnspython-2.8.0 einops-0.8.1 email-validator-2.3.0 fastapi-0.118.2 fastapi-cli-0.0.13 fastapi-cloud-cli-0.3.1 fastrlock-0.8.3 frozendict-2.4.6 gguf-0.17.1 h11-0.16.0 hf-xet-1.1.10 httpcore-1.0.9 httptools-0.6.4 httpx-0.28.1 huggingface-hub-0.35.3 interegular-0.3.3 jiter-0.11.0 jsonschema-4.25.1 jsonschema-specifications-2025.9.1 lark-1.2.2 llguidance-0.7.30 llvmlite-0.44.0 lm-format-enforcer-0.11.3 markdown-it-py-4.0.0 mdurl-0.1.2 mistral_common-1.8.5 msgspec-0.19.0 numba-0.61.2 numpy-2.2.6 openai-2.3.0 openai-harmony-0.0.4 opencv-python-headless-4.12.0.88 outlines_core-0.2.11 partial-json-parser-0.2.1.1.post6 prometheus-fastapi-instrumentator-7.1.0 prometheus_client-0.23.1 py-cpuinfo-9.0.0 pybase64-1.4.2 pycountry-24.6.1 pydantic-2.12.0 pydantic-core-2.41.1 pydantic-extra-types-2.10.6 python-dotenv-1.1.1 python-json-logger-4.0.0 python-multipart-0.0.20 pyzmq-27.1.0 ray-2.49.2 referencing-0.36.2 regex-2025.9.18 rich-14.2.0 rich-toolkit-0.15.1 rignore-0.7.0 rpds-py-0.27.1 safetensors-0.6.2 sentencepiece-0.2.1 sentry-sdk-2.41.0 setproctitle-1.3.7 shellingham-1.5.4 sniffio-1.3.1 starlette-0.48.0 tiktoken-0.12.0 tokenizers-0.22.1 transformers-4.57.0 typer-0.19.2 typing-inspection-0.4.2 typing_extensions-4.15.0 uvicorn-0.37.0 uvloop-0.21.0 vllm-0.11.0rc2.dev157+g0ad9951c4.d20251010 watchfiles-1.1.0 websockets-15.0.1 xgrammar-0.1.25 2025-10-10T01:28:57.0224934Z 2025-10-10 01:28:57,021 [INFO] cli.lib.common.pip_helper: Installing /var/lib/jenkins/workspace/dist/external/vllm/wheels/flashinfer-python/flashinfer_python-0.2.14.post1-cp39-abi3-linux_x86_64.whl... 2025-10-10T01:28:57.0228312Z 2025-10-10 01:28:57,022 [INFO] cli.lib.common.pip_helper: pip installing packages: /opt/conda/envs/py_3.12/bin/python -m pip install /var/lib/jenkins/workspace/dist/external/vllm/wheels/flashinfer-python/flashinfer_python-0.2.14.post1-cp39-abi3-linux_x86_64.whl 2025-10-10T01:28:57.0231209Z 2025-10-10 01:28:57,022 [INFO] cli.lib.common.utils: [cmd] /opt/conda/envs/py_3.12/bin/python -m pip install /var/lib/jenkins/workspace/dist/external/vllm/wheels/flashinfer-python/flashinfer_python-0.2.14.post1-cp39-abi3-linux_x86_64.whl 2025-10-10T01:28:57.3237786Z Processing /var/lib/jenkins/workspace/dist/external/vllm/wheels/flashinfer-python/flashinfer_python-0.2.14.post1-cp39-abi3-linux_x86_64.whl 2025-10-10T01:28:57.3981018Z Requirement already satisfied: numpy in /opt/conda/envs/py_3.12/lib/python3.12/site-packages (from flashinfer-python==0.2.14.post1) (2.2.6) 2025-10-10T01:28:57.3983089Z Requirement already satisfied: torch in /opt/conda/envs/py_3.12/lib/python3.12/site-packages (from flashinfer-python==0.2.14.post1) (2.10.0a0+git344e636) 2025-10-10T01:28:57.3985406Z Requirement already satisfied: ninja in /opt/conda/envs/py_3.12/lib/python3.12/site-packages (from flashinfer-python==0.2.14.post1) (1.10.2) 2025-10-10T01:28:57.3987642Z Requirement already satisfied: requests in /opt/conda/envs/py_3.12/lib/python3.12/site-packages (from flashinfer-python==0.2.14.post1) (2.32.5) 2025-10-10T01:28:57.5340321Z Collecting cuda-python<=12.9 (from flashinfer-python==0.2.14.post1) 2025-10-10T01:28:57.5551303Z Downloading cuda_python-12.9.0-py3-none-any.whl.metadata (4.6 kB) 2025-10-10T01:28:57.6129413Z Collecting pynvml (from flashinfer-python==0.2.14.post1) 2025-10-10T01:28:57.6219407Z Downloading pynvml-13.0.1-py3-none-any.whl.metadata (5.6 kB) 2025-10-10T01:28:57.6303862Z Requirement already satisfied: einops in /opt/conda/envs/py_3.12/lib/python3.12/site-packages (from flashinfer-python==0.2.14.post1) (0.8.1) 2025-10-10T01:28:57.6595506Z Collecting packaging>=24.2 (from flashinfer-python==0.2.14.post1) 2025-10-10T01:28:57.6651473Z Downloading packaging-25.0-py3-none-any.whl.metadata (3.3 kB) 2025-10-10T01:28:57.7018152Z Collecting nvidia-cudnn-frontend>=1.13.0 (from flashinfer-python==0.2.14.post1) 2025-10-10T01:28:57.7111234Z Downloading nvidia_cudnn_frontend-1.14.1-cp312-cp312-manylinux_2_27_x86_64.manylinux_2_28_x86_64.whl.metadata (6.7 kB) 2025-10-10T01:28:57.7194909Z Requirement already satisfied: cuda-bindings~=12.9.0 in /opt/conda/envs/py_3.12/lib/python3.12/site-packages (from cuda-python<=12.9->flashinfer-python==0.2.14.post1) (12.9.2) 2025-10-10T01:28:57.7203152Z Requirement already satisfied: cuda-pathfinder~=1.1 in /opt/conda/envs/py_3.12/lib/python3.12/site-packages (from cuda-bindings~=12.9.0->cuda-python<=12.9->flashinfer-python==0.2.14.post1) (1.3.0) 2025-10-10T01:28:57.7237550Z Requirement already satisfied: filelock in /opt/conda/envs/py_3.12/lib/python3.12/site-packages (from torch->flashinfer-python==0.2.14.post1) (3.18.0) 2025-10-10T01:28:57.7241252Z Requirement already satisfied: typing-extensions>=4.10.0 in /opt/conda/envs/py_3.12/lib/python3.12/site-packages (from torch->flashinfer-python==0.2.14.post1) (4.15.0) 2025-10-10T01:28:57.7244298Z Requirement already satisfied: setuptools in /opt/conda/envs/py_3.12/lib/python3.12/site-packages (from torch->flashinfer-python==0.2.14.post1) (78.1.1) 2025-10-10T01:28:57.7247990Z Requirement already satisfied: sympy>=1.13.3 in /opt/conda/envs/py_3.12/lib/python3.12/site-packages (from torch->flashinfer-python==0.2.14.post1) (1.13.3) 2025-10-10T01:28:57.7251598Z Requirement already satisfied: networkx>=2.5.1 in /opt/conda/envs/py_3.12/lib/python3.12/site-packages (from torch->flashinfer-python==0.2.14.post1) (2.8.8) 2025-10-10T01:28:57.7255042Z Requirement already satisfied: jinja2 in /opt/conda/envs/py_3.12/lib/python3.12/site-packages (from torch->flashinfer-python==0.2.14.post1) (3.1.6) 2025-10-10T01:28:57.7258583Z Requirement already satisfied: fsspec>=0.8.5 in /opt/conda/envs/py_3.12/lib/python3.12/site-packages (from torch->flashinfer-python==0.2.14.post1) (2025.9.0) 2025-10-10T01:28:57.7371081Z Requirement already satisfied: mpmath<1.4,>=1.1.0 in /opt/conda/envs/py_3.12/lib/python3.12/site-packages (from sympy>=1.13.3->torch->flashinfer-python==0.2.14.post1) (1.3.0) 2025-10-10T01:28:57.7417980Z Requirement already satisfied: MarkupSafe>=2.0 in /opt/conda/envs/py_3.12/lib/python3.12/site-packages (from jinja2->torch->flashinfer-python==0.2.14.post1) (3.0.3) 2025-10-10T01:28:57.7656699Z Collecting nvidia-ml-py>=12.0.0 (from pynvml->flashinfer-python==0.2.14.post1) 2025-10-10T01:28:57.7700517Z Downloading nvidia_ml_py-13.580.82-py3-none-any.whl.metadata (9.6 kB) 2025-10-10T01:28:57.7789427Z Requirement already satisfied: charset_normalizer<4,>=2 in /opt/conda/envs/py_3.12/lib/python3.12/site-packages (from requests->flashinfer-python==0.2.14.post1) (3.4.3) 2025-10-10T01:28:57.7792744Z Requirement already satisfied: idna<4,>=2.5 in /opt/conda/envs/py_3.12/lib/python3.12/site-packages (from requests->flashinfer-python==0.2.14.post1) (3.10) 2025-10-10T01:28:57.7797162Z Requirement already satisfied: urllib3<3,>=1.21.1 in /opt/conda/envs/py_3.12/lib/python3.12/site-packages (from requests->flashinfer-python==0.2.14.post1) (2.5.0) 2025-10-10T01:28:57.7800856Z Requirement already satisfied: certifi>=2017.4.17 in /opt/conda/envs/py_3.12/lib/python3.12/site-packages (from requests->flashinfer-python==0.2.14.post1) (2025.10.5) 2025-10-10T01:28:57.8019796Z Downloading cuda_python-12.9.0-py3-none-any.whl (7.5 kB) 2025-10-10T01:28:57.8238835Z Downloading nvidia_cudnn_frontend-1.14.1-cp312-cp312-manylinux_2_27_x86_64.manylinux_2_28_x86_64.whl (1.8 MB) 2025-10-10T01:28:57.8720335Z [?25l ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 0.0/1.8 MB ? eta -:--:-- 2025-10-10T01:28:57.8721535Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 1.8/1.8 MB 52.0 MB/s 0:00:00 2025-10-10T01:28:57.8820907Z [?25hDownloading packaging-25.0-py3-none-any.whl (66 kB) 2025-10-10T01:28:57.8950429Z Downloading pynvml-13.0.1-py3-none-any.whl (28 kB) 2025-10-10T01:28:57.9425397Z Downloading nvidia_ml_py-13.580.82-py3-none-any.whl (49 kB) 2025-10-10T01:28:58.2361990Z Installing collected packages: nvidia-ml-py, pynvml, packaging, nvidia-cudnn-frontend, cuda-python, flashinfer-python 2025-10-10T01:28:58.3072961Z [?25l 2025-10-10T01:28:58.3073628Z  Attempting uninstall: packaging 2025-10-10T01:28:58.3120928Z 2025-10-10T01:28:58.3121568Z  Found existing installation: packaging 23.1 2025-10-10T01:28:58.3152133Z 2025-10-10T01:28:58.3153006Z  Uninstalling packaging-23.1: 2025-10-10T01:28:58.3253851Z 2025-10-10T01:28:58.3254505Z  Successfully uninstalled packaging-23.1 2025-10-10T01:28:58.4041903Z 2025-10-10T01:28:58.5719524Z  ━━━━━━━━━━━━━━━━━━━━╺━━━━━━━━━━━━━━━━━━━ 3/6 [nvidia-cudnn-frontend] 2025-10-10T01:28:58.7394699Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╺━━━━━━ 5/6 [flashinfer-python] 2025-10-10T01:28:58.9071883Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╺━━━━━━ 5/6 [flashinfer-python] 2025-10-10T01:28:59.0746520Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╺━━━━━━ 5/6 [flashinfer-python] 2025-10-10T01:28:59.2422939Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╺━━━━━━ 5/6 [flashinfer-python] 2025-10-10T01:28:59.4096683Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╺━━━━━━ 5/6 [flashinfer-python] 2025-10-10T01:28:59.5773059Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╺━━━━━━ 5/6 [flashinfer-python] 2025-10-10T01:28:59.7447726Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╺━━━━━━ 5/6 [flashinfer-python] 2025-10-10T01:28:59.9125678Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╺━━━━━━ 5/6 [flashinfer-python] 2025-10-10T01:28:59.9488496Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╺━━━━━━ 5/6 [flashinfer-python] 2025-10-10T01:28:59.9489649Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 6/6 [flashinfer-python] 2025-10-10T01:28:59.9490316Z [?25h 2025-10-10T01:28:59.9625985Z Successfully installed cuda-python-12.9.0 flashinfer-python-0.2.14.post1 nvidia-cudnn-frontend-1.14.1 nvidia-ml-py-13.580.82 packaging-25.0 pynvml-13.0.1 2025-10-10T01:29:00.0192659Z 2025-10-10 01:29:00,018 [INFO] cli.lib.core.vllm.vllm_test: Done. Installed vllm wheels 2025-10-10T01:29:00.0194902Z 2025-10-10 01:29:00,019 [INFO] cli.lib.common.pip_helper: pip installing packages: /opt/conda/envs/py_3.12/bin/python -m uv pip install -e tests/vllm_test_utils 2025-10-10T01:29:00.0196750Z 2025-10-10 01:29:00,019 [INFO] cli.lib.common.utils: [cmd] /opt/conda/envs/py_3.12/bin/python -m uv pip install -e tests/vllm_test_utils 2025-10-10T01:29:00.1369354Z Using Python 3.12.11 environment at: /opt/conda/envs/py_3.12 2025-10-10T01:29:00.8804018Z Resolved 1 package in 694ms 2025-10-10T01:29:00.8823515Z Building vllm-test-utils @ file:///var/lib/jenkins/workspace/vllm/tests/vllm_test_utils 2025-10-10T01:29:01.0609511Z Built vllm-test-utils @ file:///var/lib/jenkins/workspace/vllm/tests/vllm_test_utils 2025-10-10T01:29:01.0630285Z Prepared 1 package in 181ms 2025-10-10T01:29:01.0658140Z Installed 1 package in 2ms 2025-10-10T01:29:01.0659487Z + vllm-test-utils==0.1 (from file:///var/lib/jenkins/workspace/vllm/tests/vllm_test_utils) 2025-10-10T01:29:01.0883576Z 2025-10-10 01:29:01,087 [INFO] cli.lib.common.pip_helper: pip installing packages: /opt/conda/envs/py_3.12/bin/python -m uv pip install hf_transfer 2025-10-10T01:29:01.0885877Z 2025-10-10 01:29:01,088 [INFO] cli.lib.common.utils: [cmd] /opt/conda/envs/py_3.12/bin/python -m uv pip install hf_transfer 2025-10-10T01:29:01.1445120Z Using Python 3.12.11 environment at: /opt/conda/envs/py_3.12 2025-10-10T01:29:01.1770834Z Resolved 1 package in 25ms 2025-10-10T01:29:01.1819497Z Downloading hf-transfer (3.4MiB) 2025-10-10T01:29:01.2091445Z Downloading hf-transfer 2025-10-10T01:29:01.2094159Z Prepared 1 package in 32ms 2025-10-10T01:29:01.2213675Z Installed 1 package in 11ms 2025-10-10T01:29:01.2214680Z + hf-transfer==0.1.9 2025-10-10T01:29:01.2268783Z 2025-10-10 01:29:01,226 [INFO] cli.lib.common.utils: [cmd] /opt/conda/envs/py_3.12/bin/python use_existing_torch.py 2025-10-10T01:29:01.2652067Z >>> cleaning requirements/build.txt 2025-10-10T01:29:01.2652477Z removed: 2025-10-10T01:29:01.2653605Z <<< done cleaning requirements/build.txt 2025-10-10T01:29:01.2653905Z 2025-10-10T01:29:01.2669728Z >>> cleaning requirements/cpu-build.txt 2025-10-10T01:29:01.2670029Z removed: 2025-10-10T01:29:01.2670354Z <<< done cleaning requirements/cpu-build.txt 2025-10-10T01:29:01.2670558Z 2025-10-10T01:29:01.2670641Z >>> cleaning requirements/dev.txt 2025-10-10T01:29:01.2670918Z <<< done cleaning requirements/dev.txt 2025-10-10T01:29:01.2671154Z 2025-10-10T01:29:01.2671237Z >>> cleaning requirements/docs.txt 2025-10-10T01:29:01.2671443Z removed: 2025-10-10T01:29:01.2671616Z <<< done cleaning requirements/docs.txt 2025-10-10T01:29:01.2671775Z 2025-10-10T01:29:01.2671853Z >>> cleaning requirements/lint.txt 2025-10-10T01:29:01.2672212Z <<< done cleaning requirements/lint.txt 2025-10-10T01:29:01.2672397Z 2025-10-10T01:29:01.2672493Z >>> cleaning requirements/rocm-build.txt 2025-10-10T01:29:01.2672708Z removed: 2025-10-10T01:29:01.2672983Z <<< done cleaning requirements/rocm-build.txt 2025-10-10T01:29:01.2673178Z 2025-10-10T01:29:01.2673281Z >>> cleaning requirements/rocm-test.txt 2025-10-10T01:29:01.2673553Z <<< done cleaning requirements/rocm-test.txt 2025-10-10T01:29:01.2673713Z 2025-10-10T01:29:01.2673796Z >>> cleaning requirements/tpu.txt 2025-10-10T01:29:01.2673994Z removed: 2025-10-10T01:29:01.2674158Z <<< done cleaning requirements/tpu.txt 2025-10-10T01:29:01.2674379Z 2025-10-10T01:29:01.2674496Z >>> cleaning requirements/common.txt 2025-10-10T01:29:01.2674733Z <<< done cleaning requirements/common.txt 2025-10-10T01:29:01.2674886Z 2025-10-10T01:29:01.2674961Z >>> cleaning requirements/cpu.txt 2025-10-10T01:29:01.2675157Z removed: 2025-10-10T01:29:01.2675320Z <<< done cleaning requirements/cpu.txt 2025-10-10T01:29:01.2675465Z 2025-10-10T01:29:01.2675548Z >>> cleaning requirements/cuda.txt 2025-10-10T01:29:01.2675739Z removed: 2025-10-10T01:29:01.2675913Z <<< done cleaning requirements/cuda.txt 2025-10-10T01:29:01.2676070Z 2025-10-10T01:29:01.2676158Z >>> cleaning requirements/kv_connectors.txt 2025-10-10T01:29:01.2676421Z <<< done cleaning requirements/kv_connectors.txt 2025-10-10T01:29:01.2676595Z 2025-10-10T01:29:01.2676697Z >>> cleaning requirements/nightly_torch_test.txt 2025-10-10T01:29:01.2676989Z <<< done cleaning requirements/nightly_torch_test.txt 2025-10-10T01:29:01.2677178Z 2025-10-10T01:29:01.2677255Z >>> cleaning requirements/rocm.txt 2025-10-10T01:29:01.2677483Z <<< done cleaning requirements/rocm.txt 2025-10-10T01:29:01.2677636Z 2025-10-10T01:29:01.2677719Z >>> cleaning requirements/test.txt 2025-10-10T01:29:01.2677911Z removed: 2025-10-10T01:29:01.2678081Z <<< done cleaning requirements/test.txt 2025-10-10T01:29:01.2678243Z 2025-10-10T01:29:01.2678322Z >>> cleaning requirements/xpu.txt 2025-10-10T01:29:01.2678608Z removed: 2025-10-10T01:29:01.2678769Z <<< done cleaning requirements/xpu.txt 2025-10-10T01:29:01.2678928Z 2025-10-10T01:29:01.2679010Z >>> cleaning pyproject.toml 2025-10-10T01:29:01.2679204Z removed: 2025-10-10T01:29:01.2679377Z <<< done cleaning pyproject.toml 2025-10-10T01:29:01.2679515Z 2025-10-10T01:29:01.2734439Z 2025-10-10 01:29:01,273 [INFO] cli.lib.common.pip_helper: pip installing packages: /opt/conda/envs/py_3.12/bin/python -m uv pip install -r requirements/common.txt 2025-10-10T01:29:01.2735754Z 2025-10-10 01:29:01,273 [INFO] cli.lib.common.utils: [cmd] /opt/conda/envs/py_3.12/bin/python -m uv pip install -r requirements/common.txt 2025-10-10T01:29:01.3236723Z Using Python 3.12.11 environment at: /opt/conda/envs/py_3.12 2025-10-10T01:29:01.3398328Z Audited 50 packages in 18ms 2025-10-10T01:29:01.3427541Z 2025-10-10 01:29:01,342 [INFO] cli.lib.common.pip_helper: pip installing packages: /opt/conda/envs/py_3.12/bin/python -m uv pip install -r requirements/build.txt 2025-10-10T01:29:01.3428800Z 2025-10-10 01:29:01,342 [INFO] cli.lib.common.utils: [cmd] /opt/conda/envs/py_3.12/bin/python -m uv pip install -r requirements/build.txt 2025-10-10T01:29:01.3921763Z Using Python 3.12.11 environment at: /opt/conda/envs/py_3.12 2025-10-10T01:29:01.4660786Z Resolved 11 packages in 68ms 2025-10-10T01:29:01.4793757Z Prepared 1 package in 13ms 2025-10-10T01:29:01.4998960Z Installed 1 package in 20ms 2025-10-10T01:29:01.4999893Z + setuptools-scm==9.2.0 2025-10-10T01:29:01.5077926Z 2025-10-10 01:29:01,507 [INFO] cli.lib.core.vllm.vllm_test: generate test.txt from requirements/test.in with local torch whls 2025-10-10T01:29:01.8728530Z 2025-10-10 01:29:01,872 [INFO] cli.lib.core.vllm.vllm_test: [INFO] Updated requirements/test.in 2025-10-10T01:29:01.8732085Z 2025-10-10 01:29:01,872 [INFO] cli.lib.common.utils: [cmd] /opt/conda/envs/py_3.12/bin/python -m uv pip compile requirements/test.in -o test.txt --index-strategy unsafe-best-match --constraint snapshot_constraint.txt --torch-backend cu128 2025-10-10T01:29:04.6931131Z Resolved 324 packages in 2.76s 2025-10-10T01:29:04.6932211Z # This file was autogenerated by uv via the following command: 2025-10-10T01:29:04.6933571Z # uv pip compile requirements/test.in -o test.txt --index-strategy unsafe-best-match --constraint snapshot_constraint.txt --torch-backend cu128 2025-10-10T01:29:04.7530255Z absl-py==2.1.0 2025-10-10T01:29:04.7531359Z  # via 2025-10-10T01:29:04.7532116Z # -c snapshot_constraint.txt 2025-10-10T01:29:04.7533377Z # rouge-score 2025-10-10T01:29:04.7533920Z accelerate==1.0.1 2025-10-10T01:29:04.7534537Z  # via 2025-10-10T01:29:04.7535029Z # -c snapshot_constraint.txt 2025-10-10T01:29:04.7535446Z # lm-eval 2025-10-10T01:29:04.7535726Z # peft 2025-10-10T01:29:04.7535978Z aenum==3.1.16 2025-10-10T01:29:04.7536270Z  # via 2025-10-10T01:29:04.7536567Z # -c snapshot_constraint.txt 2025-10-10T01:29:04.7536942Z # lightly 2025-10-10T01:29:04.7537205Z affine==2.4.0 2025-10-10T01:29:04.7537511Z  # via 2025-10-10T01:29:04.7537807Z # -c snapshot_constraint.txt 2025-10-10T01:29:04.7538104Z # rasterio 2025-10-10T01:29:04.7538411Z aiohappyeyeballs==2.4.3 2025-10-10T01:29:04.7538779Z  # via 2025-10-10T01:29:04.7539056Z # -c snapshot_constraint.txt 2025-10-10T01:29:04.7539308Z # aiohttp 2025-10-10T01:29:04.7539490Z aiohttp==3.10.11 2025-10-10T01:29:04.7539773Z  # via 2025-10-10T01:29:04.7540075Z # -c snapshot_constraint.txt 2025-10-10T01:29:04.7540444Z # aiohttp-cors 2025-10-10T01:29:04.7540687Z # datasets 2025-10-10T01:29:04.7540848Z # fsspec 2025-10-10T01:29:04.7541101Z # lm-eval 2025-10-10T01:29:04.7541368Z # ray 2025-10-10T01:29:04.7541611Z aiohttp-cors==0.8.1 2025-10-10T01:29:04.7541810Z  # via 2025-10-10T01:29:04.7542021Z # -c snapshot_constraint.txt 2025-10-10T01:29:04.7542325Z # ray 2025-10-10T01:29:04.7542616Z aiosignal==1.3.1 2025-10-10T01:29:04.7542871Z  # via 2025-10-10T01:29:04.7543043Z # -c snapshot_constraint.txt 2025-10-10T01:29:04.7543270Z # aiohttp 2025-10-10T01:29:04.7543446Z albucore==0.0.16 2025-10-10T01:29:04.7543894Z  # via 2025-10-10T01:29:04.7544056Z # -c snapshot_constraint.txt 2025-10-10T01:29:04.7544288Z # terratorch 2025-10-10T01:29:04.7544474Z albumentations==1.4.6 2025-10-10T01:29:04.7544673Z  # via 2025-10-10T01:29:04.7544831Z # -c snapshot_constraint.txt 2025-10-10T01:29:04.7545057Z # terratorch 2025-10-10T01:29:04.7545227Z alembic==1.16.4 2025-10-10T01:29:04.7545394Z  # via 2025-10-10T01:29:04.7545557Z # -c snapshot_constraint.txt 2025-10-10T01:29:04.7545964Z # mlflow 2025-10-10T01:29:04.7546272Z annotated-types==0.7.0 2025-10-10T01:29:04.7546638Z  # via 2025-10-10T01:29:04.7546941Z # -c snapshot_constraint.txt 2025-10-10T01:29:04.7547364Z # pydantic 2025-10-10T01:29:04.7547662Z antlr4-python3-runtime==4.9.3 2025-10-10T01:29:04.7547926Z  # via 2025-10-10T01:29:04.7548172Z # -c snapshot_constraint.txt 2025-10-10T01:29:04.7548666Z # hydra-core 2025-10-10T01:29:04.7548936Z # omegaconf 2025-10-10T01:29:04.7549199Z anyio==4.6.2.post1 2025-10-10T01:29:04.7549391Z  # via 2025-10-10T01:29:04.7549664Z # -c snapshot_constraint.txt 2025-10-10T01:29:04.7549957Z # httpx 2025-10-10T01:29:04.7550187Z # starlette 2025-10-10T01:29:04.7550456Z argcomplete==3.5.1 2025-10-10T01:29:04.7550790Z  # via 2025-10-10T01:29:04.7550984Z # -c snapshot_constraint.txt 2025-10-10T01:29:04.7551249Z # datamodel-code-generator 2025-10-10T01:29:04.7551468Z arrow==1.3.0 2025-10-10T01:29:04.7551646Z  # via 2025-10-10T01:29:04.7551813Z # -c snapshot_constraint.txt 2025-10-10T01:29:04.7552039Z # isoduration 2025-10-10T01:29:04.7552227Z attrs==24.2.0 2025-10-10T01:29:04.7552400Z  # via 2025-10-10T01:29:04.7552572Z # -c snapshot_constraint.txt 2025-10-10T01:29:04.7552756Z # aiohttp 2025-10-10T01:29:04.7552907Z # fiona 2025-10-10T01:29:04.7553067Z # hypothesis 2025-10-10T01:29:04.7553233Z # jsonlines 2025-10-10T01:29:04.7553474Z # jsonschema 2025-10-10T01:29:04.7553768Z # pytest-subtests 2025-10-10T01:29:04.7554043Z # rasterio 2025-10-10T01:29:04.7554335Z # referencing 2025-10-10T01:29:04.7554645Z audioread==3.0.1 2025-10-10T01:29:04.7554880Z  # via 2025-10-10T01:29:04.7555159Z # -c snapshot_constraint.txt 2025-10-10T01:29:04.7555416Z # librosa 2025-10-10T01:29:04.7555589Z backoff==2.2.1 2025-10-10T01:29:04.7555872Z  # via 2025-10-10T01:29:04.7556167Z # -c snapshot_constraint.txt 2025-10-10T01:29:04.7556505Z # -r requirements/test.in 2025-10-10T01:29:04.7556800Z # schemathesis 2025-10-10T01:29:04.7557078Z bitsandbytes==0.46.1 2025-10-10T01:29:04.7557275Z  # via 2025-10-10T01:29:04.7557453Z # -c snapshot_constraint.txt 2025-10-10T01:29:04.7557842Z # -r requirements/test.in 2025-10-10T01:29:04.7558177Z # lightning 2025-10-10T01:29:04.7558361Z black==24.10.0 2025-10-10T01:29:04.7558734Z  # via 2025-10-10T01:29:04.7559003Z # -c snapshot_constraint.txt 2025-10-10T01:29:04.7559476Z # datamodel-code-generator 2025-10-10T01:29:04.7559793Z blinker==1.9.0 2025-10-10T01:29:04.7560030Z  # via 2025-10-10T01:29:04.7560255Z # -c snapshot_constraint.txt 2025-10-10T01:29:04.7560574Z # flask 2025-10-10T01:29:04.7560755Z blobfile==3.0.0 2025-10-10T01:29:04.7561061Z  # via 2025-10-10T01:29:04.7561369Z # -c snapshot_constraint.txt 2025-10-10T01:29:04.7561788Z # -r requirements/test.in 2025-10-10T01:29:04.7562145Z bm25s==0.2.13 2025-10-10T01:29:04.7562371Z  # via 2025-10-10T01:29:04.7562551Z # -c snapshot_constraint.txt 2025-10-10T01:29:04.7562775Z # mteb 2025-10-10T01:29:04.7562927Z boto3==1.35.57 2025-10-10T01:29:04.7563098Z  # via 2025-10-10T01:29:04.7563262Z # -c snapshot_constraint.txt 2025-10-10T01:29:04.7563479Z # runai-model-streamer-s3 2025-10-10T01:29:04.7563893Z # tensorizer 2025-10-10T01:29:04.7564201Z botocore==1.35.57 2025-10-10T01:29:04.7564535Z  # via 2025-10-10T01:29:04.7564829Z # -c snapshot_constraint.txt 2025-10-10T01:29:04.7565152Z # boto3 2025-10-10T01:29:04.7565398Z # s3transfer 2025-10-10T01:29:04.7565664Z bounded-pool-executor==0.0.3 2025-10-10T01:29:04.7566057Z  # via 2025-10-10T01:29:04.7566287Z # -c snapshot_constraint.txt 2025-10-10T01:29:04.7566628Z # pqdm 2025-10-10T01:29:04.7566885Z buildkite-test-collector==0.1.9 2025-10-10T01:29:04.7567107Z  # via 2025-10-10T01:29:04.7567389Z # -c snapshot_constraint.txt 2025-10-10T01:29:04.7567723Z # -r requirements/test.in 2025-10-10T01:29:04.7568017Z cachetools==5.5.2 2025-10-10T01:29:04.7568204Z  # via 2025-10-10T01:29:04.7568406Z # -c snapshot_constraint.txt 2025-10-10T01:29:04.7568770Z # google-auth 2025-10-10T01:29:04.7569015Z # mlflow-skinny 2025-10-10T01:29:04.7569354Z certifi==2024.8.30 2025-10-10T01:29:04.7569670Z  # via 2025-10-10T01:29:04.7569914Z # -c snapshot_constraint.txt 2025-10-10T01:29:04.7570226Z # fiona 2025-10-10T01:29:04.7570377Z # httpcore 2025-10-10T01:29:04.7570622Z # httpx 2025-10-10T01:29:04.7570857Z # lightly 2025-10-10T01:29:04.7571128Z # pyogrio 2025-10-10T01:29:04.7571371Z # pyproj 2025-10-10T01:29:04.7571523Z # rasterio 2025-10-10T01:29:04.7571827Z # requests 2025-10-10T01:29:04.7572097Z cffi==1.17.1 2025-10-10T01:29:04.7572384Z  # via 2025-10-10T01:29:04.7572575Z # -c snapshot_constraint.txt 2025-10-10T01:29:04.7572930Z # soundfile 2025-10-10T01:29:04.7573231Z chardet==5.2.0 2025-10-10T01:29:04.7573552Z  # via 2025-10-10T01:29:04.7573849Z # -c snapshot_constraint.txt 2025-10-10T01:29:04.7574288Z # mbstrdecoder 2025-10-10T01:29:04.7574627Z charset-normalizer==3.4.0 2025-10-10T01:29:04.7575003Z  # via 2025-10-10T01:29:04.7575265Z # -c snapshot_constraint.txt 2025-10-10T01:29:04.7575593Z # requests 2025-10-10T01:29:04.7575879Z click==8.1.7 2025-10-10T01:29:04.7576174Z  # via 2025-10-10T01:29:04.7576451Z # -c snapshot_constraint.txt 2025-10-10T01:29:04.7576725Z # black 2025-10-10T01:29:04.7576878Z # click-plugins 2025-10-10T01:29:04.7577048Z # cligj 2025-10-10T01:29:04.7577193Z # fiona 2025-10-10T01:29:04.7577336Z # flask 2025-10-10T01:29:04.7577472Z # jiwer 2025-10-10T01:29:04.7577618Z # mlflow-skinny 2025-10-10T01:29:04.7577789Z # nltk 2025-10-10T01:29:04.7578038Z # rasterio 2025-10-10T01:29:04.7578270Z # ray 2025-10-10T01:29:04.7578522Z # schemathesis 2025-10-10T01:29:04.7578744Z # typer 2025-10-10T01:29:04.7579004Z # uvicorn 2025-10-10T01:29:04.7579264Z click-plugins==1.1.1.2 2025-10-10T01:29:04.7579514Z  # via 2025-10-10T01:29:04.7579696Z # -c snapshot_constraint.txt 2025-10-10T01:29:04.7579993Z # fiona 2025-10-10T01:29:04.7580337Z # rasterio 2025-10-10T01:29:04.7580549Z cligj==0.7.2 2025-10-10T01:29:04.7580728Z  # via 2025-10-10T01:29:04.7580887Z # -c snapshot_constraint.txt 2025-10-10T01:29:04.7581087Z # fiona 2025-10-10T01:29:04.7581260Z # rasterio 2025-10-10T01:29:04.7581427Z cloudpickle==3.1.1 2025-10-10T01:29:04.7581730Z  # via 2025-10-10T01:29:04.7581997Z # -c snapshot_constraint.txt 2025-10-10T01:29:04.7582388Z # mlflow-skinny 2025-10-10T01:29:04.7582639Z colorama==0.4.6 2025-10-10T01:29:04.7582937Z  # via 2025-10-10T01:29:04.7583179Z # -c snapshot_constraint.txt 2025-10-10T01:29:04.7583498Z # sacrebleu 2025-10-10T01:29:04.7583751Z # schemathesis 2025-10-10T01:29:04.7584022Z # tqdm-multiprocess 2025-10-10T01:29:04.7584370Z colorful==0.5.6 2025-10-10T01:29:04.7584690Z  # via 2025-10-10T01:29:04.7584858Z # -c snapshot_constraint.txt 2025-10-10T01:29:04.7585075Z # ray 2025-10-10T01:29:04.7586660Z contourpy==1.3.0 2025-10-10T01:29:04.7586855Z  # via 2025-10-10T01:29:04.7587022Z # -c snapshot_constraint.txt 2025-10-10T01:29:04.7587245Z # matplotlib 2025-10-10T01:29:04.7587421Z coverage==7.10.6 2025-10-10T01:29:04.7587597Z  # via 2025-10-10T01:29:04.7587758Z # -c snapshot_constraint.txt 2025-10-10T01:29:04.7587979Z # pytest-cov 2025-10-10T01:29:04.7588152Z cramjam==2.9.0 2025-10-10T01:29:04.7588321Z  # via 2025-10-10T01:29:04.7588473Z # -c snapshot_constraint.txt 2025-10-10T01:29:04.7588702Z # fastparquet 2025-10-10T01:29:04.7588882Z cupy-cuda12x==13.6.0 2025-10-10T01:29:04.7589065Z  # via 2025-10-10T01:29:04.7589217Z # -c snapshot_constraint.txt 2025-10-10T01:29:04.7589429Z # ray 2025-10-10T01:29:04.7589582Z cycler==0.12.1 2025-10-10T01:29:04.7589751Z  # via 2025-10-10T01:29:04.7589901Z # -c snapshot_constraint.txt 2025-10-10T01:29:04.7590209Z # matplotlib 2025-10-10T01:29:04.7590392Z databricks-sdk==0.59.0 2025-10-10T01:29:04.7590582Z  # via 2025-10-10T01:29:04.7590734Z # -c snapshot_constraint.txt 2025-10-10T01:29:04.7590964Z # mlflow-skinny 2025-10-10T01:29:04.7591165Z datamodel-code-generator==0.26.3 2025-10-10T01:29:04.7591388Z  # via 2025-10-10T01:29:04.7591547Z # -c snapshot_constraint.txt 2025-10-10T01:29:04.7591793Z # -r requirements/test.in 2025-10-10T01:29:04.7592005Z dataproperty==1.0.1 2025-10-10T01:29:04.7592191Z  # via 2025-10-10T01:29:04.7592344Z # -c snapshot_constraint.txt 2025-10-10T01:29:04.7592553Z # pytablewriter 2025-10-10T01:29:04.7592753Z # tabledata 2025-10-10T01:29:04.7592925Z datasets==3.0.2 2025-10-10T01:29:04.7593095Z  # via 2025-10-10T01:29:04.7593252Z # -c snapshot_constraint.txt 2025-10-10T01:29:04.7593441Z # evaluate 2025-10-10T01:29:04.7593588Z # lm-eval 2025-10-10T01:29:04.7593757Z # mteb 2025-10-10T01:29:04.7593920Z decorator==5.1.1 2025-10-10T01:29:04.7594092Z  # via 2025-10-10T01:29:04.7594241Z # -c snapshot_constraint.txt 2025-10-10T01:29:04.7594458Z # librosa 2025-10-10T01:29:04.7594620Z decord==0.6.0 2025-10-10T01:29:04.7594787Z  # via 2025-10-10T01:29:04.7594940Z # -c snapshot_constraint.txt 2025-10-10T01:29:04.7595177Z # -r requirements/test.in 2025-10-10T01:29:04.7595373Z dill==0.3.8 2025-10-10T01:29:04.7595538Z  # via 2025-10-10T01:29:04.7595690Z # -c snapshot_constraint.txt 2025-10-10T01:29:04.7595881Z # datasets 2025-10-10T01:29:04.7596028Z # evaluate 2025-10-10T01:29:04.7596174Z # lm-eval 2025-10-10T01:29:04.7596352Z # multiprocess 2025-10-10T01:29:04.7596530Z distlib==0.3.9 2025-10-10T01:29:04.7596699Z  # via 2025-10-10T01:29:04.7596871Z # -c snapshot_constraint.txt 2025-10-10T01:29:04.7597092Z # virtualenv 2025-10-10T01:29:04.7597261Z dnspython==2.7.0 2025-10-10T01:29:04.7597436Z  # via 2025-10-10T01:29:04.7597597Z # -c snapshot_constraint.txt 2025-10-10T01:29:04.7597825Z # email-validator 2025-10-10T01:29:04.7598010Z docker==7.1.0 2025-10-10T01:29:04.7598180Z  # via 2025-10-10T01:29:04.7598338Z # -c snapshot_constraint.txt 2025-10-10T01:29:04.7598649Z # mlflow 2025-10-10T01:29:04.7598812Z docopt==0.6.2 2025-10-10T01:29:04.7598982Z  # via 2025-10-10T01:29:04.7599133Z # -c snapshot_constraint.txt 2025-10-10T01:29:04.7599363Z # num2words 2025-10-10T01:29:04.7599541Z docstring-parser==0.17.0 2025-10-10T01:29:04.7599730Z  # via 2025-10-10T01:29:04.7599901Z # -c snapshot_constraint.txt 2025-10-10T01:29:04.7600126Z # jsonargparse 2025-10-10T01:29:04.7600333Z efficientnet-pytorch==0.7.1 2025-10-10T01:29:04.7600600Z  # via segmentation-models-pytorch 2025-10-10T01:29:04.7600837Z einops==0.8.1 2025-10-10T01:29:04.7601005Z  # via 2025-10-10T01:29:04.7601174Z # -c snapshot_constraint.txt 2025-10-10T01:29:04.7601481Z # -r requirements/test.in 2025-10-10T01:29:04.7601670Z # encodec 2025-10-10T01:29:04.7601816Z # terratorch 2025-10-10T01:29:04.7601977Z # torchgeo 2025-10-10T01:29:04.7602144Z # vector-quantize-pytorch 2025-10-10T01:29:04.7602365Z # vocos 2025-10-10T01:29:04.7602514Z einx==0.3.0 2025-10-10T01:29:04.7602677Z  # via 2025-10-10T01:29:04.7602838Z # -c snapshot_constraint.txt 2025-10-10T01:29:04.7603089Z # vector-quantize-pytorch 2025-10-10T01:29:04.7603296Z email-validator==2.2.0 2025-10-10T01:29:04.7603481Z  # via 2025-10-10T01:29:04.7603639Z # -c snapshot_constraint.txt 2025-10-10T01:29:04.7603858Z # pydantic 2025-10-10T01:29:04.7604016Z encodec==0.1.1 2025-10-10T01:29:04.7604182Z  # via 2025-10-10T01:29:04.7604353Z # -c snapshot_constraint.txt 2025-10-10T01:29:04.7604571Z # vocos 2025-10-10T01:29:04.7604830Z eval-type-backport==0.2.2 2025-10-10T01:29:04.7605039Z  # via 2025-10-10T01:29:04.7605201Z # -c snapshot_constraint.txt 2025-10-10T01:29:04.7605414Z # mteb 2025-10-10T01:29:04.7605563Z evaluate==0.4.3 2025-10-10T01:29:04.7605738Z  # via 2025-10-10T01:29:04.7605898Z # -c snapshot_constraint.txt 2025-10-10T01:29:04.7606120Z # lm-eval 2025-10-10T01:29:04.7606278Z fastapi==0.116.1 2025-10-10T01:29:04.7606448Z  # via 2025-10-10T01:29:04.7606776Z # -c snapshot_constraint.txt 2025-10-10T01:29:04.7607004Z # mlflow-skinny 2025-10-10T01:29:04.7607183Z fastparquet==2024.11.0 2025-10-10T01:29:04.7607367Z  # via 2025-10-10T01:29:04.7607525Z # -c snapshot_constraint.txt 2025-10-10T01:29:04.7607749Z # genai-perf 2025-10-10T01:29:04.7607919Z fastrlock==0.8.2 2025-10-10T01:29:04.7608096Z  # via 2025-10-10T01:29:04.7608253Z # -c snapshot_constraint.txt 2025-10-10T01:29:04.7608477Z # cupy-cuda12x 2025-10-10T01:29:04.7608678Z fastsafetensors==0.1.10 2025-10-10T01:29:04.7608874Z  # via 2025-10-10T01:29:04.7609035Z # -c snapshot_constraint.txt 2025-10-10T01:29:04.7609267Z # -r requirements/test.in 2025-10-10T01:29:04.7609472Z filelock==3.16.1 2025-10-10T01:29:04.7609647Z  # via 2025-10-10T01:29:04.7609810Z # -c snapshot_constraint.txt 2025-10-10T01:29:04.7609997Z # blobfile 2025-10-10T01:29:04.7610146Z # datasets 2025-10-10T01:29:04.7610301Z # huggingface-hub 2025-10-10T01:29:04.7610464Z # ray 2025-10-10T01:29:04.7610610Z # torch 2025-10-10T01:29:04.7610762Z # transformers 2025-10-10T01:29:04.7610959Z # virtualenv 2025-10-10T01:29:04.7611122Z fiona==1.10.1 2025-10-10T01:29:04.7611280Z  # via 2025-10-10T01:29:04.7611435Z # -c snapshot_constraint.txt 2025-10-10T01:29:04.7611653Z # torchgeo 2025-10-10T01:29:04.7611814Z flask==3.1.1 2025-10-10T01:29:04.7611975Z  # via 2025-10-10T01:29:04.7612133Z # -c snapshot_constraint.txt 2025-10-10T01:29:04.7612356Z # mlflow 2025-10-10T01:29:04.7612516Z fonttools==4.55.0 2025-10-10T01:29:04.7612684Z  # via 2025-10-10T01:29:04.7612843Z # -c snapshot_constraint.txt 2025-10-10T01:29:04.7613061Z # matplotlib 2025-10-10T01:29:04.7613215Z fqdn==1.5.1 2025-10-10T01:29:04.7613388Z  # via 2025-10-10T01:29:04.7613545Z # -c snapshot_constraint.txt 2025-10-10T01:29:04.7613766Z # jsonschema 2025-10-10T01:29:04.7613924Z frozendict==2.4.6 2025-10-10T01:29:04.7614098Z  # via 2025-10-10T01:29:04.7614252Z # -c snapshot_constraint.txt 2025-10-10T01:29:04.7614464Z # einx 2025-10-10T01:29:04.7614613Z frozenlist==1.5.0 2025-10-10T01:29:04.7614786Z  # via 2025-10-10T01:29:04.7614942Z # -c snapshot_constraint.txt 2025-10-10T01:29:04.7615129Z # aiohttp 2025-10-10T01:29:04.7615302Z # aiosignal 2025-10-10T01:29:04.7615466Z fsspec==2024.9.0 2025-10-10T01:29:04.7615638Z  # via 2025-10-10T01:29:04.7615798Z # -c snapshot_constraint.txt 2025-10-10T01:29:04.7616065Z # datasets 2025-10-10T01:29:04.7616211Z # evaluate 2025-10-10T01:29:04.7616372Z # fastparquet 2025-10-10T01:29:04.7616541Z # huggingface-hub 2025-10-10T01:29:04.7616702Z # lightning 2025-10-10T01:29:04.7616862Z # pytorch-lightning 2025-10-10T01:29:04.7617073Z # torch 2025-10-10T01:29:04.7617231Z ftfy==6.3.1 2025-10-10T01:29:04.7617390Z  # via 2025-10-10T01:29:04.7617547Z # -c snapshot_constraint.txt 2025-10-10T01:29:04.7617790Z # open-clip-torch 2025-10-10T01:29:04.7617987Z genai-perf==0.0.8 2025-10-10T01:29:04.7618165Z  # via 2025-10-10T01:29:04.7618322Z # -c snapshot_constraint.txt 2025-10-10T01:29:04.7618572Z # -r requirements/test.in 2025-10-10T01:29:04.7618765Z genson==1.3.0 2025-10-10T01:29:04.7618933Z  # via 2025-10-10T01:29:04.7619091Z # -c snapshot_constraint.txt 2025-10-10T01:29:04.7619423Z # datamodel-code-generator 2025-10-10T01:29:04.7619637Z geopandas==1.0.1 2025-10-10T01:29:04.7619814Z  # via 2025-10-10T01:29:04.7619972Z # -c snapshot_constraint.txt 2025-10-10T01:29:04.7620195Z # terratorch 2025-10-10T01:29:04.7620353Z gitdb==4.0.12 2025-10-10T01:29:04.7620517Z  # via 2025-10-10T01:29:04.7620676Z # -c snapshot_constraint.txt 2025-10-10T01:29:04.7620894Z # gitpython 2025-10-10T01:29:04.7621052Z gitpython==3.1.44 2025-10-10T01:29:04.7621225Z  # via 2025-10-10T01:29:04.7621380Z # -c snapshot_constraint.txt 2025-10-10T01:29:04.7621604Z # mlflow-skinny 2025-10-10T01:29:04.7621784Z google-api-core==2.24.2 2025-10-10T01:29:04.7621971Z  # via 2025-10-10T01:29:04.7622124Z # -c snapshot_constraint.txt 2025-10-10T01:29:04.7622319Z # google-cloud-core 2025-10-10T01:29:04.7622498Z # google-cloud-storage 2025-10-10T01:29:04.7622711Z # opencensus 2025-10-10T01:29:04.7622879Z google-auth==2.40.2 2025-10-10T01:29:04.7623072Z  # via 2025-10-10T01:29:04.7623230Z # -c snapshot_constraint.txt 2025-10-10T01:29:04.7623424Z # databricks-sdk 2025-10-10T01:29:04.7623596Z # google-api-core 2025-10-10T01:29:04.7623769Z # google-cloud-core 2025-10-10T01:29:04.7624127Z # google-cloud-storage 2025-10-10T01:29:04.7624390Z # runai-model-streamer-gcs 2025-10-10T01:29:04.7624623Z google-cloud-core==2.4.3 2025-10-10T01:29:04.7624827Z  # via 2025-10-10T01:29:04.7624981Z # -c snapshot_constraint.txt 2025-10-10T01:29:04.7625225Z # google-cloud-storage 2025-10-10T01:29:04.7625439Z google-cloud-storage==3.4.0 2025-10-10T01:29:04.7625640Z  # via 2025-10-10T01:29:04.7625804Z # -c snapshot_constraint.txt 2025-10-10T01:29:04.7626052Z # runai-model-streamer-gcs 2025-10-10T01:29:04.7626268Z google-crc32c==1.7.1 2025-10-10T01:29:04.7626447Z  # via 2025-10-10T01:29:04.7626610Z # -c snapshot_constraint.txt 2025-10-10T01:29:04.7626822Z # google-cloud-storage 2025-10-10T01:29:04.7627061Z # google-resumable-media 2025-10-10T01:29:04.7627279Z google-resumable-media==2.7.2 2025-10-10T01:29:04.7627491Z  # via 2025-10-10T01:29:04.7627652Z # -c snapshot_constraint.txt 2025-10-10T01:29:04.7627890Z # google-cloud-storage 2025-10-10T01:29:04.7628103Z googleapis-common-protos==1.70.0 2025-10-10T01:29:04.7628335Z  # via 2025-10-10T01:29:04.7628501Z # -c snapshot_constraint.txt 2025-10-10T01:29:04.7628732Z # google-api-core 2025-10-10T01:29:04.7628908Z graphene==3.4.3 2025-10-10T01:29:04.7629081Z  # via 2025-10-10T01:29:04.7629237Z # -c snapshot_constraint.txt 2025-10-10T01:29:04.7629453Z # mlflow 2025-10-10T01:29:04.7629610Z graphql-core==3.2.6 2025-10-10T01:29:04.7629787Z  # via 2025-10-10T01:29:04.7629941Z # -c snapshot_constraint.txt 2025-10-10T01:29:04.7630129Z # graphene 2025-10-10T01:29:04.7630279Z # graphql-relay 2025-10-10T01:29:04.7630490Z # hypothesis-graphql 2025-10-10T01:29:04.7630913Z graphql-relay==3.2.0 2025-10-10T01:29:04.7631100Z  # via 2025-10-10T01:29:04.7631256Z # -c snapshot_constraint.txt 2025-10-10T01:29:04.7631476Z # graphene 2025-10-10T01:29:04.7631649Z greenlet==3.2.3 2025-10-10T01:29:04.7631825Z  # via 2025-10-10T01:29:04.7631978Z # -c snapshot_constraint.txt 2025-10-10T01:29:04.7632200Z # sqlalchemy 2025-10-10T01:29:04.7632373Z grpcio==1.71.0 2025-10-10T01:29:04.7632540Z  # via 2025-10-10T01:29:04.7632690Z # -c snapshot_constraint.txt 2025-10-10T01:29:04.7632903Z # ray 2025-10-10T01:29:04.7633060Z gunicorn==23.0.0 2025-10-10T01:29:04.7633226Z  # via 2025-10-10T01:29:04.7633406Z # -c snapshot_constraint.txt 2025-10-10T01:29:04.7633623Z # mlflow 2025-10-10T01:29:04.7633777Z h11==0.14.0 2025-10-10T01:29:04.7633937Z  # via 2025-10-10T01:29:04.7634097Z # -c snapshot_constraint.txt 2025-10-10T01:29:04.7634425Z # httpcore 2025-10-10T01:29:04.7634616Z # uvicorn 2025-10-10T01:29:04.7634771Z h5py==3.13.0 2025-10-10T01:29:04.7634939Z  # via 2025-10-10T01:29:04.7635107Z # -c snapshot_constraint.txt 2025-10-10T01:29:04.7635333Z # terratorch 2025-10-10T01:29:04.7635499Z harfile==0.3.0 2025-10-10T01:29:04.7635681Z  # via 2025-10-10T01:29:04.7635846Z # -c snapshot_constraint.txt 2025-10-10T01:29:04.7636075Z # schemathesis 2025-10-10T01:29:04.7636249Z hf-xet==1.1.7 2025-10-10T01:29:04.7636424Z  # via 2025-10-10T01:29:04.7636594Z # -c snapshot_constraint.txt 2025-10-10T01:29:04.7636837Z # huggingface-hub 2025-10-10T01:29:04.7637017Z hiredis==3.0.0 2025-10-10T01:29:04.7637192Z  # via 2025-10-10T01:29:04.7637350Z # -c snapshot_constraint.txt 2025-10-10T01:29:04.7637573Z # tensorizer 2025-10-10T01:29:04.7637738Z httpcore==1.0.6 2025-10-10T01:29:04.7637911Z  # via 2025-10-10T01:29:04.7638075Z # -c snapshot_constraint.txt 2025-10-10T01:29:04.7638293Z # httpx 2025-10-10T01:29:04.7638451Z httpx==0.27.2 2025-10-10T01:29:04.7638693Z  # via 2025-10-10T01:29:04.7638863Z # -c snapshot_constraint.txt 2025-10-10T01:29:04.7639066Z # -r requirements/test.in 2025-10-10T01:29:04.7639290Z # schemathesis 2025-10-10T01:29:04.7639478Z huggingface-hub==0.34.3 2025-10-10T01:29:04.7639671Z  # via 2025-10-10T01:29:04.7639825Z # -c snapshot_constraint.txt 2025-10-10T01:29:04.7640017Z # accelerate 2025-10-10T01:29:04.7640169Z # datasets 2025-10-10T01:29:04.7640319Z # evaluate 2025-10-10T01:29:04.7640473Z # open-clip-torch 2025-10-10T01:29:04.7640639Z # peft 2025-10-10T01:29:04.7640810Z # segmentation-models-pytorch 2025-10-10T01:29:04.7641028Z # sentence-transformers 2025-10-10T01:29:04.7641208Z # terratorch 2025-10-10T01:29:04.7641360Z # timm 2025-10-10T01:29:04.7641504Z # tokenizers 2025-10-10T01:29:04.7641660Z # transformers 2025-10-10T01:29:04.7641845Z # vocos 2025-10-10T01:29:04.7641999Z humanize==4.11.0 2025-10-10T01:29:04.7642175Z  # via 2025-10-10T01:29:04.7642337Z # -c snapshot_constraint.txt 2025-10-10T01:29:04.7642572Z # runai-model-streamer 2025-10-10T01:29:04.7642774Z hydra-core==1.3.2 2025-10-10T01:29:04.7642954Z  # via 2025-10-10T01:29:04.7643113Z # -c snapshot_constraint.txt 2025-10-10T01:29:04.7643299Z # lightly 2025-10-10T01:29:04.7643484Z # lightning 2025-10-10T01:29:04.7643658Z hypothesis==6.131.0 2025-10-10T01:29:04.7643837Z  # via 2025-10-10T01:29:04.7643998Z # -c snapshot_constraint.txt 2025-10-10T01:29:04.7644197Z # hypothesis-graphql 2025-10-10T01:29:04.7644399Z # hypothesis-jsonschema 2025-10-10T01:29:04.7644621Z # schemathesis 2025-10-10T01:29:04.7644808Z hypothesis-graphql==0.11.1 2025-10-10T01:29:04.7645008Z  # via 2025-10-10T01:29:04.7645170Z # -c snapshot_constraint.txt 2025-10-10T01:29:04.7645481Z # schemathesis 2025-10-10T01:29:04.7645674Z hypothesis-jsonschema==0.23.1 2025-10-10T01:29:04.7645888Z  # via 2025-10-10T01:29:04.7646044Z # -c snapshot_constraint.txt 2025-10-10T01:29:04.7646260Z # schemathesis 2025-10-10T01:29:04.7646429Z idna==3.10 2025-10-10T01:29:04.7646607Z  # via 2025-10-10T01:29:04.7646764Z # -c snapshot_constraint.txt 2025-10-10T01:29:04.7646944Z # anyio 2025-10-10T01:29:04.7647097Z # email-validator 2025-10-10T01:29:04.7647256Z # httpx 2025-10-10T01:29:04.7647399Z # jsonschema 2025-10-10T01:29:04.7647544Z # requests 2025-10-10T01:29:04.7647717Z # yarl 2025-10-10T01:29:04.7647870Z imageio==2.37.0 2025-10-10T01:29:04.7648052Z  # via 2025-10-10T01:29:04.7648204Z # -c snapshot_constraint.txt 2025-10-10T01:29:04.7648433Z # scikit-image 2025-10-10T01:29:04.7648618Z importlib-metadata==8.7.0 2025-10-10T01:29:04.7648895Z  # via 2025-10-10T01:29:04.7649067Z # -c snapshot_constraint.txt 2025-10-10T01:29:04.7649265Z # mlflow-skinny 2025-10-10T01:29:04.7649476Z # opentelemetry-api 2025-10-10T01:29:04.7649677Z importlib-resources==6.5.2 2025-10-10T01:29:04.7649871Z  # via 2025-10-10T01:29:04.7650029Z # -c snapshot_constraint.txt 2025-10-10T01:29:04.7650260Z # typeshed-client 2025-10-10T01:29:04.7650448Z inflect==5.6.2 2025-10-10T01:29:04.7650622Z  # via 2025-10-10T01:29:04.7650780Z # -c snapshot_constraint.txt 2025-10-10T01:29:04.7651028Z # datamodel-code-generator 2025-10-10T01:29:04.7651235Z iniconfig==2.0.0 2025-10-10T01:29:04.7651413Z  # via 2025-10-10T01:29:04.7651571Z # -c snapshot_constraint.txt 2025-10-10T01:29:04.7651790Z # pytest 2025-10-10T01:29:04.7651949Z isoduration==20.11.0 2025-10-10T01:29:04.7652135Z  # via 2025-10-10T01:29:04.7652291Z # -c snapshot_constraint.txt 2025-10-10T01:29:04.7652516Z # jsonschema 2025-10-10T01:29:04.7652679Z isort==5.13.2 2025-10-10T01:29:04.7652853Z  # via 2025-10-10T01:29:04.7653012Z # -c snapshot_constraint.txt 2025-10-10T01:29:04.7653262Z # datamodel-code-generator 2025-10-10T01:29:04.7653468Z itsdangerous==2.2.0 2025-10-10T01:29:04.7653653Z  # via 2025-10-10T01:29:04.7653813Z # -c snapshot_constraint.txt 2025-10-10T01:29:04.7654029Z # flask 2025-10-10T01:29:04.7654175Z jinja2==3.1.6 2025-10-10T01:29:04.7654344Z  # via 2025-10-10T01:29:04.7654503Z # -c snapshot_constraint.txt 2025-10-10T01:29:04.7654711Z # datamodel-code-generator 2025-10-10T01:29:04.7654892Z # flask 2025-10-10T01:29:04.7655030Z # mlflow 2025-10-10T01:29:04.7655212Z # torch 2025-10-10T01:29:04.7655368Z jiwer==3.0.5 2025-10-10T01:29:04.7655527Z  # via 2025-10-10T01:29:04.7655682Z # -c snapshot_constraint.txt 2025-10-10T01:29:04.7655924Z # -r requirements/test.in 2025-10-10T01:29:04.7656125Z jmespath==1.0.1 2025-10-10T01:29:04.7656312Z  # via 2025-10-10T01:29:04.7656470Z # -c snapshot_constraint.txt 2025-10-10T01:29:04.7656657Z # boto3 2025-10-10T01:29:04.7656824Z # botocore 2025-10-10T01:29:04.7656989Z joblib==1.4.2 2025-10-10T01:29:04.7657153Z  # via 2025-10-10T01:29:04.7657309Z # -c snapshot_constraint.txt 2025-10-10T01:29:04.7657491Z # librosa 2025-10-10T01:29:04.7657636Z # nltk 2025-10-10T01:29:04.7657826Z # scikit-learn 2025-10-10T01:29:04.7658007Z jsonargparse==4.35.0 2025-10-10T01:29:04.7658183Z  # via 2025-10-10T01:29:04.7658342Z # -c snapshot_constraint.txt 2025-10-10T01:29:04.7658532Z # lightning 2025-10-10T01:29:04.7658719Z # terratorch 2025-10-10T01:29:04.7658881Z jsonlines==4.0.0 2025-10-10T01:29:04.7659058Z  # via 2025-10-10T01:29:04.7659224Z # -c snapshot_constraint.txt 2025-10-10T01:29:04.7659446Z # lm-eval 2025-10-10T01:29:04.7659614Z jsonpointer==3.0.0 2025-10-10T01:29:04.7659885Z  # via 2025-10-10T01:29:04.7660044Z # -c snapshot_constraint.txt 2025-10-10T01:29:04.7660267Z # jsonschema 2025-10-10T01:29:04.7660427Z jsonschema==4.23.0 2025-10-10T01:29:04.7660602Z  # via 2025-10-10T01:29:04.7660764Z # -c snapshot_constraint.txt 2025-10-10T01:29:04.7660970Z # hypothesis-jsonschema 2025-10-10T01:29:04.7661156Z # mistral-common 2025-10-10T01:29:04.7661321Z # ray 2025-10-10T01:29:04.7661499Z # schemathesis 2025-10-10T01:29:04.7661702Z jsonschema-specifications==2024.10.1 2025-10-10T01:29:04.7661933Z  # via 2025-10-10T01:29:04.7662094Z # -c snapshot_constraint.txt 2025-10-10T01:29:04.7662320Z # jsonschema 2025-10-10T01:29:04.7662482Z junit-xml==1.9 2025-10-10T01:29:04.7662652Z  # via 2025-10-10T01:29:04.7662807Z # -c snapshot_constraint.txt 2025-10-10T01:29:04.7663028Z # schemathesis 2025-10-10T01:29:04.7663191Z kaleido==0.2.1 2025-10-10T01:29:04.7663434Z  # via 2025-10-10T01:29:04.7663610Z # -c snapshot_constraint.txt 2025-10-10T01:29:04.7663853Z # genai-perf 2025-10-10T01:29:04.7664020Z kiwisolver==1.4.7 2025-10-10T01:29:04.7664196Z  # via 2025-10-10T01:29:04.7664354Z # -c snapshot_constraint.txt 2025-10-10T01:29:04.7664577Z # matplotlib 2025-10-10T01:29:04.7664734Z kornia==0.8.1 2025-10-10T01:29:04.7664900Z  # via 2025-10-10T01:29:04.7665057Z # -c snapshot_constraint.txt 2025-10-10T01:29:04.7665274Z # torchgeo 2025-10-10T01:29:04.7665432Z kornia-rs==0.1.9 2025-10-10T01:29:04.7665604Z  # via 2025-10-10T01:29:04.7665761Z # -c snapshot_constraint.txt 2025-10-10T01:29:04.7665973Z # kornia 2025-10-10T01:29:04.7666127Z lazy-loader==0.4 2025-10-10T01:29:04.7666311Z  # via 2025-10-10T01:29:04.7666468Z # -c snapshot_constraint.txt 2025-10-10T01:29:04.7666656Z # librosa 2025-10-10T01:29:04.7666833Z # scikit-image 2025-10-10T01:29:04.7667010Z libnacl==2.1.0 2025-10-10T01:29:04.7667179Z  # via 2025-10-10T01:29:04.7667328Z # -c snapshot_constraint.txt 2025-10-10T01:29:04.7667548Z # tensorizer 2025-10-10T01:29:04.7667717Z librosa==0.10.2.post1 2025-10-10T01:29:04.7667897Z  # via 2025-10-10T01:29:04.7668052Z # -c snapshot_constraint.txt 2025-10-10T01:29:04.7668293Z # -r requirements/test.in 2025-10-10T01:29:04.7668493Z lightly==1.5.20 2025-10-10T01:29:04.7668665Z  # via 2025-10-10T01:29:04.7668816Z # -c snapshot_constraint.txt 2025-10-10T01:29:04.7669007Z # terratorch 2025-10-10T01:29:04.7669199Z # torchgeo 2025-10-10T01:29:04.7669366Z lightly-utils==0.0.2 2025-10-10T01:29:04.7669544Z  # via 2025-10-10T01:29:04.7669705Z # -c snapshot_constraint.txt 2025-10-10T01:29:04.7669925Z # lightly 2025-10-10T01:29:04.7670093Z lightning==2.5.1.post0 2025-10-10T01:29:04.7670274Z  # via 2025-10-10T01:29:04.7670442Z # -c snapshot_constraint.txt 2025-10-10T01:29:04.7670642Z # terratorch 2025-10-10T01:29:04.7670825Z # torchgeo 2025-10-10T01:29:04.7671016Z lightning-utilities==0.14.3 2025-10-10T01:29:04.7671229Z  # via 2025-10-10T01:29:04.7671389Z # -c snapshot_constraint.txt 2025-10-10T01:29:04.7671582Z # lightning 2025-10-10T01:29:04.7671744Z # pytorch-lightning 2025-10-10T01:29:04.7671959Z # torchmetrics 2025-10-10T01:29:04.7672133Z llvmlite==0.44.0 2025-10-10T01:29:04.7672314Z  # via 2025-10-10T01:29:04.7672473Z # -c snapshot_constraint.txt 2025-10-10T01:29:04.7672688Z # numba 2025-10-10T01:29:04.7686726Z lm-eval @ git+https://github.com/EleutherAI/lm-evaluation-harness.git@206b7722158f58c35b7ffcd53b035fdbdda5126d 2025-10-10T01:29:04.7687316Z  # via 2025-10-10T01:29:04.7687505Z # -c snapshot_constraint.txt 2025-10-10T01:29:04.7687780Z # -r requirements/test.in 2025-10-10T01:29:04.7687992Z lxml==5.3.0 2025-10-10T01:29:04.7688170Z  # via 2025-10-10T01:29:04.7688488Z # -c snapshot_constraint.txt 2025-10-10T01:29:04.7688700Z # blobfile 2025-10-10T01:29:04.7688898Z # sacrebleu 2025-10-10T01:29:04.7689067Z mako==1.3.10 2025-10-10T01:29:04.7689242Z  # via 2025-10-10T01:29:04.7689413Z # -c snapshot_constraint.txt 2025-10-10T01:29:04.7689640Z # alembic 2025-10-10T01:29:04.7689805Z markdown==3.8.2 2025-10-10T01:29:04.7689981Z  # via 2025-10-10T01:29:04.7690154Z # -c snapshot_constraint.txt 2025-10-10T01:29:04.7690383Z # mlflow 2025-10-10T01:29:04.7690553Z markdown-it-py==3.0.0 2025-10-10T01:29:04.7690750Z  # via 2025-10-10T01:29:04.7690922Z # -c snapshot_constraint.txt 2025-10-10T01:29:04.7691152Z # rich 2025-10-10T01:29:04.7691312Z markupsafe==3.0.1 2025-10-10T01:29:04.7691496Z  # via 2025-10-10T01:29:04.7691661Z # -c snapshot_constraint.txt 2025-10-10T01:29:04.7691853Z # flask 2025-10-10T01:29:04.7692089Z # jinja2 2025-10-10T01:29:04.7692251Z # mako 2025-10-10T01:29:04.7692434Z # werkzeug 2025-10-10T01:29:04.7692606Z matplotlib==3.9.2 2025-10-10T01:29:04.7692783Z  # via 2025-10-10T01:29:04.7692944Z # -c snapshot_constraint.txt 2025-10-10T01:29:04.7693154Z # -r requirements/test.in 2025-10-10T01:29:04.7693344Z # lightning 2025-10-10T01:29:04.7693497Z # mlflow 2025-10-10T01:29:04.7693646Z # pycocotools 2025-10-10T01:29:04.7693838Z # torchgeo 2025-10-10T01:29:04.7694008Z mbstrdecoder==1.1.3 2025-10-10T01:29:04.7694189Z  # via 2025-10-10T01:29:04.7694352Z # -c snapshot_constraint.txt 2025-10-10T01:29:04.7694561Z # dataproperty 2025-10-10T01:29:04.7694733Z # pytablewriter 2025-10-10T01:29:04.7694931Z # typepy 2025-10-10T01:29:04.7695089Z mdurl==0.1.2 2025-10-10T01:29:04.7695261Z  # via 2025-10-10T01:29:04.7695417Z # -c snapshot_constraint.txt 2025-10-10T01:29:04.7695659Z # markdown-it-py 2025-10-10T01:29:04.7695857Z mistral-common==1.8.2 2025-10-10T01:29:04.7696044Z  # via 2025-10-10T01:29:04.7696198Z # -c snapshot_constraint.txt 2025-10-10T01:29:04.7696444Z # -r requirements/test.in 2025-10-10T01:29:04.7696658Z mlflow==2.22.0 2025-10-10T01:29:04.7696833Z  # via 2025-10-10T01:29:04.7696990Z # -c snapshot_constraint.txt 2025-10-10T01:29:04.7697221Z # terratorch 2025-10-10T01:29:04.7697398Z mlflow-skinny==2.22.0 2025-10-10T01:29:04.7697586Z  # via 2025-10-10T01:29:04.7697745Z # -c snapshot_constraint.txt 2025-10-10T01:29:04.7697962Z # mlflow 2025-10-10T01:29:04.7698128Z more-itertools==10.5.0 2025-10-10T01:29:04.7698320Z  # via 2025-10-10T01:29:04.7698470Z # -c snapshot_constraint.txt 2025-10-10T01:29:04.7698692Z # lm-eval 2025-10-10T01:29:04.7698855Z mpmath==1.3.0 2025-10-10T01:29:04.7699019Z  # via 2025-10-10T01:29:04.7699176Z # -c snapshot_constraint.txt 2025-10-10T01:29:04.7699400Z # sympy 2025-10-10T01:29:04.7699581Z msgpack==1.1.0 2025-10-10T01:29:04.7699747Z  # via 2025-10-10T01:29:04.7699902Z # -c snapshot_constraint.txt 2025-10-10T01:29:04.7700091Z # librosa 2025-10-10T01:29:04.7700281Z # ray 2025-10-10T01:29:04.7700431Z mteb==1.38.11 2025-10-10T01:29:04.7700601Z  # via 2025-10-10T01:29:04.7700761Z # -c snapshot_constraint.txt 2025-10-10T01:29:04.7700999Z # -r requirements/test.in 2025-10-10T01:29:04.7701203Z multidict==6.1.0 2025-10-10T01:29:04.7701381Z  # via 2025-10-10T01:29:04.7701534Z # -c snapshot_constraint.txt 2025-10-10T01:29:04.7701729Z # aiohttp 2025-10-10T01:29:04.7701904Z # yarl 2025-10-10T01:29:04.7702064Z multiprocess==0.70.16 2025-10-10T01:29:04.7702255Z  # via 2025-10-10T01:29:04.7702415Z # -c snapshot_constraint.txt 2025-10-10T01:29:04.7702595Z # datasets 2025-10-10T01:29:04.7702779Z # evaluate 2025-10-10T01:29:04.7702941Z munch==4.0.0 2025-10-10T01:29:04.7703110Z  # via 2025-10-10T01:29:04.7703380Z # -c snapshot_constraint.txt 2025-10-10T01:29:04.7703620Z # pretrainedmodels 2025-10-10T01:29:04.7703830Z mypy-extensions==1.0.0 2025-10-10T01:29:04.7704028Z  # via 2025-10-10T01:29:04.7704186Z # -c snapshot_constraint.txt 2025-10-10T01:29:04.7704404Z # black 2025-10-10T01:29:04.7704560Z networkx==3.2.1 2025-10-10T01:29:04.7704728Z  # via 2025-10-10T01:29:04.7704885Z # -c snapshot_constraint.txt 2025-10-10T01:29:04.7705086Z # scikit-image 2025-10-10T01:29:04.7705271Z # torch 2025-10-10T01:29:04.7705421Z nltk==3.9.1 2025-10-10T01:29:04.7705586Z  # via 2025-10-10T01:29:04.7705755Z # -c snapshot_constraint.txt 2025-10-10T01:29:04.7705986Z # rouge-score 2025-10-10T01:29:04.7706158Z num2words==0.5.14 2025-10-10T01:29:04.7706334Z  # via 2025-10-10T01:29:04.7706492Z # -c snapshot_constraint.txt 2025-10-10T01:29:04.7706804Z # -r requirements/test.in 2025-10-10T01:29:04.7707005Z numba==0.61.2 2025-10-10T01:29:04.7707175Z  # via 2025-10-10T01:29:04.7707335Z # -c snapshot_constraint.txt 2025-10-10T01:29:04.7707534Z # -r requirements/test.in 2025-10-10T01:29:04.7707745Z # librosa 2025-10-10T01:29:04.7707907Z numexpr==2.10.1 2025-10-10T01:29:04.7708076Z  # via 2025-10-10T01:29:04.7708235Z # -c snapshot_constraint.txt 2025-10-10T01:29:04.7708447Z # lm-eval 2025-10-10T01:29:04.7708606Z numpy==1.26.4 2025-10-10T01:29:04.7708768Z  # via 2025-10-10T01:29:04.7708925Z # -c snapshot_constraint.txt 2025-10-10T01:29:04.7709119Z # -r requirements/test.in 2025-10-10T01:29:04.7709302Z # accelerate 2025-10-10T01:29:04.7709457Z # albucore 2025-10-10T01:29:04.7709613Z # albumentations 2025-10-10T01:29:04.7709782Z # bitsandbytes 2025-10-10T01:29:04.7709939Z # bm25s 2025-10-10T01:29:04.7710088Z # contourpy 2025-10-10T01:29:04.7710246Z # cupy-cuda12x 2025-10-10T01:29:04.7710415Z # datasets 2025-10-10T01:29:04.7710564Z # decord 2025-10-10T01:29:04.7710705Z # einx 2025-10-10T01:29:04.7710844Z # encodec 2025-10-10T01:29:04.7710994Z # evaluate 2025-10-10T01:29:04.7711144Z # fastparquet 2025-10-10T01:29:04.7711315Z # genai-perf 2025-10-10T01:29:04.7711463Z # geopandas 2025-10-10T01:29:04.7711612Z # h5py 2025-10-10T01:29:04.7711758Z # imageio 2025-10-10T01:29:04.7711906Z # librosa 2025-10-10T01:29:04.7712047Z # lightly 2025-10-10T01:29:04.7712205Z # lightly-utils 2025-10-10T01:29:04.7712374Z # matplotlib 2025-10-10T01:29:04.7712535Z # mistral-common 2025-10-10T01:29:04.7712692Z # mlflow 2025-10-10T01:29:04.7712835Z # mteb 2025-10-10T01:29:04.7712978Z # numba 2025-10-10T01:29:04.7713117Z # numexpr 2025-10-10T01:29:04.7713281Z # opencv-python-headless 2025-10-10T01:29:04.7713477Z # pandas 2025-10-10T01:29:04.7713629Z # patsy 2025-10-10T01:29:04.7713773Z # peft 2025-10-10T01:29:04.7713923Z # pycocotools 2025-10-10T01:29:04.7714081Z # pyogrio 2025-10-10T01:29:04.7714232Z # rasterio 2025-10-10T01:29:04.7714379Z # rioxarray 2025-10-10T01:29:04.7714534Z # rouge-score 2025-10-10T01:29:04.7714702Z # runai-model-streamer 2025-10-10T01:29:04.7714891Z # sacrebleu 2025-10-10T01:29:04.7715038Z # scikit-image 2025-10-10T01:29:04.7715203Z # scikit-learn 2025-10-10T01:29:04.7715362Z # scipy 2025-10-10T01:29:04.7715529Z # segmentation-models-pytorch 2025-10-10T01:29:04.7715731Z # shapely 2025-10-10T01:29:04.7715876Z # soxr 2025-10-10T01:29:04.7716022Z # statsmodels 2025-10-10T01:29:04.7716177Z # tensorboardx 2025-10-10T01:29:04.7716342Z # tensorizer 2025-10-10T01:29:04.7716496Z # tifffile 2025-10-10T01:29:04.7716644Z # torchgeo 2025-10-10T01:29:04.7716788Z # torchmetrics 2025-10-10T01:29:04.7716943Z # torchvision 2025-10-10T01:29:04.7717101Z # transformers 2025-10-10T01:29:04.7717348Z # tritonclient 2025-10-10T01:29:04.7717494Z # vocos 2025-10-10T01:29:04.7717669Z # xarray 2025-10-10T01:29:04.7717824Z omegaconf==2.3.0 2025-10-10T01:29:04.7718000Z  # via 2025-10-10T01:29:04.7718155Z # -c snapshot_constraint.txt 2025-10-10T01:29:04.7718356Z # hydra-core 2025-10-10T01:29:04.7718610Z # lightning 2025-10-10T01:29:04.7718788Z open-clip-torch==2.32.0 2025-10-10T01:29:04.7719022Z  # via -r requirements/test.in 2025-10-10T01:29:04.7719257Z opencensus==0.11.4 2025-10-10T01:29:04.7719439Z  # via 2025-10-10T01:29:04.7719598Z # -c snapshot_constraint.txt 2025-10-10T01:29:04.7719816Z # ray 2025-10-10T01:29:04.7719977Z opencensus-context==0.1.3 2025-10-10T01:29:04.7720178Z  # via 2025-10-10T01:29:04.7720329Z # -c snapshot_constraint.txt 2025-10-10T01:29:04.7720552Z # opencensus 2025-10-10T01:29:04.7720830Z opencv-python-headless==4.11.0.86 2025-10-10T01:29:04.7721060Z  # via 2025-10-10T01:29:04.7721211Z # -c snapshot_constraint.txt 2025-10-10T01:29:04.7721410Z # -r requirements/test.in 2025-10-10T01:29:04.7721594Z # albucore 2025-10-10T01:29:04.7721745Z # albumentations 2025-10-10T01:29:04.7721940Z # mistral-common 2025-10-10T01:29:04.7722145Z opentelemetry-api==1.35.0 2025-10-10T01:29:04.7722349Z  # via 2025-10-10T01:29:04.7722505Z # -c snapshot_constraint.txt 2025-10-10T01:29:04.7722696Z # mlflow-skinny 2025-10-10T01:29:04.7722895Z # opentelemetry-exporter-prometheus 2025-10-10T01:29:04.7723127Z # opentelemetry-sdk 2025-10-10T01:29:04.7723381Z # opentelemetry-semantic-conventions 2025-10-10T01:29:04.7723656Z opentelemetry-exporter-prometheus==0.56b0 2025-10-10T01:29:04.7723904Z  # via 2025-10-10T01:29:04.7724229Z # -c snapshot_constraint.txt 2025-10-10T01:29:04.7724448Z # ray 2025-10-10T01:29:04.7724625Z opentelemetry-proto==1.36.0 2025-10-10T01:29:04.7724836Z  # via 2025-10-10T01:29:04.7724998Z # -c snapshot_constraint.txt 2025-10-10T01:29:04.7725203Z # ray 2025-10-10T01:29:04.7725362Z opentelemetry-sdk==1.35.0 2025-10-10T01:29:04.7725559Z  # via 2025-10-10T01:29:04.7725725Z # -c snapshot_constraint.txt 2025-10-10T01:29:04.7725920Z # mlflow-skinny 2025-10-10T01:29:04.7726110Z # opentelemetry-exporter-prometheus 2025-10-10T01:29:04.7726352Z # ray 2025-10-10T01:29:04.7726541Z opentelemetry-semantic-conventions==0.56b0 2025-10-10T01:29:04.7726783Z  # via 2025-10-10T01:29:04.7726940Z # -c snapshot_constraint.txt 2025-10-10T01:29:04.7727174Z # opentelemetry-sdk 2025-10-10T01:29:04.7727378Z packaging==24.2 2025-10-10T01:29:04.7727551Z  # via 2025-10-10T01:29:04.7727708Z # -c snapshot_constraint.txt 2025-10-10T01:29:04.7727899Z # accelerate 2025-10-10T01:29:04.7728045Z # black 2025-10-10T01:29:04.7728204Z # datamodel-code-generator 2025-10-10T01:29:04.7728398Z # datasets 2025-10-10T01:29:04.7728544Z # evaluate 2025-10-10T01:29:04.7728696Z # fastparquet 2025-10-10T01:29:04.7728845Z # geopandas 2025-10-10T01:29:04.7728997Z # gunicorn 2025-10-10T01:29:04.7729149Z # huggingface-hub 2025-10-10T01:29:04.7729316Z # hydra-core 2025-10-10T01:29:04.7729460Z # kornia 2025-10-10T01:29:04.7729605Z # lazy-loader 2025-10-10T01:29:04.7729759Z # lightning 2025-10-10T01:29:04.7729923Z # lightning-utilities 2025-10-10T01:29:04.7730098Z # matplotlib 2025-10-10T01:29:04.7730257Z # mlflow-skinny 2025-10-10T01:29:04.7730416Z # peft 2025-10-10T01:29:04.7730564Z # plotly 2025-10-10T01:29:04.7730709Z # pooch 2025-10-10T01:29:04.7730852Z # pyogrio 2025-10-10T01:29:04.7730998Z # pytest 2025-10-10T01:29:04.7731160Z # pytest-rerunfailures 2025-10-10T01:29:04.7731359Z # pytorch-lightning 2025-10-10T01:29:04.7731536Z # ray 2025-10-10T01:29:04.7731830Z # rioxarray 2025-10-10T01:29:04.7731977Z # scikit-image 2025-10-10T01:29:04.7732138Z # statsmodels 2025-10-10T01:29:04.7732289Z # tensorboardx 2025-10-10T01:29:04.7732447Z # torchmetrics 2025-10-10T01:29:04.7732597Z # transformers 2025-10-10T01:29:04.7732761Z # typepy 2025-10-10T01:29:04.7732946Z # xarray 2025-10-10T01:29:04.7733101Z pandas==2.2.3 2025-10-10T01:29:04.7733266Z  # via 2025-10-10T01:29:04.7733426Z # -c snapshot_constraint.txt 2025-10-10T01:29:04.7733621Z # datasets 2025-10-10T01:29:04.7733772Z # evaluate 2025-10-10T01:29:04.7733915Z # fastparquet 2025-10-10T01:29:04.7734066Z # genai-perf 2025-10-10T01:29:04.7734219Z # geopandas 2025-10-10T01:29:04.7734373Z # mlflow 2025-10-10T01:29:04.7734522Z # statsmodels 2025-10-10T01:29:04.7734679Z # torchgeo 2025-10-10T01:29:04.7734854Z # xarray 2025-10-10T01:29:04.7735003Z pathspec==0.12.1 2025-10-10T01:29:04.7735295Z  # via 2025-10-10T01:29:04.7735460Z # -c snapshot_constraint.txt 2025-10-10T01:29:04.7735679Z # black 2025-10-10T01:29:04.7735831Z pathvalidate==3.2.1 2025-10-10T01:29:04.7736008Z  # via 2025-10-10T01:29:04.7736162Z # -c snapshot_constraint.txt 2025-10-10T01:29:04.7736389Z # pytablewriter 2025-10-10T01:29:04.7736562Z patsy==1.0.1 2025-10-10T01:29:04.7736727Z  # via 2025-10-10T01:29:04.7736882Z # -c snapshot_constraint.txt 2025-10-10T01:29:04.7737106Z # statsmodels 2025-10-10T01:29:04.7737274Z peft==0.16.0 2025-10-10T01:29:04.7737438Z  # via 2025-10-10T01:29:04.7737594Z # -c snapshot_constraint.txt 2025-10-10T01:29:04.7737793Z # -r requirements/test.in 2025-10-10T01:29:04.7738003Z # lm-eval 2025-10-10T01:29:04.7738161Z pillow==10.4.0 2025-10-10T01:29:04.7738327Z  # via 2025-10-10T01:29:04.7738485Z # -c snapshot_constraint.txt 2025-10-10T01:29:04.7738675Z # genai-perf 2025-10-10T01:29:04.7738839Z # imageio 2025-10-10T01:29:04.7738991Z # lightly-utils 2025-10-10T01:29:04.7739150Z # matplotlib 2025-10-10T01:29:04.7739305Z # mistral-common 2025-10-10T01:29:04.7739471Z # scikit-image 2025-10-10T01:29:04.7739654Z # segmentation-models-pytorch 2025-10-10T01:29:04.7739867Z # sentence-transformers 2025-10-10T01:29:04.7740052Z # torchgeo 2025-10-10T01:29:04.7740239Z # torchvision 2025-10-10T01:29:04.7740421Z platformdirs==4.3.6 2025-10-10T01:29:04.7740593Z  # via 2025-10-10T01:29:04.7740752Z # -c snapshot_constraint.txt 2025-10-10T01:29:04.7740940Z # black 2025-10-10T01:29:04.7741081Z # pooch 2025-10-10T01:29:04.7741254Z # virtualenv 2025-10-10T01:29:04.7741424Z plotly==5.24.1 2025-10-10T01:29:04.7741599Z  # via 2025-10-10T01:29:04.7741758Z # -c snapshot_constraint.txt 2025-10-10T01:29:04.7741974Z # genai-perf 2025-10-10T01:29:04.7742140Z pluggy==1.5.0 2025-10-10T01:29:04.7742307Z  # via 2025-10-10T01:29:04.7742468Z # -c snapshot_constraint.txt 2025-10-10T01:29:04.7742652Z # pytest 2025-10-10T01:29:04.7742833Z # pytest-cov 2025-10-10T01:29:04.7743000Z polars==1.29.0 2025-10-10T01:29:04.7743163Z  # via 2025-10-10T01:29:04.7743312Z # -c snapshot_constraint.txt 2025-10-10T01:29:04.7743524Z # mteb 2025-10-10T01:29:04.7743687Z pooch==1.8.2 2025-10-10T01:29:04.7743848Z  # via 2025-10-10T01:29:04.7744007Z # -c snapshot_constraint.txt 2025-10-10T01:29:04.7744220Z # librosa 2025-10-10T01:29:04.7744382Z portalocker==2.10.1 2025-10-10T01:29:04.7744551Z  # via 2025-10-10T01:29:04.7744703Z # -c snapshot_constraint.txt 2025-10-10T01:29:04.7744920Z # sacrebleu 2025-10-10T01:29:04.7745082Z pqdm==0.2.0 2025-10-10T01:29:04.7745238Z  # via 2025-10-10T01:29:04.7745397Z # -c snapshot_constraint.txt 2025-10-10T01:29:04.7745634Z # -r requirements/test.in 2025-10-10T01:29:04.7745845Z pretrainedmodels==0.7.4 2025-10-10T01:29:04.7746109Z  # via 2025-10-10T01:29:04.7746267Z # -c snapshot_constraint.txt 2025-10-10T01:29:04.7746536Z # segmentation-models-pytorch 2025-10-10T01:29:04.7746769Z prometheus-client==0.22.0 2025-10-10T01:29:04.7746965Z  # via 2025-10-10T01:29:04.7747123Z # -c snapshot_constraint.txt 2025-10-10T01:29:04.7747345Z # opentelemetry-exporter-prometheus 2025-10-10T01:29:04.7747587Z # ray 2025-10-10T01:29:04.7747732Z propcache==0.2.0 2025-10-10T01:29:04.7747908Z  # via 2025-10-10T01:29:04.7748065Z # -c snapshot_constraint.txt 2025-10-10T01:29:04.7748270Z # yarl 2025-10-10T01:29:04.7748427Z proto-plus==1.26.1 2025-10-10T01:29:04.7748613Z  # via 2025-10-10T01:29:04.7748775Z # -c snapshot_constraint.txt 2025-10-10T01:29:04.7749017Z # google-api-core 2025-10-10T01:29:04.7749200Z protobuf==5.28.3 2025-10-10T01:29:04.7749373Z  # via 2025-10-10T01:29:04.7749614Z # -c snapshot_constraint.txt 2025-10-10T01:29:04.7749824Z # google-api-core 2025-10-10T01:29:04.7750013Z # googleapis-common-protos 2025-10-10T01:29:04.7750213Z # mlflow-skinny 2025-10-10T01:29:04.7750384Z # opentelemetry-proto 2025-10-10T01:29:04.7750569Z # proto-plus 2025-10-10T01:29:04.7750719Z # ray 2025-10-10T01:29:04.7750868Z # tensorboardx 2025-10-10T01:29:04.7751060Z # tensorizer 2025-10-10T01:29:04.7751224Z psutil==6.1.0 2025-10-10T01:29:04.7751391Z  # via 2025-10-10T01:29:04.7751547Z # -c snapshot_constraint.txt 2025-10-10T01:29:04.7751733Z # accelerate 2025-10-10T01:29:04.7751796Z # peft 2025-10-10T01:29:04.7751885Z # tensorizer 2025-10-10T01:29:04.7751943Z py==1.11.0 2025-10-10T01:29:04.7752025Z  # via 2025-10-10T01:29:04.7752096Z # -c snapshot_constraint.txt 2025-10-10T01:29:04.7752197Z # pytest-forked 2025-10-10T01:29:04.7752258Z py-spy==0.4.0 2025-10-10T01:29:04.7752338Z  # via 2025-10-10T01:29:04.7752418Z # -c snapshot_constraint.txt 2025-10-10T01:29:04.7752497Z # ray 2025-10-10T01:29:04.7752564Z pyarrow==18.0.0 2025-10-10T01:29:04.7752640Z  # via 2025-10-10T01:29:04.7752710Z # -c snapshot_constraint.txt 2025-10-10T01:29:04.7752771Z # datasets 2025-10-10T01:29:04.7752832Z # genai-perf 2025-10-10T01:29:04.7752914Z # mlflow 2025-10-10T01:29:04.7752978Z pyasn1==0.6.1 2025-10-10T01:29:04.7753053Z  # via 2025-10-10T01:29:04.7753130Z # -c snapshot_constraint.txt 2025-10-10T01:29:04.7753193Z # pyasn1-modules 2025-10-10T01:29:04.7753272Z # rsa 2025-10-10T01:29:04.7753343Z pyasn1-modules==0.4.2 2025-10-10T01:29:04.7753419Z  # via 2025-10-10T01:29:04.7753494Z # -c snapshot_constraint.txt 2025-10-10T01:29:04.7753585Z # google-auth 2025-10-10T01:29:04.7753646Z pybind11==2.13.6 2025-10-10T01:29:04.7753727Z  # via 2025-10-10T01:29:04.7753800Z # -c snapshot_constraint.txt 2025-10-10T01:29:04.7753886Z # lm-eval 2025-10-10T01:29:04.7753955Z pycocotools==2.0.8 2025-10-10T01:29:04.7754033Z  # via 2025-10-10T01:29:04.7754110Z # -c snapshot_constraint.txt 2025-10-10T01:29:04.7754199Z # terratorch 2025-10-10T01:29:04.7754261Z pycountry==24.6.1 2025-10-10T01:29:04.7754343Z  # via 2025-10-10T01:29:04.7754412Z # -c snapshot_constraint.txt 2025-10-10T01:29:04.7754525Z # pydantic-extra-types 2025-10-10T01:29:04.7754588Z pycparser==2.22 2025-10-10T01:29:04.7754666Z  # via 2025-10-10T01:29:04.7754742Z # -c snapshot_constraint.txt 2025-10-10T01:29:04.7754823Z # cffi 2025-10-10T01:29:04.7754891Z pycryptodomex==3.22.0 2025-10-10T01:29:04.7754975Z  # via 2025-10-10T01:29:04.7755045Z # -c snapshot_constraint.txt 2025-10-10T01:29:04.7755141Z # blobfile 2025-10-10T01:29:04.7755202Z pydantic==2.11.7 2025-10-10T01:29:04.7755280Z  # via 2025-10-10T01:29:04.7755363Z # -c snapshot_constraint.txt 2025-10-10T01:29:04.7755535Z # -r requirements/test.in 2025-10-10T01:29:04.7755604Z # albumentations 2025-10-10T01:29:04.7755678Z # datamodel-code-generator 2025-10-10T01:29:04.7755736Z # fastapi 2025-10-10T01:29:04.7755796Z # lightly 2025-10-10T01:29:04.7755859Z # mistral-common 2025-10-10T01:29:04.7755920Z # mlflow-skinny 2025-10-10T01:29:04.7755982Z # mteb 2025-10-10T01:29:04.7756056Z # pydantic-extra-types 2025-10-10T01:29:04.7756145Z # ray 2025-10-10T01:29:04.7756210Z pydantic-core==2.33.2 2025-10-10T01:29:04.7756286Z  # via 2025-10-10T01:29:04.7756365Z # -c snapshot_constraint.txt 2025-10-10T01:29:04.7756451Z # pydantic 2025-10-10T01:29:04.7756527Z pydantic-extra-types==2.10.5 2025-10-10T01:29:04.7756609Z  # via 2025-10-10T01:29:04.7756679Z # -c snapshot_constraint.txt 2025-10-10T01:29:04.7756779Z # mistral-common 2025-10-10T01:29:04.7756922Z pygments==2.18.0 2025-10-10T01:29:04.7757005Z  # via 2025-10-10T01:29:04.7757082Z # -c snapshot_constraint.txt 2025-10-10T01:29:04.7757163Z # rich 2025-10-10T01:29:04.7757228Z pyogrio==0.11.0 2025-10-10T01:29:04.7757306Z  # via 2025-10-10T01:29:04.7757376Z # -c snapshot_constraint.txt 2025-10-10T01:29:04.7757469Z # geopandas 2025-10-10T01:29:04.7757531Z pyparsing==3.2.0 2025-10-10T01:29:04.7757608Z  # via 2025-10-10T01:29:04.7757683Z # -c snapshot_constraint.txt 2025-10-10T01:29:04.7757742Z # matplotlib 2025-10-10T01:29:04.7757834Z # rasterio 2025-10-10T01:29:04.7757896Z pyproj==3.7.1 2025-10-10T01:29:04.7757973Z  # via 2025-10-10T01:29:04.7758053Z # -c snapshot_constraint.txt 2025-10-10T01:29:04.7758113Z # geopandas 2025-10-10T01:29:04.7758185Z # rioxarray 2025-10-10T01:29:04.7758274Z # torchgeo 2025-10-10T01:29:04.7758340Z pyrate-limiter==3.7.0 2025-10-10T01:29:04.7758421Z  # via 2025-10-10T01:29:04.7758541Z # -c snapshot_constraint.txt 2025-10-10T01:29:04.7758639Z # schemathesis 2025-10-10T01:29:04.7758717Z pystemmer==3.0.0 2025-10-10T01:29:04.7758797Z  # via 2025-10-10T01:29:04.7758874Z # -c snapshot_constraint.txt 2025-10-10T01:29:04.7758954Z # mteb 2025-10-10T01:29:04.7759020Z pytablewriter==1.2.0 2025-10-10T01:29:04.7759100Z  # via 2025-10-10T01:29:04.7759170Z # -c snapshot_constraint.txt 2025-10-10T01:29:04.7759257Z # lm-eval 2025-10-10T01:29:04.7759317Z pytest==8.3.5 2025-10-10T01:29:04.7759393Z  # via 2025-10-10T01:29:04.7759468Z # -c snapshot_constraint.txt 2025-10-10T01:29:04.7759536Z # -r requirements/test.in 2025-10-10T01:29:04.7759608Z # buildkite-test-collector 2025-10-10T01:29:04.7759672Z # genai-perf 2025-10-10T01:29:04.7759734Z # pytest-asyncio 2025-10-10T01:29:04.7759798Z # pytest-cov 2025-10-10T01:29:04.7759860Z # pytest-forked 2025-10-10T01:29:04.7759924Z # pytest-mock 2025-10-10T01:29:04.7760002Z # pytest-rerunfailures 2025-10-10T01:29:04.7760063Z # pytest-shard 2025-10-10T01:29:04.7760131Z # pytest-subtests 2025-10-10T01:29:04.7760204Z # pytest-timeout 2025-10-10T01:29:04.7760266Z # schemathesis 2025-10-10T01:29:04.7760362Z # terratorch 2025-10-10T01:29:04.7760427Z pytest-asyncio==0.24.0 2025-10-10T01:29:04.7760502Z  # via 2025-10-10T01:29:04.7760583Z # -c snapshot_constraint.txt 2025-10-10T01:29:04.7760693Z # -r requirements/test.in 2025-10-10T01:29:04.7760760Z pytest-cov==6.3.0 2025-10-10T01:29:04.7760837Z  # via 2025-10-10T01:29:04.7760909Z # -c snapshot_constraint.txt 2025-10-10T01:29:04.7761026Z # -r requirements/test.in 2025-10-10T01:29:04.7761092Z pytest-forked==1.6.0 2025-10-10T01:29:04.7761179Z  # via 2025-10-10T01:29:04.7761250Z # -c snapshot_constraint.txt 2025-10-10T01:29:04.7761353Z # -r requirements/test.in 2025-10-10T01:29:04.7761425Z pytest-mock==3.14.0 2025-10-10T01:29:04.7761589Z  # via 2025-10-10T01:29:04.7761673Z # -c snapshot_constraint.txt 2025-10-10T01:29:04.7761773Z # genai-perf 2025-10-10T01:29:04.7761848Z pytest-rerunfailures==14.0 2025-10-10T01:29:04.7761930Z  # via 2025-10-10T01:29:04.7762001Z # -c snapshot_constraint.txt 2025-10-10T01:29:04.7762108Z # -r requirements/test.in 2025-10-10T01:29:04.7762179Z pytest-shard==0.1.2 2025-10-10T01:29:04.7762257Z  # via 2025-10-10T01:29:04.7762334Z # -c snapshot_constraint.txt 2025-10-10T01:29:04.7762437Z # -r requirements/test.in 2025-10-10T01:29:04.7762505Z pytest-subtests==0.14.1 2025-10-10T01:29:04.7762590Z  # via 2025-10-10T01:29:04.7762661Z # -c snapshot_constraint.txt 2025-10-10T01:29:04.7762759Z # schemathesis 2025-10-10T01:29:04.7762826Z pytest-timeout==2.3.1 2025-10-10T01:29:04.7762905Z  # via 2025-10-10T01:29:04.7762979Z # -c snapshot_constraint.txt 2025-10-10T01:29:04.7763155Z # -r requirements/test.in 2025-10-10T01:29:04.7763222Z python-box==7.3.2 2025-10-10T01:29:04.7763307Z  # via 2025-10-10T01:29:04.7763377Z # -c snapshot_constraint.txt 2025-10-10T01:29:04.7763471Z # terratorch 2025-10-10T01:29:04.7763543Z python-dateutil==2.9.0.post0 2025-10-10T01:29:04.7763620Z  # via 2025-10-10T01:29:04.7763697Z # -c snapshot_constraint.txt 2025-10-10T01:29:04.7763754Z # arrow 2025-10-10T01:29:04.7763817Z # botocore 2025-10-10T01:29:04.7763889Z # graphene 2025-10-10T01:29:04.7763947Z # lightly 2025-10-10T01:29:04.7764013Z # matplotlib 2025-10-10T01:29:04.7764072Z # pandas 2025-10-10T01:29:04.7764160Z # typepy 2025-10-10T01:29:04.7764236Z python-rapidjson==1.20 2025-10-10T01:29:04.7764313Z  # via 2025-10-10T01:29:04.7764387Z # -c snapshot_constraint.txt 2025-10-10T01:29:04.7764479Z # tritonclient 2025-10-10T01:29:04.7764554Z pytorch-lightning==2.5.5 2025-10-10T01:29:04.7764640Z  # via 2025-10-10T01:29:04.7764696Z # lightly 2025-10-10T01:29:04.7764784Z # lightning 2025-10-10T01:29:04.7764865Z pytrec-eval-terrier==0.5.7 2025-10-10T01:29:04.7764943Z  # via 2025-10-10T01:29:04.7765019Z # -c snapshot_constraint.txt 2025-10-10T01:29:04.7765100Z # mteb 2025-10-10T01:29:04.7765160Z pytz==2024.2 2025-10-10T01:29:04.7765241Z  # via 2025-10-10T01:29:04.7765310Z # -c snapshot_constraint.txt 2025-10-10T01:29:04.7765373Z # pandas 2025-10-10T01:29:04.7765462Z # typepy 2025-10-10T01:29:04.7765521Z pyyaml==6.0.2 2025-10-10T01:29:04.7765606Z  # via 2025-10-10T01:29:04.7765677Z # -c snapshot_constraint.txt 2025-10-10T01:29:04.7765735Z # accelerate 2025-10-10T01:29:04.7765803Z # albumentations 2025-10-10T01:29:04.7765875Z # datamodel-code-generator 2025-10-10T01:29:04.7765941Z # datasets 2025-10-10T01:29:04.7766005Z # genai-perf 2025-10-10T01:29:04.7766073Z # huggingface-hub 2025-10-10T01:29:04.7766140Z # jsonargparse 2025-10-10T01:29:04.7766199Z # lightning 2025-10-10T01:29:04.7766266Z # mlflow-skinny 2025-10-10T01:29:04.7766323Z # omegaconf 2025-10-10T01:29:04.7766380Z # peft 2025-10-10T01:29:04.7766452Z # pytorch-lightning 2025-10-10T01:29:04.7766508Z # ray 2025-10-10T01:29:04.7766569Z # responses 2025-10-10T01:29:04.7766635Z # schemathesis 2025-10-10T01:29:04.7766689Z # timm 2025-10-10T01:29:04.7766753Z # transformers 2025-10-10T01:29:04.7766837Z # vocos 2025-10-10T01:29:04.7766898Z rapidfuzz==3.12.1 2025-10-10T01:29:04.7766980Z  # via 2025-10-10T01:29:04.7767051Z # -c snapshot_constraint.txt 2025-10-10T01:29:04.7767136Z # jiwer 2025-10-10T01:29:04.7767197Z rasterio==1.4.3 2025-10-10T01:29:04.7767274Z  # via 2025-10-10T01:29:04.7767347Z # -c snapshot_constraint.txt 2025-10-10T01:29:04.7767403Z # rioxarray 2025-10-10T01:29:04.7767463Z # terratorch 2025-10-10T01:29:04.7767651Z # torchgeo 2025-10-10T01:29:04.7767710Z ray==2.48.0 2025-10-10T01:29:04.7767794Z  # via 2025-10-10T01:29:04.7767866Z # -c snapshot_constraint.txt 2025-10-10T01:29:04.7767968Z # -r requirements/test.in 2025-10-10T01:29:04.7768032Z redis==5.2.0 2025-10-10T01:29:04.7768109Z  # via 2025-10-10T01:29:04.7768179Z # -c snapshot_constraint.txt 2025-10-10T01:29:04.7768271Z # tensorizer 2025-10-10T01:29:04.7768333Z referencing==0.35.1 2025-10-10T01:29:04.7768414Z  # via 2025-10-10T01:29:04.7768485Z # -c snapshot_constraint.txt 2025-10-10T01:29:04.7768544Z # jsonschema 2025-10-10T01:29:04.7768668Z # jsonschema-specifications 2025-10-10T01:29:04.7768729Z regex==2024.9.11 2025-10-10T01:29:04.7768812Z  # via 2025-10-10T01:29:04.7768883Z # -c snapshot_constraint.txt 2025-10-10T01:29:04.7768938Z # nltk 2025-10-10T01:29:04.7769010Z # open-clip-torch 2025-10-10T01:29:04.7769153Z # sacrebleu 2025-10-10T01:29:04.7769217Z # tiktoken 2025-10-10T01:29:04.7769321Z # transformers 2025-10-10T01:29:04.7769382Z requests==2.32.3 2025-10-10T01:29:04.7769465Z  # via 2025-10-10T01:29:04.7769535Z # -c snapshot_constraint.txt 2025-10-10T01:29:04.7769608Z # buildkite-test-collector 2025-10-10T01:29:04.7769678Z # databricks-sdk 2025-10-10T01:29:04.7769734Z # datasets 2025-10-10T01:29:04.7769794Z # docker 2025-10-10T01:29:04.7769849Z # evaluate 2025-10-10T01:29:04.7769915Z # google-api-core 2025-10-10T01:29:04.7769989Z # google-cloud-storage 2025-10-10T01:29:04.7770054Z # huggingface-hub 2025-10-10T01:29:04.7770111Z # lightly 2025-10-10T01:29:04.7770175Z # lm-eval 2025-10-10T01:29:04.7770238Z # mistral-common 2025-10-10T01:29:04.7770304Z # mlflow-skinny 2025-10-10T01:29:04.7770362Z # mteb 2025-10-10T01:29:04.7770418Z # pooch 2025-10-10T01:29:04.7770482Z # ray 2025-10-10T01:29:04.7770543Z # responses 2025-10-10T01:29:04.7770608Z # schemathesis 2025-10-10T01:29:04.7770678Z # starlette-testclient 2025-10-10T01:29:04.7770737Z # tiktoken 2025-10-10T01:29:04.7770836Z # transformers 2025-10-10T01:29:04.7770897Z responses==0.25.3 2025-10-10T01:29:04.7770976Z  # via 2025-10-10T01:29:04.7771057Z # -c snapshot_constraint.txt 2025-10-10T01:29:04.7771148Z # genai-perf 2025-10-10T01:29:04.7771225Z rfc3339-validator==0.1.4 2025-10-10T01:29:04.7771302Z  # via 2025-10-10T01:29:04.7771372Z # -c snapshot_constraint.txt 2025-10-10T01:29:04.7771465Z # jsonschema 2025-10-10T01:29:04.7771525Z rfc3987==1.3.8 2025-10-10T01:29:04.7771608Z  # via 2025-10-10T01:29:04.7771678Z # -c snapshot_constraint.txt 2025-10-10T01:29:04.7771772Z # jsonschema 2025-10-10T01:29:04.7771837Z rich==13.9.4 2025-10-10T01:29:04.7771913Z  # via 2025-10-10T01:29:04.7771986Z # -c snapshot_constraint.txt 2025-10-10T01:29:04.7772055Z # genai-perf 2025-10-10T01:29:04.7772111Z # lightning 2025-10-10T01:29:04.7772174Z # mteb 2025-10-10T01:29:04.7772258Z # typer 2025-10-10T01:29:04.7772326Z rioxarray==0.19.0 2025-10-10T01:29:04.7772407Z  # via 2025-10-10T01:29:04.7772476Z # -c snapshot_constraint.txt 2025-10-10T01:29:04.7772578Z # terratorch 2025-10-10T01:29:04.7772643Z rouge-score==0.1.2 2025-10-10T01:29:04.7772720Z  # via 2025-10-10T01:29:04.7772794Z # -c snapshot_constraint.txt 2025-10-10T01:29:04.7772878Z # lm-eval 2025-10-10T01:29:04.7772939Z rpds-py==0.20.1 2025-10-10T01:29:04.7773021Z  # via 2025-10-10T01:29:04.7773091Z # -c snapshot_constraint.txt 2025-10-10T01:29:04.7773155Z # jsonschema 2025-10-10T01:29:04.7773246Z # referencing 2025-10-10T01:29:04.7773303Z rsa==4.9.1 2025-10-10T01:29:04.7773384Z  # via 2025-10-10T01:29:04.7773455Z # -c snapshot_constraint.txt 2025-10-10T01:29:04.7773547Z # google-auth 2025-10-10T01:29:04.7773691Z rtree==1.4.0 2025-10-10T01:29:04.7773769Z  # via 2025-10-10T01:29:04.7773845Z # -c snapshot_constraint.txt 2025-10-10T01:29:04.7773930Z # torchgeo 2025-10-10T01:29:04.7774004Z runai-model-streamer==0.14.0 2025-10-10T01:29:04.7774087Z  # via 2025-10-10T01:29:04.7774157Z # -c snapshot_constraint.txt 2025-10-10T01:29:04.7774266Z # -r requirements/test.in 2025-10-10T01:29:04.7774343Z runai-model-streamer-gcs==0.14.0 2025-10-10T01:29:04.7774421Z  # via 2025-10-10T01:29:04.7774496Z # -c snapshot_constraint.txt 2025-10-10T01:29:04.7774600Z # runai-model-streamer 2025-10-10T01:29:04.7774680Z runai-model-streamer-s3==0.14.0 2025-10-10T01:29:04.7774757Z  # via 2025-10-10T01:29:04.7774826Z # -c snapshot_constraint.txt 2025-10-10T01:29:04.7774944Z # runai-model-streamer 2025-10-10T01:29:04.7775007Z s3transfer==0.10.3 2025-10-10T01:29:04.7775178Z  # via 2025-10-10T01:29:04.7775261Z # -c snapshot_constraint.txt 2025-10-10T01:29:04.7775345Z # boto3 2025-10-10T01:29:04.7775411Z sacrebleu==2.4.3 2025-10-10T01:29:04.7775488Z  # via 2025-10-10T01:29:04.7775559Z # -c snapshot_constraint.txt 2025-10-10T01:29:04.7775647Z # lm-eval 2025-10-10T01:29:04.7775710Z safetensors==0.4.5 2025-10-10T01:29:04.7775791Z  # via 2025-10-10T01:29:04.7775861Z # -c snapshot_constraint.txt 2025-10-10T01:29:04.7775919Z # accelerate 2025-10-10T01:29:04.7775988Z # open-clip-torch 2025-10-10T01:29:04.7776045Z # peft 2025-10-10T01:29:04.7776102Z # timm 2025-10-10T01:29:04.7776200Z # transformers 2025-10-10T01:29:04.7776263Z schemathesis==3.39.15 2025-10-10T01:29:04.7776345Z  # via 2025-10-10T01:29:04.7776414Z # -c snapshot_constraint.txt 2025-10-10T01:29:04.7776518Z # -r requirements/test.in 2025-10-10T01:29:04.7776589Z scikit-image==0.25.2 2025-10-10T01:29:04.7776670Z  # via 2025-10-10T01:29:04.7776743Z # -c snapshot_constraint.txt 2025-10-10T01:29:04.7776855Z # albumentations 2025-10-10T01:29:04.7776921Z scikit-learn==1.5.2 2025-10-10T01:29:04.7777005Z  # via 2025-10-10T01:29:04.7777076Z # -c snapshot_constraint.txt 2025-10-10T01:29:04.7777138Z # albumentations 2025-10-10T01:29:04.7777201Z # librosa 2025-10-10T01:29:04.7777259Z # lm-eval 2025-10-10T01:29:04.7777323Z # mlflow 2025-10-10T01:29:04.7777379Z # mteb 2025-10-10T01:29:04.7777489Z # sentence-transformers 2025-10-10T01:29:04.7777560Z scipy==1.13.1 2025-10-10T01:29:04.7777636Z  # via 2025-10-10T01:29:04.7777710Z # -c snapshot_constraint.txt 2025-10-10T01:29:04.7777778Z # albumentations 2025-10-10T01:29:04.7777833Z # bm25s 2025-10-10T01:29:04.7777894Z # librosa 2025-10-10T01:29:04.7777950Z # mlflow 2025-10-10T01:29:04.7778005Z # mteb 2025-10-10T01:29:04.7778071Z # scikit-image 2025-10-10T01:29:04.7778135Z # scikit-learn 2025-10-10T01:29:04.7778212Z # sentence-transformers 2025-10-10T01:29:04.7778272Z # statsmodels 2025-10-10T01:29:04.7778356Z # vocos 2025-10-10T01:29:04.7778447Z segmentation-models-pytorch==0.4.0 2025-10-10T01:29:04.7778523Z  # via 2025-10-10T01:29:04.7778582Z # terratorch 2025-10-10T01:29:04.7778673Z # torchgeo 2025-10-10T01:29:04.7778748Z sentence-transformers==3.2.1 2025-10-10T01:29:04.7778839Z  # via 2025-10-10T01:29:04.7778913Z # -c snapshot_constraint.txt 2025-10-10T01:29:04.7778982Z # -r requirements/test.in 2025-10-10T01:29:04.7779068Z # mteb 2025-10-10T01:29:04.7779134Z sentencepiece==0.2.0 2025-10-10T01:29:04.7779215Z  # via 2025-10-10T01:29:04.7779286Z # -c snapshot_constraint.txt 2025-10-10T01:29:04.7779381Z # mistral-common 2025-10-10T01:29:04.7779450Z setuptools==77.0.3 2025-10-10T01:29:04.7779525Z  # via 2025-10-10T01:29:04.7779602Z # -c snapshot_constraint.txt 2025-10-10T01:29:04.7779753Z # lightning-utilities 2025-10-10T01:29:04.7779814Z # pytablewriter 2025-10-10T01:29:04.7779900Z # torch 2025-10-10T01:29:04.7779960Z shapely==2.1.1 2025-10-10T01:29:04.7780037Z  # via 2025-10-10T01:29:04.7780113Z # -c snapshot_constraint.txt 2025-10-10T01:29:04.7780170Z # geopandas 2025-10-10T01:29:04.7780260Z # torchgeo 2025-10-10T01:29:04.7780322Z shellingham==1.5.4 2025-10-10T01:29:04.7780398Z  # via 2025-10-10T01:29:04.7780474Z # -c snapshot_constraint.txt 2025-10-10T01:29:04.7780556Z # typer 2025-10-10T01:29:04.7780614Z six==1.16.0 2025-10-10T01:29:04.7780705Z  # via 2025-10-10T01:29:04.7780777Z # -c snapshot_constraint.txt 2025-10-10T01:29:04.7780838Z # junit-xml 2025-10-10T01:29:04.7780894Z # lightly 2025-10-10T01:29:04.7780952Z # opencensus 2025-10-10T01:29:04.7781023Z # python-dateutil 2025-10-10T01:29:04.7781155Z # rfc3339-validator 2025-10-10T01:29:04.7781226Z # rouge-score 2025-10-10T01:29:04.7781353Z # segmentation-models-pytorch 2025-10-10T01:29:04.7781415Z smart-open==7.1.0 2025-10-10T01:29:04.7781496Z  # via 2025-10-10T01:29:04.7781566Z # -c snapshot_constraint.txt 2025-10-10T01:29:04.7781643Z # ray 2025-10-10T01:29:04.7781705Z smmap==5.0.2 2025-10-10T01:29:04.7781783Z  # via 2025-10-10T01:29:04.7781857Z # -c snapshot_constraint.txt 2025-10-10T01:29:04.7781939Z # gitdb 2025-10-10T01:29:04.7781998Z sniffio==1.3.1 2025-10-10T01:29:04.7782077Z  # via 2025-10-10T01:29:04.7782147Z # -c snapshot_constraint.txt 2025-10-10T01:29:04.7782207Z # anyio 2025-10-10T01:29:04.7782291Z # httpx 2025-10-10T01:29:04.7782371Z sortedcontainers==2.4.0 2025-10-10T01:29:04.7782458Z  # via 2025-10-10T01:29:04.7782537Z # -c snapshot_constraint.txt 2025-10-10T01:29:04.7782628Z # hypothesis 2025-10-10T01:29:04.7782691Z soundfile==0.12.1 2025-10-10T01:29:04.7782771Z  # via 2025-10-10T01:29:04.7782846Z # -c snapshot_constraint.txt 2025-10-10T01:29:04.7782915Z # -r requirements/test.in 2025-10-10T01:29:04.7782971Z # librosa 2025-10-10T01:29:04.7783072Z # mistral-common 2025-10-10T01:29:04.7783131Z soxr==0.5.0.post1 2025-10-10T01:29:04.7783213Z  # via 2025-10-10T01:29:04.7783282Z # -c snapshot_constraint.txt 2025-10-10T01:29:04.7783338Z # librosa 2025-10-10T01:29:04.7783437Z # mistral-common 2025-10-10T01:29:04.7783498Z sqlalchemy==2.0.41 2025-10-10T01:29:04.7783574Z  # via 2025-10-10T01:29:04.7783649Z # -c snapshot_constraint.txt 2025-10-10T01:29:04.7783708Z # alembic 2025-10-10T01:29:04.7783796Z # mlflow 2025-10-10T01:29:04.7783857Z sqlitedict==2.1.0 2025-10-10T01:29:04.7783933Z  # via 2025-10-10T01:29:04.7784009Z # -c snapshot_constraint.txt 2025-10-10T01:29:04.7784094Z # lm-eval 2025-10-10T01:29:04.7784169Z sqlparse==0.5.3 2025-10-10T01:29:04.7784257Z  # via 2025-10-10T01:29:04.7784330Z # -c snapshot_constraint.txt 2025-10-10T01:29:04.7784433Z # mlflow-skinny 2025-10-10T01:29:04.7784497Z starlette==0.46.2 2025-10-10T01:29:04.7784574Z  # via 2025-10-10T01:29:04.7784651Z # -c snapshot_constraint.txt 2025-10-10T01:29:04.7784708Z # fastapi 2025-10-10T01:29:04.7784777Z # schemathesis 2025-10-10T01:29:04.7784885Z # starlette-testclient 2025-10-10T01:29:04.7784958Z starlette-testclient==0.4.1 2025-10-10T01:29:04.7785039Z  # via 2025-10-10T01:29:04.7785111Z # -c snapshot_constraint.txt 2025-10-10T01:29:04.7785205Z # schemathesis 2025-10-10T01:29:04.7785273Z statsmodels==0.14.4 2025-10-10T01:29:04.7785349Z  # via 2025-10-10T01:29:04.7785423Z # -c snapshot_constraint.txt 2025-10-10T01:29:04.7785513Z # genai-perf 2025-10-10T01:29:04.7785571Z sympy==1.13.3 2025-10-10T01:29:04.7785652Z  # via 2025-10-10T01:29:04.7785725Z # -c snapshot_constraint.txt 2025-10-10T01:29:04.7785880Z # einx 2025-10-10T01:29:04.7785963Z # torch 2025-10-10T01:29:04.7786025Z tabledata==1.3.3 2025-10-10T01:29:04.7786108Z  # via 2025-10-10T01:29:04.7786179Z # -c snapshot_constraint.txt 2025-10-10T01:29:04.7786276Z # pytablewriter 2025-10-10T01:29:04.7786344Z tabulate==0.9.0 2025-10-10T01:29:04.7786420Z  # via 2025-10-10T01:29:04.7786494Z # -c snapshot_constraint.txt 2025-10-10T01:29:04.7786583Z # sacrebleu 2025-10-10T01:29:04.7786641Z tblib==3.1.0 2025-10-10T01:29:04.7786724Z  # via 2025-10-10T01:29:04.7786793Z # -c snapshot_constraint.txt 2025-10-10T01:29:04.7786913Z # -r requirements/test.in 2025-10-10T01:29:04.7786974Z tcolorpy==0.1.6 2025-10-10T01:29:04.7787053Z  # via 2025-10-10T01:29:04.7787127Z # -c snapshot_constraint.txt 2025-10-10T01:29:04.7787222Z # pytablewriter 2025-10-10T01:29:04.7787354Z tenacity==9.0.0 2025-10-10T01:29:04.7787446Z  # via 2025-10-10T01:29:04.7787517Z # -c snapshot_constraint.txt 2025-10-10T01:29:04.7787576Z # lm-eval 2025-10-10T01:29:04.7787662Z # plotly 2025-10-10T01:29:04.7787724Z tensorboardx==2.6.4 2025-10-10T01:29:04.7787807Z  # via 2025-10-10T01:29:04.7787877Z # -c snapshot_constraint.txt 2025-10-10T01:29:04.7787970Z # lightning 2025-10-10T01:29:04.7788030Z tensorizer==2.10.1 2025-10-10T01:29:04.7788107Z  # via 2025-10-10T01:29:04.7788181Z # -c snapshot_constraint.txt 2025-10-10T01:29:04.7788285Z # -r requirements/test.in 2025-10-10T01:29:04.7788537Z terratorch @ git+https://github.com/IBM/terratorch.git@07184fcf91a1324f831ff521dd238d97fe350e3e 2025-10-10T01:29:04.7788663Z  # via -r requirements/test.in 2025-10-10T01:29:04.7788729Z threadpoolctl==3.5.0 2025-10-10T01:29:04.7788810Z  # via 2025-10-10T01:29:04.7788882Z # -c snapshot_constraint.txt 2025-10-10T01:29:04.7788988Z # scikit-learn 2025-10-10T01:29:04.7789062Z tifffile==2025.3.30 2025-10-10T01:29:04.7789139Z  # via 2025-10-10T01:29:04.7789213Z # -c snapshot_constraint.txt 2025-10-10T01:29:04.7789274Z # scikit-image 2025-10-10T01:29:04.7789363Z # terratorch 2025-10-10T01:29:04.7789429Z tiktoken==0.7.0 2025-10-10T01:29:04.7789506Z  # via 2025-10-10T01:29:04.7789581Z # -c snapshot_constraint.txt 2025-10-10T01:29:04.7789640Z # lm-eval 2025-10-10T01:29:04.7789734Z # mistral-common 2025-10-10T01:29:04.7789797Z timm==1.0.17 2025-10-10T01:29:04.7789877Z  # via 2025-10-10T01:29:04.7789947Z # -c snapshot_constraint.txt 2025-10-10T01:29:04.7790021Z # -r requirements/test.in 2025-10-10T01:29:04.7790097Z # open-clip-torch 2025-10-10T01:29:04.7790184Z # segmentation-models-pytorch 2025-10-10T01:29:04.7790243Z # terratorch 2025-10-10T01:29:04.7790330Z # torchgeo 2025-10-10T01:29:04.7790397Z tokenizers==0.22.0 2025-10-10T01:29:04.7790477Z  # via 2025-10-10T01:29:04.7790555Z # -c snapshot_constraint.txt 2025-10-10T01:29:04.7790623Z # -r requirements/test.in 2025-10-10T01:29:04.7790717Z # transformers 2025-10-10T01:29:04.7790782Z tomli==2.2.1 2025-10-10T01:29:04.7790859Z  # via 2025-10-10T01:29:04.7790931Z # -c snapshot_constraint.txt 2025-10-10T01:29:04.7791032Z # schemathesis 2025-10-10T01:29:04.7791092Z tomli-w==1.2.0 2025-10-10T01:29:04.7791174Z  # via 2025-10-10T01:29:04.7791242Z # -c snapshot_constraint.txt 2025-10-10T01:29:04.7791334Z # schemathesis 2025-10-10T01:29:04.7791819Z torch @ file:///var/lib/jenkins/workspace/dist/torch-2.10.0a0%2Bgit344e636-cp312-cp312-linux_x86_64.whl#sha256=61b81d5619a3dff1b95456c948bf99d131288c4916a1b2d71da5834655529eee 2025-10-10T01:29:04.7791899Z  # via 2025-10-10T01:29:04.7791974Z # -r requirements/test.in 2025-10-10T01:29:04.7792033Z # accelerate 2025-10-10T01:29:04.7792096Z # bitsandbytes 2025-10-10T01:29:04.7792257Z # efficientnet-pytorch 2025-10-10T01:29:04.7792313Z # encodec 2025-10-10T01:29:04.7792392Z # fastsafetensors 2025-10-10T01:29:04.7792447Z # kornia 2025-10-10T01:29:04.7792502Z # lightly 2025-10-10T01:29:04.7792564Z # lightning 2025-10-10T01:29:04.7792623Z # lm-eval 2025-10-10T01:29:04.7792679Z # mteb 2025-10-10T01:29:04.7792749Z # open-clip-torch 2025-10-10T01:29:04.7792807Z # peft 2025-10-10T01:29:04.7792886Z # pretrainedmodels 2025-10-10T01:29:04.7792956Z # pytorch-lightning 2025-10-10T01:29:04.7793027Z # runai-model-streamer 2025-10-10T01:29:04.7793111Z # segmentation-models-pytorch 2025-10-10T01:29:04.7793182Z # sentence-transformers 2025-10-10T01:29:04.7793244Z # tensorizer 2025-10-10T01:29:04.7793310Z # terratorch 2025-10-10T01:29:04.7793365Z # timm 2025-10-10T01:29:04.7793427Z # torchaudio 2025-10-10T01:29:04.7793484Z # torchgeo 2025-10-10T01:29:04.7793996Z # torchmetrics 2025-10-10T01:29:04.7794074Z # torchvision 2025-10-10T01:29:04.7794148Z # vector-quantize-pytorch 2025-10-10T01:29:04.7794244Z # vocos 2025-10-10T01:29:04.7794766Z torchaudio @ file:///var/lib/jenkins/workspace/dist/audio/torchaudio-2.8.0a0%2B87ff22e-cp312-cp312-linux_x86_64.whl#sha256=2eac501979e64f67c113a81572de033ce6ec78d775f6402cacf95aa651581299 2025-10-10T01:29:04.7794844Z  # via 2025-10-10T01:29:04.7794921Z # -r requirements/test.in 2025-10-10T01:29:04.7794977Z # encodec 2025-10-10T01:29:04.7795064Z # vocos 2025-10-10T01:29:04.7795122Z torchgeo==0.7.0 2025-10-10T01:29:04.7795220Z  # via terratorch 2025-10-10T01:29:04.7795289Z torchmetrics==1.8.2 2025-10-10T01:29:04.7795366Z  # via 2025-10-10T01:29:04.7795427Z # lightning 2025-10-10T01:29:04.7795502Z # pytorch-lightning 2025-10-10T01:29:04.7795560Z # terratorch 2025-10-10T01:29:04.7795651Z # torchgeo 2025-10-10T01:29:04.7796186Z torchvision @ file:///var/lib/jenkins/workspace/dist/vision/torchvision-0.22.0a0%2B966da7e-cp312-cp312-linux_x86_64.whl#sha256=79aa422fd81e11b575186dd520ce1848b0356118875ba3f51c957fae9e3f437f 2025-10-10T01:29:04.7796267Z  # via 2025-10-10T01:29:04.7796344Z # -r requirements/test.in 2025-10-10T01:29:04.7796400Z # lightly 2025-10-10T01:29:04.7796474Z # open-clip-torch 2025-10-10T01:29:04.7796540Z # pretrainedmodels 2025-10-10T01:29:04.7796617Z # segmentation-models-pytorch 2025-10-10T01:29:04.7796679Z # terratorch 2025-10-10T01:29:04.7796735Z # timm 2025-10-10T01:29:04.7796825Z # torchgeo 2025-10-10T01:29:04.7796882Z tqdm==4.66.6 2025-10-10T01:29:04.7796959Z  # via 2025-10-10T01:29:04.7797035Z # -c snapshot_constraint.txt 2025-10-10T01:29:04.7797092Z # datasets 2025-10-10T01:29:04.7797149Z # evaluate 2025-10-10T01:29:04.7797219Z # huggingface-hub 2025-10-10T01:29:04.7797274Z # lightly 2025-10-10T01:29:04.7797339Z # lightning 2025-10-10T01:29:04.7797396Z # lm-eval 2025-10-10T01:29:04.7797452Z # mteb 2025-10-10T01:29:04.7797522Z # nltk 2025-10-10T01:29:04.7797590Z # open-clip-torch 2025-10-10T01:29:04.7797647Z # peft 2025-10-10T01:29:04.7797719Z # pqdm 2025-10-10T01:29:04.7797786Z # pretrainedmodels 2025-10-10T01:29:04.7797860Z # pytorch-lightning 2025-10-10T01:29:04.7797937Z # segmentation-models-pytorch 2025-10-10T01:29:04.7798007Z # sentence-transformers 2025-10-10T01:29:04.7798078Z # tqdm-multiprocess 2025-10-10T01:29:04.7798172Z # transformers 2025-10-10T01:29:04.7798245Z tqdm-multiprocess==0.0.11 2025-10-10T01:29:04.7798324Z  # via 2025-10-10T01:29:04.7798395Z # -c snapshot_constraint.txt 2025-10-10T01:29:04.7798527Z # lm-eval 2025-10-10T01:29:04.7798593Z transformers==4.56.2 2025-10-10T01:29:04.7798672Z  # via 2025-10-10T01:29:04.7798749Z # -c snapshot_constraint.txt 2025-10-10T01:29:04.7798821Z # -r requirements/test.in 2025-10-10T01:29:04.7798972Z # genai-perf 2025-10-10T01:29:04.7799029Z # lm-eval 2025-10-10T01:29:04.7799086Z # peft 2025-10-10T01:29:04.7799165Z # sentence-transformers 2025-10-10T01:29:04.7799296Z # transformers-stream-generator 2025-10-10T01:29:04.7799385Z transformers-stream-generator==0.0.5 2025-10-10T01:29:04.7799463Z  # via 2025-10-10T01:29:04.7799535Z # -c snapshot_constraint.txt 2025-10-10T01:29:04.7799645Z # -r requirements/test.in 2025-10-10T01:29:04.7799708Z tritonclient==2.51.0 2025-10-10T01:29:04.7799791Z  # via 2025-10-10T01:29:04.7799862Z # -c snapshot_constraint.txt 2025-10-10T01:29:04.7799932Z # -r requirements/test.in 2025-10-10T01:29:04.7800028Z # genai-perf 2025-10-10T01:29:04.7800089Z typepy==1.3.2 2025-10-10T01:29:04.7800168Z  # via 2025-10-10T01:29:04.7800246Z # -c snapshot_constraint.txt 2025-10-10T01:29:04.7800311Z # dataproperty 2025-10-10T01:29:04.7800541Z # pytablewriter 2025-10-10T01:29:04.7800634Z # tabledata 2025-10-10T01:29:04.7800694Z typer==0.15.2 2025-10-10T01:29:04.7800791Z  # via 2025-10-10T01:29:04.7800865Z # -c snapshot_constraint.txt 2025-10-10T01:29:04.7800968Z # fastsafetensors 2025-10-10T01:29:04.7801055Z types-python-dateutil==2.9.0.20241206 2025-10-10T01:29:04.7801134Z  # via 2025-10-10T01:29:04.7801217Z # -c snapshot_constraint.txt 2025-10-10T01:29:04.7801299Z # arrow 2025-10-10T01:29:04.7801368Z typeshed-client==2.8.2 2025-10-10T01:29:04.7801453Z  # via 2025-10-10T01:29:04.7801523Z # -c snapshot_constraint.txt 2025-10-10T01:29:04.7801622Z # jsonargparse 2025-10-10T01:29:04.7801692Z typing-extensions==4.12.2 2025-10-10T01:29:04.7801767Z  # via 2025-10-10T01:29:04.7801844Z # -c snapshot_constraint.txt 2025-10-10T01:29:04.7801906Z # albumentations 2025-10-10T01:29:04.7801968Z # alembic 2025-10-10T01:29:04.7802028Z # fastapi 2025-10-10T01:29:04.7802089Z # graphene 2025-10-10T01:29:04.7802167Z # huggingface-hub 2025-10-10T01:29:04.7802226Z # librosa 2025-10-10T01:29:04.7802283Z # lightning 2025-10-10T01:29:04.7802361Z # lightning-utilities 2025-10-10T01:29:04.7802422Z # mistral-common 2025-10-10T01:29:04.7802488Z # mlflow-skinny 2025-10-10T01:29:04.7802544Z # mteb 2025-10-10T01:29:04.7802613Z # opentelemetry-api 2025-10-10T01:29:04.7802682Z # opentelemetry-sdk 2025-10-10T01:29:04.7802775Z # opentelemetry-semantic-conventions 2025-10-10T01:29:04.7802836Z # pqdm 2025-10-10T01:29:04.7802894Z # pydantic 2025-10-10T01:29:04.7802958Z # pydantic-core 2025-10-10T01:29:04.7803034Z # pydantic-extra-types 2025-10-10T01:29:04.7803099Z # pytorch-lightning 2025-10-10T01:29:04.7803156Z # sqlalchemy 2025-10-10T01:29:04.7803226Z # torch 2025-10-10T01:29:04.7803283Z # torchgeo 2025-10-10T01:29:04.7803347Z # typer 2025-10-10T01:29:04.7803413Z # typeshed-client 2025-10-10T01:29:04.7803520Z # typing-inspection 2025-10-10T01:29:04.7803595Z typing-inspection==0.4.1 2025-10-10T01:29:04.7803670Z  # via 2025-10-10T01:29:04.7803742Z # -c snapshot_constraint.txt 2025-10-10T01:29:04.7803835Z # pydantic 2025-10-10T01:29:04.7803894Z tzdata==2024.2 2025-10-10T01:29:04.7803975Z  # via 2025-10-10T01:29:04.7804044Z # -c snapshot_constraint.txt 2025-10-10T01:29:04.7804129Z # pandas 2025-10-10T01:29:04.7804201Z uri-template==1.3.0 2025-10-10T01:29:04.7804278Z  # via 2025-10-10T01:29:04.7804355Z # -c snapshot_constraint.txt 2025-10-10T01:29:04.7804444Z # jsonschema 2025-10-10T01:29:04.7804504Z urllib3==2.2.3 2025-10-10T01:29:04.7804588Z  # via 2025-10-10T01:29:04.7804658Z # -c snapshot_constraint.txt 2025-10-10T01:29:04.7804725Z # blobfile 2025-10-10T01:29:04.7804788Z # botocore 2025-10-10T01:29:04.7804846Z # docker 2025-10-10T01:29:04.7804986Z # lightly 2025-10-10T01:29:04.7805041Z # requests 2025-10-10T01:29:04.7805099Z # responses 2025-10-10T01:29:04.7805200Z # tritonclient 2025-10-10T01:29:04.7805259Z uvicorn==0.35.0 2025-10-10T01:29:04.7805341Z  # via 2025-10-10T01:29:04.7805412Z # -c snapshot_constraint.txt 2025-10-10T01:29:04.7805507Z # mlflow-skinny 2025-10-10T01:29:04.7805590Z vector-quantize-pytorch==1.23.2 2025-10-10T01:29:04.7805713Z  # via -r requirements/test.in 2025-10-10T01:29:04.7805776Z virtualenv==20.31.2 2025-10-10T01:29:04.7805857Z  # via 2025-10-10T01:29:04.7805928Z # -c snapshot_constraint.txt 2025-10-10T01:29:04.7806011Z # ray 2025-10-10T01:29:04.7806068Z vocos==0.1.0 2025-10-10T01:29:04.7806146Z  # via 2025-10-10T01:29:04.7806221Z # -c snapshot_constraint.txt 2025-10-10T01:29:04.7806325Z # -r requirements/test.in 2025-10-10T01:29:04.7806391Z wcwidth==0.2.13 2025-10-10T01:29:04.7806538Z  # via 2025-10-10T01:29:04.7806613Z # -c snapshot_constraint.txt 2025-10-10T01:29:04.7806700Z # ftfy 2025-10-10T01:29:04.7806763Z webcolors==24.11.1 2025-10-10T01:29:04.7806840Z  # via 2025-10-10T01:29:04.7806918Z # -c snapshot_constraint.txt 2025-10-10T01:29:04.7807008Z # jsonschema 2025-10-10T01:29:04.7807072Z werkzeug==3.1.3 2025-10-10T01:29:04.7807149Z  # via 2025-10-10T01:29:04.7807221Z # -c snapshot_constraint.txt 2025-10-10T01:29:04.7807282Z # flask 2025-10-10T01:29:04.7807374Z # schemathesis 2025-10-10T01:29:04.7807440Z word2number==1.1 2025-10-10T01:29:04.7807520Z  # via 2025-10-10T01:29:04.7807594Z # -c snapshot_constraint.txt 2025-10-10T01:29:04.7807685Z # lm-eval 2025-10-10T01:29:04.7807745Z wrapt==1.17.2 2025-10-10T01:29:04.7807822Z  # via 2025-10-10T01:29:04.7807899Z # -c snapshot_constraint.txt 2025-10-10T01:29:04.7807988Z # smart-open 2025-10-10T01:29:04.7808057Z xarray==2025.7.1 2025-10-10T01:29:04.7808138Z  # via 2025-10-10T01:29:04.7808209Z # -c snapshot_constraint.txt 2025-10-10T01:29:04.7808303Z # rioxarray 2025-10-10T01:29:04.7808365Z xxhash==3.5.0 2025-10-10T01:29:04.7808453Z  # via 2025-10-10T01:29:04.7808530Z # -c snapshot_constraint.txt 2025-10-10T01:29:04.7808589Z # datasets 2025-10-10T01:29:04.7808684Z # evaluate 2025-10-10T01:29:04.7808744Z yarl==1.17.1 2025-10-10T01:29:04.7808823Z  # via 2025-10-10T01:29:04.7808902Z # -c snapshot_constraint.txt 2025-10-10T01:29:04.7808961Z # aiohttp 2025-10-10T01:29:04.7809058Z # schemathesis 2025-10-10T01:29:04.7809120Z zipp==3.23.0 2025-10-10T01:29:04.7809206Z  # via 2025-10-10T01:29:04.7809285Z # -c snapshot_constraint.txt 2025-10-10T01:29:04.7809387Z # importlib-metadata 2025-10-10T01:29:04.7809454Z zstandard==0.23.0 2025-10-10T01:29:04.7809531Z  # via 2025-10-10T01:29:04.7809603Z # -c snapshot_constraint.txt 2025-10-10T01:29:04.7809694Z # lm-eval 2025-10-10T01:29:04.8986933Z 2025-10-10 01:29:04,898 [INFO] cli.lib.common.pip_helper: pip installing packages: /opt/conda/envs/py_3.12/bin/python -m uv pip install -r test.txt 2025-10-10T01:29:04.8987843Z 2025-10-10 01:29:04,898 [INFO] cli.lib.common.utils: [cmd] /opt/conda/envs/py_3.12/bin/python -m uv pip install -r test.txt 2025-10-10T01:29:04.9490124Z Using Python 3.12.11 environment at: /opt/conda/envs/py_3.12 2025-10-10T01:29:05.0901144Z Resolved 324 packages in 135ms 2025-10-10T01:29:05.1245480Z Updating https://github.com/IBM/terratorch.git (07184fcf91a1324f831ff521dd238d97fe350e3e) 2025-10-10T01:29:05.1246519Z Updating https://github.com/EleutherAI/lm-evaluation-harness.git (206b7722158f58c35b7ffcd53b035fdbdda5126d) 2025-10-10T01:29:05.2349762Z Downloading plotly (18.2MiB) 2025-10-10T01:29:05.2363484Z Downloading py-spy (2.6MiB) 2025-10-10T01:29:05.2364162Z Downloading networkx (1.6MiB) 2025-10-10T01:29:05.2453075Z Downloading pandas (12.1MiB) 2025-10-10T01:29:05.2455251Z Downloading grpcio (5.6MiB) 2025-10-10T01:29:05.2458341Z Downloading hf-xet (3.0MiB) 2025-10-10T01:29:05.2459280Z Downloading pydantic-core (1.9MiB) 2025-10-10T01:29:05.2462427Z Downloading fastparquet (1.7MiB) 2025-10-10T01:29:05.2626160Z Downloading transformers (11.1MiB) 2025-10-10T01:29:05.2627665Z Downloading pillow (4.3MiB) 2025-10-10T01:29:05.2630775Z Downloading virtualenv (5.8MiB) 2025-10-10T01:29:05.2632060Z Downloading opencv-python-headless (47.7MiB) 2025-10-10T01:29:05.2634816Z Downloading zstandard (5.2MiB) 2025-10-10T01:29:05.2637192Z Downloading ray (66.9MiB) 2025-10-10T01:29:05.2639730Z Downloading rasterio (21.2MiB) 2025-10-10T01:29:05.2641442Z Downloading rapidfuzz (3.0MiB) 2025-10-10T01:29:05.2644491Z Downloading numpy (17.1MiB) 2025-10-10T01:29:05.2647262Z Downloading shapely (3.0MiB) 2025-10-10T01:29:05.2649754Z Downloading sqlalchemy (3.2MiB) 2025-10-10T01:29:05.2652093Z Downloading scikit-image (14.3MiB) 2025-10-10T01:29:05.2653861Z Downloading mteb (1.5MiB) 2025-10-10T01:29:05.2656334Z Downloading pyogrio (26.4MiB) 2025-10-10T01:29:05.2657998Z Downloading python-box (4.1MiB) 2025-10-10T01:29:05.2659523Z Downloading mlflow-skinny (6.0MiB) 2025-10-10T01:29:05.2661658Z Downloading botocore (12.1MiB) 2025-10-10T01:29:05.2663804Z Downloading fiona (16.4MiB) 2025-10-10T01:29:05.2664756Z Downloading scipy (36.4MiB) 2025-10-10T01:29:05.2667337Z Downloading mistral-common (6.2MiB) 2025-10-10T01:29:05.2668572Z Downloading timm (2.4MiB) 2025-10-10T01:29:05.2671073Z Downloading matplotlib (7.9MiB) 2025-10-10T01:29:05.2672349Z Downloading scikit-learn (12.3MiB) 2025-10-10T01:29:05.2674441Z Downloading h5py (4.7MiB) 2025-10-10T01:29:05.2676437Z Downloading bitsandbytes (69.5MiB) 2025-10-10T01:29:05.2678120Z Downloading pyproj (9.1MiB) 2025-10-10T01:29:05.2680388Z Downloading statsmodels (10.2MiB) 2025-10-10T01:29:05.2681762Z Downloading kornia-rs (2.6MiB) 2025-10-10T01:29:05.2684951Z Downloading fonttools (4.7MiB) 2025-10-10T01:29:05.2685517Z Downloading tritonclient (13.3MiB) 2025-10-10T01:29:05.2687993Z Downloading mlflow (27.7MiB) 2025-10-10T01:29:05.2690430Z Downloading cramjam (2.2MiB) 2025-10-10T01:29:05.2692244Z Downloading polars (33.2MiB) 2025-10-10T01:29:05.2700438Z Downloading pycryptodomex (2.2MiB) 2025-10-10T01:29:05.2702460Z Downloading python-rapidjson (1.6MiB) 2025-10-10T01:29:05.2704526Z Downloading black (1.7MiB) 2025-10-10T01:29:05.2706453Z Downloading runai-model-streamer-gcs (4.2MiB) 2025-10-10T01:29:05.2708315Z Downloading decord (13.0MiB) 2025-10-10T01:29:05.2710309Z Downloading runai-model-streamer-s3 (5.6MiB) 2025-10-10T01:29:05.2712214Z Downloading kaleido (76.2MiB) 2025-10-10T01:29:05.2714486Z Downloading tokenizers (3.2MiB) 2025-10-10T01:29:05.2716820Z Downloading pyarrow (38.2MiB) 2025-10-10T01:29:06.0307676Z Downloading python-rapidjson 2025-10-10T01:29:06.0554202Z Downloading open-clip-torch (1.5MiB) 2025-10-10T01:29:06.0614352Z Downloading fastparquet 2025-10-10T01:29:06.0772051Z Downloading nltk (1.4MiB) 2025-10-10T01:29:06.1520965Z Downloading pydantic-core 2025-10-10T01:29:06.1685820Z Downloading kiwisolver (1.4MiB) 2025-10-10T01:29:06.1936586Z Downloading black 2025-10-10T01:29:06.2051889Z Downloading fastsafetensors (1.4MiB) 2025-10-10T01:29:06.2569264Z Downloading cramjam 2025-10-10T01:29:06.2840409Z Downloading xarray (1.3MiB) 2025-10-10T01:29:06.3009298Z Downloading pycryptodomex 2025-10-10T01:29:06.3166671Z Downloading aiohttp (1.2MiB) 2025-10-10T01:29:06.4096403Z Downloading py-spy 2025-10-10T01:29:06.4225802Z Downloading sentencepiece (1.2MiB) 2025-10-10T01:29:06.4847133Z Downloading kornia-rs 2025-10-10T01:29:06.5054390Z Downloading setuptools (1.2MiB) 2025-10-10T01:29:06.5444937Z Downloading timm 2025-10-10T01:29:06.5582148Z Downloading pygments (1.1MiB) 2025-10-10T01:29:06.5652840Z Downloading networkx 2025-10-10T01:29:06.5954250Z Downloading soundfile (1.1MiB) 2025-10-10T01:29:06.6205299Z Downloading rapidfuzz 2025-10-10T01:29:06.6379808Z Downloading tiktoken (1.0MiB) 2025-10-10T01:29:06.6452101Z Downloading hf-xet 2025-10-10T01:29:06.6790771Z Downloading kornia (1.0MiB) 2025-10-10T01:29:06.7492877Z Downloading sqlalchemy 2025-10-10T01:29:06.7560244Z Downloading tokenizers 2025-10-10T01:29:06.7880069Z Downloading shapely 2025-10-10T01:29:06.7937958Z Downloading kiwisolver 2025-10-10T01:29:06.8210097Z Downloading fastsafetensors 2025-10-10T01:29:06.8490070Z Downloading aiohttp 2025-10-10T01:29:06.8699281Z Downloading xarray 2025-10-10T01:29:06.9551574Z Downloading open-clip-torch 2025-10-10T01:29:06.9866679Z Downloading nltk 2025-10-10T01:29:07.0848837Z Downloading sentencepiece 2025-10-10T01:29:07.1498880Z Downloading soundfile 2025-10-10T01:29:07.1511547Z Downloading tiktoken 2025-10-10T01:29:07.1521463Z Downloading python-box 2025-10-10T01:29:07.1674092Z Downloading mteb 2025-10-10T01:29:07.2099058Z Downloading runai-model-streamer-gcs 2025-10-10T01:29:07.2327486Z Downloading setuptools 2025-10-10T01:29:07.2655655Z Downloading pillow 2025-10-10T01:29:07.3013462Z Downloading pygments 2025-10-10T01:29:07.3397449Z Downloading fonttools 2025-10-10T01:29:07.4065884Z Downloading kornia 2025-10-10T01:29:07.4744772Z Downloading h5py 2025-10-10T01:29:07.6437908Z Downloading zstandard 2025-10-10T01:29:07.8744636Z Downloading grpcio 2025-10-10T01:29:07.9956875Z Downloading virtualenv 2025-10-10T01:29:07.9969303Z Downloading runai-model-streamer-s3 2025-10-10T01:29:08.1688369Z Downloading mistral-common 2025-10-10T01:29:08.5388842Z Building encodec==0.1.1 2025-10-10T01:29:08.5413445Z Building pretrainedmodels==0.7.4 2025-10-10T01:29:08.5414047Z Building antlr4-python3-runtime==4.9.3 2025-10-10T01:29:08.5416257Z Building docopt==0.6.2 2025-10-10T01:29:08.5419180Z Building sqlitedict==2.1.0 2025-10-10T01:29:08.5465387Z Building efficientnet-pytorch==0.7.1 2025-10-10T01:29:08.5480686Z Building rouge-score==0.1.2 2025-10-10T01:29:08.5549683Z Building word2number==1.1 2025-10-10T01:29:08.5550833Z Building transformers-stream-generator==0.0.5 2025-10-10T01:29:08.6548311Z Updated https://github.com/IBM/terratorch.git (07184fcf91a1324f831ff521dd238d97fe350e3e) 2025-10-10T01:29:08.6566005Z Building terratorch @ git+https://github.com/IBM/terratorch.git@07184fcf91a1324f831ff521dd238d97fe350e3e 2025-10-10T01:29:08.7082476Z Downloading mlflow-skinny 2025-10-10T01:29:08.7847726Z Downloading matplotlib 2025-10-10T01:29:08.9100412Z Downloading pyproj 2025-10-10T01:29:09.0838407Z Built efficientnet-pytorch==0.7.1 2025-10-10T01:29:09.0839056Z Built docopt==0.6.2 2025-10-10T01:29:09.0898478Z Built encodec==0.1.1 2025-10-10T01:29:09.1002958Z Built sqlitedict==2.1.0 2025-10-10T01:29:09.1084430Z Built rouge-score==0.1.2 2025-10-10T01:29:09.1098533Z Built pretrainedmodels==0.7.4 2025-10-10T01:29:09.1100128Z Built word2number==1.1 2025-10-10T01:29:09.1139207Z Built transformers-stream-generator==0.0.5 2025-10-10T01:29:09.1353066Z Built antlr4-python3-runtime==4.9.3 2025-10-10T01:29:09.1792689Z Downloading statsmodels 2025-10-10T01:29:09.3051337Z Updated https://github.com/EleutherAI/lm-evaluation-harness.git (206b7722158f58c35b7ffcd53b035fdbdda5126d) 2025-10-10T01:29:09.3067276Z Building lm-eval @ git+https://github.com/EleutherAI/lm-evaluation-harness.git@206b7722158f58c35b7ffcd53b035fdbdda5126d 2025-10-10T01:29:09.4472886Z Downloading pandas 2025-10-10T01:29:09.5548531Z Downloading decord 2025-10-10T01:29:09.5646785Z Built terratorch @ git+https://github.com/IBM/terratorch.git@07184fcf91a1324f831ff521dd238d97fe350e3e 2025-10-10T01:29:09.5910654Z Downloading scikit-learn 2025-10-10T01:29:09.6752169Z Downloading tritonclient 2025-10-10T01:29:09.7134414Z Downloading botocore 2025-10-10T01:29:09.7605739Z Downloading transformers 2025-10-10T01:29:09.7996323Z Downloading scikit-image 2025-10-10T01:29:09.9612160Z Downloading fiona 2025-10-10T01:29:09.9656077Z Downloading numpy 2025-10-10T01:29:10.4168690Z Downloading rasterio 2025-10-10T01:29:10.5780845Z Downloading pyogrio 2025-10-10T01:29:10.9305753Z Downloading polars 2025-10-10T01:29:10.9726140Z Downloading mlflow 2025-10-10T01:29:11.6398330Z Downloading opencv-python-headless 2025-10-10T01:29:11.7199013Z Downloading scipy 2025-10-10T01:29:11.9137210Z Downloading pyarrow 2025-10-10T01:29:12.1933437Z Downloading kaleido 2025-10-10T01:29:12.2432984Z Downloading bitsandbytes 2025-10-10T01:29:12.3379664Z Downloading ray 2025-10-10T01:29:14.5540804Z Downloading plotly 2025-10-10T01:29:23.8363297Z Built lm-eval @ git+https://github.com/EleutherAI/lm-evaluation-harness.git@206b7722158f58c35b7ffcd53b035fdbdda5126d 2025-10-10T01:29:24.0777149Z Prepared 289 packages in 18.97s 2025-10-10T01:29:28.6412947Z Uninstalled 98 packages in 4.56s 2025-10-10T01:29:29.5508037Z Installed 289 packages in 909ms 2025-10-10T01:29:29.5911053Z - absl-py==2.3.1 2025-10-10T01:29:29.5916979Z + absl-py==2.1.0 2025-10-10T01:29:29.5918002Z + accelerate==1.0.1 2025-10-10T01:29:29.5918833Z + aenum==3.1.16 2025-10-10T01:29:29.5919479Z + affine==2.4.0 2025-10-10T01:29:29.5920166Z - aiohappyeyeballs==2.6.1 2025-10-10T01:29:29.5920923Z + aiohappyeyeballs==2.4.3 2025-10-10T01:29:29.5921607Z - aiohttp==3.13.0 2025-10-10T01:29:29.5922235Z + aiohttp==3.10.11 2025-10-10T01:29:29.5922896Z + aiohttp-cors==0.8.1 2025-10-10T01:29:29.5923591Z - aiosignal==1.4.0 2025-10-10T01:29:29.5924475Z + aiosignal==1.3.1 2025-10-10T01:29:29.5924793Z + albucore==0.0.16 2025-10-10T01:29:29.5925080Z + albumentations==1.4.6 2025-10-10T01:29:29.5925371Z + alembic==1.16.4 2025-10-10T01:29:29.5926236Z + antlr4-python3-runtime==4.9.3 2025-10-10T01:29:29.5927429Z - anyio==4.11.0 2025-10-10T01:29:29.5927804Z + anyio==4.6.2.post1 2025-10-10T01:29:29.5929117Z + argcomplete==3.5.1 2025-10-10T01:29:29.5929411Z + arrow==1.3.0 2025-10-10T01:29:29.5930277Z - attrs==25.4.0 2025-10-10T01:29:29.5930989Z + attrs==24.2.0 2025-10-10T01:29:29.5931699Z + backoff==2.2.1 2025-10-10T01:29:29.5932598Z + bitsandbytes==0.46.1 2025-10-10T01:29:29.5933553Z + black==24.10.0 2025-10-10T01:29:29.5934312Z + blinker==1.9.0 2025-10-10T01:29:29.5935044Z + blobfile==3.0.0 2025-10-10T01:29:29.5935796Z + bm25s==0.2.13 2025-10-10T01:29:29.5936484Z - boto3==1.35.42 2025-10-10T01:29:29.5937267Z + boto3==1.35.57 2025-10-10T01:29:29.5938580Z - botocore==1.35.99 2025-10-10T01:29:29.5939025Z + botocore==1.35.57 2025-10-10T01:29:29.5939852Z + bounded-pool-executor==0.0.3 2025-10-10T01:29:29.5940505Z + buildkite-test-collector==0.1.9 2025-10-10T01:29:29.5941897Z - cachetools==6.2.0 2025-10-10T01:29:29.5942371Z + cachetools==5.5.2 2025-10-10T01:29:29.5942763Z - certifi==2025.10.5 2025-10-10T01:29:29.5943674Z + certifi==2024.8.30 2025-10-10T01:29:29.5944283Z - cffi==2.0.0 2025-10-10T01:29:29.5945140Z + cffi==1.17.1 2025-10-10T01:29:29.5945952Z + chardet==5.2.0 2025-10-10T01:29:29.5946696Z - charset-normalizer==3.4.3 2025-10-10T01:29:29.5948135Z + charset-normalizer==3.4.0 2025-10-10T01:29:29.5948448Z - click==8.3.0 2025-10-10T01:29:29.5949165Z + click==8.1.7 2025-10-10T01:29:29.5950044Z + click-plugins==1.1.1.2 2025-10-10T01:29:29.5950906Z + cligj==0.7.2 2025-10-10T01:29:29.5951654Z + colorful==0.5.6 2025-10-10T01:29:29.5952480Z + contourpy==1.3.0 2025-10-10T01:29:29.5953189Z + coverage==7.10.6 2025-10-10T01:29:29.5954627Z + cramjam==2.9.0 2025-10-10T01:29:29.5955217Z + cycler==0.12.1 2025-10-10T01:29:29.5955753Z + databricks-sdk==0.59.0 2025-10-10T01:29:29.5956276Z + datamodel-code-generator==0.26.3 2025-10-10T01:29:29.5957553Z + dataproperty==1.0.1 2025-10-10T01:29:29.5958279Z + datasets==3.0.2 2025-10-10T01:29:29.5958765Z - decorator==5.2.1 2025-10-10T01:29:29.5959573Z + decorator==5.1.1 2025-10-10T01:29:29.5960536Z + decord==0.6.0 2025-10-10T01:29:29.5961228Z - dill==0.3.7 2025-10-10T01:29:29.5962231Z + dill==0.3.8 2025-10-10T01:29:29.5962911Z + distlib==0.3.9 2025-10-10T01:29:29.5963721Z - dnspython==2.8.0 2025-10-10T01:29:29.5964338Z + dnspython==2.7.0 2025-10-10T01:29:29.5965836Z + docopt==0.6.2 2025-10-10T01:29:29.5966574Z + docstring-parser==0.17.0 2025-10-10T01:29:29.5967023Z + efficientnet-pytorch==0.7.1 2025-10-10T01:29:29.5967745Z + einx==0.3.0 2025-10-10T01:29:29.5969031Z - email-validator==2.3.0 2025-10-10T01:29:29.5969450Z + email-validator==2.2.0 2025-10-10T01:29:29.5971018Z + encodec==0.1.1 2025-10-10T01:29:29.5971346Z + eval-type-backport==0.2.2 2025-10-10T01:29:29.5971648Z + evaluate==0.4.3 2025-10-10T01:29:29.5973336Z - fastapi==0.118.2 2025-10-10T01:29:29.5973821Z + fastapi==0.116.1 2025-10-10T01:29:29.5974316Z + fastparquet==2024.11.0 2025-10-10T01:29:29.5975007Z - fastrlock==0.8.3 2025-10-10T01:29:29.5975629Z + fastrlock==0.8.2 2025-10-10T01:29:29.5976654Z + fastsafetensors==0.1.10 2025-10-10T01:29:29.5977288Z - filelock==3.18.0 2025-10-10T01:29:29.5977832Z + filelock==3.16.1 2025-10-10T01:29:29.5979087Z + fiona==1.10.1 2025-10-10T01:29:29.5979978Z + flask==3.1.1 2025-10-10T01:29:29.5980330Z + fonttools==4.55.0 2025-10-10T01:29:29.5981567Z + fqdn==1.5.1 2025-10-10T01:29:29.5981987Z - frozenlist==1.8.0 2025-10-10T01:29:29.5982754Z + frozenlist==1.5.0 2025-10-10T01:29:29.5984007Z - fsspec==2025.9.0 2025-10-10T01:29:29.5984373Z + fsspec==2024.9.0 2025-10-10T01:29:29.5985568Z + ftfy==6.3.1 2025-10-10T01:29:29.5986055Z + genai-perf==0.0.8 2025-10-10T01:29:29.5987226Z + genson==1.3.0 2025-10-10T01:29:29.5987685Z + geopandas==1.0.1 2025-10-10T01:29:29.5988455Z - gitpython==3.1.45 2025-10-10T01:29:29.5989200Z + gitpython==3.1.44 2025-10-10T01:29:29.5990070Z + google-api-core==2.24.2 2025-10-10T01:29:29.5990853Z - google-auth==2.41.1 2025-10-10T01:29:29.5991246Z + google-auth==2.40.2 2025-10-10T01:29:29.5992247Z + google-cloud-core==2.4.3 2025-10-10T01:29:29.5992938Z + google-cloud-storage==3.4.0 2025-10-10T01:29:29.5993948Z + google-crc32c==1.7.1 2025-10-10T01:29:29.5994677Z + google-resumable-media==2.7.2 2025-10-10T01:29:29.5995124Z + googleapis-common-protos==1.70.0 2025-10-10T01:29:29.5996104Z + graphene==3.4.3 2025-10-10T01:29:29.5997044Z + graphql-core==3.2.6 2025-10-10T01:29:29.5997696Z + graphql-relay==3.2.0 2025-10-10T01:29:29.5998610Z + greenlet==3.2.3 2025-10-10T01:29:29.5999055Z - grpcio==1.75.1 2025-10-10T01:29:29.6000422Z + grpcio==1.71.0 2025-10-10T01:29:29.6000816Z + gunicorn==23.0.0 2025-10-10T01:29:29.6001585Z - h11==0.16.0 2025-10-10T01:29:29.6002206Z + h11==0.14.0 2025-10-10T01:29:29.6003158Z + h5py==3.13.0 2025-10-10T01:29:29.6004023Z + harfile==0.3.0 2025-10-10T01:29:29.6004606Z - hf-xet==1.1.10 2025-10-10T01:29:29.6005270Z + hf-xet==1.1.7 2025-10-10T01:29:29.6006062Z + hiredis==3.0.0 2025-10-10T01:29:29.6006915Z - httpcore==1.0.9 2025-10-10T01:29:29.6007827Z + httpcore==1.0.6 2025-10-10T01:29:29.6008133Z - httpx==0.28.1 2025-10-10T01:29:29.6009726Z + httpx==0.27.2 2025-10-10T01:29:29.6010153Z - huggingface-hub==0.35.3 2025-10-10T01:29:29.6010875Z + huggingface-hub==0.34.3 2025-10-10T01:29:29.6011433Z + humanize==4.11.0 2025-10-10T01:29:29.6012105Z + hydra-core==1.3.2 2025-10-10T01:29:29.6013462Z - hypothesis==6.56.4 2025-10-10T01:29:29.6013784Z + hypothesis==6.131.0 2025-10-10T01:29:29.6014771Z + hypothesis-graphql==0.11.1 2025-10-10T01:29:29.6015295Z + hypothesis-jsonschema==0.23.1 2025-10-10T01:29:29.6016176Z + importlib-metadata==8.7.0 2025-10-10T01:29:29.6017126Z + importlib-resources==6.5.2 2025-10-10T01:29:29.6017686Z + inflect==5.6.2 2025-10-10T01:29:29.6018467Z - iniconfig==2.1.0 2025-10-10T01:29:29.6018932Z + iniconfig==2.0.0 2025-10-10T01:29:29.6020295Z + isoduration==20.11.0 2025-10-10T01:29:29.6020626Z + isort==5.13.2 2025-10-10T01:29:29.6021476Z + itsdangerous==2.2.0 2025-10-10T01:29:29.6022266Z + jiwer==3.0.5 2025-10-10T01:29:29.6022924Z - joblib==1.5.2 2025-10-10T01:29:29.6024495Z + joblib==1.4.2 2025-10-10T01:29:29.6024808Z + jsonargparse==4.35.0 2025-10-10T01:29:29.6025607Z + jsonlines==4.0.0 2025-10-10T01:29:29.6026254Z + jsonpointer==3.0.0 2025-10-10T01:29:29.6026607Z - jsonschema==4.25.1 2025-10-10T01:29:29.6028076Z + jsonschema==4.23.0 2025-10-10T01:29:29.6028838Z - jsonschema-specifications==2025.9.1 2025-10-10T01:29:29.6029671Z + jsonschema-specifications==2024.10.1 2025-10-10T01:29:29.6030088Z + junit-xml==1.9 2025-10-10T01:29:29.6030836Z + kaleido==0.2.1 2025-10-10T01:29:29.6031364Z + kiwisolver==1.4.7 2025-10-10T01:29:29.6032143Z + kornia==0.8.1 2025-10-10T01:29:29.6033371Z + kornia-rs==0.1.9 2025-10-10T01:29:29.6033750Z + libnacl==2.1.0 2025-10-10T01:29:29.6034562Z - librosa==0.10.2 2025-10-10T01:29:29.6035238Z + librosa==0.10.2.post1 2025-10-10T01:29:29.6036021Z + lightly==1.5.20 2025-10-10T01:29:29.6037417Z + lightly-utils==0.0.2 2025-10-10T01:29:29.6038165Z + lightning==2.5.1.post0 2025-10-10T01:29:29.6038935Z + lightning-utilities==0.14.3 2025-10-10T01:29:29.6040414Z + lm-eval==0.4.9.1 (from git+https://github.com/EleutherAI/lm-evaluation-harness.git@206b7722158f58c35b7ffcd53b035fdbdda5126d) 2025-10-10T01:29:29.6041004Z + mako==1.3.10 2025-10-10T01:29:29.6041395Z - markdown==3.9 2025-10-10T01:29:29.6042578Z + markdown==3.8.2 2025-10-10T01:29:29.6042965Z - markdown-it-py==4.0.0 2025-10-10T01:29:29.6043471Z + markdown-it-py==3.0.0 2025-10-10T01:29:29.6044245Z - markupsafe==3.0.3 2025-10-10T01:29:29.6045495Z + markupsafe==3.0.1 2025-10-10T01:29:29.6046031Z + matplotlib==3.9.2 2025-10-10T01:29:29.6046928Z + mbstrdecoder==1.1.3 2025-10-10T01:29:29.6047339Z - mistral-common==1.8.5 2025-10-10T01:29:29.6048894Z + mistral-common==1.8.2 2025-10-10T01:29:29.6049292Z + mlflow==2.22.0 2025-10-10T01:29:29.6049997Z + mlflow-skinny==2.22.0 2025-10-10T01:29:29.6051014Z + more-itertools==10.5.0 2025-10-10T01:29:29.6051579Z - msgpack==1.1.2 2025-10-10T01:29:29.6052431Z + msgpack==1.1.0 2025-10-10T01:29:29.6052978Z + mteb==1.38.11 2025-10-10T01:29:29.6053898Z - multidict==6.7.0 2025-10-10T01:29:29.6054622Z + multidict==6.1.0 2025-10-10T01:29:29.6055422Z + multiprocess==0.70.16 2025-10-10T01:29:29.6056290Z + munch==4.0.0 2025-10-10T01:29:29.6057254Z - mypy-extensions==1.1.0 2025-10-10T01:29:29.6057871Z + mypy-extensions==1.0.0 2025-10-10T01:29:29.6058424Z - networkx==2.8.8 2025-10-10T01:29:29.6059755Z + networkx==3.2.1 2025-10-10T01:29:29.6060145Z + nltk==3.9.1 2025-10-10T01:29:29.6060608Z + num2words==0.5.14 2025-10-10T01:29:29.6061898Z + numexpr==2.10.1 2025-10-10T01:29:29.6062311Z - numpy==2.2.6 2025-10-10T01:29:29.6063163Z + numpy==1.26.4 2025-10-10T01:29:29.6063624Z + omegaconf==2.3.0 2025-10-10T01:29:29.6065043Z + open-clip-torch==2.32.0 2025-10-10T01:29:29.6065431Z + opencensus==0.11.4 2025-10-10T01:29:29.6066120Z + opencensus-context==0.1.3 2025-10-10T01:29:29.6066859Z - opencv-python-headless==4.12.0.88 2025-10-10T01:29:29.6067935Z + opencv-python-headless==4.11.0.86 2025-10-10T01:29:29.6068741Z + opentelemetry-api==1.35.0 2025-10-10T01:29:29.6069357Z + opentelemetry-exporter-prometheus==0.56b0 2025-10-10T01:29:29.6070072Z + opentelemetry-proto==1.36.0 2025-10-10T01:29:29.6070894Z + opentelemetry-sdk==1.35.0 2025-10-10T01:29:29.6071609Z + opentelemetry-semantic-conventions==0.56b0 2025-10-10T01:29:29.6072237Z - packaging==25.0 2025-10-10T01:29:29.6073522Z + packaging==24.2 2025-10-10T01:29:29.6073939Z - pandas==2.0.3 2025-10-10T01:29:29.6074289Z + pandas==2.2.3 2025-10-10T01:29:29.6075930Z + pathvalidate==3.2.1 2025-10-10T01:29:29.6076226Z + patsy==1.0.1 2025-10-10T01:29:29.6076812Z + peft==0.16.0 2025-10-10T01:29:29.6077358Z - pillow==11.0.0 2025-10-10T01:29:29.6078634Z + pillow==10.4.0 2025-10-10T01:29:29.6079318Z - platformdirs==4.5.0 2025-10-10T01:29:29.6079662Z + platformdirs==4.3.6 2025-10-10T01:29:29.6080659Z + plotly==5.24.1 2025-10-10T01:29:29.6081561Z - pluggy==1.6.0 2025-10-10T01:29:29.6081964Z + pluggy==1.5.0 2025-10-10T01:29:29.6083120Z + polars==1.29.0 2025-10-10T01:29:29.6083486Z + portalocker==2.10.1 2025-10-10T01:29:29.6084895Z + pqdm==0.2.0 2025-10-10T01:29:29.6085290Z + pretrainedmodels==0.7.4 2025-10-10T01:29:29.6086143Z - prometheus-client==0.23.1 2025-10-10T01:29:29.6086686Z + prometheus-client==0.22.0 2025-10-10T01:29:29.6087638Z - propcache==0.3.2 2025-10-10T01:29:29.6088332Z + propcache==0.2.0 2025-10-10T01:29:29.6089570Z + proto-plus==1.26.1 2025-10-10T01:29:29.6089986Z - protobuf==5.29.5 2025-10-10T01:29:29.6090755Z + protobuf==5.28.3 2025-10-10T01:29:29.6091585Z - psutil==7.1.0 2025-10-10T01:29:29.6092181Z + psutil==6.1.0 2025-10-10T01:29:29.6093112Z + py==1.11.0 2025-10-10T01:29:29.6093912Z + py-spy==0.4.0 2025-10-10T01:29:29.6094295Z + pyarrow==18.0.0 2025-10-10T01:29:29.6095440Z + pybind11==2.13.6 2025-10-10T01:29:29.6095993Z + pycocotools==2.0.8 2025-10-10T01:29:29.6096637Z - pycparser==2.23 2025-10-10T01:29:29.6097569Z + pycparser==2.22 2025-10-10T01:29:29.6098174Z + pycryptodomex==3.22.0 2025-10-10T01:29:29.6098851Z - pydantic==2.12.0 2025-10-10T01:29:29.6100296Z + pydantic==2.11.7 2025-10-10T01:29:29.6100846Z - pydantic-core==2.41.1 2025-10-10T01:29:29.6101533Z + pydantic-core==2.33.2 2025-10-10T01:29:29.6102114Z - pydantic-extra-types==2.10.6 2025-10-10T01:29:29.6102896Z + pydantic-extra-types==2.10.5 2025-10-10T01:29:29.6103683Z - pygments==2.15.0 2025-10-10T01:29:29.6104670Z + pygments==2.18.0 2025-10-10T01:29:29.6105219Z + pyogrio==0.11.0 2025-10-10T01:29:29.6105892Z + pyparsing==3.2.0 2025-10-10T01:29:29.6106744Z + pyproj==3.7.1 2025-10-10T01:29:29.6107553Z + pyrate-limiter==3.7.0 2025-10-10T01:29:29.6108178Z + pystemmer==3.0.0 2025-10-10T01:29:29.6109094Z + pytablewriter==1.2.0 2025-10-10T01:29:29.6109751Z - pytest==7.3.2 2025-10-10T01:29:29.6110500Z + pytest==8.3.5 2025-10-10T01:29:29.6111332Z + pytest-asyncio==0.24.0 2025-10-10T01:29:29.6112005Z + pytest-cov==6.3.0 2025-10-10T01:29:29.6112941Z + pytest-forked==1.6.0 2025-10-10T01:29:29.6113627Z + pytest-mock==3.14.0 2025-10-10T01:29:29.6114639Z + pytest-shard==0.1.2 2025-10-10T01:29:29.6115391Z - pytest-subtests==0.13.1 2025-10-10T01:29:29.6116133Z + pytest-subtests==0.14.1 2025-10-10T01:29:29.6116907Z + pytest-timeout==2.3.1 2025-10-10T01:29:29.6117284Z + python-box==7.3.2 2025-10-10T01:29:29.6118686Z + python-rapidjson==1.20 2025-10-10T01:29:29.6119113Z + pytorch-lightning==2.5.5 2025-10-10T01:29:29.6119862Z + pytrec-eval-terrier==0.5.7 2025-10-10T01:29:29.6120520Z - pytz==2025.2 2025-10-10T01:29:29.6121427Z + pytz==2024.2 2025-10-10T01:29:29.6121856Z + rapidfuzz==3.12.1 2025-10-10T01:29:29.6122822Z + rasterio==1.4.3 2025-10-10T01:29:29.6123497Z - ray==2.49.2 2025-10-10T01:29:29.6124443Z + ray==2.48.0 2025-10-10T01:29:29.6125687Z - redis==6.4.0 2025-10-10T01:29:29.6126082Z + redis==5.2.0 2025-10-10T01:29:29.6126965Z - referencing==0.36.2 2025-10-10T01:29:29.6127541Z + referencing==0.35.1 2025-10-10T01:29:29.6128853Z - regex==2025.9.18 2025-10-10T01:29:29.6129443Z + regex==2024.9.11 2025-10-10T01:29:29.6129908Z - requests==2.32.5 2025-10-10T01:29:29.6131306Z + requests==2.32.3 2025-10-10T01:29:29.6131683Z + responses==0.25.3 2025-10-10T01:29:29.6132377Z + rfc3339-validator==0.1.4 2025-10-10T01:29:29.6133363Z + rfc3987==1.3.8 2025-10-10T01:29:29.6134143Z - rich==14.2.0 2025-10-10T01:29:29.6134782Z + rich==13.9.4 2025-10-10T01:29:29.6135741Z + rioxarray==0.19.0 2025-10-10T01:29:29.6136546Z + rouge-score==0.1.2 2025-10-10T01:29:29.6137846Z - rpds-py==0.27.1 2025-10-10T01:29:29.6138214Z + rpds-py==0.20.1 2025-10-10T01:29:29.6138802Z + rtree==1.4.0 2025-10-10T01:29:29.6139542Z + runai-model-streamer==0.14.0 2025-10-10T01:29:29.6140439Z + runai-model-streamer-gcs==0.14.0 2025-10-10T01:29:29.6141208Z + runai-model-streamer-s3==0.14.0 2025-10-10T01:29:29.6141821Z - s3transfer==0.10.4 2025-10-10T01:29:29.6142608Z + s3transfer==0.10.3 2025-10-10T01:29:29.6143544Z + sacrebleu==2.4.3 2025-10-10T01:29:29.6144408Z - safetensors==0.6.2 2025-10-10T01:29:29.6145075Z + safetensors==0.4.5 2025-10-10T01:29:29.6145799Z + schemathesis==3.39.15 2025-10-10T01:29:29.6146447Z - scikit-image==0.22.0 2025-10-10T01:29:29.6147338Z + scikit-image==0.25.2 2025-10-10T01:29:29.6148288Z - scikit-learn==1.7.2 2025-10-10T01:29:29.6148887Z + scikit-learn==1.5.2 2025-10-10T01:29:29.6149788Z - scipy==1.14.1 2025-10-10T01:29:29.6150485Z + scipy==1.13.1 2025-10-10T01:29:29.6151805Z + segmentation-models-pytorch==0.4.0 2025-10-10T01:29:29.6152199Z + sentence-transformers==3.2.1 2025-10-10T01:29:29.6153233Z - sentencepiece==0.2.1 2025-10-10T01:29:29.6153635Z + sentencepiece==0.2.0 2025-10-10T01:29:29.6155029Z - setuptools==78.1.1 2025-10-10T01:29:29.6155340Z + setuptools==77.0.3 2025-10-10T01:29:29.6156102Z + shapely==2.1.1 2025-10-10T01:29:29.6156778Z - six==1.17.0 2025-10-10T01:29:29.6157615Z + six==1.16.0 2025-10-10T01:29:29.6158609Z + smart-open==7.1.0 2025-10-10T01:29:29.6159224Z - soundfile==0.13.1 2025-10-10T01:29:29.6159985Z + soundfile==0.12.1 2025-10-10T01:29:29.6160836Z - soxr==1.0.0 2025-10-10T01:29:29.6161530Z + soxr==0.5.0.post1 2025-10-10T01:29:29.6162508Z + sqlalchemy==2.0.41 2025-10-10T01:29:29.6163297Z + sqlitedict==2.1.0 2025-10-10T01:29:29.6164489Z + sqlparse==0.5.3 2025-10-10T01:29:29.6165187Z - starlette==0.48.0 2025-10-10T01:29:29.6165875Z + starlette==0.46.2 2025-10-10T01:29:29.6166860Z + starlette-testclient==0.4.1 2025-10-10T01:29:29.6167448Z + statsmodels==0.14.4 2025-10-10T01:29:29.6168394Z + tabledata==1.3.3 2025-10-10T01:29:29.6169031Z + tblib==3.1.0 2025-10-10T01:29:29.6169895Z + tcolorpy==0.1.6 2025-10-10T01:29:29.6170493Z + tenacity==9.0.0 2025-10-10T01:29:29.6171751Z + tensorboardx==2.6.4 2025-10-10T01:29:29.6172336Z + tensorizer==2.10.1 2025-10-10T01:29:29.6174201Z + terratorch==1.0.2 (from git+https://github.com/IBM/terratorch.git@07184fcf91a1324f831ff521dd238d97fe350e3e) 2025-10-10T01:29:29.6174756Z - threadpoolctl==3.6.0 2025-10-10T01:29:29.6175333Z + threadpoolctl==3.5.0 2025-10-10T01:29:29.6176134Z - tifffile==2025.10.4 2025-10-10T01:29:29.6177242Z + tifffile==2025.3.30 2025-10-10T01:29:29.6177694Z - tiktoken==0.12.0 2025-10-10T01:29:29.6178500Z + tiktoken==0.7.0 2025-10-10T01:29:29.6179033Z + timm==1.0.17 2025-10-10T01:29:29.6179882Z - tokenizers==0.22.1 2025-10-10T01:29:29.6180477Z + tokenizers==0.22.0 2025-10-10T01:29:29.6181199Z + tomli==2.2.1 2025-10-10T01:29:29.6182375Z + tomli-w==1.2.0 2025-10-10T01:29:29.6184167Z - torch==2.10.0a0+git344e636 (from file:///var/lib/jenkins/workspace/dist/torch-2.10.0a0%2Bgit344e636-cp312-cp312-linux_x86_64.whl) 2025-10-10T01:29:29.6186196Z + torch==2.10.0a0+git344e636 (from file:///var/lib/jenkins/workspace/dist/torch-2.10.0a0+git344e636-cp312-cp312-linux_x86_64.whl#sha256=61b81d5619a3dff1b95456c948bf99d131288c4916a1b2d71da5834655529eee) 2025-10-10T01:29:29.6187762Z - torchaudio==2.8.0a0+87ff22e (from file:///var/lib/jenkins/workspace/dist/audio/torchaudio-2.8.0a0%2B87ff22e-cp312-cp312-linux_x86_64.whl) 2025-10-10T01:29:29.6189623Z + torchaudio==2.8.0a0+87ff22e (from file:///var/lib/jenkins/workspace/dist/audio/torchaudio-2.8.0a0+87ff22e-cp312-cp312-linux_x86_64.whl#sha256=2eac501979e64f67c113a81572de033ce6ec78d775f6402cacf95aa651581299) 2025-10-10T01:29:29.6190782Z + torchgeo==0.7.0 2025-10-10T01:29:29.6191297Z + torchmetrics==1.8.2 2025-10-10T01:29:29.6192357Z - torchvision==0.22.0a0+966da7e (from file:///var/lib/jenkins/workspace/dist/vision/torchvision-0.22.0a0%2B966da7e-cp312-cp312-linux_x86_64.whl) 2025-10-10T01:29:29.6194004Z + torchvision==0.22.0a0+966da7e (from file:///var/lib/jenkins/workspace/dist/vision/torchvision-0.22.0a0+966da7e-cp312-cp312-linux_x86_64.whl#sha256=79aa422fd81e11b575186dd520ce1848b0356118875ba3f51c957fae9e3f437f) 2025-10-10T01:29:29.6195196Z - tqdm==4.67.1 2025-10-10T01:29:29.6195567Z + tqdm==4.66.6 2025-10-10T01:29:29.6195977Z + tqdm-multiprocess==0.0.11 2025-10-10T01:29:29.6196299Z - transformers==4.57.0 2025-10-10T01:29:29.6196717Z + transformers==4.56.2 2025-10-10T01:29:29.6197305Z + transformers-stream-generator==0.0.5 2025-10-10T01:29:29.6197790Z + tritonclient==2.51.0 2025-10-10T01:29:29.6198478Z + typepy==1.3.2 2025-10-10T01:29:29.6199012Z - typer==0.19.2 2025-10-10T01:29:29.6199353Z + typer==0.15.2 2025-10-10T01:29:29.6199883Z + types-python-dateutil==2.9.0.20241206 2025-10-10T01:29:29.6200382Z + typeshed-client==2.8.2 2025-10-10T01:29:29.6201121Z - typing-extensions==4.15.0 2025-10-10T01:29:29.6201713Z + typing-extensions==4.12.2 2025-10-10T01:29:29.6202701Z - typing-inspection==0.4.2 2025-10-10T01:29:29.6203032Z + typing-inspection==0.4.1 2025-10-10T01:29:29.6203939Z - tzdata==2025.2 2025-10-10T01:29:29.6204571Z + tzdata==2024.2 2025-10-10T01:29:29.6205739Z + uri-template==1.3.0 2025-10-10T01:29:29.6206403Z - urllib3==2.5.0 2025-10-10T01:29:29.6207172Z + urllib3==2.2.3 2025-10-10T01:29:29.6207821Z - uvicorn==0.37.0 2025-10-10T01:29:29.6208305Z + uvicorn==0.35.0 2025-10-10T01:29:29.6209220Z + vector-quantize-pytorch==1.23.2 2025-10-10T01:29:29.6209929Z + virtualenv==20.31.2 2025-10-10T01:29:29.6210680Z + vocos==0.1.0 2025-10-10T01:29:29.6211580Z + wcwidth==0.2.13 2025-10-10T01:29:29.6212499Z + webcolors==24.11.1 2025-10-10T01:29:29.6213206Z + word2number==1.1 2025-10-10T01:29:29.6213661Z - wrapt==1.17.3 2025-10-10T01:29:29.6214547Z + wrapt==1.17.2 2025-10-10T01:29:29.6215151Z + xarray==2025.7.1 2025-10-10T01:29:29.6216531Z + xxhash==3.5.0 2025-10-10T01:29:29.6216828Z - yarl==1.22.0 2025-10-10T01:29:29.6217380Z + yarl==1.17.1 2025-10-10T01:29:29.6218777Z + zipp==3.23.0 2025-10-10T01:29:29.6219293Z + zstandard==0.23.0 2025-10-10T01:29:29.7971675Z 2025-10-10 01:29:29,796 [INFO] cli.lib.core.vllm.vllm_test: Done. installed requirements for test dependencies 2025-10-10T01:29:29.7973275Z 2025-10-10 01:29:29,796 [INFO] cli.lib.core.vllm.vllm_test: Double check installed packages 2025-10-10T01:29:29.8020537Z 2025-10-10 01:29:29,801 [INFO] cli.lib.common.pip_helper: torch already exist with version: 2.10.0a0+git344e636 2025-10-10T01:29:29.8024878Z 2025-10-10 01:29:29,802 [INFO] cli.lib.common.pip_helper: xformers already exist with version: 0.0.33+5d4b92a5.d20251010 2025-10-10T01:29:29.8029726Z 2025-10-10 01:29:29,802 [INFO] cli.lib.common.pip_helper: torchvision already exist with version: 0.22.0a0+966da7e 2025-10-10T01:29:29.8034654Z 2025-10-10 01:29:29,803 [INFO] cli.lib.common.pip_helper: torchaudio already exist with version: 2.8.0a0+87ff22e 2025-10-10T01:29:29.8041043Z 2025-10-10 01:29:29,803 [INFO] cli.lib.common.pip_helper: vllm already exist with version: 0.11.0rc2.dev157+g0ad9951c4.d20251010 2025-10-10T01:29:29.8042048Z 2025-10-10 01:29:29,803 [INFO] cli.lib.core.vllm.vllm_test: Done. checked installed packages 2025-10-10T01:29:29.8042806Z 2025-10-10 01:29:29,803 [INFO] cli.lib.core.vllm.lib: run vllm tests..... 2025-10-10T01:29:29.8043651Z 2025-10-10 01:29:29,804 [INFO] cli.lib.core.vllm.lib: Running tests: Distributed Tests (2 GPUs) pytorch 2.8 release failure 2025-10-10T01:29:29.8044714Z 2025-10-10 01:29:29,804 [INFO] cli.lib.core.vllm.lib: Running step: pytest -v -s distributed/test_sequence_parallel.py 2025-10-10T01:29:29.8045497Z 2025-10-10 01:29:29,804 [INFO] cli.lib.common.utils: [shell] pytest -v -s distributed/test_sequence_parallel.py 2025-10-10T01:29:40.3067725Z INFO 10-10 01:29:40 [__init__.py:215] Automatically detected platform cuda. 2025-10-10T01:29:42.8082357Z /opt/conda/envs/py_3.12/lib/python3.12/site-packages/pytest_asyncio/plugin.py:208: PytestDeprecationWarning: The configuration option "asyncio_default_fixture_loop_scope" is unset. 2025-10-10T01:29:42.8084732Z The event loop scope for asynchronous fixtures will default to the fixture caching scope. Future versions of pytest-asyncio will default the loop scope for asynchronous fixtures to function scope. Set the default fixture loop scope explicitly in order to avoid unexpected behavior in the future. Valid fixture loop scopes are: "function", "class", "module", "package", "session" 2025-10-10T01:29:42.8086393Z 2025-10-10T01:29:42.8086686Z warnings.warn(PytestDeprecationWarning(_DEFAULT_FIXTURE_LOOP_SCOPE_UNSET)) 2025-10-10T01:29:42.8154725Z ============================= test session starts ============================== 2025-10-10T01:29:42.8155331Z platform linux -- Python 3.12.11, pytest-8.3.5, pluggy-1.5.0 -- /opt/conda/envs/py_3.12/bin/python 2025-10-10T01:29:42.8248256Z cachedir: .pytest_cache 2025-10-10T01:29:42.8249112Z hypothesis profile 'ci' -> database=None, deadline=None, print_blob=True, derandomize=True, suppress_health_check=[HealthCheck.too_slow] 2025-10-10T01:29:42.8249784Z rootdir: /var/lib/jenkins/workspace/vllm 2025-10-10T01:29:42.8250088Z configfile: pyproject.toml 2025-10-10T01:29:42.8251111Z plugins: cpp-2.3.0, flakefinder-1.1.0, rerunfailures-14.0, xdist-3.3.1, xdoctest-1.3.0, mock-3.14.0, hydra-core-1.3.2, subtests-0.14.1, buildkite-test-collector-0.1.9, hypothesis-6.131.0, cov-6.3.0, schemathesis-3.39.15, forked-1.6.0, anyio-4.6.2.post1, shard-0.1.2, timeout-2.3.1, asyncio-0.24.0 2025-10-10T01:29:42.8252189Z asyncio: mode=Mode.STRICT, default_loop_scope=None 2025-10-10T01:29:46.6219318Z collecting ...  2025-10-10T01:29:46.6223318Z collecting 20 items  2025-10-10T01:29:46.6224975Z collected 20 items  2025-10-10T01:29:46.6235371Z Running 20 items in this shard: tests/distributed/test_sequence_parallel.py::test_tp_sp_generation[meta-llama/Llama-3.2-1B-Instruct-parallel_setup0-mp-1-auto-test_options0], tests/distributed/test_sequence_parallel.py::test_tp_sp_generation[meta-llama/Llama-3.2-1B-Instruct-parallel_setup1-ray-1-auto-test_options1], tests/distributed/test_sequence_parallel.py::test_tp_sp_generation[meta-llama/Llama-3.2-1B-Instruct-parallel_setup2-mp-1-auto-test_options2], tests/distributed/test_sequence_parallel.py::test_tp_sp_generation[meta-llama/Llama-3.2-1B-Instruct-parallel_setup3-ray-1-auto-test_options3], tests/distributed/test_sequence_parallel.py::test_tp_sp_generation[meta-llama/Llama-3.2-1B-Instruct-parallel_setup4-mp-1-auto-test_options4], tests/distributed/test_sequence_parallel.py::test_tp_sp_generation[meta-llama/Llama-3.2-1B-Instruct-parallel_setup5-ray-1-auto-test_options5], tests/distributed/test_sequence_parallel.py::test_tp_sp_generation[meta-llama/Llama-3.2-1B-Instruct-parallel_setup6-mp-1-auto-test_options6], tests/distributed/test_sequence_parallel.py::test_tp_sp_generation[meta-llama/Llama-3.2-1B-Instruct-parallel_setup7-ray-1-auto-test_options7], tests/distributed/test_sequence_parallel.py::test_tp_sp_generation[meta-llama/Llama-3.2-1B-Instruct-parallel_setup8-mp-1-auto-test_options8], tests/distributed/test_sequence_parallel.py::test_tp_sp_generation[meta-llama/Llama-3.2-1B-Instruct-parallel_setup9-ray-1-auto-test_options9], tests/distributed/test_sequence_parallel.py::test_tp_sp_generation[meta-llama/Llama-3.2-1B-Instruct-parallel_setup10-mp-1-auto-test_options10], tests/distributed/test_sequence_parallel.py::test_tp_sp_generation[meta-llama/Llama-3.2-1B-Instruct-parallel_setup11-ray-1-auto-test_options11], tests/distributed/test_sequence_parallel.py::test_tp_sp_generation[meta-llama/Llama-3.2-1B-Instruct-parallel_setup12-mp-1-auto-test_options12], tests/distributed/test_sequence_parallel.py::test_tp_sp_generation[meta-llama/Llama-3.2-1B-Instruct-parallel_setup13-ray-1-auto-test_options13], tests/distributed/test_sequence_parallel.py::test_tp_sp_generation[meta-llama/Llama-3.2-1B-Instruct-parallel_setup14-mp-1-auto-test_options14], tests/distributed/test_sequence_parallel.py::test_tp_sp_generation[meta-llama/Llama-3.2-1B-Instruct-parallel_setup15-ray-1-auto-test_options15], tests/distributed/test_sequence_parallel.py::test_tp_sp_generation[RedHatAI/Meta-Llama-3.1-8B-Instruct-FP8-parallel_setup16-mp-1-auto-test_options16], tests/distributed/test_sequence_parallel.py::test_tp_sp_generation[RedHatAI/Meta-Llama-3.1-8B-Instruct-FP8-parallel_setup17-ray-1-auto-test_options17], tests/distributed/test_sequence_parallel.py::test_tp_sp_generation[RedHatAI/Meta-Llama-3.1-8B-Instruct-FP8-parallel_setup18-mp-1-auto-test_options18], tests/distributed/test_sequence_parallel.py::test_tp_sp_generation[RedHatAI/Meta-Llama-3.1-8B-Instruct-FP8-parallel_setup19-ray-1-auto-test_options19] 2025-10-10T01:29:46.6244286Z 2025-10-10T01:29:46.6378794Z distributed/test_sequence_parallel.py::test_tp_sp_generation[meta-llama/Llama-3.2-1B-Instruct-parallel_setup0-mp-1-auto-test_options0] Fork a new process to run a test 1161 2025-10-10T01:29:46.6392832Z Fork a new process to run a test 0 2025-10-10T01:29:46.7415531Z 2025-10-10T01:29:46.7418395Z tokenizer_config.json: 0% 0.00/54.5k [00:00, 'use_cudagraph': True, 'cudagraph_num_of_warmups': 1, 'cudagraph_capture_sizes': [16, 8, 4, 2], 'cudagraph_copy_inputs': False, 'full_cuda_graph': True, 'use_inductor_graph_partition': False, 'pass_config': {'enable_noop': True, 'enable_sequence_parallelism': True}, 'max_capture_size': 16, 'local_cache_dir': None} 2025-10-10T01:30:09.6885033Z (EngineCore_DP0 pid=1448) WARNING 10-10 01:30:09 [multiproc_executor.py:720] Reducing Torch parallelism from 24 threads to 1 to avoid unnecessary CPU contention. Set OMP_NUM_THREADS in the external environment to tune this value as needed. 2025-10-10T01:30:09.6888424Z (EngineCore_DP0 pid=1448) INFO 10-10 01:30:09 [shm_broadcast.py:289] vLLM message queue communication handle: Handle(local_reader_ranks=[0, 1], buffer_handle=(2, 16777216, 10, 'psm_7c6a2641'), local_subscribe_addr='ipc:///tmp/6b98a77b-9a6a-48e3-8524-313b0899715a', remote_subscribe_addr=None, remote_addr_ipv6=False) 2025-10-10T01:30:10.1701469Z /opt/conda/envs/py_3.12/lib/python3.12/site-packages/torch/cuda/__init__.py:64: FutureWarning: The pynvml package is deprecated. Please install nvidia-ml-py instead. If you did not install pynvml directly, please report this to the maintainers of the package that installed pynvml for you. 2025-10-10T01:30:10.1703825Z import pynvml # type: ignore[import] 2025-10-10T01:30:10.1885029Z /opt/conda/envs/py_3.12/lib/python3.12/site-packages/torch/cuda/__init__.py:64: FutureWarning: The pynvml package is deprecated. Please install nvidia-ml-py instead. If you did not install pynvml directly, please report this to the maintainers of the package that installed pynvml for you. 2025-10-10T01:30:10.1887234Z import pynvml # type: ignore[import] 2025-10-10T01:30:12.5066032Z INFO 10-10 01:30:12 [__init__.py:215] Automatically detected platform cuda. 2025-10-10T01:30:12.5161129Z INFO 10-10 01:30:12 [__init__.py:215] Automatically detected platform cuda. 2025-10-10T01:30:17.3745811Z INFO 10-10 01:30:17 [shm_broadcast.py:289] vLLM message queue communication handle: Handle(local_reader_ranks=[0], buffer_handle=(1, 10485760, 10, 'psm_5ddfa8e8'), local_subscribe_addr='ipc:///tmp/f007e138-950b-4b69-930d-a7528e98cc4f', remote_subscribe_addr=None, remote_addr_ipv6=False) 2025-10-10T01:30:17.3941537Z INFO 10-10 01:30:17 [shm_broadcast.py:289] vLLM message queue communication handle: Handle(local_reader_ranks=[0], buffer_handle=(1, 10485760, 10, 'psm_0ce09250'), local_subscribe_addr='ipc:///tmp/5dc321ca-892c-4088-91dc-89045ee68c67', remote_subscribe_addr=None, remote_addr_ipv6=False) 2025-10-10T01:30:17.9319045Z [Gloo] Rank 0 is connected to 1 peer ranks. Expected number of connected peer ranks is : 1 2025-10-10T01:30:17.9319685Z [Gloo] Rank 1 is connected to 1 peer ranks. Expected number of connected peer ranks is : 1 2025-10-10T01:30:17.9361683Z [Gloo] Rank 0 is connected to 1 peer ranks. Expected number of connected peer ranks is : 1 2025-10-10T01:30:17.9362274Z [Gloo] Rank 1 is connected to 1 peer ranks. Expected number of connected peer ranks is : 1 2025-10-10T01:30:17.9483580Z INFO 10-10 01:30:17 [__init__.py:1384] Found nccl from library libnccl.so.2 2025-10-10T01:30:17.9484977Z INFO 10-10 01:30:17 [__init__.py:1384] Found nccl from library libnccl.so.2 2025-10-10T01:30:17.9485825Z INFO 10-10 01:30:17 [pynccl.py:104] vLLM is using nccl==2.27.5 2025-10-10T01:30:17.9487603Z INFO 10-10 01:30:17 [pynccl.py:104] vLLM is using nccl==2.27.5 2025-10-10T01:30:18.1487191Z WARNING 10-10 01:30:18 [symm_mem.py:58] SymmMemCommunicator: Device capability 8.9 not supported, communicator is not available. 2025-10-10T01:30:18.1488785Z WARNING 10-10 01:30:18 [symm_mem.py:58] SymmMemCommunicator: Device capability 8.9 not supported, communicator is not available. 2025-10-10T01:30:18.1513091Z INFO 10-10 01:30:18 [custom_all_reduce.py:35] Skipping P2P check and trusting the driver's P2P report. 2025-10-10T01:30:18.1514147Z WARNING 10-10 01:30:18 [custom_all_reduce.py:154] Custom allreduce is disabled because your platform lacks GPU P2P capability or P2P test failed. To silence this warning, specify disable_custom_all_reduce=True explicitly. 2025-10-10T01:30:18.1515230Z INFO 10-10 01:30:18 [custom_all_reduce.py:35] Skipping P2P check and trusting the driver's P2P report. 2025-10-10T01:30:18.1516606Z WARNING 10-10 01:30:18 [custom_all_reduce.py:154] Custom allreduce is disabled because your platform lacks GPU P2P capability or P2P test failed. To silence this warning, specify disable_custom_all_reduce=True explicitly. 2025-10-10T01:30:18.2739326Z INFO 10-10 01:30:18 [shm_broadcast.py:289] vLLM message queue communication handle: Handle(local_reader_ranks=[1], buffer_handle=(1, 4194304, 6, 'psm_df2e97c7'), local_subscribe_addr='ipc:///tmp/79796c5d-2a63-41e6-a5ff-365aeeab7aa2', remote_subscribe_addr=None, remote_addr_ipv6=False) 2025-10-10T01:30:18.2772770Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-10-10T01:30:18.2773269Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-10-10T01:30:18.2780458Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-10-10T01:30:18.2788724Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-10-10T01:30:18.3532826Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-10-10T01:30:18.3540632Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-10-10T01:30:18.3551231Z [Gloo] Rank 0 is connected to 1 peer ranks. Expected number of connected peer ranks is : 1 2025-10-10T01:30:18.3551745Z [Gloo] Rank 1 is connected to 1 peer ranks. Expected number of connected peer ranks is : 1 2025-10-10T01:30:18.3553482Z INFO 10-10 01:30:18 [__init__.py:1384] Found nccl from library libnccl.so.2 2025-10-10T01:30:18.3553886Z INFO 10-10 01:30:18 [__init__.py:1384] Found nccl from library libnccl.so.2 2025-10-10T01:30:18.3554230Z INFO 10-10 01:30:18 [pynccl.py:104] vLLM is using nccl==2.27.5 2025-10-10T01:30:18.3554547Z INFO 10-10 01:30:18 [pynccl.py:104] vLLM is using nccl==2.27.5 2025-10-10T01:30:18.3850994Z INFO 10-10 01:30:18 [parallel_state.py:1208] rank 0 in world size 2 is assigned as DP rank 0, PP rank 0, TP rank 0, EP rank 0 2025-10-10T01:30:18.3851771Z INFO 10-10 01:30:18 [parallel_state.py:1208] rank 1 in world size 2 is assigned as DP rank 0, PP rank 0, TP rank 1, EP rank 1 2025-10-10T01:30:18.6355715Z INFO 10-10 01:30:18 [topk_topp_sampler.py:55] Using FlashInfer for top-p & top-k sampling. 2025-10-10T01:30:18.6420291Z INFO 10-10 01:30:18 [topk_topp_sampler.py:55] Using FlashInfer for top-p & top-k sampling. 2025-10-10T01:30:18.6679681Z (Worker_TP0 pid=1567) INFO 10-10 01:30:18 [gpu_model_runner.py:2707] Starting to load model meta-llama/Llama-3.2-1B-Instruct... 2025-10-10T01:30:18.6729483Z (Worker_TP1 pid=1568) INFO 10-10 01:30:18 [gpu_model_runner.py:2707] Starting to load model meta-llama/Llama-3.2-1B-Instruct... 2025-10-10T01:30:18.8940527Z (Worker_TP0 pid=1567) INFO 10-10 01:30:18 [gpu_model_runner.py:2739] Loading model from scratch... 2025-10-10T01:30:18.9011310Z (Worker_TP1 pid=1568) INFO 10-10 01:30:18 [gpu_model_runner.py:2739] Loading model from scratch... 2025-10-10T01:30:18.9849315Z (Worker_TP0 pid=1567) INFO 10-10 01:30:18 [cuda.py:361] Using Flash Attention backend on V1 engine. 2025-10-10T01:30:18.9912977Z (Worker_TP1 pid=1568) INFO 10-10 01:30:18 [cuda.py:361] Using Flash Attention backend on V1 engine. 2025-10-10T01:30:19.5183062Z (Worker_TP1 pid=1568) INFO 10-10 01:30:19 [gpu_model_runner.py:2758] Model loading took 1.1667 GiB and 0.168515 seconds 2025-10-10T01:30:19.5215083Z (Worker_TP0 pid=1567) INFO 10-10 01:30:19 [gpu_model_runner.py:2758] Model loading took 1.1667 GiB and 0.173364 seconds 2025-10-10T01:30:22.8749160Z (Worker_TP1 pid=1568) INFO 10-10 01:30:22 [backends.py:548] Using cache directory: /var/lib/jenkins/.cache/vllm/torch_compile_cache/31e0bc77eb/rank_1_0/backbone for vLLM's torch.compile 2025-10-10T01:30:22.8751072Z (Worker_TP1 pid=1568) INFO 10-10 01:30:22 [backends.py:559] Dynamo bytecode transform time: 3.06 s 2025-10-10T01:30:22.9436851Z (Worker_TP0 pid=1567) INFO 10-10 01:30:22 [backends.py:548] Using cache directory: /var/lib/jenkins/.cache/vllm/torch_compile_cache/31e0bc77eb/rank_0_0/backbone for vLLM's torch.compile 2025-10-10T01:30:22.9441814Z (Worker_TP0 pid=1567) INFO 10-10 01:30:22 [backends.py:559] Dynamo bytecode transform time: 3.14 s 2025-10-10T01:30:24.2342560Z (Worker_TP0 pid=1567) [rank0]:W1010 01:30:24.233000 1567 site-packages/torch/_inductor/utils.py:1594] [0/0] Not enough SMs to use max_autotune_gemm mode 2025-10-10T01:30:24.2371096Z (Worker_TP1 pid=1568) [rank1]:W1010 01:30:24.236000 1568 site-packages/torch/_inductor/utils.py:1594] [0/0] Not enough SMs to use max_autotune_gemm mode 2025-10-10T01:30:25.8107900Z (Worker_TP1 pid=1568) INFO 10-10 01:30:25 [backends.py:197] Cache the graph for dynamic shape for later use 2025-10-10T01:30:25.8147467Z (Worker_TP0 pid=1567) INFO 10-10 01:30:25 [backends.py:197] Cache the graph for dynamic shape for later use 2025-10-10T01:30:33.0759117Z (Worker_TP0 pid=1567) INFO 10-10 01:30:33 [backends.py:218] Compiling a graph for dynamic shape takes 9.88 s 2025-10-10T01:30:33.0822616Z (Worker_TP1 pid=1568) INFO 10-10 01:30:33 [backends.py:218] Compiling a graph for dynamic shape takes 9.95 s 2025-10-10T01:30:40.6585968Z (Worker_TP1 pid=1568) INFO 10-10 01:30:40 [gpu_worker.py:298] Available KV cache memory: 18.50 GiB 2025-10-10T01:30:40.6961109Z (Worker_TP0 pid=1567) INFO 10-10 01:30:40 [gpu_worker.py:298] Available KV cache memory: 18.50 GiB 2025-10-10T01:30:41.0096029Z (EngineCore_DP0 pid=1448) INFO 10-10 01:30:41 [kv_cache_utils.py:1087] GPU KV cache size: 1,212,160 tokens 2025-10-10T01:30:41.0096959Z (EngineCore_DP0 pid=1448) INFO 10-10 01:30:41 [kv_cache_utils.py:1091] Maximum concurrency for 2,048 tokens per request: 591.88x 2025-10-10T01:30:41.0098019Z (EngineCore_DP0 pid=1448) INFO 10-10 01:30:41 [kv_cache_utils.py:1087] GPU KV cache size: 1,212,160 tokens 2025-10-10T01:30:41.0098917Z (EngineCore_DP0 pid=1448) INFO 10-10 01:30:41 [kv_cache_utils.py:1091] Maximum concurrency for 2,048 tokens per request: 591.88x 2025-10-10T01:30:41.3319622Z (Worker_TP0 pid=1567) 2025-10-10T01:30:42.5469168Z Capturing CUDA graphs (mixed prefill-decode, PIECEWISE): 0% 0/4 [00:00, 'use_cudagraph': True, 'cudagraph_num_of_warmups': 1, 'cudagraph_capture_sizes': [16, 8, 4, 2, 1], 'cudagraph_copy_inputs': False, 'full_cuda_graph': True, 'use_inductor_graph_partition': False, 'pass_config': {}, 'max_capture_size': 16, 'local_cache_dir': None} 2025-10-10T01:31:23.7589851Z (EngineCore_DP0 pid=3596) WARNING 10-10 01:31:23 [multiproc_executor.py:720] Reducing Torch parallelism from 24 threads to 1 to avoid unnecessary CPU contention. Set OMP_NUM_THREADS in the external environment to tune this value as needed. 2025-10-10T01:31:23.7591485Z (EngineCore_DP0 pid=3596) INFO 10-10 01:31:23 [shm_broadcast.py:289] vLLM message queue communication handle: Handle(local_reader_ranks=[0, 1], buffer_handle=(2, 16777216, 10, 'psm_f61d7074'), local_subscribe_addr='ipc:///tmp/6725d46e-ad2b-477a-8cee-d3bf7965458a', remote_subscribe_addr=None, remote_addr_ipv6=False) 2025-10-10T01:31:24.2399609Z /opt/conda/envs/py_3.12/lib/python3.12/site-packages/torch/cuda/__init__.py:64: FutureWarning: The pynvml package is deprecated. Please install nvidia-ml-py instead. If you did not install pynvml directly, please report this to the maintainers of the package that installed pynvml for you. 2025-10-10T01:31:24.2401707Z import pynvml # type: ignore[import] 2025-10-10T01:31:24.2473694Z /opt/conda/envs/py_3.12/lib/python3.12/site-packages/torch/cuda/__init__.py:64: FutureWarning: The pynvml package is deprecated. Please install nvidia-ml-py instead. If you did not install pynvml directly, please report this to the maintainers of the package that installed pynvml for you. 2025-10-10T01:31:24.2475926Z import pynvml # type: ignore[import] 2025-10-10T01:31:26.5482247Z INFO 10-10 01:31:26 [__init__.py:215] Automatically detected platform cuda. 2025-10-10T01:31:26.5771238Z INFO 10-10 01:31:26 [__init__.py:215] Automatically detected platform cuda. 2025-10-10T01:31:31.0145451Z INFO 10-10 01:31:31 [shm_broadcast.py:289] vLLM message queue communication handle: Handle(local_reader_ranks=[0], buffer_handle=(1, 10485760, 10, 'psm_de7d1926'), local_subscribe_addr='ipc:///tmp/ee062150-8e92-40db-8eae-06414b2d3fd7', remote_subscribe_addr=None, remote_addr_ipv6=False) 2025-10-10T01:31:31.1645179Z INFO 10-10 01:31:31 [shm_broadcast.py:289] vLLM message queue communication handle: Handle(local_reader_ranks=[0], buffer_handle=(1, 10485760, 10, 'psm_ee598f49'), local_subscribe_addr='ipc:///tmp/e9237ae5-f728-4293-92d8-649c3f4eb3ff', remote_subscribe_addr=None, remote_addr_ipv6=False) 2025-10-10T01:31:31.6806128Z [Gloo] Rank 0 is connected to 1 peer ranks. Expected number of connected peer ranks is : 1 2025-10-10T01:31:31.6807330Z [Gloo] Rank 1 is connected to 1 peer ranks. Expected number of connected peer ranks is : 1 2025-10-10T01:31:31.6850176Z [Gloo] Rank 0 is connected to 1 peer ranks. Expected number of connected peer ranks is : 1 2025-10-10T01:31:31.6850809Z [Gloo] Rank 1 is connected to 1 peer ranks. Expected number of connected peer ranks is : 1 2025-10-10T01:31:31.6892974Z INFO 10-10 01:31:31 [__init__.py:1384] Found nccl from library libnccl.so.2 2025-10-10T01:31:31.6893898Z INFO 10-10 01:31:31 [__init__.py:1384] Found nccl from library libnccl.so.2 2025-10-10T01:31:31.6895034Z INFO 10-10 01:31:31 [pynccl.py:104] vLLM is using nccl==2.27.5 2025-10-10T01:31:31.6895789Z INFO 10-10 01:31:31 [pynccl.py:104] vLLM is using nccl==2.27.5 2025-10-10T01:31:31.8583418Z WARNING 10-10 01:31:31 [symm_mem.py:58] SymmMemCommunicator: Device capability 8.9 not supported, communicator is not available. 2025-10-10T01:31:31.8585122Z WARNING 10-10 01:31:31 [symm_mem.py:58] SymmMemCommunicator: Device capability 8.9 not supported, communicator is not available. 2025-10-10T01:31:31.8609968Z INFO 10-10 01:31:31 [custom_all_reduce.py:35] Skipping P2P check and trusting the driver's P2P report. 2025-10-10T01:31:31.8612652Z INFO 10-10 01:31:31 [custom_all_reduce.py:35] Skipping P2P check and trusting the driver's P2P report. 2025-10-10T01:31:31.8613675Z WARNING 10-10 01:31:31 [custom_all_reduce.py:154] Custom allreduce is disabled because your platform lacks GPU P2P capability or P2P test failed. To silence this warning, specify disable_custom_all_reduce=True explicitly. 2025-10-10T01:31:31.8615145Z WARNING 10-10 01:31:31 [custom_all_reduce.py:154] Custom allreduce is disabled because your platform lacks GPU P2P capability or P2P test failed. To silence this warning, specify disable_custom_all_reduce=True explicitly. 2025-10-10T01:31:31.9940435Z INFO 10-10 01:31:31 [shm_broadcast.py:289] vLLM message queue communication handle: Handle(local_reader_ranks=[1], buffer_handle=(1, 4194304, 6, 'psm_184a85b8'), local_subscribe_addr='ipc:///tmp/733df021-a1dd-41a8-b24d-9a04ba7a8de2', remote_subscribe_addr=None, remote_addr_ipv6=False) 2025-10-10T01:31:31.9965642Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-10-10T01:31:31.9968209Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-10-10T01:31:31.9973617Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-10-10T01:31:31.9981485Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-10-10T01:31:32.0737159Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-10-10T01:31:32.0746524Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-10-10T01:31:32.0758077Z [Gloo] Rank 0 is connected to 1 peer ranks. Expected number of connected peer ranks is : 1 2025-10-10T01:31:32.0758663Z [Gloo] Rank 1 is connected to 1 peer ranks. Expected number of connected peer ranks is : 1 2025-10-10T01:31:32.0761033Z INFO 10-10 01:31:32 [__init__.py:1384] Found nccl from library libnccl.so.2 2025-10-10T01:31:32.0761415Z INFO 10-10 01:31:32 [__init__.py:1384] Found nccl from library libnccl.so.2 2025-10-10T01:31:32.0761770Z INFO 10-10 01:31:32 [pynccl.py:104] vLLM is using nccl==2.27.5 2025-10-10T01:31:32.0762089Z INFO 10-10 01:31:32 [pynccl.py:104] vLLM is using nccl==2.27.5 2025-10-10T01:31:32.1099285Z INFO 10-10 01:31:32 [parallel_state.py:1208] rank 1 in world size 2 is assigned as DP rank 0, PP rank 0, TP rank 1, EP rank 1 2025-10-10T01:31:32.1100071Z INFO 10-10 01:31:32 [parallel_state.py:1208] rank 0 in world size 2 is assigned as DP rank 0, PP rank 0, TP rank 0, EP rank 0 2025-10-10T01:31:32.3550056Z INFO 10-10 01:31:32 [topk_topp_sampler.py:55] Using FlashInfer for top-p & top-k sampling. 2025-10-10T01:31:32.3571377Z INFO 10-10 01:31:32 [topk_topp_sampler.py:55] Using FlashInfer for top-p & top-k sampling. 2025-10-10T01:31:32.3869631Z (Worker_TP0 pid=3715) INFO 10-10 01:31:32 [gpu_model_runner.py:2707] Starting to load model meta-llama/Llama-3.2-1B-Instruct... 2025-10-10T01:31:32.3876536Z (Worker_TP1 pid=3716) INFO 10-10 01:31:32 [gpu_model_runner.py:2707] Starting to load model meta-llama/Llama-3.2-1B-Instruct... 2025-10-10T01:31:32.6097757Z (Worker_TP1 pid=3716) INFO 10-10 01:31:32 [gpu_model_runner.py:2739] Loading model from scratch... 2025-10-10T01:31:32.6189352Z (Worker_TP0 pid=3715) INFO 10-10 01:31:32 [gpu_model_runner.py:2739] Loading model from scratch... 2025-10-10T01:31:32.6995758Z (Worker_TP1 pid=3716) INFO 10-10 01:31:32 [cuda.py:361] Using Flash Attention backend on V1 engine. 2025-10-10T01:31:32.7091989Z (Worker_TP0 pid=3715) INFO 10-10 01:31:32 [cuda.py:361] Using Flash Attention backend on V1 engine. 2025-10-10T01:31:33.2311224Z (Worker_TP0 pid=3715) INFO 10-10 01:31:33 [gpu_model_runner.py:2758] Model loading took 1.1667 GiB and 0.167768 seconds 2025-10-10T01:31:33.2360403Z (Worker_TP1 pid=3716) INFO 10-10 01:31:33 [gpu_model_runner.py:2758] Model loading took 1.1667 GiB and 0.166872 seconds 2025-10-10T01:31:36.7419089Z (Worker_TP1 pid=3716) INFO 10-10 01:31:36 [backends.py:548] Using cache directory: /var/lib/jenkins/.cache/vllm/torch_compile_cache/387e456a16/rank_1_0/backbone for vLLM's torch.compile 2025-10-10T01:31:36.7422502Z (Worker_TP1 pid=3716) INFO 10-10 01:31:36 [backends.py:559] Dynamo bytecode transform time: 3.19 s 2025-10-10T01:31:36.8794798Z (Worker_TP0 pid=3715) INFO 10-10 01:31:36 [backends.py:548] Using cache directory: /var/lib/jenkins/.cache/vllm/torch_compile_cache/387e456a16/rank_0_0/backbone for vLLM's torch.compile 2025-10-10T01:31:36.8799317Z (Worker_TP0 pid=3715) INFO 10-10 01:31:36 [backends.py:559] Dynamo bytecode transform time: 3.34 s 2025-10-10T01:31:37.8825384Z (Worker_TP1 pid=3716) [rank1]:W1010 01:31:37.881000 3716 site-packages/torch/_inductor/utils.py:1594] [0/0] Not enough SMs to use max_autotune_gemm mode 2025-10-10T01:31:38.0417890Z (Worker_TP0 pid=3715) [rank0]:W1010 01:31:38.041000 3715 site-packages/torch/_inductor/utils.py:1594] [0/0] Not enough SMs to use max_autotune_gemm mode 2025-10-10T01:31:39.3187035Z (Worker_TP1 pid=3716) INFO 10-10 01:31:39 [backends.py:197] Cache the graph for dynamic shape for later use 2025-10-10T01:31:39.4661679Z (Worker_TP0 pid=3715) INFO 10-10 01:31:39 [backends.py:197] Cache the graph for dynamic shape for later use 2025-10-10T01:31:49.0242915Z (Worker_TP1 pid=3716) INFO 10-10 01:31:49 [backends.py:218] Compiling a graph for dynamic shape takes 12.12 s 2025-10-10T01:31:49.4548081Z (Worker_TP0 pid=3715) INFO 10-10 01:31:49 [backends.py:218] Compiling a graph for dynamic shape takes 12.41 s 2025-10-10T01:31:56.1968544Z (Worker_TP0 pid=3715) INFO 10-10 01:31:56 [monitor.py:32] torch.compile takes 15.75 s in total 2025-10-10T01:31:56.1969328Z (Worker_TP1 pid=3716) INFO 10-10 01:31:56 [monitor.py:32] torch.compile takes 15.32 s in total 2025-10-10T01:31:56.9340004Z (Worker_TP1 pid=3716) INFO 10-10 01:31:56 [gpu_worker.py:298] Available KV cache memory: 18.49 GiB 2025-10-10T01:31:56.9616773Z (Worker_TP0 pid=3715) INFO 10-10 01:31:56 [gpu_worker.py:298] Available KV cache memory: 18.49 GiB 2025-10-10T01:31:57.2804590Z (EngineCore_DP0 pid=3596) INFO 10-10 01:31:57 [kv_cache_utils.py:1087] GPU KV cache size: 1,211,648 tokens 2025-10-10T01:31:57.2805543Z (EngineCore_DP0 pid=3596) INFO 10-10 01:31:57 [kv_cache_utils.py:1091] Maximum concurrency for 2,048 tokens per request: 591.62x 2025-10-10T01:31:57.2806396Z (EngineCore_DP0 pid=3596) INFO 10-10 01:31:57 [kv_cache_utils.py:1087] GPU KV cache size: 1,211,648 tokens 2025-10-10T01:31:57.2807215Z (EngineCore_DP0 pid=3596) INFO 10-10 01:31:57 [kv_cache_utils.py:1091] Maximum concurrency for 2,048 tokens per request: 591.62x 2025-10-10T01:31:57.6069663Z (Worker_TP0 pid=3715) 2025-10-10T01:31:57.7344760Z Capturing CUDA graphs (mixed prefill-decode, PIECEWISE): 0% 0/5 [00:00, 'use_cudagraph': True, 'cudagraph_num_of_warmups': 1, 'cudagraph_capture_sizes': [16, 8, 4, 2], 'cudagraph_copy_inputs': False, 'full_cuda_graph': True, 'use_inductor_graph_partition': False, 'pass_config': {'enable_noop': True, 'enable_sequence_parallelism': True}, 'max_capture_size': 16, 'local_cache_dir': None} 2025-10-10T01:32:20.6997191Z (EngineCore_DP0 pid=4570) 2025-10-10 01:32:20,698 INFO worker.py:1918 -- Started a local Ray instance. View the dashboard at http://127.0.0.1:8265  2025-10-10T01:32:21.6914014Z (EngineCore_DP0 pid=4570) INFO 10-10 01:32:21 [ray_utils.py:345] No current placement group found. Creating a new placement group. 2025-10-10T01:32:21.8675063Z (EngineCore_DP0 pid=4570) INFO 10-10 01:32:21 [ray_distributed_executor.py:171] use_ray_spmd_worker: True 2025-10-10T01:32:22.5040558Z (EngineCore_DP0 pid=4570) (pid=5454) /opt/conda/envs/py_3.12/lib/python3.12/site-packages/torch/cuda/__init__.py:64: FutureWarning: The pynvml package is deprecated. Please install nvidia-ml-py instead. If you did not install pynvml directly, please report this to the maintainers of the package that installed pynvml for you. 2025-10-10T01:32:22.5042267Z (EngineCore_DP0 pid=4570) (pid=5454) import pynvml # type: ignore[import] 2025-10-10T01:32:24.7511605Z (EngineCore_DP0 pid=4570) (pid=5454) INFO 10-10 01:32:24 [__init__.py:215] Automatically detected platform cuda. 2025-10-10T01:32:26.0995488Z (EngineCore_DP0 pid=4570) INFO 10-10 01:32:26 [ray_env.py:63] RAY_NON_CARRY_OVER_ENV_VARS from config: set() 2025-10-10T01:32:26.0999512Z (EngineCore_DP0 pid=4570) INFO 10-10 01:32:26 [ray_env.py:65] Copying the following environment variables to workers: ['VLLM_USE_V1', 'HUGGING_FACE_HUB_TOKEN', 'HF_TOKEN', 'LD_LIBRARY_PATH', 'VLLM_WORKER_MULTIPROC_METHOD', 'MAX_JOBS', 'VLLM_USE_RAY_SPMD_WORKER', 'VLLM_USE_RAY_COMPILED_DAG'] 2025-10-10T01:32:26.1002693Z (EngineCore_DP0 pid=4570) INFO 10-10 01:32:26 [ray_env.py:68] If certain env vars should NOT be copied, add them to /var/lib/jenkins/.config/vllm/ray_non_carry_over_env_vars.json file 2025-10-10T01:32:29.3446296Z (EngineCore_DP0 pid=4570) (RayWorkerWrapper pid=5454) [Gloo] Rank 0 is connected to 1 peer ranks. Expected number of connected peer ranks is : 1 2025-10-10T01:32:29.3447500Z (EngineCore_DP0 pid=4570) (RayWorkerWrapper pid=5454) [Gloo] Rank 0 is connected to 1 peer ranks. Expected number of connected peer ranks is : 1 2025-10-10T01:32:29.3448591Z (EngineCore_DP0 pid=4570) (RayWorkerWrapper pid=5454) INFO 10-10 01:32:29 [__init__.py:1384] Found nccl from library libnccl.so.2 2025-10-10T01:32:29.3449554Z (EngineCore_DP0 pid=4570) (RayWorkerWrapper pid=5454) INFO 10-10 01:32:29 [pynccl.py:104] vLLM is using nccl==2.27.5 2025-10-10T01:32:29.5515562Z (EngineCore_DP0 pid=4570) (RayWorkerWrapper pid=5454) WARNING 10-10 01:32:29 [symm_mem.py:58] SymmMemCommunicator: Device capability 8.9 not supported, communicator is not available. 2025-10-10T01:32:29.5517716Z (EngineCore_DP0 pid=4570) (RayWorkerWrapper pid=5454) INFO 10-10 01:32:29 [custom_all_reduce.py:35] Skipping P2P check and trusting the driver's P2P report. 2025-10-10T01:32:29.5519513Z (EngineCore_DP0 pid=4570) (RayWorkerWrapper pid=5454) WARNING 10-10 01:32:29 [custom_all_reduce.py:154] Custom allreduce is disabled because your platform lacks GPU P2P capability or P2P test failed. To silence this warning, specify disable_custom_all_reduce=True explicitly. 2025-10-10T01:32:29.5521862Z (EngineCore_DP0 pid=4570) (RayWorkerWrapper pid=5454) INFO 10-10 01:32:29 [shm_broadcast.py:289] vLLM message queue communication handle: Handle(local_reader_ranks=[1], buffer_handle=(1, 4194304, 6, 'psm_3dbc7870'), local_subscribe_addr='ipc:///tmp/ffb255f7-f350-4ec8-91a1-1de947f6791b', remote_subscribe_addr=None, remote_addr_ipv6=False) 2025-10-10T01:32:29.6568838Z (EngineCore_DP0 pid=4570) (RayWorkerWrapper pid=5454) INFO 10-10 01:32:29 [parallel_state.py:1208] rank 0 in world size 2 is assigned as DP rank 0, PP rank 0, TP rank 0, EP rank 0 2025-10-10T01:32:29.9654472Z (EngineCore_DP0 pid=4570) (RayWorkerWrapper pid=5454) INFO 10-10 01:32:29 [topk_topp_sampler.py:55] Using FlashInfer for top-p & top-k sampling. 2025-10-10T01:32:29.9655713Z (EngineCore_DP0 pid=4570) (RayWorkerWrapper pid=5454) INFO 10-10 01:32:29 [gpu_model_runner.py:2707] Starting to load model meta-llama/Llama-3.2-1B-Instruct... 2025-10-10T01:32:29.9656717Z (EngineCore_DP0 pid=4570) (pid=5456) INFO 10-10 01:32:24 [__init__.py:215] Automatically detected platform cuda. 2025-10-10T01:32:30.2735905Z (EngineCore_DP0 pid=4570) (RayWorkerWrapper pid=5454) INFO 10-10 01:32:30 [gpu_model_runner.py:2739] Loading model from scratch... 2025-10-10T01:32:30.2737004Z (EngineCore_DP0 pid=4570) (RayWorkerWrapper pid=5454) INFO 10-10 01:32:30 [cuda.py:361] Using Flash Attention backend on V1 engine. 2025-10-10T01:32:30.8883822Z (EngineCore_DP0 pid=4570) (RayWorkerWrapper pid=5454) INFO 10-10 01:32:30 [gpu_model_runner.py:2758] Model loading took 1.1667 GiB and 0.165869 seconds 2025-10-10T01:32:34.1552434Z (EngineCore_DP0 pid=4570) (RayWorkerWrapper pid=5454) INFO 10-10 01:32:34 [backends.py:548] Using cache directory: /var/lib/jenkins/.cache/vllm/torch_compile_cache/31e0bc77eb/rank_0_0/backbone for vLLM's torch.compile 2025-10-10T01:32:34.1553863Z (EngineCore_DP0 pid=4570) (RayWorkerWrapper pid=5454) INFO 10-10 01:32:34 [backends.py:559] Dynamo bytecode transform time: 2.95 s 2025-10-10T01:32:35.1778715Z (EngineCore_DP0 pid=4570) (RayWorkerWrapper pid=5456) INFO 10-10 01:32:35 [backends.py:164] Directly load the compiled graph(s) for dynamic shape from the cache, took 0.798 s 2025-10-10T01:32:35.1781527Z (EngineCore_DP0 pid=4570) (RayWorkerWrapper pid=5456) [Gloo] Rank 1 is connected to 1 peer ranks. Expected number of connected peer ranks is : 1 [repeated 10x across cluster] (Ray deduplicates logs by default. Set RAY_DEDUP_LOGS=0 to disable log deduplication, or see https://docs.ray.io/en/master/ray-observability/user-guides/configure-logging.html#log-deduplication for more options.) 2025-10-10T01:32:35.1783612Z (EngineCore_DP0 pid=4570) (RayWorkerWrapper pid=5456) INFO 10-10 01:32:29 [__init__.py:1384] Found nccl from library libnccl.so.2 [repeated 3x across cluster] 2025-10-10T01:32:35.1784775Z (EngineCore_DP0 pid=4570) (RayWorkerWrapper pid=5456) INFO 10-10 01:32:29 [pynccl.py:104] vLLM is using nccl==2.27.5 [repeated 3x across cluster] 2025-10-10T01:32:35.1786002Z (EngineCore_DP0 pid=4570) (RayWorkerWrapper pid=5456) WARNING 10-10 01:32:29 [symm_mem.py:58] SymmMemCommunicator: Device capability 8.9 not supported, communicator is not available. 2025-10-10T01:32:35.1787023Z (EngineCore_DP0 pid=4570) (RayWorkerWrapper pid=5456) INFO 10-10 01:32:29 [custom_all_reduce.py:35] Skipping P2P check and trusting the driver's P2P report. 2025-10-10T01:32:35.1788331Z (EngineCore_DP0 pid=4570) (RayWorkerWrapper pid=5456) WARNING 10-10 01:32:29 [custom_all_reduce.py:154] Custom allreduce is disabled because your platform lacks GPU P2P capability or P2P test failed. To silence this warning, specify disable_custom_all_reduce=True explicitly. 2025-10-10T01:32:35.1789636Z (EngineCore_DP0 pid=4570) (RayWorkerWrapper pid=5456) INFO 10-10 01:32:29 [parallel_state.py:1208] rank 1 in world size 2 is assigned as DP rank 0, PP rank 0, TP rank 1, EP rank 1 2025-10-10T01:32:35.1790570Z (EngineCore_DP0 pid=4570) (RayWorkerWrapper pid=5456) INFO 10-10 01:32:29 [topk_topp_sampler.py:55] Using FlashInfer for top-p & top-k sampling. 2025-10-10T01:32:35.1791482Z (EngineCore_DP0 pid=4570) (RayWorkerWrapper pid=5456) INFO 10-10 01:32:29 [gpu_model_runner.py:2707] Starting to load model meta-llama/Llama-3.2-1B-Instruct... 2025-10-10T01:32:36.5074674Z (EngineCore_DP0 pid=4570) (RayWorkerWrapper pid=5454) INFO 10-10 01:32:36 [gpu_worker.py:298] Available KV cache memory: 18.54 GiB 2025-10-10T01:32:36.5075733Z (EngineCore_DP0 pid=4570) (RayWorkerWrapper pid=5456) INFO 10-10 01:32:30 [gpu_model_runner.py:2739] Loading model from scratch... 2025-10-10T01:32:36.5076531Z (EngineCore_DP0 pid=4570) (RayWorkerWrapper pid=5456) INFO 10-10 01:32:30 [cuda.py:361] Using Flash Attention backend on V1 engine. 2025-10-10T01:32:36.5077406Z (EngineCore_DP0 pid=4570) (RayWorkerWrapper pid=5456) INFO 10-10 01:32:30 [gpu_model_runner.py:2758] Model loading took 1.1667 GiB and 0.164817 seconds 2025-10-10T01:32:36.7551139Z (EngineCore_DP0 pid=4570) INFO 10-10 01:32:36 [kv_cache_utils.py:1087] GPU KV cache size: 1,215,360 tokens 2025-10-10T01:32:36.7552546Z (EngineCore_DP0 pid=4570) INFO 10-10 01:32:36 [kv_cache_utils.py:1091] Maximum concurrency for 2,048 tokens per request: 593.44x 2025-10-10T01:32:36.7553844Z (EngineCore_DP0 pid=4570) INFO 10-10 01:32:36 [kv_cache_utils.py:1087] GPU KV cache size: 1,215,360 tokens 2025-10-10T01:32:36.7555151Z (EngineCore_DP0 pid=4570) INFO 10-10 01:32:36 [kv_cache_utils.py:1091] Maximum concurrency for 2,048 tokens per request: 593.44x 2025-10-10T01:32:37.1215869Z (EngineCore_DP0 pid=4570) (RayWorkerWrapper pid=5454) 2025-10-10T01:32:37.1216520Z Capturing CUDA graphs (mixed prefill-decode, PIECEWISE): 0%| | 0/4 [00:00, 'use_cudagraph': True, 'cudagraph_num_of_warmups': 1, 'cudagraph_capture_sizes': [16, 8, 4, 2, 1], 'cudagraph_copy_inputs': False, 'full_cuda_graph': True, 'use_inductor_graph_partition': False, 'pass_config': {}, 'max_capture_size': 16, 'local_cache_dir': None} 2025-10-10T01:33:02.0359726Z (EngineCore_DP0 pid=9433) WARNING 10-10 01:33:02 [multiproc_executor.py:720] Reducing Torch parallelism from 24 threads to 1 to avoid unnecessary CPU contention. Set OMP_NUM_THREADS in the external environment to tune this value as needed. 2025-10-10T01:33:02.0368342Z (EngineCore_DP0 pid=9433) INFO 10-10 01:33:02 [shm_broadcast.py:289] vLLM message queue communication handle: Handle(local_reader_ranks=[0, 1], buffer_handle=(2, 16777216, 10, 'psm_640246ff'), local_subscribe_addr='ipc:///tmp/8c9dbf79-6ec2-4e51-91ec-8d0d9f3d3c1b', remote_subscribe_addr=None, remote_addr_ipv6=False) 2025-10-10T01:33:02.5233316Z /opt/conda/envs/py_3.12/lib/python3.12/site-packages/torch/cuda/__init__.py:64: FutureWarning: The pynvml package is deprecated. Please install nvidia-ml-py instead. If you did not install pynvml directly, please report this to the maintainers of the package that installed pynvml for you. 2025-10-10T01:33:02.5235370Z import pynvml # type: ignore[import] 2025-10-10T01:33:02.5236530Z /opt/conda/envs/py_3.12/lib/python3.12/site-packages/torch/cuda/__init__.py:64: FutureWarning: The pynvml package is deprecated. Please install nvidia-ml-py instead. If you did not install pynvml directly, please report this to the maintainers of the package that installed pynvml for you. 2025-10-10T01:33:02.5237470Z import pynvml # type: ignore[import] 2025-10-10T01:33:04.8559557Z INFO 10-10 01:33:04 [__init__.py:215] Automatically detected platform cuda. 2025-10-10T01:33:04.8630981Z INFO 10-10 01:33:04 [__init__.py:215] Automatically detected platform cuda. 2025-10-10T01:33:09.4042417Z INFO 10-10 01:33:09 [shm_broadcast.py:289] vLLM message queue communication handle: Handle(local_reader_ranks=[0], buffer_handle=(1, 10485760, 10, 'psm_9686336e'), local_subscribe_addr='ipc:///tmp/6c530120-a30e-40ba-8379-ddf67b825a7c', remote_subscribe_addr=None, remote_addr_ipv6=False) 2025-10-10T01:33:09.5041590Z INFO 10-10 01:33:09 [shm_broadcast.py:289] vLLM message queue communication handle: Handle(local_reader_ranks=[0], buffer_handle=(1, 10485760, 10, 'psm_25f23905'), local_subscribe_addr='ipc:///tmp/dd062902-dec9-430f-8830-78634a6eb3db', remote_subscribe_addr=None, remote_addr_ipv6=False) 2025-10-10T01:33:10.1940041Z [Gloo] Rank 0 is connected to 1 peer ranks. Expected number of connected peer ranks is : 1 2025-10-10T01:33:10.1941013Z [Gloo] Rank 1 is connected to 1 peer ranks. Expected number of connected peer ranks is : 1 2025-10-10T01:33:10.1988591Z [Gloo] Rank 1 is connected to 1 peer ranks. Expected number of connected peer ranks is : 1 2025-10-10T01:33:10.1989607Z [Gloo] Rank 0 is connected to 1 peer ranks. Expected number of connected peer ranks is : 1 2025-10-10T01:33:10.2032163Z INFO 10-10 01:33:10 [__init__.py:1384] Found nccl from library libnccl.so.2 2025-10-10T01:33:10.2032930Z INFO 10-10 01:33:10 [__init__.py:1384] Found nccl from library libnccl.so.2 2025-10-10T01:33:10.2034494Z INFO 10-10 01:33:10 [pynccl.py:104] vLLM is using nccl==2.27.5 2025-10-10T01:33:10.2035134Z INFO 10-10 01:33:10 [pynccl.py:104] vLLM is using nccl==2.27.5 2025-10-10T01:33:10.3787673Z WARNING 10-10 01:33:10 [symm_mem.py:58] SymmMemCommunicator: Device capability 8.9 not supported, communicator is not available. 2025-10-10T01:33:10.3788804Z WARNING 10-10 01:33:10 [symm_mem.py:58] SymmMemCommunicator: Device capability 8.9 not supported, communicator is not available. 2025-10-10T01:33:10.3813397Z INFO 10-10 01:33:10 [custom_all_reduce.py:35] Skipping P2P check and trusting the driver's P2P report. 2025-10-10T01:33:10.3815415Z INFO 10-10 01:33:10 [custom_all_reduce.py:35] Skipping P2P check and trusting the driver's P2P report. 2025-10-10T01:33:10.3817596Z WARNING 10-10 01:33:10 [custom_all_reduce.py:154] Custom allreduce is disabled because your platform lacks GPU P2P capability or P2P test failed. To silence this warning, specify disable_custom_all_reduce=True explicitly. 2025-10-10T01:33:10.3818993Z WARNING 10-10 01:33:10 [custom_all_reduce.py:154] Custom allreduce is disabled because your platform lacks GPU P2P capability or P2P test failed. To silence this warning, specify disable_custom_all_reduce=True explicitly. 2025-10-10T01:33:10.4741067Z INFO 10-10 01:33:10 [shm_broadcast.py:289] vLLM message queue communication handle: Handle(local_reader_ranks=[1], buffer_handle=(1, 4194304, 6, 'psm_f2765b4e'), local_subscribe_addr='ipc:///tmp/f706dedb-fd48-4948-9752-fdad47d588f4', remote_subscribe_addr=None, remote_addr_ipv6=False) 2025-10-10T01:33:10.4772427Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-10-10T01:33:10.4773678Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-10-10T01:33:10.4781076Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-10-10T01:33:10.4789092Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-10-10T01:33:10.5335649Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-10-10T01:33:10.5347204Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-10-10T01:33:10.5360242Z [Gloo] Rank 0 is connected to 1 peer ranks. Expected number of connected peer ranks is : 1 2025-10-10T01:33:10.5361038Z [Gloo] Rank 1 is connected to 1 peer ranks. Expected number of connected peer ranks is : 1 2025-10-10T01:33:10.5361998Z INFO 10-10 01:33:10 [__init__.py:1384] Found nccl from library libnccl.so.2 2025-10-10T01:33:10.5362605Z INFO 10-10 01:33:10 [__init__.py:1384] Found nccl from library libnccl.so.2 2025-10-10T01:33:10.5363137Z INFO 10-10 01:33:10 [pynccl.py:104] vLLM is using nccl==2.27.5 2025-10-10T01:33:10.5363436Z INFO 10-10 01:33:10 [pynccl.py:104] vLLM is using nccl==2.27.5 2025-10-10T01:33:10.5696395Z INFO 10-10 01:33:10 [parallel_state.py:1208] rank 0 in world size 2 is assigned as DP rank 0, PP rank 0, TP rank 0, EP rank 0 2025-10-10T01:33:10.5697514Z INFO 10-10 01:33:10 [parallel_state.py:1208] rank 1 in world size 2 is assigned as DP rank 0, PP rank 0, TP rank 1, EP rank 1 2025-10-10T01:33:10.8040180Z INFO 10-10 01:33:10 [topk_topp_sampler.py:55] Using FlashInfer for top-p & top-k sampling. 2025-10-10T01:33:10.8053997Z INFO 10-10 01:33:10 [topk_topp_sampler.py:55] Using FlashInfer for top-p & top-k sampling. 2025-10-10T01:33:10.8355748Z (Worker_TP1 pid=9553) INFO 10-10 01:33:10 [gpu_model_runner.py:2707] Starting to load model meta-llama/Llama-3.2-1B-Instruct... 2025-10-10T01:33:10.8376921Z (Worker_TP0 pid=9552) INFO 10-10 01:33:10 [gpu_model_runner.py:2707] Starting to load model meta-llama/Llama-3.2-1B-Instruct... 2025-10-10T01:33:11.0588030Z (Worker_TP1 pid=9553) INFO 10-10 01:33:11 [gpu_model_runner.py:2739] Loading model from scratch... 2025-10-10T01:33:11.0602286Z (Worker_TP0 pid=9552) INFO 10-10 01:33:11 [gpu_model_runner.py:2739] Loading model from scratch... 2025-10-10T01:33:11.1487784Z (Worker_TP1 pid=9553) INFO 10-10 01:33:11 [cuda.py:361] Using Flash Attention backend on V1 engine. 2025-10-10T01:33:11.1507850Z (Worker_TP0 pid=9552) INFO 10-10 01:33:11 [cuda.py:361] Using Flash Attention backend on V1 engine. 2025-10-10T01:33:11.6739555Z (Worker_TP1 pid=9553) INFO 10-10 01:33:11 [gpu_model_runner.py:2758] Model loading took 1.1667 GiB and 0.168290 seconds 2025-10-10T01:33:11.6753792Z (Worker_TP0 pid=9552) INFO 10-10 01:33:11 [gpu_model_runner.py:2758] Model loading took 1.1667 GiB and 0.168559 seconds 2025-10-10T01:33:15.1521679Z (Worker_TP1 pid=9553) INFO 10-10 01:33:15 [backends.py:548] Using cache directory: /var/lib/jenkins/.cache/vllm/torch_compile_cache/387e456a16/rank_1_0/backbone for vLLM's torch.compile 2025-10-10T01:33:15.1527130Z (Worker_TP1 pid=9553) INFO 10-10 01:33:15 [backends.py:559] Dynamo bytecode transform time: 3.22 s 2025-10-10T01:33:15.2307752Z (Worker_TP0 pid=9552) INFO 10-10 01:33:15 [backends.py:548] Using cache directory: /var/lib/jenkins/.cache/vllm/torch_compile_cache/387e456a16/rank_0_0/backbone for vLLM's torch.compile 2025-10-10T01:33:15.2313998Z (Worker_TP0 pid=9552) INFO 10-10 01:33:15 [backends.py:559] Dynamo bytecode transform time: 3.30 s 2025-10-10T01:33:16.2745625Z (Worker_TP1 pid=9553) INFO 10-10 01:33:16 [backends.py:164] Directly load the compiled graph(s) for dynamic shape from the cache, took 0.961 s 2025-10-10T01:33:16.3817641Z (Worker_TP0 pid=9552) INFO 10-10 01:33:16 [backends.py:164] Directly load the compiled graph(s) for dynamic shape from the cache, took 0.990 s 2025-10-10T01:33:16.9314550Z (Worker_TP1 pid=9553) INFO 10-10 01:33:16 [monitor.py:32] torch.compile takes 3.22 s in total 2025-10-10T01:33:16.9315795Z (Worker_TP0 pid=9552) INFO 10-10 01:33:16 [monitor.py:32] torch.compile takes 3.30 s in total 2025-10-10T01:33:17.6242932Z (Worker_TP1 pid=9553) INFO 10-10 01:33:17 [gpu_worker.py:298] Available KV cache memory: 18.54 GiB 2025-10-10T01:33:17.6450602Z (Worker_TP0 pid=9552) INFO 10-10 01:33:17 [gpu_worker.py:298] Available KV cache memory: 18.54 GiB 2025-10-10T01:33:17.9083518Z (EngineCore_DP0 pid=9433) INFO 10-10 01:33:17 [kv_cache_utils.py:1087] GPU KV cache size: 1,214,720 tokens 2025-10-10T01:33:17.9084980Z (EngineCore_DP0 pid=9433) INFO 10-10 01:33:17 [kv_cache_utils.py:1091] Maximum concurrency for 2,048 tokens per request: 593.12x 2025-10-10T01:33:17.9085888Z (EngineCore_DP0 pid=9433) INFO 10-10 01:33:17 [kv_cache_utils.py:1087] GPU KV cache size: 1,214,720 tokens 2025-10-10T01:33:17.9086742Z (EngineCore_DP0 pid=9433) INFO 10-10 01:33:17 [kv_cache_utils.py:1091] Maximum concurrency for 2,048 tokens per request: 593.12x 2025-10-10T01:33:18.1801841Z (Worker_TP0 pid=9552) 2025-10-10T01:33:18.3042372Z Capturing CUDA graphs (mixed prefill-decode, PIECEWISE): 0% 0/5 [00:00, 'use_cudagraph': True, 'cudagraph_num_of_warmups': 1, 'cudagraph_capture_sizes': [16, 8, 4, 2], 'cudagraph_copy_inputs': False, 'full_cuda_graph': True, 'use_inductor_graph_partition': False, 'pass_config': {'enable_noop': True, 'enable_sequence_parallelism': True}, 'max_capture_size': 16, 'local_cache_dir': None} 2025-10-10T01:33:38.1561516Z (EngineCore_DP0 pid=9929) WARNING 10-10 01:33:38 [multiproc_executor.py:720] Reducing Torch parallelism from 24 threads to 1 to avoid unnecessary CPU contention. Set OMP_NUM_THREADS in the external environment to tune this value as needed. 2025-10-10T01:33:38.1565499Z (EngineCore_DP0 pid=9929) INFO 10-10 01:33:38 [shm_broadcast.py:289] vLLM message queue communication handle: Handle(local_reader_ranks=[0, 1], buffer_handle=(2, 16777216, 10, 'psm_b92b4c5e'), local_subscribe_addr='ipc:///tmp/d101e5a9-1ce3-4c1c-829f-c551d9b5dad7', remote_subscribe_addr=None, remote_addr_ipv6=False) 2025-10-10T01:33:38.6441430Z /opt/conda/envs/py_3.12/lib/python3.12/site-packages/torch/cuda/__init__.py:64: FutureWarning: The pynvml package is deprecated. Please install nvidia-ml-py instead. If you did not install pynvml directly, please report this to the maintainers of the package that installed pynvml for you. 2025-10-10T01:33:38.6443782Z import pynvml # type: ignore[import] 2025-10-10T01:33:38.6446035Z /opt/conda/envs/py_3.12/lib/python3.12/site-packages/torch/cuda/__init__.py:64: FutureWarning: The pynvml package is deprecated. Please install nvidia-ml-py instead. If you did not install pynvml directly, please report this to the maintainers of the package that installed pynvml for you. 2025-10-10T01:33:38.6447808Z import pynvml # type: ignore[import] 2025-10-10T01:33:40.9721328Z INFO 10-10 01:33:40 [__init__.py:215] Automatically detected platform cuda. 2025-10-10T01:33:40.9767730Z INFO 10-10 01:33:40 [__init__.py:215] Automatically detected platform cuda. 2025-10-10T01:33:45.5041053Z INFO 10-10 01:33:45 [shm_broadcast.py:289] vLLM message queue communication handle: Handle(local_reader_ranks=[0], buffer_handle=(1, 10485760, 10, 'psm_7eb0c306'), local_subscribe_addr='ipc:///tmp/50df92b4-a998-49e4-9be1-eff6dc5e71dd', remote_subscribe_addr=None, remote_addr_ipv6=False) 2025-10-10T01:33:45.6340334Z INFO 10-10 01:33:45 [shm_broadcast.py:289] vLLM message queue communication handle: Handle(local_reader_ranks=[0], buffer_handle=(1, 10485760, 10, 'psm_1e395dc9'), local_subscribe_addr='ipc:///tmp/7e7a45d3-0da8-49ab-a5f0-9417c48fca6e', remote_subscribe_addr=None, remote_addr_ipv6=False) 2025-10-10T01:33:46.0556944Z [Gloo] Rank 0 is connected to 1 peer ranks. Expected number of connected peer ranks is : 1 2025-10-10T01:33:46.0557574Z [Gloo] Rank 1 is connected to 1 peer ranks. Expected number of connected peer ranks is : 1 2025-10-10T01:33:46.0602641Z [Gloo] Rank 0 is connected to 1 peer ranks. Expected number of connected peer ranks is : 1 2025-10-10T01:33:46.0603254Z [Gloo] Rank 1 is connected to 1 peer ranks. Expected number of connected peer ranks is : 1 2025-10-10T01:33:46.0655617Z INFO 10-10 01:33:46 [__init__.py:1384] Found nccl from library libnccl.so.2 2025-10-10T01:33:46.0656111Z INFO 10-10 01:33:46 [__init__.py:1384] Found nccl from library libnccl.so.2 2025-10-10T01:33:46.0659566Z INFO 10-10 01:33:46 [pynccl.py:104] vLLM is using nccl==2.27.5 2025-10-10T01:33:46.0660452Z INFO 10-10 01:33:46 [pynccl.py:104] vLLM is using nccl==2.27.5 2025-10-10T01:33:46.2573965Z WARNING 10-10 01:33:46 [symm_mem.py:58] SymmMemCommunicator: Device capability 8.9 not supported, communicator is not available. 2025-10-10T01:33:46.2575669Z WARNING 10-10 01:33:46 [symm_mem.py:58] SymmMemCommunicator: Device capability 8.9 not supported, communicator is not available. 2025-10-10T01:33:46.2595015Z INFO 10-10 01:33:46 [custom_all_reduce.py:35] Skipping P2P check and trusting the driver's P2P report. 2025-10-10T01:33:46.2596252Z INFO 10-10 01:33:46 [custom_all_reduce.py:35] Skipping P2P check and trusting the driver's P2P report. 2025-10-10T01:33:46.2598948Z WARNING 10-10 01:33:46 [custom_all_reduce.py:154] Custom allreduce is disabled because your platform lacks GPU P2P capability or P2P test failed. To silence this warning, specify disable_custom_all_reduce=True explicitly. 2025-10-10T01:33:46.2600733Z WARNING 10-10 01:33:46 [custom_all_reduce.py:154] Custom allreduce is disabled because your platform lacks GPU P2P capability or P2P test failed. To silence this warning, specify disable_custom_all_reduce=True explicitly. 2025-10-10T01:33:46.3337774Z INFO 10-10 01:33:46 [shm_broadcast.py:289] vLLM message queue communication handle: Handle(local_reader_ranks=[1], buffer_handle=(1, 4194304, 6, 'psm_e4e321e0'), local_subscribe_addr='ipc:///tmp/42823284-bcdc-41a6-8415-df3da7390cc8', remote_subscribe_addr=None, remote_addr_ipv6=False) 2025-10-10T01:33:46.3367411Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-10-10T01:33:46.3367903Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-10-10T01:33:46.3377089Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-10-10T01:33:46.3384703Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-10-10T01:33:46.4236849Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-10-10T01:33:46.4243935Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-10-10T01:33:46.4256655Z [Gloo] Rank 1 is connected to 1 peer ranks. Expected number of connected peer ranks is : 1 2025-10-10T01:33:46.4257835Z [Gloo] Rank 0 is connected to 1 peer ranks. Expected number of connected peer ranks is : 1 2025-10-10T01:33:46.4259595Z INFO 10-10 01:33:46 [__init__.py:1384] Found nccl from library libnccl.so.2 2025-10-10T01:33:46.4261173Z INFO 10-10 01:33:46 [pynccl.py:104] vLLM is using nccl==2.27.5 2025-10-10T01:33:46.4261695Z INFO 10-10 01:33:46 [__init__.py:1384] Found nccl from library libnccl.so.2 2025-10-10T01:33:46.4262106Z INFO 10-10 01:33:46 [pynccl.py:104] vLLM is using nccl==2.27.5 2025-10-10T01:33:46.4588696Z INFO 10-10 01:33:46 [parallel_state.py:1208] rank 0 in world size 2 is assigned as DP rank 0, PP rank 0, TP rank 0, EP rank 0 2025-10-10T01:33:46.4589499Z INFO 10-10 01:33:46 [parallel_state.py:1208] rank 1 in world size 2 is assigned as DP rank 0, PP rank 0, TP rank 1, EP rank 1 2025-10-10T01:33:46.7227398Z INFO 10-10 01:33:46 [topk_topp_sampler.py:55] Using FlashInfer for top-p & top-k sampling. 2025-10-10T01:33:46.7245474Z INFO 10-10 01:33:46 [topk_topp_sampler.py:55] Using FlashInfer for top-p & top-k sampling. 2025-10-10T01:33:46.7547055Z (Worker_TP1 pid=10049) INFO 10-10 01:33:46 [gpu_model_runner.py:2707] Starting to load model meta-llama/Llama-3.2-1B-Instruct... 2025-10-10T01:33:46.7561267Z (Worker_TP0 pid=10048) INFO 10-10 01:33:46 [gpu_model_runner.py:2707] Starting to load model meta-llama/Llama-3.2-1B-Instruct... 2025-10-10T01:33:46.9806569Z (Worker_TP1 pid=10049) INFO 10-10 01:33:46 [gpu_model_runner.py:2739] Loading model from scratch... 2025-10-10T01:33:46.9867775Z (Worker_TP0 pid=10048) INFO 10-10 01:33:46 [gpu_model_runner.py:2739] Loading model from scratch... 2025-10-10T01:33:47.0703140Z (Worker_TP1 pid=10049) INFO 10-10 01:33:47 [cuda.py:361] Using Flash Attention backend on V1 engine. 2025-10-10T01:33:47.0767035Z (Worker_TP0 pid=10048) INFO 10-10 01:33:47 [cuda.py:361] Using Flash Attention backend on V1 engine. 2025-10-10T01:33:47.5987903Z (Worker_TP1 pid=10049) INFO 10-10 01:33:47 [gpu_model_runner.py:2758] Model loading took 1.1667 GiB and 0.166662 seconds 2025-10-10T01:33:47.6036394Z (Worker_TP0 pid=10048) INFO 10-10 01:33:47 [gpu_model_runner.py:2758] Model loading took 1.1667 GiB and 0.167234 seconds 2025-10-10T01:33:50.8914804Z (Worker_TP1 pid=10049) INFO 10-10 01:33:50 [backends.py:548] Using cache directory: /var/lib/jenkins/.cache/vllm/torch_compile_cache/31e0bc77eb/rank_1_0/backbone for vLLM's torch.compile 2025-10-10T01:33:50.8924451Z (Worker_TP1 pid=10049) INFO 10-10 01:33:50 [backends.py:559] Dynamo bytecode transform time: 2.97 s 2025-10-10T01:33:50.9512026Z (Worker_TP0 pid=10048) INFO 10-10 01:33:50 [backends.py:548] Using cache directory: /var/lib/jenkins/.cache/vllm/torch_compile_cache/31e0bc77eb/rank_0_0/backbone for vLLM's torch.compile 2025-10-10T01:33:50.9522761Z (Worker_TP0 pid=10048) INFO 10-10 01:33:50 [backends.py:559] Dynamo bytecode transform time: 3.02 s 2025-10-10T01:33:51.9575813Z (Worker_TP1 pid=10049) INFO 10-10 01:33:51 [backends.py:164] Directly load the compiled graph(s) for dynamic shape from the cache, took 0.820 s 2025-10-10T01:33:52.0463280Z (Worker_TP0 pid=10048) INFO 10-10 01:33:52 [backends.py:164] Directly load the compiled graph(s) for dynamic shape from the cache, took 0.843 s 2025-10-10T01:33:53.2835665Z (Worker_TP1 pid=10049) INFO 10-10 01:33:53 [gpu_worker.py:298] Available KV cache memory: 18.54 GiB 2025-10-10T01:33:53.2845989Z (Worker_TP0 pid=10048) INFO 10-10 01:33:53 [gpu_worker.py:298] Available KV cache memory: 18.54 GiB 2025-10-10T01:33:53.5363802Z (EngineCore_DP0 pid=9929) INFO 10-10 01:33:53 [kv_cache_utils.py:1087] GPU KV cache size: 1,215,232 tokens 2025-10-10T01:33:53.5364816Z (EngineCore_DP0 pid=9929) INFO 10-10 01:33:53 [kv_cache_utils.py:1091] Maximum concurrency for 2,048 tokens per request: 593.38x 2025-10-10T01:33:53.5365688Z (EngineCore_DP0 pid=9929) INFO 10-10 01:33:53 [kv_cache_utils.py:1087] GPU KV cache size: 1,215,232 tokens 2025-10-10T01:33:53.5366521Z (EngineCore_DP0 pid=9929) INFO 10-10 01:33:53 [kv_cache_utils.py:1091] Maximum concurrency for 2,048 tokens per request: 593.38x 2025-10-10T01:33:53.7957788Z (Worker_TP0 pid=10048) 2025-10-10T01:33:55.6156358Z Capturing CUDA graphs (mixed prefill-decode, PIECEWISE): 0% 0/4 [00:00, 'use_cudagraph': True, 'cudagraph_num_of_warmups': 1, 'cudagraph_capture_sizes': [16, 8, 4, 2, 1], 'cudagraph_copy_inputs': False, 'full_cuda_graph': True, 'use_inductor_graph_partition': False, 'pass_config': {}, 'max_capture_size': 16, 'local_cache_dir': None} 2025-10-10T01:34:18.8936564Z (EngineCore_DP0 pid=10534) WARNING 10-10 01:34:18 [multiproc_executor.py:720] Reducing Torch parallelism from 24 threads to 1 to avoid unnecessary CPU contention. Set OMP_NUM_THREADS in the external environment to tune this value as needed. 2025-10-10T01:34:18.8944178Z (EngineCore_DP0 pid=10534) INFO 10-10 01:34:18 [shm_broadcast.py:289] vLLM message queue communication handle: Handle(local_reader_ranks=[0, 1], buffer_handle=(2, 16777216, 10, 'psm_6d0b7e77'), local_subscribe_addr='ipc:///tmp/1293d307-cb59-47ee-a4f2-db49b4d96650', remote_subscribe_addr=None, remote_addr_ipv6=False) 2025-10-10T01:34:19.3780335Z /opt/conda/envs/py_3.12/lib/python3.12/site-packages/torch/cuda/__init__.py:64: FutureWarning: The pynvml package is deprecated. Please install nvidia-ml-py instead. If you did not install pynvml directly, please report this to the maintainers of the package that installed pynvml for you. 2025-10-10T01:34:19.3782169Z import pynvml # type: ignore[import] 2025-10-10T01:34:19.3880124Z /opt/conda/envs/py_3.12/lib/python3.12/site-packages/torch/cuda/__init__.py:64: FutureWarning: The pynvml package is deprecated. Please install nvidia-ml-py instead. If you did not install pynvml directly, please report this to the maintainers of the package that installed pynvml for you. 2025-10-10T01:34:19.3882070Z import pynvml # type: ignore[import] 2025-10-10T01:34:21.6734724Z INFO 10-10 01:34:21 [__init__.py:215] Automatically detected platform cuda. 2025-10-10T01:34:21.7114106Z INFO 10-10 01:34:21 [__init__.py:215] Automatically detected platform cuda. 2025-10-10T01:34:26.1840912Z INFO 10-10 01:34:26 [shm_broadcast.py:289] vLLM message queue communication handle: Handle(local_reader_ranks=[0], buffer_handle=(1, 10485760, 10, 'psm_0056b50b'), local_subscribe_addr='ipc:///tmp/3e1da791-4734-4b41-b478-d9b921d686d2', remote_subscribe_addr=None, remote_addr_ipv6=False) 2025-10-10T01:34:26.3141025Z INFO 10-10 01:34:26 [shm_broadcast.py:289] vLLM message queue communication handle: Handle(local_reader_ranks=[0], buffer_handle=(1, 10485760, 10, 'psm_54f904eb'), local_subscribe_addr='ipc:///tmp/56c8efed-e637-4d88-aa93-4d5f4d545998', remote_subscribe_addr=None, remote_addr_ipv6=False) 2025-10-10T01:34:26.4522006Z [Gloo] Rank 0 is connected to 1 peer ranks. Expected number of connected peer ranks is : 1 2025-10-10T01:34:26.4522977Z [Gloo] Rank 1 is connected to 1 peer ranks. Expected number of connected peer ranks is : 1 2025-10-10T01:34:26.4570752Z [Gloo] Rank 1 is connected to 1 peer ranks. Expected number of connected peer ranks is : 1 2025-10-10T01:34:26.4571598Z [Gloo] Rank 0 is connected to 1 peer ranks. Expected number of connected peer ranks is : 1 2025-10-10T01:34:26.4612267Z INFO 10-10 01:34:26 [__init__.py:1384] Found nccl from library libnccl.so.2 2025-10-10T01:34:26.4614967Z INFO 10-10 01:34:26 [pynccl.py:104] vLLM is using nccl==2.27.5 2025-10-10T01:34:26.4615503Z INFO 10-10 01:34:26 [__init__.py:1384] Found nccl from library libnccl.so.2 2025-10-10T01:34:26.4619096Z INFO 10-10 01:34:26 [pynccl.py:104] vLLM is using nccl==2.27.5 2025-10-10T01:34:26.6397086Z WARNING 10-10 01:34:26 [symm_mem.py:58] SymmMemCommunicator: Device capability 8.9 not supported, communicator is not available. 2025-10-10T01:34:26.6398024Z WARNING 10-10 01:34:26 [symm_mem.py:58] SymmMemCommunicator: Device capability 8.9 not supported, communicator is not available. 2025-10-10T01:34:26.6419489Z INFO 10-10 01:34:26 [custom_all_reduce.py:35] Skipping P2P check and trusting the driver's P2P report. 2025-10-10T01:34:26.6420142Z INFO 10-10 01:34:26 [custom_all_reduce.py:35] Skipping P2P check and trusting the driver's P2P report. 2025-10-10T01:34:26.6421162Z WARNING 10-10 01:34:26 [custom_all_reduce.py:154] Custom allreduce is disabled because your platform lacks GPU P2P capability or P2P test failed. To silence this warning, specify disable_custom_all_reduce=True explicitly. 2025-10-10T01:34:26.6422547Z WARNING 10-10 01:34:26 [custom_all_reduce.py:154] Custom allreduce is disabled because your platform lacks GPU P2P capability or P2P test failed. To silence this warning, specify disable_custom_all_reduce=True explicitly. 2025-10-10T01:34:26.7241359Z INFO 10-10 01:34:26 [shm_broadcast.py:289] vLLM message queue communication handle: Handle(local_reader_ranks=[1], buffer_handle=(1, 4194304, 6, 'psm_c0e02684'), local_subscribe_addr='ipc:///tmp/bc6e2538-adfc-4377-b230-9f69ab61c7c4', remote_subscribe_addr=None, remote_addr_ipv6=False) 2025-10-10T01:34:26.7265610Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-10-10T01:34:26.7268567Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-10-10T01:34:26.7273796Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-10-10T01:34:26.7281971Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-10-10T01:34:26.7736589Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-10-10T01:34:26.7747156Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-10-10T01:34:26.7760213Z [Gloo] Rank 1 is connected to 1 peer ranks. Expected number of connected peer ranks is : 1 2025-10-10T01:34:26.7761079Z [Gloo] Rank 0 is connected to 1 peer ranks. Expected number of connected peer ranks is : 1 2025-10-10T01:34:26.7763510Z INFO 10-10 01:34:26 [__init__.py:1384] Found nccl from library libnccl.so.2 2025-10-10T01:34:26.7764358Z INFO 10-10 01:34:26 [pynccl.py:104] vLLM is using nccl==2.27.5 2025-10-10T01:34:26.7765072Z INFO 10-10 01:34:26 [__init__.py:1384] Found nccl from library libnccl.so.2 2025-10-10T01:34:26.7765460Z INFO 10-10 01:34:26 [pynccl.py:104] vLLM is using nccl==2.27.5 2025-10-10T01:34:26.8111761Z INFO 10-10 01:34:26 [parallel_state.py:1208] rank 1 in world size 2 is assigned as DP rank 0, PP rank 0, TP rank 1, EP rank 1 2025-10-10T01:34:26.8112491Z INFO 10-10 01:34:26 [parallel_state.py:1208] rank 0 in world size 2 is assigned as DP rank 0, PP rank 0, TP rank 0, EP rank 0 2025-10-10T01:34:27.0553038Z INFO 10-10 01:34:27 [topk_topp_sampler.py:55] Using FlashInfer for top-p & top-k sampling. 2025-10-10T01:34:27.0644357Z INFO 10-10 01:34:27 [topk_topp_sampler.py:55] Using FlashInfer for top-p & top-k sampling. 2025-10-10T01:34:27.0873789Z (Worker_TP1 pid=10654) INFO 10-10 01:34:27 [gpu_model_runner.py:2707] Starting to load model meta-llama/Llama-3.2-1B-Instruct... 2025-10-10T01:34:27.0963039Z (Worker_TP0 pid=10653) INFO 10-10 01:34:27 [gpu_model_runner.py:2707] Starting to load model meta-llama/Llama-3.2-1B-Instruct... 2025-10-10T01:34:27.3104845Z (Worker_TP1 pid=10654) INFO 10-10 01:34:27 [gpu_model_runner.py:2739] Loading model from scratch... 2025-10-10T01:34:27.3156336Z (Worker_TP0 pid=10653) INFO 10-10 01:34:27 [gpu_model_runner.py:2739] Loading model from scratch... 2025-10-10T01:34:27.4008191Z (Worker_TP1 pid=10654) INFO 10-10 01:34:27 [cuda.py:361] Using Flash Attention backend on V1 engine. 2025-10-10T01:34:27.4061209Z (Worker_TP0 pid=10653) INFO 10-10 01:34:27 [cuda.py:361] Using Flash Attention backend on V1 engine. 2025-10-10T01:34:27.9289094Z (Worker_TP1 pid=10654) INFO 10-10 01:34:27 [gpu_model_runner.py:2758] Model loading took 1.1667 GiB and 0.168483 seconds 2025-10-10T01:34:27.9345236Z (Worker_TP0 pid=10653) INFO 10-10 01:34:27 [gpu_model_runner.py:2758] Model loading took 1.1667 GiB and 0.167965 seconds 2025-10-10T01:34:31.4487186Z (Worker_TP0 pid=10653) INFO 10-10 01:34:31 [backends.py:548] Using cache directory: /var/lib/jenkins/.cache/vllm/torch_compile_cache/387e456a16/rank_0_0/backbone for vLLM's torch.compile 2025-10-10T01:34:31.4492022Z (Worker_TP0 pid=10653) INFO 10-10 01:34:31 [backends.py:559] Dynamo bytecode transform time: 3.21 s 2025-10-10T01:34:31.5563637Z (Worker_TP1 pid=10654) INFO 10-10 01:34:31 [backends.py:548] Using cache directory: /var/lib/jenkins/.cache/vllm/torch_compile_cache/387e456a16/rank_1_0/backbone for vLLM's torch.compile 2025-10-10T01:34:31.5569729Z (Worker_TP1 pid=10654) INFO 10-10 01:34:31 [backends.py:559] Dynamo bytecode transform time: 3.31 s 2025-10-10T01:34:32.6041575Z (Worker_TP0 pid=10653) INFO 10-10 01:34:32 [backends.py:164] Directly load the compiled graph(s) for dynamic shape from the cache, took 0.993 s 2025-10-10T01:34:32.7216676Z (Worker_TP1 pid=10654) INFO 10-10 01:34:32 [backends.py:164] Directly load the compiled graph(s) for dynamic shape from the cache, took 1.003 s 2025-10-10T01:34:33.2693386Z (Worker_TP0 pid=10653) INFO 10-10 01:34:33 [monitor.py:32] torch.compile takes 3.21 s in total 2025-10-10T01:34:33.2694892Z (Worker_TP1 pid=10654) INFO 10-10 01:34:33 [monitor.py:32] torch.compile takes 3.31 s in total 2025-10-10T01:34:33.9650441Z (Worker_TP0 pid=10653) INFO 10-10 01:34:33 [gpu_worker.py:298] Available KV cache memory: 18.54 GiB 2025-10-10T01:34:33.9784259Z (Worker_TP1 pid=10654) INFO 10-10 01:34:33 [gpu_worker.py:298] Available KV cache memory: 18.54 GiB 2025-10-10T01:34:34.2590121Z (EngineCore_DP0 pid=10534) INFO 10-10 01:34:34 [kv_cache_utils.py:1087] GPU KV cache size: 1,214,720 tokens 2025-10-10T01:34:34.2591569Z (EngineCore_DP0 pid=10534) INFO 10-10 01:34:34 [kv_cache_utils.py:1091] Maximum concurrency for 2,048 tokens per request: 593.12x 2025-10-10T01:34:34.2592434Z (EngineCore_DP0 pid=10534) INFO 10-10 01:34:34 [kv_cache_utils.py:1087] GPU KV cache size: 1,214,720 tokens 2025-10-10T01:34:34.2593293Z (EngineCore_DP0 pid=10534) INFO 10-10 01:34:34 [kv_cache_utils.py:1091] Maximum concurrency for 2,048 tokens per request: 593.12x 2025-10-10T01:34:34.5278577Z (Worker_TP0 pid=10653) 2025-10-10T01:34:34.6643971Z Capturing CUDA graphs (mixed prefill-decode, PIECEWISE): 0% 0/5 [00:00, 'use_cudagraph': True, 'cudagraph_num_of_warmups': 1, 'cudagraph_capture_sizes': [16, 8, 4, 2], 'cudagraph_copy_inputs': False, 'full_cuda_graph': True, 'use_inductor_graph_partition': False, 'pass_config': {'enable_noop': True, 'enable_sequence_parallelism': True}, 'max_capture_size': 16, 'local_cache_dir': None} 2025-10-10T01:34:56.6966126Z (EngineCore_DP0 pid=11030) 2025-10-10 01:34:56,695 INFO worker.py:1918 -- Started a local Ray instance. View the dashboard at http://127.0.0.1:8265  2025-10-10T01:34:57.7200481Z (EngineCore_DP0 pid=11030) INFO 10-10 01:34:57 [ray_utils.py:345] No current placement group found. Creating a new placement group. 2025-10-10T01:34:57.8948881Z (EngineCore_DP0 pid=11030) INFO 10-10 01:34:57 [ray_distributed_executor.py:171] use_ray_spmd_worker: True 2025-10-10T01:34:58.5088547Z (EngineCore_DP0 pid=11030) (pid=11912) /opt/conda/envs/py_3.12/lib/python3.12/site-packages/torch/cuda/__init__.py:64: FutureWarning: The pynvml package is deprecated. Please install nvidia-ml-py instead. If you did not install pynvml directly, please report this to the maintainers of the package that installed pynvml for you. 2025-10-10T01:34:58.5090223Z (EngineCore_DP0 pid=11030) (pid=11912) import pynvml # type: ignore[import] 2025-10-10T01:35:00.7565731Z (EngineCore_DP0 pid=11030) (pid=11912) INFO 10-10 01:35:00 [__init__.py:215] Automatically detected platform cuda. 2025-10-10T01:35:02.1129759Z (EngineCore_DP0 pid=11030) INFO 10-10 01:35:02 [ray_env.py:63] RAY_NON_CARRY_OVER_ENV_VARS from config: set() 2025-10-10T01:35:02.1131296Z (EngineCore_DP0 pid=11030) INFO 10-10 01:35:02 [ray_env.py:65] Copying the following environment variables to workers: ['LD_LIBRARY_PATH', 'VLLM_USE_RAY_COMPILED_DAG', 'HUGGING_FACE_HUB_TOKEN', 'VLLM_USE_RAY_SPMD_WORKER', 'VLLM_USE_V1', 'MAX_JOBS', 'HF_TOKEN', 'VLLM_WORKER_MULTIPROC_METHOD'] 2025-10-10T01:35:02.1132884Z (EngineCore_DP0 pid=11030) INFO 10-10 01:35:02 [ray_env.py:68] If certain env vars should NOT be copied, add them to /var/lib/jenkins/.config/vllm/ray_non_carry_over_env_vars.json file 2025-10-10T01:35:05.3464040Z (EngineCore_DP0 pid=11030) (RayWorkerWrapper pid=11912) [Gloo] Rank 0 is connected to 1 peer ranks. Expected number of connected peer ranks is : 1 2025-10-10T01:35:05.3465541Z (EngineCore_DP0 pid=11030) (RayWorkerWrapper pid=11912) [Gloo] Rank 0 is connected to 1 peer ranks. Expected number of connected peer ranks is : 1 2025-10-10T01:35:05.3467316Z (EngineCore_DP0 pid=11030) (RayWorkerWrapper pid=11912) INFO 10-10 01:35:05 [__init__.py:1384] Found nccl from library libnccl.so.2 2025-10-10T01:35:05.3468288Z (EngineCore_DP0 pid=11030) (RayWorkerWrapper pid=11912) INFO 10-10 01:35:05 [pynccl.py:104] vLLM is using nccl==2.27.5 2025-10-10T01:35:05.5549396Z (EngineCore_DP0 pid=11030) (RayWorkerWrapper pid=11912) WARNING 10-10 01:35:05 [symm_mem.py:58] SymmMemCommunicator: Device capability 8.9 not supported, communicator is not available. 2025-10-10T01:35:05.5550815Z (EngineCore_DP0 pid=11030) (RayWorkerWrapper pid=11912) INFO 10-10 01:35:05 [custom_all_reduce.py:35] Skipping P2P check and trusting the driver's P2P report. 2025-10-10T01:35:05.5552518Z (EngineCore_DP0 pid=11030) (RayWorkerWrapper pid=11912) WARNING 10-10 01:35:05 [custom_all_reduce.py:154] Custom allreduce is disabled because your platform lacks GPU P2P capability or P2P test failed. To silence this warning, specify disable_custom_all_reduce=True explicitly. 2025-10-10T01:35:05.5554900Z (EngineCore_DP0 pid=11030) (RayWorkerWrapper pid=11912) INFO 10-10 01:35:05 [shm_broadcast.py:289] vLLM message queue communication handle: Handle(local_reader_ranks=[1], buffer_handle=(1, 4194304, 6, 'psm_5f64e7d5'), local_subscribe_addr='ipc:///tmp/08582529-8e2e-4bc5-a2dd-8d9b1354af9d', remote_subscribe_addr=None, remote_addr_ipv6=False) 2025-10-10T01:35:05.6598965Z (EngineCore_DP0 pid=11030) (RayWorkerWrapper pid=11912) INFO 10-10 01:35:05 [parallel_state.py:1208] rank 0 in world size 2 is assigned as DP rank 0, PP rank 0, TP rank 0, EP rank 0 2025-10-10T01:35:05.9686860Z (EngineCore_DP0 pid=11030) (RayWorkerWrapper pid=11912) INFO 10-10 01:35:05 [topk_topp_sampler.py:55] Using FlashInfer for top-p & top-k sampling. 2025-10-10T01:35:05.9688386Z (EngineCore_DP0 pid=11030) (RayWorkerWrapper pid=11912) INFO 10-10 01:35:05 [gpu_model_runner.py:2707] Starting to load model meta-llama/Llama-3.2-1B-Instruct... 2025-10-10T01:35:05.9689766Z (EngineCore_DP0 pid=11030) (pid=11918) INFO 10-10 01:35:00 [__init__.py:215] Automatically detected platform cuda. 2025-10-10T01:35:06.2764769Z (EngineCore_DP0 pid=11030) (RayWorkerWrapper pid=11912) INFO 10-10 01:35:06 [gpu_model_runner.py:2739] Loading model from scratch... 2025-10-10T01:35:06.2767681Z (EngineCore_DP0 pid=11030) (RayWorkerWrapper pid=11918) INFO 10-10 01:35:06 [cuda.py:361] Using Flash Attention backend on V1 engine. 2025-10-10T01:35:06.8916932Z (EngineCore_DP0 pid=11030) (RayWorkerWrapper pid=11912) INFO 10-10 01:35:06 [gpu_model_runner.py:2758] Model loading took 1.1667 GiB and 0.168253 seconds 2025-10-10T01:35:10.0553020Z (EngineCore_DP0 pid=11030) (RayWorkerWrapper pid=11918) INFO 10-10 01:35:09 [backends.py:548] Using cache directory: /var/lib/jenkins/.cache/vllm/torch_compile_cache/31e0bc77eb/rank_1_0/backbone for vLLM's torch.compile 2025-10-10T01:35:10.0554481Z (EngineCore_DP0 pid=11030) (RayWorkerWrapper pid=11918) INFO 10-10 01:35:09 [backends.py:559] Dynamo bytecode transform time: 2.89 s 2025-10-10T01:35:11.0792556Z (EngineCore_DP0 pid=11030) (RayWorkerWrapper pid=11918) INFO 10-10 01:35:11 [backends.py:164] Directly load the compiled graph(s) for dynamic shape from the cache, took 0.804 s 2025-10-10T01:35:11.0795412Z (EngineCore_DP0 pid=11030) (RayWorkerWrapper pid=11918) [Gloo] Rank 1 is connected to 1 peer ranks. Expected number of connected peer ranks is : 1 [repeated 10x across cluster] (Ray deduplicates logs by default. Set RAY_DEDUP_LOGS=0 to disable log deduplication, or see https://docs.ray.io/en/master/ray-observability/user-guides/configure-logging.html#log-deduplication for more options.) 2025-10-10T01:35:11.0797909Z (EngineCore_DP0 pid=11030) (RayWorkerWrapper pid=11918) INFO 10-10 01:35:05 [__init__.py:1384] Found nccl from library libnccl.so.2 [repeated 3x across cluster] 2025-10-10T01:35:11.0799208Z (EngineCore_DP0 pid=11030) (RayWorkerWrapper pid=11918) INFO 10-10 01:35:05 [pynccl.py:104] vLLM is using nccl==2.27.5 [repeated 3x across cluster] 2025-10-10T01:35:11.0800476Z (EngineCore_DP0 pid=11030) (RayWorkerWrapper pid=11918) WARNING 10-10 01:35:05 [symm_mem.py:58] SymmMemCommunicator: Device capability 8.9 not supported, communicator is not available. 2025-10-10T01:35:11.0801651Z (EngineCore_DP0 pid=11030) (RayWorkerWrapper pid=11918) INFO 10-10 01:35:05 [custom_all_reduce.py:35] Skipping P2P check and trusting the driver's P2P report. 2025-10-10T01:35:11.0803258Z (EngineCore_DP0 pid=11030) (RayWorkerWrapper pid=11918) WARNING 10-10 01:35:05 [custom_all_reduce.py:154] Custom allreduce is disabled because your platform lacks GPU P2P capability or P2P test failed. To silence this warning, specify disable_custom_all_reduce=True explicitly. 2025-10-10T01:35:11.0804845Z (EngineCore_DP0 pid=11030) (RayWorkerWrapper pid=11918) INFO 10-10 01:35:05 [parallel_state.py:1208] rank 1 in world size 2 is assigned as DP rank 0, PP rank 0, TP rank 1, EP rank 1 2025-10-10T01:35:11.0806428Z (EngineCore_DP0 pid=11030) (RayWorkerWrapper pid=11918) INFO 10-10 01:35:05 [topk_topp_sampler.py:55] Using FlashInfer for top-p & top-k sampling. 2025-10-10T01:35:11.0807496Z (EngineCore_DP0 pid=11030) (RayWorkerWrapper pid=11918) INFO 10-10 01:35:05 [gpu_model_runner.py:2707] Starting to load model meta-llama/Llama-3.2-1B-Instruct... 2025-10-10T01:35:12.5101936Z (EngineCore_DP0 pid=11030) (RayWorkerWrapper pid=11912) INFO 10-10 01:35:12 [gpu_worker.py:298] Available KV cache memory: 18.54 GiB 2025-10-10T01:35:12.5103107Z (EngineCore_DP0 pid=11030) (RayWorkerWrapper pid=11918) INFO 10-10 01:35:06 [gpu_model_runner.py:2739] Loading model from scratch... 2025-10-10T01:35:12.5104132Z (EngineCore_DP0 pid=11030) (RayWorkerWrapper pid=11912) INFO 10-10 01:35:06 [cuda.py:361] Using Flash Attention backend on V1 engine. 2025-10-10T01:35:12.5105638Z (EngineCore_DP0 pid=11030) (RayWorkerWrapper pid=11918) INFO 10-10 01:35:06 [gpu_model_runner.py:2758] Model loading took 1.1667 GiB and 0.168621 seconds 2025-10-10T01:35:12.7346150Z (EngineCore_DP0 pid=11030) INFO 10-10 01:35:12 [kv_cache_utils.py:1087] GPU KV cache size: 1,215,360 tokens 2025-10-10T01:35:12.7347131Z (EngineCore_DP0 pid=11030) INFO 10-10 01:35:12 [kv_cache_utils.py:1091] Maximum concurrency for 2,048 tokens per request: 593.44x 2025-10-10T01:35:12.7347999Z (EngineCore_DP0 pid=11030) INFO 10-10 01:35:12 [kv_cache_utils.py:1087] GPU KV cache size: 1,215,360 tokens 2025-10-10T01:35:12.7348832Z (EngineCore_DP0 pid=11030) INFO 10-10 01:35:12 [kv_cache_utils.py:1091] Maximum concurrency for 2,048 tokens per request: 593.44x 2025-10-10T01:35:13.0223701Z (EngineCore_DP0 pid=11030) (RayWorkerWrapper pid=11912) 2025-10-10T01:35:13.0224596Z Capturing CUDA graphs (mixed prefill-decode, PIECEWISE): 0%| | 0/4 [00:00, 'use_cudagraph': True, 'cudagraph_num_of_warmups': 1, 'cudagraph_capture_sizes': [16, 8, 4, 2, 1], 'cudagraph_copy_inputs': False, 'full_cuda_graph': True, 'use_inductor_graph_partition': False, 'pass_config': {}, 'max_capture_size': 16, 'local_cache_dir': None} 2025-10-10T01:35:34.7454414Z (EngineCore_DP0 pid=15770) WARNING 10-10 01:35:34 [multiproc_executor.py:720] Reducing Torch parallelism from 24 threads to 1 to avoid unnecessary CPU contention. Set OMP_NUM_THREADS in the external environment to tune this value as needed. 2025-10-10T01:35:34.7461098Z (EngineCore_DP0 pid=15770) INFO 10-10 01:35:34 [shm_broadcast.py:289] vLLM message queue communication handle: Handle(local_reader_ranks=[0, 1], buffer_handle=(2, 16777216, 10, 'psm_39033156'), local_subscribe_addr='ipc:///tmp/e3de35fe-78f2-4e23-95ed-d8778784532a', remote_subscribe_addr=None, remote_addr_ipv6=False) 2025-10-10T01:35:35.2260411Z /opt/conda/envs/py_3.12/lib/python3.12/site-packages/torch/cuda/__init__.py:64: FutureWarning: The pynvml package is deprecated. Please install nvidia-ml-py instead. If you did not install pynvml directly, please report this to the maintainers of the package that installed pynvml for you. 2025-10-10T01:35:35.2262722Z import pynvml # type: ignore[import] 2025-10-10T01:35:35.2264982Z /opt/conda/envs/py_3.12/lib/python3.12/site-packages/torch/cuda/__init__.py:64: FutureWarning: The pynvml package is deprecated. Please install nvidia-ml-py instead. If you did not install pynvml directly, please report this to the maintainers of the package that installed pynvml for you. 2025-10-10T01:35:35.2267153Z import pynvml # type: ignore[import] 2025-10-10T01:35:37.5300071Z INFO 10-10 01:35:37 [__init__.py:215] Automatically detected platform cuda. 2025-10-10T01:35:37.5454044Z INFO 10-10 01:35:37 [__init__.py:215] Automatically detected platform cuda. 2025-10-10T01:35:42.1446371Z INFO 10-10 01:35:42 [shm_broadcast.py:289] vLLM message queue communication handle: Handle(local_reader_ranks=[0], buffer_handle=(1, 10485760, 10, 'psm_114a8111'), local_subscribe_addr='ipc:///tmp/e44f5b36-b222-4d46-9b45-6fcebd799ca4', remote_subscribe_addr=None, remote_addr_ipv6=False) 2025-10-10T01:35:42.1938945Z INFO 10-10 01:35:42 [shm_broadcast.py:289] vLLM message queue communication handle: Handle(local_reader_ranks=[0], buffer_handle=(1, 10485760, 10, 'psm_89d6ec20'), local_subscribe_addr='ipc:///tmp/4ae382f9-7270-49da-88ee-e80d4dc58774', remote_subscribe_addr=None, remote_addr_ipv6=False) 2025-10-10T01:35:42.8802529Z [Gloo] Rank 0 is connected to 1 peer ranks. Expected number of connected peer ranks is : 1 2025-10-10T01:35:42.8803762Z [Gloo] Rank 1 is connected to 1 peer ranks. Expected number of connected peer ranks is : 1 2025-10-10T01:35:42.8850576Z [Gloo] Rank 1 is connected to 1 peer ranks. Expected number of connected peer ranks is : 1 2025-10-10T01:35:42.8851831Z [Gloo] Rank 0 is connected to 1 peer ranks. Expected number of connected peer ranks is : 1 2025-10-10T01:35:42.8892526Z INFO 10-10 01:35:42 [__init__.py:1384] Found nccl from library libnccl.so.2 2025-10-10T01:35:42.8893609Z INFO 10-10 01:35:42 [__init__.py:1384] Found nccl from library libnccl.so.2 2025-10-10T01:35:42.8894882Z INFO 10-10 01:35:42 [pynccl.py:104] vLLM is using nccl==2.27.5 2025-10-10T01:35:42.8896341Z INFO 10-10 01:35:42 [pynccl.py:104] vLLM is using nccl==2.27.5 2025-10-10T01:35:43.0607187Z WARNING 10-10 01:35:43 [symm_mem.py:58] SymmMemCommunicator: Device capability 8.9 not supported, communicator is not available. 2025-10-10T01:35:43.0608883Z WARNING 10-10 01:35:43 [symm_mem.py:58] SymmMemCommunicator: Device capability 8.9 not supported, communicator is not available. 2025-10-10T01:35:43.0627729Z INFO 10-10 01:35:43 [custom_all_reduce.py:35] Skipping P2P check and trusting the driver's P2P report. 2025-10-10T01:35:43.0628389Z INFO 10-10 01:35:43 [custom_all_reduce.py:35] Skipping P2P check and trusting the driver's P2P report. 2025-10-10T01:35:43.0629446Z WARNING 10-10 01:35:43 [custom_all_reduce.py:154] Custom allreduce is disabled because your platform lacks GPU P2P capability or P2P test failed. To silence this warning, specify disable_custom_all_reduce=True explicitly. 2025-10-10T01:35:43.0631292Z WARNING 10-10 01:35:43 [custom_all_reduce.py:154] Custom allreduce is disabled because your platform lacks GPU P2P capability or P2P test failed. To silence this warning, specify disable_custom_all_reduce=True explicitly. 2025-10-10T01:35:43.1540377Z INFO 10-10 01:35:43 [shm_broadcast.py:289] vLLM message queue communication handle: Handle(local_reader_ranks=[1], buffer_handle=(1, 4194304, 6, 'psm_9614a1a9'), local_subscribe_addr='ipc:///tmp/a5b2c16a-e26a-4a5e-a126-0013dc5f4314', remote_subscribe_addr=None, remote_addr_ipv6=False) 2025-10-10T01:35:43.1564980Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-10-10T01:35:43.1566149Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-10-10T01:35:43.1571541Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-10-10T01:35:43.1579407Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-10-10T01:35:43.2037385Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-10-10T01:35:43.2046021Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-10-10T01:35:43.2057079Z [Gloo] Rank 1 is connected to 1 peer ranks. Expected number of connected peer ranks is : 1 2025-10-10T01:35:43.2057584Z [Gloo] Rank 0 is connected to 1 peer ranks. Expected number of connected peer ranks is : 1 2025-10-10T01:35:43.2059642Z INFO 10-10 01:35:43 [__init__.py:1384] Found nccl from library libnccl.so.2 2025-10-10T01:35:43.2060036Z INFO 10-10 01:35:43 [__init__.py:1384] Found nccl from library libnccl.so.2 2025-10-10T01:35:43.2060385Z INFO 10-10 01:35:43 [pynccl.py:104] vLLM is using nccl==2.27.5 2025-10-10T01:35:43.2060699Z INFO 10-10 01:35:43 [pynccl.py:104] vLLM is using nccl==2.27.5 2025-10-10T01:35:43.2382756Z INFO 10-10 01:35:43 [parallel_state.py:1208] rank 0 in world size 2 is assigned as DP rank 0, PP rank 0, TP rank 0, EP rank 0 2025-10-10T01:35:43.2383383Z INFO 10-10 01:35:43 [parallel_state.py:1208] rank 1 in world size 2 is assigned as DP rank 0, PP rank 0, TP rank 1, EP rank 1 2025-10-10T01:35:43.4769433Z INFO 10-10 01:35:43 [topk_topp_sampler.py:55] Using FlashInfer for top-p & top-k sampling. 2025-10-10T01:35:43.4789464Z INFO 10-10 01:35:43 [topk_topp_sampler.py:55] Using FlashInfer for top-p & top-k sampling. 2025-10-10T01:35:43.5081683Z (Worker_TP0 pid=15889) INFO 10-10 01:35:43 [gpu_model_runner.py:2707] Starting to load model meta-llama/Llama-3.2-1B-Instruct... 2025-10-10T01:35:43.5105094Z (Worker_TP1 pid=15890) INFO 10-10 01:35:43 [gpu_model_runner.py:2707] Starting to load model meta-llama/Llama-3.2-1B-Instruct... 2025-10-10T01:35:43.7344315Z (Worker_TP0 pid=15889) INFO 10-10 01:35:43 [gpu_model_runner.py:2739] Loading model from scratch... 2025-10-10T01:35:43.7368585Z (Worker_TP1 pid=15890) INFO 10-10 01:35:43 [gpu_model_runner.py:2739] Loading model from scratch... 2025-10-10T01:35:43.8249125Z (Worker_TP0 pid=15889) INFO 10-10 01:35:43 [cuda.py:361] Using Flash Attention backend on V1 engine. 2025-10-10T01:35:43.8274989Z (Worker_TP1 pid=15890) INFO 10-10 01:35:43 [cuda.py:361] Using Flash Attention backend on V1 engine. 2025-10-10T01:35:44.3468196Z (Worker_TP0 pid=15889) INFO 10-10 01:35:44 [gpu_model_runner.py:2758] Model loading took 1.1667 GiB and 0.166101 seconds 2025-10-10T01:35:44.3544891Z (Worker_TP1 pid=15890) INFO 10-10 01:35:44 [gpu_model_runner.py:2758] Model loading took 1.1667 GiB and 0.168025 seconds 2025-10-10T01:35:47.8865449Z (Worker_TP1 pid=15890) INFO 10-10 01:35:47 [backends.py:548] Using cache directory: /var/lib/jenkins/.cache/vllm/torch_compile_cache/387e456a16/rank_1_0/backbone for vLLM's torch.compile 2025-10-10T01:35:47.8869776Z (Worker_TP1 pid=15890) INFO 10-10 01:35:47 [backends.py:559] Dynamo bytecode transform time: 3.22 s 2025-10-10T01:35:47.9543326Z (Worker_TP0 pid=15889) INFO 10-10 01:35:47 [backends.py:548] Using cache directory: /var/lib/jenkins/.cache/vllm/torch_compile_cache/387e456a16/rank_0_0/backbone for vLLM's torch.compile 2025-10-10T01:35:47.9549336Z (Worker_TP0 pid=15889) INFO 10-10 01:35:47 [backends.py:559] Dynamo bytecode transform time: 3.29 s 2025-10-10T01:35:49.0533244Z (Worker_TP1 pid=15890) INFO 10-10 01:35:49 [backends.py:164] Directly load the compiled graph(s) for dynamic shape from the cache, took 1.005 s 2025-10-10T01:35:49.0741179Z (Worker_TP0 pid=15889) INFO 10-10 01:35:49 [backends.py:164] Directly load the compiled graph(s) for dynamic shape from the cache, took 0.960 s 2025-10-10T01:35:49.6116446Z (Worker_TP1 pid=15890) INFO 10-10 01:35:49 [monitor.py:32] torch.compile takes 3.22 s in total 2025-10-10T01:35:49.6117957Z (Worker_TP0 pid=15889) INFO 10-10 01:35:49 [monitor.py:32] torch.compile takes 3.29 s in total 2025-10-10T01:35:50.2969038Z (Worker_TP0 pid=15889) INFO 10-10 01:35:50 [gpu_worker.py:298] Available KV cache memory: 18.54 GiB 2025-10-10T01:35:50.2998773Z (Worker_TP1 pid=15890) INFO 10-10 01:35:50 [gpu_worker.py:298] Available KV cache memory: 18.54 GiB 2025-10-10T01:35:50.5581627Z (EngineCore_DP0 pid=15770) INFO 10-10 01:35:50 [kv_cache_utils.py:1087] GPU KV cache size: 1,214,720 tokens 2025-10-10T01:35:50.5582958Z (EngineCore_DP0 pid=15770) INFO 10-10 01:35:50 [kv_cache_utils.py:1091] Maximum concurrency for 2,048 tokens per request: 593.12x 2025-10-10T01:35:50.5583828Z (EngineCore_DP0 pid=15770) INFO 10-10 01:35:50 [kv_cache_utils.py:1087] GPU KV cache size: 1,214,720 tokens 2025-10-10T01:35:50.5584674Z (EngineCore_DP0 pid=15770) INFO 10-10 01:35:50 [kv_cache_utils.py:1091] Maximum concurrency for 2,048 tokens per request: 593.12x 2025-10-10T01:35:50.8210764Z (Worker_TP0 pid=15889) 2025-10-10T01:35:50.9478529Z Capturing CUDA graphs (mixed prefill-decode, PIECEWISE): 0% 0/5 [00:00, 'use_cudagraph': True, 'cudagraph_num_of_warmups': 1, 'cudagraph_capture_sizes': [16, 8, 4, 2], 'cudagraph_copy_inputs': False, 'full_cuda_graph': True, 'use_inductor_graph_partition': False, 'pass_config': {'enable_noop': True, 'enable_sequence_parallelism': True}, 'max_capture_size': 16, 'local_cache_dir': None} 2025-10-10T01:36:10.9458514Z (EngineCore_DP0 pid=16266) WARNING 10-10 01:36:10 [multiproc_executor.py:720] Reducing Torch parallelism from 24 threads to 1 to avoid unnecessary CPU contention. Set OMP_NUM_THREADS in the external environment to tune this value as needed. 2025-10-10T01:36:10.9462608Z (EngineCore_DP0 pid=16266) INFO 10-10 01:36:10 [shm_broadcast.py:289] vLLM message queue communication handle: Handle(local_reader_ranks=[0, 1, 2, 3], buffer_handle=(4, 16777216, 10, 'psm_058705f4'), local_subscribe_addr='ipc:///tmp/1b1e3fcf-14d0-4909-86ed-58cb9e67899d', remote_subscribe_addr=None, remote_addr_ipv6=False) 2025-10-10T01:36:11.4321578Z /opt/conda/envs/py_3.12/lib/python3.12/site-packages/torch/cuda/__init__.py:64: FutureWarning: The pynvml package is deprecated. Please install nvidia-ml-py instead. If you did not install pynvml directly, please report this to the maintainers of the package that installed pynvml for you. 2025-10-10T01:36:11.4324374Z import pynvml # type: ignore[import] 2025-10-10T01:36:11.4326838Z /opt/conda/envs/py_3.12/lib/python3.12/site-packages/torch/cuda/__init__.py:64: FutureWarning: The pynvml package is deprecated. Please install nvidia-ml-py instead. If you did not install pynvml directly, please report this to the maintainers of the package that installed pynvml for you. 2025-10-10T01:36:11.4329060Z import pynvml # type: ignore[import] 2025-10-10T01:36:11.4457777Z /opt/conda/envs/py_3.12/lib/python3.12/site-packages/torch/cuda/__init__.py:64: FutureWarning: The pynvml package is deprecated. Please install nvidia-ml-py instead. If you did not install pynvml directly, please report this to the maintainers of the package that installed pynvml for you. 2025-10-10T01:36:11.4460019Z import pynvml # type: ignore[import] 2025-10-10T01:36:11.5706820Z /opt/conda/envs/py_3.12/lib/python3.12/site-packages/torch/cuda/__init__.py:64: FutureWarning: The pynvml package is deprecated. Please install nvidia-ml-py instead. If you did not install pynvml directly, please report this to the maintainers of the package that installed pynvml for you. 2025-10-10T01:36:11.5708037Z import pynvml # type: ignore[import] 2025-10-10T01:36:13.7824355Z INFO 10-10 01:36:13 [__init__.py:215] Automatically detected platform cuda. 2025-10-10T01:36:13.8227930Z INFO 10-10 01:36:13 [__init__.py:215] Automatically detected platform cuda. 2025-10-10T01:36:13.8242630Z INFO 10-10 01:36:13 [__init__.py:215] Automatically detected platform cuda. 2025-10-10T01:36:13.8958082Z INFO 10-10 01:36:13 [__init__.py:215] Automatically detected platform cuda. 2025-10-10T01:36:18.3537022Z INFO 10-10 01:36:18 [shm_broadcast.py:289] vLLM message queue communication handle: Handle(local_reader_ranks=[0], buffer_handle=(1, 10485760, 10, 'psm_1ca90f80'), local_subscribe_addr='ipc:///tmp/e240df5a-f080-4a17-9d33-e7b82f10bf8d', remote_subscribe_addr=None, remote_addr_ipv6=False) 2025-10-10T01:36:18.4762630Z INFO 10-10 01:36:18 [shm_broadcast.py:289] vLLM message queue communication handle: Handle(local_reader_ranks=[0], buffer_handle=(1, 10485760, 10, 'psm_db779f68'), local_subscribe_addr='ipc:///tmp/8411783e-e098-40dc-a177-323c5f25de0e', remote_subscribe_addr=None, remote_addr_ipv6=False) 2025-10-10T01:36:18.7540102Z INFO 10-10 01:36:18 [shm_broadcast.py:289] vLLM message queue communication handle: Handle(local_reader_ranks=[0], buffer_handle=(1, 10485760, 10, 'psm_f55dbecd'), local_subscribe_addr='ipc:///tmp/7c676eed-765c-4a27-b994-7fdc7564500d', remote_subscribe_addr=None, remote_addr_ipv6=False) 2025-10-10T01:36:18.7542010Z INFO 10-10 01:36:18 [shm_broadcast.py:289] vLLM message queue communication handle: Handle(local_reader_ranks=[0], buffer_handle=(1, 10485760, 10, 'psm_a22d3dba'), local_subscribe_addr='ipc:///tmp/5fbc26f8-47d1-45d4-8ef9-eb98138e110c', remote_subscribe_addr=None, remote_addr_ipv6=False) 2025-10-10T01:36:19.3174287Z [Gloo] Rank 1 is connected to 3 peer ranks. Expected number of connected peer ranks is : 3 2025-10-10T01:36:19.3174942Z [Gloo] Rank 0 is connected to 3 peer ranks. Expected number of connected peer ranks is : 3 2025-10-10T01:36:19.3175513Z [Gloo] Rank 2 is connected to 3 peer ranks. Expected number of connected peer ranks is : 3 2025-10-10T01:36:19.3176211Z [Gloo] Rank 3 is connected to 3 peer ranks. Expected number of connected peer ranks is : 3 2025-10-10T01:36:19.3221157Z [Gloo] Rank 1 is connected to 1 peer ranks. Expected number of connected peer ranks is : 1 2025-10-10T01:36:19.3221818Z [Gloo] Rank 0 is connected to 1 peer ranks. Expected number of connected peer ranks is : 1 2025-10-10T01:36:19.3222995Z [Gloo] Rank 0 is connected to 1 peer ranks. Expected number of connected peer ranks is : 1 2025-10-10T01:36:19.3223543Z [Gloo] Rank 1 is connected to 1 peer ranks. Expected number of connected peer ranks is : 1 2025-10-10T01:36:19.3265594Z INFO 10-10 01:36:19 [__init__.py:1384] Found nccl from library libnccl.so.2 2025-10-10T01:36:19.3266533Z INFO 10-10 01:36:19 [__init__.py:1384] Found nccl from library libnccl.so.2 2025-10-10T01:36:19.3267344Z INFO 10-10 01:36:19 [__init__.py:1384] Found nccl from library libnccl.so.2 2025-10-10T01:36:19.3268115Z INFO 10-10 01:36:19 [__init__.py:1384] Found nccl from library libnccl.so.2 2025-10-10T01:36:19.3268898Z INFO 10-10 01:36:19 [pynccl.py:104] vLLM is using nccl==2.27.5 2025-10-10T01:36:19.3269632Z INFO 10-10 01:36:19 [pynccl.py:104] vLLM is using nccl==2.27.5 2025-10-10T01:36:19.3270349Z INFO 10-10 01:36:19 [pynccl.py:104] vLLM is using nccl==2.27.5 2025-10-10T01:36:19.3271675Z INFO 10-10 01:36:19 [pynccl.py:104] vLLM is using nccl==2.27.5 2025-10-10T01:36:19.5107539Z WARNING 10-10 01:36:19 [symm_mem.py:58] SymmMemCommunicator: Device capability 8.9 not supported, communicator is not available. 2025-10-10T01:36:19.5108414Z WARNING 10-10 01:36:19 [symm_mem.py:58] SymmMemCommunicator: Device capability 8.9 not supported, communicator is not available. 2025-10-10T01:36:19.5131445Z INFO 10-10 01:36:19 [custom_all_reduce.py:35] Skipping P2P check and trusting the driver's P2P report. 2025-10-10T01:36:19.5132104Z INFO 10-10 01:36:19 [custom_all_reduce.py:35] Skipping P2P check and trusting the driver's P2P report. 2025-10-10T01:36:19.5133156Z WARNING 10-10 01:36:19 [custom_all_reduce.py:154] Custom allreduce is disabled because your platform lacks GPU P2P capability or P2P test failed. To silence this warning, specify disable_custom_all_reduce=True explicitly. 2025-10-10T01:36:19.5134930Z WARNING 10-10 01:36:19 [custom_all_reduce.py:154] Custom allreduce is disabled because your platform lacks GPU P2P capability or P2P test failed. To silence this warning, specify disable_custom_all_reduce=True explicitly. 2025-10-10T01:36:19.5149568Z INFO 10-10 01:36:19 [shm_broadcast.py:289] vLLM message queue communication handle: Handle(local_reader_ranks=[1], buffer_handle=(1, 4194304, 6, 'psm_e6f81d25'), local_subscribe_addr='ipc:///tmp/3f3a36ed-7923-4e7e-bd63-59489098d4bb', remote_subscribe_addr=None, remote_addr_ipv6=False) 2025-10-10T01:36:19.5932271Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-10-10T01:36:19.5933410Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-10-10T01:36:19.6099116Z WARNING 10-10 01:36:19 [symm_mem.py:58] SymmMemCommunicator: Device capability 8.9 not supported, communicator is not available. 2025-10-10T01:36:19.6100737Z WARNING 10-10 01:36:19 [symm_mem.py:58] SymmMemCommunicator: Device capability 8.9 not supported, communicator is not available. 2025-10-10T01:36:19.6123715Z INFO 10-10 01:36:19 [custom_all_reduce.py:35] Skipping P2P check and trusting the driver's P2P report. 2025-10-10T01:36:19.6125242Z INFO 10-10 01:36:19 [custom_all_reduce.py:35] Skipping P2P check and trusting the driver's P2P report. 2025-10-10T01:36:19.6127167Z WARNING 10-10 01:36:19 [custom_all_reduce.py:154] Custom allreduce is disabled because your platform lacks GPU P2P capability or P2P test failed. To silence this warning, specify disable_custom_all_reduce=True explicitly. 2025-10-10T01:36:19.6129803Z WARNING 10-10 01:36:19 [custom_all_reduce.py:154] Custom allreduce is disabled because your platform lacks GPU P2P capability or P2P test failed. To silence this warning, specify disable_custom_all_reduce=True explicitly. 2025-10-10T01:36:19.6142972Z INFO 10-10 01:36:19 [shm_broadcast.py:289] vLLM message queue communication handle: Handle(local_reader_ranks=[1], buffer_handle=(1, 4194304, 6, 'psm_df741e07'), local_subscribe_addr='ipc:///tmp/5fd33ca7-4b2e-4d27-b73f-403c0def07a6', remote_subscribe_addr=None, remote_addr_ipv6=False) 2025-10-10T01:36:19.6162916Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-10-10T01:36:19.6178639Z [Gloo] Rank 1 is connected to 1 peer ranks. Expected number of connected peer ranks is : 1 2025-10-10T01:36:19.6179768Z [Gloo] Rank 0 is connected to 1 peer ranks. Expected number of connected peer ranks is : 1 2025-10-10T01:36:19.6181115Z INFO 10-10 01:36:19 [__init__.py:1384] Found nccl from library libnccl.so.2 2025-10-10T01:36:19.6181921Z INFO 10-10 01:36:19 [pynccl.py:104] vLLM is using nccl==2.27.5 2025-10-10T01:36:19.6182741Z INFO 10-10 01:36:19 [__init__.py:1384] Found nccl from library libnccl.so.2 2025-10-10T01:36:19.6183564Z INFO 10-10 01:36:19 [pynccl.py:104] vLLM is using nccl==2.27.5 2025-10-10T01:36:19.6531292Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-10-10T01:36:19.6532477Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-10-10T01:36:19.6936095Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-10-10T01:36:19.6952559Z [Gloo] Rank 1 is connected to 1 peer ranks. Expected number of connected peer ranks is : 1 2025-10-10T01:36:19.6953197Z [Gloo] Rank 0 is connected to 1 peer ranks. Expected number of connected peer ranks is : 1 2025-10-10T01:36:19.6955772Z INFO 10-10 01:36:19 [__init__.py:1384] Found nccl from library libnccl.so.2 2025-10-10T01:36:19.6956226Z INFO 10-10 01:36:19 [pynccl.py:104] vLLM is using nccl==2.27.5 2025-10-10T01:36:19.6956676Z INFO 10-10 01:36:19 [__init__.py:1384] Found nccl from library libnccl.so.2 2025-10-10T01:36:19.6957078Z INFO 10-10 01:36:19 [pynccl.py:104] vLLM is using nccl==2.27.5 2025-10-10T01:36:19.7297573Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-10-10T01:36:19.7298395Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-10-10T01:36:19.7309732Z [Gloo] Rank 1 is connected to 1 peer ranks. Expected number of connected peer ranks is : 1 2025-10-10T01:36:19.7310327Z [Gloo] Rank 0 is connected to 1 peer ranks. Expected number of connected peer ranks is : 1 2025-10-10T01:36:19.7311237Z [Gloo] Rank 1 is connected to 1 peer ranks. Expected number of connected peer ranks is : 1 2025-10-10T01:36:19.7311790Z [Gloo] Rank 0 is connected to 1 peer ranks. Expected number of connected peer ranks is : 1 2025-10-10T01:36:19.7312366Z INFO 10-10 01:36:19 [__init__.py:1384] Found nccl from library libnccl.so.2 2025-10-10T01:36:19.7312872Z INFO 10-10 01:36:19 [pynccl.py:104] vLLM is using nccl==2.27.5 2025-10-10T01:36:19.7313372Z INFO 10-10 01:36:19 [__init__.py:1384] Found nccl from library libnccl.so.2 2025-10-10T01:36:19.7313747Z INFO 10-10 01:36:19 [pynccl.py:104] vLLM is using nccl==2.27.5 2025-10-10T01:36:19.7314135Z INFO 10-10 01:36:19 [__init__.py:1384] Found nccl from library libnccl.so.2 2025-10-10T01:36:19.7314536Z INFO 10-10 01:36:19 [pynccl.py:104] vLLM is using nccl==2.27.5 2025-10-10T01:36:19.7315702Z INFO 10-10 01:36:19 [__init__.py:1384] Found nccl from library libnccl.so.2 2025-10-10T01:36:19.7316515Z INFO 10-10 01:36:19 [pynccl.py:104] vLLM is using nccl==2.27.5 2025-10-10T01:36:19.7735081Z INFO 10-10 01:36:19 [parallel_state.py:1208] rank 2 in world size 4 is assigned as DP rank 0, PP rank 1, TP rank 0, EP rank 0 2025-10-10T01:36:19.7736081Z INFO 10-10 01:36:19 [parallel_state.py:1208] rank 3 in world size 4 is assigned as DP rank 0, PP rank 1, TP rank 1, EP rank 1 2025-10-10T01:36:19.7742376Z INFO 10-10 01:36:19 [parallel_state.py:1208] rank 1 in world size 4 is assigned as DP rank 0, PP rank 0, TP rank 1, EP rank 1 2025-10-10T01:36:19.7743099Z INFO 10-10 01:36:19 [parallel_state.py:1208] rank 0 in world size 4 is assigned as DP rank 0, PP rank 0, TP rank 0, EP rank 0 2025-10-10T01:36:20.0325143Z INFO 10-10 01:36:20 [topk_topp_sampler.py:55] Using FlashInfer for top-p & top-k sampling. 2025-10-10T01:36:20.0428993Z INFO 10-10 01:36:20 [topk_topp_sampler.py:55] Using FlashInfer for top-p & top-k sampling. 2025-10-10T01:36:20.0494556Z INFO 10-10 01:36:20 [topk_topp_sampler.py:55] Using FlashInfer for top-p & top-k sampling. 2025-10-10T01:36:20.0631081Z (Worker_PP0_TP0 pid=16385) INFO 10-10 01:36:20 [gpu_model_runner.py:2707] Starting to load model meta-llama/Llama-3.2-1B-Instruct... 2025-10-10T01:36:20.0714101Z INFO 10-10 01:36:20 [topk_topp_sampler.py:55] Using FlashInfer for top-p & top-k sampling. 2025-10-10T01:36:20.0730411Z (Worker_PP1_TP0 pid=16387) INFO 10-10 01:36:20 [gpu_model_runner.py:2707] Starting to load model meta-llama/Llama-3.2-1B-Instruct... 2025-10-10T01:36:20.0797898Z (Worker_PP0_TP1 pid=16386) INFO 10-10 01:36:20 [gpu_model_runner.py:2707] Starting to load model meta-llama/Llama-3.2-1B-Instruct... 2025-10-10T01:36:20.1013043Z (Worker_PP1_TP1 pid=16388) INFO 10-10 01:36:20 [gpu_model_runner.py:2707] Starting to load model meta-llama/Llama-3.2-1B-Instruct... 2025-10-10T01:36:20.2888232Z (Worker_PP0_TP0 pid=16385) INFO 10-10 01:36:20 [gpu_model_runner.py:2739] Loading model from scratch... 2025-10-10T01:36:20.3085238Z (Worker_PP1_TP0 pid=16387) INFO 10-10 01:36:20 [gpu_model_runner.py:2739] Loading model from scratch... 2025-10-10T01:36:20.3457509Z (Worker_PP0_TP1 pid=16386) INFO 10-10 01:36:20 [gpu_model_runner.py:2739] Loading model from scratch... 2025-10-10T01:36:20.3660301Z (Worker_PP1_TP1 pid=16388) INFO 10-10 01:36:20 [gpu_model_runner.py:2739] Loading model from scratch... 2025-10-10T01:36:20.3783522Z (Worker_PP0_TP0 pid=16385) INFO 10-10 01:36:20 [cuda.py:361] Using Flash Attention backend on V1 engine. 2025-10-10T01:36:20.3995452Z (Worker_PP1_TP0 pid=16387) INFO 10-10 01:36:20 [cuda.py:361] Using Flash Attention backend on V1 engine. 2025-10-10T01:36:20.4346376Z (Worker_PP0_TP1 pid=16386) INFO 10-10 01:36:20 [cuda.py:361] Using Flash Attention backend on V1 engine. 2025-10-10T01:36:20.4541239Z (Worker_PP1_TP1 pid=16388) INFO 10-10 01:36:20 [cuda.py:361] Using Flash Attention backend on V1 engine. 2025-10-10T01:36:20.8912988Z (Worker_PP0_TP0 pid=16385) INFO 10-10 01:36:20 [gpu_model_runner.py:2758] Model loading took 0.7135 GiB and 0.152407 seconds 2025-10-10T01:36:20.9116015Z (Worker_PP1_TP0 pid=16387) INFO 10-10 01:36:20 [gpu_model_runner.py:2758] Model loading took 0.7135 GiB and 0.153105 seconds 2025-10-10T01:36:21.0301269Z (Worker_PP0_TP1 pid=16386) INFO 10-10 01:36:21 [gpu_model_runner.py:2758] Model loading took 0.7135 GiB and 0.153337 seconds 2025-10-10T01:36:21.0650238Z (Worker_PP1_TP1 pid=16388) INFO 10-10 01:36:21 [gpu_model_runner.py:2758] Model loading took 0.7135 GiB and 0.150787 seconds 2025-10-10T01:36:23.3621335Z (Worker_PP1_TP0 pid=16387) INFO 10-10 01:36:23 [backends.py:548] Using cache directory: /var/lib/jenkins/.cache/vllm/torch_compile_cache/8cedb68d99/rank_2_0/backbone for vLLM's torch.compile 2025-10-10T01:36:23.3623367Z (Worker_PP1_TP0 pid=16387) INFO 10-10 01:36:23 [backends.py:559] Dynamo bytecode transform time: 2.08 s 2025-10-10T01:36:23.4216906Z (Worker_PP0_TP1 pid=16386) INFO 10-10 01:36:23 [backends.py:548] Using cache directory: /var/lib/jenkins/.cache/vllm/torch_compile_cache/5280ab4e2b/rank_1_0/backbone for vLLM's torch.compile 2025-10-10T01:36:23.4221948Z (Worker_PP0_TP1 pid=16386) INFO 10-10 01:36:23 [backends.py:559] Dynamo bytecode transform time: 2.09 s 2025-10-10T01:36:23.4444704Z (Worker_PP0_TP0 pid=16385) INFO 10-10 01:36:23 [backends.py:548] Using cache directory: /var/lib/jenkins/.cache/vllm/torch_compile_cache/5280ab4e2b/rank_0_0/backbone for vLLM's torch.compile 2025-10-10T01:36:23.4449178Z (Worker_PP0_TP0 pid=16385) INFO 10-10 01:36:23 [backends.py:559] Dynamo bytecode transform time: 2.15 s 2025-10-10T01:36:23.4809564Z (Worker_PP1_TP1 pid=16388) INFO 10-10 01:36:23 [backends.py:548] Using cache directory: /var/lib/jenkins/.cache/vllm/torch_compile_cache/8cedb68d99/rank_3_0/backbone for vLLM's torch.compile 2025-10-10T01:36:23.4814416Z (Worker_PP1_TP1 pid=16388) INFO 10-10 01:36:23 [backends.py:559] Dynamo bytecode transform time: 2.15 s 2025-10-10T01:36:24.0958711Z (Worker_PP1_TP0 pid=16387) [rank2]:W1010 01:36:24.094000 16387 site-packages/torch/_inductor/utils.py:1594] [0/0] Not enough SMs to use max_autotune_gemm mode 2025-10-10T01:36:24.2374873Z (Worker_PP1_TP1 pid=16388) [rank3]:W1010 01:36:24.236000 16388 site-packages/torch/_inductor/utils.py:1594] [0/0] Not enough SMs to use max_autotune_gemm mode 2025-10-10T01:36:24.2806523Z (Worker_PP0_TP1 pid=16386) [rank1]:W1010 01:36:24.279000 16386 site-packages/torch/_inductor/utils.py:1594] [0/0] Not enough SMs to use max_autotune_gemm mode 2025-10-10T01:36:24.3128431Z (Worker_PP0_TP0 pid=16385) [rank0]:W1010 01:36:24.311000 16385 site-packages/torch/_inductor/utils.py:1594] [0/0] Not enough SMs to use max_autotune_gemm mode 2025-10-10T01:36:25.4969972Z (Worker_PP1_TP0 pid=16387) INFO 10-10 01:36:25 [backends.py:197] Cache the graph for dynamic shape for later use 2025-10-10T01:36:25.6290779Z (Worker_PP1_TP1 pid=16388) INFO 10-10 01:36:25 [backends.py:197] Cache the graph for dynamic shape for later use 2025-10-10T01:36:25.9061351Z (Worker_PP0_TP1 pid=16386) INFO 10-10 01:36:25 [backends.py:197] Cache the graph for dynamic shape for later use 2025-10-10T01:36:25.9329166Z (Worker_PP0_TP0 pid=16385) INFO 10-10 01:36:25 [backends.py:197] Cache the graph for dynamic shape for later use 2025-10-10T01:36:29.4090149Z (Worker_PP0_TP1 pid=16386) INFO 10-10 01:36:29 [backends.py:218] Compiling a graph for dynamic shape takes 5.78 s 2025-10-10T01:36:29.4292502Z (Worker_PP0_TP0 pid=16385) INFO 10-10 01:36:29 [backends.py:218] Compiling a graph for dynamic shape takes 5.78 s 2025-10-10T01:36:29.6259530Z (Worker_PP1_TP0 pid=16387) INFO 10-10 01:36:29 [backends.py:218] Compiling a graph for dynamic shape takes 6.07 s 2025-10-10T01:36:29.7210151Z (Worker_PP1_TP1 pid=16388) INFO 10-10 01:36:29 [backends.py:218] Compiling a graph for dynamic shape takes 6.04 s 2025-10-10T01:36:33.3019889Z (Worker_PP0_TP0 pid=16385) INFO 10-10 01:36:33 [gpu_worker.py:298] Available KV cache memory: 18.97 GiB 2025-10-10T01:36:33.3090901Z (Worker_PP0_TP1 pid=16386) INFO 10-10 01:36:33 [gpu_worker.py:298] Available KV cache memory: 18.97 GiB 2025-10-10T01:36:33.8112495Z (Worker_PP1_TP1 pid=16388) INFO 10-10 01:36:33 [gpu_worker.py:298] Available KV cache memory: 18.94 GiB 2025-10-10T01:36:33.8121138Z (Worker_PP1_TP0 pid=16387) INFO 10-10 01:36:33 [gpu_worker.py:298] Available KV cache memory: 18.94 GiB 2025-10-10T01:36:34.1336783Z (EngineCore_DP0 pid=16266) INFO 10-10 01:36:34 [kv_cache_utils.py:1087] GPU KV cache size: 2,486,016 tokens 2025-10-10T01:36:34.1337729Z (EngineCore_DP0 pid=16266) INFO 10-10 01:36:34 [kv_cache_utils.py:1091] Maximum concurrency for 2,048 tokens per request: 1213.88x 2025-10-10T01:36:34.1338606Z (EngineCore_DP0 pid=16266) INFO 10-10 01:36:34 [kv_cache_utils.py:1087] GPU KV cache size: 2,486,016 tokens 2025-10-10T01:36:34.1339478Z (EngineCore_DP0 pid=16266) INFO 10-10 01:36:34 [kv_cache_utils.py:1091] Maximum concurrency for 2,048 tokens per request: 1213.88x 2025-10-10T01:36:34.1340298Z (EngineCore_DP0 pid=16266) INFO 10-10 01:36:34 [kv_cache_utils.py:1087] GPU KV cache size: 2,482,944 tokens 2025-10-10T01:36:34.1341109Z (EngineCore_DP0 pid=16266) INFO 10-10 01:36:34 [kv_cache_utils.py:1091] Maximum concurrency for 2,048 tokens per request: 1212.38x 2025-10-10T01:36:34.1341937Z (EngineCore_DP0 pid=16266) INFO 10-10 01:36:34 [kv_cache_utils.py:1087] GPU KV cache size: 2,482,944 tokens 2025-10-10T01:36:34.1342747Z (EngineCore_DP0 pid=16266) INFO 10-10 01:36:34 [kv_cache_utils.py:1091] Maximum concurrency for 2,048 tokens per request: 1212.38x 2025-10-10T01:36:34.4645402Z (Worker_PP0_TP0 pid=16385) 2025-10-10T01:36:35.4805966Z Capturing CUDA graphs (mixed prefill-decode, PIECEWISE): 0% 0/4 [00:00, 'use_cudagraph': True, 'cudagraph_num_of_warmups': 1, 'cudagraph_capture_sizes': [16, 8, 4, 2, 1], 'cudagraph_copy_inputs': False, 'full_cuda_graph': True, 'use_inductor_graph_partition': False, 'pass_config': {}, 'max_capture_size': 16, 'local_cache_dir': None} 2025-10-10T01:37:22.0028496Z (EngineCore_DP0 pid=20625) WARNING 10-10 01:37:22 [multiproc_executor.py:720] Reducing Torch parallelism from 24 threads to 1 to avoid unnecessary CPU contention. Set OMP_NUM_THREADS in the external environment to tune this value as needed. 2025-10-10T01:37:22.0033092Z (EngineCore_DP0 pid=20625) INFO 10-10 01:37:22 [shm_broadcast.py:289] vLLM message queue communication handle: Handle(local_reader_ranks=[0, 1], buffer_handle=(2, 16777216, 10, 'psm_2a511396'), local_subscribe_addr='ipc:///tmp/d88bd1b7-66c3-46cd-b91f-884c43060e57', remote_subscribe_addr=None, remote_addr_ipv6=False) 2025-10-10T01:37:22.4817018Z /opt/conda/envs/py_3.12/lib/python3.12/site-packages/torch/cuda/__init__.py:64: FutureWarning: The pynvml package is deprecated. Please install nvidia-ml-py instead. If you did not install pynvml directly, please report this to the maintainers of the package that installed pynvml for you. 2025-10-10T01:37:22.4819346Z import pynvml # type: ignore[import] 2025-10-10T01:37:22.4976154Z /opt/conda/envs/py_3.12/lib/python3.12/site-packages/torch/cuda/__init__.py:64: FutureWarning: The pynvml package is deprecated. Please install nvidia-ml-py instead. If you did not install pynvml directly, please report this to the maintainers of the package that installed pynvml for you. 2025-10-10T01:37:22.4978344Z import pynvml # type: ignore[import] 2025-10-10T01:37:24.8049095Z INFO 10-10 01:37:24 [__init__.py:215] Automatically detected platform cuda. 2025-10-10T01:37:24.8215657Z INFO 10-10 01:37:24 [__init__.py:215] Automatically detected platform cuda. 2025-10-10T01:37:29.3946196Z INFO 10-10 01:37:29 [shm_broadcast.py:289] vLLM message queue communication handle: Handle(local_reader_ranks=[0], buffer_handle=(1, 10485760, 10, 'psm_73aad806'), local_subscribe_addr='ipc:///tmp/3b54d9fd-8c77-43fc-b60b-9ff2c4988f1e', remote_subscribe_addr=None, remote_addr_ipv6=False) 2025-10-10T01:37:29.4542689Z INFO 10-10 01:37:29 [shm_broadcast.py:289] vLLM message queue communication handle: Handle(local_reader_ranks=[0], buffer_handle=(1, 10485760, 10, 'psm_fc260ee3'), local_subscribe_addr='ipc:///tmp/5b819e48-cb69-4fbc-8400-8e5f04a8cf19', remote_subscribe_addr=None, remote_addr_ipv6=False) 2025-10-10T01:37:29.9426525Z [Gloo] Rank 1 is connected to 1 peer ranks. Expected number of connected peer ranks is : 1 2025-10-10T01:37:29.9427730Z [Gloo] Rank 0 is connected to 1 peer ranks. Expected number of connected peer ranks is : 1 2025-10-10T01:37:29.9469049Z [Gloo] Rank 1 is connected to 1 peer ranks. Expected number of connected peer ranks is : 1 2025-10-10T01:37:29.9470138Z [Gloo] Rank 0 is connected to 1 peer ranks. Expected number of connected peer ranks is : 1 2025-10-10T01:37:29.9512948Z INFO 10-10 01:37:29 [__init__.py:1384] Found nccl from library libnccl.so.2 2025-10-10T01:37:29.9513872Z INFO 10-10 01:37:29 [__init__.py:1384] Found nccl from library libnccl.so.2 2025-10-10T01:37:29.9514726Z INFO 10-10 01:37:29 [pynccl.py:104] vLLM is using nccl==2.27.5 2025-10-10T01:37:29.9515453Z INFO 10-10 01:37:29 [pynccl.py:104] vLLM is using nccl==2.27.5 2025-10-10T01:37:30.1198215Z WARNING 10-10 01:37:30 [symm_mem.py:58] SymmMemCommunicator: Device capability 8.9 not supported, communicator is not available. 2025-10-10T01:37:30.1199153Z WARNING 10-10 01:37:30 [symm_mem.py:58] SymmMemCommunicator: Device capability 8.9 not supported, communicator is not available. 2025-10-10T01:37:30.1231330Z INFO 10-10 01:37:30 [custom_all_reduce.py:35] Skipping P2P check and trusting the driver's P2P report. 2025-10-10T01:37:30.1233361Z WARNING 10-10 01:37:30 [custom_all_reduce.py:154] Custom allreduce is disabled because your platform lacks GPU P2P capability or P2P test failed. To silence this warning, specify disable_custom_all_reduce=True explicitly. 2025-10-10T01:37:30.1235348Z INFO 10-10 01:37:30 [custom_all_reduce.py:35] Skipping P2P check and trusting the driver's P2P report. 2025-10-10T01:37:30.1237735Z WARNING 10-10 01:37:30 [custom_all_reduce.py:154] Custom allreduce is disabled because your platform lacks GPU P2P capability or P2P test failed. To silence this warning, specify disable_custom_all_reduce=True explicitly. 2025-10-10T01:37:30.1839705Z INFO 10-10 01:37:30 [shm_broadcast.py:289] vLLM message queue communication handle: Handle(local_reader_ranks=[1], buffer_handle=(1, 4194304, 6, 'psm_d066c1da'), local_subscribe_addr='ipc:///tmp/5535bbe7-1a2f-4af4-828a-3ac1f2ab6879', remote_subscribe_addr=None, remote_addr_ipv6=False) 2025-10-10T01:37:30.1866190Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-10-10T01:37:30.1871326Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-10-10T01:37:30.1872700Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-10-10T01:37:30.1881913Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-10-10T01:37:30.2635800Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-10-10T01:37:30.2647753Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-10-10T01:37:30.2659321Z [Gloo] Rank 1 is connected to 1 peer ranks. Expected number of connected peer ranks is : 1 2025-10-10T01:37:30.2659921Z [Gloo] Rank 0 is connected to 1 peer ranks. Expected number of connected peer ranks is : 1 2025-10-10T01:37:30.2661080Z INFO 10-10 01:37:30 [__init__.py:1384] Found nccl from library libnccl.so.2 2025-10-10T01:37:30.2663364Z INFO 10-10 01:37:30 [pynccl.py:104] vLLM is using nccl==2.27.5 2025-10-10T01:37:30.2663814Z INFO 10-10 01:37:30 [__init__.py:1384] Found nccl from library libnccl.so.2 2025-10-10T01:37:30.2664234Z INFO 10-10 01:37:30 [pynccl.py:104] vLLM is using nccl==2.27.5 2025-10-10T01:37:30.2998918Z INFO 10-10 01:37:30 [parallel_state.py:1208] rank 1 in world size 2 is assigned as DP rank 0, PP rank 0, TP rank 1, EP rank 1 2025-10-10T01:37:30.2999698Z INFO 10-10 01:37:30 [parallel_state.py:1208] rank 0 in world size 2 is assigned as DP rank 0, PP rank 0, TP rank 0, EP rank 0 2025-10-10T01:37:30.5510927Z INFO 10-10 01:37:30 [topk_topp_sampler.py:55] Using FlashInfer for top-p & top-k sampling. 2025-10-10T01:37:30.5518340Z INFO 10-10 01:37:30 [topk_topp_sampler.py:55] Using FlashInfer for top-p & top-k sampling. 2025-10-10T01:37:30.5829904Z (Worker_TP1 pid=20745) INFO 10-10 01:37:30 [gpu_model_runner.py:2707] Starting to load model meta-llama/Llama-3.2-1B-Instruct... 2025-10-10T01:37:30.5830836Z (Worker_TP0 pid=20744) INFO 10-10 01:37:30 [gpu_model_runner.py:2707] Starting to load model meta-llama/Llama-3.2-1B-Instruct... 2025-10-10T01:37:30.8043251Z (Worker_TP1 pid=20745) INFO 10-10 01:37:30 [gpu_model_runner.py:2739] Loading model from scratch... 2025-10-10T01:37:30.8055383Z (Worker_TP0 pid=20744) INFO 10-10 01:37:30 [gpu_model_runner.py:2739] Loading model from scratch... 2025-10-10T01:37:30.8937788Z (Worker_TP1 pid=20745) INFO 10-10 01:37:30 [cuda.py:361] Using Flash Attention backend on V1 engine. 2025-10-10T01:37:30.8960118Z (Worker_TP0 pid=20744) INFO 10-10 01:37:30 [cuda.py:361] Using Flash Attention backend on V1 engine. 2025-10-10T01:37:31.4225820Z (Worker_TP1 pid=20745) INFO 10-10 01:37:31 [gpu_model_runner.py:2758] Model loading took 1.1667 GiB and 0.166180 seconds 2025-10-10T01:37:31.4435839Z (Worker_TP0 pid=20744) INFO 10-10 01:37:31 [gpu_model_runner.py:2758] Model loading took 1.1667 GiB and 0.168294 seconds 2025-10-10T01:37:34.9078668Z (Worker_TP1 pid=20745) INFO 10-10 01:37:34 [backends.py:548] Using cache directory: /var/lib/jenkins/.cache/vllm/torch_compile_cache/387e456a16/rank_1_0/backbone for vLLM's torch.compile 2025-10-10T01:37:34.9083028Z (Worker_TP1 pid=20745) INFO 10-10 01:37:34 [backends.py:559] Dynamo bytecode transform time: 3.17 s 2025-10-10T01:37:35.0690964Z (Worker_TP0 pid=20744) INFO 10-10 01:37:35 [backends.py:548] Using cache directory: /var/lib/jenkins/.cache/vllm/torch_compile_cache/387e456a16/rank_0_0/backbone for vLLM's torch.compile 2025-10-10T01:37:35.0698604Z (Worker_TP0 pid=20744) INFO 10-10 01:37:35 [backends.py:559] Dynamo bytecode transform time: 3.33 s 2025-10-10T01:37:36.0416524Z (Worker_TP1 pid=20745) INFO 10-10 01:37:36 [backends.py:164] Directly load the compiled graph(s) for dynamic shape from the cache, took 0.976 s 2025-10-10T01:37:36.2143607Z (Worker_TP0 pid=20744) INFO 10-10 01:37:36 [backends.py:164] Directly load the compiled graph(s) for dynamic shape from the cache, took 0.980 s 2025-10-10T01:37:36.7588321Z (Worker_TP1 pid=20745) INFO 10-10 01:37:36 [monitor.py:32] torch.compile takes 3.17 s in total 2025-10-10T01:37:36.7590150Z (Worker_TP0 pid=20744) INFO 10-10 01:37:36 [monitor.py:32] torch.compile takes 3.33 s in total 2025-10-10T01:37:37.4546444Z (Worker_TP0 pid=20744) INFO 10-10 01:37:37 [gpu_worker.py:298] Available KV cache memory: 18.54 GiB 2025-10-10T01:37:37.4563477Z (Worker_TP1 pid=20745) INFO 10-10 01:37:37 [gpu_worker.py:298] Available KV cache memory: 18.54 GiB 2025-10-10T01:37:37.7259602Z (EngineCore_DP0 pid=20625) INFO 10-10 01:37:37 [kv_cache_utils.py:1087] GPU KV cache size: 1,214,720 tokens 2025-10-10T01:37:37.7260543Z (EngineCore_DP0 pid=20625) INFO 10-10 01:37:37 [kv_cache_utils.py:1091] Maximum concurrency for 2,048 tokens per request: 593.12x 2025-10-10T01:37:37.7261389Z (EngineCore_DP0 pid=20625) INFO 10-10 01:37:37 [kv_cache_utils.py:1087] GPU KV cache size: 1,214,720 tokens 2025-10-10T01:37:37.7262226Z (EngineCore_DP0 pid=20625) INFO 10-10 01:37:37 [kv_cache_utils.py:1091] Maximum concurrency for 2,048 tokens per request: 593.12x 2025-10-10T01:37:37.9908166Z (Worker_TP0 pid=20744) 2025-10-10T01:37:38.1237713Z Capturing CUDA graphs (mixed prefill-decode, PIECEWISE): 0% 0/5 [00:00, 'use_cudagraph': True, 'cudagraph_num_of_warmups': 1, 'cudagraph_capture_sizes': [16, 8, 4, 2], 'cudagraph_copy_inputs': False, 'full_cuda_graph': True, 'use_inductor_graph_partition': False, 'pass_config': {'enable_noop': True, 'enable_sequence_parallelism': True}, 'max_capture_size': 16, 'local_cache_dir': None} 2025-10-10T01:37:59.9595749Z (EngineCore_DP0 pid=21121) 2025-10-10 01:37:59,958 INFO worker.py:1918 -- Started a local Ray instance. View the dashboard at http://127.0.0.1:8265  2025-10-10T01:38:01.1516295Z (EngineCore_DP0 pid=21121) INFO 10-10 01:38:01 [ray_utils.py:345] No current placement group found. Creating a new placement group. 2025-10-10T01:38:01.3227226Z (EngineCore_DP0 pid=21121) INFO 10-10 01:38:01 [ray_distributed_executor.py:171] use_ray_spmd_worker: True 2025-10-10T01:38:02.0316031Z (EngineCore_DP0 pid=21121) (pid=22003) /opt/conda/envs/py_3.12/lib/python3.12/site-packages/torch/cuda/__init__.py:64: FutureWarning: The pynvml package is deprecated. Please install nvidia-ml-py instead. If you did not install pynvml directly, please report this to the maintainers of the package that installed pynvml for you. 2025-10-10T01:38:02.0317809Z (EngineCore_DP0 pid=21121) (pid=22003) import pynvml # type: ignore[import] 2025-10-10T01:38:04.3934316Z (EngineCore_DP0 pid=21121) (pid=22003) INFO 10-10 01:38:04 [__init__.py:215] Automatically detected platform cuda. 2025-10-10T01:38:05.7435927Z (EngineCore_DP0 pid=21121) INFO 10-10 01:38:05 [ray_env.py:63] RAY_NON_CARRY_OVER_ENV_VARS from config: set() 2025-10-10T01:38:05.7437317Z (EngineCore_DP0 pid=21121) INFO 10-10 01:38:05 [ray_env.py:65] Copying the following environment variables to workers: ['VLLM_USE_RAY_COMPILED_DAG', 'LD_LIBRARY_PATH', 'VLLM_WORKER_MULTIPROC_METHOD', 'HUGGING_FACE_HUB_TOKEN', 'VLLM_USE_V1', 'MAX_JOBS', 'HF_TOKEN', 'VLLM_USE_RAY_SPMD_WORKER'] 2025-10-10T01:38:05.7438710Z (EngineCore_DP0 pid=21121) INFO 10-10 01:38:05 [ray_env.py:68] If certain env vars should NOT be copied, add them to /var/lib/jenkins/.config/vllm/ray_non_carry_over_env_vars.json file 2025-10-10T01:38:09.8012603Z (EngineCore_DP0 pid=21121) (RayWorkerWrapper pid=22003) [Gloo] Rank 0 is connected to 3 peer ranks. Expected number of connected peer ranks is : 3 2025-10-10T01:38:09.8014822Z (EngineCore_DP0 pid=21121) (pid=22016) INFO 10-10 01:38:04 [__init__.py:215] Automatically detected platform cuda. [repeated 3x across cluster] (Ray deduplicates logs by default. Set RAY_DEDUP_LOGS=0 to disable log deduplication, or see https://docs.ray.io/en/master/ray-observability/user-guides/configure-logging.html#log-deduplication for more options.) 2025-10-10T01:38:09.9064861Z (EngineCore_DP0 pid=21121) (RayWorkerWrapper pid=22003) INFO 10-10 01:38:09 [__init__.py:1384] Found nccl from library libnccl.so.2 2025-10-10T01:38:09.9066959Z (EngineCore_DP0 pid=21121) (RayWorkerWrapper pid=22003) INFO 10-10 01:38:09 [pynccl.py:104] vLLM is using nccl==2.27.5 2025-10-10T01:38:10.2164177Z (EngineCore_DP0 pid=21121) (RayWorkerWrapper pid=22003) WARNING 10-10 01:38:10 [symm_mem.py:58] SymmMemCommunicator: Device capability 8.9 not supported, communicator is not available. 2025-10-10T01:38:10.2165635Z (EngineCore_DP0 pid=21121) (RayWorkerWrapper pid=22003) INFO 10-10 01:38:10 [custom_all_reduce.py:35] Skipping P2P check and trusting the driver's P2P report. 2025-10-10T01:38:10.2167797Z (EngineCore_DP0 pid=21121) (RayWorkerWrapper pid=22003) WARNING 10-10 01:38:10 [custom_all_reduce.py:154] Custom allreduce is disabled because your platform lacks GPU P2P capability or P2P test failed. To silence this warning, specify disable_custom_all_reduce=True explicitly. 2025-10-10T01:38:10.2170246Z (EngineCore_DP0 pid=21121) (RayWorkerWrapper pid=22003) INFO 10-10 01:38:10 [shm_broadcast.py:289] vLLM message queue communication handle: Handle(local_reader_ranks=[1], buffer_handle=(1, 4194304, 6, 'psm_a6e4f08a'), local_subscribe_addr='ipc:///tmp/0b53f572-3671-48cc-9404-fdc7d2d7c404', remote_subscribe_addr=None, remote_addr_ipv6=False) 2025-10-10T01:38:10.3215550Z (EngineCore_DP0 pid=21121) (RayWorkerWrapper pid=22003) INFO 10-10 01:38:10 [parallel_state.py:1208] rank 0 in world size 4 is assigned as DP rank 0, PP rank 0, TP rank 0, EP rank 0 2025-10-10T01:38:10.5299749Z (EngineCore_DP0 pid=21121) (RayWorkerWrapper pid=22003) INFO 10-10 01:38:10 [topk_topp_sampler.py:55] Using FlashInfer for top-p & top-k sampling. 2025-10-10T01:38:10.6370562Z (EngineCore_DP0 pid=21121) (RayWorkerWrapper pid=22003) INFO 10-10 01:38:10 [gpu_model_runner.py:2707] Starting to load model meta-llama/Llama-3.2-1B-Instruct... 2025-10-10T01:38:10.8457325Z (EngineCore_DP0 pid=21121) (RayWorkerWrapper pid=22003) INFO 10-10 01:38:10 [gpu_model_runner.py:2739] Loading model from scratch... 2025-10-10T01:38:10.9527122Z (EngineCore_DP0 pid=21121) (RayWorkerWrapper pid=22003) INFO 10-10 01:38:10 [cuda.py:361] Using Flash Attention backend on V1 engine. 2025-10-10T01:38:11.4690188Z (EngineCore_DP0 pid=21121) (RayWorkerWrapper pid=22011) INFO 10-10 01:38:11 [gpu_model_runner.py:2758] Model loading took 0.7135 GiB and 0.154882 seconds 2025-10-10T01:38:13.8297358Z (EngineCore_DP0 pid=21121) (RayWorkerWrapper pid=22011) INFO 10-10 01:38:13 [backends.py:548] Using cache directory: /var/lib/jenkins/.cache/vllm/torch_compile_cache/8cedb68d99/rank_3_0/backbone for vLLM's torch.compile 2025-10-10T01:38:13.8298709Z (EngineCore_DP0 pid=21121) (RayWorkerWrapper pid=22011) INFO 10-10 01:38:13 [backends.py:559] Dynamo bytecode transform time: 2.00 s 2025-10-10T01:38:14.4468210Z (EngineCore_DP0 pid=21121) (RayWorkerWrapper pid=22011) INFO 10-10 01:38:14 [backends.py:164] Directly load the compiled graph(s) for dynamic shape from the cache, took 0.403 s 2025-10-10T01:38:15.5736143Z (EngineCore_DP0 pid=21121) (RayWorkerWrapper pid=22003) INFO 10-10 01:38:15 [gpu_worker.py:298] Available KV cache memory: 19.02 GiB 2025-10-10T01:38:15.5737713Z (EngineCore_DP0 pid=21121) (RayWorkerWrapper pid=22016) [Gloo] Rank 0 is connected to 1 peer ranks. Expected number of connected peer ranks is : 1 [repeated 23x across cluster] 2025-10-10T01:38:15.5739611Z (EngineCore_DP0 pid=21121) (RayWorkerWrapper pid=22016) INFO 10-10 01:38:10 [__init__.py:1384] Found nccl from library libnccl.so.2 [repeated 11x across cluster] 2025-10-10T01:38:15.5742830Z (EngineCore_DP0 pid=21121) (RayWorkerWrapper pid=22016) INFO 10-10 01:38:10 [pynccl.py:104] vLLM is using nccl==2.27.5 [repeated 11x across cluster] 2025-10-10T01:38:15.5744122Z (EngineCore_DP0 pid=21121) (RayWorkerWrapper pid=22016) WARNING 10-10 01:38:10 [symm_mem.py:58] SymmMemCommunicator: Device capability 8.9 not supported, communicator is not available. [repeated 3x across cluster] 2025-10-10T01:38:15.5745438Z (EngineCore_DP0 pid=21121) (RayWorkerWrapper pid=22016) INFO 10-10 01:38:10 [custom_all_reduce.py:35] Skipping P2P check and trusting the driver's P2P report. [repeated 3x across cluster] 2025-10-10T01:38:15.5747184Z (EngineCore_DP0 pid=21121) (RayWorkerWrapper pid=22016) WARNING 10-10 01:38:10 [custom_all_reduce.py:154] Custom allreduce is disabled because your platform lacks GPU P2P capability or P2P test failed. To silence this warning, specify disable_custom_all_reduce=True explicitly. [repeated 3x across cluster] 2025-10-10T01:38:15.5749164Z (EngineCore_DP0 pid=21121) (RayWorkerWrapper pid=22016) INFO 10-10 01:38:10 [shm_broadcast.py:289] vLLM message queue communication handle: Handle(local_reader_ranks=[1], buffer_handle=(1, 4194304, 6, 'psm_5455dd29'), local_subscribe_addr='ipc:///tmp/e2bfeb85-9d31-48eb-9027-4a1f0dbc9c14', remote_subscribe_addr=None, remote_addr_ipv6=False) 2025-10-10T01:38:15.5750731Z (EngineCore_DP0 pid=21121) (RayWorkerWrapper pid=22016) INFO 10-10 01:38:10 [parallel_state.py:1208] rank 2 in world size 4 is assigned as DP rank 0, PP rank 1, TP rank 0, EP rank 0 [repeated 3x across cluster] 2025-10-10T01:38:15.5751870Z (EngineCore_DP0 pid=21121) (RayWorkerWrapper pid=22016) INFO 10-10 01:38:10 [topk_topp_sampler.py:55] Using FlashInfer for top-p & top-k sampling. [repeated 3x across cluster] 2025-10-10T01:38:15.8200576Z (EngineCore_DP0 pid=21121) INFO 10-10 01:38:15 [kv_cache_utils.py:1087] GPU KV cache size: 2,492,416 tokens 2025-10-10T01:38:15.8204351Z (EngineCore_DP0 pid=21121) INFO 10-10 01:38:15 [kv_cache_utils.py:1091] Maximum concurrency for 2,048 tokens per request: 1217.00x 2025-10-10T01:38:15.8205173Z (EngineCore_DP0 pid=21121) INFO 10-10 01:38:15 [kv_cache_utils.py:1087] GPU KV cache size: 2,492,416 tokens 2025-10-10T01:38:15.8205985Z (EngineCore_DP0 pid=21121) INFO 10-10 01:38:15 [kv_cache_utils.py:1091] Maximum concurrency for 2,048 tokens per request: 1217.00x 2025-10-10T01:38:15.8206661Z (EngineCore_DP0 pid=21121) INFO 10-10 01:38:15 [kv_cache_utils.py:1087] GPU KV cache size: 2,489,088 tokens 2025-10-10T01:38:15.8207386Z (EngineCore_DP0 pid=21121) INFO 10-10 01:38:15 [kv_cache_utils.py:1091] Maximum concurrency for 2,048 tokens per request: 1215.38x 2025-10-10T01:38:15.8208067Z (EngineCore_DP0 pid=21121) INFO 10-10 01:38:15 [kv_cache_utils.py:1087] GPU KV cache size: 2,489,088 tokens 2025-10-10T01:38:15.8208736Z (EngineCore_DP0 pid=21121) INFO 10-10 01:38:15 [kv_cache_utils.py:1091] Maximum concurrency for 2,048 tokens per request: 1215.38x 2025-10-10T01:38:16.1889279Z (EngineCore_DP0 pid=21121) (RayWorkerWrapper pid=22003) 2025-10-10T01:38:16.1891245Z Capturing CUDA graphs (mixed prefill-decode, PIECEWISE): 0%| | 0/4 [00:00, 'use_cudagraph': True, 'cudagraph_num_of_warmups': 1, 'cudagraph_capture_sizes': [16, 8, 4, 2, 1], 'cudagraph_copy_inputs': False, 'full_cuda_graph': True, 'use_inductor_graph_partition': False, 'pass_config': {}, 'max_capture_size': 16, 'local_cache_dir': None} 2025-10-10T01:38:49.3205035Z (EngineCore_DP0 pid=26559) WARNING 10-10 01:38:49 [multiproc_executor.py:720] Reducing Torch parallelism from 24 threads to 1 to avoid unnecessary CPU contention. Set OMP_NUM_THREADS in the external environment to tune this value as needed. 2025-10-10T01:38:49.3212690Z (EngineCore_DP0 pid=26559) INFO 10-10 01:38:49 [shm_broadcast.py:289] vLLM message queue communication handle: Handle(local_reader_ranks=[0, 1], buffer_handle=(2, 16777216, 10, 'psm_5077c005'), local_subscribe_addr='ipc:///tmp/3a117999-e727-4cdb-b029-4eda39065c73', remote_subscribe_addr=None, remote_addr_ipv6=False) 2025-10-10T01:38:49.7984570Z /opt/conda/envs/py_3.12/lib/python3.12/site-packages/torch/cuda/__init__.py:64: FutureWarning: The pynvml package is deprecated. Please install nvidia-ml-py instead. If you did not install pynvml directly, please report this to the maintainers of the package that installed pynvml for you. 2025-10-10T01:38:49.7985570Z import pynvml # type: ignore[import] 2025-10-10T01:38:49.8177998Z /opt/conda/envs/py_3.12/lib/python3.12/site-packages/torch/cuda/__init__.py:64: FutureWarning: The pynvml package is deprecated. Please install nvidia-ml-py instead. If you did not install pynvml directly, please report this to the maintainers of the package that installed pynvml for you. 2025-10-10T01:38:49.8179194Z import pynvml # type: ignore[import] 2025-10-10T01:38:52.1262885Z INFO 10-10 01:38:52 [__init__.py:215] Automatically detected platform cuda. 2025-10-10T01:38:52.1466236Z INFO 10-10 01:38:52 [__init__.py:215] Automatically detected platform cuda. 2025-10-10T01:38:56.7844120Z INFO 10-10 01:38:56 [shm_broadcast.py:289] vLLM message queue communication handle: Handle(local_reader_ranks=[0], buffer_handle=(1, 10485760, 10, 'psm_654b0446'), local_subscribe_addr='ipc:///tmp/1fd3be53-5a45-4830-afd6-0370820fa43a', remote_subscribe_addr=None, remote_addr_ipv6=False) 2025-10-10T01:38:56.7846019Z INFO 10-10 01:38:56 [shm_broadcast.py:289] vLLM message queue communication handle: Handle(local_reader_ranks=[0], buffer_handle=(1, 10485760, 10, 'psm_abbddc42'), local_subscribe_addr='ipc:///tmp/67605d5e-e9c7-45d7-a6d1-ceae6cf2766a', remote_subscribe_addr=None, remote_addr_ipv6=False) 2025-10-10T01:38:56.9934305Z [Gloo] Rank 0 is connected to 1 peer ranks. Expected number of connected peer ranks is : 1 2025-10-10T01:38:56.9934952Z [Gloo] Rank 1 is connected to 1 peer ranks. Expected number of connected peer ranks is : 1 2025-10-10T01:38:56.9983352Z [Gloo] Rank 0 is connected to 1 peer ranks. Expected number of connected peer ranks is : 1 2025-10-10T01:38:56.9983970Z [Gloo] Rank 1 is connected to 1 peer ranks. Expected number of connected peer ranks is : 1 2025-10-10T01:38:57.0028590Z INFO 10-10 01:38:57 [__init__.py:1384] Found nccl from library libnccl.so.2 2025-10-10T01:38:57.0029083Z INFO 10-10 01:38:57 [__init__.py:1384] Found nccl from library libnccl.so.2 2025-10-10T01:38:57.0029523Z INFO 10-10 01:38:57 [pynccl.py:104] vLLM is using nccl==2.27.5 2025-10-10T01:38:57.0030614Z INFO 10-10 01:38:57 [pynccl.py:104] vLLM is using nccl==2.27.5 2025-10-10T01:38:57.1761108Z WARNING 10-10 01:38:57 [symm_mem.py:58] SymmMemCommunicator: Device capability 8.9 not supported, communicator is not available. 2025-10-10T01:38:57.1761983Z WARNING 10-10 01:38:57 [symm_mem.py:58] SymmMemCommunicator: Device capability 8.9 not supported, communicator is not available. 2025-10-10T01:38:57.1787581Z INFO 10-10 01:38:57 [custom_all_reduce.py:35] Skipping P2P check and trusting the driver's P2P report. 2025-10-10T01:38:57.1788284Z INFO 10-10 01:38:57 [custom_all_reduce.py:35] Skipping P2P check and trusting the driver's P2P report. 2025-10-10T01:38:57.1789337Z WARNING 10-10 01:38:57 [custom_all_reduce.py:154] Custom allreduce is disabled because your platform lacks GPU P2P capability or P2P test failed. To silence this warning, specify disable_custom_all_reduce=True explicitly. 2025-10-10T01:38:57.1790736Z WARNING 10-10 01:38:57 [custom_all_reduce.py:154] Custom allreduce is disabled because your platform lacks GPU P2P capability or P2P test failed. To silence this warning, specify disable_custom_all_reduce=True explicitly. 2025-10-10T01:38:57.2434122Z INFO 10-10 01:38:57 [shm_broadcast.py:289] vLLM message queue communication handle: Handle(local_reader_ranks=[1], buffer_handle=(1, 4194304, 6, 'psm_a5513766'), local_subscribe_addr='ipc:///tmp/6b7baeb7-7abc-406c-8349-8890a224c4eb', remote_subscribe_addr=None, remote_addr_ipv6=False) 2025-10-10T01:38:57.2457295Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-10-10T01:38:57.2459280Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-10-10T01:38:57.2467448Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-10-10T01:38:57.2474932Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-10-10T01:38:57.3336484Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-10-10T01:38:57.3347419Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-10-10T01:38:57.3360429Z [Gloo] Rank 1 is connected to 1 peer ranks. Expected number of connected peer ranks is : 1 2025-10-10T01:38:57.3360918Z [Gloo] Rank 0 is connected to 1 peer ranks. Expected number of connected peer ranks is : 1 2025-10-10T01:38:57.3363460Z INFO 10-10 01:38:57 [__init__.py:1384] Found nccl from library libnccl.so.2 2025-10-10T01:38:57.3364399Z INFO 10-10 01:38:57 [pynccl.py:104] vLLM is using nccl==2.27.5 2025-10-10T01:38:57.3364796Z INFO 10-10 01:38:57 [__init__.py:1384] Found nccl from library libnccl.so.2 2025-10-10T01:38:57.3365152Z INFO 10-10 01:38:57 [pynccl.py:104] vLLM is using nccl==2.27.5 2025-10-10T01:38:57.3706426Z INFO 10-10 01:38:57 [parallel_state.py:1208] rank 1 in world size 2 is assigned as DP rank 0, PP rank 0, TP rank 1, EP rank 1 2025-10-10T01:38:57.3707211Z INFO 10-10 01:38:57 [parallel_state.py:1208] rank 0 in world size 2 is assigned as DP rank 0, PP rank 0, TP rank 0, EP rank 0 2025-10-10T01:38:57.6281721Z INFO 10-10 01:38:57 [topk_topp_sampler.py:55] Using FlashInfer for top-p & top-k sampling. 2025-10-10T01:38:57.6361111Z INFO 10-10 01:38:57 [topk_topp_sampler.py:55] Using FlashInfer for top-p & top-k sampling. 2025-10-10T01:38:57.6595240Z (Worker_TP0 pid=26678) INFO 10-10 01:38:57 [gpu_model_runner.py:2707] Starting to load model meta-llama/Llama-3.2-1B-Instruct... 2025-10-10T01:38:57.6672571Z (Worker_TP1 pid=26679) INFO 10-10 01:38:57 [gpu_model_runner.py:2707] Starting to load model meta-llama/Llama-3.2-1B-Instruct... 2025-10-10T01:38:57.8844440Z (Worker_TP0 pid=26678) INFO 10-10 01:38:57 [gpu_model_runner.py:2739] Loading model from scratch... 2025-10-10T01:38:57.8923310Z (Worker_TP1 pid=26679) INFO 10-10 01:38:57 [gpu_model_runner.py:2739] Loading model from scratch... 2025-10-10T01:38:57.9751056Z (Worker_TP0 pid=26678) INFO 10-10 01:38:57 [cuda.py:361] Using Flash Attention backend on V1 engine. 2025-10-10T01:38:57.9817831Z (Worker_TP1 pid=26679) INFO 10-10 01:38:57 [cuda.py:361] Using Flash Attention backend on V1 engine. 2025-10-10T01:38:58.5193070Z (Worker_TP1 pid=26679) INFO 10-10 01:38:58 [gpu_model_runner.py:2758] Model loading took 1.1667 GiB and 0.166132 seconds 2025-10-10T01:38:58.5216476Z (Worker_TP0 pid=26678) INFO 10-10 01:38:58 [gpu_model_runner.py:2758] Model loading took 1.1667 GiB and 0.167070 seconds 2025-10-10T01:39:02.1292559Z (Worker_TP1 pid=26679) INFO 10-10 01:39:02 [backends.py:548] Using cache directory: /var/lib/jenkins/.cache/vllm/torch_compile_cache/387e456a16/rank_1_0/backbone for vLLM's torch.compile 2025-10-10T01:39:02.1297497Z (Worker_TP1 pid=26679) INFO 10-10 01:39:02 [backends.py:559] Dynamo bytecode transform time: 3.32 s 2025-10-10T01:39:02.2063954Z (Worker_TP0 pid=26678) INFO 10-10 01:39:02 [backends.py:548] Using cache directory: /var/lib/jenkins/.cache/vllm/torch_compile_cache/387e456a16/rank_0_0/backbone for vLLM's torch.compile 2025-10-10T01:39:02.2070698Z (Worker_TP0 pid=26678) INFO 10-10 01:39:02 [backends.py:559] Dynamo bytecode transform time: 3.37 s 2025-10-10T01:39:03.2768067Z (Worker_TP1 pid=26679) INFO 10-10 01:39:03 [backends.py:164] Directly load the compiled graph(s) for dynamic shape from the cache, took 0.988 s 2025-10-10T01:39:03.3622002Z (Worker_TP0 pid=26678) INFO 10-10 01:39:03 [backends.py:164] Directly load the compiled graph(s) for dynamic shape from the cache, took 0.994 s 2025-10-10T01:39:03.9061488Z (Worker_TP1 pid=26679) INFO 10-10 01:39:03 [monitor.py:32] torch.compile takes 3.32 s in total 2025-10-10T01:39:03.9063030Z (Worker_TP0 pid=26678) INFO 10-10 01:39:03 [monitor.py:32] torch.compile takes 3.37 s in total 2025-10-10T01:39:04.6194933Z (Worker_TP1 pid=26679) INFO 10-10 01:39:04 [gpu_worker.py:298] Available KV cache memory: 18.54 GiB 2025-10-10T01:39:04.6245466Z (Worker_TP0 pid=26678) INFO 10-10 01:39:04 [gpu_worker.py:298] Available KV cache memory: 18.54 GiB 2025-10-10T01:39:04.9008583Z (EngineCore_DP0 pid=26559) INFO 10-10 01:39:04 [kv_cache_utils.py:1087] GPU KV cache size: 1,214,720 tokens 2025-10-10T01:39:04.9009524Z (EngineCore_DP0 pid=26559) INFO 10-10 01:39:04 [kv_cache_utils.py:1091] Maximum concurrency for 2,048 tokens per request: 593.12x 2025-10-10T01:39:04.9010537Z (EngineCore_DP0 pid=26559) INFO 10-10 01:39:04 [kv_cache_utils.py:1087] GPU KV cache size: 1,214,720 tokens 2025-10-10T01:39:04.9011259Z (EngineCore_DP0 pid=26559) INFO 10-10 01:39:04 [kv_cache_utils.py:1091] Maximum concurrency for 2,048 tokens per request: 593.12x 2025-10-10T01:39:05.1827415Z (Worker_TP0 pid=26678) 2025-10-10T01:39:05.3098946Z Capturing CUDA graphs (mixed prefill-decode, PIECEWISE): 0% 0/5 [00:00, 'use_cudagraph': True, 'cudagraph_num_of_warmups': 1, 'cudagraph_capture_sizes': [16, 8, 4, 2], 'cudagraph_copy_inputs': False, 'full_cuda_graph': True, 'use_inductor_graph_partition': False, 'pass_config': {'enable_noop': True, 'enable_sequence_parallelism': True}, 'max_capture_size': 16, 'local_cache_dir': None} 2025-10-10T01:39:25.8101268Z (EngineCore_DP0 pid=27055) WARNING 10-10 01:39:25 [multiproc_executor.py:720] Reducing Torch parallelism from 24 threads to 1 to avoid unnecessary CPU contention. Set OMP_NUM_THREADS in the external environment to tune this value as needed. 2025-10-10T01:39:25.8105018Z (EngineCore_DP0 pid=27055) INFO 10-10 01:39:25 [shm_broadcast.py:289] vLLM message queue communication handle: Handle(local_reader_ranks=[0, 1, 2, 3], buffer_handle=(4, 16777216, 10, 'psm_8d03ec67'), local_subscribe_addr='ipc:///tmp/e2dc4d7b-c92c-4ee1-a0de-8d3706e6d86a', remote_subscribe_addr=None, remote_addr_ipv6=False) 2025-10-10T01:39:26.3137269Z /opt/conda/envs/py_3.12/lib/python3.12/site-packages/torch/cuda/__init__.py:64: FutureWarning: The pynvml package is deprecated. Please install nvidia-ml-py instead. If you did not install pynvml directly, please report this to the maintainers of the package that installed pynvml for you. 2025-10-10T01:39:26.3138531Z import pynvml # type: ignore[import] 2025-10-10T01:39:26.3324930Z /opt/conda/envs/py_3.12/lib/python3.12/site-packages/torch/cuda/__init__.py:64: FutureWarning: The pynvml package is deprecated. Please install nvidia-ml-py instead. If you did not install pynvml directly, please report this to the maintainers of the package that installed pynvml for you. 2025-10-10T01:39:26.3326982Z import pynvml # type: ignore[import] 2025-10-10T01:39:26.3388953Z /opt/conda/envs/py_3.12/lib/python3.12/site-packages/torch/cuda/__init__.py:64: FutureWarning: The pynvml package is deprecated. Please install nvidia-ml-py instead. If you did not install pynvml directly, please report this to the maintainers of the package that installed pynvml for you. 2025-10-10T01:39:26.3391227Z import pynvml # type: ignore[import] 2025-10-10T01:39:26.4257073Z /opt/conda/envs/py_3.12/lib/python3.12/site-packages/torch/cuda/__init__.py:64: FutureWarning: The pynvml package is deprecated. Please install nvidia-ml-py instead. If you did not install pynvml directly, please report this to the maintainers of the package that installed pynvml for you. 2025-10-10T01:39:26.4258306Z import pynvml # type: ignore[import] 2025-10-10T01:39:28.6907388Z INFO 10-10 01:39:28 [__init__.py:215] Automatically detected platform cuda. 2025-10-10T01:39:28.6940586Z INFO 10-10 01:39:28 [__init__.py:215] Automatically detected platform cuda. 2025-10-10T01:39:28.7009770Z INFO 10-10 01:39:28 [__init__.py:215] Automatically detected platform cuda. 2025-10-10T01:39:28.7563239Z INFO 10-10 01:39:28 [__init__.py:215] Automatically detected platform cuda. 2025-10-10T01:39:33.3139553Z INFO 10-10 01:39:33 [shm_broadcast.py:289] vLLM message queue communication handle: Handle(local_reader_ranks=[0], buffer_handle=(1, 10485760, 10, 'psm_87ce5bea'), local_subscribe_addr='ipc:///tmp/db36f46e-a5d8-45e4-9bbb-ad7a505071a7', remote_subscribe_addr=None, remote_addr_ipv6=False) 2025-10-10T01:39:33.3319472Z INFO 10-10 01:39:33 [shm_broadcast.py:289] vLLM message queue communication handle: Handle(local_reader_ranks=[0], buffer_handle=(1, 10485760, 10, 'psm_ad30a623'), local_subscribe_addr='ipc:///tmp/9dc8f2a0-4b8e-4b52-b566-499862733535', remote_subscribe_addr=None, remote_addr_ipv6=False) 2025-10-10T01:39:33.5646285Z INFO 10-10 01:39:33 [shm_broadcast.py:289] vLLM message queue communication handle: Handle(local_reader_ranks=[0], buffer_handle=(1, 10485760, 10, 'psm_22f78612'), local_subscribe_addr='ipc:///tmp/931b521a-3c70-400a-a2e8-672774e68abf', remote_subscribe_addr=None, remote_addr_ipv6=False) 2025-10-10T01:39:33.5837483Z INFO 10-10 01:39:33 [shm_broadcast.py:289] vLLM message queue communication handle: Handle(local_reader_ranks=[0], buffer_handle=(1, 10485760, 10, 'psm_05fe577e'), local_subscribe_addr='ipc:///tmp/09f012cd-360a-4bd7-b478-5b1450f0661e', remote_subscribe_addr=None, remote_addr_ipv6=False) 2025-10-10T01:39:33.7774413Z [Gloo] Rank 0 is connected to 3 peer ranks. Expected number of connected peer ranks is : 3 2025-10-10T01:39:33.7775716Z [Gloo] Rank 1 is connected to 3 peer ranks. Expected number of connected peer ranks is : 3 2025-10-10T01:39:33.7776779Z [Gloo] Rank 2 is connected to 3 peer ranks. Expected number of connected peer ranks is : 3 2025-10-10T01:39:33.7777809Z [Gloo] Rank 3 is connected to 3 peer ranks. Expected number of connected peer ranks is : 3 2025-10-10T01:39:33.7820424Z [Gloo] Rank 0 is connected to 1 peer ranks. Expected number of connected peer ranks is : 1 2025-10-10T01:39:33.7821073Z [Gloo] Rank 1 is connected to 1 peer ranks. Expected number of connected peer ranks is : 1 2025-10-10T01:39:33.7822328Z [Gloo] Rank 1 is connected to 1 peer ranks. Expected number of connected peer ranks is : 1 2025-10-10T01:39:33.7822930Z [Gloo] Rank 0 is connected to 1 peer ranks. Expected number of connected peer ranks is : 1 2025-10-10T01:39:33.7863568Z INFO 10-10 01:39:33 [__init__.py:1384] Found nccl from library libnccl.so.2 2025-10-10T01:39:33.7864350Z INFO 10-10 01:39:33 [__init__.py:1384] Found nccl from library libnccl.so.2 2025-10-10T01:39:33.7865014Z INFO 10-10 01:39:33 [__init__.py:1384] Found nccl from library libnccl.so.2 2025-10-10T01:39:33.7865958Z INFO 10-10 01:39:33 [pynccl.py:104] vLLM is using nccl==2.27.5 2025-10-10T01:39:33.7866776Z INFO 10-10 01:39:33 [pynccl.py:104] vLLM is using nccl==2.27.5 2025-10-10T01:39:33.7867519Z INFO 10-10 01:39:33 [pynccl.py:104] vLLM is using nccl==2.27.5 2025-10-10T01:39:33.7868201Z INFO 10-10 01:39:33 [__init__.py:1384] Found nccl from library libnccl.so.2 2025-10-10T01:39:33.7870599Z INFO 10-10 01:39:33 [pynccl.py:104] vLLM is using nccl==2.27.5 2025-10-10T01:39:33.9716063Z WARNING 10-10 01:39:33 [symm_mem.py:58] SymmMemCommunicator: Device capability 8.9 not supported, communicator is not available. 2025-10-10T01:39:33.9717728Z WARNING 10-10 01:39:33 [symm_mem.py:58] SymmMemCommunicator: Device capability 8.9 not supported, communicator is not available. 2025-10-10T01:39:33.9745199Z INFO 10-10 01:39:33 [custom_all_reduce.py:35] Skipping P2P check and trusting the driver's P2P report. 2025-10-10T01:39:33.9745961Z INFO 10-10 01:39:33 [custom_all_reduce.py:35] Skipping P2P check and trusting the driver's P2P report. 2025-10-10T01:39:33.9747253Z WARNING 10-10 01:39:33 [custom_all_reduce.py:154] Custom allreduce is disabled because your platform lacks GPU P2P capability or P2P test failed. To silence this warning, specify disable_custom_all_reduce=True explicitly. 2025-10-10T01:39:33.9748930Z WARNING 10-10 01:39:33 [custom_all_reduce.py:154] Custom allreduce is disabled because your platform lacks GPU P2P capability or P2P test failed. To silence this warning, specify disable_custom_all_reduce=True explicitly. 2025-10-10T01:39:33.9766242Z INFO 10-10 01:39:33 [shm_broadcast.py:289] vLLM message queue communication handle: Handle(local_reader_ranks=[1], buffer_handle=(1, 4194304, 6, 'psm_61457f46'), local_subscribe_addr='ipc:///tmp/c68ac652-4c82-4e56-95ee-2a840963e7aa', remote_subscribe_addr=None, remote_addr_ipv6=False) 2025-10-10T01:39:34.0499452Z WARNING 10-10 01:39:34 [symm_mem.py:58] SymmMemCommunicator: Device capability 8.9 not supported, communicator is not available. 2025-10-10T01:39:34.0501823Z WARNING 10-10 01:39:34 [symm_mem.py:58] SymmMemCommunicator: Device capability 8.9 not supported, communicator is not available. 2025-10-10T01:39:34.0520357Z INFO 10-10 01:39:34 [custom_all_reduce.py:35] Skipping P2P check and trusting the driver's P2P report. 2025-10-10T01:39:34.0520976Z INFO 10-10 01:39:34 [custom_all_reduce.py:35] Skipping P2P check and trusting the driver's P2P report. 2025-10-10T01:39:34.0521802Z WARNING 10-10 01:39:34 [custom_all_reduce.py:154] Custom allreduce is disabled because your platform lacks GPU P2P capability or P2P test failed. To silence this warning, specify disable_custom_all_reduce=True explicitly. 2025-10-10T01:39:34.0523117Z WARNING 10-10 01:39:34 [custom_all_reduce.py:154] Custom allreduce is disabled because your platform lacks GPU P2P capability or P2P test failed. To silence this warning, specify disable_custom_all_reduce=True explicitly. 2025-10-10T01:39:34.0538879Z INFO 10-10 01:39:34 [shm_broadcast.py:289] vLLM message queue communication handle: Handle(local_reader_ranks=[1], buffer_handle=(1, 4194304, 6, 'psm_80d79a7c'), local_subscribe_addr='ipc:///tmp/15cd5ae6-9913-4747-872d-9f4ff38a88b3', remote_subscribe_addr=None, remote_addr_ipv6=False) 2025-10-10T01:39:34.0568909Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-10-10T01:39:34.0736301Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-10-10T01:39:34.0736923Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-10-10T01:39:34.0753090Z [Gloo] Rank 1 is connected to 1 peer ranks. Expected number of connected peer ranks is : 1 2025-10-10T01:39:34.0754093Z [Gloo] Rank 0 is connected to 1 peer ranks. Expected number of connected peer ranks is : 1 2025-10-10T01:39:34.0756423Z INFO 10-10 01:39:34 [__init__.py:1384] Found nccl from library libnccl.so.2 2025-10-10T01:39:34.0757170Z INFO 10-10 01:39:34 [pynccl.py:104] vLLM is using nccl==2.27.5 2025-10-10T01:39:34.0758881Z INFO 10-10 01:39:34 [__init__.py:1384] Found nccl from library libnccl.so.2 2025-10-10T01:39:34.0759662Z INFO 10-10 01:39:34 [pynccl.py:104] vLLM is using nccl==2.27.5 2025-10-10T01:39:34.1099816Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-10-10T01:39:34.1100750Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-10-10T01:39:34.1435711Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-10-10T01:39:34.1452075Z [Gloo] Rank 1 is connected to 1 peer ranks. Expected number of connected peer ranks is : 1 2025-10-10T01:39:34.1452655Z [Gloo] Rank 0 is connected to 1 peer ranks. Expected number of connected peer ranks is : 1 2025-10-10T01:39:34.1455038Z INFO 10-10 01:39:34 [__init__.py:1384] Found nccl from library libnccl.so.2 2025-10-10T01:39:34.1455800Z INFO 10-10 01:39:34 [pynccl.py:104] vLLM is using nccl==2.27.5 2025-10-10T01:39:34.1456329Z INFO 10-10 01:39:34 [__init__.py:1384] Found nccl from library libnccl.so.2 2025-10-10T01:39:34.1456764Z INFO 10-10 01:39:34 [pynccl.py:104] vLLM is using nccl==2.27.5 2025-10-10T01:39:34.1803838Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-10-10T01:39:34.1804455Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-10-10T01:39:34.1816788Z [Gloo] Rank 1 is connected to 1 peer ranks. Expected number of connected peer ranks is : 1 2025-10-10T01:39:34.1818067Z [Gloo] Rank 0 is connected to 1 peer ranks. Expected number of connected peer ranks is : 1 2025-10-10T01:39:34.1819410Z [Gloo] Rank 0 is connected to 1 peer ranks. Expected number of connected peer ranks is : 1 2025-10-10T01:39:34.1820679Z [Gloo] Rank 1 is connected to 1 peer ranks. Expected number of connected peer ranks is : 1 2025-10-10T01:39:34.1821746Z INFO 10-10 01:39:34 [__init__.py:1384] Found nccl from library libnccl.so.2 2025-10-10T01:39:34.1822601Z INFO 10-10 01:39:34 [pynccl.py:104] vLLM is using nccl==2.27.5 2025-10-10T01:39:34.1823027Z INFO 10-10 01:39:34 [__init__.py:1384] Found nccl from library libnccl.so.2 2025-10-10T01:39:34.1823440Z INFO 10-10 01:39:34 [pynccl.py:104] vLLM is using nccl==2.27.5 2025-10-10T01:39:34.1823870Z INFO 10-10 01:39:34 [__init__.py:1384] Found nccl from library libnccl.so.2 2025-10-10T01:39:34.1824545Z INFO 10-10 01:39:34 [__init__.py:1384] Found nccl from library libnccl.so.2 2025-10-10T01:39:34.1824945Z INFO 10-10 01:39:34 [pynccl.py:104] vLLM is using nccl==2.27.5 2025-10-10T01:39:34.1825305Z INFO 10-10 01:39:34 [pynccl.py:104] vLLM is using nccl==2.27.5 2025-10-10T01:39:34.2261205Z INFO 10-10 01:39:34 [parallel_state.py:1208] rank 2 in world size 4 is assigned as DP rank 0, PP rank 1, TP rank 0, EP rank 0 2025-10-10T01:39:34.2262209Z INFO 10-10 01:39:34 [parallel_state.py:1208] rank 3 in world size 4 is assigned as DP rank 0, PP rank 1, TP rank 1, EP rank 1 2025-10-10T01:39:34.2268060Z INFO 10-10 01:39:34 [parallel_state.py:1208] rank 0 in world size 4 is assigned as DP rank 0, PP rank 0, TP rank 0, EP rank 0 2025-10-10T01:39:34.2268818Z INFO 10-10 01:39:34 [parallel_state.py:1208] rank 1 in world size 4 is assigned as DP rank 0, PP rank 0, TP rank 1, EP rank 1 2025-10-10T01:39:34.4920617Z INFO 10-10 01:39:34 [topk_topp_sampler.py:55] Using FlashInfer for top-p & top-k sampling. 2025-10-10T01:39:34.5016043Z INFO 10-10 01:39:34 [topk_topp_sampler.py:55] Using FlashInfer for top-p & top-k sampling. 2025-10-10T01:39:34.5034793Z INFO 10-10 01:39:34 [topk_topp_sampler.py:55] Using FlashInfer for top-p & top-k sampling. 2025-10-10T01:39:34.5062572Z INFO 10-10 01:39:34 [topk_topp_sampler.py:55] Using FlashInfer for top-p & top-k sampling. 2025-10-10T01:39:34.5220176Z (Worker_PP1_TP1 pid=27177) INFO 10-10 01:39:34 [gpu_model_runner.py:2707] Starting to load model meta-llama/Llama-3.2-1B-Instruct... 2025-10-10T01:39:34.5319854Z (Worker_PP0_TP1 pid=27175) INFO 10-10 01:39:34 [gpu_model_runner.py:2707] Starting to load model meta-llama/Llama-3.2-1B-Instruct... 2025-10-10T01:39:34.5341998Z (Worker_PP0_TP0 pid=27174) INFO 10-10 01:39:34 [gpu_model_runner.py:2707] Starting to load model meta-llama/Llama-3.2-1B-Instruct... 2025-10-10T01:39:34.5357566Z (Worker_PP1_TP0 pid=27176) INFO 10-10 01:39:34 [gpu_model_runner.py:2707] Starting to load model meta-llama/Llama-3.2-1B-Instruct... 2025-10-10T01:39:34.7496235Z (Worker_PP1_TP1 pid=27177) INFO 10-10 01:39:34 [gpu_model_runner.py:2739] Loading model from scratch... 2025-10-10T01:39:34.7686074Z (Worker_PP0_TP1 pid=27175) INFO 10-10 01:39:34 [gpu_model_runner.py:2739] Loading model from scratch... 2025-10-10T01:39:34.8007984Z (Worker_PP0_TP0 pid=27174) INFO 10-10 01:39:34 [gpu_model_runner.py:2739] Loading model from scratch... 2025-10-10T01:39:34.8028870Z (Worker_PP1_TP0 pid=27176) INFO 10-10 01:39:34 [gpu_model_runner.py:2739] Loading model from scratch... 2025-10-10T01:39:34.8398086Z (Worker_PP1_TP1 pid=27177) INFO 10-10 01:39:34 [cuda.py:361] Using Flash Attention backend on V1 engine. 2025-10-10T01:39:34.8585226Z (Worker_PP0_TP1 pid=27175) INFO 10-10 01:39:34 [cuda.py:361] Using Flash Attention backend on V1 engine. 2025-10-10T01:39:34.8910140Z (Worker_PP0_TP0 pid=27174) INFO 10-10 01:39:34 [cuda.py:361] Using Flash Attention backend on V1 engine. 2025-10-10T01:39:34.8922344Z (Worker_PP1_TP0 pid=27176) INFO 10-10 01:39:34 [cuda.py:361] Using Flash Attention backend on V1 engine. 2025-10-10T01:39:35.3524817Z (Worker_PP1_TP1 pid=27177) INFO 10-10 01:39:35 [gpu_model_runner.py:2758] Model loading took 0.7135 GiB and 0.154711 seconds 2025-10-10T01:39:35.3745864Z (Worker_PP0_TP1 pid=27175) INFO 10-10 01:39:35 [gpu_model_runner.py:2758] Model loading took 0.7135 GiB and 0.152659 seconds 2025-10-10T01:39:35.4872336Z (Worker_PP1_TP0 pid=27176) INFO 10-10 01:39:35 [gpu_model_runner.py:2758] Model loading took 0.7135 GiB and 0.152562 seconds 2025-10-10T01:39:35.4914486Z (Worker_PP0_TP0 pid=27174) INFO 10-10 01:39:35 [gpu_model_runner.py:2758] Model loading took 0.7135 GiB and 0.154636 seconds 2025-10-10T01:39:37.7793158Z (Worker_PP1_TP1 pid=27177) INFO 10-10 01:39:37 [backends.py:548] Using cache directory: /var/lib/jenkins/.cache/vllm/torch_compile_cache/8cedb68d99/rank_3_0/backbone for vLLM's torch.compile 2025-10-10T01:39:37.7799354Z (Worker_PP1_TP1 pid=27177) INFO 10-10 01:39:37 [backends.py:559] Dynamo bytecode transform time: 2.06 s 2025-10-10T01:39:37.8355289Z (Worker_PP0_TP1 pid=27175) INFO 10-10 01:39:37 [backends.py:548] Using cache directory: /var/lib/jenkins/.cache/vllm/torch_compile_cache/5280ab4e2b/rank_1_0/backbone for vLLM's torch.compile 2025-10-10T01:39:37.8362992Z (Worker_PP0_TP1 pid=27175) INFO 10-10 01:39:37 [backends.py:559] Dynamo bytecode transform time: 2.12 s 2025-10-10T01:39:37.8776242Z (Worker_PP1_TP0 pid=27176) INFO 10-10 01:39:37 [backends.py:548] Using cache directory: /var/lib/jenkins/.cache/vllm/torch_compile_cache/8cedb68d99/rank_2_0/backbone for vLLM's torch.compile 2025-10-10T01:39:37.8783684Z (Worker_PP1_TP0 pid=27176) INFO 10-10 01:39:37 [backends.py:559] Dynamo bytecode transform time: 2.11 s 2025-10-10T01:39:37.9451529Z (Worker_PP0_TP0 pid=27174) INFO 10-10 01:39:37 [backends.py:548] Using cache directory: /var/lib/jenkins/.cache/vllm/torch_compile_cache/5280ab4e2b/rank_0_0/backbone for vLLM's torch.compile 2025-10-10T01:39:37.9459426Z (Worker_PP0_TP0 pid=27174) INFO 10-10 01:39:37 [backends.py:559] Dynamo bytecode transform time: 2.18 s 2025-10-10T01:39:38.3685549Z (Worker_PP1_TP1 pid=27177) INFO 10-10 01:39:38 [backends.py:164] Directly load the compiled graph(s) for dynamic shape from the cache, took 0.397 s 2025-10-10T01:39:38.4291050Z (Worker_PP0_TP1 pid=27175) INFO 10-10 01:39:38 [backends.py:164] Directly load the compiled graph(s) for dynamic shape from the cache, took 0.395 s 2025-10-10T01:39:38.4878567Z (Worker_PP1_TP0 pid=27176) INFO 10-10 01:39:38 [backends.py:164] Directly load the compiled graph(s) for dynamic shape from the cache, took 0.415 s 2025-10-10T01:39:38.5682033Z (Worker_PP0_TP0 pid=27174) INFO 10-10 01:39:38 [backends.py:164] Directly load the compiled graph(s) for dynamic shape from the cache, took 0.414 s 2025-10-10T01:39:39.4587966Z (Worker_PP1_TP0 pid=27176) INFO 10-10 01:39:39 [gpu_worker.py:298] Available KV cache memory: 18.99 GiB 2025-10-10T01:39:39.4900813Z (Worker_PP0_TP1 pid=27175) INFO 10-10 01:39:39 [gpu_worker.py:298] Available KV cache memory: 19.02 GiB 2025-10-10T01:39:39.5349952Z (Worker_PP1_TP1 pid=27177) INFO 10-10 01:39:39 [gpu_worker.py:298] Available KV cache memory: 18.99 GiB 2025-10-10T01:39:39.5641150Z (Worker_PP0_TP0 pid=27174) INFO 10-10 01:39:39 [gpu_worker.py:298] Available KV cache memory: 19.02 GiB 2025-10-10T01:39:39.8530518Z (EngineCore_DP0 pid=27055) INFO 10-10 01:39:39 [kv_cache_utils.py:1087] GPU KV cache size: 2,492,416 tokens 2025-10-10T01:39:39.8531949Z (EngineCore_DP0 pid=27055) INFO 10-10 01:39:39 [kv_cache_utils.py:1091] Maximum concurrency for 2,048 tokens per request: 1217.00x 2025-10-10T01:39:39.8533227Z (EngineCore_DP0 pid=27055) INFO 10-10 01:39:39 [kv_cache_utils.py:1087] GPU KV cache size: 2,492,416 tokens 2025-10-10T01:39:39.8534481Z (EngineCore_DP0 pid=27055) INFO 10-10 01:39:39 [kv_cache_utils.py:1091] Maximum concurrency for 2,048 tokens per request: 1217.00x 2025-10-10T01:39:39.8535655Z (EngineCore_DP0 pid=27055) INFO 10-10 01:39:39 [kv_cache_utils.py:1087] GPU KV cache size: 2,489,088 tokens 2025-10-10T01:39:39.8536466Z (EngineCore_DP0 pid=27055) INFO 10-10 01:39:39 [kv_cache_utils.py:1091] Maximum concurrency for 2,048 tokens per request: 1215.38x 2025-10-10T01:39:39.8537576Z (EngineCore_DP0 pid=27055) INFO 10-10 01:39:39 [kv_cache_utils.py:1087] GPU KV cache size: 2,489,088 tokens 2025-10-10T01:39:39.8538270Z (EngineCore_DP0 pid=27055) INFO 10-10 01:39:39 [kv_cache_utils.py:1091] Maximum concurrency for 2,048 tokens per request: 1215.38x 2025-10-10T01:39:40.1489299Z (Worker_PP0_TP0 pid=27174) 2025-10-10T01:39:41.6333392Z Capturing CUDA graphs (mixed prefill-decode, PIECEWISE): 0% 0/4 [00:00, 'use_cudagraph': True, 'cudagraph_num_of_warmups': 1, 'cudagraph_capture_sizes': [16, 8, 4, 2, 1], 'cudagraph_copy_inputs': False, 'full_cuda_graph': True, 'use_inductor_graph_partition': False, 'pass_config': {}, 'max_capture_size': 16, 'local_cache_dir': None} 2025-10-10T01:40:07.2170761Z (EngineCore_DP0 pid=27802) WARNING 10-10 01:40:07 [multiproc_executor.py:720] Reducing Torch parallelism from 24 threads to 1 to avoid unnecessary CPU contention. Set OMP_NUM_THREADS in the external environment to tune this value as needed. 2025-10-10T01:40:07.2173874Z (EngineCore_DP0 pid=27802) INFO 10-10 01:40:07 [shm_broadcast.py:289] vLLM message queue communication handle: Handle(local_reader_ranks=[0, 1], buffer_handle=(2, 16777216, 10, 'psm_ce850334'), local_subscribe_addr='ipc:///tmp/220fc42d-7cf7-4938-947e-6ad0ab8db316', remote_subscribe_addr=None, remote_addr_ipv6=False) 2025-10-10T01:40:07.7008648Z /opt/conda/envs/py_3.12/lib/python3.12/site-packages/torch/cuda/__init__.py:64: FutureWarning: The pynvml package is deprecated. Please install nvidia-ml-py instead. If you did not install pynvml directly, please report this to the maintainers of the package that installed pynvml for you. 2025-10-10T01:40:07.7010995Z import pynvml # type: ignore[import] 2025-10-10T01:40:07.7012491Z /opt/conda/envs/py_3.12/lib/python3.12/site-packages/torch/cuda/__init__.py:64: FutureWarning: The pynvml package is deprecated. Please install nvidia-ml-py instead. If you did not install pynvml directly, please report this to the maintainers of the package that installed pynvml for you. 2025-10-10T01:40:07.7013395Z import pynvml # type: ignore[import] 2025-10-10T01:40:10.0339002Z INFO 10-10 01:40:10 [__init__.py:215] Automatically detected platform cuda. 2025-10-10T01:40:10.0560338Z INFO 10-10 01:40:10 [__init__.py:215] Automatically detected platform cuda. 2025-10-10T01:40:14.5845959Z INFO 10-10 01:40:14 [shm_broadcast.py:289] vLLM message queue communication handle: Handle(local_reader_ranks=[0], buffer_handle=(1, 10485760, 10, 'psm_edc2460c'), local_subscribe_addr='ipc:///tmp/69d769ad-b5ae-4170-95a8-e52c3d836f31', remote_subscribe_addr=None, remote_addr_ipv6=False) 2025-10-10T01:40:14.6542067Z INFO 10-10 01:40:14 [shm_broadcast.py:289] vLLM message queue communication handle: Handle(local_reader_ranks=[0], buffer_handle=(1, 10485760, 10, 'psm_7e96499b'), local_subscribe_addr='ipc:///tmp/c8412cda-6209-4846-a5d8-7955c7ebb665', remote_subscribe_addr=None, remote_addr_ipv6=False) 2025-10-10T01:40:14.8004370Z [Gloo] Rank 1 is connected to 1 peer ranks. Expected number of connected peer ranks is : 1 2025-10-10T01:40:14.8005062Z [Gloo] Rank 0 is connected to 1 peer ranks. Expected number of connected peer ranks is : 1 2025-10-10T01:40:14.8052626Z [Gloo] Rank 1 is connected to 1 peer ranks. Expected number of connected peer ranks is : 1 2025-10-10T01:40:14.8053246Z [Gloo] Rank 0 is connected to 1 peer ranks. Expected number of connected peer ranks is : 1 2025-10-10T01:40:14.8094454Z INFO 10-10 01:40:14 [__init__.py:1384] Found nccl from library libnccl.so.2 2025-10-10T01:40:14.8095290Z INFO 10-10 01:40:14 [__init__.py:1384] Found nccl from library libnccl.so.2 2025-10-10T01:40:14.8096612Z INFO 10-10 01:40:14 [pynccl.py:104] vLLM is using nccl==2.27.5 2025-10-10T01:40:14.8097842Z INFO 10-10 01:40:14 [pynccl.py:104] vLLM is using nccl==2.27.5 2025-10-10T01:40:14.9825829Z WARNING 10-10 01:40:14 [symm_mem.py:58] SymmMemCommunicator: Device capability 8.9 not supported, communicator is not available. 2025-10-10T01:40:14.9826720Z WARNING 10-10 01:40:14 [symm_mem.py:58] SymmMemCommunicator: Device capability 8.9 not supported, communicator is not available. 2025-10-10T01:40:14.9851507Z INFO 10-10 01:40:14 [custom_all_reduce.py:35] Skipping P2P check and trusting the driver's P2P report. 2025-10-10T01:40:14.9852159Z INFO 10-10 01:40:14 [custom_all_reduce.py:35] Skipping P2P check and trusting the driver's P2P report. 2025-10-10T01:40:14.9853583Z WARNING 10-10 01:40:14 [custom_all_reduce.py:154] Custom allreduce is disabled because your platform lacks GPU P2P capability or P2P test failed. To silence this warning, specify disable_custom_all_reduce=True explicitly. 2025-10-10T01:40:14.9855070Z WARNING 10-10 01:40:14 [custom_all_reduce.py:154] Custom allreduce is disabled because your platform lacks GPU P2P capability or P2P test failed. To silence this warning, specify disable_custom_all_reduce=True explicitly. 2025-10-10T01:40:15.0940275Z INFO 10-10 01:40:15 [shm_broadcast.py:289] vLLM message queue communication handle: Handle(local_reader_ranks=[1], buffer_handle=(1, 4194304, 6, 'psm_97df56fa'), local_subscribe_addr='ipc:///tmp/efdede24-5fda-4b11-a263-740f34f42393', remote_subscribe_addr=None, remote_addr_ipv6=False) 2025-10-10T01:40:15.0970823Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-10-10T01:40:15.0971417Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-10-10T01:40:15.0978901Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-10-10T01:40:15.0987229Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-10-10T01:40:15.1736419Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-10-10T01:40:15.1747081Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-10-10T01:40:15.1758214Z [Gloo] Rank 0 is connected to 1 peer ranks. Expected number of connected peer ranks is : 1 2025-10-10T01:40:15.1758942Z [Gloo] Rank 1 is connected to 1 peer ranks. Expected number of connected peer ranks is : 1 2025-10-10T01:40:15.1761285Z INFO 10-10 01:40:15 [__init__.py:1384] Found nccl from library libnccl.so.2 2025-10-10T01:40:15.1761749Z INFO 10-10 01:40:15 [__init__.py:1384] Found nccl from library libnccl.so.2 2025-10-10T01:40:15.1762173Z INFO 10-10 01:40:15 [pynccl.py:104] vLLM is using nccl==2.27.5 2025-10-10T01:40:15.1762555Z INFO 10-10 01:40:15 [pynccl.py:104] vLLM is using nccl==2.27.5 2025-10-10T01:40:15.2113793Z INFO 10-10 01:40:15 [parallel_state.py:1208] rank 0 in world size 2 is assigned as DP rank 0, PP rank 0, TP rank 0, EP rank 0 2025-10-10T01:40:15.2114588Z INFO 10-10 01:40:15 [parallel_state.py:1208] rank 1 in world size 2 is assigned as DP rank 0, PP rank 0, TP rank 1, EP rank 1 2025-10-10T01:40:15.4621379Z INFO 10-10 01:40:15 [topk_topp_sampler.py:55] Using FlashInfer for top-p & top-k sampling. 2025-10-10T01:40:15.4662876Z INFO 10-10 01:40:15 [topk_topp_sampler.py:55] Using FlashInfer for top-p & top-k sampling. 2025-10-10T01:40:15.4948562Z (Worker_TP0 pid=27921) INFO 10-10 01:40:15 [gpu_model_runner.py:2707] Starting to load model meta-llama/Llama-3.2-1B-Instruct... 2025-10-10T01:40:15.4973992Z (Worker_TP1 pid=27922) INFO 10-10 01:40:15 [gpu_model_runner.py:2707] Starting to load model meta-llama/Llama-3.2-1B-Instruct... 2025-10-10T01:40:15.7176376Z (Worker_TP0 pid=27921) INFO 10-10 01:40:15 [gpu_model_runner.py:2739] Loading model from scratch... 2025-10-10T01:40:15.7264751Z (Worker_TP1 pid=27922) INFO 10-10 01:40:15 [gpu_model_runner.py:2739] Loading model from scratch... 2025-10-10T01:40:15.8071211Z (Worker_TP0 pid=27921) INFO 10-10 01:40:15 [cuda.py:361] Using Flash Attention backend on V1 engine. 2025-10-10T01:40:15.8165218Z (Worker_TP1 pid=27922) INFO 10-10 01:40:15 [cuda.py:361] Using Flash Attention backend on V1 engine. 2025-10-10T01:40:16.3348108Z (Worker_TP0 pid=27921) INFO 10-10 01:40:16 [gpu_model_runner.py:2758] Model loading took 1.1667 GiB and 0.167108 seconds 2025-10-10T01:40:16.3537898Z (Worker_TP1 pid=27922) INFO 10-10 01:40:16 [gpu_model_runner.py:2758] Model loading took 1.1667 GiB and 0.167649 seconds 2025-10-10T01:40:19.8993038Z (Worker_TP1 pid=27922) INFO 10-10 01:40:19 [backends.py:548] Using cache directory: /var/lib/jenkins/.cache/vllm/torch_compile_cache/387e456a16/rank_1_0/backbone for vLLM's torch.compile 2025-10-10T01:40:19.8997774Z (Worker_TP1 pid=27922) INFO 10-10 01:40:19 [backends.py:559] Dynamo bytecode transform time: 3.24 s 2025-10-10T01:40:19.9517427Z (Worker_TP0 pid=27921) INFO 10-10 01:40:19 [backends.py:548] Using cache directory: /var/lib/jenkins/.cache/vllm/torch_compile_cache/387e456a16/rank_0_0/backbone for vLLM's torch.compile 2025-10-10T01:40:19.9524252Z (Worker_TP0 pid=27921) INFO 10-10 01:40:19 [backends.py:559] Dynamo bytecode transform time: 3.30 s 2025-10-10T01:40:21.0320160Z (Worker_TP1 pid=27922) INFO 10-10 01:40:21 [backends.py:164] Directly load the compiled graph(s) for dynamic shape from the cache, took 0.972 s 2025-10-10T01:40:21.0934178Z (Worker_TP0 pid=27921) INFO 10-10 01:40:21 [backends.py:164] Directly load the compiled graph(s) for dynamic shape from the cache, took 0.981 s 2025-10-10T01:40:21.6311433Z (Worker_TP1 pid=27922) INFO 10-10 01:40:21 [monitor.py:32] torch.compile takes 3.24 s in total 2025-10-10T01:40:21.6312289Z (Worker_TP0 pid=27921) INFO 10-10 01:40:21 [monitor.py:32] torch.compile takes 3.30 s in total 2025-10-10T01:40:22.3213411Z (Worker_TP0 pid=27921) INFO 10-10 01:40:22 [gpu_worker.py:298] Available KV cache memory: 18.54 GiB 2025-10-10T01:40:22.3218198Z (Worker_TP1 pid=27922) INFO 10-10 01:40:22 [gpu_worker.py:298] Available KV cache memory: 18.54 GiB 2025-10-10T01:40:22.5879144Z (EngineCore_DP0 pid=27802) INFO 10-10 01:40:22 [kv_cache_utils.py:1087] GPU KV cache size: 1,214,720 tokens 2025-10-10T01:40:22.5880115Z (EngineCore_DP0 pid=27802) INFO 10-10 01:40:22 [kv_cache_utils.py:1091] Maximum concurrency for 2,048 tokens per request: 593.12x 2025-10-10T01:40:22.5880965Z (EngineCore_DP0 pid=27802) INFO 10-10 01:40:22 [kv_cache_utils.py:1087] GPU KV cache size: 1,214,720 tokens 2025-10-10T01:40:22.5881802Z (EngineCore_DP0 pid=27802) INFO 10-10 01:40:22 [kv_cache_utils.py:1091] Maximum concurrency for 2,048 tokens per request: 593.12x 2025-10-10T01:40:22.8539029Z (Worker_TP0 pid=27921) 2025-10-10T01:40:22.9749793Z Capturing CUDA graphs (mixed prefill-decode, PIECEWISE): 0% 0/5 [00:00, 'use_cudagraph': True, 'cudagraph_num_of_warmups': 1, 'cudagraph_capture_sizes': [16, 8, 4, 2], 'cudagraph_copy_inputs': False, 'full_cuda_graph': True, 'use_inductor_graph_partition': False, 'pass_config': {'enable_noop': True, 'enable_sequence_parallelism': True}, 'max_capture_size': 16, 'local_cache_dir': None} 2025-10-10T01:40:45.0238047Z (EngineCore_DP0 pid=28298) 2025-10-10 01:40:45,023 INFO worker.py:1918 -- Started a local Ray instance. View the dashboard at http://127.0.0.1:8265  2025-10-10T01:40:45.8356041Z (EngineCore_DP0 pid=28298) INFO 10-10 01:40:45 [ray_utils.py:345] No current placement group found. Creating a new placement group. 2025-10-10T01:40:46.0099351Z (EngineCore_DP0 pid=28298) INFO 10-10 01:40:46 [ray_distributed_executor.py:171] use_ray_spmd_worker: True 2025-10-10T01:40:46.6297175Z (EngineCore_DP0 pid=28298) (pid=29185) /opt/conda/envs/py_3.12/lib/python3.12/site-packages/torch/cuda/__init__.py:64: FutureWarning: The pynvml package is deprecated. Please install nvidia-ml-py instead. If you did not install pynvml directly, please report this to the maintainers of the package that installed pynvml for you. 2025-10-10T01:40:46.6298776Z (EngineCore_DP0 pid=28298) (pid=29185) import pynvml # type: ignore[import] 2025-10-10T01:40:48.8760403Z (EngineCore_DP0 pid=28298) (pid=29180) INFO 10-10 01:40:48 [__init__.py:215] Automatically detected platform cuda. 2025-10-10T01:40:50.3076570Z (EngineCore_DP0 pid=28298) INFO 10-10 01:40:50 [ray_env.py:63] RAY_NON_CARRY_OVER_ENV_VARS from config: set() 2025-10-10T01:40:50.3078689Z (EngineCore_DP0 pid=28298) INFO 10-10 01:40:50 [ray_env.py:65] Copying the following environment variables to workers: ['LD_LIBRARY_PATH', 'VLLM_WORKER_MULTIPROC_METHOD', 'VLLM_USE_RAY_COMPILED_DAG', 'HF_TOKEN', 'MAX_JOBS', 'HUGGING_FACE_HUB_TOKEN', 'VLLM_USE_V1', 'VLLM_USE_RAY_SPMD_WORKER'] 2025-10-10T01:40:50.3080332Z (EngineCore_DP0 pid=28298) INFO 10-10 01:40:50 [ray_env.py:68] If certain env vars should NOT be copied, add them to /var/lib/jenkins/.config/vllm/ray_non_carry_over_env_vars.json file 2025-10-10T01:40:53.8764372Z (EngineCore_DP0 pid=28298) (RayWorkerWrapper pid=29185) [Gloo] Rank 0 is connected to 3 peer ranks. Expected number of connected peer ranks is : 3 2025-10-10T01:40:53.8767656Z (EngineCore_DP0 pid=28298) (pid=29190) INFO 10-10 01:40:48 [__init__.py:215] Automatically detected platform cuda. [repeated 3x across cluster] (Ray deduplicates logs by default. Set RAY_DEDUP_LOGS=0 to disable log deduplication, or see https://docs.ray.io/en/master/ray-observability/user-guides/configure-logging.html#log-deduplication for more options.) 2025-10-10T01:40:53.9820026Z (EngineCore_DP0 pid=28298) (RayWorkerWrapper pid=29185) INFO 10-10 01:40:53 [__init__.py:1384] Found nccl from library libnccl.so.2 2025-10-10T01:40:53.9822025Z (EngineCore_DP0 pid=28298) (RayWorkerWrapper pid=29185) INFO 10-10 01:40:53 [pynccl.py:104] vLLM is using nccl==2.27.5 2025-10-10T01:40:54.2931009Z (EngineCore_DP0 pid=28298) (RayWorkerWrapper pid=29185) WARNING 10-10 01:40:54 [symm_mem.py:58] SymmMemCommunicator: Device capability 8.9 not supported, communicator is not available. 2025-10-10T01:40:54.2932455Z (EngineCore_DP0 pid=28298) (RayWorkerWrapper pid=29185) INFO 10-10 01:40:54 [custom_all_reduce.py:35] Skipping P2P check and trusting the driver's P2P report. 2025-10-10T01:40:54.2934140Z (EngineCore_DP0 pid=28298) (RayWorkerWrapper pid=29185) WARNING 10-10 01:40:54 [custom_all_reduce.py:154] Custom allreduce is disabled because your platform lacks GPU P2P capability or P2P test failed. To silence this warning, specify disable_custom_all_reduce=True explicitly. 2025-10-10T01:40:54.2936477Z (EngineCore_DP0 pid=28298) (RayWorkerWrapper pid=29185) INFO 10-10 01:40:54 [shm_broadcast.py:289] vLLM message queue communication handle: Handle(local_reader_ranks=[1], buffer_handle=(1, 4194304, 6, 'psm_0f0892b5'), local_subscribe_addr='ipc:///tmp/ee38e51b-371c-48a2-85da-09c6b1053eb5', remote_subscribe_addr=None, remote_addr_ipv6=False) 2025-10-10T01:40:54.3982693Z (EngineCore_DP0 pid=28298) (RayWorkerWrapper pid=29185) INFO 10-10 01:40:54 [parallel_state.py:1208] rank 0 in world size 4 is assigned as DP rank 0, PP rank 0, TP rank 0, EP rank 0 2025-10-10T01:40:54.7079150Z (EngineCore_DP0 pid=28298) (RayWorkerWrapper pid=29185) INFO 10-10 01:40:54 [topk_topp_sampler.py:55] Using FlashInfer for top-p & top-k sampling. 2025-10-10T01:40:54.7080483Z (EngineCore_DP0 pid=28298) (RayWorkerWrapper pid=29185) INFO 10-10 01:40:54 [gpu_model_runner.py:2707] Starting to load model meta-llama/Llama-3.2-1B-Instruct... 2025-10-10T01:40:54.9147583Z (EngineCore_DP0 pid=28298) (RayWorkerWrapper pid=29190) INFO 10-10 01:40:54 [gpu_model_runner.py:2739] Loading model from scratch... 2025-10-10T01:40:55.0185972Z (EngineCore_DP0 pid=28298) (RayWorkerWrapper pid=29185) INFO 10-10 01:40:55 [cuda.py:361] Using Flash Attention backend on V1 engine. 2025-10-10T01:40:55.5310018Z (EngineCore_DP0 pid=28298) (RayWorkerWrapper pid=29190) INFO 10-10 01:40:55 [gpu_model_runner.py:2758] Model loading took 0.7135 GiB and 0.155415 seconds 2025-10-10T01:40:57.9805981Z (EngineCore_DP0 pid=28298) (RayWorkerWrapper pid=29185) INFO 10-10 01:40:57 [backends.py:548] Using cache directory: /var/lib/jenkins/.cache/vllm/torch_compile_cache/5280ab4e2b/rank_0_0/backbone for vLLM's torch.compile 2025-10-10T01:40:57.9807836Z (EngineCore_DP0 pid=28298) (RayWorkerWrapper pid=29185) INFO 10-10 01:40:57 [backends.py:559] Dynamo bytecode transform time: 2.05 s 2025-10-10T01:40:58.5951925Z (EngineCore_DP0 pid=28298) (RayWorkerWrapper pid=29185) INFO 10-10 01:40:58 [backends.py:164] Directly load the compiled graph(s) for dynamic shape from the cache, took 0.409 s 2025-10-10T01:40:59.6203704Z (EngineCore_DP0 pid=28298) (RayWorkerWrapper pid=29185) INFO 10-10 01:40:59 [gpu_worker.py:298] Available KV cache memory: 19.02 GiB 2025-10-10T01:40:59.6205893Z (EngineCore_DP0 pid=28298) (RayWorkerWrapper pid=29180) [Gloo] Rank 1 is connected to 1 peer ranks. Expected number of connected peer ranks is : 1 [repeated 23x across cluster] 2025-10-10T01:40:59.6207364Z (EngineCore_DP0 pid=28298) (RayWorkerWrapper pid=29180) INFO 10-10 01:40:54 [__init__.py:1384] Found nccl from library libnccl.so.2 [repeated 11x across cluster] 2025-10-10T01:40:59.6208563Z (EngineCore_DP0 pid=28298) (RayWorkerWrapper pid=29180) INFO 10-10 01:40:54 [pynccl.py:104] vLLM is using nccl==2.27.5 [repeated 11x across cluster] 2025-10-10T01:40:59.6209904Z (EngineCore_DP0 pid=28298) (RayWorkerWrapper pid=29180) WARNING 10-10 01:40:54 [symm_mem.py:58] SymmMemCommunicator: Device capability 8.9 not supported, communicator is not available. [repeated 3x across cluster] 2025-10-10T01:40:59.6211160Z (EngineCore_DP0 pid=28298) (RayWorkerWrapper pid=29180) INFO 10-10 01:40:54 [custom_all_reduce.py:35] Skipping P2P check and trusting the driver's P2P report. [repeated 3x across cluster] 2025-10-10T01:40:59.6212723Z (EngineCore_DP0 pid=28298) (RayWorkerWrapper pid=29180) WARNING 10-10 01:40:54 [custom_all_reduce.py:154] Custom allreduce is disabled because your platform lacks GPU P2P capability or P2P test failed. To silence this warning, specify disable_custom_all_reduce=True explicitly. [repeated 3x across cluster] 2025-10-10T01:40:59.6214686Z (EngineCore_DP0 pid=28298) (RayWorkerWrapper pid=29183) INFO 10-10 01:40:54 [shm_broadcast.py:289] vLLM message queue communication handle: Handle(local_reader_ranks=[1], buffer_handle=(1, 4194304, 6, 'psm_a3ea99ff'), local_subscribe_addr='ipc:///tmp/dde49f4b-ac78-4839-91ba-fc5820614c41', remote_subscribe_addr=None, remote_addr_ipv6=False) 2025-10-10T01:40:59.6216254Z (EngineCore_DP0 pid=28298) (RayWorkerWrapper pid=29180) INFO 10-10 01:40:54 [parallel_state.py:1208] rank 3 in world size 4 is assigned as DP rank 0, PP rank 1, TP rank 1, EP rank 1 [repeated 3x across cluster] 2025-10-10T01:40:59.9093158Z (EngineCore_DP0 pid=28298) INFO 10-10 01:40:59 [kv_cache_utils.py:1087] GPU KV cache size: 2,492,416 tokens 2025-10-10T01:40:59.9094254Z (EngineCore_DP0 pid=28298) INFO 10-10 01:40:59 [kv_cache_utils.py:1091] Maximum concurrency for 2,048 tokens per request: 1217.00x 2025-10-10T01:40:59.9095124Z (EngineCore_DP0 pid=28298) INFO 10-10 01:40:59 [kv_cache_utils.py:1087] GPU KV cache size: 2,492,416 tokens 2025-10-10T01:40:59.9095955Z (EngineCore_DP0 pid=28298) INFO 10-10 01:40:59 [kv_cache_utils.py:1091] Maximum concurrency for 2,048 tokens per request: 1217.00x 2025-10-10T01:40:59.9096788Z (EngineCore_DP0 pid=28298) INFO 10-10 01:40:59 [kv_cache_utils.py:1087] GPU KV cache size: 2,489,088 tokens 2025-10-10T01:40:59.9097596Z (EngineCore_DP0 pid=28298) INFO 10-10 01:40:59 [kv_cache_utils.py:1091] Maximum concurrency for 2,048 tokens per request: 1215.38x 2025-10-10T01:40:59.9098400Z (EngineCore_DP0 pid=28298) INFO 10-10 01:40:59 [kv_cache_utils.py:1087] GPU KV cache size: 2,489,088 tokens 2025-10-10T01:40:59.9099661Z (EngineCore_DP0 pid=28298) INFO 10-10 01:40:59 [kv_cache_utils.py:1091] Maximum concurrency for 2,048 tokens per request: 1215.38x 2025-10-10T01:41:00.2351971Z (EngineCore_DP0 pid=28298) (RayWorkerWrapper pid=29185) 2025-10-10T01:41:00.2352642Z Capturing CUDA graphs (mixed prefill-decode, PIECEWISE): 0%| | 0/4 [00:00, 'use_cudagraph': True, 'cudagraph_num_of_warmups': 1, 'cudagraph_capture_sizes': [16, 8, 4, 2, 1], 'cudagraph_copy_inputs': False, 'full_cuda_graph': True, 'use_inductor_graph_partition': False, 'pass_config': {}, 'max_capture_size': 16, 'local_cache_dir': None} 2025-10-10T01:41:22.8691004Z (EngineCore_DP0 pid=33348) WARNING 10-10 01:41:22 [multiproc_executor.py:720] Reducing Torch parallelism from 24 threads to 1 to avoid unnecessary CPU contention. Set OMP_NUM_THREADS in the external environment to tune this value as needed. 2025-10-10T01:41:22.8695329Z (EngineCore_DP0 pid=33348) INFO 10-10 01:41:22 [shm_broadcast.py:289] vLLM message queue communication handle: Handle(local_reader_ranks=[0, 1], buffer_handle=(2, 16777216, 10, 'psm_e4637896'), local_subscribe_addr='ipc:///tmp/6b35f921-565a-44f2-b427-11e102ca1707', remote_subscribe_addr=None, remote_addr_ipv6=False) 2025-10-10T01:41:23.3480095Z /opt/conda/envs/py_3.12/lib/python3.12/site-packages/torch/cuda/__init__.py:64: FutureWarning: The pynvml package is deprecated. Please install nvidia-ml-py instead. If you did not install pynvml directly, please report this to the maintainers of the package that installed pynvml for you. 2025-10-10T01:41:23.3481384Z import pynvml # type: ignore[import] 2025-10-10T01:41:23.3483027Z /opt/conda/envs/py_3.12/lib/python3.12/site-packages/torch/cuda/__init__.py:64: FutureWarning: The pynvml package is deprecated. Please install nvidia-ml-py instead. If you did not install pynvml directly, please report this to the maintainers of the package that installed pynvml for you. 2025-10-10T01:41:23.3484255Z import pynvml # type: ignore[import] 2025-10-10T01:41:25.6633252Z INFO 10-10 01:41:25 [__init__.py:215] Automatically detected platform cuda. 2025-10-10T01:41:25.7064861Z INFO 10-10 01:41:25 [__init__.py:215] Automatically detected platform cuda. 2025-10-10T01:41:30.2442868Z INFO 10-10 01:41:30 [shm_broadcast.py:289] vLLM message queue communication handle: Handle(local_reader_ranks=[0], buffer_handle=(1, 10485760, 10, 'psm_b8453746'), local_subscribe_addr='ipc:///tmp/0642a402-32e3-4ed8-aedc-cadb52fac751', remote_subscribe_addr=None, remote_addr_ipv6=False) 2025-10-10T01:41:30.2640085Z INFO 10-10 01:41:30 [shm_broadcast.py:289] vLLM message queue communication handle: Handle(local_reader_ranks=[0], buffer_handle=(1, 10485760, 10, 'psm_841e33d1'), local_subscribe_addr='ipc:///tmp/78b43694-9d89-4d0c-9ec9-a985f1d573a3', remote_subscribe_addr=None, remote_addr_ipv6=False) 2025-10-10T01:41:30.4645600Z [Gloo] Rank 1 is connected to 1 peer ranks. Expected number of connected peer ranks is : 1 2025-10-10T01:41:30.4646257Z [Gloo] Rank 0 is connected to 1 peer ranks. Expected number of connected peer ranks is : 1 2025-10-10T01:41:30.4689054Z [Gloo] Rank 1 is connected to 1 peer ranks. Expected number of connected peer ranks is : 1 2025-10-10T01:41:30.4689625Z [Gloo] Rank 0 is connected to 1 peer ranks. Expected number of connected peer ranks is : 1 2025-10-10T01:41:30.4733555Z INFO 10-10 01:41:30 [__init__.py:1384] Found nccl from library libnccl.so.2 2025-10-10T01:41:30.4734053Z INFO 10-10 01:41:30 [__init__.py:1384] Found nccl from library libnccl.so.2 2025-10-10T01:41:30.4734473Z INFO 10-10 01:41:30 [pynccl.py:104] vLLM is using nccl==2.27.5 2025-10-10T01:41:30.4736759Z INFO 10-10 01:41:30 [pynccl.py:104] vLLM is using nccl==2.27.5 2025-10-10T01:41:30.6440672Z WARNING 10-10 01:41:30 [symm_mem.py:58] SymmMemCommunicator: Device capability 8.9 not supported, communicator is not available. 2025-10-10T01:41:30.6441745Z WARNING 10-10 01:41:30 [symm_mem.py:58] SymmMemCommunicator: Device capability 8.9 not supported, communicator is not available. 2025-10-10T01:41:30.6460853Z INFO 10-10 01:41:30 [custom_all_reduce.py:35] Skipping P2P check and trusting the driver's P2P report. 2025-10-10T01:41:30.6462097Z INFO 10-10 01:41:30 [custom_all_reduce.py:35] Skipping P2P check and trusting the driver's P2P report. 2025-10-10T01:41:30.6464077Z WARNING 10-10 01:41:30 [custom_all_reduce.py:154] Custom allreduce is disabled because your platform lacks GPU P2P capability or P2P test failed. To silence this warning, specify disable_custom_all_reduce=True explicitly. 2025-10-10T01:41:30.6466838Z WARNING 10-10 01:41:30 [custom_all_reduce.py:154] Custom allreduce is disabled because your platform lacks GPU P2P capability or P2P test failed. To silence this warning, specify disable_custom_all_reduce=True explicitly. 2025-10-10T01:41:30.7440613Z INFO 10-10 01:41:30 [shm_broadcast.py:289] vLLM message queue communication handle: Handle(local_reader_ranks=[1], buffer_handle=(1, 4194304, 6, 'psm_2a4393bc'), local_subscribe_addr='ipc:///tmp/3b7c438b-ffa5-4bcc-b22e-5ae2536bb82c', remote_subscribe_addr=None, remote_addr_ipv6=False) 2025-10-10T01:41:30.7474477Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-10-10T01:41:30.7475092Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-10-10T01:41:30.7484180Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-10-10T01:41:30.7494747Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-10-10T01:41:30.8132785Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-10-10T01:41:30.8148322Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-10-10T01:41:30.8159190Z [Gloo] Rank 1 is connected to 1 peer ranks. Expected number of connected peer ranks is : 1 2025-10-10T01:41:30.8159819Z [Gloo] Rank 0 is connected to 1 peer ranks. Expected number of connected peer ranks is : 1 2025-10-10T01:41:30.8164370Z INFO 10-10 01:41:30 [__init__.py:1384] Found nccl from library libnccl.so.2 2025-10-10T01:41:30.8165290Z INFO 10-10 01:41:30 [__init__.py:1384] Found nccl from library libnccl.so.2 2025-10-10T01:41:30.8166103Z INFO 10-10 01:41:30 [pynccl.py:104] vLLM is using nccl==2.27.5 2025-10-10T01:41:30.8166816Z INFO 10-10 01:41:30 [pynccl.py:104] vLLM is using nccl==2.27.5 2025-10-10T01:41:30.8485381Z INFO 10-10 01:41:30 [parallel_state.py:1208] rank 0 in world size 2 is assigned as DP rank 0, PP rank 0, TP rank 0, EP rank 0 2025-10-10T01:41:30.8486141Z INFO 10-10 01:41:30 [parallel_state.py:1208] rank 1 in world size 2 is assigned as DP rank 0, PP rank 0, TP rank 1, EP rank 1 2025-10-10T01:41:31.1043555Z INFO 10-10 01:41:31 [topk_topp_sampler.py:55] Using FlashInfer for top-p & top-k sampling. 2025-10-10T01:41:31.1137360Z INFO 10-10 01:41:31 [topk_topp_sampler.py:55] Using FlashInfer for top-p & top-k sampling. 2025-10-10T01:41:31.1371773Z (Worker_TP0 pid=33467) INFO 10-10 01:41:31 [gpu_model_runner.py:2707] Starting to load model meta-llama/Llama-3.2-1B-Instruct... 2025-10-10T01:41:31.1458475Z (Worker_TP1 pid=33468) INFO 10-10 01:41:31 [gpu_model_runner.py:2707] Starting to load model meta-llama/Llama-3.2-1B-Instruct... 2025-10-10T01:41:31.3643094Z (Worker_TP0 pid=33467) INFO 10-10 01:41:31 [gpu_model_runner.py:2739] Loading model from scratch... 2025-10-10T01:41:31.3743394Z (Worker_TP1 pid=33468) INFO 10-10 01:41:31 [gpu_model_runner.py:2739] Loading model from scratch... 2025-10-10T01:41:31.4547816Z (Worker_TP0 pid=33467) INFO 10-10 01:41:31 [cuda.py:361] Using Flash Attention backend on V1 engine. 2025-10-10T01:41:31.4637431Z (Worker_TP1 pid=33468) INFO 10-10 01:41:31 [cuda.py:361] Using Flash Attention backend on V1 engine. 2025-10-10T01:41:31.9958061Z (Worker_TP0 pid=33467) INFO 10-10 01:41:31 [gpu_model_runner.py:2758] Model loading took 1.1667 GiB and 0.169668 seconds 2025-10-10T01:41:32.0026266Z (Worker_TP1 pid=33468) INFO 10-10 01:41:32 [gpu_model_runner.py:2758] Model loading took 1.1667 GiB and 0.167378 seconds 2025-10-10T01:41:35.5767878Z (Worker_TP1 pid=33468) INFO 10-10 01:41:35 [backends.py:548] Using cache directory: /var/lib/jenkins/.cache/vllm/torch_compile_cache/387e456a16/rank_1_0/backbone for vLLM's torch.compile 2025-10-10T01:41:35.5772719Z (Worker_TP1 pid=33468) INFO 10-10 01:41:35 [backends.py:559] Dynamo bytecode transform time: 3.26 s 2025-10-10T01:41:35.6196856Z (Worker_TP0 pid=33467) INFO 10-10 01:41:35 [backends.py:548] Using cache directory: /var/lib/jenkins/.cache/vllm/torch_compile_cache/387e456a16/rank_0_0/backbone for vLLM's torch.compile 2025-10-10T01:41:35.6203311Z (Worker_TP0 pid=33467) INFO 10-10 01:41:35 [backends.py:559] Dynamo bytecode transform time: 3.30 s 2025-10-10T01:41:36.7526941Z (Worker_TP1 pid=33468) INFO 10-10 01:41:36 [backends.py:164] Directly load the compiled graph(s) for dynamic shape from the cache, took 1.013 s 2025-10-10T01:41:36.8057126Z (Worker_TP0 pid=33467) INFO 10-10 01:41:36 [backends.py:164] Directly load the compiled graph(s) for dynamic shape from the cache, took 1.021 s 2025-10-10T01:41:37.3551930Z (Worker_TP1 pid=33468) INFO 10-10 01:41:37 [monitor.py:32] torch.compile takes 3.26 s in total 2025-10-10T01:41:37.3554619Z (Worker_TP0 pid=33467) INFO 10-10 01:41:37 [monitor.py:32] torch.compile takes 3.30 s in total 2025-10-10T01:41:38.0469317Z (Worker_TP1 pid=33468) INFO 10-10 01:41:38 [gpu_worker.py:298] Available KV cache memory: 18.54 GiB 2025-10-10T01:41:38.0605406Z (Worker_TP0 pid=33467) INFO 10-10 01:41:38 [gpu_worker.py:298] Available KV cache memory: 18.54 GiB 2025-10-10T01:41:38.3211398Z (EngineCore_DP0 pid=33348) INFO 10-10 01:41:38 [kv_cache_utils.py:1087] GPU KV cache size: 1,214,720 tokens 2025-10-10T01:41:38.3212435Z (EngineCore_DP0 pid=33348) INFO 10-10 01:41:38 [kv_cache_utils.py:1091] Maximum concurrency for 2,048 tokens per request: 593.12x 2025-10-10T01:41:38.3213331Z (EngineCore_DP0 pid=33348) INFO 10-10 01:41:38 [kv_cache_utils.py:1087] GPU KV cache size: 1,214,720 tokens 2025-10-10T01:41:38.3214185Z (EngineCore_DP0 pid=33348) INFO 10-10 01:41:38 [kv_cache_utils.py:1091] Maximum concurrency for 2,048 tokens per request: 593.12x 2025-10-10T01:41:38.5900472Z (Worker_TP0 pid=33467) 2025-10-10T01:41:38.7140744Z Capturing CUDA graphs (mixed prefill-decode, PIECEWISE): 0% 0/5 [00:00, 'use_cudagraph': False, 'cudagraph_num_of_warmups': 0, 'cudagraph_capture_sizes': [], 'cudagraph_copy_inputs': False, 'full_cuda_graph': False, 'use_inductor_graph_partition': False, 'pass_config': {'enable_noop': True, 'enable_sequence_parallelism': True}, 'max_capture_size': 0, 'local_cache_dir': None} 2025-10-10T01:41:59.1704776Z (EngineCore_DP0 pid=33844) WARNING 10-10 01:41:59 [multiproc_executor.py:720] Reducing Torch parallelism from 24 threads to 1 to avoid unnecessary CPU contention. Set OMP_NUM_THREADS in the external environment to tune this value as needed. 2025-10-10T01:41:59.1711292Z (EngineCore_DP0 pid=33844) INFO 10-10 01:41:59 [shm_broadcast.py:289] vLLM message queue communication handle: Handle(local_reader_ranks=[0, 1], buffer_handle=(2, 16777216, 10, 'psm_07512b14'), local_subscribe_addr='ipc:///tmp/c61a825f-cebb-44a5-b31f-deb5152ff628', remote_subscribe_addr=None, remote_addr_ipv6=False) 2025-10-10T01:41:59.6481848Z /opt/conda/envs/py_3.12/lib/python3.12/site-packages/torch/cuda/__init__.py:64: FutureWarning: The pynvml package is deprecated. Please install nvidia-ml-py instead. If you did not install pynvml directly, please report this to the maintainers of the package that installed pynvml for you. 2025-10-10T01:41:59.6484162Z import pynvml # type: ignore[import] 2025-10-10T01:41:59.6683315Z /opt/conda/envs/py_3.12/lib/python3.12/site-packages/torch/cuda/__init__.py:64: FutureWarning: The pynvml package is deprecated. Please install nvidia-ml-py instead. If you did not install pynvml directly, please report this to the maintainers of the package that installed pynvml for you. 2025-10-10T01:41:59.6685541Z import pynvml # type: ignore[import] 2025-10-10T01:42:01.9830648Z INFO 10-10 01:42:01 [__init__.py:215] Automatically detected platform cuda. 2025-10-10T01:42:02.0058324Z INFO 10-10 01:42:02 [__init__.py:215] Automatically detected platform cuda. 2025-10-10T01:42:06.6245288Z INFO 10-10 01:42:06 [shm_broadcast.py:289] vLLM message queue communication handle: Handle(local_reader_ranks=[0], buffer_handle=(1, 10485760, 10, 'psm_d3e66e60'), local_subscribe_addr='ipc:///tmp/e13e10b1-5344-4b88-b383-eba904e89a74', remote_subscribe_addr=None, remote_addr_ipv6=False) 2025-10-10T01:42:06.6840768Z INFO 10-10 01:42:06 [shm_broadcast.py:289] vLLM message queue communication handle: Handle(local_reader_ranks=[0], buffer_handle=(1, 10485760, 10, 'psm_817b7920'), local_subscribe_addr='ipc:///tmp/38b06aa5-bbc1-4ef9-bef8-83db4666f274', remote_subscribe_addr=None, remote_addr_ipv6=False) 2025-10-10T01:42:07.0917819Z [Gloo] Rank 0 is connected to 1 peer ranks. Expected number of connected peer ranks is : 1 2025-10-10T01:42:07.0919232Z [Gloo] Rank 1 is connected to 1 peer ranks. Expected number of connected peer ranks is : 1 2025-10-10T01:42:07.0964936Z [Gloo] Rank 0 is connected to 1 peer ranks. Expected number of connected peer ranks is : 1 2025-10-10T01:42:07.0965534Z [Gloo] Rank 1 is connected to 1 peer ranks. Expected number of connected peer ranks is : 1 2025-10-10T01:42:07.1007390Z INFO 10-10 01:42:07 [__init__.py:1384] Found nccl from library libnccl.so.2 2025-10-10T01:42:07.1008181Z INFO 10-10 01:42:07 [__init__.py:1384] Found nccl from library libnccl.so.2 2025-10-10T01:42:07.1012340Z INFO 10-10 01:42:07 [pynccl.py:104] vLLM is using nccl==2.27.5 2025-10-10T01:42:07.1012851Z INFO 10-10 01:42:07 [pynccl.py:104] vLLM is using nccl==2.27.5 2025-10-10T01:42:07.2747699Z WARNING 10-10 01:42:07 [symm_mem.py:58] SymmMemCommunicator: Device capability 8.9 not supported, communicator is not available. 2025-10-10T01:42:07.2749298Z WARNING 10-10 01:42:07 [symm_mem.py:58] SymmMemCommunicator: Device capability 8.9 not supported, communicator is not available. 2025-10-10T01:42:07.2772682Z INFO 10-10 01:42:07 [custom_all_reduce.py:35] Skipping P2P check and trusting the driver's P2P report. 2025-10-10T01:42:07.2773736Z INFO 10-10 01:42:07 [custom_all_reduce.py:35] Skipping P2P check and trusting the driver's P2P report. 2025-10-10T01:42:07.2775070Z WARNING 10-10 01:42:07 [custom_all_reduce.py:154] Custom allreduce is disabled because your platform lacks GPU P2P capability or P2P test failed. To silence this warning, specify disable_custom_all_reduce=True explicitly. 2025-10-10T01:42:07.2776479Z WARNING 10-10 01:42:07 [custom_all_reduce.py:154] Custom allreduce is disabled because your platform lacks GPU P2P capability or P2P test failed. To silence this warning, specify disable_custom_all_reduce=True explicitly. 2025-10-10T01:42:07.3540796Z INFO 10-10 01:42:07 [shm_broadcast.py:289] vLLM message queue communication handle: Handle(local_reader_ranks=[1], buffer_handle=(1, 4194304, 6, 'psm_6c38f64d'), local_subscribe_addr='ipc:///tmp/4b71efb0-09f0-4596-ae05-f2ae1d309d57', remote_subscribe_addr=None, remote_addr_ipv6=False) 2025-10-10T01:42:07.3571409Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-10-10T01:42:07.3572575Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-10-10T01:42:07.3578801Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-10-10T01:42:07.3587332Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-10-10T01:42:07.4136551Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-10-10T01:42:07.4146913Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-10-10T01:42:07.4159376Z [Gloo] Rank 0 is connected to 1 peer ranks. Expected number of connected peer ranks is : 1 2025-10-10T01:42:07.4159973Z [Gloo] Rank 1 is connected to 1 peer ranks. Expected number of connected peer ranks is : 1 2025-10-10T01:42:07.4162217Z INFO 10-10 01:42:07 [__init__.py:1384] Found nccl from library libnccl.so.2 2025-10-10T01:42:07.4162774Z INFO 10-10 01:42:07 [__init__.py:1384] Found nccl from library libnccl.so.2 2025-10-10T01:42:07.4163276Z INFO 10-10 01:42:07 [pynccl.py:104] vLLM is using nccl==2.27.5 2025-10-10T01:42:07.4515306Z INFO 10-10 01:42:07 [pynccl.py:104] vLLM is using nccl==2.27.5 2025-10-10T01:42:07.4516409Z INFO 10-10 01:42:07 [parallel_state.py:1208] rank 0 in world size 2 is assigned as DP rank 0, PP rank 0, TP rank 0, EP rank 0 2025-10-10T01:42:07.4517745Z INFO 10-10 01:42:07 [parallel_state.py:1208] rank 1 in world size 2 is assigned as DP rank 0, PP rank 0, TP rank 1, EP rank 1 2025-10-10T01:42:07.7052516Z INFO 10-10 01:42:07 [topk_topp_sampler.py:55] Using FlashInfer for top-p & top-k sampling. 2025-10-10T01:42:07.7060832Z INFO 10-10 01:42:07 [topk_topp_sampler.py:55] Using FlashInfer for top-p & top-k sampling. 2025-10-10T01:42:07.7373588Z (Worker_TP1 pid=33964) INFO 10-10 01:42:07 [gpu_model_runner.py:2707] Starting to load model meta-llama/Llama-3.2-1B-Instruct... 2025-10-10T01:42:07.7388320Z (Worker_TP0 pid=33963) INFO 10-10 01:42:07 [gpu_model_runner.py:2707] Starting to load model meta-llama/Llama-3.2-1B-Instruct... 2025-10-10T01:42:07.9605883Z (Worker_TP1 pid=33964) INFO 10-10 01:42:07 [gpu_model_runner.py:2739] Loading model from scratch... 2025-10-10T01:42:07.9661270Z (Worker_TP0 pid=33963) INFO 10-10 01:42:07 [gpu_model_runner.py:2739] Loading model from scratch... 2025-10-10T01:42:08.0503133Z (Worker_TP1 pid=33964) INFO 10-10 01:42:08 [cuda.py:361] Using Flash Attention backend on V1 engine. 2025-10-10T01:42:08.0562451Z (Worker_TP0 pid=33963) INFO 10-10 01:42:08 [cuda.py:361] Using Flash Attention backend on V1 engine. 2025-10-10T01:42:08.5839730Z (Worker_TP1 pid=33964) INFO 10-10 01:42:08 [gpu_model_runner.py:2758] Model loading took 1.1667 GiB and 0.167492 seconds 2025-10-10T01:42:08.5868404Z (Worker_TP0 pid=33963) INFO 10-10 01:42:08 [gpu_model_runner.py:2758] Model loading took 1.1667 GiB and 0.167876 seconds 2025-10-10T01:42:11.7975781Z (Worker_TP0 pid=33963) INFO 10-10 01:42:11 [backends.py:548] Using cache directory: /var/lib/jenkins/.cache/vllm/torch_compile_cache/31e0bc77eb/rank_0_0/backbone for vLLM's torch.compile 2025-10-10T01:42:11.7985813Z (Worker_TP0 pid=33963) INFO 10-10 01:42:11 [backends.py:559] Dynamo bytecode transform time: 2.90 s 2025-10-10T01:42:11.8391044Z (Worker_TP1 pid=33964) INFO 10-10 01:42:11 [backends.py:548] Using cache directory: /var/lib/jenkins/.cache/vllm/torch_compile_cache/31e0bc77eb/rank_1_0/backbone for vLLM's torch.compile 2025-10-10T01:42:11.8401420Z (Worker_TP1 pid=33964) INFO 10-10 01:42:11 [backends.py:559] Dynamo bytecode transform time: 2.93 s 2025-10-10T01:42:12.8615502Z (Worker_TP0 pid=33963) INFO 10-10 01:42:12 [backends.py:164] Directly load the compiled graph(s) for dynamic shape from the cache, took 0.807 s 2025-10-10T01:42:12.8680608Z (Worker_TP1 pid=33964) INFO 10-10 01:42:12 [backends.py:164] Directly load the compiled graph(s) for dynamic shape from the cache, took 0.779 s 2025-10-10T01:42:14.1033931Z (Worker_TP1 pid=33964) INFO 10-10 01:42:14 [gpu_worker.py:298] Available KV cache memory: 18.54 GiB 2025-10-10T01:42:14.1195739Z (Worker_TP0 pid=33963) INFO 10-10 01:42:14 [gpu_worker.py:298] Available KV cache memory: 18.54 GiB 2025-10-10T01:42:14.3786383Z (EngineCore_DP0 pid=33844) INFO 10-10 01:42:14 [kv_cache_utils.py:1087] GPU KV cache size: 1,215,232 tokens 2025-10-10T01:42:14.3787304Z (EngineCore_DP0 pid=33844) INFO 10-10 01:42:14 [kv_cache_utils.py:1091] Maximum concurrency for 2,048 tokens per request: 593.38x 2025-10-10T01:42:14.3788133Z (EngineCore_DP0 pid=33844) INFO 10-10 01:42:14 [kv_cache_utils.py:1087] GPU KV cache size: 1,215,232 tokens 2025-10-10T01:42:14.3788956Z (EngineCore_DP0 pid=33844) INFO 10-10 01:42:14 [kv_cache_utils.py:1091] Maximum concurrency for 2,048 tokens per request: 593.38x 2025-10-10T01:42:14.3857039Z (Worker_TP1 pid=33964) INFO 10-10 01:42:14 [gpu_worker.py:332] Compile and warming up model for size 8 2025-10-10T01:42:14.3857794Z (Worker_TP0 pid=33963) INFO 10-10 01:42:14 [gpu_worker.py:332] Compile and warming up model for size 8 2025-10-10T01:42:15.1562822Z (Worker_TP1 pid=33964) INFO 10-10 01:42:15 [backends.py:168] Directly load the compiled graph(s) for shape 8 from the cache, took 0.770 s 2025-10-10T01:42:15.1615273Z (Worker_TP0 pid=33963) INFO 10-10 01:42:15 [backends.py:168] Directly load the compiled graph(s) for shape 8 from the cache, took 0.775 s 2025-10-10T01:42:15.1621713Z (Worker_TP1 pid=33964) INFO 10-10 01:42:15 [gpu_worker.py:332] Compile and warming up model for size 4 2025-10-10T01:42:15.1622859Z (Worker_TP0 pid=33963) INFO 10-10 01:42:15 [gpu_worker.py:332] Compile and warming up model for size 4 2025-10-10T01:42:15.8195620Z (Worker_TP1 pid=33964) INFO 10-10 01:42:15 [backends.py:168] Directly load the compiled graph(s) for shape 4 from the cache, took 0.657 s 2025-10-10T01:42:15.8196596Z (Worker_TP1 pid=33964) INFO 10-10 01:42:15 [monitor.py:32] torch.compile takes 2.93 s in total 2025-10-10T01:42:15.8224147Z (Worker_TP0 pid=33963) INFO 10-10 01:42:15 [backends.py:168] Directly load the compiled graph(s) for shape 4 from the cache, took 0.660 s 2025-10-10T01:42:15.8225044Z (Worker_TP0 pid=33963) INFO 10-10 01:42:15 [monitor.py:32] torch.compile takes 2.90 s in total 2025-10-10T01:42:15.8351317Z (EngineCore_DP0 pid=33844) INFO 10-10 01:42:15 [core.py:211] init engine (profile, create kv cache, warmup model) took 7.15 seconds 2025-10-10T01:42:16.3485020Z (EngineCore_DP0 pid=33844) INFO 10-10 01:42:16 [vllm.py:345] Cudagraph is disabled under eager mode 2025-10-10T01:42:16.5337462Z (APIServer pid=33722) INFO 10-10 01:42:16 [loggers.py:147] Engine 000: vllm cache_config_info with initialization after num_gpu_blocks is: 75952 2025-10-10T01:42:16.5708519Z (EngineCore_DP0 pid=33844) INFO 10-10 01:42:16 [gc_utils.py:41] GC Debug Config. enabled:False,top_objects:-1 2025-10-10T01:42:16.5928097Z (APIServer pid=33722) INFO 10-10 01:42:16 [api_server.py:1634] Supported_tasks: ['generate'] 2025-10-10T01:42:16.6140540Z (APIServer pid=33722) WARNING 10-10 01:42:16 [model.py:1424] Default sampling parameters have been overridden by the model's Hugging Face generation config recommended from the model creator. If this is not intended, please relaunch vLLM instance with `--generation-config vllm`. 2025-10-10T01:42:16.6142145Z (APIServer pid=33722) INFO 10-10 01:42:16 [serving_responses.py:140] Using default chat sampling params from model: {'temperature': 0.6, 'top_p': 0.9} 2025-10-10T01:42:16.6413984Z (APIServer pid=33722) INFO 10-10 01:42:16 [serving_chat.py:139] Using default chat sampling params from model: {'temperature': 0.6, 'top_p': 0.9} 2025-10-10T01:42:16.6698323Z (APIServer pid=33722) INFO 10-10 01:42:16 [serving_completion.py:76] Using default completion sampling params from model: {'temperature': 0.6, 'top_p': 0.9} 2025-10-10T01:42:16.6699684Z (APIServer pid=33722) INFO 10-10 01:42:16 [api_server.py:1912] Starting vLLM API server 0 on http://0.0.0.0:43109 2025-10-10T01:42:16.6700528Z (APIServer pid=33722) INFO 10-10 01:42:16 [launcher.py:34] Available routes are: 2025-10-10T01:42:16.6701714Z (APIServer pid=33722) INFO 10-10 01:42:16 [launcher.py:42] Route: /openapi.json, Methods: HEAD, GET 2025-10-10T01:42:16.6702867Z (APIServer pid=33722) INFO 10-10 01:42:16 [launcher.py:42] Route: /docs, Methods: HEAD, GET 2025-10-10T01:42:16.6703632Z (APIServer pid=33722) INFO 10-10 01:42:16 [launcher.py:42] Route: /docs/oauth2-redirect, Methods: HEAD, GET 2025-10-10T01:42:16.6704338Z (APIServer pid=33722) INFO 10-10 01:42:16 [launcher.py:42] Route: /redoc, Methods: HEAD, GET 2025-10-10T01:42:16.6705012Z (APIServer pid=33722) INFO 10-10 01:42:16 [launcher.py:42] Route: /health, Methods: GET 2025-10-10T01:42:16.6705664Z (APIServer pid=33722) INFO 10-10 01:42:16 [launcher.py:42] Route: /load, Methods: GET 2025-10-10T01:42:16.6706258Z (APIServer pid=33722) INFO 10-10 01:42:16 [launcher.py:42] Route: /ping, Methods: POST 2025-10-10T01:42:16.6706762Z (APIServer pid=33722) INFO 10-10 01:42:16 [launcher.py:42] Route: /ping, Methods: GET 2025-10-10T01:42:16.6707276Z (APIServer pid=33722) INFO 10-10 01:42:16 [launcher.py:42] Route: /tokenize, Methods: POST 2025-10-10T01:42:16.6707813Z (APIServer pid=33722) INFO 10-10 01:42:16 [launcher.py:42] Route: /detokenize, Methods: POST 2025-10-10T01:42:16.6708339Z (APIServer pid=33722) INFO 10-10 01:42:16 [launcher.py:42] Route: /v1/models, Methods: GET 2025-10-10T01:42:16.6708854Z (APIServer pid=33722) INFO 10-10 01:42:16 [launcher.py:42] Route: /version, Methods: GET 2025-10-10T01:42:16.6709387Z (APIServer pid=33722) INFO 10-10 01:42:16 [launcher.py:42] Route: /v1/responses, Methods: POST 2025-10-10T01:42:16.6709989Z (APIServer pid=33722) INFO 10-10 01:42:16 [launcher.py:42] Route: /v1/responses/{response_id}, Methods: GET 2025-10-10T01:42:16.6710861Z (APIServer pid=33722) INFO 10-10 01:42:16 [launcher.py:42] Route: /v1/responses/{response_id}/cancel, Methods: POST 2025-10-10T01:42:16.6711486Z (APIServer pid=33722) INFO 10-10 01:42:16 [launcher.py:42] Route: /v1/chat/completions, Methods: POST 2025-10-10T01:42:16.6712054Z (APIServer pid=33722) INFO 10-10 01:42:16 [launcher.py:42] Route: /v1/completions, Methods: POST 2025-10-10T01:42:16.6712607Z (APIServer pid=33722) INFO 10-10 01:42:16 [launcher.py:42] Route: /v1/embeddings, Methods: POST 2025-10-10T01:42:16.6713141Z (APIServer pid=33722) INFO 10-10 01:42:16 [launcher.py:42] Route: /pooling, Methods: POST 2025-10-10T01:42:16.6713663Z (APIServer pid=33722) INFO 10-10 01:42:16 [launcher.py:42] Route: /classify, Methods: POST 2025-10-10T01:42:16.6714326Z (APIServer pid=33722) INFO 10-10 01:42:16 [launcher.py:42] Route: /score, Methods: POST 2025-10-10T01:42:16.6714877Z (APIServer pid=33722) INFO 10-10 01:42:16 [launcher.py:42] Route: /v1/score, Methods: POST 2025-10-10T01:42:16.6715447Z (APIServer pid=33722) INFO 10-10 01:42:16 [launcher.py:42] Route: /v1/audio/transcriptions, Methods: POST 2025-10-10T01:42:16.6716039Z (APIServer pid=33722) INFO 10-10 01:42:16 [launcher.py:42] Route: /v1/audio/translations, Methods: POST 2025-10-10T01:42:16.6716576Z (APIServer pid=33722) INFO 10-10 01:42:16 [launcher.py:42] Route: /rerank, Methods: POST 2025-10-10T01:42:16.6717097Z (APIServer pid=33722) INFO 10-10 01:42:16 [launcher.py:42] Route: /v1/rerank, Methods: POST 2025-10-10T01:42:16.6717621Z (APIServer pid=33722) INFO 10-10 01:42:16 [launcher.py:42] Route: /v2/rerank, Methods: POST 2025-10-10T01:42:16.6718303Z (APIServer pid=33722) INFO 10-10 01:42:16 [launcher.py:42] Route: /scale_elastic_ep, Methods: POST 2025-10-10T01:42:16.6719013Z (APIServer pid=33722) INFO 10-10 01:42:16 [launcher.py:42] Route: /is_scaling_elastic_ep, Methods: POST 2025-10-10T01:42:16.6719582Z (APIServer pid=33722) INFO 10-10 01:42:16 [launcher.py:42] Route: /invocations, Methods: POST 2025-10-10T01:42:16.6720110Z (APIServer pid=33722) INFO 10-10 01:42:16 [launcher.py:42] Route: /metrics, Methods: GET 2025-10-10T01:42:16.6856487Z (APIServer pid=33722) INFO: Started server process [33722] 2025-10-10T01:42:16.6857143Z (APIServer pid=33722) INFO: Waiting for application startup. 2025-10-10T01:42:16.9071232Z (APIServer pid=33722) INFO: Application startup complete. 2025-10-10T01:42:17.4066901Z (APIServer pid=33722) INFO: 127.0.0.1:50126 - "GET /health HTTP/1.1" 200 OK 2025-10-10T01:42:17.6270692Z (APIServer pid=33722) INFO: 127.0.0.1:50130 - "GET /v1/models HTTP/1.1" 200 OK 2025-10-10T01:42:17.7349960Z (APIServer pid=33722) INFO: 127.0.0.1:50130 - "POST /v1/completions HTTP/1.1" 200 OK 2025-10-10T01:42:17.7910647Z (APIServer pid=33722) INFO: 127.0.0.1:50130 - "POST /v1/completions HTTP/1.1" 200 OK 2025-10-10T01:42:17.9175221Z (APIServer pid=33722) INFO: 127.0.0.1:50130 - "POST /v1/completions HTTP/1.1" 200 OK 2025-10-10T01:42:17.9981522Z (APIServer pid=33722) INFO: 127.0.0.1:50130 - "POST /v1/completions HTTP/1.1" 200 OK 2025-10-10T01:42:18.0663130Z (APIServer pid=33722) INFO: 127.0.0.1:50130 - "POST /v1/completions HTTP/1.1" 200 OK 2025-10-10T01:42:18.0738832Z (APIServer pid=33722) INFO: 127.0.0.1:50130 - "POST /v1/completions HTTP/1.1" 200 OK 2025-10-10T01:42:18.1456835Z (APIServer pid=33722) INFO 10-10 01:42:18 [launcher.py:99] Shutting down FastAPI HTTP server. 2025-10-10T01:42:18.1459734Z (Worker_TP0 pid=33963) INFO 10-10 01:42:18 [multiproc_executor.py:558] Parent process exited, terminating worker 2025-10-10T01:42:18.1461097Z (Worker_TP0 pid=33963) INFO 10-10 01:42:18 [multiproc_executor.py:599] WorkerProc shutting down. 2025-10-10T01:42:18.1464189Z (Worker_TP1 pid=33964) INFO 10-10 01:42:18 [multiproc_executor.py:558] Parent process exited, terminating worker 2025-10-10T01:42:20.5630371Z (APIServer pid=33722) INFO: Shutting down 2025-10-10T01:42:20.6635055Z (APIServer pid=33722) INFO: Waiting for application shutdown. 2025-10-10T01:42:20.6636328Z (APIServer pid=33722) INFO: Application shutdown complete. 2025-10-10T01:42:21.0652662Z INFO 10-10 01:42:21 [model.py:551] Resolved architecture: LlamaForCausalLM 2025-10-10T01:42:21.0654134Z WARNING 10-10 01:42:21 [model.py:1768] Casting torch.bfloat16 to torch.float16. 2025-10-10T01:42:21.0655027Z INFO 10-10 01:42:21 [model.py:1545] Using max model len 2048 2025-10-10T01:42:21.0657378Z Launching RemoteOpenAIServer with: vllm serve meta-llama/Llama-3.2-1B-Instruct --dtype float16 --max-model-len 2048 --max-num-seqs 8 --enforce-eager --tokenizer-mode auto --tensor-parallel-size 2 --distributed-executor-backend mp --load-format dummy --port 49977 --seed 0 2025-10-10T01:42:21.5755116Z /opt/conda/envs/py_3.12/lib/python3.12/site-packages/torch/cuda/__init__.py:64: FutureWarning: The pynvml package is deprecated. Please install nvidia-ml-py instead. If you did not install pynvml directly, please report this to the maintainers of the package that installed pynvml for you. 2025-10-10T01:42:21.5757404Z import pynvml # type: ignore[import] 2025-10-10T01:42:23.8097730Z INFO 10-10 01:42:23 [__init__.py:215] Automatically detected platform cuda. 2025-10-10T01:42:27.7343069Z (APIServer pid=34169) INFO 10-10 01:42:27 [api_server.py:1839] vLLM API server version 0.11.0rc2.dev157+g0ad9951c4.d20251010 2025-10-10T01:42:27.7366315Z (APIServer pid=34169) INFO 10-10 01:42:27 [utils.py:233] non-default args: {'model_tag': 'meta-llama/Llama-3.2-1B-Instruct', 'port': 49977, 'model': 'meta-llama/Llama-3.2-1B-Instruct', 'dtype': 'float16', 'seed': 0, 'max_model_len': 2048, 'enforce_eager': True, 'load_format': 'dummy', 'distributed_executor_backend': 'mp', 'tensor_parallel_size': 2, 'max_num_seqs': 8} 2025-10-10T01:42:27.8752956Z (APIServer pid=34169) INFO 10-10 01:42:27 [model.py:551] Resolved architecture: LlamaForCausalLM 2025-10-10T01:42:27.8753647Z (APIServer pid=34169) `torch_dtype` is deprecated! Use `dtype` instead! 2025-10-10T01:42:27.8754361Z (APIServer pid=34169) WARNING 10-10 01:42:27 [model.py:1768] Casting torch.bfloat16 to torch.float16. 2025-10-10T01:42:27.8755058Z (APIServer pid=34169) INFO 10-10 01:42:27 [model.py:1545] Using max model len 2048 2025-10-10T01:42:28.0346085Z (APIServer pid=34169) INFO 10-10 01:42:28 [scheduler.py:222] Chunked prefill is enabled with max_num_batched_tokens=2048. 2025-10-10T01:42:28.0352299Z (APIServer pid=34169) INFO 10-10 01:42:28 [vllm.py:345] Cudagraph is disabled under eager mode 2025-10-10T01:42:28.9921025Z /opt/conda/envs/py_3.12/lib/python3.12/site-packages/torch/cuda/__init__.py:64: FutureWarning: The pynvml package is deprecated. Please install nvidia-ml-py instead. If you did not install pynvml directly, please report this to the maintainers of the package that installed pynvml for you. 2025-10-10T01:42:28.9923358Z import pynvml # type: ignore[import] 2025-10-10T01:42:31.3124644Z INFO 10-10 01:42:31 [__init__.py:215] Automatically detected platform cuda. 2025-10-10T01:42:35.2938857Z (EngineCore_DP0 pid=34291) INFO 10-10 01:42:35 [core.py:648] Waiting for init message from front-end. 2025-10-10T01:42:35.3203755Z (EngineCore_DP0 pid=34291) INFO 10-10 01:42:35 [core.py:78] Initializing a V1 LLM engine (v0.11.0rc2.dev157+g0ad9951c4.d20251010) with config: model='meta-llama/Llama-3.2-1B-Instruct', speculative_config=None, tokenizer='meta-llama/Llama-3.2-1B-Instruct', skip_tokenizer_init=False, tokenizer_mode=auto, revision=None, tokenizer_revision=None, trust_remote_code=False, dtype=torch.float16, max_seq_len=2048, download_dir=None, load_format=dummy, tensor_parallel_size=2, pipeline_parallel_size=1, data_parallel_size=1, disable_custom_all_reduce=False, quantization=None, enforce_eager=True, kv_cache_dtype=auto, device_config=cuda, structured_outputs_config=StructuredOutputsConfig(backend='auto', disable_fallback=False, disable_any_whitespace=False, disable_additional_properties=False, reasoning_parser=''), observability_config=ObservabilityConfig(show_hidden_metrics_for_version=None, otlp_traces_endpoint=None, collect_detailed_traces=None), seed=0, served_model_name=meta-llama/Llama-3.2-1B-Instruct, enable_prefix_caching=True, chunked_prefill_enabled=True, pooler_config=None, compilation_config={'level': 0, 'debug_dump_path': None, 'cache_dir': '', 'backend': '', 'custom_ops': [], 'splitting_ops': None, 'use_inductor': True, 'compile_sizes': [], 'inductor_compile_config': {'enable_auto_functionalized_v2': False}, 'inductor_passes': {}, 'cudagraph_mode': , 'use_cudagraph': False, 'cudagraph_num_of_warmups': 0, 'cudagraph_capture_sizes': [], 'cudagraph_copy_inputs': False, 'full_cuda_graph': False, 'use_inductor_graph_partition': False, 'pass_config': {}, 'max_capture_size': 0, 'local_cache_dir': None} 2025-10-10T01:42:35.3209074Z (EngineCore_DP0 pid=34291) WARNING 10-10 01:42:35 [multiproc_executor.py:720] Reducing Torch parallelism from 24 threads to 1 to avoid unnecessary CPU contention. Set OMP_NUM_THREADS in the external environment to tune this value as needed. 2025-10-10T01:42:35.3217418Z (EngineCore_DP0 pid=34291) INFO 10-10 01:42:35 [shm_broadcast.py:289] vLLM message queue communication handle: Handle(local_reader_ranks=[0, 1], buffer_handle=(2, 16777216, 10, 'psm_9f5c630f'), local_subscribe_addr='ipc:///tmp/066c3573-b502-49fc-875c-3f7a12557b74', remote_subscribe_addr=None, remote_addr_ipv6=False) 2025-10-10T01:42:35.8039181Z /opt/conda/envs/py_3.12/lib/python3.12/site-packages/torch/cuda/__init__.py:64: FutureWarning: The pynvml package is deprecated. Please install nvidia-ml-py instead. If you did not install pynvml directly, please report this to the maintainers of the package that installed pynvml for you. 2025-10-10T01:42:35.8041493Z import pynvml # type: ignore[import] 2025-10-10T01:42:35.8176963Z /opt/conda/envs/py_3.12/lib/python3.12/site-packages/torch/cuda/__init__.py:64: FutureWarning: The pynvml package is deprecated. Please install nvidia-ml-py instead. If you did not install pynvml directly, please report this to the maintainers of the package that installed pynvml for you. 2025-10-10T01:42:35.8179183Z import pynvml # type: ignore[import] 2025-10-10T01:42:38.1276947Z INFO 10-10 01:42:38 [__init__.py:215] Automatically detected platform cuda. 2025-10-10T01:42:38.1609338Z INFO 10-10 01:42:38 [__init__.py:215] Automatically detected platform cuda. 2025-10-10T01:42:42.7344459Z INFO 10-10 01:42:42 [shm_broadcast.py:289] vLLM message queue communication handle: Handle(local_reader_ranks=[0], buffer_handle=(1, 10485760, 10, 'psm_243fb6fe'), local_subscribe_addr='ipc:///tmp/9e3465c6-a976-443f-8611-d9c72c0dbdb9', remote_subscribe_addr=None, remote_addr_ipv6=False) 2025-10-10T01:42:42.7740397Z INFO 10-10 01:42:42 [shm_broadcast.py:289] vLLM message queue communication handle: Handle(local_reader_ranks=[0], buffer_handle=(1, 10485760, 10, 'psm_d4b010d9'), local_subscribe_addr='ipc:///tmp/2c470b6e-94d8-4b2f-a09b-811359a1790f', remote_subscribe_addr=None, remote_addr_ipv6=False) 2025-10-10T01:42:42.9508069Z [Gloo] Rank 0 is connected to 1 peer ranks. Expected number of connected peer ranks is : 1 2025-10-10T01:42:42.9508710Z [Gloo] Rank 1 is connected to 1 peer ranks. Expected number of connected peer ranks is : 1 2025-10-10T01:42:42.9555031Z [Gloo] Rank 1 is connected to 1 peer ranks. Expected number of connected peer ranks is : 1 2025-10-10T01:42:42.9556019Z [Gloo] Rank 0 is connected to 1 peer ranks. Expected number of connected peer ranks is : 1 2025-10-10T01:42:42.9597153Z INFO 10-10 01:42:42 [__init__.py:1384] Found nccl from library libnccl.so.2 2025-10-10T01:42:42.9601602Z INFO 10-10 01:42:42 [__init__.py:1384] Found nccl from library libnccl.so.2 2025-10-10T01:42:42.9602053Z INFO 10-10 01:42:42 [pynccl.py:104] vLLM is using nccl==2.27.5 2025-10-10T01:42:42.9602455Z INFO 10-10 01:42:42 [pynccl.py:104] vLLM is using nccl==2.27.5 2025-10-10T01:42:43.1310012Z WARNING 10-10 01:42:43 [symm_mem.py:58] SymmMemCommunicator: Device capability 8.9 not supported, communicator is not available. 2025-10-10T01:42:43.1310890Z WARNING 10-10 01:42:43 [symm_mem.py:58] SymmMemCommunicator: Device capability 8.9 not supported, communicator is not available. 2025-10-10T01:42:43.1340068Z INFO 10-10 01:42:43 [custom_all_reduce.py:35] Skipping P2P check and trusting the driver's P2P report. 2025-10-10T01:42:43.1340770Z INFO 10-10 01:42:43 [custom_all_reduce.py:35] Skipping P2P check and trusting the driver's P2P report. 2025-10-10T01:42:43.1341794Z WARNING 10-10 01:42:43 [custom_all_reduce.py:154] Custom allreduce is disabled because your platform lacks GPU P2P capability or P2P test failed. To silence this warning, specify disable_custom_all_reduce=True explicitly. 2025-10-10T01:42:43.1343184Z WARNING 10-10 01:42:43 [custom_all_reduce.py:154] Custom allreduce is disabled because your platform lacks GPU P2P capability or P2P test failed. To silence this warning, specify disable_custom_all_reduce=True explicitly. 2025-10-10T01:42:43.2242253Z INFO 10-10 01:42:43 [shm_broadcast.py:289] vLLM message queue communication handle: Handle(local_reader_ranks=[1], buffer_handle=(1, 4194304, 6, 'psm_10feeac5'), local_subscribe_addr='ipc:///tmp/25c15128-793a-4c6f-84f8-3f40ca1acbcb', remote_subscribe_addr=None, remote_addr_ipv6=False) 2025-10-10T01:42:43.2272961Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-10-10T01:42:43.2273674Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-10-10T01:42:43.2281100Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-10-10T01:42:43.2289480Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-10-10T01:42:43.2936233Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-10-10T01:42:43.2949826Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-10-10T01:42:43.2959808Z [Gloo] Rank 0 is connected to 1 peer ranks. Expected number of connected peer ranks is : 1 2025-10-10T01:42:43.2960390Z [Gloo] Rank 1 is connected to 1 peer ranks. Expected number of connected peer ranks is : 1 2025-10-10T01:42:43.2964307Z INFO 10-10 01:42:43 [__init__.py:1384] Found nccl from library libnccl.so.2 2025-10-10T01:42:43.2964800Z INFO 10-10 01:42:43 [__init__.py:1384] Found nccl from library libnccl.so.2 2025-10-10T01:42:43.2965214Z INFO 10-10 01:42:43 [pynccl.py:104] vLLM is using nccl==2.27.5 2025-10-10T01:42:43.2965593Z INFO 10-10 01:42:43 [pynccl.py:104] vLLM is using nccl==2.27.5 2025-10-10T01:42:43.3313507Z INFO 10-10 01:42:43 [parallel_state.py:1208] rank 0 in world size 2 is assigned as DP rank 0, PP rank 0, TP rank 0, EP rank 0 2025-10-10T01:42:43.3314263Z INFO 10-10 01:42:43 [parallel_state.py:1208] rank 1 in world size 2 is assigned as DP rank 0, PP rank 0, TP rank 1, EP rank 1 2025-10-10T01:42:43.5805976Z INFO 10-10 01:42:43 [topk_topp_sampler.py:55] Using FlashInfer for top-p & top-k sampling. 2025-10-10T01:42:43.5821409Z INFO 10-10 01:42:43 [topk_topp_sampler.py:55] Using FlashInfer for top-p & top-k sampling. 2025-10-10T01:42:43.6114500Z (Worker_TP1 pid=34411) INFO 10-10 01:42:43 [gpu_model_runner.py:2707] Starting to load model meta-llama/Llama-3.2-1B-Instruct... 2025-10-10T01:42:43.6145294Z (Worker_TP0 pid=34410) INFO 10-10 01:42:43 [gpu_model_runner.py:2707] Starting to load model meta-llama/Llama-3.2-1B-Instruct... 2025-10-10T01:42:43.8348828Z (Worker_TP0 pid=34410) INFO 10-10 01:42:43 [gpu_model_runner.py:2739] Loading model from scratch... 2025-10-10T01:42:43.8382213Z (Worker_TP1 pid=34411) INFO 10-10 01:42:43 [gpu_model_runner.py:2739] Loading model from scratch... 2025-10-10T01:42:43.9254652Z (Worker_TP0 pid=34410) INFO 10-10 01:42:43 [cuda.py:361] Using Flash Attention backend on V1 engine. 2025-10-10T01:42:43.9284442Z (Worker_TP1 pid=34411) INFO 10-10 01:42:43 [cuda.py:361] Using Flash Attention backend on V1 engine. 2025-10-10T01:42:44.3987194Z (Worker_TP0 pid=34410) INFO 10-10 01:42:44 [gpu_model_runner.py:2758] Model loading took 1.1667 GiB and 0.119300 seconds 2025-10-10T01:42:44.4064367Z (Worker_TP1 pid=34411) INFO 10-10 01:42:44 [gpu_model_runner.py:2758] Model loading took 1.1667 GiB and 0.119551 seconds 2025-10-10T01:42:51.7483413Z (Worker_TP0 pid=34410) INFO 10-10 01:42:51 [gpu_worker.py:298] Available KV cache memory: 18.53 GiB 2025-10-10T01:42:51.7857111Z (Worker_TP1 pid=34411) INFO 10-10 01:42:51 [gpu_worker.py:298] Available KV cache memory: 18.53 GiB 2025-10-10T01:42:52.0264897Z (EngineCore_DP0 pid=34291) INFO 10-10 01:42:52 [kv_cache_utils.py:1087] GPU KV cache size: 1,214,464 tokens 2025-10-10T01:42:52.0265863Z (EngineCore_DP0 pid=34291) INFO 10-10 01:42:52 [kv_cache_utils.py:1091] Maximum concurrency for 2,048 tokens per request: 593.00x 2025-10-10T01:42:52.0266733Z (EngineCore_DP0 pid=34291) INFO 10-10 01:42:52 [kv_cache_utils.py:1087] GPU KV cache size: 1,214,464 tokens 2025-10-10T01:42:52.0267589Z (EngineCore_DP0 pid=34291) INFO 10-10 01:42:52 [kv_cache_utils.py:1091] Maximum concurrency for 2,048 tokens per request: 593.00x 2025-10-10T01:42:53.1773129Z (EngineCore_DP0 pid=34291) INFO 10-10 01:42:53 [core.py:211] init engine (profile, create kv cache, warmup model) took 8.70 seconds 2025-10-10T01:42:53.7147176Z (EngineCore_DP0 pid=34291) INFO 10-10 01:42:53 [vllm.py:345] Cudagraph is disabled under eager mode 2025-10-10T01:42:53.8991487Z (APIServer pid=34169) INFO 10-10 01:42:53 [loggers.py:147] Engine 000: vllm cache_config_info with initialization after num_gpu_blocks is: 75904 2025-10-10T01:42:53.9347040Z (EngineCore_DP0 pid=34291) INFO 10-10 01:42:53 [gc_utils.py:41] GC Debug Config. enabled:False,top_objects:-1 2025-10-10T01:42:53.9566023Z (APIServer pid=34169) INFO 10-10 01:42:53 [api_server.py:1634] Supported_tasks: ['generate'] 2025-10-10T01:42:53.9807262Z (APIServer pid=34169) WARNING 10-10 01:42:53 [model.py:1424] Default sampling parameters have been overridden by the model's Hugging Face generation config recommended from the model creator. If this is not intended, please relaunch vLLM instance with `--generation-config vllm`. 2025-10-10T01:42:53.9808969Z (APIServer pid=34169) INFO 10-10 01:42:53 [serving_responses.py:140] Using default chat sampling params from model: {'temperature': 0.6, 'top_p': 0.9} 2025-10-10T01:42:54.0031923Z (APIServer pid=34169) INFO 10-10 01:42:54 [serving_chat.py:139] Using default chat sampling params from model: {'temperature': 0.6, 'top_p': 0.9} 2025-10-10T01:42:54.0276797Z (APIServer pid=34169) INFO 10-10 01:42:54 [serving_completion.py:76] Using default completion sampling params from model: {'temperature': 0.6, 'top_p': 0.9} 2025-10-10T01:42:54.0277829Z (APIServer pid=34169) INFO 10-10 01:42:54 [api_server.py:1912] Starting vLLM API server 0 on http://0.0.0.0:49977 2025-10-10T01:42:54.0278703Z (APIServer pid=34169) INFO 10-10 01:42:54 [launcher.py:34] Available routes are: 2025-10-10T01:42:54.0279390Z (APIServer pid=34169) INFO 10-10 01:42:54 [launcher.py:42] Route: /openapi.json, Methods: HEAD, GET 2025-10-10T01:42:54.0280456Z (APIServer pid=34169) INFO 10-10 01:42:54 [launcher.py:42] Route: /docs, Methods: HEAD, GET 2025-10-10T01:42:54.0281178Z (APIServer pid=34169) INFO 10-10 01:42:54 [launcher.py:42] Route: /docs/oauth2-redirect, Methods: HEAD, GET 2025-10-10T01:42:54.0281879Z (APIServer pid=34169) INFO 10-10 01:42:54 [launcher.py:42] Route: /redoc, Methods: HEAD, GET 2025-10-10T01:42:54.0282860Z (APIServer pid=34169) INFO 10-10 01:42:54 [launcher.py:42] Route: /health, Methods: GET 2025-10-10T01:42:54.0283713Z (APIServer pid=34169) INFO 10-10 01:42:54 [launcher.py:42] Route: /load, Methods: GET 2025-10-10T01:42:54.0284400Z (APIServer pid=34169) INFO 10-10 01:42:54 [launcher.py:42] Route: /ping, Methods: POST 2025-10-10T01:42:54.0285037Z (APIServer pid=34169) INFO 10-10 01:42:54 [launcher.py:42] Route: /ping, Methods: GET 2025-10-10T01:42:54.0288073Z (APIServer pid=34169) INFO 10-10 01:42:54 [launcher.py:42] Route: /tokenize, Methods: POST 2025-10-10T01:42:54.0288815Z (APIServer pid=34169) INFO 10-10 01:42:54 [launcher.py:42] Route: /detokenize, Methods: POST 2025-10-10T01:42:54.0289345Z (APIServer pid=34169) INFO 10-10 01:42:54 [launcher.py:42] Route: /v1/models, Methods: GET 2025-10-10T01:42:54.0289860Z (APIServer pid=34169) INFO 10-10 01:42:54 [launcher.py:42] Route: /version, Methods: GET 2025-10-10T01:42:54.0290413Z (APIServer pid=34169) INFO 10-10 01:42:54 [launcher.py:42] Route: /v1/responses, Methods: POST 2025-10-10T01:42:54.0291018Z (APIServer pid=34169) INFO 10-10 01:42:54 [launcher.py:42] Route: /v1/responses/{response_id}, Methods: GET 2025-10-10T01:42:54.0291649Z (APIServer pid=34169) INFO 10-10 01:42:54 [launcher.py:42] Route: /v1/responses/{response_id}/cancel, Methods: POST 2025-10-10T01:42:54.0292267Z (APIServer pid=34169) INFO 10-10 01:42:54 [launcher.py:42] Route: /v1/chat/completions, Methods: POST 2025-10-10T01:42:54.0292835Z (APIServer pid=34169) INFO 10-10 01:42:54 [launcher.py:42] Route: /v1/completions, Methods: POST 2025-10-10T01:42:54.0293387Z (APIServer pid=34169) INFO 10-10 01:42:54 [launcher.py:42] Route: /v1/embeddings, Methods: POST 2025-10-10T01:42:54.0293928Z (APIServer pid=34169) INFO 10-10 01:42:54 [launcher.py:42] Route: /pooling, Methods: POST 2025-10-10T01:42:54.0294463Z (APIServer pid=34169) INFO 10-10 01:42:54 [launcher.py:42] Route: /classify, Methods: POST 2025-10-10T01:42:54.0294979Z (APIServer pid=34169) INFO 10-10 01:42:54 [launcher.py:42] Route: /score, Methods: POST 2025-10-10T01:42:54.0295491Z (APIServer pid=34169) INFO 10-10 01:42:54 [launcher.py:42] Route: /v1/score, Methods: POST 2025-10-10T01:42:54.0296058Z (APIServer pid=34169) INFO 10-10 01:42:54 [launcher.py:42] Route: /v1/audio/transcriptions, Methods: POST 2025-10-10T01:42:54.0296657Z (APIServer pid=34169) INFO 10-10 01:42:54 [launcher.py:42] Route: /v1/audio/translations, Methods: POST 2025-10-10T01:42:54.0297214Z (APIServer pid=34169) INFO 10-10 01:42:54 [launcher.py:42] Route: /rerank, Methods: POST 2025-10-10T01:42:54.0297749Z (APIServer pid=34169) INFO 10-10 01:42:54 [launcher.py:42] Route: /v1/rerank, Methods: POST 2025-10-10T01:42:54.0298276Z (APIServer pid=34169) INFO 10-10 01:42:54 [launcher.py:42] Route: /v2/rerank, Methods: POST 2025-10-10T01:42:54.0298834Z (APIServer pid=34169) INFO 10-10 01:42:54 [launcher.py:42] Route: /scale_elastic_ep, Methods: POST 2025-10-10T01:42:54.0299415Z (APIServer pid=34169) INFO 10-10 01:42:54 [launcher.py:42] Route: /is_scaling_elastic_ep, Methods: POST 2025-10-10T01:42:54.0299975Z (APIServer pid=34169) INFO 10-10 01:42:54 [launcher.py:42] Route: /invocations, Methods: POST 2025-10-10T01:42:54.0300505Z (APIServer pid=34169) INFO 10-10 01:42:54 [launcher.py:42] Route: /metrics, Methods: GET 2025-10-10T01:42:54.0433276Z (APIServer pid=34169) INFO: Started server process [34169] 2025-10-10T01:42:54.0433924Z (APIServer pid=34169) INFO: Waiting for application startup. 2025-10-10T01:42:54.2721467Z (APIServer pid=34169) INFO: Application startup complete. 2025-10-10T01:42:54.6495613Z (APIServer pid=34169) INFO: 127.0.0.1:45788 - "GET /health HTTP/1.1" 200 OK 2025-10-10T01:42:54.6586384Z (APIServer pid=34169) INFO: 127.0.0.1:45790 - "GET /v1/models HTTP/1.1" 200 OK 2025-10-10T01:42:55.9137159Z (APIServer pid=34169) INFO: 127.0.0.1:45790 - "POST /v1/completions HTTP/1.1" 200 OK 2025-10-10T01:42:55.9704359Z (APIServer pid=34169) INFO: 127.0.0.1:45790 - "POST /v1/completions HTTP/1.1" 200 OK 2025-10-10T01:42:56.0580715Z (APIServer pid=34169) INFO: 127.0.0.1:45790 - "POST /v1/completions HTTP/1.1" 200 OK 2025-10-10T01:42:56.1422001Z (APIServer pid=34169) INFO: 127.0.0.1:45790 - "POST /v1/completions HTTP/1.1" 200 OK 2025-10-10T01:42:56.2128875Z (APIServer pid=34169) INFO: 127.0.0.1:45790 - "POST /v1/completions HTTP/1.1" 200 OK 2025-10-10T01:42:56.2206509Z (APIServer pid=34169) INFO: 127.0.0.1:45790 - "POST /v1/completions HTTP/1.1" 200 OK 2025-10-10T01:42:56.2960784Z (APIServer pid=34169) INFO 10-10 01:42:56 [launcher.py:99] Shutting down FastAPI HTTP server. 2025-10-10T01:42:56.2963482Z (Worker_TP0 pid=34410) INFO 10-10 01:42:56 [multiproc_executor.py:558] Parent process exited, terminating worker 2025-10-10T01:42:56.2964299Z (Worker_TP0 pid=34410) INFO 10-10 01:42:56 [multiproc_executor.py:599] WorkerProc shutting down. 2025-10-10T01:42:56.2970814Z (Worker_TP1 pid=34411) INFO 10-10 01:42:56 [multiproc_executor.py:558] Parent process exited, terminating worker 2025-10-10T01:42:56.2971626Z (Worker_TP1 pid=34411) INFO 10-10 01:42:56 [multiproc_executor.py:599] WorkerProc shutting down. 2025-10-10T01:42:58.7167362Z (APIServer pid=34169) INFO: Shutting down 2025-10-10T01:42:58.8171289Z (APIServer pid=34169) INFO: Waiting for application shutdown. 2025-10-10T01:42:58.8173889Z (APIServer pid=34169) INFO: Application shutdown complete. 2025-10-10T01:42:59.4410850Z PASSED 2025-10-10T01:42:59.4525673Z distributed/test_sequence_parallel.py::test_tp_sp_generation[meta-llama/Llama-3.2-1B-Instruct-parallel_setup9-ray-1-auto-test_options9] Fork a new process to run a test 35331 2025-10-10T01:42:59.4543147Z Fork a new process to run a test 0 2025-10-10T01:43:00.0612801Z INFO 10-10 01:43:00 [model.py:551] Resolved architecture: LlamaForCausalLM 2025-10-10T01:43:00.0613751Z `torch_dtype` is deprecated! Use `dtype` instead! 2025-10-10T01:43:00.0614583Z WARNING 10-10 01:43:00 [model.py:1768] Casting torch.bfloat16 to torch.float16. 2025-10-10T01:43:00.0615448Z INFO 10-10 01:43:00 [model.py:1545] Using max model len 2048 2025-10-10T01:43:00.0619374Z Launching RemoteOpenAIServer with: vllm serve meta-llama/Llama-3.2-1B-Instruct --dtype float16 --max-model-len 2048 --max-num-seqs 8 --enforce-eager --tokenizer-mode auto --tensor-parallel-size 2 --pipeline-parallel-size 1 --distributed-executor-backend ray --compilation_config {"level": 3, "custom_ops": ["+rms_norm"], "compile_sizes": [4, 8], "pass_config": {"enable_sequence_parallelism": true, "enable_fusion": false, "enable_noop": true}} --load-format dummy --port 36523 --seed 0 2025-10-10T01:43:00.5722981Z /opt/conda/envs/py_3.12/lib/python3.12/site-packages/torch/cuda/__init__.py:64: FutureWarning: The pynvml package is deprecated. Please install nvidia-ml-py instead. If you did not install pynvml directly, please report this to the maintainers of the package that installed pynvml for you. 2025-10-10T01:43:00.5726704Z import pynvml # type: ignore[import] 2025-10-10T01:43:02.7876051Z INFO 10-10 01:43:02 [__init__.py:215] Automatically detected platform cuda. 2025-10-10T01:43:06.7193859Z (APIServer pid=35380) INFO 10-10 01:43:06 [api_server.py:1839] vLLM API server version 0.11.0rc2.dev157+g0ad9951c4.d20251010 2025-10-10T01:43:06.7218551Z (APIServer pid=35380) INFO 10-10 01:43:06 [utils.py:233] non-default args: {'model_tag': 'meta-llama/Llama-3.2-1B-Instruct', 'port': 36523, 'model': 'meta-llama/Llama-3.2-1B-Instruct', 'dtype': 'float16', 'seed': 0, 'max_model_len': 2048, 'enforce_eager': True, 'load_format': 'dummy', 'distributed_executor_backend': 'ray', 'tensor_parallel_size': 2, 'max_num_seqs': 8, 'compilation_config': {'level': 3, 'debug_dump_path': None, 'cache_dir': '', 'backend': '', 'custom_ops': ['+rms_norm'], 'splitting_ops': None, 'use_inductor': True, 'compile_sizes': [4, 8], 'inductor_compile_config': {'enable_auto_functionalized_v2': False}, 'inductor_passes': {}, 'cudagraph_mode': None, 'use_cudagraph': True, 'cudagraph_num_of_warmups': 0, 'cudagraph_capture_sizes': None, 'cudagraph_copy_inputs': False, 'full_cuda_graph': False, 'use_inductor_graph_partition': False, 'pass_config': {'enable_noop': True, 'enable_sequence_parallelism': True}, 'max_capture_size': None, 'local_cache_dir': None}} 2025-10-10T01:43:06.8761499Z (APIServer pid=35380) INFO 10-10 01:43:06 [model.py:551] Resolved architecture: LlamaForCausalLM 2025-10-10T01:43:06.8762225Z (APIServer pid=35380) `torch_dtype` is deprecated! Use `dtype` instead! 2025-10-10T01:43:06.8762917Z (APIServer pid=35380) WARNING 10-10 01:43:06 [model.py:1768] Casting torch.bfloat16 to torch.float16. 2025-10-10T01:43:06.8767406Z (APIServer pid=35380) INFO 10-10 01:43:06 [model.py:1545] Using max model len 2048 2025-10-10T01:43:07.0371949Z (APIServer pid=35380) INFO 10-10 01:43:07 [scheduler.py:222] Chunked prefill is enabled with max_num_batched_tokens=2048. 2025-10-10T01:43:07.0375951Z (APIServer pid=35380) INFO 10-10 01:43:07 [vllm.py:345] Cudagraph is disabled under eager mode 2025-10-10T01:43:08.0618653Z /opt/conda/envs/py_3.12/lib/python3.12/site-packages/torch/cuda/__init__.py:64: FutureWarning: The pynvml package is deprecated. Please install nvidia-ml-py instead. If you did not install pynvml directly, please report this to the maintainers of the package that installed pynvml for you. 2025-10-10T01:43:08.0621102Z import pynvml # type: ignore[import] 2025-10-10T01:43:10.3539087Z INFO 10-10 01:43:10 [__init__.py:215] Automatically detected platform cuda. 2025-10-10T01:43:14.4443818Z (EngineCore_DP0 pid=35502) INFO 10-10 01:43:14 [core.py:648] Waiting for init message from front-end. 2025-10-10T01:43:14.4702321Z (EngineCore_DP0 pid=35502) INFO 10-10 01:43:14 [core.py:78] Initializing a V1 LLM engine (v0.11.0rc2.dev157+g0ad9951c4.d20251010) with config: model='meta-llama/Llama-3.2-1B-Instruct', speculative_config=None, tokenizer='meta-llama/Llama-3.2-1B-Instruct', skip_tokenizer_init=False, tokenizer_mode=auto, revision=None, tokenizer_revision=None, trust_remote_code=False, dtype=torch.float16, max_seq_len=2048, download_dir=None, load_format=dummy, tensor_parallel_size=2, pipeline_parallel_size=1, data_parallel_size=1, disable_custom_all_reduce=False, quantization=None, enforce_eager=True, kv_cache_dtype=auto, device_config=cuda, structured_outputs_config=StructuredOutputsConfig(backend='auto', disable_fallback=False, disable_any_whitespace=False, disable_additional_properties=False, reasoning_parser=''), observability_config=ObservabilityConfig(show_hidden_metrics_for_version=None, otlp_traces_endpoint=None, collect_detailed_traces=None), seed=0, served_model_name=meta-llama/Llama-3.2-1B-Instruct, enable_prefix_caching=True, chunked_prefill_enabled=True, pooler_config=None, compilation_config={'level': 3, 'debug_dump_path': None, 'cache_dir': '', 'backend': '', 'custom_ops': ['+rms_norm', '+rms_norm'], 'splitting_ops': ['vllm.unified_attention', 'vllm.unified_attention_with_output', 'vllm.mamba_mixer2', 'vllm.mamba_mixer', 'vllm.short_conv', 'vllm.linear_attention', 'vllm.plamo2_mamba_mixer', 'vllm.gdn_attention', 'vllm.sparse_attn_indexer'], 'use_inductor': True, 'compile_sizes': [8, 4], 'inductor_compile_config': {'enable_auto_functionalized_v2': False}, 'inductor_passes': {}, 'cudagraph_mode': , 'use_cudagraph': False, 'cudagraph_num_of_warmups': 0, 'cudagraph_capture_sizes': [], 'cudagraph_copy_inputs': False, 'full_cuda_graph': False, 'use_inductor_graph_partition': False, 'pass_config': {'enable_noop': True, 'enable_sequence_parallelism': True}, 'max_capture_size': 0, 'local_cache_dir': None} 2025-10-10T01:43:16.3758237Z (EngineCore_DP0 pid=35502) 2025-10-10 01:43:16,375 INFO worker.py:1918 -- Started a local Ray instance. View the dashboard at http://127.0.0.1:8265  2025-10-10T01:43:17.2367097Z (EngineCore_DP0 pid=35502) INFO 10-10 01:43:17 [ray_utils.py:345] No current placement group found. Creating a new placement group. 2025-10-10T01:43:17.4113666Z (EngineCore_DP0 pid=35502) INFO 10-10 01:43:17 [ray_distributed_executor.py:171] use_ray_spmd_worker: True 2025-10-10T01:43:18.2544336Z (EngineCore_DP0 pid=35502) (pid=36391) /opt/conda/envs/py_3.12/lib/python3.12/site-packages/torch/cuda/__init__.py:64: FutureWarning: The pynvml package is deprecated. Please install nvidia-ml-py instead. If you did not install pynvml directly, please report this to the maintainers of the package that installed pynvml for you. 2025-10-10T01:43:18.2545949Z (EngineCore_DP0 pid=35502) (pid=36391) import pynvml # type: ignore[import] 2025-10-10T01:43:20.3979494Z (EngineCore_DP0 pid=35502) (pid=36385) INFO 10-10 01:43:20 [__init__.py:215] Automatically detected platform cuda. 2025-10-10T01:43:21.8222262Z (EngineCore_DP0 pid=35502) INFO 10-10 01:43:21 [ray_env.py:63] RAY_NON_CARRY_OVER_ENV_VARS from config: set() 2025-10-10T01:43:21.8223925Z (EngineCore_DP0 pid=35502) INFO 10-10 01:43:21 [ray_env.py:65] Copying the following environment variables to workers: ['VLLM_USE_RAY_COMPILED_DAG', 'MAX_JOBS', 'VLLM_WORKER_MULTIPROC_METHOD', 'HUGGING_FACE_HUB_TOKEN', 'VLLM_USE_V1', 'HF_TOKEN', 'VLLM_USE_RAY_SPMD_WORKER', 'LD_LIBRARY_PATH'] 2025-10-10T01:43:21.8225809Z (EngineCore_DP0 pid=35502) INFO 10-10 01:43:21 [ray_env.py:68] If certain env vars should NOT be copied, add them to /var/lib/jenkins/.config/vllm/ray_non_carry_over_env_vars.json file 2025-10-10T01:43:25.2964712Z (EngineCore_DP0 pid=35502) (RayWorkerWrapper pid=36391) [Gloo] Rank 0 is connected to 1 peer ranks. Expected number of connected peer ranks is : 1 2025-10-10T01:43:25.2965984Z (EngineCore_DP0 pid=35502) (RayWorkerWrapper pid=36391) [Gloo] Rank 0 is connected to 1 peer ranks. Expected number of connected peer ranks is : 1 2025-10-10T01:43:25.2967089Z (EngineCore_DP0 pid=35502) (RayWorkerWrapper pid=36391) INFO 10-10 01:43:25 [__init__.py:1384] Found nccl from library libnccl.so.2 2025-10-10T01:43:25.2968071Z (EngineCore_DP0 pid=35502) (RayWorkerWrapper pid=36391) INFO 10-10 01:43:25 [pynccl.py:104] vLLM is using nccl==2.27.5 2025-10-10T01:43:25.5040007Z (EngineCore_DP0 pid=35502) (RayWorkerWrapper pid=36391) WARNING 10-10 01:43:25 [symm_mem.py:58] SymmMemCommunicator: Device capability 8.9 not supported, communicator is not available. 2025-10-10T01:43:25.5041381Z (EngineCore_DP0 pid=35502) (RayWorkerWrapper pid=36391) INFO 10-10 01:43:25 [custom_all_reduce.py:35] Skipping P2P check and trusting the driver's P2P report. 2025-10-10T01:43:25.5043054Z (EngineCore_DP0 pid=35502) (RayWorkerWrapper pid=36391) WARNING 10-10 01:43:25 [custom_all_reduce.py:154] Custom allreduce is disabled because your platform lacks GPU P2P capability or P2P test failed. To silence this warning, specify disable_custom_all_reduce=True explicitly. 2025-10-10T01:43:25.5045768Z (EngineCore_DP0 pid=35502) (RayWorkerWrapper pid=36391) INFO 10-10 01:43:25 [shm_broadcast.py:289] vLLM message queue communication handle: Handle(local_reader_ranks=[1], buffer_handle=(1, 4194304, 6, 'psm_9a6c0243'), local_subscribe_addr='ipc:///tmp/92e3f2b7-1f00-4798-8f69-2f5b0fef3b58', remote_subscribe_addr=None, remote_addr_ipv6=False) 2025-10-10T01:43:25.5047358Z (EngineCore_DP0 pid=35502) (pid=36391) INFO 10-10 01:43:20 [__init__.py:215] Automatically detected platform cuda. 2025-10-10T01:43:25.6143566Z (EngineCore_DP0 pid=35502) (RayWorkerWrapper pid=36391) INFO 10-10 01:43:25 [parallel_state.py:1208] rank 0 in world size 2 is assigned as DP rank 0, PP rank 0, TP rank 0, EP rank 0 2025-10-10T01:43:25.8222416Z (EngineCore_DP0 pid=35502) (RayWorkerWrapper pid=36391) INFO 10-10 01:43:25 [topk_topp_sampler.py:55] Using FlashInfer for top-p & top-k sampling. 2025-10-10T01:43:25.9279166Z (EngineCore_DP0 pid=35502) (RayWorkerWrapper pid=36391) INFO 10-10 01:43:25 [gpu_model_runner.py:2707] Starting to load model meta-llama/Llama-3.2-1B-Instruct... 2025-10-10T01:43:26.1346618Z (EngineCore_DP0 pid=35502) (RayWorkerWrapper pid=36391) INFO 10-10 01:43:26 [gpu_model_runner.py:2739] Loading model from scratch... 2025-10-10T01:43:26.2392280Z (EngineCore_DP0 pid=35502) (RayWorkerWrapper pid=36391) INFO 10-10 01:43:26 [cuda.py:361] Using Flash Attention backend on V1 engine. 2025-10-10T01:43:26.8536080Z (EngineCore_DP0 pid=35502) (RayWorkerWrapper pid=36391) INFO 10-10 01:43:26 [gpu_model_runner.py:2758] Model loading took 1.1667 GiB and 0.169474 seconds 2025-10-10T01:43:29.9097829Z (EngineCore_DP0 pid=35502) (RayWorkerWrapper pid=36391) INFO 10-10 01:43:29 [backends.py:548] Using cache directory: /var/lib/jenkins/.cache/vllm/torch_compile_cache/31e0bc77eb/rank_0_0/backbone for vLLM's torch.compile 2025-10-10T01:43:29.9100023Z (EngineCore_DP0 pid=35502) (RayWorkerWrapper pid=36391) INFO 10-10 01:43:29 [backends.py:559] Dynamo bytecode transform time: 2.87 s 2025-10-10T01:43:31.0345404Z (EngineCore_DP0 pid=35502) (RayWorkerWrapper pid=36385) INFO 10-10 01:43:30 [backends.py:164] Directly load the compiled graph(s) for dynamic shape from the cache, took 0.882 s 2025-10-10T01:43:31.0348510Z (EngineCore_DP0 pid=35502) (RayWorkerWrapper pid=36385) [Gloo] Rank 1 is connected to 1 peer ranks. Expected number of connected peer ranks is : 1 [repeated 10x across cluster] (Ray deduplicates logs by default. Set RAY_DEDUP_LOGS=0 to disable log deduplication, or see https://docs.ray.io/en/master/ray-observability/user-guides/configure-logging.html#log-deduplication for more options.) 2025-10-10T01:43:31.0350710Z (EngineCore_DP0 pid=35502) (RayWorkerWrapper pid=36385) INFO 10-10 01:43:25 [__init__.py:1384] Found nccl from library libnccl.so.2 [repeated 3x across cluster] 2025-10-10T01:43:31.0351900Z (EngineCore_DP0 pid=35502) (RayWorkerWrapper pid=36385) INFO 10-10 01:43:25 [pynccl.py:104] vLLM is using nccl==2.27.5 [repeated 3x across cluster] 2025-10-10T01:43:31.0352926Z (EngineCore_DP0 pid=35502) (RayWorkerWrapper pid=36385) WARNING 10-10 01:43:25 [symm_mem.py:58] SymmMemCommunicator: Device capability 8.9 not supported, communicator is not available. 2025-10-10T01:43:31.0353980Z (EngineCore_DP0 pid=35502) (RayWorkerWrapper pid=36385) INFO 10-10 01:43:25 [custom_all_reduce.py:35] Skipping P2P check and trusting the driver's P2P report. 2025-10-10T01:43:31.0355301Z (EngineCore_DP0 pid=35502) (RayWorkerWrapper pid=36385) WARNING 10-10 01:43:25 [custom_all_reduce.py:154] Custom allreduce is disabled because your platform lacks GPU P2P capability or P2P test failed. To silence this warning, specify disable_custom_all_reduce=True explicitly. 2025-10-10T01:43:31.0356935Z (EngineCore_DP0 pid=35502) (RayWorkerWrapper pid=36385) INFO 10-10 01:43:25 [parallel_state.py:1208] rank 1 in world size 2 is assigned as DP rank 0, PP rank 0, TP rank 1, EP rank 1 2025-10-10T01:43:31.0357892Z (EngineCore_DP0 pid=35502) (RayWorkerWrapper pid=36385) INFO 10-10 01:43:25 [topk_topp_sampler.py:55] Using FlashInfer for top-p & top-k sampling. 2025-10-10T01:43:31.0358924Z (EngineCore_DP0 pid=35502) (RayWorkerWrapper pid=36385) INFO 10-10 01:43:25 [gpu_model_runner.py:2707] Starting to load model meta-llama/Llama-3.2-1B-Instruct... 2025-10-10T01:43:31.1383637Z (EngineCore_DP0 pid=35502) (RayWorkerWrapper pid=36385) INFO 10-10 01:43:26 [gpu_model_runner.py:2739] Loading model from scratch... 2025-10-10T01:43:32.3645775Z (EngineCore_DP0 pid=35502) (RayWorkerWrapper pid=36391) INFO 10-10 01:43:32 [gpu_worker.py:298] Available KV cache memory: 18.54 GiB 2025-10-10T01:43:32.3647100Z (EngineCore_DP0 pid=35502) (RayWorkerWrapper pid=36385) INFO 10-10 01:43:26 [cuda.py:361] Using Flash Attention backend on V1 engine. 2025-10-10T01:43:32.3648221Z (EngineCore_DP0 pid=35502) (RayWorkerWrapper pid=36385) INFO 10-10 01:43:26 [gpu_model_runner.py:2758] Model loading took 1.1667 GiB and 0.167603 seconds 2025-10-10T01:43:32.6401727Z (EngineCore_DP0 pid=35502) INFO 10-10 01:43:32 [kv_cache_utils.py:1087] GPU KV cache size: 1,215,360 tokens 2025-10-10T01:43:32.6403018Z (EngineCore_DP0 pid=35502) INFO 10-10 01:43:32 [kv_cache_utils.py:1091] Maximum concurrency for 2,048 tokens per request: 593.44x 2025-10-10T01:43:32.6403746Z (EngineCore_DP0 pid=35502) INFO 10-10 01:43:32 [kv_cache_utils.py:1087] GPU KV cache size: 1,215,360 tokens 2025-10-10T01:43:32.6404455Z (EngineCore_DP0 pid=35502) INFO 10-10 01:43:32 [kv_cache_utils.py:1091] Maximum concurrency for 2,048 tokens per request: 593.44x 2025-10-10T01:43:32.6726007Z (EngineCore_DP0 pid=35502) (RayWorkerWrapper pid=36391) INFO 10-10 01:43:32 [gpu_worker.py:332] Compile and warming up model for size 8 2025-10-10T01:43:33.7944756Z (EngineCore_DP0 pid=35502) (RayWorkerWrapper pid=36391) INFO 10-10 01:43:33 [backends.py:168] Directly load the compiled graph(s) for shape 8 from the cache, took 1.055 s 2025-10-10T01:43:34.3867792Z (EngineCore_DP0 pid=35502) INFO 10-10 01:43:34 [core.py:211] init engine (profile, create kv cache, warmup model) took 7.61 seconds 2025-10-10T01:43:34.4077184Z (EngineCore_DP0 pid=35502) (RayWorkerWrapper pid=36391) INFO 10-10 01:43:34 [monitor.py:32] torch.compile takes 2.87 s in total 2025-10-10T01:43:35.1517542Z (EngineCore_DP0 pid=35502) INFO 10-10 01:43:35 [vllm.py:345] Cudagraph is disabled under eager mode 2025-10-10T01:43:35.3225039Z (APIServer pid=35380) INFO 10-10 01:43:35 [loggers.py:147] Engine 000: vllm cache_config_info with initialization after num_gpu_blocks is: 75960 2025-10-10T01:43:35.4039915Z (EngineCore_DP0 pid=35502) INFO 10-10 01:43:35 [gc_utils.py:41] GC Debug Config. enabled:False,top_objects:-1 2025-10-10T01:43:35.4287542Z (APIServer pid=35380) INFO 10-10 01:43:35 [api_server.py:1634] Supported_tasks: ['generate'] 2025-10-10T01:43:35.4535261Z (APIServer pid=35380) WARNING 10-10 01:43:35 [model.py:1424] Default sampling parameters have been overridden by the model's Hugging Face generation config recommended from the model creator. If this is not intended, please relaunch vLLM instance with `--generation-config vllm`. 2025-10-10T01:43:35.4537051Z (APIServer pid=35380) INFO 10-10 01:43:35 [serving_responses.py:140] Using default chat sampling params from model: {'temperature': 0.6, 'top_p': 0.9} 2025-10-10T01:43:35.4881637Z (APIServer pid=35380) INFO 10-10 01:43:35 [serving_chat.py:139] Using default chat sampling params from model: {'temperature': 0.6, 'top_p': 0.9} 2025-10-10T01:43:35.5180993Z (APIServer pid=35380) INFO 10-10 01:43:35 [serving_completion.py:76] Using default completion sampling params from model: {'temperature': 0.6, 'top_p': 0.9} 2025-10-10T01:43:35.5184420Z (APIServer pid=35380) INFO 10-10 01:43:35 [api_server.py:1912] Starting vLLM API server 0 on http://0.0.0.0:36523 2025-10-10T01:43:35.5185062Z (APIServer pid=35380) INFO 10-10 01:43:35 [launcher.py:34] Available routes are: 2025-10-10T01:43:35.5185623Z (APIServer pid=35380) INFO 10-10 01:43:35 [launcher.py:42] Route: /openapi.json, Methods: HEAD, GET 2025-10-10T01:43:35.5186193Z (APIServer pid=35380) INFO 10-10 01:43:35 [launcher.py:42] Route: /docs, Methods: HEAD, GET 2025-10-10T01:43:35.5187035Z (APIServer pid=35380) INFO 10-10 01:43:35 [launcher.py:42] Route: /docs/oauth2-redirect, Methods: HEAD, GET 2025-10-10T01:43:35.5187809Z (APIServer pid=35380) INFO 10-10 01:43:35 [launcher.py:42] Route: /redoc, Methods: HEAD, GET 2025-10-10T01:43:35.5188348Z (APIServer pid=35380) INFO 10-10 01:43:35 [launcher.py:42] Route: /health, Methods: GET 2025-10-10T01:43:35.5188862Z (APIServer pid=35380) INFO 10-10 01:43:35 [launcher.py:42] Route: /load, Methods: GET 2025-10-10T01:43:35.5189391Z (APIServer pid=35380) INFO 10-10 01:43:35 [launcher.py:42] Route: /ping, Methods: POST 2025-10-10T01:43:35.5189928Z (APIServer pid=35380) INFO 10-10 01:43:35 [launcher.py:42] Route: /ping, Methods: GET 2025-10-10T01:43:35.5190457Z (APIServer pid=35380) INFO 10-10 01:43:35 [launcher.py:42] Route: /tokenize, Methods: POST 2025-10-10T01:43:35.5191002Z (APIServer pid=35380) INFO 10-10 01:43:35 [launcher.py:42] Route: /detokenize, Methods: POST 2025-10-10T01:43:35.5191542Z (APIServer pid=35380) INFO 10-10 01:43:35 [launcher.py:42] Route: /v1/models, Methods: GET 2025-10-10T01:43:35.5192066Z (APIServer pid=35380) INFO 10-10 01:43:35 [launcher.py:42] Route: /version, Methods: GET 2025-10-10T01:43:35.5192607Z (APIServer pid=35380) INFO 10-10 01:43:35 [launcher.py:42] Route: /v1/responses, Methods: POST 2025-10-10T01:43:35.5193203Z (APIServer pid=35380) INFO 10-10 01:43:35 [launcher.py:42] Route: /v1/responses/{response_id}, Methods: GET 2025-10-10T01:43:35.5193856Z (APIServer pid=35380) INFO 10-10 01:43:35 [launcher.py:42] Route: /v1/responses/{response_id}/cancel, Methods: POST 2025-10-10T01:43:35.5194487Z (APIServer pid=35380) INFO 10-10 01:43:35 [launcher.py:42] Route: /v1/chat/completions, Methods: POST 2025-10-10T01:43:35.5195066Z (APIServer pid=35380) INFO 10-10 01:43:35 [launcher.py:42] Route: /v1/completions, Methods: POST 2025-10-10T01:43:35.5195628Z (APIServer pid=35380) INFO 10-10 01:43:35 [launcher.py:42] Route: /v1/embeddings, Methods: POST 2025-10-10T01:43:35.5196175Z (APIServer pid=35380) INFO 10-10 01:43:35 [launcher.py:42] Route: /pooling, Methods: POST 2025-10-10T01:43:35.5196703Z (APIServer pid=35380) INFO 10-10 01:43:35 [launcher.py:42] Route: /classify, Methods: POST 2025-10-10T01:43:35.5197228Z (APIServer pid=35380) INFO 10-10 01:43:35 [launcher.py:42] Route: /score, Methods: POST 2025-10-10T01:43:35.5197752Z (APIServer pid=35380) INFO 10-10 01:43:35 [launcher.py:42] Route: /v1/score, Methods: POST 2025-10-10T01:43:35.5198325Z (APIServer pid=35380) INFO 10-10 01:43:35 [launcher.py:42] Route: /v1/audio/transcriptions, Methods: POST 2025-10-10T01:43:35.5199020Z (APIServer pid=35380) INFO 10-10 01:43:35 [launcher.py:42] Route: /v1/audio/translations, Methods: POST 2025-10-10T01:43:35.5199583Z (APIServer pid=35380) INFO 10-10 01:43:35 [launcher.py:42] Route: /rerank, Methods: POST 2025-10-10T01:43:35.5200267Z (APIServer pid=35380) INFO 10-10 01:43:35 [launcher.py:42] Route: /v1/rerank, Methods: POST 2025-10-10T01:43:35.5200802Z (APIServer pid=35380) INFO 10-10 01:43:35 [launcher.py:42] Route: /v2/rerank, Methods: POST 2025-10-10T01:43:35.5201360Z (APIServer pid=35380) INFO 10-10 01:43:35 [launcher.py:42] Route: /scale_elastic_ep, Methods: POST 2025-10-10T01:43:35.5201958Z (APIServer pid=35380) INFO 10-10 01:43:35 [launcher.py:42] Route: /is_scaling_elastic_ep, Methods: POST 2025-10-10T01:43:35.5202528Z (APIServer pid=35380) INFO 10-10 01:43:35 [launcher.py:42] Route: /invocations, Methods: POST 2025-10-10T01:43:35.5203066Z (APIServer pid=35380) INFO 10-10 01:43:35 [launcher.py:42] Route: /metrics, Methods: GET 2025-10-10T01:43:35.5336550Z (APIServer pid=35380) INFO: Started server process [35380] 2025-10-10T01:43:35.5338173Z (APIServer pid=35380) INFO: Waiting for application startup. 2025-10-10T01:43:35.7655667Z (APIServer pid=35380) INFO: Application startup complete. 2025-10-10T01:43:36.1544322Z (APIServer pid=35380) INFO: 127.0.0.1:53404 - "GET /health HTTP/1.1" 200 OK 2025-10-10T01:43:36.3835627Z (APIServer pid=35380) INFO: 127.0.0.1:53410 - "GET /v1/models HTTP/1.1" 200 OK 2025-10-10T01:43:36.4148219Z (EngineCore_DP0 pid=35502) INFO 10-10 01:43:36 [ray_distributed_executor.py:552] RAY_CGRAPH_get_timeout is set to 300 2025-10-10T01:43:36.4149198Z (EngineCore_DP0 pid=35502) INFO 10-10 01:43:36 [ray_distributed_executor.py:554] VLLM_USE_RAY_COMPILED_DAG_CHANNEL_TYPE = auto 2025-10-10T01:43:36.4150148Z (EngineCore_DP0 pid=35502) INFO 10-10 01:43:36 [ray_distributed_executor.py:556] VLLM_USE_RAY_COMPILED_DAG_OVERLAP_COMM = False 2025-10-10T01:43:36.4154368Z (EngineCore_DP0 pid=35502) INFO 10-10 01:43:36 [ray_distributed_executor.py:621] Using RayPPCommunicator (which wraps vLLM _PP GroupCoordinator) for Ray Compiled Graph communication. 2025-10-10T01:43:36.5882510Z (APIServer pid=35380) INFO: 127.0.0.1:53410 - "POST /v1/completions HTTP/1.1" 200 OK 2025-10-10T01:43:36.6466258Z (APIServer pid=35380) INFO: 127.0.0.1:53410 - "POST /v1/completions HTTP/1.1" 200 OK 2025-10-10T01:43:36.7841832Z (APIServer pid=35380) INFO: 127.0.0.1:53410 - "POST /v1/completions HTTP/1.1" 200 OK 2025-10-10T01:43:36.8558958Z (APIServer pid=35380) INFO: 127.0.0.1:53410 - "POST /v1/completions HTTP/1.1" 200 OK 2025-10-10T01:43:36.9130959Z (APIServer pid=35380) INFO: 127.0.0.1:53410 - "POST /v1/completions HTTP/1.1" 200 OK 2025-10-10T01:43:36.9211204Z (APIServer pid=35380) INFO: 127.0.0.1:53410 - "POST /v1/completions HTTP/1.1" 200 OK 2025-10-10T01:43:37.0048280Z (APIServer pid=35380) INFO 10-10 01:43:37 [launcher.py:99] Shutting down FastAPI HTTP server. 2025-10-10T01:43:37.0049346Z *** SIGTERM received at time=1760060616 on cpu 8 *** 2025-10-10T01:43:37.0098843Z PC: @ 0x7fd7f250f117 (unknown) (unknown) 2025-10-10T01:43:37.0100005Z @ 0x7fd7f24c0520 (unknown) (unknown) 2025-10-10T01:43:37.0100476Z [2025-10-10 01:43:37,009 E 35502 35502] logging.cc:501: *** SIGTERM received at time=1760060616 on cpu 8 *** 2025-10-10T01:43:37.0102547Z [2025-10-10 01:43:37,010 E 35502 35502] logging.cc:501: PC: @ 0x7fd7f250f117 (unknown) (unknown) 2025-10-10T01:43:37.0103790Z [2025-10-10 01:43:37,010 E 35502 35502] logging.cc:501: @ 0x7fd7f24c0520 (unknown) (unknown) 2025-10-10T01:43:37.0105648Z (EngineCore_DP0 pid=35502) INFO 10-10 01:43:37 [ray_distributed_executor.py:122] Shutting down Ray distributed executor. If you see error log from logging.cc regarding SIGTERM received, please ignore because this is the expected termination process in Ray. 2025-10-10T01:43:37.0108783Z (EngineCore_DP0 pid=35502) 2025-10-10 01:43:37,010 INFO compiled_dag_node.py:2171 -- Tearing down compiled DAG 2025-10-10T01:43:37.0109651Z (EngineCore_DP0 pid=35502) 2025-10-10 01:43:37,010 INFO compiled_dag_node.py:2176 -- Cancelling compiled worker on actor: Actor(RayWorkerWrapper, f32ba175f0b8e9f3e0883efd01000000) 2025-10-10T01:43:37.0110668Z (EngineCore_DP0 pid=35502) 2025-10-10 01:43:37,010 INFO compiled_dag_node.py:2176 -- Cancelling compiled worker on actor: Actor(RayWorkerWrapper, 266b270618a8e6d9d4036aba01000000) 2025-10-10T01:43:37.0151753Z (EngineCore_DP0 pid=35502) 2025-10-10 01:43:37,014 INFO compiled_dag_node.py:2198 -- Waiting for worker tasks to exit 2025-10-10T01:43:37.0153255Z (EngineCore_DP0 pid=35502) 2025-10-10 01:43:37,015 INFO compiled_dag_node.py:2201 -- Teardown complete 2025-10-10T01:43:37.5179583Z (EngineCore_DP0 pid=35502) (RayWorkerWrapper pid=36385) INFO 10-10 01:43:29 [backends.py:548] Using cache directory: /var/lib/jenkins/.cache/vllm/torch_compile_cache/31e0bc77eb/rank_1_0/backbone for vLLM's torch.compile 2025-10-10T01:43:37.5181025Z (EngineCore_DP0 pid=35502) (RayWorkerWrapper pid=36385) INFO 10-10 01:43:29 [backends.py:559] Dynamo bytecode transform time: 2.82 s 2025-10-10T01:43:37.5182236Z (EngineCore_DP0 pid=35502) (RayWorkerWrapper pid=36391) INFO 10-10 01:43:31 [backends.py:164] Directly load the compiled graph(s) for dynamic shape from the cache, took 0.933 s 2025-10-10T01:43:37.5183367Z (EngineCore_DP0 pid=35502) (RayWorkerWrapper pid=36385) INFO 10-10 01:43:32 [gpu_worker.py:298] Available KV cache memory: 18.54 GiB 2025-10-10T01:43:37.5184746Z (EngineCore_DP0 pid=35502) (RayWorkerWrapper pid=36385) INFO 10-10 01:43:33 [gpu_worker.py:332] Compile and warming up model for size 4 [repeated 3x across cluster] 2025-10-10T01:43:37.5185983Z (EngineCore_DP0 pid=35502) (RayWorkerWrapper pid=36385) INFO 10-10 01:43:34 [backends.py:168] Directly load the compiled graph(s) for shape 4 from the cache, took 0.666 s [repeated 3x across cluster] 2025-10-10T01:43:37.5186962Z (EngineCore_DP0 pid=35502) (RayWorkerWrapper pid=36385) INFO 10-10 01:43:34 [monitor.py:32] torch.compile takes 2.82 s in total 2025-10-10T01:43:37.5188396Z (EngineCore_DP0 pid=35502) (pid=36385) /opt/conda/envs/py_3.12/lib/python3.12/site-packages/torch/cuda/__init__.py:64: FutureWarning: The pynvml package is deprecated. Please install nvidia-ml-py instead. If you did not install pynvml directly, please report this to the maintainers of the package that installed pynvml for you. 2025-10-10T01:43:37.5189611Z (EngineCore_DP0 pid=35502) (pid=36385) import pynvml # type: ignore[import] 2025-10-10T01:43:39.3812443Z (APIServer pid=35380) INFO: Shutting down 2025-10-10T01:43:39.4817332Z (APIServer pid=35380) INFO: Waiting for application shutdown. 2025-10-10T01:43:39.4818478Z (APIServer pid=35380) INFO: Application shutdown complete. 2025-10-10T01:43:39.8282512Z INFO 10-10 01:43:39 [model.py:551] Resolved architecture: LlamaForCausalLM 2025-10-10T01:43:39.8283065Z WARNING 10-10 01:43:39 [model.py:1768] Casting torch.bfloat16 to torch.float16. 2025-10-10T01:43:39.8283695Z INFO 10-10 01:43:39 [model.py:1545] Using max model len 2048 2025-10-10T01:43:39.8285511Z Launching RemoteOpenAIServer with: vllm serve meta-llama/Llama-3.2-1B-Instruct --dtype float16 --max-model-len 2048 --max-num-seqs 8 --enforce-eager --tokenizer-mode auto --tensor-parallel-size 2 --distributed-executor-backend mp --load-format dummy --port 43789 --seed 0 2025-10-10T01:43:40.3671413Z /opt/conda/envs/py_3.12/lib/python3.12/site-packages/torch/cuda/__init__.py:64: FutureWarning: The pynvml package is deprecated. Please install nvidia-ml-py instead. If you did not install pynvml directly, please report this to the maintainers of the package that installed pynvml for you. 2025-10-10T01:43:40.3674379Z import pynvml # type: ignore[import] 2025-10-10T01:43:42.5639192Z INFO 10-10 01:43:42 [__init__.py:215] Automatically detected platform cuda. 2025-10-10T01:43:46.4471650Z (APIServer pid=40136) INFO 10-10 01:43:46 [api_server.py:1839] vLLM API server version 0.11.0rc2.dev157+g0ad9951c4.d20251010 2025-10-10T01:43:46.4494523Z (APIServer pid=40136) INFO 10-10 01:43:46 [utils.py:233] non-default args: {'model_tag': 'meta-llama/Llama-3.2-1B-Instruct', 'port': 43789, 'model': 'meta-llama/Llama-3.2-1B-Instruct', 'dtype': 'float16', 'seed': 0, 'max_model_len': 2048, 'enforce_eager': True, 'load_format': 'dummy', 'distributed_executor_backend': 'mp', 'tensor_parallel_size': 2, 'max_num_seqs': 8} 2025-10-10T01:43:46.6081819Z (APIServer pid=40136) INFO 10-10 01:43:46 [model.py:551] Resolved architecture: LlamaForCausalLM 2025-10-10T01:43:46.6082560Z (APIServer pid=40136) `torch_dtype` is deprecated! Use `dtype` instead! 2025-10-10T01:43:46.6083248Z (APIServer pid=40136) WARNING 10-10 01:43:46 [model.py:1768] Casting torch.bfloat16 to torch.float16. 2025-10-10T01:43:46.6083946Z (APIServer pid=40136) INFO 10-10 01:43:46 [model.py:1545] Using max model len 2048 2025-10-10T01:43:46.7634939Z (APIServer pid=40136) INFO 10-10 01:43:46 [scheduler.py:222] Chunked prefill is enabled with max_num_batched_tokens=2048. 2025-10-10T01:43:46.7640387Z (APIServer pid=40136) INFO 10-10 01:43:46 [vllm.py:345] Cudagraph is disabled under eager mode 2025-10-10T01:43:47.7548979Z /opt/conda/envs/py_3.12/lib/python3.12/site-packages/torch/cuda/__init__.py:64: FutureWarning: The pynvml package is deprecated. Please install nvidia-ml-py instead. If you did not install pynvml directly, please report this to the maintainers of the package that installed pynvml for you. 2025-10-10T01:43:47.7550881Z import pynvml # type: ignore[import] 2025-10-10T01:43:50.0615494Z INFO 10-10 01:43:50 [__init__.py:215] Automatically detected platform cuda. 2025-10-10T01:43:54.0289504Z (EngineCore_DP0 pid=40258) INFO 10-10 01:43:54 [core.py:648] Waiting for init message from front-end. 2025-10-10T01:43:54.0563458Z (EngineCore_DP0 pid=40258) INFO 10-10 01:43:54 [core.py:78] Initializing a V1 LLM engine (v0.11.0rc2.dev157+g0ad9951c4.d20251010) with config: model='meta-llama/Llama-3.2-1B-Instruct', speculative_config=None, tokenizer='meta-llama/Llama-3.2-1B-Instruct', skip_tokenizer_init=False, tokenizer_mode=auto, revision=None, tokenizer_revision=None, trust_remote_code=False, dtype=torch.float16, max_seq_len=2048, download_dir=None, load_format=dummy, tensor_parallel_size=2, pipeline_parallel_size=1, data_parallel_size=1, disable_custom_all_reduce=False, quantization=None, enforce_eager=True, kv_cache_dtype=auto, device_config=cuda, structured_outputs_config=StructuredOutputsConfig(backend='auto', disable_fallback=False, disable_any_whitespace=False, disable_additional_properties=False, reasoning_parser=''), observability_config=ObservabilityConfig(show_hidden_metrics_for_version=None, otlp_traces_endpoint=None, collect_detailed_traces=None), seed=0, served_model_name=meta-llama/Llama-3.2-1B-Instruct, enable_prefix_caching=True, chunked_prefill_enabled=True, pooler_config=None, compilation_config={'level': 0, 'debug_dump_path': None, 'cache_dir': '', 'backend': '', 'custom_ops': [], 'splitting_ops': None, 'use_inductor': True, 'compile_sizes': [], 'inductor_compile_config': {'enable_auto_functionalized_v2': False}, 'inductor_passes': {}, 'cudagraph_mode': , 'use_cudagraph': False, 'cudagraph_num_of_warmups': 0, 'cudagraph_capture_sizes': [], 'cudagraph_copy_inputs': False, 'full_cuda_graph': False, 'use_inductor_graph_partition': False, 'pass_config': {}, 'max_capture_size': 0, 'local_cache_dir': None} 2025-10-10T01:43:54.0572990Z (EngineCore_DP0 pid=40258) WARNING 10-10 01:43:54 [multiproc_executor.py:720] Reducing Torch parallelism from 24 threads to 1 to avoid unnecessary CPU contention. Set OMP_NUM_THREADS in the external environment to tune this value as needed. 2025-10-10T01:43:54.0580624Z (EngineCore_DP0 pid=40258) INFO 10-10 01:43:54 [shm_broadcast.py:289] vLLM message queue communication handle: Handle(local_reader_ranks=[0, 1], buffer_handle=(2, 16777216, 10, 'psm_bef2abc3'), local_subscribe_addr='ipc:///tmp/4a6ca3f1-e0a9-43a6-a41b-16e249983cee', remote_subscribe_addr=None, remote_addr_ipv6=False) 2025-10-10T01:43:54.5475165Z /opt/conda/envs/py_3.12/lib/python3.12/site-packages/torch/cuda/__init__.py:64: FutureWarning: The pynvml package is deprecated. Please install nvidia-ml-py instead. If you did not install pynvml directly, please report this to the maintainers of the package that installed pynvml for you. 2025-10-10T01:43:54.5477922Z import pynvml # type: ignore[import] 2025-10-10T01:43:54.5480316Z /opt/conda/envs/py_3.12/lib/python3.12/site-packages/torch/cuda/__init__.py:64: FutureWarning: The pynvml package is deprecated. Please install nvidia-ml-py instead. If you did not install pynvml directly, please report this to the maintainers of the package that installed pynvml for you. 2025-10-10T01:43:54.5481530Z import pynvml # type: ignore[import] 2025-10-10T01:43:56.8436674Z INFO 10-10 01:43:56 [__init__.py:215] Automatically detected platform cuda. 2025-10-10T01:43:56.8835616Z INFO 10-10 01:43:56 [__init__.py:215] Automatically detected platform cuda. 2025-10-10T01:44:01.3343842Z INFO 10-10 01:44:01 [shm_broadcast.py:289] vLLM message queue communication handle: Handle(local_reader_ranks=[0], buffer_handle=(1, 10485760, 10, 'psm_dadc679a'), local_subscribe_addr='ipc:///tmp/e763a749-d8ba-428f-9caf-c6900cf2ccb9', remote_subscribe_addr=None, remote_addr_ipv6=False) 2025-10-10T01:44:01.3839836Z INFO 10-10 01:44:01 [shm_broadcast.py:289] vLLM message queue communication handle: Handle(local_reader_ranks=[0], buffer_handle=(1, 10485760, 10, 'psm_9bb0a5c6'), local_subscribe_addr='ipc:///tmp/e2a01af6-b4e4-4e2a-88b0-51e31be6c00c', remote_subscribe_addr=None, remote_addr_ipv6=False) 2025-10-10T01:44:02.1325483Z [Gloo] Rank 0 is connected to 1 peer ranks. Expected number of connected peer ranks is : 1 2025-10-10T01:44:02.1326137Z [Gloo] Rank 1 is connected to 1 peer ranks. Expected number of connected peer ranks is : 1 2025-10-10T01:44:02.1365349Z [Gloo] Rank 0 is connected to 1 peer ranks. Expected number of connected peer ranks is : 1 2025-10-10T01:44:02.1365964Z [Gloo] Rank 1 is connected to 1 peer ranks. Expected number of connected peer ranks is : 1 2025-10-10T01:44:02.1407982Z INFO 10-10 01:44:02 [__init__.py:1384] Found nccl from library libnccl.so.2 2025-10-10T01:44:02.1408474Z INFO 10-10 01:44:02 [__init__.py:1384] Found nccl from library libnccl.so.2 2025-10-10T01:44:02.1410398Z INFO 10-10 01:44:02 [pynccl.py:104] vLLM is using nccl==2.27.5 2025-10-10T01:44:02.1410820Z INFO 10-10 01:44:02 [pynccl.py:104] vLLM is using nccl==2.27.5 2025-10-10T01:44:02.3114273Z WARNING 10-10 01:44:02 [symm_mem.py:58] SymmMemCommunicator: Device capability 8.9 not supported, communicator is not available. 2025-10-10T01:44:02.3115992Z WARNING 10-10 01:44:02 [symm_mem.py:58] SymmMemCommunicator: Device capability 8.9 not supported, communicator is not available. 2025-10-10T01:44:02.3136225Z INFO 10-10 01:44:02 [custom_all_reduce.py:35] Skipping P2P check and trusting the driver's P2P report. 2025-10-10T01:44:02.3136871Z INFO 10-10 01:44:02 [custom_all_reduce.py:35] Skipping P2P check and trusting the driver's P2P report. 2025-10-10T01:44:02.3137919Z WARNING 10-10 01:44:02 [custom_all_reduce.py:154] Custom allreduce is disabled because your platform lacks GPU P2P capability or P2P test failed. To silence this warning, specify disable_custom_all_reduce=True explicitly. 2025-10-10T01:44:02.3139378Z WARNING 10-10 01:44:02 [custom_all_reduce.py:154] Custom allreduce is disabled because your platform lacks GPU P2P capability or P2P test failed. To silence this warning, specify disable_custom_all_reduce=True explicitly. 2025-10-10T01:44:02.4341842Z INFO 10-10 01:44:02 [shm_broadcast.py:289] vLLM message queue communication handle: Handle(local_reader_ranks=[1], buffer_handle=(1, 4194304, 6, 'psm_1b32aefb'), local_subscribe_addr='ipc:///tmp/06b85683-a0d9-44df-9ded-5945e45f9e03', remote_subscribe_addr=None, remote_addr_ipv6=False) 2025-10-10T01:44:02.4366632Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-10-10T01:44:02.4367139Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-10-10T01:44:02.4374577Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-10-10T01:44:02.4381859Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-10-10T01:44:02.4837092Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-10-10T01:44:02.4845987Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-10-10T01:44:02.4858520Z [Gloo] Rank 0 is connected to 1 peer ranks. Expected number of connected peer ranks is : 1 2025-10-10T01:44:02.4859673Z [Gloo] Rank 1 is connected to 1 peer ranks. Expected number of connected peer ranks is : 1 2025-10-10T01:44:02.4860708Z INFO 10-10 01:44:02 [__init__.py:1384] Found nccl from library libnccl.so.2 2025-10-10T01:44:02.4861566Z INFO 10-10 01:44:02 [__init__.py:1384] Found nccl from library libnccl.so.2 2025-10-10T01:44:02.4862344Z INFO 10-10 01:44:02 [pynccl.py:104] vLLM is using nccl==2.27.5 2025-10-10T01:44:02.4863239Z INFO 10-10 01:44:02 [pynccl.py:104] vLLM is using nccl==2.27.5 2025-10-10T01:44:02.5217254Z INFO 10-10 01:44:02 [parallel_state.py:1208] rank 1 in world size 2 is assigned as DP rank 0, PP rank 0, TP rank 1, EP rank 1 2025-10-10T01:44:02.5218309Z INFO 10-10 01:44:02 [parallel_state.py:1208] rank 0 in world size 2 is assigned as DP rank 0, PP rank 0, TP rank 0, EP rank 0 2025-10-10T01:44:02.7609927Z INFO 10-10 01:44:02 [topk_topp_sampler.py:55] Using FlashInfer for top-p & top-k sampling. 2025-10-10T01:44:02.7610576Z INFO 10-10 01:44:02 [topk_topp_sampler.py:55] Using FlashInfer for top-p & top-k sampling. 2025-10-10T01:44:02.7932584Z (Worker_TP1 pid=40378) INFO 10-10 01:44:02 [gpu_model_runner.py:2707] Starting to load model meta-llama/Llama-3.2-1B-Instruct... 2025-10-10T01:44:02.7935518Z (Worker_TP0 pid=40377) INFO 10-10 01:44:02 [gpu_model_runner.py:2707] Starting to load model meta-llama/Llama-3.2-1B-Instruct... 2025-10-10T01:44:03.0132588Z (Worker_TP1 pid=40378) INFO 10-10 01:44:03 [gpu_model_runner.py:2739] Loading model from scratch... 2025-10-10T01:44:03.0192174Z (Worker_TP0 pid=40377) INFO 10-10 01:44:03 [gpu_model_runner.py:2739] Loading model from scratch... 2025-10-10T01:44:03.1028833Z (Worker_TP1 pid=40378) INFO 10-10 01:44:03 [cuda.py:361] Using Flash Attention backend on V1 engine. 2025-10-10T01:44:03.1091971Z (Worker_TP0 pid=40377) INFO 10-10 01:44:03 [cuda.py:361] Using Flash Attention backend on V1 engine. 2025-10-10T01:44:03.5871335Z (Worker_TP1 pid=40378) INFO 10-10 01:44:03 [gpu_model_runner.py:2758] Model loading took 1.1667 GiB and 0.118115 seconds 2025-10-10T01:44:03.6175669Z (Worker_TP0 pid=40377) INFO 10-10 01:44:03 [gpu_model_runner.py:2758] Model loading took 1.1667 GiB and 0.119042 seconds 2025-10-10T01:44:08.5481324Z (Worker_TP0 pid=40377) INFO 10-10 01:44:08 [gpu_worker.py:298] Available KV cache memory: 18.53 GiB 2025-10-10T01:44:08.5539827Z (Worker_TP1 pid=40378) INFO 10-10 01:44:08 [gpu_worker.py:298] Available KV cache memory: 18.53 GiB 2025-10-10T01:44:08.7753129Z (EngineCore_DP0 pid=40258) INFO 10-10 01:44:08 [kv_cache_utils.py:1087] GPU KV cache size: 1,214,464 tokens 2025-10-10T01:44:08.7754092Z (EngineCore_DP0 pid=40258) INFO 10-10 01:44:08 [kv_cache_utils.py:1091] Maximum concurrency for 2,048 tokens per request: 593.00x 2025-10-10T01:44:08.7755472Z (EngineCore_DP0 pid=40258) INFO 10-10 01:44:08 [kv_cache_utils.py:1087] GPU KV cache size: 1,214,464 tokens 2025-10-10T01:44:08.7756307Z (EngineCore_DP0 pid=40258) INFO 10-10 01:44:08 [kv_cache_utils.py:1091] Maximum concurrency for 2,048 tokens per request: 593.00x 2025-10-10T01:44:08.8979182Z (EngineCore_DP0 pid=40258) INFO 10-10 01:44:08 [core.py:211] init engine (profile, create kv cache, warmup model) took 5.21 seconds 2025-10-10T01:44:09.4040113Z (EngineCore_DP0 pid=40258) INFO 10-10 01:44:09 [vllm.py:345] Cudagraph is disabled under eager mode 2025-10-10T01:44:09.5886607Z (APIServer pid=40136) INFO 10-10 01:44:09 [loggers.py:147] Engine 000: vllm cache_config_info with initialization after num_gpu_blocks is: 75904 2025-10-10T01:44:09.6177976Z (EngineCore_DP0 pid=40258) INFO 10-10 01:44:09 [gc_utils.py:41] GC Debug Config. enabled:False,top_objects:-1 2025-10-10T01:44:09.6392005Z (APIServer pid=40136) INFO 10-10 01:44:09 [api_server.py:1634] Supported_tasks: ['generate'] 2025-10-10T01:44:09.6685017Z (APIServer pid=40136) WARNING 10-10 01:44:09 [model.py:1424] Default sampling parameters have been overridden by the model's Hugging Face generation config recommended from the model creator. If this is not intended, please relaunch vLLM instance with `--generation-config vllm`. 2025-10-10T01:44:09.6686750Z (APIServer pid=40136) INFO 10-10 01:44:09 [serving_responses.py:140] Using default chat sampling params from model: {'temperature': 0.6, 'top_p': 0.9} 2025-10-10T01:44:09.6912630Z (APIServer pid=40136) INFO 10-10 01:44:09 [serving_chat.py:139] Using default chat sampling params from model: {'temperature': 0.6, 'top_p': 0.9} 2025-10-10T01:44:09.7195423Z (APIServer pid=40136) INFO 10-10 01:44:09 [serving_completion.py:76] Using default completion sampling params from model: {'temperature': 0.6, 'top_p': 0.9} 2025-10-10T01:44:09.7196503Z (APIServer pid=40136) INFO 10-10 01:44:09 [api_server.py:1912] Starting vLLM API server 0 on http://0.0.0.0:43789 2025-10-10T01:44:09.7197235Z (APIServer pid=40136) INFO 10-10 01:44:09 [launcher.py:34] Available routes are: 2025-10-10T01:44:09.7197925Z (APIServer pid=40136) INFO 10-10 01:44:09 [launcher.py:42] Route: /openapi.json, Methods: HEAD, GET 2025-10-10T01:44:09.7198718Z (APIServer pid=40136) INFO 10-10 01:44:09 [launcher.py:42] Route: /docs, Methods: HEAD, GET 2025-10-10T01:44:09.7199441Z (APIServer pid=40136) INFO 10-10 01:44:09 [launcher.py:42] Route: /docs/oauth2-redirect, Methods: HEAD, GET 2025-10-10T01:44:09.7200188Z (APIServer pid=40136) INFO 10-10 01:44:09 [launcher.py:42] Route: /redoc, Methods: HEAD, GET 2025-10-10T01:44:09.7200856Z (APIServer pid=40136) INFO 10-10 01:44:09 [launcher.py:42] Route: /health, Methods: GET 2025-10-10T01:44:09.7201498Z (APIServer pid=40136) INFO 10-10 01:44:09 [launcher.py:42] Route: /load, Methods: GET 2025-10-10T01:44:09.7202145Z (APIServer pid=40136) INFO 10-10 01:44:09 [launcher.py:42] Route: /ping, Methods: POST 2025-10-10T01:44:09.7202771Z (APIServer pid=40136) INFO 10-10 01:44:09 [launcher.py:42] Route: /ping, Methods: GET 2025-10-10T01:44:09.7203411Z (APIServer pid=40136) INFO 10-10 01:44:09 [launcher.py:42] Route: /tokenize, Methods: POST 2025-10-10T01:44:09.7204076Z (APIServer pid=40136) INFO 10-10 01:44:09 [launcher.py:42] Route: /detokenize, Methods: POST 2025-10-10T01:44:09.7204722Z (APIServer pid=40136) INFO 10-10 01:44:09 [launcher.py:42] Route: /v1/models, Methods: GET 2025-10-10T01:44:09.7205347Z (APIServer pid=40136) INFO 10-10 01:44:09 [launcher.py:42] Route: /version, Methods: GET 2025-10-10T01:44:09.7205891Z (APIServer pid=40136) INFO 10-10 01:44:09 [launcher.py:42] Route: /v1/responses, Methods: POST 2025-10-10T01:44:09.7206764Z (APIServer pid=40136) INFO 10-10 01:44:09 [launcher.py:42] Route: /v1/responses/{response_id}, Methods: GET 2025-10-10T01:44:09.7207394Z (APIServer pid=40136) INFO 10-10 01:44:09 [launcher.py:42] Route: /v1/responses/{response_id}/cancel, Methods: POST 2025-10-10T01:44:09.7208015Z (APIServer pid=40136) INFO 10-10 01:44:09 [launcher.py:42] Route: /v1/chat/completions, Methods: POST 2025-10-10T01:44:09.7208578Z (APIServer pid=40136) INFO 10-10 01:44:09 [launcher.py:42] Route: /v1/completions, Methods: POST 2025-10-10T01:44:09.7209120Z (APIServer pid=40136) INFO 10-10 01:44:09 [launcher.py:42] Route: /v1/embeddings, Methods: POST 2025-10-10T01:44:09.7209651Z (APIServer pid=40136) INFO 10-10 01:44:09 [launcher.py:42] Route: /pooling, Methods: POST 2025-10-10T01:44:09.7210321Z (APIServer pid=40136) INFO 10-10 01:44:09 [launcher.py:42] Route: /classify, Methods: POST 2025-10-10T01:44:09.7210875Z (APIServer pid=40136) INFO 10-10 01:44:09 [launcher.py:42] Route: /score, Methods: POST 2025-10-10T01:44:09.7211396Z (APIServer pid=40136) INFO 10-10 01:44:09 [launcher.py:42] Route: /v1/score, Methods: POST 2025-10-10T01:44:09.7211983Z (APIServer pid=40136) INFO 10-10 01:44:09 [launcher.py:42] Route: /v1/audio/transcriptions, Methods: POST 2025-10-10T01:44:09.7212593Z (APIServer pid=40136) INFO 10-10 01:44:09 [launcher.py:42] Route: /v1/audio/translations, Methods: POST 2025-10-10T01:44:09.7213145Z (APIServer pid=40136) INFO 10-10 01:44:09 [launcher.py:42] Route: /rerank, Methods: POST 2025-10-10T01:44:09.7213666Z (APIServer pid=40136) INFO 10-10 01:44:09 [launcher.py:42] Route: /v1/rerank, Methods: POST 2025-10-10T01:44:09.7214180Z (APIServer pid=40136) INFO 10-10 01:44:09 [launcher.py:42] Route: /v2/rerank, Methods: POST 2025-10-10T01:44:09.7214734Z (APIServer pid=40136) INFO 10-10 01:44:09 [launcher.py:42] Route: /scale_elastic_ep, Methods: POST 2025-10-10T01:44:09.7215319Z (APIServer pid=40136) INFO 10-10 01:44:09 [launcher.py:42] Route: /is_scaling_elastic_ep, Methods: POST 2025-10-10T01:44:09.7215890Z (APIServer pid=40136) INFO 10-10 01:44:09 [launcher.py:42] Route: /invocations, Methods: POST 2025-10-10T01:44:09.7216414Z (APIServer pid=40136) INFO 10-10 01:44:09 [launcher.py:42] Route: /metrics, Methods: GET 2025-10-10T01:44:09.7347441Z (APIServer pid=40136) INFO: Started server process [40136] 2025-10-10T01:44:09.7348094Z (APIServer pid=40136) INFO: Waiting for application startup. 2025-10-10T01:44:09.9612519Z (APIServer pid=40136) INFO: Application startup complete. 2025-10-10T01:44:10.4089967Z (APIServer pid=40136) INFO: 127.0.0.1:56048 - "GET /health HTTP/1.1" 200 OK 2025-10-10T01:44:10.4176861Z (APIServer pid=40136) INFO: 127.0.0.1:56064 - "GET /v1/models HTTP/1.1" 200 OK 2025-10-10T01:44:10.5891411Z (APIServer pid=40136) INFO: 127.0.0.1:56064 - "POST /v1/completions HTTP/1.1" 200 OK 2025-10-10T01:44:10.6441161Z (APIServer pid=40136) INFO: 127.0.0.1:56064 - "POST /v1/completions HTTP/1.1" 200 OK 2025-10-10T01:44:10.7401210Z (APIServer pid=40136) INFO: 127.0.0.1:56064 - "POST /v1/completions HTTP/1.1" 200 OK 2025-10-10T01:44:10.8229515Z (APIServer pid=40136) INFO: 127.0.0.1:56064 - "POST /v1/completions HTTP/1.1" 200 OK 2025-10-10T01:44:10.8927729Z (APIServer pid=40136) INFO: 127.0.0.1:56064 - "POST /v1/completions HTTP/1.1" 200 OK 2025-10-10T01:44:10.9004945Z (APIServer pid=40136) INFO: 127.0.0.1:56064 - "POST /v1/completions HTTP/1.1" 200 OK 2025-10-10T01:44:10.9756139Z (APIServer pid=40136) INFO 10-10 01:44:10 [launcher.py:99] Shutting down FastAPI HTTP server. 2025-10-10T01:44:10.9758956Z (Worker_TP0 pid=40377) INFO 10-10 01:44:10 [multiproc_executor.py:558] Parent process exited, terminating worker 2025-10-10T01:44:10.9760292Z (Worker_TP0 pid=40377) INFO 10-10 01:44:10 [multiproc_executor.py:599] WorkerProc shutting down. 2025-10-10T01:44:10.9763279Z (Worker_TP1 pid=40378) INFO 10-10 01:44:10 [multiproc_executor.py:558] Parent process exited, terminating worker 2025-10-10T01:44:10.9764617Z (Worker_TP1 pid=40378) INFO 10-10 01:44:10 [multiproc_executor.py:599] WorkerProc shutting down. 2025-10-10T01:44:12.8923111Z (APIServer pid=40136) INFO: Shutting down 2025-10-10T01:44:12.9928779Z (APIServer pid=40136) INFO: Waiting for application shutdown. 2025-10-10T01:44:12.9930432Z (APIServer pid=40136) INFO: Application shutdown complete. 2025-10-10T01:44:13.6167028Z PASSED 2025-10-10T01:44:13.6281105Z distributed/test_sequence_parallel.py::test_tp_sp_generation[meta-llama/Llama-3.2-1B-Instruct-parallel_setup10-mp-1-auto-test_options10] Fork a new process to run a test 40895 2025-10-10T01:44:13.6297959Z Fork a new process to run a test 0 2025-10-10T01:44:14.2819104Z INFO 10-10 01:44:14 [model.py:551] Resolved architecture: LlamaForCausalLM 2025-10-10T01:44:14.2819622Z `torch_dtype` is deprecated! Use `dtype` instead! 2025-10-10T01:44:14.2820588Z WARNING 10-10 01:44:14 [model.py:1768] Casting torch.bfloat16 to torch.float16. 2025-10-10T01:44:14.2821035Z INFO 10-10 01:44:14 [model.py:1545] Using max model len 2048 2025-10-10T01:44:14.2825880Z Launching RemoteOpenAIServer with: vllm serve meta-llama/Llama-3.2-1B-Instruct --dtype float16 --max-model-len 2048 --max-num-seqs 8 --enable-chunked-prefill --enforce-eager --tokenizer-mode auto --tensor-parallel-size 2 --pipeline-parallel-size 1 --distributed-executor-backend mp --compilation_config {"level": 3, "custom_ops": ["+rms_norm"], "compile_sizes": [4, 8], "pass_config": {"enable_sequence_parallelism": true, "enable_fusion": false, "enable_noop": true}} --load-format dummy --port 58147 --seed 0 2025-10-10T01:44:14.7910373Z /opt/conda/envs/py_3.12/lib/python3.12/site-packages/torch/cuda/__init__.py:64: FutureWarning: The pynvml package is deprecated. Please install nvidia-ml-py instead. If you did not install pynvml directly, please report this to the maintainers of the package that installed pynvml for you. 2025-10-10T01:44:14.7912713Z import pynvml # type: ignore[import] 2025-10-10T01:44:17.0002662Z INFO 10-10 01:44:16 [__init__.py:215] Automatically detected platform cuda. 2025-10-10T01:44:20.9080934Z (APIServer pid=40944) INFO 10-10 01:44:20 [api_server.py:1839] vLLM API server version 0.11.0rc2.dev157+g0ad9951c4.d20251010 2025-10-10T01:44:20.9105134Z (APIServer pid=40944) INFO 10-10 01:44:20 [utils.py:233] non-default args: {'model_tag': 'meta-llama/Llama-3.2-1B-Instruct', 'port': 58147, 'model': 'meta-llama/Llama-3.2-1B-Instruct', 'dtype': 'float16', 'seed': 0, 'max_model_len': 2048, 'enforce_eager': True, 'load_format': 'dummy', 'distributed_executor_backend': 'mp', 'tensor_parallel_size': 2, 'max_num_seqs': 8, 'enable_chunked_prefill': True, 'compilation_config': {'level': 3, 'debug_dump_path': None, 'cache_dir': '', 'backend': '', 'custom_ops': ['+rms_norm'], 'splitting_ops': None, 'use_inductor': True, 'compile_sizes': [4, 8], 'inductor_compile_config': {'enable_auto_functionalized_v2': False}, 'inductor_passes': {}, 'cudagraph_mode': None, 'use_cudagraph': True, 'cudagraph_num_of_warmups': 0, 'cudagraph_capture_sizes': None, 'cudagraph_copy_inputs': False, 'full_cuda_graph': False, 'use_inductor_graph_partition': False, 'pass_config': {'enable_noop': True, 'enable_sequence_parallelism': True}, 'max_capture_size': None, 'local_cache_dir': None}} 2025-10-10T01:44:21.1144626Z (APIServer pid=40944) INFO 10-10 01:44:21 [model.py:551] Resolved architecture: LlamaForCausalLM 2025-10-10T01:44:21.1146019Z (APIServer pid=40944) `torch_dtype` is deprecated! Use `dtype` instead! 2025-10-10T01:44:21.1147375Z (APIServer pid=40944) WARNING 10-10 01:44:21 [model.py:1768] Casting torch.bfloat16 to torch.float16. 2025-10-10T01:44:21.1148715Z (APIServer pid=40944) INFO 10-10 01:44:21 [model.py:1545] Using max model len 2048 2025-10-10T01:44:21.2806484Z (APIServer pid=40944) INFO 10-10 01:44:21 [scheduler.py:222] Chunked prefill is enabled with max_num_batched_tokens=2048. 2025-10-10T01:44:21.2811433Z (APIServer pid=40944) INFO 10-10 01:44:21 [vllm.py:345] Cudagraph is disabled under eager mode 2025-10-10T01:44:22.2523827Z /opt/conda/envs/py_3.12/lib/python3.12/site-packages/torch/cuda/__init__.py:64: FutureWarning: The pynvml package is deprecated. Please install nvidia-ml-py instead. If you did not install pynvml directly, please report this to the maintainers of the package that installed pynvml for you. 2025-10-10T01:44:22.2526785Z import pynvml # type: ignore[import] 2025-10-10T01:44:24.5586028Z INFO 10-10 01:44:24 [__init__.py:215] Automatically detected platform cuda. 2025-10-10T01:44:28.5413480Z (EngineCore_DP0 pid=41066) INFO 10-10 01:44:28 [core.py:648] Waiting for init message from front-end. 2025-10-10T01:44:28.5694616Z (EngineCore_DP0 pid=41066) INFO 10-10 01:44:28 [core.py:78] Initializing a V1 LLM engine (v0.11.0rc2.dev157+g0ad9951c4.d20251010) with config: model='meta-llama/Llama-3.2-1B-Instruct', speculative_config=None, tokenizer='meta-llama/Llama-3.2-1B-Instruct', skip_tokenizer_init=False, tokenizer_mode=auto, revision=None, tokenizer_revision=None, trust_remote_code=False, dtype=torch.float16, max_seq_len=2048, download_dir=None, load_format=dummy, tensor_parallel_size=2, pipeline_parallel_size=1, data_parallel_size=1, disable_custom_all_reduce=False, quantization=None, enforce_eager=True, kv_cache_dtype=auto, device_config=cuda, structured_outputs_config=StructuredOutputsConfig(backend='auto', disable_fallback=False, disable_any_whitespace=False, disable_additional_properties=False, reasoning_parser=''), observability_config=ObservabilityConfig(show_hidden_metrics_for_version=None, otlp_traces_endpoint=None, collect_detailed_traces=None), seed=0, served_model_name=meta-llama/Llama-3.2-1B-Instruct, enable_prefix_caching=True, chunked_prefill_enabled=True, pooler_config=None, compilation_config={'level': 3, 'debug_dump_path': None, 'cache_dir': '', 'backend': '', 'custom_ops': ['+rms_norm', '+rms_norm'], 'splitting_ops': ['vllm.unified_attention', 'vllm.unified_attention_with_output', 'vllm.mamba_mixer2', 'vllm.mamba_mixer', 'vllm.short_conv', 'vllm.linear_attention', 'vllm.plamo2_mamba_mixer', 'vllm.gdn_attention', 'vllm.sparse_attn_indexer'], 'use_inductor': True, 'compile_sizes': [8, 4], 'inductor_compile_config': {'enable_auto_functionalized_v2': False}, 'inductor_passes': {}, 'cudagraph_mode': , 'use_cudagraph': False, 'cudagraph_num_of_warmups': 0, 'cudagraph_capture_sizes': [], 'cudagraph_copy_inputs': False, 'full_cuda_graph': False, 'use_inductor_graph_partition': False, 'pass_config': {'enable_noop': True, 'enable_sequence_parallelism': True}, 'max_capture_size': 0, 'local_cache_dir': None} 2025-10-10T01:44:28.5700482Z (EngineCore_DP0 pid=41066) WARNING 10-10 01:44:28 [multiproc_executor.py:720] Reducing Torch parallelism from 24 threads to 1 to avoid unnecessary CPU contention. Set OMP_NUM_THREADS in the external environment to tune this value as needed. 2025-10-10T01:44:28.5702159Z (EngineCore_DP0 pid=41066) INFO 10-10 01:44:28 [shm_broadcast.py:289] vLLM message queue communication handle: Handle(local_reader_ranks=[0, 1], buffer_handle=(2, 16777216, 10, 'psm_37846cd2'), local_subscribe_addr='ipc:///tmp/6583d62a-dbcc-4a7e-a401-ad10289abd55', remote_subscribe_addr=None, remote_addr_ipv6=False) 2025-10-10T01:44:29.0465607Z /opt/conda/envs/py_3.12/lib/python3.12/site-packages/torch/cuda/__init__.py:64: FutureWarning: The pynvml package is deprecated. Please install nvidia-ml-py instead. If you did not install pynvml directly, please report this to the maintainers of the package that installed pynvml for you. 2025-10-10T01:44:29.0467945Z import pynvml # type: ignore[import] 2025-10-10T01:44:29.0682374Z /opt/conda/envs/py_3.12/lib/python3.12/site-packages/torch/cuda/__init__.py:64: FutureWarning: The pynvml package is deprecated. Please install nvidia-ml-py instead. If you did not install pynvml directly, please report this to the maintainers of the package that installed pynvml for you. 2025-10-10T01:44:29.0684602Z import pynvml # type: ignore[import] 2025-10-10T01:44:31.3730892Z INFO 10-10 01:44:31 [__init__.py:215] Automatically detected platform cuda. 2025-10-10T01:44:31.3802659Z INFO 10-10 01:44:31 [__init__.py:215] Automatically detected platform cuda. 2025-10-10T01:44:35.8845736Z INFO 10-10 01:44:35 [shm_broadcast.py:289] vLLM message queue communication handle: Handle(local_reader_ranks=[0], buffer_handle=(1, 10485760, 10, 'psm_cdbda97d'), local_subscribe_addr='ipc:///tmp/cd1ad103-444a-4159-bacf-6af2ec6dc50b', remote_subscribe_addr=None, remote_addr_ipv6=False) 2025-10-10T01:44:35.9842154Z INFO 10-10 01:44:35 [shm_broadcast.py:289] vLLM message queue communication handle: Handle(local_reader_ranks=[0], buffer_handle=(1, 10485760, 10, 'psm_375d4e64'), local_subscribe_addr='ipc:///tmp/2918ff95-f429-4944-b221-5aa97a4fd819', remote_subscribe_addr=None, remote_addr_ipv6=False) 2025-10-10T01:44:36.5802512Z [Gloo] Rank 1 is connected to 1 peer ranks. Expected number of connected peer ranks is : 1 2025-10-10T01:44:36.5803793Z [Gloo] Rank 0 is connected to 1 peer ranks. Expected number of connected peer ranks is : 1 2025-10-10T01:44:36.5850699Z [Gloo] Rank 0 is connected to 1 peer ranks. Expected number of connected peer ranks is : 1 2025-10-10T01:44:36.5851362Z [Gloo] Rank 1 is connected to 1 peer ranks. Expected number of connected peer ranks is : 1 2025-10-10T01:44:36.5893408Z INFO 10-10 01:44:36 [__init__.py:1384] Found nccl from library libnccl.so.2 2025-10-10T01:44:36.5894391Z INFO 10-10 01:44:36 [__init__.py:1384] Found nccl from library libnccl.so.2 2025-10-10T01:44:36.5895348Z INFO 10-10 01:44:36 [pynccl.py:104] vLLM is using nccl==2.27.5 2025-10-10T01:44:36.5895959Z INFO 10-10 01:44:36 [pynccl.py:104] vLLM is using nccl==2.27.5 2025-10-10T01:44:36.7613125Z WARNING 10-10 01:44:36 [symm_mem.py:58] SymmMemCommunicator: Device capability 8.9 not supported, communicator is not available. 2025-10-10T01:44:36.7613993Z WARNING 10-10 01:44:36 [symm_mem.py:58] SymmMemCommunicator: Device capability 8.9 not supported, communicator is not available. 2025-10-10T01:44:36.7637862Z INFO 10-10 01:44:36 [custom_all_reduce.py:35] Skipping P2P check and trusting the driver's P2P report. 2025-10-10T01:44:36.7638557Z INFO 10-10 01:44:36 [custom_all_reduce.py:35] Skipping P2P check and trusting the driver's P2P report. 2025-10-10T01:44:36.7639399Z WARNING 10-10 01:44:36 [custom_all_reduce.py:154] Custom allreduce is disabled because your platform lacks GPU P2P capability or P2P test failed. To silence this warning, specify disable_custom_all_reduce=True explicitly. 2025-10-10T01:44:36.7640534Z WARNING 10-10 01:44:36 [custom_all_reduce.py:154] Custom allreduce is disabled because your platform lacks GPU P2P capability or P2P test failed. To silence this warning, specify disable_custom_all_reduce=True explicitly. 2025-10-10T01:44:36.8741102Z INFO 10-10 01:44:36 [shm_broadcast.py:289] vLLM message queue communication handle: Handle(local_reader_ranks=[1], buffer_handle=(1, 4194304, 6, 'psm_e46aa7c6'), local_subscribe_addr='ipc:///tmp/fda1e6c0-4789-455b-a579-365eec56e945', remote_subscribe_addr=None, remote_addr_ipv6=False) 2025-10-10T01:44:36.8769125Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-10-10T01:44:36.8769645Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-10-10T01:44:36.8778358Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-10-10T01:44:36.8787114Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-10-10T01:44:36.9536519Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-10-10T01:44:36.9550111Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-10-10T01:44:36.9558149Z [Gloo] Rank 0 is connected to 1 peer ranks. Expected number of connected peer ranks is : 1 2025-10-10T01:44:36.9558708Z [Gloo] Rank 1 is connected to 1 peer ranks. Expected number of connected peer ranks is : 1 2025-10-10T01:44:36.9561277Z INFO 10-10 01:44:36 [__init__.py:1384] Found nccl from library libnccl.so.2 2025-10-10T01:44:36.9561674Z INFO 10-10 01:44:36 [__init__.py:1384] Found nccl from library libnccl.so.2 2025-10-10T01:44:36.9562258Z INFO 10-10 01:44:36 [pynccl.py:104] vLLM is using nccl==2.27.5 2025-10-10T01:44:36.9564356Z INFO 10-10 01:44:36 [pynccl.py:104] vLLM is using nccl==2.27.5 2025-10-10T01:44:36.9886480Z INFO 10-10 01:44:36 [parallel_state.py:1208] rank 1 in world size 2 is assigned as DP rank 0, PP rank 0, TP rank 1, EP rank 1 2025-10-10T01:44:36.9887999Z INFO 10-10 01:44:36 [parallel_state.py:1208] rank 0 in world size 2 is assigned as DP rank 0, PP rank 0, TP rank 0, EP rank 0 2025-10-10T01:44:37.2271209Z INFO 10-10 01:44:37 [topk_topp_sampler.py:55] Using FlashInfer for top-p & top-k sampling. 2025-10-10T01:44:37.2338799Z INFO 10-10 01:44:37 [topk_topp_sampler.py:55] Using FlashInfer for top-p & top-k sampling. 2025-10-10T01:44:37.2589650Z (Worker_TP0 pid=41185) INFO 10-10 01:44:37 [gpu_model_runner.py:2707] Starting to load model meta-llama/Llama-3.2-1B-Instruct... 2025-10-10T01:44:37.2652240Z (Worker_TP1 pid=41186) INFO 10-10 01:44:37 [gpu_model_runner.py:2707] Starting to load model meta-llama/Llama-3.2-1B-Instruct... 2025-10-10T01:44:37.4794088Z (Worker_TP0 pid=41185) INFO 10-10 01:44:37 [gpu_model_runner.py:2739] Loading model from scratch... 2025-10-10T01:44:37.4974335Z (Worker_TP1 pid=41186) INFO 10-10 01:44:37 [gpu_model_runner.py:2739] Loading model from scratch... 2025-10-10T01:44:37.5693132Z (Worker_TP0 pid=41185) INFO 10-10 01:44:37 [cuda.py:361] Using Flash Attention backend on V1 engine. 2025-10-10T01:44:37.5876588Z (Worker_TP1 pid=41186) INFO 10-10 01:44:37 [cuda.py:361] Using Flash Attention backend on V1 engine. 2025-10-10T01:44:38.0924135Z (Worker_TP0 pid=41185) INFO 10-10 01:44:38 [gpu_model_runner.py:2758] Model loading took 1.1667 GiB and 0.167343 seconds 2025-10-10T01:44:38.1221846Z (Worker_TP1 pid=41186) INFO 10-10 01:44:38 [gpu_model_runner.py:2758] Model loading took 1.1667 GiB and 0.167296 seconds 2025-10-10T01:44:41.3064918Z (Worker_TP1 pid=41186) INFO 10-10 01:44:41 [backends.py:548] Using cache directory: /var/lib/jenkins/.cache/vllm/torch_compile_cache/31e0bc77eb/rank_1_0/backbone for vLLM's torch.compile 2025-10-10T01:44:41.3075843Z (Worker_TP1 pid=41186) INFO 10-10 01:44:41 [backends.py:559] Dynamo bytecode transform time: 2.89 s 2025-10-10T01:44:41.3504147Z (Worker_TP0 pid=41185) INFO 10-10 01:44:41 [backends.py:548] Using cache directory: /var/lib/jenkins/.cache/vllm/torch_compile_cache/31e0bc77eb/rank_0_0/backbone for vLLM's torch.compile 2025-10-10T01:44:41.3514700Z (Worker_TP0 pid=41185) INFO 10-10 01:44:41 [backends.py:559] Dynamo bytecode transform time: 2.94 s 2025-10-10T01:44:42.4056434Z (Worker_TP1 pid=41186) INFO 10-10 01:44:42 [backends.py:164] Directly load the compiled graph(s) for dynamic shape from the cache, took 0.838 s 2025-10-10T01:44:42.4232950Z (Worker_TP0 pid=41185) INFO 10-10 01:44:42 [backends.py:164] Directly load the compiled graph(s) for dynamic shape from the cache, took 0.817 s 2025-10-10T01:44:43.6574668Z (Worker_TP0 pid=41185) INFO 10-10 01:44:43 [gpu_worker.py:298] Available KV cache memory: 18.54 GiB 2025-10-10T01:44:43.6926388Z (Worker_TP1 pid=41186) INFO 10-10 01:44:43 [gpu_worker.py:298] Available KV cache memory: 18.54 GiB 2025-10-10T01:44:43.9713582Z (EngineCore_DP0 pid=41066) INFO 10-10 01:44:43 [kv_cache_utils.py:1087] GPU KV cache size: 1,215,232 tokens 2025-10-10T01:44:43.9714510Z (EngineCore_DP0 pid=41066) INFO 10-10 01:44:43 [kv_cache_utils.py:1091] Maximum concurrency for 2,048 tokens per request: 593.38x 2025-10-10T01:44:43.9715397Z (EngineCore_DP0 pid=41066) INFO 10-10 01:44:43 [kv_cache_utils.py:1087] GPU KV cache size: 1,215,232 tokens 2025-10-10T01:44:43.9716273Z (EngineCore_DP0 pid=41066) INFO 10-10 01:44:43 [kv_cache_utils.py:1091] Maximum concurrency for 2,048 tokens per request: 593.38x 2025-10-10T01:44:43.9782684Z (Worker_TP0 pid=41185) INFO 10-10 01:44:43 [gpu_worker.py:332] Compile and warming up model for size 8 2025-10-10T01:44:43.9783501Z (Worker_TP1 pid=41186) INFO 10-10 01:44:43 [gpu_worker.py:332] Compile and warming up model for size 8 2025-10-10T01:44:44.4915629Z (Worker_TP1 pid=41186) INFO 10-10 01:44:44 [backends.py:168] Directly load the compiled graph(s) for shape 8 from the cache, took 0.513 s 2025-10-10T01:44:44.4969829Z (Worker_TP0 pid=41185) INFO 10-10 01:44:44 [backends.py:168] Directly load the compiled graph(s) for shape 8 from the cache, took 0.518 s 2025-10-10T01:44:44.4975907Z (Worker_TP1 pid=41186) INFO 10-10 01:44:44 [gpu_worker.py:332] Compile and warming up model for size 4 2025-10-10T01:44:44.4977079Z (Worker_TP0 pid=41185) INFO 10-10 01:44:44 [gpu_worker.py:332] Compile and warming up model for size 4 2025-10-10T01:44:45.1565060Z (Worker_TP0 pid=41185) INFO 10-10 01:44:45 [backends.py:168] Directly load the compiled graph(s) for shape 4 from the cache, took 0.658 s 2025-10-10T01:44:45.1565991Z (Worker_TP0 pid=41185) INFO 10-10 01:44:45 [monitor.py:32] torch.compile takes 2.94 s in total 2025-10-10T01:44:45.1566897Z (Worker_TP1 pid=41186) INFO 10-10 01:44:45 [backends.py:168] Directly load the compiled graph(s) for shape 4 from the cache, took 0.659 s 2025-10-10T01:44:45.1567713Z (Worker_TP1 pid=41186) INFO 10-10 01:44:45 [monitor.py:32] torch.compile takes 2.89 s in total 2025-10-10T01:44:45.1689862Z (EngineCore_DP0 pid=41066) INFO 10-10 01:44:45 [core.py:211] init engine (profile, create kv cache, warmup model) took 6.97 seconds 2025-10-10T01:44:45.6936061Z (EngineCore_DP0 pid=41066) INFO 10-10 01:44:45 [vllm.py:345] Cudagraph is disabled under eager mode 2025-10-10T01:44:45.8784895Z (APIServer pid=40944) INFO 10-10 01:44:45 [loggers.py:147] Engine 000: vllm cache_config_info with initialization after num_gpu_blocks is: 75952 2025-10-10T01:44:45.9163631Z (EngineCore_DP0 pid=41066) INFO 10-10 01:44:45 [gc_utils.py:41] GC Debug Config. enabled:False,top_objects:-1 2025-10-10T01:44:45.9380386Z (APIServer pid=40944) INFO 10-10 01:44:45 [api_server.py:1634] Supported_tasks: ['generate'] 2025-10-10T01:44:45.9611333Z (APIServer pid=40944) WARNING 10-10 01:44:45 [model.py:1424] Default sampling parameters have been overridden by the model's Hugging Face generation config recommended from the model creator. If this is not intended, please relaunch vLLM instance with `--generation-config vllm`. 2025-10-10T01:44:45.9612953Z (APIServer pid=40944) INFO 10-10 01:44:45 [serving_responses.py:140] Using default chat sampling params from model: {'temperature': 0.6, 'top_p': 0.9} 2025-10-10T01:44:45.9859288Z (APIServer pid=40944) INFO 10-10 01:44:45 [serving_chat.py:139] Using default chat sampling params from model: {'temperature': 0.6, 'top_p': 0.9} 2025-10-10T01:44:46.0169395Z (APIServer pid=40944) INFO 10-10 01:44:46 [serving_completion.py:76] Using default completion sampling params from model: {'temperature': 0.6, 'top_p': 0.9} 2025-10-10T01:44:46.0170745Z (APIServer pid=40944) INFO 10-10 01:44:46 [api_server.py:1912] Starting vLLM API server 0 on http://0.0.0.0:58147 2025-10-10T01:44:46.0171490Z (APIServer pid=40944) INFO 10-10 01:44:46 [launcher.py:34] Available routes are: 2025-10-10T01:44:46.0172165Z (APIServer pid=40944) INFO 10-10 01:44:46 [launcher.py:42] Route: /openapi.json, Methods: HEAD, GET 2025-10-10T01:44:46.0172861Z (APIServer pid=40944) INFO 10-10 01:44:46 [launcher.py:42] Route: /docs, Methods: HEAD, GET 2025-10-10T01:44:46.0173575Z (APIServer pid=40944) INFO 10-10 01:44:46 [launcher.py:42] Route: /docs/oauth2-redirect, Methods: HEAD, GET 2025-10-10T01:44:46.0174290Z (APIServer pid=40944) INFO 10-10 01:44:46 [launcher.py:42] Route: /redoc, Methods: HEAD, GET 2025-10-10T01:44:46.0174946Z (APIServer pid=40944) INFO 10-10 01:44:46 [launcher.py:42] Route: /health, Methods: GET 2025-10-10T01:44:46.0175810Z (APIServer pid=40944) INFO 10-10 01:44:46 [launcher.py:42] Route: /load, Methods: GET 2025-10-10T01:44:46.0176494Z (APIServer pid=40944) INFO 10-10 01:44:46 [launcher.py:42] Route: /ping, Methods: POST 2025-10-10T01:44:46.0177169Z (APIServer pid=40944) INFO 10-10 01:44:46 [launcher.py:42] Route: /ping, Methods: GET 2025-10-10T01:44:46.0177771Z (APIServer pid=40944) INFO 10-10 01:44:46 [launcher.py:42] Route: /tokenize, Methods: POST 2025-10-10T01:44:46.0178317Z (APIServer pid=40944) INFO 10-10 01:44:46 [launcher.py:42] Route: /detokenize, Methods: POST 2025-10-10T01:44:46.0178871Z (APIServer pid=40944) INFO 10-10 01:44:46 [launcher.py:42] Route: /v1/models, Methods: GET 2025-10-10T01:44:46.0179391Z (APIServer pid=40944) INFO 10-10 01:44:46 [launcher.py:42] Route: /version, Methods: GET 2025-10-10T01:44:46.0179934Z (APIServer pid=40944) INFO 10-10 01:44:46 [launcher.py:42] Route: /v1/responses, Methods: POST 2025-10-10T01:44:46.0180529Z (APIServer pid=40944) INFO 10-10 01:44:46 [launcher.py:42] Route: /v1/responses/{response_id}, Methods: GET 2025-10-10T01:44:46.0181167Z (APIServer pid=40944) INFO 10-10 01:44:46 [launcher.py:42] Route: /v1/responses/{response_id}/cancel, Methods: POST 2025-10-10T01:44:46.0181796Z (APIServer pid=40944) INFO 10-10 01:44:46 [launcher.py:42] Route: /v1/chat/completions, Methods: POST 2025-10-10T01:44:46.0182377Z (APIServer pid=40944) INFO 10-10 01:44:46 [launcher.py:42] Route: /v1/completions, Methods: POST 2025-10-10T01:44:46.0182936Z (APIServer pid=40944) INFO 10-10 01:44:46 [launcher.py:42] Route: /v1/embeddings, Methods: POST 2025-10-10T01:44:46.0183476Z (APIServer pid=40944) INFO 10-10 01:44:46 [launcher.py:42] Route: /pooling, Methods: POST 2025-10-10T01:44:46.0184009Z (APIServer pid=40944) INFO 10-10 01:44:46 [launcher.py:42] Route: /classify, Methods: POST 2025-10-10T01:44:46.0184535Z (APIServer pid=40944) INFO 10-10 01:44:46 [launcher.py:42] Route: /score, Methods: POST 2025-10-10T01:44:46.0185065Z (APIServer pid=40944) INFO 10-10 01:44:46 [launcher.py:42] Route: /v1/score, Methods: POST 2025-10-10T01:44:46.0185644Z (APIServer pid=40944) INFO 10-10 01:44:46 [launcher.py:42] Route: /v1/audio/transcriptions, Methods: POST 2025-10-10T01:44:46.0186258Z (APIServer pid=40944) INFO 10-10 01:44:46 [launcher.py:42] Route: /v1/audio/translations, Methods: POST 2025-10-10T01:44:46.0186816Z (APIServer pid=40944) INFO 10-10 01:44:46 [launcher.py:42] Route: /rerank, Methods: POST 2025-10-10T01:44:46.0187350Z (APIServer pid=40944) INFO 10-10 01:44:46 [launcher.py:42] Route: /v1/rerank, Methods: POST 2025-10-10T01:44:46.0187880Z (APIServer pid=40944) INFO 10-10 01:44:46 [launcher.py:42] Route: /v2/rerank, Methods: POST 2025-10-10T01:44:46.0188434Z (APIServer pid=40944) INFO 10-10 01:44:46 [launcher.py:42] Route: /scale_elastic_ep, Methods: POST 2025-10-10T01:44:46.0189148Z (APIServer pid=40944) INFO 10-10 01:44:46 [launcher.py:42] Route: /is_scaling_elastic_ep, Methods: POST 2025-10-10T01:44:46.0189715Z (APIServer pid=40944) INFO 10-10 01:44:46 [launcher.py:42] Route: /invocations, Methods: POST 2025-10-10T01:44:46.0190243Z (APIServer pid=40944) INFO 10-10 01:44:46 [launcher.py:42] Route: /metrics, Methods: GET 2025-10-10T01:44:46.0324384Z (APIServer pid=40944) INFO: Started server process [40944] 2025-10-10T01:44:46.0325075Z (APIServer pid=40944) INFO: Waiting for application startup. 2025-10-10T01:44:46.2668850Z (APIServer pid=40944) INFO: Application startup complete. 2025-10-10T01:44:46.3662878Z (APIServer pid=40944) INFO: 127.0.0.1:39624 - "GET /health HTTP/1.1" 200 OK 2025-10-10T01:44:46.5862818Z (APIServer pid=40944) INFO: 127.0.0.1:39630 - "GET /v1/models HTTP/1.1" 200 OK 2025-10-10T01:44:46.6926201Z (APIServer pid=40944) INFO: 127.0.0.1:39630 - "POST /v1/completions HTTP/1.1" 200 OK 2025-10-10T01:44:46.7483245Z (APIServer pid=40944) INFO: 127.0.0.1:39630 - "POST /v1/completions HTTP/1.1" 200 OK 2025-10-10T01:44:46.8741428Z (APIServer pid=40944) INFO: 127.0.0.1:39630 - "POST /v1/completions HTTP/1.1" 200 OK 2025-10-10T01:44:46.9537242Z (APIServer pid=40944) INFO: 127.0.0.1:39630 - "POST /v1/completions HTTP/1.1" 200 OK 2025-10-10T01:44:47.0202144Z (APIServer pid=40944) INFO: 127.0.0.1:39630 - "POST /v1/completions HTTP/1.1" 200 OK 2025-10-10T01:44:47.0276124Z (APIServer pid=40944) INFO: 127.0.0.1:39630 - "POST /v1/completions HTTP/1.1" 200 OK 2025-10-10T01:44:47.0866334Z (APIServer pid=40944) INFO 10-10 01:44:47 [launcher.py:99] Shutting down FastAPI HTTP server. 2025-10-10T01:44:47.0869958Z (Worker_TP0 pid=41185) INFO 10-10 01:44:47 [multiproc_executor.py:558] Parent process exited, terminating worker 2025-10-10T01:44:47.0870967Z (Worker_TP0 pid=41185) INFO 10-10 01:44:47 [multiproc_executor.py:599] WorkerProc shutting down. 2025-10-10T01:44:47.0874463Z (Worker_TP1 pid=41186) INFO 10-10 01:44:47 [multiproc_executor.py:558] Parent process exited, terminating worker 2025-10-10T01:44:47.0875836Z (Worker_TP1 pid=41186) INFO 10-10 01:44:47 [multiproc_executor.py:599] WorkerProc shutting down. 2025-10-10T01:44:49.1079969Z (APIServer pid=40944) INFO: Shutting down 2025-10-10T01:44:49.2085450Z (APIServer pid=40944) INFO: Waiting for application shutdown. 2025-10-10T01:44:49.2086749Z (APIServer pid=40944) INFO: Application shutdown complete. 2025-10-10T01:44:49.5667259Z INFO 10-10 01:44:49 [model.py:551] Resolved architecture: LlamaForCausalLM 2025-10-10T01:44:49.5667803Z WARNING 10-10 01:44:49 [model.py:1768] Casting torch.bfloat16 to torch.float16. 2025-10-10T01:44:49.5668272Z INFO 10-10 01:44:49 [model.py:1545] Using max model len 2048 2025-10-10T01:44:49.5669556Z Launching RemoteOpenAIServer with: vllm serve meta-llama/Llama-3.2-1B-Instruct --dtype float16 --max-model-len 2048 --max-num-seqs 8 --enable-chunked-prefill --enforce-eager --tokenizer-mode auto --tensor-parallel-size 2 --distributed-executor-backend mp --load-format dummy --port 51117 --seed 0 2025-10-10T01:44:50.0806428Z /opt/conda/envs/py_3.12/lib/python3.12/site-packages/torch/cuda/__init__.py:64: FutureWarning: The pynvml package is deprecated. Please install nvidia-ml-py instead. If you did not install pynvml directly, please report this to the maintainers of the package that installed pynvml for you. 2025-10-10T01:44:50.0807651Z import pynvml # type: ignore[import] 2025-10-10T01:44:52.3272704Z INFO 10-10 01:44:52 [__init__.py:215] Automatically detected platform cuda. 2025-10-10T01:44:56.2760825Z (APIServer pid=41391) INFO 10-10 01:44:56 [api_server.py:1839] vLLM API server version 0.11.0rc2.dev157+g0ad9951c4.d20251010 2025-10-10T01:44:56.2780528Z (APIServer pid=41391) INFO 10-10 01:44:56 [utils.py:233] non-default args: {'model_tag': 'meta-llama/Llama-3.2-1B-Instruct', 'port': 51117, 'model': 'meta-llama/Llama-3.2-1B-Instruct', 'dtype': 'float16', 'seed': 0, 'max_model_len': 2048, 'enforce_eager': True, 'load_format': 'dummy', 'distributed_executor_backend': 'mp', 'tensor_parallel_size': 2, 'max_num_seqs': 8, 'enable_chunked_prefill': True} 2025-10-10T01:44:56.5636740Z (APIServer pid=41391) INFO 10-10 01:44:56 [model.py:551] Resolved architecture: LlamaForCausalLM 2025-10-10T01:44:56.5637455Z (APIServer pid=41391) `torch_dtype` is deprecated! Use `dtype` instead! 2025-10-10T01:44:56.5638554Z (APIServer pid=41391) WARNING 10-10 01:44:56 [model.py:1768] Casting torch.bfloat16 to torch.float16. 2025-10-10T01:44:56.5639312Z (APIServer pid=41391) INFO 10-10 01:44:56 [model.py:1545] Using max model len 2048 2025-10-10T01:44:56.7199147Z (APIServer pid=41391) INFO 10-10 01:44:56 [scheduler.py:222] Chunked prefill is enabled with max_num_batched_tokens=2048. 2025-10-10T01:44:56.7204003Z (APIServer pid=41391) INFO 10-10 01:44:56 [vllm.py:345] Cudagraph is disabled under eager mode 2025-10-10T01:44:57.6811740Z /opt/conda/envs/py_3.12/lib/python3.12/site-packages/torch/cuda/__init__.py:64: FutureWarning: The pynvml package is deprecated. Please install nvidia-ml-py instead. If you did not install pynvml directly, please report this to the maintainers of the package that installed pynvml for you. 2025-10-10T01:44:57.6813000Z import pynvml # type: ignore[import] 2025-10-10T01:44:59.9711596Z INFO 10-10 01:44:59 [__init__.py:215] Automatically detected platform cuda. 2025-10-10T01:45:03.9188835Z (EngineCore_DP0 pid=41513) INFO 10-10 01:45:03 [core.py:648] Waiting for init message from front-end. 2025-10-10T01:45:03.9456449Z (EngineCore_DP0 pid=41513) INFO 10-10 01:45:03 [core.py:78] Initializing a V1 LLM engine (v0.11.0rc2.dev157+g0ad9951c4.d20251010) with config: model='meta-llama/Llama-3.2-1B-Instruct', speculative_config=None, tokenizer='meta-llama/Llama-3.2-1B-Instruct', skip_tokenizer_init=False, tokenizer_mode=auto, revision=None, tokenizer_revision=None, trust_remote_code=False, dtype=torch.float16, max_seq_len=2048, download_dir=None, load_format=dummy, tensor_parallel_size=2, pipeline_parallel_size=1, data_parallel_size=1, disable_custom_all_reduce=False, quantization=None, enforce_eager=True, kv_cache_dtype=auto, device_config=cuda, structured_outputs_config=StructuredOutputsConfig(backend='auto', disable_fallback=False, disable_any_whitespace=False, disable_additional_properties=False, reasoning_parser=''), observability_config=ObservabilityConfig(show_hidden_metrics_for_version=None, otlp_traces_endpoint=None, collect_detailed_traces=None), seed=0, served_model_name=meta-llama/Llama-3.2-1B-Instruct, enable_prefix_caching=True, chunked_prefill_enabled=True, pooler_config=None, compilation_config={'level': 0, 'debug_dump_path': None, 'cache_dir': '', 'backend': '', 'custom_ops': [], 'splitting_ops': None, 'use_inductor': True, 'compile_sizes': [], 'inductor_compile_config': {'enable_auto_functionalized_v2': False}, 'inductor_passes': {}, 'cudagraph_mode': , 'use_cudagraph': False, 'cudagraph_num_of_warmups': 0, 'cudagraph_capture_sizes': [], 'cudagraph_copy_inputs': False, 'full_cuda_graph': False, 'use_inductor_graph_partition': False, 'pass_config': {}, 'max_capture_size': 0, 'local_cache_dir': None} 2025-10-10T01:45:03.9461673Z (EngineCore_DP0 pid=41513) WARNING 10-10 01:45:03 [multiproc_executor.py:720] Reducing Torch parallelism from 24 threads to 1 to avoid unnecessary CPU contention. Set OMP_NUM_THREADS in the external environment to tune this value as needed. 2025-10-10T01:45:03.9467465Z (EngineCore_DP0 pid=41513) INFO 10-10 01:45:03 [shm_broadcast.py:289] vLLM message queue communication handle: Handle(local_reader_ranks=[0, 1], buffer_handle=(2, 16777216, 10, 'psm_6d62985e'), local_subscribe_addr='ipc:///tmp/6036564f-626d-43b0-89fd-9eaac6726de3', remote_subscribe_addr=None, remote_addr_ipv6=False) 2025-10-10T01:45:04.4239564Z /opt/conda/envs/py_3.12/lib/python3.12/site-packages/torch/cuda/__init__.py:64: FutureWarning: The pynvml package is deprecated. Please install nvidia-ml-py instead. If you did not install pynvml directly, please report this to the maintainers of the package that installed pynvml for you. 2025-10-10T01:45:04.4241870Z import pynvml # type: ignore[import] 2025-10-10T01:45:04.4383650Z /opt/conda/envs/py_3.12/lib/python3.12/site-packages/torch/cuda/__init__.py:64: FutureWarning: The pynvml package is deprecated. Please install nvidia-ml-py instead. If you did not install pynvml directly, please report this to the maintainers of the package that installed pynvml for you. 2025-10-10T01:45:04.4385725Z import pynvml # type: ignore[import] 2025-10-10T01:45:06.7298239Z INFO 10-10 01:45:06 [__init__.py:215] Automatically detected platform cuda. 2025-10-10T01:45:06.7862793Z INFO 10-10 01:45:06 [__init__.py:215] Automatically detected platform cuda. 2025-10-10T01:45:11.2843463Z INFO 10-10 01:45:11 [shm_broadcast.py:289] vLLM message queue communication handle: Handle(local_reader_ranks=[0], buffer_handle=(1, 10485760, 10, 'psm_479bf19f'), local_subscribe_addr='ipc:///tmp/21a4f7b6-6f06-49f0-b987-eeb04c3bb04d', remote_subscribe_addr=None, remote_addr_ipv6=False) 2025-10-10T01:45:11.3642223Z INFO 10-10 01:45:11 [shm_broadcast.py:289] vLLM message queue communication handle: Handle(local_reader_ranks=[0], buffer_handle=(1, 10485760, 10, 'psm_d9ced4a2'), local_subscribe_addr='ipc:///tmp/f6dc4b3e-135c-495d-b0c6-c108eeff01e0', remote_subscribe_addr=None, remote_addr_ipv6=False) 2025-10-10T01:45:11.8470290Z [Gloo] Rank 1 is connected to 1 peer ranks. Expected number of connected peer ranks is : 1 2025-10-10T01:45:11.8471514Z [Gloo] Rank 0 is connected to 1 peer ranks. Expected number of connected peer ranks is : 1 2025-10-10T01:45:11.8528443Z [Gloo] Rank 1 is connected to 1 peer ranks. Expected number of connected peer ranks is : 1 2025-10-10T01:45:11.8528968Z [Gloo] Rank 0 is connected to 1 peer ranks. Expected number of connected peer ranks is : 1 2025-10-10T01:45:11.8573080Z INFO 10-10 01:45:11 [__init__.py:1384] Found nccl from library libnccl.so.2 2025-10-10T01:45:11.8573567Z INFO 10-10 01:45:11 [__init__.py:1384] Found nccl from library libnccl.so.2 2025-10-10T01:45:11.8576344Z INFO 10-10 01:45:11 [pynccl.py:104] vLLM is using nccl==2.27.5 2025-10-10T01:45:11.8576736Z INFO 10-10 01:45:11 [pynccl.py:104] vLLM is using nccl==2.27.5 2025-10-10T01:45:12.0447286Z WARNING 10-10 01:45:12 [symm_mem.py:58] SymmMemCommunicator: Device capability 8.9 not supported, communicator is not available. 2025-10-10T01:45:12.0449050Z WARNING 10-10 01:45:12 [symm_mem.py:58] SymmMemCommunicator: Device capability 8.9 not supported, communicator is not available. 2025-10-10T01:45:12.0472440Z INFO 10-10 01:45:12 [custom_all_reduce.py:35] Skipping P2P check and trusting the driver's P2P report. 2025-10-10T01:45:12.0473717Z INFO 10-10 01:45:12 [custom_all_reduce.py:35] Skipping P2P check and trusting the driver's P2P report. 2025-10-10T01:45:12.0475651Z WARNING 10-10 01:45:12 [custom_all_reduce.py:154] Custom allreduce is disabled because your platform lacks GPU P2P capability or P2P test failed. To silence this warning, specify disable_custom_all_reduce=True explicitly. 2025-10-10T01:45:12.0478067Z WARNING 10-10 01:45:12 [custom_all_reduce.py:154] Custom allreduce is disabled because your platform lacks GPU P2P capability or P2P test failed. To silence this warning, specify disable_custom_all_reduce=True explicitly. 2025-10-10T01:45:12.1240601Z INFO 10-10 01:45:12 [shm_broadcast.py:289] vLLM message queue communication handle: Handle(local_reader_ranks=[1], buffer_handle=(1, 4194304, 6, 'psm_7aac9350'), local_subscribe_addr='ipc:///tmp/1e495749-dac6-4cf6-924d-dac965350274', remote_subscribe_addr=None, remote_addr_ipv6=False) 2025-10-10T01:45:12.1272840Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-10-10T01:45:12.1273452Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-10-10T01:45:12.1281236Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-10-10T01:45:12.1290322Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-10-10T01:45:12.1635594Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-10-10T01:45:12.1646336Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-10-10T01:45:12.1657487Z [Gloo] Rank 1 is connected to 1 peer ranks. Expected number of connected peer ranks is : 1 2025-10-10T01:45:12.1658026Z [Gloo] Rank 0 is connected to 1 peer ranks. Expected number of connected peer ranks is : 1 2025-10-10T01:45:12.1659723Z INFO 10-10 01:45:12 [__init__.py:1384] Found nccl from library libnccl.so.2 2025-10-10T01:45:12.1660124Z INFO 10-10 01:45:12 [__init__.py:1384] Found nccl from library libnccl.so.2 2025-10-10T01:45:12.1660470Z INFO 10-10 01:45:12 [pynccl.py:104] vLLM is using nccl==2.27.5 2025-10-10T01:45:12.1660800Z INFO 10-10 01:45:12 [pynccl.py:104] vLLM is using nccl==2.27.5 2025-10-10T01:45:12.1996427Z INFO 10-10 01:45:12 [parallel_state.py:1208] rank 0 in world size 2 is assigned as DP rank 0, PP rank 0, TP rank 0, EP rank 0 2025-10-10T01:45:12.1997743Z INFO 10-10 01:45:12 [parallel_state.py:1208] rank 1 in world size 2 is assigned as DP rank 0, PP rank 0, TP rank 1, EP rank 1 2025-10-10T01:45:12.4576019Z INFO 10-10 01:45:12 [topk_topp_sampler.py:55] Using FlashInfer for top-p & top-k sampling. 2025-10-10T01:45:12.4648186Z INFO 10-10 01:45:12 [topk_topp_sampler.py:55] Using FlashInfer for top-p & top-k sampling. 2025-10-10T01:45:12.4894566Z (Worker_TP0 pid=41632) INFO 10-10 01:45:12 [gpu_model_runner.py:2707] Starting to load model meta-llama/Llama-3.2-1B-Instruct... 2025-10-10T01:45:12.4962213Z (Worker_TP1 pid=41633) INFO 10-10 01:45:12 [gpu_model_runner.py:2707] Starting to load model meta-llama/Llama-3.2-1B-Instruct... 2025-10-10T01:45:12.7185208Z (Worker_TP0 pid=41632) INFO 10-10 01:45:12 [gpu_model_runner.py:2739] Loading model from scratch... 2025-10-10T01:45:12.7270927Z (Worker_TP1 pid=41633) INFO 10-10 01:45:12 [gpu_model_runner.py:2739] Loading model from scratch... 2025-10-10T01:45:12.8083694Z (Worker_TP0 pid=41632) INFO 10-10 01:45:12 [cuda.py:361] Using Flash Attention backend on V1 engine. 2025-10-10T01:45:12.8166515Z (Worker_TP1 pid=41633) INFO 10-10 01:45:12 [cuda.py:361] Using Flash Attention backend on V1 engine. 2025-10-10T01:45:13.2901897Z (Worker_TP1 pid=41633) INFO 10-10 01:45:13 [gpu_model_runner.py:2758] Model loading took 1.1667 GiB and 0.118998 seconds 2025-10-10T01:45:13.2962891Z (Worker_TP0 pid=41632) INFO 10-10 01:45:13 [gpu_model_runner.py:2758] Model loading took 1.1667 GiB and 0.119061 seconds 2025-10-10T01:45:18.2515123Z (Worker_TP1 pid=41633) INFO 10-10 01:45:18 [gpu_worker.py:298] Available KV cache memory: 18.53 GiB 2025-10-10T01:45:18.2621239Z (Worker_TP0 pid=41632) INFO 10-10 01:45:18 [gpu_worker.py:298] Available KV cache memory: 18.53 GiB 2025-10-10T01:45:18.4900267Z (EngineCore_DP0 pid=41513) INFO 10-10 01:45:18 [kv_cache_utils.py:1087] GPU KV cache size: 1,214,464 tokens 2025-10-10T01:45:18.4901233Z (EngineCore_DP0 pid=41513) INFO 10-10 01:45:18 [kv_cache_utils.py:1091] Maximum concurrency for 2,048 tokens per request: 593.00x 2025-10-10T01:45:18.4902122Z (EngineCore_DP0 pid=41513) INFO 10-10 01:45:18 [kv_cache_utils.py:1087] GPU KV cache size: 1,214,464 tokens 2025-10-10T01:45:18.4903419Z (EngineCore_DP0 pid=41513) INFO 10-10 01:45:18 [kv_cache_utils.py:1091] Maximum concurrency for 2,048 tokens per request: 593.00x 2025-10-10T01:45:18.6133378Z (EngineCore_DP0 pid=41513) INFO 10-10 01:45:18 [core.py:211] init engine (profile, create kv cache, warmup model) took 5.23 seconds 2025-10-10T01:45:19.1166534Z (EngineCore_DP0 pid=41513) INFO 10-10 01:45:19 [vllm.py:345] Cudagraph is disabled under eager mode 2025-10-10T01:45:19.3012707Z (APIServer pid=41391) INFO 10-10 01:45:19 [loggers.py:147] Engine 000: vllm cache_config_info with initialization after num_gpu_blocks is: 75904 2025-10-10T01:45:19.3306615Z (EngineCore_DP0 pid=41513) INFO 10-10 01:45:19 [gc_utils.py:41] GC Debug Config. enabled:False,top_objects:-1 2025-10-10T01:45:19.3524882Z (APIServer pid=41391) INFO 10-10 01:45:19 [api_server.py:1634] Supported_tasks: ['generate'] 2025-10-10T01:45:19.3761205Z (APIServer pid=41391) WARNING 10-10 01:45:19 [model.py:1424] Default sampling parameters have been overridden by the model's Hugging Face generation config recommended from the model creator. If this is not intended, please relaunch vLLM instance with `--generation-config vllm`. 2025-10-10T01:45:19.3762846Z (APIServer pid=41391) INFO 10-10 01:45:19 [serving_responses.py:140] Using default chat sampling params from model: {'temperature': 0.6, 'top_p': 0.9} 2025-10-10T01:45:19.3989609Z (APIServer pid=41391) INFO 10-10 01:45:19 [serving_chat.py:139] Using default chat sampling params from model: {'temperature': 0.6, 'top_p': 0.9} 2025-10-10T01:45:19.4194419Z (APIServer pid=41391) INFO 10-10 01:45:19 [serving_completion.py:76] Using default completion sampling params from model: {'temperature': 0.6, 'top_p': 0.9} 2025-10-10T01:45:19.4195503Z (APIServer pid=41391) INFO 10-10 01:45:19 [api_server.py:1912] Starting vLLM API server 0 on http://0.0.0.0:51117 2025-10-10T01:45:19.4196485Z (APIServer pid=41391) INFO 10-10 01:45:19 [launcher.py:34] Available routes are: 2025-10-10T01:45:19.4197369Z (APIServer pid=41391) INFO 10-10 01:45:19 [launcher.py:42] Route: /openapi.json, Methods: HEAD, GET 2025-10-10T01:45:19.4198225Z (APIServer pid=41391) INFO 10-10 01:45:19 [launcher.py:42] Route: /docs, Methods: HEAD, GET 2025-10-10T01:45:19.4199091Z (APIServer pid=41391) INFO 10-10 01:45:19 [launcher.py:42] Route: /docs/oauth2-redirect, Methods: HEAD, GET 2025-10-10T01:45:19.4199834Z (APIServer pid=41391) INFO 10-10 01:45:19 [launcher.py:42] Route: /redoc, Methods: HEAD, GET 2025-10-10T01:45:19.4200376Z (APIServer pid=41391) INFO 10-10 01:45:19 [launcher.py:42] Route: /health, Methods: GET 2025-10-10T01:45:19.4201136Z (APIServer pid=41391) INFO 10-10 01:45:19 [launcher.py:42] Route: /load, Methods: GET 2025-10-10T01:45:19.4201667Z (APIServer pid=41391) INFO 10-10 01:45:19 [launcher.py:42] Route: /ping, Methods: POST 2025-10-10T01:45:19.4202171Z (APIServer pid=41391) INFO 10-10 01:45:19 [launcher.py:42] Route: /ping, Methods: GET 2025-10-10T01:45:19.4202711Z (APIServer pid=41391) INFO 10-10 01:45:19 [launcher.py:42] Route: /tokenize, Methods: POST 2025-10-10T01:45:19.4203238Z (APIServer pid=41391) INFO 10-10 01:45:19 [launcher.py:42] Route: /detokenize, Methods: POST 2025-10-10T01:45:19.4203762Z (APIServer pid=41391) INFO 10-10 01:45:19 [launcher.py:42] Route: /v1/models, Methods: GET 2025-10-10T01:45:19.4204269Z (APIServer pid=41391) INFO 10-10 01:45:19 [launcher.py:42] Route: /version, Methods: GET 2025-10-10T01:45:19.4204800Z (APIServer pid=41391) INFO 10-10 01:45:19 [launcher.py:42] Route: /v1/responses, Methods: POST 2025-10-10T01:45:19.4205378Z (APIServer pid=41391) INFO 10-10 01:45:19 [launcher.py:42] Route: /v1/responses/{response_id}, Methods: GET 2025-10-10T01:45:19.4206003Z (APIServer pid=41391) INFO 10-10 01:45:19 [launcher.py:42] Route: /v1/responses/{response_id}/cancel, Methods: POST 2025-10-10T01:45:19.4206831Z (APIServer pid=41391) INFO 10-10 01:45:19 [launcher.py:42] Route: /v1/chat/completions, Methods: POST 2025-10-10T01:45:19.4207416Z (APIServer pid=41391) INFO 10-10 01:45:19 [launcher.py:42] Route: /v1/completions, Methods: POST 2025-10-10T01:45:19.4207963Z (APIServer pid=41391) INFO 10-10 01:45:19 [launcher.py:42] Route: /v1/embeddings, Methods: POST 2025-10-10T01:45:19.4208491Z (APIServer pid=41391) INFO 10-10 01:45:19 [launcher.py:42] Route: /pooling, Methods: POST 2025-10-10T01:45:19.4209014Z (APIServer pid=41391) INFO 10-10 01:45:19 [launcher.py:42] Route: /classify, Methods: POST 2025-10-10T01:45:19.4209529Z (APIServer pid=41391) INFO 10-10 01:45:19 [launcher.py:42] Route: /score, Methods: POST 2025-10-10T01:45:19.4210155Z (APIServer pid=41391) INFO 10-10 01:45:19 [launcher.py:42] Route: /v1/score, Methods: POST 2025-10-10T01:45:19.4210736Z (APIServer pid=41391) INFO 10-10 01:45:19 [launcher.py:42] Route: /v1/audio/transcriptions, Methods: POST 2025-10-10T01:45:19.4211323Z (APIServer pid=41391) INFO 10-10 01:45:19 [launcher.py:42] Route: /v1/audio/translations, Methods: POST 2025-10-10T01:45:19.4211870Z (APIServer pid=41391) INFO 10-10 01:45:19 [launcher.py:42] Route: /rerank, Methods: POST 2025-10-10T01:45:19.4212389Z (APIServer pid=41391) INFO 10-10 01:45:19 [launcher.py:42] Route: /v1/rerank, Methods: POST 2025-10-10T01:45:19.4212923Z (APIServer pid=41391) INFO 10-10 01:45:19 [launcher.py:42] Route: /v2/rerank, Methods: POST 2025-10-10T01:45:19.4213471Z (APIServer pid=41391) INFO 10-10 01:45:19 [launcher.py:42] Route: /scale_elastic_ep, Methods: POST 2025-10-10T01:45:19.4214046Z (APIServer pid=41391) INFO 10-10 01:45:19 [launcher.py:42] Route: /is_scaling_elastic_ep, Methods: POST 2025-10-10T01:45:19.4214604Z (APIServer pid=41391) INFO 10-10 01:45:19 [launcher.py:42] Route: /invocations, Methods: POST 2025-10-10T01:45:19.4215118Z (APIServer pid=41391) INFO 10-10 01:45:19 [launcher.py:42] Route: /metrics, Methods: GET 2025-10-10T01:45:19.4350572Z (APIServer pid=41391) INFO: Started server process [41391] 2025-10-10T01:45:19.4351228Z (APIServer pid=41391) INFO: Waiting for application startup. 2025-10-10T01:45:19.6762466Z (APIServer pid=41391) INFO: Application startup complete. 2025-10-10T01:45:20.1468316Z (APIServer pid=41391) INFO: 127.0.0.1:34878 - "GET /health HTTP/1.1" 200 OK 2025-10-10T01:45:20.1559414Z (APIServer pid=41391) INFO: 127.0.0.1:34880 - "GET /v1/models HTTP/1.1" 200 OK 2025-10-10T01:45:20.3306299Z (APIServer pid=41391) INFO: 127.0.0.1:34880 - "POST /v1/completions HTTP/1.1" 200 OK 2025-10-10T01:45:20.3884925Z (APIServer pid=41391) INFO: 127.0.0.1:34880 - "POST /v1/completions HTTP/1.1" 200 OK 2025-10-10T01:45:20.4858346Z (APIServer pid=41391) INFO: 127.0.0.1:34880 - "POST /v1/completions HTTP/1.1" 200 OK 2025-10-10T01:45:20.5617588Z (APIServer pid=41391) INFO: 127.0.0.1:34880 - "POST /v1/completions HTTP/1.1" 200 OK 2025-10-10T01:45:20.6334720Z (APIServer pid=41391) INFO: 127.0.0.1:34880 - "POST /v1/completions HTTP/1.1" 200 OK 2025-10-10T01:45:20.6412903Z (APIServer pid=41391) INFO: 127.0.0.1:34880 - "POST /v1/completions HTTP/1.1" 200 OK 2025-10-10T01:45:20.7078222Z (APIServer pid=41391) INFO 10-10 01:45:20 [launcher.py:99] Shutting down FastAPI HTTP server. 2025-10-10T01:45:20.7080704Z (Worker_TP0 pid=41632) INFO 10-10 01:45:20 [multiproc_executor.py:558] Parent process exited, terminating worker 2025-10-10T01:45:20.7081886Z (Worker_TP0 pid=41632) INFO 10-10 01:45:20 [multiproc_executor.py:599] WorkerProc shutting down. 2025-10-10T01:45:20.7084617Z (Worker_TP1 pid=41633) INFO 10-10 01:45:20 [multiproc_executor.py:558] Parent process exited, terminating worker 2025-10-10T01:45:20.7086283Z (Worker_TP1 pid=41633) INFO 10-10 01:45:20 [multiproc_executor.py:599] WorkerProc shutting down. 2025-10-10T01:45:22.6260655Z (APIServer pid=41391) INFO: Shutting down 2025-10-10T01:45:22.7268556Z (APIServer pid=41391) INFO: Waiting for application shutdown. 2025-10-10T01:45:22.7269892Z (APIServer pid=41391) INFO: Application shutdown complete. 2025-10-10T01:45:23.3560341Z PASSED 2025-10-10T01:45:23.3702820Z distributed/test_sequence_parallel.py::test_tp_sp_generation[meta-llama/Llama-3.2-1B-Instruct-parallel_setup11-ray-1-auto-test_options11] Fork a new process to run a test 42150 2025-10-10T01:45:23.3716829Z Fork a new process to run a test 0 2025-10-10T01:45:24.0110765Z INFO 10-10 01:45:24 [model.py:551] Resolved architecture: LlamaForCausalLM 2025-10-10T01:45:24.0111227Z `torch_dtype` is deprecated! Use `dtype` instead! 2025-10-10T01:45:24.0112099Z WARNING 10-10 01:45:24 [model.py:1768] Casting torch.bfloat16 to torch.float16. 2025-10-10T01:45:24.0113058Z INFO 10-10 01:45:24 [model.py:1545] Using max model len 2048 2025-10-10T01:45:24.0118052Z Launching RemoteOpenAIServer with: vllm serve meta-llama/Llama-3.2-1B-Instruct --dtype float16 --max-model-len 2048 --max-num-seqs 8 --enable-chunked-prefill --enforce-eager --tokenizer-mode auto --tensor-parallel-size 2 --pipeline-parallel-size 1 --distributed-executor-backend ray --compilation_config {"level": 3, "custom_ops": ["+rms_norm"], "compile_sizes": [4, 8], "pass_config": {"enable_sequence_parallelism": true, "enable_fusion": false, "enable_noop": true}} --load-format dummy --port 37755 --seed 0 2025-10-10T01:45:24.5245045Z /opt/conda/envs/py_3.12/lib/python3.12/site-packages/torch/cuda/__init__.py:64: FutureWarning: The pynvml package is deprecated. Please install nvidia-ml-py instead. If you did not install pynvml directly, please report this to the maintainers of the package that installed pynvml for you. 2025-10-10T01:45:24.5247369Z import pynvml # type: ignore[import] 2025-10-10T01:45:26.7569415Z INFO 10-10 01:45:26 [__init__.py:215] Automatically detected platform cuda. 2025-10-10T01:45:30.7095009Z (APIServer pid=42199) INFO 10-10 01:45:30 [api_server.py:1839] vLLM API server version 0.11.0rc2.dev157+g0ad9951c4.d20251010 2025-10-10T01:45:30.7122946Z (APIServer pid=42199) INFO 10-10 01:45:30 [utils.py:233] non-default args: {'model_tag': 'meta-llama/Llama-3.2-1B-Instruct', 'port': 37755, 'model': 'meta-llama/Llama-3.2-1B-Instruct', 'dtype': 'float16', 'seed': 0, 'max_model_len': 2048, 'enforce_eager': True, 'load_format': 'dummy', 'distributed_executor_backend': 'ray', 'tensor_parallel_size': 2, 'max_num_seqs': 8, 'enable_chunked_prefill': True, 'compilation_config': {'level': 3, 'debug_dump_path': None, 'cache_dir': '', 'backend': '', 'custom_ops': ['+rms_norm'], 'splitting_ops': None, 'use_inductor': True, 'compile_sizes': [4, 8], 'inductor_compile_config': {'enable_auto_functionalized_v2': False}, 'inductor_passes': {}, 'cudagraph_mode': None, 'use_cudagraph': True, 'cudagraph_num_of_warmups': 0, 'cudagraph_capture_sizes': None, 'cudagraph_copy_inputs': False, 'full_cuda_graph': False, 'use_inductor_graph_partition': False, 'pass_config': {'enable_noop': True, 'enable_sequence_parallelism': True}, 'max_capture_size': None, 'local_cache_dir': None}} 2025-10-10T01:45:30.8536841Z (APIServer pid=42199) INFO 10-10 01:45:30 [model.py:551] Resolved architecture: LlamaForCausalLM 2025-10-10T01:45:30.8537883Z (APIServer pid=42199) `torch_dtype` is deprecated! Use `dtype` instead! 2025-10-10T01:45:30.8539354Z (APIServer pid=42199) WARNING 10-10 01:45:30 [model.py:1768] Casting torch.bfloat16 to torch.float16. 2025-10-10T01:45:30.8540452Z (APIServer pid=42199) INFO 10-10 01:45:30 [model.py:1545] Using max model len 2048 2025-10-10T01:45:31.0146178Z (APIServer pid=42199) INFO 10-10 01:45:31 [scheduler.py:222] Chunked prefill is enabled with max_num_batched_tokens=2048. 2025-10-10T01:45:31.0152336Z (APIServer pid=42199) INFO 10-10 01:45:31 [vllm.py:345] Cudagraph is disabled under eager mode 2025-10-10T01:45:31.9804153Z /opt/conda/envs/py_3.12/lib/python3.12/site-packages/torch/cuda/__init__.py:64: FutureWarning: The pynvml package is deprecated. Please install nvidia-ml-py instead. If you did not install pynvml directly, please report this to the maintainers of the package that installed pynvml for you. 2025-10-10T01:45:31.9805587Z import pynvml # type: ignore[import] 2025-10-10T01:45:34.2800088Z INFO 10-10 01:45:34 [__init__.py:215] Automatically detected platform cuda. 2025-10-10T01:45:38.3673888Z (EngineCore_DP0 pid=42321) INFO 10-10 01:45:38 [core.py:648] Waiting for init message from front-end. 2025-10-10T01:45:38.3949086Z (EngineCore_DP0 pid=42321) INFO 10-10 01:45:38 [core.py:78] Initializing a V1 LLM engine (v0.11.0rc2.dev157+g0ad9951c4.d20251010) with config: model='meta-llama/Llama-3.2-1B-Instruct', speculative_config=None, tokenizer='meta-llama/Llama-3.2-1B-Instruct', skip_tokenizer_init=False, tokenizer_mode=auto, revision=None, tokenizer_revision=None, trust_remote_code=False, dtype=torch.float16, max_seq_len=2048, download_dir=None, load_format=dummy, tensor_parallel_size=2, pipeline_parallel_size=1, data_parallel_size=1, disable_custom_all_reduce=False, quantization=None, enforce_eager=True, kv_cache_dtype=auto, device_config=cuda, structured_outputs_config=StructuredOutputsConfig(backend='auto', disable_fallback=False, disable_any_whitespace=False, disable_additional_properties=False, reasoning_parser=''), observability_config=ObservabilityConfig(show_hidden_metrics_for_version=None, otlp_traces_endpoint=None, collect_detailed_traces=None), seed=0, served_model_name=meta-llama/Llama-3.2-1B-Instruct, enable_prefix_caching=True, chunked_prefill_enabled=True, pooler_config=None, compilation_config={'level': 3, 'debug_dump_path': None, 'cache_dir': '', 'backend': '', 'custom_ops': ['+rms_norm', '+rms_norm'], 'splitting_ops': ['vllm.unified_attention', 'vllm.unified_attention_with_output', 'vllm.mamba_mixer2', 'vllm.mamba_mixer', 'vllm.short_conv', 'vllm.linear_attention', 'vllm.plamo2_mamba_mixer', 'vllm.gdn_attention', 'vllm.sparse_attn_indexer'], 'use_inductor': True, 'compile_sizes': [8, 4], 'inductor_compile_config': {'enable_auto_functionalized_v2': False}, 'inductor_passes': {}, 'cudagraph_mode': , 'use_cudagraph': False, 'cudagraph_num_of_warmups': 0, 'cudagraph_capture_sizes': [], 'cudagraph_copy_inputs': False, 'full_cuda_graph': False, 'use_inductor_graph_partition': False, 'pass_config': {'enable_noop': True, 'enable_sequence_parallelism': True}, 'max_capture_size': 0, 'local_cache_dir': None} 2025-10-10T01:45:40.2959277Z (EngineCore_DP0 pid=42321) 2025-10-10 01:45:40,295 INFO worker.py:1918 -- Started a local Ray instance. View the dashboard at http://127.0.0.1:8265  2025-10-10T01:45:41.3330323Z (EngineCore_DP0 pid=42321) INFO 10-10 01:45:41 [ray_utils.py:345] No current placement group found. Creating a new placement group. 2025-10-10T01:45:41.5044112Z (EngineCore_DP0 pid=42321) INFO 10-10 01:45:41 [ray_distributed_executor.py:171] use_ray_spmd_worker: True 2025-10-10T01:45:42.0961397Z (EngineCore_DP0 pid=42321) (pid=43207) /opt/conda/envs/py_3.12/lib/python3.12/site-packages/torch/cuda/__init__.py:64: FutureWarning: The pynvml package is deprecated. Please install nvidia-ml-py instead. If you did not install pynvml directly, please report this to the maintainers of the package that installed pynvml for you. 2025-10-10T01:45:42.0963466Z (EngineCore_DP0 pid=42321) (pid=43207) import pynvml # type: ignore[import] 2025-10-10T01:45:44.3414278Z (EngineCore_DP0 pid=42321) (pid=43207) INFO 10-10 01:45:44 [__init__.py:215] Automatically detected platform cuda. 2025-10-10T01:45:45.7149854Z (EngineCore_DP0 pid=42321) INFO 10-10 01:45:45 [ray_env.py:63] RAY_NON_CARRY_OVER_ENV_VARS from config: set() 2025-10-10T01:45:45.7151106Z (EngineCore_DP0 pid=42321) INFO 10-10 01:45:45 [ray_env.py:65] Copying the following environment variables to workers: ['VLLM_WORKER_MULTIPROC_METHOD', 'HF_TOKEN', 'VLLM_USE_RAY_COMPILED_DAG', 'MAX_JOBS', 'VLLM_USE_RAY_SPMD_WORKER', 'VLLM_USE_V1', 'HUGGING_FACE_HUB_TOKEN', 'LD_LIBRARY_PATH'] 2025-10-10T01:45:45.7154246Z (EngineCore_DP0 pid=42321) INFO 10-10 01:45:45 [ray_env.py:68] If certain env vars should NOT be copied, add them to /var/lib/jenkins/.config/vllm/ray_non_carry_over_env_vars.json file 2025-10-10T01:45:48.8260728Z (EngineCore_DP0 pid=42321) (RayWorkerWrapper pid=43207) [Gloo] Rank 0 is connected to 1 peer ranks. Expected number of connected peer ranks is : 1 2025-10-10T01:45:48.9313132Z (EngineCore_DP0 pid=42321) (RayWorkerWrapper pid=43207) INFO 10-10 01:45:48 [__init__.py:1384] Found nccl from library libnccl.so.2 2025-10-10T01:45:48.9314610Z (EngineCore_DP0 pid=42321) (RayWorkerWrapper pid=43207) INFO 10-10 01:45:48 [pynccl.py:104] vLLM is using nccl==2.27.5 2025-10-10T01:45:49.1387757Z (EngineCore_DP0 pid=42321) (RayWorkerWrapper pid=43207) WARNING 10-10 01:45:49 [symm_mem.py:58] SymmMemCommunicator: Device capability 8.9 not supported, communicator is not available. 2025-10-10T01:45:49.1389113Z (EngineCore_DP0 pid=42321) (RayWorkerWrapper pid=43207) INFO 10-10 01:45:49 [custom_all_reduce.py:35] Skipping P2P check and trusting the driver's P2P report. 2025-10-10T01:45:49.1390718Z (EngineCore_DP0 pid=42321) (RayWorkerWrapper pid=43207) WARNING 10-10 01:45:49 [custom_all_reduce.py:154] Custom allreduce is disabled because your platform lacks GPU P2P capability or P2P test failed. To silence this warning, specify disable_custom_all_reduce=True explicitly. 2025-10-10T01:45:49.1392631Z (EngineCore_DP0 pid=42321) (RayWorkerWrapper pid=43207) INFO 10-10 01:45:49 [shm_broadcast.py:289] vLLM message queue communication handle: Handle(local_reader_ranks=[1], buffer_handle=(1, 4194304, 6, 'psm_ec37fa3d'), local_subscribe_addr='ipc:///tmp/97d3bdf9-a577-496b-8b4d-ddf61e181ac8', remote_subscribe_addr=None, remote_addr_ipv6=False) 2025-10-10T01:45:49.2443283Z (EngineCore_DP0 pid=42321) (RayWorkerWrapper pid=43207) INFO 10-10 01:45:49 [parallel_state.py:1208] rank 0 in world size 2 is assigned as DP rank 0, PP rank 0, TP rank 0, EP rank 0 2025-10-10T01:45:49.5530533Z (EngineCore_DP0 pid=42321) (RayWorkerWrapper pid=43207) INFO 10-10 01:45:49 [topk_topp_sampler.py:55] Using FlashInfer for top-p & top-k sampling. 2025-10-10T01:45:49.5531821Z (EngineCore_DP0 pid=42321) (RayWorkerWrapper pid=43207) INFO 10-10 01:45:49 [gpu_model_runner.py:2707] Starting to load model meta-llama/Llama-3.2-1B-Instruct... 2025-10-10T01:45:49.5532845Z (EngineCore_DP0 pid=42321) (pid=43212) INFO 10-10 01:45:44 [__init__.py:215] Automatically detected platform cuda. 2025-10-10T01:45:49.8621770Z (EngineCore_DP0 pid=42321) (RayWorkerWrapper pid=43207) INFO 10-10 01:45:49 [gpu_model_runner.py:2739] Loading model from scratch... 2025-10-10T01:45:49.9668067Z (EngineCore_DP0 pid=42321) (RayWorkerWrapper pid=43207) INFO 10-10 01:45:49 [cuda.py:361] Using Flash Attention backend on V1 engine. 2025-10-10T01:45:50.4792282Z (EngineCore_DP0 pid=42321) (RayWorkerWrapper pid=43207) INFO 10-10 01:45:50 [gpu_model_runner.py:2758] Model loading took 1.1667 GiB and 0.167131 seconds 2025-10-10T01:45:53.5384558Z (EngineCore_DP0 pid=42321) (RayWorkerWrapper pid=43212) INFO 10-10 01:45:53 [backends.py:548] Using cache directory: /var/lib/jenkins/.cache/vllm/torch_compile_cache/31e0bc77eb/rank_1_0/backbone for vLLM's torch.compile 2025-10-10T01:45:53.5386007Z (EngineCore_DP0 pid=42321) (RayWorkerWrapper pid=43212) INFO 10-10 01:45:53 [backends.py:559] Dynamo bytecode transform time: 2.82 s 2025-10-10T01:45:54.7671208Z (EngineCore_DP0 pid=42321) (RayWorkerWrapper pid=43212) INFO 10-10 01:45:54 [backends.py:164] Directly load the compiled graph(s) for dynamic shape from the cache, took 0.896 s 2025-10-10T01:45:54.7675482Z (EngineCore_DP0 pid=42321) (RayWorkerWrapper pid=43212) [Gloo] Rank 1 is connected to 1 peer ranks. Expected number of connected peer ranks is : 1 [repeated 11x across cluster] (Ray deduplicates logs by default. Set RAY_DEDUP_LOGS=0 to disable log deduplication, or see https://docs.ray.io/en/master/ray-observability/user-guides/configure-logging.html#log-deduplication for more options.) 2025-10-10T01:45:54.7678688Z (EngineCore_DP0 pid=42321) (RayWorkerWrapper pid=43212) INFO 10-10 01:45:49 [__init__.py:1384] Found nccl from library libnccl.so.2 [repeated 3x across cluster] 2025-10-10T01:45:54.7679919Z (EngineCore_DP0 pid=42321) (RayWorkerWrapper pid=43212) INFO 10-10 01:45:49 [pynccl.py:104] vLLM is using nccl==2.27.5 [repeated 3x across cluster] 2025-10-10T01:45:54.7681181Z (EngineCore_DP0 pid=42321) (RayWorkerWrapper pid=43212) WARNING 10-10 01:45:49 [symm_mem.py:58] SymmMemCommunicator: Device capability 8.9 not supported, communicator is not available. 2025-10-10T01:45:54.7682200Z (EngineCore_DP0 pid=42321) (RayWorkerWrapper pid=43212) INFO 10-10 01:45:49 [custom_all_reduce.py:35] Skipping P2P check and trusting the driver's P2P report. 2025-10-10T01:45:54.7683545Z (EngineCore_DP0 pid=42321) (RayWorkerWrapper pid=43212) WARNING 10-10 01:45:49 [custom_all_reduce.py:154] Custom allreduce is disabled because your platform lacks GPU P2P capability or P2P test failed. To silence this warning, specify disable_custom_all_reduce=True explicitly. 2025-10-10T01:45:54.7684873Z (EngineCore_DP0 pid=42321) (RayWorkerWrapper pid=43212) INFO 10-10 01:45:49 [parallel_state.py:1208] rank 1 in world size 2 is assigned as DP rank 0, PP rank 0, TP rank 1, EP rank 1 2025-10-10T01:45:54.7685846Z (EngineCore_DP0 pid=42321) (RayWorkerWrapper pid=43212) INFO 10-10 01:45:49 [topk_topp_sampler.py:55] Using FlashInfer for top-p & top-k sampling. 2025-10-10T01:45:54.7686774Z (EngineCore_DP0 pid=42321) (RayWorkerWrapper pid=43212) INFO 10-10 01:45:49 [gpu_model_runner.py:2707] Starting to load model meta-llama/Llama-3.2-1B-Instruct... 2025-10-10T01:45:54.8705719Z (EngineCore_DP0 pid=42321) (RayWorkerWrapper pid=43212) INFO 10-10 01:45:49 [gpu_model_runner.py:2739] Loading model from scratch... 2025-10-10T01:45:56.0963572Z (EngineCore_DP0 pid=42321) (RayWorkerWrapper pid=43207) INFO 10-10 01:45:56 [gpu_worker.py:298] Available KV cache memory: 18.54 GiB 2025-10-10T01:45:56.0964510Z (EngineCore_DP0 pid=42321) (RayWorkerWrapper pid=43212) INFO 10-10 01:45:49 [cuda.py:361] Using Flash Attention backend on V1 engine. 2025-10-10T01:45:56.0965371Z (EngineCore_DP0 pid=42321) (RayWorkerWrapper pid=43212) INFO 10-10 01:45:50 [gpu_model_runner.py:2758] Model loading took 1.1667 GiB and 0.168023 seconds 2025-10-10T01:45:56.3828037Z (EngineCore_DP0 pid=42321) INFO 10-10 01:45:56 [kv_cache_utils.py:1087] GPU KV cache size: 1,215,360 tokens 2025-10-10T01:45:56.3829167Z (EngineCore_DP0 pid=42321) INFO 10-10 01:45:56 [kv_cache_utils.py:1091] Maximum concurrency for 2,048 tokens per request: 593.44x 2025-10-10T01:45:56.3830343Z (EngineCore_DP0 pid=42321) INFO 10-10 01:45:56 [kv_cache_utils.py:1087] GPU KV cache size: 1,215,360 tokens 2025-10-10T01:45:56.3831231Z (EngineCore_DP0 pid=42321) INFO 10-10 01:45:56 [kv_cache_utils.py:1091] Maximum concurrency for 2,048 tokens per request: 593.44x 2025-10-10T01:45:56.4050681Z (EngineCore_DP0 pid=42321) (RayWorkerWrapper pid=43207) INFO 10-10 01:45:56 [gpu_worker.py:332] Compile and warming up model for size 8 2025-10-10T01:45:56.9173366Z (EngineCore_DP0 pid=42321) (RayWorkerWrapper pid=43212) INFO 10-10 01:45:56 [backends.py:168] Directly load the compiled graph(s) for shape 8 from the cache, took 0.519 s 2025-10-10T01:45:57.9046615Z (EngineCore_DP0 pid=42321) INFO 10-10 01:45:57 [core.py:211] init engine (profile, create kv cache, warmup model) took 7.44 seconds 2025-10-10T01:45:57.9398935Z (EngineCore_DP0 pid=42321) (RayWorkerWrapper pid=43207) INFO 10-10 01:45:57 [monitor.py:32] torch.compile takes 2.87 s in total 2025-10-10T01:45:58.6722171Z (EngineCore_DP0 pid=42321) INFO 10-10 01:45:58 [vllm.py:345] Cudagraph is disabled under eager mode 2025-10-10T01:45:58.7943823Z (APIServer pid=42199) INFO 10-10 01:45:58 [loggers.py:147] Engine 000: vllm cache_config_info with initialization after num_gpu_blocks is: 75960 2025-10-10T01:45:58.9065804Z (EngineCore_DP0 pid=42321) INFO 10-10 01:45:58 [gc_utils.py:41] GC Debug Config. enabled:False,top_objects:-1 2025-10-10T01:45:58.9310352Z (APIServer pid=42199) INFO 10-10 01:45:58 [api_server.py:1634] Supported_tasks: ['generate'] 2025-10-10T01:45:58.9600121Z (APIServer pid=42199) WARNING 10-10 01:45:58 [model.py:1424] Default sampling parameters have been overridden by the model's Hugging Face generation config recommended from the model creator. If this is not intended, please relaunch vLLM instance with `--generation-config vllm`. 2025-10-10T01:45:58.9603570Z (APIServer pid=42199) INFO 10-10 01:45:58 [serving_responses.py:140] Using default chat sampling params from model: {'temperature': 0.6, 'top_p': 0.9} 2025-10-10T01:45:58.9827224Z (APIServer pid=42199) INFO 10-10 01:45:58 [serving_chat.py:139] Using default chat sampling params from model: {'temperature': 0.6, 'top_p': 0.9} 2025-10-10T01:45:59.0133496Z (APIServer pid=42199) INFO 10-10 01:45:59 [serving_completion.py:76] Using default completion sampling params from model: {'temperature': 0.6, 'top_p': 0.9} 2025-10-10T01:45:59.0135138Z (APIServer pid=42199) INFO 10-10 01:45:59 [api_server.py:1912] Starting vLLM API server 0 on http://0.0.0.0:37755 2025-10-10T01:45:59.0136128Z (APIServer pid=42199) INFO 10-10 01:45:59 [launcher.py:34] Available routes are: 2025-10-10T01:45:59.0136937Z (APIServer pid=42199) INFO 10-10 01:45:59 [launcher.py:42] Route: /openapi.json, Methods: HEAD, GET 2025-10-10T01:45:59.0137807Z (APIServer pid=42199) INFO 10-10 01:45:59 [launcher.py:42] Route: /docs, Methods: HEAD, GET 2025-10-10T01:45:59.0138540Z (APIServer pid=42199) INFO 10-10 01:45:59 [launcher.py:42] Route: /docs/oauth2-redirect, Methods: HEAD, GET 2025-10-10T01:45:59.0139273Z (APIServer pid=42199) INFO 10-10 01:45:59 [launcher.py:42] Route: /redoc, Methods: HEAD, GET 2025-10-10T01:45:59.0139929Z (APIServer pid=42199) INFO 10-10 01:45:59 [launcher.py:42] Route: /health, Methods: GET 2025-10-10T01:45:59.0140555Z (APIServer pid=42199) INFO 10-10 01:45:59 [launcher.py:42] Route: /load, Methods: GET 2025-10-10T01:45:59.0141184Z (APIServer pid=42199) INFO 10-10 01:45:59 [launcher.py:42] Route: /ping, Methods: POST 2025-10-10T01:45:59.0141809Z (APIServer pid=42199) INFO 10-10 01:45:59 [launcher.py:42] Route: /ping, Methods: GET 2025-10-10T01:45:59.0142456Z (APIServer pid=42199) INFO 10-10 01:45:59 [launcher.py:42] Route: /tokenize, Methods: POST 2025-10-10T01:45:59.0143295Z (APIServer pid=42199) INFO 10-10 01:45:59 [launcher.py:42] Route: /detokenize, Methods: POST 2025-10-10T01:45:59.0143821Z (APIServer pid=42199) INFO 10-10 01:45:59 [launcher.py:42] Route: /v1/models, Methods: GET 2025-10-10T01:45:59.0144349Z (APIServer pid=42199) INFO 10-10 01:45:59 [launcher.py:42] Route: /version, Methods: GET 2025-10-10T01:45:59.0144913Z (APIServer pid=42199) INFO 10-10 01:45:59 [launcher.py:42] Route: /v1/responses, Methods: POST 2025-10-10T01:45:59.0145502Z (APIServer pid=42199) INFO 10-10 01:45:59 [launcher.py:42] Route: /v1/responses/{response_id}, Methods: GET 2025-10-10T01:45:59.0146135Z (APIServer pid=42199) INFO 10-10 01:45:59 [launcher.py:42] Route: /v1/responses/{response_id}/cancel, Methods: POST 2025-10-10T01:45:59.0146757Z (APIServer pid=42199) INFO 10-10 01:45:59 [launcher.py:42] Route: /v1/chat/completions, Methods: POST 2025-10-10T01:45:59.0147487Z (APIServer pid=42199) INFO 10-10 01:45:59 [launcher.py:42] Route: /v1/completions, Methods: POST 2025-10-10T01:45:59.0148202Z (APIServer pid=42199) INFO 10-10 01:45:59 [launcher.py:42] Route: /v1/embeddings, Methods: POST 2025-10-10T01:45:59.0148755Z (APIServer pid=42199) INFO 10-10 01:45:59 [launcher.py:42] Route: /pooling, Methods: POST 2025-10-10T01:45:59.0149286Z (APIServer pid=42199) INFO 10-10 01:45:59 [launcher.py:42] Route: /classify, Methods: POST 2025-10-10T01:45:59.0149810Z (APIServer pid=42199) INFO 10-10 01:45:59 [launcher.py:42] Route: /score, Methods: POST 2025-10-10T01:45:59.0150312Z (APIServer pid=42199) INFO 10-10 01:45:59 [launcher.py:42] Route: /v1/score, Methods: POST 2025-10-10T01:45:59.0150881Z (APIServer pid=42199) INFO 10-10 01:45:59 [launcher.py:42] Route: /v1/audio/transcriptions, Methods: POST 2025-10-10T01:45:59.0151484Z (APIServer pid=42199) INFO 10-10 01:45:59 [launcher.py:42] Route: /v1/audio/translations, Methods: POST 2025-10-10T01:45:59.0152047Z (APIServer pid=42199) INFO 10-10 01:45:59 [launcher.py:42] Route: /rerank, Methods: POST 2025-10-10T01:45:59.0152577Z (APIServer pid=42199) INFO 10-10 01:45:59 [launcher.py:42] Route: /v1/rerank, Methods: POST 2025-10-10T01:45:59.0153102Z (APIServer pid=42199) INFO 10-10 01:45:59 [launcher.py:42] Route: /v2/rerank, Methods: POST 2025-10-10T01:45:59.0153647Z (APIServer pid=42199) INFO 10-10 01:45:59 [launcher.py:42] Route: /scale_elastic_ep, Methods: POST 2025-10-10T01:45:59.0154237Z (APIServer pid=42199) INFO 10-10 01:45:59 [launcher.py:42] Route: /is_scaling_elastic_ep, Methods: POST 2025-10-10T01:45:59.0154795Z (APIServer pid=42199) INFO 10-10 01:45:59 [launcher.py:42] Route: /invocations, Methods: POST 2025-10-10T01:45:59.0155316Z (APIServer pid=42199) INFO 10-10 01:45:59 [launcher.py:42] Route: /metrics, Methods: GET 2025-10-10T01:45:59.0285442Z (APIServer pid=42199) INFO: Started server process [42199] 2025-10-10T01:45:59.0286704Z (APIServer pid=42199) INFO: Waiting for application startup. 2025-10-10T01:45:59.2639807Z (APIServer pid=42199) INFO: Application startup complete. 2025-10-10T01:45:59.6062285Z (APIServer pid=42199) INFO: 127.0.0.1:60254 - "GET /health HTTP/1.1" 200 OK 2025-10-10T01:45:59.8259047Z (APIServer pid=42199) INFO: 127.0.0.1:60256 - "GET /v1/models HTTP/1.1" 200 OK 2025-10-10T01:45:59.8578247Z (EngineCore_DP0 pid=42321) INFO 10-10 01:45:59 [ray_distributed_executor.py:552] RAY_CGRAPH_get_timeout is set to 300 2025-10-10T01:45:59.8580097Z (EngineCore_DP0 pid=42321) INFO 10-10 01:45:59 [ray_distributed_executor.py:554] VLLM_USE_RAY_COMPILED_DAG_CHANNEL_TYPE = auto 2025-10-10T01:45:59.8581906Z (EngineCore_DP0 pid=42321) INFO 10-10 01:45:59 [ray_distributed_executor.py:556] VLLM_USE_RAY_COMPILED_DAG_OVERLAP_COMM = False 2025-10-10T01:45:59.8585000Z (EngineCore_DP0 pid=42321) INFO 10-10 01:45:59 [ray_distributed_executor.py:621] Using RayPPCommunicator (which wraps vLLM _PP GroupCoordinator) for Ray Compiled Graph communication. 2025-10-10T01:46:00.0340726Z (APIServer pid=42199) INFO: 127.0.0.1:60256 - "POST /v1/completions HTTP/1.1" 200 OK 2025-10-10T01:46:00.0934117Z (APIServer pid=42199) INFO: 127.0.0.1:60256 - "POST /v1/completions HTTP/1.1" 200 OK 2025-10-10T01:46:00.2404508Z (APIServer pid=42199) INFO: 127.0.0.1:60256 - "POST /v1/completions HTTP/1.1" 200 OK 2025-10-10T01:46:00.3251459Z (APIServer pid=42199) INFO: 127.0.0.1:60256 - "POST /v1/completions HTTP/1.1" 200 OK 2025-10-10T01:46:00.3934144Z (APIServer pid=42199) INFO: 127.0.0.1:60256 - "POST /v1/completions HTTP/1.1" 200 OK 2025-10-10T01:46:00.4010247Z (APIServer pid=42199) INFO: 127.0.0.1:60256 - "POST /v1/completions HTTP/1.1" 200 OK 2025-10-10T01:46:00.4837042Z (APIServer pid=42199) INFO 10-10 01:46:00 [launcher.py:99] Shutting down FastAPI HTTP server. 2025-10-10T01:46:00.4842681Z *** SIGTERM received at time=1760060760 on cpu 37 *** 2025-10-10T01:46:00.4888051Z PC: @ 0x7f3a8c266117 (unknown) (unknown) 2025-10-10T01:46:00.4889099Z @ 0x7f3a8c217520 (unknown) (unknown) 2025-10-10T01:46:00.4889583Z [2025-10-10 01:46:00,488 E 42321 42321] logging.cc:501: *** SIGTERM received at time=1760060760 on cpu 37 *** 2025-10-10T01:46:00.4892548Z [2025-10-10 01:46:00,488 E 42321 42321] logging.cc:501: PC: @ 0x7f3a8c266117 (unknown) (unknown) 2025-10-10T01:46:00.4893405Z [2025-10-10 01:46:00,489 E 42321 42321] logging.cc:501: @ 0x7f3a8c217520 (unknown) (unknown) 2025-10-10T01:46:00.4895268Z (EngineCore_DP0 pid=42321) INFO 10-10 01:46:00 [ray_distributed_executor.py:122] Shutting down Ray distributed executor. If you see error log from logging.cc regarding SIGTERM received, please ignore because this is the expected termination process in Ray. 2025-10-10T01:46:00.4896622Z (EngineCore_DP0 pid=42321) 2025-10-10 01:46:00,489 INFO compiled_dag_node.py:2171 -- Tearing down compiled DAG 2025-10-10T01:46:00.4897558Z (EngineCore_DP0 pid=42321) 2025-10-10 01:46:00,489 INFO compiled_dag_node.py:2176 -- Cancelling compiled worker on actor: Actor(RayWorkerWrapper, 936cb2a3c68a7054c76f483301000000) 2025-10-10T01:46:00.4898618Z (EngineCore_DP0 pid=42321) 2025-10-10 01:46:00,489 INFO compiled_dag_node.py:2176 -- Cancelling compiled worker on actor: Actor(RayWorkerWrapper, a993222b453a575343870a1f01000000) 2025-10-10T01:46:00.4939121Z (EngineCore_DP0 pid=42321) 2025-10-10 01:46:00,493 INFO compiled_dag_node.py:2198 -- Waiting for worker tasks to exit 2025-10-10T01:46:00.4940471Z (EngineCore_DP0 pid=42321) 2025-10-10 01:46:00,493 INFO compiled_dag_node.py:2201 -- Teardown complete 2025-10-10T01:46:00.9963842Z (EngineCore_DP0 pid=42321) (RayWorkerWrapper pid=43207) INFO 10-10 01:45:53 [backends.py:548] Using cache directory: /var/lib/jenkins/.cache/vllm/torch_compile_cache/31e0bc77eb/rank_0_0/backbone for vLLM's torch.compile 2025-10-10T01:46:00.9965339Z (EngineCore_DP0 pid=42321) (RayWorkerWrapper pid=43207) INFO 10-10 01:45:53 [backends.py:559] Dynamo bytecode transform time: 2.87 s 2025-10-10T01:46:00.9966300Z (EngineCore_DP0 pid=42321) (RayWorkerWrapper pid=43207) INFO 10-10 01:45:54 [backends.py:164] Directly load the compiled graph(s) for dynamic shape from the cache, took 0.975 s 2025-10-10T01:46:00.9967233Z (EngineCore_DP0 pid=42321) (RayWorkerWrapper pid=43212) INFO 10-10 01:45:56 [gpu_worker.py:298] Available KV cache memory: 18.54 GiB 2025-10-10T01:46:00.9968338Z (EngineCore_DP0 pid=42321) (RayWorkerWrapper pid=43212) INFO 10-10 01:45:56 [gpu_worker.py:332] Compile and warming up model for size 4 [repeated 3x across cluster] 2025-10-10T01:46:00.9969599Z (EngineCore_DP0 pid=42321) (RayWorkerWrapper pid=43212) INFO 10-10 01:45:57 [backends.py:168] Directly load the compiled graph(s) for shape 4 from the cache, took 0.970 s [repeated 3x across cluster] 2025-10-10T01:46:00.9970655Z (EngineCore_DP0 pid=42321) (RayWorkerWrapper pid=43212) INFO 10-10 01:45:57 [monitor.py:32] torch.compile takes 2.82 s in total 2025-10-10T01:46:00.9972298Z (EngineCore_DP0 pid=42321) (pid=43212) /opt/conda/envs/py_3.12/lib/python3.12/site-packages/torch/cuda/__init__.py:64: FutureWarning: The pynvml package is deprecated. Please install nvidia-ml-py instead. If you did not install pynvml directly, please report this to the maintainers of the package that installed pynvml for you. 2025-10-10T01:46:00.9973522Z (EngineCore_DP0 pid=42321) (pid=43212) import pynvml # type: ignore[import] 2025-10-10T01:46:02.9912637Z (APIServer pid=42199) INFO: Shutting down 2025-10-10T01:46:03.0919851Z (APIServer pid=42199) INFO: Waiting for application shutdown. 2025-10-10T01:46:03.0921265Z (APIServer pid=42199) INFO: Application shutdown complete. 2025-10-10T01:46:03.5106546Z INFO 10-10 01:46:03 [model.py:551] Resolved architecture: LlamaForCausalLM 2025-10-10T01:46:03.5107106Z WARNING 10-10 01:46:03 [model.py:1768] Casting torch.bfloat16 to torch.float16. 2025-10-10T01:46:03.5107572Z INFO 10-10 01:46:03 [model.py:1545] Using max model len 2048 2025-10-10T01:46:03.5108919Z Launching RemoteOpenAIServer with: vllm serve meta-llama/Llama-3.2-1B-Instruct --dtype float16 --max-model-len 2048 --max-num-seqs 8 --enable-chunked-prefill --enforce-eager --tokenizer-mode auto --tensor-parallel-size 2 --distributed-executor-backend mp --load-format dummy --port 60373 --seed 0 2025-10-10T01:46:04.0202312Z /opt/conda/envs/py_3.12/lib/python3.12/site-packages/torch/cuda/__init__.py:64: FutureWarning: The pynvml package is deprecated. Please install nvidia-ml-py instead. If you did not install pynvml directly, please report this to the maintainers of the package that installed pynvml for you. 2025-10-10T01:46:04.0204715Z import pynvml # type: ignore[import] 2025-10-10T01:46:06.2206122Z INFO 10-10 01:46:06 [__init__.py:215] Automatically detected platform cuda. 2025-10-10T01:46:10.1186100Z (APIServer pid=46963) INFO 10-10 01:46:10 [api_server.py:1839] vLLM API server version 0.11.0rc2.dev157+g0ad9951c4.d20251010 2025-10-10T01:46:10.1208893Z (APIServer pid=46963) INFO 10-10 01:46:10 [utils.py:233] non-default args: {'model_tag': 'meta-llama/Llama-3.2-1B-Instruct', 'port': 60373, 'model': 'meta-llama/Llama-3.2-1B-Instruct', 'dtype': 'float16', 'seed': 0, 'max_model_len': 2048, 'enforce_eager': True, 'load_format': 'dummy', 'distributed_executor_backend': 'mp', 'tensor_parallel_size': 2, 'max_num_seqs': 8, 'enable_chunked_prefill': True} 2025-10-10T01:46:10.2591175Z (APIServer pid=46963) INFO 10-10 01:46:10 [model.py:551] Resolved architecture: LlamaForCausalLM 2025-10-10T01:46:10.2592654Z (APIServer pid=46963) `torch_dtype` is deprecated! Use `dtype` instead! 2025-10-10T01:46:10.2594078Z (APIServer pid=46963) WARNING 10-10 01:46:10 [model.py:1768] Casting torch.bfloat16 to torch.float16. 2025-10-10T01:46:10.2595391Z (APIServer pid=46963) INFO 10-10 01:46:10 [model.py:1545] Using max model len 2048 2025-10-10T01:46:10.4193972Z (APIServer pid=46963) INFO 10-10 01:46:10 [scheduler.py:222] Chunked prefill is enabled with max_num_batched_tokens=2048. 2025-10-10T01:46:10.4198867Z (APIServer pid=46963) INFO 10-10 01:46:10 [vllm.py:345] Cudagraph is disabled under eager mode 2025-10-10T01:46:11.3819747Z /opt/conda/envs/py_3.12/lib/python3.12/site-packages/torch/cuda/__init__.py:64: FutureWarning: The pynvml package is deprecated. Please install nvidia-ml-py instead. If you did not install pynvml directly, please report this to the maintainers of the package that installed pynvml for you. 2025-10-10T01:46:11.3821884Z import pynvml # type: ignore[import] 2025-10-10T01:46:13.6698493Z INFO 10-10 01:46:13 [__init__.py:215] Automatically detected platform cuda. 2025-10-10T01:46:17.6019174Z (EngineCore_DP0 pid=47085) INFO 10-10 01:46:17 [core.py:648] Waiting for init message from front-end. 2025-10-10T01:46:17.6294766Z (EngineCore_DP0 pid=47085) INFO 10-10 01:46:17 [core.py:78] Initializing a V1 LLM engine (v0.11.0rc2.dev157+g0ad9951c4.d20251010) with config: model='meta-llama/Llama-3.2-1B-Instruct', speculative_config=None, tokenizer='meta-llama/Llama-3.2-1B-Instruct', skip_tokenizer_init=False, tokenizer_mode=auto, revision=None, tokenizer_revision=None, trust_remote_code=False, dtype=torch.float16, max_seq_len=2048, download_dir=None, load_format=dummy, tensor_parallel_size=2, pipeline_parallel_size=1, data_parallel_size=1, disable_custom_all_reduce=False, quantization=None, enforce_eager=True, kv_cache_dtype=auto, device_config=cuda, structured_outputs_config=StructuredOutputsConfig(backend='auto', disable_fallback=False, disable_any_whitespace=False, disable_additional_properties=False, reasoning_parser=''), observability_config=ObservabilityConfig(show_hidden_metrics_for_version=None, otlp_traces_endpoint=None, collect_detailed_traces=None), seed=0, served_model_name=meta-llama/Llama-3.2-1B-Instruct, enable_prefix_caching=True, chunked_prefill_enabled=True, pooler_config=None, compilation_config={'level': 0, 'debug_dump_path': None, 'cache_dir': '', 'backend': '', 'custom_ops': [], 'splitting_ops': None, 'use_inductor': True, 'compile_sizes': [], 'inductor_compile_config': {'enable_auto_functionalized_v2': False}, 'inductor_passes': {}, 'cudagraph_mode': , 'use_cudagraph': False, 'cudagraph_num_of_warmups': 0, 'cudagraph_capture_sizes': [], 'cudagraph_copy_inputs': False, 'full_cuda_graph': False, 'use_inductor_graph_partition': False, 'pass_config': {}, 'max_capture_size': 0, 'local_cache_dir': None} 2025-10-10T01:46:17.6301042Z (EngineCore_DP0 pid=47085) WARNING 10-10 01:46:17 [multiproc_executor.py:720] Reducing Torch parallelism from 24 threads to 1 to avoid unnecessary CPU contention. Set OMP_NUM_THREADS in the external environment to tune this value as needed. 2025-10-10T01:46:17.6308608Z (EngineCore_DP0 pid=47085) INFO 10-10 01:46:17 [shm_broadcast.py:289] vLLM message queue communication handle: Handle(local_reader_ranks=[0, 1], buffer_handle=(2, 16777216, 10, 'psm_ba4e59bd'), local_subscribe_addr='ipc:///tmp/3e272603-fb99-43ad-a5d4-c7b55afcd87b', remote_subscribe_addr=None, remote_addr_ipv6=False) 2025-10-10T01:46:18.1134504Z /opt/conda/envs/py_3.12/lib/python3.12/site-packages/torch/cuda/__init__.py:64: FutureWarning: The pynvml package is deprecated. Please install nvidia-ml-py instead. If you did not install pynvml directly, please report this to the maintainers of the package that installed pynvml for you. 2025-10-10T01:46:18.1136303Z import pynvml # type: ignore[import] 2025-10-10T01:46:18.1138032Z /opt/conda/envs/py_3.12/lib/python3.12/site-packages/torch/cuda/__init__.py:64: FutureWarning: The pynvml package is deprecated. Please install nvidia-ml-py instead. If you did not install pynvml directly, please report this to the maintainers of the package that installed pynvml for you. 2025-10-10T01:46:18.1139733Z import pynvml # type: ignore[import] 2025-10-10T01:46:20.4431548Z INFO 10-10 01:46:20 [__init__.py:215] Automatically detected platform cuda. 2025-10-10T01:46:20.4775152Z INFO 10-10 01:46:20 [__init__.py:215] Automatically detected platform cuda. 2025-10-10T01:46:25.1145329Z INFO 10-10 01:46:25 [shm_broadcast.py:289] vLLM message queue communication handle: Handle(local_reader_ranks=[0], buffer_handle=(1, 10485760, 10, 'psm_35acc1c5'), local_subscribe_addr='ipc:///tmp/0cc3e5fd-6b30-4abf-b735-b33ad947bd5a', remote_subscribe_addr=None, remote_addr_ipv6=False) 2025-10-10T01:46:25.1340460Z INFO 10-10 01:46:25 [shm_broadcast.py:289] vLLM message queue communication handle: Handle(local_reader_ranks=[0], buffer_handle=(1, 10485760, 10, 'psm_89e5d8be'), local_subscribe_addr='ipc:///tmp/38805b40-677d-474b-b3d4-f92a85d5f977', remote_subscribe_addr=None, remote_addr_ipv6=False) 2025-10-10T01:46:25.3348441Z [Gloo] Rank 1 is connected to 1 peer ranks. Expected number of connected peer ranks is : 1 2025-10-10T01:46:25.3349694Z [Gloo] Rank 0 is connected to 1 peer ranks. Expected number of connected peer ranks is : 1 2025-10-10T01:46:25.3397405Z [Gloo] Rank 0 is connected to 1 peer ranks. Expected number of connected peer ranks is : 1 2025-10-10T01:46:25.3398035Z [Gloo] Rank 1 is connected to 1 peer ranks. Expected number of connected peer ranks is : 1 2025-10-10T01:46:25.3441908Z INFO 10-10 01:46:25 [__init__.py:1384] Found nccl from library libnccl.so.2 2025-10-10T01:46:25.3442907Z INFO 10-10 01:46:25 [pynccl.py:104] vLLM is using nccl==2.27.5 2025-10-10T01:46:25.3443853Z INFO 10-10 01:46:25 [__init__.py:1384] Found nccl from library libnccl.so.2 2025-10-10T01:46:25.3445871Z INFO 10-10 01:46:25 [pynccl.py:104] vLLM is using nccl==2.27.5 2025-10-10T01:46:25.5186697Z WARNING 10-10 01:46:25 [symm_mem.py:58] SymmMemCommunicator: Device capability 8.9 not supported, communicator is not available. 2025-10-10T01:46:25.5188461Z WARNING 10-10 01:46:25 [symm_mem.py:58] SymmMemCommunicator: Device capability 8.9 not supported, communicator is not available. 2025-10-10T01:46:25.5211411Z INFO 10-10 01:46:25 [custom_all_reduce.py:35] Skipping P2P check and trusting the driver's P2P report. 2025-10-10T01:46:25.5212696Z INFO 10-10 01:46:25 [custom_all_reduce.py:35] Skipping P2P check and trusting the driver's P2P report. 2025-10-10T01:46:25.5213849Z WARNING 10-10 01:46:25 [custom_all_reduce.py:154] Custom allreduce is disabled because your platform lacks GPU P2P capability or P2P test failed. To silence this warning, specify disable_custom_all_reduce=True explicitly. 2025-10-10T01:46:25.5215260Z WARNING 10-10 01:46:25 [custom_all_reduce.py:154] Custom allreduce is disabled because your platform lacks GPU P2P capability or P2P test failed. To silence this warning, specify disable_custom_all_reduce=True explicitly. 2025-10-10T01:46:25.5839352Z INFO 10-10 01:46:25 [shm_broadcast.py:289] vLLM message queue communication handle: Handle(local_reader_ranks=[1], buffer_handle=(1, 4194304, 6, 'psm_3754758f'), local_subscribe_addr='ipc:///tmp/8e9f72d0-1353-4b79-8c16-8a4795df76c9', remote_subscribe_addr=None, remote_addr_ipv6=False) 2025-10-10T01:46:25.5869205Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-10-10T01:46:25.5869795Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-10-10T01:46:25.5876535Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-10-10T01:46:25.5886305Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-10-10T01:46:25.6736590Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-10-10T01:46:25.6753542Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-10-10T01:46:25.6761721Z [Gloo] Rank 1 is connected to 1 peer ranks. Expected number of connected peer ranks is : 1 2025-10-10T01:46:25.6762296Z [Gloo] Rank 0 is connected to 1 peer ranks. Expected number of connected peer ranks is : 1 2025-10-10T01:46:25.6764521Z INFO 10-10 01:46:25 [__init__.py:1384] Found nccl from library libnccl.so.2 2025-10-10T01:46:25.6765009Z INFO 10-10 01:46:25 [pynccl.py:104] vLLM is using nccl==2.27.5 2025-10-10T01:46:25.6765445Z INFO 10-10 01:46:25 [__init__.py:1384] Found nccl from library libnccl.so.2 2025-10-10T01:46:25.6765867Z INFO 10-10 01:46:25 [pynccl.py:104] vLLM is using nccl==2.27.5 2025-10-10T01:46:25.7114775Z INFO 10-10 01:46:25 [parallel_state.py:1208] rank 0 in world size 2 is assigned as DP rank 0, PP rank 0, TP rank 0, EP rank 0 2025-10-10T01:46:25.7115551Z INFO 10-10 01:46:25 [parallel_state.py:1208] rank 1 in world size 2 is assigned as DP rank 0, PP rank 0, TP rank 1, EP rank 1 2025-10-10T01:46:25.9565607Z INFO 10-10 01:46:25 [topk_topp_sampler.py:55] Using FlashInfer for top-p & top-k sampling. 2025-10-10T01:46:25.9663120Z INFO 10-10 01:46:25 [topk_topp_sampler.py:55] Using FlashInfer for top-p & top-k sampling. 2025-10-10T01:46:25.9882875Z (Worker_TP1 pid=47205) INFO 10-10 01:46:25 [gpu_model_runner.py:2707] Starting to load model meta-llama/Llama-3.2-1B-Instruct... 2025-10-10T01:46:25.9979076Z (Worker_TP0 pid=47204) INFO 10-10 01:46:25 [gpu_model_runner.py:2707] Starting to load model meta-llama/Llama-3.2-1B-Instruct... 2025-10-10T01:46:26.2100685Z (Worker_TP1 pid=47205) INFO 10-10 01:46:26 [gpu_model_runner.py:2739] Loading model from scratch... 2025-10-10T01:46:26.2217864Z (Worker_TP0 pid=47204) INFO 10-10 01:46:26 [gpu_model_runner.py:2739] Loading model from scratch... 2025-10-10T01:46:26.2998331Z (Worker_TP1 pid=47205) INFO 10-10 01:46:26 [cuda.py:361] Using Flash Attention backend on V1 engine. 2025-10-10T01:46:26.3125525Z (Worker_TP0 pid=47204) INFO 10-10 01:46:26 [cuda.py:361] Using Flash Attention backend on V1 engine. 2025-10-10T01:46:26.7776575Z (Worker_TP1 pid=47205) INFO 10-10 01:46:26 [gpu_model_runner.py:2758] Model loading took 1.1667 GiB and 0.119249 seconds 2025-10-10T01:46:26.7977095Z (Worker_TP0 pid=47204) INFO 10-10 01:46:26 [gpu_model_runner.py:2758] Model loading took 1.1667 GiB and 0.119182 seconds 2025-10-10T01:46:31.7235675Z (Worker_TP0 pid=47204) INFO 10-10 01:46:31 [gpu_worker.py:298] Available KV cache memory: 18.53 GiB 2025-10-10T01:46:31.7342805Z (Worker_TP1 pid=47205) INFO 10-10 01:46:31 [gpu_worker.py:298] Available KV cache memory: 18.53 GiB 2025-10-10T01:46:31.9609898Z (EngineCore_DP0 pid=47085) INFO 10-10 01:46:31 [kv_cache_utils.py:1087] GPU KV cache size: 1,214,464 tokens 2025-10-10T01:46:31.9610850Z (EngineCore_DP0 pid=47085) INFO 10-10 01:46:31 [kv_cache_utils.py:1091] Maximum concurrency for 2,048 tokens per request: 593.00x 2025-10-10T01:46:31.9611709Z (EngineCore_DP0 pid=47085) INFO 10-10 01:46:31 [kv_cache_utils.py:1087] GPU KV cache size: 1,214,464 tokens 2025-10-10T01:46:31.9612578Z (EngineCore_DP0 pid=47085) INFO 10-10 01:46:31 [kv_cache_utils.py:1091] Maximum concurrency for 2,048 tokens per request: 593.00x 2025-10-10T01:46:32.0842158Z (EngineCore_DP0 pid=47085) INFO 10-10 01:46:32 [core.py:211] init engine (profile, create kv cache, warmup model) took 5.21 seconds 2025-10-10T01:46:32.5703966Z (EngineCore_DP0 pid=47085) INFO 10-10 01:46:32 [vllm.py:345] Cudagraph is disabled under eager mode 2025-10-10T01:46:32.7551037Z (APIServer pid=46963) INFO 10-10 01:46:32 [loggers.py:147] Engine 000: vllm cache_config_info with initialization after num_gpu_blocks is: 75904 2025-10-10T01:46:32.7857077Z (EngineCore_DP0 pid=47085) INFO 10-10 01:46:32 [gc_utils.py:41] GC Debug Config. enabled:False,top_objects:-1 2025-10-10T01:46:32.8067918Z (APIServer pid=46963) INFO 10-10 01:46:32 [api_server.py:1634] Supported_tasks: ['generate'] 2025-10-10T01:46:32.8290810Z (APIServer pid=46963) WARNING 10-10 01:46:32 [model.py:1424] Default sampling parameters have been overridden by the model's Hugging Face generation config recommended from the model creator. If this is not intended, please relaunch vLLM instance with `--generation-config vllm`. 2025-10-10T01:46:32.8292421Z (APIServer pid=46963) INFO 10-10 01:46:32 [serving_responses.py:140] Using default chat sampling params from model: {'temperature': 0.6, 'top_p': 0.9} 2025-10-10T01:46:32.8510290Z (APIServer pid=46963) INFO 10-10 01:46:32 [serving_chat.py:139] Using default chat sampling params from model: {'temperature': 0.6, 'top_p': 0.9} 2025-10-10T01:46:32.8714766Z (APIServer pid=46963) INFO 10-10 01:46:32 [serving_completion.py:76] Using default completion sampling params from model: {'temperature': 0.6, 'top_p': 0.9} 2025-10-10T01:46:32.8715856Z (APIServer pid=46963) INFO 10-10 01:46:32 [api_server.py:1912] Starting vLLM API server 0 on http://0.0.0.0:60373 2025-10-10T01:46:32.8716666Z (APIServer pid=46963) INFO 10-10 01:46:32 [launcher.py:34] Available routes are: 2025-10-10T01:46:32.8717400Z (APIServer pid=46963) INFO 10-10 01:46:32 [launcher.py:42] Route: /openapi.json, Methods: GET, HEAD 2025-10-10T01:46:32.8718112Z (APIServer pid=46963) INFO 10-10 01:46:32 [launcher.py:42] Route: /docs, Methods: GET, HEAD 2025-10-10T01:46:32.8718936Z (APIServer pid=46963) INFO 10-10 01:46:32 [launcher.py:42] Route: /docs/oauth2-redirect, Methods: GET, HEAD 2025-10-10T01:46:32.8721500Z (APIServer pid=46963) INFO 10-10 01:46:32 [launcher.py:42] Route: /redoc, Methods: GET, HEAD 2025-10-10T01:46:32.8722230Z (APIServer pid=46963) INFO 10-10 01:46:32 [launcher.py:42] Route: /health, Methods: GET 2025-10-10T01:46:32.8722915Z (APIServer pid=46963) INFO 10-10 01:46:32 [launcher.py:42] Route: /load, Methods: GET 2025-10-10T01:46:32.8723666Z (APIServer pid=46963) INFO 10-10 01:46:32 [launcher.py:42] Route: /ping, Methods: POST 2025-10-10T01:46:32.8724715Z (APIServer pid=46963) INFO 10-10 01:46:32 [launcher.py:42] Route: /ping, Methods: GET 2025-10-10T01:46:32.8725293Z (APIServer pid=46963) INFO 10-10 01:46:32 [launcher.py:42] Route: /tokenize, Methods: POST 2025-10-10T01:46:32.8725854Z (APIServer pid=46963) INFO 10-10 01:46:32 [launcher.py:42] Route: /detokenize, Methods: POST 2025-10-10T01:46:32.8726422Z (APIServer pid=46963) INFO 10-10 01:46:32 [launcher.py:42] Route: /v1/models, Methods: GET 2025-10-10T01:46:32.8726960Z (APIServer pid=46963) INFO 10-10 01:46:32 [launcher.py:42] Route: /version, Methods: GET 2025-10-10T01:46:32.8727493Z (APIServer pid=46963) INFO 10-10 01:46:32 [launcher.py:42] Route: /v1/responses, Methods: POST 2025-10-10T01:46:32.8728098Z (APIServer pid=46963) INFO 10-10 01:46:32 [launcher.py:42] Route: /v1/responses/{response_id}, Methods: GET 2025-10-10T01:46:32.8728733Z (APIServer pid=46963) INFO 10-10 01:46:32 [launcher.py:42] Route: /v1/responses/{response_id}/cancel, Methods: POST 2025-10-10T01:46:32.8729352Z (APIServer pid=46963) INFO 10-10 01:46:32 [launcher.py:42] Route: /v1/chat/completions, Methods: POST 2025-10-10T01:46:32.8729942Z (APIServer pid=46963) INFO 10-10 01:46:32 [launcher.py:42] Route: /v1/completions, Methods: POST 2025-10-10T01:46:32.8730500Z (APIServer pid=46963) INFO 10-10 01:46:32 [launcher.py:42] Route: /v1/embeddings, Methods: POST 2025-10-10T01:46:32.8731059Z (APIServer pid=46963) INFO 10-10 01:46:32 [launcher.py:42] Route: /pooling, Methods: POST 2025-10-10T01:46:32.8731611Z (APIServer pid=46963) INFO 10-10 01:46:32 [launcher.py:42] Route: /classify, Methods: POST 2025-10-10T01:46:32.8732147Z (APIServer pid=46963) INFO 10-10 01:46:32 [launcher.py:42] Route: /score, Methods: POST 2025-10-10T01:46:32.8732681Z (APIServer pid=46963) INFO 10-10 01:46:32 [launcher.py:42] Route: /v1/score, Methods: POST 2025-10-10T01:46:32.8733271Z (APIServer pid=46963) INFO 10-10 01:46:32 [launcher.py:42] Route: /v1/audio/transcriptions, Methods: POST 2025-10-10T01:46:32.8733862Z (APIServer pid=46963) INFO 10-10 01:46:32 [launcher.py:42] Route: /v1/audio/translations, Methods: POST 2025-10-10T01:46:32.8734410Z (APIServer pid=46963) INFO 10-10 01:46:32 [launcher.py:42] Route: /rerank, Methods: POST 2025-10-10T01:46:32.8734938Z (APIServer pid=46963) INFO 10-10 01:46:32 [launcher.py:42] Route: /v1/rerank, Methods: POST 2025-10-10T01:46:32.8735575Z (APIServer pid=46963) INFO 10-10 01:46:32 [launcher.py:42] Route: /v2/rerank, Methods: POST 2025-10-10T01:46:32.8736126Z (APIServer pid=46963) INFO 10-10 01:46:32 [launcher.py:42] Route: /scale_elastic_ep, Methods: POST 2025-10-10T01:46:32.8736714Z (APIServer pid=46963) INFO 10-10 01:46:32 [launcher.py:42] Route: /is_scaling_elastic_ep, Methods: POST 2025-10-10T01:46:32.8737285Z (APIServer pid=46963) INFO 10-10 01:46:32 [launcher.py:42] Route: /invocations, Methods: POST 2025-10-10T01:46:32.8737821Z (APIServer pid=46963) INFO 10-10 01:46:32 [launcher.py:42] Route: /metrics, Methods: GET 2025-10-10T01:46:32.8868077Z (APIServer pid=46963) INFO: Started server process [46963] 2025-10-10T01:46:32.8868742Z (APIServer pid=46963) INFO: Waiting for application startup. 2025-10-10T01:46:33.1144108Z (APIServer pid=46963) INFO: Application startup complete. 2025-10-10T01:46:33.5865990Z (APIServer pid=46963) INFO: 127.0.0.1:52160 - "GET /health HTTP/1.1" 200 OK 2025-10-10T01:46:33.5966829Z (APIServer pid=46963) INFO: 127.0.0.1:52168 - "GET /v1/models HTTP/1.1" 200 OK 2025-10-10T01:46:33.8221049Z (APIServer pid=46963) INFO: 127.0.0.1:52168 - "POST /v1/completions HTTP/1.1" 200 OK 2025-10-10T01:46:33.8792330Z (APIServer pid=46963) INFO: 127.0.0.1:52168 - "POST /v1/completions HTTP/1.1" 200 OK 2025-10-10T01:46:33.9766987Z (APIServer pid=46963) INFO: 127.0.0.1:52168 - "POST /v1/completions HTTP/1.1" 200 OK 2025-10-10T01:46:34.0614153Z (APIServer pid=46963) INFO: 127.0.0.1:52168 - "POST /v1/completions HTTP/1.1" 200 OK 2025-10-10T01:46:34.1339221Z (APIServer pid=46963) INFO: 127.0.0.1:52168 - "POST /v1/completions HTTP/1.1" 200 OK 2025-10-10T01:46:34.1417196Z (APIServer pid=46963) INFO: 127.0.0.1:52168 - "POST /v1/completions HTTP/1.1" 200 OK 2025-10-10T01:46:34.2191895Z (APIServer pid=46963) INFO 10-10 01:46:34 [launcher.py:99] Shutting down FastAPI HTTP server. 2025-10-10T01:46:34.2195367Z (Worker_TP0 pid=47204) INFO 10-10 01:46:34 [multiproc_executor.py:558] Parent process exited, terminating worker 2025-10-10T01:46:34.2197518Z (Worker_TP0 pid=47204) INFO 10-10 01:46:34 [multiproc_executor.py:599] WorkerProc shutting down. 2025-10-10T01:46:34.2199006Z (Worker_TP1 pid=47205) INFO 10-10 01:46:34 [multiproc_executor.py:558] Parent process exited, terminating worker 2025-10-10T01:46:34.2200297Z (Worker_TP1 pid=47205) INFO 10-10 01:46:34 [multiproc_executor.py:599] WorkerProc shutting down. 2025-10-10T01:46:36.0378554Z (APIServer pid=46963) INFO: Shutting down 2025-10-10T01:46:36.1373879Z (APIServer pid=46963) INFO: Waiting for application shutdown. 2025-10-10T01:46:36.1375659Z (APIServer pid=46963) INFO: Application shutdown complete. 2025-10-10T01:46:36.7674169Z PASSED 2025-10-10T01:46:36.7817556Z distributed/test_sequence_parallel.py::test_tp_sp_generation[meta-llama/Llama-3.2-1B-Instruct-parallel_setup12-mp-1-auto-test_options12] Fork a new process to run a test 47722 2025-10-10T01:46:36.7832325Z Fork a new process to run a test 0 2025-10-10T01:46:37.4215282Z INFO 10-10 01:46:37 [model.py:551] Resolved architecture: LlamaForCausalLM 2025-10-10T01:46:37.4216245Z `torch_dtype` is deprecated! Use `dtype` instead! 2025-10-10T01:46:37.4217243Z WARNING 10-10 01:46:37 [model.py:1768] Casting torch.bfloat16 to torch.float16. 2025-10-10T01:46:37.4217900Z INFO 10-10 01:46:37 [model.py:1545] Using max model len 2048 2025-10-10T01:46:37.4220370Z Launching RemoteOpenAIServer with: vllm serve meta-llama/Llama-3.2-1B-Instruct --dtype float16 --max-model-len 2048 --max-num-seqs 8 --enforce-eager --tokenizer-mode auto --tensor-parallel-size 2 --pipeline-parallel-size 2 --distributed-executor-backend mp --compilation_config {"level": 3, "custom_ops": ["+rms_norm"], "compile_sizes": [4, 8], "pass_config": {"enable_sequence_parallelism": true, "enable_fusion": false, "enable_noop": true}} --load-format dummy --port 46161 --seed 0 2025-10-10T01:46:37.9288601Z /opt/conda/envs/py_3.12/lib/python3.12/site-packages/torch/cuda/__init__.py:64: FutureWarning: The pynvml package is deprecated. Please install nvidia-ml-py instead. If you did not install pynvml directly, please report this to the maintainers of the package that installed pynvml for you. 2025-10-10T01:46:37.9290901Z import pynvml # type: ignore[import] 2025-10-10T01:46:40.1369462Z INFO 10-10 01:46:40 [__init__.py:215] Automatically detected platform cuda. 2025-10-10T01:46:44.0658210Z (APIServer pid=47771) INFO 10-10 01:46:44 [api_server.py:1839] vLLM API server version 0.11.0rc2.dev157+g0ad9951c4.d20251010 2025-10-10T01:46:44.0680789Z (APIServer pid=47771) INFO 10-10 01:46:44 [utils.py:233] non-default args: {'model_tag': 'meta-llama/Llama-3.2-1B-Instruct', 'port': 46161, 'model': 'meta-llama/Llama-3.2-1B-Instruct', 'dtype': 'float16', 'seed': 0, 'max_model_len': 2048, 'enforce_eager': True, 'load_format': 'dummy', 'distributed_executor_backend': 'mp', 'pipeline_parallel_size': 2, 'tensor_parallel_size': 2, 'max_num_seqs': 8, 'compilation_config': {'level': 3, 'debug_dump_path': None, 'cache_dir': '', 'backend': '', 'custom_ops': ['+rms_norm'], 'splitting_ops': None, 'use_inductor': True, 'compile_sizes': [4, 8], 'inductor_compile_config': {'enable_auto_functionalized_v2': False}, 'inductor_passes': {}, 'cudagraph_mode': None, 'use_cudagraph': True, 'cudagraph_num_of_warmups': 0, 'cudagraph_capture_sizes': None, 'cudagraph_copy_inputs': False, 'full_cuda_graph': False, 'use_inductor_graph_partition': False, 'pass_config': {'enable_noop': True, 'enable_sequence_parallelism': True}, 'max_capture_size': None, 'local_cache_dir': None}} 2025-10-10T01:46:44.2610727Z (APIServer pid=47771) INFO 10-10 01:46:44 [model.py:551] Resolved architecture: LlamaForCausalLM 2025-10-10T01:46:44.2611424Z (APIServer pid=47771) `torch_dtype` is deprecated! Use `dtype` instead! 2025-10-10T01:46:44.2612141Z (APIServer pid=47771) WARNING 10-10 01:46:44 [model.py:1768] Casting torch.bfloat16 to torch.float16. 2025-10-10T01:46:44.2615061Z (APIServer pid=47771) INFO 10-10 01:46:44 [model.py:1545] Using max model len 2048 2025-10-10T01:46:44.4265029Z (APIServer pid=47771) INFO 10-10 01:46:44 [scheduler.py:222] Chunked prefill is enabled with max_num_batched_tokens=2048. 2025-10-10T01:46:44.4272287Z (APIServer pid=47771) INFO 10-10 01:46:44 [vllm.py:345] Cudagraph is disabled under eager mode 2025-10-10T01:46:45.4061531Z /opt/conda/envs/py_3.12/lib/python3.12/site-packages/torch/cuda/__init__.py:64: FutureWarning: The pynvml package is deprecated. Please install nvidia-ml-py instead. If you did not install pynvml directly, please report this to the maintainers of the package that installed pynvml for you. 2025-10-10T01:46:45.4063389Z import pynvml # type: ignore[import] 2025-10-10T01:46:47.7161574Z INFO 10-10 01:46:47 [__init__.py:215] Automatically detected platform cuda. 2025-10-10T01:46:51.6627161Z (EngineCore_DP0 pid=47893) INFO 10-10 01:46:51 [core.py:648] Waiting for init message from front-end. 2025-10-10T01:46:51.6893454Z (EngineCore_DP0 pid=47893) INFO 10-10 01:46:51 [core.py:78] Initializing a V1 LLM engine (v0.11.0rc2.dev157+g0ad9951c4.d20251010) with config: model='meta-llama/Llama-3.2-1B-Instruct', speculative_config=None, tokenizer='meta-llama/Llama-3.2-1B-Instruct', skip_tokenizer_init=False, tokenizer_mode=auto, revision=None, tokenizer_revision=None, trust_remote_code=False, dtype=torch.float16, max_seq_len=2048, download_dir=None, load_format=dummy, tensor_parallel_size=2, pipeline_parallel_size=2, data_parallel_size=1, disable_custom_all_reduce=False, quantization=None, enforce_eager=True, kv_cache_dtype=auto, device_config=cuda, structured_outputs_config=StructuredOutputsConfig(backend='auto', disable_fallback=False, disable_any_whitespace=False, disable_additional_properties=False, reasoning_parser=''), observability_config=ObservabilityConfig(show_hidden_metrics_for_version=None, otlp_traces_endpoint=None, collect_detailed_traces=None), seed=0, served_model_name=meta-llama/Llama-3.2-1B-Instruct, enable_prefix_caching=True, chunked_prefill_enabled=True, pooler_config=None, compilation_config={'level': 3, 'debug_dump_path': None, 'cache_dir': '', 'backend': '', 'custom_ops': ['+rms_norm', '+rms_norm'], 'splitting_ops': ['vllm.unified_attention', 'vllm.unified_attention_with_output', 'vllm.mamba_mixer2', 'vllm.mamba_mixer', 'vllm.short_conv', 'vllm.linear_attention', 'vllm.plamo2_mamba_mixer', 'vllm.gdn_attention', 'vllm.sparse_attn_indexer'], 'use_inductor': True, 'compile_sizes': [8, 4], 'inductor_compile_config': {'enable_auto_functionalized_v2': False}, 'inductor_passes': {}, 'cudagraph_mode': , 'use_cudagraph': False, 'cudagraph_num_of_warmups': 0, 'cudagraph_capture_sizes': [], 'cudagraph_copy_inputs': False, 'full_cuda_graph': False, 'use_inductor_graph_partition': False, 'pass_config': {'enable_noop': True, 'enable_sequence_parallelism': True}, 'max_capture_size': 0, 'local_cache_dir': None} 2025-10-10T01:46:51.6899538Z (EngineCore_DP0 pid=47893) WARNING 10-10 01:46:51 [multiproc_executor.py:720] Reducing Torch parallelism from 24 threads to 1 to avoid unnecessary CPU contention. Set OMP_NUM_THREADS in the external environment to tune this value as needed. 2025-10-10T01:46:51.6906740Z (EngineCore_DP0 pid=47893) INFO 10-10 01:46:51 [shm_broadcast.py:289] vLLM message queue communication handle: Handle(local_reader_ranks=[0, 1, 2, 3], buffer_handle=(4, 16777216, 10, 'psm_a7eb02df'), local_subscribe_addr='ipc:///tmp/a451a29b-a49a-4626-a2ed-630fab19a090', remote_subscribe_addr=None, remote_addr_ipv6=False) 2025-10-10T01:46:52.1814793Z /opt/conda/envs/py_3.12/lib/python3.12/site-packages/torch/cuda/__init__.py:64: FutureWarning: The pynvml package is deprecated. Please install nvidia-ml-py instead. If you did not install pynvml directly, please report this to the maintainers of the package that installed pynvml for you. 2025-10-10T01:46:52.1817184Z import pynvml # type: ignore[import] 2025-10-10T01:46:52.1819414Z /opt/conda/envs/py_3.12/lib/python3.12/site-packages/torch/cuda/__init__.py:64: FutureWarning: The pynvml package is deprecated. Please install nvidia-ml-py instead. If you did not install pynvml directly, please report this to the maintainers of the package that installed pynvml for you. 2025-10-10T01:46:52.1821594Z import pynvml # type: ignore[import] 2025-10-10T01:46:52.1849287Z /opt/conda/envs/py_3.12/lib/python3.12/site-packages/torch/cuda/__init__.py:64: FutureWarning: The pynvml package is deprecated. Please install nvidia-ml-py instead. If you did not install pynvml directly, please report this to the maintainers of the package that installed pynvml for you. 2025-10-10T01:46:52.1850458Z import pynvml # type: ignore[import] 2025-10-10T01:46:52.3417236Z /opt/conda/envs/py_3.12/lib/python3.12/site-packages/torch/cuda/__init__.py:64: FutureWarning: The pynvml package is deprecated. Please install nvidia-ml-py instead. If you did not install pynvml directly, please report this to the maintainers of the package that installed pynvml for you. 2025-10-10T01:46:52.3418500Z import pynvml # type: ignore[import] 2025-10-10T01:46:54.5284515Z INFO 10-10 01:46:54 [__init__.py:215] Automatically detected platform cuda. 2025-10-10T01:46:54.5560040Z INFO 10-10 01:46:54 [__init__.py:215] Automatically detected platform cuda. 2025-10-10T01:46:54.5839895Z INFO 10-10 01:46:54 [__init__.py:215] Automatically detected platform cuda. 2025-10-10T01:46:54.6822805Z INFO 10-10 01:46:54 [__init__.py:215] Automatically detected platform cuda. 2025-10-10T01:46:59.0740454Z INFO 10-10 01:46:59 [shm_broadcast.py:289] vLLM message queue communication handle: Handle(local_reader_ranks=[0], buffer_handle=(1, 10485760, 10, 'psm_ce5b09ad'), local_subscribe_addr='ipc:///tmp/ce8471da-4532-4e05-b29a-c614eca6b5d6', remote_subscribe_addr=None, remote_addr_ipv6=False) 2025-10-10T01:46:59.3426664Z INFO 10-10 01:46:59 [shm_broadcast.py:289] vLLM message queue communication handle: Handle(local_reader_ranks=[0], buffer_handle=(1, 10485760, 10, 'psm_b075a8bc'), local_subscribe_addr='ipc:///tmp/027493c2-4a69-4f44-b72f-dc07eb752fb1', remote_subscribe_addr=None, remote_addr_ipv6=False) 2025-10-10T01:46:59.5040094Z INFO 10-10 01:46:59 [shm_broadcast.py:289] vLLM message queue communication handle: Handle(local_reader_ranks=[0], buffer_handle=(1, 10485760, 10, 'psm_0f53b068'), local_subscribe_addr='ipc:///tmp/ad81a2d2-30f0-4543-9ab4-1e6879865fad', remote_subscribe_addr=None, remote_addr_ipv6=False) 2025-10-10T01:46:59.5043759Z INFO 10-10 01:46:59 [shm_broadcast.py:289] vLLM message queue communication handle: Handle(local_reader_ranks=[0], buffer_handle=(1, 10485760, 10, 'psm_367bfc83'), local_subscribe_addr='ipc:///tmp/91e45e59-9efc-4fc8-be95-a2d04e6ff8b6', remote_subscribe_addr=None, remote_addr_ipv6=False) 2025-10-10T01:47:00.0433702Z [Gloo] Rank 0 is connected to 3 peer ranks. Expected number of connected peer ranks is : 3 2025-10-10T01:47:00.0435370Z [Gloo] Rank 1 is connected to 3 peer ranks. Expected number of connected peer ranks is : 3 2025-10-10T01:47:00.0435963Z [Gloo] Rank 3 is connected to 3 peer ranks. Expected number of connected peer ranks is : 3 2025-10-10T01:47:00.0436508Z [Gloo] Rank 2 is connected to 3 peer ranks. Expected number of connected peer ranks is : 3 2025-10-10T01:47:00.0483870Z [Gloo] Rank 1 is connected to 1 peer ranks. Expected number of connected peer ranks is : 1 2025-10-10T01:47:00.0484559Z [Gloo] Rank 0 is connected to 1 peer ranks. Expected number of connected peer ranks is : 1 2025-10-10T01:47:00.0485870Z [Gloo] Rank 0 is connected to 1 peer ranks. Expected number of connected peer ranks is : 1 2025-10-10T01:47:00.0486498Z [Gloo] Rank 1 is connected to 1 peer ranks. Expected number of connected peer ranks is : 1 2025-10-10T01:47:00.0527288Z INFO 10-10 01:47:00 [__init__.py:1384] Found nccl from library libnccl.so.2 2025-10-10T01:47:00.0528511Z INFO 10-10 01:47:00 [__init__.py:1384] Found nccl from library libnccl.so.2 2025-10-10T01:47:00.0529652Z INFO 10-10 01:47:00 [__init__.py:1384] Found nccl from library libnccl.so.2 2025-10-10T01:47:00.0530659Z INFO 10-10 01:47:00 [pynccl.py:104] vLLM is using nccl==2.27.5 2025-10-10T01:47:00.0531639Z INFO 10-10 01:47:00 [pynccl.py:104] vLLM is using nccl==2.27.5 2025-10-10T01:47:00.0532350Z INFO 10-10 01:47:00 [pynccl.py:104] vLLM is using nccl==2.27.5 2025-10-10T01:47:00.0533131Z INFO 10-10 01:47:00 [__init__.py:1384] Found nccl from library libnccl.so.2 2025-10-10T01:47:00.0533937Z INFO 10-10 01:47:00 [pynccl.py:104] vLLM is using nccl==2.27.5 2025-10-10T01:47:00.2429430Z WARNING 10-10 01:47:00 [symm_mem.py:58] SymmMemCommunicator: Device capability 8.9 not supported, communicator is not available. 2025-10-10T01:47:00.2431709Z WARNING 10-10 01:47:00 [symm_mem.py:58] SymmMemCommunicator: Device capability 8.9 not supported, communicator is not available. 2025-10-10T01:47:00.2448738Z INFO 10-10 01:47:00 [custom_all_reduce.py:35] Skipping P2P check and trusting the driver's P2P report. 2025-10-10T01:47:00.2449405Z INFO 10-10 01:47:00 [custom_all_reduce.py:35] Skipping P2P check and trusting the driver's P2P report. 2025-10-10T01:47:00.2450439Z WARNING 10-10 01:47:00 [custom_all_reduce.py:154] Custom allreduce is disabled because your platform lacks GPU P2P capability or P2P test failed. To silence this warning, specify disable_custom_all_reduce=True explicitly. 2025-10-10T01:47:00.2451832Z WARNING 10-10 01:47:00 [custom_all_reduce.py:154] Custom allreduce is disabled because your platform lacks GPU P2P capability or P2P test failed. To silence this warning, specify disable_custom_all_reduce=True explicitly. 2025-10-10T01:47:00.2468647Z INFO 10-10 01:47:00 [shm_broadcast.py:289] vLLM message queue communication handle: Handle(local_reader_ranks=[1], buffer_handle=(1, 4194304, 6, 'psm_5e0e6399'), local_subscribe_addr='ipc:///tmp/00e7bf98-22bc-442a-8ca2-7fa3a5d09f2f', remote_subscribe_addr=None, remote_addr_ipv6=False) 2025-10-10T01:47:00.3333699Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-10-10T01:47:00.3334978Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-10-10T01:47:00.3504221Z WARNING 10-10 01:47:00 [symm_mem.py:58] SymmMemCommunicator: Device capability 8.9 not supported, communicator is not available. 2025-10-10T01:47:00.3505264Z WARNING 10-10 01:47:00 [symm_mem.py:58] SymmMemCommunicator: Device capability 8.9 not supported, communicator is not available. 2025-10-10T01:47:00.3531292Z INFO 10-10 01:47:00 [custom_all_reduce.py:35] Skipping P2P check and trusting the driver's P2P report. 2025-10-10T01:47:00.3532024Z INFO 10-10 01:47:00 [custom_all_reduce.py:35] Skipping P2P check and trusting the driver's P2P report. 2025-10-10T01:47:00.3533050Z WARNING 10-10 01:47:00 [custom_all_reduce.py:154] Custom allreduce is disabled because your platform lacks GPU P2P capability or P2P test failed. To silence this warning, specify disable_custom_all_reduce=True explicitly. 2025-10-10T01:47:00.3534591Z WARNING 10-10 01:47:00 [custom_all_reduce.py:154] Custom allreduce is disabled because your platform lacks GPU P2P capability or P2P test failed. To silence this warning, specify disable_custom_all_reduce=True explicitly. 2025-10-10T01:47:00.3549931Z INFO 10-10 01:47:00 [shm_broadcast.py:289] vLLM message queue communication handle: Handle(local_reader_ranks=[1], buffer_handle=(1, 4194304, 6, 'psm_44e012e3'), local_subscribe_addr='ipc:///tmp/c172c773-ece1-44af-810f-c6fcd8a8317c', remote_subscribe_addr=None, remote_addr_ipv6=False) 2025-10-10T01:47:00.3571001Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-10-10T01:47:00.3584665Z [Gloo] Rank 1 is connected to 1 peer ranks. Expected number of connected peer ranks is : 1 2025-10-10T01:47:00.3585456Z [Gloo] Rank 0 is connected to 1 peer ranks. Expected number of connected peer ranks is : 1 2025-10-10T01:47:00.3587445Z INFO 10-10 01:47:00 [__init__.py:1384] Found nccl from library libnccl.so.2 2025-10-10T01:47:00.3588047Z INFO 10-10 01:47:00 [pynccl.py:104] vLLM is using nccl==2.27.5 2025-10-10T01:47:00.3591162Z INFO 10-10 01:47:00 [__init__.py:1384] Found nccl from library libnccl.so.2 2025-10-10T01:47:00.3591725Z INFO 10-10 01:47:00 [pynccl.py:104] vLLM is using nccl==2.27.5 2025-10-10T01:47:00.3940325Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-10-10T01:47:00.3941288Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-10-10T01:47:00.4837083Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-10-10T01:47:00.4851710Z [Gloo] Rank 0 is connected to 1 peer ranks. Expected number of connected peer ranks is : 1 2025-10-10T01:47:00.4852534Z [Gloo] Rank 1 is connected to 1 peer ranks. Expected number of connected peer ranks is : 1 2025-10-10T01:47:00.4853561Z INFO 10-10 01:47:00 [__init__.py:1384] Found nccl from library libnccl.so.2 2025-10-10T01:47:00.4854152Z INFO 10-10 01:47:00 [pynccl.py:104] vLLM is using nccl==2.27.5 2025-10-10T01:47:00.4856320Z INFO 10-10 01:47:00 [__init__.py:1384] Found nccl from library libnccl.so.2 2025-10-10T01:47:00.4857053Z INFO 10-10 01:47:00 [pynccl.py:104] vLLM is using nccl==2.27.5 2025-10-10T01:47:00.5190534Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-10-10T01:47:00.5191515Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-10-10T01:47:00.5201210Z [Gloo] Rank 1 is connected to 1 peer ranks. Expected number of connected peer ranks is : 1 2025-10-10T01:47:00.5202170Z [Gloo] Rank 0 is connected to 1 peer ranks. Expected number of connected peer ranks is : 1 2025-10-10T01:47:00.5202625Z [Gloo] Rank 1 is connected to 1 peer ranks. Expected number of connected peer ranks is : 1 2025-10-10T01:47:00.5203064Z [Gloo] Rank 0 is connected to 1 peer ranks. Expected number of connected peer ranks is : 1 2025-10-10T01:47:00.5203927Z INFO 10-10 01:47:00 [__init__.py:1384] Found nccl from library libnccl.so.2 2025-10-10T01:47:00.5204553Z INFO 10-10 01:47:00 [pynccl.py:104] vLLM is using nccl==2.27.5 2025-10-10T01:47:00.5205232Z INFO 10-10 01:47:00 [__init__.py:1384] Found nccl from library libnccl.so.2 2025-10-10T01:47:00.5205955Z INFO 10-10 01:47:00 [pynccl.py:104] vLLM is using nccl==2.27.5 2025-10-10T01:47:00.5206419Z INFO 10-10 01:47:00 [__init__.py:1384] Found nccl from library libnccl.so.2 2025-10-10T01:47:00.5207031Z INFO 10-10 01:47:00 [pynccl.py:104] vLLM is using nccl==2.27.5 2025-10-10T01:47:00.5207968Z INFO 10-10 01:47:00 [__init__.py:1384] Found nccl from library libnccl.so.2 2025-10-10T01:47:00.5208646Z INFO 10-10 01:47:00 [pynccl.py:104] vLLM is using nccl==2.27.5 2025-10-10T01:47:00.5642433Z INFO 10-10 01:47:00 [parallel_state.py:1208] rank 2 in world size 4 is assigned as DP rank 0, PP rank 1, TP rank 0, EP rank 0 2025-10-10T01:47:00.5643481Z INFO 10-10 01:47:00 [parallel_state.py:1208] rank 3 in world size 4 is assigned as DP rank 0, PP rank 1, TP rank 1, EP rank 1 2025-10-10T01:47:00.5649997Z INFO 10-10 01:47:00 [parallel_state.py:1208] rank 1 in world size 4 is assigned as DP rank 0, PP rank 0, TP rank 1, EP rank 1 2025-10-10T01:47:00.5651064Z INFO 10-10 01:47:00 [parallel_state.py:1208] rank 0 in world size 4 is assigned as DP rank 0, PP rank 0, TP rank 0, EP rank 0 2025-10-10T01:47:00.8291243Z INFO 10-10 01:47:00 [topk_topp_sampler.py:55] Using FlashInfer for top-p & top-k sampling. 2025-10-10T01:47:00.8383735Z INFO 10-10 01:47:00 [topk_topp_sampler.py:55] Using FlashInfer for top-p & top-k sampling. 2025-10-10T01:47:00.8395973Z INFO 10-10 01:47:00 [topk_topp_sampler.py:55] Using FlashInfer for top-p & top-k sampling. 2025-10-10T01:47:00.8405631Z INFO 10-10 01:47:00 [topk_topp_sampler.py:55] Using FlashInfer for top-p & top-k sampling. 2025-10-10T01:47:00.8593422Z (Worker_PP1_TP1 pid=48015) INFO 10-10 01:47:00 [gpu_model_runner.py:2707] Starting to load model meta-llama/Llama-3.2-1B-Instruct... 2025-10-10T01:47:00.8684192Z (Worker_PP0_TP1 pid=48013) INFO 10-10 01:47:00 [gpu_model_runner.py:2707] Starting to load model meta-llama/Llama-3.2-1B-Instruct... 2025-10-10T01:47:00.8705907Z (Worker_PP1_TP0 pid=48014) INFO 10-10 01:47:00 [gpu_model_runner.py:2707] Starting to load model meta-llama/Llama-3.2-1B-Instruct... 2025-10-10T01:47:00.8712342Z (Worker_PP0_TP0 pid=48012) INFO 10-10 01:47:00 [gpu_model_runner.py:2707] Starting to load model meta-llama/Llama-3.2-1B-Instruct... 2025-10-10T01:47:01.0858521Z (Worker_PP1_TP1 pid=48015) INFO 10-10 01:47:01 [gpu_model_runner.py:2739] Loading model from scratch... 2025-10-10T01:47:01.0987620Z (Worker_PP0_TP1 pid=48013) INFO 10-10 01:47:01 [gpu_model_runner.py:2739] Loading model from scratch... 2025-10-10T01:47:01.1347843Z (Worker_PP1_TP0 pid=48014) INFO 10-10 01:47:01 [gpu_model_runner.py:2739] Loading model from scratch... 2025-10-10T01:47:01.1379150Z (Worker_PP0_TP0 pid=48012) INFO 10-10 01:47:01 [gpu_model_runner.py:2739] Loading model from scratch... 2025-10-10T01:47:01.1760500Z (Worker_PP1_TP1 pid=48015) INFO 10-10 01:47:01 [cuda.py:361] Using Flash Attention backend on V1 engine. 2025-10-10T01:47:01.1894779Z (Worker_PP0_TP1 pid=48013) INFO 10-10 01:47:01 [cuda.py:361] Using Flash Attention backend on V1 engine. 2025-10-10T01:47:01.2253009Z (Worker_PP1_TP0 pid=48014) INFO 10-10 01:47:01 [cuda.py:361] Using Flash Attention backend on V1 engine. 2025-10-10T01:47:01.2272021Z (Worker_PP0_TP0 pid=48012) INFO 10-10 01:47:01 [cuda.py:361] Using Flash Attention backend on V1 engine. 2025-10-10T01:47:01.6907221Z (Worker_PP1_TP1 pid=48015) INFO 10-10 01:47:01 [gpu_model_runner.py:2758] Model loading took 0.7135 GiB and 0.153505 seconds 2025-10-10T01:47:01.7055222Z (Worker_PP0_TP1 pid=48013) INFO 10-10 01:47:01 [gpu_model_runner.py:2758] Model loading took 0.7135 GiB and 0.154244 seconds 2025-10-10T01:47:01.8101891Z (Worker_PP0_TP0 pid=48012) INFO 10-10 01:47:01 [gpu_model_runner.py:2758] Model loading took 0.7135 GiB and 0.151648 seconds 2025-10-10T01:47:01.8120700Z (Worker_PP1_TP0 pid=48014) INFO 10-10 01:47:01 [gpu_model_runner.py:2758] Model loading took 0.7135 GiB and 0.155349 seconds 2025-10-10T01:47:03.9840189Z (Worker_PP1_TP1 pid=48015) INFO 10-10 01:47:03 [backends.py:548] Using cache directory: /var/lib/jenkins/.cache/vllm/torch_compile_cache/8cedb68d99/rank_3_0/backbone for vLLM's torch.compile 2025-10-10T01:47:03.9846667Z (Worker_PP1_TP1 pid=48015) INFO 10-10 01:47:03 [backends.py:559] Dynamo bytecode transform time: 1.95 s 2025-10-10T01:47:04.0703855Z (Worker_PP1_TP0 pid=48014) INFO 10-10 01:47:04 [backends.py:548] Using cache directory: /var/lib/jenkins/.cache/vllm/torch_compile_cache/8cedb68d99/rank_2_0/backbone for vLLM's torch.compile 2025-10-10T01:47:04.0711327Z (Worker_PP1_TP0 pid=48014) INFO 10-10 01:47:04 [backends.py:559] Dynamo bytecode transform time: 2.00 s 2025-10-10T01:47:04.0814703Z (Worker_PP0_TP1 pid=48013) INFO 10-10 01:47:04 [backends.py:548] Using cache directory: /var/lib/jenkins/.cache/vllm/torch_compile_cache/5280ab4e2b/rank_1_0/backbone for vLLM's torch.compile 2025-10-10T01:47:04.0821943Z (Worker_PP0_TP1 pid=48013) INFO 10-10 01:47:04 [backends.py:559] Dynamo bytecode transform time: 2.05 s 2025-10-10T01:47:04.1341681Z (Worker_PP0_TP0 pid=48012) INFO 10-10 01:47:04 [backends.py:548] Using cache directory: /var/lib/jenkins/.cache/vllm/torch_compile_cache/5280ab4e2b/rank_0_0/backbone for vLLM's torch.compile 2025-10-10T01:47:04.1348722Z (Worker_PP0_TP0 pid=48012) INFO 10-10 01:47:04 [backends.py:559] Dynamo bytecode transform time: 2.06 s 2025-10-10T01:47:04.5891026Z (Worker_PP1_TP1 pid=48015) INFO 10-10 01:47:04 [backends.py:164] Directly load the compiled graph(s) for dynamic shape from the cache, took 0.410 s 2025-10-10T01:47:04.6807384Z (Worker_PP0_TP1 pid=48013) INFO 10-10 01:47:04 [backends.py:164] Directly load the compiled graph(s) for dynamic shape from the cache, took 0.393 s 2025-10-10T01:47:04.6864169Z (Worker_PP1_TP0 pid=48014) INFO 10-10 01:47:04 [backends.py:164] Directly load the compiled graph(s) for dynamic shape from the cache, took 0.415 s 2025-10-10T01:47:04.7578643Z (Worker_PP0_TP0 pid=48012) INFO 10-10 01:47:04 [backends.py:164] Directly load the compiled graph(s) for dynamic shape from the cache, took 0.411 s 2025-10-10T01:47:05.6704724Z (Worker_PP0_TP1 pid=48013) INFO 10-10 01:47:05 [gpu_worker.py:298] Available KV cache memory: 19.02 GiB 2025-10-10T01:47:05.6822688Z (Worker_PP0_TP0 pid=48012) INFO 10-10 01:47:05 [gpu_worker.py:298] Available KV cache memory: 19.02 GiB 2025-10-10T01:47:05.7329096Z (Worker_PP1_TP0 pid=48014) INFO 10-10 01:47:05 [gpu_worker.py:298] Available KV cache memory: 18.99 GiB 2025-10-10T01:47:05.7351578Z (Worker_PP1_TP1 pid=48015) INFO 10-10 01:47:05 [gpu_worker.py:298] Available KV cache memory: 18.99 GiB 2025-10-10T01:47:06.0295795Z (EngineCore_DP0 pid=47893) INFO 10-10 01:47:06 [kv_cache_utils.py:1087] GPU KV cache size: 2,492,416 tokens 2025-10-10T01:47:06.0296760Z (EngineCore_DP0 pid=47893) INFO 10-10 01:47:06 [kv_cache_utils.py:1091] Maximum concurrency for 2,048 tokens per request: 1217.00x 2025-10-10T01:47:06.0297738Z (EngineCore_DP0 pid=47893) INFO 10-10 01:47:06 [kv_cache_utils.py:1087] GPU KV cache size: 2,492,416 tokens 2025-10-10T01:47:06.0298552Z (EngineCore_DP0 pid=47893) INFO 10-10 01:47:06 [kv_cache_utils.py:1091] Maximum concurrency for 2,048 tokens per request: 1217.00x 2025-10-10T01:47:06.0299542Z (EngineCore_DP0 pid=47893) INFO 10-10 01:47:06 [kv_cache_utils.py:1087] GPU KV cache size: 2,489,088 tokens 2025-10-10T01:47:06.0300211Z (EngineCore_DP0 pid=47893) INFO 10-10 01:47:06 [kv_cache_utils.py:1091] Maximum concurrency for 2,048 tokens per request: 1215.38x 2025-10-10T01:47:06.0300912Z (EngineCore_DP0 pid=47893) INFO 10-10 01:47:06 [kv_cache_utils.py:1087] GPU KV cache size: 2,489,088 tokens 2025-10-10T01:47:06.0301578Z (EngineCore_DP0 pid=47893) INFO 10-10 01:47:06 [kv_cache_utils.py:1091] Maximum concurrency for 2,048 tokens per request: 1215.38x 2025-10-10T01:47:06.0371181Z (Worker_PP0_TP0 pid=48012) INFO 10-10 01:47:06 [gpu_worker.py:332] Compile and warming up model for size 8 2025-10-10T01:47:06.0372218Z (Worker_PP0_TP1 pid=48013) INFO 10-10 01:47:06 [gpu_worker.py:332] Compile and warming up model for size 8 2025-10-10T01:47:06.0372979Z (Worker_PP1_TP1 pid=48015) INFO 10-10 01:47:06 [gpu_worker.py:332] Compile and warming up model for size 8 2025-10-10T01:47:06.0373757Z (Worker_PP1_TP0 pid=48014) INFO 10-10 01:47:06 [gpu_worker.py:332] Compile and warming up model for size 8 2025-10-10T01:47:06.4445856Z (Worker_PP1_TP0 pid=48014) INFO 10-10 01:47:06 [backends.py:168] Directly load the compiled graph(s) for shape 8 from the cache, took 0.407 s 2025-10-10T01:47:06.4462308Z (Worker_PP1_TP1 pid=48015) INFO 10-10 01:47:06 [backends.py:168] Directly load the compiled graph(s) for shape 8 from the cache, took 0.409 s 2025-10-10T01:47:06.4468845Z (Worker_PP1_TP0 pid=48014) INFO 10-10 01:47:06 [gpu_worker.py:332] Compile and warming up model for size 4 2025-10-10T01:47:06.4469679Z (Worker_PP1_TP1 pid=48015) INFO 10-10 01:47:06 [gpu_worker.py:332] Compile and warming up model for size 4 2025-10-10T01:47:06.4830874Z (Worker_PP0_TP0 pid=48012) INFO 10-10 01:47:06 [backends.py:168] Directly load the compiled graph(s) for shape 8 from the cache, took 0.444 s 2025-10-10T01:47:06.4831806Z (Worker_PP0_TP0 pid=48012) INFO 10-10 01:47:06 [gpu_worker.py:332] Compile and warming up model for size 4 2025-10-10T01:47:06.4873456Z (Worker_PP0_TP1 pid=48013) INFO 10-10 01:47:06 [backends.py:168] Directly load the compiled graph(s) for shape 8 from the cache, took 0.450 s 2025-10-10T01:47:06.4879095Z (Worker_PP0_TP1 pid=48013) INFO 10-10 01:47:06 [gpu_worker.py:332] Compile and warming up model for size 4 2025-10-10T01:47:07.3417980Z (Worker_PP0_TP1 pid=48013) INFO 10-10 01:47:07 [backends.py:168] Directly load the compiled graph(s) for shape 4 from the cache, took 0.853 s 2025-10-10T01:47:07.3418922Z (Worker_PP0_TP1 pid=48013) INFO 10-10 01:47:07 [monitor.py:32] torch.compile takes 2.05 s in total 2025-10-10T01:47:07.3470585Z (Worker_PP0_TP0 pid=48012) INFO 10-10 01:47:07 [backends.py:168] Directly load the compiled graph(s) for shape 4 from the cache, took 0.864 s 2025-10-10T01:47:07.3471490Z (Worker_PP0_TP0 pid=48012) INFO 10-10 01:47:07 [monitor.py:32] torch.compile takes 2.06 s in total 2025-10-10T01:47:07.6519027Z (Worker_PP1_TP1 pid=48015) INFO 10-10 01:47:07 [backends.py:168] Directly load the compiled graph(s) for shape 4 from the cache, took 1.204 s 2025-10-10T01:47:07.6520003Z (Worker_PP1_TP1 pid=48015) INFO 10-10 01:47:07 [monitor.py:32] torch.compile takes 1.95 s in total 2025-10-10T01:47:07.6520877Z (Worker_PP1_TP0 pid=48014) INFO 10-10 01:47:07 [backends.py:168] Directly load the compiled graph(s) for shape 4 from the cache, took 1.205 s 2025-10-10T01:47:07.6521725Z (Worker_PP1_TP0 pid=48014) INFO 10-10 01:47:07 [monitor.py:32] torch.compile takes 2.00 s in total 2025-10-10T01:47:07.6596803Z (EngineCore_DP0 pid=47893) INFO 10-10 01:47:07 [core.py:211] init engine (profile, create kv cache, warmup model) took 5.85 seconds 2025-10-10T01:47:08.4116982Z (EngineCore_DP0 pid=47893) INFO 10-10 01:47:08 [core.py:150] Batch queue is enabled with size 2 2025-10-10T01:47:08.4146458Z (EngineCore_DP0 pid=47893) INFO 10-10 01:47:08 [vllm.py:345] Cudagraph is disabled under eager mode 2025-10-10T01:47:08.5995092Z (APIServer pid=47771) INFO 10-10 01:47:08 [loggers.py:147] Engine 000: vllm cache_config_info with initialization after num_gpu_blocks is: 155568 2025-10-10T01:47:08.6593111Z (EngineCore_DP0 pid=47893) INFO 10-10 01:47:08 [gc_utils.py:41] GC Debug Config. enabled:False,top_objects:-1 2025-10-10T01:47:08.6895919Z (APIServer pid=47771) INFO 10-10 01:47:08 [api_server.py:1634] Supported_tasks: ['generate'] 2025-10-10T01:47:08.7215513Z (APIServer pid=47771) WARNING 10-10 01:47:08 [model.py:1424] Default sampling parameters have been overridden by the model's Hugging Face generation config recommended from the model creator. If this is not intended, please relaunch vLLM instance with `--generation-config vllm`. 2025-10-10T01:47:08.7217140Z (APIServer pid=47771) INFO 10-10 01:47:08 [serving_responses.py:140] Using default chat sampling params from model: {'temperature': 0.6, 'top_p': 0.9} 2025-10-10T01:47:08.7558780Z (APIServer pid=47771) INFO 10-10 01:47:08 [serving_chat.py:139] Using default chat sampling params from model: {'temperature': 0.6, 'top_p': 0.9} 2025-10-10T01:47:08.7750597Z (APIServer pid=47771) INFO 10-10 01:47:08 [serving_completion.py:76] Using default completion sampling params from model: {'temperature': 0.6, 'top_p': 0.9} 2025-10-10T01:47:08.7751820Z (APIServer pid=47771) INFO 10-10 01:47:08 [api_server.py:1912] Starting vLLM API server 0 on http://0.0.0.0:46161 2025-10-10T01:47:08.7752589Z (APIServer pid=47771) INFO 10-10 01:47:08 [launcher.py:34] Available routes are: 2025-10-10T01:47:08.7753293Z (APIServer pid=47771) INFO 10-10 01:47:08 [launcher.py:42] Route: /openapi.json, Methods: HEAD, GET 2025-10-10T01:47:08.7754020Z (APIServer pid=47771) INFO 10-10 01:47:08 [launcher.py:42] Route: /docs, Methods: HEAD, GET 2025-10-10T01:47:08.7754754Z (APIServer pid=47771) INFO 10-10 01:47:08 [launcher.py:42] Route: /docs/oauth2-redirect, Methods: HEAD, GET 2025-10-10T01:47:08.7755480Z (APIServer pid=47771) INFO 10-10 01:47:08 [launcher.py:42] Route: /redoc, Methods: HEAD, GET 2025-10-10T01:47:08.7756140Z (APIServer pid=47771) INFO 10-10 01:47:08 [launcher.py:42] Route: /health, Methods: GET 2025-10-10T01:47:08.7756788Z (APIServer pid=47771) INFO 10-10 01:47:08 [launcher.py:42] Route: /load, Methods: GET 2025-10-10T01:47:08.7757435Z (APIServer pid=47771) INFO 10-10 01:47:08 [launcher.py:42] Route: /ping, Methods: POST 2025-10-10T01:47:08.7757968Z (APIServer pid=47771) INFO 10-10 01:47:08 [launcher.py:42] Route: /ping, Methods: GET 2025-10-10T01:47:08.7758592Z (APIServer pid=47771) INFO 10-10 01:47:08 [launcher.py:42] Route: /tokenize, Methods: POST 2025-10-10T01:47:08.7759150Z (APIServer pid=47771) INFO 10-10 01:47:08 [launcher.py:42] Route: /detokenize, Methods: POST 2025-10-10T01:47:08.7759685Z (APIServer pid=47771) INFO 10-10 01:47:08 [launcher.py:42] Route: /v1/models, Methods: GET 2025-10-10T01:47:08.7760213Z (APIServer pid=47771) INFO 10-10 01:47:08 [launcher.py:42] Route: /version, Methods: GET 2025-10-10T01:47:08.7760756Z (APIServer pid=47771) INFO 10-10 01:47:08 [launcher.py:42] Route: /v1/responses, Methods: POST 2025-10-10T01:47:08.7761330Z (APIServer pid=47771) INFO 10-10 01:47:08 [launcher.py:42] Route: /v1/responses/{response_id}, Methods: GET 2025-10-10T01:47:08.7761975Z (APIServer pid=47771) INFO 10-10 01:47:08 [launcher.py:42] Route: /v1/responses/{response_id}/cancel, Methods: POST 2025-10-10T01:47:08.7762730Z (APIServer pid=47771) INFO 10-10 01:47:08 [launcher.py:42] Route: /v1/chat/completions, Methods: POST 2025-10-10T01:47:08.7763319Z (APIServer pid=47771) INFO 10-10 01:47:08 [launcher.py:42] Route: /v1/completions, Methods: POST 2025-10-10T01:47:08.7763886Z (APIServer pid=47771) INFO 10-10 01:47:08 [launcher.py:42] Route: /v1/embeddings, Methods: POST 2025-10-10T01:47:08.7764433Z (APIServer pid=47771) INFO 10-10 01:47:08 [launcher.py:42] Route: /pooling, Methods: POST 2025-10-10T01:47:08.7764962Z (APIServer pid=47771) INFO 10-10 01:47:08 [launcher.py:42] Route: /classify, Methods: POST 2025-10-10T01:47:08.7765499Z (APIServer pid=47771) INFO 10-10 01:47:08 [launcher.py:42] Route: /score, Methods: POST 2025-10-10T01:47:08.7766016Z (APIServer pid=47771) INFO 10-10 01:47:08 [launcher.py:42] Route: /v1/score, Methods: POST 2025-10-10T01:47:08.7766704Z (APIServer pid=47771) INFO 10-10 01:47:08 [launcher.py:42] Route: /v1/audio/transcriptions, Methods: POST 2025-10-10T01:47:08.7767323Z (APIServer pid=47771) INFO 10-10 01:47:08 [launcher.py:42] Route: /v1/audio/translations, Methods: POST 2025-10-10T01:47:08.7767880Z (APIServer pid=47771) INFO 10-10 01:47:08 [launcher.py:42] Route: /rerank, Methods: POST 2025-10-10T01:47:08.7768417Z (APIServer pid=47771) INFO 10-10 01:47:08 [launcher.py:42] Route: /v1/rerank, Methods: POST 2025-10-10T01:47:08.7769022Z (APIServer pid=47771) INFO 10-10 01:47:08 [launcher.py:42] Route: /v2/rerank, Methods: POST 2025-10-10T01:47:08.7769568Z (APIServer pid=47771) INFO 10-10 01:47:08 [launcher.py:42] Route: /scale_elastic_ep, Methods: POST 2025-10-10T01:47:08.7770151Z (APIServer pid=47771) INFO 10-10 01:47:08 [launcher.py:42] Route: /is_scaling_elastic_ep, Methods: POST 2025-10-10T01:47:08.7770716Z (APIServer pid=47771) INFO 10-10 01:47:08 [launcher.py:42] Route: /invocations, Methods: POST 2025-10-10T01:47:08.7771265Z (APIServer pid=47771) INFO 10-10 01:47:08 [launcher.py:42] Route: /metrics, Methods: GET 2025-10-10T01:47:08.7975035Z (APIServer pid=47771) INFO: Started server process [47771] 2025-10-10T01:47:08.7975708Z (APIServer pid=47771) INFO: Waiting for application startup. 2025-10-10T01:47:09.0278894Z (APIServer pid=47771) INFO: Application startup complete. 2025-10-10T01:47:09.5023435Z (APIServer pid=47771) INFO: 127.0.0.1:49124 - "GET /health HTTP/1.1" 200 OK 2025-10-10T01:47:09.7190460Z (APIServer pid=47771) INFO: 127.0.0.1:49132 - "GET /v1/models HTTP/1.1" 200 OK 2025-10-10T01:47:09.7499122Z (Worker_PP0_TP1 pid=48013) /opt/conda/envs/py_3.12/lib/python3.12/site-packages/vllm/distributed/parallel_state.py:516: UserWarning: The given buffer is not writable, and PyTorch does not support non-writable tensors. This means you can write to the underlying (supposedly non-writable) buffer using the tensor. You may want to copy the buffer to protect its data or make it writable before converting it to a tensor. This type of warning will be suppressed for the rest of this program. (Triggered internally at /var/lib/jenkins/workspace/torch/csrc/utils/tensor_new.cpp:1581.) 2025-10-10T01:47:09.7502049Z (Worker_PP0_TP1 pid=48013) object_tensor = torch.frombuffer(pickle.dumps(obj), dtype=torch.uint8) 2025-10-10T01:47:09.7505299Z (Worker_PP0_TP0 pid=48012) /opt/conda/envs/py_3.12/lib/python3.12/site-packages/vllm/distributed/parallel_state.py:516: UserWarning: The given buffer is not writable, and PyTorch does not support non-writable tensors. This means you can write to the underlying (supposedly non-writable) buffer using the tensor. You may want to copy the buffer to protect its data or make it writable before converting it to a tensor. This type of warning will be suppressed for the rest of this program. (Triggered internally at /var/lib/jenkins/workspace/torch/csrc/utils/tensor_new.cpp:1581.) 2025-10-10T01:47:09.7508311Z (Worker_PP0_TP0 pid=48012) object_tensor = torch.frombuffer(pickle.dumps(obj), dtype=torch.uint8) 2025-10-10T01:47:09.7509441Z [rank1]:[W1010 01:47:09.427396059 ProcessGroupNCCL.cpp:4063] Warning: An unbatched P2P op (send/recv) was called on this ProcessGroup with size 2. In lazy initialization mode, this will result in a new 2-rank NCCL communicator to be created. (function operator()) 2025-10-10T01:47:09.7510958Z [rank3]:[W1010 01:47:09.427508282 ProcessGroupNCCL.cpp:4063] Warning: An unbatched P2P op (send/recv) was called on this ProcessGroup with size 2. In lazy initialization mode, this will result in a new 2-rank NCCL communicator to be created. (function operator()) 2025-10-10T01:47:09.7512673Z [rank0]:[W1010 01:47:09.427730968 ProcessGroupNCCL.cpp:4063] Warning: An unbatched P2P op (send/recv) was called on this ProcessGroup with size 2. In lazy initialization mode, this will result in a new 2-rank NCCL communicator to be created. (function operator()) 2025-10-10T01:47:09.7514178Z [rank2]:[W1010 01:47:09.427859351 ProcessGroupNCCL.cpp:4063] Warning: An unbatched P2P op (send/recv) was called on this ProcessGroup with size 2. In lazy initialization mode, this will result in a new 2-rank NCCL communicator to be created. (function operator()) 2025-10-10T01:47:09.8761966Z (APIServer pid=47771) INFO: 127.0.0.1:49132 - "POST /v1/completions HTTP/1.1" 200 OK 2025-10-10T01:47:09.9410403Z (APIServer pid=47771) INFO: 127.0.0.1:49132 - "POST /v1/completions HTTP/1.1" 200 OK 2025-10-10T01:47:10.0839503Z (APIServer pid=47771) INFO: 127.0.0.1:49132 - "POST /v1/completions HTTP/1.1" 200 OK 2025-10-10T01:47:10.1751356Z (APIServer pid=47771) INFO: 127.0.0.1:49132 - "POST /v1/completions HTTP/1.1" 200 OK 2025-10-10T01:47:10.2516744Z (APIServer pid=47771) INFO: 127.0.0.1:49132 - "POST /v1/completions HTTP/1.1" 200 OK 2025-10-10T01:47:10.2594121Z (APIServer pid=47771) INFO: 127.0.0.1:49132 - "POST /v1/completions HTTP/1.1" 200 OK 2025-10-10T01:47:10.3447425Z (APIServer pid=47771) INFO 10-10 01:47:10 [launcher.py:99] Shutting down FastAPI HTTP server. 2025-10-10T01:47:10.3451523Z (Worker_PP0_TP0 pid=48012) INFO 10-10 01:47:10 [multiproc_executor.py:558] Parent process exited, terminating worker 2025-10-10T01:47:10.3452389Z (Worker_PP0_TP0 pid=48012) INFO 10-10 01:47:10 [multiproc_executor.py:599] WorkerProc shutting down. 2025-10-10T01:47:10.3455064Z (Worker_PP0_TP1 pid=48013) INFO 10-10 01:47:10 [multiproc_executor.py:558] Parent process exited, terminating worker 2025-10-10T01:47:10.3456582Z (Worker_PP0_TP1 pid=48013) INFO 10-10 01:47:10 [multiproc_executor.py:599] WorkerProc shutting down. 2025-10-10T01:47:10.3457952Z (Worker_PP1_TP0 pid=48014) INFO 10-10 01:47:10 [multiproc_executor.py:558] Parent process exited, terminating worker 2025-10-10T01:47:10.3459882Z (Worker_PP1_TP0 pid=48014) INFO 10-10 01:47:10 [multiproc_executor.py:599] WorkerProc shutting down. 2025-10-10T01:47:10.3461290Z (Worker_PP1_TP1 pid=48015) INFO 10-10 01:47:10 [multiproc_executor.py:558] Parent process exited, terminating worker 2025-10-10T01:47:10.3463389Z (Worker_PP1_TP1 pid=48015) INFO 10-10 01:47:10 [multiproc_executor.py:599] WorkerProc shutting down. 2025-10-10T01:47:13.0669495Z (APIServer pid=47771) INFO: Shutting down 2025-10-10T01:47:13.1663463Z (APIServer pid=47771) INFO: Waiting for application shutdown. 2025-10-10T01:47:13.1664735Z (APIServer pid=47771) INFO: Application shutdown complete. 2025-10-10T01:47:13.5342738Z INFO 10-10 01:47:13 [model.py:551] Resolved architecture: LlamaForCausalLM 2025-10-10T01:47:13.5344166Z WARNING 10-10 01:47:13 [model.py:1768] Casting torch.bfloat16 to torch.float16. 2025-10-10T01:47:13.5345066Z INFO 10-10 01:47:13 [model.py:1545] Using max model len 2048 2025-10-10T01:47:13.5347327Z Launching RemoteOpenAIServer with: vllm serve meta-llama/Llama-3.2-1B-Instruct --dtype float16 --max-model-len 2048 --max-num-seqs 8 --enforce-eager --tokenizer-mode auto --tensor-parallel-size 2 --distributed-executor-backend mp --load-format dummy --port 37267 --seed 0 2025-10-10T01:47:14.0484797Z /opt/conda/envs/py_3.12/lib/python3.12/site-packages/torch/cuda/__init__.py:64: FutureWarning: The pynvml package is deprecated. Please install nvidia-ml-py instead. If you did not install pynvml directly, please report this to the maintainers of the package that installed pynvml for you. 2025-10-10T01:47:14.0487085Z import pynvml # type: ignore[import] 2025-10-10T01:47:16.2588220Z INFO 10-10 01:47:16 [__init__.py:215] Automatically detected platform cuda. 2025-10-10T01:47:20.1719567Z (APIServer pid=48454) INFO 10-10 01:47:20 [api_server.py:1839] vLLM API server version 0.11.0rc2.dev157+g0ad9951c4.d20251010 2025-10-10T01:47:20.1741137Z (APIServer pid=48454) INFO 10-10 01:47:20 [utils.py:233] non-default args: {'model_tag': 'meta-llama/Llama-3.2-1B-Instruct', 'port': 37267, 'model': 'meta-llama/Llama-3.2-1B-Instruct', 'dtype': 'float16', 'seed': 0, 'max_model_len': 2048, 'enforce_eager': True, 'load_format': 'dummy', 'distributed_executor_backend': 'mp', 'tensor_parallel_size': 2, 'max_num_seqs': 8} 2025-10-10T01:47:20.3242663Z (APIServer pid=48454) INFO 10-10 01:47:20 [model.py:551] Resolved architecture: LlamaForCausalLM 2025-10-10T01:47:20.3243998Z (APIServer pid=48454) `torch_dtype` is deprecated! Use `dtype` instead! 2025-10-10T01:47:20.3245345Z (APIServer pid=48454) WARNING 10-10 01:47:20 [model.py:1768] Casting torch.bfloat16 to torch.float16. 2025-10-10T01:47:20.3246653Z (APIServer pid=48454) INFO 10-10 01:47:20 [model.py:1545] Using max model len 2048 2025-10-10T01:47:20.4839220Z (APIServer pid=48454) INFO 10-10 01:47:20 [scheduler.py:222] Chunked prefill is enabled with max_num_batched_tokens=2048. 2025-10-10T01:47:20.4845696Z (APIServer pid=48454) INFO 10-10 01:47:20 [vllm.py:345] Cudagraph is disabled under eager mode 2025-10-10T01:47:21.4560511Z /opt/conda/envs/py_3.12/lib/python3.12/site-packages/torch/cuda/__init__.py:64: FutureWarning: The pynvml package is deprecated. Please install nvidia-ml-py instead. If you did not install pynvml directly, please report this to the maintainers of the package that installed pynvml for you. 2025-10-10T01:47:21.4562797Z import pynvml # type: ignore[import] 2025-10-10T01:47:23.7565570Z INFO 10-10 01:47:23 [__init__.py:215] Automatically detected platform cuda. 2025-10-10T01:47:27.7303778Z (EngineCore_DP0 pid=48576) INFO 10-10 01:47:27 [core.py:648] Waiting for init message from front-end. 2025-10-10T01:47:27.7574063Z (EngineCore_DP0 pid=48576) INFO 10-10 01:47:27 [core.py:78] Initializing a V1 LLM engine (v0.11.0rc2.dev157+g0ad9951c4.d20251010) with config: model='meta-llama/Llama-3.2-1B-Instruct', speculative_config=None, tokenizer='meta-llama/Llama-3.2-1B-Instruct', skip_tokenizer_init=False, tokenizer_mode=auto, revision=None, tokenizer_revision=None, trust_remote_code=False, dtype=torch.float16, max_seq_len=2048, download_dir=None, load_format=dummy, tensor_parallel_size=2, pipeline_parallel_size=1, data_parallel_size=1, disable_custom_all_reduce=False, quantization=None, enforce_eager=True, kv_cache_dtype=auto, device_config=cuda, structured_outputs_config=StructuredOutputsConfig(backend='auto', disable_fallback=False, disable_any_whitespace=False, disable_additional_properties=False, reasoning_parser=''), observability_config=ObservabilityConfig(show_hidden_metrics_for_version=None, otlp_traces_endpoint=None, collect_detailed_traces=None), seed=0, served_model_name=meta-llama/Llama-3.2-1B-Instruct, enable_prefix_caching=True, chunked_prefill_enabled=True, pooler_config=None, compilation_config={'level': 0, 'debug_dump_path': None, 'cache_dir': '', 'backend': '', 'custom_ops': [], 'splitting_ops': None, 'use_inductor': True, 'compile_sizes': [], 'inductor_compile_config': {'enable_auto_functionalized_v2': False}, 'inductor_passes': {}, 'cudagraph_mode': , 'use_cudagraph': False, 'cudagraph_num_of_warmups': 0, 'cudagraph_capture_sizes': [], 'cudagraph_copy_inputs': False, 'full_cuda_graph': False, 'use_inductor_graph_partition': False, 'pass_config': {}, 'max_capture_size': 0, 'local_cache_dir': None} 2025-10-10T01:47:27.7583790Z (EngineCore_DP0 pid=48576) WARNING 10-10 01:47:27 [multiproc_executor.py:720] Reducing Torch parallelism from 24 threads to 1 to avoid unnecessary CPU contention. Set OMP_NUM_THREADS in the external environment to tune this value as needed. 2025-10-10T01:47:27.7592222Z (EngineCore_DP0 pid=48576) INFO 10-10 01:47:27 [shm_broadcast.py:289] vLLM message queue communication handle: Handle(local_reader_ranks=[0, 1], buffer_handle=(2, 16777216, 10, 'psm_655f4356'), local_subscribe_addr='ipc:///tmp/c1637d28-734f-4764-ad89-0ac7a2cbf6a5', remote_subscribe_addr=None, remote_addr_ipv6=False) 2025-10-10T01:47:28.2372612Z /opt/conda/envs/py_3.12/lib/python3.12/site-packages/torch/cuda/__init__.py:64: FutureWarning: The pynvml package is deprecated. Please install nvidia-ml-py instead. If you did not install pynvml directly, please report this to the maintainers of the package that installed pynvml for you. 2025-10-10T01:47:28.2374843Z import pynvml # type: ignore[import] 2025-10-10T01:47:28.2375786Z /opt/conda/envs/py_3.12/lib/python3.12/site-packages/torch/cuda/__init__.py:64: FutureWarning: The pynvml package is deprecated. Please install nvidia-ml-py instead. If you did not install pynvml directly, please report this to the maintainers of the package that installed pynvml for you. 2025-10-10T01:47:28.2376711Z import pynvml # type: ignore[import] 2025-10-10T01:47:30.5551068Z INFO 10-10 01:47:30 [__init__.py:215] Automatically detected platform cuda. 2025-10-10T01:47:30.5659235Z INFO 10-10 01:47:30 [__init__.py:215] Automatically detected platform cuda. 2025-10-10T01:47:35.1542815Z INFO 10-10 01:47:35 [shm_broadcast.py:289] vLLM message queue communication handle: Handle(local_reader_ranks=[0], buffer_handle=(1, 10485760, 10, 'psm_aae0a4d8'), local_subscribe_addr='ipc:///tmp/702794cf-6b43-4ec3-9435-e8a518e1c96c', remote_subscribe_addr=None, remote_addr_ipv6=False) 2025-10-10T01:47:35.1739336Z INFO 10-10 01:47:35 [shm_broadcast.py:289] vLLM message queue communication handle: Handle(local_reader_ranks=[0], buffer_handle=(1, 10485760, 10, 'psm_349236e1'), local_subscribe_addr='ipc:///tmp/a9888f99-3f02-4f9f-b81b-864a69aada73', remote_subscribe_addr=None, remote_addr_ipv6=False) 2025-10-10T01:47:35.3706438Z [Gloo] Rank 1 is connected to 1 peer ranks. Expected number of connected peer ranks is : 1 2025-10-10T01:47:35.3707463Z [Gloo] Rank 0 is connected to 1 peer ranks. Expected number of connected peer ranks is : 1 2025-10-10T01:47:35.3748779Z [Gloo] Rank 0 is connected to 1 peer ranks. Expected number of connected peer ranks is : 1 2025-10-10T01:47:35.3749958Z [Gloo] Rank 1 is connected to 1 peer ranks. Expected number of connected peer ranks is : 1 2025-10-10T01:47:35.3791505Z INFO 10-10 01:47:35 [__init__.py:1384] Found nccl from library libnccl.so.2 2025-10-10T01:47:35.3791980Z INFO 10-10 01:47:35 [__init__.py:1384] Found nccl from library libnccl.so.2 2025-10-10T01:47:35.3793153Z INFO 10-10 01:47:35 [pynccl.py:104] vLLM is using nccl==2.27.5 2025-10-10T01:47:35.3793736Z INFO 10-10 01:47:35 [pynccl.py:104] vLLM is using nccl==2.27.5 2025-10-10T01:47:35.5506648Z WARNING 10-10 01:47:35 [symm_mem.py:58] SymmMemCommunicator: Device capability 8.9 not supported, communicator is not available. 2025-10-10T01:47:35.5507737Z WARNING 10-10 01:47:35 [symm_mem.py:58] SymmMemCommunicator: Device capability 8.9 not supported, communicator is not available. 2025-10-10T01:47:35.5528912Z INFO 10-10 01:47:35 [custom_all_reduce.py:35] Skipping P2P check and trusting the driver's P2P report. 2025-10-10T01:47:35.5529617Z INFO 10-10 01:47:35 [custom_all_reduce.py:35] Skipping P2P check and trusting the driver's P2P report. 2025-10-10T01:47:35.5530691Z WARNING 10-10 01:47:35 [custom_all_reduce.py:154] Custom allreduce is disabled because your platform lacks GPU P2P capability or P2P test failed. To silence this warning, specify disable_custom_all_reduce=True explicitly. 2025-10-10T01:47:35.5532114Z WARNING 10-10 01:47:35 [custom_all_reduce.py:154] Custom allreduce is disabled because your platform lacks GPU P2P capability or P2P test failed. To silence this warning, specify disable_custom_all_reduce=True explicitly. 2025-10-10T01:47:35.6141124Z INFO 10-10 01:47:35 [shm_broadcast.py:289] vLLM message queue communication handle: Handle(local_reader_ranks=[1], buffer_handle=(1, 4194304, 6, 'psm_3a0ea300'), local_subscribe_addr='ipc:///tmp/5ba69a4a-9d60-4e52-b058-be332c44258a', remote_subscribe_addr=None, remote_addr_ipv6=False) 2025-10-10T01:47:35.6166408Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-10-10T01:47:35.6167609Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-10-10T01:47:35.6172768Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-10-10T01:47:35.6180753Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-10-10T01:47:35.6734908Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-10-10T01:47:35.6743897Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-10-10T01:47:35.6755465Z [Gloo] Rank 1 is connected to 1 peer ranks. Expected number of connected peer ranks is : 1 2025-10-10T01:47:35.6756079Z [Gloo] Rank 0 is connected to 1 peer ranks. Expected number of connected peer ranks is : 1 2025-10-10T01:47:35.6758059Z INFO 10-10 01:47:35 [__init__.py:1384] Found nccl from library libnccl.so.2 2025-10-10T01:47:35.6758675Z INFO 10-10 01:47:35 [__init__.py:1384] Found nccl from library libnccl.so.2 2025-10-10T01:47:35.6759157Z INFO 10-10 01:47:35 [pynccl.py:104] vLLM is using nccl==2.27.5 2025-10-10T01:47:35.6759487Z INFO 10-10 01:47:35 [pynccl.py:104] vLLM is using nccl==2.27.5 2025-10-10T01:47:35.7091204Z INFO 10-10 01:47:35 [parallel_state.py:1208] rank 1 in world size 2 is assigned as DP rank 0, PP rank 0, TP rank 1, EP rank 1 2025-10-10T01:47:35.7091960Z INFO 10-10 01:47:35 [parallel_state.py:1208] rank 0 in world size 2 is assigned as DP rank 0, PP rank 0, TP rank 0, EP rank 0 2025-10-10T01:47:35.9511168Z INFO 10-10 01:47:35 [topk_topp_sampler.py:55] Using FlashInfer for top-p & top-k sampling. 2025-10-10T01:47:35.9529557Z INFO 10-10 01:47:35 [topk_topp_sampler.py:55] Using FlashInfer for top-p & top-k sampling. 2025-10-10T01:47:35.9828742Z (Worker_TP0 pid=48695) INFO 10-10 01:47:35 [gpu_model_runner.py:2707] Starting to load model meta-llama/Llama-3.2-1B-Instruct... 2025-10-10T01:47:35.9846347Z (Worker_TP1 pid=48696) INFO 10-10 01:47:35 [gpu_model_runner.py:2707] Starting to load model meta-llama/Llama-3.2-1B-Instruct... 2025-10-10T01:47:36.2066209Z (Worker_TP1 pid=48696) INFO 10-10 01:47:36 [gpu_model_runner.py:2739] Loading model from scratch... 2025-10-10T01:47:36.2155686Z (Worker_TP0 pid=48695) INFO 10-10 01:47:36 [gpu_model_runner.py:2739] Loading model from scratch... 2025-10-10T01:47:36.2963098Z (Worker_TP1 pid=48696) INFO 10-10 01:47:36 [cuda.py:361] Using Flash Attention backend on V1 engine. 2025-10-10T01:47:36.3053722Z (Worker_TP0 pid=48695) INFO 10-10 01:47:36 [cuda.py:361] Using Flash Attention backend on V1 engine. 2025-10-10T01:47:36.7779061Z (Worker_TP1 pid=48696) INFO 10-10 01:47:36 [gpu_model_runner.py:2758] Model loading took 1.1667 GiB and 0.119670 seconds 2025-10-10T01:47:36.8027671Z (Worker_TP0 pid=48695) INFO 10-10 01:47:36 [gpu_model_runner.py:2758] Model loading took 1.1667 GiB and 0.119350 seconds 2025-10-10T01:47:41.8157792Z (Worker_TP1 pid=48696) INFO 10-10 01:47:41 [gpu_worker.py:298] Available KV cache memory: 18.53 GiB 2025-10-10T01:47:41.8287407Z (Worker_TP0 pid=48695) INFO 10-10 01:47:41 [gpu_worker.py:298] Available KV cache memory: 18.53 GiB 2025-10-10T01:47:42.0656649Z (EngineCore_DP0 pid=48576) INFO 10-10 01:47:42 [kv_cache_utils.py:1087] GPU KV cache size: 1,214,464 tokens 2025-10-10T01:47:42.0657843Z (EngineCore_DP0 pid=48576) INFO 10-10 01:47:42 [kv_cache_utils.py:1091] Maximum concurrency for 2,048 tokens per request: 593.00x 2025-10-10T01:47:42.0658891Z (EngineCore_DP0 pid=48576) INFO 10-10 01:47:42 [kv_cache_utils.py:1087] GPU KV cache size: 1,214,464 tokens 2025-10-10T01:47:42.0660192Z (EngineCore_DP0 pid=48576) INFO 10-10 01:47:42 [kv_cache_utils.py:1091] Maximum concurrency for 2,048 tokens per request: 593.00x 2025-10-10T01:47:42.1890470Z (EngineCore_DP0 pid=48576) INFO 10-10 01:47:42 [core.py:211] init engine (profile, create kv cache, warmup model) took 5.32 seconds 2025-10-10T01:47:42.8546145Z (EngineCore_DP0 pid=48576) INFO 10-10 01:47:42 [vllm.py:345] Cudagraph is disabled under eager mode 2025-10-10T01:47:43.0396471Z (APIServer pid=48454) INFO 10-10 01:47:43 [loggers.py:147] Engine 000: vllm cache_config_info with initialization after num_gpu_blocks is: 75904 2025-10-10T01:47:43.0755574Z (EngineCore_DP0 pid=48576) INFO 10-10 01:47:43 [gc_utils.py:41] GC Debug Config. enabled:False,top_objects:-1 2025-10-10T01:47:43.0969780Z (APIServer pid=48454) INFO 10-10 01:47:43 [api_server.py:1634] Supported_tasks: ['generate'] 2025-10-10T01:47:43.1272183Z (APIServer pid=48454) WARNING 10-10 01:47:43 [model.py:1424] Default sampling parameters have been overridden by the model's Hugging Face generation config recommended from the model creator. If this is not intended, please relaunch vLLM instance with `--generation-config vllm`. 2025-10-10T01:47:43.1273781Z (APIServer pid=48454) INFO 10-10 01:47:43 [serving_responses.py:140] Using default chat sampling params from model: {'temperature': 0.6, 'top_p': 0.9} 2025-10-10T01:47:43.1514294Z (APIServer pid=48454) INFO 10-10 01:47:43 [serving_chat.py:139] Using default chat sampling params from model: {'temperature': 0.6, 'top_p': 0.9} 2025-10-10T01:47:43.1733225Z (APIServer pid=48454) INFO 10-10 01:47:43 [serving_completion.py:76] Using default completion sampling params from model: {'temperature': 0.6, 'top_p': 0.9} 2025-10-10T01:47:43.1734260Z (APIServer pid=48454) INFO 10-10 01:47:43 [api_server.py:1912] Starting vLLM API server 0 on http://0.0.0.0:37267 2025-10-10T01:47:43.1734992Z (APIServer pid=48454) INFO 10-10 01:47:43 [launcher.py:34] Available routes are: 2025-10-10T01:47:43.1735683Z (APIServer pid=48454) INFO 10-10 01:47:43 [launcher.py:42] Route: /openapi.json, Methods: GET, HEAD 2025-10-10T01:47:43.1736409Z (APIServer pid=48454) INFO 10-10 01:47:43 [launcher.py:42] Route: /docs, Methods: GET, HEAD 2025-10-10T01:47:43.1737122Z (APIServer pid=48454) INFO 10-10 01:47:43 [launcher.py:42] Route: /docs/oauth2-redirect, Methods: GET, HEAD 2025-10-10T01:47:43.1737831Z (APIServer pid=48454) INFO 10-10 01:47:43 [launcher.py:42] Route: /redoc, Methods: GET, HEAD 2025-10-10T01:47:43.1738475Z (APIServer pid=48454) INFO 10-10 01:47:43 [launcher.py:42] Route: /health, Methods: GET 2025-10-10T01:47:43.1739123Z (APIServer pid=48454) INFO 10-10 01:47:43 [launcher.py:42] Route: /load, Methods: GET 2025-10-10T01:47:43.1739636Z (APIServer pid=48454) INFO 10-10 01:47:43 [launcher.py:42] Route: /ping, Methods: POST 2025-10-10T01:47:43.1740141Z (APIServer pid=48454) INFO 10-10 01:47:43 [launcher.py:42] Route: /ping, Methods: GET 2025-10-10T01:47:43.1740842Z (APIServer pid=48454) INFO 10-10 01:47:43 [launcher.py:42] Route: /tokenize, Methods: POST 2025-10-10T01:47:43.1741384Z (APIServer pid=48454) INFO 10-10 01:47:43 [launcher.py:42] Route: /detokenize, Methods: POST 2025-10-10T01:47:43.1741920Z (APIServer pid=48454) INFO 10-10 01:47:43 [launcher.py:42] Route: /v1/models, Methods: GET 2025-10-10T01:47:43.1742459Z (APIServer pid=48454) INFO 10-10 01:47:43 [launcher.py:42] Route: /version, Methods: GET 2025-10-10T01:47:43.1743010Z (APIServer pid=48454) INFO 10-10 01:47:43 [launcher.py:42] Route: /v1/responses, Methods: POST 2025-10-10T01:47:43.1743604Z (APIServer pid=48454) INFO 10-10 01:47:43 [launcher.py:42] Route: /v1/responses/{response_id}, Methods: GET 2025-10-10T01:47:43.1744397Z (APIServer pid=48454) INFO 10-10 01:47:43 [launcher.py:42] Route: /v1/responses/{response_id}/cancel, Methods: POST 2025-10-10T01:47:43.1745049Z (APIServer pid=48454) INFO 10-10 01:47:43 [launcher.py:42] Route: /v1/chat/completions, Methods: POST 2025-10-10T01:47:43.1745629Z (APIServer pid=48454) INFO 10-10 01:47:43 [launcher.py:42] Route: /v1/completions, Methods: POST 2025-10-10T01:47:43.1746201Z (APIServer pid=48454) INFO 10-10 01:47:43 [launcher.py:42] Route: /v1/embeddings, Methods: POST 2025-10-10T01:47:43.1746809Z (APIServer pid=48454) INFO 10-10 01:47:43 [launcher.py:42] Route: /pooling, Methods: POST 2025-10-10T01:47:43.1747356Z (APIServer pid=48454) INFO 10-10 01:47:43 [launcher.py:42] Route: /classify, Methods: POST 2025-10-10T01:47:43.1747912Z (APIServer pid=48454) INFO 10-10 01:47:43 [launcher.py:42] Route: /score, Methods: POST 2025-10-10T01:47:43.1748443Z (APIServer pid=48454) INFO 10-10 01:47:43 [launcher.py:42] Route: /v1/score, Methods: POST 2025-10-10T01:47:43.1749048Z (APIServer pid=48454) INFO 10-10 01:47:43 [launcher.py:42] Route: /v1/audio/transcriptions, Methods: POST 2025-10-10T01:47:43.1749676Z (APIServer pid=48454) INFO 10-10 01:47:43 [launcher.py:42] Route: /v1/audio/translations, Methods: POST 2025-10-10T01:47:43.1750243Z (APIServer pid=48454) INFO 10-10 01:47:43 [launcher.py:42] Route: /rerank, Methods: POST 2025-10-10T01:47:43.1750785Z (APIServer pid=48454) INFO 10-10 01:47:43 [launcher.py:42] Route: /v1/rerank, Methods: POST 2025-10-10T01:47:43.1751329Z (APIServer pid=48454) INFO 10-10 01:47:43 [launcher.py:42] Route: /v2/rerank, Methods: POST 2025-10-10T01:47:43.1751887Z (APIServer pid=48454) INFO 10-10 01:47:43 [launcher.py:42] Route: /scale_elastic_ep, Methods: POST 2025-10-10T01:47:43.1752476Z (APIServer pid=48454) INFO 10-10 01:47:43 [launcher.py:42] Route: /is_scaling_elastic_ep, Methods: POST 2025-10-10T01:47:43.1753044Z (APIServer pid=48454) INFO 10-10 01:47:43 [launcher.py:42] Route: /invocations, Methods: POST 2025-10-10T01:47:43.1753593Z (APIServer pid=48454) INFO 10-10 01:47:43 [launcher.py:42] Route: /metrics, Methods: GET 2025-10-10T01:47:43.1895106Z (APIServer pid=48454) INFO: Started server process [48454] 2025-10-10T01:47:43.1895759Z (APIServer pid=48454) INFO: Waiting for application startup. 2025-10-10T01:47:43.4131633Z (APIServer pid=48454) INFO: Application startup complete. 2025-10-10T01:47:43.6105096Z (APIServer pid=48454) INFO: 127.0.0.1:57460 - "GET /health HTTP/1.1" 200 OK 2025-10-10T01:47:43.6194518Z (APIServer pid=48454) INFO: 127.0.0.1:57462 - "GET /v1/models HTTP/1.1" 200 OK 2025-10-10T01:47:43.7936870Z (APIServer pid=48454) INFO: 127.0.0.1:57462 - "POST /v1/completions HTTP/1.1" 200 OK 2025-10-10T01:47:43.8514047Z (APIServer pid=48454) INFO: 127.0.0.1:57462 - "POST /v1/completions HTTP/1.1" 200 OK 2025-10-10T01:47:43.9417821Z (APIServer pid=48454) INFO: 127.0.0.1:57462 - "POST /v1/completions HTTP/1.1" 200 OK 2025-10-10T01:47:44.0173204Z (APIServer pid=48454) INFO: 127.0.0.1:57462 - "POST /v1/completions HTTP/1.1" 200 OK 2025-10-10T01:47:44.0887886Z (APIServer pid=48454) INFO: 127.0.0.1:57462 - "POST /v1/completions HTTP/1.1" 200 OK 2025-10-10T01:47:44.0969565Z (APIServer pid=48454) INFO: 127.0.0.1:57462 - "POST /v1/completions HTTP/1.1" 200 OK 2025-10-10T01:47:44.1642010Z (APIServer pid=48454) INFO 10-10 01:47:44 [launcher.py:99] Shutting down FastAPI HTTP server. 2025-10-10T01:47:44.1645806Z (Worker_TP0 pid=48695) INFO 10-10 01:47:44 [multiproc_executor.py:558] Parent process exited, terminating worker 2025-10-10T01:47:44.1646911Z (Worker_TP0 pid=48695) INFO 10-10 01:47:44 [multiproc_executor.py:599] WorkerProc shutting down. 2025-10-10T01:47:44.1648920Z (Worker_TP1 pid=48696) INFO 10-10 01:47:44 [multiproc_executor.py:558] Parent process exited, terminating worker 2025-10-10T01:47:44.1650935Z (Worker_TP1 pid=48696) INFO 10-10 01:47:44 [multiproc_executor.py:599] WorkerProc shutting down. 2025-10-10T01:47:46.1847980Z (APIServer pid=48454) INFO: Shutting down 2025-10-10T01:47:46.2842907Z (APIServer pid=48454) INFO: Waiting for application shutdown. 2025-10-10T01:47:46.2845143Z (APIServer pid=48454) INFO: Application shutdown complete. 2025-10-10T01:47:46.9147543Z PASSED 2025-10-10T01:47:46.9290120Z distributed/test_sequence_parallel.py::test_tp_sp_generation[meta-llama/Llama-3.2-1B-Instruct-parallel_setup13-ray-1-auto-test_options13] Fork a new process to run a test 49213 2025-10-10T01:47:46.9304373Z Fork a new process to run a test 0 2025-10-10T01:47:47.5379988Z INFO 10-10 01:47:47 [model.py:551] Resolved architecture: LlamaForCausalLM 2025-10-10T01:47:47.5380921Z `torch_dtype` is deprecated! Use `dtype` instead! 2025-10-10T01:47:47.5381776Z WARNING 10-10 01:47:47 [model.py:1768] Casting torch.bfloat16 to torch.float16. 2025-10-10T01:47:47.5382176Z INFO 10-10 01:47:47 [model.py:1545] Using max model len 2048 2025-10-10T01:47:47.5384151Z Launching RemoteOpenAIServer with: vllm serve meta-llama/Llama-3.2-1B-Instruct --dtype float16 --max-model-len 2048 --max-num-seqs 8 --enforce-eager --tokenizer-mode auto --tensor-parallel-size 2 --pipeline-parallel-size 2 --distributed-executor-backend ray --compilation_config {"level": 3, "custom_ops": ["+rms_norm"], "compile_sizes": [4, 8], "pass_config": {"enable_sequence_parallelism": true, "enable_fusion": false, "enable_noop": true}} --load-format dummy --port 49799 --seed 0 2025-10-10T01:47:48.0476568Z /opt/conda/envs/py_3.12/lib/python3.12/site-packages/torch/cuda/__init__.py:64: FutureWarning: The pynvml package is deprecated. Please install nvidia-ml-py instead. If you did not install pynvml directly, please report this to the maintainers of the package that installed pynvml for you. 2025-10-10T01:47:48.0479029Z import pynvml # type: ignore[import] 2025-10-10T01:47:50.2557071Z INFO 10-10 01:47:50 [__init__.py:215] Automatically detected platform cuda. 2025-10-10T01:47:54.1992292Z (APIServer pid=49262) INFO 10-10 01:47:54 [api_server.py:1839] vLLM API server version 0.11.0rc2.dev157+g0ad9951c4.d20251010 2025-10-10T01:47:54.2020495Z (APIServer pid=49262) INFO 10-10 01:47:54 [utils.py:233] non-default args: {'model_tag': 'meta-llama/Llama-3.2-1B-Instruct', 'port': 49799, 'model': 'meta-llama/Llama-3.2-1B-Instruct', 'dtype': 'float16', 'seed': 0, 'max_model_len': 2048, 'enforce_eager': True, 'load_format': 'dummy', 'distributed_executor_backend': 'ray', 'pipeline_parallel_size': 2, 'tensor_parallel_size': 2, 'max_num_seqs': 8, 'compilation_config': {'level': 3, 'debug_dump_path': None, 'cache_dir': '', 'backend': '', 'custom_ops': ['+rms_norm'], 'splitting_ops': None, 'use_inductor': True, 'compile_sizes': [4, 8], 'inductor_compile_config': {'enable_auto_functionalized_v2': False}, 'inductor_passes': {}, 'cudagraph_mode': None, 'use_cudagraph': True, 'cudagraph_num_of_warmups': 0, 'cudagraph_capture_sizes': None, 'cudagraph_copy_inputs': False, 'full_cuda_graph': False, 'use_inductor_graph_partition': False, 'pass_config': {'enable_noop': True, 'enable_sequence_parallelism': True}, 'max_capture_size': None, 'local_cache_dir': None}} 2025-10-10T01:47:54.3448726Z (APIServer pid=49262) INFO 10-10 01:47:54 [model.py:551] Resolved architecture: LlamaForCausalLM 2025-10-10T01:47:54.3450768Z (APIServer pid=49262) `torch_dtype` is deprecated! Use `dtype` instead! 2025-10-10T01:47:54.3452629Z (APIServer pid=49262) WARNING 10-10 01:47:54 [model.py:1768] Casting torch.bfloat16 to torch.float16. 2025-10-10T01:47:54.3453847Z (APIServer pid=49262) INFO 10-10 01:47:54 [model.py:1545] Using max model len 2048 2025-10-10T01:47:54.5047734Z (APIServer pid=49262) INFO 10-10 01:47:54 [scheduler.py:222] Chunked prefill is enabled with max_num_batched_tokens=2048. 2025-10-10T01:47:54.5053653Z (APIServer pid=49262) INFO 10-10 01:47:54 [vllm.py:345] Cudagraph is disabled under eager mode 2025-10-10T01:47:55.4723107Z /opt/conda/envs/py_3.12/lib/python3.12/site-packages/torch/cuda/__init__.py:64: FutureWarning: The pynvml package is deprecated. Please install nvidia-ml-py instead. If you did not install pynvml directly, please report this to the maintainers of the package that installed pynvml for you. 2025-10-10T01:47:55.4725716Z import pynvml # type: ignore[import] 2025-10-10T01:47:57.7809846Z INFO 10-10 01:47:57 [__init__.py:215] Automatically detected platform cuda. 2025-10-10T01:48:01.8728084Z (EngineCore_DP0 pid=49384) INFO 10-10 01:48:01 [core.py:648] Waiting for init message from front-end. 2025-10-10T01:48:01.9002835Z (EngineCore_DP0 pid=49384) INFO 10-10 01:48:01 [core.py:78] Initializing a V1 LLM engine (v0.11.0rc2.dev157+g0ad9951c4.d20251010) with config: model='meta-llama/Llama-3.2-1B-Instruct', speculative_config=None, tokenizer='meta-llama/Llama-3.2-1B-Instruct', skip_tokenizer_init=False, tokenizer_mode=auto, revision=None, tokenizer_revision=None, trust_remote_code=False, dtype=torch.float16, max_seq_len=2048, download_dir=None, load_format=dummy, tensor_parallel_size=2, pipeline_parallel_size=2, data_parallel_size=1, disable_custom_all_reduce=False, quantization=None, enforce_eager=True, kv_cache_dtype=auto, device_config=cuda, structured_outputs_config=StructuredOutputsConfig(backend='auto', disable_fallback=False, disable_any_whitespace=False, disable_additional_properties=False, reasoning_parser=''), observability_config=ObservabilityConfig(show_hidden_metrics_for_version=None, otlp_traces_endpoint=None, collect_detailed_traces=None), seed=0, served_model_name=meta-llama/Llama-3.2-1B-Instruct, enable_prefix_caching=True, chunked_prefill_enabled=True, pooler_config=None, compilation_config={'level': 3, 'debug_dump_path': None, 'cache_dir': '', 'backend': '', 'custom_ops': ['+rms_norm', '+rms_norm'], 'splitting_ops': ['vllm.unified_attention', 'vllm.unified_attention_with_output', 'vllm.mamba_mixer2', 'vllm.mamba_mixer', 'vllm.short_conv', 'vllm.linear_attention', 'vllm.plamo2_mamba_mixer', 'vllm.gdn_attention', 'vllm.sparse_attn_indexer'], 'use_inductor': True, 'compile_sizes': [8, 4], 'inductor_compile_config': {'enable_auto_functionalized_v2': False}, 'inductor_passes': {}, 'cudagraph_mode': , 'use_cudagraph': False, 'cudagraph_num_of_warmups': 0, 'cudagraph_capture_sizes': [], 'cudagraph_copy_inputs': False, 'full_cuda_graph': False, 'use_inductor_graph_partition': False, 'pass_config': {'enable_noop': True, 'enable_sequence_parallelism': True}, 'max_capture_size': 0, 'local_cache_dir': None} 2025-10-10T01:48:03.8072761Z (EngineCore_DP0 pid=49384) 2025-10-10 01:48:03,806 INFO worker.py:1918 -- Started a local Ray instance. View the dashboard at http://127.0.0.1:8265  2025-10-10T01:48:04.6408955Z (EngineCore_DP0 pid=49384) INFO 10-10 01:48:04 [ray_utils.py:345] No current placement group found. Creating a new placement group. 2025-10-10T01:48:04.8140776Z (EngineCore_DP0 pid=49384) INFO 10-10 01:48:04 [ray_distributed_executor.py:171] use_ray_spmd_worker: True 2025-10-10T01:48:05.4315758Z (EngineCore_DP0 pid=49384) (pid=50270) /opt/conda/envs/py_3.12/lib/python3.12/site-packages/torch/cuda/__init__.py:64: FutureWarning: The pynvml package is deprecated. Please install nvidia-ml-py instead. If you did not install pynvml directly, please report this to the maintainers of the package that installed pynvml for you. 2025-10-10T01:48:05.4317967Z (EngineCore_DP0 pid=49384) (pid=50270) import pynvml # type: ignore[import] 2025-10-10T01:48:07.6765442Z (EngineCore_DP0 pid=49384) (pid=50270) INFO 10-10 01:48:07 [__init__.py:215] Automatically detected platform cuda. 2025-10-10T01:48:09.0648792Z (EngineCore_DP0 pid=49384) INFO 10-10 01:48:09 [ray_env.py:63] RAY_NON_CARRY_OVER_ENV_VARS from config: set() 2025-10-10T01:48:09.0650565Z (EngineCore_DP0 pid=49384) INFO 10-10 01:48:09 [ray_env.py:65] Copying the following environment variables to workers: ['VLLM_WORKER_MULTIPROC_METHOD', 'VLLM_USE_RAY_SPMD_WORKER', 'MAX_JOBS', 'VLLM_USE_RAY_COMPILED_DAG', 'LD_LIBRARY_PATH', 'HF_TOKEN', 'HUGGING_FACE_HUB_TOKEN', 'VLLM_USE_V1'] 2025-10-10T01:48:09.0652486Z (EngineCore_DP0 pid=49384) INFO 10-10 01:48:09 [ray_env.py:68] If certain env vars should NOT be copied, add them to /var/lib/jenkins/.config/vllm/ray_non_carry_over_env_vars.json file 2025-10-10T01:48:12.6763595Z (EngineCore_DP0 pid=49384) (RayWorkerWrapper pid=50270) [Gloo] Rank 0 is connected to 3 peer ranks. Expected number of connected peer ranks is : 3 2025-10-10T01:48:12.6764842Z (EngineCore_DP0 pid=49384) (RayWorkerWrapper pid=50270) [Gloo] Rank 0 is connected to 1 peer ranks. Expected number of connected peer ranks is : 1 2025-10-10T01:48:12.6765892Z (EngineCore_DP0 pid=49384) (RayWorkerWrapper pid=50270) INFO 10-10 01:48:12 [__init__.py:1384] Found nccl from library libnccl.so.2 2025-10-10T01:48:12.6766857Z (EngineCore_DP0 pid=49384) (RayWorkerWrapper pid=50270) INFO 10-10 01:48:12 [pynccl.py:104] vLLM is using nccl==2.27.5 2025-10-10T01:48:12.6768841Z (EngineCore_DP0 pid=49384) (pid=50269) INFO 10-10 01:48:07 [__init__.py:215] Automatically detected platform cuda. [repeated 3x across cluster] (Ray deduplicates logs by default. Set RAY_DEDUP_LOGS=0 to disable log deduplication, or see https://docs.ray.io/en/master/ray-observability/user-guides/configure-logging.html#log-deduplication for more options.) 2025-10-10T01:48:13.0873586Z (EngineCore_DP0 pid=49384) (RayWorkerWrapper pid=50270) WARNING 10-10 01:48:12 [symm_mem.py:58] SymmMemCommunicator: Device capability 8.9 not supported, communicator is not available. 2025-10-10T01:48:13.0875232Z (EngineCore_DP0 pid=49384) (RayWorkerWrapper pid=50270) INFO 10-10 01:48:12 [custom_all_reduce.py:35] Skipping P2P check and trusting the driver's P2P report. 2025-10-10T01:48:13.0876930Z (EngineCore_DP0 pid=49384) (RayWorkerWrapper pid=50270) WARNING 10-10 01:48:12 [custom_all_reduce.py:154] Custom allreduce is disabled because your platform lacks GPU P2P capability or P2P test failed. To silence this warning, specify disable_custom_all_reduce=True explicitly. 2025-10-10T01:48:13.0879748Z (EngineCore_DP0 pid=49384) (RayWorkerWrapper pid=50270) INFO 10-10 01:48:12 [shm_broadcast.py:289] vLLM message queue communication handle: Handle(local_reader_ranks=[1], buffer_handle=(1, 4194304, 6, 'psm_6a60e47d'), local_subscribe_addr='ipc:///tmp/4ba59284-1778-4828-b3eb-fbf403091e08', remote_subscribe_addr=None, remote_addr_ipv6=False) 2025-10-10T01:48:13.1929733Z (EngineCore_DP0 pid=49384) (RayWorkerWrapper pid=50270) INFO 10-10 01:48:13 [parallel_state.py:1208] rank 0 in world size 4 is assigned as DP rank 0, PP rank 0, TP rank 0, EP rank 0 2025-10-10T01:48:13.4001542Z (EngineCore_DP0 pid=49384) (RayWorkerWrapper pid=50270) INFO 10-10 01:48:13 [topk_topp_sampler.py:55] Using FlashInfer for top-p & top-k sampling. 2025-10-10T01:48:13.5056294Z (EngineCore_DP0 pid=49384) (RayWorkerWrapper pid=50270) INFO 10-10 01:48:13 [gpu_model_runner.py:2707] Starting to load model meta-llama/Llama-3.2-1B-Instruct... 2025-10-10T01:48:13.7133418Z (EngineCore_DP0 pid=49384) (RayWorkerWrapper pid=50270) INFO 10-10 01:48:13 [gpu_model_runner.py:2739] Loading model from scratch... 2025-10-10T01:48:13.8188626Z (EngineCore_DP0 pid=49384) (RayWorkerWrapper pid=50270) INFO 10-10 01:48:13 [cuda.py:361] Using Flash Attention backend on V1 engine. 2025-10-10T01:48:14.3321484Z (EngineCore_DP0 pid=49384) (RayWorkerWrapper pid=50270) INFO 10-10 01:48:14 [gpu_model_runner.py:2758] Model loading took 0.7135 GiB and 0.156223 seconds 2025-10-10T01:48:16.6801772Z (EngineCore_DP0 pid=49384) (RayWorkerWrapper pid=50270) INFO 10-10 01:48:16 [backends.py:548] Using cache directory: /var/lib/jenkins/.cache/vllm/torch_compile_cache/5280ab4e2b/rank_0_0/backbone for vLLM's torch.compile 2025-10-10T01:48:16.6803887Z (EngineCore_DP0 pid=49384) (RayWorkerWrapper pid=50270) INFO 10-10 01:48:16 [backends.py:559] Dynamo bytecode transform time: 2.02 s 2025-10-10T01:48:17.2999700Z (EngineCore_DP0 pid=49384) (RayWorkerWrapper pid=50269) INFO 10-10 01:48:17 [backends.py:164] Directly load the compiled graph(s) for dynamic shape from the cache, took 0.440 s 2025-10-10T01:48:18.3293748Z (EngineCore_DP0 pid=49384) (RayWorkerWrapper pid=50270) INFO 10-10 01:48:18 [gpu_worker.py:298] Available KV cache memory: 19.02 GiB 2025-10-10T01:48:18.3295097Z (EngineCore_DP0 pid=49384) (RayWorkerWrapper pid=50269) [Gloo] Rank 1 is connected to 1 peer ranks. Expected number of connected peer ranks is : 1 [repeated 22x across cluster] 2025-10-10T01:48:18.3296419Z (EngineCore_DP0 pid=49384) (RayWorkerWrapper pid=50269) INFO 10-10 01:48:13 [__init__.py:1384] Found nccl from library libnccl.so.2 [repeated 11x across cluster] 2025-10-10T01:48:18.3297625Z (EngineCore_DP0 pid=49384) (RayWorkerWrapper pid=50269) INFO 10-10 01:48:13 [pynccl.py:104] vLLM is using nccl==2.27.5 [repeated 11x across cluster] 2025-10-10T01:48:18.3299052Z (EngineCore_DP0 pid=49384) (RayWorkerWrapper pid=50269) WARNING 10-10 01:48:12 [symm_mem.py:58] SymmMemCommunicator: Device capability 8.9 not supported, communicator is not available. [repeated 3x across cluster] 2025-10-10T01:48:18.3300725Z (EngineCore_DP0 pid=49384) (RayWorkerWrapper pid=50269) INFO 10-10 01:48:12 [custom_all_reduce.py:35] Skipping P2P check and trusting the driver's P2P report. [repeated 3x across cluster] 2025-10-10T01:48:18.3302345Z (EngineCore_DP0 pid=49384) (RayWorkerWrapper pid=50269) WARNING 10-10 01:48:12 [custom_all_reduce.py:154] Custom allreduce is disabled because your platform lacks GPU P2P capability or P2P test failed. To silence this warning, specify disable_custom_all_reduce=True explicitly. [repeated 3x across cluster] 2025-10-10T01:48:18.3304288Z (EngineCore_DP0 pid=49384) (RayWorkerWrapper pid=50267) INFO 10-10 01:48:12 [shm_broadcast.py:289] vLLM message queue communication handle: Handle(local_reader_ranks=[1], buffer_handle=(1, 4194304, 6, 'psm_dde7f306'), local_subscribe_addr='ipc:///tmp/19167edc-ac44-4eee-8b70-f535fa903edf', remote_subscribe_addr=None, remote_addr_ipv6=False) 2025-10-10T01:48:18.3306106Z (EngineCore_DP0 pid=49384) (RayWorkerWrapper pid=50269) INFO 10-10 01:48:13 [parallel_state.py:1208] rank 3 in world size 4 is assigned as DP rank 0, PP rank 1, TP rank 1, EP rank 1 [repeated 3x across cluster] 2025-10-10T01:48:18.4343558Z (EngineCore_DP0 pid=49384) (RayWorkerWrapper pid=50269) INFO 10-10 01:48:13 [topk_topp_sampler.py:55] Using FlashInfer for top-p & top-k sampling. [repeated 3x across cluster] 2025-10-10T01:48:18.7262098Z (EngineCore_DP0 pid=49384) INFO 10-10 01:48:18 [kv_cache_utils.py:1087] GPU KV cache size: 2,492,416 tokens 2025-10-10T01:48:18.7263097Z (EngineCore_DP0 pid=49384) INFO 10-10 01:48:18 [kv_cache_utils.py:1091] Maximum concurrency for 2,048 tokens per request: 1217.00x 2025-10-10T01:48:18.7264411Z (EngineCore_DP0 pid=49384) INFO 10-10 01:48:18 [kv_cache_utils.py:1087] GPU KV cache size: 2,492,416 tokens 2025-10-10T01:48:18.7265342Z (EngineCore_DP0 pid=49384) INFO 10-10 01:48:18 [kv_cache_utils.py:1091] Maximum concurrency for 2,048 tokens per request: 1217.00x 2025-10-10T01:48:18.7266209Z (EngineCore_DP0 pid=49384) INFO 10-10 01:48:18 [kv_cache_utils.py:1087] GPU KV cache size: 2,489,088 tokens 2025-10-10T01:48:18.7267047Z (EngineCore_DP0 pid=49384) INFO 10-10 01:48:18 [kv_cache_utils.py:1091] Maximum concurrency for 2,048 tokens per request: 1215.38x 2025-10-10T01:48:18.7268030Z (EngineCore_DP0 pid=49384) INFO 10-10 01:48:18 [kv_cache_utils.py:1087] GPU KV cache size: 2,489,088 tokens 2025-10-10T01:48:18.7268876Z (EngineCore_DP0 pid=49384) INFO 10-10 01:48:18 [kv_cache_utils.py:1091] Maximum concurrency for 2,048 tokens per request: 1215.38x 2025-10-10T01:48:18.7443577Z (EngineCore_DP0 pid=49384) (RayWorkerWrapper pid=50270) INFO 10-10 01:48:18 [gpu_worker.py:332] Compile and warming up model for size 8 2025-10-10T01:48:18.7445558Z (EngineCore_DP0 pid=49384) (RayWorkerWrapper pid=50269) INFO 10-10 01:48:13 [gpu_model_runner.py:2707] Starting to load model meta-llama/Llama-3.2-1B-Instruct... [repeated 3x across cluster] 2025-10-10T01:48:18.7447621Z (EngineCore_DP0 pid=49384) (RayWorkerWrapper pid=50269) INFO 10-10 01:48:13 [gpu_model_runner.py:2739] Loading model from scratch... [repeated 3x across cluster] 2025-10-10T01:48:19.1552444Z (EngineCore_DP0 pid=49384) (RayWorkerWrapper pid=50267) INFO 10-10 01:48:19 [backends.py:168] Directly load the compiled graph(s) for shape 8 from the cache, took 0.407 s 2025-10-10T01:48:19.1553855Z (EngineCore_DP0 pid=49384) (RayWorkerWrapper pid=50269) INFO 10-10 01:48:13 [cuda.py:361] Using Flash Attention backend on V1 engine. [repeated 3x across cluster] 2025-10-10T01:48:19.9747373Z (EngineCore_DP0 pid=49384) (RayWorkerWrapper pid=50267) INFO 10-10 01:48:19 [monitor.py:32] torch.compile takes 2.03 s in total 2025-10-10T01:48:19.9748841Z (EngineCore_DP0 pid=49384) (RayWorkerWrapper pid=50267) INFO 10-10 01:48:14 [gpu_model_runner.py:2758] Model loading took 0.7135 GiB and 0.155221 seconds [repeated 3x across cluster] 2025-10-10T01:48:20.4622899Z (EngineCore_DP0 pid=49384) INFO 10-10 01:48:20 [core.py:211] init engine (profile, create kv cache, warmup model) took 6.07 seconds 2025-10-10T01:48:21.2874060Z (EngineCore_DP0 pid=49384) INFO 10-10 01:48:21 [core.py:150] Batch queue is enabled with size 2 2025-10-10T01:48:21.2914258Z (EngineCore_DP0 pid=49384) INFO 10-10 01:48:21 [vllm.py:345] Cudagraph is disabled under eager mode 2025-10-10T01:48:21.4669809Z (APIServer pid=49262) INFO 10-10 01:48:21 [loggers.py:147] Engine 000: vllm cache_config_info with initialization after num_gpu_blocks is: 155568 2025-10-10T01:48:21.5378557Z (EngineCore_DP0 pid=49384) INFO 10-10 01:48:21 [gc_utils.py:41] GC Debug Config. enabled:False,top_objects:-1 2025-10-10T01:48:21.5625053Z (APIServer pid=49262) INFO 10-10 01:48:21 [api_server.py:1634] Supported_tasks: ['generate'] 2025-10-10T01:48:21.5834225Z (APIServer pid=49262) WARNING 10-10 01:48:21 [model.py:1424] Default sampling parameters have been overridden by the model's Hugging Face generation config recommended from the model creator. If this is not intended, please relaunch vLLM instance with `--generation-config vllm`. 2025-10-10T01:48:21.5835877Z (APIServer pid=49262) INFO 10-10 01:48:21 [serving_responses.py:140] Using default chat sampling params from model: {'temperature': 0.6, 'top_p': 0.9} 2025-10-10T01:48:21.6111937Z (APIServer pid=49262) INFO 10-10 01:48:21 [serving_chat.py:139] Using default chat sampling params from model: {'temperature': 0.6, 'top_p': 0.9} 2025-10-10T01:48:21.6543260Z (APIServer pid=49262) INFO 10-10 01:48:21 [serving_completion.py:76] Using default completion sampling params from model: {'temperature': 0.6, 'top_p': 0.9} 2025-10-10T01:48:21.6544150Z (APIServer pid=49262) INFO 10-10 01:48:21 [api_server.py:1912] Starting vLLM API server 0 on http://0.0.0.0:49799 2025-10-10T01:48:21.6544772Z (APIServer pid=49262) INFO 10-10 01:48:21 [launcher.py:34] Available routes are: 2025-10-10T01:48:21.6545433Z (APIServer pid=49262) INFO 10-10 01:48:21 [launcher.py:42] Route: /openapi.json, Methods: GET, HEAD 2025-10-10T01:48:21.6545999Z (APIServer pid=49262) INFO 10-10 01:48:21 [launcher.py:42] Route: /docs, Methods: GET, HEAD 2025-10-10T01:48:21.6546572Z (APIServer pid=49262) INFO 10-10 01:48:21 [launcher.py:42] Route: /docs/oauth2-redirect, Methods: GET, HEAD 2025-10-10T01:48:21.6547135Z (APIServer pid=49262) INFO 10-10 01:48:21 [launcher.py:42] Route: /redoc, Methods: GET, HEAD 2025-10-10T01:48:21.6547663Z (APIServer pid=49262) INFO 10-10 01:48:21 [launcher.py:42] Route: /health, Methods: GET 2025-10-10T01:48:21.6548187Z (APIServer pid=49262) INFO 10-10 01:48:21 [launcher.py:42] Route: /load, Methods: GET 2025-10-10T01:48:21.6548701Z (APIServer pid=49262) INFO 10-10 01:48:21 [launcher.py:42] Route: /ping, Methods: POST 2025-10-10T01:48:21.6549203Z (APIServer pid=49262) INFO 10-10 01:48:21 [launcher.py:42] Route: /ping, Methods: GET 2025-10-10T01:48:21.6549719Z (APIServer pid=49262) INFO 10-10 01:48:21 [launcher.py:42] Route: /tokenize, Methods: POST 2025-10-10T01:48:21.6550265Z (APIServer pid=49262) INFO 10-10 01:48:21 [launcher.py:42] Route: /detokenize, Methods: POST 2025-10-10T01:48:21.6550935Z (APIServer pid=49262) INFO 10-10 01:48:21 [launcher.py:42] Route: /v1/models, Methods: GET 2025-10-10T01:48:21.6551462Z (APIServer pid=49262) INFO 10-10 01:48:21 [launcher.py:42] Route: /version, Methods: GET 2025-10-10T01:48:21.6551986Z (APIServer pid=49262) INFO 10-10 01:48:21 [launcher.py:42] Route: /v1/responses, Methods: POST 2025-10-10T01:48:21.6552583Z (APIServer pid=49262) INFO 10-10 01:48:21 [launcher.py:42] Route: /v1/responses/{response_id}, Methods: GET 2025-10-10T01:48:21.6553225Z (APIServer pid=49262) INFO 10-10 01:48:21 [launcher.py:42] Route: /v1/responses/{response_id}/cancel, Methods: POST 2025-10-10T01:48:21.6553857Z (APIServer pid=49262) INFO 10-10 01:48:21 [launcher.py:42] Route: /v1/chat/completions, Methods: POST 2025-10-10T01:48:21.6554441Z (APIServer pid=49262) INFO 10-10 01:48:21 [launcher.py:42] Route: /v1/completions, Methods: POST 2025-10-10T01:48:21.6555007Z (APIServer pid=49262) INFO 10-10 01:48:21 [launcher.py:42] Route: /v1/embeddings, Methods: POST 2025-10-10T01:48:21.6555544Z (APIServer pid=49262) INFO 10-10 01:48:21 [launcher.py:42] Route: /pooling, Methods: POST 2025-10-10T01:48:21.6556085Z (APIServer pid=49262) INFO 10-10 01:48:21 [launcher.py:42] Route: /classify, Methods: POST 2025-10-10T01:48:21.6556714Z (APIServer pid=49262) INFO 10-10 01:48:21 [launcher.py:42] Route: /score, Methods: POST 2025-10-10T01:48:21.6557227Z (APIServer pid=49262) INFO 10-10 01:48:21 [launcher.py:42] Route: /v1/score, Methods: POST 2025-10-10T01:48:21.6557797Z (APIServer pid=49262) INFO 10-10 01:48:21 [launcher.py:42] Route: /v1/audio/transcriptions, Methods: POST 2025-10-10T01:48:21.6558494Z (APIServer pid=49262) INFO 10-10 01:48:21 [launcher.py:42] Route: /v1/audio/translations, Methods: POST 2025-10-10T01:48:21.6559065Z (APIServer pid=49262) INFO 10-10 01:48:21 [launcher.py:42] Route: /rerank, Methods: POST 2025-10-10T01:48:21.6559584Z (APIServer pid=49262) INFO 10-10 01:48:21 [launcher.py:42] Route: /v1/rerank, Methods: POST 2025-10-10T01:48:21.6560208Z (APIServer pid=49262) INFO 10-10 01:48:21 [launcher.py:42] Route: /v2/rerank, Methods: POST 2025-10-10T01:48:21.6560777Z (APIServer pid=49262) INFO 10-10 01:48:21 [launcher.py:42] Route: /scale_elastic_ep, Methods: POST 2025-10-10T01:48:21.6561357Z (APIServer pid=49262) INFO 10-10 01:48:21 [launcher.py:42] Route: /is_scaling_elastic_ep, Methods: POST 2025-10-10T01:48:21.6561925Z (APIServer pid=49262) INFO 10-10 01:48:21 [launcher.py:42] Route: /invocations, Methods: POST 2025-10-10T01:48:21.6562506Z (APIServer pid=49262) INFO 10-10 01:48:21 [launcher.py:42] Route: /metrics, Methods: GET 2025-10-10T01:48:21.6714582Z (APIServer pid=49262) INFO: Started server process [49262] 2025-10-10T01:48:21.6715854Z (APIServer pid=49262) INFO: Waiting for application startup. 2025-10-10T01:48:21.9267773Z (APIServer pid=49262) INFO: Application startup complete. 2025-10-10T01:48:22.1299500Z (APIServer pid=49262) INFO: 127.0.0.1:49914 - "GET /health HTTP/1.1" 200 OK 2025-10-10T01:48:22.3539237Z (APIServer pid=49262) INFO: 127.0.0.1:49930 - "GET /v1/models HTTP/1.1" 200 OK 2025-10-10T01:48:22.3841823Z (EngineCore_DP0 pid=49384) INFO 10-10 01:48:22 [ray_distributed_executor.py:552] RAY_CGRAPH_get_timeout is set to 300 2025-10-10T01:48:22.3842808Z (EngineCore_DP0 pid=49384) INFO 10-10 01:48:22 [ray_distributed_executor.py:554] VLLM_USE_RAY_COMPILED_DAG_CHANNEL_TYPE = auto 2025-10-10T01:48:22.3843810Z (EngineCore_DP0 pid=49384) INFO 10-10 01:48:22 [ray_distributed_executor.py:556] VLLM_USE_RAY_COMPILED_DAG_OVERLAP_COMM = False 2025-10-10T01:48:22.3850251Z (EngineCore_DP0 pid=49384) INFO 10-10 01:48:22 [ray_distributed_executor.py:621] Using RayPPCommunicator (which wraps vLLM _PP GroupCoordinator) for Ray Compiled Graph communication. 2025-10-10T01:48:22.4167852Z (EngineCore_DP0 pid=49384) 2025-10-10 01:48:22,416 INFO torch_tensor_accelerator_channel.py:807 -- Creating communicator group 74ec3d36-5592-44e9-a2b6-f5afe49974b5 on actors: [Actor(RayWorkerWrapper, 69899a5eb9d2f9a8633aca7401000000), Actor(RayWorkerWrapper, 9da07693b0da3b36b5a3c80801000000), Actor(RayWorkerWrapper, 6d787a8d48143a32ac631fa001000000), Actor(RayWorkerWrapper, dac2e454b84cfd70f89bf85d01000000)] 2025-10-10T01:48:22.4274046Z (EngineCore_DP0 pid=49384) (RayWorkerWrapper pid=50270) /opt/conda/envs/py_3.12/lib/python3.12/site-packages/vllm/distributed/device_communicators/ray_communicator.py:107: UserWarning: The given buffer is not writable, and PyTorch does not support non-writable tensors. This means you can write to the underlying (supposedly non-writable) buffer using the tensor. You may want to copy the buffer to protect its data or make it writable before converting it to a tensor. This type of warning will be suppressed for the rest of this program. (Triggered internally at /var/lib/jenkins/workspace/torch/csrc/utils/tensor_new.cpp:1581.) 2025-10-10T01:48:22.4276805Z (EngineCore_DP0 pid=49384) (RayWorkerWrapper pid=50270) actor_id_tensor = torch.frombuffer( 2025-10-10T01:48:22.4279098Z (EngineCore_DP0 pid=49384) (pid=50269) /opt/conda/envs/py_3.12/lib/python3.12/site-packages/torch/cuda/__init__.py:64: FutureWarning: The pynvml package is deprecated. Please install nvidia-ml-py instead. If you did not install pynvml directly, please report this to the maintainers of the package that installed pynvml for you. [repeated 3x across cluster] 2025-10-10T01:48:22.4281055Z (EngineCore_DP0 pid=49384) (pid=50269) import pynvml # type: ignore[import] [repeated 3x across cluster] 2025-10-10T01:48:22.4721047Z (EngineCore_DP0 pid=49384) 2025-10-10 01:48:22,471 INFO torch_tensor_accelerator_channel.py:833 -- Communicator group initialized. 2025-10-10T01:48:22.6972503Z (APIServer pid=49262) INFO: 127.0.0.1:49930 - "POST /v1/completions HTTP/1.1" 200 OK 2025-10-10T01:48:22.7658794Z (APIServer pid=49262) INFO: 127.0.0.1:49930 - "POST /v1/completions HTTP/1.1" 200 OK 2025-10-10T01:48:22.9229113Z (APIServer pid=49262) INFO: 127.0.0.1:49930 - "POST /v1/completions HTTP/1.1" 200 OK 2025-10-10T01:48:23.0041559Z (APIServer pid=49262) INFO: 127.0.0.1:49930 - "POST /v1/completions HTTP/1.1" 200 OK 2025-10-10T01:48:23.0678334Z (APIServer pid=49262) INFO: 127.0.0.1:49930 - "POST /v1/completions HTTP/1.1" 200 OK 2025-10-10T01:48:23.0760704Z (APIServer pid=49262) INFO: 127.0.0.1:49930 - "POST /v1/completions HTTP/1.1" 200 OK 2025-10-10T01:48:23.1859105Z (APIServer pid=49262) INFO 10-10 01:48:23 [launcher.py:99] Shutting down FastAPI HTTP server. 2025-10-10T01:48:23.1859669Z *** SIGTERM received at time=1760060903 on cpu 1 *** 2025-10-10T01:48:23.1906220Z PC: @ 0x7f191303f117 (unknown) (unknown) 2025-10-10T01:48:23.1907669Z @ 0x7f1912ff0520 (unknown) (unknown) 2025-10-10T01:48:23.1908147Z [2025-10-10 01:48:23,190 E 49384 49384] logging.cc:501: *** SIGTERM received at time=1760060903 on cpu 1 *** 2025-10-10T01:48:23.1909909Z [2025-10-10 01:48:23,190 E 49384 49384] logging.cc:501: PC: @ 0x7f191303f117 (unknown) (unknown) 2025-10-10T01:48:23.1911325Z [2025-10-10 01:48:23,190 E 49384 49384] logging.cc:501: @ 0x7f1912ff0520 (unknown) (unknown) 2025-10-10T01:48:23.1913436Z (EngineCore_DP0 pid=49384) INFO 10-10 01:48:23 [ray_distributed_executor.py:122] Shutting down Ray distributed executor. If you see error log from logging.cc regarding SIGTERM received, please ignore because this is the expected termination process in Ray. 2025-10-10T01:48:23.1914520Z (EngineCore_DP0 pid=49384) 2025-10-10 01:48:23,191 INFO compiled_dag_node.py:2171 -- Tearing down compiled DAG 2025-10-10T01:48:23.1915718Z (EngineCore_DP0 pid=49384) 2025-10-10 01:48:23,191 INFO compiled_dag_node.py:2176 -- Cancelling compiled worker on actor: Actor(RayWorkerWrapper, dac2e454b84cfd70f89bf85d01000000) 2025-10-10T01:48:23.1916914Z (EngineCore_DP0 pid=49384) 2025-10-10 01:48:23,191 INFO compiled_dag_node.py:2176 -- Cancelling compiled worker on actor: Actor(RayWorkerWrapper, 69899a5eb9d2f9a8633aca7401000000) 2025-10-10T01:48:23.1917943Z (EngineCore_DP0 pid=49384) 2025-10-10 01:48:23,191 INFO compiled_dag_node.py:2176 -- Cancelling compiled worker on actor: Actor(RayWorkerWrapper, 6d787a8d48143a32ac631fa001000000) 2025-10-10T01:48:23.1919025Z (EngineCore_DP0 pid=49384) 2025-10-10 01:48:23,191 INFO compiled_dag_node.py:2176 -- Cancelling compiled worker on actor: Actor(RayWorkerWrapper, 9da07693b0da3b36b5a3c80801000000) 2025-10-10T01:48:23.1974472Z (EngineCore_DP0 pid=49384) 2025-10-10 01:48:23,197 INFO compiled_dag_node.py:2198 -- Waiting for worker tasks to exit 2025-10-10T01:48:23.1977425Z (EngineCore_DP0 pid=49384) 2025-10-10 01:48:23,197 INFO compiled_dag_node.py:2201 -- Teardown complete 2025-10-10T01:48:23.7007078Z (EngineCore_DP0 pid=49384) (RayWorkerWrapper pid=50267) INFO 10-10 01:48:16 [backends.py:548] Using cache directory: /var/lib/jenkins/.cache/vllm/torch_compile_cache/8cedb68d99/rank_2_0/backbone for vLLM's torch.compile [repeated 3x across cluster] 2025-10-10T01:48:23.7008916Z (EngineCore_DP0 pid=49384) (RayWorkerWrapper pid=50267) INFO 10-10 01:48:16 [backends.py:559] Dynamo bytecode transform time: 2.03 s [repeated 3x across cluster] 2025-10-10T01:48:23.7010085Z (EngineCore_DP0 pid=49384) (RayWorkerWrapper pid=50267) INFO 10-10 01:48:17 [backends.py:164] Directly load the compiled graph(s) for dynamic shape from the cache, took 0.449 s [repeated 3x across cluster] 2025-10-10T01:48:23.7012949Z (EngineCore_DP0 pid=49384) (RayWorkerWrapper pid=50267) INFO 10-10 01:48:18 [gpu_worker.py:298] Available KV cache memory: 18.99 GiB [repeated 3x across cluster] 2025-10-10T01:48:23.7014146Z (EngineCore_DP0 pid=49384) (RayWorkerWrapper pid=50274) INFO 10-10 01:48:19 [gpu_worker.py:332] Compile and warming up model for size 4 [repeated 7x across cluster] 2025-10-10T01:48:23.7015369Z (EngineCore_DP0 pid=49384) (RayWorkerWrapper pid=50274) INFO 10-10 01:48:20 [backends.py:168] Directly load the compiled graph(s) for shape 4 from the cache, took 1.265 s [repeated 7x across cluster] 2025-10-10T01:48:23.7016439Z (EngineCore_DP0 pid=49384) (RayWorkerWrapper pid=50274) INFO 10-10 01:48:20 [monitor.py:32] torch.compile takes 1.99 s in total [repeated 3x across cluster] 2025-10-10T01:48:23.7019469Z (EngineCore_DP0 pid=49384) (RayWorkerWrapper pid=50269) /opt/conda/envs/py_3.12/lib/python3.12/site-packages/vllm/distributed/device_communicators/ray_communicator.py:107: UserWarning: The given buffer is not writable, and PyTorch does not support non-writable tensors. This means you can write to the underlying (supposedly non-writable) buffer using the tensor. You may want to copy the buffer to protect its data or make it writable before converting it to a tensor. This type of warning will be suppressed for the rest of this program. (Triggered internally at /var/lib/jenkins/workspace/torch/csrc/utils/tensor_new.cpp:1581.) [repeated 3x across cluster] 2025-10-10T01:48:23.7022112Z (EngineCore_DP0 pid=49384) (RayWorkerWrapper pid=50269) actor_id_tensor = torch.frombuffer( [repeated 3x across cluster] 2025-10-10T01:48:25.5927387Z (APIServer pid=49262) INFO: Shutting down 2025-10-10T01:48:25.6924333Z (APIServer pid=49262) INFO: Waiting for application shutdown. 2025-10-10T01:48:25.6925141Z (APIServer pid=49262) INFO: Application shutdown complete. 2025-10-10T01:48:26.0548118Z INFO 10-10 01:48:26 [model.py:551] Resolved architecture: LlamaForCausalLM 2025-10-10T01:48:26.0548674Z WARNING 10-10 01:48:26 [model.py:1768] Casting torch.bfloat16 to torch.float16. 2025-10-10T01:48:26.0549171Z INFO 10-10 01:48:26 [model.py:1545] Using max model len 2048 2025-10-10T01:48:26.0550373Z Launching RemoteOpenAIServer with: vllm serve meta-llama/Llama-3.2-1B-Instruct --dtype float16 --max-model-len 2048 --max-num-seqs 8 --enforce-eager --tokenizer-mode auto --tensor-parallel-size 2 --distributed-executor-backend mp --load-format dummy --port 46985 --seed 0 2025-10-10T01:48:26.5665494Z /opt/conda/envs/py_3.12/lib/python3.12/site-packages/torch/cuda/__init__.py:64: FutureWarning: The pynvml package is deprecated. Please install nvidia-ml-py instead. If you did not install pynvml directly, please report this to the maintainers of the package that installed pynvml for you. 2025-10-10T01:48:26.5666727Z import pynvml # type: ignore[import] 2025-10-10T01:48:28.7899315Z INFO 10-10 01:48:28 [__init__.py:215] Automatically detected platform cuda. 2025-10-10T01:48:32.7370198Z (APIServer pid=54183) INFO 10-10 01:48:32 [api_server.py:1839] vLLM API server version 0.11.0rc2.dev157+g0ad9951c4.d20251010 2025-10-10T01:48:32.7392724Z (APIServer pid=54183) INFO 10-10 01:48:32 [utils.py:233] non-default args: {'model_tag': 'meta-llama/Llama-3.2-1B-Instruct', 'port': 46985, 'model': 'meta-llama/Llama-3.2-1B-Instruct', 'dtype': 'float16', 'seed': 0, 'max_model_len': 2048, 'enforce_eager': True, 'load_format': 'dummy', 'distributed_executor_backend': 'mp', 'tensor_parallel_size': 2, 'max_num_seqs': 8} 2025-10-10T01:48:32.8892925Z (APIServer pid=54183) INFO 10-10 01:48:32 [model.py:551] Resolved architecture: LlamaForCausalLM 2025-10-10T01:48:32.8893638Z (APIServer pid=54183) `torch_dtype` is deprecated! Use `dtype` instead! 2025-10-10T01:48:32.8894626Z (APIServer pid=54183) WARNING 10-10 01:48:32 [model.py:1768] Casting torch.bfloat16 to torch.float16. 2025-10-10T01:48:32.8895383Z (APIServer pid=54183) INFO 10-10 01:48:32 [model.py:1545] Using max model len 2048 2025-10-10T01:48:33.0526725Z (APIServer pid=54183) INFO 10-10 01:48:33 [scheduler.py:222] Chunked prefill is enabled with max_num_batched_tokens=2048. 2025-10-10T01:48:33.0532325Z (APIServer pid=54183) INFO 10-10 01:48:33 [vllm.py:345] Cudagraph is disabled under eager mode 2025-10-10T01:48:34.0166632Z /opt/conda/envs/py_3.12/lib/python3.12/site-packages/torch/cuda/__init__.py:64: FutureWarning: The pynvml package is deprecated. Please install nvidia-ml-py instead. If you did not install pynvml directly, please report this to the maintainers of the package that installed pynvml for you. 2025-10-10T01:48:34.0168636Z import pynvml # type: ignore[import] 2025-10-10T01:48:36.3051377Z INFO 10-10 01:48:36 [__init__.py:215] Automatically detected platform cuda. 2025-10-10T01:48:40.2666486Z (EngineCore_DP0 pid=54305) INFO 10-10 01:48:40 [core.py:648] Waiting for init message from front-end. 2025-10-10T01:48:40.2936838Z (EngineCore_DP0 pid=54305) INFO 10-10 01:48:40 [core.py:78] Initializing a V1 LLM engine (v0.11.0rc2.dev157+g0ad9951c4.d20251010) with config: model='meta-llama/Llama-3.2-1B-Instruct', speculative_config=None, tokenizer='meta-llama/Llama-3.2-1B-Instruct', skip_tokenizer_init=False, tokenizer_mode=auto, revision=None, tokenizer_revision=None, trust_remote_code=False, dtype=torch.float16, max_seq_len=2048, download_dir=None, load_format=dummy, tensor_parallel_size=2, pipeline_parallel_size=1, data_parallel_size=1, disable_custom_all_reduce=False, quantization=None, enforce_eager=True, kv_cache_dtype=auto, device_config=cuda, structured_outputs_config=StructuredOutputsConfig(backend='auto', disable_fallback=False, disable_any_whitespace=False, disable_additional_properties=False, reasoning_parser=''), observability_config=ObservabilityConfig(show_hidden_metrics_for_version=None, otlp_traces_endpoint=None, collect_detailed_traces=None), seed=0, served_model_name=meta-llama/Llama-3.2-1B-Instruct, enable_prefix_caching=True, chunked_prefill_enabled=True, pooler_config=None, compilation_config={'level': 0, 'debug_dump_path': None, 'cache_dir': '', 'backend': '', 'custom_ops': [], 'splitting_ops': None, 'use_inductor': True, 'compile_sizes': [], 'inductor_compile_config': {'enable_auto_functionalized_v2': False}, 'inductor_passes': {}, 'cudagraph_mode': , 'use_cudagraph': False, 'cudagraph_num_of_warmups': 0, 'cudagraph_capture_sizes': [], 'cudagraph_copy_inputs': False, 'full_cuda_graph': False, 'use_inductor_graph_partition': False, 'pass_config': {}, 'max_capture_size': 0, 'local_cache_dir': None} 2025-10-10T01:48:40.2942333Z (EngineCore_DP0 pid=54305) WARNING 10-10 01:48:40 [multiproc_executor.py:720] Reducing Torch parallelism from 24 threads to 1 to avoid unnecessary CPU contention. Set OMP_NUM_THREADS in the external environment to tune this value as needed. 2025-10-10T01:48:40.2948525Z (EngineCore_DP0 pid=54305) INFO 10-10 01:48:40 [shm_broadcast.py:289] vLLM message queue communication handle: Handle(local_reader_ranks=[0, 1], buffer_handle=(2, 16777216, 10, 'psm_957bd46d'), local_subscribe_addr='ipc:///tmp/6c727705-6a2f-4256-9adf-d965439a6b5d', remote_subscribe_addr=None, remote_addr_ipv6=False) 2025-10-10T01:48:40.7708037Z /opt/conda/envs/py_3.12/lib/python3.12/site-packages/torch/cuda/__init__.py:64: FutureWarning: The pynvml package is deprecated. Please install nvidia-ml-py instead. If you did not install pynvml directly, please report this to the maintainers of the package that installed pynvml for you. 2025-10-10T01:48:40.7710183Z import pynvml # type: ignore[import] 2025-10-10T01:48:40.7879829Z /opt/conda/envs/py_3.12/lib/python3.12/site-packages/torch/cuda/__init__.py:64: FutureWarning: The pynvml package is deprecated. Please install nvidia-ml-py instead. If you did not install pynvml directly, please report this to the maintainers of the package that installed pynvml for you. 2025-10-10T01:48:40.7881754Z import pynvml # type: ignore[import] 2025-10-10T01:48:43.0942642Z INFO 10-10 01:48:43 [__init__.py:215] Automatically detected platform cuda. 2025-10-10T01:48:43.1075024Z INFO 10-10 01:48:43 [__init__.py:215] Automatically detected platform cuda. 2025-10-10T01:48:47.6343910Z INFO 10-10 01:48:47 [shm_broadcast.py:289] vLLM message queue communication handle: Handle(local_reader_ranks=[0], buffer_handle=(1, 10485760, 10, 'psm_a23f94ce'), local_subscribe_addr='ipc:///tmp/afab1bb9-2163-4f9a-9026-2ffb6631748c', remote_subscribe_addr=None, remote_addr_ipv6=False) 2025-10-10T01:48:47.6844393Z INFO 10-10 01:48:47 [shm_broadcast.py:289] vLLM message queue communication handle: Handle(local_reader_ranks=[0], buffer_handle=(1, 10485760, 10, 'psm_2478b4e4'), local_subscribe_addr='ipc:///tmp/fe09de23-c274-46b2-95e6-fa02fa42ff18', remote_subscribe_addr=None, remote_addr_ipv6=False) 2025-10-10T01:48:47.8594312Z [Gloo] Rank 0 is connected to 1 peer ranks. Expected number of connected peer ranks is : 1 2025-10-10T01:48:47.8594975Z [Gloo] Rank 1 is connected to 1 peer ranks. Expected number of connected peer ranks is : 1 2025-10-10T01:48:47.8640687Z [Gloo] Rank 0 is connected to 1 peer ranks. Expected number of connected peer ranks is : 1 2025-10-10T01:48:47.8641348Z [Gloo] Rank 1 is connected to 1 peer ranks. Expected number of connected peer ranks is : 1 2025-10-10T01:48:47.8681627Z INFO 10-10 01:48:47 [__init__.py:1384] Found nccl from library libnccl.so.2 2025-10-10T01:48:47.8685824Z INFO 10-10 01:48:47 [__init__.py:1384] Found nccl from library libnccl.so.2 2025-10-10T01:48:47.8686799Z INFO 10-10 01:48:47 [pynccl.py:104] vLLM is using nccl==2.27.5 2025-10-10T01:48:47.8687575Z INFO 10-10 01:48:47 [pynccl.py:104] vLLM is using nccl==2.27.5 2025-10-10T01:48:48.0425746Z WARNING 10-10 01:48:48 [symm_mem.py:58] SymmMemCommunicator: Device capability 8.9 not supported, communicator is not available. 2025-10-10T01:48:48.0427416Z WARNING 10-10 01:48:48 [symm_mem.py:58] SymmMemCommunicator: Device capability 8.9 not supported, communicator is not available. 2025-10-10T01:48:48.0448985Z INFO 10-10 01:48:48 [custom_all_reduce.py:35] Skipping P2P check and trusting the driver's P2P report. 2025-10-10T01:48:48.0450267Z INFO 10-10 01:48:48 [custom_all_reduce.py:35] Skipping P2P check and trusting the driver's P2P report. 2025-10-10T01:48:48.0452512Z WARNING 10-10 01:48:48 [custom_all_reduce.py:154] Custom allreduce is disabled because your platform lacks GPU P2P capability or P2P test failed. To silence this warning, specify disable_custom_all_reduce=True explicitly. 2025-10-10T01:48:48.0454299Z WARNING 10-10 01:48:48 [custom_all_reduce.py:154] Custom allreduce is disabled because your platform lacks GPU P2P capability or P2P test failed. To silence this warning, specify disable_custom_all_reduce=True explicitly. 2025-10-10T01:48:48.1141372Z INFO 10-10 01:48:48 [shm_broadcast.py:289] vLLM message queue communication handle: Handle(local_reader_ranks=[1], buffer_handle=(1, 4194304, 6, 'psm_517c2f22'), local_subscribe_addr='ipc:///tmp/da7faab9-2c93-4614-8c2d-272e8df2509d', remote_subscribe_addr=None, remote_addr_ipv6=False) 2025-10-10T01:48:48.1170824Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-10-10T01:48:48.1172025Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-10-10T01:48:48.1177371Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-10-10T01:48:48.1186735Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-10-10T01:48:48.1837133Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-10-10T01:48:48.1846000Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-10-10T01:48:48.1860225Z [Gloo] Rank 0 is connected to 1 peer ranks. Expected number of connected peer ranks is : 1 2025-10-10T01:48:48.1862443Z [Gloo] Rank 1 is connected to 1 peer ranks. Expected number of connected peer ranks is : 1 2025-10-10T01:48:48.1863649Z INFO 10-10 01:48:48 [__init__.py:1384] Found nccl from library libnccl.so.2 2025-10-10T01:48:48.1864075Z INFO 10-10 01:48:48 [__init__.py:1384] Found nccl from library libnccl.so.2 2025-10-10T01:48:48.1864655Z INFO 10-10 01:48:48 [pynccl.py:104] vLLM is using nccl==2.27.5 2025-10-10T01:48:48.1865136Z INFO 10-10 01:48:48 [pynccl.py:104] vLLM is using nccl==2.27.5 2025-10-10T01:48:48.2202183Z INFO 10-10 01:48:48 [parallel_state.py:1208] rank 0 in world size 2 is assigned as DP rank 0, PP rank 0, TP rank 0, EP rank 0 2025-10-10T01:48:48.2203693Z INFO 10-10 01:48:48 [parallel_state.py:1208] rank 1 in world size 2 is assigned as DP rank 0, PP rank 0, TP rank 1, EP rank 1 2025-10-10T01:48:48.4835442Z INFO 10-10 01:48:48 [topk_topp_sampler.py:55] Using FlashInfer for top-p & top-k sampling. 2025-10-10T01:48:48.4842366Z INFO 10-10 01:48:48 [topk_topp_sampler.py:55] Using FlashInfer for top-p & top-k sampling. 2025-10-10T01:48:48.5155912Z (Worker_TP0 pid=54424) INFO 10-10 01:48:48 [gpu_model_runner.py:2707] Starting to load model meta-llama/Llama-3.2-1B-Instruct... 2025-10-10T01:48:48.5167228Z (Worker_TP1 pid=54425) INFO 10-10 01:48:48 [gpu_model_runner.py:2707] Starting to load model meta-llama/Llama-3.2-1B-Instruct... 2025-10-10T01:48:48.7429712Z (Worker_TP1 pid=54425) INFO 10-10 01:48:48 [gpu_model_runner.py:2739] Loading model from scratch... 2025-10-10T01:48:48.7431398Z (Worker_TP0 pid=54424) INFO 10-10 01:48:48 [gpu_model_runner.py:2739] Loading model from scratch... 2025-10-10T01:48:48.8329469Z (Worker_TP1 pid=54425) INFO 10-10 01:48:48 [cuda.py:361] Using Flash Attention backend on V1 engine. 2025-10-10T01:48:48.8333719Z (Worker_TP0 pid=54424) INFO 10-10 01:48:48 [cuda.py:361] Using Flash Attention backend on V1 engine. 2025-10-10T01:48:49.3162609Z (Worker_TP1 pid=54425) INFO 10-10 01:48:49 [gpu_model_runner.py:2758] Model loading took 1.1667 GiB and 0.119468 seconds 2025-10-10T01:48:49.3177057Z (Worker_TP0 pid=54424) INFO 10-10 01:48:49 [gpu_model_runner.py:2758] Model loading took 1.1667 GiB and 0.120012 seconds 2025-10-10T01:48:54.3379830Z (Worker_TP0 pid=54424) INFO 10-10 01:48:54 [gpu_worker.py:298] Available KV cache memory: 18.53 GiB 2025-10-10T01:48:54.3613899Z (Worker_TP1 pid=54425) INFO 10-10 01:48:54 [gpu_worker.py:298] Available KV cache memory: 18.53 GiB 2025-10-10T01:48:54.6019059Z (EngineCore_DP0 pid=54305) INFO 10-10 01:48:54 [kv_cache_utils.py:1087] GPU KV cache size: 1,214,464 tokens 2025-10-10T01:48:54.6019997Z (EngineCore_DP0 pid=54305) INFO 10-10 01:48:54 [kv_cache_utils.py:1091] Maximum concurrency for 2,048 tokens per request: 593.00x 2025-10-10T01:48:54.6020845Z (EngineCore_DP0 pid=54305) INFO 10-10 01:48:54 [kv_cache_utils.py:1087] GPU KV cache size: 1,214,464 tokens 2025-10-10T01:48:54.6021745Z (EngineCore_DP0 pid=54305) INFO 10-10 01:48:54 [kv_cache_utils.py:1091] Maximum concurrency for 2,048 tokens per request: 593.00x 2025-10-10T01:48:54.7251868Z (EngineCore_DP0 pid=54305) INFO 10-10 01:48:54 [core.py:211] init engine (profile, create kv cache, warmup model) took 5.31 seconds 2025-10-10T01:48:55.2244593Z (EngineCore_DP0 pid=54305) INFO 10-10 01:48:55 [vllm.py:345] Cudagraph is disabled under eager mode 2025-10-10T01:48:55.4091311Z (APIServer pid=54183) INFO 10-10 01:48:55 [loggers.py:147] Engine 000: vllm cache_config_info with initialization after num_gpu_blocks is: 75904 2025-10-10T01:48:55.4430951Z (EngineCore_DP0 pid=54305) INFO 10-10 01:48:55 [gc_utils.py:41] GC Debug Config. enabled:False,top_objects:-1 2025-10-10T01:48:55.4644151Z (APIServer pid=54183) INFO 10-10 01:48:55 [api_server.py:1634] Supported_tasks: ['generate'] 2025-10-10T01:48:55.4944217Z (APIServer pid=54183) WARNING 10-10 01:48:55 [model.py:1424] Default sampling parameters have been overridden by the model's Hugging Face generation config recommended from the model creator. If this is not intended, please relaunch vLLM instance with `--generation-config vllm`. 2025-10-10T01:48:55.4946080Z (APIServer pid=54183) INFO 10-10 01:48:55 [serving_responses.py:140] Using default chat sampling params from model: {'temperature': 0.6, 'top_p': 0.9} 2025-10-10T01:48:55.5165226Z (APIServer pid=54183) INFO 10-10 01:48:55 [serving_chat.py:139] Using default chat sampling params from model: {'temperature': 0.6, 'top_p': 0.9} 2025-10-10T01:48:55.5481467Z (APIServer pid=54183) INFO 10-10 01:48:55 [serving_completion.py:76] Using default completion sampling params from model: {'temperature': 0.6, 'top_p': 0.9} 2025-10-10T01:48:55.5483557Z (APIServer pid=54183) INFO 10-10 01:48:55 [api_server.py:1912] Starting vLLM API server 0 on http://0.0.0.0:46985 2025-10-10T01:48:55.5484942Z (APIServer pid=54183) INFO 10-10 01:48:55 [launcher.py:34] Available routes are: 2025-10-10T01:48:55.5485545Z (APIServer pid=54183) INFO 10-10 01:48:55 [launcher.py:42] Route: /openapi.json, Methods: GET, HEAD 2025-10-10T01:48:55.5486106Z (APIServer pid=54183) INFO 10-10 01:48:55 [launcher.py:42] Route: /docs, Methods: GET, HEAD 2025-10-10T01:48:55.5486682Z (APIServer pid=54183) INFO 10-10 01:48:55 [launcher.py:42] Route: /docs/oauth2-redirect, Methods: GET, HEAD 2025-10-10T01:48:55.5487257Z (APIServer pid=54183) INFO 10-10 01:48:55 [launcher.py:42] Route: /redoc, Methods: GET, HEAD 2025-10-10T01:48:55.5487797Z (APIServer pid=54183) INFO 10-10 01:48:55 [launcher.py:42] Route: /health, Methods: GET 2025-10-10T01:48:55.5488300Z (APIServer pid=54183) INFO 10-10 01:48:55 [launcher.py:42] Route: /load, Methods: GET 2025-10-10T01:48:55.5488833Z (APIServer pid=54183) INFO 10-10 01:48:55 [launcher.py:42] Route: /ping, Methods: POST 2025-10-10T01:48:55.5489342Z (APIServer pid=54183) INFO 10-10 01:48:55 [launcher.py:42] Route: /ping, Methods: GET 2025-10-10T01:48:55.5489861Z (APIServer pid=54183) INFO 10-10 01:48:55 [launcher.py:42] Route: /tokenize, Methods: POST 2025-10-10T01:48:55.5490406Z (APIServer pid=54183) INFO 10-10 01:48:55 [launcher.py:42] Route: /detokenize, Methods: POST 2025-10-10T01:48:55.5490938Z (APIServer pid=54183) INFO 10-10 01:48:55 [launcher.py:42] Route: /v1/models, Methods: GET 2025-10-10T01:48:55.5491455Z (APIServer pid=54183) INFO 10-10 01:48:55 [launcher.py:42] Route: /version, Methods: GET 2025-10-10T01:48:55.5491987Z (APIServer pid=54183) INFO 10-10 01:48:55 [launcher.py:42] Route: /v1/responses, Methods: POST 2025-10-10T01:48:55.5492575Z (APIServer pid=54183) INFO 10-10 01:48:55 [launcher.py:42] Route: /v1/responses/{response_id}, Methods: GET 2025-10-10T01:48:55.5493215Z (APIServer pid=54183) INFO 10-10 01:48:55 [launcher.py:42] Route: /v1/responses/{response_id}/cancel, Methods: POST 2025-10-10T01:48:55.5493978Z (APIServer pid=54183) INFO 10-10 01:48:55 [launcher.py:42] Route: /v1/chat/completions, Methods: POST 2025-10-10T01:48:55.5494561Z (APIServer pid=54183) INFO 10-10 01:48:55 [launcher.py:42] Route: /v1/completions, Methods: POST 2025-10-10T01:48:55.5495127Z (APIServer pid=54183) INFO 10-10 01:48:55 [launcher.py:42] Route: /v1/embeddings, Methods: POST 2025-10-10T01:48:55.5495669Z (APIServer pid=54183) INFO 10-10 01:48:55 [launcher.py:42] Route: /pooling, Methods: POST 2025-10-10T01:48:55.5496202Z (APIServer pid=54183) INFO 10-10 01:48:55 [launcher.py:42] Route: /classify, Methods: POST 2025-10-10T01:48:55.5496724Z (APIServer pid=54183) INFO 10-10 01:48:55 [launcher.py:42] Route: /score, Methods: POST 2025-10-10T01:48:55.5497361Z (APIServer pid=54183) INFO 10-10 01:48:55 [launcher.py:42] Route: /v1/score, Methods: POST 2025-10-10T01:48:55.5497945Z (APIServer pid=54183) INFO 10-10 01:48:55 [launcher.py:42] Route: /v1/audio/transcriptions, Methods: POST 2025-10-10T01:48:55.5498546Z (APIServer pid=54183) INFO 10-10 01:48:55 [launcher.py:42] Route: /v1/audio/translations, Methods: POST 2025-10-10T01:48:55.5499093Z (APIServer pid=54183) INFO 10-10 01:48:55 [launcher.py:42] Route: /rerank, Methods: POST 2025-10-10T01:48:55.5499682Z (APIServer pid=54183) INFO 10-10 01:48:55 [launcher.py:42] Route: /v1/rerank, Methods: POST 2025-10-10T01:48:55.5500208Z (APIServer pid=54183) INFO 10-10 01:48:55 [launcher.py:42] Route: /v2/rerank, Methods: POST 2025-10-10T01:48:55.5500762Z (APIServer pid=54183) INFO 10-10 01:48:55 [launcher.py:42] Route: /scale_elastic_ep, Methods: POST 2025-10-10T01:48:55.5501354Z (APIServer pid=54183) INFO 10-10 01:48:55 [launcher.py:42] Route: /is_scaling_elastic_ep, Methods: POST 2025-10-10T01:48:55.5501925Z (APIServer pid=54183) INFO 10-10 01:48:55 [launcher.py:42] Route: /invocations, Methods: POST 2025-10-10T01:48:55.5502459Z (APIServer pid=54183) INFO 10-10 01:48:55 [launcher.py:42] Route: /metrics, Methods: GET 2025-10-10T01:48:55.5641598Z (APIServer pid=54183) INFO: Started server process [54183] 2025-10-10T01:48:55.5642877Z (APIServer pid=54183) INFO: Waiting for application startup. 2025-10-10T01:48:55.8085374Z (APIServer pid=54183) INFO: Application startup complete. 2025-10-10T01:48:56.1306644Z (APIServer pid=54183) INFO: 127.0.0.1:55872 - "GET /health HTTP/1.1" 200 OK 2025-10-10T01:48:56.1393652Z (APIServer pid=54183) INFO: 127.0.0.1:55878 - "GET /v1/models HTTP/1.1" 200 OK 2025-10-10T01:48:56.3120157Z (APIServer pid=54183) INFO: 127.0.0.1:55878 - "POST /v1/completions HTTP/1.1" 200 OK 2025-10-10T01:48:56.3690483Z (APIServer pid=54183) INFO: 127.0.0.1:55878 - "POST /v1/completions HTTP/1.1" 200 OK 2025-10-10T01:48:56.4655693Z (APIServer pid=54183) INFO: 127.0.0.1:55878 - "POST /v1/completions HTTP/1.1" 200 OK 2025-10-10T01:48:56.5499267Z (APIServer pid=54183) INFO: 127.0.0.1:55878 - "POST /v1/completions HTTP/1.1" 200 OK 2025-10-10T01:48:56.6207523Z (APIServer pid=54183) INFO: 127.0.0.1:55878 - "POST /v1/completions HTTP/1.1" 200 OK 2025-10-10T01:48:56.6283553Z (APIServer pid=54183) INFO: 127.0.0.1:55878 - "POST /v1/completions HTTP/1.1" 200 OK 2025-10-10T01:48:56.6964178Z (APIServer pid=54183) INFO 10-10 01:48:56 [launcher.py:99] Shutting down FastAPI HTTP server. 2025-10-10T01:48:56.6967792Z (Worker_TP0 pid=54424) INFO 10-10 01:48:56 [multiproc_executor.py:558] Parent process exited, terminating worker 2025-10-10T01:48:56.6969264Z (Worker_TP0 pid=54424) INFO 10-10 01:48:56 [multiproc_executor.py:599] WorkerProc shutting down. 2025-10-10T01:48:56.6972843Z (Worker_TP1 pid=54425) INFO 10-10 01:48:56 [multiproc_executor.py:558] Parent process exited, terminating worker 2025-10-10T01:48:56.6974634Z (Worker_TP1 pid=54425) INFO 10-10 01:48:56 [multiproc_executor.py:599] WorkerProc shutting down. 2025-10-10T01:48:58.7149064Z (APIServer pid=54183) INFO: Shutting down 2025-10-10T01:48:58.8143905Z (APIServer pid=54183) INFO: Waiting for application shutdown. 2025-10-10T01:48:58.8145296Z (APIServer pid=54183) INFO: Application shutdown complete. 2025-10-10T01:48:59.4439240Z PASSED 2025-10-10T01:48:59.4582843Z distributed/test_sequence_parallel.py::test_tp_sp_generation[meta-llama/Llama-3.2-1B-Instruct-parallel_setup14-mp-1-auto-test_options14] Fork a new process to run a test 54942 2025-10-10T01:48:59.4598732Z Fork a new process to run a test 0 2025-10-10T01:49:00.1352558Z INFO 10-10 01:49:00 [model.py:551] Resolved architecture: LlamaForCausalLM 2025-10-10T01:49:00.1353470Z `torch_dtype` is deprecated! Use `dtype` instead! 2025-10-10T01:49:00.1354301Z WARNING 10-10 01:49:00 [model.py:1768] Casting torch.bfloat16 to torch.float16. 2025-10-10T01:49:00.1355594Z INFO 10-10 01:49:00 [model.py:1545] Using max model len 2048 2025-10-10T01:49:00.1358884Z Launching RemoteOpenAIServer with: vllm serve meta-llama/Llama-3.2-1B-Instruct --dtype float16 --max-model-len 2048 --max-num-seqs 8 --enable-chunked-prefill --enforce-eager --tokenizer-mode auto --tensor-parallel-size 2 --pipeline-parallel-size 2 --distributed-executor-backend mp --compilation_config {"level": 3, "custom_ops": ["+rms_norm"], "compile_sizes": [4, 8], "pass_config": {"enable_sequence_parallelism": true, "enable_fusion": false, "enable_noop": true}} --load-format dummy --port 52029 --seed 0 2025-10-10T01:49:00.6438011Z /opt/conda/envs/py_3.12/lib/python3.12/site-packages/torch/cuda/__init__.py:64: FutureWarning: The pynvml package is deprecated. Please install nvidia-ml-py instead. If you did not install pynvml directly, please report this to the maintainers of the package that installed pynvml for you. 2025-10-10T01:49:00.6439269Z import pynvml # type: ignore[import] 2025-10-10T01:49:02.8500556Z INFO 10-10 01:49:02 [__init__.py:215] Automatically detected platform cuda. 2025-10-10T01:49:06.7458748Z (APIServer pid=54991) INFO 10-10 01:49:06 [api_server.py:1839] vLLM API server version 0.11.0rc2.dev157+g0ad9951c4.d20251010 2025-10-10T01:49:06.7486547Z (APIServer pid=54991) INFO 10-10 01:49:06 [utils.py:233] non-default args: {'model_tag': 'meta-llama/Llama-3.2-1B-Instruct', 'port': 52029, 'model': 'meta-llama/Llama-3.2-1B-Instruct', 'dtype': 'float16', 'seed': 0, 'max_model_len': 2048, 'enforce_eager': True, 'load_format': 'dummy', 'distributed_executor_backend': 'mp', 'pipeline_parallel_size': 2, 'tensor_parallel_size': 2, 'max_num_seqs': 8, 'enable_chunked_prefill': True, 'compilation_config': {'level': 3, 'debug_dump_path': None, 'cache_dir': '', 'backend': '', 'custom_ops': ['+rms_norm'], 'splitting_ops': None, 'use_inductor': True, 'compile_sizes': [4, 8], 'inductor_compile_config': {'enable_auto_functionalized_v2': False}, 'inductor_passes': {}, 'cudagraph_mode': None, 'use_cudagraph': True, 'cudagraph_num_of_warmups': 0, 'cudagraph_capture_sizes': None, 'cudagraph_copy_inputs': False, 'full_cuda_graph': False, 'use_inductor_graph_partition': False, 'pass_config': {'enable_noop': True, 'enable_sequence_parallelism': True}, 'max_capture_size': None, 'local_cache_dir': None}} 2025-10-10T01:49:06.8814618Z (APIServer pid=54991) INFO 10-10 01:49:06 [model.py:551] Resolved architecture: LlamaForCausalLM 2025-10-10T01:49:06.8816069Z (APIServer pid=54991) `torch_dtype` is deprecated! Use `dtype` instead! 2025-10-10T01:49:06.8817484Z (APIServer pid=54991) WARNING 10-10 01:49:06 [model.py:1768] Casting torch.bfloat16 to torch.float16. 2025-10-10T01:49:06.8819483Z (APIServer pid=54991) INFO 10-10 01:49:06 [model.py:1545] Using max model len 2048 2025-10-10T01:49:07.0415982Z (APIServer pid=54991) INFO 10-10 01:49:07 [scheduler.py:222] Chunked prefill is enabled with max_num_batched_tokens=2048. 2025-10-10T01:49:07.0420972Z (APIServer pid=54991) INFO 10-10 01:49:07 [vllm.py:345] Cudagraph is disabled under eager mode 2025-10-10T01:49:08.0086446Z /opt/conda/envs/py_3.12/lib/python3.12/site-packages/torch/cuda/__init__.py:64: FutureWarning: The pynvml package is deprecated. Please install nvidia-ml-py instead. If you did not install pynvml directly, please report this to the maintainers of the package that installed pynvml for you. 2025-10-10T01:49:08.0088766Z import pynvml # type: ignore[import] 2025-10-10T01:49:10.3149452Z INFO 10-10 01:49:10 [__init__.py:215] Automatically detected platform cuda. 2025-10-10T01:49:14.2969225Z (EngineCore_DP0 pid=55113) INFO 10-10 01:49:14 [core.py:648] Waiting for init message from front-end. 2025-10-10T01:49:14.3237929Z (EngineCore_DP0 pid=55113) INFO 10-10 01:49:14 [core.py:78] Initializing a V1 LLM engine (v0.11.0rc2.dev157+g0ad9951c4.d20251010) with config: model='meta-llama/Llama-3.2-1B-Instruct', speculative_config=None, tokenizer='meta-llama/Llama-3.2-1B-Instruct', skip_tokenizer_init=False, tokenizer_mode=auto, revision=None, tokenizer_revision=None, trust_remote_code=False, dtype=torch.float16, max_seq_len=2048, download_dir=None, load_format=dummy, tensor_parallel_size=2, pipeline_parallel_size=2, data_parallel_size=1, disable_custom_all_reduce=False, quantization=None, enforce_eager=True, kv_cache_dtype=auto, device_config=cuda, structured_outputs_config=StructuredOutputsConfig(backend='auto', disable_fallback=False, disable_any_whitespace=False, disable_additional_properties=False, reasoning_parser=''), observability_config=ObservabilityConfig(show_hidden_metrics_for_version=None, otlp_traces_endpoint=None, collect_detailed_traces=None), seed=0, served_model_name=meta-llama/Llama-3.2-1B-Instruct, enable_prefix_caching=True, chunked_prefill_enabled=True, pooler_config=None, compilation_config={'level': 3, 'debug_dump_path': None, 'cache_dir': '', 'backend': '', 'custom_ops': ['+rms_norm', '+rms_norm'], 'splitting_ops': ['vllm.unified_attention', 'vllm.unified_attention_with_output', 'vllm.mamba_mixer2', 'vllm.mamba_mixer', 'vllm.short_conv', 'vllm.linear_attention', 'vllm.plamo2_mamba_mixer', 'vllm.gdn_attention', 'vllm.sparse_attn_indexer'], 'use_inductor': True, 'compile_sizes': [8, 4], 'inductor_compile_config': {'enable_auto_functionalized_v2': False}, 'inductor_passes': {}, 'cudagraph_mode': , 'use_cudagraph': False, 'cudagraph_num_of_warmups': 0, 'cudagraph_capture_sizes': [], 'cudagraph_copy_inputs': False, 'full_cuda_graph': False, 'use_inductor_graph_partition': False, 'pass_config': {'enable_noop': True, 'enable_sequence_parallelism': True}, 'max_capture_size': 0, 'local_cache_dir': None} 2025-10-10T01:49:14.3244575Z (EngineCore_DP0 pid=55113) WARNING 10-10 01:49:14 [multiproc_executor.py:720] Reducing Torch parallelism from 24 threads to 1 to avoid unnecessary CPU contention. Set OMP_NUM_THREADS in the external environment to tune this value as needed. 2025-10-10T01:49:14.3249554Z (EngineCore_DP0 pid=55113) INFO 10-10 01:49:14 [shm_broadcast.py:289] vLLM message queue communication handle: Handle(local_reader_ranks=[0, 1, 2, 3], buffer_handle=(4, 16777216, 10, 'psm_9674ab4c'), local_subscribe_addr='ipc:///tmp/18af2baf-967b-418a-8aac-01cca3f2d694', remote_subscribe_addr=None, remote_addr_ipv6=False) 2025-10-10T01:49:14.8221744Z /opt/conda/envs/py_3.12/lib/python3.12/site-packages/torch/cuda/__init__.py:64: FutureWarning: The pynvml package is deprecated. Please install nvidia-ml-py instead. If you did not install pynvml directly, please report this to the maintainers of the package that installed pynvml for you. 2025-10-10T01:49:14.8223315Z import pynvml # type: ignore[import] 2025-10-10T01:49:14.8224701Z /opt/conda/envs/py_3.12/lib/python3.12/site-packages/torch/cuda/__init__.py:64: FutureWarning: The pynvml package is deprecated. Please install nvidia-ml-py instead. If you did not install pynvml directly, please report this to the maintainers of the package that installed pynvml for you. 2025-10-10T01:49:14.8225875Z import pynvml # type: ignore[import] 2025-10-10T01:49:14.8227018Z /opt/conda/envs/py_3.12/lib/python3.12/site-packages/torch/cuda/__init__.py:64: FutureWarning: The pynvml package is deprecated. Please install nvidia-ml-py instead. If you did not install pynvml directly, please report this to the maintainers of the package that installed pynvml for you. 2025-10-10T01:49:14.8228142Z import pynvml # type: ignore[import] 2025-10-10T01:49:14.9707140Z /opt/conda/envs/py_3.12/lib/python3.12/site-packages/torch/cuda/__init__.py:64: FutureWarning: The pynvml package is deprecated. Please install nvidia-ml-py instead. If you did not install pynvml directly, please report this to the maintainers of the package that installed pynvml for you. 2025-10-10T01:49:14.9708403Z import pynvml # type: ignore[import] 2025-10-10T01:49:17.1592892Z INFO 10-10 01:49:17 [__init__.py:215] Automatically detected platform cuda. 2025-10-10T01:49:17.1979375Z INFO 10-10 01:49:17 [__init__.py:215] Automatically detected platform cuda. 2025-10-10T01:49:17.2089706Z INFO 10-10 01:49:17 [__init__.py:215] Automatically detected platform cuda. 2025-10-10T01:49:17.2946184Z INFO 10-10 01:49:17 [__init__.py:215] Automatically detected platform cuda. 2025-10-10T01:49:21.6937316Z INFO 10-10 01:49:21 [shm_broadcast.py:289] vLLM message queue communication handle: Handle(local_reader_ranks=[0], buffer_handle=(1, 10485760, 10, 'psm_7ce6a69d'), local_subscribe_addr='ipc:///tmp/1b8fbed7-ed06-4c3c-aacc-2ed22e49e0b9', remote_subscribe_addr=None, remote_addr_ipv6=False) 2025-10-10T01:49:21.8140775Z INFO 10-10 01:49:21 [shm_broadcast.py:289] vLLM message queue communication handle: Handle(local_reader_ranks=[0], buffer_handle=(1, 10485760, 10, 'psm_28f0c9f1'), local_subscribe_addr='ipc:///tmp/0362fb7b-c9e1-4ef1-96fe-a8fce3893146', remote_subscribe_addr=None, remote_addr_ipv6=False) 2025-10-10T01:49:21.9698901Z INFO 10-10 01:49:21 [shm_broadcast.py:289] vLLM message queue communication handle: Handle(local_reader_ranks=[0], buffer_handle=(1, 10485760, 10, 'psm_364fee71'), local_subscribe_addr='ipc:///tmp/65754233-0f31-4cdd-b16c-8b55eb20d1d5', remote_subscribe_addr=None, remote_addr_ipv6=False) 2025-10-10T01:49:22.0040942Z INFO 10-10 01:49:22 [shm_broadcast.py:289] vLLM message queue communication handle: Handle(local_reader_ranks=[0], buffer_handle=(1, 10485760, 10, 'psm_28d3ee09'), local_subscribe_addr='ipc:///tmp/54383da8-3c1c-4a1b-b4b1-073672c66029', remote_subscribe_addr=None, remote_addr_ipv6=False) 2025-10-10T01:49:22.6807309Z [Gloo] Rank 0 is connected to 3 peer ranks. Expected number of connected peer ranks is : 3 2025-10-10T01:49:22.6807959Z [Gloo] Rank 2 is connected to 3 peer ranks. Expected number of connected peer ranks is : 3 2025-10-10T01:49:22.6808524Z [Gloo] Rank 1 is connected to 3 peer ranks. Expected number of connected peer ranks is : 3 2025-10-10T01:49:22.6809124Z [Gloo] Rank 3 is connected to 3 peer ranks. Expected number of connected peer ranks is : 3 2025-10-10T01:49:22.6859030Z [Gloo] Rank 0 is connected to 1 peer ranks. Expected number of connected peer ranks is : 1 2025-10-10T01:49:22.6859661Z [Gloo] Rank 1 is connected to 1 peer ranks. Expected number of connected peer ranks is : 1 2025-10-10T01:49:22.6860256Z [Gloo] Rank 1 is connected to 1 peer ranks. Expected number of connected peer ranks is : 1 2025-10-10T01:49:22.6860802Z [Gloo] Rank 0 is connected to 1 peer ranks. Expected number of connected peer ranks is : 1 2025-10-10T01:49:22.6901428Z INFO 10-10 01:49:22 [__init__.py:1384] Found nccl from library libnccl.so.2 2025-10-10T01:49:22.6902416Z INFO 10-10 01:49:22 [__init__.py:1384] Found nccl from library libnccl.so.2 2025-10-10T01:49:22.6903352Z INFO 10-10 01:49:22 [__init__.py:1384] Found nccl from library libnccl.so.2 2025-10-10T01:49:22.6904026Z INFO 10-10 01:49:22 [pynccl.py:104] vLLM is using nccl==2.27.5 2025-10-10T01:49:22.6904611Z INFO 10-10 01:49:22 [pynccl.py:104] vLLM is using nccl==2.27.5 2025-10-10T01:49:22.6905262Z INFO 10-10 01:49:22 [pynccl.py:104] vLLM is using nccl==2.27.5 2025-10-10T01:49:22.6907353Z INFO 10-10 01:49:22 [__init__.py:1384] Found nccl from library libnccl.so.2 2025-10-10T01:49:22.6908534Z INFO 10-10 01:49:22 [pynccl.py:104] vLLM is using nccl==2.27.5 2025-10-10T01:49:22.8800013Z WARNING 10-10 01:49:22 [symm_mem.py:58] SymmMemCommunicator: Device capability 8.9 not supported, communicator is not available. 2025-10-10T01:49:22.8800883Z WARNING 10-10 01:49:22 [symm_mem.py:58] SymmMemCommunicator: Device capability 8.9 not supported, communicator is not available. 2025-10-10T01:49:22.8823373Z INFO 10-10 01:49:22 [custom_all_reduce.py:35] Skipping P2P check and trusting the driver's P2P report. 2025-10-10T01:49:22.8824268Z INFO 10-10 01:49:22 [custom_all_reduce.py:35] Skipping P2P check and trusting the driver's P2P report. 2025-10-10T01:49:22.8825308Z WARNING 10-10 01:49:22 [custom_all_reduce.py:154] Custom allreduce is disabled because your platform lacks GPU P2P capability or P2P test failed. To silence this warning, specify disable_custom_all_reduce=True explicitly. 2025-10-10T01:49:22.8826860Z WARNING 10-10 01:49:22 [custom_all_reduce.py:154] Custom allreduce is disabled because your platform lacks GPU P2P capability or P2P test failed. To silence this warning, specify disable_custom_all_reduce=True explicitly. 2025-10-10T01:49:22.8840794Z INFO 10-10 01:49:22 [shm_broadcast.py:289] vLLM message queue communication handle: Handle(local_reader_ranks=[1], buffer_handle=(1, 4194304, 6, 'psm_c209c39f'), local_subscribe_addr='ipc:///tmp/1b986787-287d-435c-9438-b244f3ffb505', remote_subscribe_addr=None, remote_addr_ipv6=False) 2025-10-10T01:49:22.9377775Z WARNING 10-10 01:49:22 [symm_mem.py:58] SymmMemCommunicator: Device capability 8.9 not supported, communicator is not available. 2025-10-10T01:49:22.9379136Z WARNING 10-10 01:49:22 [symm_mem.py:58] SymmMemCommunicator: Device capability 8.9 not supported, communicator is not available. 2025-10-10T01:49:22.9399995Z INFO 10-10 01:49:22 [custom_all_reduce.py:35] Skipping P2P check and trusting the driver's P2P report. 2025-10-10T01:49:22.9400690Z INFO 10-10 01:49:22 [custom_all_reduce.py:35] Skipping P2P check and trusting the driver's P2P report. 2025-10-10T01:49:22.9402050Z WARNING 10-10 01:49:22 [custom_all_reduce.py:154] Custom allreduce is disabled because your platform lacks GPU P2P capability or P2P test failed. To silence this warning, specify disable_custom_all_reduce=True explicitly. 2025-10-10T01:49:22.9403844Z WARNING 10-10 01:49:22 [custom_all_reduce.py:154] Custom allreduce is disabled because your platform lacks GPU P2P capability or P2P test failed. To silence this warning, specify disable_custom_all_reduce=True explicitly. 2025-10-10T01:49:22.9418776Z INFO 10-10 01:49:22 [shm_broadcast.py:289] vLLM message queue communication handle: Handle(local_reader_ranks=[1], buffer_handle=(1, 4194304, 6, 'psm_32a1376f'), local_subscribe_addr='ipc:///tmp/9113c184-2d4e-4a49-8a54-9197117990fc', remote_subscribe_addr=None, remote_addr_ipv6=False) 2025-10-10T01:49:22.9444281Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-10-10T01:49:22.9734395Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-10-10T01:49:22.9735641Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-10-10T01:49:22.9752386Z [Gloo] Rank 1 is connected to 1 peer ranks. Expected number of connected peer ranks is : 1 2025-10-10T01:49:22.9752949Z [Gloo] Rank 0 is connected to 1 peer ranks. Expected number of connected peer ranks is : 1 2025-10-10T01:49:22.9755982Z INFO 10-10 01:49:22 [__init__.py:1384] Found nccl from library libnccl.so.2 2025-10-10T01:49:22.9756649Z INFO 10-10 01:49:22 [__init__.py:1384] Found nccl from library libnccl.so.2 2025-10-10T01:49:22.9757054Z INFO 10-10 01:49:22 [pynccl.py:104] vLLM is using nccl==2.27.5 2025-10-10T01:49:22.9757432Z INFO 10-10 01:49:22 [pynccl.py:104] vLLM is using nccl==2.27.5 2025-10-10T01:49:23.0085803Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-10-10T01:49:23.0086470Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-10-10T01:49:23.0635785Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-10-10T01:49:23.0652605Z [Gloo] Rank 0 is connected to 1 peer ranks. Expected number of connected peer ranks is : 1 2025-10-10T01:49:23.0653240Z [Gloo] Rank 1 is connected to 1 peer ranks. Expected number of connected peer ranks is : 1 2025-10-10T01:49:23.0656931Z INFO 10-10 01:49:23 [__init__.py:1384] Found nccl from library libnccl.so.2 2025-10-10T01:49:23.0657738Z INFO 10-10 01:49:23 [pynccl.py:104] vLLM is using nccl==2.27.5 2025-10-10T01:49:23.0658208Z INFO 10-10 01:49:23 [__init__.py:1384] Found nccl from library libnccl.so.2 2025-10-10T01:49:23.0658651Z INFO 10-10 01:49:23 [pynccl.py:104] vLLM is using nccl==2.27.5 2025-10-10T01:49:23.1003153Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-10-10T01:49:23.1004053Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-10-10T01:49:23.1014157Z [Gloo] Rank 1 is connected to 1 peer ranks. Expected number of connected peer ranks is : 1 2025-10-10T01:49:23.1014759Z [Gloo] Rank 1 is connected to 1 peer ranks. Expected number of connected peer ranks is : 1 2025-10-10T01:49:23.1015386Z [Gloo] Rank 0 is connected to 1 peer ranks. Expected number of connected peer ranks is : 1 2025-10-10T01:49:23.1016261Z [Gloo] Rank 0 is connected to 1 peer ranks. Expected number of connected peer ranks is : 1 2025-10-10T01:49:23.1017199Z INFO 10-10 01:49:23 [__init__.py:1384] Found nccl from library libnccl.so.2 2025-10-10T01:49:23.1017947Z INFO 10-10 01:49:23 [__init__.py:1384] Found nccl from library libnccl.so.2 2025-10-10T01:49:23.1018365Z INFO 10-10 01:49:23 [pynccl.py:104] vLLM is using nccl==2.27.5 2025-10-10T01:49:23.1018786Z INFO 10-10 01:49:23 [pynccl.py:104] vLLM is using nccl==2.27.5 2025-10-10T01:49:23.1019183Z INFO 10-10 01:49:23 [__init__.py:1384] Found nccl from library libnccl.so.2 2025-10-10T01:49:23.1019617Z INFO 10-10 01:49:23 [__init__.py:1384] Found nccl from library libnccl.so.2 2025-10-10T01:49:23.1020008Z INFO 10-10 01:49:23 [pynccl.py:104] vLLM is using nccl==2.27.5 2025-10-10T01:49:23.1020375Z INFO 10-10 01:49:23 [pynccl.py:104] vLLM is using nccl==2.27.5 2025-10-10T01:49:23.1477924Z INFO 10-10 01:49:23 [parallel_state.py:1208] rank 3 in world size 4 is assigned as DP rank 0, PP rank 1, TP rank 1, EP rank 1 2025-10-10T01:49:23.1478852Z INFO 10-10 01:49:23 [parallel_state.py:1208] rank 2 in world size 4 is assigned as DP rank 0, PP rank 1, TP rank 0, EP rank 0 2025-10-10T01:49:23.1479589Z INFO 10-10 01:49:23 [parallel_state.py:1208] rank 0 in world size 4 is assigned as DP rank 0, PP rank 0, TP rank 0, EP rank 0 2025-10-10T01:49:23.1480280Z INFO 10-10 01:49:23 [parallel_state.py:1208] rank 1 in world size 4 is assigned as DP rank 0, PP rank 0, TP rank 1, EP rank 1 2025-10-10T01:49:23.4176990Z INFO 10-10 01:49:23 [topk_topp_sampler.py:55] Using FlashInfer for top-p & top-k sampling. 2025-10-10T01:49:23.4230082Z INFO 10-10 01:49:23 [topk_topp_sampler.py:55] Using FlashInfer for top-p & top-k sampling. 2025-10-10T01:49:23.4247623Z INFO 10-10 01:49:23 [topk_topp_sampler.py:55] Using FlashInfer for top-p & top-k sampling. 2025-10-10T01:49:23.4264232Z INFO 10-10 01:49:23 [topk_topp_sampler.py:55] Using FlashInfer for top-p & top-k sampling. 2025-10-10T01:49:23.4483335Z (Worker_PP1_TP0 pid=55234) INFO 10-10 01:49:23 [gpu_model_runner.py:2707] Starting to load model meta-llama/Llama-3.2-1B-Instruct... 2025-10-10T01:49:23.4538899Z (Worker_PP0_TP0 pid=55232) INFO 10-10 01:49:23 [gpu_model_runner.py:2707] Starting to load model meta-llama/Llama-3.2-1B-Instruct... 2025-10-10T01:49:23.4551924Z (Worker_PP0_TP1 pid=55233) INFO 10-10 01:49:23 [gpu_model_runner.py:2707] Starting to load model meta-llama/Llama-3.2-1B-Instruct... 2025-10-10T01:49:23.4565642Z (Worker_PP1_TP1 pid=55235) INFO 10-10 01:49:23 [gpu_model_runner.py:2707] Starting to load model meta-llama/Llama-3.2-1B-Instruct... 2025-10-10T01:49:23.6821376Z (Worker_PP1_TP0 pid=55234) INFO 10-10 01:49:23 [gpu_model_runner.py:2739] Loading model from scratch... 2025-10-10T01:49:23.6878842Z (Worker_PP0_TP0 pid=55232) INFO 10-10 01:49:23 [gpu_model_runner.py:2739] Loading model from scratch... 2025-10-10T01:49:23.7142658Z (Worker_PP1_TP1 pid=55235) INFO 10-10 01:49:23 [gpu_model_runner.py:2739] Loading model from scratch... 2025-10-10T01:49:23.7187794Z (Worker_PP0_TP1 pid=55233) INFO 10-10 01:49:23 [gpu_model_runner.py:2739] Loading model from scratch... 2025-10-10T01:49:23.7731719Z (Worker_PP1_TP0 pid=55234) INFO 10-10 01:49:23 [cuda.py:361] Using Flash Attention backend on V1 engine. 2025-10-10T01:49:23.7784592Z (Worker_PP0_TP0 pid=55232) INFO 10-10 01:49:23 [cuda.py:361] Using Flash Attention backend on V1 engine. 2025-10-10T01:49:23.8054550Z (Worker_PP1_TP1 pid=55235) INFO 10-10 01:49:23 [cuda.py:361] Using Flash Attention backend on V1 engine. 2025-10-10T01:49:23.8089140Z (Worker_PP0_TP1 pid=55233) INFO 10-10 01:49:23 [cuda.py:361] Using Flash Attention backend on V1 engine. 2025-10-10T01:49:24.2923575Z (Worker_PP0_TP0 pid=55232) INFO 10-10 01:49:24 [gpu_model_runner.py:2758] Model loading took 0.7135 GiB and 0.153732 seconds 2025-10-10T01:49:24.2945231Z (Worker_PP1_TP0 pid=55234) INFO 10-10 01:49:24 [gpu_model_runner.py:2758] Model loading took 0.7135 GiB and 0.153136 seconds 2025-10-10T01:49:24.3774408Z (Worker_PP0_TP1 pid=55233) INFO 10-10 01:49:24 [gpu_model_runner.py:2758] Model loading took 0.7135 GiB and 0.152785 seconds 2025-10-10T01:49:24.3934806Z (Worker_PP1_TP1 pid=55235) INFO 10-10 01:49:24 [gpu_model_runner.py:2758] Model loading took 0.7135 GiB and 0.156465 seconds 2025-10-10T01:49:26.6037348Z (Worker_PP1_TP0 pid=55234) INFO 10-10 01:49:26 [backends.py:548] Using cache directory: /var/lib/jenkins/.cache/vllm/torch_compile_cache/8cedb68d99/rank_2_0/backbone for vLLM's torch.compile 2025-10-10T01:49:26.6043445Z (Worker_PP1_TP0 pid=55234) INFO 10-10 01:49:26 [backends.py:559] Dynamo bytecode transform time: 1.98 s 2025-10-10T01:49:26.6332779Z (Worker_PP1_TP1 pid=55235) INFO 10-10 01:49:26 [backends.py:548] Using cache directory: /var/lib/jenkins/.cache/vllm/torch_compile_cache/8cedb68d99/rank_3_0/backbone for vLLM's torch.compile 2025-10-10T01:49:26.6340782Z (Worker_PP1_TP1 pid=55235) INFO 10-10 01:49:26 [backends.py:559] Dynamo bytecode transform time: 1.98 s 2025-10-10T01:49:26.6408900Z (Worker_PP0_TP0 pid=55232) INFO 10-10 01:49:26 [backends.py:548] Using cache directory: /var/lib/jenkins/.cache/vllm/torch_compile_cache/5280ab4e2b/rank_0_0/backbone for vLLM's torch.compile 2025-10-10T01:49:26.6416492Z (Worker_PP0_TP0 pid=55232) INFO 10-10 01:49:26 [backends.py:559] Dynamo bytecode transform time: 2.03 s 2025-10-10T01:49:26.6791253Z (Worker_PP0_TP1 pid=55233) INFO 10-10 01:49:26 [backends.py:548] Using cache directory: /var/lib/jenkins/.cache/vllm/torch_compile_cache/5280ab4e2b/rank_1_0/backbone for vLLM's torch.compile 2025-10-10T01:49:26.6798483Z (Worker_PP0_TP1 pid=55233) INFO 10-10 01:49:26 [backends.py:559] Dynamo bytecode transform time: 2.03 s 2025-10-10T01:49:27.1827930Z (Worker_PP1_TP0 pid=55234) INFO 10-10 01:49:27 [backends.py:164] Directly load the compiled graph(s) for dynamic shape from the cache, took 0.387 s 2025-10-10T01:49:27.2325050Z (Worker_PP0_TP0 pid=55232) INFO 10-10 01:49:27 [backends.py:164] Directly load the compiled graph(s) for dynamic shape from the cache, took 0.389 s 2025-10-10T01:49:27.2336036Z (Worker_PP1_TP1 pid=55235) INFO 10-10 01:49:27 [backends.py:164] Directly load the compiled graph(s) for dynamic shape from the cache, took 0.404 s 2025-10-10T01:49:27.2887058Z (Worker_PP0_TP1 pid=55233) INFO 10-10 01:49:27 [backends.py:164] Directly load the compiled graph(s) for dynamic shape from the cache, took 0.402 s 2025-10-10T01:49:28.2199560Z (Worker_PP0_TP0 pid=55232) INFO 10-10 01:49:28 [gpu_worker.py:298] Available KV cache memory: 19.02 GiB 2025-10-10T01:49:28.2262676Z (Worker_PP1_TP0 pid=55234) INFO 10-10 01:49:28 [gpu_worker.py:298] Available KV cache memory: 18.99 GiB 2025-10-10T01:49:28.2839139Z (Worker_PP0_TP1 pid=55233) INFO 10-10 01:49:28 [gpu_worker.py:298] Available KV cache memory: 19.02 GiB 2025-10-10T01:49:28.3015379Z (Worker_PP1_TP1 pid=55235) INFO 10-10 01:49:28 [gpu_worker.py:298] Available KV cache memory: 18.99 GiB 2025-10-10T01:49:28.5913366Z (EngineCore_DP0 pid=55113) INFO 10-10 01:49:28 [kv_cache_utils.py:1087] GPU KV cache size: 2,492,416 tokens 2025-10-10T01:49:28.5914812Z (EngineCore_DP0 pid=55113) INFO 10-10 01:49:28 [kv_cache_utils.py:1091] Maximum concurrency for 2,048 tokens per request: 1217.00x 2025-10-10T01:49:28.5916482Z (EngineCore_DP0 pid=55113) INFO 10-10 01:49:28 [kv_cache_utils.py:1087] GPU KV cache size: 2,492,416 tokens 2025-10-10T01:49:28.5917623Z (EngineCore_DP0 pid=55113) INFO 10-10 01:49:28 [kv_cache_utils.py:1091] Maximum concurrency for 2,048 tokens per request: 1217.00x 2025-10-10T01:49:28.5918911Z (EngineCore_DP0 pid=55113) INFO 10-10 01:49:28 [kv_cache_utils.py:1087] GPU KV cache size: 2,489,088 tokens 2025-10-10T01:49:28.5920136Z (EngineCore_DP0 pid=55113) INFO 10-10 01:49:28 [kv_cache_utils.py:1091] Maximum concurrency for 2,048 tokens per request: 1215.38x 2025-10-10T01:49:28.5921371Z (EngineCore_DP0 pid=55113) INFO 10-10 01:49:28 [kv_cache_utils.py:1087] GPU KV cache size: 2,489,088 tokens 2025-10-10T01:49:28.5922575Z (EngineCore_DP0 pid=55113) INFO 10-10 01:49:28 [kv_cache_utils.py:1091] Maximum concurrency for 2,048 tokens per request: 1215.38x 2025-10-10T01:49:28.5988133Z (Worker_PP0_TP0 pid=55232) INFO 10-10 01:49:28 [gpu_worker.py:332] Compile and warming up model for size 8 2025-10-10T01:49:28.5989184Z (Worker_PP1_TP1 pid=55235) INFO 10-10 01:49:28 [gpu_worker.py:332] Compile and warming up model for size 8 2025-10-10T01:49:28.5990269Z (Worker_PP1_TP0 pid=55234) INFO 10-10 01:49:28 [gpu_worker.py:332] Compile and warming up model for size 8 2025-10-10T01:49:28.5991394Z (Worker_PP0_TP1 pid=55233) INFO 10-10 01:49:28 [gpu_worker.py:332] Compile and warming up model for size 8 2025-10-10T01:49:29.0064584Z (Worker_PP1_TP0 pid=55234) INFO 10-10 01:49:29 [backends.py:168] Directly load the compiled graph(s) for shape 8 from the cache, took 0.407 s 2025-10-10T01:49:29.0099566Z (Worker_PP1_TP1 pid=55235) INFO 10-10 01:49:29 [backends.py:168] Directly load the compiled graph(s) for shape 8 from the cache, took 0.411 s 2025-10-10T01:49:29.0105825Z (Worker_PP1_TP0 pid=55234) INFO 10-10 01:49:29 [gpu_worker.py:332] Compile and warming up model for size 4 2025-10-10T01:49:29.0107241Z (Worker_PP1_TP1 pid=55235) INFO 10-10 01:49:29 [gpu_worker.py:332] Compile and warming up model for size 4 2025-10-10T01:49:29.0454325Z (Worker_PP0_TP0 pid=55232) INFO 10-10 01:49:29 [backends.py:168] Directly load the compiled graph(s) for shape 8 from the cache, took 0.446 s 2025-10-10T01:49:29.0459807Z (Worker_PP0_TP0 pid=55232) INFO 10-10 01:49:29 [gpu_worker.py:332] Compile and warming up model for size 4 2025-10-10T01:49:29.0509054Z (Worker_PP0_TP1 pid=55233) INFO 10-10 01:49:29 [backends.py:168] Directly load the compiled graph(s) for shape 8 from the cache, took 0.452 s 2025-10-10T01:49:29.0514733Z (Worker_PP0_TP1 pid=55233) INFO 10-10 01:49:29 [gpu_worker.py:332] Compile and warming up model for size 4 2025-10-10T01:49:29.1729315Z (Worker_PP0_TP0 pid=55232) INFO 10-10 01:49:29 [backends.py:168] Directly load the compiled graph(s) for shape 4 from the cache, took 0.127 s 2025-10-10T01:49:29.1730192Z (Worker_PP0_TP0 pid=55232) INFO 10-10 01:49:29 [monitor.py:32] torch.compile takes 2.03 s in total 2025-10-10T01:49:29.1761309Z (Worker_PP0_TP1 pid=55233) INFO 10-10 01:49:29 [backends.py:168] Directly load the compiled graph(s) for shape 4 from the cache, took 0.124 s 2025-10-10T01:49:29.1762163Z (Worker_PP0_TP1 pid=55233) INFO 10-10 01:49:29 [monitor.py:32] torch.compile takes 2.03 s in total 2025-10-10T01:49:30.5153308Z (Worker_PP1_TP0 pid=55234) INFO 10-10 01:49:30 [backends.py:168] Directly load the compiled graph(s) for shape 4 from the cache, took 1.504 s 2025-10-10T01:49:30.5154351Z (Worker_PP1_TP0 pid=55234) INFO 10-10 01:49:30 [monitor.py:32] torch.compile takes 1.98 s in total 2025-10-10T01:49:30.5161592Z (Worker_PP1_TP1 pid=55235) INFO 10-10 01:49:30 [backends.py:168] Directly load the compiled graph(s) for shape 4 from the cache, took 1.505 s 2025-10-10T01:49:30.5162372Z (Worker_PP1_TP1 pid=55235) INFO 10-10 01:49:30 [monitor.py:32] torch.compile takes 1.98 s in total 2025-10-10T01:49:30.5238882Z (EngineCore_DP0 pid=55113) INFO 10-10 01:49:30 [core.py:211] init engine (profile, create kv cache, warmup model) took 6.13 seconds 2025-10-10T01:49:31.2577422Z (EngineCore_DP0 pid=55113) INFO 10-10 01:49:31 [core.py:150] Batch queue is enabled with size 2 2025-10-10T01:49:31.2606840Z (EngineCore_DP0 pid=55113) INFO 10-10 01:49:31 [vllm.py:345] Cudagraph is disabled under eager mode 2025-10-10T01:49:31.4450092Z (APIServer pid=54991) INFO 10-10 01:49:31 [loggers.py:147] Engine 000: vllm cache_config_info with initialization after num_gpu_blocks is: 155568 2025-10-10T01:49:31.5251262Z (EngineCore_DP0 pid=55113) INFO 10-10 01:49:31 [gc_utils.py:41] GC Debug Config. enabled:False,top_objects:-1 2025-10-10T01:49:31.5465557Z (APIServer pid=54991) INFO 10-10 01:49:31 [api_server.py:1634] Supported_tasks: ['generate'] 2025-10-10T01:49:31.5731659Z (APIServer pid=54991) WARNING 10-10 01:49:31 [model.py:1424] Default sampling parameters have been overridden by the model's Hugging Face generation config recommended from the model creator. If this is not intended, please relaunch vLLM instance with `--generation-config vllm`. 2025-10-10T01:49:31.5733209Z (APIServer pid=54991) INFO 10-10 01:49:31 [serving_responses.py:140] Using default chat sampling params from model: {'temperature': 0.6, 'top_p': 0.9} 2025-10-10T01:49:31.5976088Z (APIServer pid=54991) INFO 10-10 01:49:31 [serving_chat.py:139] Using default chat sampling params from model: {'temperature': 0.6, 'top_p': 0.9} 2025-10-10T01:49:31.6222304Z (APIServer pid=54991) INFO 10-10 01:49:31 [serving_completion.py:76] Using default completion sampling params from model: {'temperature': 0.6, 'top_p': 0.9} 2025-10-10T01:49:31.6223354Z (APIServer pid=54991) INFO 10-10 01:49:31 [api_server.py:1912] Starting vLLM API server 0 on http://0.0.0.0:52029 2025-10-10T01:49:31.6224303Z (APIServer pid=54991) INFO 10-10 01:49:31 [launcher.py:34] Available routes are: 2025-10-10T01:49:31.6224992Z (APIServer pid=54991) INFO 10-10 01:49:31 [launcher.py:42] Route: /openapi.json, Methods: GET, HEAD 2025-10-10T01:49:31.6225709Z (APIServer pid=54991) INFO 10-10 01:49:31 [launcher.py:42] Route: /docs, Methods: GET, HEAD 2025-10-10T01:49:31.6226447Z (APIServer pid=54991) INFO 10-10 01:49:31 [launcher.py:42] Route: /docs/oauth2-redirect, Methods: GET, HEAD 2025-10-10T01:49:31.6227117Z (APIServer pid=54991) INFO 10-10 01:49:31 [launcher.py:42] Route: /redoc, Methods: GET, HEAD 2025-10-10T01:49:31.6227879Z (APIServer pid=54991) INFO 10-10 01:49:31 [launcher.py:42] Route: /health, Methods: GET 2025-10-10T01:49:31.6228385Z (APIServer pid=54991) INFO 10-10 01:49:31 [launcher.py:42] Route: /load, Methods: GET 2025-10-10T01:49:31.6228912Z (APIServer pid=54991) INFO 10-10 01:49:31 [launcher.py:42] Route: /ping, Methods: POST 2025-10-10T01:49:31.6229458Z (APIServer pid=54991) INFO 10-10 01:49:31 [launcher.py:42] Route: /ping, Methods: GET 2025-10-10T01:49:31.6229988Z (APIServer pid=54991) INFO 10-10 01:49:31 [launcher.py:42] Route: /tokenize, Methods: POST 2025-10-10T01:49:31.6230529Z (APIServer pid=54991) INFO 10-10 01:49:31 [launcher.py:42] Route: /detokenize, Methods: POST 2025-10-10T01:49:31.6231244Z (APIServer pid=54991) INFO 10-10 01:49:31 [launcher.py:42] Route: /v1/models, Methods: GET 2025-10-10T01:49:31.6231787Z (APIServer pid=54991) INFO 10-10 01:49:31 [launcher.py:42] Route: /version, Methods: GET 2025-10-10T01:49:31.6232319Z (APIServer pid=54991) INFO 10-10 01:49:31 [launcher.py:42] Route: /v1/responses, Methods: POST 2025-10-10T01:49:31.6232896Z (APIServer pid=54991) INFO 10-10 01:49:31 [launcher.py:42] Route: /v1/responses/{response_id}, Methods: GET 2025-10-10T01:49:31.6233601Z (APIServer pid=54991) INFO 10-10 01:49:31 [launcher.py:42] Route: /v1/responses/{response_id}/cancel, Methods: POST 2025-10-10T01:49:31.6234209Z (APIServer pid=54991) INFO 10-10 01:49:31 [launcher.py:42] Route: /v1/chat/completions, Methods: POST 2025-10-10T01:49:31.6234771Z (APIServer pid=54991) INFO 10-10 01:49:31 [launcher.py:42] Route: /v1/completions, Methods: POST 2025-10-10T01:49:31.6235319Z (APIServer pid=54991) INFO 10-10 01:49:31 [launcher.py:42] Route: /v1/embeddings, Methods: POST 2025-10-10T01:49:31.6235872Z (APIServer pid=54991) INFO 10-10 01:49:31 [launcher.py:42] Route: /pooling, Methods: POST 2025-10-10T01:49:31.6236407Z (APIServer pid=54991) INFO 10-10 01:49:31 [launcher.py:42] Route: /classify, Methods: POST 2025-10-10T01:49:31.6236936Z (APIServer pid=54991) INFO 10-10 01:49:31 [launcher.py:42] Route: /score, Methods: POST 2025-10-10T01:49:31.6237456Z (APIServer pid=54991) INFO 10-10 01:49:31 [launcher.py:42] Route: /v1/score, Methods: POST 2025-10-10T01:49:31.6238027Z (APIServer pid=54991) INFO 10-10 01:49:31 [launcher.py:42] Route: /v1/audio/transcriptions, Methods: POST 2025-10-10T01:49:31.6238714Z (APIServer pid=54991) INFO 10-10 01:49:31 [launcher.py:42] Route: /v1/audio/translations, Methods: POST 2025-10-10T01:49:31.6239262Z (APIServer pid=54991) INFO 10-10 01:49:31 [launcher.py:42] Route: /rerank, Methods: POST 2025-10-10T01:49:31.6239802Z (APIServer pid=54991) INFO 10-10 01:49:31 [launcher.py:42] Route: /v1/rerank, Methods: POST 2025-10-10T01:49:31.6240338Z (APIServer pid=54991) INFO 10-10 01:49:31 [launcher.py:42] Route: /v2/rerank, Methods: POST 2025-10-10T01:49:31.6240889Z (APIServer pid=54991) INFO 10-10 01:49:31 [launcher.py:42] Route: /scale_elastic_ep, Methods: POST 2025-10-10T01:49:31.6241479Z (APIServer pid=54991) INFO 10-10 01:49:31 [launcher.py:42] Route: /is_scaling_elastic_ep, Methods: POST 2025-10-10T01:49:31.6242050Z (APIServer pid=54991) INFO 10-10 01:49:31 [launcher.py:42] Route: /invocations, Methods: POST 2025-10-10T01:49:31.6242577Z (APIServer pid=54991) INFO 10-10 01:49:31 [launcher.py:42] Route: /metrics, Methods: GET 2025-10-10T01:49:31.6375812Z (APIServer pid=54991) INFO: Started server process [54991] 2025-10-10T01:49:31.6376488Z (APIServer pid=54991) INFO: Waiting for application startup. 2025-10-10T01:49:31.8732456Z (APIServer pid=54991) INFO: Application startup complete. 2025-10-10T01:49:32.2153588Z (APIServer pid=54991) INFO: 127.0.0.1:60726 - "GET /health HTTP/1.1" 200 OK 2025-10-10T01:49:32.4329639Z (APIServer pid=54991) INFO: 127.0.0.1:60728 - "GET /v1/models HTTP/1.1" 200 OK 2025-10-10T01:49:32.4636128Z (Worker_PP0_TP1 pid=55233) /opt/conda/envs/py_3.12/lib/python3.12/site-packages/vllm/distributed/parallel_state.py:516: UserWarning: The given buffer is not writable, and PyTorch does not support non-writable tensors. This means you can write to the underlying (supposedly non-writable) buffer using the tensor. You may want to copy the buffer to protect its data or make it writable before converting it to a tensor. This type of warning will be suppressed for the rest of this program. (Triggered internally at /var/lib/jenkins/workspace/torch/csrc/utils/tensor_new.cpp:1581.) 2025-10-10T01:49:32.4639265Z (Worker_PP0_TP1 pid=55233) object_tensor = torch.frombuffer(pickle.dumps(obj), dtype=torch.uint8) 2025-10-10T01:49:32.4642011Z (Worker_PP0_TP0 pid=55232) /opt/conda/envs/py_3.12/lib/python3.12/site-packages/vllm/distributed/parallel_state.py:516: UserWarning: The given buffer is not writable, and PyTorch does not support non-writable tensors. This means you can write to the underlying (supposedly non-writable) buffer using the tensor. You may want to copy the buffer to protect its data or make it writable before converting it to a tensor. This type of warning will be suppressed for the rest of this program. (Triggered internally at /var/lib/jenkins/workspace/torch/csrc/utils/tensor_new.cpp:1581.) 2025-10-10T01:49:32.4644469Z (Worker_PP0_TP0 pid=55232) object_tensor = torch.frombuffer(pickle.dumps(obj), dtype=torch.uint8) 2025-10-10T01:49:32.4645665Z [rank1]:[W1010 01:49:32.141127791 ProcessGroupNCCL.cpp:4063] Warning: An unbatched P2P op (send/recv) was called on this ProcessGroup with size 2. In lazy initialization mode, this will result in a new 2-rank NCCL communicator to be created. (function operator()) 2025-10-10T01:49:32.4647200Z [rank3]:[W1010 01:49:32.141255534 ProcessGroupNCCL.cpp:4063] Warning: An unbatched P2P op (send/recv) was called on this ProcessGroup with size 2. In lazy initialization mode, this will result in a new 2-rank NCCL communicator to be created. (function operator()) 2025-10-10T01:49:32.4648526Z [rank0]:[W1010 01:49:32.141498950 ProcessGroupNCCL.cpp:4063] Warning: An unbatched P2P op (send/recv) was called on this ProcessGroup with size 2. In lazy initialization mode, this will result in a new 2-rank NCCL communicator to be created. (function operator()) 2025-10-10T01:49:32.4649828Z [rank2]:[W1010 01:49:32.141593423 ProcessGroupNCCL.cpp:4063] Warning: An unbatched P2P op (send/recv) was called on this ProcessGroup with size 2. In lazy initialization mode, this will result in a new 2-rank NCCL communicator to be created. (function operator()) 2025-10-10T01:49:32.5880251Z (APIServer pid=54991) INFO: 127.0.0.1:60728 - "POST /v1/completions HTTP/1.1" 200 OK 2025-10-10T01:49:32.6524351Z (APIServer pid=54991) INFO: 127.0.0.1:60728 - "POST /v1/completions HTTP/1.1" 200 OK 2025-10-10T01:49:32.7948848Z (APIServer pid=54991) INFO: 127.0.0.1:60728 - "POST /v1/completions HTTP/1.1" 200 OK 2025-10-10T01:49:32.8863712Z (APIServer pid=54991) INFO: 127.0.0.1:60728 - "POST /v1/completions HTTP/1.1" 200 OK 2025-10-10T01:49:32.9620618Z (APIServer pid=54991) INFO: 127.0.0.1:60728 - "POST /v1/completions HTTP/1.1" 200 OK 2025-10-10T01:49:32.9696827Z (APIServer pid=54991) INFO: 127.0.0.1:60728 - "POST /v1/completions HTTP/1.1" 200 OK 2025-10-10T01:49:33.0552059Z (APIServer pid=54991) INFO 10-10 01:49:33 [launcher.py:99] Shutting down FastAPI HTTP server. 2025-10-10T01:49:33.0554676Z (Worker_PP0_TP0 pid=55232) INFO 10-10 01:49:33 [multiproc_executor.py:558] Parent process exited, terminating worker 2025-10-10T01:49:33.0556153Z (Worker_PP0_TP0 pid=55232) INFO 10-10 01:49:33 [multiproc_executor.py:599] WorkerProc shutting down. 2025-10-10T01:49:33.0558685Z (Worker_PP0_TP1 pid=55233) INFO 10-10 01:49:33 [multiproc_executor.py:558] Parent process exited, terminating worker 2025-10-10T01:49:33.0561123Z (Worker_PP0_TP1 pid=55233) INFO 10-10 01:49:33 [multiproc_executor.py:599] WorkerProc shutting down. 2025-10-10T01:49:33.0561785Z (Worker_PP1_TP0 pid=55234) INFO 10-10 01:49:33 [multiproc_executor.py:558] Parent process exited, terminating worker 2025-10-10T01:49:33.0564147Z (Worker_PP1_TP0 pid=55234) INFO 10-10 01:49:33 [multiproc_executor.py:599] WorkerProc shutting down. 2025-10-10T01:49:33.0564989Z (Worker_PP1_TP1 pid=55235) INFO 10-10 01:49:33 [multiproc_executor.py:558] Parent process exited, terminating worker 2025-10-10T01:49:33.0565669Z (Worker_PP1_TP1 pid=55235) INFO 10-10 01:49:33 [multiproc_executor.py:599] WorkerProc shutting down. 2025-10-10T01:49:35.6740108Z (APIServer pid=54991) INFO: Shutting down 2025-10-10T01:49:35.7744396Z (APIServer pid=54991) INFO: Waiting for application shutdown. 2025-10-10T01:49:35.7746588Z (APIServer pid=54991) INFO: Application shutdown complete. 2025-10-10T01:49:36.1411974Z INFO 10-10 01:49:36 [model.py:551] Resolved architecture: LlamaForCausalLM 2025-10-10T01:49:36.1413007Z WARNING 10-10 01:49:36 [model.py:1768] Casting torch.bfloat16 to torch.float16. 2025-10-10T01:49:36.1413835Z INFO 10-10 01:49:36 [model.py:1545] Using max model len 2048 2025-10-10T01:49:36.1416317Z Launching RemoteOpenAIServer with: vllm serve meta-llama/Llama-3.2-1B-Instruct --dtype float16 --max-model-len 2048 --max-num-seqs 8 --enable-chunked-prefill --enforce-eager --tokenizer-mode auto --tensor-parallel-size 2 --distributed-executor-backend mp --load-format dummy --port 37661 --seed 0 2025-10-10T01:49:36.6514933Z /opt/conda/envs/py_3.12/lib/python3.12/site-packages/torch/cuda/__init__.py:64: FutureWarning: The pynvml package is deprecated. Please install nvidia-ml-py instead. If you did not install pynvml directly, please report this to the maintainers of the package that installed pynvml for you. 2025-10-10T01:49:36.6517215Z import pynvml # type: ignore[import] 2025-10-10T01:49:38.8564541Z INFO 10-10 01:49:38 [__init__.py:215] Automatically detected platform cuda. 2025-10-10T01:49:42.8104102Z (APIServer pid=55660) INFO 10-10 01:49:42 [api_server.py:1839] vLLM API server version 0.11.0rc2.dev157+g0ad9951c4.d20251010 2025-10-10T01:49:42.8127694Z (APIServer pid=55660) INFO 10-10 01:49:42 [utils.py:233] non-default args: {'model_tag': 'meta-llama/Llama-3.2-1B-Instruct', 'port': 37661, 'model': 'meta-llama/Llama-3.2-1B-Instruct', 'dtype': 'float16', 'seed': 0, 'max_model_len': 2048, 'enforce_eager': True, 'load_format': 'dummy', 'distributed_executor_backend': 'mp', 'tensor_parallel_size': 2, 'max_num_seqs': 8, 'enable_chunked_prefill': True} 2025-10-10T01:49:42.9486891Z (APIServer pid=55660) INFO 10-10 01:49:42 [model.py:551] Resolved architecture: LlamaForCausalLM 2025-10-10T01:49:42.9488662Z (APIServer pid=55660) `torch_dtype` is deprecated! Use `dtype` instead! 2025-10-10T01:49:42.9490031Z (APIServer pid=55660) WARNING 10-10 01:49:42 [model.py:1768] Casting torch.bfloat16 to torch.float16. 2025-10-10T01:49:42.9491274Z (APIServer pid=55660) INFO 10-10 01:49:42 [model.py:1545] Using max model len 2048 2025-10-10T01:49:43.1057878Z (APIServer pid=55660) INFO 10-10 01:49:43 [scheduler.py:222] Chunked prefill is enabled with max_num_batched_tokens=2048. 2025-10-10T01:49:43.1062656Z (APIServer pid=55660) INFO 10-10 01:49:43 [vllm.py:345] Cudagraph is disabled under eager mode 2025-10-10T01:49:44.0923655Z /opt/conda/envs/py_3.12/lib/python3.12/site-packages/torch/cuda/__init__.py:64: FutureWarning: The pynvml package is deprecated. Please install nvidia-ml-py instead. If you did not install pynvml directly, please report this to the maintainers of the package that installed pynvml for you. 2025-10-10T01:49:44.0926133Z import pynvml # type: ignore[import] 2025-10-10T01:49:46.4054495Z INFO 10-10 01:49:46 [__init__.py:215] Automatically detected platform cuda. 2025-10-10T01:49:50.3798192Z (EngineCore_DP0 pid=55782) INFO 10-10 01:49:50 [core.py:648] Waiting for init message from front-end. 2025-10-10T01:49:50.4066499Z (EngineCore_DP0 pid=55782) INFO 10-10 01:49:50 [core.py:78] Initializing a V1 LLM engine (v0.11.0rc2.dev157+g0ad9951c4.d20251010) with config: model='meta-llama/Llama-3.2-1B-Instruct', speculative_config=None, tokenizer='meta-llama/Llama-3.2-1B-Instruct', skip_tokenizer_init=False, tokenizer_mode=auto, revision=None, tokenizer_revision=None, trust_remote_code=False, dtype=torch.float16, max_seq_len=2048, download_dir=None, load_format=dummy, tensor_parallel_size=2, pipeline_parallel_size=1, data_parallel_size=1, disable_custom_all_reduce=False, quantization=None, enforce_eager=True, kv_cache_dtype=auto, device_config=cuda, structured_outputs_config=StructuredOutputsConfig(backend='auto', disable_fallback=False, disable_any_whitespace=False, disable_additional_properties=False, reasoning_parser=''), observability_config=ObservabilityConfig(show_hidden_metrics_for_version=None, otlp_traces_endpoint=None, collect_detailed_traces=None), seed=0, served_model_name=meta-llama/Llama-3.2-1B-Instruct, enable_prefix_caching=True, chunked_prefill_enabled=True, pooler_config=None, compilation_config={'level': 0, 'debug_dump_path': None, 'cache_dir': '', 'backend': '', 'custom_ops': [], 'splitting_ops': None, 'use_inductor': True, 'compile_sizes': [], 'inductor_compile_config': {'enable_auto_functionalized_v2': False}, 'inductor_passes': {}, 'cudagraph_mode': , 'use_cudagraph': False, 'cudagraph_num_of_warmups': 0, 'cudagraph_capture_sizes': [], 'cudagraph_copy_inputs': False, 'full_cuda_graph': False, 'use_inductor_graph_partition': False, 'pass_config': {}, 'max_capture_size': 0, 'local_cache_dir': None} 2025-10-10T01:49:50.4071810Z (EngineCore_DP0 pid=55782) WARNING 10-10 01:49:50 [multiproc_executor.py:720] Reducing Torch parallelism from 24 threads to 1 to avoid unnecessary CPU contention. Set OMP_NUM_THREADS in the external environment to tune this value as needed. 2025-10-10T01:49:50.4076018Z (EngineCore_DP0 pid=55782) INFO 10-10 01:49:50 [shm_broadcast.py:289] vLLM message queue communication handle: Handle(local_reader_ranks=[0, 1], buffer_handle=(2, 16777216, 10, 'psm_7d75c2da'), local_subscribe_addr='ipc:///tmp/75051007-3337-49ae-a0ef-6d8dfbf2d8c3', remote_subscribe_addr=None, remote_addr_ipv6=False) 2025-10-10T01:49:50.8873270Z /opt/conda/envs/py_3.12/lib/python3.12/site-packages/torch/cuda/__init__.py:64: FutureWarning: The pynvml package is deprecated. Please install nvidia-ml-py instead. If you did not install pynvml directly, please report this to the maintainers of the package that installed pynvml for you. 2025-10-10T01:49:50.8874829Z import pynvml # type: ignore[import] 2025-10-10T01:49:50.9093567Z /opt/conda/envs/py_3.12/lib/python3.12/site-packages/torch/cuda/__init__.py:64: FutureWarning: The pynvml package is deprecated. Please install nvidia-ml-py instead. If you did not install pynvml directly, please report this to the maintainers of the package that installed pynvml for you. 2025-10-10T01:49:50.9094941Z import pynvml # type: ignore[import] 2025-10-10T01:49:53.2170510Z INFO 10-10 01:49:53 [__init__.py:215] Automatically detected platform cuda. 2025-10-10T01:49:53.2313420Z INFO 10-10 01:49:53 [__init__.py:215] Automatically detected platform cuda. 2025-10-10T01:49:57.8246401Z INFO 10-10 01:49:57 [shm_broadcast.py:289] vLLM message queue communication handle: Handle(local_reader_ranks=[0], buffer_handle=(1, 10485760, 10, 'psm_1491684b'), local_subscribe_addr='ipc:///tmp/39b79fc6-3c7a-4b04-a0ee-8966c46f5d2a', remote_subscribe_addr=None, remote_addr_ipv6=False) 2025-10-10T01:49:57.8641521Z INFO 10-10 01:49:57 [shm_broadcast.py:289] vLLM message queue communication handle: Handle(local_reader_ranks=[0], buffer_handle=(1, 10485760, 10, 'psm_a390e3ed'), local_subscribe_addr='ipc:///tmp/e1e5ee52-0956-4566-9779-7a1b24ef5ad8', remote_subscribe_addr=None, remote_addr_ipv6=False) 2025-10-10T01:49:58.6159219Z [Gloo] Rank 0 is connected to 1 peer ranks. Expected number of connected peer ranks is : 1 2025-10-10T01:49:58.6159902Z [Gloo] Rank 1 is connected to 1 peer ranks. Expected number of connected peer ranks is : 1 2025-10-10T01:49:58.6207167Z [Gloo] Rank 1 is connected to 1 peer ranks. Expected number of connected peer ranks is : 1 2025-10-10T01:49:58.6207883Z [Gloo] Rank 0 is connected to 1 peer ranks. Expected number of connected peer ranks is : 1 2025-10-10T01:49:58.6254486Z INFO 10-10 01:49:58 [__init__.py:1384] Found nccl from library libnccl.so.2 2025-10-10T01:49:58.6255000Z INFO 10-10 01:49:58 [pynccl.py:104] vLLM is using nccl==2.27.5 2025-10-10T01:49:58.6258007Z INFO 10-10 01:49:58 [__init__.py:1384] Found nccl from library libnccl.so.2 2025-10-10T01:49:58.6259343Z INFO 10-10 01:49:58 [pynccl.py:104] vLLM is using nccl==2.27.5 2025-10-10T01:49:58.7998119Z WARNING 10-10 01:49:58 [symm_mem.py:58] SymmMemCommunicator: Device capability 8.9 not supported, communicator is not available. 2025-10-10T01:49:58.7999109Z WARNING 10-10 01:49:58 [symm_mem.py:58] SymmMemCommunicator: Device capability 8.9 not supported, communicator is not available. 2025-10-10T01:49:58.8019397Z INFO 10-10 01:49:58 [custom_all_reduce.py:35] Skipping P2P check and trusting the driver's P2P report. 2025-10-10T01:49:58.8020047Z INFO 10-10 01:49:58 [custom_all_reduce.py:35] Skipping P2P check and trusting the driver's P2P report. 2025-10-10T01:49:58.8021080Z WARNING 10-10 01:49:58 [custom_all_reduce.py:154] Custom allreduce is disabled because your platform lacks GPU P2P capability or P2P test failed. To silence this warning, specify disable_custom_all_reduce=True explicitly. 2025-10-10T01:49:58.8022482Z WARNING 10-10 01:49:58 [custom_all_reduce.py:154] Custom allreduce is disabled because your platform lacks GPU P2P capability or P2P test failed. To silence this warning, specify disable_custom_all_reduce=True explicitly. 2025-10-10T01:49:58.9040990Z INFO 10-10 01:49:58 [shm_broadcast.py:289] vLLM message queue communication handle: Handle(local_reader_ranks=[1], buffer_handle=(1, 4194304, 6, 'psm_12f26273'), local_subscribe_addr='ipc:///tmp/c9d4096d-4c2c-4946-84a2-734569eec309', remote_subscribe_addr=None, remote_addr_ipv6=False) 2025-10-10T01:49:58.9076691Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-10-10T01:49:58.9077310Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-10-10T01:49:58.9088305Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-10-10T01:49:58.9096178Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-10-10T01:49:59.0135488Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-10-10T01:49:59.0148165Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-10-10T01:49:59.0157363Z [Gloo] Rank 1 is connected to 1 peer ranks. Expected number of connected peer ranks is : 1 2025-10-10T01:49:59.0157946Z [Gloo] Rank 0 is connected to 1 peer ranks. Expected number of connected peer ranks is : 1 2025-10-10T01:49:59.0159811Z INFO 10-10 01:49:59 [__init__.py:1384] Found nccl from library libnccl.so.2 2025-10-10T01:49:59.0160251Z INFO 10-10 01:49:59 [pynccl.py:104] vLLM is using nccl==2.27.5 2025-10-10T01:49:59.0160831Z INFO 10-10 01:49:59 [__init__.py:1384] Found nccl from library libnccl.so.2 2025-10-10T01:49:59.0161274Z INFO 10-10 01:49:59 [pynccl.py:104] vLLM is using nccl==2.27.5 2025-10-10T01:49:59.0500590Z INFO 10-10 01:49:59 [parallel_state.py:1208] rank 1 in world size 2 is assigned as DP rank 0, PP rank 0, TP rank 1, EP rank 1 2025-10-10T01:49:59.0501354Z INFO 10-10 01:49:59 [parallel_state.py:1208] rank 0 in world size 2 is assigned as DP rank 0, PP rank 0, TP rank 0, EP rank 0 2025-10-10T01:49:59.2985377Z INFO 10-10 01:49:59 [topk_topp_sampler.py:55] Using FlashInfer for top-p & top-k sampling. 2025-10-10T01:49:59.3028380Z INFO 10-10 01:49:59 [topk_topp_sampler.py:55] Using FlashInfer for top-p & top-k sampling. 2025-10-10T01:49:59.3299929Z (Worker_TP1 pid=55902) INFO 10-10 01:49:59 [gpu_model_runner.py:2707] Starting to load model meta-llama/Llama-3.2-1B-Instruct... 2025-10-10T01:49:59.3346109Z (Worker_TP0 pid=55901) INFO 10-10 01:49:59 [gpu_model_runner.py:2707] Starting to load model meta-llama/Llama-3.2-1B-Instruct... 2025-10-10T01:49:59.5502560Z (Worker_TP1 pid=55902) INFO 10-10 01:49:59 [gpu_model_runner.py:2739] Loading model from scratch... 2025-10-10T01:49:59.5564921Z (Worker_TP0 pid=55901) INFO 10-10 01:49:59 [gpu_model_runner.py:2739] Loading model from scratch... 2025-10-10T01:49:59.6391918Z (Worker_TP1 pid=55902) INFO 10-10 01:49:59 [cuda.py:361] Using Flash Attention backend on V1 engine. 2025-10-10T01:49:59.6466078Z (Worker_TP0 pid=55901) INFO 10-10 01:49:59 [cuda.py:361] Using Flash Attention backend on V1 engine. 2025-10-10T01:50:00.1237938Z (Worker_TP1 pid=55902) INFO 10-10 01:50:00 [gpu_model_runner.py:2758] Model loading took 1.1667 GiB and 0.118589 seconds 2025-10-10T01:50:00.1363707Z (Worker_TP0 pid=55901) INFO 10-10 01:50:00 [gpu_model_runner.py:2758] Model loading took 1.1667 GiB and 0.120107 seconds 2025-10-10T01:50:05.0487231Z (Worker_TP1 pid=55902) INFO 10-10 01:50:05 [gpu_worker.py:298] Available KV cache memory: 18.53 GiB 2025-10-10T01:50:05.0730865Z (Worker_TP0 pid=55901) INFO 10-10 01:50:05 [gpu_worker.py:298] Available KV cache memory: 18.53 GiB 2025-10-10T01:50:05.3113481Z (EngineCore_DP0 pid=55782) INFO 10-10 01:50:05 [kv_cache_utils.py:1087] GPU KV cache size: 1,214,464 tokens 2025-10-10T01:50:05.3114443Z (EngineCore_DP0 pid=55782) INFO 10-10 01:50:05 [kv_cache_utils.py:1091] Maximum concurrency for 2,048 tokens per request: 593.00x 2025-10-10T01:50:05.3115372Z (EngineCore_DP0 pid=55782) INFO 10-10 01:50:05 [kv_cache_utils.py:1087] GPU KV cache size: 1,214,464 tokens 2025-10-10T01:50:05.3116227Z (EngineCore_DP0 pid=55782) INFO 10-10 01:50:05 [kv_cache_utils.py:1091] Maximum concurrency for 2,048 tokens per request: 593.00x 2025-10-10T01:50:05.4361570Z (EngineCore_DP0 pid=55782) INFO 10-10 01:50:05 [core.py:211] init engine (profile, create kv cache, warmup model) took 5.23 seconds 2025-10-10T01:50:05.9366243Z (EngineCore_DP0 pid=55782) INFO 10-10 01:50:05 [vllm.py:345] Cudagraph is disabled under eager mode 2025-10-10T01:50:06.1216789Z (APIServer pid=55660) INFO 10-10 01:50:06 [loggers.py:147] Engine 000: vllm cache_config_info with initialization after num_gpu_blocks is: 75904 2025-10-10T01:50:06.1563214Z (EngineCore_DP0 pid=55782) INFO 10-10 01:50:06 [gc_utils.py:41] GC Debug Config. enabled:False,top_objects:-1 2025-10-10T01:50:06.1780674Z (APIServer pid=55660) INFO 10-10 01:50:06 [api_server.py:1634] Supported_tasks: ['generate'] 2025-10-10T01:50:06.2513785Z (APIServer pid=55660) WARNING 10-10 01:50:06 [model.py:1424] Default sampling parameters have been overridden by the model's Hugging Face generation config recommended from the model creator. If this is not intended, please relaunch vLLM instance with `--generation-config vllm`. 2025-10-10T01:50:06.2515434Z (APIServer pid=55660) INFO 10-10 01:50:06 [serving_responses.py:140] Using default chat sampling params from model: {'temperature': 0.6, 'top_p': 0.9} 2025-10-10T01:50:06.2780269Z (APIServer pid=55660) INFO 10-10 01:50:06 [serving_chat.py:139] Using default chat sampling params from model: {'temperature': 0.6, 'top_p': 0.9} 2025-10-10T01:50:06.2981161Z (APIServer pid=55660) INFO 10-10 01:50:06 [serving_completion.py:76] Using default completion sampling params from model: {'temperature': 0.6, 'top_p': 0.9} 2025-10-10T01:50:06.2982202Z (APIServer pid=55660) INFO 10-10 01:50:06 [api_server.py:1912] Starting vLLM API server 0 on http://0.0.0.0:37661 2025-10-10T01:50:06.2982965Z (APIServer pid=55660) INFO 10-10 01:50:06 [launcher.py:34] Available routes are: 2025-10-10T01:50:06.2983640Z (APIServer pid=55660) INFO 10-10 01:50:06 [launcher.py:42] Route: /openapi.json, Methods: GET, HEAD 2025-10-10T01:50:06.2984352Z (APIServer pid=55660) INFO 10-10 01:50:06 [launcher.py:42] Route: /docs, Methods: GET, HEAD 2025-10-10T01:50:06.2985312Z (APIServer pid=55660) INFO 10-10 01:50:06 [launcher.py:42] Route: /docs/oauth2-redirect, Methods: GET, HEAD 2025-10-10T01:50:06.2986089Z (APIServer pid=55660) INFO 10-10 01:50:06 [launcher.py:42] Route: /redoc, Methods: GET, HEAD 2025-10-10T01:50:06.2986761Z (APIServer pid=55660) INFO 10-10 01:50:06 [launcher.py:42] Route: /health, Methods: GET 2025-10-10T01:50:06.2987425Z (APIServer pid=55660) INFO 10-10 01:50:06 [launcher.py:42] Route: /load, Methods: GET 2025-10-10T01:50:06.2988214Z (APIServer pid=55660) INFO 10-10 01:50:06 [launcher.py:42] Route: /ping, Methods: POST 2025-10-10T01:50:06.2988953Z (APIServer pid=55660) INFO 10-10 01:50:06 [launcher.py:42] Route: /ping, Methods: GET 2025-10-10T01:50:06.2989699Z (APIServer pid=55660) INFO 10-10 01:50:06 [launcher.py:42] Route: /tokenize, Methods: POST 2025-10-10T01:50:06.2990296Z (APIServer pid=55660) INFO 10-10 01:50:06 [launcher.py:42] Route: /detokenize, Methods: POST 2025-10-10T01:50:06.2990829Z (APIServer pid=55660) INFO 10-10 01:50:06 [launcher.py:42] Route: /v1/models, Methods: GET 2025-10-10T01:50:06.2991370Z (APIServer pid=55660) INFO 10-10 01:50:06 [launcher.py:42] Route: /version, Methods: GET 2025-10-10T01:50:06.2991910Z (APIServer pid=55660) INFO 10-10 01:50:06 [launcher.py:42] Route: /v1/responses, Methods: POST 2025-10-10T01:50:06.2992485Z (APIServer pid=55660) INFO 10-10 01:50:06 [launcher.py:42] Route: /v1/responses/{response_id}, Methods: GET 2025-10-10T01:50:06.2993115Z (APIServer pid=55660) INFO 10-10 01:50:06 [launcher.py:42] Route: /v1/responses/{response_id}/cancel, Methods: POST 2025-10-10T01:50:06.2993752Z (APIServer pid=55660) INFO 10-10 01:50:06 [launcher.py:42] Route: /v1/chat/completions, Methods: POST 2025-10-10T01:50:06.2994322Z (APIServer pid=55660) INFO 10-10 01:50:06 [launcher.py:42] Route: /v1/completions, Methods: POST 2025-10-10T01:50:06.2994890Z (APIServer pid=55660) INFO 10-10 01:50:06 [launcher.py:42] Route: /v1/embeddings, Methods: POST 2025-10-10T01:50:06.2995443Z (APIServer pid=55660) INFO 10-10 01:50:06 [launcher.py:42] Route: /pooling, Methods: POST 2025-10-10T01:50:06.2995960Z (APIServer pid=55660) INFO 10-10 01:50:06 [launcher.py:42] Route: /classify, Methods: POST 2025-10-10T01:50:06.2996507Z (APIServer pid=55660) INFO 10-10 01:50:06 [launcher.py:42] Route: /score, Methods: POST 2025-10-10T01:50:06.2997035Z (APIServer pid=55660) INFO 10-10 01:50:06 [launcher.py:42] Route: /v1/score, Methods: POST 2025-10-10T01:50:06.2997627Z (APIServer pid=55660) INFO 10-10 01:50:06 [launcher.py:42] Route: /v1/audio/transcriptions, Methods: POST 2025-10-10T01:50:06.2998235Z (APIServer pid=55660) INFO 10-10 01:50:06 [launcher.py:42] Route: /v1/audio/translations, Methods: POST 2025-10-10T01:50:06.2998887Z (APIServer pid=55660) INFO 10-10 01:50:06 [launcher.py:42] Route: /rerank, Methods: POST 2025-10-10T01:50:06.2999433Z (APIServer pid=55660) INFO 10-10 01:50:06 [launcher.py:42] Route: /v1/rerank, Methods: POST 2025-10-10T01:50:06.3000054Z (APIServer pid=55660) INFO 10-10 01:50:06 [launcher.py:42] Route: /v2/rerank, Methods: POST 2025-10-10T01:50:06.3000601Z (APIServer pid=55660) INFO 10-10 01:50:06 [launcher.py:42] Route: /scale_elastic_ep, Methods: POST 2025-10-10T01:50:06.3001183Z (APIServer pid=55660) INFO 10-10 01:50:06 [launcher.py:42] Route: /is_scaling_elastic_ep, Methods: POST 2025-10-10T01:50:06.3001734Z (APIServer pid=55660) INFO 10-10 01:50:06 [launcher.py:42] Route: /invocations, Methods: POST 2025-10-10T01:50:06.3002266Z (APIServer pid=55660) INFO 10-10 01:50:06 [launcher.py:42] Route: /metrics, Methods: GET 2025-10-10T01:50:06.3139540Z (APIServer pid=55660) INFO: Started server process [55660] 2025-10-10T01:50:06.3140364Z (APIServer pid=55660) INFO: Waiting for application startup. 2025-10-10T01:50:06.5472018Z (APIServer pid=55660) INFO: Application startup complete. 2025-10-10T01:50:06.7163450Z (APIServer pid=55660) INFO: 127.0.0.1:50460 - "GET /health HTTP/1.1" 200 OK 2025-10-10T01:50:06.7256212Z (APIServer pid=55660) INFO: 127.0.0.1:50474 - "GET /v1/models HTTP/1.1" 200 OK 2025-10-10T01:50:06.9019341Z (APIServer pid=55660) INFO: 127.0.0.1:50474 - "POST /v1/completions HTTP/1.1" 200 OK 2025-10-10T01:50:06.9588021Z (APIServer pid=55660) INFO: 127.0.0.1:50474 - "POST /v1/completions HTTP/1.1" 200 OK 2025-10-10T01:50:07.0489059Z (APIServer pid=55660) INFO: 127.0.0.1:50474 - "POST /v1/completions HTTP/1.1" 200 OK 2025-10-10T01:50:07.1346853Z (APIServer pid=55660) INFO: 127.0.0.1:50474 - "POST /v1/completions HTTP/1.1" 200 OK 2025-10-10T01:50:07.2066424Z (APIServer pid=55660) INFO: 127.0.0.1:50474 - "POST /v1/completions HTTP/1.1" 200 OK 2025-10-10T01:50:07.2142914Z (APIServer pid=55660) INFO: 127.0.0.1:50474 - "POST /v1/completions HTTP/1.1" 200 OK 2025-10-10T01:50:07.2918360Z (APIServer pid=55660) INFO 10-10 01:50:07 [launcher.py:99] Shutting down FastAPI HTTP server. 2025-10-10T01:50:07.2922023Z (Worker_TP0 pid=55901) INFO 10-10 01:50:07 [multiproc_executor.py:558] Parent process exited, terminating worker 2025-10-10T01:50:07.2922933Z (Worker_TP0 pid=55901) INFO 10-10 01:50:07 [multiproc_executor.py:599] WorkerProc shutting down. 2025-10-10T01:50:07.2926137Z (Worker_TP1 pid=55902) INFO 10-10 01:50:07 [multiproc_executor.py:558] Parent process exited, terminating worker 2025-10-10T01:50:07.2927391Z (Worker_TP1 pid=55902) INFO 10-10 01:50:07 [multiproc_executor.py:599] WorkerProc shutting down. 2025-10-10T01:50:09.4108405Z (APIServer pid=55660) INFO: Shutting down 2025-10-10T01:50:09.5114151Z (APIServer pid=55660) INFO: Waiting for application shutdown. 2025-10-10T01:50:09.5116793Z (APIServer pid=55660) INFO: Application shutdown complete. 2025-10-10T01:50:10.1428194Z PASSED 2025-10-10T01:50:10.1578825Z distributed/test_sequence_parallel.py::test_tp_sp_generation[meta-llama/Llama-3.2-1B-Instruct-parallel_setup15-ray-1-auto-test_options15] Fork a new process to run a test 56419 2025-10-10T01:50:10.1598064Z Fork a new process to run a test 0 2025-10-10T01:50:10.8491009Z INFO 10-10 01:50:10 [model.py:551] Resolved architecture: LlamaForCausalLM 2025-10-10T01:50:10.8491598Z `torch_dtype` is deprecated! Use `dtype` instead! 2025-10-10T01:50:10.8492036Z WARNING 10-10 01:50:10 [model.py:1768] Casting torch.bfloat16 to torch.float16. 2025-10-10T01:50:10.8492540Z INFO 10-10 01:50:10 [model.py:1545] Using max model len 2048 2025-10-10T01:50:10.8495084Z Launching RemoteOpenAIServer with: vllm serve meta-llama/Llama-3.2-1B-Instruct --dtype float16 --max-model-len 2048 --max-num-seqs 8 --enable-chunked-prefill --enforce-eager --tokenizer-mode auto --tensor-parallel-size 2 --pipeline-parallel-size 2 --distributed-executor-backend ray --compilation_config {"level": 3, "custom_ops": ["+rms_norm"], "compile_sizes": [4, 8], "pass_config": {"enable_sequence_parallelism": true, "enable_fusion": false, "enable_noop": true}} --load-format dummy --port 58845 --seed 0 2025-10-10T01:50:11.3592183Z /opt/conda/envs/py_3.12/lib/python3.12/site-packages/torch/cuda/__init__.py:64: FutureWarning: The pynvml package is deprecated. Please install nvidia-ml-py instead. If you did not install pynvml directly, please report this to the maintainers of the package that installed pynvml for you. 2025-10-10T01:50:11.3594516Z import pynvml # type: ignore[import] 2025-10-10T01:50:13.5658291Z INFO 10-10 01:50:13 [__init__.py:215] Automatically detected platform cuda. 2025-10-10T01:50:17.4790648Z (APIServer pid=56468) INFO 10-10 01:50:17 [api_server.py:1839] vLLM API server version 0.11.0rc2.dev157+g0ad9951c4.d20251010 2025-10-10T01:50:17.4815315Z (APIServer pid=56468) INFO 10-10 01:50:17 [utils.py:233] non-default args: {'model_tag': 'meta-llama/Llama-3.2-1B-Instruct', 'port': 58845, 'model': 'meta-llama/Llama-3.2-1B-Instruct', 'dtype': 'float16', 'seed': 0, 'max_model_len': 2048, 'enforce_eager': True, 'load_format': 'dummy', 'distributed_executor_backend': 'ray', 'pipeline_parallel_size': 2, 'tensor_parallel_size': 2, 'max_num_seqs': 8, 'enable_chunked_prefill': True, 'compilation_config': {'level': 3, 'debug_dump_path': None, 'cache_dir': '', 'backend': '', 'custom_ops': ['+rms_norm'], 'splitting_ops': None, 'use_inductor': True, 'compile_sizes': [4, 8], 'inductor_compile_config': {'enable_auto_functionalized_v2': False}, 'inductor_passes': {}, 'cudagraph_mode': None, 'use_cudagraph': True, 'cudagraph_num_of_warmups': 0, 'cudagraph_capture_sizes': None, 'cudagraph_copy_inputs': False, 'full_cuda_graph': False, 'use_inductor_graph_partition': False, 'pass_config': {'enable_noop': True, 'enable_sequence_parallelism': True}, 'max_capture_size': None, 'local_cache_dir': None}} 2025-10-10T01:50:17.6434871Z (APIServer pid=56468) INFO 10-10 01:50:17 [model.py:551] Resolved architecture: LlamaForCausalLM 2025-10-10T01:50:17.6435599Z (APIServer pid=56468) `torch_dtype` is deprecated! Use `dtype` instead! 2025-10-10T01:50:17.6436297Z (APIServer pid=56468) WARNING 10-10 01:50:17 [model.py:1768] Casting torch.bfloat16 to torch.float16. 2025-10-10T01:50:17.6437478Z (APIServer pid=56468) INFO 10-10 01:50:17 [model.py:1545] Using max model len 2048 2025-10-10T01:50:17.8028583Z (APIServer pid=56468) INFO 10-10 01:50:17 [scheduler.py:222] Chunked prefill is enabled with max_num_batched_tokens=2048. 2025-10-10T01:50:17.8034059Z (APIServer pid=56468) INFO 10-10 01:50:17 [vllm.py:345] Cudagraph is disabled under eager mode 2025-10-10T01:50:18.7624757Z /opt/conda/envs/py_3.12/lib/python3.12/site-packages/torch/cuda/__init__.py:64: FutureWarning: The pynvml package is deprecated. Please install nvidia-ml-py instead. If you did not install pynvml directly, please report this to the maintainers of the package that installed pynvml for you. 2025-10-10T01:50:18.7626032Z import pynvml # type: ignore[import] 2025-10-10T01:50:21.0577724Z INFO 10-10 01:50:21 [__init__.py:215] Automatically detected platform cuda. 2025-10-10T01:50:25.1504131Z (EngineCore_DP0 pid=56590) INFO 10-10 01:50:25 [core.py:648] Waiting for init message from front-end. 2025-10-10T01:50:25.1763726Z (EngineCore_DP0 pid=56590) INFO 10-10 01:50:25 [core.py:78] Initializing a V1 LLM engine (v0.11.0rc2.dev157+g0ad9951c4.d20251010) with config: model='meta-llama/Llama-3.2-1B-Instruct', speculative_config=None, tokenizer='meta-llama/Llama-3.2-1B-Instruct', skip_tokenizer_init=False, tokenizer_mode=auto, revision=None, tokenizer_revision=None, trust_remote_code=False, dtype=torch.float16, max_seq_len=2048, download_dir=None, load_format=dummy, tensor_parallel_size=2, pipeline_parallel_size=2, data_parallel_size=1, disable_custom_all_reduce=False, quantization=None, enforce_eager=True, kv_cache_dtype=auto, device_config=cuda, structured_outputs_config=StructuredOutputsConfig(backend='auto', disable_fallback=False, disable_any_whitespace=False, disable_additional_properties=False, reasoning_parser=''), observability_config=ObservabilityConfig(show_hidden_metrics_for_version=None, otlp_traces_endpoint=None, collect_detailed_traces=None), seed=0, served_model_name=meta-llama/Llama-3.2-1B-Instruct, enable_prefix_caching=True, chunked_prefill_enabled=True, pooler_config=None, compilation_config={'level': 3, 'debug_dump_path': None, 'cache_dir': '', 'backend': '', 'custom_ops': ['+rms_norm', '+rms_norm'], 'splitting_ops': ['vllm.unified_attention', 'vllm.unified_attention_with_output', 'vllm.mamba_mixer2', 'vllm.mamba_mixer', 'vllm.short_conv', 'vllm.linear_attention', 'vllm.plamo2_mamba_mixer', 'vllm.gdn_attention', 'vllm.sparse_attn_indexer'], 'use_inductor': True, 'compile_sizes': [8, 4], 'inductor_compile_config': {'enable_auto_functionalized_v2': False}, 'inductor_passes': {}, 'cudagraph_mode': , 'use_cudagraph': False, 'cudagraph_num_of_warmups': 0, 'cudagraph_capture_sizes': [], 'cudagraph_copy_inputs': False, 'full_cuda_graph': False, 'use_inductor_graph_partition': False, 'pass_config': {'enable_noop': True, 'enable_sequence_parallelism': True}, 'max_capture_size': 0, 'local_cache_dir': None} 2025-10-10T01:50:27.1002827Z (EngineCore_DP0 pid=56590) 2025-10-10 01:50:27,099 INFO worker.py:1918 -- Started a local Ray instance. View the dashboard at http://127.0.0.1:8265  2025-10-10T01:50:27.9726513Z (EngineCore_DP0 pid=56590) INFO 10-10 01:50:27 [ray_utils.py:345] No current placement group found. Creating a new placement group. 2025-10-10T01:50:28.1434785Z (EngineCore_DP0 pid=56590) INFO 10-10 01:50:28 [ray_distributed_executor.py:171] use_ray_spmd_worker: True 2025-10-10T01:50:28.7490911Z (EngineCore_DP0 pid=56590) (pid=57475) /opt/conda/envs/py_3.12/lib/python3.12/site-packages/torch/cuda/__init__.py:64: FutureWarning: The pynvml package is deprecated. Please install nvidia-ml-py instead. If you did not install pynvml directly, please report this to the maintainers of the package that installed pynvml for you. 2025-10-10T01:50:28.7492718Z (EngineCore_DP0 pid=56590) (pid=57475) import pynvml # type: ignore[import] 2025-10-10T01:50:30.9965868Z (EngineCore_DP0 pid=56590) (pid=57475) INFO 10-10 01:50:30 [__init__.py:215] Automatically detected platform cuda. 2025-10-10T01:50:32.3969742Z (EngineCore_DP0 pid=56590) INFO 10-10 01:50:32 [ray_env.py:63] RAY_NON_CARRY_OVER_ENV_VARS from config: set() 2025-10-10T01:50:32.3971233Z (EngineCore_DP0 pid=56590) INFO 10-10 01:50:32 [ray_env.py:65] Copying the following environment variables to workers: ['HF_TOKEN', 'LD_LIBRARY_PATH', 'VLLM_USE_V1', 'VLLM_USE_RAY_SPMD_WORKER', 'MAX_JOBS', 'VLLM_WORKER_MULTIPROC_METHOD', 'VLLM_USE_RAY_COMPILED_DAG', 'HUGGING_FACE_HUB_TOKEN'] 2025-10-10T01:50:32.3972546Z (EngineCore_DP0 pid=56590) INFO 10-10 01:50:32 [ray_env.py:68] If certain env vars should NOT be copied, add them to /var/lib/jenkins/.config/vllm/ray_non_carry_over_env_vars.json file 2025-10-10T01:50:36.3031963Z (EngineCore_DP0 pid=56590) (RayWorkerWrapper pid=57475) [Gloo] Rank 1 is connected to 3 peer ranks. Expected number of connected peer ranks is : 3 2025-10-10T01:50:36.3033241Z (EngineCore_DP0 pid=56590) (RayWorkerWrapper pid=57475) [Gloo] Rank 1 is connected to 1 peer ranks. Expected number of connected peer ranks is : 1 2025-10-10T01:50:36.3034353Z (EngineCore_DP0 pid=56590) (RayWorkerWrapper pid=57475) INFO 10-10 01:50:36 [__init__.py:1384] Found nccl from library libnccl.so.2 2025-10-10T01:50:36.3035827Z (EngineCore_DP0 pid=56590) (RayWorkerWrapper pid=57475) INFO 10-10 01:50:36 [pynccl.py:104] vLLM is using nccl==2.27.5 2025-10-10T01:50:36.3037789Z (EngineCore_DP0 pid=56590) (pid=57479) INFO 10-10 01:50:30 [__init__.py:215] Automatically detected platform cuda. [repeated 3x across cluster] (Ray deduplicates logs by default. Set RAY_DEDUP_LOGS=0 to disable log deduplication, or see https://docs.ray.io/en/master/ray-observability/user-guides/configure-logging.html#log-deduplication for more options.) 2025-10-10T01:50:36.6133353Z (EngineCore_DP0 pid=56590) (RayWorkerWrapper pid=57475) WARNING 10-10 01:50:36 [symm_mem.py:58] SymmMemCommunicator: Device capability 8.9 not supported, communicator is not available. 2025-10-10T01:50:36.6135073Z (EngineCore_DP0 pid=56590) (RayWorkerWrapper pid=57475) INFO 10-10 01:50:36 [custom_all_reduce.py:35] Skipping P2P check and trusting the driver's P2P report. 2025-10-10T01:50:36.6136809Z (EngineCore_DP0 pid=56590) (RayWorkerWrapper pid=57475) WARNING 10-10 01:50:36 [custom_all_reduce.py:154] Custom allreduce is disabled because your platform lacks GPU P2P capability or P2P test failed. To silence this warning, specify disable_custom_all_reduce=True explicitly. 2025-10-10T01:50:36.6139297Z (EngineCore_DP0 pid=56590) (RayWorkerWrapper pid=57477) INFO 10-10 01:50:36 [shm_broadcast.py:289] vLLM message queue communication handle: Handle(local_reader_ranks=[1], buffer_handle=(1, 4194304, 6, 'psm_6349c77d'), local_subscribe_addr='ipc:///tmp/5fec1fda-e938-4137-b512-95ffd0ddc821', remote_subscribe_addr=None, remote_addr_ipv6=False) 2025-10-10T01:50:36.7192720Z (EngineCore_DP0 pid=56590) (RayWorkerWrapper pid=57475) INFO 10-10 01:50:36 [parallel_state.py:1208] rank 1 in world size 4 is assigned as DP rank 0, PP rank 0, TP rank 1, EP rank 1 2025-10-10T01:50:37.0282868Z (EngineCore_DP0 pid=56590) (RayWorkerWrapper pid=57475) INFO 10-10 01:50:36 [topk_topp_sampler.py:55] Using FlashInfer for top-p & top-k sampling. 2025-10-10T01:50:37.0284417Z (EngineCore_DP0 pid=56590) (RayWorkerWrapper pid=57475) INFO 10-10 01:50:37 [gpu_model_runner.py:2707] Starting to load model meta-llama/Llama-3.2-1B-Instruct... 2025-10-10T01:50:37.3375123Z (EngineCore_DP0 pid=56590) (RayWorkerWrapper pid=57475) INFO 10-10 01:50:37 [gpu_model_runner.py:2739] Loading model from scratch... 2025-10-10T01:50:37.4430311Z (EngineCore_DP0 pid=56590) (RayWorkerWrapper pid=57475) INFO 10-10 01:50:37 [cuda.py:361] Using Flash Attention backend on V1 engine. 2025-10-10T01:50:37.9559867Z (EngineCore_DP0 pid=56590) (RayWorkerWrapper pid=57472) INFO 10-10 01:50:37 [gpu_model_runner.py:2758] Model loading took 0.7135 GiB and 0.155632 seconds 2025-10-10T01:50:40.2013879Z (EngineCore_DP0 pid=56590) (RayWorkerWrapper pid=57472) INFO 10-10 01:50:40 [backends.py:548] Using cache directory: /var/lib/jenkins/.cache/vllm/torch_compile_cache/8cedb68d99/rank_2_0/backbone for vLLM's torch.compile 2025-10-10T01:50:40.2015393Z (EngineCore_DP0 pid=56590) (RayWorkerWrapper pid=57472) INFO 10-10 01:50:40 [backends.py:559] Dynamo bytecode transform time: 1.95 s 2025-10-10T01:50:40.9223506Z (EngineCore_DP0 pid=56590) (RayWorkerWrapper pid=57472) INFO 10-10 01:50:40 [backends.py:164] Directly load the compiled graph(s) for dynamic shape from the cache, took 0.456 s 2025-10-10T01:50:42.0493430Z (EngineCore_DP0 pid=56590) (RayWorkerWrapper pid=57475) INFO 10-10 01:50:42 [gpu_worker.py:298] Available KV cache memory: 19.02 GiB 2025-10-10T01:50:42.0494821Z (EngineCore_DP0 pid=56590) (RayWorkerWrapper pid=57479) [Gloo] Rank 1 is connected to 1 peer ranks. Expected number of connected peer ranks is : 1 [repeated 22x across cluster] 2025-10-10T01:50:42.0496483Z (EngineCore_DP0 pid=56590) (RayWorkerWrapper pid=57479) INFO 10-10 01:50:36 [__init__.py:1384] Found nccl from library libnccl.so.2 [repeated 11x across cluster] 2025-10-10T01:50:42.0497675Z (EngineCore_DP0 pid=56590) (RayWorkerWrapper pid=57479) INFO 10-10 01:50:36 [pynccl.py:104] vLLM is using nccl==2.27.5 [repeated 11x across cluster] 2025-10-10T01:50:42.0499108Z (EngineCore_DP0 pid=56590) (RayWorkerWrapper pid=57479) WARNING 10-10 01:50:36 [symm_mem.py:58] SymmMemCommunicator: Device capability 8.9 not supported, communicator is not available. [repeated 3x across cluster] 2025-10-10T01:50:42.0500887Z (EngineCore_DP0 pid=56590) (RayWorkerWrapper pid=57479) INFO 10-10 01:50:36 [custom_all_reduce.py:35] Skipping P2P check and trusting the driver's P2P report. [repeated 3x across cluster] 2025-10-10T01:50:42.0502450Z (EngineCore_DP0 pid=56590) (RayWorkerWrapper pid=57479) WARNING 10-10 01:50:36 [custom_all_reduce.py:154] Custom allreduce is disabled because your platform lacks GPU P2P capability or P2P test failed. To silence this warning, specify disable_custom_all_reduce=True explicitly. [repeated 3x across cluster] 2025-10-10T01:50:42.0504522Z (EngineCore_DP0 pid=56590) (RayWorkerWrapper pid=57472) INFO 10-10 01:50:36 [shm_broadcast.py:289] vLLM message queue communication handle: Handle(local_reader_ranks=[1], buffer_handle=(1, 4194304, 6, 'psm_efbc1a64'), local_subscribe_addr='ipc:///tmp/ba98ab07-f416-4548-a264-365e02b191c0', remote_subscribe_addr=None, remote_addr_ipv6=False) 2025-10-10T01:50:42.0506100Z (EngineCore_DP0 pid=56590) (RayWorkerWrapper pid=57479) INFO 10-10 01:50:36 [parallel_state.py:1208] rank 3 in world size 4 is assigned as DP rank 0, PP rank 1, TP rank 1, EP rank 1 [repeated 3x across cluster] 2025-10-10T01:50:42.0507222Z (EngineCore_DP0 pid=56590) (RayWorkerWrapper pid=57479) INFO 10-10 01:50:36 [topk_topp_sampler.py:55] Using FlashInfer for top-p & top-k sampling. [repeated 3x across cluster] 2025-10-10T01:50:42.0508354Z (EngineCore_DP0 pid=56590) (RayWorkerWrapper pid=57479) INFO 10-10 01:50:37 [gpu_model_runner.py:2707] Starting to load model meta-llama/Llama-3.2-1B-Instruct... [repeated 3x across cluster] 2025-10-10T01:50:42.3125981Z (EngineCore_DP0 pid=56590) INFO 10-10 01:50:42 [kv_cache_utils.py:1087] GPU KV cache size: 2,492,416 tokens 2025-10-10T01:50:42.3126970Z (EngineCore_DP0 pid=56590) INFO 10-10 01:50:42 [kv_cache_utils.py:1091] Maximum concurrency for 2,048 tokens per request: 1217.00x 2025-10-10T01:50:42.3127837Z (EngineCore_DP0 pid=56590) INFO 10-10 01:50:42 [kv_cache_utils.py:1087] GPU KV cache size: 2,492,416 tokens 2025-10-10T01:50:42.3128721Z (EngineCore_DP0 pid=56590) INFO 10-10 01:50:42 [kv_cache_utils.py:1091] Maximum concurrency for 2,048 tokens per request: 1217.00x 2025-10-10T01:50:42.3129594Z (EngineCore_DP0 pid=56590) INFO 10-10 01:50:42 [kv_cache_utils.py:1087] GPU KV cache size: 2,489,088 tokens 2025-10-10T01:50:42.3130411Z (EngineCore_DP0 pid=56590) INFO 10-10 01:50:42 [kv_cache_utils.py:1091] Maximum concurrency for 2,048 tokens per request: 1215.38x 2025-10-10T01:50:42.3131245Z (EngineCore_DP0 pid=56590) INFO 10-10 01:50:42 [kv_cache_utils.py:1087] GPU KV cache size: 2,489,088 tokens 2025-10-10T01:50:42.3131908Z (EngineCore_DP0 pid=56590) INFO 10-10 01:50:42 [kv_cache_utils.py:1091] Maximum concurrency for 2,048 tokens per request: 1215.38x 2025-10-10T01:50:42.3580866Z (EngineCore_DP0 pid=56590) (RayWorkerWrapper pid=57475) INFO 10-10 01:50:42 [gpu_worker.py:332] Compile and warming up model for size 8 2025-10-10T01:50:42.3582646Z (EngineCore_DP0 pid=56590) (RayWorkerWrapper pid=57479) INFO 10-10 01:50:37 [gpu_model_runner.py:2739] Loading model from scratch... [repeated 3x across cluster] 2025-10-10T01:50:42.7692349Z (EngineCore_DP0 pid=56590) (RayWorkerWrapper pid=57477) INFO 10-10 01:50:42 [backends.py:168] Directly load the compiled graph(s) for shape 8 from the cache, took 0.444 s 2025-10-10T01:50:42.7693745Z (EngineCore_DP0 pid=56590) (RayWorkerWrapper pid=57479) INFO 10-10 01:50:37 [cuda.py:361] Using Flash Attention backend on V1 engine. [repeated 3x across cluster] 2025-10-10T01:50:42.9779235Z (EngineCore_DP0 pid=56590) (RayWorkerWrapper pid=57475) INFO 10-10 01:50:42 [monitor.py:32] torch.compile takes 2.02 s in total 2025-10-10T01:50:42.9782161Z (EngineCore_DP0 pid=56590) (RayWorkerWrapper pid=57477) INFO 10-10 01:50:38 [gpu_model_runner.py:2758] Model loading took 0.7135 GiB and 0.155419 seconds [repeated 3x across cluster] 2025-10-10T01:50:43.0857033Z (EngineCore_DP0 pid=56590) INFO 10-10 01:50:43 [core.py:211] init engine (profile, create kv cache, warmup model) took 5.08 seconds 2025-10-10T01:50:43.8952506Z (EngineCore_DP0 pid=56590) INFO 10-10 01:50:43 [core.py:150] Batch queue is enabled with size 2 2025-10-10T01:50:43.8995334Z (EngineCore_DP0 pid=56590) INFO 10-10 01:50:43 [vllm.py:345] Cudagraph is disabled under eager mode 2025-10-10T01:50:44.0727089Z (APIServer pid=56468) INFO 10-10 01:50:44 [loggers.py:147] Engine 000: vllm cache_config_info with initialization after num_gpu_blocks is: 155568 2025-10-10T01:50:44.1543576Z (EngineCore_DP0 pid=56590) INFO 10-10 01:50:44 [gc_utils.py:41] GC Debug Config. enabled:False,top_objects:-1 2025-10-10T01:50:44.1787177Z (APIServer pid=56468) INFO 10-10 01:50:44 [api_server.py:1634] Supported_tasks: ['generate'] 2025-10-10T01:50:44.2027850Z (APIServer pid=56468) WARNING 10-10 01:50:44 [model.py:1424] Default sampling parameters have been overridden by the model's Hugging Face generation config recommended from the model creator. If this is not intended, please relaunch vLLM instance with `--generation-config vllm`. 2025-10-10T01:50:44.2029524Z (APIServer pid=56468) INFO 10-10 01:50:44 [serving_responses.py:140] Using default chat sampling params from model: {'temperature': 0.6, 'top_p': 0.9} 2025-10-10T01:50:44.2285286Z (APIServer pid=56468) INFO 10-10 01:50:44 [serving_chat.py:139] Using default chat sampling params from model: {'temperature': 0.6, 'top_p': 0.9} 2025-10-10T01:50:44.2479146Z (APIServer pid=56468) INFO 10-10 01:50:44 [serving_completion.py:76] Using default completion sampling params from model: {'temperature': 0.6, 'top_p': 0.9} 2025-10-10T01:50:44.2480205Z (APIServer pid=56468) INFO 10-10 01:50:44 [api_server.py:1912] Starting vLLM API server 0 on http://0.0.0.0:58845 2025-10-10T01:50:44.2481009Z (APIServer pid=56468) INFO 10-10 01:50:44 [launcher.py:34] Available routes are: 2025-10-10T01:50:44.2481792Z (APIServer pid=56468) INFO 10-10 01:50:44 [launcher.py:42] Route: /openapi.json, Methods: GET, HEAD 2025-10-10T01:50:44.2482647Z (APIServer pid=56468) INFO 10-10 01:50:44 [launcher.py:42] Route: /docs, Methods: GET, HEAD 2025-10-10T01:50:44.2483430Z (APIServer pid=56468) INFO 10-10 01:50:44 [launcher.py:42] Route: /docs/oauth2-redirect, Methods: GET, HEAD 2025-10-10T01:50:44.2484112Z (APIServer pid=56468) INFO 10-10 01:50:44 [launcher.py:42] Route: /redoc, Methods: GET, HEAD 2025-10-10T01:50:44.2484678Z (APIServer pid=56468) INFO 10-10 01:50:44 [launcher.py:42] Route: /health, Methods: GET 2025-10-10T01:50:44.2485185Z (APIServer pid=56468) INFO 10-10 01:50:44 [launcher.py:42] Route: /load, Methods: GET 2025-10-10T01:50:44.2485708Z (APIServer pid=56468) INFO 10-10 01:50:44 [launcher.py:42] Route: /ping, Methods: POST 2025-10-10T01:50:44.2486234Z (APIServer pid=56468) INFO 10-10 01:50:44 [launcher.py:42] Route: /ping, Methods: GET 2025-10-10T01:50:44.2487071Z (APIServer pid=56468) INFO 10-10 01:50:44 [launcher.py:42] Route: /tokenize, Methods: POST 2025-10-10T01:50:44.2487625Z (APIServer pid=56468) INFO 10-10 01:50:44 [launcher.py:42] Route: /detokenize, Methods: POST 2025-10-10T01:50:44.2488182Z (APIServer pid=56468) INFO 10-10 01:50:44 [launcher.py:42] Route: /v1/models, Methods: GET 2025-10-10T01:50:44.2488696Z (APIServer pid=56468) INFO 10-10 01:50:44 [launcher.py:42] Route: /version, Methods: GET 2025-10-10T01:50:44.2489238Z (APIServer pid=56468) INFO 10-10 01:50:44 [launcher.py:42] Route: /v1/responses, Methods: POST 2025-10-10T01:50:44.2489825Z (APIServer pid=56468) INFO 10-10 01:50:44 [launcher.py:42] Route: /v1/responses/{response_id}, Methods: GET 2025-10-10T01:50:44.2490640Z (APIServer pid=56468) INFO 10-10 01:50:44 [launcher.py:42] Route: /v1/responses/{response_id}/cancel, Methods: POST 2025-10-10T01:50:44.2491281Z (APIServer pid=56468) INFO 10-10 01:50:44 [launcher.py:42] Route: /v1/chat/completions, Methods: POST 2025-10-10T01:50:44.2491841Z (APIServer pid=56468) INFO 10-10 01:50:44 [launcher.py:42] Route: /v1/completions, Methods: POST 2025-10-10T01:50:44.2492394Z (APIServer pid=56468) INFO 10-10 01:50:44 [launcher.py:42] Route: /v1/embeddings, Methods: POST 2025-10-10T01:50:44.2493021Z (APIServer pid=56468) INFO 10-10 01:50:44 [launcher.py:42] Route: /pooling, Methods: POST 2025-10-10T01:50:44.2493543Z (APIServer pid=56468) INFO 10-10 01:50:44 [launcher.py:42] Route: /classify, Methods: POST 2025-10-10T01:50:44.2494055Z (APIServer pid=56468) INFO 10-10 01:50:44 [launcher.py:42] Route: /score, Methods: POST 2025-10-10T01:50:44.2494564Z (APIServer pid=56468) INFO 10-10 01:50:44 [launcher.py:42] Route: /v1/score, Methods: POST 2025-10-10T01:50:44.2495147Z (APIServer pid=56468) INFO 10-10 01:50:44 [launcher.py:42] Route: /v1/audio/transcriptions, Methods: POST 2025-10-10T01:50:44.2495775Z (APIServer pid=56468) INFO 10-10 01:50:44 [launcher.py:42] Route: /v1/audio/translations, Methods: POST 2025-10-10T01:50:44.2496323Z (APIServer pid=56468) INFO 10-10 01:50:44 [launcher.py:42] Route: /rerank, Methods: POST 2025-10-10T01:50:44.2496849Z (APIServer pid=56468) INFO 10-10 01:50:44 [launcher.py:42] Route: /v1/rerank, Methods: POST 2025-10-10T01:50:44.2497376Z (APIServer pid=56468) INFO 10-10 01:50:44 [launcher.py:42] Route: /v2/rerank, Methods: POST 2025-10-10T01:50:44.2497922Z (APIServer pid=56468) INFO 10-10 01:50:44 [launcher.py:42] Route: /scale_elastic_ep, Methods: POST 2025-10-10T01:50:44.2498495Z (APIServer pid=56468) INFO 10-10 01:50:44 [launcher.py:42] Route: /is_scaling_elastic_ep, Methods: POST 2025-10-10T01:50:44.2499076Z (APIServer pid=56468) INFO 10-10 01:50:44 [launcher.py:42] Route: /invocations, Methods: POST 2025-10-10T01:50:44.2499628Z (APIServer pid=56468) INFO 10-10 01:50:44 [launcher.py:42] Route: /metrics, Methods: GET 2025-10-10T01:50:44.2636381Z (APIServer pid=56468) INFO: Started server process [56468] 2025-10-10T01:50:44.2636905Z (APIServer pid=56468) INFO: Waiting for application startup. 2025-10-10T01:50:44.5084277Z (APIServer pid=56468) INFO: Application startup complete. 2025-10-10T01:50:44.9387217Z (APIServer pid=56468) INFO: 127.0.0.1:49556 - "GET /health HTTP/1.1" 200 OK 2025-10-10T01:50:45.1594554Z (APIServer pid=56468) INFO: 127.0.0.1:49560 - "GET /v1/models HTTP/1.1" 200 OK 2025-10-10T01:50:45.1912282Z (EngineCore_DP0 pid=56590) INFO 10-10 01:50:45 [ray_distributed_executor.py:552] RAY_CGRAPH_get_timeout is set to 300 2025-10-10T01:50:45.1913882Z (EngineCore_DP0 pid=56590) INFO 10-10 01:50:45 [ray_distributed_executor.py:554] VLLM_USE_RAY_COMPILED_DAG_CHANNEL_TYPE = auto 2025-10-10T01:50:45.1915093Z (EngineCore_DP0 pid=56590) INFO 10-10 01:50:45 [ray_distributed_executor.py:556] VLLM_USE_RAY_COMPILED_DAG_OVERLAP_COMM = False 2025-10-10T01:50:45.1919297Z (EngineCore_DP0 pid=56590) INFO 10-10 01:50:45 [ray_distributed_executor.py:621] Using RayPPCommunicator (which wraps vLLM _PP GroupCoordinator) for Ray Compiled Graph communication. 2025-10-10T01:50:45.2207723Z (EngineCore_DP0 pid=56590) 2025-10-10 01:50:45,220 INFO torch_tensor_accelerator_channel.py:807 -- Creating communicator group 1a2fa722-be85-47ba-9c1f-07af0896ec20 on actors: [Actor(RayWorkerWrapper, 6a8c6e5ea16e9b7db2ca721701000000), Actor(RayWorkerWrapper, fe3a86aa8ab65c20c2afa14501000000), Actor(RayWorkerWrapper, 8c1aefa2cc697983c9b941ff01000000), Actor(RayWorkerWrapper, 45cdd89c6587c46f5ba5f44301000000)] 2025-10-10T01:50:45.2752949Z (EngineCore_DP0 pid=56590) 2025-10-10 01:50:45,274 INFO torch_tensor_accelerator_channel.py:833 -- Communicator group initialized. 2025-10-10T01:50:45.3297948Z (EngineCore_DP0 pid=56590) (RayWorkerWrapper pid=57475) /opt/conda/envs/py_3.12/lib/python3.12/site-packages/vllm/distributed/device_communicators/ray_communicator.py:107: UserWarning: The given buffer is not writable, and PyTorch does not support non-writable tensors. This means you can write to the underlying (supposedly non-writable) buffer using the tensor. You may want to copy the buffer to protect its data or make it writable before converting it to a tensor. This type of warning will be suppressed for the rest of this program. (Triggered internally at /var/lib/jenkins/workspace/torch/csrc/utils/tensor_new.cpp:1581.) 2025-10-10T01:50:45.3300939Z (EngineCore_DP0 pid=56590) (RayWorkerWrapper pid=57475) actor_id_tensor = torch.frombuffer( 2025-10-10T01:50:45.3302787Z (EngineCore_DP0 pid=56590) (pid=57479) /opt/conda/envs/py_3.12/lib/python3.12/site-packages/torch/cuda/__init__.py:64: FutureWarning: The pynvml package is deprecated. Please install nvidia-ml-py instead. If you did not install pynvml directly, please report this to the maintainers of the package that installed pynvml for you. [repeated 3x across cluster] 2025-10-10T01:50:45.3304211Z (EngineCore_DP0 pid=56590) (pid=57479) import pynvml # type: ignore[import] [repeated 3x across cluster] 2025-10-10T01:50:45.5042471Z (APIServer pid=56468) INFO: 127.0.0.1:49560 - "POST /v1/completions HTTP/1.1" 200 OK 2025-10-10T01:50:45.5733075Z (APIServer pid=56468) INFO: 127.0.0.1:49560 - "POST /v1/completions HTTP/1.1" 200 OK 2025-10-10T01:50:45.7268289Z (APIServer pid=56468) INFO: 127.0.0.1:49560 - "POST /v1/completions HTTP/1.1" 200 OK 2025-10-10T01:50:45.8083806Z (APIServer pid=56468) INFO: 127.0.0.1:49560 - "POST /v1/completions HTTP/1.1" 200 OK 2025-10-10T01:50:45.8720265Z (APIServer pid=56468) INFO: 127.0.0.1:49560 - "POST /v1/completions HTTP/1.1" 200 OK 2025-10-10T01:50:45.8794811Z (APIServer pid=56468) INFO: 127.0.0.1:49560 - "POST /v1/completions HTTP/1.1" 200 OK 2025-10-10T01:50:45.9739607Z (APIServer pid=56468) INFO 10-10 01:50:45 [launcher.py:99] Shutting down FastAPI HTTP server. 2025-10-10T01:50:45.9740185Z *** SIGTERM received at time=1760061045 on cpu 41 *** 2025-10-10T01:50:45.9786709Z PC: @ 0x7f6d07e99117 (unknown) (unknown) 2025-10-10T01:50:45.9787970Z @ 0x7f6d07e4a520 (unknown) (unknown) 2025-10-10T01:50:45.9788467Z [2025-10-10 01:50:45,978 E 56590 56590] logging.cc:501: *** SIGTERM received at time=1760061045 on cpu 41 *** 2025-10-10T01:50:45.9790517Z [2025-10-10 01:50:45,978 E 56590 56590] logging.cc:501: PC: @ 0x7f6d07e99117 (unknown) (unknown) 2025-10-10T01:50:45.9792040Z [2025-10-10 01:50:45,979 E 56590 56590] logging.cc:501: @ 0x7f6d07e4a520 (unknown) (unknown) 2025-10-10T01:50:45.9795338Z (EngineCore_DP0 pid=56590) INFO 10-10 01:50:45 [ray_distributed_executor.py:122] Shutting down Ray distributed executor. If you see error log from logging.cc regarding SIGTERM received, please ignore because this is the expected termination process in Ray. 2025-10-10T01:50:45.9796451Z (EngineCore_DP0 pid=56590) 2025-10-10 01:50:45,979 INFO compiled_dag_node.py:2171 -- Tearing down compiled DAG 2025-10-10T01:50:45.9797499Z (EngineCore_DP0 pid=56590) 2025-10-10 01:50:45,979 INFO compiled_dag_node.py:2176 -- Cancelling compiled worker on actor: Actor(RayWorkerWrapper, 45cdd89c6587c46f5ba5f44301000000) 2025-10-10T01:50:45.9798824Z (EngineCore_DP0 pid=56590) 2025-10-10 01:50:45,979 INFO compiled_dag_node.py:2176 -- Cancelling compiled worker on actor: Actor(RayWorkerWrapper, fe3a86aa8ab65c20c2afa14501000000) 2025-10-10T01:50:45.9799885Z (EngineCore_DP0 pid=56590) 2025-10-10 01:50:45,979 INFO compiled_dag_node.py:2176 -- Cancelling compiled worker on actor: Actor(RayWorkerWrapper, 8c1aefa2cc697983c9b941ff01000000) 2025-10-10T01:50:45.9800908Z (EngineCore_DP0 pid=56590) 2025-10-10 01:50:45,979 INFO compiled_dag_node.py:2176 -- Cancelling compiled worker on actor: Actor(RayWorkerWrapper, 6a8c6e5ea16e9b7db2ca721701000000) 2025-10-10T01:50:45.9852023Z (EngineCore_DP0 pid=56590) 2025-10-10 01:50:45,984 INFO compiled_dag_node.py:2198 -- Waiting for worker tasks to exit 2025-10-10T01:50:45.9854594Z (EngineCore_DP0 pid=56590) 2025-10-10 01:50:45,985 INFO compiled_dag_node.py:2201 -- Teardown complete 2025-10-10T01:50:46.4885957Z (EngineCore_DP0 pid=56590) (RayWorkerWrapper pid=57479) INFO 10-10 01:50:40 [backends.py:548] Using cache directory: /var/lib/jenkins/.cache/vllm/torch_compile_cache/8cedb68d99/rank_3_0/backbone for vLLM's torch.compile [repeated 3x across cluster] 2025-10-10T01:50:46.4889559Z (EngineCore_DP0 pid=56590) (RayWorkerWrapper pid=57477) INFO 10-10 01:50:40 [backends.py:559] Dynamo bytecode transform time: 2.02 s [repeated 3x across cluster] 2025-10-10T01:50:46.4892341Z (EngineCore_DP0 pid=56590) (RayWorkerWrapper pid=57477) INFO 10-10 01:50:40 [backends.py:164] Directly load the compiled graph(s) for dynamic shape from the cache, took 0.475 s [repeated 3x across cluster] 2025-10-10T01:50:46.4894057Z (EngineCore_DP0 pid=56590) (RayWorkerWrapper pid=57479) INFO 10-10 01:50:41 [gpu_worker.py:298] Available KV cache memory: 18.99 GiB [repeated 3x across cluster] 2025-10-10T01:50:46.4895058Z (EngineCore_DP0 pid=56590) (RayWorkerWrapper pid=57475) INFO 10-10 01:50:42 [gpu_worker.py:332] Compile and warming up model for size 4 [repeated 7x across cluster] 2025-10-10T01:50:46.4896179Z (EngineCore_DP0 pid=56590) (RayWorkerWrapper pid=57479) INFO 10-10 01:50:43 [backends.py:168] Directly load the compiled graph(s) for shape 4 from the cache, took 0.343 s [repeated 7x across cluster] 2025-10-10T01:50:46.4897262Z (EngineCore_DP0 pid=56590) (RayWorkerWrapper pid=57479) INFO 10-10 01:50:43 [monitor.py:32] torch.compile takes 1.97 s in total [repeated 3x across cluster] 2025-10-10T01:50:46.4899955Z (EngineCore_DP0 pid=56590) (RayWorkerWrapper pid=57479) /opt/conda/envs/py_3.12/lib/python3.12/site-packages/vllm/distributed/device_communicators/ray_communicator.py:107: UserWarning: The given buffer is not writable, and PyTorch does not support non-writable tensors. This means you can write to the underlying (supposedly non-writable) buffer using the tensor. You may want to copy the buffer to protect its data or make it writable before converting it to a tensor. This type of warning will be suppressed for the rest of this program. (Triggered internally at /var/lib/jenkins/workspace/torch/csrc/utils/tensor_new.cpp:1581.) [repeated 3x across cluster] 2025-10-10T01:50:46.4902472Z (EngineCore_DP0 pid=56590) (RayWorkerWrapper pid=57479) actor_id_tensor = torch.frombuffer( [repeated 3x across cluster] 2025-10-10T01:50:48.3890892Z (APIServer pid=56468) INFO: Shutting down 2025-10-10T01:50:48.4893096Z (APIServer pid=56468) INFO: Waiting for application shutdown. 2025-10-10T01:50:48.4893771Z (APIServer pid=56468) INFO: Application shutdown complete. 2025-10-10T01:50:48.8372005Z INFO 10-10 01:50:48 [model.py:551] Resolved architecture: LlamaForCausalLM 2025-10-10T01:50:48.8372691Z WARNING 10-10 01:50:48 [model.py:1768] Casting torch.bfloat16 to torch.float16. 2025-10-10T01:50:48.8373704Z INFO 10-10 01:50:48 [model.py:1545] Using max model len 2048 2025-10-10T01:50:48.8375325Z Launching RemoteOpenAIServer with: vllm serve meta-llama/Llama-3.2-1B-Instruct --dtype float16 --max-model-len 2048 --max-num-seqs 8 --enable-chunked-prefill --enforce-eager --tokenizer-mode auto --tensor-parallel-size 2 --distributed-executor-backend mp --load-format dummy --port 54833 --seed 0 2025-10-10T01:50:49.3712141Z /opt/conda/envs/py_3.12/lib/python3.12/site-packages/torch/cuda/__init__.py:64: FutureWarning: The pynvml package is deprecated. Please install nvidia-ml-py instead. If you did not install pynvml directly, please report this to the maintainers of the package that installed pynvml for you. 2025-10-10T01:50:49.3714469Z import pynvml # type: ignore[import] 2025-10-10T01:50:51.5812167Z INFO 10-10 01:50:51 [__init__.py:215] Automatically detected platform cuda. 2025-10-10T01:50:55.5065984Z (APIServer pid=61359) INFO 10-10 01:50:55 [api_server.py:1839] vLLM API server version 0.11.0rc2.dev157+g0ad9951c4.d20251010 2025-10-10T01:50:55.5089605Z (APIServer pid=61359) INFO 10-10 01:50:55 [utils.py:233] non-default args: {'model_tag': 'meta-llama/Llama-3.2-1B-Instruct', 'port': 54833, 'model': 'meta-llama/Llama-3.2-1B-Instruct', 'dtype': 'float16', 'seed': 0, 'max_model_len': 2048, 'enforce_eager': True, 'load_format': 'dummy', 'distributed_executor_backend': 'mp', 'tensor_parallel_size': 2, 'max_num_seqs': 8, 'enable_chunked_prefill': True} 2025-10-10T01:50:55.6520138Z (APIServer pid=61359) INFO 10-10 01:50:55 [model.py:551] Resolved architecture: LlamaForCausalLM 2025-10-10T01:50:55.6520835Z (APIServer pid=61359) `torch_dtype` is deprecated! Use `dtype` instead! 2025-10-10T01:50:55.6521545Z (APIServer pid=61359) WARNING 10-10 01:50:55 [model.py:1768] Casting torch.bfloat16 to torch.float16. 2025-10-10T01:50:55.6524243Z (APIServer pid=61359) INFO 10-10 01:50:55 [model.py:1545] Using max model len 2048 2025-10-10T01:50:55.8294590Z (APIServer pid=61359) INFO 10-10 01:50:55 [scheduler.py:222] Chunked prefill is enabled with max_num_batched_tokens=2048. 2025-10-10T01:50:55.8300613Z (APIServer pid=61359) INFO 10-10 01:50:55 [vllm.py:345] Cudagraph is disabled under eager mode 2025-10-10T01:50:56.8102489Z /opt/conda/envs/py_3.12/lib/python3.12/site-packages/torch/cuda/__init__.py:64: FutureWarning: The pynvml package is deprecated. Please install nvidia-ml-py instead. If you did not install pynvml directly, please report this to the maintainers of the package that installed pynvml for you. 2025-10-10T01:50:56.8104774Z import pynvml # type: ignore[import] 2025-10-10T01:50:59.1169704Z INFO 10-10 01:50:59 [__init__.py:215] Automatically detected platform cuda. 2025-10-10T01:51:03.0956688Z (EngineCore_DP0 pid=61481) INFO 10-10 01:51:03 [core.py:648] Waiting for init message from front-end. 2025-10-10T01:51:03.1226796Z (EngineCore_DP0 pid=61481) INFO 10-10 01:51:03 [core.py:78] Initializing a V1 LLM engine (v0.11.0rc2.dev157+g0ad9951c4.d20251010) with config: model='meta-llama/Llama-3.2-1B-Instruct', speculative_config=None, tokenizer='meta-llama/Llama-3.2-1B-Instruct', skip_tokenizer_init=False, tokenizer_mode=auto, revision=None, tokenizer_revision=None, trust_remote_code=False, dtype=torch.float16, max_seq_len=2048, download_dir=None, load_format=dummy, tensor_parallel_size=2, pipeline_parallel_size=1, data_parallel_size=1, disable_custom_all_reduce=False, quantization=None, enforce_eager=True, kv_cache_dtype=auto, device_config=cuda, structured_outputs_config=StructuredOutputsConfig(backend='auto', disable_fallback=False, disable_any_whitespace=False, disable_additional_properties=False, reasoning_parser=''), observability_config=ObservabilityConfig(show_hidden_metrics_for_version=None, otlp_traces_endpoint=None, collect_detailed_traces=None), seed=0, served_model_name=meta-llama/Llama-3.2-1B-Instruct, enable_prefix_caching=True, chunked_prefill_enabled=True, pooler_config=None, compilation_config={'level': 0, 'debug_dump_path': None, 'cache_dir': '', 'backend': '', 'custom_ops': [], 'splitting_ops': None, 'use_inductor': True, 'compile_sizes': [], 'inductor_compile_config': {'enable_auto_functionalized_v2': False}, 'inductor_passes': {}, 'cudagraph_mode': , 'use_cudagraph': False, 'cudagraph_num_of_warmups': 0, 'cudagraph_capture_sizes': [], 'cudagraph_copy_inputs': False, 'full_cuda_graph': False, 'use_inductor_graph_partition': False, 'pass_config': {}, 'max_capture_size': 0, 'local_cache_dir': None} 2025-10-10T01:51:03.1238078Z (EngineCore_DP0 pid=61481) WARNING 10-10 01:51:03 [multiproc_executor.py:720] Reducing Torch parallelism from 24 threads to 1 to avoid unnecessary CPU contention. Set OMP_NUM_THREADS in the external environment to tune this value as needed. 2025-10-10T01:51:03.1240187Z (EngineCore_DP0 pid=61481) INFO 10-10 01:51:03 [shm_broadcast.py:289] vLLM message queue communication handle: Handle(local_reader_ranks=[0, 1], buffer_handle=(2, 16777216, 10, 'psm_66d89163'), local_subscribe_addr='ipc:///tmp/39d70b9e-d520-42a7-82e6-dbdd98354821', remote_subscribe_addr=None, remote_addr_ipv6=False) 2025-10-10T01:51:03.6034955Z /opt/conda/envs/py_3.12/lib/python3.12/site-packages/torch/cuda/__init__.py:64: FutureWarning: The pynvml package is deprecated. Please install nvidia-ml-py instead. If you did not install pynvml directly, please report this to the maintainers of the package that installed pynvml for you. 2025-10-10T01:51:03.6037242Z import pynvml # type: ignore[import] 2025-10-10T01:51:03.6177788Z /opt/conda/envs/py_3.12/lib/python3.12/site-packages/torch/cuda/__init__.py:64: FutureWarning: The pynvml package is deprecated. Please install nvidia-ml-py instead. If you did not install pynvml directly, please report this to the maintainers of the package that installed pynvml for you. 2025-10-10T01:51:03.6179839Z import pynvml # type: ignore[import] 2025-10-10T01:51:05.9363240Z INFO 10-10 01:51:05 [__init__.py:215] Automatically detected platform cuda. 2025-10-10T01:51:05.9588387Z INFO 10-10 01:51:05 [__init__.py:215] Automatically detected platform cuda. 2025-10-10T01:51:10.4645564Z INFO 10-10 01:51:10 [shm_broadcast.py:289] vLLM message queue communication handle: Handle(local_reader_ranks=[0], buffer_handle=(1, 10485760, 10, 'psm_76107c2c'), local_subscribe_addr='ipc:///tmp/6feb2ab9-3089-447c-91e6-a375748e1feb', remote_subscribe_addr=None, remote_addr_ipv6=False) 2025-10-10T01:51:10.5339882Z INFO 10-10 01:51:10 [shm_broadcast.py:289] vLLM message queue communication handle: Handle(local_reader_ranks=[0], buffer_handle=(1, 10485760, 10, 'psm_0aecb023'), local_subscribe_addr='ipc:///tmp/bea529d6-b03b-4774-909d-7e4da10c8963', remote_subscribe_addr=None, remote_addr_ipv6=False) 2025-10-10T01:51:10.6822422Z [Gloo] Rank 0 is connected to 1 peer ranks. Expected number of connected peer ranks is : 1 2025-10-10T01:51:10.6823097Z [Gloo] Rank 1 is connected to 1 peer ranks. Expected number of connected peer ranks is : 1 2025-10-10T01:51:10.6863026Z [Gloo] Rank 0 is connected to 1 peer ranks. Expected number of connected peer ranks is : 1 2025-10-10T01:51:10.6863992Z [Gloo] Rank 1 is connected to 1 peer ranks. Expected number of connected peer ranks is : 1 2025-10-10T01:51:10.6905739Z INFO 10-10 01:51:10 [__init__.py:1384] Found nccl from library libnccl.so.2 2025-10-10T01:51:10.6906671Z INFO 10-10 01:51:10 [__init__.py:1384] Found nccl from library libnccl.so.2 2025-10-10T01:51:10.6907823Z INFO 10-10 01:51:10 [pynccl.py:104] vLLM is using nccl==2.27.5 2025-10-10T01:51:10.6908561Z INFO 10-10 01:51:10 [pynccl.py:104] vLLM is using nccl==2.27.5 2025-10-10T01:51:10.8651320Z WARNING 10-10 01:51:10 [symm_mem.py:58] SymmMemCommunicator: Device capability 8.9 not supported, communicator is not available. 2025-10-10T01:51:10.8653048Z WARNING 10-10 01:51:10 [symm_mem.py:58] SymmMemCommunicator: Device capability 8.9 not supported, communicator is not available. 2025-10-10T01:51:10.8679526Z INFO 10-10 01:51:10 [custom_all_reduce.py:35] Skipping P2P check and trusting the driver's P2P report. 2025-10-10T01:51:10.8681295Z INFO 10-10 01:51:10 [custom_all_reduce.py:35] Skipping P2P check and trusting the driver's P2P report. 2025-10-10T01:51:10.8682506Z WARNING 10-10 01:51:10 [custom_all_reduce.py:154] Custom allreduce is disabled because your platform lacks GPU P2P capability or P2P test failed. To silence this warning, specify disable_custom_all_reduce=True explicitly. 2025-10-10T01:51:10.8683902Z WARNING 10-10 01:51:10 [custom_all_reduce.py:154] Custom allreduce is disabled because your platform lacks GPU P2P capability or P2P test failed. To silence this warning, specify disable_custom_all_reduce=True explicitly. 2025-10-10T01:51:10.9440242Z INFO 10-10 01:51:10 [shm_broadcast.py:289] vLLM message queue communication handle: Handle(local_reader_ranks=[1], buffer_handle=(1, 4194304, 6, 'psm_293ce13d'), local_subscribe_addr='ipc:///tmp/b291dd5c-edc6-44a0-a672-c0f6941adb9f', remote_subscribe_addr=None, remote_addr_ipv6=False) 2025-10-10T01:51:10.9467641Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-10-10T01:51:10.9468856Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-10-10T01:51:10.9475785Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-10-10T01:51:10.9484719Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-10-10T01:51:11.0237130Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-10-10T01:51:11.0245610Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-10-10T01:51:11.0258367Z [Gloo] Rank 0 is connected to 1 peer ranks. Expected number of connected peer ranks is : 1 2025-10-10T01:51:11.0259546Z [Gloo] Rank 1 is connected to 1 peer ranks. Expected number of connected peer ranks is : 1 2025-10-10T01:51:11.0260934Z INFO 10-10 01:51:11 [__init__.py:1384] Found nccl from library libnccl.so.2 2025-10-10T01:51:11.0261815Z INFO 10-10 01:51:11 [__init__.py:1384] Found nccl from library libnccl.so.2 2025-10-10T01:51:11.0262649Z INFO 10-10 01:51:11 [pynccl.py:104] vLLM is using nccl==2.27.5 2025-10-10T01:51:11.0263430Z INFO 10-10 01:51:11 [pynccl.py:104] vLLM is using nccl==2.27.5 2025-10-10T01:51:11.0608988Z INFO 10-10 01:51:11 [parallel_state.py:1208] rank 0 in world size 2 is assigned as DP rank 0, PP rank 0, TP rank 0, EP rank 0 2025-10-10T01:51:11.0610475Z INFO 10-10 01:51:11 [parallel_state.py:1208] rank 1 in world size 2 is assigned as DP rank 0, PP rank 0, TP rank 1, EP rank 1 2025-10-10T01:51:11.3107278Z INFO 10-10 01:51:11 [topk_topp_sampler.py:55] Using FlashInfer for top-p & top-k sampling. 2025-10-10T01:51:11.3109951Z INFO 10-10 01:51:11 [topk_topp_sampler.py:55] Using FlashInfer for top-p & top-k sampling. 2025-10-10T01:51:11.3421562Z (Worker_TP1 pid=61601) INFO 10-10 01:51:11 [gpu_model_runner.py:2707] Starting to load model meta-llama/Llama-3.2-1B-Instruct... 2025-10-10T01:51:11.3428080Z (Worker_TP0 pid=61600) INFO 10-10 01:51:11 [gpu_model_runner.py:2707] Starting to load model meta-llama/Llama-3.2-1B-Instruct... 2025-10-10T01:51:11.5696421Z (Worker_TP1 pid=61601) INFO 10-10 01:51:11 [gpu_model_runner.py:2739] Loading model from scratch... 2025-10-10T01:51:11.5697326Z (Worker_TP0 pid=61600) INFO 10-10 01:51:11 [gpu_model_runner.py:2739] Loading model from scratch... 2025-10-10T01:51:11.6593926Z (Worker_TP1 pid=61601) INFO 10-10 01:51:11 [cuda.py:361] Using Flash Attention backend on V1 engine. 2025-10-10T01:51:11.6602644Z (Worker_TP0 pid=61600) INFO 10-10 01:51:11 [cuda.py:361] Using Flash Attention backend on V1 engine. 2025-10-10T01:51:12.1379825Z (Worker_TP1 pid=61601) INFO 10-10 01:51:12 [gpu_model_runner.py:2758] Model loading took 1.1667 GiB and 0.118853 seconds 2025-10-10T01:51:12.1395112Z (Worker_TP0 pid=61600) INFO 10-10 01:51:12 [gpu_model_runner.py:2758] Model loading took 1.1667 GiB and 0.119702 seconds 2025-10-10T01:51:17.0478105Z (Worker_TP1 pid=61601) INFO 10-10 01:51:17 [gpu_worker.py:298] Available KV cache memory: 18.53 GiB 2025-10-10T01:51:17.0578324Z (Worker_TP0 pid=61600) INFO 10-10 01:51:17 [gpu_worker.py:298] Available KV cache memory: 18.53 GiB 2025-10-10T01:51:17.2842081Z (EngineCore_DP0 pid=61481) INFO 10-10 01:51:17 [kv_cache_utils.py:1087] GPU KV cache size: 1,214,464 tokens 2025-10-10T01:51:17.2843298Z (EngineCore_DP0 pid=61481) INFO 10-10 01:51:17 [kv_cache_utils.py:1091] Maximum concurrency for 2,048 tokens per request: 593.00x 2025-10-10T01:51:17.2853433Z (EngineCore_DP0 pid=61481) INFO 10-10 01:51:17 [kv_cache_utils.py:1087] GPU KV cache size: 1,214,464 tokens 2025-10-10T01:51:17.2854436Z (EngineCore_DP0 pid=61481) INFO 10-10 01:51:17 [kv_cache_utils.py:1091] Maximum concurrency for 2,048 tokens per request: 593.00x 2025-10-10T01:51:17.4069283Z (EngineCore_DP0 pid=61481) INFO 10-10 01:51:17 [core.py:211] init engine (profile, create kv cache, warmup model) took 5.21 seconds 2025-10-10T01:51:17.9122590Z (EngineCore_DP0 pid=61481) INFO 10-10 01:51:17 [vllm.py:345] Cudagraph is disabled under eager mode 2025-10-10T01:51:18.0967513Z (APIServer pid=61359) INFO 10-10 01:51:18 [loggers.py:147] Engine 000: vllm cache_config_info with initialization after num_gpu_blocks is: 75904 2025-10-10T01:51:18.1392125Z (EngineCore_DP0 pid=61481) INFO 10-10 01:51:18 [gc_utils.py:41] GC Debug Config. enabled:False,top_objects:-1 2025-10-10T01:51:18.1610183Z (APIServer pid=61359) INFO 10-10 01:51:18 [api_server.py:1634] Supported_tasks: ['generate'] 2025-10-10T01:51:18.1865667Z (APIServer pid=61359) WARNING 10-10 01:51:18 [model.py:1424] Default sampling parameters have been overridden by the model's Hugging Face generation config recommended from the model creator. If this is not intended, please relaunch vLLM instance with `--generation-config vllm`. 2025-10-10T01:51:18.1867279Z (APIServer pid=61359) INFO 10-10 01:51:18 [serving_responses.py:140] Using default chat sampling params from model: {'temperature': 0.6, 'top_p': 0.9} 2025-10-10T01:51:18.2082795Z (APIServer pid=61359) INFO 10-10 01:51:18 [serving_chat.py:139] Using default chat sampling params from model: {'temperature': 0.6, 'top_p': 0.9} 2025-10-10T01:51:18.2360597Z (APIServer pid=61359) INFO 10-10 01:51:18 [serving_completion.py:76] Using default completion sampling params from model: {'temperature': 0.6, 'top_p': 0.9} 2025-10-10T01:51:18.2362657Z (APIServer pid=61359) INFO 10-10 01:51:18 [api_server.py:1912] Starting vLLM API server 0 on http://0.0.0.0:54833 2025-10-10T01:51:18.2364043Z (APIServer pid=61359) INFO 10-10 01:51:18 [launcher.py:34] Available routes are: 2025-10-10T01:51:18.2364703Z (APIServer pid=61359) INFO 10-10 01:51:18 [launcher.py:42] Route: /openapi.json, Methods: HEAD, GET 2025-10-10T01:51:18.2365264Z (APIServer pid=61359) INFO 10-10 01:51:18 [launcher.py:42] Route: /docs, Methods: HEAD, GET 2025-10-10T01:51:18.2366140Z (APIServer pid=61359) INFO 10-10 01:51:18 [launcher.py:42] Route: /docs/oauth2-redirect, Methods: HEAD, GET 2025-10-10T01:51:18.2366736Z (APIServer pid=61359) INFO 10-10 01:51:18 [launcher.py:42] Route: /redoc, Methods: HEAD, GET 2025-10-10T01:51:18.2367281Z (APIServer pid=61359) INFO 10-10 01:51:18 [launcher.py:42] Route: /health, Methods: GET 2025-10-10T01:51:18.2367798Z (APIServer pid=61359) INFO 10-10 01:51:18 [launcher.py:42] Route: /load, Methods: GET 2025-10-10T01:51:18.2368305Z (APIServer pid=61359) INFO 10-10 01:51:18 [launcher.py:42] Route: /ping, Methods: POST 2025-10-10T01:51:18.2368827Z (APIServer pid=61359) INFO 10-10 01:51:18 [launcher.py:42] Route: /ping, Methods: GET 2025-10-10T01:51:18.2369345Z (APIServer pid=61359) INFO 10-10 01:51:18 [launcher.py:42] Route: /tokenize, Methods: POST 2025-10-10T01:51:18.2370046Z (APIServer pid=61359) INFO 10-10 01:51:18 [launcher.py:42] Route: /detokenize, Methods: POST 2025-10-10T01:51:18.2370591Z (APIServer pid=61359) INFO 10-10 01:51:18 [launcher.py:42] Route: /v1/models, Methods: GET 2025-10-10T01:51:18.2371102Z (APIServer pid=61359) INFO 10-10 01:51:18 [launcher.py:42] Route: /version, Methods: GET 2025-10-10T01:51:18.2371727Z (APIServer pid=61359) INFO 10-10 01:51:18 [launcher.py:42] Route: /v1/responses, Methods: POST 2025-10-10T01:51:18.2372313Z (APIServer pid=61359) INFO 10-10 01:51:18 [launcher.py:42] Route: /v1/responses/{response_id}, Methods: GET 2025-10-10T01:51:18.2372947Z (APIServer pid=61359) INFO 10-10 01:51:18 [launcher.py:42] Route: /v1/responses/{response_id}/cancel, Methods: POST 2025-10-10T01:51:18.2373566Z (APIServer pid=61359) INFO 10-10 01:51:18 [launcher.py:42] Route: /v1/chat/completions, Methods: POST 2025-10-10T01:51:18.2374135Z (APIServer pid=61359) INFO 10-10 01:51:18 [launcher.py:42] Route: /v1/completions, Methods: POST 2025-10-10T01:51:18.2374692Z (APIServer pid=61359) INFO 10-10 01:51:18 [launcher.py:42] Route: /v1/embeddings, Methods: POST 2025-10-10T01:51:18.2375233Z (APIServer pid=61359) INFO 10-10 01:51:18 [launcher.py:42] Route: /pooling, Methods: POST 2025-10-10T01:51:18.2375764Z (APIServer pid=61359) INFO 10-10 01:51:18 [launcher.py:42] Route: /classify, Methods: POST 2025-10-10T01:51:18.2376284Z (APIServer pid=61359) INFO 10-10 01:51:18 [launcher.py:42] Route: /score, Methods: POST 2025-10-10T01:51:18.2376795Z (APIServer pid=61359) INFO 10-10 01:51:18 [launcher.py:42] Route: /v1/score, Methods: POST 2025-10-10T01:51:18.2377362Z (APIServer pid=61359) INFO 10-10 01:51:18 [launcher.py:42] Route: /v1/audio/transcriptions, Methods: POST 2025-10-10T01:51:18.2377962Z (APIServer pid=61359) INFO 10-10 01:51:18 [launcher.py:42] Route: /v1/audio/translations, Methods: POST 2025-10-10T01:51:18.2378530Z (APIServer pid=61359) INFO 10-10 01:51:18 [launcher.py:42] Route: /rerank, Methods: POST 2025-10-10T01:51:18.2379056Z (APIServer pid=61359) INFO 10-10 01:51:18 [launcher.py:42] Route: /v1/rerank, Methods: POST 2025-10-10T01:51:18.2379576Z (APIServer pid=61359) INFO 10-10 01:51:18 [launcher.py:42] Route: /v2/rerank, Methods: POST 2025-10-10T01:51:18.2380126Z (APIServer pid=61359) INFO 10-10 01:51:18 [launcher.py:42] Route: /scale_elastic_ep, Methods: POST 2025-10-10T01:51:18.2380700Z (APIServer pid=61359) INFO 10-10 01:51:18 [launcher.py:42] Route: /is_scaling_elastic_ep, Methods: POST 2025-10-10T01:51:18.2381261Z (APIServer pid=61359) INFO 10-10 01:51:18 [launcher.py:42] Route: /invocations, Methods: POST 2025-10-10T01:51:18.2381790Z (APIServer pid=61359) INFO 10-10 01:51:18 [launcher.py:42] Route: /metrics, Methods: GET 2025-10-10T01:51:18.2516205Z (APIServer pid=61359) INFO: Started server process [61359] 2025-10-10T01:51:18.2517674Z (APIServer pid=61359) INFO: Waiting for application startup. 2025-10-10T01:51:18.4811263Z (APIServer pid=61359) INFO: Application startup complete. 2025-10-10T01:51:18.9087258Z (APIServer pid=61359) INFO: 127.0.0.1:50536 - "GET /health HTTP/1.1" 200 OK 2025-10-10T01:51:18.9173042Z (APIServer pid=61359) INFO: 127.0.0.1:50540 - "GET /v1/models HTTP/1.1" 200 OK 2025-10-10T01:51:19.0903102Z (APIServer pid=61359) INFO: 127.0.0.1:50540 - "POST /v1/completions HTTP/1.1" 200 OK 2025-10-10T01:51:19.1468822Z (APIServer pid=61359) INFO: 127.0.0.1:50540 - "POST /v1/completions HTTP/1.1" 200 OK 2025-10-10T01:51:19.2342288Z (APIServer pid=61359) INFO: 127.0.0.1:50540 - "POST /v1/completions HTTP/1.1" 200 OK 2025-10-10T01:51:19.3178114Z (APIServer pid=61359) INFO: 127.0.0.1:50540 - "POST /v1/completions HTTP/1.1" 200 OK 2025-10-10T01:51:19.3882389Z (APIServer pid=61359) INFO: 127.0.0.1:50540 - "POST /v1/completions HTTP/1.1" 200 OK 2025-10-10T01:51:19.3959433Z (APIServer pid=61359) INFO: 127.0.0.1:50540 - "POST /v1/completions HTTP/1.1" 200 OK 2025-10-10T01:51:19.4727186Z (APIServer pid=61359) INFO 10-10 01:51:19 [launcher.py:99] Shutting down FastAPI HTTP server. 2025-10-10T01:51:19.4730236Z (Worker_TP0 pid=61600) INFO 10-10 01:51:19 [multiproc_executor.py:558] Parent process exited, terminating worker 2025-10-10T01:51:19.4732412Z (Worker_TP0 pid=61600) INFO 10-10 01:51:19 [multiproc_executor.py:599] WorkerProc shutting down. 2025-10-10T01:51:19.4734586Z (Worker_TP1 pid=61601) INFO 10-10 01:51:19 [multiproc_executor.py:558] Parent process exited, terminating worker 2025-10-10T01:51:19.4735671Z (Worker_TP1 pid=61601) INFO 10-10 01:51:19 [multiproc_executor.py:599] WorkerProc shutting down. 2025-10-10T01:51:21.3955940Z (APIServer pid=61359) INFO: Shutting down 2025-10-10T01:51:21.4960763Z (APIServer pid=61359) INFO: Waiting for application shutdown. 2025-10-10T01:51:21.4962037Z (APIServer pid=61359) INFO: Application shutdown complete. 2025-10-10T01:51:22.1179844Z PASSED 2025-10-10T01:51:22.1292712Z distributed/test_sequence_parallel.py::test_tp_sp_generation[RedHatAI/Meta-Llama-3.1-8B-Instruct-FP8-parallel_setup16-mp-1-auto-test_options16] Fork a new process to run a test 62118 2025-10-10T01:51:22.1309960Z Fork a new process to run a test 0 2025-10-10T01:51:22.1794557Z 2025-10-10T01:51:22.1800517Z tokenizer_config.json: 0.00B [00:00, ?B/s] 2025-10-10T01:51:22.1801219Z tokenizer_config.json: 55.4kB [00:00, 99.1MB/s] 2025-10-10T01:51:22.2483571Z 2025-10-10T01:51:22.2841321Z tokenizer.json: 0.00B [00:00, ?B/s] 2025-10-10T01:51:22.2842037Z tokenizer.json: 9.09MB [00:00, 255MB/s] 2025-10-10T01:51:22.3597132Z 2025-10-10T01:51:22.3598781Z special_tokens_map.json: 0% 0.00/325 [00:00, 'use_cudagraph': False, 'cudagraph_num_of_warmups': 0, 'cudagraph_capture_sizes': [], 'cudagraph_copy_inputs': False, 'full_cuda_graph': False, 'use_inductor_graph_partition': False, 'pass_config': {'enable_noop': True, 'enable_sequence_parallelism': True}, 'max_capture_size': 0, 'local_cache_dir': None} 2025-10-10T01:51:37.7036097Z (EngineCore_DP0 pid=62292) WARNING 10-10 01:51:37 [multiproc_executor.py:720] Reducing Torch parallelism from 24 threads to 1 to avoid unnecessary CPU contention. Set OMP_NUM_THREADS in the external environment to tune this value as needed. 2025-10-10T01:51:37.7040212Z (EngineCore_DP0 pid=62292) INFO 10-10 01:51:37 [shm_broadcast.py:289] vLLM message queue communication handle: Handle(local_reader_ranks=[0, 1], buffer_handle=(2, 16777216, 10, 'psm_5741d6b2'), local_subscribe_addr='ipc:///tmp/b28d215b-ff2f-4147-8ac5-66ca318a7f15', remote_subscribe_addr=None, remote_addr_ipv6=False) 2025-10-10T01:51:38.1823816Z /opt/conda/envs/py_3.12/lib/python3.12/site-packages/torch/cuda/__init__.py:64: FutureWarning: The pynvml package is deprecated. Please install nvidia-ml-py instead. If you did not install pynvml directly, please report this to the maintainers of the package that installed pynvml for you. 2025-10-10T01:51:38.1827007Z import pynvml # type: ignore[import] 2025-10-10T01:51:38.1871172Z /opt/conda/envs/py_3.12/lib/python3.12/site-packages/torch/cuda/__init__.py:64: FutureWarning: The pynvml package is deprecated. Please install nvidia-ml-py instead. If you did not install pynvml directly, please report this to the maintainers of the package that installed pynvml for you. 2025-10-10T01:51:38.1873409Z import pynvml # type: ignore[import] 2025-10-10T01:51:40.5115462Z INFO 10-10 01:51:40 [__init__.py:215] Automatically detected platform cuda. 2025-10-10T01:51:40.5246846Z INFO 10-10 01:51:40 [__init__.py:215] Automatically detected platform cuda. 2025-10-10T01:51:45.1746292Z INFO 10-10 01:51:45 [shm_broadcast.py:289] vLLM message queue communication handle: Handle(local_reader_ranks=[0], buffer_handle=(1, 10485760, 10, 'psm_50bed27c'), local_subscribe_addr='ipc:///tmp/5f267bb1-f640-4287-bb98-078383ba2193', remote_subscribe_addr=None, remote_addr_ipv6=False) 2025-10-10T01:51:45.1748283Z INFO 10-10 01:51:45 [shm_broadcast.py:289] vLLM message queue communication handle: Handle(local_reader_ranks=[0], buffer_handle=(1, 10485760, 10, 'psm_9dd4969d'), local_subscribe_addr='ipc:///tmp/1621fb8c-8b8d-4c10-bd50-e11d6aa9ca7e', remote_subscribe_addr=None, remote_addr_ipv6=False) 2025-10-10T01:51:45.8784148Z [Gloo] Rank 0 is connected to 1 peer ranks. Expected number of connected peer ranks is : 1 2025-10-10T01:51:45.8785695Z [Gloo] Rank 1 is connected to 1 peer ranks. Expected number of connected peer ranks is : 1 2025-10-10T01:51:45.8831104Z [Gloo] Rank 1 is connected to 1 peer ranks. Expected number of connected peer ranks is : 1 2025-10-10T01:51:45.8832252Z [Gloo] Rank 0 is connected to 1 peer ranks. Expected number of connected peer ranks is : 1 2025-10-10T01:51:45.8872467Z INFO 10-10 01:51:45 [__init__.py:1384] Found nccl from library libnccl.so.2 2025-10-10T01:51:45.8874211Z INFO 10-10 01:51:45 [pynccl.py:104] vLLM is using nccl==2.27.5 2025-10-10T01:51:45.8878710Z INFO 10-10 01:51:45 [__init__.py:1384] Found nccl from library libnccl.so.2 2025-10-10T01:51:45.8881022Z INFO 10-10 01:51:45 [pynccl.py:104] vLLM is using nccl==2.27.5 2025-10-10T01:51:46.0612386Z WARNING 10-10 01:51:46 [symm_mem.py:58] SymmMemCommunicator: Device capability 8.9 not supported, communicator is not available. 2025-10-10T01:51:46.0614718Z WARNING 10-10 01:51:46 [symm_mem.py:58] SymmMemCommunicator: Device capability 8.9 not supported, communicator is not available. 2025-10-10T01:51:46.0641286Z INFO 10-10 01:51:46 [custom_all_reduce.py:35] Skipping P2P check and trusting the driver's P2P report. 2025-10-10T01:51:46.0642510Z INFO 10-10 01:51:46 [custom_all_reduce.py:35] Skipping P2P check and trusting the driver's P2P report. 2025-10-10T01:51:46.0644751Z WARNING 10-10 01:51:46 [custom_all_reduce.py:154] Custom allreduce is disabled because your platform lacks GPU P2P capability or P2P test failed. To silence this warning, specify disable_custom_all_reduce=True explicitly. 2025-10-10T01:51:46.0646394Z WARNING 10-10 01:51:46 [custom_all_reduce.py:154] Custom allreduce is disabled because your platform lacks GPU P2P capability or P2P test failed. To silence this warning, specify disable_custom_all_reduce=True explicitly. 2025-10-10T01:51:46.1439990Z INFO 10-10 01:51:46 [shm_broadcast.py:289] vLLM message queue communication handle: Handle(local_reader_ranks=[1], buffer_handle=(1, 4194304, 6, 'psm_e3691955'), local_subscribe_addr='ipc:///tmp/8f7396ea-4a6b-4eec-955c-6e9929347202', remote_subscribe_addr=None, remote_addr_ipv6=False) 2025-10-10T01:51:46.1471840Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-10-10T01:51:46.1472950Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-10-10T01:51:46.1480788Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-10-10T01:51:46.1487633Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-10-10T01:51:46.1936607Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-10-10T01:51:46.1947216Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-10-10T01:51:46.1958769Z [Gloo] Rank 0 is connected to 1 peer ranks. Expected number of connected peer ranks is : 1 2025-10-10T01:51:46.1959800Z [Gloo] Rank 1 is connected to 1 peer ranks. Expected number of connected peer ranks is : 1 2025-10-10T01:51:46.1960903Z INFO 10-10 01:51:46 [__init__.py:1384] Found nccl from library libnccl.so.2 2025-10-10T01:51:46.1961385Z INFO 10-10 01:51:46 [__init__.py:1384] Found nccl from library libnccl.so.2 2025-10-10T01:51:46.1961808Z INFO 10-10 01:51:46 [pynccl.py:104] vLLM is using nccl==2.27.5 2025-10-10T01:51:46.1962205Z INFO 10-10 01:51:46 [pynccl.py:104] vLLM is using nccl==2.27.5 2025-10-10T01:51:46.2295359Z INFO 10-10 01:51:46 [parallel_state.py:1208] rank 0 in world size 2 is assigned as DP rank 0, PP rank 0, TP rank 0, EP rank 0 2025-10-10T01:51:46.2296107Z INFO 10-10 01:51:46 [parallel_state.py:1208] rank 1 in world size 2 is assigned as DP rank 0, PP rank 0, TP rank 1, EP rank 1 2025-10-10T01:51:46.4713876Z INFO 10-10 01:51:46 [topk_topp_sampler.py:55] Using FlashInfer for top-p & top-k sampling. 2025-10-10T01:51:46.4762157Z INFO 10-10 01:51:46 [topk_topp_sampler.py:55] Using FlashInfer for top-p & top-k sampling. 2025-10-10T01:51:46.5077205Z (Worker_TP1 pid=62412) INFO 10-10 01:51:46 [gpu_model_runner.py:2707] Starting to load model RedHatAI/Meta-Llama-3.1-8B-Instruct-FP8... 2025-10-10T01:51:46.5125062Z (Worker_TP0 pid=62411) INFO 10-10 01:51:46 [gpu_model_runner.py:2707] Starting to load model RedHatAI/Meta-Llama-3.1-8B-Instruct-FP8... 2025-10-10T01:51:46.7393382Z (Worker_TP1 pid=62412) INFO 10-10 01:51:46 [gpu_model_runner.py:2739] Loading model from scratch... 2025-10-10T01:51:46.7432077Z (Worker_TP0 pid=62411) INFO 10-10 01:51:46 [gpu_model_runner.py:2739] Loading model from scratch... 2025-10-10T01:51:46.8307976Z (Worker_TP1 pid=62412) INFO 10-10 01:51:46 [cuda.py:361] Using Flash Attention backend on V1 engine. 2025-10-10T01:51:46.8349656Z (Worker_TP0 pid=62411) INFO 10-10 01:51:46 [cuda.py:361] Using Flash Attention backend on V1 engine. 2025-10-10T01:51:47.7426261Z (Worker_TP0 pid=62411) INFO 10-10 01:51:47 [gpu_model_runner.py:2758] Model loading took 4.2624 GiB and 0.536474 seconds 2025-10-10T01:51:47.7464601Z (Worker_TP1 pid=62412) INFO 10-10 01:51:47 [gpu_model_runner.py:2758] Model loading took 4.2624 GiB and 0.540515 seconds 2025-10-10T01:51:55.1462197Z (Worker_TP0 pid=62411) INFO 10-10 01:51:55 [backends.py:548] Using cache directory: /var/lib/jenkins/.cache/vllm/torch_compile_cache/a29f583b26/rank_0_0/backbone for vLLM's torch.compile 2025-10-10T01:51:55.1464805Z (Worker_TP0 pid=62411) INFO 10-10 01:51:55 [backends.py:559] Dynamo bytecode transform time: 7.11 s 2025-10-10T01:51:55.2276305Z (Worker_TP1 pid=62412) INFO 10-10 01:51:55 [backends.py:548] Using cache directory: /var/lib/jenkins/.cache/vllm/torch_compile_cache/a29f583b26/rank_1_0/backbone for vLLM's torch.compile 2025-10-10T01:51:55.2281308Z (Worker_TP1 pid=62412) INFO 10-10 01:51:55 [backends.py:559] Dynamo bytecode transform time: 7.19 s 2025-10-10T01:51:57.8649807Z (Worker_TP0 pid=62411) INFO 10-10 01:51:57 [backends.py:197] Cache the graph for dynamic shape for later use 2025-10-10T01:51:57.9811934Z (Worker_TP1 pid=62412) INFO 10-10 01:51:57 [backends.py:197] Cache the graph for dynamic shape for later use 2025-10-10T01:52:24.2854989Z (Worker_TP0 pid=62411) INFO 10-10 01:52:24 [backends.py:218] Compiling a graph for dynamic shape takes 28.67 s 2025-10-10T01:52:25.8859303Z (Worker_TP1 pid=62412) INFO 10-10 01:52:25 [backends.py:218] Compiling a graph for dynamic shape takes 30.19 s 2025-10-10T01:52:47.1171859Z (Worker_TP1 pid=62412) INFO 10-10 01:52:47 [gpu_worker.py:298] Available KV cache memory: 15.29 GiB 2025-10-10T01:52:47.1261901Z (Worker_TP0 pid=62411) INFO 10-10 01:52:47 [gpu_worker.py:298] Available KV cache memory: 15.29 GiB 2025-10-10T01:52:47.5019933Z (EngineCore_DP0 pid=62292) INFO 10-10 01:52:47 [kv_cache_utils.py:1087] GPU KV cache size: 250,528 tokens 2025-10-10T01:52:47.5020867Z (EngineCore_DP0 pid=62292) INFO 10-10 01:52:47 [kv_cache_utils.py:1091] Maximum concurrency for 2,048 tokens per request: 122.33x 2025-10-10T01:52:47.5021904Z (EngineCore_DP0 pid=62292) INFO 10-10 01:52:47 [kv_cache_utils.py:1087] GPU KV cache size: 250,528 tokens 2025-10-10T01:52:47.5022866Z (EngineCore_DP0 pid=62292) INFO 10-10 01:52:47 [kv_cache_utils.py:1091] Maximum concurrency for 2,048 tokens per request: 122.33x 2025-10-10T01:52:47.5138130Z (Worker_TP1 pid=62412) INFO 10-10 01:52:47 [gpu_worker.py:332] Compile and warming up model for size 8 2025-10-10T01:52:47.5138976Z (Worker_TP0 pid=62411) INFO 10-10 01:52:47 [gpu_worker.py:332] Compile and warming up model for size 8 2025-10-10T01:52:48.7043190Z (Worker_TP0 pid=62411) INFO 10-10 01:52:48 [backends.py:200] Cache the graph of shape 8 for later use 2025-10-10T01:52:48.7236743Z (Worker_TP1 pid=62412) INFO 10-10 01:52:48 [backends.py:200] Cache the graph of shape 8 for later use 2025-10-10T01:53:22.1102716Z (Worker_TP1 pid=62412) INFO 10-10 01:53:22 [backends.py:221] Compiling a graph for shape 8 takes 34.60 s 2025-10-10T01:53:22.7194833Z (Worker_TP0 pid=62411) INFO 10-10 01:53:22 [backends.py:221] Compiling a graph for shape 8 takes 35.21 s 2025-10-10T01:53:22.8035738Z (Worker_TP1 pid=62412) INFO 10-10 01:53:22 [gpu_worker.py:332] Compile and warming up model for size 4 2025-10-10T01:53:22.8038739Z (Worker_TP0 pid=62411) INFO 10-10 01:53:22 [gpu_worker.py:332] Compile and warming up model for size 4 2025-10-10T01:53:23.9748504Z (Worker_TP0 pid=62411) INFO 10-10 01:53:23 [backends.py:200] Cache the graph of shape 4 for later use 2025-10-10T01:53:23.9959283Z (Worker_TP1 pid=62412) INFO 10-10 01:53:23 [backends.py:200] Cache the graph of shape 4 for later use 2025-10-10T01:53:47.5143500Z (EngineCore_DP0 pid=62292) INFO 10-10 01:53:47 [shm_broadcast.py:466] No available shared memory broadcast block found in 60 seconds. This typically happens when some processes are hanging or doing some time-consuming work (e.g. compilation). 2025-10-10T01:53:59.0402148Z (Worker_TP0 pid=62411) INFO 10-10 01:53:59 [backends.py:221] Compiling a graph for shape 4 takes 36.24 s 2025-10-10T01:53:59.0422776Z (Worker_TP0 pid=62411) INFO 10-10 01:53:59 [monitor.py:32] torch.compile takes 107.23 s in total 2025-10-10T01:53:59.0876268Z (Worker_TP1 pid=62412) INFO 10-10 01:53:59 [backends.py:221] Compiling a graph for shape 4 takes 36.28 s 2025-10-10T01:53:59.0896400Z (Worker_TP1 pid=62412) INFO 10-10 01:53:59 [monitor.py:32] torch.compile takes 108.25 s in total 2025-10-10T01:53:59.1161362Z (EngineCore_DP0 pid=62292) INFO 10-10 01:53:59 [core.py:211] init engine (profile, create kv cache, warmup model) took 131.31 seconds 2025-10-10T01:53:59.6353296Z (EngineCore_DP0 pid=62292) INFO 10-10 01:53:59 [vllm.py:345] Cudagraph is disabled under eager mode 2025-10-10T01:53:59.8201150Z (APIServer pid=62168) INFO 10-10 01:53:59 [loggers.py:147] Engine 000: vllm cache_config_info with initialization after num_gpu_blocks is: 15658 2025-10-10T01:53:59.8620532Z (EngineCore_DP0 pid=62292) INFO 10-10 01:53:59 [gc_utils.py:41] GC Debug Config. enabled:False,top_objects:-1 2025-10-10T01:53:59.8841799Z (APIServer pid=62168) INFO 10-10 01:53:59 [api_server.py:1634] Supported_tasks: ['generate'] 2025-10-10T01:53:59.9319100Z (APIServer pid=62168) WARNING 10-10 01:53:59 [model.py:1424] Default sampling parameters have been overridden by the model's Hugging Face generation config recommended from the model creator. If this is not intended, please relaunch vLLM instance with `--generation-config vllm`. 2025-10-10T01:53:59.9320755Z (APIServer pid=62168) INFO 10-10 01:53:59 [serving_responses.py:140] Using default chat sampling params from model: {'temperature': 0.6, 'top_p': 0.9} 2025-10-10T01:53:59.9640866Z (APIServer pid=62168) INFO 10-10 01:53:59 [serving_chat.py:139] Using default chat sampling params from model: {'temperature': 0.6, 'top_p': 0.9} 2025-10-10T01:53:59.9997380Z (APIServer pid=62168) INFO 10-10 01:53:59 [serving_completion.py:76] Using default completion sampling params from model: {'temperature': 0.6, 'top_p': 0.9} 2025-10-10T01:53:59.9998825Z (APIServer pid=62168) INFO 10-10 01:53:59 [api_server.py:1912] Starting vLLM API server 0 on http://0.0.0.0:52647 2025-10-10T01:53:59.9999946Z (APIServer pid=62168) INFO 10-10 01:53:59 [launcher.py:34] Available routes are: 2025-10-10T01:54:00.0000896Z (APIServer pid=62168) INFO 10-10 01:53:59 [launcher.py:42] Route: /openapi.json, Methods: HEAD, GET 2025-10-10T01:54:00.0001629Z (APIServer pid=62168) INFO 10-10 01:53:59 [launcher.py:42] Route: /docs, Methods: HEAD, GET 2025-10-10T01:54:00.0002363Z (APIServer pid=62168) INFO 10-10 01:53:59 [launcher.py:42] Route: /docs/oauth2-redirect, Methods: HEAD, GET 2025-10-10T01:54:00.0003515Z (APIServer pid=62168) INFO 10-10 01:53:59 [launcher.py:42] Route: /redoc, Methods: HEAD, GET 2025-10-10T01:54:00.0004244Z (APIServer pid=62168) INFO 10-10 01:53:59 [launcher.py:42] Route: /health, Methods: GET 2025-10-10T01:54:00.0004932Z (APIServer pid=62168) INFO 10-10 01:53:59 [launcher.py:42] Route: /load, Methods: GET 2025-10-10T01:54:00.0005454Z (APIServer pid=62168) INFO 10-10 01:53:59 [launcher.py:42] Route: /ping, Methods: POST 2025-10-10T01:54:00.0005961Z (APIServer pid=62168) INFO 10-10 01:53:59 [launcher.py:42] Route: /ping, Methods: GET 2025-10-10T01:54:00.0006479Z (APIServer pid=62168) INFO 10-10 01:53:59 [launcher.py:42] Route: /tokenize, Methods: POST 2025-10-10T01:54:00.0007019Z (APIServer pid=62168) INFO 10-10 01:53:59 [launcher.py:42] Route: /detokenize, Methods: POST 2025-10-10T01:54:00.0007748Z (APIServer pid=62168) INFO 10-10 01:53:59 [launcher.py:42] Route: /v1/models, Methods: GET 2025-10-10T01:54:00.0008331Z (APIServer pid=62168) INFO 10-10 01:53:59 [launcher.py:42] Route: /version, Methods: GET 2025-10-10T01:54:00.0008904Z (APIServer pid=62168) INFO 10-10 01:53:59 [launcher.py:42] Route: /v1/responses, Methods: POST 2025-10-10T01:54:00.0009534Z (APIServer pid=62168) INFO 10-10 01:53:59 [launcher.py:42] Route: /v1/responses/{response_id}, Methods: GET 2025-10-10T01:54:00.0010266Z (APIServer pid=62168) INFO 10-10 01:53:59 [launcher.py:42] Route: /v1/responses/{response_id}/cancel, Methods: POST 2025-10-10T01:54:00.0010891Z (APIServer pid=62168) INFO 10-10 01:53:59 [launcher.py:42] Route: /v1/chat/completions, Methods: POST 2025-10-10T01:54:00.0011465Z (APIServer pid=62168) INFO 10-10 01:54:00 [launcher.py:42] Route: /v1/completions, Methods: POST 2025-10-10T01:54:00.0012015Z (APIServer pid=62168) INFO 10-10 01:54:00 [launcher.py:42] Route: /v1/embeddings, Methods: POST 2025-10-10T01:54:00.0012557Z (APIServer pid=62168) INFO 10-10 01:54:00 [launcher.py:42] Route: /pooling, Methods: POST 2025-10-10T01:54:00.0013089Z (APIServer pid=62168) INFO 10-10 01:54:00 [launcher.py:42] Route: /classify, Methods: POST 2025-10-10T01:54:00.0013611Z (APIServer pid=62168) INFO 10-10 01:54:00 [launcher.py:42] Route: /score, Methods: POST 2025-10-10T01:54:00.0014131Z (APIServer pid=62168) INFO 10-10 01:54:00 [launcher.py:42] Route: /v1/score, Methods: POST 2025-10-10T01:54:00.0014698Z (APIServer pid=62168) INFO 10-10 01:54:00 [launcher.py:42] Route: /v1/audio/transcriptions, Methods: POST 2025-10-10T01:54:00.0015333Z (APIServer pid=62168) INFO 10-10 01:54:00 [launcher.py:42] Route: /v1/audio/translations, Methods: POST 2025-10-10T01:54:00.0015887Z (APIServer pid=62168) INFO 10-10 01:54:00 [launcher.py:42] Route: /rerank, Methods: POST 2025-10-10T01:54:00.0016419Z (APIServer pid=62168) INFO 10-10 01:54:00 [launcher.py:42] Route: /v1/rerank, Methods: POST 2025-10-10T01:54:00.0016952Z (APIServer pid=62168) INFO 10-10 01:54:00 [launcher.py:42] Route: /v2/rerank, Methods: POST 2025-10-10T01:54:00.0017519Z (APIServer pid=62168) INFO 10-10 01:54:00 [launcher.py:42] Route: /scale_elastic_ep, Methods: POST 2025-10-10T01:54:00.0018126Z (APIServer pid=62168) INFO 10-10 01:54:00 [launcher.py:42] Route: /is_scaling_elastic_ep, Methods: POST 2025-10-10T01:54:00.0018714Z (APIServer pid=62168) INFO 10-10 01:54:00 [launcher.py:42] Route: /invocations, Methods: POST 2025-10-10T01:54:00.0019320Z (APIServer pid=62168) INFO 10-10 01:54:00 [launcher.py:42] Route: /metrics, Methods: GET 2025-10-10T01:54:00.0154893Z (APIServer pid=62168) INFO: Started server process [62168] 2025-10-10T01:54:00.0155523Z (APIServer pid=62168) INFO: Waiting for application startup. 2025-10-10T01:54:00.2475301Z (APIServer pid=62168) INFO: Application startup complete. 2025-10-10T01:54:00.4415315Z (APIServer pid=62168) INFO: 127.0.0.1:42508 - "GET /health HTTP/1.1" 200 OK 2025-10-10T01:54:00.6615567Z (APIServer pid=62168) INFO: 127.0.0.1:42520 - "GET /v1/models HTTP/1.1" 200 OK 2025-10-10T01:54:00.8254021Z (APIServer pid=62168) INFO: 127.0.0.1:42520 - "POST /v1/completions HTTP/1.1" 200 OK 2025-10-10T01:54:00.9425745Z (APIServer pid=62168) INFO: 127.0.0.1:42520 - "POST /v1/completions HTTP/1.1" 200 OK 2025-10-10T01:54:01.1259800Z (APIServer pid=62168) INFO: 127.0.0.1:42520 - "POST /v1/completions HTTP/1.1" 200 OK 2025-10-10T01:54:01.2799238Z (APIServer pid=62168) INFO: 127.0.0.1:42520 - "POST /v1/completions HTTP/1.1" 200 OK 2025-10-10T01:54:01.4212008Z (APIServer pid=62168) INFO: 127.0.0.1:42520 - "POST /v1/completions HTTP/1.1" 200 OK 2025-10-10T01:54:01.4288640Z (APIServer pid=62168) INFO: 127.0.0.1:42520 - "POST /v1/completions HTTP/1.1" 200 OK 2025-10-10T01:54:01.5499866Z (APIServer pid=62168) INFO 10-10 01:54:01 [launcher.py:99] Shutting down FastAPI HTTP server. 2025-10-10T01:54:01.5502875Z (Worker_TP0 pid=62411) INFO 10-10 01:54:01 [multiproc_executor.py:558] Parent process exited, terminating worker 2025-10-10T01:54:01.5504963Z (Worker_TP0 pid=62411) INFO 10-10 01:54:01 [multiproc_executor.py:599] WorkerProc shutting down. 2025-10-10T01:54:01.5509793Z (Worker_TP1 pid=62412) INFO 10-10 01:54:01 [multiproc_executor.py:558] Parent process exited, terminating worker 2025-10-10T01:54:04.8728477Z (APIServer pid=62168) INFO: Shutting down 2025-10-10T01:54:04.9735229Z (APIServer pid=62168) INFO: Waiting for application shutdown. 2025-10-10T01:54:04.9736556Z (APIServer pid=62168) INFO: Application shutdown complete. 2025-10-10T01:54:05.3673452Z INFO 10-10 01:54:05 [model.py:551] Resolved architecture: LlamaForCausalLM 2025-10-10T01:54:05.3674004Z WARNING 10-10 01:54:05 [model.py:1768] Casting torch.bfloat16 to torch.float16. 2025-10-10T01:54:05.3674480Z INFO 10-10 01:54:05 [model.py:1545] Using max model len 2048 2025-10-10T01:54:05.3679938Z Launching RemoteOpenAIServer with: vllm serve RedHatAI/Meta-Llama-3.1-8B-Instruct-FP8 --dtype float16 --max-model-len 2048 --max-num-seqs 8 --enforce-eager --tokenizer-mode auto --tensor-parallel-size 2 --distributed-executor-backend mp --load-format dummy --port 46331 --seed 0 2025-10-10T01:54:05.8827425Z /opt/conda/envs/py_3.12/lib/python3.12/site-packages/torch/cuda/__init__.py:64: FutureWarning: The pynvml package is deprecated. Please install nvidia-ml-py instead. If you did not install pynvml directly, please report this to the maintainers of the package that installed pynvml for you. 2025-10-10T01:54:05.8828659Z import pynvml # type: ignore[import] 2025-10-10T01:54:08.1014717Z INFO 10-10 01:54:08 [__init__.py:215] Automatically detected platform cuda. 2025-10-10T01:54:12.0256708Z (APIServer pid=66473) INFO 10-10 01:54:12 [api_server.py:1839] vLLM API server version 0.11.0rc2.dev157+g0ad9951c4.d20251010 2025-10-10T01:54:12.0279450Z (APIServer pid=66473) INFO 10-10 01:54:12 [utils.py:233] non-default args: {'model_tag': 'RedHatAI/Meta-Llama-3.1-8B-Instruct-FP8', 'port': 46331, 'model': 'RedHatAI/Meta-Llama-3.1-8B-Instruct-FP8', 'dtype': 'float16', 'seed': 0, 'max_model_len': 2048, 'enforce_eager': True, 'load_format': 'dummy', 'distributed_executor_backend': 'mp', 'tensor_parallel_size': 2, 'max_num_seqs': 8} 2025-10-10T01:54:12.2303568Z (APIServer pid=66473) INFO 10-10 01:54:12 [model.py:551] Resolved architecture: LlamaForCausalLM 2025-10-10T01:54:12.2306099Z (APIServer pid=66473) `torch_dtype` is deprecated! Use `dtype` instead! 2025-10-10T01:54:12.2306894Z (APIServer pid=66473) WARNING 10-10 01:54:12 [model.py:1768] Casting torch.bfloat16 to torch.float16. 2025-10-10T01:54:12.2307585Z (APIServer pid=66473) INFO 10-10 01:54:12 [model.py:1545] Using max model len 2048 2025-10-10T01:54:12.5631642Z (APIServer pid=66473) INFO 10-10 01:54:12 [scheduler.py:222] Chunked prefill is enabled with max_num_batched_tokens=2048. 2025-10-10T01:54:12.6002596Z (APIServer pid=66473) INFO 10-10 01:54:12 [vllm.py:345] Cudagraph is disabled under eager mode 2025-10-10T01:54:13.7101220Z /opt/conda/envs/py_3.12/lib/python3.12/site-packages/torch/cuda/__init__.py:64: FutureWarning: The pynvml package is deprecated. Please install nvidia-ml-py instead. If you did not install pynvml directly, please report this to the maintainers of the package that installed pynvml for you. 2025-10-10T01:54:13.7102493Z import pynvml # type: ignore[import] 2025-10-10T01:54:16.0038906Z INFO 10-10 01:54:16 [__init__.py:215] Automatically detected platform cuda. 2025-10-10T01:54:20.1022043Z (EngineCore_DP0 pid=66596) INFO 10-10 01:54:20 [core.py:648] Waiting for init message from front-end. 2025-10-10T01:54:20.1292233Z (EngineCore_DP0 pid=66596) INFO 10-10 01:54:20 [core.py:78] Initializing a V1 LLM engine (v0.11.0rc2.dev157+g0ad9951c4.d20251010) with config: model='RedHatAI/Meta-Llama-3.1-8B-Instruct-FP8', speculative_config=None, tokenizer='RedHatAI/Meta-Llama-3.1-8B-Instruct-FP8', skip_tokenizer_init=False, tokenizer_mode=auto, revision=None, tokenizer_revision=None, trust_remote_code=False, dtype=torch.float16, max_seq_len=2048, download_dir=None, load_format=dummy, tensor_parallel_size=2, pipeline_parallel_size=1, data_parallel_size=1, disable_custom_all_reduce=False, quantization=compressed-tensors, enforce_eager=True, kv_cache_dtype=auto, device_config=cuda, structured_outputs_config=StructuredOutputsConfig(backend='auto', disable_fallback=False, disable_any_whitespace=False, disable_additional_properties=False, reasoning_parser=''), observability_config=ObservabilityConfig(show_hidden_metrics_for_version=None, otlp_traces_endpoint=None, collect_detailed_traces=None), seed=0, served_model_name=RedHatAI/Meta-Llama-3.1-8B-Instruct-FP8, enable_prefix_caching=True, chunked_prefill_enabled=True, pooler_config=None, compilation_config={'level': 0, 'debug_dump_path': None, 'cache_dir': '', 'backend': '', 'custom_ops': [], 'splitting_ops': None, 'use_inductor': True, 'compile_sizes': [], 'inductor_compile_config': {'enable_auto_functionalized_v2': False}, 'inductor_passes': {}, 'cudagraph_mode': , 'use_cudagraph': False, 'cudagraph_num_of_warmups': 0, 'cudagraph_capture_sizes': [], 'cudagraph_copy_inputs': False, 'full_cuda_graph': False, 'use_inductor_graph_partition': False, 'pass_config': {}, 'max_capture_size': 0, 'local_cache_dir': None} 2025-10-10T01:54:20.1297839Z (EngineCore_DP0 pid=66596) WARNING 10-10 01:54:20 [multiproc_executor.py:720] Reducing Torch parallelism from 24 threads to 1 to avoid unnecessary CPU contention. Set OMP_NUM_THREADS in the external environment to tune this value as needed. 2025-10-10T01:54:20.1302850Z (EngineCore_DP0 pid=66596) INFO 10-10 01:54:20 [shm_broadcast.py:289] vLLM message queue communication handle: Handle(local_reader_ranks=[0, 1], buffer_handle=(2, 16777216, 10, 'psm_3f3d3839'), local_subscribe_addr='ipc:///tmp/6a38c46b-9eba-42c8-936e-d8af2260deed', remote_subscribe_addr=None, remote_addr_ipv6=False) 2025-10-10T01:54:20.6373706Z /opt/conda/envs/py_3.12/lib/python3.12/site-packages/torch/cuda/__init__.py:64: FutureWarning: The pynvml package is deprecated. Please install nvidia-ml-py instead. If you did not install pynvml directly, please report this to the maintainers of the package that installed pynvml for you. 2025-10-10T01:54:20.6374990Z import pynvml # type: ignore[import] 2025-10-10T01:54:20.6537731Z /opt/conda/envs/py_3.12/lib/python3.12/site-packages/torch/cuda/__init__.py:64: FutureWarning: The pynvml package is deprecated. Please install nvidia-ml-py instead. If you did not install pynvml directly, please report this to the maintainers of the package that installed pynvml for you. 2025-10-10T01:54:20.6538935Z import pynvml # type: ignore[import] 2025-10-10T01:54:22.9648665Z INFO 10-10 01:54:22 [__init__.py:215] Automatically detected platform cuda. 2025-10-10T01:54:22.9797856Z INFO 10-10 01:54:22 [__init__.py:215] Automatically detected platform cuda. 2025-10-10T01:54:27.5440326Z INFO 10-10 01:54:27 [shm_broadcast.py:289] vLLM message queue communication handle: Handle(local_reader_ranks=[0], buffer_handle=(1, 10485760, 10, 'psm_23a6c275'), local_subscribe_addr='ipc:///tmp/6b21eb48-14ab-4f26-bfd4-bb4875f03e6c', remote_subscribe_addr=None, remote_addr_ipv6=False) 2025-10-10T01:54:27.6240788Z INFO 10-10 01:54:27 [shm_broadcast.py:289] vLLM message queue communication handle: Handle(local_reader_ranks=[0], buffer_handle=(1, 10485760, 10, 'psm_78017392'), local_subscribe_addr='ipc:///tmp/c0fcd89c-4488-42cb-8e71-3a6c4161f4a2', remote_subscribe_addr=None, remote_addr_ipv6=False) 2025-10-10T01:54:28.3032936Z [Gloo] Rank 0 is connected to 1 peer ranks. Expected number of connected peer ranks is : 1 2025-10-10T01:54:28.3034167Z [Gloo] Rank 1 is connected to 1 peer ranks. Expected number of connected peer ranks is : 1 2025-10-10T01:54:28.3075209Z [Gloo] Rank 1 is connected to 1 peer ranks. Expected number of connected peer ranks is : 1 2025-10-10T01:54:28.3076387Z [Gloo] Rank 0 is connected to 1 peer ranks. Expected number of connected peer ranks is : 1 2025-10-10T01:54:28.3117267Z INFO 10-10 01:54:28 [__init__.py:1384] Found nccl from library libnccl.so.2 2025-10-10T01:54:28.3118162Z INFO 10-10 01:54:28 [__init__.py:1384] Found nccl from library libnccl.so.2 2025-10-10T01:54:28.3119580Z INFO 10-10 01:54:28 [pynccl.py:104] vLLM is using nccl==2.27.5 2025-10-10T01:54:28.3121169Z INFO 10-10 01:54:28 [pynccl.py:104] vLLM is using nccl==2.27.5 2025-10-10T01:54:28.4849189Z WARNING 10-10 01:54:28 [symm_mem.py:58] SymmMemCommunicator: Device capability 8.9 not supported, communicator is not available. 2025-10-10T01:54:28.4851097Z WARNING 10-10 01:54:28 [symm_mem.py:58] SymmMemCommunicator: Device capability 8.9 not supported, communicator is not available. 2025-10-10T01:54:28.4872482Z INFO 10-10 01:54:28 [custom_all_reduce.py:35] Skipping P2P check and trusting the driver's P2P report. 2025-10-10T01:54:28.4873368Z WARNING 10-10 01:54:28 [custom_all_reduce.py:154] Custom allreduce is disabled because your platform lacks GPU P2P capability or P2P test failed. To silence this warning, specify disable_custom_all_reduce=True explicitly. 2025-10-10T01:54:28.4874217Z INFO 10-10 01:54:28 [custom_all_reduce.py:35] Skipping P2P check and trusting the driver's P2P report. 2025-10-10T01:54:28.4875074Z WARNING 10-10 01:54:28 [custom_all_reduce.py:154] Custom allreduce is disabled because your platform lacks GPU P2P capability or P2P test failed. To silence this warning, specify disable_custom_all_reduce=True explicitly. 2025-10-10T01:54:28.5940291Z INFO 10-10 01:54:28 [shm_broadcast.py:289] vLLM message queue communication handle: Handle(local_reader_ranks=[1], buffer_handle=(1, 4194304, 6, 'psm_8e00a4ed'), local_subscribe_addr='ipc:///tmp/3654f4e6-836a-4a45-abaf-5b5c6ce5c38e', remote_subscribe_addr=None, remote_addr_ipv6=False) 2025-10-10T01:54:28.5966818Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-10-10T01:54:28.5967317Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-10-10T01:54:28.5975816Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-10-10T01:54:28.5987293Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-10-10T01:54:28.6737175Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-10-10T01:54:28.6744119Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-10-10T01:54:28.6754556Z [Gloo] Rank 0 is connected to 1 peer ranks. Expected number of connected peer ranks is : 1 2025-10-10T01:54:28.6755049Z [Gloo] Rank 1 is connected to 1 peer ranks. Expected number of connected peer ranks is : 1 2025-10-10T01:54:28.6756996Z INFO 10-10 01:54:28 [__init__.py:1384] Found nccl from library libnccl.so.2 2025-10-10T01:54:28.6757383Z INFO 10-10 01:54:28 [__init__.py:1384] Found nccl from library libnccl.so.2 2025-10-10T01:54:28.6757729Z INFO 10-10 01:54:28 [pynccl.py:104] vLLM is using nccl==2.27.5 2025-10-10T01:54:28.6758036Z INFO 10-10 01:54:28 [pynccl.py:104] vLLM is using nccl==2.27.5 2025-10-10T01:54:28.7099907Z INFO 10-10 01:54:28 [parallel_state.py:1208] rank 1 in world size 2 is assigned as DP rank 0, PP rank 0, TP rank 1, EP rank 1 2025-10-10T01:54:28.7100986Z INFO 10-10 01:54:28 [parallel_state.py:1208] rank 0 in world size 2 is assigned as DP rank 0, PP rank 0, TP rank 0, EP rank 0 2025-10-10T01:54:28.9488003Z INFO 10-10 01:54:28 [topk_topp_sampler.py:55] Using FlashInfer for top-p & top-k sampling. 2025-10-10T01:54:28.9543876Z INFO 10-10 01:54:28 [topk_topp_sampler.py:55] Using FlashInfer for top-p & top-k sampling. 2025-10-10T01:54:28.9846582Z (Worker_TP1 pid=66716) INFO 10-10 01:54:28 [gpu_model_runner.py:2707] Starting to load model RedHatAI/Meta-Llama-3.1-8B-Instruct-FP8... 2025-10-10T01:54:28.9902570Z (Worker_TP0 pid=66715) INFO 10-10 01:54:28 [gpu_model_runner.py:2707] Starting to load model RedHatAI/Meta-Llama-3.1-8B-Instruct-FP8... 2025-10-10T01:54:29.2080766Z (Worker_TP1 pid=66716) INFO 10-10 01:54:29 [gpu_model_runner.py:2739] Loading model from scratch... 2025-10-10T01:54:29.2204883Z (Worker_TP0 pid=66715) INFO 10-10 01:54:29 [gpu_model_runner.py:2739] Loading model from scratch... 2025-10-10T01:54:29.2994023Z (Worker_TP1 pid=66716) INFO 10-10 01:54:29 [cuda.py:361] Using Flash Attention backend on V1 engine. 2025-10-10T01:54:29.3118849Z (Worker_TP0 pid=66715) INFO 10-10 01:54:29 [cuda.py:361] Using Flash Attention backend on V1 engine. 2025-10-10T01:54:30.1532057Z (Worker_TP1 pid=66716) INFO 10-10 01:54:30 [gpu_model_runner.py:2758] Model loading took 4.2624 GiB and 0.490438 seconds 2025-10-10T01:54:30.1763742Z (Worker_TP0 pid=66715) INFO 10-10 01:54:30 [gpu_model_runner.py:2758] Model loading took 4.2624 GiB and 0.490396 seconds 2025-10-10T01:54:35.6052727Z (Worker_TP0 pid=66715) INFO 10-10 01:54:35 [gpu_worker.py:298] Available KV cache memory: 15.32 GiB 2025-10-10T01:54:35.6093374Z (Worker_TP1 pid=66716) INFO 10-10 01:54:35 [gpu_worker.py:298] Available KV cache memory: 15.32 GiB 2025-10-10T01:54:35.8449611Z (EngineCore_DP0 pid=66596) INFO 10-10 01:54:35 [kv_cache_utils.py:1087] GPU KV cache size: 250,976 tokens 2025-10-10T01:54:35.8450626Z (EngineCore_DP0 pid=66596) INFO 10-10 01:54:35 [kv_cache_utils.py:1091] Maximum concurrency for 2,048 tokens per request: 122.55x 2025-10-10T01:54:35.8451544Z (EngineCore_DP0 pid=66596) INFO 10-10 01:54:35 [kv_cache_utils.py:1087] GPU KV cache size: 250,976 tokens 2025-10-10T01:54:35.8452550Z (EngineCore_DP0 pid=66596) INFO 10-10 01:54:35 [kv_cache_utils.py:1091] Maximum concurrency for 2,048 tokens per request: 122.55x 2025-10-10T01:54:35.9764091Z (EngineCore_DP0 pid=66596) INFO 10-10 01:54:35 [core.py:211] init engine (profile, create kv cache, warmup model) took 5.72 seconds 2025-10-10T01:54:36.5366698Z (EngineCore_DP0 pid=66596) INFO 10-10 01:54:36 [vllm.py:345] Cudagraph is disabled under eager mode 2025-10-10T01:54:36.7216181Z (APIServer pid=66473) INFO 10-10 01:54:36 [loggers.py:147] Engine 000: vllm cache_config_info with initialization after num_gpu_blocks is: 15686 2025-10-10T01:54:36.7498356Z (EngineCore_DP0 pid=66596) INFO 10-10 01:54:36 [gc_utils.py:41] GC Debug Config. enabled:False,top_objects:-1 2025-10-10T01:54:36.7716623Z (APIServer pid=66473) INFO 10-10 01:54:36 [api_server.py:1634] Supported_tasks: ['generate'] 2025-10-10T01:54:36.8006668Z (APIServer pid=66473) WARNING 10-10 01:54:36 [model.py:1424] Default sampling parameters have been overridden by the model's Hugging Face generation config recommended from the model creator. If this is not intended, please relaunch vLLM instance with `--generation-config vllm`. 2025-10-10T01:54:36.8008313Z (APIServer pid=66473) INFO 10-10 01:54:36 [serving_responses.py:140] Using default chat sampling params from model: {'temperature': 0.6, 'top_p': 0.9} 2025-10-10T01:54:36.8343907Z (APIServer pid=66473) INFO 10-10 01:54:36 [serving_chat.py:139] Using default chat sampling params from model: {'temperature': 0.6, 'top_p': 0.9} 2025-10-10T01:54:36.8690934Z (APIServer pid=66473) INFO 10-10 01:54:36 [serving_completion.py:76] Using default completion sampling params from model: {'temperature': 0.6, 'top_p': 0.9} 2025-10-10T01:54:36.8692163Z (APIServer pid=66473) INFO 10-10 01:54:36 [api_server.py:1912] Starting vLLM API server 0 on http://0.0.0.0:46331 2025-10-10T01:54:36.8693070Z (APIServer pid=66473) INFO 10-10 01:54:36 [launcher.py:34] Available routes are: 2025-10-10T01:54:36.8693853Z (APIServer pid=66473) INFO 10-10 01:54:36 [launcher.py:42] Route: /openapi.json, Methods: GET, HEAD 2025-10-10T01:54:36.8694674Z (APIServer pid=66473) INFO 10-10 01:54:36 [launcher.py:42] Route: /docs, Methods: GET, HEAD 2025-10-10T01:54:36.8695412Z (APIServer pid=66473) INFO 10-10 01:54:36 [launcher.py:42] Route: /docs/oauth2-redirect, Methods: GET, HEAD 2025-10-10T01:54:36.8696142Z (APIServer pid=66473) INFO 10-10 01:54:36 [launcher.py:42] Route: /redoc, Methods: GET, HEAD 2025-10-10T01:54:36.8696801Z (APIServer pid=66473) INFO 10-10 01:54:36 [launcher.py:42] Route: /health, Methods: GET 2025-10-10T01:54:36.8697441Z (APIServer pid=66473) INFO 10-10 01:54:36 [launcher.py:42] Route: /load, Methods: GET 2025-10-10T01:54:36.8698068Z (APIServer pid=66473) INFO 10-10 01:54:36 [launcher.py:42] Route: /ping, Methods: POST 2025-10-10T01:54:36.8698739Z (APIServer pid=66473) INFO 10-10 01:54:36 [launcher.py:42] Route: /ping, Methods: GET 2025-10-10T01:54:36.8699392Z (APIServer pid=66473) INFO 10-10 01:54:36 [launcher.py:42] Route: /tokenize, Methods: POST 2025-10-10T01:54:36.8700062Z (APIServer pid=66473) INFO 10-10 01:54:36 [launcher.py:42] Route: /detokenize, Methods: POST 2025-10-10T01:54:36.8700721Z (APIServer pid=66473) INFO 10-10 01:54:36 [launcher.py:42] Route: /v1/models, Methods: GET 2025-10-10T01:54:36.8701359Z (APIServer pid=66473) INFO 10-10 01:54:36 [launcher.py:42] Route: /version, Methods: GET 2025-10-10T01:54:36.8702029Z (APIServer pid=66473) INFO 10-10 01:54:36 [launcher.py:42] Route: /v1/responses, Methods: POST 2025-10-10T01:54:36.8702707Z (APIServer pid=66473) INFO 10-10 01:54:36 [launcher.py:42] Route: /v1/responses/{response_id}, Methods: GET 2025-10-10T01:54:36.8703349Z (APIServer pid=66473) INFO 10-10 01:54:36 [launcher.py:42] Route: /v1/responses/{response_id}/cancel, Methods: POST 2025-10-10T01:54:36.8703972Z (APIServer pid=66473) INFO 10-10 01:54:36 [launcher.py:42] Route: /v1/chat/completions, Methods: POST 2025-10-10T01:54:36.8704553Z (APIServer pid=66473) INFO 10-10 01:54:36 [launcher.py:42] Route: /v1/completions, Methods: POST 2025-10-10T01:54:36.8705102Z (APIServer pid=66473) INFO 10-10 01:54:36 [launcher.py:42] Route: /v1/embeddings, Methods: POST 2025-10-10T01:54:36.8705641Z (APIServer pid=66473) INFO 10-10 01:54:36 [launcher.py:42] Route: /pooling, Methods: POST 2025-10-10T01:54:36.8706189Z (APIServer pid=66473) INFO 10-10 01:54:36 [launcher.py:42] Route: /classify, Methods: POST 2025-10-10T01:54:36.8706788Z (APIServer pid=66473) INFO 10-10 01:54:36 [launcher.py:42] Route: /score, Methods: POST 2025-10-10T01:54:36.8707311Z (APIServer pid=66473) INFO 10-10 01:54:36 [launcher.py:42] Route: /v1/score, Methods: POST 2025-10-10T01:54:36.8707876Z (APIServer pid=66473) INFO 10-10 01:54:36 [launcher.py:42] Route: /v1/audio/transcriptions, Methods: POST 2025-10-10T01:54:36.8708479Z (APIServer pid=66473) INFO 10-10 01:54:36 [launcher.py:42] Route: /v1/audio/translations, Methods: POST 2025-10-10T01:54:36.8709028Z (APIServer pid=66473) INFO 10-10 01:54:36 [launcher.py:42] Route: /rerank, Methods: POST 2025-10-10T01:54:36.8709561Z (APIServer pid=66473) INFO 10-10 01:54:36 [launcher.py:42] Route: /v1/rerank, Methods: POST 2025-10-10T01:54:36.8710096Z (APIServer pid=66473) INFO 10-10 01:54:36 [launcher.py:42] Route: /v2/rerank, Methods: POST 2025-10-10T01:54:36.8710722Z (APIServer pid=66473) INFO 10-10 01:54:36 [launcher.py:42] Route: /scale_elastic_ep, Methods: POST 2025-10-10T01:54:36.8711311Z (APIServer pid=66473) INFO 10-10 01:54:36 [launcher.py:42] Route: /is_scaling_elastic_ep, Methods: POST 2025-10-10T01:54:36.8711885Z (APIServer pid=66473) INFO 10-10 01:54:36 [launcher.py:42] Route: /invocations, Methods: POST 2025-10-10T01:54:36.8712428Z (APIServer pid=66473) INFO 10-10 01:54:36 [launcher.py:42] Route: /metrics, Methods: GET 2025-10-10T01:54:36.8852504Z (APIServer pid=66473) INFO: Started server process [66473] 2025-10-10T01:54:36.8853142Z (APIServer pid=66473) INFO: Waiting for application startup. 2025-10-10T01:54:37.1218428Z (APIServer pid=66473) INFO: Application startup complete. 2025-10-10T01:54:37.4493740Z (APIServer pid=66473) INFO: 127.0.0.1:58950 - "GET /health HTTP/1.1" 200 OK 2025-10-10T01:54:37.4587832Z (APIServer pid=66473) INFO: 127.0.0.1:58960 - "GET /v1/models HTTP/1.1" 200 OK 2025-10-10T01:54:37.6845875Z (APIServer pid=66473) INFO: 127.0.0.1:58960 - "POST /v1/completions HTTP/1.1" 200 OK 2025-10-10T01:54:37.7973159Z (APIServer pid=66473) INFO: 127.0.0.1:58960 - "POST /v1/completions HTTP/1.1" 200 OK 2025-10-10T01:54:37.9386485Z (APIServer pid=66473) INFO: 127.0.0.1:58960 - "POST /v1/completions HTTP/1.1" 200 OK 2025-10-10T01:54:38.0865769Z (APIServer pid=66473) INFO: 127.0.0.1:58960 - "POST /v1/completions HTTP/1.1" 200 OK 2025-10-10T01:54:38.2264351Z (APIServer pid=66473) INFO: 127.0.0.1:58960 - "POST /v1/completions HTTP/1.1" 200 OK 2025-10-10T01:54:38.2339473Z (APIServer pid=66473) INFO: 127.0.0.1:58960 - "POST /v1/completions HTTP/1.1" 200 OK 2025-10-10T01:54:38.3782955Z (APIServer pid=66473) INFO 10-10 01:54:38 [launcher.py:99] Shutting down FastAPI HTTP server. 2025-10-10T01:54:38.3786991Z (Worker_TP0 pid=66715) INFO 10-10 01:54:38 [multiproc_executor.py:558] Parent process exited, terminating worker 2025-10-10T01:54:38.3788935Z (Worker_TP0 pid=66715) INFO 10-10 01:54:38 [multiproc_executor.py:599] WorkerProc shutting down. 2025-10-10T01:54:38.3790506Z (Worker_TP1 pid=66716) INFO 10-10 01:54:38 [multiproc_executor.py:558] Parent process exited, terminating worker 2025-10-10T01:54:38.3792649Z (Worker_TP1 pid=66716) INFO 10-10 01:54:38 [multiproc_executor.py:599] WorkerProc shutting down. 2025-10-10T01:54:40.4995740Z (APIServer pid=66473) INFO: Shutting down 2025-10-10T01:54:40.6000975Z (APIServer pid=66473) INFO: Waiting for application shutdown. 2025-10-10T01:54:40.6002253Z (APIServer pid=66473) INFO: Application shutdown complete. 2025-10-10T01:54:41.2265193Z PASSED 2025-10-10T01:54:41.2407400Z distributed/test_sequence_parallel.py::test_tp_sp_generation[RedHatAI/Meta-Llama-3.1-8B-Instruct-FP8-parallel_setup17-ray-1-auto-test_options17] Fork a new process to run a test 67233 2025-10-10T01:54:41.2420464Z Fork a new process to run a test 0 2025-10-10T01:54:41.8806346Z INFO 10-10 01:54:41 [model.py:551] Resolved architecture: LlamaForCausalLM 2025-10-10T01:54:41.8807219Z `torch_dtype` is deprecated! Use `dtype` instead! 2025-10-10T01:54:41.8808057Z WARNING 10-10 01:54:41 [model.py:1768] Casting torch.bfloat16 to torch.float16. 2025-10-10T01:54:41.8808929Z INFO 10-10 01:54:41 [model.py:1545] Using max model len 2048 2025-10-10T01:54:41.9849837Z Launching RemoteOpenAIServer with: vllm serve RedHatAI/Meta-Llama-3.1-8B-Instruct-FP8 --dtype float16 --max-model-len 2048 --max-num-seqs 8 --enforce-eager --tokenizer-mode auto --tensor-parallel-size 2 --pipeline-parallel-size 1 --distributed-executor-backend ray --compilation_config {"level": 3, "custom_ops": ["+rms_norm"], "compile_sizes": [4, 8], "pass_config": {"enable_sequence_parallelism": true, "enable_fusion": false, "enable_noop": true}} --load-format dummy --port 48041 --seed 0 2025-10-10T01:54:42.4941305Z /opt/conda/envs/py_3.12/lib/python3.12/site-packages/torch/cuda/__init__.py:64: FutureWarning: The pynvml package is deprecated. Please install nvidia-ml-py instead. If you did not install pynvml directly, please report this to the maintainers of the package that installed pynvml for you. 2025-10-10T01:54:42.4943890Z import pynvml # type: ignore[import] 2025-10-10T01:54:44.7009858Z INFO 10-10 01:54:44 [__init__.py:215] Automatically detected platform cuda. 2025-10-10T01:54:48.6058539Z (APIServer pid=67282) INFO 10-10 01:54:48 [api_server.py:1839] vLLM API server version 0.11.0rc2.dev157+g0ad9951c4.d20251010 2025-10-10T01:54:48.6084605Z (APIServer pid=67282) INFO 10-10 01:54:48 [utils.py:233] non-default args: {'model_tag': 'RedHatAI/Meta-Llama-3.1-8B-Instruct-FP8', 'port': 48041, 'model': 'RedHatAI/Meta-Llama-3.1-8B-Instruct-FP8', 'dtype': 'float16', 'seed': 0, 'max_model_len': 2048, 'enforce_eager': True, 'load_format': 'dummy', 'distributed_executor_backend': 'ray', 'tensor_parallel_size': 2, 'max_num_seqs': 8, 'compilation_config': {'level': 3, 'debug_dump_path': None, 'cache_dir': '', 'backend': '', 'custom_ops': ['+rms_norm'], 'splitting_ops': None, 'use_inductor': True, 'compile_sizes': [4, 8], 'inductor_compile_config': {'enable_auto_functionalized_v2': False}, 'inductor_passes': {}, 'cudagraph_mode': None, 'use_cudagraph': True, 'cudagraph_num_of_warmups': 0, 'cudagraph_capture_sizes': None, 'cudagraph_copy_inputs': False, 'full_cuda_graph': False, 'use_inductor_graph_partition': False, 'pass_config': {'enable_noop': True, 'enable_sequence_parallelism': True}, 'max_capture_size': None, 'local_cache_dir': None}} 2025-10-10T01:54:48.7971591Z (APIServer pid=67282) INFO 10-10 01:54:48 [model.py:551] Resolved architecture: LlamaForCausalLM 2025-10-10T01:54:48.7972997Z (APIServer pid=67282) `torch_dtype` is deprecated! Use `dtype` instead! 2025-10-10T01:54:48.7974383Z (APIServer pid=67282) WARNING 10-10 01:54:48 [model.py:1768] Casting torch.bfloat16 to torch.float16. 2025-10-10T01:54:48.7975761Z (APIServer pid=67282) INFO 10-10 01:54:48 [model.py:1545] Using max model len 2048 2025-10-10T01:54:49.1271271Z (APIServer pid=67282) INFO 10-10 01:54:49 [scheduler.py:222] Chunked prefill is enabled with max_num_batched_tokens=2048. 2025-10-10T01:54:49.1633086Z (APIServer pid=67282) INFO 10-10 01:54:49 [vllm.py:345] Cudagraph is disabled under eager mode 2025-10-10T01:54:50.1641317Z /opt/conda/envs/py_3.12/lib/python3.12/site-packages/torch/cuda/__init__.py:64: FutureWarning: The pynvml package is deprecated. Please install nvidia-ml-py instead. If you did not install pynvml directly, please report this to the maintainers of the package that installed pynvml for you. 2025-10-10T01:54:50.1644121Z import pynvml # type: ignore[import] 2025-10-10T01:54:52.4471989Z INFO 10-10 01:54:52 [__init__.py:215] Automatically detected platform cuda. 2025-10-10T01:54:56.6662545Z (EngineCore_DP0 pid=67405) INFO 10-10 01:54:56 [core.py:648] Waiting for init message from front-end. 2025-10-10T01:54:56.6929885Z (EngineCore_DP0 pid=67405) INFO 10-10 01:54:56 [core.py:78] Initializing a V1 LLM engine (v0.11.0rc2.dev157+g0ad9951c4.d20251010) with config: model='RedHatAI/Meta-Llama-3.1-8B-Instruct-FP8', speculative_config=None, tokenizer='RedHatAI/Meta-Llama-3.1-8B-Instruct-FP8', skip_tokenizer_init=False, tokenizer_mode=auto, revision=None, tokenizer_revision=None, trust_remote_code=False, dtype=torch.float16, max_seq_len=2048, download_dir=None, load_format=dummy, tensor_parallel_size=2, pipeline_parallel_size=1, data_parallel_size=1, disable_custom_all_reduce=False, quantization=compressed-tensors, enforce_eager=True, kv_cache_dtype=auto, device_config=cuda, structured_outputs_config=StructuredOutputsConfig(backend='auto', disable_fallback=False, disable_any_whitespace=False, disable_additional_properties=False, reasoning_parser=''), observability_config=ObservabilityConfig(show_hidden_metrics_for_version=None, otlp_traces_endpoint=None, collect_detailed_traces=None), seed=0, served_model_name=RedHatAI/Meta-Llama-3.1-8B-Instruct-FP8, enable_prefix_caching=True, chunked_prefill_enabled=True, pooler_config=None, compilation_config={'level': 3, 'debug_dump_path': None, 'cache_dir': '', 'backend': '', 'custom_ops': ['+rms_norm', '+rms_norm'], 'splitting_ops': ['vllm.unified_attention', 'vllm.unified_attention_with_output', 'vllm.mamba_mixer2', 'vllm.mamba_mixer', 'vllm.short_conv', 'vllm.linear_attention', 'vllm.plamo2_mamba_mixer', 'vllm.gdn_attention', 'vllm.sparse_attn_indexer'], 'use_inductor': True, 'compile_sizes': [8, 4], 'inductor_compile_config': {'enable_auto_functionalized_v2': False}, 'inductor_passes': {}, 'cudagraph_mode': , 'use_cudagraph': False, 'cudagraph_num_of_warmups': 0, 'cudagraph_capture_sizes': [], 'cudagraph_copy_inputs': False, 'full_cuda_graph': False, 'use_inductor_graph_partition': False, 'pass_config': {'enable_noop': True, 'enable_sequence_parallelism': True}, 'max_capture_size': 0, 'local_cache_dir': None} 2025-10-10T01:54:58.6036923Z (EngineCore_DP0 pid=67405) 2025-10-10 01:54:58,602 INFO worker.py:1918 -- Started a local Ray instance. View the dashboard at http://127.0.0.1:8265  2025-10-10T01:54:59.4547216Z (EngineCore_DP0 pid=67405) INFO 10-10 01:54:59 [ray_utils.py:345] No current placement group found. Creating a new placement group. 2025-10-10T01:54:59.6264299Z (EngineCore_DP0 pid=67405) INFO 10-10 01:54:59 [ray_distributed_executor.py:171] use_ray_spmd_worker: True 2025-10-10T01:55:00.1727997Z (EngineCore_DP0 pid=67405) (pid=68289) /opt/conda/envs/py_3.12/lib/python3.12/site-packages/torch/cuda/__init__.py:64: FutureWarning: The pynvml package is deprecated. Please install nvidia-ml-py instead. If you did not install pynvml directly, please report this to the maintainers of the package that installed pynvml for you. 2025-10-10T01:55:00.1729641Z (EngineCore_DP0 pid=67405) (pid=68289) import pynvml # type: ignore[import] 2025-10-10T01:55:02.4188773Z (EngineCore_DP0 pid=67405) (pid=68289) INFO 10-10 01:55:02 [__init__.py:215] Automatically detected platform cuda. 2025-10-10T01:55:05.4195454Z (EngineCore_DP0 pid=67405) INFO 10-10 01:55:05 [ray_env.py:63] RAY_NON_CARRY_OVER_ENV_VARS from config: set() 2025-10-10T01:55:05.4197752Z (EngineCore_DP0 pid=67405) INFO 10-10 01:55:05 [ray_env.py:65] Copying the following environment variables to workers: ['VLLM_USE_RAY_COMPILED_DAG', 'VLLM_USE_V1', 'HUGGING_FACE_HUB_TOKEN', 'LD_LIBRARY_PATH', 'VLLM_WORKER_MULTIPROC_METHOD', 'HF_TOKEN', 'VLLM_USE_RAY_SPMD_WORKER', 'MAX_JOBS'] 2025-10-10T01:55:05.4200351Z (EngineCore_DP0 pid=67405) INFO 10-10 01:55:05 [ray_env.py:68] If certain env vars should NOT be copied, add them to /var/lib/jenkins/.config/vllm/ray_non_carry_over_env_vars.json file 2025-10-10T01:55:07.3178757Z (EngineCore_DP0 pid=67405) (RayWorkerWrapper pid=68289) [Gloo] Rank 1 is connected to 1 peer ranks. Expected number of connected peer ranks is : 1 2025-10-10T01:55:07.3179771Z (EngineCore_DP0 pid=67405) (RayWorkerWrapper pid=68289) [Gloo] Rank 1 is connected to 1 peer ranks. Expected number of connected peer ranks is : 1 2025-10-10T01:55:07.3180622Z (EngineCore_DP0 pid=67405) (RayWorkerWrapper pid=68289) INFO 10-10 01:55:07 [__init__.py:1384] Found nccl from library libnccl.so.2 2025-10-10T01:55:07.3181386Z (EngineCore_DP0 pid=67405) (RayWorkerWrapper pid=68289) INFO 10-10 01:55:07 [pynccl.py:104] vLLM is using nccl==2.27.5 2025-10-10T01:55:07.5263907Z (EngineCore_DP0 pid=67405) (RayWorkerWrapper pid=68289) WARNING 10-10 01:55:07 [symm_mem.py:58] SymmMemCommunicator: Device capability 8.9 not supported, communicator is not available. 2025-10-10T01:55:07.5265351Z (EngineCore_DP0 pid=67405) (RayWorkerWrapper pid=68289) INFO 10-10 01:55:07 [custom_all_reduce.py:35] Skipping P2P check and trusting the driver's P2P report. 2025-10-10T01:55:07.5267107Z (EngineCore_DP0 pid=67405) (RayWorkerWrapper pid=68289) WARNING 10-10 01:55:07 [custom_all_reduce.py:154] Custom allreduce is disabled because your platform lacks GPU P2P capability or P2P test failed. To silence this warning, specify disable_custom_all_reduce=True explicitly. 2025-10-10T01:55:07.5268844Z (EngineCore_DP0 pid=67405) (pid=68290) INFO 10-10 01:55:02 [__init__.py:215] Automatically detected platform cuda. 2025-10-10T01:55:07.5290004Z (EngineCore_DP0 pid=67405) (RayWorkerWrapper pid=68290) INFO 10-10 01:55:07 [shm_broadcast.py:289] vLLM message queue communication handle: Handle(local_reader_ranks=[1], buffer_handle=(1, 4194304, 6, 'psm_bd4e7fc6'), local_subscribe_addr='ipc:///tmp/40efa5d7-93b8-4f21-b701-764d33ba53f5', remote_subscribe_addr=None, remote_addr_ipv6=False) 2025-10-10T01:55:07.7386013Z (EngineCore_DP0 pid=67405) (RayWorkerWrapper pid=68289) INFO 10-10 01:55:07 [parallel_state.py:1208] rank 1 in world size 2 is assigned as DP rank 0, PP rank 0, TP rank 1, EP rank 1 2025-10-10T01:55:07.9472383Z (EngineCore_DP0 pid=67405) (RayWorkerWrapper pid=68289) INFO 10-10 01:55:07 [topk_topp_sampler.py:55] Using FlashInfer for top-p & top-k sampling. 2025-10-10T01:55:08.0533230Z (EngineCore_DP0 pid=67405) (RayWorkerWrapper pid=68289) INFO 10-10 01:55:07 [gpu_model_runner.py:2707] Starting to load model RedHatAI/Meta-Llama-3.1-8B-Instruct-FP8... 2025-10-10T01:55:08.2603067Z (EngineCore_DP0 pid=67405) (RayWorkerWrapper pid=68289) INFO 10-10 01:55:08 [gpu_model_runner.py:2739] Loading model from scratch... 2025-10-10T01:55:08.3649827Z (EngineCore_DP0 pid=67405) (RayWorkerWrapper pid=68289) INFO 10-10 01:55:08 [cuda.py:361] Using Flash Attention backend on V1 engine. 2025-10-10T01:55:09.2857917Z (EngineCore_DP0 pid=67405) (RayWorkerWrapper pid=68289) INFO 10-10 01:55:09 [gpu_model_runner.py:2758] Model loading took 4.2624 GiB and 0.543191 seconds 2025-10-10T01:55:16.7260572Z (EngineCore_DP0 pid=67405) (RayWorkerWrapper pid=68289) INFO 10-10 01:55:16 [backends.py:548] Using cache directory: /var/lib/jenkins/.cache/vllm/torch_compile_cache/a29f583b26/rank_1_0/backbone for vLLM's torch.compile 2025-10-10T01:55:16.7261737Z (EngineCore_DP0 pid=67405) (RayWorkerWrapper pid=68289) INFO 10-10 01:55:16 [backends.py:559] Dynamo bytecode transform time: 7.19 s 2025-10-10T01:55:16.7263478Z (EngineCore_DP0 pid=67405) (RayWorkerWrapper pid=68290) [Gloo] Rank 0 is connected to 1 peer ranks. Expected number of connected peer ranks is : 1 [repeated 10x across cluster] (Ray deduplicates logs by default. Set RAY_DEDUP_LOGS=0 to disable log deduplication, or see https://docs.ray.io/en/master/ray-observability/user-guides/configure-logging.html#log-deduplication for more options.) 2025-10-10T01:55:16.7265333Z (EngineCore_DP0 pid=67405) (RayWorkerWrapper pid=68290) INFO 10-10 01:55:07 [__init__.py:1384] Found nccl from library libnccl.so.2 [repeated 3x across cluster] 2025-10-10T01:55:16.7266322Z (EngineCore_DP0 pid=67405) (RayWorkerWrapper pid=68290) INFO 10-10 01:55:07 [pynccl.py:104] vLLM is using nccl==2.27.5 [repeated 3x across cluster] 2025-10-10T01:55:16.7267335Z (EngineCore_DP0 pid=67405) (RayWorkerWrapper pid=68290) WARNING 10-10 01:55:07 [symm_mem.py:58] SymmMemCommunicator: Device capability 8.9 not supported, communicator is not available. 2025-10-10T01:55:16.7268534Z (EngineCore_DP0 pid=67405) (RayWorkerWrapper pid=68290) INFO 10-10 01:55:07 [custom_all_reduce.py:35] Skipping P2P check and trusting the driver's P2P report. 2025-10-10T01:55:16.7269912Z (EngineCore_DP0 pid=67405) (RayWorkerWrapper pid=68290) WARNING 10-10 01:55:07 [custom_all_reduce.py:154] Custom allreduce is disabled because your platform lacks GPU P2P capability or P2P test failed. To silence this warning, specify disable_custom_all_reduce=True explicitly. 2025-10-10T01:55:16.7271336Z (EngineCore_DP0 pid=67405) (RayWorkerWrapper pid=68290) INFO 10-10 01:55:07 [parallel_state.py:1208] rank 0 in world size 2 is assigned as DP rank 0, PP rank 0, TP rank 0, EP rank 0 2025-10-10T01:55:16.7272296Z (EngineCore_DP0 pid=67405) (RayWorkerWrapper pid=68290) INFO 10-10 01:55:07 [topk_topp_sampler.py:55] Using FlashInfer for top-p & top-k sampling. 2025-10-10T01:55:16.7273283Z (EngineCore_DP0 pid=67405) (RayWorkerWrapper pid=68290) INFO 10-10 01:55:07 [gpu_model_runner.py:2707] Starting to load model RedHatAI/Meta-Llama-3.1-8B-Instruct-FP8... 2025-10-10T01:55:16.7274181Z (EngineCore_DP0 pid=67405) (RayWorkerWrapper pid=68290) INFO 10-10 01:55:08 [gpu_model_runner.py:2739] Loading model from scratch... 2025-10-10T01:55:16.7274980Z (EngineCore_DP0 pid=67405) (RayWorkerWrapper pid=68290) INFO 10-10 01:55:08 [cuda.py:361] Using Flash Attention backend on V1 engine. 2025-10-10T01:55:16.7275876Z (EngineCore_DP0 pid=67405) (RayWorkerWrapper pid=68290) INFO 10-10 01:55:09 [gpu_model_runner.py:2758] Model loading took 4.2624 GiB and 0.545789 seconds 2025-10-10T01:55:19.5846014Z (EngineCore_DP0 pid=67405) (RayWorkerWrapper pid=68289) INFO 10-10 01:55:19 [backends.py:164] Directly load the compiled graph(s) for dynamic shape from the cache, took 2.354 s 2025-10-10T01:55:22.3400016Z (EngineCore_DP0 pid=67405) (RayWorkerWrapper pid=68289) INFO 10-10 01:55:22 [gpu_worker.py:298] Available KV cache memory: 15.34 GiB 2025-10-10T01:55:22.3401548Z (EngineCore_DP0 pid=67405) (RayWorkerWrapper pid=68290) INFO 10-10 01:55:16 [backends.py:548] Using cache directory: /var/lib/jenkins/.cache/vllm/torch_compile_cache/a29f583b26/rank_0_0/backbone for vLLM's torch.compile 2025-10-10T01:55:22.3402913Z (EngineCore_DP0 pid=67405) (RayWorkerWrapper pid=68290) INFO 10-10 01:55:16 [backends.py:559] Dynamo bytecode transform time: 7.33 s 2025-10-10T01:55:22.6208617Z (EngineCore_DP0 pid=67405) INFO 10-10 01:55:22 [kv_cache_utils.py:1087] GPU KV cache size: 251,360 tokens 2025-10-10T01:55:22.6210350Z (EngineCore_DP0 pid=67405) INFO 10-10 01:55:22 [kv_cache_utils.py:1091] Maximum concurrency for 2,048 tokens per request: 122.73x 2025-10-10T01:55:22.6211634Z (EngineCore_DP0 pid=67405) INFO 10-10 01:55:22 [kv_cache_utils.py:1087] GPU KV cache size: 251,360 tokens 2025-10-10T01:55:22.6212745Z (EngineCore_DP0 pid=67405) INFO 10-10 01:55:22 [kv_cache_utils.py:1091] Maximum concurrency for 2,048 tokens per request: 122.73x 2025-10-10T01:55:22.6509241Z (EngineCore_DP0 pid=67405) (RayWorkerWrapper pid=68289) INFO 10-10 01:55:22 [gpu_worker.py:332] Compile and warming up model for size 8 2025-10-10T01:55:37.7474996Z (EngineCore_DP0 pid=67405) (RayWorkerWrapper pid=68289) INFO 10-10 01:55:37 [backends.py:168] Directly load the compiled graph(s) for shape 8 from the cache, took 15.010 s 2025-10-10T01:55:37.7477677Z (EngineCore_DP0 pid=67405) (RayWorkerWrapper pid=68290) INFO 10-10 01:55:19 [backends.py:164] Directly load the compiled graph(s) for dynamic shape from the cache, took 2.358 s 2025-10-10T01:55:37.7480082Z (EngineCore_DP0 pid=67405) (RayWorkerWrapper pid=68290) INFO 10-10 01:55:22 [gpu_worker.py:298] Available KV cache memory: 15.34 GiB 2025-10-10T01:55:37.7482835Z (EngineCore_DP0 pid=67405) (RayWorkerWrapper pid=68289) INFO 10-10 01:55:37 [gpu_worker.py:332] Compile and warming up model for size 4 [repeated 2x across cluster] 2025-10-10T01:55:57.2348105Z (EngineCore_DP0 pid=67405) INFO 10-10 01:55:57 [core.py:211] init engine (profile, create kv cache, warmup model) took 47.96 seconds 2025-10-10T01:55:57.2389243Z (EngineCore_DP0 pid=67405) (RayWorkerWrapper pid=68289) INFO 10-10 01:55:57 [monitor.py:32] torch.compile takes 7.19 s in total 2025-10-10T01:55:57.2390892Z (EngineCore_DP0 pid=67405) (RayWorkerWrapper pid=68289) INFO 10-10 01:55:57 [backends.py:168] Directly load the compiled graph(s) for shape 4 from the cache, took 19.547 s [repeated 2x across cluster] 2025-10-10T01:55:57.2392151Z (EngineCore_DP0 pid=67405) (RayWorkerWrapper pid=68290) INFO 10-10 01:55:37 [gpu_worker.py:332] Compile and warming up model for size 4 2025-10-10T01:55:57.7374380Z (EngineCore_DP0 pid=67405) INFO 10-10 01:55:57 [vllm.py:345] Cudagraph is disabled under eager mode 2025-10-10T01:55:57.8917262Z (APIServer pid=67282) INFO 10-10 01:55:57 [loggers.py:147] Engine 000: vllm cache_config_info with initialization after num_gpu_blocks is: 15710 2025-10-10T01:55:58.0137922Z (EngineCore_DP0 pid=67405) INFO 10-10 01:55:58 [gc_utils.py:41] GC Debug Config. enabled:False,top_objects:-1 2025-10-10T01:55:58.0382245Z (APIServer pid=67282) INFO 10-10 01:55:58 [api_server.py:1634] Supported_tasks: ['generate'] 2025-10-10T01:55:58.0744961Z (APIServer pid=67282) WARNING 10-10 01:55:58 [model.py:1424] Default sampling parameters have been overridden by the model's Hugging Face generation config recommended from the model creator. If this is not intended, please relaunch vLLM instance with `--generation-config vllm`. 2025-10-10T01:55:58.0746605Z (APIServer pid=67282) INFO 10-10 01:55:58 [serving_responses.py:140] Using default chat sampling params from model: {'temperature': 0.6, 'top_p': 0.9} 2025-10-10T01:55:58.1020044Z (APIServer pid=67282) INFO 10-10 01:55:58 [serving_chat.py:139] Using default chat sampling params from model: {'temperature': 0.6, 'top_p': 0.9} 2025-10-10T01:55:58.2291851Z (APIServer pid=67282) INFO 10-10 01:55:58 [serving_completion.py:76] Using default completion sampling params from model: {'temperature': 0.6, 'top_p': 0.9} 2025-10-10T01:55:58.2293110Z (APIServer pid=67282) INFO 10-10 01:55:58 [api_server.py:1912] Starting vLLM API server 0 on http://0.0.0.0:48041 2025-10-10T01:55:58.2293826Z (APIServer pid=67282) INFO 10-10 01:55:58 [launcher.py:34] Available routes are: 2025-10-10T01:55:58.2294570Z (APIServer pid=67282) INFO 10-10 01:55:58 [launcher.py:42] Route: /openapi.json, Methods: HEAD, GET 2025-10-10T01:55:58.2295293Z (APIServer pid=67282) INFO 10-10 01:55:58 [launcher.py:42] Route: /docs, Methods: HEAD, GET 2025-10-10T01:55:58.2296361Z (APIServer pid=67282) INFO 10-10 01:55:58 [launcher.py:42] Route: /docs/oauth2-redirect, Methods: HEAD, GET 2025-10-10T01:55:58.2297059Z (APIServer pid=67282) INFO 10-10 01:55:58 [launcher.py:42] Route: /redoc, Methods: HEAD, GET 2025-10-10T01:55:58.2297708Z (APIServer pid=67282) INFO 10-10 01:55:58 [launcher.py:42] Route: /health, Methods: GET 2025-10-10T01:55:58.2298341Z (APIServer pid=67282) INFO 10-10 01:55:58 [launcher.py:42] Route: /load, Methods: GET 2025-10-10T01:55:58.2298971Z (APIServer pid=67282) INFO 10-10 01:55:58 [launcher.py:42] Route: /ping, Methods: POST 2025-10-10T01:55:58.2299591Z (APIServer pid=67282) INFO 10-10 01:55:58 [launcher.py:42] Route: /ping, Methods: GET 2025-10-10T01:55:58.2300219Z (APIServer pid=67282) INFO 10-10 01:55:58 [launcher.py:42] Route: /tokenize, Methods: POST 2025-10-10T01:55:58.2301075Z (APIServer pid=67282) INFO 10-10 01:55:58 [launcher.py:42] Route: /detokenize, Methods: POST 2025-10-10T01:55:58.2301668Z (APIServer pid=67282) INFO 10-10 01:55:58 [launcher.py:42] Route: /v1/models, Methods: GET 2025-10-10T01:55:58.2302195Z (APIServer pid=67282) INFO 10-10 01:55:58 [launcher.py:42] Route: /version, Methods: GET 2025-10-10T01:55:58.2302722Z (APIServer pid=67282) INFO 10-10 01:55:58 [launcher.py:42] Route: /v1/responses, Methods: POST 2025-10-10T01:55:58.2303400Z (APIServer pid=67282) INFO 10-10 01:55:58 [launcher.py:42] Route: /v1/responses/{response_id}, Methods: GET 2025-10-10T01:55:58.2304028Z (APIServer pid=67282) INFO 10-10 01:55:58 [launcher.py:42] Route: /v1/responses/{response_id}/cancel, Methods: POST 2025-10-10T01:55:58.2304639Z (APIServer pid=67282) INFO 10-10 01:55:58 [launcher.py:42] Route: /v1/chat/completions, Methods: POST 2025-10-10T01:55:58.2305208Z (APIServer pid=67282) INFO 10-10 01:55:58 [launcher.py:42] Route: /v1/completions, Methods: POST 2025-10-10T01:55:58.2305782Z (APIServer pid=67282) INFO 10-10 01:55:58 [launcher.py:42] Route: /v1/embeddings, Methods: POST 2025-10-10T01:55:58.2306330Z (APIServer pid=67282) INFO 10-10 01:55:58 [launcher.py:42] Route: /pooling, Methods: POST 2025-10-10T01:55:58.2306851Z (APIServer pid=67282) INFO 10-10 01:55:58 [launcher.py:42] Route: /classify, Methods: POST 2025-10-10T01:55:58.2307368Z (APIServer pid=67282) INFO 10-10 01:55:58 [launcher.py:42] Route: /score, Methods: POST 2025-10-10T01:55:58.2307880Z (APIServer pid=67282) INFO 10-10 01:55:58 [launcher.py:42] Route: /v1/score, Methods: POST 2025-10-10T01:55:58.2308450Z (APIServer pid=67282) INFO 10-10 01:55:58 [launcher.py:42] Route: /v1/audio/transcriptions, Methods: POST 2025-10-10T01:55:58.2309042Z (APIServer pid=67282) INFO 10-10 01:55:58 [launcher.py:42] Route: /v1/audio/translations, Methods: POST 2025-10-10T01:55:58.2309583Z (APIServer pid=67282) INFO 10-10 01:55:58 [launcher.py:42] Route: /rerank, Methods: POST 2025-10-10T01:55:58.2310108Z (APIServer pid=67282) INFO 10-10 01:55:58 [launcher.py:42] Route: /v1/rerank, Methods: POST 2025-10-10T01:55:58.2310640Z (APIServer pid=67282) INFO 10-10 01:55:58 [launcher.py:42] Route: /v2/rerank, Methods: POST 2025-10-10T01:55:58.2311182Z (APIServer pid=67282) INFO 10-10 01:55:58 [launcher.py:42] Route: /scale_elastic_ep, Methods: POST 2025-10-10T01:55:58.2311757Z (APIServer pid=67282) INFO 10-10 01:55:58 [launcher.py:42] Route: /is_scaling_elastic_ep, Methods: POST 2025-10-10T01:55:58.2312310Z (APIServer pid=67282) INFO 10-10 01:55:58 [launcher.py:42] Route: /invocations, Methods: POST 2025-10-10T01:55:58.2312834Z (APIServer pid=67282) INFO 10-10 01:55:58 [launcher.py:42] Route: /metrics, Methods: GET 2025-10-10T01:55:58.2452975Z (APIServer pid=67282) INFO: Started server process [67282] 2025-10-10T01:55:58.2453748Z (APIServer pid=67282) INFO: Waiting for application startup. 2025-10-10T01:55:58.4935888Z (APIServer pid=67282) INFO: Application startup complete. 2025-10-10T01:55:58.6760090Z (APIServer pid=67282) INFO: 127.0.0.1:60694 - "GET /health HTTP/1.1" 200 OK 2025-10-10T01:55:58.9061398Z (APIServer pid=67282) INFO: 127.0.0.1:60710 - "GET /v1/models HTTP/1.1" 200 OK 2025-10-10T01:55:58.9373050Z (EngineCore_DP0 pid=67405) INFO 10-10 01:55:58 [ray_distributed_executor.py:552] RAY_CGRAPH_get_timeout is set to 300 2025-10-10T01:55:58.9374946Z (EngineCore_DP0 pid=67405) INFO 10-10 01:55:58 [ray_distributed_executor.py:554] VLLM_USE_RAY_COMPILED_DAG_CHANNEL_TYPE = auto 2025-10-10T01:55:58.9377251Z (EngineCore_DP0 pid=67405) INFO 10-10 01:55:58 [ray_distributed_executor.py:556] VLLM_USE_RAY_COMPILED_DAG_OVERLAP_COMM = False 2025-10-10T01:55:58.9380657Z (EngineCore_DP0 pid=67405) INFO 10-10 01:55:58 [ray_distributed_executor.py:621] Using RayPPCommunicator (which wraps vLLM _PP GroupCoordinator) for Ray Compiled Graph communication. 2025-10-10T01:55:59.1711172Z (APIServer pid=67282) INFO: 127.0.0.1:60710 - "POST /v1/completions HTTP/1.1" 200 OK 2025-10-10T01:55:59.2924670Z (APIServer pid=67282) INFO: 127.0.0.1:60710 - "POST /v1/completions HTTP/1.1" 200 OK 2025-10-10T01:55:59.4949035Z (APIServer pid=67282) INFO: 127.0.0.1:60710 - "POST /v1/completions HTTP/1.1" 200 OK 2025-10-10T01:55:59.6218405Z (APIServer pid=67282) INFO: 127.0.0.1:60710 - "POST /v1/completions HTTP/1.1" 200 OK 2025-10-10T01:55:59.7418093Z (APIServer pid=67282) INFO: 127.0.0.1:60710 - "POST /v1/completions HTTP/1.1" 200 OK 2025-10-10T01:55:59.7502431Z (APIServer pid=67282) INFO: 127.0.0.1:60710 - "POST /v1/completions HTTP/1.1" 200 OK 2025-10-10T01:55:59.8959447Z (APIServer pid=67282) INFO 10-10 01:55:59 [launcher.py:99] Shutting down FastAPI HTTP server. 2025-10-10T01:55:59.8960151Z *** SIGTERM received at time=1760061359 on cpu 31 *** 2025-10-10T01:55:59.9008454Z PC: @ 0x7ff73a965117 (unknown) (unknown) 2025-10-10T01:55:59.9008929Z @ 0x7ff73a916520 (unknown) (unknown) 2025-10-10T01:55:59.9009579Z [2025-10-10 01:55:59,900 E 67405 67405] logging.cc:501: *** SIGTERM received at time=1760061359 on cpu 31 *** 2025-10-10T01:55:59.9012661Z [2025-10-10 01:55:59,900 E 67405 67405] logging.cc:501: PC: @ 0x7ff73a965117 (unknown) (unknown) 2025-10-10T01:55:59.9014266Z [2025-10-10 01:55:59,901 E 67405 67405] logging.cc:501: @ 0x7ff73a916520 (unknown) (unknown) 2025-10-10T01:55:59.9015779Z (EngineCore_DP0 pid=67405) INFO 10-10 01:55:59 [ray_distributed_executor.py:122] Shutting down Ray distributed executor. If you see error log from logging.cc regarding SIGTERM received, please ignore because this is the expected termination process in Ray. 2025-10-10T01:55:59.9017020Z (EngineCore_DP0 pid=67405) 2025-10-10 01:55:59,901 INFO compiled_dag_node.py:2171 -- Tearing down compiled DAG 2025-10-10T01:55:59.9019382Z (EngineCore_DP0 pid=67405) 2025-10-10 01:55:59,901 INFO compiled_dag_node.py:2176 -- Cancelling compiled worker on actor: Actor(RayWorkerWrapper, 4ab9e0fea7fb089d5b1e455a01000000) 2025-10-10T01:55:59.9020668Z (EngineCore_DP0 pid=67405) 2025-10-10 01:55:59,901 INFO compiled_dag_node.py:2176 -- Cancelling compiled worker on actor: Actor(RayWorkerWrapper, bf2ce6f31a3caf32233fd68b01000000) 2025-10-10T01:55:59.9063923Z (EngineCore_DP0 pid=67405) 2025-10-10 01:55:59,906 INFO compiled_dag_node.py:2198 -- Waiting for worker tasks to exit 2025-10-10T01:55:59.9065161Z (EngineCore_DP0 pid=67405) 2025-10-10 01:55:59,906 INFO compiled_dag_node.py:2201 -- Teardown complete 2025-10-10T01:56:00.4097299Z (EngineCore_DP0 pid=67405) (RayWorkerWrapper pid=68290) INFO 10-10 01:55:57 [monitor.py:32] torch.compile takes 7.33 s in total 2025-10-10T01:56:00.4098951Z (EngineCore_DP0 pid=67405) (RayWorkerWrapper pid=68290) INFO 10-10 01:55:57 [backends.py:168] Directly load the compiled graph(s) for shape 4 from the cache, took 19.541 s 2025-10-10T01:56:00.4100785Z (EngineCore_DP0 pid=67405) (pid=68290) /opt/conda/envs/py_3.12/lib/python3.12/site-packages/torch/cuda/__init__.py:64: FutureWarning: The pynvml package is deprecated. Please install nvidia-ml-py instead. If you did not install pynvml directly, please report this to the maintainers of the package that installed pynvml for you. 2025-10-10T01:56:00.4102547Z (EngineCore_DP0 pid=67405) (pid=68290) import pynvml # type: ignore[import] 2025-10-10T01:56:02.4248869Z (APIServer pid=67282) INFO: Shutting down 2025-10-10T01:56:02.5242852Z (APIServer pid=67282) INFO: Waiting for application shutdown. 2025-10-10T01:56:02.5253084Z (APIServer pid=67282) INFO: Application shutdown complete. 2025-10-10T01:56:02.8801180Z INFO 10-10 01:56:02 [model.py:551] Resolved architecture: LlamaForCausalLM 2025-10-10T01:56:02.8802818Z WARNING 10-10 01:56:02 [model.py:1768] Casting torch.bfloat16 to torch.float16. 2025-10-10T01:56:02.8803631Z INFO 10-10 01:56:02 [model.py:1545] Using max model len 2048 2025-10-10T01:56:02.8807974Z Launching RemoteOpenAIServer with: vllm serve RedHatAI/Meta-Llama-3.1-8B-Instruct-FP8 --dtype float16 --max-model-len 2048 --max-num-seqs 8 --enforce-eager --tokenizer-mode auto --tensor-parallel-size 2 --distributed-executor-backend mp --load-format dummy --port 41393 --seed 0 2025-10-10T01:56:03.3961289Z /opt/conda/envs/py_3.12/lib/python3.12/site-packages/torch/cuda/__init__.py:64: FutureWarning: The pynvml package is deprecated. Please install nvidia-ml-py instead. If you did not install pynvml directly, please report this to the maintainers of the package that installed pynvml for you. 2025-10-10T01:56:03.3963310Z import pynvml # type: ignore[import] 2025-10-10T01:56:05.6505081Z INFO 10-10 01:56:05 [__init__.py:215] Automatically detected platform cuda. 2025-10-10T01:56:09.6480449Z (APIServer pid=72502) INFO 10-10 01:56:09 [api_server.py:1839] vLLM API server version 0.11.0rc2.dev157+g0ad9951c4.d20251010 2025-10-10T01:56:09.6503321Z (APIServer pid=72502) INFO 10-10 01:56:09 [utils.py:233] non-default args: {'model_tag': 'RedHatAI/Meta-Llama-3.1-8B-Instruct-FP8', 'port': 41393, 'model': 'RedHatAI/Meta-Llama-3.1-8B-Instruct-FP8', 'dtype': 'float16', 'seed': 0, 'max_model_len': 2048, 'enforce_eager': True, 'load_format': 'dummy', 'distributed_executor_backend': 'mp', 'tensor_parallel_size': 2, 'max_num_seqs': 8} 2025-10-10T01:56:09.8599450Z (APIServer pid=72502) INFO 10-10 01:56:09 [model.py:551] Resolved architecture: LlamaForCausalLM 2025-10-10T01:56:09.8600249Z (APIServer pid=72502) `torch_dtype` is deprecated! Use `dtype` instead! 2025-10-10T01:56:09.8601340Z (APIServer pid=72502) WARNING 10-10 01:56:09 [model.py:1768] Casting torch.bfloat16 to torch.float16. 2025-10-10T01:56:09.8603442Z (APIServer pid=72502) INFO 10-10 01:56:09 [model.py:1545] Using max model len 2048 2025-10-10T01:56:10.2412165Z (APIServer pid=72502) INFO 10-10 01:56:10 [scheduler.py:222] Chunked prefill is enabled with max_num_batched_tokens=2048. 2025-10-10T01:56:10.2777907Z (APIServer pid=72502) INFO 10-10 01:56:10 [vllm.py:345] Cudagraph is disabled under eager mode 2025-10-10T01:56:11.2865344Z /opt/conda/envs/py_3.12/lib/python3.12/site-packages/torch/cuda/__init__.py:64: FutureWarning: The pynvml package is deprecated. Please install nvidia-ml-py instead. If you did not install pynvml directly, please report this to the maintainers of the package that installed pynvml for you. 2025-10-10T01:56:11.2867213Z import pynvml # type: ignore[import] 2025-10-10T01:56:13.6111010Z INFO 10-10 01:56:13 [__init__.py:215] Automatically detected platform cuda. 2025-10-10T01:56:17.6962411Z (EngineCore_DP0 pid=72625) INFO 10-10 01:56:17 [core.py:648] Waiting for init message from front-end. 2025-10-10T01:56:17.7230031Z (EngineCore_DP0 pid=72625) INFO 10-10 01:56:17 [core.py:78] Initializing a V1 LLM engine (v0.11.0rc2.dev157+g0ad9951c4.d20251010) with config: model='RedHatAI/Meta-Llama-3.1-8B-Instruct-FP8', speculative_config=None, tokenizer='RedHatAI/Meta-Llama-3.1-8B-Instruct-FP8', skip_tokenizer_init=False, tokenizer_mode=auto, revision=None, tokenizer_revision=None, trust_remote_code=False, dtype=torch.float16, max_seq_len=2048, download_dir=None, load_format=dummy, tensor_parallel_size=2, pipeline_parallel_size=1, data_parallel_size=1, disable_custom_all_reduce=False, quantization=compressed-tensors, enforce_eager=True, kv_cache_dtype=auto, device_config=cuda, structured_outputs_config=StructuredOutputsConfig(backend='auto', disable_fallback=False, disable_any_whitespace=False, disable_additional_properties=False, reasoning_parser=''), observability_config=ObservabilityConfig(show_hidden_metrics_for_version=None, otlp_traces_endpoint=None, collect_detailed_traces=None), seed=0, served_model_name=RedHatAI/Meta-Llama-3.1-8B-Instruct-FP8, enable_prefix_caching=True, chunked_prefill_enabled=True, pooler_config=None, compilation_config={'level': 0, 'debug_dump_path': None, 'cache_dir': '', 'backend': '', 'custom_ops': [], 'splitting_ops': None, 'use_inductor': True, 'compile_sizes': [], 'inductor_compile_config': {'enable_auto_functionalized_v2': False}, 'inductor_passes': {}, 'cudagraph_mode': , 'use_cudagraph': False, 'cudagraph_num_of_warmups': 0, 'cudagraph_capture_sizes': [], 'cudagraph_copy_inputs': False, 'full_cuda_graph': False, 'use_inductor_graph_partition': False, 'pass_config': {}, 'max_capture_size': 0, 'local_cache_dir': None} 2025-10-10T01:56:17.7235508Z (EngineCore_DP0 pid=72625) WARNING 10-10 01:56:17 [multiproc_executor.py:720] Reducing Torch parallelism from 24 threads to 1 to avoid unnecessary CPU contention. Set OMP_NUM_THREADS in the external environment to tune this value as needed. 2025-10-10T01:56:17.7244408Z (EngineCore_DP0 pid=72625) INFO 10-10 01:56:17 [shm_broadcast.py:289] vLLM message queue communication handle: Handle(local_reader_ranks=[0, 1], buffer_handle=(2, 16777216, 10, 'psm_b74992a5'), local_subscribe_addr='ipc:///tmp/97e147f9-acf3-4ba4-a3f1-bbff0f1db707', remote_subscribe_addr=None, remote_addr_ipv6=False) 2025-10-10T01:56:18.2059963Z /opt/conda/envs/py_3.12/lib/python3.12/site-packages/torch/cuda/__init__.py:64: FutureWarning: The pynvml package is deprecated. Please install nvidia-ml-py instead. If you did not install pynvml directly, please report this to the maintainers of the package that installed pynvml for you. 2025-10-10T01:56:18.2062367Z import pynvml # type: ignore[import] 2025-10-10T01:56:18.2064579Z /opt/conda/envs/py_3.12/lib/python3.12/site-packages/torch/cuda/__init__.py:64: FutureWarning: The pynvml package is deprecated. Please install nvidia-ml-py instead. If you did not install pynvml directly, please report this to the maintainers of the package that installed pynvml for you. 2025-10-10T01:56:18.2066744Z import pynvml # type: ignore[import] 2025-10-10T01:56:20.5448467Z INFO 10-10 01:56:20 [__init__.py:215] Automatically detected platform cuda. 2025-10-10T01:56:20.5503016Z INFO 10-10 01:56:20 [__init__.py:215] Automatically detected platform cuda. 2025-10-10T01:56:25.1745245Z INFO 10-10 01:56:25 [shm_broadcast.py:289] vLLM message queue communication handle: Handle(local_reader_ranks=[0], buffer_handle=(1, 10485760, 10, 'psm_969f9ba1'), local_subscribe_addr='ipc:///tmp/e665f4b0-b09b-4e66-a3e4-c7e613726268', remote_subscribe_addr=None, remote_addr_ipv6=False) 2025-10-10T01:56:25.2038999Z INFO 10-10 01:56:25 [shm_broadcast.py:289] vLLM message queue communication handle: Handle(local_reader_ranks=[0], buffer_handle=(1, 10485760, 10, 'psm_46ef07fc'), local_subscribe_addr='ipc:///tmp/f860fa6b-db3f-4a8e-97ea-9decab9d0723', remote_subscribe_addr=None, remote_addr_ipv6=False) 2025-10-10T01:56:25.9242493Z [Gloo] Rank 0 is connected to 1 peer ranks. Expected number of connected peer ranks is : 1 2025-10-10T01:56:25.9243735Z [Gloo] Rank 1 is connected to 1 peer ranks. Expected number of connected peer ranks is : 1 2025-10-10T01:56:25.9283725Z [Gloo] Rank 0 is connected to 1 peer ranks. Expected number of connected peer ranks is : 1 2025-10-10T01:56:25.9284887Z [Gloo] Rank 1 is connected to 1 peer ranks. Expected number of connected peer ranks is : 1 2025-10-10T01:56:25.9326971Z INFO 10-10 01:56:25 [__init__.py:1384] Found nccl from library libnccl.so.2 2025-10-10T01:56:25.9327842Z INFO 10-10 01:56:25 [__init__.py:1384] Found nccl from library libnccl.so.2 2025-10-10T01:56:25.9329782Z INFO 10-10 01:56:25 [pynccl.py:104] vLLM is using nccl==2.27.5 2025-10-10T01:56:25.9330721Z INFO 10-10 01:56:25 [pynccl.py:104] vLLM is using nccl==2.27.5 2025-10-10T01:56:26.1027059Z WARNING 10-10 01:56:26 [symm_mem.py:58] SymmMemCommunicator: Device capability 8.9 not supported, communicator is not available. 2025-10-10T01:56:26.1028746Z WARNING 10-10 01:56:26 [symm_mem.py:58] SymmMemCommunicator: Device capability 8.9 not supported, communicator is not available. 2025-10-10T01:56:26.1050945Z INFO 10-10 01:56:26 [custom_all_reduce.py:35] Skipping P2P check and trusting the driver's P2P report. 2025-10-10T01:56:26.1052175Z INFO 10-10 01:56:26 [custom_all_reduce.py:35] Skipping P2P check and trusting the driver's P2P report. 2025-10-10T01:56:26.1054116Z WARNING 10-10 01:56:26 [custom_all_reduce.py:154] Custom allreduce is disabled because your platform lacks GPU P2P capability or P2P test failed. To silence this warning, specify disable_custom_all_reduce=True explicitly. 2025-10-10T01:56:26.1056882Z WARNING 10-10 01:56:26 [custom_all_reduce.py:154] Custom allreduce is disabled because your platform lacks GPU P2P capability or P2P test failed. To silence this warning, specify disable_custom_all_reduce=True explicitly. 2025-10-10T01:56:26.1740769Z INFO 10-10 01:56:26 [shm_broadcast.py:289] vLLM message queue communication handle: Handle(local_reader_ranks=[1], buffer_handle=(1, 4194304, 6, 'psm_193ff72d'), local_subscribe_addr='ipc:///tmp/eae93a63-26c6-431d-b430-fc86062f16ee', remote_subscribe_addr=None, remote_addr_ipv6=False) 2025-10-10T01:56:26.1769429Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-10-10T01:56:26.1770634Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-10-10T01:56:26.1776150Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-10-10T01:56:26.1784042Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-10-10T01:56:26.2136732Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-10-10T01:56:26.2144839Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-10-10T01:56:26.2156491Z [Gloo] Rank 0 is connected to 1 peer ranks. Expected number of connected peer ranks is : 1 2025-10-10T01:56:26.2157639Z [Gloo] Rank 1 is connected to 1 peer ranks. Expected number of connected peer ranks is : 1 2025-10-10T01:56:26.2159233Z INFO 10-10 01:56:26 [__init__.py:1384] Found nccl from library libnccl.so.2 2025-10-10T01:56:26.2160203Z INFO 10-10 01:56:26 [pynccl.py:104] vLLM is using nccl==2.27.5 2025-10-10T01:56:26.2160968Z INFO 10-10 01:56:26 [__init__.py:1384] Found nccl from library libnccl.so.2 2025-10-10T01:56:26.2161322Z INFO 10-10 01:56:26 [pynccl.py:104] vLLM is using nccl==2.27.5 2025-10-10T01:56:26.2495520Z INFO 10-10 01:56:26 [parallel_state.py:1208] rank 0 in world size 2 is assigned as DP rank 0, PP rank 0, TP rank 0, EP rank 0 2025-10-10T01:56:26.2497073Z INFO 10-10 01:56:26 [parallel_state.py:1208] rank 1 in world size 2 is assigned as DP rank 0, PP rank 0, TP rank 1, EP rank 1 2025-10-10T01:56:26.4949209Z INFO 10-10 01:56:26 [topk_topp_sampler.py:55] Using FlashInfer for top-p & top-k sampling. 2025-10-10T01:56:26.4978716Z INFO 10-10 01:56:26 [topk_topp_sampler.py:55] Using FlashInfer for top-p & top-k sampling. 2025-10-10T01:56:26.5313700Z (Worker_TP0 pid=72744) INFO 10-10 01:56:26 [gpu_model_runner.py:2707] Starting to load model RedHatAI/Meta-Llama-3.1-8B-Instruct-FP8... 2025-10-10T01:56:26.5337928Z (Worker_TP1 pid=72745) INFO 10-10 01:56:26 [gpu_model_runner.py:2707] Starting to load model RedHatAI/Meta-Llama-3.1-8B-Instruct-FP8... 2025-10-10T01:56:26.7580241Z (Worker_TP1 pid=72745) INFO 10-10 01:56:26 [gpu_model_runner.py:2739] Loading model from scratch... 2025-10-10T01:56:26.7581038Z (Worker_TP0 pid=72744) INFO 10-10 01:56:26 [gpu_model_runner.py:2739] Loading model from scratch... 2025-10-10T01:56:26.8496168Z (Worker_TP0 pid=72744) INFO 10-10 01:56:26 [cuda.py:361] Using Flash Attention backend on V1 engine. 2025-10-10T01:56:26.8499978Z (Worker_TP1 pid=72745) INFO 10-10 01:56:26 [cuda.py:361] Using Flash Attention backend on V1 engine. 2025-10-10T01:56:27.7016783Z (Worker_TP0 pid=72744) INFO 10-10 01:56:27 [gpu_model_runner.py:2758] Model loading took 4.2624 GiB and 0.489674 seconds 2025-10-10T01:56:27.7168435Z (Worker_TP1 pid=72745) INFO 10-10 01:56:27 [gpu_model_runner.py:2758] Model loading took 4.2624 GiB and 0.490846 seconds 2025-10-10T01:56:33.1423200Z (Worker_TP0 pid=72744) INFO 10-10 01:56:33 [gpu_worker.py:298] Available KV cache memory: 15.32 GiB 2025-10-10T01:56:33.1680442Z (Worker_TP1 pid=72745) INFO 10-10 01:56:33 [gpu_worker.py:298] Available KV cache memory: 15.32 GiB 2025-10-10T01:56:33.4024739Z (EngineCore_DP0 pid=72625) INFO 10-10 01:56:33 [kv_cache_utils.py:1087] GPU KV cache size: 250,976 tokens 2025-10-10T01:56:33.4025700Z (EngineCore_DP0 pid=72625) INFO 10-10 01:56:33 [kv_cache_utils.py:1091] Maximum concurrency for 2,048 tokens per request: 122.55x 2025-10-10T01:56:33.4026567Z (EngineCore_DP0 pid=72625) INFO 10-10 01:56:33 [kv_cache_utils.py:1087] GPU KV cache size: 250,976 tokens 2025-10-10T01:56:33.4027402Z (EngineCore_DP0 pid=72625) INFO 10-10 01:56:33 [kv_cache_utils.py:1091] Maximum concurrency for 2,048 tokens per request: 122.55x 2025-10-10T01:56:33.5316340Z (EngineCore_DP0 pid=72625) INFO 10-10 01:56:33 [core.py:211] init engine (profile, create kv cache, warmup model) took 5.75 seconds 2025-10-10T01:56:34.0010625Z (EngineCore_DP0 pid=72625) INFO 10-10 01:56:34 [vllm.py:345] Cudagraph is disabled under eager mode 2025-10-10T01:56:34.1858288Z (APIServer pid=72502) INFO 10-10 01:56:34 [loggers.py:147] Engine 000: vllm cache_config_info with initialization after num_gpu_blocks is: 15686 2025-10-10T01:56:34.2231179Z (EngineCore_DP0 pid=72625) INFO 10-10 01:56:34 [gc_utils.py:41] GC Debug Config. enabled:False,top_objects:-1 2025-10-10T01:56:34.2444107Z (APIServer pid=72502) INFO 10-10 01:56:34 [api_server.py:1634] Supported_tasks: ['generate'] 2025-10-10T01:56:34.2752713Z (APIServer pid=72502) WARNING 10-10 01:56:34 [model.py:1424] Default sampling parameters have been overridden by the model's Hugging Face generation config recommended from the model creator. If this is not intended, please relaunch vLLM instance with `--generation-config vllm`. 2025-10-10T01:56:34.2755321Z (APIServer pid=72502) INFO 10-10 01:56:34 [serving_responses.py:140] Using default chat sampling params from model: {'temperature': 0.6, 'top_p': 0.9} 2025-10-10T01:56:34.3107968Z (APIServer pid=72502) INFO 10-10 01:56:34 [serving_chat.py:139] Using default chat sampling params from model: {'temperature': 0.6, 'top_p': 0.9} 2025-10-10T01:56:34.3446121Z (APIServer pid=72502) INFO 10-10 01:56:34 [serving_completion.py:76] Using default completion sampling params from model: {'temperature': 0.6, 'top_p': 0.9} 2025-10-10T01:56:34.3449457Z (APIServer pid=72502) INFO 10-10 01:56:34 [api_server.py:1912] Starting vLLM API server 0 on http://0.0.0.0:41393 2025-10-10T01:56:34.3450925Z (APIServer pid=72502) INFO 10-10 01:56:34 [launcher.py:34] Available routes are: 2025-10-10T01:56:34.3452260Z (APIServer pid=72502) INFO 10-10 01:56:34 [launcher.py:42] Route: /openapi.json, Methods: GET, HEAD 2025-10-10T01:56:34.3453613Z (APIServer pid=72502) INFO 10-10 01:56:34 [launcher.py:42] Route: /docs, Methods: GET, HEAD 2025-10-10T01:56:34.3454415Z (APIServer pid=72502) INFO 10-10 01:56:34 [launcher.py:42] Route: /docs/oauth2-redirect, Methods: GET, HEAD 2025-10-10T01:56:34.3454992Z (APIServer pid=72502) INFO 10-10 01:56:34 [launcher.py:42] Route: /redoc, Methods: GET, HEAD 2025-10-10T01:56:34.3455743Z (APIServer pid=72502) INFO 10-10 01:56:34 [launcher.py:42] Route: /health, Methods: GET 2025-10-10T01:56:34.3456292Z (APIServer pid=72502) INFO 10-10 01:56:34 [launcher.py:42] Route: /load, Methods: GET 2025-10-10T01:56:34.3456807Z (APIServer pid=72502) INFO 10-10 01:56:34 [launcher.py:42] Route: /ping, Methods: POST 2025-10-10T01:56:34.3457320Z (APIServer pid=72502) INFO 10-10 01:56:34 [launcher.py:42] Route: /ping, Methods: GET 2025-10-10T01:56:34.3457932Z (APIServer pid=72502) INFO 10-10 01:56:34 [launcher.py:42] Route: /tokenize, Methods: POST 2025-10-10T01:56:34.3458474Z (APIServer pid=72502) INFO 10-10 01:56:34 [launcher.py:42] Route: /detokenize, Methods: POST 2025-10-10T01:56:34.3459015Z (APIServer pid=72502) INFO 10-10 01:56:34 [launcher.py:42] Route: /v1/models, Methods: GET 2025-10-10T01:56:34.3459549Z (APIServer pid=72502) INFO 10-10 01:56:34 [launcher.py:42] Route: /version, Methods: GET 2025-10-10T01:56:34.3460096Z (APIServer pid=72502) INFO 10-10 01:56:34 [launcher.py:42] Route: /v1/responses, Methods: POST 2025-10-10T01:56:34.3460695Z (APIServer pid=72502) INFO 10-10 01:56:34 [launcher.py:42] Route: /v1/responses/{response_id}, Methods: GET 2025-10-10T01:56:34.3461331Z (APIServer pid=72502) INFO 10-10 01:56:34 [launcher.py:42] Route: /v1/responses/{response_id}/cancel, Methods: POST 2025-10-10T01:56:34.3461971Z (APIServer pid=72502) INFO 10-10 01:56:34 [launcher.py:42] Route: /v1/chat/completions, Methods: POST 2025-10-10T01:56:34.3462555Z (APIServer pid=72502) INFO 10-10 01:56:34 [launcher.py:42] Route: /v1/completions, Methods: POST 2025-10-10T01:56:34.3463124Z (APIServer pid=72502) INFO 10-10 01:56:34 [launcher.py:42] Route: /v1/embeddings, Methods: POST 2025-10-10T01:56:34.3463661Z (APIServer pid=72502) INFO 10-10 01:56:34 [launcher.py:42] Route: /pooling, Methods: POST 2025-10-10T01:56:34.3464234Z (APIServer pid=72502) INFO 10-10 01:56:34 [launcher.py:42] Route: /classify, Methods: POST 2025-10-10T01:56:34.3464784Z (APIServer pid=72502) INFO 10-10 01:56:34 [launcher.py:42] Route: /score, Methods: POST 2025-10-10T01:56:34.3465304Z (APIServer pid=72502) INFO 10-10 01:56:34 [launcher.py:42] Route: /v1/score, Methods: POST 2025-10-10T01:56:34.3465879Z (APIServer pid=72502) INFO 10-10 01:56:34 [launcher.py:42] Route: /v1/audio/transcriptions, Methods: POST 2025-10-10T01:56:34.3466480Z (APIServer pid=72502) INFO 10-10 01:56:34 [launcher.py:42] Route: /v1/audio/translations, Methods: POST 2025-10-10T01:56:34.3467031Z (APIServer pid=72502) INFO 10-10 01:56:34 [launcher.py:42] Route: /rerank, Methods: POST 2025-10-10T01:56:34.3467557Z (APIServer pid=72502) INFO 10-10 01:56:34 [launcher.py:42] Route: /v1/rerank, Methods: POST 2025-10-10T01:56:34.3468097Z (APIServer pid=72502) INFO 10-10 01:56:34 [launcher.py:42] Route: /v2/rerank, Methods: POST 2025-10-10T01:56:34.3468683Z (APIServer pid=72502) INFO 10-10 01:56:34 [launcher.py:42] Route: /scale_elastic_ep, Methods: POST 2025-10-10T01:56:34.3469345Z (APIServer pid=72502) INFO 10-10 01:56:34 [launcher.py:42] Route: /is_scaling_elastic_ep, Methods: POST 2025-10-10T01:56:34.3469926Z (APIServer pid=72502) INFO 10-10 01:56:34 [launcher.py:42] Route: /invocations, Methods: POST 2025-10-10T01:56:34.3470470Z (APIServer pid=72502) INFO 10-10 01:56:34 [launcher.py:42] Route: /metrics, Methods: GET 2025-10-10T01:56:34.3601779Z (APIServer pid=72502) INFO: Started server process [72502] 2025-10-10T01:56:34.3603028Z (APIServer pid=72502) INFO: Waiting for application startup. 2025-10-10T01:56:34.6130837Z (APIServer pid=72502) INFO: Application startup complete. 2025-10-10T01:56:34.9616855Z (APIServer pid=72502) INFO: 127.0.0.1:56976 - "GET /health HTTP/1.1" 200 OK 2025-10-10T01:56:34.9705527Z (APIServer pid=72502) INFO: 127.0.0.1:56990 - "GET /v1/models HTTP/1.1" 200 OK 2025-10-10T01:56:35.2009262Z (APIServer pid=72502) INFO: 127.0.0.1:56990 - "POST /v1/completions HTTP/1.1" 200 OK 2025-10-10T01:56:35.3162748Z (APIServer pid=72502) INFO: 127.0.0.1:56990 - "POST /v1/completions HTTP/1.1" 200 OK 2025-10-10T01:56:35.4611726Z (APIServer pid=72502) INFO: 127.0.0.1:56990 - "POST /v1/completions HTTP/1.1" 200 OK 2025-10-10T01:56:35.6135451Z (APIServer pid=72502) INFO: 127.0.0.1:56990 - "POST /v1/completions HTTP/1.1" 200 OK 2025-10-10T01:56:35.7564146Z (APIServer pid=72502) INFO: 127.0.0.1:56990 - "POST /v1/completions HTTP/1.1" 200 OK 2025-10-10T01:56:35.7638792Z (APIServer pid=72502) INFO: 127.0.0.1:56990 - "POST /v1/completions HTTP/1.1" 200 OK 2025-10-10T01:56:35.9106742Z (APIServer pid=72502) INFO 10-10 01:56:35 [launcher.py:99] Shutting down FastAPI HTTP server. 2025-10-10T01:56:35.9109691Z (Worker_TP0 pid=72744) INFO 10-10 01:56:35 [multiproc_executor.py:558] Parent process exited, terminating worker 2025-10-10T01:56:35.9111243Z (Worker_TP0 pid=72744) INFO 10-10 01:56:35 [multiproc_executor.py:599] WorkerProc shutting down. 2025-10-10T01:56:35.9114474Z (Worker_TP1 pid=72745) INFO 10-10 01:56:35 [multiproc_executor.py:558] Parent process exited, terminating worker 2025-10-10T01:56:35.9115379Z (Worker_TP1 pid=72745) INFO 10-10 01:56:35 [multiproc_executor.py:599] WorkerProc shutting down. 2025-10-10T01:56:38.0335694Z (APIServer pid=72502) INFO: Shutting down 2025-10-10T01:56:38.1339287Z (APIServer pid=72502) INFO: Waiting for application shutdown. 2025-10-10T01:56:38.1340633Z (APIServer pid=72502) INFO: Application shutdown complete. 2025-10-10T01:56:38.7585664Z PASSED 2025-10-10T01:56:38.7701502Z distributed/test_sequence_parallel.py::test_tp_sp_generation[RedHatAI/Meta-Llama-3.1-8B-Instruct-FP8-parallel_setup18-mp-1-auto-test_options18] Fork a new process to run a test 73262 2025-10-10T01:56:38.7716956Z Fork a new process to run a test 0 2025-10-10T01:56:39.4499256Z INFO 10-10 01:56:39 [model.py:551] Resolved architecture: LlamaForCausalLM 2025-10-10T01:56:39.4500151Z `torch_dtype` is deprecated! Use `dtype` instead! 2025-10-10T01:56:39.4501724Z WARNING 10-10 01:56:39 [model.py:1768] Casting torch.bfloat16 to torch.float16. 2025-10-10T01:56:39.4502578Z INFO 10-10 01:56:39 [model.py:1545] Using max model len 2048 2025-10-10T01:56:39.5544664Z Launching RemoteOpenAIServer with: vllm serve RedHatAI/Meta-Llama-3.1-8B-Instruct-FP8 --dtype float16 --max-model-len 2048 --max-num-seqs 8 --enforce-eager --tokenizer-mode auto --tensor-parallel-size 2 --pipeline-parallel-size 1 --distributed-executor-backend mp --compilation_config {"level": 3, "custom_ops": ["+rms_norm"], "compile_sizes": [4, 8], "pass_config": {"enable_sequence_parallelism": true, "enable_fusion": true, "enable_noop": true}} --load-format dummy --port 44817 --seed 0 2025-10-10T01:56:40.0632392Z /opt/conda/envs/py_3.12/lib/python3.12/site-packages/torch/cuda/__init__.py:64: FutureWarning: The pynvml package is deprecated. Please install nvidia-ml-py instead. If you did not install pynvml directly, please report this to the maintainers of the package that installed pynvml for you. 2025-10-10T01:56:40.0633635Z import pynvml # type: ignore[import] 2025-10-10T01:56:42.2974684Z INFO 10-10 01:56:42 [__init__.py:215] Automatically detected platform cuda. 2025-10-10T01:56:46.2499476Z (APIServer pid=73311) INFO 10-10 01:56:46 [api_server.py:1839] vLLM API server version 0.11.0rc2.dev157+g0ad9951c4.d20251010 2025-10-10T01:56:46.2525702Z (APIServer pid=73311) INFO 10-10 01:56:46 [utils.py:233] non-default args: {'model_tag': 'RedHatAI/Meta-Llama-3.1-8B-Instruct-FP8', 'port': 44817, 'model': 'RedHatAI/Meta-Llama-3.1-8B-Instruct-FP8', 'dtype': 'float16', 'seed': 0, 'max_model_len': 2048, 'enforce_eager': True, 'load_format': 'dummy', 'distributed_executor_backend': 'mp', 'tensor_parallel_size': 2, 'max_num_seqs': 8, 'compilation_config': {'level': 3, 'debug_dump_path': None, 'cache_dir': '', 'backend': '', 'custom_ops': ['+rms_norm'], 'splitting_ops': None, 'use_inductor': True, 'compile_sizes': [4, 8], 'inductor_compile_config': {'enable_auto_functionalized_v2': False}, 'inductor_passes': {}, 'cudagraph_mode': None, 'use_cudagraph': True, 'cudagraph_num_of_warmups': 0, 'cudagraph_capture_sizes': None, 'cudagraph_copy_inputs': False, 'full_cuda_graph': False, 'use_inductor_graph_partition': False, 'pass_config': {'enable_fusion': True, 'enable_noop': True, 'enable_sequence_parallelism': True}, 'max_capture_size': None, 'local_cache_dir': None}} 2025-10-10T01:56:46.5744557Z (APIServer pid=73311) INFO 10-10 01:56:46 [model.py:551] Resolved architecture: LlamaForCausalLM 2025-10-10T01:56:46.5745271Z (APIServer pid=73311) `torch_dtype` is deprecated! Use `dtype` instead! 2025-10-10T01:56:46.5746005Z (APIServer pid=73311) WARNING 10-10 01:56:46 [model.py:1768] Casting torch.bfloat16 to torch.float16. 2025-10-10T01:56:46.5749355Z (APIServer pid=73311) INFO 10-10 01:56:46 [model.py:1545] Using max model len 2048 2025-10-10T01:56:46.9085241Z (APIServer pid=73311) INFO 10-10 01:56:46 [scheduler.py:222] Chunked prefill is enabled with max_num_batched_tokens=2048. 2025-10-10T01:56:46.9453160Z (APIServer pid=73311) INFO 10-10 01:56:46 [vllm.py:345] Cudagraph is disabled under eager mode 2025-10-10T01:56:47.9365489Z /opt/conda/envs/py_3.12/lib/python3.12/site-packages/torch/cuda/__init__.py:64: FutureWarning: The pynvml package is deprecated. Please install nvidia-ml-py instead. If you did not install pynvml directly, please report this to the maintainers of the package that installed pynvml for you. 2025-10-10T01:56:47.9367859Z import pynvml # type: ignore[import] 2025-10-10T01:56:50.2390540Z INFO 10-10 01:56:50 [__init__.py:215] Automatically detected platform cuda. 2025-10-10T01:56:54.3451141Z (EngineCore_DP0 pid=73434) INFO 10-10 01:56:54 [core.py:648] Waiting for init message from front-end. 2025-10-10T01:56:54.3725339Z (EngineCore_DP0 pid=73434) INFO 10-10 01:56:54 [core.py:78] Initializing a V1 LLM engine (v0.11.0rc2.dev157+g0ad9951c4.d20251010) with config: model='RedHatAI/Meta-Llama-3.1-8B-Instruct-FP8', speculative_config=None, tokenizer='RedHatAI/Meta-Llama-3.1-8B-Instruct-FP8', skip_tokenizer_init=False, tokenizer_mode=auto, revision=None, tokenizer_revision=None, trust_remote_code=False, dtype=torch.float16, max_seq_len=2048, download_dir=None, load_format=dummy, tensor_parallel_size=2, pipeline_parallel_size=1, data_parallel_size=1, disable_custom_all_reduce=False, quantization=compressed-tensors, enforce_eager=True, kv_cache_dtype=auto, device_config=cuda, structured_outputs_config=StructuredOutputsConfig(backend='auto', disable_fallback=False, disable_any_whitespace=False, disable_additional_properties=False, reasoning_parser=''), observability_config=ObservabilityConfig(show_hidden_metrics_for_version=None, otlp_traces_endpoint=None, collect_detailed_traces=None), seed=0, served_model_name=RedHatAI/Meta-Llama-3.1-8B-Instruct-FP8, enable_prefix_caching=True, chunked_prefill_enabled=True, pooler_config=None, compilation_config={'level': 3, 'debug_dump_path': None, 'cache_dir': '', 'backend': '', 'custom_ops': ['+rms_norm', '+rms_norm'], 'splitting_ops': ['vllm.unified_attention', 'vllm.unified_attention_with_output', 'vllm.mamba_mixer2', 'vllm.mamba_mixer', 'vllm.short_conv', 'vllm.linear_attention', 'vllm.plamo2_mamba_mixer', 'vllm.gdn_attention', 'vllm.sparse_attn_indexer'], 'use_inductor': True, 'compile_sizes': [8, 4], 'inductor_compile_config': {'enable_auto_functionalized_v2': False}, 'inductor_passes': {}, 'cudagraph_mode': , 'use_cudagraph': False, 'cudagraph_num_of_warmups': 0, 'cudagraph_capture_sizes': [], 'cudagraph_copy_inputs': False, 'full_cuda_graph': False, 'use_inductor_graph_partition': False, 'pass_config': {'enable_fusion': True, 'enable_noop': True, 'enable_sequence_parallelism': True}, 'max_capture_size': 0, 'local_cache_dir': None} 2025-10-10T01:56:54.3731674Z (EngineCore_DP0 pid=73434) WARNING 10-10 01:56:54 [multiproc_executor.py:720] Reducing Torch parallelism from 24 threads to 1 to avoid unnecessary CPU contention. Set OMP_NUM_THREADS in the external environment to tune this value as needed. 2025-10-10T01:56:54.3740567Z (EngineCore_DP0 pid=73434) INFO 10-10 01:56:54 [shm_broadcast.py:289] vLLM message queue communication handle: Handle(local_reader_ranks=[0, 1], buffer_handle=(2, 16777216, 10, 'psm_8a89b531'), local_subscribe_addr='ipc:///tmp/b336d22f-8e05-466a-b8e7-f69e413272a7', remote_subscribe_addr=None, remote_addr_ipv6=False) 2025-10-10T01:56:54.8625544Z /opt/conda/envs/py_3.12/lib/python3.12/site-packages/torch/cuda/__init__.py:64: FutureWarning: The pynvml package is deprecated. Please install nvidia-ml-py instead. If you did not install pynvml directly, please report this to the maintainers of the package that installed pynvml for you. 2025-10-10T01:56:54.8626780Z import pynvml # type: ignore[import] 2025-10-10T01:56:54.8785679Z /opt/conda/envs/py_3.12/lib/python3.12/site-packages/torch/cuda/__init__.py:64: FutureWarning: The pynvml package is deprecated. Please install nvidia-ml-py instead. If you did not install pynvml directly, please report this to the maintainers of the package that installed pynvml for you. 2025-10-10T01:56:54.8786888Z import pynvml # type: ignore[import] 2025-10-10T01:56:57.1997310Z INFO 10-10 01:56:57 [__init__.py:215] Automatically detected platform cuda. 2025-10-10T01:56:57.2051803Z INFO 10-10 01:56:57 [__init__.py:215] Automatically detected platform cuda. 2025-10-10T01:57:01.9146686Z INFO 10-10 01:57:01 [shm_broadcast.py:289] vLLM message queue communication handle: Handle(local_reader_ranks=[0], buffer_handle=(1, 10485760, 10, 'psm_03d22815'), local_subscribe_addr='ipc:///tmp/6fc31a2e-94b7-4fa1-a2c2-3d1c1b671da0', remote_subscribe_addr=None, remote_addr_ipv6=False) 2025-10-10T01:57:01.9541583Z INFO 10-10 01:57:01 [shm_broadcast.py:289] vLLM message queue communication handle: Handle(local_reader_ranks=[0], buffer_handle=(1, 10485760, 10, 'psm_2663a775'), local_subscribe_addr='ipc:///tmp/3b8a2ed7-7fb5-40f8-b956-c55e8fe1f8a4', remote_subscribe_addr=None, remote_addr_ipv6=False) 2025-10-10T01:57:02.1351368Z [Gloo] Rank 1 is connected to 1 peer ranks. Expected number of connected peer ranks is : 1 2025-10-10T01:57:02.1352659Z [Gloo] Rank 0 is connected to 1 peer ranks. Expected number of connected peer ranks is : 1 2025-10-10T01:57:02.1398278Z [Gloo] Rank 0 is connected to 1 peer ranks. Expected number of connected peer ranks is : 1 2025-10-10T01:57:02.1399594Z [Gloo] Rank 1 is connected to 1 peer ranks. Expected number of connected peer ranks is : 1 2025-10-10T01:57:02.1440770Z INFO 10-10 01:57:02 [__init__.py:1384] Found nccl from library libnccl.so.2 2025-10-10T01:57:02.1442629Z INFO 10-10 01:57:02 [__init__.py:1384] Found nccl from library libnccl.so.2 2025-10-10T01:57:02.1443069Z INFO 10-10 01:57:02 [pynccl.py:104] vLLM is using nccl==2.27.5 2025-10-10T01:57:02.1444338Z INFO 10-10 01:57:02 [pynccl.py:104] vLLM is using nccl==2.27.5 2025-10-10T01:57:02.3187900Z WARNING 10-10 01:57:02 [symm_mem.py:58] SymmMemCommunicator: Device capability 8.9 not supported, communicator is not available. 2025-10-10T01:57:02.3189513Z WARNING 10-10 01:57:02 [symm_mem.py:58] SymmMemCommunicator: Device capability 8.9 not supported, communicator is not available. 2025-10-10T01:57:02.3218412Z INFO 10-10 01:57:02 [custom_all_reduce.py:35] Skipping P2P check and trusting the driver's P2P report. 2025-10-10T01:57:02.3219326Z INFO 10-10 01:57:02 [custom_all_reduce.py:35] Skipping P2P check and trusting the driver's P2P report. 2025-10-10T01:57:02.3220373Z WARNING 10-10 01:57:02 [custom_all_reduce.py:154] Custom allreduce is disabled because your platform lacks GPU P2P capability or P2P test failed. To silence this warning, specify disable_custom_all_reduce=True explicitly. 2025-10-10T01:57:02.3221768Z WARNING 10-10 01:57:02 [custom_all_reduce.py:154] Custom allreduce is disabled because your platform lacks GPU P2P capability or P2P test failed. To silence this warning, specify disable_custom_all_reduce=True explicitly. 2025-10-10T01:57:02.4240380Z INFO 10-10 01:57:02 [shm_broadcast.py:289] vLLM message queue communication handle: Handle(local_reader_ranks=[1], buffer_handle=(1, 4194304, 6, 'psm_9fc71cb4'), local_subscribe_addr='ipc:///tmp/6a73f225-680c-4c8b-9967-36da8945defa', remote_subscribe_addr=None, remote_addr_ipv6=False) 2025-10-10T01:57:02.4270314Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-10-10T01:57:02.4270944Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-10-10T01:57:02.4278140Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-10-10T01:57:02.4286551Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-10-10T01:57:02.5235569Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-10-10T01:57:02.5244486Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-10-10T01:57:02.5256466Z [Gloo] Rank 1 is connected to 1 peer ranks. Expected number of connected peer ranks is : 1 2025-10-10T01:57:02.5257063Z [Gloo] Rank 0 is connected to 1 peer ranks. Expected number of connected peer ranks is : 1 2025-10-10T01:57:02.5259209Z INFO 10-10 01:57:02 [__init__.py:1384] Found nccl from library libnccl.so.2 2025-10-10T01:57:02.5259674Z INFO 10-10 01:57:02 [pynccl.py:104] vLLM is using nccl==2.27.5 2025-10-10T01:57:02.5260122Z INFO 10-10 01:57:02 [__init__.py:1384] Found nccl from library libnccl.so.2 2025-10-10T01:57:02.5260525Z INFO 10-10 01:57:02 [pynccl.py:104] vLLM is using nccl==2.27.5 2025-10-10T01:57:02.5612051Z INFO 10-10 01:57:02 [parallel_state.py:1208] rank 0 in world size 2 is assigned as DP rank 0, PP rank 0, TP rank 0, EP rank 0 2025-10-10T01:57:02.5612815Z INFO 10-10 01:57:02 [parallel_state.py:1208] rank 1 in world size 2 is assigned as DP rank 0, PP rank 0, TP rank 1, EP rank 1 2025-10-10T01:57:02.8220097Z INFO 10-10 01:57:02 [topk_topp_sampler.py:55] Using FlashInfer for top-p & top-k sampling. 2025-10-10T01:57:02.8286610Z INFO 10-10 01:57:02 [topk_topp_sampler.py:55] Using FlashInfer for top-p & top-k sampling. 2025-10-10T01:57:02.8582782Z (Worker_TP1 pid=73554) INFO 10-10 01:57:02 [gpu_model_runner.py:2707] Starting to load model RedHatAI/Meta-Llama-3.1-8B-Instruct-FP8... 2025-10-10T01:57:02.8648182Z (Worker_TP0 pid=73553) INFO 10-10 01:57:02 [gpu_model_runner.py:2707] Starting to load model RedHatAI/Meta-Llama-3.1-8B-Instruct-FP8... 2025-10-10T01:57:03.0895137Z (Worker_TP1 pid=73554) INFO 10-10 01:57:03 [gpu_model_runner.py:2739] Loading model from scratch... 2025-10-10T01:57:03.1006620Z (Worker_TP0 pid=73553) INFO 10-10 01:57:03 [gpu_model_runner.py:2739] Loading model from scratch... 2025-10-10T01:57:03.1811049Z (Worker_TP1 pid=73554) INFO 10-10 01:57:03 [cuda.py:361] Using Flash Attention backend on V1 engine. 2025-10-10T01:57:03.1917233Z (Worker_TP0 pid=73553) INFO 10-10 01:57:03 [cuda.py:361] Using Flash Attention backend on V1 engine. 2025-10-10T01:57:04.0995690Z (Worker_TP1 pid=73554) INFO 10-10 01:57:04 [gpu_model_runner.py:2758] Model loading took 4.2624 GiB and 0.543121 seconds 2025-10-10T01:57:04.1001290Z (Worker_TP0 pid=73553) INFO 10-10 01:57:04 [gpu_model_runner.py:2758] Model loading took 4.2624 GiB and 0.535453 seconds 2025-10-10T01:57:11.5537502Z (Worker_TP0 pid=73553) INFO 10-10 01:57:11 [backends.py:548] Using cache directory: /var/lib/jenkins/.cache/vllm/torch_compile_cache/6f893e84df/rank_0_0/backbone for vLLM's torch.compile 2025-10-10T01:57:11.5540423Z (Worker_TP0 pid=73553) INFO 10-10 01:57:11 [backends.py:559] Dynamo bytecode transform time: 7.13 s 2025-10-10T01:57:11.7903474Z (Worker_TP1 pid=73554) INFO 10-10 01:57:11 [backends.py:548] Using cache directory: /var/lib/jenkins/.cache/vllm/torch_compile_cache/6f893e84df/rank_1_0/backbone for vLLM's torch.compile 2025-10-10T01:57:11.7909127Z (Worker_TP1 pid=73554) INFO 10-10 01:57:11 [backends.py:559] Dynamo bytecode transform time: 7.35 s 2025-10-10T01:57:14.4036586Z (Worker_TP0 pid=73553) INFO 10-10 01:57:14 [backends.py:197] Cache the graph for dynamic shape for later use 2025-10-10T01:57:14.6827981Z (Worker_TP1 pid=73554) INFO 10-10 01:57:14 [backends.py:197] Cache the graph for dynamic shape for later use 2025-10-10T01:57:41.8550615Z (Worker_TP0 pid=73553) INFO 10-10 01:57:41 [backends.py:218] Compiling a graph for dynamic shape takes 29.74 s 2025-10-10T01:57:42.9467882Z (Worker_TP1 pid=73554) INFO 10-10 01:57:42 [backends.py:218] Compiling a graph for dynamic shape takes 30.59 s 2025-10-10T01:58:04.2054796Z (EngineCore_DP0 pid=73434) INFO 10-10 01:58:04 [shm_broadcast.py:466] No available shared memory broadcast block found in 60 seconds. This typically happens when some processes are hanging or doing some time-consuming work (e.g. compilation). 2025-10-10T01:58:04.2345248Z (Worker_TP1 pid=73554) INFO 10-10 01:58:04 [gpu_worker.py:298] Available KV cache memory: 15.29 GiB 2025-10-10T01:58:04.2631747Z (Worker_TP0 pid=73553) INFO 10-10 01:58:04 [gpu_worker.py:298] Available KV cache memory: 15.29 GiB 2025-10-10T01:58:04.6547518Z (EngineCore_DP0 pid=73434) INFO 10-10 01:58:04 [kv_cache_utils.py:1087] GPU KV cache size: 250,528 tokens 2025-10-10T01:58:04.6548455Z (EngineCore_DP0 pid=73434) INFO 10-10 01:58:04 [kv_cache_utils.py:1091] Maximum concurrency for 2,048 tokens per request: 122.33x 2025-10-10T01:58:04.6549320Z (EngineCore_DP0 pid=73434) INFO 10-10 01:58:04 [kv_cache_utils.py:1087] GPU KV cache size: 250,528 tokens 2025-10-10T01:58:04.6550146Z (EngineCore_DP0 pid=73434) INFO 10-10 01:58:04 [kv_cache_utils.py:1091] Maximum concurrency for 2,048 tokens per request: 122.33x 2025-10-10T01:58:04.6669255Z (Worker_TP1 pid=73554) INFO 10-10 01:58:04 [gpu_worker.py:332] Compile and warming up model for size 8 2025-10-10T01:58:04.6670092Z (Worker_TP0 pid=73553) INFO 10-10 01:58:04 [gpu_worker.py:332] Compile and warming up model for size 8 2025-10-10T01:58:05.8959905Z (Worker_TP0 pid=73553) INFO 10-10 01:58:05 [backends.py:200] Cache the graph of shape 8 for later use 2025-10-10T01:58:05.9067434Z (Worker_TP1 pid=73554) INFO 10-10 01:58:05 [backends.py:200] Cache the graph of shape 8 for later use 2025-10-10T01:58:36.0494513Z (Worker_TP0 pid=73553) INFO 10-10 01:58:36 [backends.py:221] Compiling a graph for shape 8 takes 31.38 s 2025-10-10T01:58:36.1885820Z (Worker_TP1 pid=73554) INFO 10-10 01:58:36 [backends.py:221] Compiling a graph for shape 8 takes 31.52 s 2025-10-10T01:58:36.2731267Z (Worker_TP0 pid=73553) INFO 10-10 01:58:36 [gpu_worker.py:332] Compile and warming up model for size 4 2025-10-10T01:58:36.2746605Z (Worker_TP1 pid=73554) INFO 10-10 01:58:36 [gpu_worker.py:332] Compile and warming up model for size 4 2025-10-10T01:58:37.4793855Z (Worker_TP0 pid=73553) INFO 10-10 01:58:37 [backends.py:200] Cache the graph of shape 4 for later use 2025-10-10T01:58:37.4816072Z (Worker_TP1 pid=73554) INFO 10-10 01:58:37 [backends.py:200] Cache the graph of shape 4 for later use 2025-10-10T01:59:04.6672519Z (EngineCore_DP0 pid=73434) INFO 10-10 01:59:04 [shm_broadcast.py:466] No available shared memory broadcast block found in 60 seconds. This typically happens when some processes are hanging or doing some time-consuming work (e.g. compilation). 2025-10-10T01:59:19.5446674Z (Worker_TP1 pid=73554) INFO 10-10 01:59:19 [backends.py:221] Compiling a graph for shape 4 takes 43.27 s 2025-10-10T01:59:19.5466959Z (Worker_TP1 pid=73554) INFO 10-10 01:59:19 [monitor.py:32] torch.compile takes 112.72 s in total 2025-10-10T01:59:19.5468511Z (Worker_TP0 pid=73553) INFO 10-10 01:59:19 [backends.py:221] Compiling a graph for shape 4 takes 43.27 s 2025-10-10T01:59:19.5492448Z (Worker_TP0 pid=73553) INFO 10-10 01:59:19 [monitor.py:32] torch.compile takes 111.52 s in total 2025-10-10T01:59:19.5749385Z (EngineCore_DP0 pid=73434) INFO 10-10 01:59:19 [core.py:211] init engine (profile, create kv cache, warmup model) took 135.37 seconds 2025-10-10T01:59:20.0949523Z (EngineCore_DP0 pid=73434) INFO 10-10 01:59:20 [vllm.py:345] Cudagraph is disabled under eager mode 2025-10-10T01:59:20.2797229Z (APIServer pid=73311) INFO 10-10 01:59:20 [loggers.py:147] Engine 000: vllm cache_config_info with initialization after num_gpu_blocks is: 15658 2025-10-10T01:59:20.3157805Z (EngineCore_DP0 pid=73434) INFO 10-10 01:59:20 [gc_utils.py:41] GC Debug Config. enabled:False,top_objects:-1 2025-10-10T01:59:20.3367886Z (APIServer pid=73311) INFO 10-10 01:59:20 [api_server.py:1634] Supported_tasks: ['generate'] 2025-10-10T01:59:20.3657706Z (APIServer pid=73311) WARNING 10-10 01:59:20 [model.py:1424] Default sampling parameters have been overridden by the model's Hugging Face generation config recommended from the model creator. If this is not intended, please relaunch vLLM instance with `--generation-config vllm`. 2025-10-10T01:59:20.3660869Z (APIServer pid=73311) INFO 10-10 01:59:20 [serving_responses.py:140] Using default chat sampling params from model: {'temperature': 0.6, 'top_p': 0.9} 2025-10-10T01:59:20.3957551Z (APIServer pid=73311) INFO 10-10 01:59:20 [serving_chat.py:139] Using default chat sampling params from model: {'temperature': 0.6, 'top_p': 0.9} 2025-10-10T01:59:20.4377091Z (APIServer pid=73311) INFO 10-10 01:59:20 [serving_completion.py:76] Using default completion sampling params from model: {'temperature': 0.6, 'top_p': 0.9} 2025-10-10T01:59:20.4379273Z (APIServer pid=73311) INFO 10-10 01:59:20 [api_server.py:1912] Starting vLLM API server 0 on http://0.0.0.0:44817 2025-10-10T01:59:20.4381791Z (APIServer pid=73311) INFO 10-10 01:59:20 [launcher.py:34] Available routes are: 2025-10-10T01:59:20.4382489Z (APIServer pid=73311) INFO 10-10 01:59:20 [launcher.py:42] Route: /openapi.json, Methods: HEAD, GET 2025-10-10T01:59:20.4383058Z (APIServer pid=73311) INFO 10-10 01:59:20 [launcher.py:42] Route: /docs, Methods: HEAD, GET 2025-10-10T01:59:20.4383647Z (APIServer pid=73311) INFO 10-10 01:59:20 [launcher.py:42] Route: /docs/oauth2-redirect, Methods: HEAD, GET 2025-10-10T01:59:20.4384440Z (APIServer pid=73311) INFO 10-10 01:59:20 [launcher.py:42] Route: /redoc, Methods: HEAD, GET 2025-10-10T01:59:20.4384959Z (APIServer pid=73311) INFO 10-10 01:59:20 [launcher.py:42] Route: /health, Methods: GET 2025-10-10T01:59:20.4385482Z (APIServer pid=73311) INFO 10-10 01:59:20 [launcher.py:42] Route: /load, Methods: GET 2025-10-10T01:59:20.4385985Z (APIServer pid=73311) INFO 10-10 01:59:20 [launcher.py:42] Route: /ping, Methods: POST 2025-10-10T01:59:20.4386472Z (APIServer pid=73311) INFO 10-10 01:59:20 [launcher.py:42] Route: /ping, Methods: GET 2025-10-10T01:59:20.4386982Z (APIServer pid=73311) INFO 10-10 01:59:20 [launcher.py:42] Route: /tokenize, Methods: POST 2025-10-10T01:59:20.4387512Z (APIServer pid=73311) INFO 10-10 01:59:20 [launcher.py:42] Route: /detokenize, Methods: POST 2025-10-10T01:59:20.4388210Z (APIServer pid=73311) INFO 10-10 01:59:20 [launcher.py:42] Route: /v1/models, Methods: GET 2025-10-10T01:59:20.4388739Z (APIServer pid=73311) INFO 10-10 01:59:20 [launcher.py:42] Route: /version, Methods: GET 2025-10-10T01:59:20.4389275Z (APIServer pid=73311) INFO 10-10 01:59:20 [launcher.py:42] Route: /v1/responses, Methods: POST 2025-10-10T01:59:20.4389857Z (APIServer pid=73311) INFO 10-10 01:59:20 [launcher.py:42] Route: /v1/responses/{response_id}, Methods: GET 2025-10-10T01:59:20.4390581Z (APIServer pid=73311) INFO 10-10 01:59:20 [launcher.py:42] Route: /v1/responses/{response_id}/cancel, Methods: POST 2025-10-10T01:59:20.4391194Z (APIServer pid=73311) INFO 10-10 01:59:20 [launcher.py:42] Route: /v1/chat/completions, Methods: POST 2025-10-10T01:59:20.4391758Z (APIServer pid=73311) INFO 10-10 01:59:20 [launcher.py:42] Route: /v1/completions, Methods: POST 2025-10-10T01:59:20.4392312Z (APIServer pid=73311) INFO 10-10 01:59:20 [launcher.py:42] Route: /v1/embeddings, Methods: POST 2025-10-10T01:59:20.4392851Z (APIServer pid=73311) INFO 10-10 01:59:20 [launcher.py:42] Route: /pooling, Methods: POST 2025-10-10T01:59:20.4393376Z (APIServer pid=73311) INFO 10-10 01:59:20 [launcher.py:42] Route: /classify, Methods: POST 2025-10-10T01:59:20.4393897Z (APIServer pid=73311) INFO 10-10 01:59:20 [launcher.py:42] Route: /score, Methods: POST 2025-10-10T01:59:20.4394414Z (APIServer pid=73311) INFO 10-10 01:59:20 [launcher.py:42] Route: /v1/score, Methods: POST 2025-10-10T01:59:20.4394981Z (APIServer pid=73311) INFO 10-10 01:59:20 [launcher.py:42] Route: /v1/audio/transcriptions, Methods: POST 2025-10-10T01:59:20.4395567Z (APIServer pid=73311) INFO 10-10 01:59:20 [launcher.py:42] Route: /v1/audio/translations, Methods: POST 2025-10-10T01:59:20.4396110Z (APIServer pid=73311) INFO 10-10 01:59:20 [launcher.py:42] Route: /rerank, Methods: POST 2025-10-10T01:59:20.4396625Z (APIServer pid=73311) INFO 10-10 01:59:20 [launcher.py:42] Route: /v1/rerank, Methods: POST 2025-10-10T01:59:20.4397148Z (APIServer pid=73311) INFO 10-10 01:59:20 [launcher.py:42] Route: /v2/rerank, Methods: POST 2025-10-10T01:59:20.4397688Z (APIServer pid=73311) INFO 10-10 01:59:20 [launcher.py:42] Route: /scale_elastic_ep, Methods: POST 2025-10-10T01:59:20.4398251Z (APIServer pid=73311) INFO 10-10 01:59:20 [launcher.py:42] Route: /is_scaling_elastic_ep, Methods: POST 2025-10-10T01:59:20.4398932Z (APIServer pid=73311) INFO 10-10 01:59:20 [launcher.py:42] Route: /invocations, Methods: POST 2025-10-10T01:59:20.4399454Z (APIServer pid=73311) INFO 10-10 01:59:20 [launcher.py:42] Route: /metrics, Methods: GET 2025-10-10T01:59:20.4530747Z (APIServer pid=73311) INFO: Started server process [73311] 2025-10-10T01:59:20.4532019Z (APIServer pid=73311) INFO: Waiting for application startup. 2025-10-10T01:59:20.6963292Z (APIServer pid=73311) INFO: Application startup complete. 2025-10-10T01:59:20.9140268Z (APIServer pid=73311) INFO: 127.0.0.1:45506 - "GET /health HTTP/1.1" 200 OK 2025-10-10T01:59:21.1385217Z (APIServer pid=73311) INFO: 127.0.0.1:45516 - "GET /v1/models HTTP/1.1" 200 OK 2025-10-10T01:59:21.3016550Z (APIServer pid=73311) INFO: 127.0.0.1:45516 - "POST /v1/completions HTTP/1.1" 200 OK 2025-10-10T01:59:21.4207711Z (APIServer pid=73311) INFO: 127.0.0.1:45516 - "POST /v1/completions HTTP/1.1" 200 OK 2025-10-10T01:59:21.6114475Z (APIServer pid=73311) INFO: 127.0.0.1:45516 - "POST /v1/completions HTTP/1.1" 200 OK 2025-10-10T01:59:21.7641158Z (APIServer pid=73311) INFO: 127.0.0.1:45516 - "POST /v1/completions HTTP/1.1" 200 OK 2025-10-10T01:59:21.9053236Z (APIServer pid=73311) INFO: 127.0.0.1:45516 - "POST /v1/completions HTTP/1.1" 200 OK 2025-10-10T01:59:21.9129398Z (APIServer pid=73311) INFO: 127.0.0.1:45516 - "POST /v1/completions HTTP/1.1" 200 OK 2025-10-10T01:59:22.0348565Z (APIServer pid=73311) INFO 10-10 01:59:22 [launcher.py:99] Shutting down FastAPI HTTP server. 2025-10-10T01:59:22.0351775Z (Worker_TP0 pid=73553) INFO 10-10 01:59:22 [multiproc_executor.py:558] Parent process exited, terminating worker 2025-10-10T01:59:22.0353639Z (Worker_TP0 pid=73553) INFO 10-10 01:59:22 [multiproc_executor.py:599] WorkerProc shutting down. 2025-10-10T01:59:22.0354650Z (Worker_TP1 pid=73554) INFO 10-10 01:59:22 [multiproc_executor.py:558] Parent process exited, terminating worker 2025-10-10T01:59:22.0356517Z (Worker_TP1 pid=73554) INFO 10-10 01:59:22 [multiproc_executor.py:599] WorkerProc shutting down. 2025-10-10T01:59:25.3577953Z (APIServer pid=73311) INFO: Shutting down 2025-10-10T01:59:25.4572503Z (APIServer pid=73311) INFO: Waiting for application shutdown. 2025-10-10T01:59:25.4574369Z (APIServer pid=73311) INFO: Application shutdown complete. 2025-10-10T01:59:25.9425577Z INFO 10-10 01:59:25 [model.py:551] Resolved architecture: LlamaForCausalLM 2025-10-10T01:59:25.9426651Z WARNING 10-10 01:59:25 [model.py:1768] Casting torch.bfloat16 to torch.float16. 2025-10-10T01:59:25.9427508Z INFO 10-10 01:59:25 [model.py:1545] Using max model len 2048 2025-10-10T01:59:25.9433710Z Launching RemoteOpenAIServer with: vllm serve RedHatAI/Meta-Llama-3.1-8B-Instruct-FP8 --dtype float16 --max-model-len 2048 --max-num-seqs 8 --enforce-eager --tokenizer-mode auto --tensor-parallel-size 2 --distributed-executor-backend mp --load-format dummy --port 38063 --seed 0 2025-10-10T01:59:26.4591189Z /opt/conda/envs/py_3.12/lib/python3.12/site-packages/torch/cuda/__init__.py:64: FutureWarning: The pynvml package is deprecated. Please install nvidia-ml-py instead. If you did not install pynvml directly, please report this to the maintainers of the package that installed pynvml for you. 2025-10-10T01:59:26.4592436Z import pynvml # type: ignore[import] 2025-10-10T01:59:28.6956049Z INFO 10-10 01:59:28 [__init__.py:215] Automatically detected platform cuda. 2025-10-10T01:59:32.6528441Z (APIServer pid=77664) INFO 10-10 01:59:32 [api_server.py:1839] vLLM API server version 0.11.0rc2.dev157+g0ad9951c4.d20251010 2025-10-10T01:59:32.6550460Z (APIServer pid=77664) INFO 10-10 01:59:32 [utils.py:233] non-default args: {'model_tag': 'RedHatAI/Meta-Llama-3.1-8B-Instruct-FP8', 'port': 38063, 'model': 'RedHatAI/Meta-Llama-3.1-8B-Instruct-FP8', 'dtype': 'float16', 'seed': 0, 'max_model_len': 2048, 'enforce_eager': True, 'load_format': 'dummy', 'distributed_executor_backend': 'mp', 'tensor_parallel_size': 2, 'max_num_seqs': 8} 2025-10-10T01:59:32.8761408Z (APIServer pid=77664) INFO 10-10 01:59:32 [model.py:551] Resolved architecture: LlamaForCausalLM 2025-10-10T01:59:32.8762393Z (APIServer pid=77664) `torch_dtype` is deprecated! Use `dtype` instead! 2025-10-10T01:59:32.8763226Z (APIServer pid=77664) WARNING 10-10 01:59:32 [model.py:1768] Casting torch.bfloat16 to torch.float16. 2025-10-10T01:59:32.8764286Z (APIServer pid=77664) INFO 10-10 01:59:32 [model.py:1545] Using max model len 2048 2025-10-10T01:59:33.2081636Z (APIServer pid=77664) INFO 10-10 01:59:33 [scheduler.py:222] Chunked prefill is enabled with max_num_batched_tokens=2048. 2025-10-10T01:59:33.2453049Z (APIServer pid=77664) INFO 10-10 01:59:33 [vllm.py:345] Cudagraph is disabled under eager mode 2025-10-10T01:59:34.2225770Z /opt/conda/envs/py_3.12/lib/python3.12/site-packages/torch/cuda/__init__.py:64: FutureWarning: The pynvml package is deprecated. Please install nvidia-ml-py instead. If you did not install pynvml directly, please report this to the maintainers of the package that installed pynvml for you. 2025-10-10T01:59:34.2227060Z import pynvml # type: ignore[import] 2025-10-10T01:59:36.5448684Z INFO 10-10 01:59:36 [__init__.py:215] Automatically detected platform cuda. 2025-10-10T01:59:40.6651418Z (EngineCore_DP0 pid=77787) INFO 10-10 01:59:40 [core.py:648] Waiting for init message from front-end. 2025-10-10T01:59:40.6921839Z (EngineCore_DP0 pid=77787) INFO 10-10 01:59:40 [core.py:78] Initializing a V1 LLM engine (v0.11.0rc2.dev157+g0ad9951c4.d20251010) with config: model='RedHatAI/Meta-Llama-3.1-8B-Instruct-FP8', speculative_config=None, tokenizer='RedHatAI/Meta-Llama-3.1-8B-Instruct-FP8', skip_tokenizer_init=False, tokenizer_mode=auto, revision=None, tokenizer_revision=None, trust_remote_code=False, dtype=torch.float16, max_seq_len=2048, download_dir=None, load_format=dummy, tensor_parallel_size=2, pipeline_parallel_size=1, data_parallel_size=1, disable_custom_all_reduce=False, quantization=compressed-tensors, enforce_eager=True, kv_cache_dtype=auto, device_config=cuda, structured_outputs_config=StructuredOutputsConfig(backend='auto', disable_fallback=False, disable_any_whitespace=False, disable_additional_properties=False, reasoning_parser=''), observability_config=ObservabilityConfig(show_hidden_metrics_for_version=None, otlp_traces_endpoint=None, collect_detailed_traces=None), seed=0, served_model_name=RedHatAI/Meta-Llama-3.1-8B-Instruct-FP8, enable_prefix_caching=True, chunked_prefill_enabled=True, pooler_config=None, compilation_config={'level': 0, 'debug_dump_path': None, 'cache_dir': '', 'backend': '', 'custom_ops': [], 'splitting_ops': None, 'use_inductor': True, 'compile_sizes': [], 'inductor_compile_config': {'enable_auto_functionalized_v2': False}, 'inductor_passes': {}, 'cudagraph_mode': , 'use_cudagraph': False, 'cudagraph_num_of_warmups': 0, 'cudagraph_capture_sizes': [], 'cudagraph_copy_inputs': False, 'full_cuda_graph': False, 'use_inductor_graph_partition': False, 'pass_config': {}, 'max_capture_size': 0, 'local_cache_dir': None} 2025-10-10T01:59:40.6928152Z (EngineCore_DP0 pid=77787) WARNING 10-10 01:59:40 [multiproc_executor.py:720] Reducing Torch parallelism from 24 threads to 1 to avoid unnecessary CPU contention. Set OMP_NUM_THREADS in the external environment to tune this value as needed. 2025-10-10T01:59:40.6930062Z (EngineCore_DP0 pid=77787) INFO 10-10 01:59:40 [shm_broadcast.py:289] vLLM message queue communication handle: Handle(local_reader_ranks=[0, 1], buffer_handle=(2, 16777216, 10, 'psm_f3b75dca'), local_subscribe_addr='ipc:///tmp/9efab43a-a956-4f32-9b3a-c5c94dd09a97', remote_subscribe_addr=None, remote_addr_ipv6=False) 2025-10-10T01:59:41.1734730Z /opt/conda/envs/py_3.12/lib/python3.12/site-packages/torch/cuda/__init__.py:64: FutureWarning: The pynvml package is deprecated. Please install nvidia-ml-py instead. If you did not install pynvml directly, please report this to the maintainers of the package that installed pynvml for you. 2025-10-10T01:59:41.1736561Z import pynvml # type: ignore[import] 2025-10-10T01:59:41.1782824Z /opt/conda/envs/py_3.12/lib/python3.12/site-packages/torch/cuda/__init__.py:64: FutureWarning: The pynvml package is deprecated. Please install nvidia-ml-py instead. If you did not install pynvml directly, please report this to the maintainers of the package that installed pynvml for you. 2025-10-10T01:59:41.1783995Z import pynvml # type: ignore[import] 2025-10-10T01:59:43.5048499Z INFO 10-10 01:59:43 [__init__.py:215] Automatically detected platform cuda. 2025-10-10T01:59:43.5174380Z INFO 10-10 01:59:43 [__init__.py:215] Automatically detected platform cuda. 2025-10-10T01:59:48.1847267Z INFO 10-10 01:59:48 [shm_broadcast.py:289] vLLM message queue communication handle: Handle(local_reader_ranks=[0], buffer_handle=(1, 10485760, 10, 'psm_a92be37e'), local_subscribe_addr='ipc:///tmp/61bc5899-b3d2-49c9-884c-150f5ee63671', remote_subscribe_addr=None, remote_addr_ipv6=False) 2025-10-10T01:59:48.2941244Z INFO 10-10 01:59:48 [shm_broadcast.py:289] vLLM message queue communication handle: Handle(local_reader_ranks=[0], buffer_handle=(1, 10485760, 10, 'psm_50b56f3e'), local_subscribe_addr='ipc:///tmp/3f82cc8b-4788-48cc-ba7e-f1d274f3c503', remote_subscribe_addr=None, remote_addr_ipv6=False) 2025-10-10T01:59:48.4331926Z [Gloo] Rank 0 is connected to 1 peer ranks. Expected number of connected peer ranks is : 1 2025-10-10T01:59:48.4332735Z [Gloo] Rank 1 is connected to 1 peer ranks. Expected number of connected peer ranks is : 1 2025-10-10T01:59:48.4370702Z [Gloo] Rank 1 is connected to 1 peer ranks. Expected number of connected peer ranks is : 1 2025-10-10T01:59:48.4371208Z [Gloo] Rank 0 is connected to 1 peer ranks. Expected number of connected peer ranks is : 1 2025-10-10T01:59:48.4412836Z INFO 10-10 01:59:48 [__init__.py:1384] Found nccl from library libnccl.so.2 2025-10-10T01:59:48.4414056Z INFO 10-10 01:59:48 [__init__.py:1384] Found nccl from library libnccl.so.2 2025-10-10T01:59:48.4414512Z INFO 10-10 01:59:48 [pynccl.py:104] vLLM is using nccl==2.27.5 2025-10-10T01:59:48.4417235Z INFO 10-10 01:59:48 [pynccl.py:104] vLLM is using nccl==2.27.5 2025-10-10T01:59:48.6157450Z WARNING 10-10 01:59:48 [symm_mem.py:58] SymmMemCommunicator: Device capability 8.9 not supported, communicator is not available. 2025-10-10T01:59:48.6159298Z WARNING 10-10 01:59:48 [symm_mem.py:58] SymmMemCommunicator: Device capability 8.9 not supported, communicator is not available. 2025-10-10T01:59:48.6181474Z INFO 10-10 01:59:48 [custom_all_reduce.py:35] Skipping P2P check and trusting the driver's P2P report. 2025-10-10T01:59:48.6182714Z INFO 10-10 01:59:48 [custom_all_reduce.py:35] Skipping P2P check and trusting the driver's P2P report. 2025-10-10T01:59:48.6185485Z WARNING 10-10 01:59:48 [custom_all_reduce.py:154] Custom allreduce is disabled because your platform lacks GPU P2P capability or P2P test failed. To silence this warning, specify disable_custom_all_reduce=True explicitly. 2025-10-10T01:59:48.6188223Z WARNING 10-10 01:59:48 [custom_all_reduce.py:154] Custom allreduce is disabled because your platform lacks GPU P2P capability or P2P test failed. To silence this warning, specify disable_custom_all_reduce=True explicitly. 2025-10-10T01:59:48.6639386Z INFO 10-10 01:59:48 [shm_broadcast.py:289] vLLM message queue communication handle: Handle(local_reader_ranks=[1], buffer_handle=(1, 4194304, 6, 'psm_a42c6414'), local_subscribe_addr='ipc:///tmp/7445ec39-cc9b-4964-9ead-d11672579c17', remote_subscribe_addr=None, remote_addr_ipv6=False) 2025-10-10T01:59:48.6670450Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-10-10T01:59:48.6671848Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-10-10T01:59:48.6678905Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-10-10T01:59:48.6685106Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-10-10T01:59:48.7036765Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-10-10T01:59:48.7044684Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-10-10T01:59:48.7055001Z [Gloo] Rank 0 is connected to 1 peer ranks. Expected number of connected peer ranks is : 1 2025-10-10T01:59:48.7055461Z [Gloo] Rank 1 is connected to 1 peer ranks. Expected number of connected peer ranks is : 1 2025-10-10T01:59:48.7059687Z INFO 10-10 01:59:48 [__init__.py:1384] Found nccl from library libnccl.so.2 2025-10-10T01:59:48.7060064Z INFO 10-10 01:59:48 [__init__.py:1384] Found nccl from library libnccl.so.2 2025-10-10T01:59:48.7060412Z INFO 10-10 01:59:48 [pynccl.py:104] vLLM is using nccl==2.27.5 2025-10-10T01:59:48.7060728Z INFO 10-10 01:59:48 [pynccl.py:104] vLLM is using nccl==2.27.5 2025-10-10T01:59:48.7389228Z INFO 10-10 01:59:48 [parallel_state.py:1208] rank 1 in world size 2 is assigned as DP rank 0, PP rank 0, TP rank 1, EP rank 1 2025-10-10T01:59:48.7389970Z INFO 10-10 01:59:48 [parallel_state.py:1208] rank 0 in world size 2 is assigned as DP rank 0, PP rank 0, TP rank 0, EP rank 0 2025-10-10T01:59:49.0053499Z INFO 10-10 01:59:49 [topk_topp_sampler.py:55] Using FlashInfer for top-p & top-k sampling. 2025-10-10T01:59:49.0142867Z INFO 10-10 01:59:49 [topk_topp_sampler.py:55] Using FlashInfer for top-p & top-k sampling. 2025-10-10T01:59:49.0412541Z (Worker_TP0 pid=77906) INFO 10-10 01:59:49 [gpu_model_runner.py:2707] Starting to load model RedHatAI/Meta-Llama-3.1-8B-Instruct-FP8... 2025-10-10T01:59:49.0503624Z (Worker_TP1 pid=77907) INFO 10-10 01:59:49 [gpu_model_runner.py:2707] Starting to load model RedHatAI/Meta-Llama-3.1-8B-Instruct-FP8... 2025-10-10T01:59:49.2728596Z (Worker_TP0 pid=77906) INFO 10-10 01:59:49 [gpu_model_runner.py:2739] Loading model from scratch... 2025-10-10T01:59:49.2875157Z (Worker_TP1 pid=77907) INFO 10-10 01:59:49 [gpu_model_runner.py:2739] Loading model from scratch... 2025-10-10T01:59:49.3646107Z (Worker_TP0 pid=77906) INFO 10-10 01:59:49 [cuda.py:361] Using Flash Attention backend on V1 engine. 2025-10-10T01:59:49.3787807Z (Worker_TP1 pid=77907) INFO 10-10 01:59:49 [cuda.py:361] Using Flash Attention backend on V1 engine. 2025-10-10T01:59:50.2282711Z (Worker_TP0 pid=77906) INFO 10-10 01:59:50 [gpu_model_runner.py:2758] Model loading took 4.2624 GiB and 0.492568 seconds 2025-10-10T01:59:50.2454392Z (Worker_TP1 pid=77907) INFO 10-10 01:59:50 [gpu_model_runner.py:2758] Model loading took 4.2624 GiB and 0.493969 seconds 2025-10-10T01:59:55.6586618Z (Worker_TP1 pid=77907) INFO 10-10 01:59:55 [gpu_worker.py:298] Available KV cache memory: 15.32 GiB 2025-10-10T01:59:55.6654878Z (Worker_TP0 pid=77906) INFO 10-10 01:59:55 [gpu_worker.py:298] Available KV cache memory: 15.32 GiB 2025-10-10T01:59:55.9028195Z (EngineCore_DP0 pid=77787) INFO 10-10 01:59:55 [kv_cache_utils.py:1087] GPU KV cache size: 250,976 tokens 2025-10-10T01:59:55.9029171Z (EngineCore_DP0 pid=77787) INFO 10-10 01:59:55 [kv_cache_utils.py:1091] Maximum concurrency for 2,048 tokens per request: 122.55x 2025-10-10T01:59:55.9030016Z (EngineCore_DP0 pid=77787) INFO 10-10 01:59:55 [kv_cache_utils.py:1087] GPU KV cache size: 250,976 tokens 2025-10-10T01:59:55.9030841Z (EngineCore_DP0 pid=77787) INFO 10-10 01:59:55 [kv_cache_utils.py:1091] Maximum concurrency for 2,048 tokens per request: 122.55x 2025-10-10T01:59:56.0339110Z (EngineCore_DP0 pid=77787) INFO 10-10 01:59:56 [core.py:211] init engine (profile, create kv cache, warmup model) took 5.75 seconds 2025-10-10T01:59:56.5022086Z (EngineCore_DP0 pid=77787) INFO 10-10 01:59:56 [vllm.py:345] Cudagraph is disabled under eager mode 2025-10-10T01:59:56.6869131Z (APIServer pid=77664) INFO 10-10 01:59:56 [loggers.py:147] Engine 000: vllm cache_config_info with initialization after num_gpu_blocks is: 15686 2025-10-10T01:59:56.7223678Z (EngineCore_DP0 pid=77787) INFO 10-10 01:59:56 [gc_utils.py:41] GC Debug Config. enabled:False,top_objects:-1 2025-10-10T01:59:56.7436521Z (APIServer pid=77664) INFO 10-10 01:59:56 [api_server.py:1634] Supported_tasks: ['generate'] 2025-10-10T01:59:56.7786781Z (APIServer pid=77664) WARNING 10-10 01:59:56 [model.py:1424] Default sampling parameters have been overridden by the model's Hugging Face generation config recommended from the model creator. If this is not intended, please relaunch vLLM instance with `--generation-config vllm`. 2025-10-10T01:59:56.7788370Z (APIServer pid=77664) INFO 10-10 01:59:56 [serving_responses.py:140] Using default chat sampling params from model: {'temperature': 0.6, 'top_p': 0.9} 2025-10-10T01:59:56.8059011Z (APIServer pid=77664) INFO 10-10 01:59:56 [serving_chat.py:139] Using default chat sampling params from model: {'temperature': 0.6, 'top_p': 0.9} 2025-10-10T01:59:56.8439768Z (APIServer pid=77664) INFO 10-10 01:59:56 [serving_completion.py:76] Using default completion sampling params from model: {'temperature': 0.6, 'top_p': 0.9} 2025-10-10T01:59:56.8441038Z (APIServer pid=77664) INFO 10-10 01:59:56 [api_server.py:1912] Starting vLLM API server 0 on http://0.0.0.0:38063 2025-10-10T01:59:56.8442158Z (APIServer pid=77664) INFO 10-10 01:59:56 [launcher.py:34] Available routes are: 2025-10-10T01:59:56.8443672Z (APIServer pid=77664) INFO 10-10 01:59:56 [launcher.py:42] Route: /openapi.json, Methods: GET, HEAD 2025-10-10T01:59:56.8444940Z (APIServer pid=77664) INFO 10-10 01:59:56 [launcher.py:42] Route: /docs, Methods: GET, HEAD 2025-10-10T01:59:56.8445816Z (APIServer pid=77664) INFO 10-10 01:59:56 [launcher.py:42] Route: /docs/oauth2-redirect, Methods: GET, HEAD 2025-10-10T01:59:56.8446650Z (APIServer pid=77664) INFO 10-10 01:59:56 [launcher.py:42] Route: /redoc, Methods: GET, HEAD 2025-10-10T01:59:56.8447355Z (APIServer pid=77664) INFO 10-10 01:59:56 [launcher.py:42] Route: /health, Methods: GET 2025-10-10T01:59:56.8448085Z (APIServer pid=77664) INFO 10-10 01:59:56 [launcher.py:42] Route: /load, Methods: GET 2025-10-10T01:59:56.8448619Z (APIServer pid=77664) INFO 10-10 01:59:56 [launcher.py:42] Route: /ping, Methods: POST 2025-10-10T01:59:56.8449130Z (APIServer pid=77664) INFO 10-10 01:59:56 [launcher.py:42] Route: /ping, Methods: GET 2025-10-10T01:59:56.8449657Z (APIServer pid=77664) INFO 10-10 01:59:56 [launcher.py:42] Route: /tokenize, Methods: POST 2025-10-10T01:59:56.8450196Z (APIServer pid=77664) INFO 10-10 01:59:56 [launcher.py:42] Route: /detokenize, Methods: POST 2025-10-10T01:59:56.8450844Z (APIServer pid=77664) INFO 10-10 01:59:56 [launcher.py:42] Route: /v1/models, Methods: GET 2025-10-10T01:59:56.8451377Z (APIServer pid=77664) INFO 10-10 01:59:56 [launcher.py:42] Route: /version, Methods: GET 2025-10-10T01:59:56.8451923Z (APIServer pid=77664) INFO 10-10 01:59:56 [launcher.py:42] Route: /v1/responses, Methods: POST 2025-10-10T01:59:56.8452506Z (APIServer pid=77664) INFO 10-10 01:59:56 [launcher.py:42] Route: /v1/responses/{response_id}, Methods: GET 2025-10-10T01:59:56.8453154Z (APIServer pid=77664) INFO 10-10 01:59:56 [launcher.py:42] Route: /v1/responses/{response_id}/cancel, Methods: POST 2025-10-10T01:59:56.8453768Z (APIServer pid=77664) INFO 10-10 01:59:56 [launcher.py:42] Route: /v1/chat/completions, Methods: POST 2025-10-10T01:59:56.8454348Z (APIServer pid=77664) INFO 10-10 01:59:56 [launcher.py:42] Route: /v1/completions, Methods: POST 2025-10-10T01:59:56.8454905Z (APIServer pid=77664) INFO 10-10 01:59:56 [launcher.py:42] Route: /v1/embeddings, Methods: POST 2025-10-10T01:59:56.8455443Z (APIServer pid=77664) INFO 10-10 01:59:56 [launcher.py:42] Route: /pooling, Methods: POST 2025-10-10T01:59:56.8456001Z (APIServer pid=77664) INFO 10-10 01:59:56 [launcher.py:42] Route: /classify, Methods: POST 2025-10-10T01:59:56.8456655Z (APIServer pid=77664) INFO 10-10 01:59:56 [launcher.py:42] Route: /score, Methods: POST 2025-10-10T01:59:56.8457175Z (APIServer pid=77664) INFO 10-10 01:59:56 [launcher.py:42] Route: /v1/score, Methods: POST 2025-10-10T01:59:56.8457750Z (APIServer pid=77664) INFO 10-10 01:59:56 [launcher.py:42] Route: /v1/audio/transcriptions, Methods: POST 2025-10-10T01:59:56.8458351Z (APIServer pid=77664) INFO 10-10 01:59:56 [launcher.py:42] Route: /v1/audio/translations, Methods: POST 2025-10-10T01:59:56.8458899Z (APIServer pid=77664) INFO 10-10 01:59:56 [launcher.py:42] Route: /rerank, Methods: POST 2025-10-10T01:59:56.8459413Z (APIServer pid=77664) INFO 10-10 01:59:56 [launcher.py:42] Route: /v1/rerank, Methods: POST 2025-10-10T01:59:56.8460015Z (APIServer pid=77664) INFO 10-10 01:59:56 [launcher.py:42] Route: /v2/rerank, Methods: POST 2025-10-10T01:59:56.8460566Z (APIServer pid=77664) INFO 10-10 01:59:56 [launcher.py:42] Route: /scale_elastic_ep, Methods: POST 2025-10-10T01:59:56.8461165Z (APIServer pid=77664) INFO 10-10 01:59:56 [launcher.py:42] Route: /is_scaling_elastic_ep, Methods: POST 2025-10-10T01:59:56.8461726Z (APIServer pid=77664) INFO 10-10 01:59:56 [launcher.py:42] Route: /invocations, Methods: POST 2025-10-10T01:59:56.8462287Z (APIServer pid=77664) INFO 10-10 01:59:56 [launcher.py:42] Route: /metrics, Methods: GET 2025-10-10T01:59:56.8596320Z (APIServer pid=77664) INFO: Started server process [77664] 2025-10-10T01:59:56.8596980Z (APIServer pid=77664) INFO: Waiting for application startup. 2025-10-10T01:59:57.1230664Z (APIServer pid=77664) INFO: Application startup complete. 2025-10-10T01:59:57.5203608Z (APIServer pid=77664) INFO: 127.0.0.1:44912 - "GET /health HTTP/1.1" 200 OK 2025-10-10T01:59:57.5295504Z (APIServer pid=77664) INFO: 127.0.0.1:44926 - "GET /v1/models HTTP/1.1" 200 OK 2025-10-10T01:59:57.7609168Z (APIServer pid=77664) INFO: 127.0.0.1:44926 - "POST /v1/completions HTTP/1.1" 200 OK 2025-10-10T01:59:57.8781040Z (APIServer pid=77664) INFO: 127.0.0.1:44926 - "POST /v1/completions HTTP/1.1" 200 OK 2025-10-10T01:59:58.0255775Z (APIServer pid=77664) INFO: 127.0.0.1:44926 - "POST /v1/completions HTTP/1.1" 200 OK 2025-10-10T01:59:58.1799614Z (APIServer pid=77664) INFO: 127.0.0.1:44926 - "POST /v1/completions HTTP/1.1" 200 OK 2025-10-10T01:59:58.3235245Z (APIServer pid=77664) INFO: 127.0.0.1:44926 - "POST /v1/completions HTTP/1.1" 200 OK 2025-10-10T01:59:58.3310109Z (APIServer pid=77664) INFO: 127.0.0.1:44926 - "POST /v1/completions HTTP/1.1" 200 OK 2025-10-10T01:59:58.4582322Z (APIServer pid=77664) INFO 10-10 01:59:58 [launcher.py:99] Shutting down FastAPI HTTP server. 2025-10-10T01:59:58.4585661Z (Worker_TP0 pid=77906) INFO 10-10 01:59:58 [multiproc_executor.py:558] Parent process exited, terminating worker 2025-10-10T01:59:58.4587159Z (Worker_TP0 pid=77906) INFO 10-10 01:59:58 [multiproc_executor.py:599] WorkerProc shutting down. 2025-10-10T01:59:58.4592204Z (Worker_TP1 pid=77907) INFO 10-10 01:59:58 [multiproc_executor.py:558] Parent process exited, terminating worker 2025-10-10T01:59:58.4593535Z (Worker_TP1 pid=77907) INFO 10-10 01:59:58 [multiproc_executor.py:599] WorkerProc shutting down. 2025-10-10T02:00:00.5804283Z (APIServer pid=77664) INFO: Shutting down 2025-10-10T02:00:00.6809361Z (APIServer pid=77664) INFO: Waiting for application shutdown. 2025-10-10T02:00:00.6811086Z (APIServer pid=77664) INFO: Application shutdown complete. 2025-10-10T02:00:01.3164469Z PASSED 2025-10-10T02:00:01.3304199Z distributed/test_sequence_parallel.py::test_tp_sp_generation[RedHatAI/Meta-Llama-3.1-8B-Instruct-FP8-parallel_setup19-ray-1-auto-test_options19] Fork a new process to run a test 78424 2025-10-10T02:00:01.3321755Z Fork a new process to run a test 0 2025-10-10T02:00:02.2368468Z INFO 10-10 02:00:02 [model.py:551] Resolved architecture: LlamaForCausalLM 2025-10-10T02:00:02.2368932Z `torch_dtype` is deprecated! Use `dtype` instead! 2025-10-10T02:00:02.2370071Z WARNING 10-10 02:00:02 [model.py:1768] Casting torch.bfloat16 to torch.float16. 2025-10-10T02:00:02.2370786Z INFO 10-10 02:00:02 [model.py:1545] Using max model len 2048 2025-10-10T02:00:02.3410154Z Launching RemoteOpenAIServer with: vllm serve RedHatAI/Meta-Llama-3.1-8B-Instruct-FP8 --dtype float16 --max-model-len 2048 --max-num-seqs 8 --enforce-eager --tokenizer-mode auto --tensor-parallel-size 2 --pipeline-parallel-size 1 --distributed-executor-backend ray --compilation_config {"level": 3, "custom_ops": ["+rms_norm"], "compile_sizes": [4, 8], "pass_config": {"enable_sequence_parallelism": true, "enable_fusion": true, "enable_noop": true}} --load-format dummy --port 60053 --seed 0 2025-10-10T02:00:02.8489800Z /opt/conda/envs/py_3.12/lib/python3.12/site-packages/torch/cuda/__init__.py:64: FutureWarning: The pynvml package is deprecated. Please install nvidia-ml-py instead. If you did not install pynvml directly, please report this to the maintainers of the package that installed pynvml for you. 2025-10-10T02:00:02.8491641Z import pynvml # type: ignore[import] 2025-10-10T02:00:05.0588856Z INFO 10-10 02:00:05 [__init__.py:215] Automatically detected platform cuda. 2025-10-10T02:00:08.9744845Z (APIServer pid=78473) INFO 10-10 02:00:08 [api_server.py:1839] vLLM API server version 0.11.0rc2.dev157+g0ad9951c4.d20251010 2025-10-10T02:00:08.9766776Z (APIServer pid=78473) INFO 10-10 02:00:08 [utils.py:233] non-default args: {'model_tag': 'RedHatAI/Meta-Llama-3.1-8B-Instruct-FP8', 'port': 60053, 'model': 'RedHatAI/Meta-Llama-3.1-8B-Instruct-FP8', 'dtype': 'float16', 'seed': 0, 'max_model_len': 2048, 'enforce_eager': True, 'load_format': 'dummy', 'distributed_executor_backend': 'ray', 'tensor_parallel_size': 2, 'max_num_seqs': 8, 'compilation_config': {'level': 3, 'debug_dump_path': None, 'cache_dir': '', 'backend': '', 'custom_ops': ['+rms_norm'], 'splitting_ops': None, 'use_inductor': True, 'compile_sizes': [4, 8], 'inductor_compile_config': {'enable_auto_functionalized_v2': False}, 'inductor_passes': {}, 'cudagraph_mode': None, 'use_cudagraph': True, 'cudagraph_num_of_warmups': 0, 'cudagraph_capture_sizes': None, 'cudagraph_copy_inputs': False, 'full_cuda_graph': False, 'use_inductor_graph_partition': False, 'pass_config': {'enable_fusion': True, 'enable_noop': True, 'enable_sequence_parallelism': True}, 'max_capture_size': None, 'local_cache_dir': None}} 2025-10-10T02:00:09.7054240Z (APIServer pid=78473) INFO 10-10 02:00:09 [model.py:551] Resolved architecture: LlamaForCausalLM 2025-10-10T02:00:09.7055386Z (APIServer pid=78473) `torch_dtype` is deprecated! Use `dtype` instead! 2025-10-10T02:00:09.7056177Z (APIServer pid=78473) WARNING 10-10 02:00:09 [model.py:1768] Casting torch.bfloat16 to torch.float16. 2025-10-10T02:00:09.7059003Z (APIServer pid=78473) INFO 10-10 02:00:09 [model.py:1545] Using max model len 2048 2025-10-10T02:00:10.1703355Z (APIServer pid=78473) INFO 10-10 02:00:10 [scheduler.py:222] Chunked prefill is enabled with max_num_batched_tokens=2048. 2025-10-10T02:00:10.2072136Z (APIServer pid=78473) INFO 10-10 02:00:10 [vllm.py:345] Cudagraph is disabled under eager mode 2025-10-10T02:00:11.8804018Z /opt/conda/envs/py_3.12/lib/python3.12/site-packages/torch/cuda/__init__.py:64: FutureWarning: The pynvml package is deprecated. Please install nvidia-ml-py instead. If you did not install pynvml directly, please report this to the maintainers of the package that installed pynvml for you. 2025-10-10T02:00:11.8805565Z import pynvml # type: ignore[import] 2025-10-10T02:00:14.1862011Z INFO 10-10 02:00:14 [__init__.py:215] Automatically detected platform cuda. 2025-10-10T02:00:18.4462709Z (EngineCore_DP0 pid=78596) INFO 10-10 02:00:18 [core.py:648] Waiting for init message from front-end. 2025-10-10T02:00:18.4727386Z (EngineCore_DP0 pid=78596) INFO 10-10 02:00:18 [core.py:78] Initializing a V1 LLM engine (v0.11.0rc2.dev157+g0ad9951c4.d20251010) with config: model='RedHatAI/Meta-Llama-3.1-8B-Instruct-FP8', speculative_config=None, tokenizer='RedHatAI/Meta-Llama-3.1-8B-Instruct-FP8', skip_tokenizer_init=False, tokenizer_mode=auto, revision=None, tokenizer_revision=None, trust_remote_code=False, dtype=torch.float16, max_seq_len=2048, download_dir=None, load_format=dummy, tensor_parallel_size=2, pipeline_parallel_size=1, data_parallel_size=1, disable_custom_all_reduce=False, quantization=compressed-tensors, enforce_eager=True, kv_cache_dtype=auto, device_config=cuda, structured_outputs_config=StructuredOutputsConfig(backend='auto', disable_fallback=False, disable_any_whitespace=False, disable_additional_properties=False, reasoning_parser=''), observability_config=ObservabilityConfig(show_hidden_metrics_for_version=None, otlp_traces_endpoint=None, collect_detailed_traces=None), seed=0, served_model_name=RedHatAI/Meta-Llama-3.1-8B-Instruct-FP8, enable_prefix_caching=True, chunked_prefill_enabled=True, pooler_config=None, compilation_config={'level': 3, 'debug_dump_path': None, 'cache_dir': '', 'backend': '', 'custom_ops': ['+rms_norm', '+rms_norm'], 'splitting_ops': ['vllm.unified_attention', 'vllm.unified_attention_with_output', 'vllm.mamba_mixer2', 'vllm.mamba_mixer', 'vllm.short_conv', 'vllm.linear_attention', 'vllm.plamo2_mamba_mixer', 'vllm.gdn_attention', 'vllm.sparse_attn_indexer'], 'use_inductor': True, 'compile_sizes': [8, 4], 'inductor_compile_config': {'enable_auto_functionalized_v2': False}, 'inductor_passes': {}, 'cudagraph_mode': , 'use_cudagraph': False, 'cudagraph_num_of_warmups': 0, 'cudagraph_capture_sizes': [], 'cudagraph_copy_inputs': False, 'full_cuda_graph': False, 'use_inductor_graph_partition': False, 'pass_config': {'enable_fusion': True, 'enable_noop': True, 'enable_sequence_parallelism': True}, 'max_capture_size': 0, 'local_cache_dir': None} 2025-10-10T02:00:20.3855041Z (EngineCore_DP0 pid=78596) 2025-10-10 02:00:20,384 INFO worker.py:1918 -- Started a local Ray instance. View the dashboard at http://127.0.0.1:8265  2025-10-10T02:00:21.4682498Z (EngineCore_DP0 pid=78596) INFO 10-10 02:00:21 [ray_utils.py:345] No current placement group found. Creating a new placement group. 2025-10-10T02:00:21.6414804Z (EngineCore_DP0 pid=78596) INFO 10-10 02:00:21 [ray_distributed_executor.py:171] use_ray_spmd_worker: True 2025-10-10T02:00:22.2966954Z (EngineCore_DP0 pid=78596) (pid=79482) /opt/conda/envs/py_3.12/lib/python3.12/site-packages/torch/cuda/__init__.py:64: FutureWarning: The pynvml package is deprecated. Please install nvidia-ml-py instead. If you did not install pynvml directly, please report this to the maintainers of the package that installed pynvml for you. 2025-10-10T02:00:22.2968611Z (EngineCore_DP0 pid=78596) (pid=79482) import pynvml # type: ignore[import] 2025-10-10T02:00:24.5426925Z (EngineCore_DP0 pid=78596) (pid=79482) INFO 10-10 02:00:24 [__init__.py:215] Automatically detected platform cuda. 2025-10-10T02:00:27.5091317Z (EngineCore_DP0 pid=78596) INFO 10-10 02:00:27 [ray_env.py:63] RAY_NON_CARRY_OVER_ENV_VARS from config: set() 2025-10-10T02:00:27.5092982Z (EngineCore_DP0 pid=78596) INFO 10-10 02:00:27 [ray_env.py:65] Copying the following environment variables to workers: ['VLLM_USE_RAY_COMPILED_DAG', 'MAX_JOBS', 'HF_TOKEN', 'VLLM_USE_V1', 'VLLM_WORKER_MULTIPROC_METHOD', 'VLLM_USE_RAY_SPMD_WORKER', 'LD_LIBRARY_PATH', 'HUGGING_FACE_HUB_TOKEN'] 2025-10-10T02:00:27.5094722Z (EngineCore_DP0 pid=78596) INFO 10-10 02:00:27 [ray_env.py:68] If certain env vars should NOT be copied, add them to /var/lib/jenkins/.config/vllm/ray_non_carry_over_env_vars.json file 2025-10-10T02:00:29.5409856Z (EngineCore_DP0 pid=78596) (RayWorkerWrapper pid=79482) [Gloo] Rank 0 is connected to 1 peer ranks. Expected number of connected peer ranks is : 1 2025-10-10T02:00:29.5430927Z (EngineCore_DP0 pid=78596) (pid=79486) INFO 10-10 02:00:24 [__init__.py:215] Automatically detected platform cuda. 2025-10-10T02:00:29.6482831Z (EngineCore_DP0 pid=78596) (RayWorkerWrapper pid=79482) INFO 10-10 02:00:29 [__init__.py:1384] Found nccl from library libnccl.so.2 2025-10-10T02:00:29.6485686Z (EngineCore_DP0 pid=78596) (RayWorkerWrapper pid=79482) INFO 10-10 02:00:29 [pynccl.py:104] vLLM is using nccl==2.27.5 2025-10-10T02:00:29.8554880Z (EngineCore_DP0 pid=78596) (RayWorkerWrapper pid=79482) WARNING 10-10 02:00:29 [symm_mem.py:58] SymmMemCommunicator: Device capability 8.9 not supported, communicator is not available. 2025-10-10T02:00:29.8556811Z (EngineCore_DP0 pid=78596) (RayWorkerWrapper pid=79482) INFO 10-10 02:00:29 [custom_all_reduce.py:35] Skipping P2P check and trusting the driver's P2P report. 2025-10-10T02:00:29.8558353Z (EngineCore_DP0 pid=78596) (RayWorkerWrapper pid=79482) WARNING 10-10 02:00:29 [custom_all_reduce.py:154] Custom allreduce is disabled because your platform lacks GPU P2P capability or P2P test failed. To silence this warning, specify disable_custom_all_reduce=True explicitly. 2025-10-10T02:00:29.8560327Z (EngineCore_DP0 pid=78596) (RayWorkerWrapper pid=79482) INFO 10-10 02:00:29 [shm_broadcast.py:289] vLLM message queue communication handle: Handle(local_reader_ranks=[1], buffer_handle=(1, 4194304, 6, 'psm_6f2584dc'), local_subscribe_addr='ipc:///tmp/4aad0d30-1a59-4d27-b5e6-9a20e8cdb95d', remote_subscribe_addr=None, remote_addr_ipv6=False) 2025-10-10T02:00:29.9603420Z (EngineCore_DP0 pid=78596) (RayWorkerWrapper pid=79482) INFO 10-10 02:00:29 [parallel_state.py:1208] rank 0 in world size 2 is assigned as DP rank 0, PP rank 0, TP rank 0, EP rank 0 2025-10-10T02:00:30.2689014Z (EngineCore_DP0 pid=78596) (RayWorkerWrapper pid=79482) INFO 10-10 02:00:30 [topk_topp_sampler.py:55] Using FlashInfer for top-p & top-k sampling. 2025-10-10T02:00:30.2691545Z (EngineCore_DP0 pid=78596) (RayWorkerWrapper pid=79482) INFO 10-10 02:00:30 [gpu_model_runner.py:2707] Starting to load model RedHatAI/Meta-Llama-3.1-8B-Instruct-FP8... 2025-10-10T02:00:30.4748295Z (EngineCore_DP0 pid=78596) (RayWorkerWrapper pid=79482) INFO 10-10 02:00:30 [gpu_model_runner.py:2739] Loading model from scratch... 2025-10-10T02:00:30.5791581Z (EngineCore_DP0 pid=78596) (RayWorkerWrapper pid=79482) INFO 10-10 02:00:30 [cuda.py:361] Using Flash Attention backend on V1 engine. 2025-10-10T02:00:31.6003906Z (EngineCore_DP0 pid=78596) (RayWorkerWrapper pid=79482) INFO 10-10 02:00:31 [gpu_model_runner.py:2758] Model loading took 4.2624 GiB and 0.539961 seconds 2025-10-10T02:00:38.9339403Z (EngineCore_DP0 pid=78596) (RayWorkerWrapper pid=79482) INFO 10-10 02:00:38 [backends.py:548] Using cache directory: /var/lib/jenkins/.cache/vllm/torch_compile_cache/6f893e84df/rank_0_0/backbone for vLLM's torch.compile 2025-10-10T02:00:38.9340559Z (EngineCore_DP0 pid=78596) (RayWorkerWrapper pid=79482) INFO 10-10 02:00:38 [backends.py:559] Dynamo bytecode transform time: 7.09 s 2025-10-10T02:00:38.9342283Z (EngineCore_DP0 pid=78596) (RayWorkerWrapper pid=79486) [Gloo] Rank 1 is connected to 1 peer ranks. Expected number of connected peer ranks is : 1 [repeated 11x across cluster] (Ray deduplicates logs by default. Set RAY_DEDUP_LOGS=0 to disable log deduplication, or see https://docs.ray.io/en/master/ray-observability/user-guides/configure-logging.html#log-deduplication for more options.) 2025-10-10T02:00:38.9344148Z (EngineCore_DP0 pid=78596) (RayWorkerWrapper pid=79486) INFO 10-10 02:00:29 [__init__.py:1384] Found nccl from library libnccl.so.2 [repeated 3x across cluster] 2025-10-10T02:00:38.9345084Z (EngineCore_DP0 pid=78596) (RayWorkerWrapper pid=79486) INFO 10-10 02:00:29 [pynccl.py:104] vLLM is using nccl==2.27.5 [repeated 3x across cluster] 2025-10-10T02:00:38.9346087Z (EngineCore_DP0 pid=78596) (RayWorkerWrapper pid=79486) WARNING 10-10 02:00:29 [symm_mem.py:58] SymmMemCommunicator: Device capability 8.9 not supported, communicator is not available. 2025-10-10T02:00:38.9348412Z (EngineCore_DP0 pid=78596) (RayWorkerWrapper pid=79486) INFO 10-10 02:00:29 [custom_all_reduce.py:35] Skipping P2P check and trusting the driver's P2P report. 2025-10-10T02:00:38.9349790Z (EngineCore_DP0 pid=78596) (RayWorkerWrapper pid=79486) WARNING 10-10 02:00:29 [custom_all_reduce.py:154] Custom allreduce is disabled because your platform lacks GPU P2P capability or P2P test failed. To silence this warning, specify disable_custom_all_reduce=True explicitly. 2025-10-10T02:00:38.9351203Z (EngineCore_DP0 pid=78596) (RayWorkerWrapper pid=79486) INFO 10-10 02:00:29 [parallel_state.py:1208] rank 1 in world size 2 is assigned as DP rank 0, PP rank 0, TP rank 1, EP rank 1 2025-10-10T02:00:38.9352143Z (EngineCore_DP0 pid=78596) (RayWorkerWrapper pid=79486) INFO 10-10 02:00:30 [topk_topp_sampler.py:55] Using FlashInfer for top-p & top-k sampling. 2025-10-10T02:00:38.9353080Z (EngineCore_DP0 pid=78596) (RayWorkerWrapper pid=79486) INFO 10-10 02:00:30 [gpu_model_runner.py:2707] Starting to load model RedHatAI/Meta-Llama-3.1-8B-Instruct-FP8... 2025-10-10T02:00:38.9353955Z (EngineCore_DP0 pid=78596) (RayWorkerWrapper pid=79486) INFO 10-10 02:00:30 [gpu_model_runner.py:2739] Loading model from scratch... 2025-10-10T02:00:38.9354726Z (EngineCore_DP0 pid=78596) (RayWorkerWrapper pid=79486) INFO 10-10 02:00:30 [cuda.py:361] Using Flash Attention backend on V1 engine. 2025-10-10T02:00:38.9355563Z (EngineCore_DP0 pid=78596) (RayWorkerWrapper pid=79486) INFO 10-10 02:00:31 [gpu_model_runner.py:2758] Model loading took 4.2624 GiB and 0.544833 seconds 2025-10-10T02:00:41.8914537Z (EngineCore_DP0 pid=78596) (RayWorkerWrapper pid=79482) INFO 10-10 02:00:41 [backends.py:164] Directly load the compiled graph(s) for dynamic shape from the cache, took 2.401 s 2025-10-10T02:00:44.5461884Z (EngineCore_DP0 pid=78596) (RayWorkerWrapper pid=79482) INFO 10-10 02:00:44 [gpu_worker.py:298] Available KV cache memory: 15.34 GiB 2025-10-10T02:00:44.5463371Z (EngineCore_DP0 pid=78596) (RayWorkerWrapper pid=79486) INFO 10-10 02:00:38 [backends.py:548] Using cache directory: /var/lib/jenkins/.cache/vllm/torch_compile_cache/6f893e84df/rank_1_0/backbone for vLLM's torch.compile 2025-10-10T02:00:44.5464743Z (EngineCore_DP0 pid=78596) (RayWorkerWrapper pid=79486) INFO 10-10 02:00:38 [backends.py:559] Dynamo bytecode transform time: 7.21 s 2025-10-10T02:00:44.8754619Z (EngineCore_DP0 pid=78596) INFO 10-10 02:00:44 [kv_cache_utils.py:1087] GPU KV cache size: 251,360 tokens 2025-10-10T02:00:44.8755570Z (EngineCore_DP0 pid=78596) INFO 10-10 02:00:44 [kv_cache_utils.py:1091] Maximum concurrency for 2,048 tokens per request: 122.73x 2025-10-10T02:00:44.8756454Z (EngineCore_DP0 pid=78596) INFO 10-10 02:00:44 [kv_cache_utils.py:1087] GPU KV cache size: 251,360 tokens 2025-10-10T02:00:44.8757312Z (EngineCore_DP0 pid=78596) INFO 10-10 02:00:44 [kv_cache_utils.py:1091] Maximum concurrency for 2,048 tokens per request: 122.73x 2025-10-10T02:00:44.9561164Z (EngineCore_DP0 pid=78596) (RayWorkerWrapper pid=79482) INFO 10-10 02:00:44 [gpu_worker.py:332] Compile and warming up model for size 8 2025-10-10T02:00:56.2859277Z (EngineCore_DP0 pid=78596) (RayWorkerWrapper pid=79482) INFO 10-10 02:00:56 [backends.py:168] Directly load the compiled graph(s) for shape 8 from the cache, took 11.341 s 2025-10-10T02:00:56.2860698Z (EngineCore_DP0 pid=78596) (RayWorkerWrapper pid=79486) INFO 10-10 02:00:41 [backends.py:164] Directly load the compiled graph(s) for dynamic shape from the cache, took 2.369 s 2025-10-10T02:00:56.2861845Z (EngineCore_DP0 pid=78596) (RayWorkerWrapper pid=79486) INFO 10-10 02:00:44 [gpu_worker.py:298] Available KV cache memory: 15.34 GiB 2025-10-10T02:00:56.2863299Z (EngineCore_DP0 pid=78596) (RayWorkerWrapper pid=79482) INFO 10-10 02:00:56 [gpu_worker.py:332] Compile and warming up model for size 4 [repeated 2x across cluster] 2025-10-10T02:01:22.4821227Z (EngineCore_DP0 pid=78596) (RayWorkerWrapper pid=79486) INFO 10-10 02:01:22 [monitor.py:32] torch.compile takes 7.21 s in total 2025-10-10T02:01:22.4822580Z (EngineCore_DP0 pid=78596) (RayWorkerWrapper pid=79486) INFO 10-10 02:01:22 [backends.py:168] Directly load the compiled graph(s) for shape 4 from the cache, took 26.181 s [repeated 2x across cluster] 2025-10-10T02:01:22.4823896Z (EngineCore_DP0 pid=78596) (RayWorkerWrapper pid=79486) INFO 10-10 02:00:56 [gpu_worker.py:332] Compile and warming up model for size 4 2025-10-10T02:01:22.6599306Z (EngineCore_DP0 pid=78596) INFO 10-10 02:01:22 [core.py:211] init engine (profile, create kv cache, warmup model) took 51.13 seconds 2025-10-10T02:01:23.1510424Z (EngineCore_DP0 pid=78596) INFO 10-10 02:01:23 [vllm.py:345] Cudagraph is disabled under eager mode 2025-10-10T02:01:23.3369069Z (APIServer pid=78473) INFO 10-10 02:01:23 [loggers.py:147] Engine 000: vllm cache_config_info with initialization after num_gpu_blocks is: 15710 2025-10-10T02:01:23.4274124Z (EngineCore_DP0 pid=78596) INFO 10-10 02:01:23 [gc_utils.py:41] GC Debug Config. enabled:False,top_objects:-1 2025-10-10T02:01:23.4519563Z (APIServer pid=78473) INFO 10-10 02:01:23 [api_server.py:1634] Supported_tasks: ['generate'] 2025-10-10T02:01:23.4912164Z (APIServer pid=78473) WARNING 10-10 02:01:23 [model.py:1424] Default sampling parameters have been overridden by the model's Hugging Face generation config recommended from the model creator. If this is not intended, please relaunch vLLM instance with `--generation-config vllm`. 2025-10-10T02:01:23.4915271Z (APIServer pid=78473) INFO 10-10 02:01:23 [serving_responses.py:140] Using default chat sampling params from model: {'temperature': 0.6, 'top_p': 0.9} 2025-10-10T02:01:23.5219516Z (APIServer pid=78473) INFO 10-10 02:01:23 [serving_chat.py:139] Using default chat sampling params from model: {'temperature': 0.6, 'top_p': 0.9} 2025-10-10T02:01:23.5608033Z (APIServer pid=78473) INFO 10-10 02:01:23 [serving_completion.py:76] Using default completion sampling params from model: {'temperature': 0.6, 'top_p': 0.9} 2025-10-10T02:01:23.5610070Z (APIServer pid=78473) INFO 10-10 02:01:23 [api_server.py:1912] Starting vLLM API server 0 on http://0.0.0.0:60053 2025-10-10T02:01:23.5611481Z (APIServer pid=78473) INFO 10-10 02:01:23 [launcher.py:34] Available routes are: 2025-10-10T02:01:23.5612822Z (APIServer pid=78473) INFO 10-10 02:01:23 [launcher.py:42] Route: /openapi.json, Methods: HEAD, GET 2025-10-10T02:01:23.5614450Z (APIServer pid=78473) INFO 10-10 02:01:23 [launcher.py:42] Route: /docs, Methods: HEAD, GET 2025-10-10T02:01:23.5615588Z (APIServer pid=78473) INFO 10-10 02:01:23 [launcher.py:42] Route: /docs/oauth2-redirect, Methods: HEAD, GET 2025-10-10T02:01:23.5616402Z (APIServer pid=78473) INFO 10-10 02:01:23 [launcher.py:42] Route: /redoc, Methods: HEAD, GET 2025-10-10T02:01:23.5616922Z (APIServer pid=78473) INFO 10-10 02:01:23 [launcher.py:42] Route: /health, Methods: GET 2025-10-10T02:01:23.5617414Z (APIServer pid=78473) INFO 10-10 02:01:23 [launcher.py:42] Route: /load, Methods: GET 2025-10-10T02:01:23.5617929Z (APIServer pid=78473) INFO 10-10 02:01:23 [launcher.py:42] Route: /ping, Methods: POST 2025-10-10T02:01:23.5618429Z (APIServer pid=78473) INFO 10-10 02:01:23 [launcher.py:42] Route: /ping, Methods: GET 2025-10-10T02:01:23.5618951Z (APIServer pid=78473) INFO 10-10 02:01:23 [launcher.py:42] Route: /tokenize, Methods: POST 2025-10-10T02:01:23.5619701Z (APIServer pid=78473) INFO 10-10 02:01:23 [launcher.py:42] Route: /detokenize, Methods: POST 2025-10-10T02:01:23.5620251Z (APIServer pid=78473) INFO 10-10 02:01:23 [launcher.py:42] Route: /v1/models, Methods: GET 2025-10-10T02:01:23.5620778Z (APIServer pid=78473) INFO 10-10 02:01:23 [launcher.py:42] Route: /version, Methods: GET 2025-10-10T02:01:23.5621313Z (APIServer pid=78473) INFO 10-10 02:01:23 [launcher.py:42] Route: /v1/responses, Methods: POST 2025-10-10T02:01:23.5621991Z (APIServer pid=78473) INFO 10-10 02:01:23 [launcher.py:42] Route: /v1/responses/{response_id}, Methods: GET 2025-10-10T02:01:23.5622629Z (APIServer pid=78473) INFO 10-10 02:01:23 [launcher.py:42] Route: /v1/responses/{response_id}/cancel, Methods: POST 2025-10-10T02:01:23.5623246Z (APIServer pid=78473) INFO 10-10 02:01:23 [launcher.py:42] Route: /v1/chat/completions, Methods: POST 2025-10-10T02:01:23.5623818Z (APIServer pid=78473) INFO 10-10 02:01:23 [launcher.py:42] Route: /v1/completions, Methods: POST 2025-10-10T02:01:23.5624716Z (APIServer pid=78473) INFO 10-10 02:01:23 [launcher.py:42] Route: /v1/embeddings, Methods: POST 2025-10-10T02:01:23.5625279Z (APIServer pid=78473) INFO 10-10 02:01:23 [launcher.py:42] Route: /pooling, Methods: POST 2025-10-10T02:01:23.5625819Z (APIServer pid=78473) INFO 10-10 02:01:23 [launcher.py:42] Route: /classify, Methods: POST 2025-10-10T02:01:23.5626330Z (APIServer pid=78473) INFO 10-10 02:01:23 [launcher.py:42] Route: /score, Methods: POST 2025-10-10T02:01:23.5626848Z (APIServer pid=78473) INFO 10-10 02:01:23 [launcher.py:42] Route: /v1/score, Methods: POST 2025-10-10T02:01:23.5627432Z (APIServer pid=78473) INFO 10-10 02:01:23 [launcher.py:42] Route: /v1/audio/transcriptions, Methods: POST 2025-10-10T02:01:23.5628033Z (APIServer pid=78473) INFO 10-10 02:01:23 [launcher.py:42] Route: /v1/audio/translations, Methods: POST 2025-10-10T02:01:23.5628679Z (APIServer pid=78473) INFO 10-10 02:01:23 [launcher.py:42] Route: /rerank, Methods: POST 2025-10-10T02:01:23.5629211Z (APIServer pid=78473) INFO 10-10 02:01:23 [launcher.py:42] Route: /v1/rerank, Methods: POST 2025-10-10T02:01:23.5629728Z (APIServer pid=78473) INFO 10-10 02:01:23 [launcher.py:42] Route: /v2/rerank, Methods: POST 2025-10-10T02:01:23.5630274Z (APIServer pid=78473) INFO 10-10 02:01:23 [launcher.py:42] Route: /scale_elastic_ep, Methods: POST 2025-10-10T02:01:23.5630850Z (APIServer pid=78473) INFO 10-10 02:01:23 [launcher.py:42] Route: /is_scaling_elastic_ep, Methods: POST 2025-10-10T02:01:23.5631404Z (APIServer pid=78473) INFO 10-10 02:01:23 [launcher.py:42] Route: /invocations, Methods: POST 2025-10-10T02:01:23.5631938Z (APIServer pid=78473) INFO 10-10 02:01:23 [launcher.py:42] Route: /metrics, Methods: GET 2025-10-10T02:01:23.5769634Z (APIServer pid=78473) INFO: Started server process [78473] 2025-10-10T02:01:23.5770873Z (APIServer pid=78473) INFO: Waiting for application startup. 2025-10-10T02:01:23.8198267Z (APIServer pid=78473) INFO: Application startup complete. 2025-10-10T02:01:24.0470059Z (APIServer pid=78473) INFO: 127.0.0.1:60362 - "GET /health HTTP/1.1" 200 OK 2025-10-10T02:01:24.2766204Z (APIServer pid=78473) INFO: 127.0.0.1:60366 - "GET /v1/models HTTP/1.1" 200 OK 2025-10-10T02:01:24.3078829Z (EngineCore_DP0 pid=78596) INFO 10-10 02:01:24 [ray_distributed_executor.py:552] RAY_CGRAPH_get_timeout is set to 300 2025-10-10T02:01:24.3080666Z (EngineCore_DP0 pid=78596) INFO 10-10 02:01:24 [ray_distributed_executor.py:554] VLLM_USE_RAY_COMPILED_DAG_CHANNEL_TYPE = auto 2025-10-10T02:01:24.3082456Z (EngineCore_DP0 pid=78596) INFO 10-10 02:01:24 [ray_distributed_executor.py:556] VLLM_USE_RAY_COMPILED_DAG_OVERLAP_COMM = False 2025-10-10T02:01:24.3085915Z (EngineCore_DP0 pid=78596) INFO 10-10 02:01:24 [ray_distributed_executor.py:621] Using RayPPCommunicator (which wraps vLLM _PP GroupCoordinator) for Ray Compiled Graph communication. 2025-10-10T02:01:24.5381041Z (APIServer pid=78473) INFO: 127.0.0.1:60366 - "POST /v1/completions HTTP/1.1" 200 OK 2025-10-10T02:01:24.6588398Z (APIServer pid=78473) INFO: 127.0.0.1:60366 - "POST /v1/completions HTTP/1.1" 200 OK 2025-10-10T02:01:24.8547809Z (APIServer pid=78473) INFO: 127.0.0.1:60366 - "POST /v1/completions HTTP/1.1" 200 OK 2025-10-10T02:01:25.0114525Z (APIServer pid=78473) INFO: 127.0.0.1:60366 - "POST /v1/completions HTTP/1.1" 200 OK 2025-10-10T02:01:25.1338488Z (APIServer pid=78473) INFO: 127.0.0.1:60366 - "POST /v1/completions HTTP/1.1" 200 OK 2025-10-10T02:01:25.1422674Z (APIServer pid=78473) INFO: 127.0.0.1:60366 - "POST /v1/completions HTTP/1.1" 200 OK 2025-10-10T02:01:25.3137064Z (APIServer pid=78473) INFO 10-10 02:01:25 [launcher.py:99] Shutting down FastAPI HTTP server. 2025-10-10T02:01:25.3137609Z *** SIGTERM received at time=1760061685 on cpu 35 *** 2025-10-10T02:01:25.3187512Z PC: @ 0x7f3d2d595117 (unknown) (unknown) 2025-10-10T02:01:25.3188547Z @ 0x7f3d2d546520 (unknown) (unknown) 2025-10-10T02:01:25.3189026Z [2025-10-10 02:01:25,318 E 78596 78596] logging.cc:501: *** SIGTERM received at time=1760061685 on cpu 35 *** 2025-10-10T02:01:25.3191030Z [2025-10-10 02:01:25,318 E 78596 78596] logging.cc:501: PC: @ 0x7f3d2d595117 (unknown) (unknown) 2025-10-10T02:01:25.3192684Z [2025-10-10 02:01:25,319 E 78596 78596] logging.cc:501: @ 0x7f3d2d546520 (unknown) (unknown) 2025-10-10T02:01:25.3194672Z (EngineCore_DP0 pid=78596) INFO 10-10 02:01:25 [ray_distributed_executor.py:122] Shutting down Ray distributed executor. If you see error log from logging.cc regarding SIGTERM received, please ignore because this is the expected termination process in Ray. 2025-10-10T02:01:25.3195757Z (EngineCore_DP0 pid=78596) 2025-10-10 02:01:25,319 INFO compiled_dag_node.py:2171 -- Tearing down compiled DAG 2025-10-10T02:01:25.3196871Z (EngineCore_DP0 pid=78596) 2025-10-10 02:01:25,319 INFO compiled_dag_node.py:2176 -- Cancelling compiled worker on actor: Actor(RayWorkerWrapper, 469fcc2145828043bdce29a001000000) 2025-10-10T02:01:25.3197892Z (EngineCore_DP0 pid=78596) 2025-10-10 02:01:25,319 INFO compiled_dag_node.py:2176 -- Cancelling compiled worker on actor: Actor(RayWorkerWrapper, 4e8351978c62aa76c14d0d6c01000000) 2025-10-10T02:01:25.3235493Z (EngineCore_DP0 pid=78596) 2025-10-10 02:01:25,323 INFO compiled_dag_node.py:2198 -- Waiting for worker tasks to exit 2025-10-10T02:01:25.3236570Z (EngineCore_DP0 pid=78596) 2025-10-10 02:01:25,323 INFO compiled_dag_node.py:2201 -- Teardown complete 2025-10-10T02:01:25.8272487Z (EngineCore_DP0 pid=78596) (RayWorkerWrapper pid=79482) INFO 10-10 02:01:22 [monitor.py:32] torch.compile takes 7.09 s in total 2025-10-10T02:01:25.8275186Z (EngineCore_DP0 pid=78596) (RayWorkerWrapper pid=79482) INFO 10-10 02:01:22 [backends.py:168] Directly load the compiled graph(s) for shape 4 from the cache, took 26.393 s 2025-10-10T02:01:25.8277379Z (EngineCore_DP0 pid=78596) (pid=79486) /opt/conda/envs/py_3.12/lib/python3.12/site-packages/torch/cuda/__init__.py:64: FutureWarning: The pynvml package is deprecated. Please install nvidia-ml-py instead. If you did not install pynvml directly, please report this to the maintainers of the package that installed pynvml for you. 2025-10-10T02:01:25.8278817Z (EngineCore_DP0 pid=78596) (pid=79486) import pynvml # type: ignore[import] 2025-10-10T02:01:27.8137778Z (APIServer pid=78473) INFO: Shutting down 2025-10-10T02:01:27.9141258Z (APIServer pid=78473) INFO: Waiting for application shutdown. 2025-10-10T02:01:27.9142973Z (APIServer pid=78473) INFO: Application shutdown complete. 2025-10-10T02:01:28.3040267Z INFO 10-10 02:01:28 [model.py:551] Resolved architecture: LlamaForCausalLM 2025-10-10T02:01:28.3040815Z WARNING 10-10 02:01:28 [model.py:1768] Casting torch.bfloat16 to torch.float16. 2025-10-10T02:01:28.3041561Z INFO 10-10 02:01:28 [model.py:1545] Using max model len 2048 2025-10-10T02:01:28.3047405Z Launching RemoteOpenAIServer with: vllm serve RedHatAI/Meta-Llama-3.1-8B-Instruct-FP8 --dtype float16 --max-model-len 2048 --max-num-seqs 8 --enforce-eager --tokenizer-mode auto --tensor-parallel-size 2 --distributed-executor-backend mp --load-format dummy --port 45607 --seed 0 2025-10-10T02:01:28.8188623Z /opt/conda/envs/py_3.12/lib/python3.12/site-packages/torch/cuda/__init__.py:64: FutureWarning: The pynvml package is deprecated. Please install nvidia-ml-py instead. If you did not install pynvml directly, please report this to the maintainers of the package that installed pynvml for you. 2025-10-10T02:01:28.8189892Z import pynvml # type: ignore[import] 2025-10-10T02:01:31.0404885Z INFO 10-10 02:01:31 [__init__.py:215] Automatically detected platform cuda. 2025-10-10T02:01:34.9698285Z (APIServer pid=83717) INFO 10-10 02:01:34 [api_server.py:1839] vLLM API server version 0.11.0rc2.dev157+g0ad9951c4.d20251010 2025-10-10T02:01:34.9720192Z (APIServer pid=83717) INFO 10-10 02:01:34 [utils.py:233] non-default args: {'model_tag': 'RedHatAI/Meta-Llama-3.1-8B-Instruct-FP8', 'port': 45607, 'model': 'RedHatAI/Meta-Llama-3.1-8B-Instruct-FP8', 'dtype': 'float16', 'seed': 0, 'max_model_len': 2048, 'enforce_eager': True, 'load_format': 'dummy', 'distributed_executor_backend': 'mp', 'tensor_parallel_size': 2, 'max_num_seqs': 8} 2025-10-10T02:01:35.1752860Z (APIServer pid=83717) INFO 10-10 02:01:35 [model.py:551] Resolved architecture: LlamaForCausalLM 2025-10-10T02:01:35.1754227Z (APIServer pid=83717) `torch_dtype` is deprecated! Use `dtype` instead! 2025-10-10T02:01:35.1756148Z (APIServer pid=83717) WARNING 10-10 02:01:35 [model.py:1768] Casting torch.bfloat16 to torch.float16. 2025-10-10T02:01:35.1757634Z (APIServer pid=83717) INFO 10-10 02:01:35 [model.py:1545] Using max model len 2048 2025-10-10T02:01:35.5239455Z (APIServer pid=83717) INFO 10-10 02:01:35 [scheduler.py:222] Chunked prefill is enabled with max_num_batched_tokens=2048. 2025-10-10T02:01:35.5602671Z (APIServer pid=83717) INFO 10-10 02:01:35 [vllm.py:345] Cudagraph is disabled under eager mode 2025-10-10T02:01:36.5390782Z /opt/conda/envs/py_3.12/lib/python3.12/site-packages/torch/cuda/__init__.py:64: FutureWarning: The pynvml package is deprecated. Please install nvidia-ml-py instead. If you did not install pynvml directly, please report this to the maintainers of the package that installed pynvml for you. 2025-10-10T02:01:36.5392301Z import pynvml # type: ignore[import] 2025-10-10T02:01:38.8429956Z INFO 10-10 02:01:38 [__init__.py:215] Automatically detected platform cuda. 2025-10-10T02:01:42.9522124Z (EngineCore_DP0 pid=83840) INFO 10-10 02:01:42 [core.py:648] Waiting for init message from front-end. 2025-10-10T02:01:42.9794612Z (EngineCore_DP0 pid=83840) INFO 10-10 02:01:42 [core.py:78] Initializing a V1 LLM engine (v0.11.0rc2.dev157+g0ad9951c4.d20251010) with config: model='RedHatAI/Meta-Llama-3.1-8B-Instruct-FP8', speculative_config=None, tokenizer='RedHatAI/Meta-Llama-3.1-8B-Instruct-FP8', skip_tokenizer_init=False, tokenizer_mode=auto, revision=None, tokenizer_revision=None, trust_remote_code=False, dtype=torch.float16, max_seq_len=2048, download_dir=None, load_format=dummy, tensor_parallel_size=2, pipeline_parallel_size=1, data_parallel_size=1, disable_custom_all_reduce=False, quantization=compressed-tensors, enforce_eager=True, kv_cache_dtype=auto, device_config=cuda, structured_outputs_config=StructuredOutputsConfig(backend='auto', disable_fallback=False, disable_any_whitespace=False, disable_additional_properties=False, reasoning_parser=''), observability_config=ObservabilityConfig(show_hidden_metrics_for_version=None, otlp_traces_endpoint=None, collect_detailed_traces=None), seed=0, served_model_name=RedHatAI/Meta-Llama-3.1-8B-Instruct-FP8, enable_prefix_caching=True, chunked_prefill_enabled=True, pooler_config=None, compilation_config={'level': 0, 'debug_dump_path': None, 'cache_dir': '', 'backend': '', 'custom_ops': [], 'splitting_ops': None, 'use_inductor': True, 'compile_sizes': [], 'inductor_compile_config': {'enable_auto_functionalized_v2': False}, 'inductor_passes': {}, 'cudagraph_mode': , 'use_cudagraph': False, 'cudagraph_num_of_warmups': 0, 'cudagraph_capture_sizes': [], 'cudagraph_copy_inputs': False, 'full_cuda_graph': False, 'use_inductor_graph_partition': False, 'pass_config': {}, 'max_capture_size': 0, 'local_cache_dir': None} 2025-10-10T02:01:42.9800116Z (EngineCore_DP0 pid=83840) WARNING 10-10 02:01:42 [multiproc_executor.py:720] Reducing Torch parallelism from 24 threads to 1 to avoid unnecessary CPU contention. Set OMP_NUM_THREADS in the external environment to tune this value as needed. 2025-10-10T02:01:42.9804454Z (EngineCore_DP0 pid=83840) INFO 10-10 02:01:42 [shm_broadcast.py:289] vLLM message queue communication handle: Handle(local_reader_ranks=[0, 1], buffer_handle=(2, 16777216, 10, 'psm_046859e0'), local_subscribe_addr='ipc:///tmp/0a76e694-1087-44f9-8354-41a2ab79fd7d', remote_subscribe_addr=None, remote_addr_ipv6=False) 2025-10-10T02:01:43.4657987Z /opt/conda/envs/py_3.12/lib/python3.12/site-packages/torch/cuda/__init__.py:64: FutureWarning: The pynvml package is deprecated. Please install nvidia-ml-py instead. If you did not install pynvml directly, please report this to the maintainers of the package that installed pynvml for you. 2025-10-10T02:01:43.4659234Z import pynvml # type: ignore[import] 2025-10-10T02:01:43.4886773Z /opt/conda/envs/py_3.12/lib/python3.12/site-packages/torch/cuda/__init__.py:64: FutureWarning: The pynvml package is deprecated. Please install nvidia-ml-py instead. If you did not install pynvml directly, please report this to the maintainers of the package that installed pynvml for you. 2025-10-10T02:01:43.4887947Z import pynvml # type: ignore[import] 2025-10-10T02:01:45.8023390Z INFO 10-10 02:01:45 [__init__.py:215] Automatically detected platform cuda. 2025-10-10T02:01:45.8025089Z INFO 10-10 02:01:45 [__init__.py:215] Automatically detected platform cuda. 2025-10-10T02:01:50.4645278Z INFO 10-10 02:01:50 [shm_broadcast.py:289] vLLM message queue communication handle: Handle(local_reader_ranks=[0], buffer_handle=(1, 10485760, 10, 'psm_80549df5'), local_subscribe_addr='ipc:///tmp/a54f7edb-7ab3-4814-a33c-e1005678fae4', remote_subscribe_addr=None, remote_addr_ipv6=False) 2025-10-10T02:01:50.4647237Z INFO 10-10 02:01:50 [shm_broadcast.py:289] vLLM message queue communication handle: Handle(local_reader_ranks=[0], buffer_handle=(1, 10485760, 10, 'psm_f527dcb3'), local_subscribe_addr='ipc:///tmp/3a7de55d-ed59-4b4e-af46-99d7e277c90e', remote_subscribe_addr=None, remote_addr_ipv6=False) 2025-10-10T02:01:50.6718209Z [Gloo] Rank 0 is connected to 1 peer ranks. Expected number of connected peer ranks is : 1 2025-10-10T02:01:50.6719582Z [Gloo] Rank 1 is connected to 1 peer ranks. Expected number of connected peer ranks is : 1 2025-10-10T02:01:50.6757475Z [Gloo] Rank 0 is connected to 1 peer ranks. Expected number of connected peer ranks is : 1 2025-10-10T02:01:50.6758223Z [Gloo] Rank 1 is connected to 1 peer ranks. Expected number of connected peer ranks is : 1 2025-10-10T02:01:50.6800839Z INFO 10-10 02:01:50 [__init__.py:1384] Found nccl from library libnccl.so.2 2025-10-10T02:01:50.6803986Z INFO 10-10 02:01:50 [pynccl.py:104] vLLM is using nccl==2.27.5 2025-10-10T02:01:50.6804704Z INFO 10-10 02:01:50 [__init__.py:1384] Found nccl from library libnccl.so.2 2025-10-10T02:01:50.6806199Z INFO 10-10 02:01:50 [pynccl.py:104] vLLM is using nccl==2.27.5 2025-10-10T02:01:50.8553238Z WARNING 10-10 02:01:50 [symm_mem.py:58] SymmMemCommunicator: Device capability 8.9 not supported, communicator is not available. 2025-10-10T02:01:50.8554142Z WARNING 10-10 02:01:50 [symm_mem.py:58] SymmMemCommunicator: Device capability 8.9 not supported, communicator is not available. 2025-10-10T02:01:50.8577887Z INFO 10-10 02:01:50 [custom_all_reduce.py:35] Skipping P2P check and trusting the driver's P2P report. 2025-10-10T02:01:50.8578555Z INFO 10-10 02:01:50 [custom_all_reduce.py:35] Skipping P2P check and trusting the driver's P2P report. 2025-10-10T02:01:50.8579610Z WARNING 10-10 02:01:50 [custom_all_reduce.py:154] Custom allreduce is disabled because your platform lacks GPU P2P capability or P2P test failed. To silence this warning, specify disable_custom_all_reduce=True explicitly. 2025-10-10T02:01:50.8581022Z WARNING 10-10 02:01:50 [custom_all_reduce.py:154] Custom allreduce is disabled because your platform lacks GPU P2P capability or P2P test failed. To silence this warning, specify disable_custom_all_reduce=True explicitly. 2025-10-10T02:01:50.9441700Z INFO 10-10 02:01:50 [shm_broadcast.py:289] vLLM message queue communication handle: Handle(local_reader_ranks=[1], buffer_handle=(1, 4194304, 6, 'psm_cbc4c0db'), local_subscribe_addr='ipc:///tmp/2f1267ba-f015-49d2-9f9c-e55e7a79de9e', remote_subscribe_addr=None, remote_addr_ipv6=False) 2025-10-10T02:01:50.9466784Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-10-10T02:01:50.9467955Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-10-10T02:01:50.9478184Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-10-10T02:01:50.9486368Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-10-10T02:01:51.0138420Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-10-10T02:01:51.0147134Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-10-10T02:01:51.0159759Z [Gloo] Rank 1 is connected to 1 peer ranks. Expected number of connected peer ranks is : 1 2025-10-10T02:01:51.0160503Z [Gloo] Rank 0 is connected to 1 peer ranks. Expected number of connected peer ranks is : 1 2025-10-10T02:01:51.0162597Z INFO 10-10 02:01:51 [__init__.py:1384] Found nccl from library libnccl.so.2 2025-10-10T02:01:51.0163071Z INFO 10-10 02:01:51 [pynccl.py:104] vLLM is using nccl==2.27.5 2025-10-10T02:01:51.0163618Z INFO 10-10 02:01:51 [__init__.py:1384] Found nccl from library libnccl.so.2 2025-10-10T02:01:51.0164152Z INFO 10-10 02:01:51 [pynccl.py:104] vLLM is using nccl==2.27.5 2025-10-10T02:01:51.0507145Z INFO 10-10 02:01:51 [parallel_state.py:1208] rank 0 in world size 2 is assigned as DP rank 0, PP rank 0, TP rank 0, EP rank 0 2025-10-10T02:01:51.0507934Z INFO 10-10 02:01:51 [parallel_state.py:1208] rank 1 in world size 2 is assigned as DP rank 0, PP rank 0, TP rank 1, EP rank 1 2025-10-10T02:01:51.3101177Z INFO 10-10 02:01:51 [topk_topp_sampler.py:55] Using FlashInfer for top-p & top-k sampling. 2025-10-10T02:01:51.3112118Z INFO 10-10 02:01:51 [topk_topp_sampler.py:55] Using FlashInfer for top-p & top-k sampling. 2025-10-10T02:01:51.3463368Z (Worker_TP0 pid=83959) INFO 10-10 02:01:51 [gpu_model_runner.py:2707] Starting to load model RedHatAI/Meta-Llama-3.1-8B-Instruct-FP8... 2025-10-10T02:01:51.3465253Z (Worker_TP1 pid=83960) INFO 10-10 02:01:51 [gpu_model_runner.py:2707] Starting to load model RedHatAI/Meta-Llama-3.1-8B-Instruct-FP8... 2025-10-10T02:01:51.5762487Z (Worker_TP1 pid=83960) INFO 10-10 02:01:51 [gpu_model_runner.py:2739] Loading model from scratch... 2025-10-10T02:01:51.5823724Z (Worker_TP0 pid=83959) INFO 10-10 02:01:51 [gpu_model_runner.py:2739] Loading model from scratch... 2025-10-10T02:01:51.6678760Z (Worker_TP1 pid=83960) INFO 10-10 02:01:51 [cuda.py:361] Using Flash Attention backend on V1 engine. 2025-10-10T02:01:51.6741101Z (Worker_TP0 pid=83959) INFO 10-10 02:01:51 [cuda.py:361] Using Flash Attention backend on V1 engine. 2025-10-10T02:01:52.5248975Z (Worker_TP1 pid=83960) INFO 10-10 02:01:52 [gpu_model_runner.py:2758] Model loading took 4.2624 GiB and 0.493203 seconds 2025-10-10T02:01:52.5372854Z (Worker_TP0 pid=83959) INFO 10-10 02:01:52 [gpu_model_runner.py:2758] Model loading took 4.2624 GiB and 0.489406 seconds 2025-10-10T02:01:57.9613641Z (Worker_TP1 pid=83960) INFO 10-10 02:01:57 [gpu_worker.py:298] Available KV cache memory: 15.32 GiB 2025-10-10T02:01:57.9818751Z (Worker_TP0 pid=83959) INFO 10-10 02:01:57 [gpu_worker.py:298] Available KV cache memory: 15.32 GiB 2025-10-10T02:01:58.2144107Z (EngineCore_DP0 pid=83840) INFO 10-10 02:01:58 [kv_cache_utils.py:1087] GPU KV cache size: 250,976 tokens 2025-10-10T02:01:58.2145089Z (EngineCore_DP0 pid=83840) INFO 10-10 02:01:58 [kv_cache_utils.py:1091] Maximum concurrency for 2,048 tokens per request: 122.55x 2025-10-10T02:01:58.2145873Z (EngineCore_DP0 pid=83840) INFO 10-10 02:01:58 [kv_cache_utils.py:1087] GPU KV cache size: 250,976 tokens 2025-10-10T02:01:58.2146535Z (EngineCore_DP0 pid=83840) INFO 10-10 02:01:58 [kv_cache_utils.py:1091] Maximum concurrency for 2,048 tokens per request: 122.55x 2025-10-10T02:01:58.3427537Z (EngineCore_DP0 pid=83840) INFO 10-10 02:01:58 [core.py:211] init engine (profile, create kv cache, warmup model) took 5.76 seconds 2025-10-10T02:01:58.8102785Z (EngineCore_DP0 pid=83840) INFO 10-10 02:01:58 [vllm.py:345] Cudagraph is disabled under eager mode 2025-10-10T02:01:58.9950864Z (APIServer pid=83717) INFO 10-10 02:01:58 [loggers.py:147] Engine 000: vllm cache_config_info with initialization after num_gpu_blocks is: 15686 2025-10-10T02:01:59.0312553Z (EngineCore_DP0 pid=83840) INFO 10-10 02:01:59 [gc_utils.py:41] GC Debug Config. enabled:False,top_objects:-1 2025-10-10T02:01:59.0526872Z (APIServer pid=83717) INFO 10-10 02:01:59 [api_server.py:1634] Supported_tasks: ['generate'] 2025-10-10T02:01:59.0856155Z (APIServer pid=83717) WARNING 10-10 02:01:59 [model.py:1424] Default sampling parameters have been overridden by the model's Hugging Face generation config recommended from the model creator. If this is not intended, please relaunch vLLM instance with `--generation-config vllm`. 2025-10-10T02:01:59.0857790Z (APIServer pid=83717) INFO 10-10 02:01:59 [serving_responses.py:140] Using default chat sampling params from model: {'temperature': 0.6, 'top_p': 0.9} 2025-10-10T02:01:59.1182745Z (APIServer pid=83717) INFO 10-10 02:01:59 [serving_chat.py:139] Using default chat sampling params from model: {'temperature': 0.6, 'top_p': 0.9} 2025-10-10T02:01:59.1546319Z (APIServer pid=83717) INFO 10-10 02:01:59 [serving_completion.py:76] Using default completion sampling params from model: {'temperature': 0.6, 'top_p': 0.9} 2025-10-10T02:01:59.1548166Z (APIServer pid=83717) INFO 10-10 02:01:59 [api_server.py:1912] Starting vLLM API server 0 on http://0.0.0.0:45607 2025-10-10T02:01:59.1549215Z (APIServer pid=83717) INFO 10-10 02:01:59 [launcher.py:34] Available routes are: 2025-10-10T02:01:59.1550446Z (APIServer pid=83717) INFO 10-10 02:01:59 [launcher.py:42] Route: /openapi.json, Methods: GET, HEAD 2025-10-10T02:01:59.1551646Z (APIServer pid=83717) INFO 10-10 02:01:59 [launcher.py:42] Route: /docs, Methods: GET, HEAD 2025-10-10T02:01:59.1552872Z (APIServer pid=83717) INFO 10-10 02:01:59 [launcher.py:42] Route: /docs/oauth2-redirect, Methods: GET, HEAD 2025-10-10T02:01:59.1554089Z (APIServer pid=83717) INFO 10-10 02:01:59 [launcher.py:42] Route: /redoc, Methods: GET, HEAD 2025-10-10T02:01:59.1555475Z (APIServer pid=83717) INFO 10-10 02:01:59 [launcher.py:42] Route: /health, Methods: GET 2025-10-10T02:01:59.1556166Z (APIServer pid=83717) INFO 10-10 02:01:59 [launcher.py:42] Route: /load, Methods: GET 2025-10-10T02:01:59.1556679Z (APIServer pid=83717) INFO 10-10 02:01:59 [launcher.py:42] Route: /ping, Methods: POST 2025-10-10T02:01:59.1557188Z (APIServer pid=83717) INFO 10-10 02:01:59 [launcher.py:42] Route: /ping, Methods: GET 2025-10-10T02:01:59.1557800Z (APIServer pid=83717) INFO 10-10 02:01:59 [launcher.py:42] Route: /tokenize, Methods: POST 2025-10-10T02:01:59.1558355Z (APIServer pid=83717) INFO 10-10 02:01:59 [launcher.py:42] Route: /detokenize, Methods: POST 2025-10-10T02:01:59.1559005Z (APIServer pid=83717) INFO 10-10 02:01:59 [launcher.py:42] Route: /v1/models, Methods: GET 2025-10-10T02:01:59.1559512Z (APIServer pid=83717) INFO 10-10 02:01:59 [launcher.py:42] Route: /version, Methods: GET 2025-10-10T02:01:59.1560046Z (APIServer pid=83717) INFO 10-10 02:01:59 [launcher.py:42] Route: /v1/responses, Methods: POST 2025-10-10T02:01:59.1560634Z (APIServer pid=83717) INFO 10-10 02:01:59 [launcher.py:42] Route: /v1/responses/{response_id}, Methods: GET 2025-10-10T02:01:59.1561283Z (APIServer pid=83717) INFO 10-10 02:01:59 [launcher.py:42] Route: /v1/responses/{response_id}/cancel, Methods: POST 2025-10-10T02:01:59.1561894Z (APIServer pid=83717) INFO 10-10 02:01:59 [launcher.py:42] Route: /v1/chat/completions, Methods: POST 2025-10-10T02:01:59.1562459Z (APIServer pid=83717) INFO 10-10 02:01:59 [launcher.py:42] Route: /v1/completions, Methods: POST 2025-10-10T02:01:59.1563003Z (APIServer pid=83717) INFO 10-10 02:01:59 [launcher.py:42] Route: /v1/embeddings, Methods: POST 2025-10-10T02:01:59.1563540Z (APIServer pid=83717) INFO 10-10 02:01:59 [launcher.py:42] Route: /pooling, Methods: POST 2025-10-10T02:01:59.1564069Z (APIServer pid=83717) INFO 10-10 02:01:59 [launcher.py:42] Route: /classify, Methods: POST 2025-10-10T02:01:59.1564585Z (APIServer pid=83717) INFO 10-10 02:01:59 [launcher.py:42] Route: /score, Methods: POST 2025-10-10T02:01:59.1565097Z (APIServer pid=83717) INFO 10-10 02:01:59 [launcher.py:42] Route: /v1/score, Methods: POST 2025-10-10T02:01:59.1565668Z (APIServer pid=83717) INFO 10-10 02:01:59 [launcher.py:42] Route: /v1/audio/transcriptions, Methods: POST 2025-10-10T02:01:59.1566262Z (APIServer pid=83717) INFO 10-10 02:01:59 [launcher.py:42] Route: /v1/audio/translations, Methods: POST 2025-10-10T02:01:59.1566808Z (APIServer pid=83717) INFO 10-10 02:01:59 [launcher.py:42] Route: /rerank, Methods: POST 2025-10-10T02:01:59.1567326Z (APIServer pid=83717) INFO 10-10 02:01:59 [launcher.py:42] Route: /v1/rerank, Methods: POST 2025-10-10T02:01:59.1567841Z (APIServer pid=83717) INFO 10-10 02:01:59 [launcher.py:42] Route: /v2/rerank, Methods: POST 2025-10-10T02:01:59.1568388Z (APIServer pid=83717) INFO 10-10 02:01:59 [launcher.py:42] Route: /scale_elastic_ep, Methods: POST 2025-10-10T02:01:59.1569029Z (APIServer pid=83717) INFO 10-10 02:01:59 [launcher.py:42] Route: /is_scaling_elastic_ep, Methods: POST 2025-10-10T02:01:59.1569582Z (APIServer pid=83717) INFO 10-10 02:01:59 [launcher.py:42] Route: /invocations, Methods: POST 2025-10-10T02:01:59.1570108Z (APIServer pid=83717) INFO 10-10 02:01:59 [launcher.py:42] Route: /metrics, Methods: GET 2025-10-10T02:01:59.1700560Z (APIServer pid=83717) INFO: Started server process [83717] 2025-10-10T02:01:59.1701581Z (APIServer pid=83717) INFO: Waiting for application startup. 2025-10-10T02:01:59.4034133Z (APIServer pid=83717) INFO: Application startup complete. 2025-10-10T02:01:59.8850427Z (APIServer pid=83717) INFO: 127.0.0.1:35706 - "GET /health HTTP/1.1" 200 OK 2025-10-10T02:01:59.8939803Z (APIServer pid=83717) INFO: 127.0.0.1:35708 - "GET /v1/models HTTP/1.1" 200 OK 2025-10-10T02:02:00.1243240Z (APIServer pid=83717) INFO: 127.0.0.1:35708 - "POST /v1/completions HTTP/1.1" 200 OK 2025-10-10T02:02:00.2407642Z (APIServer pid=83717) INFO: 127.0.0.1:35708 - "POST /v1/completions HTTP/1.1" 200 OK 2025-10-10T02:02:00.3951184Z (APIServer pid=83717) INFO: 127.0.0.1:35708 - "POST /v1/completions HTTP/1.1" 200 OK 2025-10-10T02:02:00.5484002Z (APIServer pid=83717) INFO: 127.0.0.1:35708 - "POST /v1/completions HTTP/1.1" 200 OK 2025-10-10T02:02:00.6912619Z (APIServer pid=83717) INFO: 127.0.0.1:35708 - "POST /v1/completions HTTP/1.1" 200 OK 2025-10-10T02:02:00.6986368Z (APIServer pid=83717) INFO: 127.0.0.1:35708 - "POST /v1/completions HTTP/1.1" 200 OK 2025-10-10T02:02:00.8272314Z (APIServer pid=83717) INFO 10-10 02:02:00 [launcher.py:99] Shutting down FastAPI HTTP server. 2025-10-10T02:02:00.8275129Z (Worker_TP0 pid=83959) INFO 10-10 02:02:00 [multiproc_executor.py:558] Parent process exited, terminating worker 2025-10-10T02:02:00.8276479Z (Worker_TP0 pid=83959) INFO 10-10 02:02:00 [multiproc_executor.py:599] WorkerProc shutting down. 2025-10-10T02:02:00.8279616Z (Worker_TP1 pid=83960) INFO 10-10 02:02:00 [multiproc_executor.py:558] Parent process exited, terminating worker 2025-10-10T02:02:00.8280803Z (Worker_TP1 pid=83960) INFO 10-10 02:02:00 [multiproc_executor.py:599] WorkerProc shutting down. 2025-10-10T02:02:02.8518809Z (APIServer pid=83717) INFO: Shutting down 2025-10-10T02:02:02.9514303Z (APIServer pid=83717) INFO: Waiting for application shutdown. 2025-10-10T02:02:02.9516438Z (APIServer pid=83717) INFO: Application shutdown complete. 2025-10-10T02:02:03.5801145Z PASSED 2025-10-10T02:02:03.5801372Z 2025-10-10T02:02:03.5801646Z =============================== warnings summary =============================== 2025-10-10T02:02:03.5802273Z ../../../../../../opt/conda/envs/py_3.12/lib/python3.12/site-packages/torch/cuda/__init__.py:64 2025-10-10T02:02:03.5803753Z /opt/conda/envs/py_3.12/lib/python3.12/site-packages/torch/cuda/__init__.py:64: FutureWarning: The pynvml package is deprecated. Please install nvidia-ml-py instead. If you did not install pynvml directly, please report this to the maintainers of the package that installed pynvml for you. 2025-10-10T02:02:03.5804953Z import pynvml # type: ignore[import] 2025-10-10T02:02:03.5805162Z 2025-10-10T02:02:03.5805471Z ../../../../../../opt/conda/envs/py_3.12/lib/python3.12/site-packages/schemathesis/generation/coverage.py:305 2025-10-10T02:02:03.5806971Z /opt/conda/envs/py_3.12/lib/python3.12/site-packages/schemathesis/generation/coverage.py:305: DeprecationWarning: jsonschema.exceptions.RefResolutionError is deprecated as of version 4.18.0. If you wish to catch potential reference resolution errors, directly catch referencing.exceptions.Unresolvable. 2025-10-10T02:02:03.5808445Z ref_error: type[Exception] = jsonschema.RefResolutionError, 2025-10-10T02:02:03.5808672Z 2025-10-10T02:02:03.5808822Z tests/distributed/test_sequence_parallel.py: 20 warnings 2025-10-10T02:02:03.5809487Z /var/lib/jenkins/workspace/vllm/tests/utils.py:829: DeprecationWarning: This process (pid=1044) is multi-threaded, use of fork() may lead to deadlocks in the child. 2025-10-10T02:02:03.5810180Z pid = os.fork() 2025-10-10T02:02:03.5810286Z 2025-10-10T02:02:03.5810471Z -- Docs: https://docs.pytest.org/en/stable/how-to/capture-warnings.html 2025-10-10T02:02:03.5811525Z ================= 20 passed, 22 warnings in 1940.77s (0:32:20) ================= 2025-10-10T02:02:05.1672229Z 2025-10-10 02:02:05,166 [INFO] cli.lib.core.vllm.lib: Finish running step: pytest -v -s distributed/test_sequence_parallel.py 2025-10-10T02:02:05.1673335Z 2025-10-10 02:02:05,166 [INFO] cli.lib.core.vllm.lib: Done. All tests passed 2025-10-10T02:02:05.1674099Z 2025-10-10 02:02:05,166 [INFO] cli.lib.core.vllm.vllm_test: Double check installed packages 2025-10-10T02:02:05.1678530Z 2025-10-10 02:02:05,167 [INFO] cli.lib.common.pip_helper: torch already exist with version: 2.10.0a0+git344e636 2025-10-10T02:02:05.1681526Z 2025-10-10 02:02:05,167 [INFO] cli.lib.common.pip_helper: xformers already exist with version: 0.0.33+5d4b92a5.d20251010 2025-10-10T02:02:05.1684911Z 2025-10-10 02:02:05,168 [INFO] cli.lib.common.pip_helper: torchvision already exist with version: 0.22.0a0+966da7e 2025-10-10T02:02:05.1688699Z 2025-10-10 02:02:05,168 [INFO] cli.lib.common.pip_helper: torchaudio already exist with version: 2.8.0a0+87ff22e 2025-10-10T02:02:05.1694017Z 2025-10-10 02:02:05,169 [INFO] cli.lib.common.pip_helper: vllm already exist with version: 0.11.0rc2.dev157+g0ad9951c4.d20251010 2025-10-10T02:02:05.1694721Z 2025-10-10 02:02:05,169 [INFO] cli.lib.core.vllm.vllm_test: Done. checked installed packages 2025-10-10T02:02:05.1935910Z + sccache_epilogue 2025-10-10T02:02:05.1936472Z + echo '::group::Sccache Compilation Log' 2025-10-10T02:02:05.1937687Z ##[group]Sccache Compilation Log 2025-10-10T02:02:05.1938449Z + echo '=================== sccache compilation log ===================' 2025-10-10T02:02:05.1939315Z =================== sccache compilation log =================== 2025-10-10T02:02:05.1939962Z + python /var/lib/jenkins/workspace/.ci/pytorch/print_sccache_log.py /var/lib/jenkins/sccache_error.log 2025-10-10T02:02:05.2397535Z + echo '=========== If your build fails, please take a look at the log above for possible reasons ===========' 2025-10-10T02:02:05.2398992Z =========== If your build fails, please take a look at the log above for possible reasons =========== 2025-10-10T02:02:05.2399858Z + sccache --show-stats 2025-10-10T02:02:05.2425166Z Compile requests 90 2025-10-10T02:02:05.2425809Z Compile requests executed 0 2025-10-10T02:02:05.2426330Z Cache hits 0 2025-10-10T02:02:05.2426821Z Cache misses 0 2025-10-10T02:02:05.2427491Z Cache hits rate - 2025-10-10T02:02:05.2428240Z Cache timeouts 0 2025-10-10T02:02:05.2428869Z Cache read errors 0 2025-10-10T02:02:05.2429368Z Forced recaches 0 2025-10-10T02:02:05.2429677Z Cache write errors 0 2025-10-10T02:02:05.2429961Z Cache errors 0 2025-10-10T02:02:05.2430227Z Compilations 0 2025-10-10T02:02:05.2430492Z Compilation failures 0 2025-10-10T02:02:05.2430772Z Non-cacheable compilations 0 2025-10-10T02:02:05.2431045Z Non-cacheable calls 0 2025-10-10T02:02:05.2431325Z Non-compilation calls 90 2025-10-10T02:02:05.2431772Z Unsupported compiler calls 0 2025-10-10T02:02:05.2432053Z Average cache write 0.000 s 2025-10-10T02:02:05.2432357Z Average compiler 0.000 s 2025-10-10T02:02:05.2432647Z Average cache read hit 0.000 s 2025-10-10T02:02:05.2432944Z Failed distributed compilations 0 2025-10-10T02:02:05.2433348Z Cache location s3, name: ossci-compiler-cache-circleci-v2, prefix: / 2025-10-10T02:02:05.2433764Z Version (client) 0.10.0 2025-10-10T02:02:05.2434051Z + sccache --stop-server 2025-10-10T02:02:05.2447935Z Stopping sccache server... 2025-10-10T02:02:05.2452432Z Compile requests 90 2025-10-10T02:02:05.2453012Z Compile requests executed 0 2025-10-10T02:02:05.2453510Z Cache hits 0 2025-10-10T02:02:05.2454028Z Cache misses 0 2025-10-10T02:02:05.2454534Z Cache hits rate - 2025-10-10T02:02:05.2455627Z Cache timeouts 0 2025-10-10T02:02:05.2456368Z Cache read errors 0 2025-10-10T02:02:05.2456870Z Forced recaches 0 2025-10-10T02:02:05.2457386Z Cache write errors 0 2025-10-10T02:02:05.2457965Z Cache errors 0 2025-10-10T02:02:05.2458582Z Compilations 0 2025-10-10T02:02:05.2459349Z Compilation failures 0 2025-10-10T02:02:05.2459669Z Non-cacheable compilations 0 2025-10-10T02:02:05.2459936Z Non-cacheable calls 0 2025-10-10T02:02:05.2460167Z Non-compilation calls 90 2025-10-10T02:02:05.2460385Z Unsupported compiler calls 0 2025-10-10T02:02:05.2460604Z Average cache write 0.000 s 2025-10-10T02:02:05.2460833Z Average compiler 0.000 s 2025-10-10T02:02:05.2461058Z Average cache read hit 0.000 s 2025-10-10T02:02:05.2461303Z Failed distributed compilations 0 2025-10-10T02:02:05.2461623Z Cache location s3, name: ossci-compiler-cache-circleci-v2, prefix: / 2025-10-10T02:02:05.2461944Z Version (client) 0.10.0 2025-10-10T02:02:05.2462175Z + echo ::endgroup:: 2025-10-10T02:02:05.2462493Z ##[endgroup] 2025-10-10T02:02:05.2462655Z + cleanup_workspace 2025-10-10T02:02:05.2463007Z + echo 'sudo may print the following warning message that can be ignored. The chown command will still run.' 2025-10-10T02:02:05.2463569Z sudo may print the following warning message that can be ignored. The chown command will still run. 2025-10-10T02:02:05.2464020Z + echo ' sudo: setrlimit(RLIMIT_STACK): Operation not permitted' 2025-10-10T02:02:05.2464350Z sudo: setrlimit(RLIMIT_STACK): Operation not permitted 2025-10-10T02:02:05.2464738Z + echo 'For more details refer to https://github.com/sudo-project/sudo/issues/42' 2025-10-10T02:02:05.2465150Z For more details refer to https://github.com/sudo-project/sudo/issues/42 2025-10-10T02:02:05.2465496Z + sudo chown -R 1000 /var/lib/jenkins/workspace 2025-10-10T02:02:06.2789728Z ##[group]Run pytorch/test-infra/.github/actions/upload-benchmark-results@main 2025-10-10T02:02:06.2790109Z with: 2025-10-10T02:02:06.2790307Z benchmark-results-dir: test/test-reports 2025-10-10T02:02:06.2790551Z dry-run: false 2025-10-10T02:02:06.2790739Z schema-version: v3 2025-10-10T02:02:06.2791116Z github-token: *** 2025-10-10T02:02:06.2791305Z env: 2025-10-10T02:02:06.2791469Z GIT_DEFAULT_BRANCH: main 2025-10-10T02:02:06.2791714Z GPU_FLAG: --gpus all -e NVIDIA_DRIVER_CAPABILITIES=all 2025-10-10T02:02:06.2792124Z DOCKER_CONTAINER_ID: 92bc90bdf047b31a0d3a4858f15294a82b80fed2ff3cd44e76b438de5ef8fa55 2025-10-10T02:02:06.2792486Z ##[endgroup] 2025-10-10T02:02:06.2818468Z ##[group]Run set -eux 2025-10-10T02:02:06.2818689Z set -eux 2025-10-10T02:02:06.2818847Z  2025-10-10T02:02:06.2819007Z if [[ -n "" ]]; then 2025-10-10T02:02:06.2819205Z  source "" 2025-10-10T02:02:06.2819372Z fi 2025-10-10T02:02:06.2819791Z python3 -mpip install boto3==1.35.33 psutil==7.0.0 pynvml==12.0.0 2025-10-10T02:02:06.2820095Z  2025-10-10T02:02:06.2820248Z DEVICE_NAME="" 2025-10-10T02:02:06.2820445Z DEVICE_TYPE="" 2025-10-10T02:02:06.2820624Z  2025-10-10T02:02:06.2820795Z if command -v nvidia-smi; then 2025-10-10T02:02:06.2821125Z  # NB: I'm using PyTorch here to get the device name, however, it needs to 2025-10-10T02:02:06.2821534Z  # install the correct version of PyTorch manually for now. Any PyTorch 2025-10-10T02:02:06.2821916Z  # version is fine, I just use 2.7.1 to satify PYPIDEP linter 2025-10-10T02:02:06.2822239Z  python3 -mpip install torch==2.7.1 2025-10-10T02:02:06.2822482Z elif command -v rocminfo; then 2025-10-10T02:02:06.2822790Z  # NB: Installing torch on ROCm runner with pip here causes CI to fail 2025-10-10T02:02:06.2823188Z  # with a memoryview is too large error only on MI300 runners. Is pip 2025-10-10T02:02:06.2823592Z  # version on ROCm runner there too old? As a workaround, let's use the 2025-10-10T02:02:06.2824150Z  # GPU device name coming from rocminfo instead 2025-10-10T02:02:06.2824465Z  DEVICE_NAME=rocm 2025-10-10T02:02:06.2824819Z  DEVICE_TYPE=$(rocminfo | grep "Marketing Name" | tail -n1 | awk -F':' '{print $2}' | xargs) 2025-10-10T02:02:06.2825297Z fi 2025-10-10T02:02:06.2825454Z  2025-10-10T02:02:06.2825645Z echo "DEVICE_NAME=$DEVICE_NAME" >> $GITHUB_ENV 2025-10-10T02:02:06.2825938Z echo "DEVICE_TYPE=$DEVICE_TYPE" >> $GITHUB_ENV 2025-10-10T02:02:06.2838296Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2025-10-10T02:02:06.2838675Z env: 2025-10-10T02:02:06.2838834Z GIT_DEFAULT_BRANCH: main 2025-10-10T02:02:06.2839077Z GPU_FLAG: --gpus all -e NVIDIA_DRIVER_CAPABILITIES=all 2025-10-10T02:02:06.2839485Z DOCKER_CONTAINER_ID: 92bc90bdf047b31a0d3a4858f15294a82b80fed2ff3cd44e76b438de5ef8fa55 2025-10-10T02:02:06.2839845Z ##[endgroup] 2025-10-10T02:02:06.2881382Z + [[ -n '' ]] 2025-10-10T02:02:06.2882095Z + python3 -mpip install boto3==1.35.33 psutil==7.0.0 pynvml==12.0.0 2025-10-10T02:02:06.5061832Z Defaulting to user installation because normal site-packages is not writeable 2025-10-10T02:02:07.5690506Z Collecting boto3==1.35.33 2025-10-10T02:02:07.5876255Z Downloading boto3-1.35.33-py3-none-any.whl (139 kB) 2025-10-10T02:02:07.8594189Z Collecting psutil==7.0.0 2025-10-10T02:02:07.8646284Z Downloading psutil-7.0.0-cp36-abi3-manylinux_2_12_x86_64.manylinux2010_x86_64.manylinux_2_17_x86_64.manylinux2014_x86_64.whl (277 kB) 2025-10-10T02:02:07.8946154Z Collecting pynvml==12.0.0 2025-10-10T02:02:07.8989385Z Downloading pynvml-12.0.0-py3-none-any.whl (26 kB) 2025-10-10T02:02:07.9072223Z Requirement already satisfied: jmespath<2.0.0,>=0.7.1 in /usr/lib/python3.9/site-packages (from boto3==1.35.33) (0.10.0) 2025-10-10T02:02:08.9426118Z Collecting botocore<1.36.0,>=1.35.33 2025-10-10T02:02:08.9480029Z Downloading botocore-1.35.99-py3-none-any.whl (13.3 MB) 2025-10-10T02:02:09.1157461Z Collecting s3transfer<0.11.0,>=0.10.0 2025-10-10T02:02:09.1203712Z Downloading s3transfer-0.10.4-py3-none-any.whl (83 kB) 2025-10-10T02:02:09.1654558Z Collecting nvidia-ml-py<13.0.0a0,>=12.0.0 2025-10-10T02:02:09.1700218Z Downloading nvidia_ml_py-12.575.51-py3-none-any.whl (47 kB) 2025-10-10T02:02:09.1782079Z Requirement already satisfied: urllib3<1.27,>=1.25.4 in /usr/lib/python3.9/site-packages (from botocore<1.36.0,>=1.35.33->boto3==1.35.33) (1.25.10) 2025-10-10T02:02:09.1787221Z Requirement already satisfied: python-dateutil<3.0.0,>=2.1 in /usr/lib/python3.9/site-packages (from botocore<1.36.0,>=1.35.33->boto3==1.35.33) (2.8.1) 2025-10-10T02:02:09.2966751Z Requirement already satisfied: six>=1.5 in /usr/lib/python3.9/site-packages (from python-dateutil<3.0.0,>=2.1->botocore<1.36.0,>=1.35.33->boto3==1.35.33) (1.15.0) 2025-10-10T02:02:09.4119495Z Installing collected packages: botocore, s3transfer, nvidia-ml-py, pynvml, psutil, boto3 2025-10-10T02:02:09.9464059Z Attempting uninstall: nvidia-ml-py 2025-10-10T02:02:09.9466340Z Found existing installation: nvidia-ml-py 11.525.84 2025-10-10T02:02:09.9485954Z Uninstalling nvidia-ml-py-11.525.84: 2025-10-10T02:02:09.9719789Z Successfully uninstalled nvidia-ml-py-11.525.84 2025-10-10T02:02:10.0220437Z Attempting uninstall: psutil 2025-10-10T02:02:10.0222517Z Found existing installation: psutil 5.9.8 2025-10-10T02:02:10.0301809Z Uninstalling psutil-5.9.8: 2025-10-10T02:02:10.0308739Z Successfully uninstalled psutil-5.9.8 2025-10-10T02:02:10.1802593Z Successfully installed boto3-1.35.33 botocore-1.35.99 nvidia-ml-py-12.575.51 psutil-7.0.0 pynvml-12.0.0 s3transfer-0.10.4 2025-10-10T02:02:10.2721449Z + DEVICE_NAME= 2025-10-10T02:02:10.2721903Z + DEVICE_TYPE= 2025-10-10T02:02:10.2722312Z + command -v nvidia-smi 2025-10-10T02:02:10.2722793Z + python3 -mpip install torch==2.7.1 2025-10-10T02:02:10.2723348Z /usr/bin/nvidia-smi 2025-10-10T02:02:10.4910770Z Defaulting to user installation because normal site-packages is not writeable 2025-10-10T02:02:10.7308883Z Collecting torch==2.7.1 2025-10-10T02:02:10.8070914Z Downloading torch-2.7.1-cp39-cp39-manylinux_2_28_x86_64.whl (821.1 MB) 2025-10-10T02:02:31.1947171Z Collecting nvidia-curand-cu12==10.3.7.77 2025-10-10T02:02:31.2063097Z Downloading nvidia_curand_cu12-10.3.7.77-py3-none-manylinux2014_x86_64.manylinux_2_17_x86_64.whl (56.3 MB) 2025-10-10T02:02:32.2305149Z Collecting nvidia-cusolver-cu12==11.7.1.2 2025-10-10T02:02:32.2394174Z Downloading nvidia_cusolver_cu12-11.7.1.2-py3-none-manylinux2014_x86_64.manylinux_2_17_x86_64.whl (158.2 MB) 2025-10-10T02:02:35.6302974Z Collecting nvidia-cufft-cu12==11.3.0.4 2025-10-10T02:02:35.6361845Z Downloading nvidia_cufft_cu12-11.3.0.4-py3-none-manylinux2014_x86_64.manylinux_2_17_x86_64.whl (200.2 MB) 2025-10-10T02:02:39.1111679Z Collecting networkx 2025-10-10T02:02:39.1156949Z Downloading networkx-3.2.1-py3-none-any.whl (1.6 MB) 2025-10-10T02:02:39.1686454Z Collecting nvidia-cuda-runtime-cu12==12.6.77 2025-10-10T02:02:39.1754943Z Downloading nvidia_cuda_runtime_cu12-12.6.77-py3-none-manylinux2014_x86_64.manylinux_2_17_x86_64.whl (897 kB) 2025-10-10T02:02:39.2185600Z Collecting nvidia-nccl-cu12==2.26.2 2025-10-10T02:02:39.2271256Z Downloading nvidia_nccl_cu12-2.26.2-py3-none-manylinux2014_x86_64.manylinux_2_17_x86_64.whl (201.3 MB) 2025-10-10T02:02:42.6388353Z Collecting filelock 2025-10-10T02:02:42.6435707Z Downloading filelock-3.19.1-py3-none-any.whl (15 kB) 2025-10-10T02:02:42.6473099Z Requirement already satisfied: typing-extensions>=4.10.0 in /home/ec2-user/.local/lib/python3.9/site-packages (from torch==2.7.1) (4.15.0) 2025-10-10T02:02:42.6759068Z Collecting nvidia-nvtx-cu12==12.6.77 2025-10-10T02:02:42.6837567Z Downloading nvidia_nvtx_cu12-12.6.77-py3-none-manylinux2014_x86_64.manylinux_2_17_x86_64.whl (89 kB) 2025-10-10T02:02:42.7060332Z Collecting nvidia-cufile-cu12==1.11.1.6 2025-10-10T02:02:42.7156730Z Downloading nvidia_cufile_cu12-1.11.1.6-py3-none-manylinux2014_x86_64.manylinux_2_17_x86_64.whl (1.1 MB) 2025-10-10T02:02:42.7551600Z Collecting nvidia-cuda-nvrtc-cu12==12.6.77 2025-10-10T02:02:42.7675876Z Downloading nvidia_cuda_nvrtc_cu12-12.6.77-py3-none-manylinux2014_x86_64.whl (23.7 MB) 2025-10-10T02:02:43.0747448Z Collecting nvidia-cuda-cupti-cu12==12.6.80 2025-10-10T02:02:43.1054809Z Downloading nvidia_cuda_cupti_cu12-12.6.80-py3-none-manylinux2014_x86_64.manylinux_2_17_x86_64.whl (8.9 MB) 2025-10-10T02:02:43.3150498Z Collecting nvidia-nvjitlink-cu12==12.6.85 2025-10-10T02:02:43.3241701Z Downloading nvidia_nvjitlink_cu12-12.6.85-py3-none-manylinux2010_x86_64.manylinux_2_12_x86_64.whl (19.7 MB) 2025-10-10T02:02:43.6468071Z Collecting triton==3.3.1 2025-10-10T02:02:43.6545885Z Downloading triton-3.3.1-cp39-cp39-manylinux_2_27_x86_64.manylinux_2_28_x86_64.whl (155.6 MB) 2025-10-10T02:02:46.7283347Z Collecting nvidia-cudnn-cu12==9.5.1.17 2025-10-10T02:02:46.8614013Z Downloading nvidia_cudnn_cu12-9.5.1.17-py3-none-manylinux_2_28_x86_64.whl (571.0 MB) 2025-10-10T02:02:58.5557575Z Collecting nvidia-cusparselt-cu12==0.6.3 2025-10-10T02:02:58.5598187Z Downloading nvidia_cusparselt_cu12-0.6.3-py3-none-manylinux2014_x86_64.whl (156.8 MB) 2025-10-10T02:03:01.2066312Z Collecting fsspec 2025-10-10T02:03:01.2112680Z Downloading fsspec-2025.9.0-py3-none-any.whl (199 kB) 2025-10-10T02:03:01.2458038Z Collecting nvidia-cusparse-cu12==12.5.4.2 2025-10-10T02:03:01.2542628Z Downloading nvidia_cusparse_cu12-12.5.4.2-py3-none-manylinux2014_x86_64.manylinux_2_17_x86_64.whl (216.6 MB) 2025-10-10T02:03:04.6879112Z Collecting sympy>=1.13.3 2025-10-10T02:03:04.6929436Z Downloading sympy-1.14.0-py3-none-any.whl (6.3 MB) 2025-10-10T02:03:04.7942398Z Collecting nvidia-cublas-cu12==12.6.4.1 2025-10-10T02:03:04.7993988Z Downloading nvidia_cublas_cu12-12.6.4.1-py3-none-manylinux2014_x86_64.manylinux_2_17_x86_64.whl (393.1 MB) 2025-10-10T02:03:12.3153698Z Requirement already satisfied: jinja2 in /usr/lib/python3.9/site-packages (from torch==2.7.1) (2.11.3) 2025-10-10T02:03:12.3394732Z Requirement already satisfied: setuptools>=40.8.0 in /usr/lib/python3.9/site-packages (from triton==3.3.1->torch==2.7.1) (59.6.0) 2025-10-10T02:03:12.3671930Z Collecting mpmath<1.4,>=1.1.0 2025-10-10T02:03:12.3717034Z Downloading mpmath-1.3.0-py3-none-any.whl (536 kB) 2025-10-10T02:03:12.4513199Z Requirement already satisfied: MarkupSafe>=0.23 in /usr/lib64/python3.9/site-packages (from jinja2->torch==2.7.1) (1.1.1) 2025-10-10T02:03:12.7508098Z Installing collected packages: nvidia-nvjitlink-cu12, nvidia-cusparse-cu12, nvidia-cublas-cu12, mpmath, triton, sympy, nvidia-nvtx-cu12, nvidia-nccl-cu12, nvidia-cusparselt-cu12, nvidia-cusolver-cu12, nvidia-curand-cu12, nvidia-cufile-cu12, nvidia-cufft-cu12, nvidia-cudnn-cu12, nvidia-cuda-runtime-cu12, nvidia-cuda-nvrtc-cu12, nvidia-cuda-cupti-cu12, networkx, fsspec, filelock, torch 2025-10-10T02:03:20.4269072Z WARNING: The scripts proton and proton-viewer are installed in '/home/ec2-user/.local/bin' which is not on PATH. 2025-10-10T02:03:20.4269938Z Consider adding this directory to PATH or, if you prefer to suppress this warning, use --no-warn-script-location. 2025-10-10T02:03:23.8850550Z WARNING: The script isympy is installed in '/home/ec2-user/.local/bin' which is not on PATH. 2025-10-10T02:03:23.8851333Z Consider adding this directory to PATH or, if you prefer to suppress this warning, use --no-warn-script-location. 2025-10-10T02:03:50.4551956Z WARNING: The scripts torchfrtrace and torchrun are installed in '/home/ec2-user/.local/bin' which is not on PATH. 2025-10-10T02:03:50.4552832Z Consider adding this directory to PATH or, if you prefer to suppress this warning, use --no-warn-script-location. 2025-10-10T02:03:50.6494605Z Successfully installed filelock-3.19.1 fsspec-2025.9.0 mpmath-1.3.0 networkx-3.2.1 nvidia-cublas-cu12-12.6.4.1 nvidia-cuda-cupti-cu12-12.6.80 nvidia-cuda-nvrtc-cu12-12.6.77 nvidia-cuda-runtime-cu12-12.6.77 nvidia-cudnn-cu12-9.5.1.17 nvidia-cufft-cu12-11.3.0.4 nvidia-cufile-cu12-1.11.1.6 nvidia-curand-cu12-10.3.7.77 nvidia-cusolver-cu12-11.7.1.2 nvidia-cusparse-cu12-12.5.4.2 nvidia-cusparselt-cu12-0.6.3 nvidia-nccl-cu12-2.26.2 nvidia-nvjitlink-cu12-12.6.85 nvidia-nvtx-cu12-12.6.77 sympy-1.14.0 torch-2.7.1 triton-3.3.1 2025-10-10T02:03:51.1491606Z + echo DEVICE_NAME= 2025-10-10T02:03:51.1493898Z + echo DEVICE_TYPE= 2025-10-10T02:03:51.1524812Z ##[group]Run set -eux 2025-10-10T02:03:51.1525021Z set -eux 2025-10-10T02:03:51.1525225Z  2025-10-10T02:03:51.1525404Z if [[ -z "${GITHUB_TOKEN}" ]]; then 2025-10-10T02:03:51.1525669Z  echo "Missing github-token input" 2025-10-10T02:03:51.1525892Z  exit 1 2025-10-10T02:03:51.1526049Z fi 2025-10-10T02:03:51.1535659Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2025-10-10T02:03:51.1536000Z env: 2025-10-10T02:03:51.1536164Z GIT_DEFAULT_BRANCH: main 2025-10-10T02:03:51.1536415Z GPU_FLAG: --gpus all -e NVIDIA_DRIVER_CAPABILITIES=all 2025-10-10T02:03:51.1536999Z DOCKER_CONTAINER_ID: 92bc90bdf047b31a0d3a4858f15294a82b80fed2ff3cd44e76b438de5ef8fa55 2025-10-10T02:03:51.1537360Z DEVICE_NAME: 2025-10-10T02:03:51.1537530Z DEVICE_TYPE: 2025-10-10T02:03:51.1537924Z GITHUB_TOKEN: *** 2025-10-10T02:03:51.1538108Z ##[endgroup] 2025-10-10T02:03:51.1577613Z + [[ -z *** ]] 2025-10-10T02:03:51.1686662Z ##[group]Run pytorch/test-infra/.github/actions/get-workflow-job-id@main 2025-10-10T02:03:51.1686986Z with: 2025-10-10T02:03:51.1687275Z github-token: *** 2025-10-10T02:03:51.1687453Z env: 2025-10-10T02:03:51.1687610Z GIT_DEFAULT_BRANCH: main 2025-10-10T02:03:51.1687852Z GPU_FLAG: --gpus all -e NVIDIA_DRIVER_CAPABILITIES=all 2025-10-10T02:03:51.1688269Z DOCKER_CONTAINER_ID: 92bc90bdf047b31a0d3a4858f15294a82b80fed2ff3cd44e76b438de5ef8fa55 2025-10-10T02:03:51.1688625Z DEVICE_NAME: 2025-10-10T02:03:51.1688789Z DEVICE_TYPE: 2025-10-10T02:03:51.1688954Z ##[endgroup] 2025-10-10T02:03:51.1873644Z ##[group]Run set -eux 2025-10-10T02:03:51.1873842Z set -eux 2025-10-10T02:03:51.1874026Z  2025-10-10T02:03:51.1874367Z python3 "${GITHUB_ACTION_PATH}/../../scripts/get_workflow_job_id.py" "${GITHUB_RUN_ID}" "${RUNNER_NAME}" 2025-10-10T02:03:51.1882095Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2025-10-10T02:03:51.1882383Z env: 2025-10-10T02:03:51.1882546Z GIT_DEFAULT_BRANCH: main 2025-10-10T02:03:51.1882951Z GPU_FLAG: --gpus all -e NVIDIA_DRIVER_CAPABILITIES=all 2025-10-10T02:03:51.1883358Z DOCKER_CONTAINER_ID: 92bc90bdf047b31a0d3a4858f15294a82b80fed2ff3cd44e76b438de5ef8fa55 2025-10-10T02:03:51.1883726Z DEVICE_NAME: 2025-10-10T02:03:51.1883893Z DEVICE_TYPE: 2025-10-10T02:03:51.1884171Z GITHUB_TOKEN: *** 2025-10-10T02:03:51.1884357Z ##[endgroup] 2025-10-10T02:03:51.1916747Z + python3 /home/ec2-user/actions-runner/_work/_actions/pytorch/test-infra/main/.github/actions/get-workflow-job-id/../../scripts/get_workflow_job_id.py 18392306035 i-0b2d5251fda19aeb9 2025-10-10T02:03:51.8343611Z setting job-id=52408504942 2025-10-10T02:03:51.8344760Z setting job-name=ci-vllm-test / test (vllm_distributed_test_28_failure_test, 1, 1, linux.g6.12xlarge.nvidia.gpu) 2025-10-10T02:03:51.8561931Z ##[group]Run set -eux 2025-10-10T02:03:51.8562150Z set -eux 2025-10-10T02:03:51.8562312Z  2025-10-10T02:03:51.8562481Z if [[ -n "" ]]; then 2025-10-10T02:03:51.8562687Z  source "" 2025-10-10T02:03:51.8562863Z fi 2025-10-10T02:03:51.8563013Z  2025-10-10T02:03:51.8563294Z python3 "${GITHUB_ACTION_PATH}/../../scripts/benchmarks/gather_metadata.py" \ 2025-10-10T02:03:51.8563685Z  --schema-version "${SCHEMA_VERSION}" \ 2025-10-10T02:03:51.8563938Z  --repo "${REPO}" \ 2025-10-10T02:03:51.8564161Z  --head-branch "${HEAD_BRANCH}" \ 2025-10-10T02:03:51.8564412Z  --head-sha "${HEAD_SHA}" \ 2025-10-10T02:03:51.8564677Z  --workflow-id "${WORKFLOW_RUN_ID}" \ 2025-10-10T02:03:51.8564934Z  --run-attempt "${RUN_ATTEMPT}" \ 2025-10-10T02:03:51.8565394Z  --job-id "${JOB_ID}" \ 2025-10-10T02:03:51.8565629Z  --job-name "${JOB_NAME}" 2025-10-10T02:03:51.8573489Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2025-10-10T02:03:51.8573772Z env: 2025-10-10T02:03:51.8573927Z GIT_DEFAULT_BRANCH: main 2025-10-10T02:03:51.8574207Z GPU_FLAG: --gpus all -e NVIDIA_DRIVER_CAPABILITIES=all 2025-10-10T02:03:51.8574705Z DOCKER_CONTAINER_ID: 92bc90bdf047b31a0d3a4858f15294a82b80fed2ff3cd44e76b438de5ef8fa55 2025-10-10T02:03:51.8575119Z DEVICE_NAME: 2025-10-10T02:03:51.8575303Z DEVICE_TYPE: 2025-10-10T02:03:51.8575495Z SCHEMA_VERSION: v3 2025-10-10T02:03:51.8575710Z REPO: pytorch/pytorch 2025-10-10T02:03:51.8575933Z HEAD_BRANCH: refs/heads/main 2025-10-10T02:03:51.8576158Z HEAD_SHA: 344e6365a0068c2d2847fcec0c55dd53291d475e 2025-10-10T02:03:51.8576403Z WORKFLOW_RUN_ID: 18392306035 2025-10-10T02:03:51.8576611Z RUN_ATTEMPT: 1 2025-10-10T02:03:51.8576777Z JOB_ID: 52408504942 2025-10-10T02:03:51.8577261Z JOB_NAME: ci-vllm-test / test (vllm_distributed_test_28_failure_test, 1, 1, linux.g6.12xlarge.nvidia.gpu) 2025-10-10T02:03:51.8577665Z ##[endgroup] 2025-10-10T02:03:51.8609318Z + [[ -n '' ]] 2025-10-10T02:03:51.8612746Z + python3 /home/ec2-user/actions-runner/_work/_actions/pytorch/test-infra/main/.github/actions/upload-benchmark-results/../../scripts/benchmarks/gather_metadata.py --schema-version v3 --repo pytorch/pytorch --head-branch refs/heads/main --head-sha 344e6365a0068c2d2847fcec0c55dd53291d475e --workflow-id 18392306035 --run-attempt 1 --job-id 52408504942 --job-name 'ci-vllm-test / test (vllm_distributed_test_28_failure_test, 1, 1, linux.g6.12xlarge.nvidia.gpu)' 2025-10-10T02:03:51.9283054Z ##[group]Run set -eux 2025-10-10T02:03:51.9283278Z set -eux 2025-10-10T02:03:51.9283443Z  2025-10-10T02:03:51.9283601Z if [[ -n "" ]]; then 2025-10-10T02:03:51.9283802Z  source "" 2025-10-10T02:03:51.9283976Z fi 2025-10-10T02:03:51.9284149Z  2025-10-10T02:03:51.9284439Z python3 "${GITHUB_ACTION_PATH}/../../scripts/benchmarks/gather_runners_info.py" 2025-10-10T02:03:51.9292182Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2025-10-10T02:03:51.9292463Z env: 2025-10-10T02:03:51.9292620Z GIT_DEFAULT_BRANCH: main 2025-10-10T02:03:51.9292863Z GPU_FLAG: --gpus all -e NVIDIA_DRIVER_CAPABILITIES=all 2025-10-10T02:03:51.9293416Z DOCKER_CONTAINER_ID: 92bc90bdf047b31a0d3a4858f15294a82b80fed2ff3cd44e76b438de5ef8fa55 2025-10-10T02:03:51.9293788Z DEVICE_NAME: 2025-10-10T02:03:51.9293971Z DEVICE_TYPE: 2025-10-10T02:03:51.9294133Z ##[endgroup] 2025-10-10T02:03:51.9325630Z + [[ -n '' ]] 2025-10-10T02:03:51.9327092Z + python3 /home/ec2-user/actions-runner/_work/_actions/pytorch/test-infra/main/.github/actions/upload-benchmark-results/../../scripts/benchmarks/gather_runners_info.py 2025-10-10T02:03:52.8011703Z /home/ec2-user/.local/lib/python3.9/site-packages/torch/_subclasses/functional_tensor.py:276: UserWarning: Failed to initialize NumPy: No module named 'numpy' (Triggered internally at /pytorch/torch/csrc/utils/tensor_numpy.cpp:81.) 2025-10-10T02:03:52.8013821Z cpu = _conversion_method_template(device=torch.device("cpu")) 2025-10-10T02:03:54.1455039Z ##[group]Run set -eux 2025-10-10T02:03:54.1455251Z set -eux 2025-10-10T02:03:54.1455413Z  2025-10-10T02:03:54.1455607Z # TODO (huydhn): Implement this part 2025-10-10T02:03:54.1455889Z echo "dependencies={}" >> "${GITHUB_OUTPUT}" 2025-10-10T02:03:54.1464562Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2025-10-10T02:03:54.1464853Z env: 2025-10-10T02:03:54.1465011Z GIT_DEFAULT_BRANCH: main 2025-10-10T02:03:54.1465259Z GPU_FLAG: --gpus all -e NVIDIA_DRIVER_CAPABILITIES=all 2025-10-10T02:03:54.1465676Z DOCKER_CONTAINER_ID: 92bc90bdf047b31a0d3a4858f15294a82b80fed2ff3cd44e76b438de5ef8fa55 2025-10-10T02:03:54.1466049Z DEVICE_NAME: 2025-10-10T02:03:54.1466218Z DEVICE_TYPE: 2025-10-10T02:03:54.1466570Z ##[endgroup] 2025-10-10T02:03:54.1499611Z + echo 'dependencies={}' 2025-10-10T02:03:54.1634276Z ##[group]Run set -eux 2025-10-10T02:03:54.1634498Z set -eux 2025-10-10T02:03:54.1634659Z  2025-10-10T02:03:54.1634829Z if [[ -n "" ]]; then 2025-10-10T02:03:54.1635065Z  source "" 2025-10-10T02:03:54.1635267Z fi 2025-10-10T02:03:54.1635461Z  2025-10-10T02:03:54.1635683Z if [[ ! -d "${BENCHMARK_RESULTS_DIR}" ]]; then 2025-10-10T02:03:54.1636082Z  echo "${BENCHMARK_RESULTS_DIR} does not exist, skipping" 2025-10-10T02:03:54.1636498Z  # We don't want the job to fail if the directory doesn't exist 2025-10-10T02:03:54.1636798Z  exit 0 2025-10-10T02:03:54.1636952Z fi 2025-10-10T02:03:54.1637106Z  2025-10-10T02:03:54.1637272Z if [[ "${DRY_RUN}" == "true" ]]; then 2025-10-10T02:03:54.1637616Z  python3 "${GITHUB_ACTION_PATH}/../../scripts/upload_benchmark_results.py" \ 2025-10-10T02:03:54.1638175Z  --benchmark-results-dir "${BENCHMARK_RESULTS_DIR}" \ 2025-10-10T02:03:54.1638581Z  --metadata "${BENCHMARK_METADATA}" \ 2025-10-10T02:03:54.1638839Z  --runners "${RUNNER_INFO}" \ 2025-10-10T02:03:54.1639091Z  --dependencies "${DEPENDENCIES}" \ 2025-10-10T02:03:54.1639338Z  --dry-run 2025-10-10T02:03:54.1639519Z else 2025-10-10T02:03:54.1639803Z  python3 "${GITHUB_ACTION_PATH}/../../scripts/upload_benchmark_results.py" \ 2025-10-10T02:03:54.1640196Z  --benchmark-results-dir "${BENCHMARK_RESULTS_DIR}" \ 2025-10-10T02:03:54.1640512Z  --metadata "${BENCHMARK_METADATA}" \ 2025-10-10T02:03:54.1640766Z  --runners "${RUNNER_INFO}" \ 2025-10-10T02:03:54.1641017Z  --dependencies "${DEPENDENCIES}" 2025-10-10T02:03:54.1641244Z fi 2025-10-10T02:03:54.1649342Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2025-10-10T02:03:54.1649641Z env: 2025-10-10T02:03:54.1649809Z GIT_DEFAULT_BRANCH: main 2025-10-10T02:03:54.1650057Z GPU_FLAG: --gpus all -e NVIDIA_DRIVER_CAPABILITIES=all 2025-10-10T02:03:54.1650464Z DOCKER_CONTAINER_ID: 92bc90bdf047b31a0d3a4858f15294a82b80fed2ff3cd44e76b438de5ef8fa55 2025-10-10T02:03:54.1650834Z DEVICE_NAME: 2025-10-10T02:03:54.1650996Z DEVICE_TYPE: 2025-10-10T02:03:54.1651285Z BENCHMARK_RESULTS_DIR: test/test-reports 2025-10-10T02:03:54.1651524Z DRY_RUN: false 2025-10-10T02:03:54.1652495Z BENCHMARK_METADATA: {"timestamp": 1760061831, "schema_version": "v3", "name": "ci-vllm-test / test (vllm_distributed_test_28_failure_test, 1, 1, linux.g6.12xlarge.nvidia.gpu)", "repo": "pytorch/pytorch", "head_branch": "refs/heads/main", "head_sha": "344e6365a0068c2d2847fcec0c55dd53291d475e", "workflow_id": 18392306035, "run_attempt": 1, "job_id": 52408504942} 2025-10-10T02:03:54.1653846Z RUNNER_INFO: [{"cpu_info": "x86_64", "cpu_count": 48, "avail_mem_in_gb": 181, "extra_info": {"hostname": "ip-10-0-6-129.ec2.internal"}, "name": "cuda", "type": "NVIDIA L4", "gpu_count": 4, "avail_gpu_mem_in_gb": 88}] 2025-10-10T02:03:54.1654405Z DEPENDENCIES: {} 2025-10-10T02:03:54.1654572Z ##[endgroup] 2025-10-10T02:03:54.1689475Z + [[ -n '' ]] 2025-10-10T02:03:54.1689997Z test/test-reports does not exist, skipping 2025-10-10T02:03:54.1690707Z + [[ ! -d test/test-reports ]] 2025-10-10T02:03:54.1691371Z + echo 'test/test-reports does not exist, skipping' 2025-10-10T02:03:54.1691935Z + exit 0 2025-10-10T02:03:54.1855655Z ##[group]Run cat test/**/*_toprint.log || true 2025-10-10T02:03:54.1855978Z cat test/**/*_toprint.log || true 2025-10-10T02:03:54.1862952Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2025-10-10T02:03:54.1863233Z env: 2025-10-10T02:03:54.1863393Z GIT_DEFAULT_BRANCH: main 2025-10-10T02:03:54.1863637Z GPU_FLAG: --gpus all -e NVIDIA_DRIVER_CAPABILITIES=all 2025-10-10T02:03:54.1864033Z DOCKER_CONTAINER_ID: 92bc90bdf047b31a0d3a4858f15294a82b80fed2ff3cd44e76b438de5ef8fa55 2025-10-10T02:03:54.1864402Z DEVICE_NAME: 2025-10-10T02:03:54.1864566Z DEVICE_TYPE: 2025-10-10T02:03:54.1864724Z ##[endgroup] 2025-10-10T02:03:54.1963627Z cat: 'test/**/*_toprint.log': No such file or directory 2025-10-10T02:03:54.2044466Z ##[group]Run kill "$MONITOR_SCRIPT_PID" 2025-10-10T02:03:54.2044760Z kill "$MONITOR_SCRIPT_PID" 2025-10-10T02:03:54.2052066Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2025-10-10T02:03:54.2052350Z env: 2025-10-10T02:03:54.2052514Z GIT_DEFAULT_BRANCH: main 2025-10-10T02:03:54.2052766Z GPU_FLAG: --gpus all -e NVIDIA_DRIVER_CAPABILITIES=all 2025-10-10T02:03:54.2053173Z DOCKER_CONTAINER_ID: 92bc90bdf047b31a0d3a4858f15294a82b80fed2ff3cd44e76b438de5ef8fa55 2025-10-10T02:03:54.2053536Z DEVICE_NAME: 2025-10-10T02:03:54.2053700Z DEVICE_TYPE: 2025-10-10T02:03:54.2053872Z MONITOR_SCRIPT_PID: 67405 2025-10-10T02:03:54.2054055Z ##[endgroup] 2025-10-10T02:03:54.2201622Z Prepare all required actions 2025-10-10T02:03:54.2202004Z Getting action download info 2025-10-10T02:03:54.3726731Z Download action repository 'seemethere/upload-artifact-s3@v5' (SHA:baba72d0712b404f646cebe0730933554ebce96a) 2025-10-10T02:03:54.9338084Z Download action repository 'actions/upload-artifact@v4' (SHA:ea165f8d65b6e75b540449e92b4886f43607fa02) 2025-10-10T02:03:59.5161025Z ##[group]Run ./.github/actions/upload-test-artifacts 2025-10-10T02:03:59.5161356Z with: 2025-10-10T02:03:59.5161727Z file-suffix: test-vllm_distributed_test_28_failure_test-1-1-linux.g6.12xlarge.nvidia.gpu_52408504942 2025-10-10T02:03:59.5162132Z s3-bucket: gha-artifacts 2025-10-10T02:03:59.5162334Z env: 2025-10-10T02:03:59.5162491Z GIT_DEFAULT_BRANCH: main 2025-10-10T02:03:59.5162731Z GPU_FLAG: --gpus all -e NVIDIA_DRIVER_CAPABILITIES=all 2025-10-10T02:03:59.5163130Z DOCKER_CONTAINER_ID: 92bc90bdf047b31a0d3a4858f15294a82b80fed2ff3cd44e76b438de5ef8fa55 2025-10-10T02:03:59.5163512Z DEVICE_NAME: 2025-10-10T02:03:59.5163681Z DEVICE_TYPE: 2025-10-10T02:03:59.5163967Z ##[endgroup] 2025-10-10T02:03:59.5277498Z ##[group]Run # Remove any previous test jsons if they exist 2025-10-10T02:03:59.5277851Z # Remove any previous test jsons if they exist 2025-10-10T02:03:59.5278120Z rm -f test-jsons-*.zip 2025-10-10T02:03:59.5278428Z zip -r "test-jsons-${FILE_SUFFIX}.zip" test/test-reports -i '*.json' 2025-10-10T02:03:59.5286718Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2025-10-10T02:03:59.5287127Z env: 2025-10-10T02:03:59.5287299Z GIT_DEFAULT_BRANCH: main 2025-10-10T02:03:59.5287544Z GPU_FLAG: --gpus all -e NVIDIA_DRIVER_CAPABILITIES=all 2025-10-10T02:03:59.5287937Z DOCKER_CONTAINER_ID: 92bc90bdf047b31a0d3a4858f15294a82b80fed2ff3cd44e76b438de5ef8fa55 2025-10-10T02:03:59.5288296Z DEVICE_NAME: 2025-10-10T02:03:59.5288586Z DEVICE_TYPE: 2025-10-10T02:03:59.5288940Z FILE_SUFFIX: test-vllm_distributed_test_28_failure_test-1-1-linux.g6.12xlarge.nvidia.gpu_52408504942 2025-10-10T02:03:59.5289335Z ##[endgroup] 2025-10-10T02:03:59.5976500Z zip warning: name not matched: test/test-reports 2025-10-10T02:03:59.5978140Z zip warning: zip file empty 2025-10-10T02:03:59.6115977Z ##[group]Run # Remove any previous test reports if they exist 2025-10-10T02:03:59.6116337Z # Remove any previous test reports if they exist 2025-10-10T02:03:59.6116614Z rm -f test-reports-*.zip 2025-10-10T02:03:59.6116952Z zip -r "test-reports-${FILE_SUFFIX}.zip" test/test-reports -i '*.xml' -i '*.csv' 2025-10-10T02:03:59.6125525Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2025-10-10T02:03:59.6125812Z env: 2025-10-10T02:03:59.6125971Z GIT_DEFAULT_BRANCH: main 2025-10-10T02:03:59.6126239Z GPU_FLAG: --gpus all -e NVIDIA_DRIVER_CAPABILITIES=all 2025-10-10T02:03:59.6126646Z DOCKER_CONTAINER_ID: 92bc90bdf047b31a0d3a4858f15294a82b80fed2ff3cd44e76b438de5ef8fa55 2025-10-10T02:03:59.6127017Z DEVICE_NAME: 2025-10-10T02:03:59.6127213Z DEVICE_TYPE: 2025-10-10T02:03:59.6127559Z FILE_SUFFIX: test-vllm_distributed_test_28_failure_test-1-1-linux.g6.12xlarge.nvidia.gpu_52408504942 2025-10-10T02:03:59.6148370Z ##[endgroup] 2025-10-10T02:03:59.6231378Z zip warning: name not matched: test/test-reports 2025-10-10T02:03:59.6232378Z zip warning: zip file empty 2025-10-10T02:03:59.6295713Z ##[group]Run # Remove any previous usage logs if they exist 2025-10-10T02:03:59.6296051Z # Remove any previous usage logs if they exist 2025-10-10T02:03:59.6296320Z rm -f logs-*.zip 2025-10-10T02:03:59.6296578Z zip "logs-${FILE_SUFFIX}.zip" 'usage_log.txt' || true 2025-10-10T02:03:59.6296941Z zip -r "logs-${FILE_SUFFIX}.zip" test/test-reports -i '*.log' || true 2025-10-10T02:03:59.6303977Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2025-10-10T02:03:59.6304256Z env: 2025-10-10T02:03:59.6304415Z GIT_DEFAULT_BRANCH: main 2025-10-10T02:03:59.6304660Z GPU_FLAG: --gpus all -e NVIDIA_DRIVER_CAPABILITIES=all 2025-10-10T02:03:59.6305254Z DOCKER_CONTAINER_ID: 92bc90bdf047b31a0d3a4858f15294a82b80fed2ff3cd44e76b438de5ef8fa55 2025-10-10T02:03:59.6305727Z DEVICE_NAME: 2025-10-10T02:03:59.6305894Z DEVICE_TYPE: 2025-10-10T02:03:59.6306253Z FILE_SUFFIX: test-vllm_distributed_test_28_failure_test-1-1-linux.g6.12xlarge.nvidia.gpu_52408504942 2025-10-10T02:03:59.6306634Z ##[endgroup] 2025-10-10T02:03:59.6489393Z adding: usage_log.txt (deflated 96%) 2025-10-10T02:03:59.6507794Z zip warning: name not matched: test/test-reports 2025-10-10T02:03:59.6508077Z 2025-10-10T02:03:59.6508496Z zip error: Nothing to do! (logs-test-vllm_distributed_test_28_failure_test-1-1-linux.g6.12xlarge.nvidia.gpu_52408504942.zip) 2025-10-10T02:03:59.6566673Z ##[group]Run # Remove any previous debugging artifacts if they exist 2025-10-10T02:03:59.6567071Z # Remove any previous debugging artifacts if they exist 2025-10-10T02:03:59.6567386Z rm -f debug-*.zip 2025-10-10T02:03:59.6567657Z if [ -d 'test/debug' ]; then 2025-10-10T02:03:59.6567956Z  zip -r "debug-${FILE_SUFFIX}.zip" test/debug 2025-10-10T02:03:59.6568261Z fi 2025-10-10T02:03:59.6575783Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2025-10-10T02:03:59.6576079Z env: 2025-10-10T02:03:59.6576240Z GIT_DEFAULT_BRANCH: main 2025-10-10T02:03:59.6576490Z GPU_FLAG: --gpus all -e NVIDIA_DRIVER_CAPABILITIES=all 2025-10-10T02:03:59.6576901Z DOCKER_CONTAINER_ID: 92bc90bdf047b31a0d3a4858f15294a82b80fed2ff3cd44e76b438de5ef8fa55 2025-10-10T02:03:59.6577381Z DEVICE_NAME: 2025-10-10T02:03:59.6577539Z DEVICE_TYPE: 2025-10-10T02:03:59.6577884Z FILE_SUFFIX: test-vllm_distributed_test_28_failure_test-1-1-linux.g6.12xlarge.nvidia.gpu_52408504942 2025-10-10T02:03:59.6578274Z ##[endgroup] 2025-10-10T02:03:59.6689071Z ##[group]Run seemethere/upload-artifact-s3@v5 2025-10-10T02:03:59.6689334Z with: 2025-10-10T02:03:59.6689512Z s3-bucket: gha-artifacts 2025-10-10T02:03:59.6689783Z s3-prefix: pytorch/pytorch/18392306035/1/artifact 2025-10-10T02:03:59.6690048Z retention-days: 14 2025-10-10T02:03:59.6690230Z if-no-files-found: warn 2025-10-10T02:03:59.6690453Z path: test-jsons-*.zip 2025-10-10T02:03:59.6690656Z name: artifact 2025-10-10T02:03:59.6690857Z region: us-east-1 2025-10-10T02:03:59.6691030Z env: 2025-10-10T02:03:59.6691192Z GIT_DEFAULT_BRANCH: main 2025-10-10T02:03:59.6691438Z GPU_FLAG: --gpus all -e NVIDIA_DRIVER_CAPABILITIES=all 2025-10-10T02:03:59.6691850Z DOCKER_CONTAINER_ID: 92bc90bdf047b31a0d3a4858f15294a82b80fed2ff3cd44e76b438de5ef8fa55 2025-10-10T02:03:59.6692209Z DEVICE_NAME: 2025-10-10T02:03:59.6692377Z DEVICE_TYPE: 2025-10-10T02:03:59.6692541Z ##[endgroup] 2025-10-10T02:04:00.1201882Z NOTE: s3-prefix specified, ignoring name parameter 2025-10-10T02:04:00.1202301Z With the provided path, there will be 1 file uploaded 2025-10-10T02:04:00.1202732Z Uploading to s3 prefix: pytorch/pytorch/18392306035/1/artifact 2025-10-10T02:04:00.1272482Z Starting upload of test-jsons-test-vllm_distributed_test_28_failure_test-1-1-linux.g6.12xlarge.nvidia.gpu_52408504942.zip 2025-10-10T02:04:00.2308239Z Finished upload of test-jsons-test-vllm_distributed_test_28_failure_test-1-1-linux.g6.12xlarge.nvidia.gpu_52408504942.zip 2025-10-10T02:04:00.3671643Z ##[group]Run seemethere/upload-artifact-s3@v5 2025-10-10T02:04:00.3671912Z with: 2025-10-10T02:04:00.3672089Z s3-bucket: gha-artifacts 2025-10-10T02:04:00.3672346Z s3-prefix: pytorch/pytorch/18392306035/1/artifact 2025-10-10T02:04:00.3672634Z retention-days: 14 2025-10-10T02:04:00.3672819Z if-no-files-found: error 2025-10-10T02:04:00.3673048Z path: test-reports-*.zip 2025-10-10T02:04:00.3673253Z name: artifact 2025-10-10T02:04:00.3673426Z region: us-east-1 2025-10-10T02:04:00.3673588Z env: 2025-10-10T02:04:00.3673748Z GIT_DEFAULT_BRANCH: main 2025-10-10T02:04:00.3674006Z GPU_FLAG: --gpus all -e NVIDIA_DRIVER_CAPABILITIES=all 2025-10-10T02:04:00.3674414Z DOCKER_CONTAINER_ID: 92bc90bdf047b31a0d3a4858f15294a82b80fed2ff3cd44e76b438de5ef8fa55 2025-10-10T02:04:00.3674989Z DEVICE_NAME: 2025-10-10T02:04:00.3675169Z DEVICE_TYPE: 2025-10-10T02:04:00.3675436Z ##[endgroup] 2025-10-10T02:04:00.9469607Z NOTE: s3-prefix specified, ignoring name parameter 2025-10-10T02:04:00.9470444Z With the provided path, there will be 1 file uploaded 2025-10-10T02:04:00.9471226Z Uploading to s3 prefix: pytorch/pytorch/18392306035/1/artifact 2025-10-10T02:04:00.9538877Z Starting upload of test-reports-test-vllm_distributed_test_28_failure_test-1-1-linux.g6.12xlarge.nvidia.gpu_52408504942.zip 2025-10-10T02:04:01.0638881Z Finished upload of test-reports-test-vllm_distributed_test_28_failure_test-1-1-linux.g6.12xlarge.nvidia.gpu_52408504942.zip 2025-10-10T02:04:01.1009699Z ##[group]Run seemethere/upload-artifact-s3@v5 2025-10-10T02:04:01.1009953Z with: 2025-10-10T02:04:01.1010117Z s3-bucket: gha-artifacts 2025-10-10T02:04:01.1010361Z s3-prefix: pytorch/pytorch/18392306035/1/artifact 2025-10-10T02:04:01.1010618Z retention-days: 14 2025-10-10T02:04:01.1010818Z if-no-files-found: ignore 2025-10-10T02:04:01.1011016Z path: logs-*.zip 2025-10-10T02:04:01.1011185Z name: artifact 2025-10-10T02:04:01.1011360Z region: us-east-1 2025-10-10T02:04:01.1011514Z env: 2025-10-10T02:04:01.1011678Z GIT_DEFAULT_BRANCH: main 2025-10-10T02:04:01.1011922Z GPU_FLAG: --gpus all -e NVIDIA_DRIVER_CAPABILITIES=all 2025-10-10T02:04:01.1012319Z DOCKER_CONTAINER_ID: 92bc90bdf047b31a0d3a4858f15294a82b80fed2ff3cd44e76b438de5ef8fa55 2025-10-10T02:04:01.1012799Z DEVICE_NAME: 2025-10-10T02:04:01.1012961Z DEVICE_TYPE: 2025-10-10T02:04:01.1013121Z ##[endgroup] 2025-10-10T02:04:01.3965954Z NOTE: s3-prefix specified, ignoring name parameter 2025-10-10T02:04:01.3966421Z With the provided path, there will be 1 file uploaded 2025-10-10T02:04:01.3966834Z Uploading to s3 prefix: pytorch/pytorch/18392306035/1/artifact 2025-10-10T02:04:01.4035031Z Starting upload of logs-test-vllm_distributed_test_28_failure_test-1-1-linux.g6.12xlarge.nvidia.gpu_52408504942.zip 2025-10-10T02:04:01.5318311Z Finished upload of logs-test-vllm_distributed_test_28_failure_test-1-1-linux.g6.12xlarge.nvidia.gpu_52408504942.zip 2025-10-10T02:04:01.5728913Z ##[group]Run seemethere/upload-artifact-s3@v5 2025-10-10T02:04:01.5729163Z with: 2025-10-10T02:04:01.5729337Z s3-bucket: gha-artifacts 2025-10-10T02:04:01.5729590Z s3-prefix: pytorch/pytorch/18392306035/1/artifact 2025-10-10T02:04:01.5729847Z retention-days: 14 2025-10-10T02:04:01.5730028Z if-no-files-found: ignore 2025-10-10T02:04:01.5730242Z path: debug-*.zip 2025-10-10T02:04:01.5730414Z name: artifact 2025-10-10T02:04:01.5730577Z region: us-east-1 2025-10-10T02:04:01.5730730Z env: 2025-10-10T02:04:01.5730890Z GIT_DEFAULT_BRANCH: main 2025-10-10T02:04:01.5731135Z GPU_FLAG: --gpus all -e NVIDIA_DRIVER_CAPABILITIES=all 2025-10-10T02:04:01.5731530Z DOCKER_CONTAINER_ID: 92bc90bdf047b31a0d3a4858f15294a82b80fed2ff3cd44e76b438de5ef8fa55 2025-10-10T02:04:01.5731876Z DEVICE_NAME: 2025-10-10T02:04:01.5732043Z DEVICE_TYPE: 2025-10-10T02:04:01.5732218Z ##[endgroup] 2025-10-10T02:04:01.8611919Z No files were found with the provided path: debug-*.zip. No artifacts will be uploaded. 2025-10-10T02:04:01.8967570Z ##[group]Run # shellcheck disable=SC2156 2025-10-10T02:04:01.8967967Z # shellcheck disable=SC2156 2025-10-10T02:04:01.8968409Z find . -iname "core.[1-9]*" -exec docker exec "${DOCKER_CONTAINER_ID}" sh -c "gdb python {} -ex 'bt' -ex 'q'" \; 2025-10-10T02:04:01.8977248Z shell: /usr/bin/bash -e {0} 2025-10-10T02:04:01.8977475Z env: 2025-10-10T02:04:01.8977637Z GIT_DEFAULT_BRANCH: main 2025-10-10T02:04:01.8977883Z GPU_FLAG: --gpus all -e NVIDIA_DRIVER_CAPABILITIES=all 2025-10-10T02:04:01.8978308Z DOCKER_CONTAINER_ID: 92bc90bdf047b31a0d3a4858f15294a82b80fed2ff3cd44e76b438de5ef8fa55 2025-10-10T02:04:01.8978680Z DEVICE_NAME: 2025-10-10T02:04:01.8978856Z DEVICE_TYPE: 2025-10-10T02:04:01.8979017Z ##[endgroup] 2025-10-10T02:04:02.3816504Z Prepare all required actions 2025-10-10T02:04:02.3816877Z Getting action download info 2025-10-10T02:04:02.5374721Z ##[group]Run ./.github/actions/upload-utilization-stats 2025-10-10T02:04:02.5375083Z with: 2025-10-10T02:04:02.5375243Z job_id: 52408504942 2025-10-10T02:04:02.5375620Z job_name: ci-vllm-test / test (vllm_distributed_test_28_failure_test, 1, 1, linux.g6.12xlarge.nvidia.gpu) 2025-10-10T02:04:02.5376016Z workflow_name: vllm-test 2025-10-10T02:04:02.5376208Z workflow_run_id: 18392306035 2025-10-10T02:04:02.5376405Z workflow_attempt: 1 2025-10-10T02:04:02.5376586Z env: 2025-10-10T02:04:02.5376738Z GIT_DEFAULT_BRANCH: main 2025-10-10T02:04:02.5376969Z GPU_FLAG: --gpus all -e NVIDIA_DRIVER_CAPABILITIES=all 2025-10-10T02:04:02.5377366Z DOCKER_CONTAINER_ID: 92bc90bdf047b31a0d3a4858f15294a82b80fed2ff3cd44e76b438de5ef8fa55 2025-10-10T02:04:02.5377717Z DEVICE_NAME: 2025-10-10T02:04:02.5377883Z DEVICE_TYPE: 2025-10-10T02:04:02.5378037Z ##[endgroup] 2025-10-10T02:04:02.5565755Z ##[group]Run echo "workflow_id: 18392306035" 2025-10-10T02:04:02.5566043Z echo "workflow_id: 18392306035" 2025-10-10T02:04:02.5566279Z echo "workflow_attempt: 1" 2025-10-10T02:04:02.5566502Z echo "workflow_Name: vllm-test" 2025-10-10T02:04:02.5566726Z echo "job_id: 52408504942" 2025-10-10T02:04:02.5567145Z echo "job_name: ci-vllm-test / test (vllm_distributed_test_28_failure_test, 1, 1, linux.g6.12xlarge.nvidia.gpu)" 2025-10-10T02:04:02.5567581Z echo "artifact_prefix: " 2025-10-10T02:04:02.5567800Z python3 --version 2025-10-10T02:04:02.5575918Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2025-10-10T02:04:02.5576192Z env: 2025-10-10T02:04:02.5576349Z GIT_DEFAULT_BRANCH: main 2025-10-10T02:04:02.5576593Z GPU_FLAG: --gpus all -e NVIDIA_DRIVER_CAPABILITIES=all 2025-10-10T02:04:02.5576981Z DOCKER_CONTAINER_ID: 92bc90bdf047b31a0d3a4858f15294a82b80fed2ff3cd44e76b438de5ef8fa55 2025-10-10T02:04:02.5577368Z DEVICE_NAME: 2025-10-10T02:04:02.5577536Z DEVICE_TYPE: 2025-10-10T02:04:02.5577693Z ##[endgroup] 2025-10-10T02:04:02.5611236Z workflow_id: 18392306035 2025-10-10T02:04:02.5611792Z workflow_attempt: 1 2025-10-10T02:04:02.5612259Z workflow_Name: vllm-test 2025-10-10T02:04:02.5612722Z job_id: 52408504942 2025-10-10T02:04:02.5613569Z job_name: ci-vllm-test / test (vllm_distributed_test_28_failure_test, 1, 1, linux.g6.12xlarge.nvidia.gpu) 2025-10-10T02:04:02.5614524Z artifact_prefix: 2025-10-10T02:04:02.5631657Z Python 3.9.23 2025-10-10T02:04:02.5898690Z ##[group]Run nick-fields/retry@v3.0.0 2025-10-10T02:04:02.5898943Z with: 2025-10-10T02:04:02.5899103Z shell: bash 2025-10-10T02:04:02.5899273Z timeout_minutes: 5 2025-10-10T02:04:02.5899458Z max_attempts: 5 2025-10-10T02:04:02.5899647Z retry_wait_seconds: 30 2025-10-10T02:04:02.5900075Z command: set -eu python3 -m pip install python-dateutil==2.8.2 boto3==1.35.42 pandas==2.1.3 dataclasses_json==0.6.7 2025-10-10T02:04:02.5900554Z polling_interval_seconds: 1 2025-10-10T02:04:02.5900791Z warning_on_retry: true 2025-10-10T02:04:02.5900994Z continue_on_error: false 2025-10-10T02:04:02.5901207Z env: 2025-10-10T02:04:02.5901389Z GIT_DEFAULT_BRANCH: main 2025-10-10T02:04:02.5901656Z GPU_FLAG: --gpus all -e NVIDIA_DRIVER_CAPABILITIES=all 2025-10-10T02:04:02.5902066Z DOCKER_CONTAINER_ID: 92bc90bdf047b31a0d3a4858f15294a82b80fed2ff3cd44e76b438de5ef8fa55 2025-10-10T02:04:02.5902435Z DEVICE_NAME: 2025-10-10T02:04:02.5902605Z DEVICE_TYPE: 2025-10-10T02:04:02.5902765Z ##[endgroup] 2025-10-10T02:04:02.9175487Z Defaulting to user installation because normal site-packages is not writeable 2025-10-10T02:04:03.0567529Z Collecting python-dateutil==2.8.2 2025-10-10T02:04:03.0767418Z Downloading python_dateutil-2.8.2-py2.py3-none-any.whl (247 kB) 2025-10-10T02:04:04.0926542Z Collecting boto3==1.35.42 2025-10-10T02:04:04.0980791Z Downloading boto3-1.35.42-py3-none-any.whl (139 kB) 2025-10-10T02:04:04.6776287Z Collecting pandas==2.1.3 2025-10-10T02:04:04.6837349Z Downloading pandas-2.1.3-cp39-cp39-manylinux_2_17_x86_64.manylinux2014_x86_64.whl (12.3 MB) 2025-10-10T02:04:05.0096949Z Requirement already satisfied: dataclasses_json==0.6.7 in /home/ec2-user/.local/lib/python3.9/site-packages (0.6.7) 2025-10-10T02:04:05.0112333Z Requirement already satisfied: six>=1.5 in /usr/lib/python3.9/site-packages (from python-dateutil==2.8.2) (1.15.0) 2025-10-10T02:04:05.0153942Z Requirement already satisfied: jmespath<2.0.0,>=0.7.1 in /usr/lib/python3.9/site-packages (from boto3==1.35.42) (0.10.0) 2025-10-10T02:04:05.0158740Z Requirement already satisfied: botocore<1.36.0,>=1.35.42 in /home/ec2-user/.local/lib/python3.9/site-packages (from boto3==1.35.42) (1.35.99) 2025-10-10T02:04:05.0161490Z Requirement already satisfied: s3transfer<0.11.0,>=0.10.0 in /home/ec2-user/.local/lib/python3.9/site-packages (from boto3==1.35.42) (0.10.4) 2025-10-10T02:04:05.1406599Z Collecting tzdata>=2022.1 2025-10-10T02:04:05.1456639Z Downloading tzdata-2025.2-py2.py3-none-any.whl (347 kB) 2025-10-10T02:04:06.0237337Z Collecting numpy<2,>=1.22.4 2025-10-10T02:04:06.0304316Z Downloading numpy-1.26.4-cp39-cp39-manylinux_2_17_x86_64.manylinux2014_x86_64.whl (18.2 MB) 2025-10-10T02:04:06.4642206Z Requirement already satisfied: pytz>=2020.1 in /usr/lib/python3.9/site-packages (from pandas==2.1.3) (2022.7.1) 2025-10-10T02:04:06.4678277Z Requirement already satisfied: typing-inspect<1,>=0.4.0 in /home/ec2-user/.local/lib/python3.9/site-packages (from dataclasses_json==0.6.7) (0.9.0) 2025-10-10T02:04:06.4681340Z Requirement already satisfied: marshmallow<4.0.0,>=3.18.0 in /home/ec2-user/.local/lib/python3.9/site-packages (from dataclasses_json==0.6.7) (3.26.1) 2025-10-10T02:04:06.4760358Z Requirement already satisfied: urllib3<1.27,>=1.25.4 in /usr/lib/python3.9/site-packages (from botocore<1.36.0,>=1.35.42->boto3==1.35.42) (1.25.10) 2025-10-10T02:04:06.4839934Z Requirement already satisfied: packaging>=17.0 in /home/ec2-user/.local/lib/python3.9/site-packages (from marshmallow<4.0.0,>=3.18.0->dataclasses_json==0.6.7) (25.0) 2025-10-10T02:04:06.4933467Z Requirement already satisfied: typing-extensions>=3.7.4 in /home/ec2-user/.local/lib/python3.9/site-packages (from typing-inspect<1,>=0.4.0->dataclasses_json==0.6.7) (4.15.0) 2025-10-10T02:04:06.4936441Z Requirement already satisfied: mypy-extensions>=0.3.0 in /home/ec2-user/.local/lib/python3.9/site-packages (from typing-inspect<1,>=0.4.0->dataclasses_json==0.6.7) (1.1.0) 2025-10-10T02:04:06.7671989Z Installing collected packages: python-dateutil, tzdata, numpy, pandas, boto3 2025-10-10T02:04:11.9574405Z Attempting uninstall: boto3 2025-10-10T02:04:11.9575127Z Found existing installation: boto3 1.35.33 2025-10-10T02:04:11.9685893Z Uninstalling boto3-1.35.33: 2025-10-10T02:04:11.9700898Z Successfully uninstalled boto3-1.35.33 2025-10-10T02:04:12.2063285Z Successfully installed boto3-1.35.42 numpy-1.26.4 pandas-2.1.3 python-dateutil-2.8.2 tzdata-2025.2 2025-10-10T02:04:12.6695778Z Command completed after 1 attempt(s). 2025-10-10T02:04:12.7021311Z ##[group]Run python3 -m tools.stats.upload_utilization_stats.upload_utilization_stats \ 2025-10-10T02:04:12.7021846Z python3 -m tools.stats.upload_utilization_stats.upload_utilization_stats \ 2025-10-10T02:04:12.7022231Z  --workflow-run-id "18392306035" \ 2025-10-10T02:04:12.7022496Z  --workflow-name "vllm-test" \ 2025-10-10T02:04:12.7022736Z  --workflow-run-attempt "1" \ 2025-10-10T02:04:12.7022968Z  --job-id "52408504942" \ 2025-10-10T02:04:12.7023375Z  --job-name "ci-vllm-test / test (vllm_distributed_test_28_failure_test, 1, 1, linux.g6.12xlarge.nvidia.gpu)" \ 2025-10-10T02:04:12.7023805Z  --local-path "" \ 2025-10-10T02:04:12.7024251Z  --artifact-prefix "" 2025-10-10T02:04:12.7032197Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2025-10-10T02:04:12.7032478Z env: 2025-10-10T02:04:12.7032639Z GIT_DEFAULT_BRANCH: main 2025-10-10T02:04:12.7033071Z GPU_FLAG: --gpus all -e NVIDIA_DRIVER_CAPABILITIES=all 2025-10-10T02:04:12.7033495Z DOCKER_CONTAINER_ID: 92bc90bdf047b31a0d3a4858f15294a82b80fed2ff3cd44e76b438de5ef8fa55 2025-10-10T02:04:12.7033850Z DEVICE_NAME: 2025-10-10T02:04:12.7034008Z DEVICE_TYPE: 2025-10-10T02:04:12.7034277Z ##[endgroup] 2025-10-10T02:04:15.5279497Z repo: pytorch/pytorch 2025-10-10T02:04:15.5279861Z Search for test log in s3 bucket: ossci-utilization 2025-10-10T02:04:15.5280479Z Downloading logs-test-vllm_distributed_test_28_failure_test-1-1-linux.g6.12xlarge.nvidia.gpu_52408504942.zip 2025-10-10T02:04:15.5281378Z extracting usage_log.txt from zip file logs-test-vllm_distributed_test_28_failure_test-1-1-linux.g6.12xlarge.nvidia.gpu_52408504942.zip 2025-10-10T02:04:15.5282028Z Converted Log Model: UtilizationMetadata: 2025-10-10T02:04:15.5283511Z UtilizationMetadata(level='metadata', workflow_id='18392306035', job_id='52408504942', workflow_name='vllm-test', job_name='ci-vllm-test / test (vllm_distributed_test_28_failure_test, 1, 1, linux.g6.12xlarge.nvidia.gpu)', usage_collect_interval=1.0, data_model_version=1.5, start_at=1760059590, gpu_count=4, cpu_count=48, gpu_type='pynvml', error=None) 2025-10-10T02:04:15.5284686Z [Db Segments] detected pytest cmd: 13, generated segments: 13 2025-10-10T02:04:15.5284979Z [db model] Peek db timeseries 2025-10-10T02:04:15.5285167Z :{ 2025-10-10T02:04:15.5285317Z "created_at": 1760061854, 2025-10-10T02:04:15.5285506Z "type": "utilization", 2025-10-10T02:04:15.5285684Z "tags": [ 2025-10-10T02:04:15.5285830Z "record" 2025-10-10T02:04:15.5286010Z ], 2025-10-10T02:04:15.5286161Z "time_stamp": 1760059590, 2025-10-10T02:04:15.5286708Z "repo": "pytorch/pytorch", 2025-10-10T02:04:15.5286907Z "workflow_id": 18392306035, 2025-10-10T02:04:15.5287097Z "run_attempt": 1, 2025-10-10T02:04:15.5287263Z "job_id": 52408504942, 2025-10-10T02:04:15.5287450Z "workflow_name": "vllm-test", 2025-10-10T02:04:15.5287846Z "job_name": "ci-vllm-test / test (vllm_distributed_test_28_failure_test, 1, 1, linux.g6.12xlarge.nvidia.gpu)", 2025-10-10T02:04:15.5288233Z "json_data": "{}" 2025-10-10T02:04:15.5288387Z } 2025-10-10T02:04:15.5288737Z Writing 1 documents to S3 ossci-utilization/util_metadata/v_1.5/pytorch/pytorch/18392306035/1/52408504942/metadata 2025-10-10T02:04:15.5289374Z Done! Finish writing document to S3 ossci-utilization/util_metadata/v_1.5/pytorch/pytorch/18392306035/1/52408504942/metadata 2025-10-10T02:04:15.5290030Z Writing 449 documents to S3 ossci-utilization/util_timeseries/v_1.5/pytorch/pytorch/18392306035/1/52408504942/time_series 2025-10-10T02:04:15.5290712Z Done! Finish writing document to S3 ossci-utilization/util_timeseries/v_1.5/pytorch/pytorch/18392306035/1/52408504942/time_series 2025-10-10T02:04:15.6379720Z ##[group]Run pytorch/test-infra/.github/actions/teardown-linux@main 2025-10-10T02:04:15.6380104Z with: 2025-10-10T02:04:15.6380258Z env: 2025-10-10T02:04:15.6380411Z GIT_DEFAULT_BRANCH: main 2025-10-10T02:04:15.6380674Z GPU_FLAG: --gpus all -e NVIDIA_DRIVER_CAPABILITIES=all 2025-10-10T02:04:15.6381074Z DOCKER_CONTAINER_ID: 92bc90bdf047b31a0d3a4858f15294a82b80fed2ff3cd44e76b438de5ef8fa55 2025-10-10T02:04:15.6381433Z DEVICE_NAME: 2025-10-10T02:04:15.6381602Z DEVICE_TYPE: 2025-10-10T02:04:15.6381776Z ##[endgroup] 2025-10-10T02:04:15.6522216Z ##[group]Run set -eou pipefail 2025-10-10T02:04:15.6522578Z set -eou pipefail 2025-10-10T02:04:15.6522782Z  2025-10-10T02:04:15.6523049Z echo "Holding runner for 2 hours until all ssh sessions have logged out" 2025-10-10T02:04:15.6523372Z for _ in $(seq 1440); do 2025-10-10T02:04:15.6523617Z  # Break if no ssh session exists anymore 2025-10-10T02:04:15.6523888Z  if [ "$(who)" = "" ]; then 2025-10-10T02:04:15.6524389Z  break 2025-10-10T02:04:15.6524564Z  fi 2025-10-10T02:04:15.6524728Z  echo "." 2025-10-10T02:04:15.6524900Z  sleep 5 2025-10-10T02:04:15.6525060Z done 2025-10-10T02:04:15.6534166Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2025-10-10T02:04:15.6534448Z env: 2025-10-10T02:04:15.6534612Z GIT_DEFAULT_BRANCH: main 2025-10-10T02:04:15.6534861Z GPU_FLAG: --gpus all -e NVIDIA_DRIVER_CAPABILITIES=all 2025-10-10T02:04:15.6535366Z DOCKER_CONTAINER_ID: 92bc90bdf047b31a0d3a4858f15294a82b80fed2ff3cd44e76b438de5ef8fa55 2025-10-10T02:04:15.6535724Z DEVICE_NAME: 2025-10-10T02:04:15.6535887Z DEVICE_TYPE: 2025-10-10T02:04:15.6536040Z ##[endgroup] 2025-10-10T02:04:15.6566895Z Holding runner for 2 hours until all ssh sessions have logged out 2025-10-10T02:04:15.7792052Z ##[group]Run # ignore expansion of "docker ps -q" since it could be empty 2025-10-10T02:04:15.7792486Z # ignore expansion of "docker ps -q" since it could be empty 2025-10-10T02:04:15.7792795Z # shellcheck disable=SC2046 2025-10-10T02:04:15.7793051Z docker stop $(docker ps -q) || true 2025-10-10T02:04:15.7793299Z # Prune all of the docker images 2025-10-10T02:04:15.7793537Z docker system prune -af 2025-10-10T02:04:15.7800784Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2025-10-10T02:04:15.7801088Z env: 2025-10-10T02:04:15.7801252Z GIT_DEFAULT_BRANCH: main 2025-10-10T02:04:15.7801511Z GPU_FLAG: --gpus all -e NVIDIA_DRIVER_CAPABILITIES=all 2025-10-10T02:04:15.7801927Z DOCKER_CONTAINER_ID: 92bc90bdf047b31a0d3a4858f15294a82b80fed2ff3cd44e76b438de5ef8fa55 2025-10-10T02:04:15.7802289Z DEVICE_NAME: 2025-10-10T02:04:15.7802461Z DEVICE_TYPE: 2025-10-10T02:04:15.7802627Z ##[endgroup] 2025-10-10T02:04:28.7878291Z 92bc90bdf047 2025-10-10T02:04:37.0133707Z Deleted Containers: 2025-10-10T02:04:37.0137799Z 92bc90bdf047b31a0d3a4858f15294a82b80fed2ff3cd44e76b438de5ef8fa55 2025-10-10T02:04:37.0138485Z 2025-10-10T02:04:48.5870059Z Deleted Images: 2025-10-10T02:04:48.5870944Z untagged: 308535385114.dkr.ecr.us-east-1.amazonaws.com/pytorch/ci-image:pytorch-linux-jammy-cuda12.8-cudnn9-py3.12-gcc11-vllm-d8be0384e085f551506bd739678109fa0f5ee7ac 2025-10-10T02:04:48.5872159Z untagged: 308535385114.dkr.ecr.us-east-1.amazonaws.com/pytorch/ci-image@sha256:dbf8617838c93f0ebd67b8b8bb938f5947e45ccc171ead9c6da61e625c32803f 2025-10-10T02:04:48.5872996Z deleted: sha256:fb0d053588b3e7d110638a9f678211e76aa4515ed1bad57389e38e2ab079e7b1 2025-10-10T02:04:48.5873603Z deleted: sha256:839691e084b3ec6bcbfe66d9e2f381fd843ef9f96565a69ff80aa86506df4ab2 2025-10-10T02:04:48.5874272Z deleted: sha256:14b10d5599eb456683f22fba693584a0ef542f8590cc214f13f622cdde3ffda4 2025-10-10T02:04:48.5874961Z deleted: sha256:4571fb4b2d2c700ea9adf9affdc419a0ae177bb9383a07181383b9d17371480e 2025-10-10T02:04:48.5875579Z deleted: sha256:d70dcb9b01054c38ce00200391740bcebd3a1cd72afe53327b81536008898606 2025-10-10T02:04:48.5876425Z deleted: sha256:8b8e5f7b43e9465220aafb7425efb40f3bfc014cc63b0364f673a990df3aa535 2025-10-10T02:04:48.5876914Z deleted: sha256:7d48d66ee4d478c78c3a6adadf587cf95bd748322ce32fff444f5d33ef2e0684 2025-10-10T02:04:48.5877370Z deleted: sha256:e4cd8739685cdd499ba08e895a29d365b8cb699f94cbb6b67e4e3619050a6dcc 2025-10-10T02:04:48.5877814Z deleted: sha256:559b9e576cde4002024cdd33e43847270549d69ec51a9446714cc1116826c5c4 2025-10-10T02:04:48.5878251Z deleted: sha256:d35159fdcc78c1037a6857000f6d1f7369c8dbe2b4e525aa0d66ee6ab394a829 2025-10-10T02:04:48.5878804Z deleted: sha256:1c1e45032bdc2f01b125fa0145c205376fce6aa2f8c2fbe5d7d286d8eb8efde6 2025-10-10T02:04:48.5879245Z deleted: sha256:e731b5134da5b9c47ce1696070037429ca15f364df4814a33bed968d293675ad 2025-10-10T02:04:48.5879674Z deleted: sha256:76e8028e50dc41cd6016284c7f7ab521b5b26a1f040406e7acfdd45d783aec82 2025-10-10T02:04:48.5880127Z deleted: sha256:da7adf94e8e4e4a5f8d1743254a6f6940542d2be33bf27fbf5390e8327ff9ec8 2025-10-10T02:04:48.5880580Z deleted: sha256:9f739ff4c0b7b0b7e859d014c1429bf8aeae2fbc74c4f95a729db4e18b6016f1 2025-10-10T02:04:48.5881016Z deleted: sha256:da101355a6a1035e5d8e120d6fcf20d6134887b145b5ba47cefc89942896b034 2025-10-10T02:04:48.5881458Z deleted: sha256:ec55a4fb03fa5b68218c181b9f8b5b48718b3b7a9f2688bc2a2ddebbc7e702d2 2025-10-10T02:04:48.5881896Z deleted: sha256:fc582e3da7f990556f9a46cc65e276a17e7b968253b500622a9fc411601644c1 2025-10-10T02:04:48.5882329Z deleted: sha256:3877d9b3300249f1fc15fa5be7f75e7368e3c302dfe721ad3ca5fb18d64c638f 2025-10-10T02:04:48.5882885Z deleted: sha256:e135fcbb872ffee1b10ef2430436bcdc6070f80bada4f6f29df4ff43ce22c5bf 2025-10-10T02:04:48.5883317Z deleted: sha256:b43b13192327ca33b41058e4ec301e99afe2545681946ef769a7a61cdcfbeca2 2025-10-10T02:04:48.5883753Z deleted: sha256:62563a12c0c16c746a7463f31b20031f14af18e2b07de16bed2dd6ae497cdda8 2025-10-10T02:04:48.5884192Z deleted: sha256:669aeb2091c6f16bf25899dca441b227b2f3f71c81b4f866752eea98afe33da0 2025-10-10T02:04:48.5884636Z deleted: sha256:662ed9daf3bbd7c75e1c6bbb332d7c26cbde354c93a15878b7e02beb0f34c678 2025-10-10T02:04:48.5885084Z deleted: sha256:6fde4e64ae77bf05993c494e6298fa4110b08631d69aab7a4772a15a8a442548 2025-10-10T02:04:48.5885516Z deleted: sha256:d9ead953554539f20688ecd6386499b2d82362b1b1c6b25021d0d7a69eec50e6 2025-10-10T02:04:48.5885943Z deleted: sha256:432a1c437c100c04fa7d6a8f39066ab56dde5a4861b64594b9c118d8b7b0675f 2025-10-10T02:04:48.5886378Z deleted: sha256:c30afe8220d7a2284c68b6064ac32cb056bfb4e4d40e3107107ce4599b5fec2e 2025-10-10T02:04:48.5886819Z deleted: sha256:88e9ee3ca3e2f5eec2be33087de509650ec96fdb43809c7230e0bb08750d7add 2025-10-10T02:04:48.5887263Z deleted: sha256:ca8d91823917321f63efd16926ca25c471fbd891b76c7b96783bb21013847838 2025-10-10T02:04:48.5887693Z deleted: sha256:e5757fc7a2da09412211d63b4fc54ce031023eae16fd5107165b5dd74ec3c6e7 2025-10-10T02:04:48.5888122Z deleted: sha256:908f42a23148aab74de493ebd7e97eea7865858e220b8aaa349f5d890479aa6e 2025-10-10T02:04:48.5888659Z deleted: sha256:c1e16b697497a8789e43bf5454ce8bdc219a4f1a0b4c684bb83098848349ee8c 2025-10-10T02:04:48.5889084Z deleted: sha256:6a220b588e3aa2555918233ab65a6abd91cb6764d772471c10c1c5d79f13cb72 2025-10-10T02:04:48.5889525Z deleted: sha256:ed65823401bbea17ab5c2b8ecc4e84c3d047fdf86c4044acd37cfe917849e506 2025-10-10T02:04:48.5889966Z deleted: sha256:a5b1e92804e52197e5e17c1f13907e802b3957d5f99011689965c433d58821f9 2025-10-10T02:04:48.5890392Z deleted: sha256:253e1d7fa0e015cdf247416952711f482e14604902a5d2b8cce9993a2a3b3d24 2025-10-10T02:04:48.5890819Z deleted: sha256:ca5ad77fb21d67464112b110aef81544ac5d43b2309cd12d61bed9bbfdc42e15 2025-10-10T02:04:48.5891254Z deleted: sha256:e94ee3c9b54243395ef87c3a1eec31ed7c6a31d37673b2d322816e08dbd1a118 2025-10-10T02:04:48.5891691Z deleted: sha256:89315ca20f065b6c143f279a913c3619669070f03be2bb6b2b03caef18998874 2025-10-10T02:04:48.5892132Z deleted: sha256:9aeae45b7bab4d904ee27a4e663b08e16e1be8e8a1c28c11328e8d666b482fe9 2025-10-10T02:04:48.5892566Z deleted: sha256:e52815a14f0f75252efae6743f2fce3814b200777535d74ab664e555e18886ff 2025-10-10T02:04:48.5893095Z deleted: sha256:e9f6ee51325b8f72704d3927dc62ccf70c72973e5c861b586404da7556ec4529 2025-10-10T02:04:48.5893540Z deleted: sha256:14132363c1ac9a5b1ab218a98d9209b5b5d523214bccbe6a9dae201d834208e8 2025-10-10T02:04:48.5893987Z deleted: sha256:4692931cca4fcefe71348bf3cffb47a7ceededdd131123f163f831d4bae1eac1 2025-10-10T02:04:48.5894422Z deleted: sha256:7ce9ee103794300058a597fc76e03268edc096ea2f1593770afaa73016f9f0e8 2025-10-10T02:04:48.5894849Z deleted: sha256:fb39599ef882320a257b5924b0a0ac99d8c62c3396cfe4508ff3d37556427421 2025-10-10T02:04:48.5895272Z deleted: sha256:8c676994914d959dc5ef16fe1d754c180a79c310513ebab96484cb97c28d99ad 2025-10-10T02:04:48.5895701Z deleted: sha256:d8d663334001dfecc055ccb2048c53a115763719fb55aaac728c5a5ec52e9354 2025-10-10T02:04:48.5896142Z deleted: sha256:cb9aec32a0cd5a45d9dae9fc89636bd5d1b7235a09352e85d88ed801a8c1a7bb 2025-10-10T02:04:48.5896584Z deleted: sha256:1bf54ca08eab0479e04daeb4450cd5359bda2139918d55576dbc435ef789c461 2025-10-10T02:04:48.5897021Z deleted: sha256:5d71c10e7be7172bf9b45248d8507e82c3119364e52c950940839da3984c4d5b 2025-10-10T02:04:48.5897436Z deleted: sha256:a9176111bd8471aa8d2d4f7d409a1547c59b74024372144b537a5f718350f74a 2025-10-10T02:04:48.5897878Z deleted: sha256:0946a9532ee6742ae519c31a401f2a013a94ddfe1bddcf6bff363a0c4910fe91 2025-10-10T02:04:48.5898321Z deleted: sha256:8e6353dc1ffcc20dda69135631eaa1aad42e0038e8db456b7d9d3eea4f111d75 2025-10-10T02:04:48.5898758Z deleted: sha256:b1b4a94efc01eaf3460de231b1685c299a122913e4e5df86b8b9d2a23d831437 2025-10-10T02:04:48.5899185Z deleted: sha256:2259025413435c0ff1a9a070523ae61e62c78530989b2c3330f0d23704a137b3 2025-10-10T02:04:48.5899651Z deleted: sha256:2acfa88bcefc179cc04eb4c414d36a47f7ec8444c41777613bbb3c0f2ccdaa5d 2025-10-10T02:04:48.5900102Z deleted: sha256:f98daf1fa3b1d8bcbef8cd4cea137f6d08fab572ecc5906c981e0dfe1dd76d42 2025-10-10T02:04:48.5900549Z deleted: sha256:3ffcf927608899854a2da49be3d0e41d77dd8d580b2b834338811fe9ad80fdc0 2025-10-10T02:04:48.5900984Z deleted: sha256:2059085fe5ac39f2b0398089556ef08128c7f87c71d9c6cc7940bf1e87de4b59 2025-10-10T02:04:48.5901415Z deleted: sha256:9ef72a5b6d2dfa7e843ab3735e33120bfcd1a9eaf5c602aebff0f9c4eb0984a8 2025-10-10T02:04:48.5901853Z deleted: sha256:90d3b2693815ff0e9a39f0733433a3043f62368ad99f513bc78adbaca85ffade 2025-10-10T02:04:48.5902299Z deleted: sha256:4bbbd89c196ee1904d74b95160dcbff15155c7824a4673698e1bf3c72b20873e 2025-10-10T02:04:48.5902735Z deleted: sha256:1a6b88de508df43a76a0243f7365abef52e51fa52b6e85a21cbb4d9b4844afe8 2025-10-10T02:04:48.5903167Z deleted: sha256:3080b6827f691d38c1ccada1ad766a71d4e605694b24f2a6264ac696b06d3952 2025-10-10T02:04:48.5903596Z deleted: sha256:c38bce02ca739cd758f8a6123addd2b5e45469761e91ac7f4391de364e02e87b 2025-10-10T02:04:48.5904028Z deleted: sha256:638cb10820f55ac2920396c01be87e60717f24fb04b4575eeaaf9e3da18169a1 2025-10-10T02:04:48.5904462Z deleted: sha256:fad077106111685ce7c6301a52c492f3533f691f213b2ad692821c52899b7405 2025-10-10T02:04:48.5904932Z deleted: sha256:6dbc7a719203586dc5dca848bff9ec50c54bb5478984eb9f4a1fa058e382da65 2025-10-10T02:04:48.5905373Z deleted: sha256:a33f103e4f676b2b0ef58e9428818918162447fb2bf885b60b0d1a2dd0814845 2025-10-10T02:04:48.5905804Z deleted: sha256:c7acd74e94da39045757348e6edb610ae80743b66fefb53fc5ba40746b6d620b 2025-10-10T02:04:48.5906234Z deleted: sha256:fc85ea9d45a7e990c9b3b2305376bd4b869b126167ff95a211eaef592b9595c5 2025-10-10T02:04:48.5906668Z deleted: sha256:01baeac660bbb677ba72b007e65d679ff1183c683767ef2ffb137a2c52804186 2025-10-10T02:04:48.5907102Z deleted: sha256:5af13118677cad47c318ab0e651bbbbe5745cb06de7e1edc1609bc8e64fc0b34 2025-10-10T02:04:48.5907538Z deleted: sha256:45a23f76e9d32ccc4cf19e7249e666bbff02ac5de3ac3e2d4da8d757d1fb2b91 2025-10-10T02:04:48.5907973Z deleted: sha256:767e56ba346ae714b6e6b816baa839051145ed78cfa0e4524a86cc287b0c4b00 2025-10-10T02:04:48.5908337Z untagged: public.ecr.aws/docker/library/python:3.13 2025-10-10T02:04:48.5908826Z untagged: public.ecr.aws/docker/library/python@sha256:4889af0e45f04b7c5dd741421a1280919499d38d3125d714b69fa86b23b1052a 2025-10-10T02:04:48.5909456Z deleted: sha256:6c82e3449d7794702180419555c0a0e1687ea79a0c665b250436286924681a55 2025-10-10T02:04:48.5909875Z deleted: sha256:68a0419cb3069ed43905ab41b911f2b7248601df62c854ae65e8c8a0342dbb30 2025-10-10T02:04:48.5910311Z deleted: sha256:b258354078ead7184c2f6d72eb3d5db1855162c0f80d164c09e794b21f30f48b 2025-10-10T02:04:48.5910751Z deleted: sha256:3e65e3c281dedcfdb54cb848bd29efd0e832cf5f29dec4b6b9849cd7420266cb 2025-10-10T02:04:48.5911192Z deleted: sha256:42f4cd5b256627f333ad4537462aac85c359e741da4f02d1cb68600c128841c5 2025-10-10T02:04:48.5911617Z deleted: sha256:4e7df8e345c749980c75fd48e7b2ef15e63dc912b467ffa446284f0dbcc5aa33 2025-10-10T02:04:48.5912046Z deleted: sha256:345f9c4d6fe93d61688b6f1a607137261d7983d3788b5d88e8791b6ebeb8a920 2025-10-10T02:04:48.5912476Z deleted: sha256:a5ec5ec9d16c5551ce8889cbc03af0609b92cf8a8d60b32e72a7eabb8378eaec 2025-10-10T02:04:48.5912737Z 2025-10-10T02:04:48.5912825Z Total reclaimed space: 43GB 2025-10-10T02:04:48.6003978Z Post job cleanup. 2025-10-10T02:04:48.6048753Z Post job cleanup. 2025-10-10T02:04:48.6963444Z [command]/usr/bin/git version 2025-10-10T02:04:48.7016071Z git version 2.50.1 2025-10-10T02:04:48.7056964Z Copying '/home/ec2-user/.gitconfig' to '/home/ec2-user/actions-runner/_work/_temp/991b24f6-80f4-41e6-b92a-9c3a99c9b79f/.gitconfig' 2025-10-10T02:04:48.7076451Z Temporarily overriding HOME='/home/ec2-user/actions-runner/_work/_temp/991b24f6-80f4-41e6-b92a-9c3a99c9b79f' before making global git config changes 2025-10-10T02:04:48.7079923Z Adding repository directory to the temporary git global config as a safe directory 2025-10-10T02:04:48.7087515Z [command]/usr/bin/git config --global --add safe.directory /home/ec2-user/actions-runner/_work/pytorch/pytorch 2025-10-10T02:04:48.7137900Z [command]/usr/bin/git config --local --name-only --get-regexp core\.sshCommand 2025-10-10T02:04:48.7200926Z [command]/usr/bin/git submodule foreach --recursive sh -c "git config --local --name-only --get-regexp 'core\.sshCommand' && git config --local --unset-all 'core.sshCommand' || :" 2025-10-10T02:04:48.7690516Z Entering 'android/libs/fbjni' 2025-10-10T02:04:48.7784127Z Entering 'third_party/FP16' 2025-10-10T02:04:48.7874210Z Entering 'third_party/FXdiv' 2025-10-10T02:04:48.7969953Z Entering 'third_party/NNPACK' 2025-10-10T02:04:48.8057306Z Entering 'third_party/NVTX' 2025-10-10T02:04:48.8147082Z Entering 'third_party/VulkanMemoryAllocator' 2025-10-10T02:04:48.8235334Z Entering 'third_party/XNNPACK' 2025-10-10T02:04:48.8333936Z Entering 'third_party/aiter' 2025-10-10T02:04:48.8421982Z Entering 'third_party/aiter/3rdparty/composable_kernel' 2025-10-10T02:04:48.8518700Z Entering 'third_party/benchmark' 2025-10-10T02:04:48.8605315Z Entering 'third_party/composable_kernel' 2025-10-10T02:04:48.8699858Z Entering 'third_party/cpp-httplib' 2025-10-10T02:04:48.8785308Z Entering 'third_party/cpuinfo' 2025-10-10T02:04:48.8875571Z Entering 'third_party/cudnn_frontend' 2025-10-10T02:04:48.8972184Z Entering 'third_party/cutlass' 2025-10-10T02:04:48.9072007Z Entering 'third_party/fbgemm' 2025-10-10T02:04:48.9160995Z Entering 'third_party/fbgemm/external/asmjit' 2025-10-10T02:04:48.9256630Z Entering 'third_party/fbgemm/external/composable_kernel' 2025-10-10T02:04:48.9364059Z Entering 'third_party/fbgemm/external/cpuinfo' 2025-10-10T02:04:48.9445095Z Entering 'third_party/fbgemm/external/cutlass' 2025-10-10T02:04:48.9543213Z Entering 'third_party/fbgemm/external/googletest' 2025-10-10T02:04:48.9629762Z Entering 'third_party/fbgemm/external/hipify_torch' 2025-10-10T02:04:48.9709114Z Entering 'third_party/fbgemm/external/json' 2025-10-10T02:04:48.9796535Z Entering 'third_party/flash-attention' 2025-10-10T02:04:48.9890534Z Entering 'third_party/flash-attention/csrc/composable_kernel' 2025-10-10T02:04:48.9987646Z Entering 'third_party/flash-attention/csrc/cutlass' 2025-10-10T02:04:49.0087282Z Entering 'third_party/flatbuffers' 2025-10-10T02:04:49.0175539Z Entering 'third_party/fmt' 2025-10-10T02:04:49.0271692Z Entering 'third_party/gemmlowp/gemmlowp' 2025-10-10T02:04:49.0364529Z Entering 'third_party/gloo' 2025-10-10T02:04:49.0454715Z Entering 'third_party/googletest' 2025-10-10T02:04:49.0549151Z Entering 'third_party/ideep' 2025-10-10T02:04:49.0641504Z Entering 'third_party/ideep/mkl-dnn' 2025-10-10T02:04:49.0743252Z Entering 'third_party/ittapi' 2025-10-10T02:04:49.0837333Z Entering 'third_party/kineto' 2025-10-10T02:04:49.0916951Z Entering 'third_party/kineto/libkineto/third_party/dynolog' 2025-10-10T02:04:49.0996890Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/DCGM' 2025-10-10T02:04:49.1088212Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/cpr' 2025-10-10T02:04:49.1173651Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/fmt' 2025-10-10T02:04:49.1263327Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/gflags' 2025-10-10T02:04:49.1351093Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/gflags/doc' 2025-10-10T02:04:49.1446112Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/glog' 2025-10-10T02:04:49.1529978Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/googletest' 2025-10-10T02:04:49.1621991Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/json' 2025-10-10T02:04:49.1703551Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/pfs' 2025-10-10T02:04:49.1805013Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/prometheus-cpp' 2025-10-10T02:04:49.1888558Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/prometheus-cpp/3rdparty/civetweb' 2025-10-10T02:04:49.1977707Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/prometheus-cpp/3rdparty/googletest' 2025-10-10T02:04:49.2086560Z Entering 'third_party/kineto/libkineto/third_party/fmt' 2025-10-10T02:04:49.2181011Z Entering 'third_party/kineto/libkineto/third_party/googletest' 2025-10-10T02:04:49.2264602Z Entering 'third_party/kleidiai' 2025-10-10T02:04:49.2363448Z Entering 'third_party/mimalloc' 2025-10-10T02:04:49.2454099Z Entering 'third_party/nlohmann' 2025-10-10T02:04:49.2548938Z Entering 'third_party/onnx' 2025-10-10T02:04:49.2657523Z Entering 'third_party/onnx/third_party/pybind11' 2025-10-10T02:04:49.2748443Z Entering 'third_party/opentelemetry-cpp' 2025-10-10T02:04:49.2840872Z Entering 'third_party/opentelemetry-cpp/third_party/benchmark' 2025-10-10T02:04:49.2930009Z Entering 'third_party/opentelemetry-cpp/third_party/googletest' 2025-10-10T02:04:49.3014873Z Entering 'third_party/opentelemetry-cpp/third_party/ms-gsl' 2025-10-10T02:04:49.3106060Z Entering 'third_party/opentelemetry-cpp/third_party/nlohmann-json' 2025-10-10T02:04:49.3195629Z Entering 'third_party/opentelemetry-cpp/third_party/opentelemetry-proto' 2025-10-10T02:04:49.3282942Z Entering 'third_party/opentelemetry-cpp/third_party/opentracing-cpp' 2025-10-10T02:04:49.3372881Z Entering 'third_party/opentelemetry-cpp/third_party/prometheus-cpp' 2025-10-10T02:04:49.3460079Z Entering 'third_party/opentelemetry-cpp/third_party/prometheus-cpp/3rdparty/civetweb' 2025-10-10T02:04:49.3545275Z Entering 'third_party/opentelemetry-cpp/third_party/prometheus-cpp/3rdparty/googletest' 2025-10-10T02:04:49.3638303Z Entering 'third_party/opentelemetry-cpp/tools/vcpkg' 2025-10-10T02:04:49.3747182Z Entering 'third_party/pocketfft' 2025-10-10T02:04:49.3842394Z Entering 'third_party/protobuf' 2025-10-10T02:04:49.3935058Z Entering 'third_party/protobuf/third_party/benchmark' 2025-10-10T02:04:49.4020714Z Entering 'third_party/protobuf/third_party/googletest' 2025-10-10T02:04:49.4110509Z Entering 'third_party/psimd' 2025-10-10T02:04:49.4199089Z Entering 'third_party/pthreadpool' 2025-10-10T02:04:49.4281321Z Entering 'third_party/pybind11' 2025-10-10T02:04:49.4373658Z Entering 'third_party/python-peachpy' 2025-10-10T02:04:49.4461437Z Entering 'third_party/sleef' 2025-10-10T02:04:49.4553233Z Entering 'third_party/tensorpipe' 2025-10-10T02:04:49.4638787Z Entering 'third_party/tensorpipe/third_party/googletest' 2025-10-10T02:04:49.4720562Z Entering 'third_party/tensorpipe/third_party/libnop' 2025-10-10T02:04:49.4801610Z Entering 'third_party/tensorpipe/third_party/libuv' 2025-10-10T02:04:49.4880690Z Entering 'third_party/tensorpipe/third_party/pybind11' 2025-10-10T02:04:49.4960457Z Entering 'third_party/tensorpipe/third_party/pybind11/tools/clang' 2025-10-10T02:04:49.5085172Z [command]/usr/bin/git config --local --name-only --get-regexp http\.https\:\/\/github\.com\/\.extraheader 2025-10-10T02:04:49.5112115Z http.https://github.com/.extraheader 2025-10-10T02:04:49.5127094Z [command]/usr/bin/git config --local --unset-all http.https://github.com/.extraheader 2025-10-10T02:04:49.5163316Z [command]/usr/bin/git submodule foreach --recursive sh -c "git config --local --name-only --get-regexp 'http\.https\:\/\/github\.com\/\.extraheader' && git config --local --unset-all 'http.https://github.com/.extraheader' || :" 2025-10-10T02:04:49.5633750Z Entering 'android/libs/fbjni' 2025-10-10T02:04:49.5690715Z http.https://github.com/.extraheader 2025-10-10T02:04:49.5745549Z Entering 'third_party/FP16' 2025-10-10T02:04:49.5807438Z http.https://github.com/.extraheader 2025-10-10T02:04:49.5861696Z Entering 'third_party/FXdiv' 2025-10-10T02:04:49.5927085Z http.https://github.com/.extraheader 2025-10-10T02:04:49.5983627Z Entering 'third_party/NNPACK' 2025-10-10T02:04:49.6037327Z http.https://github.com/.extraheader 2025-10-10T02:04:49.6095459Z Entering 'third_party/NVTX' 2025-10-10T02:04:49.6157139Z http.https://github.com/.extraheader 2025-10-10T02:04:49.6215574Z Entering 'third_party/VulkanMemoryAllocator' 2025-10-10T02:04:49.6269141Z http.https://github.com/.extraheader 2025-10-10T02:04:49.6325146Z Entering 'third_party/XNNPACK' 2025-10-10T02:04:49.6379726Z http.https://github.com/.extraheader 2025-10-10T02:04:49.6451898Z Entering 'third_party/aiter' 2025-10-10T02:04:49.6510255Z http.https://github.com/.extraheader 2025-10-10T02:04:49.6571741Z Entering 'third_party/aiter/3rdparty/composable_kernel' 2025-10-10T02:04:49.6628993Z http.https://github.com/.extraheader 2025-10-10T02:04:49.6700441Z Entering 'third_party/benchmark' 2025-10-10T02:04:49.6755960Z http.https://github.com/.extraheader 2025-10-10T02:04:49.6811952Z Entering 'third_party/composable_kernel' 2025-10-10T02:04:49.6868458Z http.https://github.com/.extraheader 2025-10-10T02:04:49.6944021Z Entering 'third_party/cpp-httplib' 2025-10-10T02:04:49.6999478Z http.https://github.com/.extraheader 2025-10-10T02:04:49.7057966Z Entering 'third_party/cpuinfo' 2025-10-10T02:04:49.7115101Z http.https://github.com/.extraheader 2025-10-10T02:04:49.7172178Z Entering 'third_party/cudnn_frontend' 2025-10-10T02:04:49.7229987Z http.https://github.com/.extraheader 2025-10-10T02:04:49.7289402Z Entering 'third_party/cutlass' 2025-10-10T02:04:49.7341569Z http.https://github.com/.extraheader 2025-10-10T02:04:49.7406157Z Entering 'third_party/fbgemm' 2025-10-10T02:04:49.7460774Z http.https://github.com/.extraheader 2025-10-10T02:04:49.7525327Z Entering 'third_party/fbgemm/external/asmjit' 2025-10-10T02:04:49.7577965Z http.https://github.com/.extraheader 2025-10-10T02:04:49.7642690Z Entering 'third_party/fbgemm/external/composable_kernel' 2025-10-10T02:04:49.7699781Z http.https://github.com/.extraheader 2025-10-10T02:04:49.7771559Z Entering 'third_party/fbgemm/external/cpuinfo' 2025-10-10T02:04:49.7828367Z http.https://github.com/.extraheader 2025-10-10T02:04:49.7889814Z Entering 'third_party/fbgemm/external/cutlass' 2025-10-10T02:04:49.7940042Z http.https://github.com/.extraheader 2025-10-10T02:04:49.8005918Z Entering 'third_party/fbgemm/external/googletest' 2025-10-10T02:04:49.8057704Z http.https://github.com/.extraheader 2025-10-10T02:04:49.8116850Z Entering 'third_party/fbgemm/external/hipify_torch' 2025-10-10T02:04:49.8176623Z http.https://github.com/.extraheader 2025-10-10T02:04:49.8235494Z Entering 'third_party/fbgemm/external/json' 2025-10-10T02:04:49.8290878Z http.https://github.com/.extraheader 2025-10-10T02:04:49.8356261Z Entering 'third_party/flash-attention' 2025-10-10T02:04:49.8408405Z http.https://github.com/.extraheader 2025-10-10T02:04:49.8470925Z Entering 'third_party/flash-attention/csrc/composable_kernel' 2025-10-10T02:04:49.8529987Z http.https://github.com/.extraheader 2025-10-10T02:04:49.8588224Z Entering 'third_party/flash-attention/csrc/cutlass' 2025-10-10T02:04:49.8640543Z http.https://github.com/.extraheader 2025-10-10T02:04:49.8706798Z Entering 'third_party/flatbuffers' 2025-10-10T02:04:49.8767761Z http.https://github.com/.extraheader 2025-10-10T02:04:49.8833966Z Entering 'third_party/fmt' 2025-10-10T02:04:49.8890212Z http.https://github.com/.extraheader 2025-10-10T02:04:49.8951510Z Entering 'third_party/gemmlowp/gemmlowp' 2025-10-10T02:04:49.9009612Z http.https://github.com/.extraheader 2025-10-10T02:04:49.9072274Z Entering 'third_party/gloo' 2025-10-10T02:04:49.9129516Z http.https://github.com/.extraheader 2025-10-10T02:04:49.9188078Z Entering 'third_party/googletest' 2025-10-10T02:04:49.9240273Z http.https://github.com/.extraheader 2025-10-10T02:04:49.9301931Z Entering 'third_party/ideep' 2025-10-10T02:04:49.9359238Z http.https://github.com/.extraheader 2025-10-10T02:04:49.9414979Z Entering 'third_party/ideep/mkl-dnn' 2025-10-10T02:04:49.9469192Z http.https://github.com/.extraheader 2025-10-10T02:04:49.9544133Z Entering 'third_party/ittapi' 2025-10-10T02:04:49.9599792Z http.https://github.com/.extraheader 2025-10-10T02:04:49.9654643Z Entering 'third_party/kineto' 2025-10-10T02:04:49.9717406Z http.https://github.com/.extraheader 2025-10-10T02:04:49.9772293Z Entering 'third_party/kineto/libkineto/third_party/dynolog' 2025-10-10T02:04:49.9830591Z http.https://github.com/.extraheader 2025-10-10T02:04:49.9888039Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/DCGM' 2025-10-10T02:04:49.9940658Z http.https://github.com/.extraheader 2025-10-10T02:04:49.9997362Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/cpr' 2025-10-10T02:04:50.0059599Z http.https://github.com/.extraheader 2025-10-10T02:04:50.0113474Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/fmt' 2025-10-10T02:04:50.0170609Z http.https://github.com/.extraheader 2025-10-10T02:04:50.0233883Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/gflags' 2025-10-10T02:04:50.0290049Z http.https://github.com/.extraheader 2025-10-10T02:04:50.0348919Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/gflags/doc' 2025-10-10T02:04:50.0409272Z http.https://github.com/.extraheader 2025-10-10T02:04:50.0479591Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/glog' 2025-10-10T02:04:50.0538778Z http.https://github.com/.extraheader 2025-10-10T02:04:50.0599223Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/googletest' 2025-10-10T02:04:50.0650590Z http.https://github.com/.extraheader 2025-10-10T02:04:50.0707446Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/json' 2025-10-10T02:04:50.0767888Z http.https://github.com/.extraheader 2025-10-10T02:04:50.0828228Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/pfs' 2025-10-10T02:04:50.0888009Z http.https://github.com/.extraheader 2025-10-10T02:04:50.0947278Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/prometheus-cpp' 2025-10-10T02:04:50.1007475Z http.https://github.com/.extraheader 2025-10-10T02:04:50.1063433Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/prometheus-cpp/3rdparty/civetweb' 2025-10-10T02:04:50.1120793Z http.https://github.com/.extraheader 2025-10-10T02:04:50.1186378Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/prometheus-cpp/3rdparty/googletest' 2025-10-10T02:04:50.1242342Z http.https://github.com/.extraheader 2025-10-10T02:04:50.1311546Z Entering 'third_party/kineto/libkineto/third_party/fmt' 2025-10-10T02:04:50.1367031Z http.https://github.com/.extraheader 2025-10-10T02:04:50.1427824Z Entering 'third_party/kineto/libkineto/third_party/googletest' 2025-10-10T02:04:50.1479504Z http.https://github.com/.extraheader 2025-10-10T02:04:50.1544052Z Entering 'third_party/kleidiai' 2025-10-10T02:04:50.1600456Z http.https://github.com/.extraheader 2025-10-10T02:04:50.1660047Z Entering 'third_party/mimalloc' 2025-10-10T02:04:50.1716540Z http.https://github.com/.extraheader 2025-10-10T02:04:50.1770261Z Entering 'third_party/nlohmann' 2025-10-10T02:04:50.1828612Z http.https://github.com/.extraheader 2025-10-10T02:04:50.1892468Z Entering 'third_party/onnx' 2025-10-10T02:04:50.1949931Z http.https://github.com/.extraheader 2025-10-10T02:04:50.2020942Z Entering 'third_party/onnx/third_party/pybind11' 2025-10-10T02:04:50.2077857Z http.https://github.com/.extraheader 2025-10-10T02:04:50.2138756Z Entering 'third_party/opentelemetry-cpp' 2025-10-10T02:04:50.2198617Z http.https://github.com/.extraheader 2025-10-10T02:04:50.2252846Z Entering 'third_party/opentelemetry-cpp/third_party/benchmark' 2025-10-10T02:04:50.2316492Z http.https://github.com/.extraheader 2025-10-10T02:04:50.2375757Z Entering 'third_party/opentelemetry-cpp/third_party/googletest' 2025-10-10T02:04:50.2428029Z http.https://github.com/.extraheader 2025-10-10T02:04:50.2487526Z Entering 'third_party/opentelemetry-cpp/third_party/ms-gsl' 2025-10-10T02:04:50.2541657Z http.https://github.com/.extraheader 2025-10-10T02:04:50.2601711Z Entering 'third_party/opentelemetry-cpp/third_party/nlohmann-json' 2025-10-10T02:04:50.2658712Z http.https://github.com/.extraheader 2025-10-10T02:04:50.2725140Z Entering 'third_party/opentelemetry-cpp/third_party/opentelemetry-proto' 2025-10-10T02:04:50.2778215Z http.https://github.com/.extraheader 2025-10-10T02:04:50.2834745Z Entering 'third_party/opentelemetry-cpp/third_party/opentracing-cpp' 2025-10-10T02:04:50.2889947Z http.https://github.com/.extraheader 2025-10-10T02:04:50.2947631Z Entering 'third_party/opentelemetry-cpp/third_party/prometheus-cpp' 2025-10-10T02:04:50.3006705Z http.https://github.com/.extraheader 2025-10-10T02:04:50.3060473Z Entering 'third_party/opentelemetry-cpp/third_party/prometheus-cpp/3rdparty/civetweb' 2025-10-10T02:04:50.3116520Z http.https://github.com/.extraheader 2025-10-10T02:04:50.3179574Z Entering 'third_party/opentelemetry-cpp/third_party/prometheus-cpp/3rdparty/googletest' 2025-10-10T02:04:50.3229145Z http.https://github.com/.extraheader 2025-10-10T02:04:50.3292704Z Entering 'third_party/opentelemetry-cpp/tools/vcpkg' 2025-10-10T02:04:50.3350640Z http.https://github.com/.extraheader 2025-10-10T02:04:50.3434240Z Entering 'third_party/pocketfft' 2025-10-10T02:04:50.3490546Z http.https://github.com/.extraheader 2025-10-10T02:04:50.3554414Z Entering 'third_party/protobuf' 2025-10-10T02:04:50.3616985Z http.https://github.com/.extraheader 2025-10-10T02:04:50.3675184Z Entering 'third_party/protobuf/third_party/benchmark' 2025-10-10T02:04:50.3730727Z http.https://github.com/.extraheader 2025-10-10T02:04:50.3792246Z Entering 'third_party/protobuf/third_party/googletest' 2025-10-10T02:04:50.3849921Z http.https://github.com/.extraheader 2025-10-10T02:04:50.3913762Z Entering 'third_party/psimd' 2025-10-10T02:04:50.3978780Z http.https://github.com/.extraheader 2025-10-10T02:04:50.4038436Z Entering 'third_party/pthreadpool' 2025-10-10T02:04:50.4097415Z http.https://github.com/.extraheader 2025-10-10T02:04:50.4155709Z Entering 'third_party/pybind11' 2025-10-10T02:04:50.4216842Z http.https://github.com/.extraheader 2025-10-10T02:04:50.4287160Z Entering 'third_party/python-peachpy' 2025-10-10T02:04:50.4348225Z http.https://github.com/.extraheader 2025-10-10T02:04:50.4406467Z Entering 'third_party/sleef' 2025-10-10T02:04:50.4464208Z http.https://github.com/.extraheader 2025-10-10T02:04:50.4526735Z Entering 'third_party/tensorpipe' 2025-10-10T02:04:50.4581712Z http.https://github.com/.extraheader 2025-10-10T02:04:50.4633876Z Entering 'third_party/tensorpipe/third_party/googletest' 2025-10-10T02:04:50.4689805Z http.https://github.com/.extraheader 2025-10-10T02:04:50.4752335Z Entering 'third_party/tensorpipe/third_party/libnop' 2025-10-10T02:04:50.4809723Z http.https://github.com/.extraheader 2025-10-10T02:04:50.4871640Z Entering 'third_party/tensorpipe/third_party/libuv' 2025-10-10T02:04:50.4929939Z http.https://github.com/.extraheader 2025-10-10T02:04:50.4988993Z Entering 'third_party/tensorpipe/third_party/pybind11' 2025-10-10T02:04:50.5047201Z http.https://github.com/.extraheader 2025-10-10T02:04:50.5108169Z Entering 'third_party/tensorpipe/third_party/pybind11/tools/clang' 2025-10-10T02:04:50.5165147Z http.https://github.com/.extraheader 2025-10-10T02:04:50.5351611Z A job completed hook has been configured by the self-hosted runner administrator 2025-10-10T02:04:50.5387823Z ##[group]Run '/home/ec2-user/runner-scripts/after_job.sh' 2025-10-10T02:04:50.5394707Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2025-10-10T02:04:50.5395048Z ##[endgroup] 2025-10-10T02:04:50.5523901Z [!ALERT!] Swap in detected! [!ALERT!] 2025-10-10T02:05:00.7105750Z [!ALERT!] Swap out detected [!ALERT!] 2025-10-10T02:05:18.1343248Z Cleaning up orphan processes