2025-11-03T16:56:52.0292638Z Current runner version: '2.329.0' 2025-11-03T16:56:52.0298236Z Runner name: 'i-0aa7ffe8f5edfd321' 2025-11-03T16:56:52.0298968Z Runner group name: 'default' 2025-11-03T16:56:52.0299757Z Machine name: 'ip-10-0-2-76' 2025-11-03T16:56:52.0302571Z ##[group]GITHUB_TOKEN Permissions 2025-11-03T16:56:52.0304707Z Contents: read 2025-11-03T16:56:52.0305557Z Metadata: read 2025-11-03T16:56:52.0306126Z ##[endgroup] 2025-11-03T16:56:52.0308005Z Secret source: Actions 2025-11-03T16:56:52.0308859Z Prepare workflow directory 2025-11-03T16:56:52.0776513Z Prepare all required actions 2025-11-03T16:56:52.0810195Z Getting action download info 2025-11-03T16:56:52.4272472Z Download action repository 'pytorch/test-infra@main' (SHA:62d0084d5e2a6926f8884d6714e7fe289eb96af5) 2025-11-03T16:56:54.9235787Z Download action repository 'pytorch/pytorch@main' (SHA:104b86861862ccc47f6df6528347a017bef813ad) 2025-11-03T16:57:11.5132465Z Download action repository 'actions/setup-python@a26af69be951a213d495a4c3e4e4022e16d87065' (SHA:a26af69be951a213d495a4c3e4e4022e16d87065) 2025-11-03T16:57:11.8543928Z Download action repository 'aws-actions/configure-aws-credentials@ececac1a45f3b08a01d2dd070d28d111c5fe6722' (SHA:ececac1a45f3b08a01d2dd070d28d111c5fe6722) 2025-11-03T16:57:12.0888195Z Download action repository 'aws-actions/amazon-ecr-login@062b18b96a7aff071d4dc91bc00c4c1a7945b076' (SHA:062b18b96a7aff071d4dc91bc00c4c1a7945b076) 2025-11-03T16:57:12.2985346Z Download action repository 'seemethere/download-artifact-s3@1da556a7aa0a088e3153970611f6c432d58e80e6' (SHA:1da556a7aa0a088e3153970611f6c432d58e80e6) 2025-11-03T16:57:12.5420630Z Download action repository 'seemethere/upload-artifact-s3@baba72d0712b404f646cebe0730933554ebce96a' (SHA:baba72d0712b404f646cebe0730933554ebce96a) 2025-11-03T16:57:12.7835504Z Getting action download info 2025-11-03T16:57:12.9798634Z Download action repository 'actions/checkout@v4' (SHA:08eba0b27e820071cde6df949e0beb9ba4906955) 2025-11-03T16:57:13.2288106Z Getting action download info 2025-11-03T16:57:13.3805526Z Download action repository 'nick-fields/retry@v3.0.0' (SHA:7152eba30c6575329ac0576536151aca5a72780e) 2025-11-03T16:57:13.5704309Z Getting action download info 2025-11-03T16:57:13.7041623Z Download action repository 'nick-fields/retry@3e91a01664abd3c5cd539100d10d33b9c5b68482' (SHA:3e91a01664abd3c5cd539100d10d33b9c5b68482) 2025-11-03T16:57:13.8854309Z Getting action download info 2025-11-03T16:57:14.0551922Z Uses: pytorch/pytorch/.github/workflows/_linux-test.yml@refs/heads/main (3f6538febd727b782e6e13cfd026a309fb14351d) 2025-11-03T16:57:14.0555512Z ##[group] Inputs 2025-11-03T16:57:14.0555855Z build-environment: linux-jammy-cuda12.8-py3.12-gcc11 2025-11-03T16:57:14.0560931Z test-matrix: {"include": [{"config": "vllm_basic_correctness_test", "shard": 1, "num_shards": 1, "runner": "linux.g6.4xlarge.experimental.nvidia.gpu"}, {"config": "vllm_basic_models_test", "shard": 1, "num_shards": 1, "runner": "linux.g6.4xlarge.experimental.nvidia.gpu"}, {"config": "vllm_entrypoints_test", "shard": 1, "num_shards": 1, "runner": "linux.g6.4xlarge.experimental.nvidia.gpu"}, {"config": "vllm_regression_test", "shard": 1, "num_shards": 1, "runner": "linux.g6.4xlarge.experimental.nvidia.gpu"}, {"config": "vllm_multi_model_processor_test", "shard": 1, "num_shards": 1, "runner": "linux.g6.4xlarge.experimental.nvidia.gpu"}, {"config": "vllm_pytorch_compilation_unit_tests", "shard": 1, "num_shards": 1, "runner": "linux.g6.4xlarge.experimental.nvidia.gpu"}, {"config": "vllm_lora_28_failure_test", "shard": 1, "num_shards": 1, "runner": "linux.g6.4xlarge.experimental.nvidia.gpu"}, {"config": "vllm_multi_model_test_28_failure_test", "shard": 1, "num_shards": 1, "runner": "linux.g6.4xlarge.experimental.nvidia.gpu"}, {"config": "vllm_language_model_test_extended_generation_28_failure_test", "shard": 1, "num_shards": 1, "runner": "linux.g6.4xlarge.experimental.nvidia.gpu"}, {"config": "vllm_distributed_test_2_gpu_28_failure_test", "shard": 1, "num_shards": 1, "runner": "linux.g6.4xlarge.experimental.nvidia.gpu"}, {"config": "vllm_lora_test", "shard": 0, "num_shards": 4, "runner": "linux.g6.4xlarge.experimental.nvidia.gpu"}, {"config": "vllm_lora_test", "shard": 1, "num_shards": 4, "runner": "linux.g6.4xlarge.experimental.nvidia.gpu"}, {"config": "vllm_lora_test", "shard": 2, "num_shards": 4, "runner": "linux.g6.4xlarge.experimental.nvidia.gpu"}, {"config": "vllm_lora_test", "shard": 3, "num_shards": 4, "runner": "linux.g6.4xlarge.experimental.nvidia.gpu"}, {"config": "vllm_lora_tp_test_distributed", "shard": 1, "num_shards": 1, "runner": "linux.g6.12xlarge.nvidia.gpu"}, {"config": "vllm_distributed_test_28_failure_test", "shard": 1, "num_shards": 1, "runner": "linux.g6.12xlarge.nvidia.gpu"}]} 2025-11-03T16:57:14.0566370Z docker-image: 308535385114.dkr.ecr.us-east-1.amazonaws.com/pytorch/ci-image:pytorch-linux-jammy-cuda12.8-cudnn9-py3.12-gcc11-vllm-30147fb242f88ab8f13e14b76ab186d189d8605c 2025-11-03T16:57:14.0567018Z sync-tag: 2025-11-03T16:57:14.0567698Z timeout-minutes: 240 2025-11-03T16:57:14.0567924Z use-gha: 2025-11-03T16:57:14.0568099Z dashboard-tag: 2025-11-03T16:57:14.0568287Z s3-bucket: gha-artifacts 2025-11-03T16:57:14.0568488Z aws-role-to-assume: 2025-11-03T16:57:14.0568982Z disable-monitor: false 2025-11-03T16:57:14.0569207Z monitor-log-interval: 5 2025-11-03T16:57:14.0569436Z monitor-data-collect-interval: 1 2025-11-03T16:57:14.0569670Z ##[endgroup] 2025-11-03T16:57:14.0570038Z Complete job name: ci-vllm-test / test (vllm_basic_models_test, 1, 1, linux.g6.4xlarge.experimental.nvidia.gpu) 2025-11-03T16:57:14.1205598Z A job started hook has been configured by the self-hosted runner administrator 2025-11-03T16:57:14.1302335Z ##[group]Run '/home/ec2-user/runner-scripts/before_job.sh' 2025-11-03T16:57:14.1312559Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2025-11-03T16:57:14.1313106Z ##[endgroup] 2025-11-03T16:57:15.4369266Z Runner Type: linux.g6.4xlarge.experimental.nvidia.gpu 2025-11-03T16:57:15.4369775Z Instance Type: g6.4xlarge 2025-11-03T16:57:15.4369984Z AMI Name: unknown 2025-11-03T16:57:15.4408899Z AMI ID: ami-08982f1c5bf93d976 2025-11-03T16:57:20.3358854Z ##[group]Run pytorch/test-infra/.github/actions/setup-ssh@main 2025-11-03T16:57:20.3359211Z with: 2025-11-03T16:57:20.3359720Z github-secret: *** 2025-11-03T16:57:20.3360258Z instructions: All testing is done inside the container, to start an interactive session run: docker exec -it $(docker container ps --format '{{.ID}}') bash 2025-11-03T16:57:20.3360808Z activate-with-label: false 2025-11-03T16:57:20.3361011Z label: with-ssh 2025-11-03T16:57:20.3361196Z remove-existing-keys: true 2025-11-03T16:57:20.3361400Z fail-silently: true 2025-11-03T16:57:20.3361621Z env: 2025-11-03T16:57:20.3361769Z GIT_DEFAULT_BRANCH: main 2025-11-03T16:57:20.3361958Z ##[endgroup] 2025-11-03T16:57:20.4689718Z Please see https://github.com/pytorch/pytorch/wiki/Debugging-using-with-ssh-for-Github-Actions for more info. 2025-11-03T16:57:20.4690804Z Not on pull request and ciflow reference could not be extracted, skipping adding ssh keys 2025-11-03T16:57:20.4853334Z ##[group]Run pytorch/pytorch/.github/actions/checkout-pytorch@main 2025-11-03T16:57:20.4853688Z with: 2025-11-03T16:57:20.4853856Z no-sudo: true 2025-11-03T16:57:20.4854038Z submodules: recursive 2025-11-03T16:57:20.4854249Z fetch-depth: 0 2025-11-03T16:57:20.4854410Z env: 2025-11-03T16:57:20.4854567Z GIT_DEFAULT_BRANCH: main 2025-11-03T16:57:20.4854764Z ##[endgroup] 2025-11-03T16:57:20.4929781Z ##[group]Run echo "IN_CONTAINER_RUNNER=$(if [ -f /.inarc ] || [ -f /.incontainer ]; then echo true ; else echo false; fi)" >> "$GITHUB_OUTPUT" 2025-11-03T16:57:20.4930495Z echo "IN_CONTAINER_RUNNER=$(if [ -f /.inarc ] || [ -f /.incontainer ]; then echo true ; else echo false; fi)" >> "$GITHUB_OUTPUT" 2025-11-03T16:57:20.4940623Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2025-11-03T16:57:20.4940925Z env: 2025-11-03T16:57:20.4941124Z GIT_DEFAULT_BRANCH: main 2025-11-03T16:57:20.4941346Z ##[endgroup] 2025-11-03T16:57:20.5040796Z ##[group]Run # Use all available CPUs for fetching 2025-11-03T16:57:20.5041317Z # Use all available CPUs for fetching 2025-11-03T16:57:20.5041579Z cd "${GITHUB_WORKSPACE}" 2025-11-03T16:57:20.5041833Z git config --global fetch.parallel 0 2025-11-03T16:57:20.5042115Z git config --global submodule.fetchJobs 0 2025-11-03T16:57:20.5042350Z  2025-11-03T16:57:20.5042608Z # Clean workspace. The default checkout action should also do this, but 2025-11-03T16:57:20.5042941Z # do it here as well just in case 2025-11-03T16:57:20.5043170Z if [[ -d .git ]]; then 2025-11-03T16:57:20.5043383Z  if [ -z "${NO_SUDO}" ]; then 2025-11-03T16:57:20.5043596Z  sudo git clean -ffdx 2025-11-03T16:57:20.5043795Z  else 2025-11-03T16:57:20.5043965Z  git clean -ffdx 2025-11-03T16:57:20.5044152Z  fi 2025-11-03T16:57:20.5044300Z fi 2025-11-03T16:57:20.5051772Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2025-11-03T16:57:20.5052069Z env: 2025-11-03T16:57:20.5052235Z GIT_DEFAULT_BRANCH: main 2025-11-03T16:57:20.5052419Z NO_SUDO: true 2025-11-03T16:57:20.5052650Z ##[endgroup] 2025-11-03T16:57:20.5207314Z ##[group]Run actions/checkout@v4 2025-11-03T16:57:20.5207541Z with: 2025-11-03T16:57:20.5207741Z ref: 3f6538febd727b782e6e13cfd026a309fb14351d 2025-11-03T16:57:20.5208001Z fetch-depth: 0 2025-11-03T16:57:20.5208179Z submodules: recursive 2025-11-03T16:57:20.5208370Z show-progress: false 2025-11-03T16:57:20.5208569Z repository: pytorch/pytorch 2025-11-03T16:57:20.5208890Z token: *** 2025-11-03T16:57:20.5209062Z ssh-strict: true 2025-11-03T16:57:20.5209227Z ssh-user: git 2025-11-03T16:57:20.5209404Z persist-credentials: true 2025-11-03T16:57:20.5209604Z clean: true 2025-11-03T16:57:20.5209792Z sparse-checkout-cone-mode: true 2025-11-03T16:57:20.5210005Z fetch-tags: false 2025-11-03T16:57:20.5210184Z lfs: false 2025-11-03T16:57:20.5210354Z set-safe-directory: true 2025-11-03T16:57:20.5210557Z env: 2025-11-03T16:57:20.5210714Z GIT_DEFAULT_BRANCH: main 2025-11-03T16:57:20.5210903Z ##[endgroup] 2025-11-03T16:57:20.6219287Z Syncing repository: pytorch/pytorch 2025-11-03T16:57:20.6220815Z ##[group]Getting Git version info 2025-11-03T16:57:20.6221662Z Working directory is '/home/ec2-user/actions-runner/_work/pytorch/pytorch' 2025-11-03T16:57:20.6222209Z [command]/usr/bin/git version 2025-11-03T16:57:20.6425428Z git version 2.50.1 2025-11-03T16:57:20.6449370Z ##[endgroup] 2025-11-03T16:57:20.6458692Z Copying '/home/ec2-user/.gitconfig' to '/home/ec2-user/actions-runner/_work/_temp/6bb9a450-d502-47cd-a096-5f859c7ad967/.gitconfig' 2025-11-03T16:57:20.6479981Z Temporarily overriding HOME='/home/ec2-user/actions-runner/_work/_temp/6bb9a450-d502-47cd-a096-5f859c7ad967' before making global git config changes 2025-11-03T16:57:20.6480716Z Adding repository directory to the temporary git global config as a safe directory 2025-11-03T16:57:20.6484876Z [command]/usr/bin/git config --global --add safe.directory /home/ec2-user/actions-runner/_work/pytorch/pytorch 2025-11-03T16:57:20.6538959Z Deleting the contents of '/home/ec2-user/actions-runner/_work/pytorch/pytorch' 2025-11-03T16:57:20.6542423Z ##[group]Initializing the repository 2025-11-03T16:57:20.6546336Z [command]/usr/bin/git init /home/ec2-user/actions-runner/_work/pytorch/pytorch 2025-11-03T16:57:20.6630169Z hint: Using 'master' as the name for the initial branch. This default branch name 2025-11-03T16:57:20.6630742Z hint: is subject to change. To configure the initial branch name to use in all 2025-11-03T16:57:20.6631275Z hint: of your new repositories, which will suppress this warning, call: 2025-11-03T16:57:20.6631654Z hint: 2025-11-03T16:57:20.6631930Z hint: git config --global init.defaultBranch 2025-11-03T16:57:20.6632250Z hint: 2025-11-03T16:57:20.6632560Z hint: Names commonly chosen instead of 'master' are 'main', 'trunk' and 2025-11-03T16:57:20.6633080Z hint: 'development'. The just-created branch can be renamed via this command: 2025-11-03T16:57:20.6633718Z hint: 2025-11-03T16:57:20.6633935Z hint: git branch -m 2025-11-03T16:57:20.6634200Z hint: 2025-11-03T16:57:20.6634526Z hint: Disable this message with "git config set advice.defaultBranchName false" 2025-11-03T16:57:20.6641021Z Initialized empty Git repository in /home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/ 2025-11-03T16:57:20.6650455Z [command]/usr/bin/git remote add origin https://github.com/pytorch/pytorch 2025-11-03T16:57:20.6695354Z ##[endgroup] 2025-11-03T16:57:20.6695795Z ##[group]Disabling automatic garbage collection 2025-11-03T16:57:20.6698714Z [command]/usr/bin/git config --local gc.auto 0 2025-11-03T16:57:20.6727819Z ##[endgroup] 2025-11-03T16:57:20.6728189Z ##[group]Setting up auth 2025-11-03T16:57:20.6733443Z [command]/usr/bin/git config --local --name-only --get-regexp core\.sshCommand 2025-11-03T16:57:20.6761538Z [command]/usr/bin/git submodule foreach --recursive sh -c "git config --local --name-only --get-regexp 'core\.sshCommand' && git config --local --unset-all 'core.sshCommand' || :" 2025-11-03T16:57:20.7168982Z [command]/usr/bin/git config --local --name-only --get-regexp http\.https\:\/\/github\.com\/\.extraheader 2025-11-03T16:57:20.7197865Z [command]/usr/bin/git submodule foreach --recursive sh -c "git config --local --name-only --get-regexp 'http\.https\:\/\/github\.com\/\.extraheader' && git config --local --unset-all 'http.https://github.com/.extraheader' || :" 2025-11-03T16:57:20.7553087Z [command]/usr/bin/git config --local http.https://github.com/.extraheader AUTHORIZATION: basic *** 2025-11-03T16:57:20.7600402Z ##[endgroup] 2025-11-03T16:57:20.7600848Z ##[group]Fetching the repository 2025-11-03T16:57:20.7607366Z [command]/usr/bin/git -c protocol.version=2 fetch --prune --no-recurse-submodules origin +refs/heads/*:refs/remotes/origin/* +refs/tags/*:refs/tags/* 2025-11-03T16:58:01.1690351Z From https://github.com/pytorch/pytorch 2025-11-03T16:58:01.1690948Z * [new branch] 2.6.0.dev20241004+ -> origin/2.6.0.dev20241004+ 2025-11-03T16:58:01.1691864Z * [new branch] AaronWang04_addmmfusion_perftest -> origin/AaronWang04_addmmfusion_perftest 2025-11-03T16:58:01.1692511Z * [new branch] DyVariaSourceRepr -> origin/DyVariaSourceRepr 2025-11-03T16:58:01.1693714Z * [new branch] DynamoFixGit -> origin/DynamoFixGit 2025-11-03T16:58:01.1696328Z * [new branch] DynamoVariaT -> origin/DynamoVariaT 2025-11-03T16:58:01.1699199Z * [new branch] Flamefire-misplaced-TestFailure -> origin/Flamefire-misplaced-TestFailure 2025-11-03T16:58:01.1701932Z * [new branch] HDCharles-2.6.0-release-notes -> origin/HDCharles-2.6.0-release-notes 2025-11-03T16:58:01.1705618Z * [new branch] IvanKobzarev/stack/1 -> origin/IvanKobzarev/stack/1 2025-11-03T16:58:01.1708056Z * [new branch] IvanKobzarev/stack/2 -> origin/IvanKobzarev/stack/2 2025-11-03T16:58:01.1710576Z * [new branch] NicoshevSVE128 -> origin/NicoshevSVE128 2025-11-03T16:58:01.1712661Z * [new branch] PR-AOTInductorNoneBug -> origin/PR-AOTInductorNoneBug 2025-11-03T16:58:01.1714991Z * [new branch] PR-AOTInductorNoneBugFix -> origin/PR-AOTInductorNoneBugFix 2025-11-03T16:58:01.1717184Z * [new branch] PR-FixConfigsIssue -> origin/PR-FixConfigsIssue 2025-11-03T16:58:01.1719626Z * [new branch] PR-NoneBugFix-viable -> origin/PR-NoneBugFix-viable 2025-11-03T16:58:01.1722061Z * [new branch] PR-ResetToZero -> origin/PR-ResetToZero 2025-11-03T16:58:01.1724477Z * [new branch] Update-Flash-Packaging -> origin/Update-Flash-Packaging 2025-11-03T16:58:01.1727274Z * [new branch] VLA_exp -> origin/VLA_exp 2025-11-03T16:58:01.1729494Z * [new branch] VariaT -> origin/VariaT 2025-11-03T16:58:01.1731970Z * [new branch] add-pyrefly-to-lintrunner -> origin/add-pyrefly-to-lintrunner 2025-11-03T16:58:01.1734124Z * [new branch] add_conv3d_ops -> origin/add_conv3d_ops 2025-11-03T16:58:01.1736350Z * [new branch] add_op_to_dashboard -> origin/add_op_to_dashboard 2025-11-03T16:58:01.1738609Z * [new branch] addmm-heuristic -> origin/addmm-heuristic 2025-11-03T16:58:01.1740837Z * [new branch] addmm10000rtx -> origin/addmm10000rtx 2025-11-03T16:58:01.1743825Z * [new branch] adi/onednn_aarch64 -> origin/adi/onednn_aarch64 2025-11-03T16:58:01.1745984Z * [new branch] adi/test -> origin/adi/test 2025-11-03T16:58:01.1748156Z * [new branch] adi/test_bgemm -> origin/adi/test_bgemm 2025-11-03T16:58:01.1750353Z * [new branch] adi/test_fusions -> origin/adi/test_fusions 2025-11-03T16:58:01.1752426Z * [new branch] adi/test_onednn -> origin/adi/test_onednn 2025-11-03T16:58:01.1754750Z * [new branch] adi/test_onednn_v3.9 -> origin/adi/test_onednn_v3.9 2025-11-03T16:58:01.1756963Z * [new branch] adi/test_presve_change -> origin/adi/test_presve_change 2025-11-03T16:58:01.1758958Z * [new branch] adi/test_timm -> origin/adi/test_timm 2025-11-03T16:58:01.1761429Z * [new branch] adi/testpresve_change -> origin/adi/testpresve_change 2025-11-03T16:58:01.1765074Z * [new branch] aditew01/test/vec_bf16 -> origin/aditew01/test/vec_bf16 2025-11-03T16:58:01.1767180Z * [new branch] ah-globalfeedback-hook -> origin/ah-globalfeedback-hook 2025-11-03T16:58:01.1769409Z * [new branch] alt-disable -> origin/alt-disable 2025-11-03T16:58:01.1772343Z * [new branch] angelayi/allow_fake -> origin/angelayi/allow_fake 2025-11-03T16:58:01.1774574Z * [new branch] angelayi/aoti_additional_files -> origin/angelayi/aoti_additional_files 2025-11-03T16:58:01.1776806Z * [new branch] angelayi/benchmark -> origin/angelayi/benchmark 2025-11-03T16:58:01.1778911Z * [new branch] angelayi/change_pytree_serialization -> origin/angelayi/change_pytree_serialization 2025-11-03T16:58:01.1781021Z * [new branch] angelayi/const_folder -> origin/angelayi/const_folder 2025-11-03T16:58:01.1783366Z * [new branch] angelayi/cpp_loader -> origin/angelayi/cpp_loader 2025-11-03T16:58:01.1785494Z * [new branch] angelayi/joint_kwargs -> origin/angelayi/joint_kwargs 2025-11-03T16:58:01.1787975Z * [new branch] angelayi/no_so_weight -> origin/angelayi/no_so_weight 2025-11-03T16:58:01.1790526Z * [new branch] angelayi/opaque_obj_v2 -> origin/angelayi/opaque_obj_v2 2025-11-03T16:58:01.1793120Z * [new branch] angelayi/post_grad -> origin/angelayi/post_grad 2025-11-03T16:58:01.1795169Z * [new branch] angelayi/pytree -> origin/angelayi/pytree 2025-11-03T16:58:01.1797421Z * [new branch] angelayi/scan_layers -> origin/angelayi/scan_layers 2025-11-03T16:58:01.1799634Z * [new branch] angelayi/state_dict -> origin/angelayi/state_dict 2025-11-03T16:58:01.1801953Z * [new branch] angelayi/static_input_indices -> origin/angelayi/static_input_indices 2025-11-03T16:58:01.1804150Z * [new branch] angelayi/symint_input -> origin/angelayi/symint_input 2025-11-03T16:58:01.1806273Z * [new branch] angelayi/symm_mem -> origin/angelayi/symm_mem 2025-11-03T16:58:01.1808550Z * [new branch] angelayi/test_cpp -> origin/angelayi/test_cpp 2025-11-03T16:58:01.1810704Z * [new branch] angelayi/torch_size -> origin/angelayi/torch_size 2025-11-03T16:58:01.1812916Z * [new branch] annotate_before_eliminate -> origin/annotate_before_eliminate 2025-11-03T16:58:01.1815039Z * [new branch] annotate_fn -> origin/annotate_fn 2025-11-03T16:58:01.1817272Z * [new branch] annotation_bw -> origin/annotation_bw 2025-11-03T16:58:01.1819448Z * [new branch] annotation_bw_1 -> origin/annotation_bw_1 2025-11-03T16:58:01.1821742Z * [new branch] annotation_dynamo -> origin/annotation_dynamo 2025-11-03T16:58:01.1824000Z * [new branch] annotation_log -> origin/annotation_log 2025-11-03T16:58:01.1826266Z * [new branch] annotation_replay -> origin/annotation_replay 2025-11-03T16:58:01.1828415Z * [new branch] annotation_submod -> origin/annotation_submod 2025-11-03T16:58:01.1830616Z * [new branch] aot_eager_stack_trace -> origin/aot_eager_stack_trace 2025-11-03T16:58:01.1832841Z * [new branch] aoti-cuda-alloc -> origin/aoti-cuda-alloc 2025-11-03T16:58:01.1835025Z * [new branch] aoti_fqn_name_interface -> origin/aoti_fqn_name_interface 2025-11-03T16:58:01.1837146Z * [new branch] aoti_metal_shimify -> origin/aoti_metal_shimify 2025-11-03T16:58:01.1839318Z * [new branch] aoti_package_weights_binary -> origin/aoti_package_weights_binary 2025-11-03T16:58:01.1841484Z * [new branch] aoti_shim_library_list -> origin/aoti_shim_library_list 2025-11-03T16:58:01.1843669Z * [new branch] aoti_target_windows -> origin/aoti_target_windows 2025-11-03T16:58:01.1845853Z * [new branch] aoti_weight_sharing -> origin/aoti_weight_sharing 2025-11-03T16:58:01.1849383Z * [new branch] arsh/feat/inductor_check_profiling -> origin/arsh/feat/inductor_check_profiling 2025-11-03T16:58:01.1851571Z * [new branch] async_tp -> origin/async_tp 2025-11-03T16:58:01.1853962Z * [new branch] atalman-inductor-perf-cu124 -> origin/atalman-inductor-perf-cu124 2025-11-03T16:58:01.1856196Z * [new branch] atalman-inductor-perf-cu124.1 -> origin/atalman-inductor-perf-cu124.1 2025-11-03T16:58:01.1858425Z * [new branch] atalman-patch-1 -> origin/atalman-patch-1 2025-11-03T16:58:01.1860740Z * [new branch] atalman-patch-2 -> origin/atalman-patch-2 2025-11-03T16:58:01.1863234Z * [new branch] atalman-patch-3 -> origin/atalman-patch-3 2025-11-03T16:58:01.1865446Z * [new branch] atalman-patch-5 -> origin/atalman-patch-5 2025-11-03T16:58:01.1867741Z * [new branch] atalman_inductor_2.3.0 -> origin/atalman_inductor_2.3.0 2025-11-03T16:58:01.1870004Z * [new branch] atalman_inductor_2.3.1 -> origin/atalman_inductor_2.3.1 2025-11-03T16:58:01.1872349Z * [new branch] atalman_inductor_2.4.0 -> origin/atalman_inductor_2.4.0 2025-11-03T16:58:01.1874427Z * [new branch] atalman_inductor_2.4.x -> origin/atalman_inductor_2.4.x 2025-11-03T16:58:01.1876603Z * [new branch] attention_benchmark -> origin/attention_benchmark 2025-11-03T16:58:01.1878907Z * [new branch] attention_benchmarking_clean -> origin/attention_benchmarking_clean 2025-11-03T16:58:01.1881651Z * [new branch] bahuang/dt_fix_scalar_add -> origin/bahuang/dt_fix_scalar_add 2025-11-03T16:58:01.1883862Z * [new branch] bahuang/fix_debug_mode -> origin/bahuang/fix_debug_mode 2025-11-03T16:58:01.1885848Z * [new branch] bahuang/fix_expand -> origin/bahuang/fix_expand 2025-11-03T16:58:01.1887987Z * [new branch] bahuang/test -> origin/bahuang/test 2025-11-03T16:58:01.1890993Z * [new branch] base/1.5 -> origin/base/1.5 2025-11-03T16:58:01.1893335Z * [new branch] batching_sdpa_efficient_attention -> origin/batching_sdpa_efficient_attention 2025-11-03T16:58:01.1895610Z * [new branch] bench_scaled_mm_ops -> origin/bench_scaled_mm_ops 2025-11-03T16:58:01.1897857Z * [new branch] benchmark-updates -> origin/benchmark-updates 2025-11-03T16:58:01.1900020Z * [new branch] benchmarking-script -> origin/benchmarking-script 2025-11-03T16:58:01.1903100Z * [new branch] bertmaher/pinbump26 -> origin/bertmaher/pinbump26 2025-11-03T16:58:01.1906058Z * [new branch] bertrand/cutlass -> origin/bertrand/cutlass 2025-11-03T16:58:01.1908817Z * [new branch] bf/cg-backend -> origin/bf/cg-backend 2025-11-03T16:58:01.1911074Z * [new branch] bf/cg-partition-custom-op-mutation -> origin/bf/cg-partition-custom-op-mutation 2025-11-03T16:58:01.1913038Z * [new branch] bf/cg-remove-check -> origin/bf/cg-remove-check 2025-11-03T16:58:01.1915204Z * [new branch] bf/clean-hf -> origin/bf/clean-hf 2025-11-03T16:58:01.1917267Z * [new branch] bf/clean-timm -> origin/bf/clean-timm 2025-11-03T16:58:01.1919403Z * [new branch] bf/clean-torchbench -> origin/bf/clean-torchbench 2025-11-03T16:58:01.1922017Z * [new branch] bf/clean-torchbench-hf -> origin/bf/clean-torchbench-hf 2025-11-03T16:58:01.1924414Z * [new branch] bf/cudagraph -> origin/bf/cudagraph 2025-11-03T16:58:01.1927267Z * [new branch] bf/cudagraph-disable-input-mutation -> origin/bf/cudagraph-disable-input-mutation 2025-11-03T16:58:01.1929695Z * [new branch] bf/cudagraph-enable-input-mutation-support-benchmark -> origin/bf/cudagraph-enable-input-mutation-support-benchmark 2025-11-03T16:58:01.1931738Z * [new branch] bf/cudagraph-partition -> origin/bf/cudagraph-partition 2025-11-03T16:58:01.1933875Z * [new branch] bf/donated-buffer-bench -> origin/bf/donated-buffer-bench 2025-11-03T16:58:01.1936081Z * [new branch] bf/dynamo-partition -> origin/bf/dynamo-partition 2025-11-03T16:58:01.1938222Z * [new branch] bf/pa-non-divisible -> origin/bf/pa-non-divisible 2025-11-03T16:58:01.1940455Z * [new branch] bf/partition-cache-free-symbols -> origin/bf/partition-cache-free-symbols 2025-11-03T16:58:01.1942771Z * [new branch] bf/partition-custom-rules -> origin/bf/partition-custom-rules 2025-11-03T16:58:01.1944902Z * [new branch] bf/partition-memory-plan -> origin/bf/partition-memory-plan 2025-11-03T16:58:01.1947205Z * [new branch] bf/partition-move-cpu -> origin/bf/partition-move-cpu 2025-11-03T16:58:01.1949370Z * [new branch] bf/partition-view-fallback -> origin/bf/partition-view-fallback 2025-11-03T16:58:01.1951648Z * [new branch] bf/remove-check-55b0c39d -> origin/bf/remove-check-55b0c39d 2025-11-03T16:58:01.1953656Z * [new branch] bf16_support_per_channel -> origin/bf16_support_per_channel 2025-11-03T16:58:01.1955814Z * [new branch] bisect_perf_hf_T5_3acc6eac492 -> origin/bisect_perf_hf_T5_3acc6eac492 2025-11-03T16:58:01.1957909Z * [new branch] bisect_perf_hf_T5_3fcf66f61fb -> origin/bisect_perf_hf_T5_3fcf66f61fb 2025-11-03T16:58:01.1960042Z * [new branch] bisect_perf_hf_T5_4009d154129 -> origin/bisect_perf_hf_T5_4009d154129 2025-11-03T16:58:01.1962497Z * [new branch] bisect_perf_hf_T5_40d0740e73d -> origin/bisect_perf_hf_T5_40d0740e73d 2025-11-03T16:58:01.1964682Z * [new branch] bisect_perf_hf_T5_5268754e -> origin/bisect_perf_hf_T5_5268754e 2025-11-03T16:58:01.1966840Z * [new branch] bisect_perf_hf_T5_7d89a8d385c -> origin/bisect_perf_hf_T5_7d89a8d385c 2025-11-03T16:58:01.1968958Z * [new branch] bisect_perf_hf_T5_b7a25c1ee7c -> origin/bisect_perf_hf_T5_b7a25c1ee7c 2025-11-03T16:58:01.1971122Z * [new branch] bisect_perf_hf_T5_c25b201583f -> origin/bisect_perf_hf_T5_c25b201583f 2025-11-03T16:58:01.1973299Z * [new branch] bisect_perf_hf_T5_c93e57efac0 -> origin/bisect_perf_hf_T5_c93e57efac0 2025-11-03T16:58:01.1975397Z * [new branch] bisect_perf_hf_T5_ca9813ea149 -> origin/bisect_perf_hf_T5_ca9813ea149 2025-11-03T16:58:01.1977528Z * [new branch] bisect_perf_hf_T5_d65f194a -> origin/bisect_perf_hf_T5_d65f194a 2025-11-03T16:58:01.1979640Z * [new branch] bisect_perf_hf_T5_da94ab0b -> origin/bisect_perf_hf_T5_da94ab0b 2025-11-03T16:58:01.1981860Z * [new branch] bisect_perf_hf_T5_da94ab0b_new -> origin/bisect_perf_hf_T5_da94ab0b_new 2025-11-03T16:58:01.1984031Z * [new branch] bisect_perf_hf_T5_db4e8a1d8a8 -> origin/bisect_perf_hf_T5_db4e8a1d8a8 2025-11-03T16:58:01.1986131Z * [new branch] bisect_perf_hf_T5_e0d97e936a2 -> origin/bisect_perf_hf_T5_e0d97e936a2 2025-11-03T16:58:01.1988243Z * [new branch] bisect_perf_hf_T5_f23621ec563 -> origin/bisect_perf_hf_T5_f23621ec563 2025-11-03T16:58:01.1991105Z * [new branch] bowbao/wip_prs -> origin/bowbao/wip_prs 2025-11-03T16:58:01.1993974Z * [new branch] brister/break_dynamic_scalar -> origin/brister/break_dynamic_scalar 2025-11-03T16:58:01.1996050Z * [new branch] brister/cast_rank0 -> origin/brister/cast_rank0 2025-11-03T16:58:01.1998075Z * [new branch] brister/item_fx -> origin/brister/item_fx 2025-11-03T16:58:01.2000218Z * [new branch] brister/test_inductor_all_fx -> origin/brister/test_inductor_all_fx 2025-11-03T16:58:01.2002445Z * [new branch] brister/tiled_reduction_no_numel_check -> origin/brister/tiled_reduction_no_numel_check 2025-11-03T16:58:01.2004517Z * [new branch] bwd-backup -> origin/bwd-backup 2025-11-03T16:58:01.2006780Z * [new branch] c57382a49 -> origin/c57382a49 2025-11-03T16:58:01.2008901Z * [new branch] ca_0431d47eaa -> origin/ca_0431d47eaa 2025-11-03T16:58:01.2011052Z * [new branch] ca_fix_0431d47eaa -> origin/ca_fix_0431d47eaa 2025-11-03T16:58:01.2013966Z * [new branch] camyllh/test_setup_hooks_push -> origin/camyllh/test_setup_hooks_push 2025-11-03T16:58:01.2016191Z * [new branch] cherry-pick-152361-by-pytorch_bot_bot_ -> origin/cherry-pick-152361-by-pytorch_bot_bot_ 2025-11-03T16:58:01.2018347Z * [new branch] cherry-pick-157453-by-pytorch_bot_bot_ -> origin/cherry-pick-157453-by-pytorch_bot_bot_ 2025-11-03T16:58:01.2020619Z * [new branch] cherry-pick-157513-by-pytorch_bot_bot_ -> origin/cherry-pick-157513-by-pytorch_bot_bot_ 2025-11-03T16:58:01.2023707Z * [new branch] cherry-pick-157695-by-pytorch_bot_bot_ -> origin/cherry-pick-157695-by-pytorch_bot_bot_ 2025-11-03T16:58:01.2026025Z * [new branch] cherry-pick-157732-by-pytorch_bot_bot_ -> origin/cherry-pick-157732-by-pytorch_bot_bot_ 2025-11-03T16:58:01.2028142Z * [new branch] cherry-pick-159969-by-pytorch_bot_bot_ -> origin/cherry-pick-159969-by-pytorch_bot_bot_ 2025-11-03T16:58:01.2030309Z * [new branch] cherry-pick-160586-by-pytorch_bot_bot_ -> origin/cherry-pick-160586-by-pytorch_bot_bot_ 2025-11-03T16:58:01.2032893Z * [new branch] cherry-pick-162194-by-pytorch_bot_bot_ -> origin/cherry-pick-162194-by-pytorch_bot_bot_ 2025-11-03T16:58:01.2035272Z * [new branch] cherry-pick-162693-by-pytorch_bot_bot_ -> origin/cherry-pick-162693-by-pytorch_bot_bot_ 2025-11-03T16:58:01.2037577Z * [new branch] cherry-pick-162764-by-pytorch_bot_bot_ -> origin/cherry-pick-162764-by-pytorch_bot_bot_ 2025-11-03T16:58:01.2039762Z * [new branch] cherry-pick-163029-by-pytorch_bot_bot_ -> origin/cherry-pick-163029-by-pytorch_bot_bot_ 2025-11-03T16:58:01.2042106Z * [new branch] cherry-pick-163776-by-pytorch_bot_bot_ -> origin/cherry-pick-163776-by-pytorch_bot_bot_ 2025-11-03T16:58:01.2044369Z * [new branch] cherry-pick-164774-by-pytorch_bot_bot_ -> origin/cherry-pick-164774-by-pytorch_bot_bot_ 2025-11-03T16:58:01.2046727Z * [new branch] cherry-pick-164870-by-pytorch_bot_bot_ -> origin/cherry-pick-164870-by-pytorch_bot_bot_ 2025-11-03T16:58:01.2048997Z * [new branch] cherry-pick-164946-by-pytorch_bot_bot_ -> origin/cherry-pick-164946-by-pytorch_bot_bot_ 2025-11-03T16:58:01.2051203Z * [new branch] cherry-pick-165013-by-pytorch_bot_bot_ -> origin/cherry-pick-165013-by-pytorch_bot_bot_ 2025-11-03T16:58:01.2053321Z * [new branch] cherry-pick-165465-by-pytorch_bot_bot_ -> origin/cherry-pick-165465-by-pytorch_bot_bot_ 2025-11-03T16:58:01.2055547Z * [new branch] cherry-pick-165665-by-pytorch_bot_bot_ -> origin/cherry-pick-165665-by-pytorch_bot_bot_ 2025-11-03T16:58:01.2057791Z * [new branch] cherry-pick-165794-by-pytorch_bot_bot_ -> origin/cherry-pick-165794-by-pytorch_bot_bot_ 2025-11-03T16:58:01.2059986Z * [new branch] cherry-pick-166181-by-pytorch_bot_bot_ -> origin/cherry-pick-166181-by-pytorch_bot_bot_ 2025-11-03T16:58:01.2062299Z * [new branch] chuanqi129-patch-1 -> origin/chuanqi129-patch-1 2025-11-03T16:58:01.2064508Z * [new branch] ci_attn -> origin/ci_attn 2025-11-03T16:58:01.2066708Z * [new branch] codegen_trace -> origin/codegen_trace 2025-11-03T16:58:01.2068877Z * [new branch] codex-testing -> origin/codex-testing 2025-11-03T16:58:01.2071794Z * [new branch] codex/add-check_memory_overlap-helper-functions -> origin/codex/add-check_memory_overlap-helper-functions 2025-11-03T16:58:01.2073944Z * [new branch] codex/create-test-for-tensor-memory-leak-in-cudagraph -> origin/codex/create-test-for-tensor-memory-leak-in-cudagraph 2025-11-03T16:58:01.2076350Z * [new branch] codex/enhance-cuda.matmul-with-allow_splitk-argument -> origin/codex/enhance-cuda.matmul-with-allow_splitk-argument 2025-11-03T16:58:01.2078695Z * [new branch] codex/fix-issue-121219-in-pytorch -> origin/codex/fix-issue-121219-in-pytorch 2025-11-03T16:58:01.2081077Z * [new branch] codex/refactor-dimension-handling-in-shape.cu -> origin/codex/refactor-dimension-handling-in-shape.cu 2025-11-03T16:58:01.2083276Z * [new branch] codex/refactor-lintrunner-config-to-use-uv-run -> origin/codex/refactor-lintrunner-config-to-use-uv-run 2025-11-03T16:58:01.2085358Z * [new branch] context_test -> origin/context_test 2025-11-03T16:58:01.2088342Z * [new branch] copilot/code-style-cleanup-python-pip -> origin/copilot/code-style-cleanup-python-pip 2025-11-03T16:58:01.2091004Z * [new branch] cpio/fix_new_ami_tests -> origin/cpio/fix_new_ami_tests 2025-11-03T16:58:01.2093293Z * [new branch] cpp-docs-dependency-upgrade -> origin/cpp-docs-dependency-upgrade 2025-11-03T16:58:01.2095339Z * [new branch] crcrpar-patch-1 -> origin/crcrpar-patch-1 2025-11-03T16:58:01.2098148Z * [new branch] csl/always_produce_xml -> origin/csl/always_produce_xml 2025-11-03T16:58:01.2100257Z * [new branch] csl/build_test_more_procs -> origin/csl/build_test_more_procs 2025-11-03T16:58:01.2102351Z * [new branch] csl/build_test_more_procs2 -> origin/csl/build_test_more_procs2 2025-11-03T16:58:01.2104470Z * [new branch] csl/debug_tests_larger_runner -> origin/csl/debug_tests_larger_runner 2025-11-03T16:58:01.2106815Z * [new branch] csl/disable_test_leaking_memory -> origin/csl/disable_test_leaking_memory 2025-11-03T16:58:01.2109320Z * [new branch] csl/failed_consistently_null -> origin/csl/failed_consistently_null 2025-11-03T16:58:01.2111793Z * [new branch] csl/fix_internal_graph_executor -> origin/csl/fix_internal_graph_executor 2025-11-03T16:58:01.2113925Z * [new branch] csl/fix_periodic_debug_1 -> origin/csl/fix_periodic_debug_1 2025-11-03T16:58:01.2116172Z * [new branch] csl/inductor_h100_nightly -> origin/csl/inductor_h100_nightly 2025-11-03T16:58:01.2118325Z * [new branch] csl/katex -> origin/csl/katex 2025-11-03T16:58:01.2120560Z * [new branch] csl/keep_going_manual -> origin/csl/keep_going_manual 2025-11-03T16:58:01.2122952Z * [new branch] csl/larger_runner -> origin/csl/larger_runner 2025-11-03T16:58:01.2125149Z * [new branch] csl/lint_testing -> origin/csl/lint_testing 2025-11-03T16:58:01.2127234Z * [new branch] csl/lint_thing -> origin/csl/lint_thing 2025-11-03T16:58:01.2129473Z * [new branch] csl/lintrunner_stuff -> origin/csl/lintrunner_stuff 2025-11-03T16:58:01.2131687Z * [new branch] csl/mps_sharding -> origin/csl/mps_sharding 2025-11-03T16:58:01.2133875Z * [new branch] csl/multistage_docker -> origin/csl/multistage_docker 2025-11-03T16:58:01.2136050Z * [new branch] csl/no_keep_goin_rocm -> origin/csl/no_keep_goin_rocm 2025-11-03T16:58:01.2138247Z * [new branch] csl/nogpu_inductor_caching_test -> origin/csl/nogpu_inductor_caching_test 2025-11-03T16:58:01.2140405Z * [new branch] csl/nogpu_on_no_gpu -> origin/csl/nogpu_on_no_gpu 2025-11-03T16:58:01.2142779Z * [new branch] csl/periodic_disable -> origin/csl/periodic_disable 2025-11-03T16:58:01.2144971Z * [new branch] csl/print_something_for_log_classifier -> origin/csl/print_something_for_log_classifier 2025-11-03T16:58:01.2147026Z * [new branch] csl/print_timing -> origin/csl/print_timing 2025-11-03T16:58:01.2149234Z * [new branch] csl/remove_maybe_unused_var -> origin/csl/remove_maybe_unused_var 2025-11-03T16:58:01.2151407Z * [new branch] csl/remove_run_parallel -> origin/csl/remove_run_parallel 2025-11-03T16:58:01.2153481Z * [new branch] csl/remove_unused_vars -> origin/csl/remove_unused_vars 2025-11-03T16:58:01.2155623Z * [new branch] csl/revert_no_rule -> origin/csl/revert_no_rule 2025-11-03T16:58:01.2157813Z * [new branch] csl/revert_open -> origin/csl/revert_open 2025-11-03T16:58:01.2159941Z * [new branch] csl/run_full_lint_trunk -> origin/csl/run_full_lint_trunk 2025-11-03T16:58:01.2162167Z * [new branch] csl/skip_build -> origin/csl/skip_build 2025-11-03T16:58:01.2164348Z * [new branch] csl/smaller_avx_amx_runenrs -> origin/csl/smaller_avx_amx_runenrs 2025-11-03T16:58:01.2166482Z * [new branch] csl/td_job_level -> origin/csl/td_job_level 2025-11-03T16:58:01.2168822Z * [new branch] csl/test_cuda_build_large_runner -> origin/csl/test_cuda_build_large_runner 2025-11-03T16:58:01.2170985Z * [new branch] csl/test_owners_autograd_dispatch_nn -> origin/csl/test_owners_autograd_dispatch_nn 2025-11-03T16:58:01.2173101Z * [new branch] csl/test_owners_higher_confidence -> origin/csl/test_owners_higher_confidence 2025-11-03T16:58:01.2175240Z * [new branch] csl/upload_all_data -> origin/csl/upload_all_data 2025-11-03T16:58:01.2177348Z * [new branch] csl/win_cpp_tests -> origin/csl/win_cpp_tests 2025-11-03T16:58:01.2179564Z * [new branch] csl/win_sccache -> origin/csl/win_sccache 2025-11-03T16:58:01.2181920Z * [new branch] csl/workflow_consistency_linter -> origin/csl/workflow_consistency_linter 2025-11-03T16:58:01.2184244Z * [new branch] csl/workflow_sync_more_files -> origin/csl/workflow_sync_more_files 2025-11-03T16:58:01.2186245Z * [new branch] csl/xml_stuff -> origin/csl/xml_stuff 2025-11-03T16:58:01.2188566Z * [new branch] cublasltrelax2 -> origin/cublasltrelax2 2025-11-03T16:58:01.2190676Z * [new branch] cublasrelax2 -> origin/cublasrelax2 2025-11-03T16:58:01.2192843Z * [new branch] custom_lowering_dict -> origin/custom_lowering_dict 2025-11-03T16:58:01.2195935Z * [new branch] daxia6/2.8o3 -> origin/daxia6/2.8o3 2025-11-03T16:58:01.2198150Z * [new branch] debug-guard -> origin/debug-guard 2025-11-03T16:58:01.2200319Z * [new branch] delete-quant-docs -> origin/delete-quant-docs 2025-11-03T16:58:01.2206001Z * [new branch] dependabot/pip/dot-ci/docker/ci_commit_pins/main/transformers-4.57.0 -> origin/dependabot/pip/dot-ci/docker/ci_commit_pins/main/transformers-4.57.0 2025-11-03T16:58:01.2208396Z * [new branch] dependabot/pip/dot-ci/docker/ci_commit_pins/main/transformers-4.57.1 -> origin/dependabot/pip/dot-ci/docker/ci_commit_pins/main/transformers-4.57.1 2025-11-03T16:58:01.2210945Z * [new branch] desertfire/test_cpp_wrapper -> origin/desertfire/test_cpp_wrapper 2025-11-03T16:58:01.2213120Z * [new branch] desertfire/triton-cpu-for-aarch64 -> origin/desertfire/triton-cpu-for-aarch64 2025-11-03T16:58:01.2216291Z * [new branch] dev/dhruva/flex_attn_opt -> origin/dev/dhruva/flex_attn_opt 2025-11-03T16:58:01.2219421Z * [new branch] dev/joona/MPSNDArrayAdd -> origin/dev/joona/MPSNDArrayAdd 2025-11-03T16:58:01.2221894Z * [new branch] dev/joona/Unranked -> origin/dev/joona/Unranked 2025-11-03T16:58:01.2224265Z * [new branch] dev/joona/cat -> origin/dev/joona/cat 2025-11-03T16:58:01.2226467Z * [new branch] dev/joona/embeddingbag -> origin/dev/joona/embeddingbag 2025-11-03T16:58:01.2228824Z * [new branch] dev/joona/getTensorsString -> origin/dev/joona/getTensorsString 2025-11-03T16:58:01.2231162Z * [new branch] dev/joona/mps_linear_macos14 -> origin/dev/joona/mps_linear_macos14 2025-11-03T16:58:01.2233405Z * [new branch] dev/joona/sdpa -> origin/dev/joona/sdpa 2025-11-03T16:58:01.2235684Z * [new branch] dev/joona/type_inf -> origin/dev/joona/type_inf 2025-11-03T16:58:01.2237852Z * [new branch] dev/joona/upsize3d -> origin/dev/joona/upsize3d 2025-11-03T16:58:01.2239985Z * [new branch] disp_counter -> origin/disp_counter 2025-11-03T16:58:01.2242146Z * [new branch] document -> origin/document 2025-11-03T16:58:01.2244324Z * [new branch] document-apis -> origin/document-apis 2025-11-03T16:58:01.2246445Z * [new branch] dynamo_stack_trace -> origin/dynamo_stack_trace 2025-11-03T16:58:01.2248759Z * [new branch] eager_model_benchmarks -> origin/eager_model_benchmarks 2025-11-03T16:58:01.2251498Z * [new branch] embg/test_inductor_ci_128B -> origin/embg/test_inductor_ci_128B 2025-11-03T16:58:01.2253674Z * [new branch] embg/test_inductor_ci_base -> origin/embg/test_inductor_ci_base 2025-11-03T16:58:01.2255832Z * [new branch] embg/test_inductor_ci_control -> origin/embg/test_inductor_ci_control 2025-11-03T16:58:01.2257924Z * [new branch] embg/triton_l2_prefetch_128B -> origin/embg/triton_l2_prefetch_128B 2025-11-03T16:58:01.2260238Z * [new branch] embg/triton_l2_prefetch_256B -> origin/embg/triton_l2_prefetch_256B 2025-11-03T16:58:01.2263317Z * [new branch] exclamaforte/amd-ma -> origin/exclamaforte/amd-ma 2025-11-03T16:58:01.2265397Z * [new branch] exclamaforte/combo-kernels-perf-run -> origin/exclamaforte/combo-kernels-perf-run 2025-11-03T16:58:01.2267470Z * [new branch] exclamaforte/do_bench_refactor -> origin/exclamaforte/do_bench_refactor 2025-11-03T16:58:01.2269659Z * [new branch] exclamaforte/enable-mem-dep-fusion -> origin/exclamaforte/enable-mem-dep-fusion 2025-11-03T16:58:01.2271894Z * [new branch] exclamaforte/fix-exhaustive-autotuning -> origin/exclamaforte/fix-exhaustive-autotuning 2025-11-03T16:58:01.2274373Z * [new branch] exclamaforte/fix-trace-parsing-fx-svg -> origin/exclamaforte/fix-trace-parsing-fx-svg 2025-11-03T16:58:01.2276833Z * [new branch] exclamaforte/force-pointwise-cat-perf-run -> origin/exclamaforte/force-pointwise-cat-perf-run 2025-11-03T16:58:01.2278883Z * [new branch] exclamaforte/fusion-data -> origin/exclamaforte/fusion-data 2025-11-03T16:58:01.2281116Z * [new branch] exclamaforte/gemm-benchmark-run -> origin/exclamaforte/gemm-benchmark-run 2025-11-03T16:58:01.2283224Z * [new branch] exclamaforte/gemm-export-model -> origin/exclamaforte/gemm-export-model 2025-11-03T16:58:01.2285399Z * [new branch] exclamaforte/gemm-model -> origin/exclamaforte/gemm-model 2025-11-03T16:58:01.2287679Z * [new branch] exclamaforte/gemm-model-all-data-collection -> origin/exclamaforte/gemm-model-all-data-collection 2025-11-03T16:58:01.2289715Z * [new branch] exclamaforte/gemm-to-amd -> origin/exclamaforte/gemm-to-amd 2025-11-03T16:58:01.2291922Z * [new branch] exclamaforte/just-gemm-model -> origin/exclamaforte/just-gemm-model 2025-11-03T16:58:01.2294236Z * [new branch] exclamaforte/just-gemm-model-no-refactor -> origin/exclamaforte/just-gemm-model-no-refactor 2025-11-03T16:58:01.2296322Z * [new branch] exclamaforte/profile-diff-algo -> origin/exclamaforte/profile-diff-algo 2025-11-03T16:58:01.2298485Z * [new branch] exclamaforte/profiler-visualization -> origin/exclamaforte/profiler-visualization 2025-11-03T16:58:01.2300685Z * [new branch] exclamaforte/test_cpp_wrapper_mode -> origin/exclamaforte/test_cpp_wrapper_mode 2025-11-03T16:58:01.2303042Z * [new branch] exclamaforte/update-autotune-configs -> origin/exclamaforte/update-autotune-configs 2025-11-03T16:58:01.2305256Z * [new branch] exclamaforte/update-autotune-configs-2 -> origin/exclamaforte/update-autotune-configs-2 2025-11-03T16:58:01.2308014Z * [new branch] exclamforte/gemm-model-final -> origin/exclamforte/gemm-model-final 2025-11-03T16:58:01.2310209Z * [new branch] exec -> origin/exec 2025-11-03T16:58:01.2312503Z * [new branch] experimental-mosaic -> origin/experimental-mosaic 2025-11-03T16:58:01.2314668Z * [new branch] export-D58091437 -> origin/export-D58091437 2025-11-03T16:58:01.2316924Z * [new branch] export-D61047529 -> origin/export-D61047529 2025-11-03T16:58:01.2319114Z * [new branch] export-D71412006 -> origin/export-D71412006 2025-11-03T16:58:01.2321606Z * [new branch] export-D73042989 -> origin/export-D73042989 2025-11-03T16:58:01.2323871Z * [new branch] export-D76797250 -> origin/export-D76797250 2025-11-03T16:58:01.2326576Z * [new branch] export-D78375400 -> origin/export-D78375400 2025-11-03T16:58:01.2328678Z * [new branch] export-D78431305 -> origin/export-D78431305 2025-11-03T16:58:01.2330809Z * [new branch] export-D78822171 -> origin/export-D78822171 2025-11-03T16:58:01.2332955Z * [new branch] export-D78822351 -> origin/export-D78822351 2025-11-03T16:58:01.2335080Z * [new branch] export-D78822507 -> origin/export-D78822507 2025-11-03T16:58:01.2337366Z * [new branch] export-D78826994 -> origin/export-D78826994 2025-11-03T16:58:01.2339497Z * [new branch] export-D78894324 -> origin/export-D78894324 2025-11-03T16:58:01.2341763Z * [new branch] export-D78929245 -> origin/export-D78929245 2025-11-03T16:58:01.2344514Z * [new branch] export-D78934925 -> origin/export-D78934925 2025-11-03T16:58:01.2346394Z * [new branch] export-D78953203 -> origin/export-D78953203 2025-11-03T16:58:01.2348558Z * [new branch] export-D78953229 -> origin/export-D78953229 2025-11-03T16:58:01.2350611Z * [new branch] export-D78957093 -> origin/export-D78957093 2025-11-03T16:58:01.2352810Z * [new branch] export-D78957389 -> origin/export-D78957389 2025-11-03T16:58:01.2355079Z * [new branch] export-D78996107 -> origin/export-D78996107 2025-11-03T16:58:01.2357092Z * [new branch] export-D79026433 -> origin/export-D79026433 2025-11-03T16:58:01.2359268Z * [new branch] export-D79319835 -> origin/export-D79319835 2025-11-03T16:58:01.2361364Z * [new branch] export-D79328456 -> origin/export-D79328456 2025-11-03T16:58:01.2363548Z * [new branch] export-D80823877 -> origin/export-D80823877 2025-11-03T16:58:01.2365880Z * [new branch] export-D80958642 -> origin/export-D80958642 2025-11-03T16:58:01.2368001Z * [new branch] export-D81054193 -> origin/export-D81054193 2025-11-03T16:58:01.2370145Z * [new branch] export-D81204584 -> origin/export-D81204584 2025-11-03T16:58:01.2372367Z * [new branch] export-D81429090 -> origin/export-D81429090 2025-11-03T16:58:01.2374565Z * [new branch] export-D81698719 -> origin/export-D81698719 2025-11-03T16:58:01.2376837Z * [new branch] export-D82250826 -> origin/export-D82250826 2025-11-03T16:58:01.2378957Z * [new branch] export-D82253817 -> origin/export-D82253817 2025-11-03T16:58:01.2381181Z * [new branch] export-D83390563 -> origin/export-D83390563 2025-11-03T16:58:01.2383523Z * [new branch] export-D83391942 -> origin/export-D83391942 2025-11-03T16:58:01.2385633Z * [new branch] export-D83395610 -> origin/export-D83395610 2025-11-03T16:58:01.2387746Z * [new branch] export-D83541846 -> origin/export-D83541846 2025-11-03T16:58:01.2390004Z * [new branch] export-D83591083 -> origin/export-D83591083 2025-11-03T16:58:01.2392927Z * [new branch] export-D83609850 -> origin/export-D83609850 2025-11-03T16:58:01.2395080Z * [new branch] export-D83627170 -> origin/export-D83627170 2025-11-03T16:58:01.2397217Z * [new branch] export-D83714690 -> origin/export-D83714690 2025-11-03T16:58:01.2399418Z * [new branch] export-D83766701 -> origin/export-D83766701 2025-11-03T16:58:01.2401545Z * [new branch] export-D83768878 -> origin/export-D83768878 2025-11-03T16:58:01.2403687Z * [new branch] export-D83769447 -> origin/export-D83769447 2025-11-03T16:58:01.2405870Z * [new branch] export-D84009392 -> origin/export-D84009392 2025-11-03T16:58:01.2408247Z * [new branch] export-D84025878 -> origin/export-D84025878 2025-11-03T16:58:01.2410378Z * [new branch] export-D84089824 -> origin/export-D84089824 2025-11-03T16:58:01.2412579Z * [new branch] export-D84098898 -> origin/export-D84098898 2025-11-03T16:58:01.2414937Z * [new branch] export-D84103213 -> origin/export-D84103213 2025-11-03T16:58:01.2417204Z * [new branch] export-D84169910 -> origin/export-D84169910 2025-11-03T16:58:01.2419329Z * [new branch] export-D84213020 -> origin/export-D84213020 2025-11-03T16:58:01.2421869Z * [new branch] export-D84373821 -> origin/export-D84373821 2025-11-03T16:58:01.2424875Z * [new branch] export-D84522373 -> origin/export-D84522373 2025-11-03T16:58:01.2426857Z * [new branch] export-D84612194 -> origin/export-D84612194 2025-11-03T16:58:01.2428591Z * [new branch] export-D84890985 -> origin/export-D84890985 2025-11-03T16:58:01.2430349Z * [new branch] export-D85122326 -> origin/export-D85122326 2025-11-03T16:58:01.2432410Z * [new branch] export-D85745809 -> origin/export-D85745809 2025-11-03T16:58:01.2434362Z * [new branch] exported-model-train-idempotent -> origin/exported-model-train-idempotent 2025-11-03T16:58:01.2436167Z * [new branch] ezyang-titan-october -> origin/ezyang-titan-october 2025-11-03T16:58:01.2438004Z * [new branch] ezyang-titan-october2 -> origin/ezyang-titan-october2 2025-11-03T16:58:01.2439741Z * [new branch] ezyang-war -> origin/ezyang-war 2025-11-03T16:58:01.2442126Z * [new branch] ezyang/wip-aot-descriptors -> origin/ezyang/wip-aot-descriptors 2025-11-03T16:58:01.2443747Z * [new branch] fa_u8_brgemm -> origin/fa_u8_brgemm 2025-11-03T16:58:01.2445803Z * [new branch] fadeputr-fix-fbgemm_genai-build -> origin/fadeputr-fix-fbgemm_genai-build 2025-11-03T16:58:01.2448157Z * [new branch] fadeputr/sequence_fbgemm -> origin/fadeputr/sequence_fbgemm 2025-11-03T16:58:01.2449936Z * [new branch] fastmath_baseline -> origin/fastmath_baseline 2025-11-03T16:58:01.2452445Z * [new branch] fbcode/warm -> origin/fbcode/warm 2025-11-03T16:58:01.2454336Z * [new branch] fca -> origin/fca 2025-11-03T16:58:01.2456091Z * [new branch] fca2_ca5984c -> origin/fca2_ca5984c 2025-11-03T16:58:01.2457836Z * [new branch] fca5 -> origin/fca5 2025-11-03T16:58:01.2460292Z * [new branch] feature/forkserver-numa -> origin/feature/forkserver-numa 2025-11-03T16:58:01.2462106Z * [new branch] feature/justknobs-cpp -> origin/feature/justknobs-cpp 2025-11-03T16:58:01.2463968Z * [new branch] feature/numa-forkserver -> origin/feature/numa-forkserver 2025-11-03T16:58:01.2466012Z * [new branch] ffast_math_baseline -> origin/ffast_math_baseline 2025-11-03T16:58:01.2468423Z * [new branch] ffast_math_target -> origin/ffast_math_target 2025-11-03T16:58:01.2471007Z * [new branch] findhao/base_commit -> origin/findhao/base_commit 2025-11-03T16:58:01.2472661Z * [new branch] findhao/base_commit1 -> origin/findhao/base_commit1 2025-11-03T16:58:01.2474332Z * [new branch] findhao/multistream2 -> origin/findhao/multistream2 2025-11-03T16:58:01.2476119Z * [new branch] findhao/multistream5 -> origin/findhao/multistream5 2025-11-03T16:58:01.2477769Z * [new branch] findhao/multistream6 -> origin/findhao/multistream6 2025-11-03T16:58:01.2479408Z * [new branch] findhao/operatorbench3 -> origin/findhao/operatorbench3 2025-11-03T16:58:01.2481281Z * [new branch] findhao/operatorbench5 -> origin/findhao/operatorbench5 2025-11-03T16:58:01.2482973Z * [new branch] findhao/tritonparse -> origin/findhao/tritonparse 2025-11-03T16:58:01.2484837Z * [new branch] fix-ck-gemm-template-format -> origin/fix-ck-gemm-template-format 2025-11-03T16:58:01.2486614Z * [new branch] fix-config-ignore -> origin/fix-config-ignore 2025-11-03T16:58:01.2488557Z * [new branch] fix-dict-guard -> origin/fix-dict-guard 2025-11-03T16:58:01.2490461Z * [new branch] fix-rlease-feature-template -> origin/fix-rlease-feature-template 2025-11-03T16:58:01.2492252Z * [new branch] fix_addmm_issue -> origin/fix_addmm_issue 2025-11-03T16:58:01.2494271Z * [new branch] fix_fx_graph_print_space -> origin/fix_fx_graph_print_space 2025-11-03T16:58:01.2496072Z * [new branch] fix_nvrtc_discovery -> origin/fix_nvrtc_discovery 2025-11-03T16:58:01.2497819Z * [new branch] fix_op_runner -> origin/fix_op_runner 2025-11-03T16:58:01.2499558Z * [new branch] fix_ubn_159469 -> origin/fix_ubn_159469 2025-11-03T16:58:01.2501307Z * [new branch] fixbugh100 -> origin/fixbugh100 2025-11-03T16:58:01.2503357Z * [new branch] fixes-triage -> origin/fixes-triage 2025-11-03T16:58:01.2505130Z * [new branch] fixflashinfer -> origin/fixflashinfer 2025-11-03T16:58:01.2506832Z * [new branch] fixvllmoct -> origin/fixvllmoct 2025-11-03T16:58:01.2508682Z * [new branch] flash_decoding_cpu -> origin/flash_decoding_cpu 2025-11-03T16:58:01.2510493Z * [new branch] flex-flash -> origin/flex-flash 2025-11-03T16:58:01.2512487Z * [new branch] flex_attention_functorch_grad -> origin/flex_attention_functorch_grad 2025-11-03T16:58:01.2514182Z * [new branch] flex_flash -> origin/flex_flash 2025-11-03T16:58:01.2516760Z * [new branch] fmassa/fix_memeff_sharding_rule -> origin/fmassa/fix_memeff_sharding_rule 2025-11-03T16:58:01.2518484Z * [new branch] forkserver_fix -> origin/forkserver_fix 2025-11-03T16:58:01.2520352Z * [new branch] fsdp2_trace_rules -> origin/fsdp2_trace_rules 2025-11-03T16:58:01.2522459Z * [new branch] fx_cpp -> origin/fx_cpp 2025-11-03T16:58:01.2524884Z * [new branch] fy/fix-win -> origin/fy/fix-win 2025-11-03T16:58:01.2527399Z * [new branch] georgehong/cmakelists-patch -> origin/georgehong/cmakelists-patch 2025-11-03T16:58:01.2530939Z * [new branch] gh/AlnisM/1/base -> origin/gh/AlnisM/1/base 2025-11-03T16:58:01.2532739Z * [new branch] gh/AlnisM/1/head -> origin/gh/AlnisM/1/head 2025-11-03T16:58:01.2535770Z * [new branch] gh/ColinPeppler/94/base -> origin/gh/ColinPeppler/94/base 2025-11-03T16:58:01.2537933Z * [new branch] gh/ColinPeppler/94/head -> origin/gh/ColinPeppler/94/head 2025-11-03T16:58:01.2539362Z * [new branch] gh/ColinPeppler/94/orig -> origin/gh/ColinPeppler/94/orig 2025-11-03T16:58:01.2542109Z * [new branch] gh/ColinPeppler/95/base -> origin/gh/ColinPeppler/95/base 2025-11-03T16:58:01.2543772Z * [new branch] gh/ColinPeppler/95/head -> origin/gh/ColinPeppler/95/head 2025-11-03T16:58:01.2545679Z * [new branch] gh/ColinPeppler/95/orig -> origin/gh/ColinPeppler/95/orig 2025-11-03T16:58:01.2548438Z * [new branch] gh/EikanWang/67/base -> origin/gh/EikanWang/67/base 2025-11-03T16:58:01.2550171Z * [new branch] gh/EikanWang/67/head -> origin/gh/EikanWang/67/head 2025-11-03T16:58:01.2552524Z * [new branch] gh/EikanWang/68/base -> origin/gh/EikanWang/68/base 2025-11-03T16:58:01.2554299Z * [new branch] gh/EikanWang/68/head -> origin/gh/EikanWang/68/head 2025-11-03T16:58:01.2556040Z * [new branch] gh/EikanWang/68/orig -> origin/gh/EikanWang/68/orig 2025-11-03T16:58:01.2559139Z * [new branch] gh/Gasoonjia/1/base -> origin/gh/Gasoonjia/1/base 2025-11-03T16:58:01.2560891Z * [new branch] gh/Gasoonjia/1/head -> origin/gh/Gasoonjia/1/head 2025-11-03T16:58:01.2563777Z * [new branch] gh/H-Huang/131/base -> origin/gh/H-Huang/131/base 2025-11-03T16:58:01.2565556Z * [new branch] gh/H-Huang/131/head -> origin/gh/H-Huang/131/head 2025-11-03T16:58:01.2567248Z * [new branch] gh/H-Huang/131/orig -> origin/gh/H-Huang/131/orig 2025-11-03T16:58:01.2569667Z * [new branch] gh/H-Huang/132/base -> origin/gh/H-Huang/132/base 2025-11-03T16:58:01.2571385Z * [new branch] gh/H-Huang/132/head -> origin/gh/H-Huang/132/head 2025-11-03T16:58:01.2573070Z * [new branch] gh/H-Huang/132/orig -> origin/gh/H-Huang/132/orig 2025-11-03T16:58:01.2575490Z * [new branch] gh/H-Huang/180/base -> origin/gh/H-Huang/180/base 2025-11-03T16:58:01.2577218Z * [new branch] gh/H-Huang/180/head -> origin/gh/H-Huang/180/head 2025-11-03T16:58:01.2578923Z * [new branch] gh/H-Huang/180/orig -> origin/gh/H-Huang/180/orig 2025-11-03T16:58:01.2581151Z * [new branch] gh/H-Huang/182/base -> origin/gh/H-Huang/182/base 2025-11-03T16:58:01.2583092Z * [new branch] gh/H-Huang/182/head -> origin/gh/H-Huang/182/head 2025-11-03T16:58:01.2584770Z * [new branch] gh/H-Huang/182/orig -> origin/gh/H-Huang/182/orig 2025-11-03T16:58:01.2587126Z * [new branch] gh/H-Huang/212/base -> origin/gh/H-Huang/212/base 2025-11-03T16:58:01.2588919Z * [new branch] gh/H-Huang/212/head -> origin/gh/H-Huang/212/head 2025-11-03T16:58:01.2590664Z * [new branch] gh/H-Huang/212/orig -> origin/gh/H-Huang/212/orig 2025-11-03T16:58:01.2593246Z * [new branch] gh/H-Huang/214/base -> origin/gh/H-Huang/214/base 2025-11-03T16:58:01.2594877Z * [new branch] gh/H-Huang/214/head -> origin/gh/H-Huang/214/head 2025-11-03T16:58:01.2596620Z * [new branch] gh/H-Huang/214/orig -> origin/gh/H-Huang/214/orig 2025-11-03T16:58:01.2599009Z * [new branch] gh/H-Huang/216/base -> origin/gh/H-Huang/216/base 2025-11-03T16:58:01.2600779Z * [new branch] gh/H-Huang/216/head -> origin/gh/H-Huang/216/head 2025-11-03T16:58:01.2602517Z * [new branch] gh/H-Huang/216/orig -> origin/gh/H-Huang/216/orig 2025-11-03T16:58:01.2604836Z * [new branch] gh/H-Huang/218/base -> origin/gh/H-Huang/218/base 2025-11-03T16:58:01.2606608Z * [new branch] gh/H-Huang/218/head -> origin/gh/H-Huang/218/head 2025-11-03T16:58:01.2608308Z * [new branch] gh/H-Huang/218/orig -> origin/gh/H-Huang/218/orig 2025-11-03T16:58:01.2610634Z * [new branch] gh/H-Huang/219/base -> origin/gh/H-Huang/219/base 2025-11-03T16:58:01.2612382Z * [new branch] gh/H-Huang/219/head -> origin/gh/H-Huang/219/head 2025-11-03T16:58:01.2614563Z * [new branch] gh/H-Huang/219/orig -> origin/gh/H-Huang/219/orig 2025-11-03T16:58:01.2617035Z * [new branch] gh/H-Huang/220/base -> origin/gh/H-Huang/220/base 2025-11-03T16:58:01.2618684Z * [new branch] gh/H-Huang/220/head -> origin/gh/H-Huang/220/head 2025-11-03T16:58:01.2620367Z * [new branch] gh/H-Huang/220/orig -> origin/gh/H-Huang/220/orig 2025-11-03T16:58:01.2623119Z * [new branch] gh/H-Huang/221/base -> origin/gh/H-Huang/221/base 2025-11-03T16:58:01.2624837Z * [new branch] gh/H-Huang/221/head -> origin/gh/H-Huang/221/head 2025-11-03T16:58:01.2626595Z * [new branch] gh/H-Huang/221/orig -> origin/gh/H-Huang/221/orig 2025-11-03T16:58:01.2628928Z * [new branch] gh/H-Huang/222/base -> origin/gh/H-Huang/222/base 2025-11-03T16:58:01.2630621Z * [new branch] gh/H-Huang/222/head -> origin/gh/H-Huang/222/head 2025-11-03T16:58:01.2632339Z * [new branch] gh/H-Huang/222/orig -> origin/gh/H-Huang/222/orig 2025-11-03T16:58:01.2634787Z * [new branch] gh/H-Huang/223/base -> origin/gh/H-Huang/223/base 2025-11-03T16:58:01.2636755Z * [new branch] gh/H-Huang/223/head -> origin/gh/H-Huang/223/head 2025-11-03T16:58:01.2638432Z * [new branch] gh/H-Huang/223/orig -> origin/gh/H-Huang/223/orig 2025-11-03T16:58:01.2640815Z * [new branch] gh/H-Huang/224/base -> origin/gh/H-Huang/224/base 2025-11-03T16:58:01.2642521Z * [new branch] gh/H-Huang/224/head -> origin/gh/H-Huang/224/head 2025-11-03T16:58:01.2644289Z * [new branch] gh/H-Huang/224/orig -> origin/gh/H-Huang/224/orig 2025-11-03T16:58:01.2646631Z * [new branch] gh/H-Huang/225/base -> origin/gh/H-Huang/225/base 2025-11-03T16:58:01.2648356Z * [new branch] gh/H-Huang/225/head -> origin/gh/H-Huang/225/head 2025-11-03T16:58:01.2650049Z * [new branch] gh/H-Huang/225/orig -> origin/gh/H-Huang/225/orig 2025-11-03T16:58:01.2652399Z * [new branch] gh/H-Huang/226/base -> origin/gh/H-Huang/226/base 2025-11-03T16:58:01.2654133Z * [new branch] gh/H-Huang/226/head -> origin/gh/H-Huang/226/head 2025-11-03T16:58:01.2655933Z * [new branch] gh/H-Huang/226/orig -> origin/gh/H-Huang/226/orig 2025-11-03T16:58:01.2658883Z * [new branch] gh/H-Huang/227/base -> origin/gh/H-Huang/227/base 2025-11-03T16:58:01.2660603Z * [new branch] gh/H-Huang/227/head -> origin/gh/H-Huang/227/head 2025-11-03T16:58:01.2662429Z * [new branch] gh/H-Huang/227/orig -> origin/gh/H-Huang/227/orig 2025-11-03T16:58:01.2665006Z * [new branch] gh/H-Huang/228/base -> origin/gh/H-Huang/228/base 2025-11-03T16:58:01.2666758Z * [new branch] gh/H-Huang/228/head -> origin/gh/H-Huang/228/head 2025-11-03T16:58:01.2668540Z * [new branch] gh/H-Huang/228/orig -> origin/gh/H-Huang/228/orig 2025-11-03T16:58:01.2671437Z * [new branch] gh/IvanKobzarev/150/base -> origin/gh/IvanKobzarev/150/base 2025-11-03T16:58:01.2673204Z * [new branch] gh/IvanKobzarev/150/head -> origin/gh/IvanKobzarev/150/head 2025-11-03T16:58:01.2674960Z * [new branch] gh/IvanKobzarev/150/orig -> origin/gh/IvanKobzarev/150/orig 2025-11-03T16:58:01.2677360Z * [new branch] gh/IvanKobzarev/151/base -> origin/gh/IvanKobzarev/151/base 2025-11-03T16:58:01.2679148Z * [new branch] gh/IvanKobzarev/151/head -> origin/gh/IvanKobzarev/151/head 2025-11-03T16:58:01.2680865Z * [new branch] gh/IvanKobzarev/151/orig -> origin/gh/IvanKobzarev/151/orig 2025-11-03T16:58:01.2683391Z * [new branch] gh/IvanKobzarev/156/base -> origin/gh/IvanKobzarev/156/base 2025-11-03T16:58:01.2685119Z * [new branch] gh/IvanKobzarev/156/head -> origin/gh/IvanKobzarev/156/head 2025-11-03T16:58:01.2687073Z * [new branch] gh/IvanKobzarev/156/orig -> origin/gh/IvanKobzarev/156/orig 2025-11-03T16:58:01.2689339Z * [new branch] gh/IvanKobzarev/157/base -> origin/gh/IvanKobzarev/157/base 2025-11-03T16:58:01.2691136Z * [new branch] gh/IvanKobzarev/157/head -> origin/gh/IvanKobzarev/157/head 2025-11-03T16:58:01.2692910Z * [new branch] gh/IvanKobzarev/157/orig -> origin/gh/IvanKobzarev/157/orig 2025-11-03T16:58:01.2695417Z * [new branch] gh/IvanKobzarev/158/base -> origin/gh/IvanKobzarev/158/base 2025-11-03T16:58:01.2697143Z * [new branch] gh/IvanKobzarev/158/head -> origin/gh/IvanKobzarev/158/head 2025-11-03T16:58:01.2698854Z * [new branch] gh/IvanKobzarev/158/orig -> origin/gh/IvanKobzarev/158/orig 2025-11-03T16:58:01.2701246Z * [new branch] gh/IvanKobzarev/159/base -> origin/gh/IvanKobzarev/159/base 2025-11-03T16:58:01.2703166Z * [new branch] gh/IvanKobzarev/159/head -> origin/gh/IvanKobzarev/159/head 2025-11-03T16:58:01.2704865Z * [new branch] gh/IvanKobzarev/159/orig -> origin/gh/IvanKobzarev/159/orig 2025-11-03T16:58:01.2707251Z * [new branch] gh/IvanKobzarev/161/base -> origin/gh/IvanKobzarev/161/base 2025-11-03T16:58:01.2708984Z * [new branch] gh/IvanKobzarev/161/head -> origin/gh/IvanKobzarev/161/head 2025-11-03T16:58:01.2710741Z * [new branch] gh/IvanKobzarev/161/orig -> origin/gh/IvanKobzarev/161/orig 2025-11-03T16:58:01.2713100Z * [new branch] gh/IvanKobzarev/162/base -> origin/gh/IvanKobzarev/162/base 2025-11-03T16:58:01.2714917Z * [new branch] gh/IvanKobzarev/162/head -> origin/gh/IvanKobzarev/162/head 2025-11-03T16:58:01.2716685Z * [new branch] gh/IvanKobzarev/162/orig -> origin/gh/IvanKobzarev/162/orig 2025-11-03T16:58:01.2719184Z * [new branch] gh/IvanKobzarev/163/base -> origin/gh/IvanKobzarev/163/base 2025-11-03T16:58:01.2720955Z * [new branch] gh/IvanKobzarev/163/head -> origin/gh/IvanKobzarev/163/head 2025-11-03T16:58:01.2722935Z * [new branch] gh/IvanKobzarev/163/orig -> origin/gh/IvanKobzarev/163/orig 2025-11-03T16:58:01.2725546Z * [new branch] gh/IvanKobzarev/164/base -> origin/gh/IvanKobzarev/164/base 2025-11-03T16:58:01.2727180Z * [new branch] gh/IvanKobzarev/164/head -> origin/gh/IvanKobzarev/164/head 2025-11-03T16:58:01.2728890Z * [new branch] gh/IvanKobzarev/164/orig -> origin/gh/IvanKobzarev/164/orig 2025-11-03T16:58:01.2731291Z * [new branch] gh/IvanKobzarev/165/base -> origin/gh/IvanKobzarev/165/base 2025-11-03T16:58:01.2733021Z * [new branch] gh/IvanKobzarev/165/head -> origin/gh/IvanKobzarev/165/head 2025-11-03T16:58:01.2734743Z * [new branch] gh/IvanKobzarev/165/orig -> origin/gh/IvanKobzarev/165/orig 2025-11-03T16:58:01.2737078Z * [new branch] gh/IvanKobzarev/166/base -> origin/gh/IvanKobzarev/166/base 2025-11-03T16:58:01.2738855Z * [new branch] gh/IvanKobzarev/166/head -> origin/gh/IvanKobzarev/166/head 2025-11-03T16:58:01.2740531Z * [new branch] gh/IvanKobzarev/166/orig -> origin/gh/IvanKobzarev/166/orig 2025-11-03T16:58:01.2743983Z * [new branch] gh/NikhilAPatel/1/base -> origin/gh/NikhilAPatel/1/base 2025-11-03T16:58:01.2745776Z * [new branch] gh/NikhilAPatel/1/head -> origin/gh/NikhilAPatel/1/head 2025-11-03T16:58:01.2748057Z * [new branch] gh/NikhilAPatel/2/base -> origin/gh/NikhilAPatel/2/base 2025-11-03T16:58:01.2749842Z * [new branch] gh/NikhilAPatel/2/head -> origin/gh/NikhilAPatel/2/head 2025-11-03T16:58:01.2752959Z * [new branch] gh/NikhilAPatel/4/base -> origin/gh/NikhilAPatel/4/base 2025-11-03T16:58:01.2755232Z * [new branch] gh/NikhilAPatel/4/head -> origin/gh/NikhilAPatel/4/head 2025-11-03T16:58:01.2758005Z * [new branch] gh/PaliC/17/base -> origin/gh/PaliC/17/base 2025-11-03T16:58:01.2759597Z * [new branch] gh/PaliC/17/head -> origin/gh/PaliC/17/head 2025-11-03T16:58:01.2761327Z * [new branch] gh/PaliC/17/orig -> origin/gh/PaliC/17/orig 2025-11-03T16:58:01.2763723Z * [new branch] gh/PaliC/18/base -> origin/gh/PaliC/18/base 2025-11-03T16:58:01.2765464Z * [new branch] gh/PaliC/18/head -> origin/gh/PaliC/18/head 2025-11-03T16:58:01.2767189Z * [new branch] gh/PaliC/18/orig -> origin/gh/PaliC/18/orig 2025-11-03T16:58:01.2769628Z * [new branch] gh/PaliC/20/base -> origin/gh/PaliC/20/base 2025-11-03T16:58:01.2771356Z * [new branch] gh/PaliC/20/head -> origin/gh/PaliC/20/head 2025-11-03T16:58:01.2773465Z * [new branch] gh/PaliC/20/orig -> origin/gh/PaliC/20/orig 2025-11-03T16:58:01.2775830Z * [new branch] gh/PaliC/21/base -> origin/gh/PaliC/21/base 2025-11-03T16:58:01.2777577Z * [new branch] gh/PaliC/21/head -> origin/gh/PaliC/21/head 2025-11-03T16:58:01.2779362Z * [new branch] gh/PaliC/21/orig -> origin/gh/PaliC/21/orig 2025-11-03T16:58:01.2781756Z * [new branch] gh/PaliC/23/base -> origin/gh/PaliC/23/base 2025-11-03T16:58:01.2783580Z * [new branch] gh/PaliC/23/head -> origin/gh/PaliC/23/head 2025-11-03T16:58:01.2785286Z * [new branch] gh/PaliC/23/orig -> origin/gh/PaliC/23/orig 2025-11-03T16:58:01.2787695Z * [new branch] gh/PaliC/24/base -> origin/gh/PaliC/24/base 2025-11-03T16:58:01.2789439Z * [new branch] gh/PaliC/24/head -> origin/gh/PaliC/24/head 2025-11-03T16:58:01.2791156Z * [new branch] gh/PaliC/24/orig -> origin/gh/PaliC/24/orig 2025-11-03T16:58:01.2793649Z * [new branch] gh/PaliC/25/head -> origin/gh/PaliC/25/head 2025-11-03T16:58:01.2809615Z * [new branch] gh/PaliC/25/next -> origin/gh/PaliC/25/next 2025-11-03T16:58:01.2810237Z * [new branch] gh/PaliC/25/orig -> origin/gh/PaliC/25/orig 2025-11-03T16:58:01.2810632Z * [new branch] gh/PaliC/26/head -> origin/gh/PaliC/26/head 2025-11-03T16:58:01.2810999Z * [new branch] gh/PaliC/26/next -> origin/gh/PaliC/26/next 2025-11-03T16:58:01.2811366Z * [new branch] gh/PaliC/26/orig -> origin/gh/PaliC/26/orig 2025-11-03T16:58:01.2811730Z * [new branch] gh/PaliC/27/next -> origin/gh/PaliC/27/next 2025-11-03T16:58:01.2812081Z * [new branch] gh/PaliC/28/head -> origin/gh/PaliC/28/head 2025-11-03T16:58:01.2812430Z * [new branch] gh/PaliC/28/next -> origin/gh/PaliC/28/next 2025-11-03T16:58:01.2812787Z * [new branch] gh/PaliC/28/orig -> origin/gh/PaliC/28/orig 2025-11-03T16:58:01.2813181Z * [new branch] gh/PaliC/29/head -> origin/gh/PaliC/29/head 2025-11-03T16:58:01.2814977Z * [new branch] gh/PaliC/29/next -> origin/gh/PaliC/29/next 2025-11-03T16:58:01.2816691Z * [new branch] gh/PaliC/29/orig -> origin/gh/PaliC/29/orig 2025-11-03T16:58:01.2818932Z * [new branch] gh/PaliC/30/head -> origin/gh/PaliC/30/head 2025-11-03T16:58:01.2820571Z * [new branch] gh/PaliC/30/next -> origin/gh/PaliC/30/next 2025-11-03T16:58:01.2822717Z * [new branch] gh/PaliC/30/orig -> origin/gh/PaliC/30/orig 2025-11-03T16:58:01.2825070Z * [new branch] gh/PaliC/31/head -> origin/gh/PaliC/31/head 2025-11-03T16:58:01.2826668Z * [new branch] gh/PaliC/31/next -> origin/gh/PaliC/31/next 2025-11-03T16:58:01.2828396Z * [new branch] gh/PaliC/31/orig -> origin/gh/PaliC/31/orig 2025-11-03T16:58:01.2831499Z * [new branch] gh/PaulZhang12/22/base -> origin/gh/PaulZhang12/22/base 2025-11-03T16:58:01.2833145Z * [new branch] gh/PaulZhang12/22/head -> origin/gh/PaulZhang12/22/head 2025-11-03T16:58:01.2834831Z * [new branch] gh/PaulZhang12/22/orig -> origin/gh/PaulZhang12/22/orig 2025-11-03T16:58:01.2837200Z * [new branch] gh/PaulZhang12/25/base -> origin/gh/PaulZhang12/25/base 2025-11-03T16:58:01.2838976Z * [new branch] gh/PaulZhang12/25/head -> origin/gh/PaulZhang12/25/head 2025-11-03T16:58:01.2840696Z * [new branch] gh/PaulZhang12/25/orig -> origin/gh/PaulZhang12/25/orig 2025-11-03T16:58:01.2843224Z * [new branch] gh/PaulZhang12/26/base -> origin/gh/PaulZhang12/26/base 2025-11-03T16:58:01.2845102Z * [new branch] gh/PaulZhang12/26/head -> origin/gh/PaulZhang12/26/head 2025-11-03T16:58:01.2846798Z * [new branch] gh/PaulZhang12/26/orig -> origin/gh/PaulZhang12/26/orig 2025-11-03T16:58:01.2849353Z * [new branch] gh/PaulZhang12/27/base -> origin/gh/PaulZhang12/27/base 2025-11-03T16:58:01.2851139Z * [new branch] gh/PaulZhang12/27/head -> origin/gh/PaulZhang12/27/head 2025-11-03T16:58:01.2852906Z * [new branch] gh/PaulZhang12/27/orig -> origin/gh/PaulZhang12/27/orig 2025-11-03T16:58:01.2855277Z * [new branch] gh/PaulZhang12/28/base -> origin/gh/PaulZhang12/28/base 2025-11-03T16:58:01.2857068Z * [new branch] gh/PaulZhang12/28/head -> origin/gh/PaulZhang12/28/head 2025-11-03T16:58:01.2858789Z * [new branch] gh/PaulZhang12/28/orig -> origin/gh/PaulZhang12/28/orig 2025-11-03T16:58:01.2861279Z * [new branch] gh/PaulZhang12/30/base -> origin/gh/PaulZhang12/30/base 2025-11-03T16:58:01.2863160Z * [new branch] gh/PaulZhang12/30/head -> origin/gh/PaulZhang12/30/head 2025-11-03T16:58:01.2864821Z * [new branch] gh/PaulZhang12/30/orig -> origin/gh/PaulZhang12/30/orig 2025-11-03T16:58:01.2867331Z * [new branch] gh/PaulZhang12/31/base -> origin/gh/PaulZhang12/31/base 2025-11-03T16:58:01.2869144Z * [new branch] gh/PaulZhang12/31/head -> origin/gh/PaulZhang12/31/head 2025-11-03T16:58:01.2870854Z * [new branch] gh/PaulZhang12/31/orig -> origin/gh/PaulZhang12/31/orig 2025-11-03T16:58:01.2873355Z * [new branch] gh/PaulZhang12/32/base -> origin/gh/PaulZhang12/32/base 2025-11-03T16:58:01.2875036Z * [new branch] gh/PaulZhang12/32/head -> origin/gh/PaulZhang12/32/head 2025-11-03T16:58:01.2876743Z * [new branch] gh/PaulZhang12/32/orig -> origin/gh/PaulZhang12/32/orig 2025-11-03T16:58:01.2880248Z * [new branch] gh/PaulZhang12/33/base -> origin/gh/PaulZhang12/33/base 2025-11-03T16:58:01.2882485Z * [new branch] gh/PaulZhang12/33/head -> origin/gh/PaulZhang12/33/head 2025-11-03T16:58:01.2884292Z * [new branch] gh/PaulZhang12/33/orig -> origin/gh/PaulZhang12/33/orig 2025-11-03T16:58:01.2886655Z * [new branch] gh/PaulZhang12/34/base -> origin/gh/PaulZhang12/34/base 2025-11-03T16:58:01.2888366Z * [new branch] gh/PaulZhang12/34/head -> origin/gh/PaulZhang12/34/head 2025-11-03T16:58:01.2890207Z * [new branch] gh/PaulZhang12/34/orig -> origin/gh/PaulZhang12/34/orig 2025-11-03T16:58:01.2893994Z * [new branch] gh/PaulZhang12/35/base -> origin/gh/PaulZhang12/35/base 2025-11-03T16:58:01.2896825Z * [new branch] gh/PaulZhang12/35/head -> origin/gh/PaulZhang12/35/head 2025-11-03T16:58:01.2898567Z * [new branch] gh/PaulZhang12/35/orig -> origin/gh/PaulZhang12/35/orig 2025-11-03T16:58:01.2901039Z * [new branch] gh/PaulZhang12/36/base -> origin/gh/PaulZhang12/36/base 2025-11-03T16:58:01.2902922Z * [new branch] gh/PaulZhang12/36/head -> origin/gh/PaulZhang12/36/head 2025-11-03T16:58:01.2904527Z * [new branch] gh/PaulZhang12/36/orig -> origin/gh/PaulZhang12/36/orig 2025-11-03T16:58:01.2906955Z * [new branch] gh/PaulZhang12/37/base -> origin/gh/PaulZhang12/37/base 2025-11-03T16:58:01.2908611Z * [new branch] gh/PaulZhang12/37/head -> origin/gh/PaulZhang12/37/head 2025-11-03T16:58:01.2910407Z * [new branch] gh/PaulZhang12/37/orig -> origin/gh/PaulZhang12/37/orig 2025-11-03T16:58:01.2912735Z * [new branch] gh/PaulZhang12/38/base -> origin/gh/PaulZhang12/38/base 2025-11-03T16:58:01.2914568Z * [new branch] gh/PaulZhang12/38/head -> origin/gh/PaulZhang12/38/head 2025-11-03T16:58:01.2916206Z * [new branch] gh/PaulZhang12/38/orig -> origin/gh/PaulZhang12/38/orig 2025-11-03T16:58:01.2918595Z * [new branch] gh/PaulZhang12/39/base -> origin/gh/PaulZhang12/39/base 2025-11-03T16:58:01.2920399Z * [new branch] gh/PaulZhang12/39/head -> origin/gh/PaulZhang12/39/head 2025-11-03T16:58:01.2922762Z * [new branch] gh/PaulZhang12/39/orig -> origin/gh/PaulZhang12/39/orig 2025-11-03T16:58:01.2925102Z * [new branch] gh/PaulZhang12/40/base -> origin/gh/PaulZhang12/40/base 2025-11-03T16:58:01.2926823Z * [new branch] gh/PaulZhang12/40/head -> origin/gh/PaulZhang12/40/head 2025-11-03T16:58:01.2928541Z * [new branch] gh/PaulZhang12/40/orig -> origin/gh/PaulZhang12/40/orig 2025-11-03T16:58:01.2930924Z * [new branch] gh/PaulZhang12/41/base -> origin/gh/PaulZhang12/41/base 2025-11-03T16:58:01.2932722Z * [new branch] gh/PaulZhang12/41/head -> origin/gh/PaulZhang12/41/head 2025-11-03T16:58:01.2934454Z * [new branch] gh/PaulZhang12/41/orig -> origin/gh/PaulZhang12/41/orig 2025-11-03T16:58:01.2937270Z * [new branch] gh/SamGinzburg/11/base -> origin/gh/SamGinzburg/11/base 2025-11-03T16:58:01.2939045Z * [new branch] gh/SamGinzburg/11/head -> origin/gh/SamGinzburg/11/head 2025-11-03T16:58:01.2942461Z * [new branch] gh/SherlockNoMad/1/base -> origin/gh/SherlockNoMad/1/base 2025-11-03T16:58:01.2944185Z * [new branch] gh/SherlockNoMad/1/head -> origin/gh/SherlockNoMad/1/head 2025-11-03T16:58:01.2946610Z * [new branch] gh/SherlockNoMad/10/base -> origin/gh/SherlockNoMad/10/base 2025-11-03T16:58:01.2948363Z * [new branch] gh/SherlockNoMad/10/head -> origin/gh/SherlockNoMad/10/head 2025-11-03T16:58:01.2950136Z * [new branch] gh/SherlockNoMad/10/orig -> origin/gh/SherlockNoMad/10/orig 2025-11-03T16:58:01.2952399Z * [new branch] gh/SherlockNoMad/11/base -> origin/gh/SherlockNoMad/11/base 2025-11-03T16:58:01.2954163Z * [new branch] gh/SherlockNoMad/11/head -> origin/gh/SherlockNoMad/11/head 2025-11-03T16:58:01.2955892Z * [new branch] gh/SherlockNoMad/11/orig -> origin/gh/SherlockNoMad/11/orig 2025-11-03T16:58:01.2958084Z * [new branch] gh/SherlockNoMad/12/base -> origin/gh/SherlockNoMad/12/base 2025-11-03T16:58:01.2959851Z * [new branch] gh/SherlockNoMad/12/head -> origin/gh/SherlockNoMad/12/head 2025-11-03T16:58:01.2961554Z * [new branch] gh/SherlockNoMad/12/orig -> origin/gh/SherlockNoMad/12/orig 2025-11-03T16:58:01.2963790Z * [new branch] gh/SherlockNoMad/13/base -> origin/gh/SherlockNoMad/13/base 2025-11-03T16:58:01.2965628Z * [new branch] gh/SherlockNoMad/13/head -> origin/gh/SherlockNoMad/13/head 2025-11-03T16:58:01.2967427Z * [new branch] gh/SherlockNoMad/13/orig -> origin/gh/SherlockNoMad/13/orig 2025-11-03T16:58:01.2969967Z * [new branch] gh/SherlockNoMad/14/base -> origin/gh/SherlockNoMad/14/base 2025-11-03T16:58:01.2972164Z * [new branch] gh/SherlockNoMad/14/head -> origin/gh/SherlockNoMad/14/head 2025-11-03T16:58:01.2973948Z * [new branch] gh/SherlockNoMad/14/orig -> origin/gh/SherlockNoMad/14/orig 2025-11-03T16:58:01.2976332Z * [new branch] gh/SherlockNoMad/15/base -> origin/gh/SherlockNoMad/15/base 2025-11-03T16:58:01.2978165Z * [new branch] gh/SherlockNoMad/15/head -> origin/gh/SherlockNoMad/15/head 2025-11-03T16:58:01.2979942Z * [new branch] gh/SherlockNoMad/15/orig -> origin/gh/SherlockNoMad/15/orig 2025-11-03T16:58:01.2982496Z * [new branch] gh/SherlockNoMad/16/base -> origin/gh/SherlockNoMad/16/base 2025-11-03T16:58:01.2984232Z * [new branch] gh/SherlockNoMad/16/head -> origin/gh/SherlockNoMad/16/head 2025-11-03T16:58:01.2985948Z * [new branch] gh/SherlockNoMad/16/orig -> origin/gh/SherlockNoMad/16/orig 2025-11-03T16:58:01.2988175Z * [new branch] gh/SherlockNoMad/17/base -> origin/gh/SherlockNoMad/17/base 2025-11-03T16:58:01.2989985Z * [new branch] gh/SherlockNoMad/17/head -> origin/gh/SherlockNoMad/17/head 2025-11-03T16:58:01.2991971Z * [new branch] gh/SherlockNoMad/17/orig -> origin/gh/SherlockNoMad/17/orig 2025-11-03T16:58:01.2994385Z * [new branch] gh/SherlockNoMad/2/base -> origin/gh/SherlockNoMad/2/base 2025-11-03T16:58:01.2996035Z * [new branch] gh/SherlockNoMad/2/head -> origin/gh/SherlockNoMad/2/head 2025-11-03T16:58:01.2998340Z * [new branch] gh/SherlockNoMad/3/base -> origin/gh/SherlockNoMad/3/base 2025-11-03T16:58:01.3000003Z * [new branch] gh/SherlockNoMad/3/head -> origin/gh/SherlockNoMad/3/head 2025-11-03T16:58:01.3002490Z * [new branch] gh/SherlockNoMad/4/base -> origin/gh/SherlockNoMad/4/base 2025-11-03T16:58:01.3004196Z * [new branch] gh/SherlockNoMad/4/head -> origin/gh/SherlockNoMad/4/head 2025-11-03T16:58:01.3006475Z * [new branch] gh/SherlockNoMad/5/base -> origin/gh/SherlockNoMad/5/base 2025-11-03T16:58:01.3008105Z * [new branch] gh/SherlockNoMad/5/head -> origin/gh/SherlockNoMad/5/head 2025-11-03T16:58:01.3011107Z * [new branch] gh/Sidharth123-cpu/24/base -> origin/gh/Sidharth123-cpu/24/base 2025-11-03T16:58:01.3013322Z * [new branch] gh/Sidharth123-cpu/25/base -> origin/gh/Sidharth123-cpu/25/base 2025-11-03T16:58:01.3015517Z * [new branch] gh/Sidharth123-cpu/26/base -> origin/gh/Sidharth123-cpu/26/base 2025-11-03T16:58:01.3018020Z * [new branch] gh/Sidharth123-cpu/27/base -> origin/gh/Sidharth123-cpu/27/base 2025-11-03T16:58:01.3020873Z * [new branch] gh/StrongerXi/1/base -> origin/gh/StrongerXi/1/base 2025-11-03T16:58:01.3023053Z * [new branch] gh/StrongerXi/1/head -> origin/gh/StrongerXi/1/head 2025-11-03T16:58:01.3025360Z * [new branch] gh/StrongerXi/136/base -> origin/gh/StrongerXi/136/base 2025-11-03T16:58:01.3027066Z * [new branch] gh/StrongerXi/136/head -> origin/gh/StrongerXi/136/head 2025-11-03T16:58:01.3028820Z * [new branch] gh/StrongerXi/136/orig -> origin/gh/StrongerXi/136/orig 2025-11-03T16:58:01.3031157Z * [new branch] gh/StrongerXi/137/base -> origin/gh/StrongerXi/137/base 2025-11-03T16:58:01.3032902Z * [new branch] gh/StrongerXi/137/head -> origin/gh/StrongerXi/137/head 2025-11-03T16:58:01.3034668Z * [new branch] gh/StrongerXi/137/orig -> origin/gh/StrongerXi/137/orig 2025-11-03T16:58:01.3036948Z * [new branch] gh/StrongerXi/138/base -> origin/gh/StrongerXi/138/base 2025-11-03T16:58:01.3038734Z * [new branch] gh/StrongerXi/138/head -> origin/gh/StrongerXi/138/head 2025-11-03T16:58:01.3040545Z * [new branch] gh/StrongerXi/138/orig -> origin/gh/StrongerXi/138/orig 2025-11-03T16:58:01.3043410Z * [new branch] gh/StrongerXi/71/base -> origin/gh/StrongerXi/71/base 2025-11-03T16:58:01.3044818Z * [new branch] gh/StrongerXi/71/head -> origin/gh/StrongerXi/71/head 2025-11-03T16:58:01.3047071Z * [new branch] gh/StrongerXi/72/base -> origin/gh/StrongerXi/72/base 2025-11-03T16:58:01.3048806Z * [new branch] gh/StrongerXi/72/head -> origin/gh/StrongerXi/72/head 2025-11-03T16:58:01.3051702Z * [new branch] gh/XilunWu/151/base -> origin/gh/XilunWu/151/base 2025-11-03T16:58:01.3053496Z * [new branch] gh/XilunWu/151/head -> origin/gh/XilunWu/151/head 2025-11-03T16:58:01.3055270Z * [new branch] gh/XilunWu/151/orig -> origin/gh/XilunWu/151/orig 2025-11-03T16:58:01.3057719Z * [new branch] gh/XilunWu/152/base -> origin/gh/XilunWu/152/base 2025-11-03T16:58:01.3059448Z * [new branch] gh/XilunWu/152/head -> origin/gh/XilunWu/152/head 2025-11-03T16:58:01.3061178Z * [new branch] gh/XilunWu/152/orig -> origin/gh/XilunWu/152/orig 2025-11-03T16:58:01.3063709Z * [new branch] gh/XilunWu/160/base -> origin/gh/XilunWu/160/base 2025-11-03T16:58:01.3065304Z * [new branch] gh/XilunWu/160/head -> origin/gh/XilunWu/160/head 2025-11-03T16:58:01.3066987Z * [new branch] gh/XilunWu/160/orig -> origin/gh/XilunWu/160/orig 2025-11-03T16:58:01.3069451Z * [new branch] gh/XilunWu/163/base -> origin/gh/XilunWu/163/base 2025-11-03T16:58:01.3071119Z * [new branch] gh/XilunWu/163/head -> origin/gh/XilunWu/163/head 2025-11-03T16:58:01.3072830Z * [new branch] gh/XilunWu/163/orig -> origin/gh/XilunWu/163/orig 2025-11-03T16:58:01.3075508Z * [new branch] gh/XilunWu/166/base -> origin/gh/XilunWu/166/base 2025-11-03T16:58:01.3077231Z * [new branch] gh/XilunWu/166/head -> origin/gh/XilunWu/166/head 2025-11-03T16:58:01.3078962Z * [new branch] gh/XilunWu/166/orig -> origin/gh/XilunWu/166/orig 2025-11-03T16:58:01.3081601Z * [new branch] gh/XilunWu/168/base -> origin/gh/XilunWu/168/base 2025-11-03T16:58:01.3083537Z * [new branch] gh/XilunWu/168/head -> origin/gh/XilunWu/168/head 2025-11-03T16:58:01.3085340Z * [new branch] gh/XilunWu/168/orig -> origin/gh/XilunWu/168/orig 2025-11-03T16:58:01.3087672Z * [new branch] gh/XilunWu/169/base -> origin/gh/XilunWu/169/base 2025-11-03T16:58:01.3089409Z * [new branch] gh/XilunWu/169/head -> origin/gh/XilunWu/169/head 2025-11-03T16:58:01.3091084Z * [new branch] gh/XilunWu/169/orig -> origin/gh/XilunWu/169/orig 2025-11-03T16:58:01.3093218Z * [new branch] gh/XilunWu/170/base -> origin/gh/XilunWu/170/base 2025-11-03T16:58:01.3094942Z * [new branch] gh/XilunWu/170/head -> origin/gh/XilunWu/170/head 2025-11-03T16:58:01.3096688Z * [new branch] gh/XilunWu/170/orig -> origin/gh/XilunWu/170/orig 2025-11-03T16:58:01.3099143Z * [new branch] gh/XilunWu/171/base -> origin/gh/XilunWu/171/base 2025-11-03T16:58:01.3100843Z * [new branch] gh/XilunWu/171/head -> origin/gh/XilunWu/171/head 2025-11-03T16:58:01.3102928Z * [new branch] gh/XilunWu/171/orig -> origin/gh/XilunWu/171/orig 2025-11-03T16:58:01.3105304Z * [new branch] gh/XilunWu/172/base -> origin/gh/XilunWu/172/base 2025-11-03T16:58:01.3107162Z * [new branch] gh/XilunWu/172/head -> origin/gh/XilunWu/172/head 2025-11-03T16:58:01.3108974Z * [new branch] gh/XilunWu/172/orig -> origin/gh/XilunWu/172/orig 2025-11-03T16:58:01.3111244Z * [new branch] gh/XilunWu/173/base -> origin/gh/XilunWu/173/base 2025-11-03T16:58:01.3113030Z * [new branch] gh/XilunWu/173/head -> origin/gh/XilunWu/173/head 2025-11-03T16:58:01.3114874Z * [new branch] gh/XilunWu/173/orig -> origin/gh/XilunWu/173/orig 2025-11-03T16:58:01.3117248Z * [new branch] gh/XilunWu/175/base -> origin/gh/XilunWu/175/base 2025-11-03T16:58:01.3118890Z * [new branch] gh/XilunWu/175/head -> origin/gh/XilunWu/175/head 2025-11-03T16:58:01.3120631Z * [new branch] gh/XilunWu/175/orig -> origin/gh/XilunWu/175/orig 2025-11-03T16:58:01.3123513Z * [new branch] gh/XilunWu/176/base -> origin/gh/XilunWu/176/base 2025-11-03T16:58:01.3125176Z * [new branch] gh/XilunWu/176/head -> origin/gh/XilunWu/176/head 2025-11-03T16:58:01.3127011Z * [new branch] gh/XilunWu/176/orig -> origin/gh/XilunWu/176/orig 2025-11-03T16:58:01.3129388Z * [new branch] gh/XilunWu/177/base -> origin/gh/XilunWu/177/base 2025-11-03T16:58:01.3131173Z * [new branch] gh/XilunWu/177/head -> origin/gh/XilunWu/177/head 2025-11-03T16:58:01.3132995Z * [new branch] gh/XilunWu/177/orig -> origin/gh/XilunWu/177/orig 2025-11-03T16:58:01.3135872Z * [new branch] gh/XuehaiPan/14/base -> origin/gh/XuehaiPan/14/base 2025-11-03T16:58:01.3137604Z * [new branch] gh/XuehaiPan/14/head -> origin/gh/XuehaiPan/14/head 2025-11-03T16:58:01.3139303Z * [new branch] gh/XuehaiPan/14/orig -> origin/gh/XuehaiPan/14/orig 2025-11-03T16:58:01.3141982Z * [new branch] gh/XuehaiPan/179/base -> origin/gh/XuehaiPan/179/base 2025-11-03T16:58:01.3143670Z * [new branch] gh/XuehaiPan/179/head -> origin/gh/XuehaiPan/179/head 2025-11-03T16:58:01.3145479Z * [new branch] gh/XuehaiPan/179/orig -> origin/gh/XuehaiPan/179/orig 2025-11-03T16:58:01.3147844Z * [new branch] gh/XuehaiPan/189/base -> origin/gh/XuehaiPan/189/base 2025-11-03T16:58:01.3149537Z * [new branch] gh/XuehaiPan/189/head -> origin/gh/XuehaiPan/189/head 2025-11-03T16:58:01.3151239Z * [new branch] gh/XuehaiPan/189/orig -> origin/gh/XuehaiPan/189/orig 2025-11-03T16:58:01.3153460Z * [new branch] gh/XuehaiPan/249/base -> origin/gh/XuehaiPan/249/base 2025-11-03T16:58:01.3155543Z * [new branch] gh/XuehaiPan/249/head -> origin/gh/XuehaiPan/249/head 2025-11-03T16:58:01.3157063Z * [new branch] gh/XuehaiPan/249/orig -> origin/gh/XuehaiPan/249/orig 2025-11-03T16:58:01.3159437Z * [new branch] gh/XuehaiPan/253/base -> origin/gh/XuehaiPan/253/base 2025-11-03T16:58:01.3161205Z * [new branch] gh/XuehaiPan/253/head -> origin/gh/XuehaiPan/253/head 2025-11-03T16:58:01.3162823Z * [new branch] gh/XuehaiPan/253/orig -> origin/gh/XuehaiPan/253/orig 2025-11-03T16:58:01.3165212Z * [new branch] gh/XuehaiPan/254/base -> origin/gh/XuehaiPan/254/base 2025-11-03T16:58:01.3167009Z * [new branch] gh/XuehaiPan/254/head -> origin/gh/XuehaiPan/254/head 2025-11-03T16:58:01.3168652Z * [new branch] gh/XuehaiPan/254/orig -> origin/gh/XuehaiPan/254/orig 2025-11-03T16:58:01.3170939Z * [new branch] gh/XuehaiPan/255/base -> origin/gh/XuehaiPan/255/base 2025-11-03T16:58:01.3172642Z * [new branch] gh/XuehaiPan/255/head -> origin/gh/XuehaiPan/255/head 2025-11-03T16:58:01.3174283Z * [new branch] gh/XuehaiPan/255/orig -> origin/gh/XuehaiPan/255/orig 2025-11-03T16:58:01.3176578Z * [new branch] gh/XuehaiPan/257/base -> origin/gh/XuehaiPan/257/base 2025-11-03T16:58:01.3178352Z * [new branch] gh/XuehaiPan/257/head -> origin/gh/XuehaiPan/257/head 2025-11-03T16:58:01.3180012Z * [new branch] gh/XuehaiPan/257/orig -> origin/gh/XuehaiPan/257/orig 2025-11-03T16:58:01.3182678Z * [new branch] gh/XuehaiPan/271/base -> origin/gh/XuehaiPan/271/base 2025-11-03T16:58:01.3184195Z * [new branch] gh/XuehaiPan/271/head -> origin/gh/XuehaiPan/271/head 2025-11-03T16:58:01.3185859Z * [new branch] gh/XuehaiPan/271/orig -> origin/gh/XuehaiPan/271/orig 2025-11-03T16:58:01.3188242Z * [new branch] gh/XuehaiPan/290/base -> origin/gh/XuehaiPan/290/base 2025-11-03T16:58:01.3190025Z * [new branch] gh/XuehaiPan/290/head -> origin/gh/XuehaiPan/290/head 2025-11-03T16:58:01.3191785Z * [new branch] gh/XuehaiPan/290/orig -> origin/gh/XuehaiPan/290/orig 2025-11-03T16:58:01.3193976Z * [new branch] gh/XuehaiPan/343/base -> origin/gh/XuehaiPan/343/base 2025-11-03T16:58:01.3195676Z * [new branch] gh/XuehaiPan/343/head -> origin/gh/XuehaiPan/343/head 2025-11-03T16:58:01.3197400Z * [new branch] gh/XuehaiPan/343/orig -> origin/gh/XuehaiPan/343/orig 2025-11-03T16:58:01.3199673Z * [new branch] gh/XuehaiPan/347/base -> origin/gh/XuehaiPan/347/base 2025-11-03T16:58:01.3201681Z * [new branch] gh/XuehaiPan/347/head -> origin/gh/XuehaiPan/347/head 2025-11-03T16:58:01.3203589Z * [new branch] gh/XuehaiPan/347/orig -> origin/gh/XuehaiPan/347/orig 2025-11-03T16:58:01.3205947Z * [new branch] gh/XuehaiPan/348/base -> origin/gh/XuehaiPan/348/base 2025-11-03T16:58:01.3207574Z * [new branch] gh/XuehaiPan/348/head -> origin/gh/XuehaiPan/348/head 2025-11-03T16:58:01.3209235Z * [new branch] gh/XuehaiPan/348/orig -> origin/gh/XuehaiPan/348/orig 2025-11-03T16:58:01.3211452Z * [new branch] gh/XuehaiPan/350/base -> origin/gh/XuehaiPan/350/base 2025-11-03T16:58:01.3213183Z * [new branch] gh/XuehaiPan/350/head -> origin/gh/XuehaiPan/350/head 2025-11-03T16:58:01.3214923Z * [new branch] gh/XuehaiPan/350/orig -> origin/gh/XuehaiPan/350/orig 2025-11-03T16:58:01.3217312Z * [new branch] gh/XuehaiPan/356/base -> origin/gh/XuehaiPan/356/base 2025-11-03T16:58:01.3219039Z * [new branch] gh/XuehaiPan/356/head -> origin/gh/XuehaiPan/356/head 2025-11-03T16:58:01.3220671Z * [new branch] gh/XuehaiPan/356/orig -> origin/gh/XuehaiPan/356/orig 2025-11-03T16:58:01.3223496Z * [new branch] gh/XuehaiPan/357/base -> origin/gh/XuehaiPan/357/base 2025-11-03T16:58:01.3225198Z * [new branch] gh/XuehaiPan/357/head -> origin/gh/XuehaiPan/357/head 2025-11-03T16:58:01.3226850Z * [new branch] gh/XuehaiPan/357/orig -> origin/gh/XuehaiPan/357/orig 2025-11-03T16:58:01.3229137Z * [new branch] gh/XuehaiPan/358/base -> origin/gh/XuehaiPan/358/base 2025-11-03T16:58:01.3230901Z * [new branch] gh/XuehaiPan/358/head -> origin/gh/XuehaiPan/358/head 2025-11-03T16:58:01.3232709Z * [new branch] gh/XuehaiPan/358/orig -> origin/gh/XuehaiPan/358/orig 2025-11-03T16:58:01.3235519Z * [new branch] gh/XuehaiPan/359/base -> origin/gh/XuehaiPan/359/base 2025-11-03T16:58:01.3237185Z * [new branch] gh/XuehaiPan/359/head -> origin/gh/XuehaiPan/359/head 2025-11-03T16:58:01.3238869Z * [new branch] gh/XuehaiPan/359/orig -> origin/gh/XuehaiPan/359/orig 2025-11-03T16:58:01.3241270Z * [new branch] gh/XuehaiPan/360/base -> origin/gh/XuehaiPan/360/base 2025-11-03T16:58:01.3243311Z * [new branch] gh/XuehaiPan/360/head -> origin/gh/XuehaiPan/360/head 2025-11-03T16:58:01.3244966Z * [new branch] gh/XuehaiPan/360/orig -> origin/gh/XuehaiPan/360/orig 2025-11-03T16:58:01.3247331Z * [new branch] gh/XuehaiPan/365/base -> origin/gh/XuehaiPan/365/base 2025-11-03T16:58:01.3249130Z * [new branch] gh/XuehaiPan/365/head -> origin/gh/XuehaiPan/365/head 2025-11-03T16:58:01.3250998Z * [new branch] gh/XuehaiPan/365/orig -> origin/gh/XuehaiPan/365/orig 2025-11-03T16:58:01.3253240Z * [new branch] gh/XuehaiPan/366/base -> origin/gh/XuehaiPan/366/base 2025-11-03T16:58:01.3254916Z * [new branch] gh/XuehaiPan/366/head -> origin/gh/XuehaiPan/366/head 2025-11-03T16:58:01.3257303Z * [new branch] gh/XuehaiPan/370/base -> origin/gh/XuehaiPan/370/base 2025-11-03T16:58:01.3258978Z * [new branch] gh/XuehaiPan/370/head -> origin/gh/XuehaiPan/370/head 2025-11-03T16:58:01.3260843Z * [new branch] gh/XuehaiPan/370/orig -> origin/gh/XuehaiPan/370/orig 2025-11-03T16:58:01.3263564Z * [new branch] gh/XuehaiPan/384/base -> origin/gh/XuehaiPan/384/base 2025-11-03T16:58:01.3265173Z * [new branch] gh/XuehaiPan/384/head -> origin/gh/XuehaiPan/384/head 2025-11-03T16:58:01.3266857Z * [new branch] gh/XuehaiPan/384/orig -> origin/gh/XuehaiPan/384/orig 2025-11-03T16:58:01.3269214Z * [new branch] gh/XuehaiPan/390/base -> origin/gh/XuehaiPan/390/base 2025-11-03T16:58:01.3271021Z * [new branch] gh/XuehaiPan/390/head -> origin/gh/XuehaiPan/390/head 2025-11-03T16:58:01.3272739Z * [new branch] gh/XuehaiPan/390/orig -> origin/gh/XuehaiPan/390/orig 2025-11-03T16:58:01.3275086Z * [new branch] gh/XuehaiPan/391/base -> origin/gh/XuehaiPan/391/base 2025-11-03T16:58:01.3276779Z * [new branch] gh/XuehaiPan/391/head -> origin/gh/XuehaiPan/391/head 2025-11-03T16:58:01.3278501Z * [new branch] gh/XuehaiPan/391/orig -> origin/gh/XuehaiPan/391/orig 2025-11-03T16:58:01.3280820Z * [new branch] gh/XuehaiPan/392/base -> origin/gh/XuehaiPan/392/base 2025-11-03T16:58:01.3282521Z * [new branch] gh/XuehaiPan/392/head -> origin/gh/XuehaiPan/392/head 2025-11-03T16:58:01.3284249Z * [new branch] gh/XuehaiPan/392/orig -> origin/gh/XuehaiPan/392/orig 2025-11-03T16:58:01.3286735Z * [new branch] gh/XuehaiPan/393/base -> origin/gh/XuehaiPan/393/base 2025-11-03T16:58:01.3288337Z * [new branch] gh/XuehaiPan/393/head -> origin/gh/XuehaiPan/393/head 2025-11-03T16:58:01.3290008Z * [new branch] gh/XuehaiPan/393/orig -> origin/gh/XuehaiPan/393/orig 2025-11-03T16:58:01.3292870Z * [new branch] gh/XuehaiPan/394/base -> origin/gh/XuehaiPan/394/base 2025-11-03T16:58:01.3294517Z * [new branch] gh/XuehaiPan/394/head -> origin/gh/XuehaiPan/394/head 2025-11-03T16:58:01.3296254Z * [new branch] gh/XuehaiPan/394/orig -> origin/gh/XuehaiPan/394/orig 2025-11-03T16:58:01.3298671Z * [new branch] gh/XuehaiPan/395/base -> origin/gh/XuehaiPan/395/base 2025-11-03T16:58:01.3300399Z * [new branch] gh/XuehaiPan/395/head -> origin/gh/XuehaiPan/395/head 2025-11-03T16:58:01.3302635Z * [new branch] gh/XuehaiPan/395/orig -> origin/gh/XuehaiPan/395/orig 2025-11-03T16:58:01.3305035Z * [new branch] gh/XuehaiPan/396/base -> origin/gh/XuehaiPan/396/base 2025-11-03T16:58:01.3306735Z * [new branch] gh/XuehaiPan/396/orig -> origin/gh/XuehaiPan/396/orig 2025-11-03T16:58:01.3309574Z * [new branch] gh/ZhiweiYan-96/39/base -> origin/gh/ZhiweiYan-96/39/base 2025-11-03T16:58:01.3311337Z * [new branch] gh/ZhiweiYan-96/39/head -> origin/gh/ZhiweiYan-96/39/head 2025-11-03T16:58:01.3313310Z * [new branch] gh/ZhiweiYan-96/39/orig -> origin/gh/ZhiweiYan-96/39/orig 2025-11-03T16:58:01.3315609Z * [new branch] gh/ZhiweiYan-96/44/base -> origin/gh/ZhiweiYan-96/44/base 2025-11-03T16:58:01.3317406Z * [new branch] gh/ZhiweiYan-96/44/head -> origin/gh/ZhiweiYan-96/44/head 2025-11-03T16:58:01.3319674Z * [new branch] gh/ZhiweiYan-96/45/base -> origin/gh/ZhiweiYan-96/45/base 2025-11-03T16:58:01.3321517Z * [new branch] gh/ZhiweiYan-96/45/head -> origin/gh/ZhiweiYan-96/45/head 2025-11-03T16:58:01.3325258Z * [new branch] gh/ZhiweiYan-96/49/base -> origin/gh/ZhiweiYan-96/49/base 2025-11-03T16:58:01.3326771Z * [new branch] gh/ZhiweiYan-96/49/head -> origin/gh/ZhiweiYan-96/49/head 2025-11-03T16:58:01.3329011Z * [new branch] gh/ZhiweiYan-96/62/base -> origin/gh/ZhiweiYan-96/62/base 2025-11-03T16:58:01.3330712Z * [new branch] gh/ZhiweiYan-96/62/head -> origin/gh/ZhiweiYan-96/62/head 2025-11-03T16:58:01.3333153Z * [new branch] gh/ZhiweiYan-96/66/base -> origin/gh/ZhiweiYan-96/66/base 2025-11-03T16:58:01.3334858Z * [new branch] gh/ZhiweiYan-96/66/head -> origin/gh/ZhiweiYan-96/66/head 2025-11-03T16:58:01.3337160Z * [new branch] gh/ZhiweiYan-96/67/base -> origin/gh/ZhiweiYan-96/67/base 2025-11-03T16:58:01.3339045Z * [new branch] gh/ZhiweiYan-96/67/head -> origin/gh/ZhiweiYan-96/67/head 2025-11-03T16:58:01.3341534Z * [new branch] gh/ZhiweiYan-96/68/base -> origin/gh/ZhiweiYan-96/68/base 2025-11-03T16:58:01.3343176Z * [new branch] gh/ZhiweiYan-96/68/head -> origin/gh/ZhiweiYan-96/68/head 2025-11-03T16:58:01.3344763Z * [new branch] gh/ZhiweiYan-96/68/orig -> origin/gh/ZhiweiYan-96/68/orig 2025-11-03T16:58:01.3347668Z * [new branch] gh/aakhundov/1/base -> origin/gh/aakhundov/1/base 2025-11-03T16:58:01.3349409Z * [new branch] gh/aakhundov/1/head -> origin/gh/aakhundov/1/head 2025-11-03T16:58:01.3351652Z * [new branch] gh/aakhundov/2/base -> origin/gh/aakhundov/2/base 2025-11-03T16:58:01.3353353Z * [new branch] gh/aakhundov/2/head -> origin/gh/aakhundov/2/head 2025-11-03T16:58:01.3355675Z * [new branch] gh/aakhundov/3/base -> origin/gh/aakhundov/3/base 2025-11-03T16:58:01.3357423Z * [new branch] gh/aakhundov/3/head -> origin/gh/aakhundov/3/head 2025-11-03T16:58:01.3359097Z * [new branch] gh/aakhundov/3/orig -> origin/gh/aakhundov/3/orig 2025-11-03T16:58:01.3361491Z * [new branch] gh/aditew01/openblas -> origin/gh/aditew01/openblas 2025-11-03T16:58:01.3363323Z * [new branch] gh/aditew01/sbgemm -> origin/gh/aditew01/sbgemm 2025-11-03T16:58:01.3365037Z * [new branch] gh/aditew01/vecbf16 -> origin/gh/aditew01/vecbf16 2025-11-03T16:58:01.3367844Z * [new branch] gh/albanD/1/base -> origin/gh/albanD/1/base 2025-11-03T16:58:01.3369483Z * [new branch] gh/albanD/1/head -> origin/gh/albanD/1/head 2025-11-03T16:58:01.3371158Z * [new branch] gh/albanD/1/orig -> origin/gh/albanD/1/orig 2025-11-03T16:58:01.3373334Z * [new branch] gh/albanD/2/base -> origin/gh/albanD/2/base 2025-11-03T16:58:01.3375041Z * [new branch] gh/albanD/2/head -> origin/gh/albanD/2/head 2025-11-03T16:58:01.3376720Z * [new branch] gh/albanD/2/orig -> origin/gh/albanD/2/orig 2025-11-03T16:58:01.3379059Z * [new branch] gh/albanD/3/base -> origin/gh/albanD/3/base 2025-11-03T16:58:01.3380793Z * [new branch] gh/albanD/3/head -> origin/gh/albanD/3/head 2025-11-03T16:58:01.3382748Z * [new branch] gh/albanD/3/orig -> origin/gh/albanD/3/orig 2025-11-03T16:58:01.3385093Z * [new branch] gh/albanD/4/base -> origin/gh/albanD/4/base 2025-11-03T16:58:01.3386737Z * [new branch] gh/albanD/4/head -> origin/gh/albanD/4/head 2025-11-03T16:58:01.3388421Z * [new branch] gh/albanD/4/orig -> origin/gh/albanD/4/orig 2025-11-03T16:58:01.3391051Z * [new branch] gh/alexbrauckmann/paddedtensor_faketensor_init -> origin/gh/alexbrauckmann/paddedtensor_faketensor_init 2025-11-03T16:58:01.3393768Z * [new branch] gh/alexsamardzic/11/base -> origin/gh/alexsamardzic/11/base 2025-11-03T16:58:01.3395428Z * [new branch] gh/alexsamardzic/11/head -> origin/gh/alexsamardzic/11/head 2025-11-03T16:58:01.3397135Z * [new branch] gh/alexsamardzic/11/orig -> origin/gh/alexsamardzic/11/orig 2025-11-03T16:58:01.3399446Z * [new branch] gh/alexsamardzic/12/base -> origin/gh/alexsamardzic/12/base 2025-11-03T16:58:01.3401127Z * [new branch] gh/alexsamardzic/12/head -> origin/gh/alexsamardzic/12/head 2025-11-03T16:58:01.3402723Z * [new branch] gh/alexsamardzic/12/orig -> origin/gh/alexsamardzic/12/orig 2025-11-03T16:58:01.3405071Z * [new branch] gh/alexsamardzic/13/base -> origin/gh/alexsamardzic/13/base 2025-11-03T16:58:01.3406788Z * [new branch] gh/alexsamardzic/13/head -> origin/gh/alexsamardzic/13/head 2025-11-03T16:58:01.3408509Z * [new branch] gh/alexsamardzic/13/orig -> origin/gh/alexsamardzic/13/orig 2025-11-03T16:58:01.3410824Z * [new branch] gh/alexsamardzic/14/base -> origin/gh/alexsamardzic/14/base 2025-11-03T16:58:01.3412682Z * [new branch] gh/alexsamardzic/14/head -> origin/gh/alexsamardzic/14/head 2025-11-03T16:58:01.3414398Z * [new branch] gh/alexsamardzic/14/orig -> origin/gh/alexsamardzic/14/orig 2025-11-03T16:58:01.3417256Z * [new branch] gh/amjames/18/base -> origin/gh/amjames/18/base 2025-11-03T16:58:01.3419031Z * [new branch] gh/amjames/18/head -> origin/gh/amjames/18/head 2025-11-03T16:58:01.3420668Z * [new branch] gh/amjames/18/orig -> origin/gh/amjames/18/orig 2025-11-03T16:58:01.3424522Z * [new branch] gh/andrewor14/35/base -> origin/gh/andrewor14/35/base 2025-11-03T16:58:01.3426228Z * [new branch] gh/andrewor14/35/head -> origin/gh/andrewor14/35/head 2025-11-03T16:58:01.3427929Z * [new branch] gh/andrewor14/35/orig -> origin/gh/andrewor14/35/orig 2025-11-03T16:58:01.3430422Z * [new branch] gh/andrewor14/50/base -> origin/gh/andrewor14/50/base 2025-11-03T16:58:01.3432255Z * [new branch] gh/andrewor14/50/head -> origin/gh/andrewor14/50/head 2025-11-03T16:58:01.3434120Z * [new branch] gh/andrewor14/50/orig -> origin/gh/andrewor14/50/orig 2025-11-03T16:58:01.3436930Z * [new branch] gh/andyanwang/30/base -> origin/gh/andyanwang/30/base 2025-11-03T16:58:01.3438791Z * [new branch] gh/andyanwang/30/orig -> origin/gh/andyanwang/30/orig 2025-11-03T16:58:01.3441132Z * [new branch] gh/andyanwang/31/base -> origin/gh/andyanwang/31/base 2025-11-03T16:58:01.3442984Z * [new branch] gh/andyanwang/31/orig -> origin/gh/andyanwang/31/orig 2025-11-03T16:58:01.3445591Z * [new branch] gh/andyanwang/32/base -> origin/gh/andyanwang/32/base 2025-11-03T16:58:01.3447314Z * [new branch] gh/andyanwang/32/head -> origin/gh/andyanwang/32/head 2025-11-03T16:58:01.3449113Z * [new branch] gh/andyanwang/32/orig -> origin/gh/andyanwang/32/orig 2025-11-03T16:58:01.3451609Z * [new branch] gh/andyanwang/39/base -> origin/gh/andyanwang/39/base 2025-11-03T16:58:01.3453341Z * [new branch] gh/andyanwang/39/head -> origin/gh/andyanwang/39/head 2025-11-03T16:58:01.3455069Z * [new branch] gh/andyanwang/39/orig -> origin/gh/andyanwang/39/orig 2025-11-03T16:58:01.3457450Z * [new branch] gh/andyanwang/40/base -> origin/gh/andyanwang/40/base 2025-11-03T16:58:01.3459312Z * [new branch] gh/andyanwang/40/head -> origin/gh/andyanwang/40/head 2025-11-03T16:58:01.3461043Z * [new branch] gh/andyanwang/40/orig -> origin/gh/andyanwang/40/orig 2025-11-03T16:58:01.3463873Z * [new branch] gh/andyanwang/41/base -> origin/gh/andyanwang/41/base 2025-11-03T16:58:01.3465873Z * [new branch] gh/andyanwang/41/head -> origin/gh/andyanwang/41/head 2025-11-03T16:58:01.3467601Z * [new branch] gh/andyanwang/41/orig -> origin/gh/andyanwang/41/orig 2025-11-03T16:58:01.3470113Z * [new branch] gh/andyanwang/42/base -> origin/gh/andyanwang/42/base 2025-11-03T16:58:01.3471844Z * [new branch] gh/andyanwang/42/head -> origin/gh/andyanwang/42/head 2025-11-03T16:58:01.3473706Z * [new branch] gh/andyanwang/42/orig -> origin/gh/andyanwang/42/orig 2025-11-03T16:58:01.3476038Z * [new branch] gh/andyanwang/43/base -> origin/gh/andyanwang/43/base 2025-11-03T16:58:01.3477944Z * [new branch] gh/andyanwang/43/head -> origin/gh/andyanwang/43/head 2025-11-03T16:58:01.3479710Z * [new branch] gh/andyanwang/43/orig -> origin/gh/andyanwang/43/orig 2025-11-03T16:58:01.3481984Z * [new branch] gh/andyanwang/44/base -> origin/gh/andyanwang/44/base 2025-11-03T16:58:01.3483567Z * [new branch] gh/andyanwang/44/head -> origin/gh/andyanwang/44/head 2025-11-03T16:58:01.3485423Z * [new branch] gh/andyanwang/44/orig -> origin/gh/andyanwang/44/orig 2025-11-03T16:58:01.3488213Z * [new branch] gh/angelayi/107/base -> origin/gh/angelayi/107/base 2025-11-03T16:58:01.3489985Z * [new branch] gh/angelayi/107/head -> origin/gh/angelayi/107/head 2025-11-03T16:58:01.3492289Z * [new branch] gh/angelayi/114/base -> origin/gh/angelayi/114/base 2025-11-03T16:58:01.3494140Z * [new branch] gh/angelayi/114/head -> origin/gh/angelayi/114/head 2025-11-03T16:58:01.3495754Z * [new branch] gh/angelayi/114/orig -> origin/gh/angelayi/114/orig 2025-11-03T16:58:01.3498213Z * [new branch] gh/angelayi/116/base -> origin/gh/angelayi/116/base 2025-11-03T16:58:01.3500000Z * [new branch] gh/angelayi/116/head -> origin/gh/angelayi/116/head 2025-11-03T16:58:01.3501676Z * [new branch] gh/angelayi/116/orig -> origin/gh/angelayi/116/orig 2025-11-03T16:58:01.3504255Z * [new branch] gh/angelayi/119/base -> origin/gh/angelayi/119/base 2025-11-03T16:58:01.3506206Z * [new branch] gh/angelayi/119/head -> origin/gh/angelayi/119/head 2025-11-03T16:58:01.3507515Z * [new branch] gh/angelayi/119/orig -> origin/gh/angelayi/119/orig 2025-11-03T16:58:01.3510000Z * [new branch] gh/angelayi/120/base -> origin/gh/angelayi/120/base 2025-11-03T16:58:01.3511605Z * [new branch] gh/angelayi/120/head -> origin/gh/angelayi/120/head 2025-11-03T16:58:01.3513430Z * [new branch] gh/angelayi/120/orig -> origin/gh/angelayi/120/orig 2025-11-03T16:58:01.3515751Z * [new branch] gh/angelayi/121/base -> origin/gh/angelayi/121/base 2025-11-03T16:58:01.3517494Z * [new branch] gh/angelayi/121/head -> origin/gh/angelayi/121/head 2025-11-03T16:58:01.3518956Z * [new branch] gh/angelayi/121/orig -> origin/gh/angelayi/121/orig 2025-11-03T16:58:01.3521580Z * [new branch] gh/angelayi/122/base -> origin/gh/angelayi/122/base 2025-11-03T16:58:01.3523463Z * [new branch] gh/angelayi/122/head -> origin/gh/angelayi/122/head 2025-11-03T16:58:01.3525143Z * [new branch] gh/angelayi/122/orig -> origin/gh/angelayi/122/orig 2025-11-03T16:58:01.3527564Z * [new branch] gh/angelayi/124/base -> origin/gh/angelayi/124/base 2025-11-03T16:58:01.3529334Z * [new branch] gh/angelayi/124/head -> origin/gh/angelayi/124/head 2025-11-03T16:58:01.3530902Z * [new branch] gh/angelayi/124/orig -> origin/gh/angelayi/124/orig 2025-11-03T16:58:01.3533677Z * [new branch] gh/angelayi/127/base -> origin/gh/angelayi/127/base 2025-11-03T16:58:01.3535109Z * [new branch] gh/angelayi/127/head -> origin/gh/angelayi/127/head 2025-11-03T16:58:01.3537128Z * [new branch] gh/angelayi/127/orig -> origin/gh/angelayi/127/orig 2025-11-03T16:58:01.3539578Z * [new branch] gh/angelayi/128/base -> origin/gh/angelayi/128/base 2025-11-03T16:58:01.3541012Z * [new branch] gh/angelayi/128/head -> origin/gh/angelayi/128/head 2025-11-03T16:58:01.3543080Z * [new branch] gh/angelayi/128/orig -> origin/gh/angelayi/128/orig 2025-11-03T16:58:01.3545604Z * [new branch] gh/angelayi/129/base -> origin/gh/angelayi/129/base 2025-11-03T16:58:01.3547292Z * [new branch] gh/angelayi/129/head -> origin/gh/angelayi/129/head 2025-11-03T16:58:01.3549350Z * [new branch] gh/angelayi/129/orig -> origin/gh/angelayi/129/orig 2025-11-03T16:58:01.3551716Z * [new branch] gh/angelayi/130/base -> origin/gh/angelayi/130/base 2025-11-03T16:58:01.3553232Z * [new branch] gh/angelayi/130/head -> origin/gh/angelayi/130/head 2025-11-03T16:58:01.3555029Z * [new branch] gh/angelayi/130/orig -> origin/gh/angelayi/130/orig 2025-11-03T16:58:01.3557925Z * [new branch] gh/anijain2305/753/base -> origin/gh/anijain2305/753/base 2025-11-03T16:58:01.3559606Z * [new branch] gh/anijain2305/753/head -> origin/gh/anijain2305/753/head 2025-11-03T16:58:01.3561391Z * [new branch] gh/anijain2305/753/orig -> origin/gh/anijain2305/753/orig 2025-11-03T16:58:01.3563841Z * [new branch] gh/anijain2305/792/base -> origin/gh/anijain2305/792/base 2025-11-03T16:58:01.3566041Z * [new branch] gh/anijain2305/792/head -> origin/gh/anijain2305/792/head 2025-11-03T16:58:01.3568566Z * [new branch] gh/anijain2305/792/orig -> origin/gh/anijain2305/792/orig 2025-11-03T16:58:01.3572047Z * [new branch] gh/anijain2305/805/base -> origin/gh/anijain2305/805/base 2025-11-03T16:58:01.3574604Z * [new branch] gh/anijain2305/805/head -> origin/gh/anijain2305/805/head 2025-11-03T16:58:01.3577147Z * [new branch] gh/anijain2305/805/orig -> origin/gh/anijain2305/805/orig 2025-11-03T16:58:01.3580524Z * [new branch] gh/anijain2305/810/base -> origin/gh/anijain2305/810/base 2025-11-03T16:58:01.3583006Z * [new branch] gh/anijain2305/810/head -> origin/gh/anijain2305/810/head 2025-11-03T16:58:01.3585302Z * [new branch] gh/anijain2305/810/orig -> origin/gh/anijain2305/810/orig 2025-11-03T16:58:01.3588520Z * [new branch] gh/anijain2305/812/base -> origin/gh/anijain2305/812/base 2025-11-03T16:58:01.3590800Z * [new branch] gh/anijain2305/812/head -> origin/gh/anijain2305/812/head 2025-11-03T16:58:01.3593163Z * [new branch] gh/anijain2305/812/orig -> origin/gh/anijain2305/812/orig 2025-11-03T16:58:01.3596372Z * [new branch] gh/anijain2305/854/base -> origin/gh/anijain2305/854/base 2025-11-03T16:58:01.3598780Z * [new branch] gh/anijain2305/854/head -> origin/gh/anijain2305/854/head 2025-11-03T16:58:01.3601144Z * [new branch] gh/anijain2305/854/orig -> origin/gh/anijain2305/854/orig 2025-11-03T16:58:01.3604750Z * [new branch] gh/anijain2305/864/base -> origin/gh/anijain2305/864/base 2025-11-03T16:58:01.3606597Z * [new branch] gh/anijain2305/864/head -> origin/gh/anijain2305/864/head 2025-11-03T16:58:01.3608299Z * [new branch] gh/anijain2305/864/orig -> origin/gh/anijain2305/864/orig 2025-11-03T16:58:01.3610875Z * [new branch] gh/anijain2305/870/base -> origin/gh/anijain2305/870/base 2025-11-03T16:58:01.3612572Z * [new branch] gh/anijain2305/870/head -> origin/gh/anijain2305/870/head 2025-11-03T16:58:01.3614347Z * [new branch] gh/anijain2305/870/orig -> origin/gh/anijain2305/870/orig 2025-11-03T16:58:01.3616662Z * [new branch] gh/anijain2305/873/base -> origin/gh/anijain2305/873/base 2025-11-03T16:58:01.3618264Z * [new branch] gh/anijain2305/873/head -> origin/gh/anijain2305/873/head 2025-11-03T16:58:01.3620101Z * [new branch] gh/anijain2305/873/orig -> origin/gh/anijain2305/873/orig 2025-11-03T16:58:01.3622808Z * [new branch] gh/anijain2305/891/base -> origin/gh/anijain2305/891/base 2025-11-03T16:58:01.3624674Z * [new branch] gh/anijain2305/891/head -> origin/gh/anijain2305/891/head 2025-11-03T16:58:01.3626480Z * [new branch] gh/anijain2305/891/orig -> origin/gh/anijain2305/891/orig 2025-11-03T16:58:01.3628898Z * [new branch] gh/anijain2305/894/base -> origin/gh/anijain2305/894/base 2025-11-03T16:58:01.3630667Z * [new branch] gh/anijain2305/894/head -> origin/gh/anijain2305/894/head 2025-11-03T16:58:01.3632324Z * [new branch] gh/anijain2305/894/orig -> origin/gh/anijain2305/894/orig 2025-11-03T16:58:01.3634737Z * [new branch] gh/anijain2305/895/base -> origin/gh/anijain2305/895/base 2025-11-03T16:58:01.3636598Z * [new branch] gh/anijain2305/895/head -> origin/gh/anijain2305/895/head 2025-11-03T16:58:01.3638099Z * [new branch] gh/anijain2305/895/orig -> origin/gh/anijain2305/895/orig 2025-11-03T16:58:01.3640586Z * [new branch] gh/anijain2305/896/base -> origin/gh/anijain2305/896/base 2025-11-03T16:58:01.3642422Z * [new branch] gh/anijain2305/896/head -> origin/gh/anijain2305/896/head 2025-11-03T16:58:01.3644008Z * [new branch] gh/anijain2305/896/orig -> origin/gh/anijain2305/896/orig 2025-11-03T16:58:01.3646634Z * [new branch] gh/anijain2305/897/base -> origin/gh/anijain2305/897/base 2025-11-03T16:58:01.3648469Z * [new branch] gh/anijain2305/897/head -> origin/gh/anijain2305/897/head 2025-11-03T16:58:01.3650242Z * [new branch] gh/anijain2305/897/orig -> origin/gh/anijain2305/897/orig 2025-11-03T16:58:01.3652714Z * [new branch] gh/anijain2305/898/base -> origin/gh/anijain2305/898/base 2025-11-03T16:58:01.3654672Z * [new branch] gh/anijain2305/898/head -> origin/gh/anijain2305/898/head 2025-11-03T16:58:01.3656418Z * [new branch] gh/anijain2305/898/orig -> origin/gh/anijain2305/898/orig 2025-11-03T16:58:01.3658736Z * [new branch] gh/anijain2305/899/base -> origin/gh/anijain2305/899/base 2025-11-03T16:58:01.3660483Z * [new branch] gh/anijain2305/899/head -> origin/gh/anijain2305/899/head 2025-11-03T16:58:01.3662315Z * [new branch] gh/anijain2305/899/orig -> origin/gh/anijain2305/899/orig 2025-11-03T16:58:01.3665774Z * [new branch] gh/anijain2305/900/base -> origin/gh/anijain2305/900/base 2025-11-03T16:58:01.3667645Z * [new branch] gh/anijain2305/900/head -> origin/gh/anijain2305/900/head 2025-11-03T16:58:01.3669344Z * [new branch] gh/anijain2305/900/orig -> origin/gh/anijain2305/900/orig 2025-11-03T16:58:01.3671610Z * [new branch] gh/anijain2305/901/base -> origin/gh/anijain2305/901/base 2025-11-03T16:58:01.3673347Z * [new branch] gh/anijain2305/901/head -> origin/gh/anijain2305/901/head 2025-11-03T16:58:01.3675169Z * [new branch] gh/anijain2305/901/orig -> origin/gh/anijain2305/901/orig 2025-11-03T16:58:01.3677592Z * [new branch] gh/anijain2305/902/base -> origin/gh/anijain2305/902/base 2025-11-03T16:58:01.3679336Z * [new branch] gh/anijain2305/902/head -> origin/gh/anijain2305/902/head 2025-11-03T16:58:01.3680988Z * [new branch] gh/anijain2305/902/orig -> origin/gh/anijain2305/902/orig 2025-11-03T16:58:01.3683610Z * [new branch] gh/anijain2305/903/base -> origin/gh/anijain2305/903/base 2025-11-03T16:58:01.3684919Z * [new branch] gh/anijain2305/903/head -> origin/gh/anijain2305/903/head 2025-11-03T16:58:01.3686927Z * [new branch] gh/anijain2305/903/orig -> origin/gh/anijain2305/903/orig 2025-11-03T16:58:01.3689365Z * [new branch] gh/anijain2305/904/base -> origin/gh/anijain2305/904/base 2025-11-03T16:58:01.3691144Z * [new branch] gh/anijain2305/904/head -> origin/gh/anijain2305/904/head 2025-11-03T16:58:01.3692777Z * [new branch] gh/anijain2305/904/orig -> origin/gh/anijain2305/904/orig 2025-11-03T16:58:01.3695255Z * [new branch] gh/anijain2305/905/base -> origin/gh/anijain2305/905/base 2025-11-03T16:58:01.3697028Z * [new branch] gh/anijain2305/905/head -> origin/gh/anijain2305/905/head 2025-11-03T16:58:01.3698629Z * [new branch] gh/anijain2305/905/orig -> origin/gh/anijain2305/905/orig 2025-11-03T16:58:01.3701281Z * [new branch] gh/anijain2305/906/base -> origin/gh/anijain2305/906/base 2025-11-03T16:58:01.3703184Z * [new branch] gh/anijain2305/906/head -> origin/gh/anijain2305/906/head 2025-11-03T16:58:01.3704619Z * [new branch] gh/anijain2305/906/orig -> origin/gh/anijain2305/906/orig 2025-11-03T16:58:01.3707114Z * [new branch] gh/anijain2305/907/base -> origin/gh/anijain2305/907/base 2025-11-03T16:58:01.3708839Z * [new branch] gh/anijain2305/907/head -> origin/gh/anijain2305/907/head 2025-11-03T16:58:01.3710326Z * [new branch] gh/anijain2305/907/orig -> origin/gh/anijain2305/907/orig 2025-11-03T16:58:01.3712949Z * [new branch] gh/anijain2305/908/base -> origin/gh/anijain2305/908/base 2025-11-03T16:58:01.3714702Z * [new branch] gh/anijain2305/908/head -> origin/gh/anijain2305/908/head 2025-11-03T16:58:01.3716416Z * [new branch] gh/anijain2305/908/orig -> origin/gh/anijain2305/908/orig 2025-11-03T16:58:01.3718917Z * [new branch] gh/anijain2305/909/base -> origin/gh/anijain2305/909/base 2025-11-03T16:58:01.3720548Z * [new branch] gh/anijain2305/909/head -> origin/gh/anijain2305/909/head 2025-11-03T16:58:01.3724294Z * [new branch] gh/anijain2305/909/orig -> origin/gh/anijain2305/909/orig 2025-11-03T16:58:01.3727005Z * [new branch] gh/anijain2305/910/base -> origin/gh/anijain2305/910/base 2025-11-03T16:58:01.3728456Z * [new branch] gh/anijain2305/910/head -> origin/gh/anijain2305/910/head 2025-11-03T16:58:01.3730309Z * [new branch] gh/anijain2305/910/orig -> origin/gh/anijain2305/910/orig 2025-11-03T16:58:01.3732740Z * [new branch] gh/anijain2305/911/base -> origin/gh/anijain2305/911/base 2025-11-03T16:58:01.3734483Z * [new branch] gh/anijain2305/911/head -> origin/gh/anijain2305/911/head 2025-11-03T16:58:01.3736215Z * [new branch] gh/anijain2305/911/orig -> origin/gh/anijain2305/911/orig 2025-11-03T16:58:01.3738666Z * [new branch] gh/anijain2305/912/base -> origin/gh/anijain2305/912/base 2025-11-03T16:58:01.3740426Z * [new branch] gh/anijain2305/912/head -> origin/gh/anijain2305/912/head 2025-11-03T16:58:01.3742267Z * [new branch] gh/anijain2305/912/orig -> origin/gh/anijain2305/912/orig 2025-11-03T16:58:01.3744656Z * [new branch] gh/anijain2305/913/base -> origin/gh/anijain2305/913/base 2025-11-03T16:58:01.3746471Z * [new branch] gh/anijain2305/913/head -> origin/gh/anijain2305/913/head 2025-11-03T16:58:01.3748254Z * [new branch] gh/anijain2305/913/orig -> origin/gh/anijain2305/913/orig 2025-11-03T16:58:01.3750810Z * [new branch] gh/anijain2305/914/base -> origin/gh/anijain2305/914/base 2025-11-03T16:58:01.3752762Z * [new branch] gh/anijain2305/914/head -> origin/gh/anijain2305/914/head 2025-11-03T16:58:01.3754131Z * [new branch] gh/anijain2305/914/orig -> origin/gh/anijain2305/914/orig 2025-11-03T16:58:01.3756622Z * [new branch] gh/anijain2305/915/base -> origin/gh/anijain2305/915/base 2025-11-03T16:58:01.3758426Z * [new branch] gh/anijain2305/915/head -> origin/gh/anijain2305/915/head 2025-11-03T16:58:01.3760101Z * [new branch] gh/anijain2305/915/orig -> origin/gh/anijain2305/915/orig 2025-11-03T16:58:01.3762362Z * [new branch] gh/anijain2305/916/base -> origin/gh/anijain2305/916/base 2025-11-03T16:58:01.3764145Z * [new branch] gh/anijain2305/916/head -> origin/gh/anijain2305/916/head 2025-11-03T16:58:01.3765819Z * [new branch] gh/anijain2305/916/orig -> origin/gh/anijain2305/916/orig 2025-11-03T16:58:01.3768122Z * [new branch] gh/anijain2305/917/base -> origin/gh/anijain2305/917/base 2025-11-03T16:58:01.3769830Z * [new branch] gh/anijain2305/917/head -> origin/gh/anijain2305/917/head 2025-11-03T16:58:01.3771575Z * [new branch] gh/anijain2305/917/orig -> origin/gh/anijain2305/917/orig 2025-11-03T16:58:01.3774030Z * [new branch] gh/anijain2305/918/base -> origin/gh/anijain2305/918/base 2025-11-03T16:58:01.3775693Z * [new branch] gh/anijain2305/918/head -> origin/gh/anijain2305/918/head 2025-11-03T16:58:01.3777686Z * [new branch] gh/anijain2305/918/orig -> origin/gh/anijain2305/918/orig 2025-11-03T16:58:01.3780081Z * [new branch] gh/anijain2305/919/base -> origin/gh/anijain2305/919/base 2025-11-03T16:58:01.3781660Z * [new branch] gh/anijain2305/919/head -> origin/gh/anijain2305/919/head 2025-11-03T16:58:01.3783976Z * [new branch] gh/anijain2305/919/orig -> origin/gh/anijain2305/919/orig 2025-11-03T16:58:01.3786188Z * [new branch] gh/anijain2305/920/base -> origin/gh/anijain2305/920/base 2025-11-03T16:58:01.3787850Z * [new branch] gh/anijain2305/920/head -> origin/gh/anijain2305/920/head 2025-11-03T16:58:01.3789648Z * [new branch] gh/anijain2305/920/orig -> origin/gh/anijain2305/920/orig 2025-11-03T16:58:01.3791982Z * [new branch] gh/anijain2305/921/base -> origin/gh/anijain2305/921/base 2025-11-03T16:58:01.3793795Z * [new branch] gh/anijain2305/921/head -> origin/gh/anijain2305/921/head 2025-11-03T16:58:01.3795518Z * [new branch] gh/anijain2305/921/orig -> origin/gh/anijain2305/921/orig 2025-11-03T16:58:01.3797896Z * [new branch] gh/anijain2305/922/base -> origin/gh/anijain2305/922/base 2025-11-03T16:58:01.3799637Z * [new branch] gh/anijain2305/922/head -> origin/gh/anijain2305/922/head 2025-11-03T16:58:01.3801380Z * [new branch] gh/anijain2305/922/orig -> origin/gh/anijain2305/922/orig 2025-11-03T16:58:01.3803697Z * [new branch] gh/anijain2305/923/base -> origin/gh/anijain2305/923/base 2025-11-03T16:58:01.3805386Z * [new branch] gh/anijain2305/923/head -> origin/gh/anijain2305/923/head 2025-11-03T16:58:01.3807122Z * [new branch] gh/anijain2305/923/orig -> origin/gh/anijain2305/923/orig 2025-11-03T16:58:01.3809389Z * [new branch] gh/anijain2305/924/base -> origin/gh/anijain2305/924/base 2025-11-03T16:58:01.3811097Z * [new branch] gh/anijain2305/924/head -> origin/gh/anijain2305/924/head 2025-11-03T16:58:01.3812790Z * [new branch] gh/anijain2305/924/orig -> origin/gh/anijain2305/924/orig 2025-11-03T16:58:01.3815241Z * [new branch] gh/anijain2305/925/base -> origin/gh/anijain2305/925/base 2025-11-03T16:58:01.3817010Z * [new branch] gh/anijain2305/925/head -> origin/gh/anijain2305/925/head 2025-11-03T16:58:01.3818735Z * [new branch] gh/anijain2305/925/orig -> origin/gh/anijain2305/925/orig 2025-11-03T16:58:01.3821255Z * [new branch] gh/anijain2305/926/base -> origin/gh/anijain2305/926/base 2025-11-03T16:58:01.3823314Z * [new branch] gh/anijain2305/926/head -> origin/gh/anijain2305/926/head 2025-11-03T16:58:01.3824972Z * [new branch] gh/anijain2305/926/orig -> origin/gh/anijain2305/926/orig 2025-11-03T16:58:01.3827380Z * [new branch] gh/anijain2305/927/base -> origin/gh/anijain2305/927/base 2025-11-03T16:58:01.3829118Z * [new branch] gh/anijain2305/927/head -> origin/gh/anijain2305/927/head 2025-11-03T16:58:01.3830798Z * [new branch] gh/anijain2305/927/orig -> origin/gh/anijain2305/927/orig 2025-11-03T16:58:01.3833260Z * [new branch] gh/anijain2305/928/base -> origin/gh/anijain2305/928/base 2025-11-03T16:58:01.3835040Z * [new branch] gh/anijain2305/928/head -> origin/gh/anijain2305/928/head 2025-11-03T16:58:01.3836814Z * [new branch] gh/anijain2305/928/orig -> origin/gh/anijain2305/928/orig 2025-11-03T16:58:01.3839148Z * [new branch] gh/anijain2305/929/base -> origin/gh/anijain2305/929/base 2025-11-03T16:58:01.3840759Z * [new branch] gh/anijain2305/929/head -> origin/gh/anijain2305/929/head 2025-11-03T16:58:01.3842426Z * [new branch] gh/anijain2305/929/orig -> origin/gh/anijain2305/929/orig 2025-11-03T16:58:01.3844917Z * [new branch] gh/anijain2305/930/base -> origin/gh/anijain2305/930/base 2025-11-03T16:58:01.3846598Z * [new branch] gh/anijain2305/930/head -> origin/gh/anijain2305/930/head 2025-11-03T16:58:01.3848302Z * [new branch] gh/anijain2305/930/orig -> origin/gh/anijain2305/930/orig 2025-11-03T16:58:01.3850778Z * [new branch] gh/anijain2305/931/base -> origin/gh/anijain2305/931/base 2025-11-03T16:58:01.3852475Z * [new branch] gh/anijain2305/931/head -> origin/gh/anijain2305/931/head 2025-11-03T16:58:01.3854131Z * [new branch] gh/anijain2305/931/orig -> origin/gh/anijain2305/931/orig 2025-11-03T16:58:01.3856585Z * [new branch] gh/anijain2305/932/base -> origin/gh/anijain2305/932/base 2025-11-03T16:58:01.3858357Z * [new branch] gh/anijain2305/932/head -> origin/gh/anijain2305/932/head 2025-11-03T16:58:01.3860181Z * [new branch] gh/anijain2305/932/orig -> origin/gh/anijain2305/932/orig 2025-11-03T16:58:01.3862861Z * [new branch] gh/anijain2305/933/base -> origin/gh/anijain2305/933/base 2025-11-03T16:58:01.3864533Z * [new branch] gh/anijain2305/933/head -> origin/gh/anijain2305/933/head 2025-11-03T16:58:01.3866189Z * [new branch] gh/anijain2305/933/orig -> origin/gh/anijain2305/933/orig 2025-11-03T16:58:01.3868691Z * [new branch] gh/anijain2305/934/base -> origin/gh/anijain2305/934/base 2025-11-03T16:58:01.3870365Z * [new branch] gh/anijain2305/934/head -> origin/gh/anijain2305/934/head 2025-11-03T16:58:01.3872076Z * [new branch] gh/anijain2305/934/orig -> origin/gh/anijain2305/934/orig 2025-11-03T16:58:01.3874539Z * [new branch] gh/anijain2305/935/base -> origin/gh/anijain2305/935/base 2025-11-03T16:58:01.3876320Z * [new branch] gh/anijain2305/935/head -> origin/gh/anijain2305/935/head 2025-11-03T16:58:01.3877947Z * [new branch] gh/anijain2305/935/orig -> origin/gh/anijain2305/935/orig 2025-11-03T16:58:01.3880275Z * [new branch] gh/anijain2305/936/base -> origin/gh/anijain2305/936/base 2025-11-03T16:58:01.3881907Z * [new branch] gh/anijain2305/936/head -> origin/gh/anijain2305/936/head 2025-11-03T16:58:01.3883580Z * [new branch] gh/anijain2305/936/orig -> origin/gh/anijain2305/936/orig 2025-11-03T16:58:01.3886130Z * [new branch] gh/anijain2305/937/base -> origin/gh/anijain2305/937/base 2025-11-03T16:58:01.3887828Z * [new branch] gh/anijain2305/937/head -> origin/gh/anijain2305/937/head 2025-11-03T16:58:01.3889517Z * [new branch] gh/anijain2305/937/orig -> origin/gh/anijain2305/937/orig 2025-11-03T16:58:01.3891913Z * [new branch] gh/anijain2305/938/base -> origin/gh/anijain2305/938/base 2025-11-03T16:58:01.3893685Z * [new branch] gh/anijain2305/938/head -> origin/gh/anijain2305/938/head 2025-11-03T16:58:01.3895398Z * [new branch] gh/anijain2305/938/orig -> origin/gh/anijain2305/938/orig 2025-11-03T16:58:01.3897735Z * [new branch] gh/anijain2305/939/base -> origin/gh/anijain2305/939/base 2025-11-03T16:58:01.3899328Z * [new branch] gh/anijain2305/939/head -> origin/gh/anijain2305/939/head 2025-11-03T16:58:01.3901073Z * [new branch] gh/anijain2305/939/orig -> origin/gh/anijain2305/939/orig 2025-11-03T16:58:01.3903597Z * [new branch] gh/anijain2305/940/base -> origin/gh/anijain2305/940/base 2025-11-03T16:58:01.3905263Z * [new branch] gh/anijain2305/940/head -> origin/gh/anijain2305/940/head 2025-11-03T16:58:01.3906943Z * [new branch] gh/anijain2305/940/orig -> origin/gh/anijain2305/940/orig 2025-11-03T16:58:01.3909897Z * [new branch] gh/anjali411/216/base -> origin/gh/anjali411/216/base 2025-11-03T16:58:01.3911587Z * [new branch] gh/anjali411/216/head -> origin/gh/anjali411/216/head 2025-11-03T16:58:01.3913422Z * [new branch] gh/anjali411/216/orig -> origin/gh/anjali411/216/orig 2025-11-03T16:58:01.3916376Z * [new branch] gh/ankitageorge/17/base -> origin/gh/ankitageorge/17/base 2025-11-03T16:58:01.3917974Z * [new branch] gh/ankitageorge/17/head -> origin/gh/ankitageorge/17/head 2025-11-03T16:58:01.3919785Z * [new branch] gh/ankitageorge/17/orig -> origin/gh/ankitageorge/17/orig 2025-11-03T16:58:01.3922853Z * [new branch] gh/anshul-si/1/base -> origin/gh/anshul-si/1/base 2025-11-03T16:58:01.3924621Z * [new branch] gh/anshul-si/1/head -> origin/gh/anshul-si/1/head 2025-11-03T16:58:01.3926851Z * [new branch] gh/anshul-si/2/base -> origin/gh/anshul-si/2/base 2025-11-03T16:58:01.3928505Z * [new branch] gh/anshul-si/2/head -> origin/gh/anshul-si/2/head 2025-11-03T16:58:01.3930686Z * [new branch] gh/anshul-si/3/base -> origin/gh/anshul-si/3/base 2025-11-03T16:58:01.3932437Z * [new branch] gh/anshul-si/3/head -> origin/gh/anshul-si/3/head 2025-11-03T16:58:01.3934671Z * [new branch] gh/anshul-si/4/base -> origin/gh/anshul-si/4/base 2025-11-03T16:58:01.3936392Z * [new branch] gh/anshul-si/4/head -> origin/gh/anshul-si/4/head 2025-11-03T16:58:01.3938633Z * [new branch] gh/anshul-si/5/base -> origin/gh/anshul-si/5/base 2025-11-03T16:58:01.3940306Z * [new branch] gh/anshul-si/5/head -> origin/gh/anshul-si/5/head 2025-11-03T16:58:01.3942851Z * [new branch] gh/anshul-si/51/base -> origin/gh/anshul-si/51/base 2025-11-03T16:58:01.3944536Z * [new branch] gh/anshul-si/51/head -> origin/gh/anshul-si/51/head 2025-11-03T16:58:01.3946238Z * [new branch] gh/anshul-si/51/orig -> origin/gh/anshul-si/51/orig 2025-11-03T16:58:01.3948429Z * [new branch] gh/anshul-si/52/base -> origin/gh/anshul-si/52/base 2025-11-03T16:58:01.3950215Z * [new branch] gh/anshul-si/52/head -> origin/gh/anshul-si/52/head 2025-11-03T16:58:01.3951920Z * [new branch] gh/anshul-si/52/orig -> origin/gh/anshul-si/52/orig 2025-11-03T16:58:01.3954415Z * [new branch] gh/anshul-si/53/base -> origin/gh/anshul-si/53/base 2025-11-03T16:58:01.3956332Z * [new branch] gh/anshul-si/53/head -> origin/gh/anshul-si/53/head 2025-11-03T16:58:01.3958427Z * [new branch] gh/anshul-si/54/base -> origin/gh/anshul-si/54/base 2025-11-03T16:58:01.3960007Z * [new branch] gh/anshul-si/54/head -> origin/gh/anshul-si/54/head 2025-11-03T16:58:01.3961682Z * [new branch] gh/anshul-si/54/orig -> origin/gh/anshul-si/54/orig 2025-11-03T16:58:01.3963909Z * [new branch] gh/anshul-si/55/base -> origin/gh/anshul-si/55/base 2025-11-03T16:58:01.3965634Z * [new branch] gh/anshul-si/55/head -> origin/gh/anshul-si/55/head 2025-11-03T16:58:01.3967348Z * [new branch] gh/anshul-si/55/orig -> origin/gh/anshul-si/55/orig 2025-11-03T16:58:01.3969737Z * [new branch] gh/anshul-si/56/base -> origin/gh/anshul-si/56/base 2025-11-03T16:58:01.3971327Z * [new branch] gh/anshul-si/56/head -> origin/gh/anshul-si/56/head 2025-11-03T16:58:01.3972953Z * [new branch] gh/anshul-si/56/orig -> origin/gh/anshul-si/56/orig 2025-11-03T16:58:01.3975277Z * [new branch] gh/anshul-si/57/base -> origin/gh/anshul-si/57/base 2025-11-03T16:58:01.3976944Z * [new branch] gh/anshul-si/57/head -> origin/gh/anshul-si/57/head 2025-11-03T16:58:01.3978789Z * [new branch] gh/anshul-si/57/orig -> origin/gh/anshul-si/57/orig 2025-11-03T16:58:01.3980926Z * [new branch] gh/anshul-si/58/base -> origin/gh/anshul-si/58/base 2025-11-03T16:58:01.3982959Z * [new branch] gh/anshul-si/58/head -> origin/gh/anshul-si/58/head 2025-11-03T16:58:01.3985148Z * [new branch] gh/anshul-si/59/base -> origin/gh/anshul-si/59/base 2025-11-03T16:58:01.3986899Z * [new branch] gh/anshul-si/59/head -> origin/gh/anshul-si/59/head 2025-11-03T16:58:01.3988633Z * [new branch] gh/anshul-si/59/orig -> origin/gh/anshul-si/59/orig 2025-11-03T16:58:01.3990903Z * [new branch] gh/anshul-si/60/base -> origin/gh/anshul-si/60/base 2025-11-03T16:58:01.3992609Z * [new branch] gh/anshul-si/60/head -> origin/gh/anshul-si/60/head 2025-11-03T16:58:01.3994299Z * [new branch] gh/anshul-si/60/orig -> origin/gh/anshul-si/60/orig 2025-11-03T16:58:01.3996708Z * [new branch] gh/anshul-si/61/base -> origin/gh/anshul-si/61/base 2025-11-03T16:58:01.3998477Z * [new branch] gh/anshul-si/61/head -> origin/gh/anshul-si/61/head 2025-11-03T16:58:01.4000238Z * [new branch] gh/anshul-si/61/orig -> origin/gh/anshul-si/61/orig 2025-11-03T16:58:01.4002494Z * [new branch] gh/anshul-si/62/base -> origin/gh/anshul-si/62/base 2025-11-03T16:58:01.4004479Z * [new branch] gh/anshul-si/62/head -> origin/gh/anshul-si/62/head 2025-11-03T16:58:01.4005867Z * [new branch] gh/anshul-si/62/orig -> origin/gh/anshul-si/62/orig 2025-11-03T16:58:01.4008023Z * [new branch] gh/anshul-si/63/base -> origin/gh/anshul-si/63/base 2025-11-03T16:58:01.4009807Z * [new branch] gh/anshul-si/63/head -> origin/gh/anshul-si/63/head 2025-11-03T16:58:01.4012788Z * [new branch] gh/anshul-si/63/orig -> origin/gh/anshul-si/63/orig 2025-11-03T16:58:01.4014385Z * [new branch] gh/anshul-si/64/base -> origin/gh/anshul-si/64/base 2025-11-03T16:58:01.4015896Z * [new branch] gh/anshul-si/64/head -> origin/gh/anshul-si/64/head 2025-11-03T16:58:01.4017582Z * [new branch] gh/anshul-si/64/orig -> origin/gh/anshul-si/64/orig 2025-11-03T16:58:01.4020250Z * [new branch] gh/anshul-si/65/base -> origin/gh/anshul-si/65/base 2025-11-03T16:58:01.4022178Z * [new branch] gh/anshul-si/65/head -> origin/gh/anshul-si/65/head 2025-11-03T16:58:01.4023920Z * [new branch] gh/anshul-si/65/orig -> origin/gh/anshul-si/65/orig 2025-11-03T16:58:01.4026884Z * [new branch] gh/aorenste/132/base -> origin/gh/aorenste/132/base 2025-11-03T16:58:01.4028527Z * [new branch] gh/aorenste/132/head -> origin/gh/aorenste/132/head 2025-11-03T16:58:01.4030994Z * [new branch] gh/aorenste/133/base -> origin/gh/aorenste/133/base 2025-11-03T16:58:01.4032744Z * [new branch] gh/aorenste/133/head -> origin/gh/aorenste/133/head 2025-11-03T16:58:01.4034396Z * [new branch] gh/aorenste/133/orig -> origin/gh/aorenste/133/orig 2025-11-03T16:58:01.4036958Z * [new branch] gh/aorenste/134/base -> origin/gh/aorenste/134/base 2025-11-03T16:58:01.4039035Z * [new branch] gh/aorenste/134/head -> origin/gh/aorenste/134/head 2025-11-03T16:58:01.4040848Z * [new branch] gh/aorenste/134/orig -> origin/gh/aorenste/134/orig 2025-11-03T16:58:01.4043247Z * [new branch] gh/aorenste/135/base -> origin/gh/aorenste/135/base 2025-11-03T16:58:01.4044929Z * [new branch] gh/aorenste/135/head -> origin/gh/aorenste/135/head 2025-11-03T16:58:01.4046660Z * [new branch] gh/aorenste/135/orig -> origin/gh/aorenste/135/orig 2025-11-03T16:58:01.4049395Z * [new branch] gh/aorenste/136/base -> origin/gh/aorenste/136/base 2025-11-03T16:58:01.4051387Z * [new branch] gh/aorenste/136/head -> origin/gh/aorenste/136/head 2025-11-03T16:58:01.4053019Z * [new branch] gh/aorenste/136/orig -> origin/gh/aorenste/136/orig 2025-11-03T16:58:01.4055511Z * [new branch] gh/aorenste/137/base -> origin/gh/aorenste/137/base 2025-11-03T16:58:01.4057265Z * [new branch] gh/aorenste/137/head -> origin/gh/aorenste/137/head 2025-11-03T16:58:01.4058975Z * [new branch] gh/aorenste/137/orig -> origin/gh/aorenste/137/orig 2025-11-03T16:58:01.4061357Z * [new branch] gh/aorenste/138/base -> origin/gh/aorenste/138/base 2025-11-03T16:58:01.4063244Z * [new branch] gh/aorenste/138/head -> origin/gh/aorenste/138/head 2025-11-03T16:58:01.4064862Z * [new branch] gh/aorenste/138/orig -> origin/gh/aorenste/138/orig 2025-11-03T16:58:01.4067304Z * [new branch] gh/aorenste/139/base -> origin/gh/aorenste/139/base 2025-11-03T16:58:01.4068948Z * [new branch] gh/aorenste/139/head -> origin/gh/aorenste/139/head 2025-11-03T16:58:01.4070682Z * [new branch] gh/aorenste/139/orig -> origin/gh/aorenste/139/orig 2025-11-03T16:58:01.4073161Z * [new branch] gh/aorenste/140/base -> origin/gh/aorenste/140/base 2025-11-03T16:58:01.4074887Z * [new branch] gh/aorenste/140/head -> origin/gh/aorenste/140/head 2025-11-03T16:58:01.4076608Z * [new branch] gh/aorenste/140/orig -> origin/gh/aorenste/140/orig 2025-11-03T16:58:01.4078907Z * [new branch] gh/aorenste/141/base -> origin/gh/aorenste/141/base 2025-11-03T16:58:01.4080606Z * [new branch] gh/aorenste/141/head -> origin/gh/aorenste/141/head 2025-11-03T16:58:01.4083006Z * [new branch] gh/aorenste/142/base -> origin/gh/aorenste/142/base 2025-11-03T16:58:01.4084871Z * [new branch] gh/aorenste/142/head -> origin/gh/aorenste/142/head 2025-11-03T16:58:01.4086541Z * [new branch] gh/aorenste/142/orig -> origin/gh/aorenste/142/orig 2025-11-03T16:58:01.4088960Z * [new branch] gh/aorenste/143/base -> origin/gh/aorenste/143/base 2025-11-03T16:58:01.4090566Z * [new branch] gh/aorenste/143/head -> origin/gh/aorenste/143/head 2025-11-03T16:58:01.4092322Z * [new branch] gh/aorenste/143/orig -> origin/gh/aorenste/143/orig 2025-11-03T16:58:01.4095026Z * [new branch] gh/aorenste/144/base -> origin/gh/aorenste/144/base 2025-11-03T16:58:01.4096700Z * [new branch] gh/aorenste/144/head -> origin/gh/aorenste/144/head 2025-11-03T16:58:01.4098393Z * [new branch] gh/aorenste/144/orig -> origin/gh/aorenste/144/orig 2025-11-03T16:58:01.4100812Z * [new branch] gh/aorenste/145/base -> origin/gh/aorenste/145/base 2025-11-03T16:58:01.4102673Z * [new branch] gh/aorenste/145/head -> origin/gh/aorenste/145/head 2025-11-03T16:58:01.4104460Z * [new branch] gh/aorenste/145/orig -> origin/gh/aorenste/145/orig 2025-11-03T16:58:01.4107313Z * [new branch] gh/avikchaudhuri/1/base -> origin/gh/avikchaudhuri/1/base 2025-11-03T16:58:01.4109085Z * [new branch] gh/avikchaudhuri/1/head -> origin/gh/avikchaudhuri/1/head 2025-11-03T16:58:01.4111268Z * [new branch] gh/avikchaudhuri/2/base -> origin/gh/avikchaudhuri/2/base 2025-11-03T16:58:01.4112984Z * [new branch] gh/avikchaudhuri/2/head -> origin/gh/avikchaudhuri/2/head 2025-11-03T16:58:01.4114622Z * [new branch] gh/avikchaudhuri/2/orig -> origin/gh/avikchaudhuri/2/orig 2025-11-03T16:58:01.4117741Z * [new branch] gh/bdhirsh/666/base -> origin/gh/bdhirsh/666/base 2025-11-03T16:58:01.4119476Z * [new branch] gh/bdhirsh/666/head -> origin/gh/bdhirsh/666/head 2025-11-03T16:58:01.4121452Z * [new branch] gh/bdhirsh/666/orig -> origin/gh/bdhirsh/666/orig 2025-11-03T16:58:01.4123775Z * [new branch] gh/bdhirsh/668/base -> origin/gh/bdhirsh/668/base 2025-11-03T16:58:01.4125479Z * [new branch] gh/bdhirsh/668/head -> origin/gh/bdhirsh/668/head 2025-11-03T16:58:01.4127106Z * [new branch] gh/bdhirsh/668/orig -> origin/gh/bdhirsh/668/orig 2025-11-03T16:58:01.4129581Z * [new branch] gh/bdhirsh/669/base -> origin/gh/bdhirsh/669/base 2025-11-03T16:58:01.4131220Z * [new branch] gh/bdhirsh/669/head -> origin/gh/bdhirsh/669/head 2025-11-03T16:58:01.4132920Z * [new branch] gh/bdhirsh/669/orig -> origin/gh/bdhirsh/669/orig 2025-11-03T16:58:01.4135422Z * [new branch] gh/bdhirsh/670/base -> origin/gh/bdhirsh/670/base 2025-11-03T16:58:01.4137193Z * [new branch] gh/bdhirsh/670/head -> origin/gh/bdhirsh/670/head 2025-11-03T16:58:01.4138892Z * [new branch] gh/bdhirsh/670/orig -> origin/gh/bdhirsh/670/orig 2025-11-03T16:58:01.4141388Z * [new branch] gh/bdhirsh/671/base -> origin/gh/bdhirsh/671/base 2025-11-03T16:58:01.4143418Z * [new branch] gh/bdhirsh/671/head -> origin/gh/bdhirsh/671/head 2025-11-03T16:58:01.4145105Z * [new branch] gh/bdhirsh/671/orig -> origin/gh/bdhirsh/671/orig 2025-11-03T16:58:01.4147484Z * [new branch] gh/bdhirsh/672/base -> origin/gh/bdhirsh/672/base 2025-11-03T16:58:01.4149375Z * [new branch] gh/bdhirsh/672/head -> origin/gh/bdhirsh/672/head 2025-11-03T16:58:01.4150956Z * [new branch] gh/bdhirsh/672/orig -> origin/gh/bdhirsh/672/orig 2025-11-03T16:58:01.4153271Z * [new branch] gh/bdhirsh/673/base -> origin/gh/bdhirsh/673/base 2025-11-03T16:58:01.4155027Z * [new branch] gh/bdhirsh/673/head -> origin/gh/bdhirsh/673/head 2025-11-03T16:58:01.4156704Z * [new branch] gh/bdhirsh/673/orig -> origin/gh/bdhirsh/673/orig 2025-11-03T16:58:01.4159192Z * [new branch] gh/bdhirsh/674/base -> origin/gh/bdhirsh/674/base 2025-11-03T16:58:01.4160925Z * [new branch] gh/bdhirsh/674/head -> origin/gh/bdhirsh/674/head 2025-11-03T16:58:01.4162686Z * [new branch] gh/bdhirsh/674/orig -> origin/gh/bdhirsh/674/orig 2025-11-03T16:58:01.4165221Z * [new branch] gh/bdhirsh/675/base -> origin/gh/bdhirsh/675/base 2025-11-03T16:58:01.4166784Z * [new branch] gh/bdhirsh/675/head -> origin/gh/bdhirsh/675/head 2025-11-03T16:58:01.4168444Z * [new branch] gh/bdhirsh/675/orig -> origin/gh/bdhirsh/675/orig 2025-11-03T16:58:01.4171310Z * [new branch] gh/benjaminglass1/101/base -> origin/gh/benjaminglass1/101/base 2025-11-03T16:58:01.4172987Z * [new branch] gh/benjaminglass1/101/head -> origin/gh/benjaminglass1/101/head 2025-11-03T16:58:01.4175173Z * [new branch] gh/benjaminglass1/101/orig -> origin/gh/benjaminglass1/101/orig 2025-11-03T16:58:01.4177148Z * [new branch] gh/benjaminglass1/102/base -> origin/gh/benjaminglass1/102/base 2025-11-03T16:58:01.4178834Z * [new branch] gh/benjaminglass1/102/head -> origin/gh/benjaminglass1/102/head 2025-11-03T16:58:01.4180526Z * [new branch] gh/benjaminglass1/102/orig -> origin/gh/benjaminglass1/102/orig 2025-11-03T16:58:01.4183112Z * [new branch] gh/benjaminglass1/106/base -> origin/gh/benjaminglass1/106/base 2025-11-03T16:58:01.4184775Z * [new branch] gh/benjaminglass1/106/head -> origin/gh/benjaminglass1/106/head 2025-11-03T16:58:01.4186513Z * [new branch] gh/benjaminglass1/106/orig -> origin/gh/benjaminglass1/106/orig 2025-11-03T16:58:01.4188799Z * [new branch] gh/benjaminglass1/107/base -> origin/gh/benjaminglass1/107/base 2025-11-03T16:58:01.4190511Z * [new branch] gh/benjaminglass1/107/head -> origin/gh/benjaminglass1/107/head 2025-11-03T16:58:01.4192199Z * [new branch] gh/benjaminglass1/107/orig -> origin/gh/benjaminglass1/107/orig 2025-11-03T16:58:01.4194491Z * [new branch] gh/benjaminglass1/108/base -> origin/gh/benjaminglass1/108/base 2025-11-03T16:58:01.4196170Z * [new branch] gh/benjaminglass1/108/head -> origin/gh/benjaminglass1/108/head 2025-11-03T16:58:01.4197840Z * [new branch] gh/benjaminglass1/108/orig -> origin/gh/benjaminglass1/108/orig 2025-11-03T16:58:01.4200210Z * [new branch] gh/benjaminglass1/79/base -> origin/gh/benjaminglass1/79/base 2025-11-03T16:58:01.4201913Z * [new branch] gh/benjaminglass1/79/head -> origin/gh/benjaminglass1/79/head 2025-11-03T16:58:01.4203629Z * [new branch] gh/benjaminglass1/79/orig -> origin/gh/benjaminglass1/79/orig 2025-11-03T16:58:01.4205928Z * [new branch] gh/benjaminglass1/86/base -> origin/gh/benjaminglass1/86/base 2025-11-03T16:58:01.4207593Z * [new branch] gh/benjaminglass1/86/head -> origin/gh/benjaminglass1/86/head 2025-11-03T16:58:01.4209397Z * [new branch] gh/benjaminglass1/86/orig -> origin/gh/benjaminglass1/86/orig 2025-11-03T16:58:01.4211727Z * [new branch] gh/benjaminglass1/97/base -> origin/gh/benjaminglass1/97/base 2025-11-03T16:58:01.4213316Z * [new branch] gh/benjaminglass1/97/head -> origin/gh/benjaminglass1/97/head 2025-11-03T16:58:01.4215008Z * [new branch] gh/benjaminglass1/97/orig -> origin/gh/benjaminglass1/97/orig 2025-11-03T16:58:01.4217854Z * [new branch] gh/bobrenjc93/570/base -> origin/gh/bobrenjc93/570/base 2025-11-03T16:58:01.4219561Z * [new branch] gh/bobrenjc93/570/head -> origin/gh/bobrenjc93/570/head 2025-11-03T16:58:01.4221404Z * [new branch] gh/bobrenjc93/570/orig -> origin/gh/bobrenjc93/570/orig 2025-11-03T16:58:01.4225738Z * [new branch] gh/bobrenjc93/604/base -> origin/gh/bobrenjc93/604/base 2025-11-03T16:58:01.4227378Z * [new branch] gh/bobrenjc93/604/head -> origin/gh/bobrenjc93/604/head 2025-11-03T16:58:01.4229126Z * [new branch] gh/bobrenjc93/604/orig -> origin/gh/bobrenjc93/604/orig 2025-11-03T16:58:01.4231588Z * [new branch] gh/bobrenjc93/625/base -> origin/gh/bobrenjc93/625/base 2025-11-03T16:58:01.4233602Z * [new branch] gh/bobrenjc93/625/head -> origin/gh/bobrenjc93/625/head 2025-11-03T16:58:01.4247168Z * [new branch] gh/bobrenjc93/625/orig -> origin/gh/bobrenjc93/625/orig 2025-11-03T16:58:01.4247636Z * [new branch] gh/bobrenjc93/626/base -> origin/gh/bobrenjc93/626/base 2025-11-03T16:58:01.4248037Z * [new branch] gh/bobrenjc93/626/head -> origin/gh/bobrenjc93/626/head 2025-11-03T16:58:01.4248430Z * [new branch] gh/bobrenjc93/626/orig -> origin/gh/bobrenjc93/626/orig 2025-11-03T16:58:01.4248806Z * [new branch] gh/bobrenjc93/627/base -> origin/gh/bobrenjc93/627/base 2025-11-03T16:58:01.4249185Z * [new branch] gh/bobrenjc93/627/head -> origin/gh/bobrenjc93/627/head 2025-11-03T16:58:01.4249578Z * [new branch] gh/bobrenjc93/627/orig -> origin/gh/bobrenjc93/627/orig 2025-11-03T16:58:01.4249972Z * [new branch] gh/bobrenjc93/630/base -> origin/gh/bobrenjc93/630/base 2025-11-03T16:58:01.4251148Z * [new branch] gh/bobrenjc93/630/head -> origin/gh/bobrenjc93/630/head 2025-11-03T16:58:01.4252956Z * [new branch] gh/bobrenjc93/630/orig -> origin/gh/bobrenjc93/630/orig 2025-11-03T16:58:01.4255159Z * [new branch] gh/bobrenjc93/631/base -> origin/gh/bobrenjc93/631/base 2025-11-03T16:58:01.4256926Z * [new branch] gh/bobrenjc93/631/head -> origin/gh/bobrenjc93/631/head 2025-11-03T16:58:01.4258600Z * [new branch] gh/bobrenjc93/631/orig -> origin/gh/bobrenjc93/631/orig 2025-11-03T16:58:01.4260952Z * [new branch] gh/bobrenjc93/632/base -> origin/gh/bobrenjc93/632/base 2025-11-03T16:58:01.4262940Z * [new branch] gh/bobrenjc93/632/head -> origin/gh/bobrenjc93/632/head 2025-11-03T16:58:01.4264551Z * [new branch] gh/bobrenjc93/632/orig -> origin/gh/bobrenjc93/632/orig 2025-11-03T16:58:01.4266815Z * [new branch] gh/bobrenjc93/633/base -> origin/gh/bobrenjc93/633/base 2025-11-03T16:58:01.4268557Z * [new branch] gh/bobrenjc93/633/head -> origin/gh/bobrenjc93/633/head 2025-11-03T16:58:01.4270254Z * [new branch] gh/bobrenjc93/633/orig -> origin/gh/bobrenjc93/633/orig 2025-11-03T16:58:01.4272425Z * [new branch] gh/bobrenjc93/634/base -> origin/gh/bobrenjc93/634/base 2025-11-03T16:58:01.4274132Z * [new branch] gh/bobrenjc93/634/head -> origin/gh/bobrenjc93/634/head 2025-11-03T16:58:01.4275820Z * [new branch] gh/bobrenjc93/634/orig -> origin/gh/bobrenjc93/634/orig 2025-11-03T16:58:01.4278036Z * [new branch] gh/bobrenjc93/635/base -> origin/gh/bobrenjc93/635/base 2025-11-03T16:58:01.4279787Z * [new branch] gh/bobrenjc93/635/head -> origin/gh/bobrenjc93/635/head 2025-11-03T16:58:01.4281485Z * [new branch] gh/bobrenjc93/635/orig -> origin/gh/bobrenjc93/635/orig 2025-11-03T16:58:01.4283790Z * [new branch] gh/bobrenjc93/636/base -> origin/gh/bobrenjc93/636/base 2025-11-03T16:58:01.4285512Z * [new branch] gh/bobrenjc93/636/head -> origin/gh/bobrenjc93/636/head 2025-11-03T16:58:01.4287179Z * [new branch] gh/bobrenjc93/636/orig -> origin/gh/bobrenjc93/636/orig 2025-11-03T16:58:01.4289535Z * [new branch] gh/bobrenjc93/637/base -> origin/gh/bobrenjc93/637/base 2025-11-03T16:58:01.4291168Z * [new branch] gh/bobrenjc93/637/head -> origin/gh/bobrenjc93/637/head 2025-11-03T16:58:01.4292902Z * [new branch] gh/bobrenjc93/637/orig -> origin/gh/bobrenjc93/637/orig 2025-11-03T16:58:01.4295213Z * [new branch] gh/bobrenjc93/638/base -> origin/gh/bobrenjc93/638/base 2025-11-03T16:58:01.4296885Z * [new branch] gh/bobrenjc93/638/head -> origin/gh/bobrenjc93/638/head 2025-11-03T16:58:01.4298874Z * [new branch] gh/bobrenjc93/638/orig -> origin/gh/bobrenjc93/638/orig 2025-11-03T16:58:01.4301122Z * [new branch] gh/bobrenjc93/639/base -> origin/gh/bobrenjc93/639/base 2025-11-03T16:58:01.4303154Z * [new branch] gh/bobrenjc93/639/head -> origin/gh/bobrenjc93/639/head 2025-11-03T16:58:01.4304826Z * [new branch] gh/bobrenjc93/639/orig -> origin/gh/bobrenjc93/639/orig 2025-11-03T16:58:01.4307110Z * [new branch] gh/bobrenjc93/640/base -> origin/gh/bobrenjc93/640/base 2025-11-03T16:58:01.4308771Z * [new branch] gh/bobrenjc93/640/head -> origin/gh/bobrenjc93/640/head 2025-11-03T16:58:01.4310445Z * [new branch] gh/bobrenjc93/640/orig -> origin/gh/bobrenjc93/640/orig 2025-11-03T16:58:01.4313214Z * [new branch] gh/bobrenjc93/641/base -> origin/gh/bobrenjc93/641/base 2025-11-03T16:58:01.4314946Z * [new branch] gh/bobrenjc93/641/head -> origin/gh/bobrenjc93/641/head 2025-11-03T16:58:01.4316673Z * [new branch] gh/bobrenjc93/641/orig -> origin/gh/bobrenjc93/641/orig 2025-11-03T16:58:01.4319083Z * [new branch] gh/bobrenjc93/642/base -> origin/gh/bobrenjc93/642/base 2025-11-03T16:58:01.4320765Z * [new branch] gh/bobrenjc93/642/head -> origin/gh/bobrenjc93/642/head 2025-11-03T16:58:01.4322795Z * [new branch] gh/bobrenjc93/642/orig -> origin/gh/bobrenjc93/642/orig 2025-11-03T16:58:01.4325081Z * [new branch] gh/bobrenjc93/643/base -> origin/gh/bobrenjc93/643/base 2025-11-03T16:58:01.4326829Z * [new branch] gh/bobrenjc93/643/head -> origin/gh/bobrenjc93/643/head 2025-11-03T16:58:01.4328519Z * [new branch] gh/bobrenjc93/643/orig -> origin/gh/bobrenjc93/643/orig 2025-11-03T16:58:01.4330844Z * [new branch] gh/bobrenjc93/644/base -> origin/gh/bobrenjc93/644/base 2025-11-03T16:58:01.4332593Z * [new branch] gh/bobrenjc93/644/head -> origin/gh/bobrenjc93/644/head 2025-11-03T16:58:01.4334276Z * [new branch] gh/bobrenjc93/644/orig -> origin/gh/bobrenjc93/644/orig 2025-11-03T16:58:01.4336640Z * [new branch] gh/bobrenjc93/645/base -> origin/gh/bobrenjc93/645/base 2025-11-03T16:58:01.4338369Z * [new branch] gh/bobrenjc93/645/head -> origin/gh/bobrenjc93/645/head 2025-11-03T16:58:01.4339987Z * [new branch] gh/bobrenjc93/645/orig -> origin/gh/bobrenjc93/645/orig 2025-11-03T16:58:01.4342568Z * [new branch] gh/bobrenjc93/646/base -> origin/gh/bobrenjc93/646/base 2025-11-03T16:58:01.4344184Z * [new branch] gh/bobrenjc93/646/head -> origin/gh/bobrenjc93/646/head 2025-11-03T16:58:01.4345843Z * [new branch] gh/bobrenjc93/646/orig -> origin/gh/bobrenjc93/646/orig 2025-11-03T16:58:01.4348416Z * [new branch] gh/bobrenjc93/647/base -> origin/gh/bobrenjc93/647/base 2025-11-03T16:58:01.4350239Z * [new branch] gh/bobrenjc93/647/head -> origin/gh/bobrenjc93/647/head 2025-11-03T16:58:01.4351942Z * [new branch] gh/bobrenjc93/647/orig -> origin/gh/bobrenjc93/647/orig 2025-11-03T16:58:01.4354794Z * [new branch] gh/bobrenjc93/648/base -> origin/gh/bobrenjc93/648/base 2025-11-03T16:58:01.4356462Z * [new branch] gh/bobrenjc93/648/head -> origin/gh/bobrenjc93/648/head 2025-11-03T16:58:01.4358116Z * [new branch] gh/bobrenjc93/648/orig -> origin/gh/bobrenjc93/648/orig 2025-11-03T16:58:01.4360525Z * [new branch] gh/bobrenjc93/649/base -> origin/gh/bobrenjc93/649/base 2025-11-03T16:58:01.4362235Z * [new branch] gh/bobrenjc93/649/head -> origin/gh/bobrenjc93/649/head 2025-11-03T16:58:01.4363938Z * [new branch] gh/bobrenjc93/649/orig -> origin/gh/bobrenjc93/649/orig 2025-11-03T16:58:01.4366481Z * [new branch] gh/bobrenjc93/650/base -> origin/gh/bobrenjc93/650/base 2025-11-03T16:58:01.4368057Z * [new branch] gh/bobrenjc93/650/head -> origin/gh/bobrenjc93/650/head 2025-11-03T16:58:01.4369676Z * [new branch] gh/bobrenjc93/650/orig -> origin/gh/bobrenjc93/650/orig 2025-11-03T16:58:01.4372479Z * [new branch] gh/bobrenjc93/651/base -> origin/gh/bobrenjc93/651/base 2025-11-03T16:58:01.4374316Z * [new branch] gh/bobrenjc93/651/head -> origin/gh/bobrenjc93/651/head 2025-11-03T16:58:01.4375990Z * [new branch] gh/bobrenjc93/651/orig -> origin/gh/bobrenjc93/651/orig 2025-11-03T16:58:01.4378277Z * [new branch] gh/bobrenjc93/652/base -> origin/gh/bobrenjc93/652/base 2025-11-03T16:58:01.4379942Z * [new branch] gh/bobrenjc93/652/head -> origin/gh/bobrenjc93/652/head 2025-11-03T16:58:01.4381894Z * [new branch] gh/bobrenjc93/652/orig -> origin/gh/bobrenjc93/652/orig 2025-11-03T16:58:01.4384496Z * [new branch] gh/bobrenjc93/653/base -> origin/gh/bobrenjc93/653/base 2025-11-03T16:58:01.4386190Z * [new branch] gh/bobrenjc93/653/head -> origin/gh/bobrenjc93/653/head 2025-11-03T16:58:01.4387842Z * [new branch] gh/bobrenjc93/653/orig -> origin/gh/bobrenjc93/653/orig 2025-11-03T16:58:01.4390267Z * [new branch] gh/bobrenjc93/654/base -> origin/gh/bobrenjc93/654/base 2025-11-03T16:58:01.4391924Z * [new branch] gh/bobrenjc93/654/head -> origin/gh/bobrenjc93/654/head 2025-11-03T16:58:01.4393619Z * [new branch] gh/bobrenjc93/654/orig -> origin/gh/bobrenjc93/654/orig 2025-11-03T16:58:01.4395986Z * [new branch] gh/bobrenjc93/655/base -> origin/gh/bobrenjc93/655/base 2025-11-03T16:58:01.4397736Z * [new branch] gh/bobrenjc93/655/head -> origin/gh/bobrenjc93/655/head 2025-11-03T16:58:01.4399514Z * [new branch] gh/bobrenjc93/655/orig -> origin/gh/bobrenjc93/655/orig 2025-11-03T16:58:01.4401835Z * [new branch] gh/bobrenjc93/656/base -> origin/gh/bobrenjc93/656/base 2025-11-03T16:58:01.4403429Z * [new branch] gh/bobrenjc93/656/head -> origin/gh/bobrenjc93/656/head 2025-11-03T16:58:01.4405152Z * [new branch] gh/bobrenjc93/656/orig -> origin/gh/bobrenjc93/656/orig 2025-11-03T16:58:01.4407491Z * [new branch] gh/bobrenjc93/657/base -> origin/gh/bobrenjc93/657/base 2025-11-03T16:58:01.4409204Z * [new branch] gh/bobrenjc93/657/head -> origin/gh/bobrenjc93/657/head 2025-11-03T16:58:01.4410910Z * [new branch] gh/bobrenjc93/657/orig -> origin/gh/bobrenjc93/657/orig 2025-11-03T16:58:01.4413238Z * [new branch] gh/bobrenjc93/658/base -> origin/gh/bobrenjc93/658/base 2025-11-03T16:58:01.4415028Z * [new branch] gh/bobrenjc93/658/head -> origin/gh/bobrenjc93/658/head 2025-11-03T16:58:01.4416673Z * [new branch] gh/bobrenjc93/658/orig -> origin/gh/bobrenjc93/658/orig 2025-11-03T16:58:01.4419056Z * [new branch] gh/bobrenjc93/659/base -> origin/gh/bobrenjc93/659/base 2025-11-03T16:58:01.4420873Z * [new branch] gh/bobrenjc93/659/head -> origin/gh/bobrenjc93/659/head 2025-11-03T16:58:01.4423075Z * [new branch] gh/bobrenjc93/659/orig -> origin/gh/bobrenjc93/659/orig 2025-11-03T16:58:01.4425955Z * [new branch] gh/bobrenjc93/660/base -> origin/gh/bobrenjc93/660/base 2025-11-03T16:58:01.4427632Z * [new branch] gh/bobrenjc93/660/head -> origin/gh/bobrenjc93/660/head 2025-11-03T16:58:01.4429264Z * [new branch] gh/bobrenjc93/660/orig -> origin/gh/bobrenjc93/660/orig 2025-11-03T16:58:01.4431637Z * [new branch] gh/bobrenjc93/661/base -> origin/gh/bobrenjc93/661/base 2025-11-03T16:58:01.4433316Z * [new branch] gh/bobrenjc93/661/head -> origin/gh/bobrenjc93/661/head 2025-11-03T16:58:01.4435204Z * [new branch] gh/bobrenjc93/661/orig -> origin/gh/bobrenjc93/661/orig 2025-11-03T16:58:01.4437465Z * [new branch] gh/bobrenjc93/662/base -> origin/gh/bobrenjc93/662/base 2025-11-03T16:58:01.4439111Z * [new branch] gh/bobrenjc93/662/head -> origin/gh/bobrenjc93/662/head 2025-11-03T16:58:01.4440757Z * [new branch] gh/bobrenjc93/662/orig -> origin/gh/bobrenjc93/662/orig 2025-11-03T16:58:01.4443137Z * [new branch] gh/bobrenjc93/663/base -> origin/gh/bobrenjc93/663/base 2025-11-03T16:58:01.4444930Z * [new branch] gh/bobrenjc93/663/head -> origin/gh/bobrenjc93/663/head 2025-11-03T16:58:01.4446680Z * [new branch] gh/bobrenjc93/663/orig -> origin/gh/bobrenjc93/663/orig 2025-11-03T16:58:01.4449559Z * [new branch] gh/bobrenjc93/664/base -> origin/gh/bobrenjc93/664/base 2025-11-03T16:58:01.4451546Z * [new branch] gh/bobrenjc93/664/head -> origin/gh/bobrenjc93/664/head 2025-11-03T16:58:01.4453264Z * [new branch] gh/bobrenjc93/664/orig -> origin/gh/bobrenjc93/664/orig 2025-11-03T16:58:01.4456424Z * [new branch] gh/bobrenjc93/665/base -> origin/gh/bobrenjc93/665/base 2025-11-03T16:58:01.4457970Z * [new branch] gh/bobrenjc93/665/head -> origin/gh/bobrenjc93/665/head 2025-11-03T16:58:01.4459647Z * [new branch] gh/bobrenjc93/665/orig -> origin/gh/bobrenjc93/665/orig 2025-11-03T16:58:01.4462396Z * [new branch] gh/bobrenjc93/666/base -> origin/gh/bobrenjc93/666/base 2025-11-03T16:58:01.4464174Z * [new branch] gh/bobrenjc93/666/head -> origin/gh/bobrenjc93/666/head 2025-11-03T16:58:01.4465883Z * [new branch] gh/bobrenjc93/666/orig -> origin/gh/bobrenjc93/666/orig 2025-11-03T16:58:01.4468295Z * [new branch] gh/bobrenjc93/667/base -> origin/gh/bobrenjc93/667/base 2025-11-03T16:58:01.4470144Z * [new branch] gh/bobrenjc93/667/head -> origin/gh/bobrenjc93/667/head 2025-11-03T16:58:01.4471858Z * [new branch] gh/bobrenjc93/667/orig -> origin/gh/bobrenjc93/667/orig 2025-11-03T16:58:01.4474560Z * [new branch] gh/bobrenjc93/668/base -> origin/gh/bobrenjc93/668/base 2025-11-03T16:58:01.4476276Z * [new branch] gh/bobrenjc93/668/head -> origin/gh/bobrenjc93/668/head 2025-11-03T16:58:01.4478089Z * [new branch] gh/bobrenjc93/668/orig -> origin/gh/bobrenjc93/668/orig 2025-11-03T16:58:01.4480474Z * [new branch] gh/bobrenjc93/669/base -> origin/gh/bobrenjc93/669/base 2025-11-03T16:58:01.4482175Z * [new branch] gh/bobrenjc93/669/head -> origin/gh/bobrenjc93/669/head 2025-11-03T16:58:01.4483878Z * [new branch] gh/bobrenjc93/669/orig -> origin/gh/bobrenjc93/669/orig 2025-11-03T16:58:01.4486304Z * [new branch] gh/bobrenjc93/670/base -> origin/gh/bobrenjc93/670/base 2025-11-03T16:58:01.4487939Z * [new branch] gh/bobrenjc93/670/head -> origin/gh/bobrenjc93/670/head 2025-11-03T16:58:01.4489694Z * [new branch] gh/bobrenjc93/670/orig -> origin/gh/bobrenjc93/670/orig 2025-11-03T16:58:01.4492539Z * [new branch] gh/bobrenjc93/671/base -> origin/gh/bobrenjc93/671/base 2025-11-03T16:58:01.4494260Z * [new branch] gh/bobrenjc93/671/head -> origin/gh/bobrenjc93/671/head 2025-11-03T16:58:01.4496038Z * [new branch] gh/bobrenjc93/671/orig -> origin/gh/bobrenjc93/671/orig 2025-11-03T16:58:01.4498432Z * [new branch] gh/bobrenjc93/672/base -> origin/gh/bobrenjc93/672/base 2025-11-03T16:58:01.4500082Z * [new branch] gh/bobrenjc93/672/head -> origin/gh/bobrenjc93/672/head 2025-11-03T16:58:01.4501798Z * [new branch] gh/bobrenjc93/672/orig -> origin/gh/bobrenjc93/672/orig 2025-11-03T16:58:01.4504520Z * [new branch] gh/bobrenjc93/673/base -> origin/gh/bobrenjc93/673/base 2025-11-03T16:58:01.4506109Z * [new branch] gh/bobrenjc93/673/head -> origin/gh/bobrenjc93/673/head 2025-11-03T16:58:01.4507777Z * [new branch] gh/bobrenjc93/673/orig -> origin/gh/bobrenjc93/673/orig 2025-11-03T16:58:01.4510016Z * [new branch] gh/bobrenjc93/674/base -> origin/gh/bobrenjc93/674/base 2025-11-03T16:58:01.4511712Z * [new branch] gh/bobrenjc93/674/head -> origin/gh/bobrenjc93/674/head 2025-11-03T16:58:01.4513476Z * [new branch] gh/bobrenjc93/674/orig -> origin/gh/bobrenjc93/674/orig 2025-11-03T16:58:01.4515725Z * [new branch] gh/bobrenjc93/675/base -> origin/gh/bobrenjc93/675/base 2025-11-03T16:58:01.4517539Z * [new branch] gh/bobrenjc93/675/head -> origin/gh/bobrenjc93/675/head 2025-11-03T16:58:01.4519194Z * [new branch] gh/bobrenjc93/675/orig -> origin/gh/bobrenjc93/675/orig 2025-11-03T16:58:01.4521506Z * [new branch] gh/bobrenjc93/676/base -> origin/gh/bobrenjc93/676/base 2025-11-03T16:58:01.4524996Z * [new branch] gh/bobrenjc93/676/head -> origin/gh/bobrenjc93/676/head 2025-11-03T16:58:01.4526639Z * [new branch] gh/bobrenjc93/676/orig -> origin/gh/bobrenjc93/676/orig 2025-11-03T16:58:01.4529055Z * [new branch] gh/bobrenjc93/677/base -> origin/gh/bobrenjc93/677/base 2025-11-03T16:58:01.4530632Z * [new branch] gh/bobrenjc93/677/head -> origin/gh/bobrenjc93/677/head 2025-11-03T16:58:01.4532303Z * [new branch] gh/bobrenjc93/677/orig -> origin/gh/bobrenjc93/677/orig 2025-11-03T16:58:01.4534678Z * [new branch] gh/bobrenjc93/678/base -> origin/gh/bobrenjc93/678/base 2025-11-03T16:58:01.4536299Z * [new branch] gh/bobrenjc93/678/head -> origin/gh/bobrenjc93/678/head 2025-11-03T16:58:01.4537961Z * [new branch] gh/bobrenjc93/678/orig -> origin/gh/bobrenjc93/678/orig 2025-11-03T16:58:01.4540448Z * [new branch] gh/bobrenjc93/679/base -> origin/gh/bobrenjc93/679/base 2025-11-03T16:58:01.4542667Z * [new branch] gh/bobrenjc93/679/head -> origin/gh/bobrenjc93/679/head 2025-11-03T16:58:01.4544565Z * [new branch] gh/bobrenjc93/679/orig -> origin/gh/bobrenjc93/679/orig 2025-11-03T16:58:01.4547474Z * [new branch] gh/c00w/23/base -> origin/gh/c00w/23/base 2025-11-03T16:58:01.4549191Z * [new branch] gh/c00w/23/head -> origin/gh/c00w/23/head 2025-11-03T16:58:01.4551517Z * [new branch] gh/c00w/53/base -> origin/gh/c00w/53/base 2025-11-03T16:58:01.4553222Z * [new branch] gh/c00w/53/head -> origin/gh/c00w/53/head 2025-11-03T16:58:01.4555087Z * [new branch] gh/c00w/53/orig -> origin/gh/c00w/53/orig 2025-11-03T16:58:01.4557354Z * [new branch] gh/c00w/54/base -> origin/gh/c00w/54/base 2025-11-03T16:58:01.4559084Z * [new branch] gh/c00w/54/head -> origin/gh/c00w/54/head 2025-11-03T16:58:01.4560750Z * [new branch] gh/c00w/54/orig -> origin/gh/c00w/54/orig 2025-11-03T16:58:01.4562998Z * [new branch] gh/c00w/55/base -> origin/gh/c00w/55/base 2025-11-03T16:58:01.4564725Z * [new branch] gh/c00w/55/head -> origin/gh/c00w/55/head 2025-11-03T16:58:01.4566410Z * [new branch] gh/c00w/55/orig -> origin/gh/c00w/55/orig 2025-11-03T16:58:01.4568722Z * [new branch] gh/c00w/56/base -> origin/gh/c00w/56/base 2025-11-03T16:58:01.4570408Z * [new branch] gh/c00w/56/head -> origin/gh/c00w/56/head 2025-11-03T16:58:01.4572118Z * [new branch] gh/c00w/56/orig -> origin/gh/c00w/56/orig 2025-11-03T16:58:01.4574576Z * [new branch] gh/c00w/57/base -> origin/gh/c00w/57/base 2025-11-03T16:58:01.4576469Z * [new branch] gh/c00w/57/head -> origin/gh/c00w/57/head 2025-11-03T16:58:01.4578653Z * [new branch] gh/c00w/57/orig -> origin/gh/c00w/57/orig 2025-11-03T16:58:01.4582065Z * [new branch] gh/c00w/58/base -> origin/gh/c00w/58/base 2025-11-03T16:58:01.4584671Z * [new branch] gh/c00w/58/head -> origin/gh/c00w/58/head 2025-11-03T16:58:01.4587100Z * [new branch] gh/c00w/58/orig -> origin/gh/c00w/58/orig 2025-11-03T16:58:01.4591390Z * [new branch] gh/clee2000/1/base -> origin/gh/clee2000/1/base 2025-11-03T16:58:01.4593824Z * [new branch] gh/clee2000/1/head -> origin/gh/clee2000/1/head 2025-11-03T16:58:01.4596095Z * [new branch] gh/clee2000/1/orig -> origin/gh/clee2000/1/orig 2025-11-03T16:58:01.4599962Z * [new branch] gh/coconutruben/1/base -> origin/gh/coconutruben/1/base 2025-11-03T16:58:01.4602474Z * [new branch] gh/coconutruben/1/head -> origin/gh/coconutruben/1/head 2025-11-03T16:58:01.4605853Z * [new branch] gh/coconutruben/20/base -> origin/gh/coconutruben/20/base 2025-11-03T16:58:01.4608239Z * [new branch] gh/coconutruben/20/head -> origin/gh/coconutruben/20/head 2025-11-03T16:58:01.4610549Z * [new branch] gh/coconutruben/20/orig -> origin/gh/coconutruben/20/orig 2025-11-03T16:58:01.4614078Z * [new branch] gh/coconutruben/55/base -> origin/gh/coconutruben/55/base 2025-11-03T16:58:01.4616369Z * [new branch] gh/coconutruben/55/head -> origin/gh/coconutruben/55/head 2025-11-03T16:58:01.4619303Z * [new branch] gh/coconutruben/55/orig -> origin/gh/coconutruben/55/orig 2025-11-03T16:58:01.4623355Z * [new branch] gh/coconutruben/57/base -> origin/gh/coconutruben/57/base 2025-11-03T16:58:01.4625766Z * [new branch] gh/coconutruben/57/head -> origin/gh/coconutruben/57/head 2025-11-03T16:58:01.4628186Z * [new branch] gh/coconutruben/57/orig -> origin/gh/coconutruben/57/orig 2025-11-03T16:58:01.4630887Z * [new branch] gh/coconutruben/70/base -> origin/gh/coconutruben/70/base 2025-11-03T16:58:01.4632762Z * [new branch] gh/coconutruben/70/head -> origin/gh/coconutruben/70/head 2025-11-03T16:58:01.4634544Z * [new branch] gh/coconutruben/70/orig -> origin/gh/coconutruben/70/orig 2025-11-03T16:58:01.4636780Z * [new branch] gh/coconutruben/71/base -> origin/gh/coconutruben/71/base 2025-11-03T16:58:01.4638584Z * [new branch] gh/coconutruben/71/head -> origin/gh/coconutruben/71/head 2025-11-03T16:58:01.4640305Z * [new branch] gh/coconutruben/71/orig -> origin/gh/coconutruben/71/orig 2025-11-03T16:58:01.4642325Z * [new branch] gh/coconutruben/72/base -> origin/gh/coconutruben/72/base 2025-11-03T16:58:01.4644075Z * [new branch] gh/coconutruben/72/head -> origin/gh/coconutruben/72/head 2025-11-03T16:58:01.4645800Z * [new branch] gh/coconutruben/72/orig -> origin/gh/coconutruben/72/orig 2025-11-03T16:58:01.4648581Z * [new branch] gh/coconutruben/73/base -> origin/gh/coconutruben/73/base 2025-11-03T16:58:01.4650300Z * [new branch] gh/coconutruben/73/head -> origin/gh/coconutruben/73/head 2025-11-03T16:58:01.4652085Z * [new branch] gh/coconutruben/73/orig -> origin/gh/coconutruben/73/orig 2025-11-03T16:58:01.4654722Z * [new branch] gh/coconutruben/74/base -> origin/gh/coconutruben/74/base 2025-11-03T16:58:01.4656549Z * [new branch] gh/coconutruben/74/head -> origin/gh/coconutruben/74/head 2025-11-03T16:58:01.4658241Z * [new branch] gh/coconutruben/74/orig -> origin/gh/coconutruben/74/orig 2025-11-03T16:58:01.4660763Z * [new branch] gh/coconutruben/79/base -> origin/gh/coconutruben/79/base 2025-11-03T16:58:01.4662531Z * [new branch] gh/coconutruben/79/head -> origin/gh/coconutruben/79/head 2025-11-03T16:58:01.4664227Z * [new branch] gh/coconutruben/79/orig -> origin/gh/coconutruben/79/orig 2025-11-03T16:58:01.4666515Z * [new branch] gh/coconutruben/80/base -> origin/gh/coconutruben/80/base 2025-11-03T16:58:01.4668259Z * [new branch] gh/coconutruben/80/head -> origin/gh/coconutruben/80/head 2025-11-03T16:58:01.4670039Z * [new branch] gh/coconutruben/80/orig -> origin/gh/coconutruben/80/orig 2025-11-03T16:58:01.4672563Z * [new branch] gh/coconutruben/81/base -> origin/gh/coconutruben/81/base 2025-11-03T16:58:01.4674312Z * [new branch] gh/coconutruben/81/head -> origin/gh/coconutruben/81/head 2025-11-03T16:58:01.4675998Z * [new branch] gh/coconutruben/81/orig -> origin/gh/coconutruben/81/orig 2025-11-03T16:58:01.4678443Z * [new branch] gh/coconutruben/82/base -> origin/gh/coconutruben/82/base 2025-11-03T16:58:01.4680138Z * [new branch] gh/coconutruben/82/head -> origin/gh/coconutruben/82/head 2025-11-03T16:58:01.4681773Z * [new branch] gh/coconutruben/82/orig -> origin/gh/coconutruben/82/orig 2025-11-03T16:58:01.4684366Z * [new branch] gh/coconutruben/83/base -> origin/gh/coconutruben/83/base 2025-11-03T16:58:01.4686036Z * [new branch] gh/coconutruben/83/head -> origin/gh/coconutruben/83/head 2025-11-03T16:58:01.4687727Z * [new branch] gh/coconutruben/83/orig -> origin/gh/coconutruben/83/orig 2025-11-03T16:58:01.4690591Z * [new branch] gh/colinchan15/1/base -> origin/gh/colinchan15/1/base 2025-11-03T16:58:01.4692183Z * [new branch] gh/colinchan15/1/head -> origin/gh/colinchan15/1/head 2025-11-03T16:58:01.4694502Z * [new branch] gh/colinchan15/2/base -> origin/gh/colinchan15/2/base 2025-11-03T16:58:01.4696136Z * [new branch] gh/colinchan15/2/head -> origin/gh/colinchan15/2/head 2025-11-03T16:58:01.4698330Z * [new branch] gh/colinchan15/3/base -> origin/gh/colinchan15/3/base 2025-11-03T16:58:01.4700010Z * [new branch] gh/colinchan15/3/head -> origin/gh/colinchan15/3/head 2025-11-03T16:58:01.4702294Z * [new branch] gh/colinchan15/6/base -> origin/gh/colinchan15/6/base 2025-11-03T16:58:01.4704088Z * [new branch] gh/colinchan15/6/head -> origin/gh/colinchan15/6/head 2025-11-03T16:58:01.4706980Z * [new branch] gh/davidberard98/382/base -> origin/gh/davidberard98/382/base 2025-11-03T16:58:01.4708791Z * [new branch] gh/davidberard98/382/head -> origin/gh/davidberard98/382/head 2025-11-03T16:58:01.4710453Z * [new branch] gh/davidberard98/382/orig -> origin/gh/davidberard98/382/orig 2025-11-03T16:58:01.4713058Z * [new branch] gh/davidberard98/386/base -> origin/gh/davidberard98/386/base 2025-11-03T16:58:01.4714803Z * [new branch] gh/davidberard98/386/head -> origin/gh/davidberard98/386/head 2025-11-03T16:58:01.4716593Z * [new branch] gh/davidberard98/386/orig -> origin/gh/davidberard98/386/orig 2025-11-03T16:58:01.4718942Z * [new branch] gh/davidberard98/391/base -> origin/gh/davidberard98/391/base 2025-11-03T16:58:01.4720739Z * [new branch] gh/davidberard98/391/head -> origin/gh/davidberard98/391/head 2025-11-03T16:58:01.4724717Z * [new branch] gh/davidberard98/391/orig -> origin/gh/davidberard98/391/orig 2025-11-03T16:58:01.4727210Z * [new branch] gh/davidberard98/392/base -> origin/gh/davidberard98/392/base 2025-11-03T16:58:01.4728872Z * [new branch] gh/davidberard98/392/head -> origin/gh/davidberard98/392/head 2025-11-03T16:58:01.4730695Z * [new branch] gh/davidberard98/392/orig -> origin/gh/davidberard98/392/orig 2025-11-03T16:58:01.4733039Z * [new branch] gh/davidberard98/399/base -> origin/gh/davidberard98/399/base 2025-11-03T16:58:01.4734784Z * [new branch] gh/davidberard98/399/head -> origin/gh/davidberard98/399/head 2025-11-03T16:58:01.4736556Z * [new branch] gh/davidberard98/399/orig -> origin/gh/davidberard98/399/orig 2025-11-03T16:58:01.4739332Z * [new branch] gh/desertfire/594/base -> origin/gh/desertfire/594/base 2025-11-03T16:58:01.4741113Z * [new branch] gh/desertfire/594/head -> origin/gh/desertfire/594/head 2025-11-03T16:58:01.4742998Z * [new branch] gh/desertfire/594/orig -> origin/gh/desertfire/594/orig 2025-11-03T16:58:01.4745237Z * [new branch] gh/desertfire/595/base -> origin/gh/desertfire/595/base 2025-11-03T16:58:01.4746977Z * [new branch] gh/desertfire/595/head -> origin/gh/desertfire/595/head 2025-11-03T16:58:01.4748684Z * [new branch] gh/desertfire/595/orig -> origin/gh/desertfire/595/orig 2025-11-03T16:58:01.4751048Z * [new branch] gh/desertfire/597/base -> origin/gh/desertfire/597/base 2025-11-03T16:58:01.4752771Z * [new branch] gh/desertfire/597/head -> origin/gh/desertfire/597/head 2025-11-03T16:58:01.4754512Z * [new branch] gh/desertfire/597/orig -> origin/gh/desertfire/597/orig 2025-11-03T16:58:01.4756799Z * [new branch] gh/desertfire/600/base -> origin/gh/desertfire/600/base 2025-11-03T16:58:01.4758460Z * [new branch] gh/desertfire/600/head -> origin/gh/desertfire/600/head 2025-11-03T16:58:01.4760300Z * [new branch] gh/desertfire/600/orig -> origin/gh/desertfire/600/orig 2025-11-03T16:58:01.4763065Z * [new branch] gh/desertfire/601/base -> origin/gh/desertfire/601/base 2025-11-03T16:58:01.4764842Z * [new branch] gh/desertfire/601/head -> origin/gh/desertfire/601/head 2025-11-03T16:58:01.4766541Z * [new branch] gh/desertfire/601/orig -> origin/gh/desertfire/601/orig 2025-11-03T16:58:01.4768880Z * [new branch] gh/desertfire/602/base -> origin/gh/desertfire/602/base 2025-11-03T16:58:01.4770505Z * [new branch] gh/desertfire/602/head -> origin/gh/desertfire/602/head 2025-11-03T16:58:01.4772120Z * [new branch] gh/desertfire/602/orig -> origin/gh/desertfire/602/orig 2025-11-03T16:58:01.4774495Z * [new branch] gh/desertfire/603/base -> origin/gh/desertfire/603/base 2025-11-03T16:58:01.4776257Z * [new branch] gh/desertfire/603/head -> origin/gh/desertfire/603/head 2025-11-03T16:58:01.4777943Z * [new branch] gh/desertfire/603/orig -> origin/gh/desertfire/603/orig 2025-11-03T16:58:01.4780111Z * [new branch] gh/desertfire/604/base -> origin/gh/desertfire/604/base 2025-11-03T16:58:01.4782256Z * [new branch] gh/desertfire/604/head -> origin/gh/desertfire/604/head 2025-11-03T16:58:01.4784175Z * [new branch] gh/desertfire/604/orig -> origin/gh/desertfire/604/orig 2025-11-03T16:58:01.4786959Z * [new branch] gh/dharakk/1/base -> origin/gh/dharakk/1/base 2025-11-03T16:58:01.4788714Z * [new branch] gh/dharakk/1/head -> origin/gh/dharakk/1/head 2025-11-03T16:58:01.4791636Z * [new branch] gh/drisspg/170/base -> origin/gh/drisspg/170/base 2025-11-03T16:58:01.4793571Z * [new branch] gh/drisspg/170/head -> origin/gh/drisspg/170/head 2025-11-03T16:58:01.4795299Z * [new branch] gh/drisspg/170/orig -> origin/gh/drisspg/170/orig 2025-11-03T16:58:01.4797614Z * [new branch] gh/drisspg/182/base -> origin/gh/drisspg/182/base 2025-11-03T16:58:01.4799305Z * [new branch] gh/drisspg/182/head -> origin/gh/drisspg/182/head 2025-11-03T16:58:01.4801755Z * [new branch] gh/drisspg/183/base -> origin/gh/drisspg/183/base 2025-11-03T16:58:01.4803485Z * [new branch] gh/drisspg/183/head -> origin/gh/drisspg/183/head 2025-11-03T16:58:01.4805528Z * [new branch] gh/drisspg/184/base -> origin/gh/drisspg/184/base 2025-11-03T16:58:01.4807292Z * [new branch] gh/drisspg/184/head -> origin/gh/drisspg/184/head 2025-11-03T16:58:01.4809802Z * [new branch] gh/drisspg/185/base -> origin/gh/drisspg/185/base 2025-11-03T16:58:01.4811318Z * [new branch] gh/drisspg/185/head -> origin/gh/drisspg/185/head 2025-11-03T16:58:01.4813775Z * [new branch] gh/drisspg/187/base -> origin/gh/drisspg/187/base 2025-11-03T16:58:01.4815467Z * [new branch] gh/drisspg/187/head -> origin/gh/drisspg/187/head 2025-11-03T16:58:01.4817139Z * [new branch] gh/drisspg/187/orig -> origin/gh/drisspg/187/orig 2025-11-03T16:58:01.4819370Z * [new branch] gh/drisspg/194/base -> origin/gh/drisspg/194/base 2025-11-03T16:58:01.4821248Z * [new branch] gh/drisspg/194/head -> origin/gh/drisspg/194/head 2025-11-03T16:58:01.4823069Z * [new branch] gh/drisspg/194/orig -> origin/gh/drisspg/194/orig 2025-11-03T16:58:01.4825249Z * [new branch] gh/drisspg/197/base -> origin/gh/drisspg/197/base 2025-11-03T16:58:01.4826928Z * [new branch] gh/drisspg/197/head -> origin/gh/drisspg/197/head 2025-11-03T16:58:01.4828604Z * [new branch] gh/drisspg/197/orig -> origin/gh/drisspg/197/orig 2025-11-03T16:58:01.4831047Z * [new branch] gh/drisspg/199/base -> origin/gh/drisspg/199/base 2025-11-03T16:58:01.4832659Z * [new branch] gh/drisspg/199/head -> origin/gh/drisspg/199/head 2025-11-03T16:58:01.4834414Z * [new branch] gh/drisspg/199/orig -> origin/gh/drisspg/199/orig 2025-11-03T16:58:01.4836675Z * [new branch] gh/drisspg/200/base -> origin/gh/drisspg/200/base 2025-11-03T16:58:01.4838356Z * [new branch] gh/drisspg/200/head -> origin/gh/drisspg/200/head 2025-11-03T16:58:01.4840048Z * [new branch] gh/drisspg/200/orig -> origin/gh/drisspg/200/orig 2025-11-03T16:58:01.4842384Z * [new branch] gh/drisspg/210/base -> origin/gh/drisspg/210/base 2025-11-03T16:58:01.4844110Z * [new branch] gh/drisspg/210/head -> origin/gh/drisspg/210/head 2025-11-03T16:58:01.4845746Z * [new branch] gh/drisspg/210/orig -> origin/gh/drisspg/210/orig 2025-11-03T16:58:01.4848031Z * [new branch] gh/drisspg/211/base -> origin/gh/drisspg/211/base 2025-11-03T16:58:01.4849697Z * [new branch] gh/drisspg/211/head -> origin/gh/drisspg/211/head 2025-11-03T16:58:01.4851356Z * [new branch] gh/drisspg/211/orig -> origin/gh/drisspg/211/orig 2025-11-03T16:58:01.4853688Z * [new branch] gh/drisspg/212/base -> origin/gh/drisspg/212/base 2025-11-03T16:58:01.4855341Z * [new branch] gh/drisspg/212/head -> origin/gh/drisspg/212/head 2025-11-03T16:58:01.4857055Z * [new branch] gh/drisspg/212/orig -> origin/gh/drisspg/212/orig 2025-11-03T16:58:01.4859335Z * [new branch] gh/drisspg/213/base -> origin/gh/drisspg/213/base 2025-11-03T16:58:01.4861042Z * [new branch] gh/drisspg/213/head -> origin/gh/drisspg/213/head 2025-11-03T16:58:01.4862948Z * [new branch] gh/drisspg/213/orig -> origin/gh/drisspg/213/orig 2025-11-03T16:58:01.4865171Z * [new branch] gh/drisspg/214/base -> origin/gh/drisspg/214/base 2025-11-03T16:58:01.4866905Z * [new branch] gh/drisspg/214/head -> origin/gh/drisspg/214/head 2025-11-03T16:58:01.4868851Z * [new branch] gh/drisspg/214/orig -> origin/gh/drisspg/214/orig 2025-11-03T16:58:01.4871096Z * [new branch] gh/drisspg/215/base -> origin/gh/drisspg/215/base 2025-11-03T16:58:01.4872716Z * [new branch] gh/drisspg/215/head -> origin/gh/drisspg/215/head 2025-11-03T16:58:01.4874396Z * [new branch] gh/drisspg/215/orig -> origin/gh/drisspg/215/orig 2025-11-03T16:58:01.4877710Z * [new branch] gh/drisspg/216/base -> origin/gh/drisspg/216/base 2025-11-03T16:58:01.4879362Z * [new branch] gh/drisspg/216/head -> origin/gh/drisspg/216/head 2025-11-03T16:58:01.4881011Z * [new branch] gh/drisspg/216/orig -> origin/gh/drisspg/216/orig 2025-11-03T16:58:01.4883409Z * [new branch] gh/drisspg/217/base -> origin/gh/drisspg/217/base 2025-11-03T16:58:01.4885128Z * [new branch] gh/drisspg/217/head -> origin/gh/drisspg/217/head 2025-11-03T16:58:01.4887015Z * [new branch] gh/drisspg/217/orig -> origin/gh/drisspg/217/orig 2025-11-03T16:58:01.4889920Z * [new branch] gh/dsjohns2/1/base -> origin/gh/dsjohns2/1/base 2025-11-03T16:58:01.4891614Z * [new branch] gh/dsjohns2/1/head -> origin/gh/dsjohns2/1/head 2025-11-03T16:58:01.4894494Z * [new branch] gh/dzmitry-huba/1/base -> origin/gh/dzmitry-huba/1/base 2025-11-03T16:58:01.4896179Z * [new branch] gh/dzmitry-huba/1/head -> origin/gh/dzmitry-huba/1/head 2025-11-03T16:58:01.4898888Z * [new branch] gh/dzmitry-huba/10/base -> origin/gh/dzmitry-huba/10/base 2025-11-03T16:58:01.4900710Z * [new branch] gh/dzmitry-huba/10/head -> origin/gh/dzmitry-huba/10/head 2025-11-03T16:58:01.4902563Z * [new branch] gh/dzmitry-huba/10/orig -> origin/gh/dzmitry-huba/10/orig 2025-11-03T16:58:01.4904846Z * [new branch] gh/dzmitry-huba/11/base -> origin/gh/dzmitry-huba/11/base 2025-11-03T16:58:01.4906614Z * [new branch] gh/dzmitry-huba/11/head -> origin/gh/dzmitry-huba/11/head 2025-11-03T16:58:01.4908381Z * [new branch] gh/dzmitry-huba/11/orig -> origin/gh/dzmitry-huba/11/orig 2025-11-03T16:58:01.4910816Z * [new branch] gh/dzmitry-huba/12/base -> origin/gh/dzmitry-huba/12/base 2025-11-03T16:58:01.4912487Z * [new branch] gh/dzmitry-huba/12/head -> origin/gh/dzmitry-huba/12/head 2025-11-03T16:58:01.4914228Z * [new branch] gh/dzmitry-huba/12/orig -> origin/gh/dzmitry-huba/12/orig 2025-11-03T16:58:01.4916402Z * [new branch] gh/dzmitry-huba/2/base -> origin/gh/dzmitry-huba/2/base 2025-11-03T16:58:01.4918241Z * [new branch] gh/dzmitry-huba/2/head -> origin/gh/dzmitry-huba/2/head 2025-11-03T16:58:01.4920484Z * [new branch] gh/dzmitry-huba/3/base -> origin/gh/dzmitry-huba/3/base 2025-11-03T16:58:01.4922479Z * [new branch] gh/dzmitry-huba/3/head -> origin/gh/dzmitry-huba/3/head 2025-11-03T16:58:01.4924700Z * [new branch] gh/dzmitry-huba/4/base -> origin/gh/dzmitry-huba/4/base 2025-11-03T16:58:01.4926329Z * [new branch] gh/dzmitry-huba/4/head -> origin/gh/dzmitry-huba/4/head 2025-11-03T16:58:01.4927954Z * [new branch] gh/dzmitry-huba/4/orig -> origin/gh/dzmitry-huba/4/orig 2025-11-03T16:58:01.4930861Z * [new branch] gh/dzmitry-huba/5/base -> origin/gh/dzmitry-huba/5/base 2025-11-03T16:58:01.4932811Z * [new branch] gh/dzmitry-huba/5/head -> origin/gh/dzmitry-huba/5/head 2025-11-03T16:58:01.4934643Z * [new branch] gh/dzmitry-huba/5/orig -> origin/gh/dzmitry-huba/5/orig 2025-11-03T16:58:01.4936767Z * [new branch] gh/dzmitry-huba/6/base -> origin/gh/dzmitry-huba/6/base 2025-11-03T16:58:01.4938531Z * [new branch] gh/dzmitry-huba/6/head -> origin/gh/dzmitry-huba/6/head 2025-11-03T16:58:01.4940494Z * [new branch] gh/dzmitry-huba/6/orig -> origin/gh/dzmitry-huba/6/orig 2025-11-03T16:58:01.4943036Z * [new branch] gh/dzmitry-huba/7/base -> origin/gh/dzmitry-huba/7/base 2025-11-03T16:58:01.4944828Z * [new branch] gh/dzmitry-huba/7/head -> origin/gh/dzmitry-huba/7/head 2025-11-03T16:58:01.4946594Z * [new branch] gh/dzmitry-huba/7/orig -> origin/gh/dzmitry-huba/7/orig 2025-11-03T16:58:01.4948945Z * [new branch] gh/dzmitry-huba/8/base -> origin/gh/dzmitry-huba/8/base 2025-11-03T16:58:01.4950669Z * [new branch] gh/dzmitry-huba/8/head -> origin/gh/dzmitry-huba/8/head 2025-11-03T16:58:01.4952299Z * [new branch] gh/dzmitry-huba/8/orig -> origin/gh/dzmitry-huba/8/orig 2025-11-03T16:58:01.4954700Z * [new branch] gh/dzmitry-huba/9/base -> origin/gh/dzmitry-huba/9/base 2025-11-03T16:58:01.4956436Z * [new branch] gh/dzmitry-huba/9/head -> origin/gh/dzmitry-huba/9/head 2025-11-03T16:58:01.4958197Z * [new branch] gh/dzmitry-huba/9/orig -> origin/gh/dzmitry-huba/9/orig 2025-11-03T16:58:01.4961027Z * [new branch] gh/eellison/808/base -> origin/gh/eellison/808/base 2025-11-03T16:58:01.4962778Z * [new branch] gh/eellison/808/head -> origin/gh/eellison/808/head 2025-11-03T16:58:01.4964574Z * [new branch] gh/eellison/808/orig -> origin/gh/eellison/808/orig 2025-11-03T16:58:01.4966863Z * [new branch] gh/eellison/809/base -> origin/gh/eellison/809/base 2025-11-03T16:58:01.4968610Z * [new branch] gh/eellison/809/head -> origin/gh/eellison/809/head 2025-11-03T16:58:01.4970257Z * [new branch] gh/eellison/809/orig -> origin/gh/eellison/809/orig 2025-11-03T16:58:01.4972484Z * [new branch] gh/eellison/822/base -> origin/gh/eellison/822/base 2025-11-03T16:58:01.4974221Z * [new branch] gh/eellison/822/head -> origin/gh/eellison/822/head 2025-11-03T16:58:01.4975939Z * [new branch] gh/eellison/822/orig -> origin/gh/eellison/822/orig 2025-11-03T16:58:01.4978482Z * [new branch] gh/eellison/823/base -> origin/gh/eellison/823/base 2025-11-03T16:58:01.4979925Z * [new branch] gh/eellison/823/head -> origin/gh/eellison/823/head 2025-11-03T16:58:01.4981737Z * [new branch] gh/eellison/823/orig -> origin/gh/eellison/823/orig 2025-11-03T16:58:01.4984294Z * [new branch] gh/eellison/836/base -> origin/gh/eellison/836/base 2025-11-03T16:58:01.4985933Z * [new branch] gh/eellison/836/head -> origin/gh/eellison/836/head 2025-11-03T16:58:01.4987716Z * [new branch] gh/eellison/836/orig -> origin/gh/eellison/836/orig 2025-11-03T16:58:01.4990128Z * [new branch] gh/eellison/837/base -> origin/gh/eellison/837/base 2025-11-03T16:58:01.4991938Z * [new branch] gh/eellison/837/head -> origin/gh/eellison/837/head 2025-11-03T16:58:01.4993821Z * [new branch] gh/eellison/837/orig -> origin/gh/eellison/837/orig 2025-11-03T16:58:01.4996100Z * [new branch] gh/eellison/838/base -> origin/gh/eellison/838/base 2025-11-03T16:58:01.4997797Z * [new branch] gh/eellison/838/head -> origin/gh/eellison/838/head 2025-11-03T16:58:01.4999487Z * [new branch] gh/eellison/838/orig -> origin/gh/eellison/838/orig 2025-11-03T16:58:01.5001751Z * [new branch] gh/eellison/839/base -> origin/gh/eellison/839/base 2025-11-03T16:58:01.5003512Z * [new branch] gh/eellison/839/head -> origin/gh/eellison/839/head 2025-11-03T16:58:01.5005319Z * [new branch] gh/eellison/839/orig -> origin/gh/eellison/839/orig 2025-11-03T16:58:01.5008026Z * [new branch] gh/eellison/840/base -> origin/gh/eellison/840/base 2025-11-03T16:58:01.5009886Z * [new branch] gh/eellison/840/head -> origin/gh/eellison/840/head 2025-11-03T16:58:01.5011472Z * [new branch] gh/eellison/840/orig -> origin/gh/eellison/840/orig 2025-11-03T16:58:01.5013723Z * [new branch] gh/eellison/841/base -> origin/gh/eellison/841/base 2025-11-03T16:58:01.5015513Z * [new branch] gh/eellison/841/head -> origin/gh/eellison/841/head 2025-11-03T16:58:01.5017663Z * [new branch] gh/eellison/841/orig -> origin/gh/eellison/841/orig 2025-11-03T16:58:01.5019830Z * [new branch] gh/eellison/842/base -> origin/gh/eellison/842/base 2025-11-03T16:58:01.5021920Z * [new branch] gh/eellison/842/head -> origin/gh/eellison/842/head 2025-11-03T16:58:01.5023612Z * [new branch] gh/eellison/842/orig -> origin/gh/eellison/842/orig 2025-11-03T16:58:01.5025944Z * [new branch] gh/eellison/843/base -> origin/gh/eellison/843/base 2025-11-03T16:58:01.5027571Z * [new branch] gh/eellison/843/head -> origin/gh/eellison/843/head 2025-11-03T16:58:01.5029179Z * [new branch] gh/eellison/843/orig -> origin/gh/eellison/843/orig 2025-11-03T16:58:01.5031358Z * [new branch] gh/eellison/844/base -> origin/gh/eellison/844/base 2025-11-03T16:58:01.5033430Z * [new branch] gh/eellison/844/head -> origin/gh/eellison/844/head 2025-11-03T16:58:01.5035144Z * [new branch] gh/eellison/844/orig -> origin/gh/eellison/844/orig 2025-11-03T16:58:01.5037390Z * [new branch] gh/eellison/845/base -> origin/gh/eellison/845/base 2025-11-03T16:58:01.5039174Z * [new branch] gh/eellison/845/head -> origin/gh/eellison/845/head 2025-11-03T16:58:01.5040852Z * [new branch] gh/eellison/845/orig -> origin/gh/eellison/845/orig 2025-11-03T16:58:01.5043144Z * [new branch] gh/eellison/846/base -> origin/gh/eellison/846/base 2025-11-03T16:58:01.5044813Z * [new branch] gh/eellison/846/head -> origin/gh/eellison/846/head 2025-11-03T16:58:01.5046528Z * [new branch] gh/eellison/846/orig -> origin/gh/eellison/846/orig 2025-11-03T16:58:01.5048822Z * [new branch] gh/eellison/847/base -> origin/gh/eellison/847/base 2025-11-03T16:58:01.5050488Z * [new branch] gh/eellison/847/head -> origin/gh/eellison/847/head 2025-11-03T16:58:01.5052201Z * [new branch] gh/eellison/847/orig -> origin/gh/eellison/847/orig 2025-11-03T16:58:01.5055184Z * [new branch] gh/eellison/848/base -> origin/gh/eellison/848/base 2025-11-03T16:58:01.5056892Z * [new branch] gh/eellison/848/head -> origin/gh/eellison/848/head 2025-11-03T16:58:01.5058639Z * [new branch] gh/eellison/848/orig -> origin/gh/eellison/848/orig 2025-11-03T16:58:01.5061299Z * [new branch] gh/eellison/849/base -> origin/gh/eellison/849/base 2025-11-03T16:58:01.5063258Z * [new branch] gh/eellison/849/head -> origin/gh/eellison/849/head 2025-11-03T16:58:01.5064919Z * [new branch] gh/eellison/849/orig -> origin/gh/eellison/849/orig 2025-11-03T16:58:01.5067246Z * [new branch] gh/eellison/850/base -> origin/gh/eellison/850/base 2025-11-03T16:58:01.5068962Z * [new branch] gh/eellison/850/head -> origin/gh/eellison/850/head 2025-11-03T16:58:01.5070713Z * [new branch] gh/eellison/850/orig -> origin/gh/eellison/850/orig 2025-11-03T16:58:01.5073442Z * [new branch] gh/eellison/851/base -> origin/gh/eellison/851/base 2025-11-03T16:58:01.5075130Z * [new branch] gh/eellison/851/head -> origin/gh/eellison/851/head 2025-11-03T16:58:01.5076896Z * [new branch] gh/eellison/851/orig -> origin/gh/eellison/851/orig 2025-11-03T16:58:01.5079361Z * [new branch] gh/eellison/852/base -> origin/gh/eellison/852/base 2025-11-03T16:58:01.5080956Z * [new branch] gh/eellison/852/head -> origin/gh/eellison/852/head 2025-11-03T16:58:01.5082639Z * [new branch] gh/eellison/852/orig -> origin/gh/eellison/852/orig 2025-11-03T16:58:01.5084981Z * [new branch] gh/eellison/853/base -> origin/gh/eellison/853/base 2025-11-03T16:58:01.5086719Z * [new branch] gh/eellison/853/head -> origin/gh/eellison/853/head 2025-11-03T16:58:01.5088414Z * [new branch] gh/eellison/853/orig -> origin/gh/eellison/853/orig 2025-11-03T16:58:01.5090919Z * [new branch] gh/eellison/854/base -> origin/gh/eellison/854/base 2025-11-03T16:58:01.5092732Z * [new branch] gh/eellison/854/head -> origin/gh/eellison/854/head 2025-11-03T16:58:01.5094565Z * [new branch] gh/eellison/854/orig -> origin/gh/eellison/854/orig 2025-11-03T16:58:01.5096897Z * [new branch] gh/eellison/855/base -> origin/gh/eellison/855/base 2025-11-03T16:58:01.5098571Z * [new branch] gh/eellison/855/head -> origin/gh/eellison/855/head 2025-11-03T16:58:01.5100294Z * [new branch] gh/eellison/855/orig -> origin/gh/eellison/855/orig 2025-11-03T16:58:01.5102736Z * [new branch] gh/eellison/856/base -> origin/gh/eellison/856/base 2025-11-03T16:58:01.5104430Z * [new branch] gh/eellison/856/head -> origin/gh/eellison/856/head 2025-11-03T16:58:01.5106117Z * [new branch] gh/eellison/856/orig -> origin/gh/eellison/856/orig 2025-11-03T16:58:01.5108497Z * [new branch] gh/eellison/857/base -> origin/gh/eellison/857/base 2025-11-03T16:58:01.5110330Z * [new branch] gh/eellison/857/head -> origin/gh/eellison/857/head 2025-11-03T16:58:01.5111996Z * [new branch] gh/eellison/857/orig -> origin/gh/eellison/857/orig 2025-11-03T16:58:01.5114375Z * [new branch] gh/eellison/858/base -> origin/gh/eellison/858/base 2025-11-03T16:58:01.5116605Z * [new branch] gh/eellison/858/head -> origin/gh/eellison/858/head 2025-11-03T16:58:01.5118332Z * [new branch] gh/eellison/858/orig -> origin/gh/eellison/858/orig 2025-11-03T16:58:01.5120803Z * [new branch] gh/eellison/859/base -> origin/gh/eellison/859/base 2025-11-03T16:58:01.5122893Z * [new branch] gh/eellison/859/head -> origin/gh/eellison/859/head 2025-11-03T16:58:01.5124583Z * [new branch] gh/eellison/859/orig -> origin/gh/eellison/859/orig 2025-11-03T16:58:01.5126854Z * [new branch] gh/eellison/860/base -> origin/gh/eellison/860/base 2025-11-03T16:58:01.5128954Z * [new branch] gh/eellison/860/head -> origin/gh/eellison/860/head 2025-11-03T16:58:01.5130775Z * [new branch] gh/eellison/860/orig -> origin/gh/eellison/860/orig 2025-11-03T16:58:01.5133067Z * [new branch] gh/eellison/861/base -> origin/gh/eellison/861/base 2025-11-03T16:58:01.5135054Z * [new branch] gh/eellison/861/head -> origin/gh/eellison/861/head 2025-11-03T16:58:01.5136786Z * [new branch] gh/eellison/861/orig -> origin/gh/eellison/861/orig 2025-11-03T16:58:01.5139567Z * [new branch] gh/etaf/147/base -> origin/gh/etaf/147/base 2025-11-03T16:58:01.5141250Z * [new branch] gh/etaf/147/head -> origin/gh/etaf/147/head 2025-11-03T16:58:01.5143992Z * [new branch] gh/etaf/154/base -> origin/gh/etaf/154/base 2025-11-03T16:58:01.5145699Z * [new branch] gh/etaf/154/head -> origin/gh/etaf/154/head 2025-11-03T16:58:01.5147322Z * [new branch] gh/etaf/154/orig -> origin/gh/etaf/154/orig 2025-11-03T16:58:01.5149705Z * [new branch] gh/etaf/156/base -> origin/gh/etaf/156/base 2025-11-03T16:58:01.5151318Z * [new branch] gh/etaf/156/head -> origin/gh/etaf/156/head 2025-11-03T16:58:01.5152971Z * [new branch] gh/etaf/156/orig -> origin/gh/etaf/156/orig 2025-11-03T16:58:01.5155564Z * [new branch] gh/etaf/157/base -> origin/gh/etaf/157/base 2025-11-03T16:58:01.5157242Z * [new branch] gh/etaf/157/head -> origin/gh/etaf/157/head 2025-11-03T16:58:01.5159005Z * [new branch] gh/etaf/157/orig -> origin/gh/etaf/157/orig 2025-11-03T16:58:01.5161191Z * [new branch] gh/etaf/158/base -> origin/gh/etaf/158/base 2025-11-03T16:58:01.5162871Z * [new branch] gh/etaf/158/head -> origin/gh/etaf/158/head 2025-11-03T16:58:01.5164626Z * [new branch] gh/etaf/158/orig -> origin/gh/etaf/158/orig 2025-11-03T16:58:01.5166986Z * [new branch] gh/etaf/159/base -> origin/gh/etaf/159/base 2025-11-03T16:58:01.5168779Z * [new branch] gh/etaf/159/head -> origin/gh/etaf/159/head 2025-11-03T16:58:01.5170502Z * [new branch] gh/etaf/159/orig -> origin/gh/etaf/159/orig 2025-11-03T16:58:01.5173051Z * [new branch] gh/etaf/160/base -> origin/gh/etaf/160/base 2025-11-03T16:58:01.5174792Z * [new branch] gh/etaf/160/head -> origin/gh/etaf/160/head 2025-11-03T16:58:01.5176671Z * [new branch] gh/etaf/160/orig -> origin/gh/etaf/160/orig 2025-11-03T16:58:01.5179003Z * [new branch] gh/etaf/161/base -> origin/gh/etaf/161/base 2025-11-03T16:58:01.5181302Z * [new branch] gh/etaf/161/head -> origin/gh/etaf/161/head 2025-11-03T16:58:01.5183876Z * [new branch] gh/etaf/161/orig -> origin/gh/etaf/161/orig 2025-11-03T16:58:01.5186987Z * [new branch] gh/etaf/162/base -> origin/gh/etaf/162/base 2025-11-03T16:58:01.5188647Z * [new branch] gh/etaf/162/head -> origin/gh/etaf/162/head 2025-11-03T16:58:01.5190299Z * [new branch] gh/etaf/162/orig -> origin/gh/etaf/162/orig 2025-11-03T16:58:01.5192749Z * [new branch] gh/etaf/166/base -> origin/gh/etaf/166/base 2025-11-03T16:58:01.5194580Z * [new branch] gh/etaf/166/head -> origin/gh/etaf/166/head 2025-11-03T16:58:01.5196293Z * [new branch] gh/etaf/166/orig -> origin/gh/etaf/166/orig 2025-11-03T16:58:01.5198700Z * [new branch] gh/etaf/167/base -> origin/gh/etaf/167/base 2025-11-03T16:58:01.5200457Z * [new branch] gh/etaf/167/head -> origin/gh/etaf/167/head 2025-11-03T16:58:01.5202126Z * [new branch] gh/etaf/167/orig -> origin/gh/etaf/167/orig 2025-11-03T16:58:01.5204523Z * [new branch] gh/etaf/168/base -> origin/gh/etaf/168/base 2025-11-03T16:58:01.5206323Z * [new branch] gh/etaf/168/head -> origin/gh/etaf/168/head 2025-11-03T16:58:01.5208112Z * [new branch] gh/etaf/168/orig -> origin/gh/etaf/168/orig 2025-11-03T16:58:01.5210485Z * [new branch] gh/etaf/171/base -> origin/gh/etaf/171/base 2025-11-03T16:58:01.5212164Z * [new branch] gh/etaf/171/head -> origin/gh/etaf/171/head 2025-11-03T16:58:01.5213845Z * [new branch] gh/etaf/171/orig -> origin/gh/etaf/171/orig 2025-11-03T16:58:01.5216172Z * [new branch] gh/etaf/172/base -> origin/gh/etaf/172/base 2025-11-03T16:58:01.5217928Z * [new branch] gh/etaf/172/head -> origin/gh/etaf/172/head 2025-11-03T16:58:01.5220140Z * [new branch] gh/etaf/172/orig -> origin/gh/etaf/172/orig 2025-11-03T16:58:01.5224179Z * [new branch] gh/exclamaforte/1/base -> origin/gh/exclamaforte/1/base 2025-11-03T16:58:01.5225706Z * [new branch] gh/exclamaforte/1/head -> origin/gh/exclamaforte/1/head 2025-11-03T16:58:01.5227892Z * [new branch] gh/exclamaforte/2/base -> origin/gh/exclamaforte/2/base 2025-11-03T16:58:01.5229612Z * [new branch] gh/exclamaforte/2/head -> origin/gh/exclamaforte/2/head 2025-11-03T16:58:01.5231854Z * [new branch] gh/exclamaforte/3/base -> origin/gh/exclamaforte/3/base 2025-11-03T16:58:01.5233712Z * [new branch] gh/exclamaforte/3/head -> origin/gh/exclamaforte/3/head 2025-11-03T16:58:01.5236011Z * [new branch] gh/exclamaforte/4/base -> origin/gh/exclamaforte/4/base 2025-11-03T16:58:01.5237686Z * [new branch] gh/exclamaforte/4/head -> origin/gh/exclamaforte/4/head 2025-11-03T16:58:01.5240552Z * [new branch] gh/ezyang/2374/base -> origin/gh/ezyang/2374/base 2025-11-03T16:58:01.5242298Z * [new branch] gh/ezyang/2374/head -> origin/gh/ezyang/2374/head 2025-11-03T16:58:01.5244104Z * [new branch] gh/ezyang/2374/orig -> origin/gh/ezyang/2374/orig 2025-11-03T16:58:01.5246419Z * [new branch] gh/ezyang/2973/base -> origin/gh/ezyang/2973/base 2025-11-03T16:58:01.5248164Z * [new branch] gh/ezyang/2973/head -> origin/gh/ezyang/2973/head 2025-11-03T16:58:01.5249898Z * [new branch] gh/ezyang/2973/orig -> origin/gh/ezyang/2973/orig 2025-11-03T16:58:01.5252178Z * [new branch] gh/ezyang/2974/base -> origin/gh/ezyang/2974/base 2025-11-03T16:58:01.5253870Z * [new branch] gh/ezyang/2974/head -> origin/gh/ezyang/2974/head 2025-11-03T16:58:01.5255580Z * [new branch] gh/ezyang/2974/orig -> origin/gh/ezyang/2974/orig 2025-11-03T16:58:01.5257880Z * [new branch] gh/ezyang/3120/base -> origin/gh/ezyang/3120/base 2025-11-03T16:58:01.5259661Z * [new branch] gh/ezyang/3120/head -> origin/gh/ezyang/3120/head 2025-11-03T16:58:01.5261371Z * [new branch] gh/ezyang/3120/orig -> origin/gh/ezyang/3120/orig 2025-11-03T16:58:01.5263685Z * [new branch] gh/ezyang/3131/base -> origin/gh/ezyang/3131/base 2025-11-03T16:58:01.5265344Z * [new branch] gh/ezyang/3131/head -> origin/gh/ezyang/3131/head 2025-11-03T16:58:01.5266979Z * [new branch] gh/ezyang/3131/orig -> origin/gh/ezyang/3131/orig 2025-11-03T16:58:01.5269317Z * [new branch] gh/ezyang/3139/base -> origin/gh/ezyang/3139/base 2025-11-03T16:58:01.5271062Z * [new branch] gh/ezyang/3139/head -> origin/gh/ezyang/3139/head 2025-11-03T16:58:01.5272809Z * [new branch] gh/ezyang/3139/orig -> origin/gh/ezyang/3139/orig 2025-11-03T16:58:01.5275213Z * [new branch] gh/ezyang/3140/base -> origin/gh/ezyang/3140/base 2025-11-03T16:58:01.5276923Z * [new branch] gh/ezyang/3140/head -> origin/gh/ezyang/3140/head 2025-11-03T16:58:01.5278638Z * [new branch] gh/ezyang/3140/orig -> origin/gh/ezyang/3140/orig 2025-11-03T16:58:01.5280943Z * [new branch] gh/ezyang/3143/base -> origin/gh/ezyang/3143/base 2025-11-03T16:58:01.5282627Z * [new branch] gh/ezyang/3143/head -> origin/gh/ezyang/3143/head 2025-11-03T16:58:01.5284372Z * [new branch] gh/ezyang/3143/orig -> origin/gh/ezyang/3143/orig 2025-11-03T16:58:01.5286679Z * [new branch] gh/ezyang/3144/base -> origin/gh/ezyang/3144/base 2025-11-03T16:58:01.5288347Z * [new branch] gh/ezyang/3144/head -> origin/gh/ezyang/3144/head 2025-11-03T16:58:01.5290074Z * [new branch] gh/ezyang/3144/orig -> origin/gh/ezyang/3144/orig 2025-11-03T16:58:01.5292572Z * [new branch] gh/ezyang/3148/base -> origin/gh/ezyang/3148/base 2025-11-03T16:58:01.5294095Z * [new branch] gh/ezyang/3148/head -> origin/gh/ezyang/3148/head 2025-11-03T16:58:01.5295720Z * [new branch] gh/ezyang/3148/orig -> origin/gh/ezyang/3148/orig 2025-11-03T16:58:01.5298111Z * [new branch] gh/ezyang/3158/base -> origin/gh/ezyang/3158/base 2025-11-03T16:58:01.5299807Z * [new branch] gh/ezyang/3158/head -> origin/gh/ezyang/3158/head 2025-11-03T16:58:01.5301575Z * [new branch] gh/ezyang/3158/orig -> origin/gh/ezyang/3158/orig 2025-11-03T16:58:01.5303952Z * [new branch] gh/ezyang/3165/base -> origin/gh/ezyang/3165/base 2025-11-03T16:58:01.5305614Z * [new branch] gh/ezyang/3165/head -> origin/gh/ezyang/3165/head 2025-11-03T16:58:01.5307298Z * [new branch] gh/ezyang/3165/orig -> origin/gh/ezyang/3165/orig 2025-11-03T16:58:01.5309578Z * [new branch] gh/ezyang/3167/base -> origin/gh/ezyang/3167/base 2025-11-03T16:58:01.5311333Z * [new branch] gh/ezyang/3167/head -> origin/gh/ezyang/3167/head 2025-11-03T16:58:01.5313462Z * [new branch] gh/ezyang/3167/orig -> origin/gh/ezyang/3167/orig 2025-11-03T16:58:01.5315754Z * [new branch] gh/ezyang/3168/base -> origin/gh/ezyang/3168/base 2025-11-03T16:58:01.5317446Z * [new branch] gh/ezyang/3168/head -> origin/gh/ezyang/3168/head 2025-11-03T16:58:01.5319170Z * [new branch] gh/ezyang/3168/orig -> origin/gh/ezyang/3168/orig 2025-11-03T16:58:01.5321535Z * [new branch] gh/ezyang/3169/base -> origin/gh/ezyang/3169/base 2025-11-03T16:58:01.5323711Z * [new branch] gh/ezyang/3169/head -> origin/gh/ezyang/3169/head 2025-11-03T16:58:01.5325353Z * [new branch] gh/ezyang/3169/orig -> origin/gh/ezyang/3169/orig 2025-11-03T16:58:01.5328121Z * [new branch] gh/ezyang/3170/base -> origin/gh/ezyang/3170/base 2025-11-03T16:58:01.5329834Z * [new branch] gh/ezyang/3170/head -> origin/gh/ezyang/3170/head 2025-11-03T16:58:01.5331640Z * [new branch] gh/ezyang/3170/orig -> origin/gh/ezyang/3170/orig 2025-11-03T16:58:01.5333948Z * [new branch] gh/ezyang/3171/base -> origin/gh/ezyang/3171/base 2025-11-03T16:58:01.5335673Z * [new branch] gh/ezyang/3171/head -> origin/gh/ezyang/3171/head 2025-11-03T16:58:01.5337427Z * [new branch] gh/ezyang/3171/orig -> origin/gh/ezyang/3171/orig 2025-11-03T16:58:01.5339761Z * [new branch] gh/ezyang/3172/base -> origin/gh/ezyang/3172/base 2025-11-03T16:58:01.5341721Z * [new branch] gh/ezyang/3172/head -> origin/gh/ezyang/3172/head 2025-11-03T16:58:01.5343433Z * [new branch] gh/ezyang/3172/orig -> origin/gh/ezyang/3172/orig 2025-11-03T16:58:01.5345882Z * [new branch] gh/ezyang/3173/base -> origin/gh/ezyang/3173/base 2025-11-03T16:58:01.5347541Z * [new branch] gh/ezyang/3173/head -> origin/gh/ezyang/3173/head 2025-11-03T16:58:01.5349187Z * [new branch] gh/ezyang/3173/orig -> origin/gh/ezyang/3173/orig 2025-11-03T16:58:01.5351600Z * [new branch] gh/ezyang/3174/base -> origin/gh/ezyang/3174/base 2025-11-03T16:58:01.5353300Z * [new branch] gh/ezyang/3174/head -> origin/gh/ezyang/3174/head 2025-11-03T16:58:01.5354982Z * [new branch] gh/ezyang/3174/orig -> origin/gh/ezyang/3174/orig 2025-11-03T16:58:01.5357434Z * [new branch] gh/ezyang/3175/base -> origin/gh/ezyang/3175/base 2025-11-03T16:58:01.5359140Z * [new branch] gh/ezyang/3175/head -> origin/gh/ezyang/3175/head 2025-11-03T16:58:01.5361002Z * [new branch] gh/ezyang/3175/orig -> origin/gh/ezyang/3175/orig 2025-11-03T16:58:01.5363330Z * [new branch] gh/ezyang/3176/base -> origin/gh/ezyang/3176/base 2025-11-03T16:58:01.5365098Z * [new branch] gh/ezyang/3176/head -> origin/gh/ezyang/3176/head 2025-11-03T16:58:01.5366688Z * [new branch] gh/ezyang/3176/orig -> origin/gh/ezyang/3176/orig 2025-11-03T16:58:01.5369076Z * [new branch] gh/ezyang/3177/base -> origin/gh/ezyang/3177/base 2025-11-03T16:58:01.5370772Z * [new branch] gh/ezyang/3177/head -> origin/gh/ezyang/3177/head 2025-11-03T16:58:01.5372477Z * [new branch] gh/ezyang/3177/orig -> origin/gh/ezyang/3177/orig 2025-11-03T16:58:01.5374797Z * [new branch] gh/ezyang/3178/base -> origin/gh/ezyang/3178/base 2025-11-03T16:58:01.5376476Z * [new branch] gh/ezyang/3178/head -> origin/gh/ezyang/3178/head 2025-11-03T16:58:01.5378226Z * [new branch] gh/ezyang/3178/orig -> origin/gh/ezyang/3178/orig 2025-11-03T16:58:01.5380540Z * [new branch] gh/ezyang/3179/base -> origin/gh/ezyang/3179/base 2025-11-03T16:58:01.5382345Z * [new branch] gh/ezyang/3179/head -> origin/gh/ezyang/3179/head 2025-11-03T16:58:01.5384061Z * [new branch] gh/ezyang/3179/orig -> origin/gh/ezyang/3179/orig 2025-11-03T16:58:01.5386481Z * [new branch] gh/ezyang/3180/base -> origin/gh/ezyang/3180/base 2025-11-03T16:58:01.5388148Z * [new branch] gh/ezyang/3180/head -> origin/gh/ezyang/3180/head 2025-11-03T16:58:01.5389841Z * [new branch] gh/ezyang/3180/orig -> origin/gh/ezyang/3180/orig 2025-11-03T16:58:01.5392310Z * [new branch] gh/ezyang/3181/base -> origin/gh/ezyang/3181/base 2025-11-03T16:58:01.5393915Z * [new branch] gh/ezyang/3181/head -> origin/gh/ezyang/3181/head 2025-11-03T16:58:01.5395623Z * [new branch] gh/ezyang/3181/orig -> origin/gh/ezyang/3181/orig 2025-11-03T16:58:01.5397957Z * [new branch] gh/ezyang/3182/base -> origin/gh/ezyang/3182/base 2025-11-03T16:58:01.5399621Z * [new branch] gh/ezyang/3182/head -> origin/gh/ezyang/3182/head 2025-11-03T16:58:01.5401376Z * [new branch] gh/ezyang/3182/orig -> origin/gh/ezyang/3182/orig 2025-11-03T16:58:01.5403630Z * [new branch] gh/ezyang/3183/base -> origin/gh/ezyang/3183/base 2025-11-03T16:58:01.5405421Z * [new branch] gh/ezyang/3183/head -> origin/gh/ezyang/3183/head 2025-11-03T16:58:01.5407090Z * [new branch] gh/ezyang/3183/orig -> origin/gh/ezyang/3183/orig 2025-11-03T16:58:01.5409423Z * [new branch] gh/ezyang/3184/base -> origin/gh/ezyang/3184/base 2025-11-03T16:58:01.5411055Z * [new branch] gh/ezyang/3184/head -> origin/gh/ezyang/3184/head 2025-11-03T16:58:01.5412777Z * [new branch] gh/ezyang/3184/orig -> origin/gh/ezyang/3184/orig 2025-11-03T16:58:01.5415466Z * [new branch] gh/ezyang/3185/base -> origin/gh/ezyang/3185/base 2025-11-03T16:58:01.5417204Z * [new branch] gh/ezyang/3185/head -> origin/gh/ezyang/3185/head 2025-11-03T16:58:01.5418915Z * [new branch] gh/ezyang/3185/orig -> origin/gh/ezyang/3185/orig 2025-11-03T16:58:01.5421352Z * [new branch] gh/ezyang/3186/base -> origin/gh/ezyang/3186/base 2025-11-03T16:58:01.5423219Z * [new branch] gh/ezyang/3186/head -> origin/gh/ezyang/3186/head 2025-11-03T16:58:01.5424898Z * [new branch] gh/ezyang/3186/orig -> origin/gh/ezyang/3186/orig 2025-11-03T16:58:01.5427268Z * [new branch] gh/ezyang/3187/base -> origin/gh/ezyang/3187/base 2025-11-03T16:58:01.5428951Z * [new branch] gh/ezyang/3187/head -> origin/gh/ezyang/3187/head 2025-11-03T16:58:01.5430897Z * [new branch] gh/ezyang/3187/orig -> origin/gh/ezyang/3187/orig 2025-11-03T16:58:01.5433161Z * [new branch] gh/ezyang/3188/base -> origin/gh/ezyang/3188/base 2025-11-03T16:58:01.5434768Z * [new branch] gh/ezyang/3188/head -> origin/gh/ezyang/3188/head 2025-11-03T16:58:01.5436488Z * [new branch] gh/ezyang/3188/orig -> origin/gh/ezyang/3188/orig 2025-11-03T16:58:01.5438969Z * [new branch] gh/ezyang/3189/base -> origin/gh/ezyang/3189/base 2025-11-03T16:58:01.5440612Z * [new branch] gh/ezyang/3189/head -> origin/gh/ezyang/3189/head 2025-11-03T16:58:01.5442406Z * [new branch] gh/ezyang/3189/orig -> origin/gh/ezyang/3189/orig 2025-11-03T16:58:01.5444736Z * [new branch] gh/ezyang/3190/base -> origin/gh/ezyang/3190/base 2025-11-03T16:58:01.5446464Z * [new branch] gh/ezyang/3190/head -> origin/gh/ezyang/3190/head 2025-11-03T16:58:01.5448080Z * [new branch] gh/ezyang/3190/orig -> origin/gh/ezyang/3190/orig 2025-11-03T16:58:01.5451014Z * [new branch] gh/fadara01/1/base -> origin/gh/fadara01/1/base 2025-11-03T16:58:01.5452718Z * [new branch] gh/fadara01/1/head -> origin/gh/fadara01/1/head 2025-11-03T16:58:01.5454362Z * [new branch] gh/fadara01/1/orig -> origin/gh/fadara01/1/orig 2025-11-03T16:58:01.5457674Z * [new branch] gh/fadara01/2/base -> origin/gh/fadara01/2/base 2025-11-03T16:58:01.5459403Z * [new branch] gh/fadara01/2/head -> origin/gh/fadara01/2/head 2025-11-03T16:58:01.5461198Z * [new branch] gh/fadara01/2/orig -> origin/gh/fadara01/2/orig 2025-11-03T16:58:01.5463620Z * [new branch] gh/fadara01/3/base -> origin/gh/fadara01/3/base 2025-11-03T16:58:01.5465303Z * [new branch] gh/fadara01/3/head -> origin/gh/fadara01/3/head 2025-11-03T16:58:01.5467264Z * [new branch] gh/fadara01/3/orig -> origin/gh/fadara01/3/orig 2025-11-03T16:58:01.5469435Z * [new branch] gh/fadara01/4/base -> origin/gh/fadara01/4/base 2025-11-03T16:58:01.5471107Z * [new branch] gh/fadara01/4/head -> origin/gh/fadara01/4/head 2025-11-03T16:58:01.5472815Z * [new branch] gh/fadara01/4/orig -> origin/gh/fadara01/4/orig 2025-11-03T16:58:01.5475098Z * [new branch] gh/fadara01/5/base -> origin/gh/fadara01/5/base 2025-11-03T16:58:01.5476779Z * [new branch] gh/fadara01/5/head -> origin/gh/fadara01/5/head 2025-11-03T16:58:01.5478499Z * [new branch] gh/fadara01/5/orig -> origin/gh/fadara01/5/orig 2025-11-03T16:58:01.5480768Z * [new branch] gh/fadara01/6/base -> origin/gh/fadara01/6/base 2025-11-03T16:58:01.5482964Z * [new branch] gh/fadara01/6/head -> origin/gh/fadara01/6/head 2025-11-03T16:58:01.5484754Z * [new branch] gh/fadara01/6/orig -> origin/gh/fadara01/6/orig 2025-11-03T16:58:01.5487459Z * [new branch] gh/fadara01/7/base -> origin/gh/fadara01/7/base 2025-11-03T16:58:01.5489180Z * [new branch] gh/fadara01/7/head -> origin/gh/fadara01/7/head 2025-11-03T16:58:01.5490706Z * [new branch] gh/fadara01/7/orig -> origin/gh/fadara01/7/orig 2025-11-03T16:58:01.5493333Z * [new branch] gh/fduwjj/182/base -> origin/gh/fduwjj/182/base 2025-11-03T16:58:01.5495042Z * [new branch] gh/fduwjj/182/head -> origin/gh/fduwjj/182/head 2025-11-03T16:58:01.5496923Z * [new branch] gh/fduwjj/182/orig -> origin/gh/fduwjj/182/orig 2025-11-03T16:58:01.5499226Z * [new branch] gh/fduwjj/185/base -> origin/gh/fduwjj/185/base 2025-11-03T16:58:01.5500958Z * [new branch] gh/fduwjj/185/head -> origin/gh/fduwjj/185/head 2025-11-03T16:58:01.5502628Z * [new branch] gh/fduwjj/185/orig -> origin/gh/fduwjj/185/orig 2025-11-03T16:58:01.5505029Z * [new branch] gh/fduwjj/206/base -> origin/gh/fduwjj/206/base 2025-11-03T16:58:01.5507495Z * [new branch] gh/fduwjj/206/head -> origin/gh/fduwjj/206/head 2025-11-03T16:58:01.5510767Z * [new branch] gh/fduwjj/206/orig -> origin/gh/fduwjj/206/orig 2025-11-03T16:58:01.5513335Z * [new branch] gh/fduwjj/211/base -> origin/gh/fduwjj/211/base 2025-11-03T16:58:01.5515677Z * [new branch] gh/fduwjj/211/head -> origin/gh/fduwjj/211/head 2025-11-03T16:58:01.5518001Z * [new branch] gh/fduwjj/211/orig -> origin/gh/fduwjj/211/orig 2025-11-03T16:58:01.5521313Z * [new branch] gh/fduwjj/212/base -> origin/gh/fduwjj/212/base 2025-11-03T16:58:01.5523705Z * [new branch] gh/fduwjj/212/head -> origin/gh/fduwjj/212/head 2025-11-03T16:58:01.5525395Z * [new branch] gh/fduwjj/212/orig -> origin/gh/fduwjj/212/orig 2025-11-03T16:58:01.5527700Z * [new branch] gh/fduwjj/213/base -> origin/gh/fduwjj/213/base 2025-11-03T16:58:01.5529397Z * [new branch] gh/fduwjj/213/head -> origin/gh/fduwjj/213/head 2025-11-03T16:58:01.5531079Z * [new branch] gh/fduwjj/213/orig -> origin/gh/fduwjj/213/orig 2025-11-03T16:58:01.5533428Z * [new branch] gh/fduwjj/217/base -> origin/gh/fduwjj/217/base 2025-11-03T16:58:01.5535268Z * [new branch] gh/fduwjj/217/head -> origin/gh/fduwjj/217/head 2025-11-03T16:58:01.5536977Z * [new branch] gh/fduwjj/217/orig -> origin/gh/fduwjj/217/orig 2025-11-03T16:58:01.5539365Z * [new branch] gh/fduwjj/219/base -> origin/gh/fduwjj/219/base 2025-11-03T16:58:01.5541073Z * [new branch] gh/fduwjj/219/head -> origin/gh/fduwjj/219/head 2025-11-03T16:58:01.5542971Z * [new branch] gh/fduwjj/219/orig -> origin/gh/fduwjj/219/orig 2025-11-03T16:58:01.5545286Z * [new branch] gh/fduwjj/220/base -> origin/gh/fduwjj/220/base 2025-11-03T16:58:01.5547020Z * [new branch] gh/fduwjj/220/head -> origin/gh/fduwjj/220/head 2025-11-03T16:58:01.5548879Z * [new branch] gh/fduwjj/220/orig -> origin/gh/fduwjj/220/orig 2025-11-03T16:58:01.5551172Z * [new branch] gh/fduwjj/221/base -> origin/gh/fduwjj/221/base 2025-11-03T16:58:01.5553026Z * [new branch] gh/fduwjj/221/head -> origin/gh/fduwjj/221/head 2025-11-03T16:58:01.5554748Z * [new branch] gh/fduwjj/221/orig -> origin/gh/fduwjj/221/orig 2025-11-03T16:58:01.5556921Z * [new branch] gh/fduwjj/222/base -> origin/gh/fduwjj/222/base 2025-11-03T16:58:01.5558674Z * [new branch] gh/fduwjj/222/head -> origin/gh/fduwjj/222/head 2025-11-03T16:58:01.5560880Z * [new branch] gh/fduwjj/222/orig -> origin/gh/fduwjj/222/orig 2025-11-03T16:58:01.5563214Z * [new branch] gh/fduwjj/223/base -> origin/gh/fduwjj/223/base 2025-11-03T16:58:01.5564862Z * [new branch] gh/fduwjj/223/head -> origin/gh/fduwjj/223/head 2025-11-03T16:58:01.5568566Z * [new branch] gh/fduwjj/223/orig -> origin/gh/fduwjj/223/orig 2025-11-03T16:58:01.5569737Z * [new branch] gh/fduwjj/224/base -> origin/gh/fduwjj/224/base 2025-11-03T16:58:01.5570886Z * [new branch] gh/fduwjj/224/head -> origin/gh/fduwjj/224/head 2025-11-03T16:58:01.5572566Z * [new branch] gh/fduwjj/224/orig -> origin/gh/fduwjj/224/orig 2025-11-03T16:58:01.5574831Z * [new branch] gh/fduwjj/225/base -> origin/gh/fduwjj/225/base 2025-11-03T16:58:01.5576686Z * [new branch] gh/fduwjj/225/head -> origin/gh/fduwjj/225/head 2025-11-03T16:58:01.5578270Z * [new branch] gh/fduwjj/225/orig -> origin/gh/fduwjj/225/orig 2025-11-03T16:58:01.5580610Z * [new branch] gh/fduwjj/226/base -> origin/gh/fduwjj/226/base 2025-11-03T16:58:01.5582454Z * [new branch] gh/fduwjj/226/head -> origin/gh/fduwjj/226/head 2025-11-03T16:58:01.5584078Z * [new branch] gh/fduwjj/226/orig -> origin/gh/fduwjj/226/orig 2025-11-03T16:58:01.5586671Z * [new branch] gh/fduwjj/227/base -> origin/gh/fduwjj/227/base 2025-11-03T16:58:01.5588183Z * [new branch] gh/fduwjj/227/head -> origin/gh/fduwjj/227/head 2025-11-03T16:58:01.5589878Z * [new branch] gh/fduwjj/227/orig -> origin/gh/fduwjj/227/orig 2025-11-03T16:58:01.5592847Z * [new branch] gh/fduwjj/228/base -> origin/gh/fduwjj/228/base 2025-11-03T16:58:01.5595126Z * [new branch] gh/fduwjj/228/head -> origin/gh/fduwjj/228/head 2025-11-03T16:58:01.5597818Z * [new branch] gh/fduwjj/228/orig -> origin/gh/fduwjj/228/orig 2025-11-03T16:58:01.5601439Z * [new branch] gh/fduwjj/229/base -> origin/gh/fduwjj/229/base 2025-11-03T16:58:01.5603792Z * [new branch] gh/fduwjj/229/head -> origin/gh/fduwjj/229/head 2025-11-03T16:58:01.5606101Z * [new branch] gh/fduwjj/229/orig -> origin/gh/fduwjj/229/orig 2025-11-03T16:58:01.5608407Z * [new branch] gh/fduwjj/230/base -> origin/gh/fduwjj/230/base 2025-11-03T16:58:01.5610205Z * [new branch] gh/fduwjj/230/head -> origin/gh/fduwjj/230/head 2025-11-03T16:58:01.5611910Z * [new branch] gh/fduwjj/230/orig -> origin/gh/fduwjj/230/orig 2025-11-03T16:58:01.5614264Z * [new branch] gh/fduwjj/231/base -> origin/gh/fduwjj/231/base 2025-11-03T16:58:01.5616040Z * [new branch] gh/fduwjj/231/head -> origin/gh/fduwjj/231/head 2025-11-03T16:58:01.5617747Z * [new branch] gh/fduwjj/231/orig -> origin/gh/fduwjj/231/orig 2025-11-03T16:58:01.5620067Z * [new branch] gh/fduwjj/232/base -> origin/gh/fduwjj/232/base 2025-11-03T16:58:01.5622819Z * [new branch] gh/fduwjj/232/head -> origin/gh/fduwjj/232/head 2025-11-03T16:58:01.5624693Z * [new branch] gh/fduwjj/232/orig -> origin/gh/fduwjj/232/orig 2025-11-03T16:58:01.5626867Z * [new branch] gh/fduwjj/233/base -> origin/gh/fduwjj/233/base 2025-11-03T16:58:01.5628539Z * [new branch] gh/fduwjj/233/head -> origin/gh/fduwjj/233/head 2025-11-03T16:58:01.5630249Z * [new branch] gh/fduwjj/233/orig -> origin/gh/fduwjj/233/orig 2025-11-03T16:58:01.5633144Z * [new branch] gh/fegin/318/base -> origin/gh/fegin/318/base 2025-11-03T16:58:01.5634932Z * [new branch] gh/fegin/318/head -> origin/gh/fegin/318/head 2025-11-03T16:58:01.5636803Z * [new branch] gh/fegin/318/orig -> origin/gh/fegin/318/orig 2025-11-03T16:58:01.5639099Z * [new branch] gh/fegin/321/base -> origin/gh/fegin/321/base 2025-11-03T16:58:01.5640779Z * [new branch] gh/fegin/321/head -> origin/gh/fegin/321/head 2025-11-03T16:58:01.5642813Z * [new branch] gh/fegin/321/orig -> origin/gh/fegin/321/orig 2025-11-03T16:58:01.5644793Z * [new branch] gh/fegin/325/base -> origin/gh/fegin/325/base 2025-11-03T16:58:01.5646425Z * [new branch] gh/fegin/325/head -> origin/gh/fegin/325/head 2025-11-03T16:58:01.5648111Z * [new branch] gh/fegin/325/orig -> origin/gh/fegin/325/orig 2025-11-03T16:58:01.5650431Z * [new branch] gh/fegin/326/base -> origin/gh/fegin/326/base 2025-11-03T16:58:01.5652243Z * [new branch] gh/fegin/326/head -> origin/gh/fegin/326/head 2025-11-03T16:58:01.5653842Z * [new branch] gh/fegin/326/orig -> origin/gh/fegin/326/orig 2025-11-03T16:58:01.5656275Z * [new branch] gh/fegin/327/base -> origin/gh/fegin/327/base 2025-11-03T16:58:01.5657901Z * [new branch] gh/fegin/327/head -> origin/gh/fegin/327/head 2025-11-03T16:58:01.5659574Z * [new branch] gh/fegin/327/orig -> origin/gh/fegin/327/orig 2025-11-03T16:58:01.5661815Z * [new branch] gh/fegin/328/base -> origin/gh/fegin/328/base 2025-11-03T16:58:01.5663672Z * [new branch] gh/fegin/328/head -> origin/gh/fegin/328/head 2025-11-03T16:58:01.5665533Z * [new branch] gh/fegin/328/orig -> origin/gh/fegin/328/orig 2025-11-03T16:58:01.5667802Z * [new branch] gh/fegin/329/base -> origin/gh/fegin/329/base 2025-11-03T16:58:01.5669496Z * [new branch] gh/fegin/329/head -> origin/gh/fegin/329/head 2025-11-03T16:58:01.5671228Z * [new branch] gh/fegin/329/orig -> origin/gh/fegin/329/orig 2025-11-03T16:58:01.5673492Z * [new branch] gh/fegin/330/base -> origin/gh/fegin/330/base 2025-11-03T16:58:01.5675149Z * [new branch] gh/fegin/330/head -> origin/gh/fegin/330/head 2025-11-03T16:58:01.5676891Z * [new branch] gh/fegin/330/orig -> origin/gh/fegin/330/orig 2025-11-03T16:58:01.5679309Z * [new branch] gh/fegin/331/base -> origin/gh/fegin/331/base 2025-11-03T16:58:01.5681068Z * [new branch] gh/fegin/331/head -> origin/gh/fegin/331/head 2025-11-03T16:58:01.5682664Z * [new branch] gh/fegin/331/orig -> origin/gh/fegin/331/orig 2025-11-03T16:58:01.5685474Z * [new branch] gh/fffrog/137/base -> origin/gh/fffrog/137/base 2025-11-03T16:58:01.5687148Z * [new branch] gh/fffrog/137/head -> origin/gh/fffrog/137/head 2025-11-03T16:58:01.5688815Z * [new branch] gh/fffrog/137/orig -> origin/gh/fffrog/137/orig 2025-11-03T16:58:01.5691096Z * [new branch] gh/fffrog/160/base -> origin/gh/fffrog/160/base 2025-11-03T16:58:01.5692749Z * [new branch] gh/fffrog/160/head -> origin/gh/fffrog/160/head 2025-11-03T16:58:01.5695012Z * [new branch] gh/fffrog/171/base -> origin/gh/fffrog/171/base 2025-11-03T16:58:01.5696678Z * [new branch] gh/fffrog/171/head -> origin/gh/fffrog/171/head 2025-11-03T16:58:01.5698320Z * [new branch] gh/fffrog/171/orig -> origin/gh/fffrog/171/orig 2025-11-03T16:58:01.5700766Z * [new branch] gh/fffrog/175/base -> origin/gh/fffrog/175/base 2025-11-03T16:58:01.5703240Z * [new branch] gh/fffrog/175/head -> origin/gh/fffrog/175/head 2025-11-03T16:58:01.5704938Z * [new branch] gh/fffrog/175/orig -> origin/gh/fffrog/175/orig 2025-11-03T16:58:01.5707271Z * [new branch] gh/fffrog/176/base -> origin/gh/fffrog/176/base 2025-11-03T16:58:01.5708949Z * [new branch] gh/fffrog/176/head -> origin/gh/fffrog/176/head 2025-11-03T16:58:01.5710641Z * [new branch] gh/fffrog/176/orig -> origin/gh/fffrog/176/orig 2025-11-03T16:58:01.5712852Z * [new branch] gh/fffrog/177/base -> origin/gh/fffrog/177/base 2025-11-03T16:58:01.5714542Z * [new branch] gh/fffrog/177/head -> origin/gh/fffrog/177/head 2025-11-03T16:58:01.5716229Z * [new branch] gh/fffrog/177/orig -> origin/gh/fffrog/177/orig 2025-11-03T16:58:01.5718493Z * [new branch] gh/fffrog/178/base -> origin/gh/fffrog/178/base 2025-11-03T16:58:01.5720324Z * [new branch] gh/fffrog/178/head -> origin/gh/fffrog/178/head 2025-11-03T16:58:01.5722236Z * [new branch] gh/fffrog/178/orig -> origin/gh/fffrog/178/orig 2025-11-03T16:58:01.5724663Z * [new branch] gh/fffrog/179/base -> origin/gh/fffrog/179/base 2025-11-03T16:58:01.5726306Z * [new branch] gh/fffrog/179/head -> origin/gh/fffrog/179/head 2025-11-03T16:58:01.5728008Z * [new branch] gh/fffrog/179/orig -> origin/gh/fffrog/179/orig 2025-11-03T16:58:01.5730339Z * [new branch] gh/fffrog/180/base -> origin/gh/fffrog/180/base 2025-11-03T16:58:01.5731983Z * [new branch] gh/fffrog/180/head -> origin/gh/fffrog/180/head 2025-11-03T16:58:01.5733692Z * [new branch] gh/fffrog/180/orig -> origin/gh/fffrog/180/orig 2025-11-03T16:58:01.5736045Z * [new branch] gh/fffrog/181/base -> origin/gh/fffrog/181/base 2025-11-03T16:58:01.5737701Z * [new branch] gh/fffrog/181/head -> origin/gh/fffrog/181/head 2025-11-03T16:58:01.5739522Z * [new branch] gh/fffrog/181/orig -> origin/gh/fffrog/181/orig 2025-11-03T16:58:01.5741849Z * [new branch] gh/fffrog/182/base -> origin/gh/fffrog/182/base 2025-11-03T16:58:01.5743584Z * [new branch] gh/fffrog/182/head -> origin/gh/fffrog/182/head 2025-11-03T16:58:01.5745225Z * [new branch] gh/fffrog/182/orig -> origin/gh/fffrog/182/orig 2025-11-03T16:58:01.5748101Z * [new branch] gh/fxdawnn/1/base -> origin/gh/fxdawnn/1/base 2025-11-03T16:58:01.5749783Z * [new branch] gh/fxdawnn/1/head -> origin/gh/fxdawnn/1/head 2025-11-03T16:58:01.5751486Z * [new branch] gh/fxdawnn/1/orig -> origin/gh/fxdawnn/1/orig 2025-11-03T16:58:01.5753796Z * [new branch] gh/fxdawnn/2/base -> origin/gh/fxdawnn/2/base 2025-11-03T16:58:01.5755535Z * [new branch] gh/fxdawnn/2/head -> origin/gh/fxdawnn/2/head 2025-11-03T16:58:01.5757215Z * [new branch] gh/fxdawnn/2/orig -> origin/gh/fxdawnn/2/orig 2025-11-03T16:58:01.5759379Z * [new branch] gh/fxdawnn/3/base -> origin/gh/fxdawnn/3/base 2025-11-03T16:58:01.5761073Z * [new branch] gh/fxdawnn/3/head -> origin/gh/fxdawnn/3/head 2025-11-03T16:58:01.5762752Z * [new branch] gh/fxdawnn/3/orig -> origin/gh/fxdawnn/3/orig 2025-11-03T16:58:01.5765161Z * [new branch] gh/fxdawnn/4/base -> origin/gh/fxdawnn/4/base 2025-11-03T16:58:01.5766781Z * [new branch] gh/fxdawnn/4/orig -> origin/gh/fxdawnn/4/orig 2025-11-03T16:58:01.5769186Z * [new branch] gh/fxdawnn/5/base -> origin/gh/fxdawnn/5/base 2025-11-03T16:58:01.5770839Z * [new branch] gh/fxdawnn/5/head -> origin/gh/fxdawnn/5/head 2025-11-03T16:58:01.5772579Z * [new branch] gh/fxdawnn/5/orig -> origin/gh/fxdawnn/5/orig 2025-11-03T16:58:01.5774922Z * [new branch] gh/fxdawnn/6/base -> origin/gh/fxdawnn/6/base 2025-11-03T16:58:01.5776602Z * [new branch] gh/fxdawnn/6/head -> origin/gh/fxdawnn/6/head 2025-11-03T16:58:01.5778295Z * [new branch] gh/fxdawnn/6/orig -> origin/gh/fxdawnn/6/orig 2025-11-03T16:58:01.5780656Z * [new branch] gh/fxdawnn/7/base -> origin/gh/fxdawnn/7/base 2025-11-03T16:58:01.5782531Z * [new branch] gh/fxdawnn/7/head -> origin/gh/fxdawnn/7/head 2025-11-03T16:58:01.5784184Z * [new branch] gh/fxdawnn/7/orig -> origin/gh/fxdawnn/7/orig 2025-11-03T16:58:01.5786966Z * [new branch] gh/gmagogsfm/2/base -> origin/gh/gmagogsfm/2/base 2025-11-03T16:58:01.5788628Z * [new branch] gh/gmagogsfm/2/head -> origin/gh/gmagogsfm/2/head 2025-11-03T16:58:01.5790483Z * [new branch] gh/gmagogsfm/2/orig -> origin/gh/gmagogsfm/2/orig 2025-11-03T16:58:01.5792888Z * [new branch] gh/gmagogsfm/3/base -> origin/gh/gmagogsfm/3/base 2025-11-03T16:58:01.5794679Z * [new branch] gh/gmagogsfm/3/head -> origin/gh/gmagogsfm/3/head 2025-11-03T16:58:01.5796294Z * [new branch] gh/gmagogsfm/3/orig -> origin/gh/gmagogsfm/3/orig 2025-11-03T16:58:01.5799094Z * [new branch] gh/guangyey/134/base -> origin/gh/guangyey/134/base 2025-11-03T16:58:01.5800812Z * [new branch] gh/guangyey/134/head -> origin/gh/guangyey/134/head 2025-11-03T16:58:01.5802899Z * [new branch] gh/guangyey/134/orig -> origin/gh/guangyey/134/orig 2025-11-03T16:58:01.5805209Z * [new branch] gh/guangyey/163/base -> origin/gh/guangyey/163/base 2025-11-03T16:58:01.5806845Z * [new branch] gh/guangyey/163/head -> origin/gh/guangyey/163/head 2025-11-03T16:58:01.5808515Z * [new branch] gh/guangyey/163/orig -> origin/gh/guangyey/163/orig 2025-11-03T16:58:01.5810829Z * [new branch] gh/guangyey/168/base -> origin/gh/guangyey/168/base 2025-11-03T16:58:01.5812512Z * [new branch] gh/guangyey/168/head -> origin/gh/guangyey/168/head 2025-11-03T16:58:01.5814328Z * [new branch] gh/guangyey/168/orig -> origin/gh/guangyey/168/orig 2025-11-03T16:58:01.5816603Z * [new branch] gh/guangyey/169/base -> origin/gh/guangyey/169/base 2025-11-03T16:58:01.5818307Z * [new branch] gh/guangyey/169/head -> origin/gh/guangyey/169/head 2025-11-03T16:58:01.5819979Z * [new branch] gh/guangyey/169/orig -> origin/gh/guangyey/169/orig 2025-11-03T16:58:01.5822557Z * [new branch] gh/guangyey/170/base -> origin/gh/guangyey/170/base 2025-11-03T16:58:01.5824259Z * [new branch] gh/guangyey/170/head -> origin/gh/guangyey/170/head 2025-11-03T16:58:01.5825949Z * [new branch] gh/guangyey/170/orig -> origin/gh/guangyey/170/orig 2025-11-03T16:58:01.5828154Z * [new branch] gh/guangyey/171/base -> origin/gh/guangyey/171/base 2025-11-03T16:58:01.5829805Z * [new branch] gh/guangyey/171/head -> origin/gh/guangyey/171/head 2025-11-03T16:58:01.5831512Z * [new branch] gh/guangyey/171/orig -> origin/gh/guangyey/171/orig 2025-11-03T16:58:01.5833702Z * [new branch] gh/guangyey/176/base -> origin/gh/guangyey/176/base 2025-11-03T16:58:01.5835425Z * [new branch] gh/guangyey/176/head -> origin/gh/guangyey/176/head 2025-11-03T16:58:01.5837175Z * [new branch] gh/guangyey/176/orig -> origin/gh/guangyey/176/orig 2025-11-03T16:58:01.5839497Z * [new branch] gh/guangyey/178/base -> origin/gh/guangyey/178/base 2025-11-03T16:58:01.5841132Z * [new branch] gh/guangyey/178/head -> origin/gh/guangyey/178/head 2025-11-03T16:58:01.5842894Z * [new branch] gh/guangyey/178/orig -> origin/gh/guangyey/178/orig 2025-11-03T16:58:01.5845709Z * [new branch] gh/guangyey/181/base -> origin/gh/guangyey/181/base 2025-11-03T16:58:01.5847403Z * [new branch] gh/guangyey/181/head -> origin/gh/guangyey/181/head 2025-11-03T16:58:01.5849051Z * [new branch] gh/guangyey/181/orig -> origin/gh/guangyey/181/orig 2025-11-03T16:58:01.5851393Z * [new branch] gh/guangyey/182/base -> origin/gh/guangyey/182/base 2025-11-03T16:58:01.5853084Z * [new branch] gh/guangyey/182/head -> origin/gh/guangyey/182/head 2025-11-03T16:58:01.5854710Z * [new branch] gh/guangyey/182/orig -> origin/gh/guangyey/182/orig 2025-11-03T16:58:01.5856938Z * [new branch] gh/guangyey/183/base -> origin/gh/guangyey/183/base 2025-11-03T16:58:01.5858831Z * [new branch] gh/guangyey/183/head -> origin/gh/guangyey/183/head 2025-11-03T16:58:01.5860517Z * [new branch] gh/guangyey/183/orig -> origin/gh/guangyey/183/orig 2025-11-03T16:58:01.5863075Z * [new branch] gh/guangyey/185/base -> origin/gh/guangyey/185/base 2025-11-03T16:58:01.5864706Z * [new branch] gh/guangyey/185/head -> origin/gh/guangyey/185/head 2025-11-03T16:58:01.5866364Z * [new branch] gh/guangyey/185/orig -> origin/gh/guangyey/185/orig 2025-11-03T16:58:01.5868637Z * [new branch] gh/guangyey/186/base -> origin/gh/guangyey/186/base 2025-11-03T16:58:01.5870290Z * [new branch] gh/guangyey/186/head -> origin/gh/guangyey/186/head 2025-11-03T16:58:01.5872208Z * [new branch] gh/guangyey/186/orig -> origin/gh/guangyey/186/orig 2025-11-03T16:58:01.5874857Z * [new branch] gh/guangyey/187/base -> origin/gh/guangyey/187/base 2025-11-03T16:58:01.5876522Z * [new branch] gh/guangyey/187/head -> origin/gh/guangyey/187/head 2025-11-03T16:58:01.5878244Z * [new branch] gh/guangyey/187/orig -> origin/gh/guangyey/187/orig 2025-11-03T16:58:01.5880495Z * [new branch] gh/guangyey/188/base -> origin/gh/guangyey/188/base 2025-11-03T16:58:01.5882159Z * [new branch] gh/guangyey/188/head -> origin/gh/guangyey/188/head 2025-11-03T16:58:01.5883869Z * [new branch] gh/guangyey/188/orig -> origin/gh/guangyey/188/orig 2025-11-03T16:58:01.5886211Z * [new branch] gh/guangyey/190/base -> origin/gh/guangyey/190/base 2025-11-03T16:58:01.5887928Z * [new branch] gh/guangyey/190/head -> origin/gh/guangyey/190/head 2025-11-03T16:58:01.5889623Z * [new branch] gh/guangyey/190/orig -> origin/gh/guangyey/190/orig 2025-11-03T16:58:01.5892344Z * [new branch] gh/guangyey/195/base -> origin/gh/guangyey/195/base 2025-11-03T16:58:01.5894289Z * [new branch] gh/guangyey/195/head -> origin/gh/guangyey/195/head 2025-11-03T16:58:01.5895934Z * [new branch] gh/guangyey/195/orig -> origin/gh/guangyey/195/orig 2025-11-03T16:58:01.5898260Z * [new branch] gh/guangyey/208/base -> origin/gh/guangyey/208/base 2025-11-03T16:58:01.5899916Z * [new branch] gh/guangyey/208/head -> origin/gh/guangyey/208/head 2025-11-03T16:58:01.5901621Z * [new branch] gh/guangyey/208/orig -> origin/gh/guangyey/208/orig 2025-11-03T16:58:01.5904160Z * [new branch] gh/guangyey/210/base -> origin/gh/guangyey/210/base 2025-11-03T16:58:01.5905834Z * [new branch] gh/guangyey/210/head -> origin/gh/guangyey/210/head 2025-11-03T16:58:01.5907600Z * [new branch] gh/guangyey/210/orig -> origin/gh/guangyey/210/orig 2025-11-03T16:58:01.5909981Z * [new branch] gh/guangyey/212/base -> origin/gh/guangyey/212/base 2025-11-03T16:58:01.5911747Z * [new branch] gh/guangyey/212/head -> origin/gh/guangyey/212/head 2025-11-03T16:58:01.5913356Z * [new branch] gh/guangyey/212/orig -> origin/gh/guangyey/212/orig 2025-11-03T16:58:01.5915705Z * [new branch] gh/guangyey/213/base -> origin/gh/guangyey/213/base 2025-11-03T16:58:01.5917383Z * [new branch] gh/guangyey/213/head -> origin/gh/guangyey/213/head 2025-11-03T16:58:01.5919220Z * [new branch] gh/guangyey/213/orig -> origin/gh/guangyey/213/orig 2025-11-03T16:58:01.5921988Z * [new branch] gh/guangyey/214/base -> origin/gh/guangyey/214/base 2025-11-03T16:58:01.5923809Z * [new branch] gh/guangyey/214/head -> origin/gh/guangyey/214/head 2025-11-03T16:58:01.5925475Z * [new branch] gh/guangyey/214/orig -> origin/gh/guangyey/214/orig 2025-11-03T16:58:01.5928562Z * [new branch] gh/guangyey/215/base -> origin/gh/guangyey/215/base 2025-11-03T16:58:01.5930136Z * [new branch] gh/guangyey/215/head -> origin/gh/guangyey/215/head 2025-11-03T16:58:01.5931974Z * [new branch] gh/guangyey/215/orig -> origin/gh/guangyey/215/orig 2025-11-03T16:58:01.5934254Z * [new branch] gh/guangyey/216/base -> origin/gh/guangyey/216/base 2025-11-03T16:58:01.5935925Z * [new branch] gh/guangyey/216/head -> origin/gh/guangyey/216/head 2025-11-03T16:58:01.5937644Z * [new branch] gh/guangyey/216/orig -> origin/gh/guangyey/216/orig 2025-11-03T16:58:01.5940060Z * [new branch] gh/guangyey/217/base -> origin/gh/guangyey/217/base 2025-11-03T16:58:01.5941829Z * [new branch] gh/guangyey/217/head -> origin/gh/guangyey/217/head 2025-11-03T16:58:01.5943578Z * [new branch] gh/guangyey/217/orig -> origin/gh/guangyey/217/orig 2025-11-03T16:58:01.5945963Z * [new branch] gh/guangyey/218/base -> origin/gh/guangyey/218/base 2025-11-03T16:58:01.5947759Z * [new branch] gh/guangyey/218/head -> origin/gh/guangyey/218/head 2025-11-03T16:58:01.5949359Z * [new branch] gh/guangyey/218/orig -> origin/gh/guangyey/218/orig 2025-11-03T16:58:01.5951712Z * [new branch] gh/guangyey/219/base -> origin/gh/guangyey/219/base 2025-11-03T16:58:01.5953334Z * [new branch] gh/guangyey/219/head -> origin/gh/guangyey/219/head 2025-11-03T16:58:01.5955243Z * [new branch] gh/guangyey/219/orig -> origin/gh/guangyey/219/orig 2025-11-03T16:58:01.5957493Z * [new branch] gh/guangyey/220/base -> origin/gh/guangyey/220/base 2025-11-03T16:58:01.5959248Z * [new branch] gh/guangyey/220/head -> origin/gh/guangyey/220/head 2025-11-03T16:58:01.5960938Z * [new branch] gh/guangyey/220/orig -> origin/gh/guangyey/220/orig 2025-11-03T16:58:01.5963320Z * [new branch] gh/guangyey/221/base -> origin/gh/guangyey/221/base 2025-11-03T16:58:01.5964976Z * [new branch] gh/guangyey/221/head -> origin/gh/guangyey/221/head 2025-11-03T16:58:01.5966651Z * [new branch] gh/guangyey/221/orig -> origin/gh/guangyey/221/orig 2025-11-03T16:58:01.5968995Z * [new branch] gh/guangyey/222/base -> origin/gh/guangyey/222/base 2025-11-03T16:58:01.5970780Z * [new branch] gh/guangyey/222/head -> origin/gh/guangyey/222/head 2025-11-03T16:58:01.5972469Z * [new branch] gh/guangyey/222/orig -> origin/gh/guangyey/222/orig 2025-11-03T16:58:01.5974812Z * [new branch] gh/guangyey/223/base -> origin/gh/guangyey/223/base 2025-11-03T16:58:01.5976524Z * [new branch] gh/guangyey/223/head -> origin/gh/guangyey/223/head 2025-11-03T16:58:01.5978355Z * [new branch] gh/guangyey/223/orig -> origin/gh/guangyey/223/orig 2025-11-03T16:58:01.5980751Z * [new branch] gh/guangyey/224/base -> origin/gh/guangyey/224/base 2025-11-03T16:58:01.5982545Z * [new branch] gh/guangyey/224/head -> origin/gh/guangyey/224/head 2025-11-03T16:58:01.5984310Z * [new branch] gh/guangyey/224/orig -> origin/gh/guangyey/224/orig 2025-11-03T16:58:01.5986683Z * [new branch] gh/guangyey/225/base -> origin/gh/guangyey/225/base 2025-11-03T16:58:01.5988337Z * [new branch] gh/guangyey/225/head -> origin/gh/guangyey/225/head 2025-11-03T16:58:01.5990053Z * [new branch] gh/guangyey/225/orig -> origin/gh/guangyey/225/orig 2025-11-03T16:58:01.5992271Z * [new branch] gh/guangyey/226/base -> origin/gh/guangyey/226/base 2025-11-03T16:58:01.5994244Z * [new branch] gh/guangyey/226/head -> origin/gh/guangyey/226/head 2025-11-03T16:58:01.5995903Z * [new branch] gh/guangyey/226/orig -> origin/gh/guangyey/226/orig 2025-11-03T16:58:01.5998155Z * [new branch] gh/guangyey/227/base -> origin/gh/guangyey/227/base 2025-11-03T16:58:01.5999843Z * [new branch] gh/guangyey/227/head -> origin/gh/guangyey/227/head 2025-11-03T16:58:01.6002086Z * [new branch] gh/guangyey/227/orig -> origin/gh/guangyey/227/orig 2025-11-03T16:58:01.6004565Z * [new branch] gh/guangyey/228/base -> origin/gh/guangyey/228/base 2025-11-03T16:58:01.6006245Z * [new branch] gh/guangyey/228/head -> origin/gh/guangyey/228/head 2025-11-03T16:58:01.6007930Z * [new branch] gh/guangyey/228/orig -> origin/gh/guangyey/228/orig 2025-11-03T16:58:01.6010341Z * [new branch] gh/guangyey/229/base -> origin/gh/guangyey/229/base 2025-11-03T16:58:01.6012028Z * [new branch] gh/guangyey/229/head -> origin/gh/guangyey/229/head 2025-11-03T16:58:01.6013782Z * [new branch] gh/guangyey/229/orig -> origin/gh/guangyey/229/orig 2025-11-03T16:58:01.6016227Z * [new branch] gh/guangyey/230/base -> origin/gh/guangyey/230/base 2025-11-03T16:58:01.6017867Z * [new branch] gh/guangyey/230/head -> origin/gh/guangyey/230/head 2025-11-03T16:58:01.6019566Z * [new branch] gh/guangyey/230/orig -> origin/gh/guangyey/230/orig 2025-11-03T16:58:01.6023979Z * [new branch] gh/guangyey/231/base -> origin/gh/guangyey/231/base 2025-11-03T16:58:01.6026101Z * [new branch] gh/guangyey/231/head -> origin/gh/guangyey/231/head 2025-11-03T16:58:01.6027861Z * [new branch] gh/guangyey/231/orig -> origin/gh/guangyey/231/orig 2025-11-03T16:58:01.6030244Z * [new branch] gh/guangyey/232/base -> origin/gh/guangyey/232/base 2025-11-03T16:58:01.6031966Z * [new branch] gh/guangyey/232/head -> origin/gh/guangyey/232/head 2025-11-03T16:58:01.6033713Z * [new branch] gh/guangyey/232/orig -> origin/gh/guangyey/232/orig 2025-11-03T16:58:01.6036082Z * [new branch] gh/guangyey/233/base -> origin/gh/guangyey/233/base 2025-11-03T16:58:01.6037711Z * [new branch] gh/guangyey/233/head -> origin/gh/guangyey/233/head 2025-11-03T16:58:01.6039638Z * [new branch] gh/guangyey/233/orig -> origin/gh/guangyey/233/orig 2025-11-03T16:58:01.6041978Z * [new branch] gh/guangyey/234/base -> origin/gh/guangyey/234/base 2025-11-03T16:58:01.6043644Z * [new branch] gh/guangyey/234/head -> origin/gh/guangyey/234/head 2025-11-03T16:58:01.6045310Z * [new branch] gh/guangyey/234/orig -> origin/gh/guangyey/234/orig 2025-11-03T16:58:01.6047592Z * [new branch] gh/guangyey/235/base -> origin/gh/guangyey/235/base 2025-11-03T16:58:01.6049292Z * [new branch] gh/guangyey/235/head -> origin/gh/guangyey/235/head 2025-11-03T16:58:01.6051030Z * [new branch] gh/guangyey/235/orig -> origin/gh/guangyey/235/orig 2025-11-03T16:58:01.6053966Z * [new branch] gh/guilhermeleobas/107/base -> origin/gh/guilhermeleobas/107/base 2025-11-03T16:58:01.6055694Z * [new branch] gh/guilhermeleobas/107/head -> origin/gh/guilhermeleobas/107/head 2025-11-03T16:58:01.6057384Z * [new branch] gh/guilhermeleobas/107/orig -> origin/gh/guilhermeleobas/107/orig 2025-11-03T16:58:01.6059648Z * [new branch] gh/guilhermeleobas/108/base -> origin/gh/guilhermeleobas/108/base 2025-11-03T16:58:01.6061381Z * [new branch] gh/guilhermeleobas/108/head -> origin/gh/guilhermeleobas/108/head 2025-11-03T16:58:01.6063237Z * [new branch] gh/guilhermeleobas/108/orig -> origin/gh/guilhermeleobas/108/orig 2025-11-03T16:58:01.6065466Z * [new branch] gh/guilhermeleobas/147/base -> origin/gh/guilhermeleobas/147/base 2025-11-03T16:58:01.6067285Z * [new branch] gh/guilhermeleobas/147/head -> origin/gh/guilhermeleobas/147/head 2025-11-03T16:58:01.6068852Z * [new branch] gh/guilhermeleobas/147/orig -> origin/gh/guilhermeleobas/147/orig 2025-11-03T16:58:01.6071403Z * [new branch] gh/guilhermeleobas/150/base -> origin/gh/guilhermeleobas/150/base 2025-11-03T16:58:01.6073145Z * [new branch] gh/guilhermeleobas/150/head -> origin/gh/guilhermeleobas/150/head 2025-11-03T16:58:01.6074849Z * [new branch] gh/guilhermeleobas/150/orig -> origin/gh/guilhermeleobas/150/orig 2025-11-03T16:58:01.6077152Z * [new branch] gh/guilhermeleobas/166/base -> origin/gh/guilhermeleobas/166/base 2025-11-03T16:58:01.6078940Z * [new branch] gh/guilhermeleobas/166/head -> origin/gh/guilhermeleobas/166/head 2025-11-03T16:58:01.6080564Z * [new branch] gh/guilhermeleobas/166/orig -> origin/gh/guilhermeleobas/166/orig 2025-11-03T16:58:01.6082862Z * [new branch] gh/guilhermeleobas/168/base -> origin/gh/guilhermeleobas/168/base 2025-11-03T16:58:01.6084631Z * [new branch] gh/guilhermeleobas/168/head -> origin/gh/guilhermeleobas/168/head 2025-11-03T16:58:01.6086294Z * [new branch] gh/guilhermeleobas/168/orig -> origin/gh/guilhermeleobas/168/orig 2025-11-03T16:58:01.6088928Z * [new branch] gh/guilhermeleobas/169/base -> origin/gh/guilhermeleobas/169/base 2025-11-03T16:58:01.6090619Z * [new branch] gh/guilhermeleobas/169/head -> origin/gh/guilhermeleobas/169/head 2025-11-03T16:58:01.6092301Z * [new branch] gh/guilhermeleobas/169/orig -> origin/gh/guilhermeleobas/169/orig 2025-11-03T16:58:01.6094606Z * [new branch] gh/guilhermeleobas/170/base -> origin/gh/guilhermeleobas/170/base 2025-11-03T16:58:01.6096363Z * [new branch] gh/guilhermeleobas/170/head -> origin/gh/guilhermeleobas/170/head 2025-11-03T16:58:01.6098207Z * [new branch] gh/guilhermeleobas/170/orig -> origin/gh/guilhermeleobas/170/orig 2025-11-03T16:58:01.6100510Z * [new branch] gh/guilhermeleobas/171/base -> origin/gh/guilhermeleobas/171/base 2025-11-03T16:58:01.6102352Z * [new branch] gh/guilhermeleobas/171/head -> origin/gh/guilhermeleobas/171/head 2025-11-03T16:58:01.6104084Z * [new branch] gh/guilhermeleobas/171/orig -> origin/gh/guilhermeleobas/171/orig 2025-11-03T16:58:01.6106349Z * [new branch] gh/guilhermeleobas/173/base -> origin/gh/guilhermeleobas/173/base 2025-11-03T16:58:01.6108009Z * [new branch] gh/guilhermeleobas/173/head -> origin/gh/guilhermeleobas/173/head 2025-11-03T16:58:01.6109660Z * [new branch] gh/guilhermeleobas/173/orig -> origin/gh/guilhermeleobas/173/orig 2025-11-03T16:58:01.6111969Z * [new branch] gh/guilhermeleobas/193/base -> origin/gh/guilhermeleobas/193/base 2025-11-03T16:58:01.6113979Z * [new branch] gh/guilhermeleobas/193/head -> origin/gh/guilhermeleobas/193/head 2025-11-03T16:58:01.6115789Z * [new branch] gh/guilhermeleobas/193/orig -> origin/gh/guilhermeleobas/193/orig 2025-11-03T16:58:01.6118519Z * [new branch] gh/guilhermeleobas/204/base -> origin/gh/guilhermeleobas/204/base 2025-11-03T16:58:01.6120285Z * [new branch] gh/guilhermeleobas/204/head -> origin/gh/guilhermeleobas/204/head 2025-11-03T16:58:01.6122384Z * [new branch] gh/guilhermeleobas/204/orig -> origin/gh/guilhermeleobas/204/orig 2025-11-03T16:58:01.6124791Z * [new branch] gh/guilhermeleobas/211/base -> origin/gh/guilhermeleobas/211/base 2025-11-03T16:58:01.6126548Z * [new branch] gh/guilhermeleobas/211/head -> origin/gh/guilhermeleobas/211/head 2025-11-03T16:58:01.6128235Z * [new branch] gh/guilhermeleobas/211/orig -> origin/gh/guilhermeleobas/211/orig 2025-11-03T16:58:01.6130574Z * [new branch] gh/guilhermeleobas/226/base -> origin/gh/guilhermeleobas/226/base 2025-11-03T16:58:01.6132414Z * [new branch] gh/guilhermeleobas/226/head -> origin/gh/guilhermeleobas/226/head 2025-11-03T16:58:01.6133978Z * [new branch] gh/guilhermeleobas/226/orig -> origin/gh/guilhermeleobas/226/orig 2025-11-03T16:58:01.6136210Z * [new branch] gh/guilhermeleobas/236/base -> origin/gh/guilhermeleobas/236/base 2025-11-03T16:58:01.6137901Z * [new branch] gh/guilhermeleobas/236/head -> origin/gh/guilhermeleobas/236/head 2025-11-03T16:58:01.6139804Z * [new branch] gh/guilhermeleobas/236/orig -> origin/gh/guilhermeleobas/236/orig 2025-11-03T16:58:01.6141942Z * [new branch] gh/guilhermeleobas/239/base -> origin/gh/guilhermeleobas/239/base 2025-11-03T16:58:01.6143731Z * [new branch] gh/guilhermeleobas/239/head -> origin/gh/guilhermeleobas/239/head 2025-11-03T16:58:01.6145429Z * [new branch] gh/guilhermeleobas/239/orig -> origin/gh/guilhermeleobas/239/orig 2025-11-03T16:58:01.6148260Z * [new branch] gh/guilhermeleobas/246/base -> origin/gh/guilhermeleobas/246/base 2025-11-03T16:58:01.6149917Z * [new branch] gh/guilhermeleobas/246/head -> origin/gh/guilhermeleobas/246/head 2025-11-03T16:58:01.6151695Z * [new branch] gh/guilhermeleobas/246/orig -> origin/gh/guilhermeleobas/246/orig 2025-11-03T16:58:01.6153963Z * [new branch] gh/guilhermeleobas/247/base -> origin/gh/guilhermeleobas/247/base 2025-11-03T16:58:01.6155717Z * [new branch] gh/guilhermeleobas/247/head -> origin/gh/guilhermeleobas/247/head 2025-11-03T16:58:01.6157522Z * [new branch] gh/guilhermeleobas/247/orig -> origin/gh/guilhermeleobas/247/orig 2025-11-03T16:58:01.6160543Z * [new branch] gh/guilhermeleobas/248/base -> origin/gh/guilhermeleobas/248/base 2025-11-03T16:58:01.6162169Z * [new branch] gh/guilhermeleobas/248/head -> origin/gh/guilhermeleobas/248/head 2025-11-03T16:58:01.6164151Z * [new branch] gh/guilhermeleobas/248/orig -> origin/gh/guilhermeleobas/248/orig 2025-11-03T16:58:01.6166442Z * [new branch] gh/guilhermeleobas/250/base -> origin/gh/guilhermeleobas/250/base 2025-11-03T16:58:01.6168162Z * [new branch] gh/guilhermeleobas/250/head -> origin/gh/guilhermeleobas/250/head 2025-11-03T16:58:01.6169929Z * [new branch] gh/guilhermeleobas/250/orig -> origin/gh/guilhermeleobas/250/orig 2025-11-03T16:58:01.6172204Z * [new branch] gh/guilhermeleobas/251/base -> origin/gh/guilhermeleobas/251/base 2025-11-03T16:58:01.6174034Z * [new branch] gh/guilhermeleobas/251/head -> origin/gh/guilhermeleobas/251/head 2025-11-03T16:58:01.6175662Z * [new branch] gh/guilhermeleobas/251/orig -> origin/gh/guilhermeleobas/251/orig 2025-11-03T16:58:01.6178042Z * [new branch] gh/guilhermeleobas/252/base -> origin/gh/guilhermeleobas/252/base 2025-11-03T16:58:01.6179741Z * [new branch] gh/guilhermeleobas/252/head -> origin/gh/guilhermeleobas/252/head 2025-11-03T16:58:01.6181541Z * [new branch] gh/guilhermeleobas/252/orig -> origin/gh/guilhermeleobas/252/orig 2025-11-03T16:58:01.6184432Z * [new branch] gh/henrylhtsang/150/base -> origin/gh/henrylhtsang/150/base 2025-11-03T16:58:01.6186166Z * [new branch] gh/henrylhtsang/150/head -> origin/gh/henrylhtsang/150/head 2025-11-03T16:58:01.6187936Z * [new branch] gh/henrylhtsang/150/orig -> origin/gh/henrylhtsang/150/orig 2025-11-03T16:58:01.6190745Z * [new branch] gh/huydhn/1/next -> origin/gh/huydhn/1/next 2025-11-03T16:58:01.6192964Z * [new branch] gh/huydhn/2/next -> origin/gh/huydhn/2/next 2025-11-03T16:58:01.6195342Z * [new branch] gh/huydhn/3/next -> origin/gh/huydhn/3/next 2025-11-03T16:58:01.6197653Z * [new branch] gh/huydhn/4/next -> origin/gh/huydhn/4/next 2025-11-03T16:58:01.6200114Z * [new branch] gh/huydhn/5/next -> origin/gh/huydhn/5/next 2025-11-03T16:58:01.6202265Z * [new branch] gh/huydhn/6/next -> origin/gh/huydhn/6/next 2025-11-03T16:58:01.6204920Z * [new branch] gh/int3/97/base -> origin/gh/int3/97/base 2025-11-03T16:58:01.6206653Z * [new branch] gh/int3/97/head -> origin/gh/int3/97/head 2025-11-03T16:58:01.6209586Z * [new branch] gh/isuruf/101/base -> origin/gh/isuruf/101/base 2025-11-03T16:58:01.6211260Z * [new branch] gh/isuruf/101/head -> origin/gh/isuruf/101/head 2025-11-03T16:58:01.6213536Z * [new branch] gh/isuruf/146/base -> origin/gh/isuruf/146/base 2025-11-03T16:58:01.6215240Z * [new branch] gh/isuruf/146/head -> origin/gh/isuruf/146/head 2025-11-03T16:58:01.6216944Z * [new branch] gh/isuruf/146/orig -> origin/gh/isuruf/146/orig 2025-11-03T16:58:01.6219770Z * [new branch] gh/isuruf/157/base -> origin/gh/isuruf/157/base 2025-11-03T16:58:01.6221677Z * [new branch] gh/isuruf/157/head -> origin/gh/isuruf/157/head 2025-11-03T16:58:01.6224882Z * [new branch] gh/isuruf/157/orig -> origin/gh/isuruf/157/orig 2025-11-03T16:58:01.6227111Z * [new branch] gh/isuruf/81/base -> origin/gh/isuruf/81/base 2025-11-03T16:58:01.6228800Z * [new branch] gh/isuruf/81/head -> origin/gh/isuruf/81/head 2025-11-03T16:58:01.6230496Z * [new branch] gh/isuruf/81/orig -> origin/gh/isuruf/81/orig 2025-11-03T16:58:01.6233358Z * [new branch] gh/jamesjwu/176/base -> origin/gh/jamesjwu/176/base 2025-11-03T16:58:01.6235019Z * [new branch] gh/jamesjwu/176/head -> origin/gh/jamesjwu/176/head 2025-11-03T16:58:01.6236701Z * [new branch] gh/jamesjwu/176/orig -> origin/gh/jamesjwu/176/orig 2025-11-03T16:58:01.6238975Z * [new branch] gh/jamesjwu/187/base -> origin/gh/jamesjwu/187/base 2025-11-03T16:58:01.6240640Z * [new branch] gh/jamesjwu/187/head -> origin/gh/jamesjwu/187/head 2025-11-03T16:58:01.6242320Z * [new branch] gh/jamesjwu/187/orig -> origin/gh/jamesjwu/187/orig 2025-11-03T16:58:01.6244779Z * [new branch] gh/jamesjwu/195/base -> origin/gh/jamesjwu/195/base 2025-11-03T16:58:01.6246450Z * [new branch] gh/jamesjwu/195/head -> origin/gh/jamesjwu/195/head 2025-11-03T16:58:01.6248067Z * [new branch] gh/jamesjwu/195/orig -> origin/gh/jamesjwu/195/orig 2025-11-03T16:58:01.6250358Z * [new branch] gh/jamesjwu/196/base -> origin/gh/jamesjwu/196/base 2025-11-03T16:58:01.6252151Z * [new branch] gh/jamesjwu/196/head -> origin/gh/jamesjwu/196/head 2025-11-03T16:58:01.6253942Z * [new branch] gh/jamesjwu/196/orig -> origin/gh/jamesjwu/196/orig 2025-11-03T16:58:01.6256188Z * [new branch] gh/jamesjwu/197/base -> origin/gh/jamesjwu/197/base 2025-11-03T16:58:01.6257912Z * [new branch] gh/jamesjwu/197/head -> origin/gh/jamesjwu/197/head 2025-11-03T16:58:01.6259672Z * [new branch] gh/jamesjwu/197/orig -> origin/gh/jamesjwu/197/orig 2025-11-03T16:58:01.6262075Z * [new branch] gh/jamesjwu/198/base -> origin/gh/jamesjwu/198/base 2025-11-03T16:58:01.6263705Z * [new branch] gh/jamesjwu/198/head -> origin/gh/jamesjwu/198/head 2025-11-03T16:58:01.6265348Z * [new branch] gh/jamesjwu/198/orig -> origin/gh/jamesjwu/198/orig 2025-11-03T16:58:01.6267644Z * [new branch] gh/jamesjwu/199/base -> origin/gh/jamesjwu/199/base 2025-11-03T16:58:01.6269394Z * [new branch] gh/jamesjwu/199/head -> origin/gh/jamesjwu/199/head 2025-11-03T16:58:01.6271069Z * [new branch] gh/jamesjwu/199/orig -> origin/gh/jamesjwu/199/orig 2025-11-03T16:58:01.6273508Z * [new branch] gh/jamesjwu/200/base -> origin/gh/jamesjwu/200/base 2025-11-03T16:58:01.6275036Z * [new branch] gh/jamesjwu/200/head -> origin/gh/jamesjwu/200/head 2025-11-03T16:58:01.6276799Z * [new branch] gh/jamesjwu/200/orig -> origin/gh/jamesjwu/200/orig 2025-11-03T16:58:01.6279117Z * [new branch] gh/jamesjwu/201/base -> origin/gh/jamesjwu/201/base 2025-11-03T16:58:01.6280814Z * [new branch] gh/jamesjwu/201/head -> origin/gh/jamesjwu/201/head 2025-11-03T16:58:01.6282498Z * [new branch] gh/jamesjwu/201/orig -> origin/gh/jamesjwu/201/orig 2025-11-03T16:58:01.6285004Z * [new branch] gh/jamesjwu/202/base -> origin/gh/jamesjwu/202/base 2025-11-03T16:58:01.6286591Z * [new branch] gh/jamesjwu/202/head -> origin/gh/jamesjwu/202/head 2025-11-03T16:58:01.6288233Z * [new branch] gh/jamesjwu/202/orig -> origin/gh/jamesjwu/202/orig 2025-11-03T16:58:01.6290612Z * [new branch] gh/jamesjwu/203/base -> origin/gh/jamesjwu/203/base 2025-11-03T16:58:01.6292272Z * [new branch] gh/jamesjwu/203/head -> origin/gh/jamesjwu/203/head 2025-11-03T16:58:01.6293906Z * [new branch] gh/jamesjwu/203/orig -> origin/gh/jamesjwu/203/orig 2025-11-03T16:58:01.6296313Z * [new branch] gh/jamesjwu/204/base -> origin/gh/jamesjwu/204/base 2025-11-03T16:58:01.6298021Z * [new branch] gh/jamesjwu/204/head -> origin/gh/jamesjwu/204/head 2025-11-03T16:58:01.6299795Z * [new branch] gh/jamesjwu/204/orig -> origin/gh/jamesjwu/204/orig 2025-11-03T16:58:01.6302906Z * [new branch] gh/jamesjwu/205/base -> origin/gh/jamesjwu/205/base 2025-11-03T16:58:01.6304529Z * [new branch] gh/jamesjwu/205/head -> origin/gh/jamesjwu/205/head 2025-11-03T16:58:01.6306226Z * [new branch] gh/jamesjwu/205/orig -> origin/gh/jamesjwu/205/orig 2025-11-03T16:58:01.6308407Z * [new branch] gh/jamesjwu/206/base -> origin/gh/jamesjwu/206/base 2025-11-03T16:58:01.6310192Z * [new branch] gh/jamesjwu/206/head -> origin/gh/jamesjwu/206/head 2025-11-03T16:58:01.6311954Z * [new branch] gh/jamesjwu/206/orig -> origin/gh/jamesjwu/206/orig 2025-11-03T16:58:01.6314391Z * [new branch] gh/jamesjwu/52/base -> origin/gh/jamesjwu/52/base 2025-11-03T16:58:01.6316086Z * [new branch] gh/jamesjwu/52/head -> origin/gh/jamesjwu/52/head 2025-11-03T16:58:01.6318239Z * [new branch] gh/jamesjwu/53/base -> origin/gh/jamesjwu/53/base 2025-11-03T16:58:01.6319894Z * [new branch] gh/jamesjwu/53/head -> origin/gh/jamesjwu/53/head 2025-11-03T16:58:01.6322438Z * [new branch] gh/jamesjwu/54/base -> origin/gh/jamesjwu/54/base 2025-11-03T16:58:01.6324160Z * [new branch] gh/jamesjwu/54/head -> origin/gh/jamesjwu/54/head 2025-11-03T16:58:01.6326930Z * [new branch] gh/jamesjwu/55/base -> origin/gh/jamesjwu/55/base 2025-11-03T16:58:01.6328616Z * [new branch] gh/jamesjwu/55/head -> origin/gh/jamesjwu/55/head 2025-11-03T16:58:01.6330842Z * [new branch] gh/jamesjwu/56/base -> origin/gh/jamesjwu/56/base 2025-11-03T16:58:01.6332496Z * [new branch] gh/jamesjwu/56/head -> origin/gh/jamesjwu/56/head 2025-11-03T16:58:01.6334783Z * [new branch] gh/jamesjwu/57/base -> origin/gh/jamesjwu/57/base 2025-11-03T16:58:01.6336425Z * [new branch] gh/jamesjwu/57/head -> origin/gh/jamesjwu/57/head 2025-11-03T16:58:01.6338681Z * [new branch] gh/jamesjwu/58/base -> origin/gh/jamesjwu/58/base 2025-11-03T16:58:01.6340353Z * [new branch] gh/jamesjwu/58/head -> origin/gh/jamesjwu/58/head 2025-11-03T16:58:01.6342874Z * [new branch] gh/jamesjwu/59/base -> origin/gh/jamesjwu/59/base 2025-11-03T16:58:01.6344503Z * [new branch] gh/jamesjwu/59/head -> origin/gh/jamesjwu/59/head 2025-11-03T16:58:01.6346900Z * [new branch] gh/jamesjwu/60/base -> origin/gh/jamesjwu/60/base 2025-11-03T16:58:01.6348518Z * [new branch] gh/jamesjwu/60/head -> origin/gh/jamesjwu/60/head 2025-11-03T16:58:01.6350731Z * [new branch] gh/jamesjwu/61/base -> origin/gh/jamesjwu/61/base 2025-11-03T16:58:01.6352775Z * [new branch] gh/jamesjwu/61/head -> origin/gh/jamesjwu/61/head 2025-11-03T16:58:01.6355054Z * [new branch] gh/jamesjwu/62/base -> origin/gh/jamesjwu/62/base 2025-11-03T16:58:01.6356700Z * [new branch] gh/jamesjwu/62/head -> origin/gh/jamesjwu/62/head 2025-11-03T16:58:01.6358914Z * [new branch] gh/jamesjwu/63/base -> origin/gh/jamesjwu/63/base 2025-11-03T16:58:01.6360596Z * [new branch] gh/jamesjwu/63/head -> origin/gh/jamesjwu/63/head 2025-11-03T16:58:01.6362980Z * [new branch] gh/jamesjwu/64/base -> origin/gh/jamesjwu/64/base 2025-11-03T16:58:01.6365107Z * [new branch] gh/jamesjwu/64/head -> origin/gh/jamesjwu/64/head 2025-11-03T16:58:01.6367524Z * [new branch] gh/jamesjwu/65/base -> origin/gh/jamesjwu/65/base 2025-11-03T16:58:01.6369138Z * [new branch] gh/jamesjwu/65/head -> origin/gh/jamesjwu/65/head 2025-11-03T16:58:01.6371929Z * [new branch] gh/janeyx99/165/base -> origin/gh/janeyx99/165/base 2025-11-03T16:58:01.6373689Z * [new branch] gh/janeyx99/165/head -> origin/gh/janeyx99/165/head 2025-11-03T16:58:01.6375430Z * [new branch] gh/janeyx99/165/orig -> origin/gh/janeyx99/165/orig 2025-11-03T16:58:01.6377671Z * [new branch] gh/janeyx99/201/base -> origin/gh/janeyx99/201/base 2025-11-03T16:58:01.6379436Z * [new branch] gh/janeyx99/201/head -> origin/gh/janeyx99/201/head 2025-11-03T16:58:01.6381072Z * [new branch] gh/janeyx99/201/orig -> origin/gh/janeyx99/201/orig 2025-11-03T16:58:01.6383801Z * [new branch] gh/janeyx99/225/base -> origin/gh/janeyx99/225/base 2025-11-03T16:58:01.6385539Z * [new branch] gh/janeyx99/225/head -> origin/gh/janeyx99/225/head 2025-11-03T16:58:01.6387231Z * [new branch] gh/janeyx99/225/orig -> origin/gh/janeyx99/225/orig 2025-11-03T16:58:01.6389559Z * [new branch] gh/janeyx99/299/base -> origin/gh/janeyx99/299/base 2025-11-03T16:58:01.6391240Z * [new branch] gh/janeyx99/299/head -> origin/gh/janeyx99/299/head 2025-11-03T16:58:01.6392895Z * [new branch] gh/janeyx99/299/orig -> origin/gh/janeyx99/299/orig 2025-11-03T16:58:01.6395599Z * [new branch] gh/janeyx99/302/base -> origin/gh/janeyx99/302/base 2025-11-03T16:58:01.6397350Z * [new branch] gh/janeyx99/302/head -> origin/gh/janeyx99/302/head 2025-11-03T16:58:01.6399483Z * [new branch] gh/janeyx99/303/base -> origin/gh/janeyx99/303/base 2025-11-03T16:58:01.6401159Z * [new branch] gh/janeyx99/303/head -> origin/gh/janeyx99/303/head 2025-11-03T16:58:01.6403369Z * [new branch] gh/janeyx99/305/base -> origin/gh/janeyx99/305/base 2025-11-03T16:58:01.6405054Z * [new branch] gh/janeyx99/305/head -> origin/gh/janeyx99/305/head 2025-11-03T16:58:01.6407244Z * [new branch] gh/janeyx99/306/base -> origin/gh/janeyx99/306/base 2025-11-03T16:58:01.6408873Z * [new branch] gh/janeyx99/306/head -> origin/gh/janeyx99/306/head 2025-11-03T16:58:01.6411310Z * [new branch] gh/janeyx99/310/base -> origin/gh/janeyx99/310/base 2025-11-03T16:58:01.6413068Z * [new branch] gh/janeyx99/310/head -> origin/gh/janeyx99/310/head 2025-11-03T16:58:01.6414671Z * [new branch] gh/janeyx99/310/orig -> origin/gh/janeyx99/310/orig 2025-11-03T16:58:01.6416910Z * [new branch] gh/janeyx99/311/base -> origin/gh/janeyx99/311/base 2025-11-03T16:58:01.6418599Z * [new branch] gh/janeyx99/311/head -> origin/gh/janeyx99/311/head 2025-11-03T16:58:01.6420384Z * [new branch] gh/janeyx99/311/orig -> origin/gh/janeyx99/311/orig 2025-11-03T16:58:01.6424314Z * [new branch] gh/janeyx99/312/base -> origin/gh/janeyx99/312/base 2025-11-03T16:58:01.6425958Z * [new branch] gh/janeyx99/312/head -> origin/gh/janeyx99/312/head 2025-11-03T16:58:01.6427635Z * [new branch] gh/janeyx99/312/orig -> origin/gh/janeyx99/312/orig 2025-11-03T16:58:01.6429814Z * [new branch] gh/janeyx99/313/base -> origin/gh/janeyx99/313/base 2025-11-03T16:58:01.6431497Z * [new branch] gh/janeyx99/313/head -> origin/gh/janeyx99/313/head 2025-11-03T16:58:01.6433164Z * [new branch] gh/janeyx99/313/orig -> origin/gh/janeyx99/313/orig 2025-11-03T16:58:01.6435563Z * [new branch] gh/janeyx99/314/base -> origin/gh/janeyx99/314/base 2025-11-03T16:58:01.6437257Z * [new branch] gh/janeyx99/314/head -> origin/gh/janeyx99/314/head 2025-11-03T16:58:01.6438932Z * [new branch] gh/janeyx99/314/orig -> origin/gh/janeyx99/314/orig 2025-11-03T16:58:01.6441242Z * [new branch] gh/janeyx99/315/base -> origin/gh/janeyx99/315/base 2025-11-03T16:58:01.6442952Z * [new branch] gh/janeyx99/315/head -> origin/gh/janeyx99/315/head 2025-11-03T16:58:01.6444622Z * [new branch] gh/janeyx99/315/orig -> origin/gh/janeyx99/315/orig 2025-11-03T16:58:01.6446931Z * [new branch] gh/janeyx99/316/base -> origin/gh/janeyx99/316/base 2025-11-03T16:58:01.6448607Z * [new branch] gh/janeyx99/316/head -> origin/gh/janeyx99/316/head 2025-11-03T16:58:01.6450296Z * [new branch] gh/janeyx99/316/orig -> origin/gh/janeyx99/316/orig 2025-11-03T16:58:01.6452582Z * [new branch] gh/janeyx99/317/base -> origin/gh/janeyx99/317/base 2025-11-03T16:58:01.6454339Z * [new branch] gh/janeyx99/317/head -> origin/gh/janeyx99/317/head 2025-11-03T16:58:01.6456003Z * [new branch] gh/janeyx99/317/orig -> origin/gh/janeyx99/317/orig 2025-11-03T16:58:01.6458912Z * [new branch] gh/janeyx99/318/base -> origin/gh/janeyx99/318/base 2025-11-03T16:58:01.6460612Z * [new branch] gh/janeyx99/318/head -> origin/gh/janeyx99/318/head 2025-11-03T16:58:01.6462608Z * [new branch] gh/janeyx99/318/orig -> origin/gh/janeyx99/318/orig 2025-11-03T16:58:01.6465650Z * [new branch] gh/janeyx99/319/base -> origin/gh/janeyx99/319/base 2025-11-03T16:58:01.6467478Z * [new branch] gh/janeyx99/319/head -> origin/gh/janeyx99/319/head 2025-11-03T16:58:01.6468887Z * [new branch] gh/janeyx99/319/orig -> origin/gh/janeyx99/319/orig 2025-11-03T16:58:01.6471379Z * [new branch] gh/janeyx99/320/base -> origin/gh/janeyx99/320/base 2025-11-03T16:58:01.6472989Z * [new branch] gh/janeyx99/320/head -> origin/gh/janeyx99/320/head 2025-11-03T16:58:01.6474642Z * [new branch] gh/janeyx99/320/orig -> origin/gh/janeyx99/320/orig 2025-11-03T16:58:01.6477115Z * [new branch] gh/janeyx99/321/base -> origin/gh/janeyx99/321/base 2025-11-03T16:58:01.6478844Z * [new branch] gh/janeyx99/321/head -> origin/gh/janeyx99/321/head 2025-11-03T16:58:01.6480577Z * [new branch] gh/janeyx99/321/orig -> origin/gh/janeyx99/321/orig 2025-11-03T16:58:01.6483193Z * [new branch] gh/janeyx99/322/base -> origin/gh/janeyx99/322/base 2025-11-03T16:58:01.6484778Z * [new branch] gh/janeyx99/322/head -> origin/gh/janeyx99/322/head 2025-11-03T16:58:01.6486409Z * [new branch] gh/janeyx99/322/orig -> origin/gh/janeyx99/322/orig 2025-11-03T16:58:01.6488848Z * [new branch] gh/janeyx99/323/base -> origin/gh/janeyx99/323/base 2025-11-03T16:58:01.6490957Z * [new branch] gh/janeyx99/323/head -> origin/gh/janeyx99/323/head 2025-11-03T16:58:01.6493485Z * [new branch] gh/janeyx99/323/orig -> origin/gh/janeyx99/323/orig 2025-11-03T16:58:01.6497303Z * [new branch] gh/janeyx99/324/base -> origin/gh/janeyx99/324/base 2025-11-03T16:58:01.6499685Z * [new branch] gh/janeyx99/324/head -> origin/gh/janeyx99/324/head 2025-11-03T16:58:01.6502141Z * [new branch] gh/janeyx99/324/orig -> origin/gh/janeyx99/324/orig 2025-11-03T16:58:01.6505603Z * [new branch] gh/janeyx99/325/base -> origin/gh/janeyx99/325/base 2025-11-03T16:58:01.6508307Z * [new branch] gh/janeyx99/325/head -> origin/gh/janeyx99/325/head 2025-11-03T16:58:01.6509676Z * [new branch] gh/janeyx99/325/orig -> origin/gh/janeyx99/325/orig 2025-11-03T16:58:01.6512073Z * [new branch] gh/janeyx99/326/base -> origin/gh/janeyx99/326/base 2025-11-03T16:58:01.6513704Z * [new branch] gh/janeyx99/326/head -> origin/gh/janeyx99/326/head 2025-11-03T16:58:01.6515447Z * [new branch] gh/janeyx99/326/orig -> origin/gh/janeyx99/326/orig 2025-11-03T16:58:01.6518024Z * [new branch] gh/janeyx99/88/base -> origin/gh/janeyx99/88/base 2025-11-03T16:58:01.6519741Z * [new branch] gh/janeyx99/88/head -> origin/gh/janeyx99/88/head 2025-11-03T16:58:01.6521542Z * [new branch] gh/janeyx99/88/orig -> origin/gh/janeyx99/88/orig 2025-11-03T16:58:01.6524697Z * [new branch] gh/jansel/360/base -> origin/gh/jansel/360/base 2025-11-03T16:58:01.6526420Z * [new branch] gh/jansel/360/head -> origin/gh/jansel/360/head 2025-11-03T16:58:01.6528717Z * [new branch] gh/jansel/451/base -> origin/gh/jansel/451/base 2025-11-03T16:58:01.6530888Z * [new branch] gh/jansel/451/head -> origin/gh/jansel/451/head 2025-11-03T16:58:01.6532735Z * [new branch] gh/jansel/451/orig -> origin/gh/jansel/451/orig 2025-11-03T16:58:01.6535122Z * [new branch] gh/jansel/462/base -> origin/gh/jansel/462/base 2025-11-03T16:58:01.6536805Z * [new branch] gh/jansel/462/head -> origin/gh/jansel/462/head 2025-11-03T16:58:01.6538490Z * [new branch] gh/jansel/462/orig -> origin/gh/jansel/462/orig 2025-11-03T16:58:01.6540787Z * [new branch] gh/jansel/531/base -> origin/gh/jansel/531/base 2025-11-03T16:58:01.6542650Z * [new branch] gh/jansel/531/head -> origin/gh/jansel/531/head 2025-11-03T16:58:01.6544328Z * [new branch] gh/jansel/531/orig -> origin/gh/jansel/531/orig 2025-11-03T16:58:01.6546640Z * [new branch] gh/jansel/533/base -> origin/gh/jansel/533/base 2025-11-03T16:58:01.6548365Z * [new branch] gh/jansel/533/head -> origin/gh/jansel/533/head 2025-11-03T16:58:01.6550010Z * [new branch] gh/jansel/533/orig -> origin/gh/jansel/533/orig 2025-11-03T16:58:01.6552219Z * [new branch] gh/jansel/549/base -> origin/gh/jansel/549/base 2025-11-03T16:58:01.6553920Z * [new branch] gh/jansel/549/head -> origin/gh/jansel/549/head 2025-11-03T16:58:01.6555726Z * [new branch] gh/jansel/549/orig -> origin/gh/jansel/549/orig 2025-11-03T16:58:01.6558210Z * [new branch] gh/jansel/550/base -> origin/gh/jansel/550/base 2025-11-03T16:58:01.6559736Z * [new branch] gh/jansel/550/head -> origin/gh/jansel/550/head 2025-11-03T16:58:01.6561348Z * [new branch] gh/jansel/550/orig -> origin/gh/jansel/550/orig 2025-11-03T16:58:01.6563709Z * [new branch] gh/jansel/551/base -> origin/gh/jansel/551/base 2025-11-03T16:58:01.6565404Z * [new branch] gh/jansel/551/head -> origin/gh/jansel/551/head 2025-11-03T16:58:01.6567138Z * [new branch] gh/jansel/551/orig -> origin/gh/jansel/551/orig 2025-11-03T16:58:01.6570011Z * [new branch] gh/jbschlosser/247/base -> origin/gh/jbschlosser/247/base 2025-11-03T16:58:01.6571761Z * [new branch] gh/jbschlosser/247/head -> origin/gh/jbschlosser/247/head 2025-11-03T16:58:01.6573420Z * [new branch] gh/jbschlosser/247/orig -> origin/gh/jbschlosser/247/orig 2025-11-03T16:58:01.6575867Z * [new branch] gh/jbschlosser/250/base -> origin/gh/jbschlosser/250/base 2025-11-03T16:58:01.6577795Z * [new branch] gh/jbschlosser/250/head -> origin/gh/jbschlosser/250/head 2025-11-03T16:58:01.6579596Z * [new branch] gh/jbschlosser/250/orig -> origin/gh/jbschlosser/250/orig 2025-11-03T16:58:01.6582399Z * [new branch] gh/jgong5/1/base -> origin/gh/jgong5/1/base 2025-11-03T16:58:01.6584082Z * [new branch] gh/jgong5/1/head -> origin/gh/jgong5/1/head 2025-11-03T16:58:01.6585793Z * [new branch] gh/jgong5/1/orig -> origin/gh/jgong5/1/orig 2025-11-03T16:58:01.6588606Z * [new branch] gh/jiayisunx/59/base -> origin/gh/jiayisunx/59/base 2025-11-03T16:58:01.6590285Z * [new branch] gh/jiayisunx/59/head -> origin/gh/jiayisunx/59/head 2025-11-03T16:58:01.6592058Z * [new branch] gh/jiayisunx/59/orig -> origin/gh/jiayisunx/59/orig 2025-11-03T16:58:01.6594301Z * [new branch] gh/jiayisunx/61/base -> origin/gh/jiayisunx/61/base 2025-11-03T16:58:01.6595985Z * [new branch] gh/jiayisunx/61/head -> origin/gh/jiayisunx/61/head 2025-11-03T16:58:01.6597697Z * [new branch] gh/jiayisunx/61/orig -> origin/gh/jiayisunx/61/orig 2025-11-03T16:58:01.6600540Z * [new branch] gh/jiayisunx/68/base -> origin/gh/jiayisunx/68/base 2025-11-03T16:58:01.6602162Z * [new branch] gh/jiayisunx/68/head -> origin/gh/jiayisunx/68/head 2025-11-03T16:58:01.6603874Z * [new branch] gh/jiayisunx/68/orig -> origin/gh/jiayisunx/68/orig 2025-11-03T16:58:01.6606134Z * [new branch] gh/jiayisunx/71/base -> origin/gh/jiayisunx/71/base 2025-11-03T16:58:01.6607812Z * [new branch] gh/jiayisunx/71/head -> origin/gh/jiayisunx/71/head 2025-11-03T16:58:01.6609494Z * [new branch] gh/jiayisunx/71/orig -> origin/gh/jiayisunx/71/orig 2025-11-03T16:58:01.6611867Z * [new branch] gh/jiayisunx/77/base -> origin/gh/jiayisunx/77/base 2025-11-03T16:58:01.6613561Z * [new branch] gh/jiayisunx/77/head -> origin/gh/jiayisunx/77/head 2025-11-03T16:58:01.6615239Z * [new branch] gh/jiayisunx/77/orig -> origin/gh/jiayisunx/77/orig 2025-11-03T16:58:01.6617521Z * [new branch] gh/jiayisunx/78/base -> origin/gh/jiayisunx/78/base 2025-11-03T16:58:01.6619220Z * [new branch] gh/jiayisunx/78/head -> origin/gh/jiayisunx/78/head 2025-11-03T16:58:01.6620871Z * [new branch] gh/jiayisunx/78/orig -> origin/gh/jiayisunx/78/orig 2025-11-03T16:58:01.6623619Z * [new branch] gh/jiayisunx/79/base -> origin/gh/jiayisunx/79/base 2025-11-03T16:58:01.6625283Z * [new branch] gh/jiayisunx/79/head -> origin/gh/jiayisunx/79/head 2025-11-03T16:58:01.6627060Z * [new branch] gh/jiayisunx/79/orig -> origin/gh/jiayisunx/79/orig 2025-11-03T16:58:01.6629336Z * [new branch] gh/jiayisunx/80/base -> origin/gh/jiayisunx/80/base 2025-11-03T16:58:01.6630988Z * [new branch] gh/jiayisunx/80/head -> origin/gh/jiayisunx/80/head 2025-11-03T16:58:01.6633091Z * [new branch] gh/jiayisunx/80/orig -> origin/gh/jiayisunx/80/orig 2025-11-03T16:58:01.6635425Z * [new branch] gh/jiayisunx/81/base -> origin/gh/jiayisunx/81/base 2025-11-03T16:58:01.6637118Z * [new branch] gh/jiayisunx/81/head -> origin/gh/jiayisunx/81/head 2025-11-03T16:58:01.6638832Z * [new branch] gh/jiayisunx/81/orig -> origin/gh/jiayisunx/81/orig 2025-11-03T16:58:01.6641203Z * [new branch] gh/jiayisunx/82/base -> origin/gh/jiayisunx/82/base 2025-11-03T16:58:01.6642938Z * [new branch] gh/jiayisunx/82/head -> origin/gh/jiayisunx/82/head 2025-11-03T16:58:01.6644678Z * [new branch] gh/jiayisunx/82/orig -> origin/gh/jiayisunx/82/orig 2025-11-03T16:58:01.6647073Z * [new branch] gh/jiayisunx/83/base -> origin/gh/jiayisunx/83/base 2025-11-03T16:58:01.6648723Z * [new branch] gh/jiayisunx/83/head -> origin/gh/jiayisunx/83/head 2025-11-03T16:58:01.6650412Z * [new branch] gh/jiayisunx/83/orig -> origin/gh/jiayisunx/83/orig 2025-11-03T16:58:01.6653084Z * [new branch] gh/jiayisunx/84/base -> origin/gh/jiayisunx/84/base 2025-11-03T16:58:01.6654822Z * [new branch] gh/jiayisunx/84/head -> origin/gh/jiayisunx/84/head 2025-11-03T16:58:01.6656506Z * [new branch] gh/jiayisunx/84/orig -> origin/gh/jiayisunx/84/orig 2025-11-03T16:58:01.6659200Z * [new branch] gh/jiayisunx/85/base -> origin/gh/jiayisunx/85/base 2025-11-03T16:58:01.6660905Z * [new branch] gh/jiayisunx/85/head -> origin/gh/jiayisunx/85/head 2025-11-03T16:58:01.6663292Z * [new branch] gh/jiayisunx/85/orig -> origin/gh/jiayisunx/85/orig 2025-11-03T16:58:01.6665596Z * [new branch] gh/jiayisunx/86/base -> origin/gh/jiayisunx/86/base 2025-11-03T16:58:01.6667244Z * [new branch] gh/jiayisunx/86/head -> origin/gh/jiayisunx/86/head 2025-11-03T16:58:01.6668863Z * [new branch] gh/jiayisunx/86/orig -> origin/gh/jiayisunx/86/orig 2025-11-03T16:58:01.6671255Z * [new branch] gh/jiayisunx/87/base -> origin/gh/jiayisunx/87/base 2025-11-03T16:58:01.6672969Z * [new branch] gh/jiayisunx/87/head -> origin/gh/jiayisunx/87/head 2025-11-03T16:58:01.6674642Z * [new branch] gh/jiayisunx/87/orig -> origin/gh/jiayisunx/87/orig 2025-11-03T16:58:01.6677311Z * [new branch] gh/jjwu@meta.com/1/base -> origin/gh/jjwu@meta.com/1/base 2025-11-03T16:58:01.6678948Z * [new branch] gh/jjwu@meta.com/1/head -> origin/gh/jjwu@meta.com/1/head 2025-11-03T16:58:01.6682686Z * [new branch] gh/karthickai/6/base -> origin/gh/karthickai/6/base 2025-11-03T16:58:01.6684803Z * [new branch] gh/karthickai/6/head -> origin/gh/karthickai/6/head 2025-11-03T16:58:01.6686608Z * [new branch] gh/karthickai/6/orig -> origin/gh/karthickai/6/orig 2025-11-03T16:58:01.6688883Z * [new branch] gh/karthickai/7/base -> origin/gh/karthickai/7/base 2025-11-03T16:58:01.6690713Z * [new branch] gh/karthickai/7/head -> origin/gh/karthickai/7/head 2025-11-03T16:58:01.6692406Z * [new branch] gh/karthickai/7/orig -> origin/gh/karthickai/7/orig 2025-11-03T16:58:01.6695302Z * [new branch] gh/karthickai/8/base -> origin/gh/karthickai/8/base 2025-11-03T16:58:01.6697033Z * [new branch] gh/karthickai/8/head -> origin/gh/karthickai/8/head 2025-11-03T16:58:01.6698888Z * [new branch] gh/karthickai/8/orig -> origin/gh/karthickai/8/orig 2025-11-03T16:58:01.6701334Z * [new branch] gh/karthickai/9/base -> origin/gh/karthickai/9/base 2025-11-03T16:58:01.6703114Z * [new branch] gh/karthickai/9/head -> origin/gh/karthickai/9/head 2025-11-03T16:58:01.6704816Z * [new branch] gh/karthickai/9/orig -> origin/gh/karthickai/9/orig 2025-11-03T16:58:01.6707612Z * [new branch] gh/kurtamohler/32/base -> origin/gh/kurtamohler/32/base 2025-11-03T16:58:01.6709308Z * [new branch] gh/kurtamohler/32/head -> origin/gh/kurtamohler/32/head 2025-11-03T16:58:01.6710970Z * [new branch] gh/kurtamohler/32/orig -> origin/gh/kurtamohler/32/orig 2025-11-03T16:58:01.6713322Z * [new branch] gh/kurtamohler/33/base -> origin/gh/kurtamohler/33/base 2025-11-03T16:58:01.6715013Z * [new branch] gh/kurtamohler/33/head -> origin/gh/kurtamohler/33/head 2025-11-03T16:58:01.6716760Z * [new branch] gh/kurtamohler/33/orig -> origin/gh/kurtamohler/33/orig 2025-11-03T16:58:01.6719091Z * [new branch] gh/kurtamohler/34/base -> origin/gh/kurtamohler/34/base 2025-11-03T16:58:01.6720739Z * [new branch] gh/kurtamohler/34/head -> origin/gh/kurtamohler/34/head 2025-11-03T16:58:01.6722819Z * [new branch] gh/kurtamohler/34/orig -> origin/gh/kurtamohler/34/orig 2025-11-03T16:58:01.6725059Z * [new branch] gh/kurtamohler/55/base -> origin/gh/kurtamohler/55/base 2025-11-03T16:58:01.6726771Z * [new branch] gh/kurtamohler/55/head -> origin/gh/kurtamohler/55/head 2025-11-03T16:58:01.6728445Z * [new branch] gh/kurtamohler/55/orig -> origin/gh/kurtamohler/55/orig 2025-11-03T16:58:01.6730781Z * [new branch] gh/kurtamohler/56/base -> origin/gh/kurtamohler/56/base 2025-11-03T16:58:01.6732442Z * [new branch] gh/kurtamohler/56/head -> origin/gh/kurtamohler/56/head 2025-11-03T16:58:01.6734124Z * [new branch] gh/kurtamohler/56/orig -> origin/gh/kurtamohler/56/orig 2025-11-03T16:58:01.6736411Z * [new branch] gh/kurtamohler/57/base -> origin/gh/kurtamohler/57/base 2025-11-03T16:58:01.6738124Z * [new branch] gh/kurtamohler/57/head -> origin/gh/kurtamohler/57/head 2025-11-03T16:58:01.6739899Z * [new branch] gh/kurtamohler/57/orig -> origin/gh/kurtamohler/57/orig 2025-11-03T16:58:01.6742398Z * [new branch] gh/kurtamohler/58/base -> origin/gh/kurtamohler/58/base 2025-11-03T16:58:01.6744112Z * [new branch] gh/kurtamohler/58/head -> origin/gh/kurtamohler/58/head 2025-11-03T16:58:01.6745788Z * [new branch] gh/kurtamohler/58/orig -> origin/gh/kurtamohler/58/orig 2025-11-03T16:58:01.6748056Z * [new branch] gh/kurtamohler/59/base -> origin/gh/kurtamohler/59/base 2025-11-03T16:58:01.6749704Z * [new branch] gh/kurtamohler/59/head -> origin/gh/kurtamohler/59/head 2025-11-03T16:58:01.6751484Z * [new branch] gh/kurtamohler/59/orig -> origin/gh/kurtamohler/59/orig 2025-11-03T16:58:01.6754552Z * [new branch] gh/kwen2501/130/base -> origin/gh/kwen2501/130/base 2025-11-03T16:58:01.6756380Z * [new branch] gh/kwen2501/130/head -> origin/gh/kwen2501/130/head 2025-11-03T16:58:01.6758576Z * [new branch] gh/kwen2501/130/orig -> origin/gh/kwen2501/130/orig 2025-11-03T16:58:01.6760986Z * [new branch] gh/kwen2501/170/base -> origin/gh/kwen2501/170/base 2025-11-03T16:58:01.6762808Z * [new branch] gh/kwen2501/170/head -> origin/gh/kwen2501/170/head 2025-11-03T16:58:01.6765196Z * [new branch] gh/kwen2501/187/base -> origin/gh/kwen2501/187/base 2025-11-03T16:58:01.6766972Z * [new branch] gh/kwen2501/187/head -> origin/gh/kwen2501/187/head 2025-11-03T16:58:01.6768910Z * [new branch] gh/kwen2501/187/orig -> origin/gh/kwen2501/187/orig 2025-11-03T16:58:01.6770987Z * [new branch] gh/kwen2501/188/base -> origin/gh/kwen2501/188/base 2025-11-03T16:58:01.6772557Z * [new branch] gh/kwen2501/188/head -> origin/gh/kwen2501/188/head 2025-11-03T16:58:01.6774216Z * [new branch] gh/kwen2501/188/orig -> origin/gh/kwen2501/188/orig 2025-11-03T16:58:01.6776524Z * [new branch] gh/kwen2501/211/base -> origin/gh/kwen2501/211/base 2025-11-03T16:58:01.6778282Z * [new branch] gh/kwen2501/211/head -> origin/gh/kwen2501/211/head 2025-11-03T16:58:01.6780636Z * [new branch] gh/kwen2501/224/base -> origin/gh/kwen2501/224/base 2025-11-03T16:58:01.6782626Z * [new branch] gh/kwen2501/224/head -> origin/gh/kwen2501/224/head 2025-11-03T16:58:01.6784322Z * [new branch] gh/kwen2501/224/orig -> origin/gh/kwen2501/224/orig 2025-11-03T16:58:01.6786670Z * [new branch] gh/kwen2501/228/base -> origin/gh/kwen2501/228/base 2025-11-03T16:58:01.6788358Z * [new branch] gh/kwen2501/228/head -> origin/gh/kwen2501/228/head 2025-11-03T16:58:01.6790014Z * [new branch] gh/kwen2501/228/orig -> origin/gh/kwen2501/228/orig 2025-11-03T16:58:01.6792626Z * [new branch] gh/kwen2501/231/base -> origin/gh/kwen2501/231/base 2025-11-03T16:58:01.6794339Z * [new branch] gh/kwen2501/231/head -> origin/gh/kwen2501/231/head 2025-11-03T16:58:01.6796043Z * [new branch] gh/kwen2501/231/orig -> origin/gh/kwen2501/231/orig 2025-11-03T16:58:01.6798393Z * [new branch] gh/kwen2501/234/base -> origin/gh/kwen2501/234/base 2025-11-03T16:58:01.6800188Z * [new branch] gh/kwen2501/234/head -> origin/gh/kwen2501/234/head 2025-11-03T16:58:01.6801881Z * [new branch] gh/kwen2501/234/orig -> origin/gh/kwen2501/234/orig 2025-11-03T16:58:01.6804292Z * [new branch] gh/kwen2501/235/base -> origin/gh/kwen2501/235/base 2025-11-03T16:58:01.6805853Z * [new branch] gh/kwen2501/235/head -> origin/gh/kwen2501/235/head 2025-11-03T16:58:01.6807656Z * [new branch] gh/kwen2501/235/orig -> origin/gh/kwen2501/235/orig 2025-11-03T16:58:01.6823028Z * [new branch] gh/kwen2501/236/base -> origin/gh/kwen2501/236/base 2025-11-03T16:58:01.6823492Z * [new branch] gh/kwen2501/236/head -> origin/gh/kwen2501/236/head 2025-11-03T16:58:01.6823728Z * [new branch] gh/kwen2501/236/orig -> origin/gh/kwen2501/236/orig 2025-11-03T16:58:01.6823925Z * [new branch] gh/kwen2501/237/base -> origin/gh/kwen2501/237/base 2025-11-03T16:58:01.6824137Z * [new branch] gh/kwen2501/237/head -> origin/gh/kwen2501/237/head 2025-11-03T16:58:01.6824481Z * [new branch] gh/kwen2501/237/orig -> origin/gh/kwen2501/237/orig 2025-11-03T16:58:01.6824802Z * [new branch] gh/kwen2501/238/base -> origin/gh/kwen2501/238/base 2025-11-03T16:58:01.6824997Z * [new branch] gh/kwen2501/238/head -> origin/gh/kwen2501/238/head 2025-11-03T16:58:01.6826831Z * [new branch] gh/kwen2501/238/orig -> origin/gh/kwen2501/238/orig 2025-11-03T16:58:01.6829191Z * [new branch] gh/kwen2501/240/base -> origin/gh/kwen2501/240/base 2025-11-03T16:58:01.6830749Z * [new branch] gh/kwen2501/240/head -> origin/gh/kwen2501/240/head 2025-11-03T16:58:01.6832324Z * [new branch] gh/kwen2501/240/orig -> origin/gh/kwen2501/240/orig 2025-11-03T16:58:01.6835427Z * [new branch] gh/kwen2501/241/base -> origin/gh/kwen2501/241/base 2025-11-03T16:58:01.6837092Z * [new branch] gh/kwen2501/241/head -> origin/gh/kwen2501/241/head 2025-11-03T16:58:01.6839024Z * [new branch] gh/kwen2501/241/orig -> origin/gh/kwen2501/241/orig 2025-11-03T16:58:01.6841192Z * [new branch] gh/kwen2501/247/base -> origin/gh/kwen2501/247/base 2025-11-03T16:58:01.6843078Z * [new branch] gh/kwen2501/247/head -> origin/gh/kwen2501/247/head 2025-11-03T16:58:01.6844732Z * [new branch] gh/kwen2501/247/orig -> origin/gh/kwen2501/247/orig 2025-11-03T16:58:01.6847600Z * [new branch] gh/kwen2501/252/base -> origin/gh/kwen2501/252/base 2025-11-03T16:58:01.6849224Z * [new branch] gh/kwen2501/252/head -> origin/gh/kwen2501/252/head 2025-11-03T16:58:01.6850890Z * [new branch] gh/kwen2501/252/orig -> origin/gh/kwen2501/252/orig 2025-11-03T16:58:01.6853340Z * [new branch] gh/kwen2501/259/base -> origin/gh/kwen2501/259/base 2025-11-03T16:58:01.6855043Z * [new branch] gh/kwen2501/259/head -> origin/gh/kwen2501/259/head 2025-11-03T16:58:01.6856825Z * [new branch] gh/kwen2501/259/orig -> origin/gh/kwen2501/259/orig 2025-11-03T16:58:01.6859385Z * [new branch] gh/kwen2501/260/base -> origin/gh/kwen2501/260/base 2025-11-03T16:58:01.6861040Z * [new branch] gh/kwen2501/260/head -> origin/gh/kwen2501/260/head 2025-11-03T16:58:01.6862846Z * [new branch] gh/kwen2501/260/orig -> origin/gh/kwen2501/260/orig 2025-11-03T16:58:01.6865185Z * [new branch] gh/kwen2501/268/base -> origin/gh/kwen2501/268/base 2025-11-03T16:58:01.6866903Z * [new branch] gh/kwen2501/268/head -> origin/gh/kwen2501/268/head 2025-11-03T16:58:01.6868603Z * [new branch] gh/kwen2501/268/orig -> origin/gh/kwen2501/268/orig 2025-11-03T16:58:01.6870975Z * [new branch] gh/kwen2501/269/base -> origin/gh/kwen2501/269/base 2025-11-03T16:58:01.6872778Z * [new branch] gh/kwen2501/269/head -> origin/gh/kwen2501/269/head 2025-11-03T16:58:01.6874417Z * [new branch] gh/kwen2501/269/orig -> origin/gh/kwen2501/269/orig 2025-11-03T16:58:01.6876771Z * [new branch] gh/kwen2501/270/base -> origin/gh/kwen2501/270/base 2025-11-03T16:58:01.6878497Z * [new branch] gh/kwen2501/270/head -> origin/gh/kwen2501/270/head 2025-11-03T16:58:01.6880158Z * [new branch] gh/kwen2501/270/orig -> origin/gh/kwen2501/270/orig 2025-11-03T16:58:01.6882660Z * [new branch] gh/kwen2501/271/base -> origin/gh/kwen2501/271/base 2025-11-03T16:58:01.6884216Z * [new branch] gh/kwen2501/271/head -> origin/gh/kwen2501/271/head 2025-11-03T16:58:01.6885846Z * [new branch] gh/kwen2501/271/orig -> origin/gh/kwen2501/271/orig 2025-11-03T16:58:01.6888238Z * [new branch] gh/kwen2501/272/base -> origin/gh/kwen2501/272/base 2025-11-03T16:58:01.6889946Z * [new branch] gh/kwen2501/272/head -> origin/gh/kwen2501/272/head 2025-11-03T16:58:01.6891615Z * [new branch] gh/kwen2501/272/orig -> origin/gh/kwen2501/272/orig 2025-11-03T16:58:01.6893984Z * [new branch] gh/kwen2501/273/base -> origin/gh/kwen2501/273/base 2025-11-03T16:58:01.6895644Z * [new branch] gh/kwen2501/273/head -> origin/gh/kwen2501/273/head 2025-11-03T16:58:01.6897361Z * [new branch] gh/kwen2501/273/orig -> origin/gh/kwen2501/273/orig 2025-11-03T16:58:01.6899732Z * [new branch] gh/kwen2501/274/base -> origin/gh/kwen2501/274/base 2025-11-03T16:58:01.6901428Z * [new branch] gh/kwen2501/274/head -> origin/gh/kwen2501/274/head 2025-11-03T16:58:01.6903225Z * [new branch] gh/kwen2501/274/orig -> origin/gh/kwen2501/274/orig 2025-11-03T16:58:01.6906070Z * [new branch] gh/kwen2501/275/base -> origin/gh/kwen2501/275/base 2025-11-03T16:58:01.6907855Z * [new branch] gh/kwen2501/275/head -> origin/gh/kwen2501/275/head 2025-11-03T16:58:01.6909520Z * [new branch] gh/kwen2501/275/orig -> origin/gh/kwen2501/275/orig 2025-11-03T16:58:01.6911805Z * [new branch] gh/kwen2501/276/base -> origin/gh/kwen2501/276/base 2025-11-03T16:58:01.6913488Z * [new branch] gh/kwen2501/276/head -> origin/gh/kwen2501/276/head 2025-11-03T16:58:01.6915141Z * [new branch] gh/kwen2501/276/orig -> origin/gh/kwen2501/276/orig 2025-11-03T16:58:01.6917518Z * [new branch] gh/kwen2501/277/base -> origin/gh/kwen2501/277/base 2025-11-03T16:58:01.6919220Z * [new branch] gh/kwen2501/277/head -> origin/gh/kwen2501/277/head 2025-11-03T16:58:01.6920930Z * [new branch] gh/kwen2501/277/orig -> origin/gh/kwen2501/277/orig 2025-11-03T16:58:01.6923766Z * [new branch] gh/kwen2501/278/base -> origin/gh/kwen2501/278/base 2025-11-03T16:58:01.6925509Z * [new branch] gh/kwen2501/278/head -> origin/gh/kwen2501/278/head 2025-11-03T16:58:01.6927317Z * [new branch] gh/kwen2501/278/orig -> origin/gh/kwen2501/278/orig 2025-11-03T16:58:01.6929681Z * [new branch] gh/kwen2501/279/base -> origin/gh/kwen2501/279/base 2025-11-03T16:58:01.6931367Z * [new branch] gh/kwen2501/279/head -> origin/gh/kwen2501/279/head 2025-11-03T16:58:01.6933131Z * [new branch] gh/kwen2501/279/orig -> origin/gh/kwen2501/279/orig 2025-11-03T16:58:01.6935861Z * [new branch] gh/laithsakka/251/base -> origin/gh/laithsakka/251/base 2025-11-03T16:58:01.6937525Z * [new branch] gh/laithsakka/251/head -> origin/gh/laithsakka/251/head 2025-11-03T16:58:01.6939200Z * [new branch] gh/laithsakka/251/orig -> origin/gh/laithsakka/251/orig 2025-11-03T16:58:01.6941567Z * [new branch] gh/laithsakka/276/base -> origin/gh/laithsakka/276/base 2025-11-03T16:58:01.6943321Z * [new branch] gh/laithsakka/276/head -> origin/gh/laithsakka/276/head 2025-11-03T16:58:01.6944977Z * [new branch] gh/laithsakka/276/orig -> origin/gh/laithsakka/276/orig 2025-11-03T16:58:01.6947899Z * [new branch] gh/laithsakka/28/base -> origin/gh/laithsakka/28/base 2025-11-03T16:58:01.6950283Z * [new branch] gh/laithsakka/285/base -> origin/gh/laithsakka/285/base 2025-11-03T16:58:01.6952057Z * [new branch] gh/laithsakka/285/head -> origin/gh/laithsakka/285/head 2025-11-03T16:58:01.6953777Z * [new branch] gh/laithsakka/285/orig -> origin/gh/laithsakka/285/orig 2025-11-03T16:58:01.6956083Z * [new branch] gh/laithsakka/287/base -> origin/gh/laithsakka/287/base 2025-11-03T16:58:01.6957788Z * [new branch] gh/laithsakka/287/head -> origin/gh/laithsakka/287/head 2025-11-03T16:58:01.6959463Z * [new branch] gh/laithsakka/287/orig -> origin/gh/laithsakka/287/orig 2025-11-03T16:58:01.6961808Z * [new branch] gh/laithsakka/289/base -> origin/gh/laithsakka/289/base 2025-11-03T16:58:01.6963684Z * [new branch] gh/laithsakka/289/head -> origin/gh/laithsakka/289/head 2025-11-03T16:58:01.6965471Z * [new branch] gh/laithsakka/289/orig -> origin/gh/laithsakka/289/orig 2025-11-03T16:58:01.6967593Z * [new branch] gh/laithsakka/29/base -> origin/gh/laithsakka/29/base 2025-11-03T16:58:01.6970061Z * [new branch] gh/laithsakka/292/base -> origin/gh/laithsakka/292/base 2025-11-03T16:58:01.6971846Z * [new branch] gh/laithsakka/292/head -> origin/gh/laithsakka/292/head 2025-11-03T16:58:01.6973654Z * [new branch] gh/laithsakka/292/orig -> origin/gh/laithsakka/292/orig 2025-11-03T16:58:01.6975860Z * [new branch] gh/laithsakka/293/base -> origin/gh/laithsakka/293/base 2025-11-03T16:58:01.6977727Z * [new branch] gh/laithsakka/293/head -> origin/gh/laithsakka/293/head 2025-11-03T16:58:01.6979289Z * [new branch] gh/laithsakka/293/orig -> origin/gh/laithsakka/293/orig 2025-11-03T16:58:01.6981553Z * [new branch] gh/laithsakka/294/base -> origin/gh/laithsakka/294/base 2025-11-03T16:58:01.6983353Z * [new branch] gh/laithsakka/294/head -> origin/gh/laithsakka/294/head 2025-11-03T16:58:01.6984952Z * [new branch] gh/laithsakka/294/orig -> origin/gh/laithsakka/294/orig 2025-11-03T16:58:01.6987602Z * [new branch] gh/laithsakka/295/base -> origin/gh/laithsakka/295/base 2025-11-03T16:58:01.6989213Z * [new branch] gh/laithsakka/295/head -> origin/gh/laithsakka/295/head 2025-11-03T16:58:01.6990814Z * [new branch] gh/laithsakka/295/orig -> origin/gh/laithsakka/295/orig 2025-11-03T16:58:01.6993194Z * [new branch] gh/laithsakka/296/base -> origin/gh/laithsakka/296/base 2025-11-03T16:58:01.6994995Z * [new branch] gh/laithsakka/296/head -> origin/gh/laithsakka/296/head 2025-11-03T16:58:01.6996673Z * [new branch] gh/laithsakka/296/orig -> origin/gh/laithsakka/296/orig 2025-11-03T16:58:01.6999085Z * [new branch] gh/laithsakka/297/base -> origin/gh/laithsakka/297/base 2025-11-03T16:58:01.7000738Z * [new branch] gh/laithsakka/297/head -> origin/gh/laithsakka/297/head 2025-11-03T16:58:01.7002534Z * [new branch] gh/laithsakka/297/orig -> origin/gh/laithsakka/297/orig 2025-11-03T16:58:01.7004851Z * [new branch] gh/laithsakka/298/base -> origin/gh/laithsakka/298/base 2025-11-03T16:58:01.7006524Z * [new branch] gh/laithsakka/298/head -> origin/gh/laithsakka/298/head 2025-11-03T16:58:01.7008229Z * [new branch] gh/laithsakka/298/orig -> origin/gh/laithsakka/298/orig 2025-11-03T16:58:01.7010501Z * [new branch] gh/laithsakka/299/base -> origin/gh/laithsakka/299/base 2025-11-03T16:58:01.7012184Z * [new branch] gh/laithsakka/299/head -> origin/gh/laithsakka/299/head 2025-11-03T16:58:01.7013863Z * [new branch] gh/laithsakka/299/orig -> origin/gh/laithsakka/299/orig 2025-11-03T16:58:01.7016059Z * [new branch] gh/laithsakka/30/base -> origin/gh/laithsakka/30/base 2025-11-03T16:58:01.7017788Z * [new branch] gh/laithsakka/30/head -> origin/gh/laithsakka/30/head 2025-11-03T16:58:01.7020257Z * [new branch] gh/laithsakka/300/base -> origin/gh/laithsakka/300/base 2025-11-03T16:58:01.7022191Z * [new branch] gh/laithsakka/300/head -> origin/gh/laithsakka/300/head 2025-11-03T16:58:01.7023905Z * [new branch] gh/laithsakka/300/orig -> origin/gh/laithsakka/300/orig 2025-11-03T16:58:01.7027129Z * [new branch] gh/laithsakka/301/base -> origin/gh/laithsakka/301/base 2025-11-03T16:58:01.7028827Z * [new branch] gh/laithsakka/301/head -> origin/gh/laithsakka/301/head 2025-11-03T16:58:01.7030533Z * [new branch] gh/laithsakka/301/orig -> origin/gh/laithsakka/301/orig 2025-11-03T16:58:01.7032958Z * [new branch] gh/laithsakka/302/base -> origin/gh/laithsakka/302/base 2025-11-03T16:58:01.7034643Z * [new branch] gh/laithsakka/302/head -> origin/gh/laithsakka/302/head 2025-11-03T16:58:01.7036249Z * [new branch] gh/laithsakka/302/orig -> origin/gh/laithsakka/302/orig 2025-11-03T16:58:01.7038490Z * [new branch] gh/laithsakka/303/base -> origin/gh/laithsakka/303/base 2025-11-03T16:58:01.7040194Z * [new branch] gh/laithsakka/303/head -> origin/gh/laithsakka/303/head 2025-11-03T16:58:01.7041899Z * [new branch] gh/laithsakka/303/orig -> origin/gh/laithsakka/303/orig 2025-11-03T16:58:01.7044374Z * [new branch] gh/laithsakka/304/base -> origin/gh/laithsakka/304/base 2025-11-03T16:58:01.7045948Z * [new branch] gh/laithsakka/304/head -> origin/gh/laithsakka/304/head 2025-11-03T16:58:01.7047583Z * [new branch] gh/laithsakka/304/orig -> origin/gh/laithsakka/304/orig 2025-11-03T16:58:01.7049807Z * [new branch] gh/laithsakka/305/base -> origin/gh/laithsakka/305/base 2025-11-03T16:58:01.7051533Z * [new branch] gh/laithsakka/305/head -> origin/gh/laithsakka/305/head 2025-11-03T16:58:01.7053119Z * [new branch] gh/laithsakka/305/orig -> origin/gh/laithsakka/305/orig 2025-11-03T16:58:01.7055390Z * [new branch] gh/laithsakka/306/base -> origin/gh/laithsakka/306/base 2025-11-03T16:58:01.7057052Z * [new branch] gh/laithsakka/306/head -> origin/gh/laithsakka/306/head 2025-11-03T16:58:01.7058662Z * [new branch] gh/laithsakka/306/orig -> origin/gh/laithsakka/306/orig 2025-11-03T16:58:01.7060874Z * [new branch] gh/laithsakka/307/base -> origin/gh/laithsakka/307/base 2025-11-03T16:58:01.7062844Z * [new branch] gh/laithsakka/307/head -> origin/gh/laithsakka/307/head 2025-11-03T16:58:01.7064513Z * [new branch] gh/laithsakka/307/orig -> origin/gh/laithsakka/307/orig 2025-11-03T16:58:01.7066813Z * [new branch] gh/laithsakka/308/base -> origin/gh/laithsakka/308/base 2025-11-03T16:58:01.7068499Z * [new branch] gh/laithsakka/308/head -> origin/gh/laithsakka/308/head 2025-11-03T16:58:01.7070290Z * [new branch] gh/laithsakka/308/orig -> origin/gh/laithsakka/308/orig 2025-11-03T16:58:01.7072659Z * [new branch] gh/laithsakka/309/base -> origin/gh/laithsakka/309/base 2025-11-03T16:58:01.7074471Z * [new branch] gh/laithsakka/309/head -> origin/gh/laithsakka/309/head 2025-11-03T16:58:01.7076147Z * [new branch] gh/laithsakka/309/orig -> origin/gh/laithsakka/309/orig 2025-11-03T16:58:01.7078408Z * [new branch] gh/laithsakka/31/base -> origin/gh/laithsakka/31/base 2025-11-03T16:58:01.7080056Z * [new branch] gh/laithsakka/31/head -> origin/gh/laithsakka/31/head 2025-11-03T16:58:01.7082440Z * [new branch] gh/laithsakka/310/base -> origin/gh/laithsakka/310/base 2025-11-03T16:58:01.7084092Z * [new branch] gh/laithsakka/310/head -> origin/gh/laithsakka/310/head 2025-11-03T16:58:01.7085837Z * [new branch] gh/laithsakka/310/orig -> origin/gh/laithsakka/310/orig 2025-11-03T16:58:01.7088409Z * [new branch] gh/laithsakka/311/base -> origin/gh/laithsakka/311/base 2025-11-03T16:58:01.7090046Z * [new branch] gh/laithsakka/311/head -> origin/gh/laithsakka/311/head 2025-11-03T16:58:01.7091779Z * [new branch] gh/laithsakka/311/orig -> origin/gh/laithsakka/311/orig 2025-11-03T16:58:01.7094272Z * [new branch] gh/laithsakka/312/base -> origin/gh/laithsakka/312/base 2025-11-03T16:58:01.7096020Z * [new branch] gh/laithsakka/312/head -> origin/gh/laithsakka/312/head 2025-11-03T16:58:01.7097706Z * [new branch] gh/laithsakka/312/orig -> origin/gh/laithsakka/312/orig 2025-11-03T16:58:01.7100191Z * [new branch] gh/laithsakka/313/base -> origin/gh/laithsakka/313/base 2025-11-03T16:58:01.7101834Z * [new branch] gh/laithsakka/313/head -> origin/gh/laithsakka/313/head 2025-11-03T16:58:01.7103625Z * [new branch] gh/laithsakka/313/orig -> origin/gh/laithsakka/313/orig 2025-11-03T16:58:01.7105998Z * [new branch] gh/laithsakka/314/base -> origin/gh/laithsakka/314/base 2025-11-03T16:58:01.7107766Z * [new branch] gh/laithsakka/314/head -> origin/gh/laithsakka/314/head 2025-11-03T16:58:01.7109564Z * [new branch] gh/laithsakka/314/orig -> origin/gh/laithsakka/314/orig 2025-11-03T16:58:01.7111930Z * [new branch] gh/laithsakka/315/base -> origin/gh/laithsakka/315/base 2025-11-03T16:58:01.7113584Z * [new branch] gh/laithsakka/315/head -> origin/gh/laithsakka/315/head 2025-11-03T16:58:01.7115263Z * [new branch] gh/laithsakka/315/orig -> origin/gh/laithsakka/315/orig 2025-11-03T16:58:01.7117741Z * [new branch] gh/laithsakka/316/base -> origin/gh/laithsakka/316/base 2025-11-03T16:58:01.7119472Z * [new branch] gh/laithsakka/316/head -> origin/gh/laithsakka/316/head 2025-11-03T16:58:01.7121332Z * [new branch] gh/laithsakka/316/orig -> origin/gh/laithsakka/316/orig 2025-11-03T16:58:01.7124006Z * [new branch] gh/laithsakka/317/base -> origin/gh/laithsakka/317/base 2025-11-03T16:58:01.7125613Z * [new branch] gh/laithsakka/317/head -> origin/gh/laithsakka/317/head 2025-11-03T16:58:01.7127304Z * [new branch] gh/laithsakka/317/orig -> origin/gh/laithsakka/317/orig 2025-11-03T16:58:01.7129705Z * [new branch] gh/laithsakka/318/base -> origin/gh/laithsakka/318/base 2025-11-03T16:58:01.7131392Z * [new branch] gh/laithsakka/318/head -> origin/gh/laithsakka/318/head 2025-11-03T16:58:01.7133027Z * [new branch] gh/laithsakka/318/orig -> origin/gh/laithsakka/318/orig 2025-11-03T16:58:01.7135496Z * [new branch] gh/laithsakka/319/base -> origin/gh/laithsakka/319/base 2025-11-03T16:58:01.7137190Z * [new branch] gh/laithsakka/319/head -> origin/gh/laithsakka/319/head 2025-11-03T16:58:01.7138840Z * [new branch] gh/laithsakka/319/orig -> origin/gh/laithsakka/319/orig 2025-11-03T16:58:01.7141185Z * [new branch] gh/laithsakka/32/base -> origin/gh/laithsakka/32/base 2025-11-03T16:58:01.7143030Z * [new branch] gh/laithsakka/32/head -> origin/gh/laithsakka/32/head 2025-11-03T16:58:01.7145508Z * [new branch] gh/laithsakka/320/base -> origin/gh/laithsakka/320/base 2025-11-03T16:58:01.7147289Z * [new branch] gh/laithsakka/320/head -> origin/gh/laithsakka/320/head 2025-11-03T16:58:01.7149000Z * [new branch] gh/laithsakka/320/orig -> origin/gh/laithsakka/320/orig 2025-11-03T16:58:01.7151368Z * [new branch] gh/laithsakka/321/base -> origin/gh/laithsakka/321/base 2025-11-03T16:58:01.7153070Z * [new branch] gh/laithsakka/321/head -> origin/gh/laithsakka/321/head 2025-11-03T16:58:01.7154814Z * [new branch] gh/laithsakka/321/orig -> origin/gh/laithsakka/321/orig 2025-11-03T16:58:01.7157599Z * [new branch] gh/liangel-02/1/base -> origin/gh/liangel-02/1/base 2025-11-03T16:58:01.7159310Z * [new branch] gh/liangel-02/1/head -> origin/gh/liangel-02/1/head 2025-11-03T16:58:01.7161010Z * [new branch] gh/liangel-02/1/orig -> origin/gh/liangel-02/1/orig 2025-11-03T16:58:01.7163270Z * [new branch] gh/liangel-02/2/base -> origin/gh/liangel-02/2/base 2025-11-03T16:58:01.7165000Z * [new branch] gh/liangel-02/2/head -> origin/gh/liangel-02/2/head 2025-11-03T16:58:01.7166671Z * [new branch] gh/liangel-02/2/orig -> origin/gh/liangel-02/2/orig 2025-11-03T16:58:01.7169600Z * [new branch] gh/liangel/1/base -> origin/gh/liangel/1/base 2025-11-03T16:58:01.7171402Z * [new branch] gh/liangel/1/head -> origin/gh/liangel/1/head 2025-11-03T16:58:01.7173091Z * [new branch] gh/liangel/1/orig -> origin/gh/liangel/1/orig 2025-11-03T16:58:01.7175441Z * [new branch] gh/liangel/2/base -> origin/gh/liangel/2/base 2025-11-03T16:58:01.7177179Z * [new branch] gh/liangel/2/head -> origin/gh/liangel/2/head 2025-11-03T16:58:01.7178993Z * [new branch] gh/liangel/2/orig -> origin/gh/liangel/2/orig 2025-11-03T16:58:01.7181403Z * [new branch] gh/liangel/3/base -> origin/gh/liangel/3/base 2025-11-03T16:58:01.7183317Z * [new branch] gh/liangel/3/head -> origin/gh/liangel/3/head 2025-11-03T16:58:01.7185000Z * [new branch] gh/liangel/3/orig -> origin/gh/liangel/3/orig 2025-11-03T16:58:01.7187793Z * [new branch] gh/liangel/4/base -> origin/gh/liangel/4/base 2025-11-03T16:58:01.7189678Z * [new branch] gh/liangel/4/head -> origin/gh/liangel/4/head 2025-11-03T16:58:01.7191390Z * [new branch] gh/liangel/4/orig -> origin/gh/liangel/4/orig 2025-11-03T16:58:01.7195667Z * [new branch] gh/lucaskabela/1/base -> origin/gh/lucaskabela/1/base 2025-11-03T16:58:01.7197368Z * [new branch] gh/lucaskabela/1/head -> origin/gh/lucaskabela/1/head 2025-11-03T16:58:01.7199653Z * [new branch] gh/lucaskabela/10/base -> origin/gh/lucaskabela/10/base 2025-11-03T16:58:01.7201347Z * [new branch] gh/lucaskabela/10/head -> origin/gh/lucaskabela/10/head 2025-11-03T16:58:01.7203059Z * [new branch] gh/lucaskabela/10/orig -> origin/gh/lucaskabela/10/orig 2025-11-03T16:58:01.7205439Z * [new branch] gh/lucaskabela/2/base -> origin/gh/lucaskabela/2/base 2025-11-03T16:58:01.7207130Z * [new branch] gh/lucaskabela/2/head -> origin/gh/lucaskabela/2/head 2025-11-03T16:58:01.7209296Z * [new branch] gh/lucaskabela/3/base -> origin/gh/lucaskabela/3/base 2025-11-03T16:58:01.7211008Z * [new branch] gh/lucaskabela/3/head -> origin/gh/lucaskabela/3/head 2025-11-03T16:58:01.7212669Z * [new branch] gh/lucaskabela/3/orig -> origin/gh/lucaskabela/3/orig 2025-11-03T16:58:01.7214808Z * [new branch] gh/lucaskabela/4/base -> origin/gh/lucaskabela/4/base 2025-11-03T16:58:01.7216497Z * [new branch] gh/lucaskabela/4/head -> origin/gh/lucaskabela/4/head 2025-11-03T16:58:01.7218242Z * [new branch] gh/lucaskabela/4/orig -> origin/gh/lucaskabela/4/orig 2025-11-03T16:58:01.7220442Z * [new branch] gh/lucaskabela/5/base -> origin/gh/lucaskabela/5/base 2025-11-03T16:58:01.7223379Z * [new branch] gh/lucaskabela/5/head -> origin/gh/lucaskabela/5/head 2025-11-03T16:58:01.7225035Z * [new branch] gh/lucaskabela/5/orig -> origin/gh/lucaskabela/5/orig 2025-11-03T16:58:01.7227262Z * [new branch] gh/lucaskabela/6/base -> origin/gh/lucaskabela/6/base 2025-11-03T16:58:01.7228907Z * [new branch] gh/lucaskabela/6/head -> origin/gh/lucaskabela/6/head 2025-11-03T16:58:01.7230562Z * [new branch] gh/lucaskabela/6/orig -> origin/gh/lucaskabela/6/orig 2025-11-03T16:58:01.7232741Z * [new branch] gh/lucaskabela/7/base -> origin/gh/lucaskabela/7/base 2025-11-03T16:58:01.7234450Z * [new branch] gh/lucaskabela/7/head -> origin/gh/lucaskabela/7/head 2025-11-03T16:58:01.7236122Z * [new branch] gh/lucaskabela/7/orig -> origin/gh/lucaskabela/7/orig 2025-11-03T16:58:01.7238250Z * [new branch] gh/lucaskabela/8/base -> origin/gh/lucaskabela/8/base 2025-11-03T16:58:01.7239932Z * [new branch] gh/lucaskabela/8/head -> origin/gh/lucaskabela/8/head 2025-11-03T16:58:01.7241633Z * [new branch] gh/lucaskabela/8/orig -> origin/gh/lucaskabela/8/orig 2025-11-03T16:58:01.7243724Z * [new branch] gh/lucaskabela/9/base -> origin/gh/lucaskabela/9/base 2025-11-03T16:58:01.7245284Z * [new branch] gh/lucaskabela/9/head -> origin/gh/lucaskabela/9/head 2025-11-03T16:58:01.7247035Z * [new branch] gh/lucaskabela/9/orig -> origin/gh/lucaskabela/9/orig 2025-11-03T16:58:01.7250073Z * [new branch] gh/lw/10/base -> origin/gh/lw/10/base 2025-11-03T16:58:01.7251601Z * [new branch] gh/lw/10/head -> origin/gh/lw/10/head 2025-11-03T16:58:01.7253290Z * [new branch] gh/lw/10/orig -> origin/gh/lw/10/orig 2025-11-03T16:58:01.7255545Z * [new branch] gh/lw/3/base -> origin/gh/lw/3/base 2025-11-03T16:58:01.7257197Z * [new branch] gh/lw/3/head -> origin/gh/lw/3/head 2025-11-03T16:58:01.7258894Z * [new branch] gh/lw/3/orig -> origin/gh/lw/3/orig 2025-11-03T16:58:01.7261146Z * [new branch] gh/lw/4/base -> origin/gh/lw/4/base 2025-11-03T16:58:01.7263041Z * [new branch] gh/lw/4/head -> origin/gh/lw/4/head 2025-11-03T16:58:01.7264703Z * [new branch] gh/lw/4/orig -> origin/gh/lw/4/orig 2025-11-03T16:58:01.7266933Z * [new branch] gh/lw/5/base -> origin/gh/lw/5/base 2025-11-03T16:58:01.7268634Z * [new branch] gh/lw/5/head -> origin/gh/lw/5/head 2025-11-03T16:58:01.7270374Z * [new branch] gh/lw/5/orig -> origin/gh/lw/5/orig 2025-11-03T16:58:01.7272683Z * [new branch] gh/lw/6/base -> origin/gh/lw/6/base 2025-11-03T16:58:01.7274353Z * [new branch] gh/lw/6/head -> origin/gh/lw/6/head 2025-11-03T16:58:01.7276021Z * [new branch] gh/lw/6/orig -> origin/gh/lw/6/orig 2025-11-03T16:58:01.7278353Z * [new branch] gh/lw/7/base -> origin/gh/lw/7/base 2025-11-03T16:58:01.7280021Z * [new branch] gh/lw/7/head -> origin/gh/lw/7/head 2025-11-03T16:58:01.7281729Z * [new branch] gh/lw/7/orig -> origin/gh/lw/7/orig 2025-11-03T16:58:01.7283956Z * [new branch] gh/lw/8/base -> origin/gh/lw/8/base 2025-11-03T16:58:01.7285630Z * [new branch] gh/lw/8/head -> origin/gh/lw/8/head 2025-11-03T16:58:01.7287315Z * [new branch] gh/lw/8/orig -> origin/gh/lw/8/orig 2025-11-03T16:58:01.7289663Z * [new branch] gh/lw/9/base -> origin/gh/lw/9/base 2025-11-03T16:58:01.7291459Z * [new branch] gh/lw/9/head -> origin/gh/lw/9/head 2025-11-03T16:58:01.7293145Z * [new branch] gh/lw/9/orig -> origin/gh/lw/9/orig 2025-11-03T16:58:01.7296024Z * [new branch] gh/maggiemoss/1/base -> origin/gh/maggiemoss/1/base 2025-11-03T16:58:01.7297689Z * [new branch] gh/maggiemoss/1/head -> origin/gh/maggiemoss/1/head 2025-11-03T16:58:01.7299389Z * [new branch] gh/maggiemoss/1/orig -> origin/gh/maggiemoss/1/orig 2025-11-03T16:58:01.7301710Z * [new branch] gh/maggiemoss/2/base -> origin/gh/maggiemoss/2/base 2025-11-03T16:58:01.7303557Z * [new branch] gh/maggiemoss/2/head -> origin/gh/maggiemoss/2/head 2025-11-03T16:58:01.7305195Z * [new branch] gh/maggiemoss/2/orig -> origin/gh/maggiemoss/2/orig 2025-11-03T16:58:01.7307943Z * [new branch] gh/malfet/14/base -> origin/gh/malfet/14/base 2025-11-03T16:58:01.7310281Z * [new branch] gh/malfet/417/base -> origin/gh/malfet/417/base 2025-11-03T16:58:01.7311972Z * [new branch] gh/malfet/417/head -> origin/gh/malfet/417/head 2025-11-03T16:58:01.7313688Z * [new branch] gh/malfet/417/orig -> origin/gh/malfet/417/orig 2025-11-03T16:58:01.7316022Z * [new branch] gh/malfet/506/base -> origin/gh/malfet/506/base 2025-11-03T16:58:01.7317716Z * [new branch] gh/malfet/506/head -> origin/gh/malfet/506/head 2025-11-03T16:58:01.7319384Z * [new branch] gh/malfet/506/orig -> origin/gh/malfet/506/orig 2025-11-03T16:58:01.7321902Z * [new branch] gh/malfet/517/base -> origin/gh/malfet/517/base 2025-11-03T16:58:01.7323754Z * [new branch] gh/malfet/517/head -> origin/gh/malfet/517/head 2025-11-03T16:58:01.7326002Z * [new branch] gh/malfet/528/base -> origin/gh/malfet/528/base 2025-11-03T16:58:01.7327675Z * [new branch] gh/malfet/528/head -> origin/gh/malfet/528/head 2025-11-03T16:58:01.7329339Z * [new branch] gh/malfet/528/orig -> origin/gh/malfet/528/orig 2025-11-03T16:58:01.7331615Z * [new branch] gh/malfet/537/base -> origin/gh/malfet/537/base 2025-11-03T16:58:01.7333281Z * [new branch] gh/malfet/537/head -> origin/gh/malfet/537/head 2025-11-03T16:58:01.7335007Z * [new branch] gh/malfet/537/orig -> origin/gh/malfet/537/orig 2025-11-03T16:58:01.7337321Z * [new branch] gh/malfet/546/base -> origin/gh/malfet/546/base 2025-11-03T16:58:01.7339220Z * [new branch] gh/malfet/546/head -> origin/gh/malfet/546/head 2025-11-03T16:58:01.7341145Z * [new branch] gh/malfet/546/orig -> origin/gh/malfet/546/orig 2025-11-03T16:58:01.7343501Z * [new branch] gh/malfet/547/base -> origin/gh/malfet/547/base 2025-11-03T16:58:01.7345173Z * [new branch] gh/malfet/547/head -> origin/gh/malfet/547/head 2025-11-03T16:58:01.7346840Z * [new branch] gh/malfet/547/orig -> origin/gh/malfet/547/orig 2025-11-03T16:58:01.7348946Z * [new branch] gh/malfet/548/base -> origin/gh/malfet/548/base 2025-11-03T16:58:01.7350672Z * [new branch] gh/malfet/548/head -> origin/gh/malfet/548/head 2025-11-03T16:58:01.7352391Z * [new branch] gh/malfet/548/orig -> origin/gh/malfet/548/orig 2025-11-03T16:58:01.7354642Z * [new branch] gh/malfet/549/base -> origin/gh/malfet/549/base 2025-11-03T16:58:01.7356359Z * [new branch] gh/malfet/549/head -> origin/gh/malfet/549/head 2025-11-03T16:58:01.7358097Z * [new branch] gh/malfet/549/orig -> origin/gh/malfet/549/orig 2025-11-03T16:58:01.7360432Z * [new branch] gh/malfet/550/base -> origin/gh/malfet/550/base 2025-11-03T16:58:01.7362162Z * [new branch] gh/malfet/550/head -> origin/gh/malfet/550/head 2025-11-03T16:58:01.7363898Z * [new branch] gh/malfet/550/orig -> origin/gh/malfet/550/orig 2025-11-03T16:58:01.7366144Z * [new branch] gh/malfet/551/base -> origin/gh/malfet/551/base 2025-11-03T16:58:01.7367838Z * [new branch] gh/malfet/551/head -> origin/gh/malfet/551/head 2025-11-03T16:58:01.7370023Z * [new branch] gh/malfet/551/orig -> origin/gh/malfet/551/orig 2025-11-03T16:58:01.7372328Z * [new branch] gh/malfet/552/base -> origin/gh/malfet/552/base 2025-11-03T16:58:01.7373981Z * [new branch] gh/malfet/552/head -> origin/gh/malfet/552/head 2025-11-03T16:58:01.7375661Z * [new branch] gh/malfet/552/orig -> origin/gh/malfet/552/orig 2025-11-03T16:58:01.7377964Z * [new branch] gh/malfet/553/base -> origin/gh/malfet/553/base 2025-11-03T16:58:01.7379704Z * [new branch] gh/malfet/553/head -> origin/gh/malfet/553/head 2025-11-03T16:58:01.7381397Z * [new branch] gh/malfet/553/orig -> origin/gh/malfet/553/orig 2025-11-03T16:58:01.7384932Z * [new branch] gh/malfet/554/base -> origin/gh/malfet/554/base 2025-11-03T16:58:01.7387100Z * [new branch] gh/malfet/554/head -> origin/gh/malfet/554/head 2025-11-03T16:58:01.7389092Z * [new branch] gh/malfet/554/orig -> origin/gh/malfet/554/orig 2025-11-03T16:58:01.7391794Z * [new branch] gh/malfet/555/base -> origin/gh/malfet/555/base 2025-11-03T16:58:01.7393695Z * [new branch] gh/malfet/555/head -> origin/gh/malfet/555/head 2025-11-03T16:58:01.7395144Z * [new branch] gh/malfet/555/orig -> origin/gh/malfet/555/orig 2025-11-03T16:58:01.7397409Z * [new branch] gh/malfet/556/base -> origin/gh/malfet/556/base 2025-11-03T16:58:01.7399071Z * [new branch] gh/malfet/556/head -> origin/gh/malfet/556/head 2025-11-03T16:58:01.7400892Z * [new branch] gh/malfet/556/orig -> origin/gh/malfet/556/orig 2025-11-03T16:58:01.7403778Z * [new branch] gh/malfet/557/base -> origin/gh/malfet/557/base 2025-11-03T16:58:01.7405488Z * [new branch] gh/malfet/557/head -> origin/gh/malfet/557/head 2025-11-03T16:58:01.7407174Z * [new branch] gh/malfet/557/orig -> origin/gh/malfet/557/orig 2025-11-03T16:58:01.7409567Z * [new branch] gh/malfet/558/base -> origin/gh/malfet/558/base 2025-11-03T16:58:01.7411292Z * [new branch] gh/malfet/558/head -> origin/gh/malfet/558/head 2025-11-03T16:58:01.7413026Z * [new branch] gh/malfet/558/orig -> origin/gh/malfet/558/orig 2025-11-03T16:58:01.7415389Z * [new branch] gh/malfet/559/base -> origin/gh/malfet/559/base 2025-11-03T16:58:01.7417092Z * [new branch] gh/malfet/559/head -> origin/gh/malfet/559/head 2025-11-03T16:58:01.7418777Z * [new branch] gh/malfet/559/orig -> origin/gh/malfet/559/orig 2025-11-03T16:58:01.7421386Z * [new branch] gh/malfet/560/base -> origin/gh/malfet/560/base 2025-11-03T16:58:01.7423377Z * [new branch] gh/malfet/560/head -> origin/gh/malfet/560/head 2025-11-03T16:58:01.7425048Z * [new branch] gh/malfet/560/orig -> origin/gh/malfet/560/orig 2025-11-03T16:58:01.7427550Z * [new branch] gh/malfet/561/base -> origin/gh/malfet/561/base 2025-11-03T16:58:01.7429132Z * [new branch] gh/malfet/561/head -> origin/gh/malfet/561/head 2025-11-03T16:58:01.7430866Z * [new branch] gh/malfet/561/orig -> origin/gh/malfet/561/orig 2025-11-03T16:58:01.7433303Z * [new branch] gh/malfet/562/base -> origin/gh/malfet/562/base 2025-11-03T16:58:01.7434977Z * [new branch] gh/malfet/562/head -> origin/gh/malfet/562/head 2025-11-03T16:58:01.7436688Z * [new branch] gh/malfet/562/orig -> origin/gh/malfet/562/orig 2025-11-03T16:58:01.7438982Z * [new branch] gh/malfet/563/base -> origin/gh/malfet/563/base 2025-11-03T16:58:01.7440796Z * [new branch] gh/malfet/563/head -> origin/gh/malfet/563/head 2025-11-03T16:58:01.7442574Z * [new branch] gh/malfet/563/orig -> origin/gh/malfet/563/orig 2025-11-03T16:58:01.7445432Z * [new branch] gh/malfet/564/base -> origin/gh/malfet/564/base 2025-11-03T16:58:01.7447657Z * [new branch] gh/malfet/564/head -> origin/gh/malfet/564/head 2025-11-03T16:58:01.7449093Z * [new branch] gh/malfet/564/orig -> origin/gh/malfet/564/orig 2025-11-03T16:58:01.7451335Z * [new branch] gh/malfet/565/base -> origin/gh/malfet/565/base 2025-11-03T16:58:01.7452992Z * [new branch] gh/malfet/565/head -> origin/gh/malfet/565/head 2025-11-03T16:58:01.7454700Z * [new branch] gh/malfet/565/orig -> origin/gh/malfet/565/orig 2025-11-03T16:58:01.7457377Z * [new branch] gh/malfet/566/base -> origin/gh/malfet/566/base 2025-11-03T16:58:01.7459306Z * [new branch] gh/malfet/566/head -> origin/gh/malfet/566/head 2025-11-03T16:58:01.7461004Z * [new branch] gh/malfet/566/orig -> origin/gh/malfet/566/orig 2025-11-03T16:58:01.7463835Z * [new branch] gh/malfet/567/base -> origin/gh/malfet/567/base 2025-11-03T16:58:01.7465363Z * [new branch] gh/malfet/567/head -> origin/gh/malfet/567/head 2025-11-03T16:58:01.7466958Z * [new branch] gh/malfet/567/orig -> origin/gh/malfet/567/orig 2025-11-03T16:58:01.7469439Z * [new branch] gh/malfet/568/base -> origin/gh/malfet/568/base 2025-11-03T16:58:01.7471152Z * [new branch] gh/malfet/568/head -> origin/gh/malfet/568/head 2025-11-03T16:58:01.7472894Z * [new branch] gh/malfet/568/orig -> origin/gh/malfet/568/orig 2025-11-03T16:58:01.7475287Z * [new branch] gh/malfet/569/base -> origin/gh/malfet/569/base 2025-11-03T16:58:01.7476964Z * [new branch] gh/malfet/569/head -> origin/gh/malfet/569/head 2025-11-03T16:58:01.7478709Z * [new branch] gh/malfet/569/orig -> origin/gh/malfet/569/orig 2025-11-03T16:58:01.7481135Z * [new branch] gh/malfet/570/base -> origin/gh/malfet/570/base 2025-11-03T16:58:01.7482871Z * [new branch] gh/malfet/570/head -> origin/gh/malfet/570/head 2025-11-03T16:58:01.7484614Z * [new branch] gh/malfet/570/orig -> origin/gh/malfet/570/orig 2025-11-03T16:58:01.7486965Z * [new branch] gh/malfet/571/base -> origin/gh/malfet/571/base 2025-11-03T16:58:01.7488648Z * [new branch] gh/malfet/571/head -> origin/gh/malfet/571/head 2025-11-03T16:58:01.7490355Z * [new branch] gh/malfet/571/orig -> origin/gh/malfet/571/orig 2025-11-03T16:58:01.7492674Z * [new branch] gh/malfet/572/base -> origin/gh/malfet/572/base 2025-11-03T16:58:01.7494377Z * [new branch] gh/malfet/572/head -> origin/gh/malfet/572/head 2025-11-03T16:58:01.7496163Z * [new branch] gh/malfet/572/orig -> origin/gh/malfet/572/orig 2025-11-03T16:58:01.7498565Z * [new branch] gh/malfet/573/base -> origin/gh/malfet/573/base 2025-11-03T16:58:01.7500273Z * [new branch] gh/malfet/573/head -> origin/gh/malfet/573/head 2025-11-03T16:58:01.7502542Z * [new branch] gh/malfet/573/orig -> origin/gh/malfet/573/orig 2025-11-03T16:58:01.7504999Z * [new branch] gh/malfet/574/base -> origin/gh/malfet/574/base 2025-11-03T16:58:01.7506683Z * [new branch] gh/malfet/574/head -> origin/gh/malfet/574/head 2025-11-03T16:58:01.7508478Z * [new branch] gh/malfet/574/orig -> origin/gh/malfet/574/orig 2025-11-03T16:58:01.7510813Z * [new branch] gh/malfet/575/base -> origin/gh/malfet/575/base 2025-11-03T16:58:01.7512542Z * [new branch] gh/malfet/575/head -> origin/gh/malfet/575/head 2025-11-03T16:58:01.7514245Z * [new branch] gh/malfet/575/orig -> origin/gh/malfet/575/orig 2025-11-03T16:58:01.7516591Z * [new branch] gh/malfet/576/base -> origin/gh/malfet/576/base 2025-11-03T16:58:01.7518304Z * [new branch] gh/malfet/576/head -> origin/gh/malfet/576/head 2025-11-03T16:58:01.7520128Z * [new branch] gh/malfet/576/orig -> origin/gh/malfet/576/orig 2025-11-03T16:58:01.7522908Z * [new branch] gh/malfet/577/base -> origin/gh/malfet/577/base 2025-11-03T16:58:01.7524592Z * [new branch] gh/malfet/577/head -> origin/gh/malfet/577/head 2025-11-03T16:58:01.7526258Z * [new branch] gh/malfet/577/orig -> origin/gh/malfet/577/orig 2025-11-03T16:58:01.7529268Z * [new branch] gh/malfet/578/base -> origin/gh/malfet/578/base 2025-11-03T16:58:01.7530958Z * [new branch] gh/malfet/578/head -> origin/gh/malfet/578/head 2025-11-03T16:58:01.7532698Z * [new branch] gh/malfet/578/orig -> origin/gh/malfet/578/orig 2025-11-03T16:58:01.7535309Z * [new branch] gh/malfet/579/base -> origin/gh/malfet/579/base 2025-11-03T16:58:01.7536898Z * [new branch] gh/malfet/579/head -> origin/gh/malfet/579/head 2025-11-03T16:58:01.7538545Z * [new branch] gh/malfet/579/orig -> origin/gh/malfet/579/orig 2025-11-03T16:58:01.7540848Z * [new branch] gh/malfet/580/base -> origin/gh/malfet/580/base 2025-11-03T16:58:01.7542668Z * [new branch] gh/malfet/580/head -> origin/gh/malfet/580/head 2025-11-03T16:58:01.7544347Z * [new branch] gh/malfet/580/orig -> origin/gh/malfet/580/orig 2025-11-03T16:58:01.7546683Z * [new branch] gh/malfet/581/base -> origin/gh/malfet/581/base 2025-11-03T16:58:01.7548359Z * [new branch] gh/malfet/581/head -> origin/gh/malfet/581/head 2025-11-03T16:58:01.7550106Z * [new branch] gh/malfet/581/orig -> origin/gh/malfet/581/orig 2025-11-03T16:58:01.7553103Z * [new branch] gh/malfet/582/base -> origin/gh/malfet/582/base 2025-11-03T16:58:01.7554711Z * [new branch] gh/malfet/582/head -> origin/gh/malfet/582/head 2025-11-03T16:58:01.7556473Z * [new branch] gh/malfet/582/orig -> origin/gh/malfet/582/orig 2025-11-03T16:58:01.7558824Z * [new branch] gh/malfet/583/base -> origin/gh/malfet/583/base 2025-11-03T16:58:01.7560457Z * [new branch] gh/malfet/583/head -> origin/gh/malfet/583/head 2025-11-03T16:58:01.7562096Z * [new branch] gh/malfet/583/orig -> origin/gh/malfet/583/orig 2025-11-03T16:58:01.7564917Z * [new branch] gh/malfet/584/base -> origin/gh/malfet/584/base 2025-11-03T16:58:01.7566624Z * [new branch] gh/malfet/584/head -> origin/gh/malfet/584/head 2025-11-03T16:58:01.7568281Z * [new branch] gh/malfet/584/orig -> origin/gh/malfet/584/orig 2025-11-03T16:58:01.7570617Z * [new branch] gh/malfet/585/base -> origin/gh/malfet/585/base 2025-11-03T16:58:01.7572308Z * [new branch] gh/malfet/585/head -> origin/gh/malfet/585/head 2025-11-03T16:58:01.7574118Z * [new branch] gh/malfet/585/orig -> origin/gh/malfet/585/orig 2025-11-03T16:58:01.7576465Z * [new branch] gh/malfet/586/base -> origin/gh/malfet/586/base 2025-11-03T16:58:01.7578162Z * [new branch] gh/malfet/586/head -> origin/gh/malfet/586/head 2025-11-03T16:58:01.7579807Z * [new branch] gh/malfet/586/orig -> origin/gh/malfet/586/orig 2025-11-03T16:58:01.7582572Z * [new branch] gh/malfet/587/base -> origin/gh/malfet/587/base 2025-11-03T16:58:01.7584232Z * [new branch] gh/malfet/587/head -> origin/gh/malfet/587/head 2025-11-03T16:58:01.7585939Z * [new branch] gh/malfet/587/orig -> origin/gh/malfet/587/orig 2025-11-03T16:58:01.7588273Z * [new branch] gh/malfet/64/base -> origin/gh/malfet/64/base 2025-11-03T16:58:01.7589948Z * [new branch] gh/malfet/64/head -> origin/gh/malfet/64/head 2025-11-03T16:58:01.7592850Z * [new branch] gh/manuelcandales/11/base -> origin/gh/manuelcandales/11/base 2025-11-03T16:58:01.7594708Z * [new branch] gh/manuelcandales/11/head -> origin/gh/manuelcandales/11/head 2025-11-03T16:58:01.7596573Z * [new branch] gh/manuelcandales/11/orig -> origin/gh/manuelcandales/11/orig 2025-11-03T16:58:01.7599619Z * [new branch] gh/markkm/1/base -> origin/gh/markkm/1/base 2025-11-03T16:58:01.7602897Z * [new branch] gh/masnesral/235/base -> origin/gh/masnesral/235/base 2025-11-03T16:58:01.7604673Z * [new branch] gh/masnesral/235/head -> origin/gh/masnesral/235/head 2025-11-03T16:58:01.7606555Z * [new branch] gh/masnesral/235/orig -> origin/gh/masnesral/235/orig 2025-11-03T16:58:01.7608747Z * [new branch] gh/masnesral/236/base -> origin/gh/masnesral/236/base 2025-11-03T16:58:01.7610428Z * [new branch] gh/masnesral/236/head -> origin/gh/masnesral/236/head 2025-11-03T16:58:01.7612090Z * [new branch] gh/masnesral/236/orig -> origin/gh/masnesral/236/orig 2025-11-03T16:58:01.7614422Z * [new branch] gh/masnesral/237/base -> origin/gh/masnesral/237/base 2025-11-03T16:58:01.7616750Z * [new branch] gh/masnesral/237/head -> origin/gh/masnesral/237/head 2025-11-03T16:58:01.7618630Z * [new branch] gh/masnesral/237/orig -> origin/gh/masnesral/237/orig 2025-11-03T16:58:01.7621239Z * [new branch] gh/masnesral/238/base -> origin/gh/masnesral/238/base 2025-11-03T16:58:01.7624828Z * [new branch] gh/masnesral/238/head -> origin/gh/masnesral/238/head 2025-11-03T16:58:01.7626519Z * [new branch] gh/masnesral/238/orig -> origin/gh/masnesral/238/orig 2025-11-03T16:58:01.7629351Z * [new branch] gh/mhorowitz/0/base -> origin/gh/mhorowitz/0/base 2025-11-03T16:58:01.7631117Z * [new branch] gh/mhorowitz/0/head -> origin/gh/mhorowitz/0/head 2025-11-03T16:58:01.7633325Z * [new branch] gh/mhorowitz/1/base -> origin/gh/mhorowitz/1/base 2025-11-03T16:58:01.7635016Z * [new branch] gh/mhorowitz/1/head -> origin/gh/mhorowitz/1/head 2025-11-03T16:58:01.7637204Z * [new branch] gh/mhorowitz/2/base -> origin/gh/mhorowitz/2/base 2025-11-03T16:58:01.7638939Z * [new branch] gh/mhorowitz/2/head -> origin/gh/mhorowitz/2/head 2025-11-03T16:58:01.7641094Z * [new branch] gh/mhorowitz/3/base -> origin/gh/mhorowitz/3/base 2025-11-03T16:58:01.7642741Z * [new branch] gh/mhorowitz/3/head -> origin/gh/mhorowitz/3/head 2025-11-03T16:58:01.7644994Z * [new branch] gh/mhorowitz/4/base -> origin/gh/mhorowitz/4/base 2025-11-03T16:58:01.7646667Z * [new branch] gh/mhorowitz/4/head -> origin/gh/mhorowitz/4/head 2025-11-03T16:58:01.7648811Z * [new branch] gh/mhorowitz/5/base -> origin/gh/mhorowitz/5/base 2025-11-03T16:58:01.7650414Z * [new branch] gh/mhorowitz/5/head -> origin/gh/mhorowitz/5/head 2025-11-03T16:58:01.7652652Z * [new branch] gh/mhorowitz/6/base -> origin/gh/mhorowitz/6/base 2025-11-03T16:58:01.7654242Z * [new branch] gh/mhorowitz/6/head -> origin/gh/mhorowitz/6/head 2025-11-03T16:58:01.7657181Z * [new branch] gh/mikaylagawarecki/234/base -> origin/gh/mikaylagawarecki/234/base 2025-11-03T16:58:01.7658905Z * [new branch] gh/mikaylagawarecki/234/head -> origin/gh/mikaylagawarecki/234/head 2025-11-03T16:58:01.7661131Z * [new branch] gh/mikaylagawarecki/235/base -> origin/gh/mikaylagawarecki/235/base 2025-11-03T16:58:01.7662965Z * [new branch] gh/mikaylagawarecki/235/head -> origin/gh/mikaylagawarecki/235/head 2025-11-03T16:58:01.7665294Z * [new branch] gh/mikaylagawarecki/236/base -> origin/gh/mikaylagawarecki/236/base 2025-11-03T16:58:01.7666894Z * [new branch] gh/mikaylagawarecki/236/head -> origin/gh/mikaylagawarecki/236/head 2025-11-03T16:58:01.7669128Z * [new branch] gh/mikaylagawarecki/237/base -> origin/gh/mikaylagawarecki/237/base 2025-11-03T16:58:01.7670734Z * [new branch] gh/mikaylagawarecki/237/head -> origin/gh/mikaylagawarecki/237/head 2025-11-03T16:58:01.7673203Z * [new branch] gh/mikaylagawarecki/238/base -> origin/gh/mikaylagawarecki/238/base 2025-11-03T16:58:01.7674883Z * [new branch] gh/mikaylagawarecki/238/head -> origin/gh/mikaylagawarecki/238/head 2025-11-03T16:58:01.7677747Z * [new branch] gh/mikaylagawarecki/336/base -> origin/gh/mikaylagawarecki/336/base 2025-11-03T16:58:01.7679349Z * [new branch] gh/mikaylagawarecki/336/head -> origin/gh/mikaylagawarecki/336/head 2025-11-03T16:58:01.7681052Z * [new branch] gh/mikaylagawarecki/336/orig -> origin/gh/mikaylagawarecki/336/orig 2025-11-03T16:58:01.7683969Z * [new branch] gh/mikaylagawarecki/337/base -> origin/gh/mikaylagawarecki/337/base 2025-11-03T16:58:01.7685805Z * [new branch] gh/mikaylagawarecki/337/head -> origin/gh/mikaylagawarecki/337/head 2025-11-03T16:58:01.7687585Z * [new branch] gh/mikaylagawarecki/337/orig -> origin/gh/mikaylagawarecki/337/orig 2025-11-03T16:58:01.7690096Z * [new branch] gh/mikaylagawarecki/341/base -> origin/gh/mikaylagawarecki/341/base 2025-11-03T16:58:01.7691797Z * [new branch] gh/mikaylagawarecki/341/head -> origin/gh/mikaylagawarecki/341/head 2025-11-03T16:58:01.7693549Z * [new branch] gh/mikaylagawarecki/341/orig -> origin/gh/mikaylagawarecki/341/orig 2025-11-03T16:58:01.7696035Z * [new branch] gh/mikaylagawarecki/342/base -> origin/gh/mikaylagawarecki/342/base 2025-11-03T16:58:01.7697568Z * [new branch] gh/mikaylagawarecki/342/head -> origin/gh/mikaylagawarecki/342/head 2025-11-03T16:58:01.7699367Z * [new branch] gh/mikaylagawarecki/342/orig -> origin/gh/mikaylagawarecki/342/orig 2025-11-03T16:58:01.7701733Z * [new branch] gh/mikaylagawarecki/343/base -> origin/gh/mikaylagawarecki/343/base 2025-11-03T16:58:01.7703593Z * [new branch] gh/mikaylagawarecki/343/head -> origin/gh/mikaylagawarecki/343/head 2025-11-03T16:58:01.7705232Z * [new branch] gh/mikaylagawarecki/343/orig -> origin/gh/mikaylagawarecki/343/orig 2025-11-03T16:58:01.7707520Z * [new branch] gh/mikaylagawarecki/344/base -> origin/gh/mikaylagawarecki/344/base 2025-11-03T16:58:01.7709274Z * [new branch] gh/mikaylagawarecki/344/head -> origin/gh/mikaylagawarecki/344/head 2025-11-03T16:58:01.7711241Z * [new branch] gh/mikaylagawarecki/344/orig -> origin/gh/mikaylagawarecki/344/orig 2025-11-03T16:58:01.7713581Z * [new branch] gh/mikaylagawarecki/345/base -> origin/gh/mikaylagawarecki/345/base 2025-11-03T16:58:01.7715208Z * [new branch] gh/mikaylagawarecki/345/head -> origin/gh/mikaylagawarecki/345/head 2025-11-03T16:58:01.7716844Z * [new branch] gh/mikaylagawarecki/345/orig -> origin/gh/mikaylagawarecki/345/orig 2025-11-03T16:58:01.7719341Z * [new branch] gh/mikaylagawarecki/346/base -> origin/gh/mikaylagawarecki/346/base 2025-11-03T16:58:01.7720950Z * [new branch] gh/mikaylagawarecki/346/head -> origin/gh/mikaylagawarecki/346/head 2025-11-03T16:58:01.7723106Z * [new branch] gh/mikaylagawarecki/346/orig -> origin/gh/mikaylagawarecki/346/orig 2025-11-03T16:58:01.7725455Z * [new branch] gh/mikaylagawarecki/347/base -> origin/gh/mikaylagawarecki/347/base 2025-11-03T16:58:01.7727133Z * [new branch] gh/mikaylagawarecki/347/head -> origin/gh/mikaylagawarecki/347/head 2025-11-03T16:58:01.7728754Z * [new branch] gh/mikaylagawarecki/347/orig -> origin/gh/mikaylagawarecki/347/orig 2025-11-03T16:58:01.7731213Z * [new branch] gh/mikaylagawarecki/348/base -> origin/gh/mikaylagawarecki/348/base 2025-11-03T16:58:01.7732503Z * [new branch] gh/mikaylagawarecki/348/head -> origin/gh/mikaylagawarecki/348/head 2025-11-03T16:58:01.7734282Z * [new branch] gh/mikaylagawarecki/348/orig -> origin/gh/mikaylagawarecki/348/orig 2025-11-03T16:58:01.7736709Z * [new branch] gh/mikaylagawarecki/349/base -> origin/gh/mikaylagawarecki/349/base 2025-11-03T16:58:01.7738356Z * [new branch] gh/mikaylagawarecki/349/head -> origin/gh/mikaylagawarecki/349/head 2025-11-03T16:58:01.7740121Z * [new branch] gh/mikaylagawarecki/349/orig -> origin/gh/mikaylagawarecki/349/orig 2025-11-03T16:58:01.7743356Z * [new branch] gh/mikaylagawarecki/350/base -> origin/gh/mikaylagawarecki/350/base 2025-11-03T16:58:01.7744882Z * [new branch] gh/mikaylagawarecki/350/head -> origin/gh/mikaylagawarecki/350/head 2025-11-03T16:58:01.7746514Z * [new branch] gh/mikaylagawarecki/350/orig -> origin/gh/mikaylagawarecki/350/orig 2025-11-03T16:58:01.7749782Z * [new branch] gh/mikaylagawarecki/351/base -> origin/gh/mikaylagawarecki/351/base 2025-11-03T16:58:01.7751597Z * [new branch] gh/mikaylagawarecki/351/head -> origin/gh/mikaylagawarecki/351/head 2025-11-03T16:58:01.7753676Z * [new branch] gh/mikaylagawarecki/351/orig -> origin/gh/mikaylagawarecki/351/orig 2025-11-03T16:58:01.7756259Z * [new branch] gh/mikaylagawarecki/352/base -> origin/gh/mikaylagawarecki/352/base 2025-11-03T16:58:01.7758112Z * [new branch] gh/mikaylagawarecki/352/head -> origin/gh/mikaylagawarecki/352/head 2025-11-03T16:58:01.7759900Z * [new branch] gh/mikaylagawarecki/352/orig -> origin/gh/mikaylagawarecki/352/orig 2025-11-03T16:58:01.7762520Z * [new branch] gh/mikaylagawarecki/353/base -> origin/gh/mikaylagawarecki/353/base 2025-11-03T16:58:01.7764418Z * [new branch] gh/mikaylagawarecki/353/head -> origin/gh/mikaylagawarecki/353/head 2025-11-03T16:58:01.7766067Z * [new branch] gh/mikaylagawarecki/353/orig -> origin/gh/mikaylagawarecki/353/orig 2025-11-03T16:58:01.7768773Z * [new branch] gh/mikaylagawarecki/354/base -> origin/gh/mikaylagawarecki/354/base 2025-11-03T16:58:01.7770313Z * [new branch] gh/mikaylagawarecki/354/head -> origin/gh/mikaylagawarecki/354/head 2025-11-03T16:58:01.7772142Z * [new branch] gh/mikaylagawarecki/354/orig -> origin/gh/mikaylagawarecki/354/orig 2025-11-03T16:58:01.7774509Z * [new branch] gh/mikaylagawarecki/355/base -> origin/gh/mikaylagawarecki/355/base 2025-11-03T16:58:01.7776696Z * [new branch] gh/mikaylagawarecki/355/head -> origin/gh/mikaylagawarecki/355/head 2025-11-03T16:58:01.7778460Z * [new branch] gh/mikaylagawarecki/355/orig -> origin/gh/mikaylagawarecki/355/orig 2025-11-03T16:58:01.7780900Z * [new branch] gh/mikaylagawarecki/356/base -> origin/gh/mikaylagawarecki/356/base 2025-11-03T16:58:01.7783131Z * [new branch] gh/mikaylagawarecki/356/head -> origin/gh/mikaylagawarecki/356/head 2025-11-03T16:58:01.7784888Z * [new branch] gh/mikaylagawarecki/356/orig -> origin/gh/mikaylagawarecki/356/orig 2025-11-03T16:58:01.7787159Z * [new branch] gh/mikaylagawarecki/357/base -> origin/gh/mikaylagawarecki/357/base 2025-11-03T16:58:01.7788948Z * [new branch] gh/mikaylagawarecki/357/head -> origin/gh/mikaylagawarecki/357/head 2025-11-03T16:58:01.7790529Z * [new branch] gh/mikaylagawarecki/357/orig -> origin/gh/mikaylagawarecki/357/orig 2025-11-03T16:58:01.7792924Z * [new branch] gh/mikaylagawarecki/358/base -> origin/gh/mikaylagawarecki/358/base 2025-11-03T16:58:01.7794555Z * [new branch] gh/mikaylagawarecki/358/head -> origin/gh/mikaylagawarecki/358/head 2025-11-03T16:58:01.7796266Z * [new branch] gh/mikaylagawarecki/358/orig -> origin/gh/mikaylagawarecki/358/orig 2025-11-03T16:58:01.7798528Z * [new branch] gh/mikaylagawarecki/359/base -> origin/gh/mikaylagawarecki/359/base 2025-11-03T16:58:01.7800327Z * [new branch] gh/mikaylagawarecki/359/head -> origin/gh/mikaylagawarecki/359/head 2025-11-03T16:58:01.7802058Z * [new branch] gh/mikaylagawarecki/359/orig -> origin/gh/mikaylagawarecki/359/orig 2025-11-03T16:58:01.7804523Z * [new branch] gh/mikaylagawarecki/360/base -> origin/gh/mikaylagawarecki/360/base 2025-11-03T16:58:01.7806191Z * [new branch] gh/mikaylagawarecki/360/head -> origin/gh/mikaylagawarecki/360/head 2025-11-03T16:58:01.7808197Z * [new branch] gh/mikaylagawarecki/360/orig -> origin/gh/mikaylagawarecki/360/orig 2025-11-03T16:58:01.7810481Z * [new branch] gh/mikaylagawarecki/361/base -> origin/gh/mikaylagawarecki/361/base 2025-11-03T16:58:01.7812360Z * [new branch] gh/mikaylagawarecki/361/head -> origin/gh/mikaylagawarecki/361/head 2025-11-03T16:58:01.7814041Z * [new branch] gh/mikaylagawarecki/361/orig -> origin/gh/mikaylagawarecki/361/orig 2025-11-03T16:58:01.7817350Z * [new branch] gh/mlazos/20/base -> origin/gh/mlazos/20/base 2025-11-03T16:58:01.7819066Z * [new branch] gh/mlazos/20/head -> origin/gh/mlazos/20/head 2025-11-03T16:58:01.7820810Z * [new branch] gh/mlazos/20/orig -> origin/gh/mlazos/20/orig 2025-11-03T16:58:01.7823554Z * [new branch] gh/mlazos/21/base -> origin/gh/mlazos/21/base 2025-11-03T16:58:01.7825300Z * [new branch] gh/mlazos/21/head -> origin/gh/mlazos/21/head 2025-11-03T16:58:01.7826994Z * [new branch] gh/mlazos/21/orig -> origin/gh/mlazos/21/orig 2025-11-03T16:58:01.7829267Z * [new branch] gh/mlazos/22/base -> origin/gh/mlazos/22/base 2025-11-03T16:58:01.7830925Z * [new branch] gh/mlazos/22/head -> origin/gh/mlazos/22/head 2025-11-03T16:58:01.7832637Z * [new branch] gh/mlazos/22/orig -> origin/gh/mlazos/22/orig 2025-11-03T16:58:01.7834907Z * [new branch] gh/mlazos/23/base -> origin/gh/mlazos/23/base 2025-11-03T16:58:01.7836657Z * [new branch] gh/mlazos/23/head -> origin/gh/mlazos/23/head 2025-11-03T16:58:01.7838384Z * [new branch] gh/mlazos/23/orig -> origin/gh/mlazos/23/orig 2025-11-03T16:58:01.7840690Z * [new branch] gh/mlazos/24/base -> origin/gh/mlazos/24/base 2025-11-03T16:58:01.7842362Z * [new branch] gh/mlazos/24/head -> origin/gh/mlazos/24/head 2025-11-03T16:58:01.7844042Z * [new branch] gh/mlazos/24/orig -> origin/gh/mlazos/24/orig 2025-11-03T16:58:01.7846503Z * [new branch] gh/mlazos/25/base -> origin/gh/mlazos/25/base 2025-11-03T16:58:01.7848225Z * [new branch] gh/mlazos/25/head -> origin/gh/mlazos/25/head 2025-11-03T16:58:01.7849955Z * [new branch] gh/mlazos/25/orig -> origin/gh/mlazos/25/orig 2025-11-03T16:58:01.7852187Z * [new branch] gh/mlazos/26/base -> origin/gh/mlazos/26/base 2025-11-03T16:58:01.7853907Z * [new branch] gh/mlazos/26/head -> origin/gh/mlazos/26/head 2025-11-03T16:58:01.7855591Z * [new branch] gh/mlazos/26/orig -> origin/gh/mlazos/26/orig 2025-11-03T16:58:01.7858364Z * [new branch] gh/mlazos/27/base -> origin/gh/mlazos/27/base 2025-11-03T16:58:01.7860073Z * [new branch] gh/mlazos/27/head -> origin/gh/mlazos/27/head 2025-11-03T16:58:01.7861878Z * [new branch] gh/mlazos/27/orig -> origin/gh/mlazos/27/orig 2025-11-03T16:58:01.7864230Z * [new branch] gh/mlazos/28/base -> origin/gh/mlazos/28/base 2025-11-03T16:58:01.7866009Z * [new branch] gh/mlazos/28/head -> origin/gh/mlazos/28/head 2025-11-03T16:58:01.7867640Z * [new branch] gh/mlazos/28/orig -> origin/gh/mlazos/28/orig 2025-11-03T16:58:01.7869924Z * [new branch] gh/mlazos/29/base -> origin/gh/mlazos/29/base 2025-11-03T16:58:01.7871612Z * [new branch] gh/mlazos/29/head -> origin/gh/mlazos/29/head 2025-11-03T16:58:01.7873360Z * [new branch] gh/mlazos/29/orig -> origin/gh/mlazos/29/orig 2025-11-03T16:58:01.7875564Z * [new branch] gh/mlazos/30/base -> origin/gh/mlazos/30/base 2025-11-03T16:58:01.7877351Z * [new branch] gh/mlazos/30/head -> origin/gh/mlazos/30/head 2025-11-03T16:58:01.7879227Z * [new branch] gh/mlazos/30/orig -> origin/gh/mlazos/30/orig 2025-11-03T16:58:01.7881383Z * [new branch] gh/mlazos/31/base -> origin/gh/mlazos/31/base 2025-11-03T16:58:01.7883023Z * [new branch] gh/mlazos/31/head -> origin/gh/mlazos/31/head 2025-11-03T16:58:01.7884668Z * [new branch] gh/mlazos/31/orig -> origin/gh/mlazos/31/orig 2025-11-03T16:58:01.7886998Z * [new branch] gh/mlazos/32/base -> origin/gh/mlazos/32/base 2025-11-03T16:58:01.7888677Z * [new branch] gh/mlazos/32/head -> origin/gh/mlazos/32/head 2025-11-03T16:58:01.7890417Z * [new branch] gh/mlazos/32/orig -> origin/gh/mlazos/32/orig 2025-11-03T16:58:01.7892780Z * [new branch] gh/mlazos/33/base -> origin/gh/mlazos/33/base 2025-11-03T16:58:01.7894507Z * [new branch] gh/mlazos/33/head -> origin/gh/mlazos/33/head 2025-11-03T16:58:01.7896193Z * [new branch] gh/mlazos/33/orig -> origin/gh/mlazos/33/orig 2025-11-03T16:58:01.7898549Z * [new branch] gh/mlazos/34/base -> origin/gh/mlazos/34/base 2025-11-03T16:58:01.7900176Z * [new branch] gh/mlazos/34/head -> origin/gh/mlazos/34/head 2025-11-03T16:58:01.7901998Z * [new branch] gh/mlazos/34/orig -> origin/gh/mlazos/34/orig 2025-11-03T16:58:01.7904172Z * [new branch] gh/mlazos/35/base -> origin/gh/mlazos/35/base 2025-11-03T16:58:01.7905818Z * [new branch] gh/mlazos/35/head -> origin/gh/mlazos/35/head 2025-11-03T16:58:01.7907478Z * [new branch] gh/mlazos/35/orig -> origin/gh/mlazos/35/orig 2025-11-03T16:58:01.7909800Z * [new branch] gh/mlazos/36/base -> origin/gh/mlazos/36/base 2025-11-03T16:58:01.7911465Z * [new branch] gh/mlazos/36/head -> origin/gh/mlazos/36/head 2025-11-03T16:58:01.7913165Z * [new branch] gh/mlazos/36/orig -> origin/gh/mlazos/36/orig 2025-11-03T16:58:01.7915465Z * [new branch] gh/mlazos/37/base -> origin/gh/mlazos/37/base 2025-11-03T16:58:01.7917182Z * [new branch] gh/mlazos/37/head -> origin/gh/mlazos/37/head 2025-11-03T16:58:01.7918919Z * [new branch] gh/mlazos/37/orig -> origin/gh/mlazos/37/orig 2025-11-03T16:58:01.7922577Z * [new branch] gh/mlazos/38/base -> origin/gh/mlazos/38/base 2025-11-03T16:58:01.7924477Z * [new branch] gh/mlazos/38/head -> origin/gh/mlazos/38/head 2025-11-03T16:58:01.7926183Z * [new branch] gh/mlazos/38/orig -> origin/gh/mlazos/38/orig 2025-11-03T16:58:01.7928606Z * [new branch] gh/mlazos/39/base -> origin/gh/mlazos/39/base 2025-11-03T16:58:01.7930330Z * [new branch] gh/mlazos/39/head -> origin/gh/mlazos/39/head 2025-11-03T16:58:01.7932079Z * [new branch] gh/mlazos/39/orig -> origin/gh/mlazos/39/orig 2025-11-03T16:58:01.7934406Z * [new branch] gh/mlazos/40/base -> origin/gh/mlazos/40/base 2025-11-03T16:58:01.7935982Z * [new branch] gh/mlazos/40/head -> origin/gh/mlazos/40/head 2025-11-03T16:58:01.7937777Z * [new branch] gh/mlazos/40/orig -> origin/gh/mlazos/40/orig 2025-11-03T16:58:01.7940166Z * [new branch] gh/mlazos/41/base -> origin/gh/mlazos/41/base 2025-11-03T16:58:01.7941893Z * [new branch] gh/mlazos/41/head -> origin/gh/mlazos/41/head 2025-11-03T16:58:01.7943534Z * [new branch] gh/mlazos/41/orig -> origin/gh/mlazos/41/orig 2025-11-03T16:58:01.7945962Z * [new branch] gh/mlazos/42/base -> origin/gh/mlazos/42/base 2025-11-03T16:58:01.7947644Z * [new branch] gh/mlazos/42/head -> origin/gh/mlazos/42/head 2025-11-03T16:58:01.7949519Z * [new branch] gh/mlazos/42/orig -> origin/gh/mlazos/42/orig 2025-11-03T16:58:01.7951681Z * [new branch] gh/mlazos/43/base -> origin/gh/mlazos/43/base 2025-11-03T16:58:01.7953362Z * [new branch] gh/mlazos/43/head -> origin/gh/mlazos/43/head 2025-11-03T16:58:01.7955002Z * [new branch] gh/mlazos/43/orig -> origin/gh/mlazos/43/orig 2025-11-03T16:58:01.7957263Z * [new branch] gh/mlazos/44/base -> origin/gh/mlazos/44/base 2025-11-03T16:58:01.7958944Z * [new branch] gh/mlazos/44/head -> origin/gh/mlazos/44/head 2025-11-03T16:58:01.7960674Z * [new branch] gh/mlazos/44/orig -> origin/gh/mlazos/44/orig 2025-11-03T16:58:01.7963081Z * [new branch] gh/mlazos/45/base -> origin/gh/mlazos/45/base 2025-11-03T16:58:01.7964828Z * [new branch] gh/mlazos/45/head -> origin/gh/mlazos/45/head 2025-11-03T16:58:01.7966559Z * [new branch] gh/mlazos/45/orig -> origin/gh/mlazos/45/orig 2025-11-03T16:58:01.7968807Z * [new branch] gh/mlazos/46/base -> origin/gh/mlazos/46/base 2025-11-03T16:58:01.7970529Z * [new branch] gh/mlazos/46/head -> origin/gh/mlazos/46/head 2025-11-03T16:58:01.7972254Z * [new branch] gh/mlazos/46/orig -> origin/gh/mlazos/46/orig 2025-11-03T16:58:01.7975078Z * [new branch] gh/mrmiywj/1/base -> origin/gh/mrmiywj/1/base 2025-11-03T16:58:01.7976841Z * [new branch] gh/mrmiywj/1/head -> origin/gh/mrmiywj/1/head 2025-11-03T16:58:01.7979771Z * [new branch] gh/muchulee8/68/base -> origin/gh/muchulee8/68/base 2025-11-03T16:58:01.7981638Z * [new branch] gh/muchulee8/68/head -> origin/gh/muchulee8/68/head 2025-11-03T16:58:01.7983262Z * [new branch] gh/muchulee8/68/orig -> origin/gh/muchulee8/68/orig 2025-11-03T16:58:01.7985466Z * [new branch] gh/muchulee8/69/base -> origin/gh/muchulee8/69/base 2025-11-03T16:58:01.7987204Z * [new branch] gh/muchulee8/69/head -> origin/gh/muchulee8/69/head 2025-11-03T16:58:01.7988889Z * [new branch] gh/muchulee8/69/orig -> origin/gh/muchulee8/69/orig 2025-11-03T16:58:01.7991128Z * [new branch] gh/muchulee8/70/base -> origin/gh/muchulee8/70/base 2025-11-03T16:58:01.7992878Z * [new branch] gh/muchulee8/70/head -> origin/gh/muchulee8/70/head 2025-11-03T16:58:01.7994583Z * [new branch] gh/muchulee8/70/orig -> origin/gh/muchulee8/70/orig 2025-11-03T16:58:01.7996765Z * [new branch] gh/muchulee8/71/base -> origin/gh/muchulee8/71/base 2025-11-03T16:58:01.7998547Z * [new branch] gh/muchulee8/71/head -> origin/gh/muchulee8/71/head 2025-11-03T16:58:01.8000243Z * [new branch] gh/muchulee8/71/orig -> origin/gh/muchulee8/71/orig 2025-11-03T16:58:01.8002534Z * [new branch] gh/muchulee8/72/base -> origin/gh/muchulee8/72/base 2025-11-03T16:58:01.8004348Z * [new branch] gh/muchulee8/72/head -> origin/gh/muchulee8/72/head 2025-11-03T16:58:01.8006004Z * [new branch] gh/muchulee8/72/orig -> origin/gh/muchulee8/72/orig 2025-11-03T16:58:01.8008560Z * [new branch] gh/muchulee8/73/base -> origin/gh/muchulee8/73/base 2025-11-03T16:58:01.8010394Z * [new branch] gh/muchulee8/73/head -> origin/gh/muchulee8/73/head 2025-11-03T16:58:01.8012184Z * [new branch] gh/muchulee8/73/orig -> origin/gh/muchulee8/73/orig 2025-11-03T16:58:01.8015102Z * [new branch] gh/naveenthangudu/1/base -> origin/gh/naveenthangudu/1/base 2025-11-03T16:58:01.8016770Z * [new branch] gh/naveenthangudu/1/head -> origin/gh/naveenthangudu/1/head 2025-11-03T16:58:01.8018954Z * [new branch] gh/naveenthangudu/1/orig -> origin/gh/naveenthangudu/1/orig 2025-11-03T16:58:01.8021349Z * [new branch] gh/naveenthangudu/2/base -> origin/gh/naveenthangudu/2/base 2025-11-03T16:58:01.8025441Z * [new branch] gh/naveenthangudu/2/head -> origin/gh/naveenthangudu/2/head 2025-11-03T16:58:01.8027581Z * [new branch] gh/naveenthangudu/2/orig -> origin/gh/naveenthangudu/2/orig 2025-11-03T16:58:01.8029892Z * [new branch] gh/naveenthangudu/3/base -> origin/gh/naveenthangudu/3/base 2025-11-03T16:58:01.8031705Z * [new branch] gh/naveenthangudu/3/head -> origin/gh/naveenthangudu/3/head 2025-11-03T16:58:01.8033688Z * [new branch] gh/naveenthangudu/3/orig -> origin/gh/naveenthangudu/3/orig 2025-11-03T16:58:01.8035950Z * [new branch] gh/naveenthangudu/4/base -> origin/gh/naveenthangudu/4/base 2025-11-03T16:58:01.8037728Z * [new branch] gh/naveenthangudu/4/head -> origin/gh/naveenthangudu/4/head 2025-11-03T16:58:01.8039578Z * [new branch] gh/naveenthangudu/4/orig -> origin/gh/naveenthangudu/4/orig 2025-11-03T16:58:01.8041917Z * [new branch] gh/naveenthangudu/5/base -> origin/gh/naveenthangudu/5/base 2025-11-03T16:58:01.8043637Z * [new branch] gh/naveenthangudu/5/head -> origin/gh/naveenthangudu/5/head 2025-11-03T16:58:01.8045518Z * [new branch] gh/naveenthangudu/5/orig -> origin/gh/naveenthangudu/5/orig 2025-11-03T16:58:01.8047733Z * [new branch] gh/naveenthangudu/6/base -> origin/gh/naveenthangudu/6/base 2025-11-03T16:58:01.8049404Z * [new branch] gh/naveenthangudu/6/head -> origin/gh/naveenthangudu/6/head 2025-11-03T16:58:01.8051018Z * [new branch] gh/naveenthangudu/6/orig -> origin/gh/naveenthangudu/6/orig 2025-11-03T16:58:01.8053334Z * [new branch] gh/naveenthangudu/7/base -> origin/gh/naveenthangudu/7/base 2025-11-03T16:58:01.8055033Z * [new branch] gh/naveenthangudu/7/head -> origin/gh/naveenthangudu/7/head 2025-11-03T16:58:01.8056655Z * [new branch] gh/naveenthangudu/7/orig -> origin/gh/naveenthangudu/7/orig 2025-11-03T16:58:01.8059372Z * [new branch] gh/naveenthangudu/8/base -> origin/gh/naveenthangudu/8/base 2025-11-03T16:58:01.8061107Z * [new branch] gh/naveenthangudu/8/head -> origin/gh/naveenthangudu/8/head 2025-11-03T16:58:01.8062985Z * [new branch] gh/naveenthangudu/8/orig -> origin/gh/naveenthangudu/8/orig 2025-11-03T16:58:01.8065346Z * [new branch] gh/naveenthangudu/9/base -> origin/gh/naveenthangudu/9/base 2025-11-03T16:58:01.8066986Z * [new branch] gh/naveenthangudu/9/head -> origin/gh/naveenthangudu/9/head 2025-11-03T16:58:01.8068775Z * [new branch] gh/naveenthangudu/9/orig -> origin/gh/naveenthangudu/9/orig 2025-11-03T16:58:01.8071522Z * [new branch] gh/nikitaved/1/base -> origin/gh/nikitaved/1/base 2025-11-03T16:58:01.8073202Z * [new branch] gh/nikitaved/1/head -> origin/gh/nikitaved/1/head 2025-11-03T16:58:01.8074885Z * [new branch] gh/nikitaved/1/orig -> origin/gh/nikitaved/1/orig 2025-11-03T16:58:01.8077216Z * [new branch] gh/nikitaved/2/base -> origin/gh/nikitaved/2/base 2025-11-03T16:58:01.8078950Z * [new branch] gh/nikitaved/2/head -> origin/gh/nikitaved/2/head 2025-11-03T16:58:01.8080698Z * [new branch] gh/nikitaved/2/orig -> origin/gh/nikitaved/2/orig 2025-11-03T16:58:01.8083130Z * [new branch] gh/nikitaved/4/base -> origin/gh/nikitaved/4/base 2025-11-03T16:58:01.8084921Z * [new branch] gh/nikitaved/4/head -> origin/gh/nikitaved/4/head 2025-11-03T16:58:01.8086783Z * [new branch] gh/nikitaved/4/orig -> origin/gh/nikitaved/4/orig 2025-11-03T16:58:01.8089650Z * [new branch] gh/nikitaved/5/base -> origin/gh/nikitaved/5/base 2025-11-03T16:58:01.8091298Z * [new branch] gh/nikitaved/5/head -> origin/gh/nikitaved/5/head 2025-11-03T16:58:01.8092935Z * [new branch] gh/nikitaved/5/orig -> origin/gh/nikitaved/5/orig 2025-11-03T16:58:01.8095289Z * [new branch] gh/nikitaved/6/base -> origin/gh/nikitaved/6/base 2025-11-03T16:58:01.8097028Z * [new branch] gh/nikitaved/6/head -> origin/gh/nikitaved/6/head 2025-11-03T16:58:01.8098743Z * [new branch] gh/nikitaved/6/orig -> origin/gh/nikitaved/6/orig 2025-11-03T16:58:01.8101071Z * [new branch] gh/nikitaved/7/base -> origin/gh/nikitaved/7/base 2025-11-03T16:58:01.8102945Z * [new branch] gh/nikitaved/7/head -> origin/gh/nikitaved/7/head 2025-11-03T16:58:01.8104671Z * [new branch] gh/nikitaved/7/orig -> origin/gh/nikitaved/7/orig 2025-11-03T16:58:01.8107462Z * [new branch] gh/nikitaved/8/base -> origin/gh/nikitaved/8/base 2025-11-03T16:58:01.8109187Z * [new branch] gh/nikitaved/8/head -> origin/gh/nikitaved/8/head 2025-11-03T16:58:01.8110923Z * [new branch] gh/nikitaved/8/orig -> origin/gh/nikitaved/8/orig 2025-11-03T16:58:01.8113667Z * [new branch] gh/oulgen/1/base -> origin/gh/oulgen/1/base 2025-11-03T16:58:01.8115356Z * [new branch] gh/oulgen/1/head -> origin/gh/oulgen/1/head 2025-11-03T16:58:01.8117053Z * [new branch] gh/oulgen/1/orig -> origin/gh/oulgen/1/orig 2025-11-03T16:58:01.8119269Z * [new branch] gh/oulgen/2/base -> origin/gh/oulgen/2/base 2025-11-03T16:58:01.8120901Z * [new branch] gh/oulgen/2/head -> origin/gh/oulgen/2/head 2025-11-03T16:58:01.8123004Z * [new branch] gh/oulgen/2/orig -> origin/gh/oulgen/2/orig 2025-11-03T16:58:01.8125240Z * [new branch] gh/oulgen/3/base -> origin/gh/oulgen/3/base 2025-11-03T16:58:01.8126912Z * [new branch] gh/oulgen/3/head -> origin/gh/oulgen/3/head 2025-11-03T16:58:01.8128552Z * [new branch] gh/oulgen/3/orig -> origin/gh/oulgen/3/orig 2025-11-03T16:58:01.8130819Z * [new branch] gh/oulgen/4/base -> origin/gh/oulgen/4/base 2025-11-03T16:58:01.8132584Z * [new branch] gh/oulgen/4/head -> origin/gh/oulgen/4/head 2025-11-03T16:58:01.8134293Z * [new branch] gh/oulgen/4/orig -> origin/gh/oulgen/4/orig 2025-11-03T16:58:01.8136560Z * [new branch] gh/oulgen/5/base -> origin/gh/oulgen/5/base 2025-11-03T16:58:01.8138248Z * [new branch] gh/oulgen/5/head -> origin/gh/oulgen/5/head 2025-11-03T16:58:01.8139937Z * [new branch] gh/oulgen/5/orig -> origin/gh/oulgen/5/orig 2025-11-03T16:58:01.8142362Z * [new branch] gh/oulgen/6/base -> origin/gh/oulgen/6/base 2025-11-03T16:58:01.8144059Z * [new branch] gh/oulgen/6/head -> origin/gh/oulgen/6/head 2025-11-03T16:58:01.8145678Z * [new branch] gh/oulgen/6/orig -> origin/gh/oulgen/6/orig 2025-11-03T16:58:01.8147930Z * [new branch] gh/oulgen/7/base -> origin/gh/oulgen/7/base 2025-11-03T16:58:01.8149588Z * [new branch] gh/oulgen/7/head -> origin/gh/oulgen/7/head 2025-11-03T16:58:01.8151306Z * [new branch] gh/oulgen/7/orig -> origin/gh/oulgen/7/orig 2025-11-03T16:58:01.8153781Z * [new branch] gh/patvig/mtia-serialization -> origin/gh/patvig/mtia-serialization 2025-11-03T16:58:01.8156797Z * [new branch] gh/pearu/108/base -> origin/gh/pearu/108/base 2025-11-03T16:58:01.8158494Z * [new branch] gh/pearu/108/head -> origin/gh/pearu/108/head 2025-11-03T16:58:01.8160453Z * [new branch] gh/pearu/108/orig -> origin/gh/pearu/108/orig 2025-11-03T16:58:01.8162662Z * [new branch] gh/pearu/109/base -> origin/gh/pearu/109/base 2025-11-03T16:58:01.8164299Z * [new branch] gh/pearu/109/head -> origin/gh/pearu/109/head 2025-11-03T16:58:01.8165999Z * [new branch] gh/pearu/109/orig -> origin/gh/pearu/109/orig 2025-11-03T16:58:01.8168311Z * [new branch] gh/pearu/110/base -> origin/gh/pearu/110/base 2025-11-03T16:58:01.8170031Z * [new branch] gh/pearu/110/head -> origin/gh/pearu/110/head 2025-11-03T16:58:01.8171696Z * [new branch] gh/pearu/110/orig -> origin/gh/pearu/110/orig 2025-11-03T16:58:01.8173963Z * [new branch] gh/pearu/111/base -> origin/gh/pearu/111/base 2025-11-03T16:58:01.8175593Z * [new branch] gh/pearu/111/head -> origin/gh/pearu/111/head 2025-11-03T16:58:01.8177489Z * [new branch] gh/pearu/111/orig -> origin/gh/pearu/111/orig 2025-11-03T16:58:01.8179807Z * [new branch] gh/pearu/112/base -> origin/gh/pearu/112/base 2025-11-03T16:58:01.8181534Z * [new branch] gh/pearu/112/head -> origin/gh/pearu/112/head 2025-11-03T16:58:01.8183368Z * [new branch] gh/pearu/112/orig -> origin/gh/pearu/112/orig 2025-11-03T16:58:01.8185527Z * [new branch] gh/pearu/115/base -> origin/gh/pearu/115/base 2025-11-03T16:58:01.8187261Z * [new branch] gh/pearu/115/head -> origin/gh/pearu/115/head 2025-11-03T16:58:01.8188907Z * [new branch] gh/pearu/115/orig -> origin/gh/pearu/115/orig 2025-11-03T16:58:01.8191179Z * [new branch] gh/pearu/116/base -> origin/gh/pearu/116/base 2025-11-03T16:58:01.8192909Z * [new branch] gh/pearu/116/head -> origin/gh/pearu/116/head 2025-11-03T16:58:01.8194551Z * [new branch] gh/pearu/116/orig -> origin/gh/pearu/116/orig 2025-11-03T16:58:01.8196877Z * [new branch] gh/pearu/117/base -> origin/gh/pearu/117/base 2025-11-03T16:58:01.8198575Z * [new branch] gh/pearu/117/head -> origin/gh/pearu/117/head 2025-11-03T16:58:01.8200896Z * [new branch] gh/pearu/117/orig -> origin/gh/pearu/117/orig 2025-11-03T16:58:01.8203129Z * [new branch] gh/pearu/118/base -> origin/gh/pearu/118/base 2025-11-03T16:58:01.8204857Z * [new branch] gh/pearu/118/head -> origin/gh/pearu/118/head 2025-11-03T16:58:01.8206574Z * [new branch] gh/pearu/118/orig -> origin/gh/pearu/118/orig 2025-11-03T16:58:01.8208928Z * [new branch] gh/pearu/119/base -> origin/gh/pearu/119/base 2025-11-03T16:58:01.8210614Z * [new branch] gh/pearu/119/head -> origin/gh/pearu/119/head 2025-11-03T16:58:01.8212313Z * [new branch] gh/pearu/119/orig -> origin/gh/pearu/119/orig 2025-11-03T16:58:01.8214807Z * [new branch] gh/pearu/120/base -> origin/gh/pearu/120/base 2025-11-03T16:58:01.8216478Z * [new branch] gh/pearu/120/head -> origin/gh/pearu/120/head 2025-11-03T16:58:01.8218603Z * [new branch] gh/pearu/120/orig -> origin/gh/pearu/120/orig 2025-11-03T16:58:01.8220915Z * [new branch] gh/pearu/121/base -> origin/gh/pearu/121/base 2025-11-03T16:58:01.8222993Z * [new branch] gh/pearu/121/head -> origin/gh/pearu/121/head 2025-11-03T16:58:01.8224695Z * [new branch] gh/pearu/121/orig -> origin/gh/pearu/121/orig 2025-11-03T16:58:01.8226958Z * [new branch] gh/pearu/122/base -> origin/gh/pearu/122/base 2025-11-03T16:58:01.8228669Z * [new branch] gh/pearu/122/head -> origin/gh/pearu/122/head 2025-11-03T16:58:01.8230494Z * [new branch] gh/pearu/122/orig -> origin/gh/pearu/122/orig 2025-11-03T16:58:01.8232711Z * [new branch] gh/pearu/138/base -> origin/gh/pearu/138/base 2025-11-03T16:58:01.8234327Z * [new branch] gh/pearu/138/head -> origin/gh/pearu/138/head 2025-11-03T16:58:01.8236020Z * [new branch] gh/pearu/138/orig -> origin/gh/pearu/138/orig 2025-11-03T16:58:01.8238299Z * [new branch] gh/pearu/139/base -> origin/gh/pearu/139/base 2025-11-03T16:58:01.8240001Z * [new branch] gh/pearu/139/head -> origin/gh/pearu/139/head 2025-11-03T16:58:01.8241687Z * [new branch] gh/pearu/139/orig -> origin/gh/pearu/139/orig 2025-11-03T16:58:01.8243928Z * [new branch] gh/pearu/140/base -> origin/gh/pearu/140/base 2025-11-03T16:58:01.8245626Z * [new branch] gh/pearu/140/head -> origin/gh/pearu/140/head 2025-11-03T16:58:01.8247386Z * [new branch] gh/pearu/140/orig -> origin/gh/pearu/140/orig 2025-11-03T16:58:01.8250183Z * [new branch] gh/pearu/141/base -> origin/gh/pearu/141/base 2025-11-03T16:58:01.8251932Z * [new branch] gh/pearu/141/head -> origin/gh/pearu/141/head 2025-11-03T16:58:01.8253588Z * [new branch] gh/pearu/141/orig -> origin/gh/pearu/141/orig 2025-11-03T16:58:01.8255945Z * [new branch] gh/pearu/142/base -> origin/gh/pearu/142/base 2025-11-03T16:58:01.8257691Z * [new branch] gh/pearu/142/head -> origin/gh/pearu/142/head 2025-11-03T16:58:01.8259366Z * [new branch] gh/pearu/142/orig -> origin/gh/pearu/142/orig 2025-11-03T16:58:01.8261898Z * [new branch] gh/pearu/143/base -> origin/gh/pearu/143/base 2025-11-03T16:58:01.8263687Z * [new branch] gh/pearu/143/head -> origin/gh/pearu/143/head 2025-11-03T16:58:01.8265423Z * [new branch] gh/pearu/143/orig -> origin/gh/pearu/143/orig 2025-11-03T16:58:01.8267767Z * [new branch] gh/pearu/144/base -> origin/gh/pearu/144/base 2025-11-03T16:58:01.8269441Z * [new branch] gh/pearu/144/head -> origin/gh/pearu/144/head 2025-11-03T16:58:01.8271217Z * [new branch] gh/pearu/144/orig -> origin/gh/pearu/144/orig 2025-11-03T16:58:01.8273509Z * [new branch] gh/pearu/145/base -> origin/gh/pearu/145/base 2025-11-03T16:58:01.8275610Z * [new branch] gh/pearu/145/head -> origin/gh/pearu/145/head 2025-11-03T16:58:01.8277285Z * [new branch] gh/pearu/145/orig -> origin/gh/pearu/145/orig 2025-11-03T16:58:01.8279683Z * [new branch] gh/pearu/146/base -> origin/gh/pearu/146/base 2025-11-03T16:58:01.8281349Z * [new branch] gh/pearu/146/head -> origin/gh/pearu/146/head 2025-11-03T16:58:01.8283695Z * [new branch] gh/pearu/147/base -> origin/gh/pearu/147/base 2025-11-03T16:58:01.8285473Z * [new branch] gh/pearu/147/head -> origin/gh/pearu/147/head 2025-11-03T16:58:01.8287135Z * [new branch] gh/pearu/147/orig -> origin/gh/pearu/147/orig 2025-11-03T16:58:01.8289959Z * [new branch] gh/pearu/148/base -> origin/gh/pearu/148/base 2025-11-03T16:58:01.8291621Z * [new branch] gh/pearu/148/head -> origin/gh/pearu/148/head 2025-11-03T16:58:01.8293280Z * [new branch] gh/pearu/148/orig -> origin/gh/pearu/148/orig 2025-11-03T16:58:01.8295626Z * [new branch] gh/pearu/149/base -> origin/gh/pearu/149/base 2025-11-03T16:58:01.8297305Z * [new branch] gh/pearu/149/head -> origin/gh/pearu/149/head 2025-11-03T16:58:01.8298985Z * [new branch] gh/pearu/149/orig -> origin/gh/pearu/149/orig 2025-11-03T16:58:01.8301611Z * [new branch] gh/pearu/150/base -> origin/gh/pearu/150/base 2025-11-03T16:58:01.8303369Z * [new branch] gh/pearu/150/head -> origin/gh/pearu/150/head 2025-11-03T16:58:01.8305080Z * [new branch] gh/pearu/150/orig -> origin/gh/pearu/150/orig 2025-11-03T16:58:01.8307797Z * [new branch] gh/pearu/56/base -> origin/gh/pearu/56/base 2025-11-03T16:58:01.8309792Z * [new branch] gh/pearu/56/head -> origin/gh/pearu/56/head 2025-11-03T16:58:01.8311506Z * [new branch] gh/pearu/56/orig -> origin/gh/pearu/56/orig 2025-11-03T16:58:01.8313963Z * [new branch] gh/pearu/97/base -> origin/gh/pearu/97/base 2025-11-03T16:58:01.8315682Z * [new branch] gh/pearu/97/head -> origin/gh/pearu/97/head 2025-11-03T16:58:01.8317391Z * [new branch] gh/pearu/97/orig -> origin/gh/pearu/97/orig 2025-11-03T16:58:01.8320277Z * [new branch] gh/pianpwk/1/base -> origin/gh/pianpwk/1/base 2025-11-03T16:58:01.8322239Z * [new branch] gh/pianpwk/1/head -> origin/gh/pianpwk/1/head 2025-11-03T16:58:01.8323936Z * [new branch] gh/pianpwk/1/orig -> origin/gh/pianpwk/1/orig 2025-11-03T16:58:01.8326432Z * [new branch] gh/pianpwk/10/base -> origin/gh/pianpwk/10/base 2025-11-03T16:58:01.8328118Z * [new branch] gh/pianpwk/10/head -> origin/gh/pianpwk/10/head 2025-11-03T16:58:01.8329754Z * [new branch] gh/pianpwk/10/orig -> origin/gh/pianpwk/10/orig 2025-11-03T16:58:01.8332443Z * [new branch] gh/pianpwk/11/base -> origin/gh/pianpwk/11/base 2025-11-03T16:58:01.8334419Z * [new branch] gh/pianpwk/11/head -> origin/gh/pianpwk/11/head 2025-11-03T16:58:01.8336256Z * [new branch] gh/pianpwk/11/orig -> origin/gh/pianpwk/11/orig 2025-11-03T16:58:01.8338651Z * [new branch] gh/pianpwk/12/base -> origin/gh/pianpwk/12/base 2025-11-03T16:58:01.8340369Z * [new branch] gh/pianpwk/12/head -> origin/gh/pianpwk/12/head 2025-11-03T16:58:01.8342218Z * [new branch] gh/pianpwk/12/orig -> origin/gh/pianpwk/12/orig 2025-11-03T16:58:01.8345309Z * [new branch] gh/pianpwk/13/base -> origin/gh/pianpwk/13/base 2025-11-03T16:58:01.8346916Z * [new branch] gh/pianpwk/13/head -> origin/gh/pianpwk/13/head 2025-11-03T16:58:01.8348600Z * [new branch] gh/pianpwk/13/orig -> origin/gh/pianpwk/13/orig 2025-11-03T16:58:01.8351384Z * [new branch] gh/pianpwk/14/base -> origin/gh/pianpwk/14/base 2025-11-03T16:58:01.8353338Z * [new branch] gh/pianpwk/14/head -> origin/gh/pianpwk/14/head 2025-11-03T16:58:01.8355052Z * [new branch] gh/pianpwk/14/orig -> origin/gh/pianpwk/14/orig 2025-11-03T16:58:01.8357317Z * [new branch] gh/pianpwk/15/base -> origin/gh/pianpwk/15/base 2025-11-03T16:58:01.8359035Z * [new branch] gh/pianpwk/15/head -> origin/gh/pianpwk/15/head 2025-11-03T16:58:01.8360803Z * [new branch] gh/pianpwk/15/orig -> origin/gh/pianpwk/15/orig 2025-11-03T16:58:01.8363406Z * [new branch] gh/pianpwk/16/base -> origin/gh/pianpwk/16/base 2025-11-03T16:58:01.8365065Z * [new branch] gh/pianpwk/16/head -> origin/gh/pianpwk/16/head 2025-11-03T16:58:01.8366826Z * [new branch] gh/pianpwk/16/orig -> origin/gh/pianpwk/16/orig 2025-11-03T16:58:01.8369412Z * [new branch] gh/pianpwk/17/base -> origin/gh/pianpwk/17/base 2025-11-03T16:58:01.8371420Z * [new branch] gh/pianpwk/17/head -> origin/gh/pianpwk/17/head 2025-11-03T16:58:01.8373200Z * [new branch] gh/pianpwk/17/orig -> origin/gh/pianpwk/17/orig 2025-11-03T16:58:01.8376110Z * [new branch] gh/pianpwk/18/base -> origin/gh/pianpwk/18/base 2025-11-03T16:58:01.8377664Z * [new branch] gh/pianpwk/18/head -> origin/gh/pianpwk/18/head 2025-11-03T16:58:01.8379779Z * [new branch] gh/pianpwk/19/base -> origin/gh/pianpwk/19/base 2025-11-03T16:58:01.8381602Z * [new branch] gh/pianpwk/19/head -> origin/gh/pianpwk/19/head 2025-11-03T16:58:01.8383694Z * [new branch] gh/pianpwk/19/orig -> origin/gh/pianpwk/19/orig 2025-11-03T16:58:01.8385923Z * [new branch] gh/pianpwk/2/base -> origin/gh/pianpwk/2/base 2025-11-03T16:58:01.8387592Z * [new branch] gh/pianpwk/2/head -> origin/gh/pianpwk/2/head 2025-11-03T16:58:01.8389381Z * [new branch] gh/pianpwk/2/orig -> origin/gh/pianpwk/2/orig 2025-11-03T16:58:01.8391701Z * [new branch] gh/pianpwk/20/base -> origin/gh/pianpwk/20/base 2025-11-03T16:58:01.8393330Z * [new branch] gh/pianpwk/20/head -> origin/gh/pianpwk/20/head 2025-11-03T16:58:01.8395081Z * [new branch] gh/pianpwk/20/orig -> origin/gh/pianpwk/20/orig 2025-11-03T16:58:01.8397285Z * [new branch] gh/pianpwk/21/base -> origin/gh/pianpwk/21/base 2025-11-03T16:58:01.8398904Z * [new branch] gh/pianpwk/21/head -> origin/gh/pianpwk/21/head 2025-11-03T16:58:01.8401036Z * [new branch] gh/pianpwk/22/base -> origin/gh/pianpwk/22/base 2025-11-03T16:58:01.8402822Z * [new branch] gh/pianpwk/22/head -> origin/gh/pianpwk/22/head 2025-11-03T16:58:01.8404467Z * [new branch] gh/pianpwk/22/orig -> origin/gh/pianpwk/22/orig 2025-11-03T16:58:01.8406990Z * [new branch] gh/pianpwk/23/base -> origin/gh/pianpwk/23/base 2025-11-03T16:58:01.8408758Z * [new branch] gh/pianpwk/23/head -> origin/gh/pianpwk/23/head 2025-11-03T16:58:01.8410437Z * [new branch] gh/pianpwk/23/orig -> origin/gh/pianpwk/23/orig 2025-11-03T16:58:01.8412758Z * [new branch] gh/pianpwk/24/base -> origin/gh/pianpwk/24/base 2025-11-03T16:58:01.8414565Z * [new branch] gh/pianpwk/24/head -> origin/gh/pianpwk/24/head 2025-11-03T16:58:01.8416176Z * [new branch] gh/pianpwk/24/orig -> origin/gh/pianpwk/24/orig 2025-11-03T16:58:01.8418424Z * [new branch] gh/pianpwk/25/base -> origin/gh/pianpwk/25/base 2025-11-03T16:58:01.8420101Z * [new branch] gh/pianpwk/25/head -> origin/gh/pianpwk/25/head 2025-11-03T16:58:01.8421903Z * [new branch] gh/pianpwk/25/orig -> origin/gh/pianpwk/25/orig 2025-11-03T16:58:01.8426402Z * [new branch] gh/pianpwk/26/base -> origin/gh/pianpwk/26/base 2025-11-03T16:58:01.8428136Z * [new branch] gh/pianpwk/26/head -> origin/gh/pianpwk/26/head 2025-11-03T16:58:01.8430265Z * [new branch] gh/pianpwk/26/orig -> origin/gh/pianpwk/26/orig 2025-11-03T16:58:01.8432735Z * [new branch] gh/pianpwk/27/base -> origin/gh/pianpwk/27/base 2025-11-03T16:58:01.8434464Z * [new branch] gh/pianpwk/27/head -> origin/gh/pianpwk/27/head 2025-11-03T16:58:01.8436160Z * [new branch] gh/pianpwk/27/orig -> origin/gh/pianpwk/27/orig 2025-11-03T16:58:01.8438676Z * [new branch] gh/pianpwk/28/base -> origin/gh/pianpwk/28/base 2025-11-03T16:58:01.8440368Z * [new branch] gh/pianpwk/28/head -> origin/gh/pianpwk/28/head 2025-11-03T16:58:01.8441967Z * [new branch] gh/pianpwk/28/orig -> origin/gh/pianpwk/28/orig 2025-11-03T16:58:01.8444339Z * [new branch] gh/pianpwk/3/base -> origin/gh/pianpwk/3/base 2025-11-03T16:58:01.8446024Z * [new branch] gh/pianpwk/3/head -> origin/gh/pianpwk/3/head 2025-11-03T16:58:01.8447851Z * [new branch] gh/pianpwk/3/orig -> origin/gh/pianpwk/3/orig 2025-11-03T16:58:01.8450118Z * [new branch] gh/pianpwk/4/base -> origin/gh/pianpwk/4/base 2025-11-03T16:58:01.8451736Z * [new branch] gh/pianpwk/4/head -> origin/gh/pianpwk/4/head 2025-11-03T16:58:01.8453413Z * [new branch] gh/pianpwk/4/orig -> origin/gh/pianpwk/4/orig 2025-11-03T16:58:01.8455694Z * [new branch] gh/pianpwk/5/base -> origin/gh/pianpwk/5/base 2025-11-03T16:58:01.8457400Z * [new branch] gh/pianpwk/5/head -> origin/gh/pianpwk/5/head 2025-11-03T16:58:01.8459491Z * [new branch] gh/pianpwk/5/orig -> origin/gh/pianpwk/5/orig 2025-11-03T16:58:01.8462104Z * [new branch] gh/pianpwk/6/base -> origin/gh/pianpwk/6/base 2025-11-03T16:58:01.8463750Z * [new branch] gh/pianpwk/6/head -> origin/gh/pianpwk/6/head 2025-11-03T16:58:01.8465489Z * [new branch] gh/pianpwk/6/orig -> origin/gh/pianpwk/6/orig 2025-11-03T16:58:01.8468818Z * [new branch] gh/pianpwk/7/base -> origin/gh/pianpwk/7/base 2025-11-03T16:58:01.8470525Z * [new branch] gh/pianpwk/7/head -> origin/gh/pianpwk/7/head 2025-11-03T16:58:01.8472289Z * [new branch] gh/pianpwk/7/orig -> origin/gh/pianpwk/7/orig 2025-11-03T16:58:01.8474564Z * [new branch] gh/pianpwk/8/base -> origin/gh/pianpwk/8/base 2025-11-03T16:58:01.8476320Z * [new branch] gh/pianpwk/8/head -> origin/gh/pianpwk/8/head 2025-11-03T16:58:01.8477982Z * [new branch] gh/pianpwk/8/orig -> origin/gh/pianpwk/8/orig 2025-11-03T16:58:01.8480659Z * [new branch] gh/pianpwk/9/base -> origin/gh/pianpwk/9/base 2025-11-03T16:58:01.8482286Z * [new branch] gh/pianpwk/9/head -> origin/gh/pianpwk/9/head 2025-11-03T16:58:01.8484018Z * [new branch] gh/pianpwk/9/orig -> origin/gh/pianpwk/9/orig 2025-11-03T16:58:01.8486853Z * [new branch] gh/qqaatw/29/head -> origin/gh/qqaatw/29/head 2025-11-03T16:58:01.8489576Z * [new branch] gh/rec/141/base -> origin/gh/rec/141/base 2025-11-03T16:58:01.8491281Z * [new branch] gh/rec/141/head -> origin/gh/rec/141/head 2025-11-03T16:58:01.8493651Z * [new branch] gh/rec/153/base -> origin/gh/rec/153/base 2025-11-03T16:58:01.8495337Z * [new branch] gh/rec/153/head -> origin/gh/rec/153/head 2025-11-03T16:58:01.8497014Z * [new branch] gh/rec/153/orig -> origin/gh/rec/153/orig 2025-11-03T16:58:01.8499317Z * [new branch] gh/rec/154/base -> origin/gh/rec/154/base 2025-11-03T16:58:01.8501041Z * [new branch] gh/rec/154/head -> origin/gh/rec/154/head 2025-11-03T16:58:01.8502846Z * [new branch] gh/rec/154/orig -> origin/gh/rec/154/orig 2025-11-03T16:58:01.8505141Z * [new branch] gh/rec/164/base -> origin/gh/rec/164/base 2025-11-03T16:58:01.8506903Z * [new branch] gh/rec/164/head -> origin/gh/rec/164/head 2025-11-03T16:58:01.8508693Z * [new branch] gh/rec/164/orig -> origin/gh/rec/164/orig 2025-11-03T16:58:01.8510939Z * [new branch] gh/rec/166/base -> origin/gh/rec/166/base 2025-11-03T16:58:01.8512606Z * [new branch] gh/rec/166/head -> origin/gh/rec/166/head 2025-11-03T16:58:01.8514431Z * [new branch] gh/rec/166/orig -> origin/gh/rec/166/orig 2025-11-03T16:58:01.8517326Z * [new branch] gh/robert-hardwick/3/base -> origin/gh/robert-hardwick/3/base 2025-11-03T16:58:01.8519333Z * [new branch] gh/robert-hardwick/3/head -> origin/gh/robert-hardwick/3/head 2025-11-03T16:58:01.8521397Z * [new branch] gh/robert-hardwick/3/orig -> origin/gh/robert-hardwick/3/orig 2025-11-03T16:58:01.8523588Z * [new branch] gh/robert-hardwick/4/base -> origin/gh/robert-hardwick/4/base 2025-11-03T16:58:01.8525379Z * [new branch] gh/robert-hardwick/4/head -> origin/gh/robert-hardwick/4/head 2025-11-03T16:58:01.8526987Z * [new branch] gh/robert-hardwick/4/orig -> origin/gh/robert-hardwick/4/orig 2025-11-03T16:58:01.8529343Z * [new branch] gh/robert-hardwick/5/base -> origin/gh/robert-hardwick/5/base 2025-11-03T16:58:01.8531057Z * [new branch] gh/robert-hardwick/5/head -> origin/gh/robert-hardwick/5/head 2025-11-03T16:58:01.8532840Z * [new branch] gh/robert-hardwick/5/orig -> origin/gh/robert-hardwick/5/orig 2025-11-03T16:58:01.8535590Z * [new branch] gh/rtimpe/1/base -> origin/gh/rtimpe/1/base 2025-11-03T16:58:01.8537260Z * [new branch] gh/rtimpe/1/head -> origin/gh/rtimpe/1/head 2025-11-03T16:58:01.8539595Z * [new branch] gh/rtimpe/11/base -> origin/gh/rtimpe/11/base 2025-11-03T16:58:01.8541357Z * [new branch] gh/rtimpe/11/head -> origin/gh/rtimpe/11/head 2025-11-03T16:58:01.8543206Z * [new branch] gh/rtimpe/11/orig -> origin/gh/rtimpe/11/orig 2025-11-03T16:58:01.8545442Z * [new branch] gh/rtimpe/16/base -> origin/gh/rtimpe/16/base 2025-11-03T16:58:01.8547071Z * [new branch] gh/rtimpe/16/head -> origin/gh/rtimpe/16/head 2025-11-03T16:58:01.8549061Z * [new branch] gh/rtimpe/16/orig -> origin/gh/rtimpe/16/orig 2025-11-03T16:58:01.8551323Z * [new branch] gh/rtimpe/19/base -> origin/gh/rtimpe/19/base 2025-11-03T16:58:01.8552909Z * [new branch] gh/rtimpe/19/head -> origin/gh/rtimpe/19/head 2025-11-03T16:58:01.8554590Z * [new branch] gh/rtimpe/19/orig -> origin/gh/rtimpe/19/orig 2025-11-03T16:58:01.8570237Z * [new branch] gh/rtimpe/2/base -> origin/gh/rtimpe/2/base 2025-11-03T16:58:01.8570436Z * [new branch] gh/rtimpe/2/head -> origin/gh/rtimpe/2/head 2025-11-03T16:58:01.8570610Z * [new branch] gh/rtimpe/20/base -> origin/gh/rtimpe/20/base 2025-11-03T16:58:01.8570762Z * [new branch] gh/rtimpe/20/head -> origin/gh/rtimpe/20/head 2025-11-03T16:58:01.8570910Z * [new branch] gh/rtimpe/20/orig -> origin/gh/rtimpe/20/orig 2025-11-03T16:58:01.8571054Z * [new branch] gh/rtimpe/21/base -> origin/gh/rtimpe/21/base 2025-11-03T16:58:01.8571195Z * [new branch] gh/rtimpe/21/head -> origin/gh/rtimpe/21/head 2025-11-03T16:58:01.8571339Z * [new branch] gh/rtimpe/21/orig -> origin/gh/rtimpe/21/orig 2025-11-03T16:58:01.8572068Z * [new branch] gh/rtimpe/3/base -> origin/gh/rtimpe/3/base 2025-11-03T16:58:01.8573936Z * [new branch] gh/rtimpe/3/head -> origin/gh/rtimpe/3/head 2025-11-03T16:58:01.8576213Z * [new branch] gh/rtimpe/4/base -> origin/gh/rtimpe/4/base 2025-11-03T16:58:01.8578455Z * [new branch] gh/rtimpe/4/head -> origin/gh/rtimpe/4/head 2025-11-03T16:58:01.8581268Z * [new branch] gh/ruisizhang123/1/base -> origin/gh/ruisizhang123/1/base 2025-11-03T16:58:01.8583086Z * [new branch] gh/ruisizhang123/1/head -> origin/gh/ruisizhang123/1/head 2025-11-03T16:58:01.8584844Z * [new branch] gh/ruisizhang123/1/orig -> origin/gh/ruisizhang123/1/orig 2025-11-03T16:58:01.8587112Z * [new branch] gh/ruisizhang123/4/base -> origin/gh/ruisizhang123/4/base 2025-11-03T16:58:01.8588816Z * [new branch] gh/ruisizhang123/4/head -> origin/gh/ruisizhang123/4/head 2025-11-03T16:58:01.8590774Z * [new branch] gh/ruisizhang123/4/orig -> origin/gh/ruisizhang123/4/orig 2025-11-03T16:58:01.8592945Z * [new branch] gh/ruisizhang123/5/base -> origin/gh/ruisizhang123/5/base 2025-11-03T16:58:01.8594655Z * [new branch] gh/ruisizhang123/5/head -> origin/gh/ruisizhang123/5/head 2025-11-03T16:58:01.8596439Z * [new branch] gh/ruisizhang123/5/orig -> origin/gh/ruisizhang123/5/orig 2025-11-03T16:58:01.8598743Z * [new branch] gh/ruisizhang123/6/base -> origin/gh/ruisizhang123/6/base 2025-11-03T16:58:01.8600496Z * [new branch] gh/ruisizhang123/6/head -> origin/gh/ruisizhang123/6/head 2025-11-03T16:58:01.8602327Z * [new branch] gh/ruisizhang123/6/orig -> origin/gh/ruisizhang123/6/orig 2025-11-03T16:58:01.8604669Z * [new branch] gh/ruisizhang123/7/base -> origin/gh/ruisizhang123/7/base 2025-11-03T16:58:01.8606324Z * [new branch] gh/ruisizhang123/7/head -> origin/gh/ruisizhang123/7/head 2025-11-03T16:58:01.8607999Z * [new branch] gh/ruisizhang123/7/orig -> origin/gh/ruisizhang123/7/orig 2025-11-03T16:58:01.8610249Z * [new branch] gh/ruisizhang123/8/base -> origin/gh/ruisizhang123/8/base 2025-11-03T16:58:01.8611906Z * [new branch] gh/ruisizhang123/8/head -> origin/gh/ruisizhang123/8/head 2025-11-03T16:58:01.8613511Z * [new branch] gh/ruisizhang123/8/orig -> origin/gh/ruisizhang123/8/orig 2025-11-03T16:58:01.8615879Z * [new branch] gh/ruisizhang123/9/base -> origin/gh/ruisizhang123/9/base 2025-11-03T16:58:01.8617608Z * [new branch] gh/ruisizhang123/9/head -> origin/gh/ruisizhang123/9/head 2025-11-03T16:58:01.8619335Z * [new branch] gh/ruisizhang123/9/orig -> origin/gh/ruisizhang123/9/orig 2025-11-03T16:58:01.8622448Z * [new branch] gh/seemethere/43/base -> origin/gh/seemethere/43/base 2025-11-03T16:58:01.8624105Z * [new branch] gh/seemethere/43/head -> origin/gh/seemethere/43/head 2025-11-03T16:58:01.8626021Z * [new branch] gh/seemethere/43/orig -> origin/gh/seemethere/43/orig 2025-11-03T16:58:01.8628613Z * [new branch] gh/seemethere/44/base -> origin/gh/seemethere/44/base 2025-11-03T16:58:01.8630254Z * [new branch] gh/seemethere/44/head -> origin/gh/seemethere/44/head 2025-11-03T16:58:01.8632453Z * [new branch] gh/seemethere/44/orig -> origin/gh/seemethere/44/orig 2025-11-03T16:58:01.8634761Z * [new branch] gh/seemethere/48/base -> origin/gh/seemethere/48/base 2025-11-03T16:58:01.8636428Z * [new branch] gh/seemethere/48/head -> origin/gh/seemethere/48/head 2025-11-03T16:58:01.8638083Z * [new branch] gh/seemethere/48/orig -> origin/gh/seemethere/48/orig 2025-11-03T16:58:01.8640363Z * [new branch] gh/seemethere/49/base -> origin/gh/seemethere/49/base 2025-11-03T16:58:01.8642220Z * [new branch] gh/seemethere/49/head -> origin/gh/seemethere/49/head 2025-11-03T16:58:01.8643874Z * [new branch] gh/seemethere/49/orig -> origin/gh/seemethere/49/orig 2025-11-03T16:58:01.8646167Z * [new branch] gh/seemethere/52/base -> origin/gh/seemethere/52/base 2025-11-03T16:58:01.8647850Z * [new branch] gh/seemethere/52/head -> origin/gh/seemethere/52/head 2025-11-03T16:58:01.8649545Z * [new branch] gh/seemethere/52/orig -> origin/gh/seemethere/52/orig 2025-11-03T16:58:01.8651814Z * [new branch] gh/seemethere/53/base -> origin/gh/seemethere/53/base 2025-11-03T16:58:01.8653506Z * [new branch] gh/seemethere/53/head -> origin/gh/seemethere/53/head 2025-11-03T16:58:01.8655636Z * [new branch] gh/seemethere/53/orig -> origin/gh/seemethere/53/orig 2025-11-03T16:58:01.8658247Z * [new branch] gh/seemethere/54/base -> origin/gh/seemethere/54/base 2025-11-03T16:58:01.8659814Z * [new branch] gh/seemethere/54/head -> origin/gh/seemethere/54/head 2025-11-03T16:58:01.8661527Z * [new branch] gh/seemethere/54/orig -> origin/gh/seemethere/54/orig 2025-11-03T16:58:01.8663805Z * [new branch] gh/seemethere/55/base -> origin/gh/seemethere/55/base 2025-11-03T16:58:01.8665475Z * [new branch] gh/seemethere/55/head -> origin/gh/seemethere/55/head 2025-11-03T16:58:01.8667171Z * [new branch] gh/seemethere/55/orig -> origin/gh/seemethere/55/orig 2025-11-03T16:58:01.8669397Z * [new branch] gh/seemethere/59/base -> origin/gh/seemethere/59/base 2025-11-03T16:58:01.8671123Z * [new branch] gh/seemethere/59/head -> origin/gh/seemethere/59/head 2025-11-03T16:58:01.8672881Z * [new branch] gh/seemethere/59/orig -> origin/gh/seemethere/59/orig 2025-11-03T16:58:01.8675651Z * [new branch] gh/seemethere/62/base -> origin/gh/seemethere/62/base 2025-11-03T16:58:01.8677346Z * [new branch] gh/seemethere/62/head -> origin/gh/seemethere/62/head 2025-11-03T16:58:01.8678996Z * [new branch] gh/seemethere/62/orig -> origin/gh/seemethere/62/orig 2025-11-03T16:58:01.8681229Z * [new branch] gh/seemethere/63/base -> origin/gh/seemethere/63/base 2025-11-03T16:58:01.8682868Z * [new branch] gh/seemethere/63/head -> origin/gh/seemethere/63/head 2025-11-03T16:58:01.8684578Z * [new branch] gh/seemethere/63/orig -> origin/gh/seemethere/63/orig 2025-11-03T16:58:01.8686928Z * [new branch] gh/seemethere/71/base -> origin/gh/seemethere/71/base 2025-11-03T16:58:01.8688656Z * [new branch] gh/seemethere/71/head -> origin/gh/seemethere/71/head 2025-11-03T16:58:01.8690350Z * [new branch] gh/seemethere/71/orig -> origin/gh/seemethere/71/orig 2025-11-03T16:58:01.8692686Z * [new branch] gh/seemethere/72/base -> origin/gh/seemethere/72/base 2025-11-03T16:58:01.8694385Z * [new branch] gh/seemethere/72/head -> origin/gh/seemethere/72/head 2025-11-03T16:58:01.8696061Z * [new branch] gh/seemethere/72/orig -> origin/gh/seemethere/72/orig 2025-11-03T16:58:01.8698263Z * [new branch] gh/seemethere/73/base -> origin/gh/seemethere/73/base 2025-11-03T16:58:01.8700037Z * [new branch] gh/seemethere/73/head -> origin/gh/seemethere/73/head 2025-11-03T16:58:01.8701852Z * [new branch] gh/seemethere/73/orig -> origin/gh/seemethere/73/orig 2025-11-03T16:58:01.8704708Z * [new branch] gh/shunting314/145/base -> origin/gh/shunting314/145/base 2025-11-03T16:58:01.8706452Z * [new branch] gh/shunting314/145/head -> origin/gh/shunting314/145/head 2025-11-03T16:58:01.8708260Z * [new branch] gh/shunting314/145/orig -> origin/gh/shunting314/145/orig 2025-11-03T16:58:01.8710845Z * [new branch] gh/shunting314/176/base -> origin/gh/shunting314/176/base 2025-11-03T16:58:01.8712897Z * [new branch] gh/shunting314/176/head -> origin/gh/shunting314/176/head 2025-11-03T16:58:01.8714654Z * [new branch] gh/shunting314/176/orig -> origin/gh/shunting314/176/orig 2025-11-03T16:58:01.8716980Z * [new branch] gh/shunting314/217/base -> origin/gh/shunting314/217/base 2025-11-03T16:58:01.8718648Z * [new branch] gh/shunting314/217/head -> origin/gh/shunting314/217/head 2025-11-03T16:58:01.8720268Z * [new branch] gh/shunting314/217/orig -> origin/gh/shunting314/217/orig 2025-11-03T16:58:01.8723003Z * [new branch] gh/shunting314/228/base -> origin/gh/shunting314/228/base 2025-11-03T16:58:01.8724692Z * [new branch] gh/shunting314/228/head -> origin/gh/shunting314/228/head 2025-11-03T16:58:01.8726562Z * [new branch] gh/shunting314/228/orig -> origin/gh/shunting314/228/orig 2025-11-03T16:58:01.8728737Z * [new branch] gh/shunting314/229/base -> origin/gh/shunting314/229/base 2025-11-03T16:58:01.8730554Z * [new branch] gh/shunting314/229/head -> origin/gh/shunting314/229/head 2025-11-03T16:58:01.8732288Z * [new branch] gh/shunting314/229/orig -> origin/gh/shunting314/229/orig 2025-11-03T16:58:01.8734672Z * [new branch] gh/shunting314/230/base -> origin/gh/shunting314/230/base 2025-11-03T16:58:01.8736430Z * [new branch] gh/shunting314/230/head -> origin/gh/shunting314/230/head 2025-11-03T16:58:01.8738142Z * [new branch] gh/shunting314/230/orig -> origin/gh/shunting314/230/orig 2025-11-03T16:58:01.8740413Z * [new branch] gh/shunting314/231/base -> origin/gh/shunting314/231/base 2025-11-03T16:58:01.8742365Z * [new branch] gh/shunting314/231/head -> origin/gh/shunting314/231/head 2025-11-03T16:58:01.8744033Z * [new branch] gh/shunting314/231/orig -> origin/gh/shunting314/231/orig 2025-11-03T16:58:01.8746294Z * [new branch] gh/shunting314/232/base -> origin/gh/shunting314/232/base 2025-11-03T16:58:01.8748092Z * [new branch] gh/shunting314/232/head -> origin/gh/shunting314/232/head 2025-11-03T16:58:01.8749727Z * [new branch] gh/shunting314/232/orig -> origin/gh/shunting314/232/orig 2025-11-03T16:58:01.8751845Z * [new branch] gh/shunting314/233/base -> origin/gh/shunting314/233/base 2025-11-03T16:58:01.8753496Z * [new branch] gh/shunting314/233/head -> origin/gh/shunting314/233/head 2025-11-03T16:58:01.8755137Z * [new branch] gh/shunting314/233/orig -> origin/gh/shunting314/233/orig 2025-11-03T16:58:01.8757747Z * [new branch] gh/shunting314/234/base -> origin/gh/shunting314/234/base 2025-11-03T16:58:01.8759608Z * [new branch] gh/shunting314/234/head -> origin/gh/shunting314/234/head 2025-11-03T16:58:01.8761449Z * [new branch] gh/shunting314/234/orig -> origin/gh/shunting314/234/orig 2025-11-03T16:58:01.8763729Z * [new branch] gh/shunting314/235/base -> origin/gh/shunting314/235/base 2025-11-03T16:58:01.8765459Z * [new branch] gh/shunting314/235/head -> origin/gh/shunting314/235/head 2025-11-03T16:58:01.8767128Z * [new branch] gh/shunting314/235/orig -> origin/gh/shunting314/235/orig 2025-11-03T16:58:01.8769491Z * [new branch] gh/shunting314/236/base -> origin/gh/shunting314/236/base 2025-11-03T16:58:01.8771262Z * [new branch] gh/shunting314/236/head -> origin/gh/shunting314/236/head 2025-11-03T16:58:01.8772904Z * [new branch] gh/shunting314/236/orig -> origin/gh/shunting314/236/orig 2025-11-03T16:58:01.8775512Z * [new branch] gh/shunting314/237/base -> origin/gh/shunting314/237/base 2025-11-03T16:58:01.8777608Z * [new branch] gh/shunting314/237/head -> origin/gh/shunting314/237/head 2025-11-03T16:58:01.8779418Z * [new branch] gh/shunting314/237/orig -> origin/gh/shunting314/237/orig 2025-11-03T16:58:01.8781908Z * [new branch] gh/shunting314/238/base -> origin/gh/shunting314/238/base 2025-11-03T16:58:01.8783719Z * [new branch] gh/shunting314/238/head -> origin/gh/shunting314/238/head 2025-11-03T16:58:01.8785479Z * [new branch] gh/shunting314/238/orig -> origin/gh/shunting314/238/orig 2025-11-03T16:58:01.8787961Z * [new branch] gh/shunting314/239/base -> origin/gh/shunting314/239/base 2025-11-03T16:58:01.8789655Z * [new branch] gh/shunting314/239/head -> origin/gh/shunting314/239/head 2025-11-03T16:58:01.8791360Z * [new branch] gh/shunting314/239/orig -> origin/gh/shunting314/239/orig 2025-11-03T16:58:01.8793750Z * [new branch] gh/shunting314/240/base -> origin/gh/shunting314/240/base 2025-11-03T16:58:01.8795470Z * [new branch] gh/shunting314/240/head -> origin/gh/shunting314/240/head 2025-11-03T16:58:01.8797136Z * [new branch] gh/shunting314/240/orig -> origin/gh/shunting314/240/orig 2025-11-03T16:58:01.8799815Z * [new branch] gh/shunting314/241/base -> origin/gh/shunting314/241/base 2025-11-03T16:58:01.8801641Z * [new branch] gh/shunting314/241/head -> origin/gh/shunting314/241/head 2025-11-03T16:58:01.8803466Z * [new branch] gh/shunting314/241/orig -> origin/gh/shunting314/241/orig 2025-11-03T16:58:01.8806044Z * [new branch] gh/shunting314/242/base -> origin/gh/shunting314/242/base 2025-11-03T16:58:01.8807778Z * [new branch] gh/shunting314/242/head -> origin/gh/shunting314/242/head 2025-11-03T16:58:01.8809523Z * [new branch] gh/shunting314/242/orig -> origin/gh/shunting314/242/orig 2025-11-03T16:58:01.8811826Z * [new branch] gh/shunting314/243/base -> origin/gh/shunting314/243/base 2025-11-03T16:58:01.8813487Z * [new branch] gh/shunting314/243/head -> origin/gh/shunting314/243/head 2025-11-03T16:58:01.8815232Z * [new branch] gh/shunting314/243/orig -> origin/gh/shunting314/243/orig 2025-11-03T16:58:01.8817727Z * [new branch] gh/shunting314/244/base -> origin/gh/shunting314/244/base 2025-11-03T16:58:01.8819397Z * [new branch] gh/shunting314/244/head -> origin/gh/shunting314/244/head 2025-11-03T16:58:01.8821234Z * [new branch] gh/shunting314/244/orig -> origin/gh/shunting314/244/orig 2025-11-03T16:58:01.8825449Z * [new branch] gh/shunting314/245/base -> origin/gh/shunting314/245/base 2025-11-03T16:58:01.8827217Z * [new branch] gh/shunting314/245/head -> origin/gh/shunting314/245/head 2025-11-03T16:58:01.8828969Z * [new branch] gh/shunting314/245/orig -> origin/gh/shunting314/245/orig 2025-11-03T16:58:01.8831502Z * [new branch] gh/shunting314/246/base -> origin/gh/shunting314/246/base 2025-11-03T16:58:01.8833433Z * [new branch] gh/shunting314/246/head -> origin/gh/shunting314/246/head 2025-11-03T16:58:01.8835089Z * [new branch] gh/shunting314/246/orig -> origin/gh/shunting314/246/orig 2025-11-03T16:58:01.8837456Z * [new branch] gh/shunting314/247/base -> origin/gh/shunting314/247/base 2025-11-03T16:58:01.8839174Z * [new branch] gh/shunting314/247/head -> origin/gh/shunting314/247/head 2025-11-03T16:58:01.8840913Z * [new branch] gh/shunting314/247/orig -> origin/gh/shunting314/247/orig 2025-11-03T16:58:01.8843464Z * [new branch] gh/shunting314/248/base -> origin/gh/shunting314/248/base 2025-11-03T16:58:01.8845160Z * [new branch] gh/shunting314/248/head -> origin/gh/shunting314/248/head 2025-11-03T16:58:01.8846822Z * [new branch] gh/shunting314/248/orig -> origin/gh/shunting314/248/orig 2025-11-03T16:58:01.8849258Z * [new branch] gh/shunting314/249/base -> origin/gh/shunting314/249/base 2025-11-03T16:58:01.8851049Z * [new branch] gh/shunting314/249/head -> origin/gh/shunting314/249/head 2025-11-03T16:58:01.8852855Z * [new branch] gh/shunting314/249/orig -> origin/gh/shunting314/249/orig 2025-11-03T16:58:01.8855366Z * [new branch] gh/shunting314/250/base -> origin/gh/shunting314/250/base 2025-11-03T16:58:01.8856981Z * [new branch] gh/shunting314/250/head -> origin/gh/shunting314/250/head 2025-11-03T16:58:01.8858764Z * [new branch] gh/shunting314/250/orig -> origin/gh/shunting314/250/orig 2025-11-03T16:58:01.8861406Z * [new branch] gh/shunting314/251/base -> origin/gh/shunting314/251/base 2025-11-03T16:58:01.8863394Z * [new branch] gh/shunting314/251/head -> origin/gh/shunting314/251/head 2025-11-03T16:58:01.8864932Z * [new branch] gh/shunting314/251/orig -> origin/gh/shunting314/251/orig 2025-11-03T16:58:01.8867350Z * [new branch] gh/shunting314/252/base -> origin/gh/shunting314/252/base 2025-11-03T16:58:01.8869075Z * [new branch] gh/shunting314/252/head -> origin/gh/shunting314/252/head 2025-11-03T16:58:01.8870735Z * [new branch] gh/shunting314/252/orig -> origin/gh/shunting314/252/orig 2025-11-03T16:58:01.8873153Z * [new branch] gh/shunting314/253/base -> origin/gh/shunting314/253/base 2025-11-03T16:58:01.8874939Z * [new branch] gh/shunting314/253/head -> origin/gh/shunting314/253/head 2025-11-03T16:58:01.8876599Z * [new branch] gh/shunting314/253/orig -> origin/gh/shunting314/253/orig 2025-11-03T16:58:01.8879258Z * [new branch] gh/shunting314/254/base -> origin/gh/shunting314/254/base 2025-11-03T16:58:01.8881004Z * [new branch] gh/shunting314/254/head -> origin/gh/shunting314/254/head 2025-11-03T16:58:01.8882754Z * [new branch] gh/shunting314/254/orig -> origin/gh/shunting314/254/orig 2025-11-03T16:58:01.8885004Z * [new branch] gh/shunting314/255/base -> origin/gh/shunting314/255/base 2025-11-03T16:58:01.8886712Z * [new branch] gh/shunting314/255/head -> origin/gh/shunting314/255/head 2025-11-03T16:58:01.8888443Z * [new branch] gh/shunting314/255/orig -> origin/gh/shunting314/255/orig 2025-11-03T16:58:01.8891287Z * [new branch] gh/silverguo/1/base -> origin/gh/silverguo/1/base 2025-11-03T16:58:01.8893016Z * [new branch] gh/silverguo/1/head -> origin/gh/silverguo/1/head 2025-11-03T16:58:01.8895177Z * [new branch] gh/silverguo/2/base -> origin/gh/silverguo/2/base 2025-11-03T16:58:01.8896811Z * [new branch] gh/silverguo/2/head -> origin/gh/silverguo/2/head 2025-11-03T16:58:01.8898999Z * [new branch] gh/silverguo/3/base -> origin/gh/silverguo/3/base 2025-11-03T16:58:01.8900770Z * [new branch] gh/silverguo/3/head -> origin/gh/silverguo/3/head 2025-11-03T16:58:01.8903505Z * [new branch] gh/silverguo/4/base -> origin/gh/silverguo/4/base 2025-11-03T16:58:01.8905192Z * [new branch] gh/silverguo/4/head -> origin/gh/silverguo/4/head 2025-11-03T16:58:01.8908009Z * [new branch] gh/sinhaanhsul/1/base -> origin/gh/sinhaanhsul/1/base 2025-11-03T16:58:01.8909720Z * [new branch] gh/sinhaanhsul/1/head -> origin/gh/sinhaanhsul/1/head 2025-11-03T16:58:01.8912490Z * [new branch] gh/slayton58/14/base -> origin/gh/slayton58/14/base 2025-11-03T16:58:01.8914192Z * [new branch] gh/slayton58/14/head -> origin/gh/slayton58/14/head 2025-11-03T16:58:01.8915920Z * [new branch] gh/slayton58/14/orig -> origin/gh/slayton58/14/orig 2025-11-03T16:58:01.8918285Z * [new branch] gh/slayton58/16/base -> origin/gh/slayton58/16/base 2025-11-03T16:58:01.8919875Z * [new branch] gh/slayton58/16/head -> origin/gh/slayton58/16/head 2025-11-03T16:58:01.8921801Z * [new branch] gh/slayton58/16/orig -> origin/gh/slayton58/16/orig 2025-11-03T16:58:01.8924375Z * [new branch] gh/slayton58/17/base -> origin/gh/slayton58/17/base 2025-11-03T16:58:01.8926179Z * [new branch] gh/slayton58/17/head -> origin/gh/slayton58/17/head 2025-11-03T16:58:01.8927912Z * [new branch] gh/slayton58/17/orig -> origin/gh/slayton58/17/orig 2025-11-03T16:58:01.8930383Z * [new branch] gh/slayton58/19/base -> origin/gh/slayton58/19/base 2025-11-03T16:58:01.8932062Z * [new branch] gh/slayton58/19/head -> origin/gh/slayton58/19/head 2025-11-03T16:58:01.8933995Z * [new branch] gh/slayton58/19/orig -> origin/gh/slayton58/19/orig 2025-11-03T16:58:01.8936167Z * [new branch] gh/slayton58/20/base -> origin/gh/slayton58/20/base 2025-11-03T16:58:01.8937856Z * [new branch] gh/slayton58/20/head -> origin/gh/slayton58/20/head 2025-11-03T16:58:01.8939560Z * [new branch] gh/slayton58/20/orig -> origin/gh/slayton58/20/orig 2025-11-03T16:58:01.8941827Z * [new branch] gh/slayton58/21/base -> origin/gh/slayton58/21/base 2025-11-03T16:58:01.8943612Z * [new branch] gh/slayton58/21/head -> origin/gh/slayton58/21/head 2025-11-03T16:58:01.8945436Z * [new branch] gh/slayton58/21/orig -> origin/gh/slayton58/21/orig 2025-11-03T16:58:01.8947595Z * [new branch] gh/slayton58/23/base -> origin/gh/slayton58/23/base 2025-11-03T16:58:01.8949242Z * [new branch] gh/slayton58/23/head -> origin/gh/slayton58/23/head 2025-11-03T16:58:01.8950991Z * [new branch] gh/slayton58/23/orig -> origin/gh/slayton58/23/orig 2025-11-03T16:58:01.8953275Z * [new branch] gh/slayton58/25/base -> origin/gh/slayton58/25/base 2025-11-03T16:58:01.8954899Z * [new branch] gh/slayton58/25/head -> origin/gh/slayton58/25/head 2025-11-03T16:58:01.8956591Z * [new branch] gh/slayton58/25/orig -> origin/gh/slayton58/25/orig 2025-11-03T16:58:01.8958995Z * [new branch] gh/slayton58/26/base -> origin/gh/slayton58/26/base 2025-11-03T16:58:01.8960696Z * [new branch] gh/slayton58/26/head -> origin/gh/slayton58/26/head 2025-11-03T16:58:01.8962366Z * [new branch] gh/slayton58/26/orig -> origin/gh/slayton58/26/orig 2025-11-03T16:58:01.8964783Z * [new branch] gh/slayton58/27/base -> origin/gh/slayton58/27/base 2025-11-03T16:58:01.8966480Z * [new branch] gh/slayton58/27/head -> origin/gh/slayton58/27/head 2025-11-03T16:58:01.8968330Z * [new branch] gh/slayton58/27/orig -> origin/gh/slayton58/27/orig 2025-11-03T16:58:01.8970592Z * [new branch] gh/slayton58/28/base -> origin/gh/slayton58/28/base 2025-11-03T16:58:01.8972224Z * [new branch] gh/slayton58/28/head -> origin/gh/slayton58/28/head 2025-11-03T16:58:01.8973985Z * [new branch] gh/slayton58/28/orig -> origin/gh/slayton58/28/orig 2025-11-03T16:58:01.8976238Z * [new branch] gh/slayton58/29/base -> origin/gh/slayton58/29/base 2025-11-03T16:58:01.8977950Z * [new branch] gh/slayton58/29/head -> origin/gh/slayton58/29/head 2025-11-03T16:58:01.8979751Z * [new branch] gh/slayton58/29/head-jeffdaily-patch-1 -> origin/gh/slayton58/29/head-jeffdaily-patch-1 2025-11-03T16:58:01.8981560Z * [new branch] gh/slayton58/29/head-jeffdaily-patch-2 -> origin/gh/slayton58/29/head-jeffdaily-patch-2 2025-11-03T16:58:01.8983341Z * [new branch] gh/slayton58/29/orig -> origin/gh/slayton58/29/orig 2025-11-03T16:58:01.8985756Z * [new branch] gh/slayton58/30/base -> origin/gh/slayton58/30/base 2025-11-03T16:58:01.8987352Z * [new branch] gh/slayton58/30/head -> origin/gh/slayton58/30/head 2025-11-03T16:58:01.8989161Z * [new branch] gh/slayton58/30/orig -> origin/gh/slayton58/30/orig 2025-11-03T16:58:01.8991653Z * [new branch] gh/slayton58/31/base -> origin/gh/slayton58/31/base 2025-11-03T16:58:01.8993262Z * [new branch] gh/slayton58/31/head -> origin/gh/slayton58/31/head 2025-11-03T16:58:01.8995072Z * [new branch] gh/slayton58/31/orig -> origin/gh/slayton58/31/orig 2025-11-03T16:58:01.8997378Z * [new branch] gh/slayton58/32/base -> origin/gh/slayton58/32/base 2025-11-03T16:58:01.8999087Z * [new branch] gh/slayton58/32/head -> origin/gh/slayton58/32/head 2025-11-03T16:58:01.9000622Z * [new branch] gh/slayton58/32/orig -> origin/gh/slayton58/32/orig 2025-11-03T16:58:01.9002930Z * [new branch] gh/slayton58/33/base -> origin/gh/slayton58/33/base 2025-11-03T16:58:01.9004652Z * [new branch] gh/slayton58/33/head -> origin/gh/slayton58/33/head 2025-11-03T16:58:01.9006472Z * [new branch] gh/slayton58/33/orig -> origin/gh/slayton58/33/orig 2025-11-03T16:58:01.9008650Z * [new branch] gh/slayton58/34/base -> origin/gh/slayton58/34/base 2025-11-03T16:58:01.9010359Z * [new branch] gh/slayton58/34/head -> origin/gh/slayton58/34/head 2025-11-03T16:58:01.9012115Z * [new branch] gh/slayton58/34/orig -> origin/gh/slayton58/34/orig 2025-11-03T16:58:01.9014366Z * [new branch] gh/slayton58/35/base -> origin/gh/slayton58/35/base 2025-11-03T16:58:01.9016099Z * [new branch] gh/slayton58/35/head -> origin/gh/slayton58/35/head 2025-11-03T16:58:01.9017816Z * [new branch] gh/slayton58/35/orig -> origin/gh/slayton58/35/orig 2025-11-03T16:58:01.9020982Z * [new branch] gh/slayton58/36/base -> origin/gh/slayton58/36/base 2025-11-03T16:58:01.9023215Z * [new branch] gh/slayton58/36/head -> origin/gh/slayton58/36/head 2025-11-03T16:58:01.9024876Z * [new branch] gh/slayton58/36/orig -> origin/gh/slayton58/36/orig 2025-11-03T16:58:01.9027147Z * [new branch] gh/slayton58/37/base -> origin/gh/slayton58/37/base 2025-11-03T16:58:01.9028949Z * [new branch] gh/slayton58/37/head -> origin/gh/slayton58/37/head 2025-11-03T16:58:01.9030726Z * [new branch] gh/slayton58/37/orig -> origin/gh/slayton58/37/orig 2025-11-03T16:58:01.9033127Z * [new branch] gh/slayton58/38/base -> origin/gh/slayton58/38/base 2025-11-03T16:58:01.9034788Z * [new branch] gh/slayton58/38/head -> origin/gh/slayton58/38/head 2025-11-03T16:58:01.9036464Z * [new branch] gh/slayton58/38/orig -> origin/gh/slayton58/38/orig 2025-11-03T16:58:01.9039026Z * [new branch] gh/slayton58/39/base -> origin/gh/slayton58/39/base 2025-11-03T16:58:01.9040759Z * [new branch] gh/slayton58/39/head -> origin/gh/slayton58/39/head 2025-11-03T16:58:01.9042450Z * [new branch] gh/slayton58/39/orig -> origin/gh/slayton58/39/orig 2025-11-03T16:58:01.9044996Z * [new branch] gh/slayton58/40/base -> origin/gh/slayton58/40/base 2025-11-03T16:58:01.9046661Z * [new branch] gh/slayton58/40/head -> origin/gh/slayton58/40/head 2025-11-03T16:58:01.9048388Z * [new branch] gh/slayton58/40/orig -> origin/gh/slayton58/40/orig 2025-11-03T16:58:01.9050740Z * [new branch] gh/slayton58/41/base -> origin/gh/slayton58/41/base 2025-11-03T16:58:01.9052479Z * [new branch] gh/slayton58/41/head -> origin/gh/slayton58/41/head 2025-11-03T16:58:01.9054196Z * [new branch] gh/slayton58/41/orig -> origin/gh/slayton58/41/orig 2025-11-03T16:58:01.9056523Z * [new branch] gh/slayton58/42/base -> origin/gh/slayton58/42/base 2025-11-03T16:58:01.9058198Z * [new branch] gh/slayton58/42/head -> origin/gh/slayton58/42/head 2025-11-03T16:58:01.9059985Z * [new branch] gh/slayton58/42/orig -> origin/gh/slayton58/42/orig 2025-11-03T16:58:01.9062474Z * [new branch] gh/slayton58/6/base -> origin/gh/slayton58/6/base 2025-11-03T16:58:01.9064193Z * [new branch] gh/slayton58/6/head -> origin/gh/slayton58/6/head 2025-11-03T16:58:01.9066391Z * [new branch] gh/slayton58/7/base -> origin/gh/slayton58/7/base 2025-11-03T16:58:01.9068252Z * [new branch] gh/slayton58/7/head -> origin/gh/slayton58/7/head 2025-11-03T16:58:01.9071021Z * [new branch] gh/soulitzer/269/base -> origin/gh/soulitzer/269/base 2025-11-03T16:58:01.9072636Z * [new branch] gh/soulitzer/269/head -> origin/gh/soulitzer/269/head 2025-11-03T16:58:01.9074380Z * [new branch] gh/soulitzer/269/orig -> origin/gh/soulitzer/269/orig 2025-11-03T16:58:01.9076815Z * [new branch] gh/soulitzer/276/base -> origin/gh/soulitzer/276/base 2025-11-03T16:58:01.9078481Z * [new branch] gh/soulitzer/276/head -> origin/gh/soulitzer/276/head 2025-11-03T16:58:01.9080209Z * [new branch] gh/soulitzer/276/orig -> origin/gh/soulitzer/276/orig 2025-11-03T16:58:01.9082759Z * [new branch] gh/soulitzer/287/base -> origin/gh/soulitzer/287/base 2025-11-03T16:58:01.9084534Z * [new branch] gh/soulitzer/287/head -> origin/gh/soulitzer/287/head 2025-11-03T16:58:01.9086332Z * [new branch] gh/soulitzer/287/orig -> origin/gh/soulitzer/287/orig 2025-11-03T16:58:01.9088794Z * [new branch] gh/soulitzer/296/base -> origin/gh/soulitzer/296/base 2025-11-03T16:58:01.9090626Z * [new branch] gh/soulitzer/296/head -> origin/gh/soulitzer/296/head 2025-11-03T16:58:01.9092419Z * [new branch] gh/soulitzer/296/orig -> origin/gh/soulitzer/296/orig 2025-11-03T16:58:01.9094717Z * [new branch] gh/soulitzer/299/base -> origin/gh/soulitzer/299/base 2025-11-03T16:58:01.9096460Z * [new branch] gh/soulitzer/299/head -> origin/gh/soulitzer/299/head 2025-11-03T16:58:01.9098232Z * [new branch] gh/soulitzer/299/orig -> origin/gh/soulitzer/299/orig 2025-11-03T16:58:01.9100555Z * [new branch] gh/soulitzer/300/base -> origin/gh/soulitzer/300/base 2025-11-03T16:58:01.9102411Z * [new branch] gh/soulitzer/300/head -> origin/gh/soulitzer/300/head 2025-11-03T16:58:01.9104076Z * [new branch] gh/soulitzer/300/orig -> origin/gh/soulitzer/300/orig 2025-11-03T16:58:01.9106492Z * [new branch] gh/soulitzer/301/base -> origin/gh/soulitzer/301/base 2025-11-03T16:58:01.9108287Z * [new branch] gh/soulitzer/301/head -> origin/gh/soulitzer/301/head 2025-11-03T16:58:01.9110020Z * [new branch] gh/soulitzer/301/orig -> origin/gh/soulitzer/301/orig 2025-11-03T16:58:01.9112806Z * [new branch] gh/soulitzer/313/base -> origin/gh/soulitzer/313/base 2025-11-03T16:58:01.9114537Z * [new branch] gh/soulitzer/313/head -> origin/gh/soulitzer/313/head 2025-11-03T16:58:01.9116248Z * [new branch] gh/soulitzer/313/orig -> origin/gh/soulitzer/313/orig 2025-11-03T16:58:01.9118600Z * [new branch] gh/soulitzer/319/base -> origin/gh/soulitzer/319/base 2025-11-03T16:58:01.9120331Z * [new branch] gh/soulitzer/319/head -> origin/gh/soulitzer/319/head 2025-11-03T16:58:01.9122424Z * [new branch] gh/soulitzer/319/orig -> origin/gh/soulitzer/319/orig 2025-11-03T16:58:01.9124887Z * [new branch] gh/soulitzer/320/base -> origin/gh/soulitzer/320/base 2025-11-03T16:58:01.9126503Z * [new branch] gh/soulitzer/320/head -> origin/gh/soulitzer/320/head 2025-11-03T16:58:01.9128124Z * [new branch] gh/soulitzer/320/orig -> origin/gh/soulitzer/320/orig 2025-11-03T16:58:01.9130531Z * [new branch] gh/soulitzer/336/base -> origin/gh/soulitzer/336/base 2025-11-03T16:58:01.9132210Z * [new branch] gh/soulitzer/336/head -> origin/gh/soulitzer/336/head 2025-11-03T16:58:01.9133929Z * [new branch] gh/soulitzer/336/orig -> origin/gh/soulitzer/336/orig 2025-11-03T16:58:01.9136213Z * [new branch] gh/soulitzer/347/base -> origin/gh/soulitzer/347/base 2025-11-03T16:58:01.9138085Z * [new branch] gh/soulitzer/347/head -> origin/gh/soulitzer/347/head 2025-11-03T16:58:01.9139582Z * [new branch] gh/soulitzer/347/orig -> origin/gh/soulitzer/347/orig 2025-11-03T16:58:01.9142235Z * [new branch] gh/soulitzer/349/base -> origin/gh/soulitzer/349/base 2025-11-03T16:58:01.9143919Z * [new branch] gh/soulitzer/349/head -> origin/gh/soulitzer/349/head 2025-11-03T16:58:01.9145668Z * [new branch] gh/soulitzer/349/orig -> origin/gh/soulitzer/349/orig 2025-11-03T16:58:01.9147919Z * [new branch] gh/soulitzer/350/base -> origin/gh/soulitzer/350/base 2025-11-03T16:58:01.9149518Z * [new branch] gh/soulitzer/350/head -> origin/gh/soulitzer/350/head 2025-11-03T16:58:01.9151244Z * [new branch] gh/soulitzer/350/orig -> origin/gh/soulitzer/350/orig 2025-11-03T16:58:01.9153668Z * [new branch] gh/soulitzer/351/base -> origin/gh/soulitzer/351/base 2025-11-03T16:58:01.9155414Z * [new branch] gh/soulitzer/351/head -> origin/gh/soulitzer/351/head 2025-11-03T16:58:01.9157171Z * [new branch] gh/soulitzer/351/orig -> origin/gh/soulitzer/351/orig 2025-11-03T16:58:01.9159479Z * [new branch] gh/soulitzer/353/base -> origin/gh/soulitzer/353/base 2025-11-03T16:58:01.9161271Z * [new branch] gh/soulitzer/353/head -> origin/gh/soulitzer/353/head 2025-11-03T16:58:01.9162893Z * [new branch] gh/soulitzer/353/orig -> origin/gh/soulitzer/353/orig 2025-11-03T16:58:01.9165855Z * [new branch] gh/soulitzer/358/base -> origin/gh/soulitzer/358/base 2025-11-03T16:58:01.9168004Z * [new branch] gh/soulitzer/358/head -> origin/gh/soulitzer/358/head 2025-11-03T16:58:01.9169768Z * [new branch] gh/soulitzer/358/orig -> origin/gh/soulitzer/358/orig 2025-11-03T16:58:01.9172443Z * [new branch] gh/soulitzer/359/base -> origin/gh/soulitzer/359/base 2025-11-03T16:58:01.9174168Z * [new branch] gh/soulitzer/359/head -> origin/gh/soulitzer/359/head 2025-11-03T16:58:01.9175923Z * [new branch] gh/soulitzer/359/orig -> origin/gh/soulitzer/359/orig 2025-11-03T16:58:01.9178294Z * [new branch] gh/soulitzer/374/base -> origin/gh/soulitzer/374/base 2025-11-03T16:58:01.9180052Z * [new branch] gh/soulitzer/374/head -> origin/gh/soulitzer/374/head 2025-11-03T16:58:01.9181846Z * [new branch] gh/soulitzer/374/orig -> origin/gh/soulitzer/374/orig 2025-11-03T16:58:01.9184259Z * [new branch] gh/soulitzer/375/base -> origin/gh/soulitzer/375/base 2025-11-03T16:58:01.9185905Z * [new branch] gh/soulitzer/375/head -> origin/gh/soulitzer/375/head 2025-11-03T16:58:01.9187571Z * [new branch] gh/soulitzer/375/orig -> origin/gh/soulitzer/375/orig 2025-11-03T16:58:01.9189936Z * [new branch] gh/soulitzer/380/base -> origin/gh/soulitzer/380/base 2025-11-03T16:58:01.9191586Z * [new branch] gh/soulitzer/380/head -> origin/gh/soulitzer/380/head 2025-11-03T16:58:01.9193435Z * [new branch] gh/soulitzer/380/orig -> origin/gh/soulitzer/380/orig 2025-11-03T16:58:01.9195777Z * [new branch] gh/soulitzer/381/base -> origin/gh/soulitzer/381/base 2025-11-03T16:58:01.9197444Z * [new branch] gh/soulitzer/381/head -> origin/gh/soulitzer/381/head 2025-11-03T16:58:01.9199138Z * [new branch] gh/soulitzer/381/orig -> origin/gh/soulitzer/381/orig 2025-11-03T16:58:01.9201480Z * [new branch] gh/soulitzer/384/base -> origin/gh/soulitzer/384/base 2025-11-03T16:58:01.9203244Z * [new branch] gh/soulitzer/384/head -> origin/gh/soulitzer/384/head 2025-11-03T16:58:01.9204848Z * [new branch] gh/soulitzer/384/orig -> origin/gh/soulitzer/384/orig 2025-11-03T16:58:01.9207267Z * [new branch] gh/soulitzer/385/base -> origin/gh/soulitzer/385/base 2025-11-03T16:58:01.9208908Z * [new branch] gh/soulitzer/385/head -> origin/gh/soulitzer/385/head 2025-11-03T16:58:01.9210571Z * [new branch] gh/soulitzer/385/orig -> origin/gh/soulitzer/385/orig 2025-11-03T16:58:01.9212902Z * [new branch] gh/soulitzer/386/base -> origin/gh/soulitzer/386/base 2025-11-03T16:58:01.9214663Z * [new branch] gh/soulitzer/386/head -> origin/gh/soulitzer/386/head 2025-11-03T16:58:01.9216302Z * [new branch] gh/soulitzer/386/orig -> origin/gh/soulitzer/386/orig 2025-11-03T16:58:01.9218651Z * [new branch] gh/soulitzer/387/base -> origin/gh/soulitzer/387/base 2025-11-03T16:58:01.9220339Z * [new branch] gh/soulitzer/387/head -> origin/gh/soulitzer/387/head 2025-11-03T16:58:01.9223731Z * [new branch] gh/soulitzer/387/orig -> origin/gh/soulitzer/387/orig 2025-11-03T16:58:01.9226617Z * [new branch] gh/swolchok/728/next -> origin/gh/swolchok/728/next 2025-11-03T16:58:01.9228909Z * [new branch] gh/swolchok/786/base -> origin/gh/swolchok/786/base 2025-11-03T16:58:01.9230557Z * [new branch] gh/swolchok/786/head -> origin/gh/swolchok/786/head 2025-11-03T16:58:01.9232359Z * [new branch] gh/swolchok/786/orig -> origin/gh/swolchok/786/orig 2025-11-03T16:58:01.9234523Z * [new branch] gh/swolchok/787/base -> origin/gh/swolchok/787/base 2025-11-03T16:58:01.9236184Z * [new branch] gh/swolchok/787/head -> origin/gh/swolchok/787/head 2025-11-03T16:58:01.9238029Z * [new branch] gh/swolchok/787/orig -> origin/gh/swolchok/787/orig 2025-11-03T16:58:01.9240579Z * [new branch] gh/swolchok/819/base -> origin/gh/swolchok/819/base 2025-11-03T16:58:01.9242167Z * [new branch] gh/swolchok/819/head -> origin/gh/swolchok/819/head 2025-11-03T16:58:01.9243854Z * [new branch] gh/swolchok/819/orig -> origin/gh/swolchok/819/orig 2025-11-03T16:58:01.9246108Z * [new branch] gh/swolchok/824/base -> origin/gh/swolchok/824/base 2025-11-03T16:58:01.9247856Z * [new branch] gh/swolchok/824/head -> origin/gh/swolchok/824/head 2025-11-03T16:58:01.9249605Z * [new branch] gh/swolchok/824/orig -> origin/gh/swolchok/824/orig 2025-11-03T16:58:01.9251922Z * [new branch] gh/swolchok/829/base -> origin/gh/swolchok/829/base 2025-11-03T16:58:01.9253574Z * [new branch] gh/swolchok/829/head -> origin/gh/swolchok/829/head 2025-11-03T16:58:01.9255279Z * [new branch] gh/swolchok/829/orig -> origin/gh/swolchok/829/orig 2025-11-03T16:58:01.9257588Z * [new branch] gh/swolchok/838/base -> origin/gh/swolchok/838/base 2025-11-03T16:58:01.9259251Z * [new branch] gh/swolchok/838/head -> origin/gh/swolchok/838/head 2025-11-03T16:58:01.9260950Z * [new branch] gh/swolchok/838/orig -> origin/gh/swolchok/838/orig 2025-11-03T16:58:01.9263495Z * [new branch] gh/swolchok/839/base -> origin/gh/swolchok/839/base 2025-11-03T16:58:01.9265136Z * [new branch] gh/swolchok/839/head -> origin/gh/swolchok/839/head 2025-11-03T16:58:01.9266766Z * [new branch] gh/swolchok/839/orig -> origin/gh/swolchok/839/orig 2025-11-03T16:58:01.9268967Z * [new branch] gh/swolchok/841/base -> origin/gh/swolchok/841/base 2025-11-03T16:58:01.9270732Z * [new branch] gh/swolchok/841/head -> origin/gh/swolchok/841/head 2025-11-03T16:58:01.9272474Z * [new branch] gh/swolchok/841/orig -> origin/gh/swolchok/841/orig 2025-11-03T16:58:01.9274777Z * [new branch] gh/swolchok/842/base -> origin/gh/swolchok/842/base 2025-11-03T16:58:01.9276635Z * [new branch] gh/swolchok/842/head -> origin/gh/swolchok/842/head 2025-11-03T16:58:01.9278215Z * [new branch] gh/swolchok/842/orig -> origin/gh/swolchok/842/orig 2025-11-03T16:58:01.9280425Z * [new branch] gh/swolchok/843/base -> origin/gh/swolchok/843/base 2025-11-03T16:58:01.9282131Z * [new branch] gh/swolchok/843/head -> origin/gh/swolchok/843/head 2025-11-03T16:58:01.9284269Z * [new branch] gh/swolchok/843/orig -> origin/gh/swolchok/843/orig 2025-11-03T16:58:01.9287085Z * [new branch] gh/swolchok/844/base -> origin/gh/swolchok/844/base 2025-11-03T16:58:01.9288812Z * [new branch] gh/swolchok/844/head -> origin/gh/swolchok/844/head 2025-11-03T16:58:01.9290687Z * [new branch] gh/swolchok/844/orig -> origin/gh/swolchok/844/orig 2025-11-03T16:58:01.9293011Z * [new branch] gh/swolchok/845/base -> origin/gh/swolchok/845/base 2025-11-03T16:58:01.9294739Z * [new branch] gh/swolchok/845/head -> origin/gh/swolchok/845/head 2025-11-03T16:58:01.9296472Z * [new branch] gh/swolchok/845/orig -> origin/gh/swolchok/845/orig 2025-11-03T16:58:01.9298794Z * [new branch] gh/swolchok/847/base -> origin/gh/swolchok/847/base 2025-11-03T16:58:01.9300532Z * [new branch] gh/swolchok/847/head -> origin/gh/swolchok/847/head 2025-11-03T16:58:01.9302350Z * [new branch] gh/swolchok/847/orig -> origin/gh/swolchok/847/orig 2025-11-03T16:58:01.9304801Z * [new branch] gh/swolchok/848/base -> origin/gh/swolchok/848/base 2025-11-03T16:58:01.9306634Z * [new branch] gh/swolchok/848/head -> origin/gh/swolchok/848/head 2025-11-03T16:58:01.9308281Z * [new branch] gh/swolchok/848/orig -> origin/gh/swolchok/848/orig 2025-11-03T16:58:01.9310684Z * [new branch] gh/swolchok/849/base -> origin/gh/swolchok/849/base 2025-11-03T16:58:01.9312326Z * [new branch] gh/swolchok/849/head -> origin/gh/swolchok/849/head 2025-11-03T16:58:01.9314206Z * [new branch] gh/swolchok/849/orig -> origin/gh/swolchok/849/orig 2025-11-03T16:58:01.9316882Z * [new branch] gh/swolchok/850/base -> origin/gh/swolchok/850/base 2025-11-03T16:58:01.9318629Z * [new branch] gh/swolchok/850/head -> origin/gh/swolchok/850/head 2025-11-03T16:58:01.9320402Z * [new branch] gh/swolchok/850/orig -> origin/gh/swolchok/850/orig 2025-11-03T16:58:01.9322998Z * [new branch] gh/swolchok/851/base -> origin/gh/swolchok/851/base 2025-11-03T16:58:01.9324756Z * [new branch] gh/swolchok/851/head -> origin/gh/swolchok/851/head 2025-11-03T16:58:01.9326501Z * [new branch] gh/swolchok/851/orig -> origin/gh/swolchok/851/orig 2025-11-03T16:58:01.9329432Z * [new branch] gh/swolchok/852/base -> origin/gh/swolchok/852/base 2025-11-03T16:58:01.9331134Z * [new branch] gh/swolchok/852/head -> origin/gh/swolchok/852/head 2025-11-03T16:58:01.9332801Z * [new branch] gh/swolchok/852/orig -> origin/gh/swolchok/852/orig 2025-11-03T16:58:01.9335214Z * [new branch] gh/swolchok/853/base -> origin/gh/swolchok/853/base 2025-11-03T16:58:01.9337067Z * [new branch] gh/swolchok/853/head -> origin/gh/swolchok/853/head 2025-11-03T16:58:01.9339827Z * [new branch] gh/swolchok/853/orig -> origin/gh/swolchok/853/orig 2025-11-03T16:58:01.9342471Z * [new branch] gh/swolchok/854/base -> origin/gh/swolchok/854/base 2025-11-03T16:58:01.9344026Z * [new branch] gh/swolchok/854/head -> origin/gh/swolchok/854/head 2025-11-03T16:58:01.9345860Z * [new branch] gh/swolchok/854/orig -> origin/gh/swolchok/854/orig 2025-11-03T16:58:01.9348381Z * [new branch] gh/swolchok/855/base -> origin/gh/swolchok/855/base 2025-11-03T16:58:01.9349927Z * [new branch] gh/swolchok/855/head -> origin/gh/swolchok/855/head 2025-11-03T16:58:01.9351626Z * [new branch] gh/swolchok/855/orig -> origin/gh/swolchok/855/orig 2025-11-03T16:58:01.9353964Z * [new branch] gh/swolchok/856/base -> origin/gh/swolchok/856/base 2025-11-03T16:58:01.9355597Z * [new branch] gh/swolchok/856/head -> origin/gh/swolchok/856/head 2025-11-03T16:58:01.9357237Z * [new branch] gh/swolchok/856/orig -> origin/gh/swolchok/856/orig 2025-11-03T16:58:01.9359676Z * [new branch] gh/swolchok/857/base -> origin/gh/swolchok/857/base 2025-11-03T16:58:01.9362029Z * [new branch] gh/swolchok/857/head -> origin/gh/swolchok/857/head 2025-11-03T16:58:01.9363682Z * [new branch] gh/swolchok/857/orig -> origin/gh/swolchok/857/orig 2025-11-03T16:58:01.9366063Z * [new branch] gh/swolchok/858/base -> origin/gh/swolchok/858/base 2025-11-03T16:58:01.9367803Z * [new branch] gh/swolchok/858/head -> origin/gh/swolchok/858/head 2025-11-03T16:58:01.9369618Z * [new branch] gh/swolchok/858/orig -> origin/gh/swolchok/858/orig 2025-11-03T16:58:01.9371924Z * [new branch] gh/swolchok/859/base -> origin/gh/swolchok/859/base 2025-11-03T16:58:01.9373691Z * [new branch] gh/swolchok/859/head -> origin/gh/swolchok/859/head 2025-11-03T16:58:01.9375431Z * [new branch] gh/swolchok/859/orig -> origin/gh/swolchok/859/orig 2025-11-03T16:58:01.9377929Z * [new branch] gh/swolchok/860/base -> origin/gh/swolchok/860/base 2025-11-03T16:58:01.9379689Z * [new branch] gh/swolchok/860/head -> origin/gh/swolchok/860/head 2025-11-03T16:58:01.9381416Z * [new branch] gh/swolchok/860/orig -> origin/gh/swolchok/860/orig 2025-11-03T16:58:01.9384185Z * [new branch] gh/swolchok/861/base -> origin/gh/swolchok/861/base 2025-11-03T16:58:01.9385858Z * [new branch] gh/swolchok/861/head -> origin/gh/swolchok/861/head 2025-11-03T16:58:01.9387560Z * [new branch] gh/swolchok/861/orig -> origin/gh/swolchok/861/orig 2025-11-03T16:58:01.9390057Z * [new branch] gh/swolchok/862/base -> origin/gh/swolchok/862/base 2025-11-03T16:58:01.9391939Z * [new branch] gh/swolchok/862/head -> origin/gh/swolchok/862/head 2025-11-03T16:58:01.9393658Z * [new branch] gh/swolchok/862/orig -> origin/gh/swolchok/862/orig 2025-11-03T16:58:01.9396198Z * [new branch] gh/swolchok/863/base -> origin/gh/swolchok/863/base 2025-11-03T16:58:01.9397993Z * [new branch] gh/swolchok/863/head -> origin/gh/swolchok/863/head 2025-11-03T16:58:01.9399718Z * [new branch] gh/swolchok/863/orig -> origin/gh/swolchok/863/orig 2025-11-03T16:58:01.9402110Z * [new branch] gh/swolchok/864/base -> origin/gh/swolchok/864/base 2025-11-03T16:58:01.9403855Z * [new branch] gh/swolchok/864/head -> origin/gh/swolchok/864/head 2025-11-03T16:58:01.9405572Z * [new branch] gh/swolchok/864/orig -> origin/gh/swolchok/864/orig 2025-11-03T16:58:01.9407808Z * [new branch] gh/swolchok/865/base -> origin/gh/swolchok/865/base 2025-11-03T16:58:01.9409503Z * [new branch] gh/swolchok/865/head -> origin/gh/swolchok/865/head 2025-11-03T16:58:01.9411155Z * [new branch] gh/swolchok/865/orig -> origin/gh/swolchok/865/orig 2025-11-03T16:58:01.9414068Z * [new branch] gh/teja-rao/4/base -> origin/gh/teja-rao/4/base 2025-11-03T16:58:01.9415864Z * [new branch] gh/teja-rao/4/head -> origin/gh/teja-rao/4/head 2025-11-03T16:58:01.9417709Z * [new branch] gh/teja-rao/4/orig -> origin/gh/teja-rao/4/orig 2025-11-03T16:58:01.9420380Z * [new branch] gh/tianyu-l/2/base -> origin/gh/tianyu-l/2/base 2025-11-03T16:58:01.9422356Z * [new branch] gh/tianyu-l/2/head -> origin/gh/tianyu-l/2/head 2025-11-03T16:58:01.9424493Z * [new branch] gh/tianyu-l/2/orig -> origin/gh/tianyu-l/2/orig 2025-11-03T16:58:01.9427539Z * [new branch] gh/tugsbayasgalan/10/base -> origin/gh/tugsbayasgalan/10/base 2025-11-03T16:58:01.9429243Z * [new branch] gh/tugsbayasgalan/10/head -> origin/gh/tugsbayasgalan/10/head 2025-11-03T16:58:01.9430955Z * [new branch] gh/tugsbayasgalan/10/orig -> origin/gh/tugsbayasgalan/10/orig 2025-11-03T16:58:01.9433396Z * [new branch] gh/tugsbayasgalan/13/base -> origin/gh/tugsbayasgalan/13/base 2025-11-03T16:58:01.9435113Z * [new branch] gh/tugsbayasgalan/13/head -> origin/gh/tugsbayasgalan/13/head 2025-11-03T16:58:01.9436843Z * [new branch] gh/tugsbayasgalan/13/orig -> origin/gh/tugsbayasgalan/13/orig 2025-11-03T16:58:01.9439257Z * [new branch] gh/tugsbayasgalan/17/base -> origin/gh/tugsbayasgalan/17/base 2025-11-03T16:58:01.9440903Z * [new branch] gh/tugsbayasgalan/17/head -> origin/gh/tugsbayasgalan/17/head 2025-11-03T16:58:01.9442897Z * [new branch] gh/tugsbayasgalan/17/orig -> origin/gh/tugsbayasgalan/17/orig 2025-11-03T16:58:01.9445502Z * [new branch] gh/tugsbayasgalan/2/base -> origin/gh/tugsbayasgalan/2/base 2025-11-03T16:58:01.9447166Z * [new branch] gh/tugsbayasgalan/2/head -> origin/gh/tugsbayasgalan/2/head 2025-11-03T16:58:01.9448908Z * [new branch] gh/tugsbayasgalan/2/orig -> origin/gh/tugsbayasgalan/2/orig 2025-11-03T16:58:01.9451481Z * [new branch] gh/tugsbayasgalan/28/base -> origin/gh/tugsbayasgalan/28/base 2025-11-03T16:58:01.9453479Z * [new branch] gh/tugsbayasgalan/28/head -> origin/gh/tugsbayasgalan/28/head 2025-11-03T16:58:01.9455614Z * [new branch] gh/tugsbayasgalan/28/orig -> origin/gh/tugsbayasgalan/28/orig 2025-11-03T16:58:01.9457791Z * [new branch] gh/tugsbayasgalan/3/base -> origin/gh/tugsbayasgalan/3/base 2025-11-03T16:58:01.9459595Z * [new branch] gh/tugsbayasgalan/3/head -> origin/gh/tugsbayasgalan/3/head 2025-11-03T16:58:01.9461267Z * [new branch] gh/tugsbayasgalan/3/orig -> origin/gh/tugsbayasgalan/3/orig 2025-11-03T16:58:01.9464478Z * [new branch] gh/tugsbayasgalan/32/base -> origin/gh/tugsbayasgalan/32/base 2025-11-03T16:58:01.9466282Z * [new branch] gh/tugsbayasgalan/32/head -> origin/gh/tugsbayasgalan/32/head 2025-11-03T16:58:01.9467921Z * [new branch] gh/tugsbayasgalan/32/orig -> origin/gh/tugsbayasgalan/32/orig 2025-11-03T16:58:01.9470337Z * [new branch] gh/tugsbayasgalan/35/base -> origin/gh/tugsbayasgalan/35/base 2025-11-03T16:58:01.9472087Z * [new branch] gh/tugsbayasgalan/35/head -> origin/gh/tugsbayasgalan/35/head 2025-11-03T16:58:01.9473968Z * [new branch] gh/tugsbayasgalan/35/orig -> origin/gh/tugsbayasgalan/35/orig 2025-11-03T16:58:01.9476223Z * [new branch] gh/tugsbayasgalan/36/base -> origin/gh/tugsbayasgalan/36/base 2025-11-03T16:58:01.9477877Z * [new branch] gh/tugsbayasgalan/36/head -> origin/gh/tugsbayasgalan/36/head 2025-11-03T16:58:01.9479570Z * [new branch] gh/tugsbayasgalan/36/orig -> origin/gh/tugsbayasgalan/36/orig 2025-11-03T16:58:01.9481903Z * [new branch] gh/tugsbayasgalan/37/base -> origin/gh/tugsbayasgalan/37/base 2025-11-03T16:58:01.9483504Z * [new branch] gh/tugsbayasgalan/37/head -> origin/gh/tugsbayasgalan/37/head 2025-11-03T16:58:01.9485203Z * [new branch] gh/tugsbayasgalan/37/orig -> origin/gh/tugsbayasgalan/37/orig 2025-11-03T16:58:01.9487818Z * [new branch] gh/tugsbayasgalan/39/base -> origin/gh/tugsbayasgalan/39/base 2025-11-03T16:58:01.9489505Z * [new branch] gh/tugsbayasgalan/39/head -> origin/gh/tugsbayasgalan/39/head 2025-11-03T16:58:01.9491345Z * [new branch] gh/tugsbayasgalan/39/orig -> origin/gh/tugsbayasgalan/39/orig 2025-11-03T16:58:01.9493648Z * [new branch] gh/tugsbayasgalan/43/base -> origin/gh/tugsbayasgalan/43/base 2025-11-03T16:58:01.9495324Z * [new branch] gh/tugsbayasgalan/43/head -> origin/gh/tugsbayasgalan/43/head 2025-11-03T16:58:01.9497021Z * [new branch] gh/tugsbayasgalan/43/orig -> origin/gh/tugsbayasgalan/43/orig 2025-11-03T16:58:01.9499426Z * [new branch] gh/tugsbayasgalan/46/base -> origin/gh/tugsbayasgalan/46/base 2025-11-03T16:58:01.9501044Z * [new branch] gh/tugsbayasgalan/46/head -> origin/gh/tugsbayasgalan/46/head 2025-11-03T16:58:01.9502931Z * [new branch] gh/tugsbayasgalan/46/orig -> origin/gh/tugsbayasgalan/46/orig 2025-11-03T16:58:01.9505663Z * [new branch] gh/tugsbayasgalan/47/base -> origin/gh/tugsbayasgalan/47/base 2025-11-03T16:58:01.9507319Z * [new branch] gh/tugsbayasgalan/47/head -> origin/gh/tugsbayasgalan/47/head 2025-11-03T16:58:01.9509031Z * [new branch] gh/tugsbayasgalan/47/orig -> origin/gh/tugsbayasgalan/47/orig 2025-11-03T16:58:01.9511307Z * [new branch] gh/tugsbayasgalan/48/base -> origin/gh/tugsbayasgalan/48/base 2025-11-03T16:58:01.9512973Z * [new branch] gh/tugsbayasgalan/48/head -> origin/gh/tugsbayasgalan/48/head 2025-11-03T16:58:01.9514644Z * [new branch] gh/tugsbayasgalan/48/orig -> origin/gh/tugsbayasgalan/48/orig 2025-11-03T16:58:01.9516788Z * [new branch] gh/tugsbayasgalan/49/base -> origin/gh/tugsbayasgalan/49/base 2025-11-03T16:58:01.9518543Z * [new branch] gh/tugsbayasgalan/49/head -> origin/gh/tugsbayasgalan/49/head 2025-11-03T16:58:01.9520263Z * [new branch] gh/tugsbayasgalan/49/orig -> origin/gh/tugsbayasgalan/49/orig 2025-11-03T16:58:01.9523503Z * [new branch] gh/tugsbayasgalan/50/base -> origin/gh/tugsbayasgalan/50/base 2025-11-03T16:58:01.9525393Z * [new branch] gh/tugsbayasgalan/50/head -> origin/gh/tugsbayasgalan/50/head 2025-11-03T16:58:01.9527092Z * [new branch] gh/tugsbayasgalan/50/orig -> origin/gh/tugsbayasgalan/50/orig 2025-11-03T16:58:01.9529702Z * [new branch] gh/tugsbayasgalan/51/base -> origin/gh/tugsbayasgalan/51/base 2025-11-03T16:58:01.9531384Z * [new branch] gh/tugsbayasgalan/51/head -> origin/gh/tugsbayasgalan/51/head 2025-11-03T16:58:01.9533038Z * [new branch] gh/tugsbayasgalan/51/orig -> origin/gh/tugsbayasgalan/51/orig 2025-11-03T16:58:01.9535710Z * [new branch] gh/tugsbayasgalan/52/base -> origin/gh/tugsbayasgalan/52/base 2025-11-03T16:58:01.9537483Z * [new branch] gh/tugsbayasgalan/52/head -> origin/gh/tugsbayasgalan/52/head 2025-11-03T16:58:01.9539162Z * [new branch] gh/tugsbayasgalan/52/orig -> origin/gh/tugsbayasgalan/52/orig 2025-11-03T16:58:01.9541681Z * [new branch] gh/tugsbayasgalan/53/base -> origin/gh/tugsbayasgalan/53/base 2025-11-03T16:58:01.9543495Z * [new branch] gh/tugsbayasgalan/53/head -> origin/gh/tugsbayasgalan/53/head 2025-11-03T16:58:01.9545141Z * [new branch] gh/tugsbayasgalan/53/orig -> origin/gh/tugsbayasgalan/53/orig 2025-11-03T16:58:01.9547418Z * [new branch] gh/tugsbayasgalan/54/base -> origin/gh/tugsbayasgalan/54/base 2025-11-03T16:58:01.9549193Z * [new branch] gh/tugsbayasgalan/54/head -> origin/gh/tugsbayasgalan/54/head 2025-11-03T16:58:01.9550928Z * [new branch] gh/tugsbayasgalan/54/orig -> origin/gh/tugsbayasgalan/54/orig 2025-11-03T16:58:01.9554113Z * [new branch] gh/tugsbayasgalan/55/base -> origin/gh/tugsbayasgalan/55/base 2025-11-03T16:58:01.9555684Z * [new branch] gh/tugsbayasgalan/55/head -> origin/gh/tugsbayasgalan/55/head 2025-11-03T16:58:01.9557408Z * [new branch] gh/tugsbayasgalan/55/orig -> origin/gh/tugsbayasgalan/55/orig 2025-11-03T16:58:01.9563078Z * [new branch] gh/tugsbayasgalan/56/base -> origin/gh/tugsbayasgalan/56/base 2025-11-03T16:58:01.9564804Z * [new branch] gh/tugsbayasgalan/56/head -> origin/gh/tugsbayasgalan/56/head 2025-11-03T16:58:01.9566492Z * [new branch] gh/tugsbayasgalan/56/orig -> origin/gh/tugsbayasgalan/56/orig 2025-11-03T16:58:01.9568911Z * [new branch] gh/tugsbayasgalan/57/base -> origin/gh/tugsbayasgalan/57/base 2025-11-03T16:58:01.9570680Z * [new branch] gh/tugsbayasgalan/57/head -> origin/gh/tugsbayasgalan/57/head 2025-11-03T16:58:01.9572311Z * [new branch] gh/tugsbayasgalan/57/orig -> origin/gh/tugsbayasgalan/57/orig 2025-11-03T16:58:01.9575026Z * [new branch] gh/tugsbayasgalan/58/base -> origin/gh/tugsbayasgalan/58/base 2025-11-03T16:58:01.9576668Z * [new branch] gh/tugsbayasgalan/58/head -> origin/gh/tugsbayasgalan/58/head 2025-11-03T16:58:01.9578393Z * [new branch] gh/tugsbayasgalan/58/orig -> origin/gh/tugsbayasgalan/58/orig 2025-11-03T16:58:01.9580843Z * [new branch] gh/tugsbayasgalan/59/base -> origin/gh/tugsbayasgalan/59/base 2025-11-03T16:58:01.9582759Z * [new branch] gh/tugsbayasgalan/59/head -> origin/gh/tugsbayasgalan/59/head 2025-11-03T16:58:01.9584459Z * [new branch] gh/tugsbayasgalan/59/orig -> origin/gh/tugsbayasgalan/59/orig 2025-11-03T16:58:01.9586836Z * [new branch] gh/tugsbayasgalan/6/base -> origin/gh/tugsbayasgalan/6/base 2025-11-03T16:58:01.9588554Z * [new branch] gh/tugsbayasgalan/6/head -> origin/gh/tugsbayasgalan/6/head 2025-11-03T16:58:01.9590244Z * [new branch] gh/tugsbayasgalan/6/orig -> origin/gh/tugsbayasgalan/6/orig 2025-11-03T16:58:01.9592510Z * [new branch] gh/tugsbayasgalan/60/base -> origin/gh/tugsbayasgalan/60/base 2025-11-03T16:58:01.9594188Z * [new branch] gh/tugsbayasgalan/60/head -> origin/gh/tugsbayasgalan/60/head 2025-11-03T16:58:01.9595919Z * [new branch] gh/tugsbayasgalan/60/orig -> origin/gh/tugsbayasgalan/60/orig 2025-11-03T16:58:01.9598357Z * [new branch] gh/tugsbayasgalan/61/base -> origin/gh/tugsbayasgalan/61/base 2025-11-03T16:58:01.9599950Z * [new branch] gh/tugsbayasgalan/61/head -> origin/gh/tugsbayasgalan/61/head 2025-11-03T16:58:01.9601651Z * [new branch] gh/tugsbayasgalan/61/orig -> origin/gh/tugsbayasgalan/61/orig 2025-11-03T16:58:01.9603805Z * [new branch] gh/tugsbayasgalan/62/base -> origin/gh/tugsbayasgalan/62/base 2025-11-03T16:58:01.9605537Z * [new branch] gh/tugsbayasgalan/62/head -> origin/gh/tugsbayasgalan/62/head 2025-11-03T16:58:01.9607217Z * [new branch] gh/tugsbayasgalan/62/orig -> origin/gh/tugsbayasgalan/62/orig 2025-11-03T16:58:01.9609556Z * [new branch] gh/tugsbayasgalan/63/base -> origin/gh/tugsbayasgalan/63/base 2025-11-03T16:58:01.9611260Z * [new branch] gh/tugsbayasgalan/63/head -> origin/gh/tugsbayasgalan/63/head 2025-11-03T16:58:01.9613002Z * [new branch] gh/tugsbayasgalan/63/orig -> origin/gh/tugsbayasgalan/63/orig 2025-11-03T16:58:01.9615511Z * [new branch] gh/tugsbayasgalan/64/base -> origin/gh/tugsbayasgalan/64/base 2025-11-03T16:58:01.9617181Z * [new branch] gh/tugsbayasgalan/64/head -> origin/gh/tugsbayasgalan/64/head 2025-11-03T16:58:01.9618817Z * [new branch] gh/tugsbayasgalan/64/orig -> origin/gh/tugsbayasgalan/64/orig 2025-11-03T16:58:01.9621408Z * [new branch] gh/tugsbayasgalan/65/base -> origin/gh/tugsbayasgalan/65/base 2025-11-03T16:58:01.9624907Z * [new branch] gh/tugsbayasgalan/65/head -> origin/gh/tugsbayasgalan/65/head 2025-11-03T16:58:01.9626954Z * [new branch] gh/tugsbayasgalan/65/orig -> origin/gh/tugsbayasgalan/65/orig 2025-11-03T16:58:01.9629830Z * [new branch] gh/tugsbayasgalan/66/base -> origin/gh/tugsbayasgalan/66/base 2025-11-03T16:58:01.9631973Z * [new branch] gh/tugsbayasgalan/66/head -> origin/gh/tugsbayasgalan/66/head 2025-11-03T16:58:01.9634017Z * [new branch] gh/tugsbayasgalan/66/orig -> origin/gh/tugsbayasgalan/66/orig 2025-11-03T16:58:01.9636294Z * [new branch] gh/tugsbayasgalan/67/base -> origin/gh/tugsbayasgalan/67/base 2025-11-03T16:58:01.9637981Z * [new branch] gh/tugsbayasgalan/67/head -> origin/gh/tugsbayasgalan/67/head 2025-11-03T16:58:01.9639784Z * [new branch] gh/tugsbayasgalan/67/orig -> origin/gh/tugsbayasgalan/67/orig 2025-11-03T16:58:01.9642363Z * [new branch] gh/tugsbayasgalan/68/base -> origin/gh/tugsbayasgalan/68/base 2025-11-03T16:58:01.9644069Z * [new branch] gh/tugsbayasgalan/68/head -> origin/gh/tugsbayasgalan/68/head 2025-11-03T16:58:01.9645817Z * [new branch] gh/tugsbayasgalan/68/orig -> origin/gh/tugsbayasgalan/68/orig 2025-11-03T16:58:01.9648112Z * [new branch] gh/tugsbayasgalan/69/base -> origin/gh/tugsbayasgalan/69/base 2025-11-03T16:58:01.9649951Z * [new branch] gh/tugsbayasgalan/69/head -> origin/gh/tugsbayasgalan/69/head 2025-11-03T16:58:01.9651642Z * [new branch] gh/tugsbayasgalan/69/orig -> origin/gh/tugsbayasgalan/69/orig 2025-11-03T16:58:01.9654035Z * [new branch] gh/tugsbayasgalan/7/base -> origin/gh/tugsbayasgalan/7/base 2025-11-03T16:58:01.9655765Z * [new branch] gh/tugsbayasgalan/7/head -> origin/gh/tugsbayasgalan/7/head 2025-11-03T16:58:01.9658486Z * [new branch] gh/tugsbayasgalan/7/orig -> origin/gh/tugsbayasgalan/7/orig 2025-11-03T16:58:01.9659947Z * [new branch] gh/tugsbayasgalan/70/base -> origin/gh/tugsbayasgalan/70/base 2025-11-03T16:58:01.9661935Z * [new branch] gh/tugsbayasgalan/70/head -> origin/gh/tugsbayasgalan/70/head 2025-11-03T16:58:01.9663765Z * [new branch] gh/tugsbayasgalan/70/orig -> origin/gh/tugsbayasgalan/70/orig 2025-11-03T16:58:01.9666364Z * [new branch] gh/tugsbayasgalan/71/base -> origin/gh/tugsbayasgalan/71/base 2025-11-03T16:58:01.9668323Z * [new branch] gh/tugsbayasgalan/71/head -> origin/gh/tugsbayasgalan/71/head 2025-11-03T16:58:01.9670053Z * [new branch] gh/tugsbayasgalan/71/orig -> origin/gh/tugsbayasgalan/71/orig 2025-11-03T16:58:01.9672514Z * [new branch] gh/tugsbayasgalan/8/base -> origin/gh/tugsbayasgalan/8/base 2025-11-03T16:58:01.9674218Z * [new branch] gh/tugsbayasgalan/8/head -> origin/gh/tugsbayasgalan/8/head 2025-11-03T16:58:01.9675933Z * [new branch] gh/tugsbayasgalan/8/orig -> origin/gh/tugsbayasgalan/8/orig 2025-11-03T16:58:01.9678334Z * [new branch] gh/tugsbayasgalan/9/base -> origin/gh/tugsbayasgalan/9/base 2025-11-03T16:58:01.9679916Z * [new branch] gh/tugsbayasgalan/9/head -> origin/gh/tugsbayasgalan/9/head 2025-11-03T16:58:01.9681649Z * [new branch] gh/tugsbayasgalan/9/orig -> origin/gh/tugsbayasgalan/9/orig 2025-11-03T16:58:01.9684514Z * [new branch] gh/v0i0/13/base -> origin/gh/v0i0/13/base 2025-11-03T16:58:01.9686152Z * [new branch] gh/v0i0/13/head -> origin/gh/v0i0/13/head 2025-11-03T16:58:01.9687882Z * [new branch] gh/v0i0/13/orig -> origin/gh/v0i0/13/orig 2025-11-03T16:58:01.9690250Z * [new branch] gh/v0i0/14/base -> origin/gh/v0i0/14/base 2025-11-03T16:58:01.9691981Z * [new branch] gh/v0i0/14/head -> origin/gh/v0i0/14/head 2025-11-03T16:58:01.9694083Z * [new branch] gh/v0i0/14/orig -> origin/gh/v0i0/14/orig 2025-11-03T16:58:01.9696825Z * [new branch] gh/vishal9-team/1/base -> origin/gh/vishal9-team/1/base 2025-11-03T16:58:01.9698523Z * [new branch] gh/vishal9-team/1/head -> origin/gh/vishal9-team/1/head 2025-11-03T16:58:01.9700648Z * [new branch] gh/vishal9-team/2/base -> origin/gh/vishal9-team/2/base 2025-11-03T16:58:01.9702498Z * [new branch] gh/vishal9-team/2/head -> origin/gh/vishal9-team/2/head 2025-11-03T16:58:01.9704211Z * [new branch] gh/vishal9-team/2/orig -> origin/gh/vishal9-team/2/orig 2025-11-03T16:58:01.9706462Z * [new branch] gh/vishal9-team/3/base -> origin/gh/vishal9-team/3/base 2025-11-03T16:58:01.9708106Z * [new branch] gh/vishal9-team/3/head -> origin/gh/vishal9-team/3/head 2025-11-03T16:58:01.9709837Z * [new branch] gh/vishal9-team/3/orig -> origin/gh/vishal9-team/3/orig 2025-11-03T16:58:01.9713309Z * [new branch] gh/vkuzo/1/next -> origin/gh/vkuzo/1/next 2025-11-03T16:58:01.9715684Z * [new branch] gh/vkuzo/2/next -> origin/gh/vkuzo/2/next 2025-11-03T16:58:01.9717936Z * [new branch] gh/vkuzo/3/next -> origin/gh/vkuzo/3/next 2025-11-03T16:58:01.9720740Z * [new branch] gh/wconstab/424/base -> origin/gh/wconstab/424/base 2025-11-03T16:58:01.9723185Z * [new branch] gh/wconstab/424/head -> origin/gh/wconstab/424/head 2025-11-03T16:58:01.9724933Z * [new branch] gh/wconstab/424/orig -> origin/gh/wconstab/424/orig 2025-11-03T16:58:01.9727266Z * [new branch] gh/wconstab/435/base -> origin/gh/wconstab/435/base 2025-11-03T16:58:01.9729013Z * [new branch] gh/wconstab/435/head -> origin/gh/wconstab/435/head 2025-11-03T16:58:01.9730855Z * [new branch] gh/wconstab/435/orig -> origin/gh/wconstab/435/orig 2025-11-03T16:58:01.9733049Z * [new branch] gh/wconstab/438/base -> origin/gh/wconstab/438/base 2025-11-03T16:58:01.9734780Z * [new branch] gh/wconstab/438/head -> origin/gh/wconstab/438/head 2025-11-03T16:58:01.9736467Z * [new branch] gh/wconstab/438/orig -> origin/gh/wconstab/438/orig 2025-11-03T16:58:01.9738858Z * [new branch] gh/wconstab/444/base -> origin/gh/wconstab/444/base 2025-11-03T16:58:01.9740608Z * [new branch] gh/wconstab/444/head -> origin/gh/wconstab/444/head 2025-11-03T16:58:01.9742486Z * [new branch] gh/wconstab/444/orig -> origin/gh/wconstab/444/orig 2025-11-03T16:58:01.9744808Z * [new branch] gh/wconstab/447/base -> origin/gh/wconstab/447/base 2025-11-03T16:58:01.9746684Z * [new branch] gh/wconstab/447/head -> origin/gh/wconstab/447/head 2025-11-03T16:58:01.9748362Z * [new branch] gh/wconstab/447/orig -> origin/gh/wconstab/447/orig 2025-11-03T16:58:01.9751243Z * [new branch] gh/wconstab/448/base -> origin/gh/wconstab/448/base 2025-11-03T16:58:01.9752949Z * [new branch] gh/wconstab/448/head -> origin/gh/wconstab/448/head 2025-11-03T16:58:01.9754893Z * [new branch] gh/wconstab/448/orig -> origin/gh/wconstab/448/orig 2025-11-03T16:58:01.9757805Z * [new branch] gh/weifengpy/34/base -> origin/gh/weifengpy/34/base 2025-11-03T16:58:01.9759555Z * [new branch] gh/weifengpy/34/head -> origin/gh/weifengpy/34/head 2025-11-03T16:58:01.9761247Z * [new branch] gh/weifengpy/34/orig -> origin/gh/weifengpy/34/orig 2025-11-03T16:58:01.9763550Z * [new branch] gh/weifengpy/35/base -> origin/gh/weifengpy/35/base 2025-11-03T16:58:01.9765143Z * [new branch] gh/weifengpy/35/head -> origin/gh/weifengpy/35/head 2025-11-03T16:58:01.9767033Z * [new branch] gh/weifengpy/35/orig -> origin/gh/weifengpy/35/orig 2025-11-03T16:58:01.9769344Z * [new branch] gh/weifengpy/36/base -> origin/gh/weifengpy/36/base 2025-11-03T16:58:01.9771063Z * [new branch] gh/weifengpy/36/head -> origin/gh/weifengpy/36/head 2025-11-03T16:58:01.9772829Z * [new branch] gh/weifengpy/36/orig -> origin/gh/weifengpy/36/orig 2025-11-03T16:58:01.9775233Z * [new branch] gh/weifengpy/37/base -> origin/gh/weifengpy/37/base 2025-11-03T16:58:01.9777069Z * [new branch] gh/weifengpy/37/head -> origin/gh/weifengpy/37/head 2025-11-03T16:58:01.9778792Z * [new branch] gh/weifengpy/37/orig -> origin/gh/weifengpy/37/orig 2025-11-03T16:58:01.9781180Z * [new branch] gh/weifengpy/38/base -> origin/gh/weifengpy/38/base 2025-11-03T16:58:01.9783527Z * [new branch] gh/weifengpy/38/head -> origin/gh/weifengpy/38/head 2025-11-03T16:58:01.9785291Z * [new branch] gh/weifengpy/38/orig -> origin/gh/weifengpy/38/orig 2025-11-03T16:58:01.9787621Z * [new branch] gh/weifengpy/39/base -> origin/gh/weifengpy/39/base 2025-11-03T16:58:01.9789322Z * [new branch] gh/weifengpy/39/head -> origin/gh/weifengpy/39/head 2025-11-03T16:58:01.9791113Z * [new branch] gh/weifengpy/39/orig -> origin/gh/weifengpy/39/orig 2025-11-03T16:58:01.9793935Z * [new branch] gh/williamwen42/250/base -> origin/gh/williamwen42/250/base 2025-11-03T16:58:01.9795715Z * [new branch] gh/williamwen42/250/head -> origin/gh/williamwen42/250/head 2025-11-03T16:58:01.9797418Z * [new branch] gh/williamwen42/250/orig -> origin/gh/williamwen42/250/orig 2025-11-03T16:58:01.9799979Z * [new branch] gh/williamwen42/278/base -> origin/gh/williamwen42/278/base 2025-11-03T16:58:01.9801655Z * [new branch] gh/williamwen42/278/head -> origin/gh/williamwen42/278/head 2025-11-03T16:58:01.9803405Z * [new branch] gh/williamwen42/278/orig -> origin/gh/williamwen42/278/orig 2025-11-03T16:58:01.9805881Z * [new branch] gh/williamwen42/279/base -> origin/gh/williamwen42/279/base 2025-11-03T16:58:01.9807786Z * [new branch] gh/williamwen42/279/head -> origin/gh/williamwen42/279/head 2025-11-03T16:58:01.9809487Z * [new branch] gh/williamwen42/279/orig -> origin/gh/williamwen42/279/orig 2025-11-03T16:58:01.9811741Z * [new branch] gh/williamwen42/282/base -> origin/gh/williamwen42/282/base 2025-11-03T16:58:01.9813453Z * [new branch] gh/williamwen42/282/head -> origin/gh/williamwen42/282/head 2025-11-03T16:58:01.9815121Z * [new branch] gh/williamwen42/282/orig -> origin/gh/williamwen42/282/orig 2025-11-03T16:58:01.9817791Z * [new branch] gh/williamwen42/287/base -> origin/gh/williamwen42/287/base 2025-11-03T16:58:01.9819585Z * [new branch] gh/williamwen42/287/head -> origin/gh/williamwen42/287/head 2025-11-03T16:58:01.9821331Z * [new branch] gh/williamwen42/287/orig -> origin/gh/williamwen42/287/orig 2025-11-03T16:58:01.9824076Z * [new branch] gh/williamwen42/288/base -> origin/gh/williamwen42/288/base 2025-11-03T16:58:01.9825817Z * [new branch] gh/williamwen42/288/head -> origin/gh/williamwen42/288/head 2025-11-03T16:58:01.9827500Z * [new branch] gh/williamwen42/288/orig -> origin/gh/williamwen42/288/orig 2025-11-03T16:58:01.9829959Z * [new branch] gh/williamwen42/290/base -> origin/gh/williamwen42/290/base 2025-11-03T16:58:01.9831615Z * [new branch] gh/williamwen42/290/head -> origin/gh/williamwen42/290/head 2025-11-03T16:58:01.9833384Z * [new branch] gh/williamwen42/290/orig -> origin/gh/williamwen42/290/orig 2025-11-03T16:58:01.9836272Z * [new branch] gh/williamwen42/296/base -> origin/gh/williamwen42/296/base 2025-11-03T16:58:01.9837878Z * [new branch] gh/williamwen42/296/head -> origin/gh/williamwen42/296/head 2025-11-03T16:58:01.9839540Z * [new branch] gh/williamwen42/296/orig -> origin/gh/williamwen42/296/orig 2025-11-03T16:58:01.9841846Z * [new branch] gh/williamwen42/297/base -> origin/gh/williamwen42/297/base 2025-11-03T16:58:01.9843511Z * [new branch] gh/williamwen42/297/head -> origin/gh/williamwen42/297/head 2025-11-03T16:58:01.9845248Z * [new branch] gh/williamwen42/297/orig -> origin/gh/williamwen42/297/orig 2025-11-03T16:58:01.9847741Z * [new branch] gh/williamwen42/305/base -> origin/gh/williamwen42/305/base 2025-11-03T16:58:01.9849483Z * [new branch] gh/williamwen42/305/head -> origin/gh/williamwen42/305/head 2025-11-03T16:58:01.9851153Z * [new branch] gh/williamwen42/305/orig -> origin/gh/williamwen42/305/orig 2025-11-03T16:58:01.9853440Z * [new branch] gh/williamwen42/306/base -> origin/gh/williamwen42/306/base 2025-11-03T16:58:01.9855250Z * [new branch] gh/williamwen42/306/head -> origin/gh/williamwen42/306/head 2025-11-03T16:58:01.9856981Z * [new branch] gh/williamwen42/306/orig -> origin/gh/williamwen42/306/orig 2025-11-03T16:58:01.9859265Z * [new branch] gh/williamwen42/307/base -> origin/gh/williamwen42/307/base 2025-11-03T16:58:01.9860983Z * [new branch] gh/williamwen42/307/head -> origin/gh/williamwen42/307/head 2025-11-03T16:58:01.9862938Z * [new branch] gh/williamwen42/307/orig -> origin/gh/williamwen42/307/orig 2025-11-03T16:58:01.9865237Z * [new branch] gh/williamwen42/308/base -> origin/gh/williamwen42/308/base 2025-11-03T16:58:01.9866928Z * [new branch] gh/williamwen42/308/head -> origin/gh/williamwen42/308/head 2025-11-03T16:58:01.9869073Z * [new branch] gh/williamwen42/308/orig -> origin/gh/williamwen42/308/orig 2025-11-03T16:58:01.9871601Z * [new branch] gh/williamwen42/309/base -> origin/gh/williamwen42/309/base 2025-11-03T16:58:01.9873335Z * [new branch] gh/williamwen42/309/head -> origin/gh/williamwen42/309/head 2025-11-03T16:58:01.9875043Z * [new branch] gh/williamwen42/309/orig -> origin/gh/williamwen42/309/orig 2025-11-03T16:58:01.9877716Z * [new branch] gh/williamwen42/310/base -> origin/gh/williamwen42/310/base 2025-11-03T16:58:01.9879503Z * [new branch] gh/williamwen42/310/head -> origin/gh/williamwen42/310/head 2025-11-03T16:58:01.9881209Z * [new branch] gh/williamwen42/310/orig -> origin/gh/williamwen42/310/orig 2025-11-03T16:58:01.9884482Z * [new branch] gh/williamwen42/311/base -> origin/gh/williamwen42/311/base 2025-11-03T16:58:01.9886265Z * [new branch] gh/williamwen42/311/head -> origin/gh/williamwen42/311/head 2025-11-03T16:58:01.9887941Z * [new branch] gh/williamwen42/311/orig -> origin/gh/williamwen42/311/orig 2025-11-03T16:58:01.9890766Z * [new branch] gh/williamwen42/312/base -> origin/gh/williamwen42/312/base 2025-11-03T16:58:01.9892616Z * [new branch] gh/williamwen42/312/head -> origin/gh/williamwen42/312/head 2025-11-03T16:58:01.9894612Z * [new branch] gh/williamwen42/312/orig -> origin/gh/williamwen42/312/orig 2025-11-03T16:58:01.9897185Z * [new branch] gh/williamwen42/313/base -> origin/gh/williamwen42/313/base 2025-11-03T16:58:01.9898928Z * [new branch] gh/williamwen42/313/head -> origin/gh/williamwen42/313/head 2025-11-03T16:58:01.9900674Z * [new branch] gh/williamwen42/313/orig -> origin/gh/williamwen42/313/orig 2025-11-03T16:58:01.9902985Z * [new branch] gh/williamwen42/314/base -> origin/gh/williamwen42/314/base 2025-11-03T16:58:01.9904811Z * [new branch] gh/williamwen42/314/head -> origin/gh/williamwen42/314/head 2025-11-03T16:58:01.9906317Z * [new branch] gh/williamwen42/314/orig -> origin/gh/williamwen42/314/orig 2025-11-03T16:58:01.9908620Z * [new branch] gh/williamwen42/315/base -> origin/gh/williamwen42/315/base 2025-11-03T16:58:01.9910328Z * [new branch] gh/williamwen42/315/head -> origin/gh/williamwen42/315/head 2025-11-03T16:58:01.9912083Z * [new branch] gh/williamwen42/315/orig -> origin/gh/williamwen42/315/orig 2025-11-03T16:58:01.9914449Z * [new branch] gh/williamwen42/316/base -> origin/gh/williamwen42/316/base 2025-11-03T16:58:01.9916229Z * [new branch] gh/williamwen42/316/head -> origin/gh/williamwen42/316/head 2025-11-03T16:58:01.9917927Z * [new branch] gh/williamwen42/316/orig -> origin/gh/williamwen42/316/orig 2025-11-03T16:58:01.9920529Z * [new branch] gh/williamwen42/317/base -> origin/gh/williamwen42/317/base 2025-11-03T16:58:01.9922460Z * [new branch] gh/williamwen42/317/head -> origin/gh/williamwen42/317/head 2025-11-03T16:58:01.9924152Z * [new branch] gh/williamwen42/317/orig -> origin/gh/williamwen42/317/orig 2025-11-03T16:58:01.9926378Z * [new branch] gh/williamwen42/318/base -> origin/gh/williamwen42/318/base 2025-11-03T16:58:01.9928234Z * [new branch] gh/williamwen42/318/head -> origin/gh/williamwen42/318/head 2025-11-03T16:58:01.9929863Z * [new branch] gh/williamwen42/318/orig -> origin/gh/williamwen42/318/orig 2025-11-03T16:58:01.9932104Z * [new branch] gh/williamwen42/319/base -> origin/gh/williamwen42/319/base 2025-11-03T16:58:01.9933878Z * [new branch] gh/williamwen42/319/head -> origin/gh/williamwen42/319/head 2025-11-03T16:58:01.9935605Z * [new branch] gh/williamwen42/319/orig -> origin/gh/williamwen42/319/orig 2025-11-03T16:58:01.9938011Z * [new branch] gh/williamwen42/320/base -> origin/gh/williamwen42/320/base 2025-11-03T16:58:01.9939700Z * [new branch] gh/williamwen42/320/head -> origin/gh/williamwen42/320/head 2025-11-03T16:58:01.9941391Z * [new branch] gh/williamwen42/320/orig -> origin/gh/williamwen42/320/orig 2025-11-03T16:58:01.9943995Z * [new branch] gh/williamwen42/321/base -> origin/gh/williamwen42/321/base 2025-11-03T16:58:01.9945708Z * [new branch] gh/williamwen42/321/head -> origin/gh/williamwen42/321/head 2025-11-03T16:58:01.9947387Z * [new branch] gh/williamwen42/321/orig -> origin/gh/williamwen42/321/orig 2025-11-03T16:58:01.9949774Z * [new branch] gh/williamwen42/322/base -> origin/gh/williamwen42/322/base 2025-11-03T16:58:01.9951478Z * [new branch] gh/williamwen42/322/head -> origin/gh/williamwen42/322/head 2025-11-03T16:58:01.9953225Z * [new branch] gh/williamwen42/322/orig -> origin/gh/williamwen42/322/orig 2025-11-03T16:58:01.9955634Z * [new branch] gh/williamwen42/323/base -> origin/gh/williamwen42/323/base 2025-11-03T16:58:01.9957342Z * [new branch] gh/williamwen42/323/head -> origin/gh/williamwen42/323/head 2025-11-03T16:58:01.9959077Z * [new branch] gh/williamwen42/323/orig -> origin/gh/williamwen42/323/orig 2025-11-03T16:58:01.9961446Z * [new branch] gh/williamwen42/324/base -> origin/gh/williamwen42/324/base 2025-11-03T16:58:01.9963394Z * [new branch] gh/williamwen42/324/head -> origin/gh/williamwen42/324/head 2025-11-03T16:58:01.9965131Z * [new branch] gh/williamwen42/324/orig -> origin/gh/williamwen42/324/orig 2025-11-03T16:58:01.9967607Z * [new branch] gh/williamwen42/325/base -> origin/gh/williamwen42/325/base 2025-11-03T16:58:01.9969280Z * [new branch] gh/williamwen42/325/head -> origin/gh/williamwen42/325/head 2025-11-03T16:58:01.9971112Z * [new branch] gh/williamwen42/325/orig -> origin/gh/williamwen42/325/orig 2025-11-03T16:58:01.9973431Z * [new branch] gh/williamwen42/326/base -> origin/gh/williamwen42/326/base 2025-11-03T16:58:01.9975191Z * [new branch] gh/williamwen42/326/head -> origin/gh/williamwen42/326/head 2025-11-03T16:58:01.9976945Z * [new branch] gh/williamwen42/326/orig -> origin/gh/williamwen42/326/orig 2025-11-03T16:58:01.9979355Z * [new branch] gh/williamwen42/327/base -> origin/gh/williamwen42/327/base 2025-11-03T16:58:01.9981118Z * [new branch] gh/williamwen42/327/head -> origin/gh/williamwen42/327/head 2025-11-03T16:58:01.9983236Z * [new branch] gh/williamwen42/327/orig -> origin/gh/williamwen42/327/orig 2025-11-03T16:58:01.9985636Z * [new branch] gh/williamwen42/328/base -> origin/gh/williamwen42/328/base 2025-11-03T16:58:01.9987385Z * [new branch] gh/williamwen42/328/head -> origin/gh/williamwen42/328/head 2025-11-03T16:58:01.9989212Z * [new branch] gh/williamwen42/328/orig -> origin/gh/williamwen42/328/orig 2025-11-03T16:58:01.9991632Z * [new branch] gh/williamwen42/329/base -> origin/gh/williamwen42/329/base 2025-11-03T16:58:01.9993315Z * [new branch] gh/williamwen42/329/head -> origin/gh/williamwen42/329/head 2025-11-03T16:58:01.9995002Z * [new branch] gh/williamwen42/329/orig -> origin/gh/williamwen42/329/orig 2025-11-03T16:58:01.9997767Z * [new branch] gh/xmfan/169/base -> origin/gh/xmfan/169/base 2025-11-03T16:58:01.9999419Z * [new branch] gh/xmfan/169/head -> origin/gh/xmfan/169/head 2025-11-03T16:58:02.0002084Z * [new branch] gh/xmfan/170/base -> origin/gh/xmfan/170/base 2025-11-03T16:58:02.0003754Z * [new branch] gh/xmfan/170/head -> origin/gh/xmfan/170/head 2025-11-03T16:58:02.0006132Z * [new branch] gh/xmfan/274/base -> origin/gh/xmfan/274/base 2025-11-03T16:58:02.0007775Z * [new branch] gh/xmfan/274/head -> origin/gh/xmfan/274/head 2025-11-03T16:58:02.0009441Z * [new branch] gh/xmfan/274/orig -> origin/gh/xmfan/274/orig 2025-11-03T16:58:02.0011775Z * [new branch] gh/xmfan/277/base -> origin/gh/xmfan/277/base 2025-11-03T16:58:02.0013539Z * [new branch] gh/xmfan/277/head -> origin/gh/xmfan/277/head 2025-11-03T16:58:02.0015194Z * [new branch] gh/xmfan/277/orig -> origin/gh/xmfan/277/orig 2025-11-03T16:58:02.0017533Z * [new branch] gh/xmfan/289/base -> origin/gh/xmfan/289/base 2025-11-03T16:58:02.0019357Z * [new branch] gh/xmfan/289/head -> origin/gh/xmfan/289/head 2025-11-03T16:58:02.0021016Z * [new branch] gh/xmfan/289/orig -> origin/gh/xmfan/289/orig 2025-11-03T16:58:02.0023913Z * [new branch] gh/xmfan/291/base -> origin/gh/xmfan/291/base 2025-11-03T16:58:02.0025663Z * [new branch] gh/xmfan/291/head -> origin/gh/xmfan/291/head 2025-11-03T16:58:02.0027355Z * [new branch] gh/xmfan/291/orig -> origin/gh/xmfan/291/orig 2025-11-03T16:58:02.0029778Z * [new branch] gh/xmfan/292/base -> origin/gh/xmfan/292/base 2025-11-03T16:58:02.0032271Z * [new branch] gh/xmfan/292/head -> origin/gh/xmfan/292/head 2025-11-03T16:58:02.0034208Z * [new branch] gh/xmfan/292/orig -> origin/gh/xmfan/292/orig 2025-11-03T16:58:02.0036705Z * [new branch] gh/xmfan/295/base -> origin/gh/xmfan/295/base 2025-11-03T16:58:02.0038774Z * [new branch] gh/xmfan/295/head -> origin/gh/xmfan/295/head 2025-11-03T16:58:02.0040462Z * [new branch] gh/xmfan/295/orig -> origin/gh/xmfan/295/orig 2025-11-03T16:58:02.0043342Z * [new branch] gh/xmfan/296/base -> origin/gh/xmfan/296/base 2025-11-03T16:58:02.0044974Z * [new branch] gh/xmfan/296/head -> origin/gh/xmfan/296/head 2025-11-03T16:58:02.0046647Z * [new branch] gh/xmfan/296/orig -> origin/gh/xmfan/296/orig 2025-11-03T16:58:02.0048971Z * [new branch] gh/xmfan/297/base -> origin/gh/xmfan/297/base 2025-11-03T16:58:02.0050760Z * [new branch] gh/xmfan/297/head -> origin/gh/xmfan/297/head 2025-11-03T16:58:02.0052351Z * [new branch] gh/xmfan/297/orig -> origin/gh/xmfan/297/orig 2025-11-03T16:58:02.0054669Z * [new branch] gh/xmfan/298/base -> origin/gh/xmfan/298/base 2025-11-03T16:58:02.0056384Z * [new branch] gh/xmfan/298/head -> origin/gh/xmfan/298/head 2025-11-03T16:58:02.0058072Z * [new branch] gh/xmfan/298/orig -> origin/gh/xmfan/298/orig 2025-11-03T16:58:02.0060381Z * [new branch] gh/xmfan/299/base -> origin/gh/xmfan/299/base 2025-11-03T16:58:02.0062386Z * [new branch] gh/xmfan/299/head -> origin/gh/xmfan/299/head 2025-11-03T16:58:02.0064299Z * [new branch] gh/xmfan/299/orig -> origin/gh/xmfan/299/orig 2025-11-03T16:58:02.0066531Z * [new branch] gh/xmfan/300/base -> origin/gh/xmfan/300/base 2025-11-03T16:58:02.0068235Z * [new branch] gh/xmfan/300/head -> origin/gh/xmfan/300/head 2025-11-03T16:58:02.0069938Z * [new branch] gh/xmfan/300/orig -> origin/gh/xmfan/300/orig 2025-11-03T16:58:02.0072182Z * [new branch] gh/xmfan/301/base -> origin/gh/xmfan/301/base 2025-11-03T16:58:02.0073781Z * [new branch] gh/xmfan/301/head -> origin/gh/xmfan/301/head 2025-11-03T16:58:02.0075456Z * [new branch] gh/xmfan/301/orig -> origin/gh/xmfan/301/orig 2025-11-03T16:58:02.0077817Z * [new branch] gh/xmfan/302/base -> origin/gh/xmfan/302/base 2025-11-03T16:58:02.0079520Z * [new branch] gh/xmfan/302/head -> origin/gh/xmfan/302/head 2025-11-03T16:58:02.0081217Z * [new branch] gh/xmfan/302/orig -> origin/gh/xmfan/302/orig 2025-11-03T16:58:02.0083577Z * [new branch] gh/xmfan/303/base -> origin/gh/xmfan/303/base 2025-11-03T16:58:02.0085423Z * [new branch] gh/xmfan/303/head -> origin/gh/xmfan/303/head 2025-11-03T16:58:02.0087320Z * [new branch] gh/xmfan/303/orig -> origin/gh/xmfan/303/orig 2025-11-03T16:58:02.0090114Z * [new branch] gh/xmfan/304/base -> origin/gh/xmfan/304/base 2025-11-03T16:58:02.0091814Z * [new branch] gh/xmfan/304/head -> origin/gh/xmfan/304/head 2025-11-03T16:58:02.0093554Z * [new branch] gh/xmfan/304/orig -> origin/gh/xmfan/304/orig 2025-11-03T16:58:02.0095806Z * [new branch] gh/xmfan/305/base -> origin/gh/xmfan/305/base 2025-11-03T16:58:02.0097483Z * [new branch] gh/xmfan/305/head -> origin/gh/xmfan/305/head 2025-11-03T16:58:02.0099221Z * [new branch] gh/xmfan/305/orig -> origin/gh/xmfan/305/orig 2025-11-03T16:58:02.0101417Z * [new branch] gh/xmfan/306/base -> origin/gh/xmfan/306/base 2025-11-03T16:58:02.0103450Z * [new branch] gh/xmfan/306/head -> origin/gh/xmfan/306/head 2025-11-03T16:58:02.0105033Z * [new branch] gh/xmfan/306/orig -> origin/gh/xmfan/306/orig 2025-11-03T16:58:02.0107504Z * [new branch] gh/xmfan/307/base -> origin/gh/xmfan/307/base 2025-11-03T16:58:02.0109670Z * [new branch] gh/xmfan/307/head -> origin/gh/xmfan/307/head 2025-11-03T16:58:02.0111380Z * [new branch] gh/xmfan/307/orig -> origin/gh/xmfan/307/orig 2025-11-03T16:58:02.0114386Z * [new branch] gh/xmfan/308/base -> origin/gh/xmfan/308/base 2025-11-03T16:58:02.0116236Z * [new branch] gh/xmfan/308/head -> origin/gh/xmfan/308/head 2025-11-03T16:58:02.0117906Z * [new branch] gh/xmfan/308/orig -> origin/gh/xmfan/308/orig 2025-11-03T16:58:02.0120296Z * [new branch] gh/xmfan/309/base -> origin/gh/xmfan/309/base 2025-11-03T16:58:02.0122631Z * [new branch] gh/xmfan/309/head -> origin/gh/xmfan/309/head 2025-11-03T16:58:02.0124431Z * [new branch] gh/xmfan/309/orig -> origin/gh/xmfan/309/orig 2025-11-03T16:58:02.0127142Z * [new branch] gh/xuanzhang816/22/base -> origin/gh/xuanzhang816/22/base 2025-11-03T16:58:02.0128823Z * [new branch] gh/xuanzhang816/22/head -> origin/gh/xuanzhang816/22/head 2025-11-03T16:58:02.0130519Z * [new branch] gh/xuanzhang816/22/orig -> origin/gh/xuanzhang816/22/orig 2025-11-03T16:58:02.0133018Z * [new branch] gh/xuanzhang816/23/base -> origin/gh/xuanzhang816/23/base 2025-11-03T16:58:02.0134735Z * [new branch] gh/xuanzhang816/23/head -> origin/gh/xuanzhang816/23/head 2025-11-03T16:58:02.0136760Z * [new branch] gh/xuanzhang816/23/orig -> origin/gh/xuanzhang816/23/orig 2025-11-03T16:58:02.0139058Z * [new branch] gh/xuanzhang816/27/base -> origin/gh/xuanzhang816/27/base 2025-11-03T16:58:02.0140819Z * [new branch] gh/xuanzhang816/27/head -> origin/gh/xuanzhang816/27/head 2025-11-03T16:58:02.0142628Z * [new branch] gh/xuanzhang816/27/orig -> origin/gh/xuanzhang816/27/orig 2025-11-03T16:58:02.0144985Z * [new branch] gh/xuanzhang816/32/base -> origin/gh/xuanzhang816/32/base 2025-11-03T16:58:02.0146635Z * [new branch] gh/xuanzhang816/32/head -> origin/gh/xuanzhang816/32/head 2025-11-03T16:58:02.0148357Z * [new branch] gh/xuanzhang816/32/orig -> origin/gh/xuanzhang816/32/orig 2025-11-03T16:58:02.0150608Z * [new branch] gh/xuanzhang816/33/base -> origin/gh/xuanzhang816/33/base 2025-11-03T16:58:02.0152284Z * [new branch] gh/xuanzhang816/33/head -> origin/gh/xuanzhang816/33/head 2025-11-03T16:58:02.0153993Z * [new branch] gh/xuanzhang816/33/orig -> origin/gh/xuanzhang816/33/orig 2025-11-03T16:58:02.0157314Z * [new branch] gh/yanbing-j/11/base -> origin/gh/yanbing-j/11/base 2025-11-03T16:58:02.0159027Z * [new branch] gh/yanbing-j/11/head -> origin/gh/yanbing-j/11/head 2025-11-03T16:58:02.0160758Z * [new branch] gh/yanbing-j/11/orig -> origin/gh/yanbing-j/11/orig 2025-11-03T16:58:02.0163129Z * [new branch] gh/yanbing-j/12/base -> origin/gh/yanbing-j/12/base 2025-11-03T16:58:02.0165089Z * [new branch] gh/yanbing-j/12/head -> origin/gh/yanbing-j/12/head 2025-11-03T16:58:02.0166797Z * [new branch] gh/yanbing-j/12/orig -> origin/gh/yanbing-j/12/orig 2025-11-03T16:58:02.0169600Z * [new branch] gh/yanbing-j/13/base -> origin/gh/yanbing-j/13/base 2025-11-03T16:58:02.0171342Z * [new branch] gh/yanbing-j/13/head -> origin/gh/yanbing-j/13/head 2025-11-03T16:58:02.0172946Z * [new branch] gh/yanbing-j/13/orig -> origin/gh/yanbing-j/13/orig 2025-11-03T16:58:02.0175217Z * [new branch] gh/yanbing-j/14/base -> origin/gh/yanbing-j/14/base 2025-11-03T16:58:02.0176967Z * [new branch] gh/yanbing-j/14/head -> origin/gh/yanbing-j/14/head 2025-11-03T16:58:02.0178679Z * [new branch] gh/yanbing-j/14/orig -> origin/gh/yanbing-j/14/orig 2025-11-03T16:58:02.0180987Z * [new branch] gh/yanbing-j/15/base -> origin/gh/yanbing-j/15/base 2025-11-03T16:58:02.0183326Z * [new branch] gh/yanbing-j/15/head -> origin/gh/yanbing-j/15/head 2025-11-03T16:58:02.0185208Z * [new branch] gh/yanbing-j/15/orig -> origin/gh/yanbing-j/15/orig 2025-11-03T16:58:02.0187298Z * [new branch] gh/yanbing-j/18/base -> origin/gh/yanbing-j/18/base 2025-11-03T16:58:02.0188982Z * [new branch] gh/yanbing-j/18/head -> origin/gh/yanbing-j/18/head 2025-11-03T16:58:02.0190741Z * [new branch] gh/yanbing-j/18/orig -> origin/gh/yanbing-j/18/orig 2025-11-03T16:58:02.0193095Z * [new branch] gh/yanbing-j/19/base -> origin/gh/yanbing-j/19/base 2025-11-03T16:58:02.0194807Z * [new branch] gh/yanbing-j/19/head -> origin/gh/yanbing-j/19/head 2025-11-03T16:58:02.0196455Z * [new branch] gh/yanbing-j/19/orig -> origin/gh/yanbing-j/19/orig 2025-11-03T16:58:02.0198807Z * [new branch] gh/yanbing-j/20/base -> origin/gh/yanbing-j/20/base 2025-11-03T16:58:02.0200478Z * [new branch] gh/yanbing-j/20/head -> origin/gh/yanbing-j/20/head 2025-11-03T16:58:02.0202171Z * [new branch] gh/yanbing-j/20/orig -> origin/gh/yanbing-j/20/orig 2025-11-03T16:58:02.0204806Z * [new branch] gh/yanbing-j/21/base -> origin/gh/yanbing-j/21/base 2025-11-03T16:58:02.0206528Z * [new branch] gh/yanbing-j/21/head -> origin/gh/yanbing-j/21/head 2025-11-03T16:58:02.0209259Z * [new branch] gh/yanbing-j/22/base -> origin/gh/yanbing-j/22/base 2025-11-03T16:58:02.0210994Z * [new branch] gh/yanbing-j/22/head -> origin/gh/yanbing-j/22/head 2025-11-03T16:58:02.0212623Z * [new branch] gh/yanbing-j/22/orig -> origin/gh/yanbing-j/22/orig 2025-11-03T16:58:02.0214843Z * [new branch] gh/yanbing-j/23/base -> origin/gh/yanbing-j/23/base 2025-11-03T16:58:02.0216520Z * [new branch] gh/yanbing-j/23/head -> origin/gh/yanbing-j/23/head 2025-11-03T16:58:02.0218247Z * [new branch] gh/yanbing-j/23/orig -> origin/gh/yanbing-j/23/orig 2025-11-03T16:58:02.0220929Z * [new branch] gh/yanbing-j/24/base -> origin/gh/yanbing-j/24/base 2025-11-03T16:58:02.0223055Z * [new branch] gh/yanbing-j/24/head -> origin/gh/yanbing-j/24/head 2025-11-03T16:58:02.0224717Z * [new branch] gh/yanbing-j/24/orig -> origin/gh/yanbing-j/24/orig 2025-11-03T16:58:02.0227123Z * [new branch] gh/yanbing-j/25/base -> origin/gh/yanbing-j/25/base 2025-11-03T16:58:02.0228984Z * [new branch] gh/yanbing-j/25/head -> origin/gh/yanbing-j/25/head 2025-11-03T16:58:02.0230668Z * [new branch] gh/yanbing-j/25/orig -> origin/gh/yanbing-j/25/orig 2025-11-03T16:58:02.0232867Z * [new branch] gh/yanbing-j/26/base -> origin/gh/yanbing-j/26/base 2025-11-03T16:58:02.0234584Z * [new branch] gh/yanbing-j/26/head -> origin/gh/yanbing-j/26/head 2025-11-03T16:58:02.0236256Z * [new branch] gh/yanbing-j/26/orig -> origin/gh/yanbing-j/26/orig 2025-11-03T16:58:02.0239042Z * [new branch] gh/yangw-dev/12/base -> origin/gh/yangw-dev/12/base 2025-11-03T16:58:02.0240733Z * [new branch] gh/yangw-dev/12/head -> origin/gh/yangw-dev/12/head 2025-11-03T16:58:02.0242474Z * [new branch] gh/yangw-dev/12/orig -> origin/gh/yangw-dev/12/orig 2025-11-03T16:58:02.0244707Z * [new branch] gh/yangw-dev/13/base -> origin/gh/yangw-dev/13/base 2025-11-03T16:58:02.0246522Z * [new branch] gh/yangw-dev/13/head -> origin/gh/yangw-dev/13/head 2025-11-03T16:58:02.0248236Z * [new branch] gh/yangw-dev/13/orig -> origin/gh/yangw-dev/13/orig 2025-11-03T16:58:02.0250649Z * [new branch] gh/yangw-dev/14/base -> origin/gh/yangw-dev/14/base 2025-11-03T16:58:02.0252304Z * [new branch] gh/yangw-dev/14/head -> origin/gh/yangw-dev/14/head 2025-11-03T16:58:02.0254105Z * [new branch] gh/yangw-dev/14/orig -> origin/gh/yangw-dev/14/orig 2025-11-03T16:58:02.0256335Z * [new branch] gh/yangw-dev/15/base -> origin/gh/yangw-dev/15/base 2025-11-03T16:58:02.0258117Z * [new branch] gh/yangw-dev/15/head -> origin/gh/yangw-dev/15/head 2025-11-03T16:58:02.0259771Z * [new branch] gh/yangw-dev/15/orig -> origin/gh/yangw-dev/15/orig 2025-11-03T16:58:02.0262165Z * [new branch] gh/yangw-dev/19/base -> origin/gh/yangw-dev/19/base 2025-11-03T16:58:02.0263959Z * [new branch] gh/yangw-dev/19/head -> origin/gh/yangw-dev/19/head 2025-11-03T16:58:02.0265658Z * [new branch] gh/yangw-dev/19/orig -> origin/gh/yangw-dev/19/orig 2025-11-03T16:58:02.0267983Z * [new branch] gh/yangw-dev/26/base -> origin/gh/yangw-dev/26/base 2025-11-03T16:58:02.0269704Z * [new branch] gh/yangw-dev/26/head -> origin/gh/yangw-dev/26/head 2025-11-03T16:58:02.0271456Z * [new branch] gh/yangw-dev/26/orig -> origin/gh/yangw-dev/26/orig 2025-11-03T16:58:02.0273805Z * [new branch] gh/yangw-dev/27/base -> origin/gh/yangw-dev/27/base 2025-11-03T16:58:02.0275504Z * [new branch] gh/yangw-dev/27/head -> origin/gh/yangw-dev/27/head 2025-11-03T16:58:02.0277202Z * [new branch] gh/yangw-dev/27/orig -> origin/gh/yangw-dev/27/orig 2025-11-03T16:58:02.0280240Z * [new branch] gh/ydwu4/283/base -> origin/gh/ydwu4/283/base 2025-11-03T16:58:02.0281966Z * [new branch] gh/ydwu4/283/head -> origin/gh/ydwu4/283/head 2025-11-03T16:58:02.0283637Z * [new branch] gh/ydwu4/283/orig -> origin/gh/ydwu4/283/orig 2025-11-03T16:58:02.0285964Z * [new branch] gh/ydwu4/292/base -> origin/gh/ydwu4/292/base 2025-11-03T16:58:02.0287665Z * [new branch] gh/ydwu4/292/head -> origin/gh/ydwu4/292/head 2025-11-03T16:58:02.0289331Z * [new branch] gh/ydwu4/292/orig -> origin/gh/ydwu4/292/orig 2025-11-03T16:58:02.0291686Z * [new branch] gh/ydwu4/294/base -> origin/gh/ydwu4/294/base 2025-11-03T16:58:02.0293414Z * [new branch] gh/ydwu4/294/head -> origin/gh/ydwu4/294/head 2025-11-03T16:58:02.0295196Z * [new branch] gh/ydwu4/294/orig -> origin/gh/ydwu4/294/orig 2025-11-03T16:58:02.0297652Z * [new branch] gh/ydwu4/295/base -> origin/gh/ydwu4/295/base 2025-11-03T16:58:02.0299366Z * [new branch] gh/ydwu4/295/head -> origin/gh/ydwu4/295/head 2025-11-03T16:58:02.0301127Z * [new branch] gh/ydwu4/295/orig -> origin/gh/ydwu4/295/orig 2025-11-03T16:58:02.0303587Z * [new branch] gh/ydwu4/296/base -> origin/gh/ydwu4/296/base 2025-11-03T16:58:02.0305139Z * [new branch] gh/ydwu4/296/head -> origin/gh/ydwu4/296/head 2025-11-03T16:58:02.0306787Z * [new branch] gh/ydwu4/296/orig -> origin/gh/ydwu4/296/orig 2025-11-03T16:58:02.0309121Z * [new branch] gh/ydwu4/306/base -> origin/gh/ydwu4/306/base 2025-11-03T16:58:02.0310933Z * [new branch] gh/ydwu4/306/head -> origin/gh/ydwu4/306/head 2025-11-03T16:58:02.0312716Z * [new branch] gh/ydwu4/306/orig -> origin/gh/ydwu4/306/orig 2025-11-03T16:58:02.0314986Z * [new branch] gh/ydwu4/312/base -> origin/gh/ydwu4/312/base 2025-11-03T16:58:02.0316680Z * [new branch] gh/ydwu4/312/head -> origin/gh/ydwu4/312/head 2025-11-03T16:58:02.0318445Z * [new branch] gh/ydwu4/312/orig -> origin/gh/ydwu4/312/orig 2025-11-03T16:58:02.0320688Z * [new branch] gh/ydwu4/322/base -> origin/gh/ydwu4/322/base 2025-11-03T16:58:02.0322755Z * [new branch] gh/ydwu4/322/head -> origin/gh/ydwu4/322/head 2025-11-03T16:58:02.0324318Z * [new branch] gh/ydwu4/322/orig -> origin/gh/ydwu4/322/orig 2025-11-03T16:58:02.0326631Z * [new branch] gh/ydwu4/326/base -> origin/gh/ydwu4/326/base 2025-11-03T16:58:02.0328367Z * [new branch] gh/ydwu4/326/head -> origin/gh/ydwu4/326/head 2025-11-03T16:58:02.0330049Z * [new branch] gh/ydwu4/326/orig -> origin/gh/ydwu4/326/orig 2025-11-03T16:58:02.0332304Z * [new branch] gh/ydwu4/327/base -> origin/gh/ydwu4/327/base 2025-11-03T16:58:02.0334071Z * [new branch] gh/ydwu4/327/head -> origin/gh/ydwu4/327/head 2025-11-03T16:58:02.0335724Z * [new branch] gh/ydwu4/327/orig -> origin/gh/ydwu4/327/orig 2025-11-03T16:58:02.0338261Z * [new branch] gh/ydwu4/328/base -> origin/gh/ydwu4/328/base 2025-11-03T16:58:02.0339953Z * [new branch] gh/ydwu4/328/head -> origin/gh/ydwu4/328/head 2025-11-03T16:58:02.0341866Z * [new branch] gh/ydwu4/328/orig -> origin/gh/ydwu4/328/orig 2025-11-03T16:58:02.0344034Z * [new branch] gh/ydwu4/329/base -> origin/gh/ydwu4/329/base 2025-11-03T16:58:02.0345698Z * [new branch] gh/ydwu4/329/head -> origin/gh/ydwu4/329/head 2025-11-03T16:58:02.0347362Z * [new branch] gh/ydwu4/329/orig -> origin/gh/ydwu4/329/orig 2025-11-03T16:58:02.0350142Z * [new branch] gh/ydwu4/330/base -> origin/gh/ydwu4/330/base 2025-11-03T16:58:02.0351766Z * [new branch] gh/ydwu4/330/head -> origin/gh/ydwu4/330/head 2025-11-03T16:58:02.0353477Z * [new branch] gh/ydwu4/330/orig -> origin/gh/ydwu4/330/orig 2025-11-03T16:58:02.0355678Z * [new branch] gh/ydwu4/331/base -> origin/gh/ydwu4/331/base 2025-11-03T16:58:02.0357393Z * [new branch] gh/ydwu4/331/head -> origin/gh/ydwu4/331/head 2025-11-03T16:58:02.0359021Z * [new branch] gh/ydwu4/331/orig -> origin/gh/ydwu4/331/orig 2025-11-03T16:58:02.0361207Z * [new branch] gh/ydwu4/332/base -> origin/gh/ydwu4/332/base 2025-11-03T16:58:02.0362964Z * [new branch] gh/ydwu4/332/head -> origin/gh/ydwu4/332/head 2025-11-03T16:58:02.0364720Z * [new branch] gh/ydwu4/332/orig -> origin/gh/ydwu4/332/orig 2025-11-03T16:58:02.0366878Z * [new branch] gh/ydwu4/333/base -> origin/gh/ydwu4/333/base 2025-11-03T16:58:02.0368675Z * [new branch] gh/ydwu4/333/head -> origin/gh/ydwu4/333/head 2025-11-03T16:58:02.0370351Z * [new branch] gh/ydwu4/333/orig -> origin/gh/ydwu4/333/orig 2025-11-03T16:58:02.0373087Z * [new branch] gh/ydwu4/334/base -> origin/gh/ydwu4/334/base 2025-11-03T16:58:02.0374826Z * [new branch] gh/ydwu4/334/head -> origin/gh/ydwu4/334/head 2025-11-03T16:58:02.0376520Z * [new branch] gh/ydwu4/334/orig -> origin/gh/ydwu4/334/orig 2025-11-03T16:58:02.0378764Z * [new branch] gh/ydwu4/335/base -> origin/gh/ydwu4/335/base 2025-11-03T16:58:02.0380539Z * [new branch] gh/ydwu4/335/head -> origin/gh/ydwu4/335/head 2025-11-03T16:58:02.0382347Z * [new branch] gh/ydwu4/335/orig -> origin/gh/ydwu4/335/orig 2025-11-03T16:58:02.0384630Z * [new branch] gh/ydwu4/336/base -> origin/gh/ydwu4/336/base 2025-11-03T16:58:02.0386297Z * [new branch] gh/ydwu4/336/head -> origin/gh/ydwu4/336/head 2025-11-03T16:58:02.0388112Z * [new branch] gh/ydwu4/336/orig -> origin/gh/ydwu4/336/orig 2025-11-03T16:58:02.0390506Z * [new branch] gh/ydwu4/337/base -> origin/gh/ydwu4/337/base 2025-11-03T16:58:02.0392322Z * [new branch] gh/ydwu4/337/head -> origin/gh/ydwu4/337/head 2025-11-03T16:58:02.0394653Z * [new branch] gh/ydwu4/337/orig -> origin/gh/ydwu4/337/orig 2025-11-03T16:58:02.0397065Z * [new branch] gh/ydwu4/338/base -> origin/gh/ydwu4/338/base 2025-11-03T16:58:02.0398803Z * [new branch] gh/ydwu4/338/head -> origin/gh/ydwu4/338/head 2025-11-03T16:58:02.0400537Z * [new branch] gh/ydwu4/338/orig -> origin/gh/ydwu4/338/orig 2025-11-03T16:58:02.0402710Z * [new branch] gh/ydwu4/339/base -> origin/gh/ydwu4/339/base 2025-11-03T16:58:02.0404453Z * [new branch] gh/ydwu4/339/head -> origin/gh/ydwu4/339/head 2025-11-03T16:58:02.0406127Z * [new branch] gh/ydwu4/339/orig -> origin/gh/ydwu4/339/orig 2025-11-03T16:58:02.0408506Z * [new branch] gh/ydwu4/340/base -> origin/gh/ydwu4/340/base 2025-11-03T16:58:02.0410256Z * [new branch] gh/ydwu4/340/head -> origin/gh/ydwu4/340/head 2025-11-03T16:58:02.0412191Z * [new branch] gh/ydwu4/340/orig -> origin/gh/ydwu4/340/orig 2025-11-03T16:58:02.0414970Z * [new branch] gh/yf225/133/base -> origin/gh/yf225/133/base 2025-11-03T16:58:02.0417207Z * [new branch] gh/yf225/133/head -> origin/gh/yf225/133/head 2025-11-03T16:58:02.0419491Z * [new branch] gh/yf225/93/base -> origin/gh/yf225/93/base 2025-11-03T16:58:02.0421363Z * [new branch] gh/yf225/93/head -> origin/gh/yf225/93/head 2025-11-03T16:58:02.0426098Z * [new branch] gh/yifuwang/152/base -> origin/gh/yifuwang/152/base 2025-11-03T16:58:02.0428125Z * [new branch] gh/yifuwang/152/head -> origin/gh/yifuwang/152/head 2025-11-03T16:58:02.0429975Z * [new branch] gh/yifuwang/152/orig -> origin/gh/yifuwang/152/orig 2025-11-03T16:58:02.0432161Z * [new branch] gh/yifuwang/195/base -> origin/gh/yifuwang/195/base 2025-11-03T16:58:02.0434028Z * [new branch] gh/yifuwang/195/head -> origin/gh/yifuwang/195/head 2025-11-03T16:58:02.0435835Z * [new branch] gh/yifuwang/195/orig -> origin/gh/yifuwang/195/orig 2025-11-03T16:58:02.0438717Z * [new branch] gh/yiming0416/1/base -> origin/gh/yiming0416/1/base 2025-11-03T16:58:02.0440464Z * [new branch] gh/yiming0416/1/head -> origin/gh/yiming0416/1/head 2025-11-03T16:58:02.0442751Z * [new branch] gh/yiming0416/2/base -> origin/gh/yiming0416/2/base 2025-11-03T16:58:02.0444357Z * [new branch] gh/yiming0416/2/head -> origin/gh/yiming0416/2/head 2025-11-03T16:58:02.0447231Z * [new branch] gh/yushangdi/1/base -> origin/gh/yushangdi/1/base 2025-11-03T16:58:02.0448939Z * [new branch] gh/yushangdi/1/head -> origin/gh/yushangdi/1/head 2025-11-03T16:58:02.0451109Z * [new branch] gh/yushangdi/2/base -> origin/gh/yushangdi/2/base 2025-11-03T16:58:02.0452727Z * [new branch] gh/yushangdi/2/head -> origin/gh/yushangdi/2/head 2025-11-03T16:58:02.0455034Z * [new branch] gh/yushangdi/3/base -> origin/gh/yushangdi/3/base 2025-11-03T16:58:02.0456715Z * [new branch] gh/yushangdi/3/head -> origin/gh/yushangdi/3/head 2025-11-03T16:58:02.0458362Z * [new branch] gh/yushangdi/3/orig -> origin/gh/yushangdi/3/orig 2025-11-03T16:58:02.0460763Z * [new branch] gh/yushangdi/4/base -> origin/gh/yushangdi/4/base 2025-11-03T16:58:02.0462772Z * [new branch] gh/yushangdi/4/head -> origin/gh/yushangdi/4/head 2025-11-03T16:58:02.0464691Z * [new branch] gh/yushangdi/4/orig -> origin/gh/yushangdi/4/orig 2025-11-03T16:58:02.0467029Z * [new branch] gh/yushangdi/5/base -> origin/gh/yushangdi/5/base 2025-11-03T16:58:02.0468963Z * [new branch] gh/yushangdi/5/head -> origin/gh/yushangdi/5/head 2025-11-03T16:58:02.0470581Z * [new branch] gh/yushangdi/5/orig -> origin/gh/yushangdi/5/orig 2025-11-03T16:58:02.0472522Z * [new branch] gh/yushangdi/6/base -> origin/gh/yushangdi/6/base 2025-11-03T16:58:02.0474118Z * [new branch] gh/yushangdi/6/head -> origin/gh/yushangdi/6/head 2025-11-03T16:58:02.0475751Z * [new branch] gh/yushangdi/6/orig -> origin/gh/yushangdi/6/orig 2025-11-03T16:58:02.0478162Z * [new branch] gh/yushangdi/7/base -> origin/gh/yushangdi/7/base 2025-11-03T16:58:02.0480148Z * [new branch] gh/yushangdi/7/head -> origin/gh/yushangdi/7/head 2025-11-03T16:58:02.0481957Z * [new branch] gh/yushangdi/7/orig -> origin/gh/yushangdi/7/orig 2025-11-03T16:58:02.0484354Z * [new branch] gh/yushangdi/8/base -> origin/gh/yushangdi/8/base 2025-11-03T16:58:02.0486073Z * [new branch] gh/yushangdi/8/head -> origin/gh/yushangdi/8/head 2025-11-03T16:58:02.0487746Z * [new branch] gh/yushangdi/8/orig -> origin/gh/yushangdi/8/orig 2025-11-03T16:58:02.0490569Z * [new branch] gh/zhxchen17/34/base -> origin/gh/zhxchen17/34/base 2025-11-03T16:58:02.0492299Z * [new branch] gh/zhxchen17/34/head -> origin/gh/zhxchen17/34/head 2025-11-03T16:58:02.0494623Z * [new branch] gh/zhxchen17/35/base -> origin/gh/zhxchen17/35/base 2025-11-03T16:58:02.0496372Z * [new branch] gh/zhxchen17/35/head -> origin/gh/zhxchen17/35/head 2025-11-03T16:58:02.0498794Z * [new branch] gh/zhxchen17/36/base -> origin/gh/zhxchen17/36/base 2025-11-03T16:58:02.0500667Z * [new branch] gh/zhxchen17/36/head -> origin/gh/zhxchen17/36/head 2025-11-03T16:58:02.0502423Z * [new branch] gh/zhxchen17/36/orig -> origin/gh/zhxchen17/36/orig 2025-11-03T16:58:02.0504670Z * [new branch] gh/zhxchen17/37/base -> origin/gh/zhxchen17/37/base 2025-11-03T16:58:02.0506424Z * [new branch] gh/zhxchen17/37/head -> origin/gh/zhxchen17/37/head 2025-11-03T16:58:02.0508115Z * [new branch] gh/zhxchen17/37/orig -> origin/gh/zhxchen17/37/orig 2025-11-03T16:58:02.0510355Z * [new branch] gh/zhxchen17/38/base -> origin/gh/zhxchen17/38/base 2025-11-03T16:58:02.0512251Z * [new branch] gh/zhxchen17/38/head -> origin/gh/zhxchen17/38/head 2025-11-03T16:58:02.0514079Z * [new branch] gh/zhxchen17/38/orig -> origin/gh/zhxchen17/38/orig 2025-11-03T16:58:02.0516845Z * [new branch] gh/zklaus/18/base -> origin/gh/zklaus/18/base 2025-11-03T16:58:02.0518562Z * [new branch] gh/zklaus/18/head -> origin/gh/zklaus/18/head 2025-11-03T16:58:02.0520309Z * [new branch] gh/zklaus/18/orig -> origin/gh/zklaus/18/orig 2025-11-03T16:58:02.0523447Z * [new branch] gh/zou3519/1197/base -> origin/gh/zou3519/1197/base 2025-11-03T16:58:02.0525040Z * [new branch] gh/zou3519/1197/head -> origin/gh/zou3519/1197/head 2025-11-03T16:58:02.0526726Z * [new branch] gh/zou3519/1197/orig -> origin/gh/zou3519/1197/orig 2025-11-03T16:58:02.0529109Z * [new branch] gh/zou3519/1198/base -> origin/gh/zou3519/1198/base 2025-11-03T16:58:02.0530852Z * [new branch] gh/zou3519/1198/head -> origin/gh/zou3519/1198/head 2025-11-03T16:58:02.0532590Z * [new branch] gh/zou3519/1198/orig -> origin/gh/zou3519/1198/orig 2025-11-03T16:58:02.0535358Z * [new branch] gh/zpcore/1/base -> origin/gh/zpcore/1/base 2025-11-03T16:58:02.0537451Z * [new branch] gh/zpcore/1/head -> origin/gh/zpcore/1/head 2025-11-03T16:58:02.0540089Z * [new branch] gh/zpcore/11/base -> origin/gh/zpcore/11/base 2025-11-03T16:58:02.0541798Z * [new branch] gh/zpcore/11/head -> origin/gh/zpcore/11/head 2025-11-03T16:58:02.0543577Z * [new branch] gh/zpcore/11/orig -> origin/gh/zpcore/11/orig 2025-11-03T16:58:02.0546313Z * [new branch] gh/zpcore/12/base -> origin/gh/zpcore/12/base 2025-11-03T16:58:02.0547994Z * [new branch] gh/zpcore/12/head -> origin/gh/zpcore/12/head 2025-11-03T16:58:02.0549755Z * [new branch] gh/zpcore/12/orig -> origin/gh/zpcore/12/orig 2025-11-03T16:58:02.0552161Z * [new branch] gh/zpcore/13/base -> origin/gh/zpcore/13/base 2025-11-03T16:58:02.0553827Z * [new branch] gh/zpcore/13/head -> origin/gh/zpcore/13/head 2025-11-03T16:58:02.0555592Z * [new branch] gh/zpcore/13/orig -> origin/gh/zpcore/13/orig 2025-11-03T16:58:02.0557946Z * [new branch] gh/zpcore/14/base -> origin/gh/zpcore/14/base 2025-11-03T16:58:02.0559696Z * [new branch] gh/zpcore/14/head -> origin/gh/zpcore/14/head 2025-11-03T16:58:02.0561377Z * [new branch] gh/zpcore/14/orig -> origin/gh/zpcore/14/orig 2025-11-03T16:58:02.0563872Z * [new branch] gh/zpcore/15/base -> origin/gh/zpcore/15/base 2025-11-03T16:58:02.0565600Z * [new branch] gh/zpcore/15/head -> origin/gh/zpcore/15/head 2025-11-03T16:58:02.0567234Z * [new branch] gh/zpcore/15/orig -> origin/gh/zpcore/15/orig 2025-11-03T16:58:02.0569745Z * [new branch] gh/zpcore/16/base -> origin/gh/zpcore/16/base 2025-11-03T16:58:02.0571529Z * [new branch] gh/zpcore/16/head -> origin/gh/zpcore/16/head 2025-11-03T16:58:02.0573260Z * [new branch] gh/zpcore/16/orig -> origin/gh/zpcore/16/orig 2025-11-03T16:58:02.0575558Z * [new branch] gh/zpcore/17/base -> origin/gh/zpcore/17/base 2025-11-03T16:58:02.0577269Z * [new branch] gh/zpcore/17/head -> origin/gh/zpcore/17/head 2025-11-03T16:58:02.0578967Z * [new branch] gh/zpcore/17/orig -> origin/gh/zpcore/17/orig 2025-11-03T16:58:02.0582046Z * [new branch] gh/zpcore/18/base -> origin/gh/zpcore/18/base 2025-11-03T16:58:02.0583791Z * [new branch] gh/zpcore/18/head -> origin/gh/zpcore/18/head 2025-11-03T16:58:02.0585505Z * [new branch] gh/zpcore/18/orig -> origin/gh/zpcore/18/orig 2025-11-03T16:58:02.0587802Z * [new branch] gh/zpcore/19/base -> origin/gh/zpcore/19/base 2025-11-03T16:58:02.0589618Z * [new branch] gh/zpcore/19/head -> origin/gh/zpcore/19/head 2025-11-03T16:58:02.0591380Z * [new branch] gh/zpcore/19/orig -> origin/gh/zpcore/19/orig 2025-11-03T16:58:02.0593790Z * [new branch] gh/zpcore/2/base -> origin/gh/zpcore/2/base 2025-11-03T16:58:02.0595482Z * [new branch] gh/zpcore/2/head -> origin/gh/zpcore/2/head 2025-11-03T16:58:02.0597969Z * [new branch] gh/zpcore/20/base -> origin/gh/zpcore/20/base 2025-11-03T16:58:02.0599673Z * [new branch] gh/zpcore/20/head -> origin/gh/zpcore/20/head 2025-11-03T16:58:02.0601412Z * [new branch] gh/zpcore/20/orig -> origin/gh/zpcore/20/orig 2025-11-03T16:58:02.0604110Z * [new branch] gh/zpcore/21/base -> origin/gh/zpcore/21/base 2025-11-03T16:58:02.0605980Z * [new branch] gh/zpcore/21/head -> origin/gh/zpcore/21/head 2025-11-03T16:58:02.0607629Z * [new branch] gh/zpcore/21/orig -> origin/gh/zpcore/21/orig 2025-11-03T16:58:02.0610109Z * [new branch] gh/zpcore/22/base -> origin/gh/zpcore/22/base 2025-11-03T16:58:02.0611954Z * [new branch] gh/zpcore/22/head -> origin/gh/zpcore/22/head 2025-11-03T16:58:02.0613637Z * [new branch] gh/zpcore/22/orig -> origin/gh/zpcore/22/orig 2025-11-03T16:58:02.0615840Z * [new branch] gh/zpcore/23/base -> origin/gh/zpcore/23/base 2025-11-03T16:58:02.0617513Z * [new branch] gh/zpcore/23/head -> origin/gh/zpcore/23/head 2025-11-03T16:58:02.0619184Z * [new branch] gh/zpcore/23/orig -> origin/gh/zpcore/23/orig 2025-11-03T16:58:02.0621362Z * [new branch] gh/zpcore/24/base -> origin/gh/zpcore/24/base 2025-11-03T16:58:02.0623353Z * [new branch] gh/zpcore/24/head -> origin/gh/zpcore/24/head 2025-11-03T16:58:02.0625055Z * [new branch] gh/zpcore/24/orig -> origin/gh/zpcore/24/orig 2025-11-03T16:58:02.0627354Z * [new branch] gh/zpcore/25/base -> origin/gh/zpcore/25/base 2025-11-03T16:58:02.0629539Z * [new branch] gh/zpcore/25/head -> origin/gh/zpcore/25/head 2025-11-03T16:58:02.0631302Z * [new branch] gh/zpcore/25/orig -> origin/gh/zpcore/25/orig 2025-11-03T16:58:02.0633443Z * [new branch] gh/zpcore/3/base -> origin/gh/zpcore/3/base 2025-11-03T16:58:02.0635070Z * [new branch] gh/zpcore/3/head -> origin/gh/zpcore/3/head 2025-11-03T16:58:02.0637233Z * [new branch] gh/zpcore/4/base -> origin/gh/zpcore/4/base 2025-11-03T16:58:02.0638996Z * [new branch] gh/zpcore/4/head -> origin/gh/zpcore/4/head 2025-11-03T16:58:02.0641848Z * [new branch] gh/zpcore/5/base -> origin/gh/zpcore/5/base 2025-11-03T16:58:02.0643897Z * [new branch] gh/zpcore/5/head -> origin/gh/zpcore/5/head 2025-11-03T16:58:02.0646316Z * [new branch] gh/zpcore/6/base -> origin/gh/zpcore/6/base 2025-11-03T16:58:02.0648003Z * [new branch] gh/zpcore/6/head -> origin/gh/zpcore/6/head 2025-11-03T16:58:02.0650217Z * [new branch] gh/zpcore/7/base -> origin/gh/zpcore/7/base 2025-11-03T16:58:02.0651905Z * [new branch] gh/zpcore/7/head -> origin/gh/zpcore/7/head 2025-11-03T16:58:02.0654067Z * [new branch] gh/zpcore/8/base -> origin/gh/zpcore/8/base 2025-11-03T16:58:02.0655724Z * [new branch] gh/zpcore/8/head -> origin/gh/zpcore/8/head 2025-11-03T16:58:02.0657613Z * [new branch] google-main -> origin/google-main 2025-11-03T16:58:02.0659376Z * [new branch] greencontext -> origin/greencontext 2025-11-03T16:58:02.0661855Z * [new branch] guangyey/config -> origin/guangyey/config 2025-11-03T16:58:02.0663555Z * [new branch] guangyey/external_stream -> origin/guangyey/external_stream 2025-11-03T16:58:02.0665167Z * [new branch] guangyey/reimport -> origin/guangyey/reimport 2025-11-03T16:58:02.0666780Z * [new branch] guangyey/test_2025 -> origin/guangyey/test_2025 2025-11-03T16:58:02.0669276Z * [new branch] guilhermeleobas/cherry-pick-55d87d9dfd9 -> origin/guilhermeleobas/cherry-pick-55d87d9dfd9 2025-11-03T16:58:02.0671494Z * [new branch] hameerabbasi/gradcheck-allclose -> origin/hameerabbasi/gradcheck-allclose 2025-11-03T16:58:02.0673725Z * [new branch] haozhe/bf16-dynamic-shape -> origin/haozhe/bf16-dynamic-shape 2025-11-03T16:58:02.0675494Z * [new branch] hc_baseline -> origin/hc_baseline 2025-11-03T16:58:02.0677270Z * [new branch] hhh_decomp_mul -> origin/hhh_decomp_mul 2025-11-03T16:58:02.0679010Z * [new branch] hhh_rand -> origin/hhh_rand 2025-11-03T16:58:02.0681279Z * [new branch] hoy/triton-PR3973 -> origin/hoy/triton-PR3973 2025-11-03T16:58:02.0684119Z * [new branch] huba/debug_mode -> origin/huba/debug_mode 2025-11-03T16:58:02.0685410Z * [new branch] huba/dtensor_equal -> origin/huba/dtensor_equal 2025-11-03T16:58:02.0687096Z * [new branch] huba/f1 -> origin/huba/f1 2025-11-03T16:58:02.0688940Z * [new branch] huba/local_tensor -> origin/huba/local_tensor 2025-11-03T16:58:02.0690906Z * [new branch] increase-asan-build-memory -> origin/increase-asan-build-memory 2025-11-03T16:58:02.0692632Z * [new branch] inductor-perf-increase-timeout -> origin/inductor-perf-increase-timeout 2025-11-03T16:58:02.0694401Z * [new branch] inductordecompfix -> origin/inductordecompfix 2025-11-03T16:58:02.0696172Z * [new branch] inlining -> origin/inlining 2025-11-03T16:58:02.0698071Z * [new branch] inlining-ezyang -> origin/inlining-ezyang 2025-11-03T16:58:02.0700009Z * [new branch] install-torchao-0.13.0 -> origin/install-torchao-0.13.0 2025-11-03T16:58:02.0701821Z * [new branch] install_free_tensors -> origin/install_free_tensors 2025-11-03T16:58:02.0703774Z * [new branch] invoke-subgraph -> origin/invoke-subgraph 2025-11-03T16:58:02.0705676Z * [new branch] issue#58739 -> origin/issue#58739 2025-11-03T16:58:02.0707912Z * [new branch] jathu/o3 -> origin/jathu/o3 2025-11-03T16:58:02.0709539Z * [new branch] jathu/sve -> origin/jathu/sve 2025-11-03T16:58:02.0712356Z * [new branch] jcaip/test-cusparselt-version-0.6.2 -> origin/jcaip/test-cusparselt-version-0.6.2 2025-11-03T16:58:02.0714208Z * [new branch] jcaip/update-cusparselt-0.6.2 -> origin/jcaip/update-cusparselt-0.6.2 2025-11-03T16:58:02.0716451Z * [new branch] jeanschmidt/autorevert_tag_cisev -> origin/jeanschmidt/autorevert_tag_cisev 2025-11-03T16:58:02.0718082Z * [new branch] jeanschmidt/autorevert_tag_cisev_2 -> origin/jeanschmidt/autorevert_tag_cisev_2 2025-11-03T16:58:02.0719943Z * [new branch] jithunnair-amd-patch-2 -> origin/jithunnair-amd-patch-2 2025-11-03T16:58:02.0721855Z * [new branch] jithunnair-amd-patch-3 -> origin/jithunnair-amd-patch-3 2025-11-03T16:58:02.0723953Z * [new branch] jithunnair-amd-patch-4 -> origin/jithunnair-amd-patch-4 2025-11-03T16:58:02.0726253Z * [new branch] justinchu/attention-tests -> origin/justinchu/attention-tests 2025-11-03T16:58:02.0727950Z * [new branch] justinchu/export-warning -> origin/justinchu/export-warning 2025-11-03T16:58:02.0729614Z * [new branch] justinchu/native-qdq -> origin/justinchu/native-qdq 2025-11-03T16:58:02.0731994Z * [new branch] justinchuby/onnx-deprecation-msg -> origin/justinchuby/onnx-deprecation-msg 2025-11-03T16:58:02.0734324Z * [new branch] kainan666/xlf_debug -> origin/kainan666/xlf_debug 2025-11-03T16:58:02.0736174Z * [new branch] kainan_test -> origin/kainan_test 2025-11-03T16:58:02.0738941Z * [new branch] leslie/test_group_gemm_epilogues -> origin/leslie/test_group_gemm_epilogues 2025-11-03T16:58:02.0741686Z * [new branch] lessw2020/fix_cutlass_cache_error -> origin/lessw2020/fix_cutlass_cache_error 2025-11-03T16:58:02.0744154Z * [new branch] liaoxuan/shm_all_reduce -> origin/liaoxuan/shm_all_reduce 2025-11-03T16:58:02.0745844Z * [new branch] liaoxuan/test_fa_disable_softmax -> origin/liaoxuan/test_fa_disable_softmax 2025-11-03T16:58:02.0747426Z * [new branch] liaoxuan/test_int8_sdpa -> origin/liaoxuan/test_int8_sdpa 2025-11-03T16:58:02.0749169Z * [new branch] lintbuilddocker -> origin/lintbuilddocker 2025-11-03T16:58:02.0751492Z * [new branch] llama4-stable -> origin/llama4-stable 2025-11-03T16:58:02.0753123Z * [new branch] logdetfix -> origin/logdetfix 2025-11-03T16:58:02.0756002Z * [new branch] lts/release/1.8 -> origin/lts/release/1.8 2025-11-03T16:58:02.0758458Z * [new branch] lucaskabela/#94773 -> origin/lucaskabela/#94773 2025-11-03T16:58:02.0760174Z * [new branch] lucaskabela/fix_164814 -> origin/lucaskabela/fix_164814 2025-11-03T16:58:02.0761937Z * [new branch] lucaskabela/fix_164823 -> origin/lucaskabela/fix_164823 2025-11-03T16:58:02.0763783Z * [new branch] lucaskabela/fix_164875 -> origin/lucaskabela/fix_164875 2025-11-03T16:58:02.0765454Z * [new branch] lucaskabela/fix_164876 -> origin/lucaskabela/fix_164876 2025-11-03T16:58:02.0767119Z * [new branch] lucaskabela/fix_error_for_named_tuples -> origin/lucaskabela/fix_error_for_named_tuples 2025-11-03T16:58:02.0768654Z * [new branch] lucaskabela/flop_counter -> origin/lucaskabela/flop_counter 2025-11-03T16:58:02.0770694Z * [new branch] lucaskabela/func_under_decomp -> origin/lucaskabela/func_under_decomp 2025-11-03T16:58:02.0772829Z * [new branch] lucaskabela/functional_in_dynamo -> origin/lucaskabela/functional_in_dynamo 2025-11-03T16:58:02.0774656Z * [new branch] lucaskabela/install_params_as_graph_attr -> origin/lucaskabela/install_params_as_graph_attr 2025-11-03T16:58:02.0776475Z * [new branch] lucaskabela/parameters_as_graph_attr -> origin/lucaskabela/parameters_as_graph_attr 2025-11-03T16:58:02.0778739Z * [new branch] lucaskabela/remove_aot_dispatcher_metadata -> origin/lucaskabela/remove_aot_dispatcher_metadata 2025-11-03T16:58:02.0780380Z * [new branch] lucaskabela/rnn_decomp -> origin/lucaskabela/rnn_decomp 2025-11-03T16:58:02.0782293Z * [new branch] lucaskabela/type_files_in_variables -> origin/lucaskabela/type_files_in_variables 2025-11-03T16:58:02.0784169Z * [new branch] lucaskabela/typing_backends -> origin/lucaskabela/typing_backends 2025-11-03T16:58:02.0785968Z * [new branch] lucaskabela/typing_variables -> origin/lucaskabela/typing_variables 2025-11-03T16:58:02.0787793Z * [new branch] lucaskabela/typinng_variables_builtin -> origin/lucaskabela/typinng_variables_builtin 2025-11-03T16:58:02.0790055Z * [new branch] lucaskablea/pickle_einops -> origin/lucaskablea/pickle_einops 2025-11-03T16:58:02.0791923Z * [new branch] main -> origin/main 2025-11-03T16:58:02.0794262Z * [new branch] main-enable-b200-distributed-tests -> origin/main-enable-b200-distributed-tests 2025-11-03T16:58:02.0796035Z * [new branch] malfet-patch-2 -> origin/malfet-patch-2 2025-11-03T16:58:02.0797912Z * [new branch] malfet-patch-3 -> origin/malfet-patch-3 2025-11-03T16:58:02.0799776Z * [new branch] malfet-patch-7 -> origin/malfet-patch-7 2025-11-03T16:58:02.0801621Z * [new branch] malfet-patch-8 -> origin/malfet-patch-8 2025-11-03T16:58:02.0803364Z * [new branch] malfet-patch-9 -> origin/malfet-patch-9 2025-11-03T16:58:02.0805746Z * [new branch] malfet/add-3.14-ci -> origin/malfet/add-3.14-ci 2025-11-03T16:58:02.0807453Z * [new branch] malfet/be-green-context -> origin/malfet/be-green-context 2025-11-03T16:58:02.0809272Z * [new branch] malfet/be-move-more-settings-to-checkout-pytorch -> origin/malfet/be-move-more-settings-to-checkout-pytorch 2025-11-03T16:58:02.0810734Z * [new branch] malfet/mps-implement-col2im -> origin/malfet/mps-implement-col2im 2025-11-03T16:58:02.0813503Z * [new branch] manuel/aoti_metal_shimify-thread_safe -> origin/manuel/aoti_metal_shimify-thread_safe 2025-11-03T16:58:02.0815248Z * [new branch] manuel/test-ops-common-allow-mps -> origin/manuel/test-ops-common-allow-mps 2025-11-03T16:58:02.0817308Z * [new branch] masnesral/metaconda -> origin/masnesral/metaconda 2025-11-03T16:58:02.0819097Z * [new branch] masnesral/pt2_internal_logging -> origin/masnesral/pt2_internal_logging 2025-11-03T16:58:02.0820718Z * [new branch] mem_profiler_stack_trace -> origin/mem_profiler_stack_trace 2025-11-03T16:58:02.0823871Z * [new branch] memory_profiler_stack -> origin/memory_profiler_stack 2025-11-03T16:58:02.0825683Z * [new branch] metascroy-patch-1 -> origin/metascroy-patch-1 2025-11-03T16:58:02.0827457Z * [new branch] mingw_posix -> origin/mingw_posix 2025-11-03T16:58:02.0829923Z * [new branch] mlazos/S429861-debug -> origin/mlazos/S429861-debug 2025-11-03T16:58:02.0831454Z * [new branch] mlazos/aa -> origin/mlazos/aa 2025-11-03T16:58:02.0833410Z * [new branch] mlazos/acts -> origin/mlazos/acts 2025-11-03T16:58:02.0834995Z * [new branch] mlazos/arg-renames -> origin/mlazos/arg-renames 2025-11-03T16:58:02.0836699Z * [new branch] mlazos/backup-test-branch -> origin/mlazos/backup-test-branch 2025-11-03T16:58:02.0838297Z * [new branch] mlazos/bad-cudagraphs -> origin/mlazos/bad-cudagraphs 2025-11-03T16:58:02.0839934Z * [new branch] mlazos/baseline -> origin/mlazos/baseline 2025-11-03T16:58:02.0842184Z * [new branch] mlazos/baseline-graph-breaks -> origin/mlazos/baseline-graph-breaks 2025-11-03T16:58:02.0844113Z * [new branch] mlazos/beta-tensor -> origin/mlazos/beta-tensor 2025-11-03T16:58:02.0846161Z * [new branch] mlazos/buffers -> origin/mlazos/buffers 2025-11-03T16:58:02.0847776Z * [new branch] mlazos/buffers2 -> origin/mlazos/buffers2 2025-11-03T16:58:02.0849491Z * [new branch] mlazos/buffers3 -> origin/mlazos/buffers3 2025-11-03T16:58:02.0851546Z * [new branch] mlazos/ck2 -> origin/mlazos/ck2 2025-11-03T16:58:02.0853381Z * [new branch] mlazos/combokernels -> origin/mlazos/combokernels 2025-11-03T16:58:02.0855141Z * [new branch] mlazos/ctx-cleanup -> origin/mlazos/ctx-cleanup 2025-11-03T16:58:02.0857301Z * [new branch] mlazos/cuda-cmd-log -> origin/mlazos/cuda-cmd-log 2025-11-03T16:58:02.0859213Z * [new branch] mlazos/cudagraph-tests -> origin/mlazos/cudagraph-tests 2025-11-03T16:58:02.0861036Z * [new branch] mlazos/cudagraphs-measurement -> origin/mlazos/cudagraphs-measurement 2025-11-03T16:58:02.0862986Z * [new branch] mlazos/cutlass-test -> origin/mlazos/cutlass-test 2025-11-03T16:58:02.0864701Z * [new branch] mlazos/cutlass-topo-bug -> origin/mlazos/cutlass-topo-bug 2025-11-03T16:58:02.0866425Z * [new branch] mlazos/dataclass-proxy -> origin/mlazos/dataclass-proxy 2025-11-03T16:58:02.0868155Z * [new branch] mlazos/dc-attrs -> origin/mlazos/dc-attrs 2025-11-03T16:58:02.0869977Z * [new branch] mlazos/dc-helion -> origin/mlazos/dc-helion 2025-11-03T16:58:02.0871691Z * [new branch] mlazos/dict-fix -> origin/mlazos/dict-fix 2025-11-03T16:58:02.0873502Z * [new branch] mlazos/disable-tf -> origin/mlazos/disable-tf 2025-11-03T16:58:02.0875151Z * [new branch] mlazos/dupe-fix -> origin/mlazos/dupe-fix 2025-11-03T16:58:02.0876940Z * [new branch] mlazos/dyn-batch -> origin/mlazos/dyn-batch 2025-11-03T16:58:02.0878653Z * [new branch] mlazos/evt -> origin/mlazos/evt 2025-11-03T16:58:02.0880601Z * [new branch] mlazos/extract-examples -> origin/mlazos/extract-examples 2025-11-03T16:58:02.0882191Z * [new branch] mlazos/foreach-op -> origin/mlazos/foreach-op 2025-11-03T16:58:02.0884001Z * [new branch] mlazos/fp8 -> origin/mlazos/fp8 2025-11-03T16:58:02.0885745Z * [new branch] mlazos/fp8-bias -> origin/mlazos/fp8-bias 2025-11-03T16:58:02.0887561Z * [new branch] mlazos/fp8-bias-fusion -> origin/mlazos/fp8-bias-fusion 2025-11-03T16:58:02.0890441Z * [new branch] mlazos/fp8-fixes -> origin/mlazos/fp8-fixes 2025-11-03T16:58:02.0891342Z * [new branch] mlazos/freezing -> origin/mlazos/freezing 2025-11-03T16:58:02.0893043Z * [new branch] mlazos/h-comp -> origin/mlazos/h-comp 2025-11-03T16:58:02.0894946Z * [new branch] mlazos/h-comp2 -> origin/mlazos/h-comp2 2025-11-03T16:58:02.0896662Z * [new branch] mlazos/hash-hop -> origin/mlazos/hash-hop 2025-11-03T16:58:02.0898422Z * [new branch] mlazos/hc -> origin/mlazos/hc 2025-11-03T16:58:02.0900166Z * [new branch] mlazos/hc-cycles -> origin/mlazos/hc-cycles 2025-11-03T16:58:02.0901968Z * [new branch] mlazos/hc-fixes -> origin/mlazos/hc-fixes 2025-11-03T16:58:02.0903751Z * [new branch] mlazos/hc-fixes3 -> origin/mlazos/hc-fixes3 2025-11-03T16:58:02.0905457Z * [new branch] mlazos/hc-fixes4 -> origin/mlazos/hc-fixes4 2025-11-03T16:58:02.0907178Z * [new branch] mlazos/hc-hf -> origin/mlazos/hc-hf 2025-11-03T16:58:02.0908970Z * [new branch] mlazos/hc-mut -> origin/mlazos/hc-mut 2025-11-03T16:58:02.0910685Z * [new branch] mlazos/hc10 -> origin/mlazos/hc10 2025-11-03T16:58:02.0912459Z * [new branch] mlazos/hc11 -> origin/mlazos/hc11 2025-11-03T16:58:02.0914250Z * [new branch] mlazos/hc12 -> origin/mlazos/hc12 2025-11-03T16:58:02.0915974Z * [new branch] mlazos/hc13 -> origin/mlazos/hc13 2025-11-03T16:58:02.0917652Z * [new branch] mlazos/hc14 -> origin/mlazos/hc14 2025-11-03T16:58:02.0919374Z * [new branch] mlazos/hc15 -> origin/mlazos/hc15 2025-11-03T16:58:02.0921452Z * [new branch] mlazos/hc2 -> origin/mlazos/hc2 2025-11-03T16:58:02.0923268Z * [new branch] mlazos/hc4 -> origin/mlazos/hc4 2025-11-03T16:58:02.0925041Z * [new branch] mlazos/hc5 -> origin/mlazos/hc5 2025-11-03T16:58:02.0926772Z * [new branch] mlazos/hc6 -> origin/mlazos/hc6 2025-11-03T16:58:02.0928511Z * [new branch] mlazos/hc7 -> origin/mlazos/hc7 2025-11-03T16:58:02.0930157Z * [new branch] mlazos/hc8 -> origin/mlazos/hc8 2025-11-03T16:58:02.0931951Z * [new branch] mlazos/hc9 -> origin/mlazos/hc9 2025-11-03T16:58:02.0933716Z * [new branch] mlazos/hc_baseline2 -> origin/mlazos/hc_baseline2 2025-11-03T16:58:02.0935392Z * [new branch] mlazos/inductor-streams -> origin/mlazos/inductor-streams 2025-11-03T16:58:02.0937225Z * [new branch] mlazos/lr-composibility -> origin/mlazos/lr-composibility 2025-11-03T16:58:02.0938745Z * [new branch] mlazos/main -> origin/mlazos/main 2025-11-03T16:58:02.0940560Z * [new branch] mlazos/main-test-enablement -> origin/mlazos/main-test-enablement 2025-11-03T16:58:02.0942388Z * [new branch] mlazos/mcg -> origin/mlazos/mcg 2025-11-03T16:58:02.0944249Z * [new branch] mlazos/mcg2 -> origin/mlazos/mcg2 2025-11-03T16:58:02.0946114Z * [new branch] mlazos/meta-guards -> origin/mlazos/meta-guards 2025-11-03T16:58:02.0948196Z * [new branch] mlazos/mlazos/ck2 -> origin/mlazos/mlazos/ck2 2025-11-03T16:58:02.0949908Z * [new branch] mlazos/mlazos/foreach-map-adam -> origin/mlazos/mlazos/foreach-map-adam 2025-11-03T16:58:02.0951752Z * [new branch] mlazos/mlazos/tf-mode-backup -> origin/mlazos/mlazos/tf-mode-backup 2025-11-03T16:58:02.0953428Z * [new branch] mlazos/mod-fix -> origin/mlazos/mod-fix 2025-11-03T16:58:02.0955203Z * [new branch] mlazos/mode-fix -> origin/mlazos/mode-fix 2025-11-03T16:58:02.0956920Z * [new branch] mlazos/more-tests -> origin/mlazos/more-tests 2025-11-03T16:58:02.0958664Z * [new branch] mlazos/offsets -> origin/mlazos/offsets 2025-11-03T16:58:02.0960464Z * [new branch] mlazos/proxy-ctors -> origin/mlazos/proxy-ctors 2025-11-03T16:58:02.0962157Z * [new branch] mlazos/quant-fix -> origin/mlazos/quant-fix 2025-11-03T16:58:02.0963975Z * [new branch] mlazos/resnet-fix -> origin/mlazos/resnet-fix 2025-11-03T16:58:02.0965884Z * [new branch] mlazos/rm-buf-names -> origin/mlazos/rm-buf-names 2025-11-03T16:58:02.0967680Z * [new branch] mlazos/rm-code -> origin/mlazos/rm-code 2025-11-03T16:58:02.0969393Z * [new branch] mlazos/rm-spam -> origin/mlazos/rm-spam 2025-11-03T16:58:02.0971108Z * [new branch] mlazos/rtp -> origin/mlazos/rtp 2025-11-03T16:58:02.0972851Z * [new branch] mlazos/static-idx-dbg -> origin/mlazos/static-idx-dbg 2025-11-03T16:58:02.0974607Z * [new branch] mlazos/static-inputs-log -> origin/mlazos/static-inputs-log 2025-11-03T16:58:02.0976718Z * [new branch] mlazos/td-fix2 -> origin/mlazos/td-fix2 2025-11-03T16:58:02.0978489Z * [new branch] mlazos/tensor-hasattr2 -> origin/mlazos/tensor-hasattr2 2025-11-03T16:58:02.0980279Z * [new branch] mlazos/test -> origin/mlazos/test 2025-11-03T16:58:02.0982151Z * [new branch] mlazos/tf-mode -> origin/mlazos/tf-mode 2025-11-03T16:58:02.0983903Z * [new branch] mlazos/tf-mode-backup2 -> origin/mlazos/tf-mode-backup2 2025-11-03T16:58:02.0985622Z * [new branch] mlazos/tf-mode-reland -> origin/mlazos/tf-mode-reland 2025-11-03T16:58:02.0987411Z * [new branch] mlazos/tf-mode-reland2 -> origin/mlazos/tf-mode-reland2 2025-11-03T16:58:02.0989119Z * [new branch] mlazos/tf-mode-reland3 -> origin/mlazos/tf-mode-reland3 2025-11-03T16:58:02.0990852Z * [new branch] mlazos/triton-no-epi -> origin/mlazos/triton-no-epi 2025-11-03T16:58:02.0992544Z * [new branch] mlazos/tune-proto -> origin/mlazos/tune-proto 2025-11-03T16:58:02.0994480Z * [new branch] mlazos/tuple-fixes -> origin/mlazos/tuple-fixes 2025-11-03T16:58:02.0996229Z * [new branch] mlazos/tuple-fixes2 -> origin/mlazos/tuple-fixes2 2025-11-03T16:58:02.0997976Z * [new branch] mlazos/tuple-handling -> origin/mlazos/tuple-handling 2025-11-03T16:58:02.0999658Z * [new branch] mlazos/user-stream-base -> origin/mlazos/user-stream-base 2025-11-03T16:58:02.1001305Z * [new branch] mlazos/user-streams -> origin/mlazos/user-streams 2025-11-03T16:58:02.1002959Z * [new branch] mlazos/user-streams-backup -> origin/mlazos/user-streams-backup 2025-11-03T16:58:02.1004697Z * [new branch] mlazos/user-streams-backup2 -> origin/mlazos/user-streams-backup2 2025-11-03T16:58:02.1006428Z * [new branch] mlazos/vary-beta -> origin/mlazos/vary-beta 2025-11-03T16:58:02.1008742Z * [new branch] mlazos/vary-beta2 -> origin/mlazos/vary-beta2 2025-11-03T16:58:02.1010398Z * [new branch] mlazos/weird-perf1 -> origin/mlazos/weird-perf1 2025-11-03T16:58:02.1012161Z * [new branch] mm_out_dtype_compile -> origin/mm_out_dtype_compile 2025-11-03T16:58:02.1013872Z * [new branch] module-shim -> origin/module-shim 2025-11-03T16:58:02.1015695Z * [new branch] move-theme-out-docker -> origin/move-theme-out-docker 2025-11-03T16:58:02.1017849Z * [new branch] move_aws_steps_inside_setup_rocm -> origin/move_aws_steps_inside_setup_rocm 2025-11-03T16:58:02.1019754Z * [new branch] msaroufim-patch-2 -> origin/msaroufim-patch-2 2025-11-03T16:58:02.1022439Z * [new branch] msaroufim/reduce -> origin/msaroufim/reduce 2025-11-03T16:58:02.1024757Z * [new branch] mtia/basic-cmake -> origin/mtia/basic-cmake 2025-11-03T16:58:02.1027191Z * [new branch] mwizak/fix-triton-block-shape -> origin/mwizak/fix-triton-block-shape 2025-11-03T16:58:02.1029054Z * [new branch] mwizak/restrict-test-mm-backend -> origin/mwizak/restrict-test-mm-backend 2025-11-03T16:58:02.1030767Z * [new branch] my_varlen_backup -> origin/my_varlen_backup 2025-11-03T16:58:02.1032492Z * [new branch] nativert_num_outputs -> origin/nativert_num_outputs 2025-11-03T16:58:02.1034515Z * [new branch] new-codegen -> origin/new-codegen 2025-11-03T16:58:02.1036335Z * [new branch] newtest-base -> origin/newtest-base 2025-11-03T16:58:02.1038708Z * [new branch] ngimel/discont_allgather -> origin/ngimel/discont_allgather 2025-11-03T16:58:02.1040278Z * [new branch] ngimel/error_index_list -> origin/ngimel/error_index_list 2025-11-03T16:58:02.1041834Z * [new branch] ngimel/gg_new -> origin/ngimel/gg_new 2025-11-03T16:58:02.1043400Z * [new branch] ngimel/nDim_fix -> origin/ngimel/nDim_fix 2025-11-03T16:58:02.1045213Z * [new branch] nightly -> origin/nightly 2025-11-03T16:58:02.1047661Z * [new branch] nikitaved/addmm_1_rowcol_lt_path_check -> origin/nikitaved/addmm_1_rowcol_lt_path_check 2025-11-03T16:58:02.1049273Z * [new branch] nikitaved/addmm_epilogue_fusions -> origin/nikitaved/addmm_epilogue_fusions 2025-11-03T16:58:02.1050918Z * [new branch] nikitaved/addmm_epilogue_fusions_2d_bias -> origin/nikitaved/addmm_epilogue_fusions_2d_bias 2025-11-03T16:58:02.1052758Z * [new branch] nikitaved/addmm_epilogue_fusions_inductor -> origin/nikitaved/addmm_epilogue_fusions_inductor 2025-11-03T16:58:02.1054733Z * [new branch] nikitaved/addmm_epilogue_fusions_scratch -> origin/nikitaved/addmm_epilogue_fusions_scratch 2025-11-03T16:58:02.1056803Z * [new branch] nikitaved/simpler_can_use_32bit_index -> origin/nikitaved/simpler_can_use_32bit_index 2025-11-03T16:58:02.1058457Z * [new branch] nikitaved/test -> origin/nikitaved/test 2025-11-03T16:58:02.1060594Z * [new branch] nmacchioni-40p-machine-balance -> origin/nmacchioni-40p-machine-balance 2025-11-03T16:58:02.1062610Z * [new branch] nmacchioni-60p-machine-balance -> origin/nmacchioni-60p-machine-balance 2025-11-03T16:58:02.1064523Z * [new branch] nmacchioni-padmm-no-compute-bound-check -> origin/nmacchioni-padmm-no-compute-bound-check 2025-11-03T16:58:02.1066256Z * [new branch] nmacchioni-patch-1 -> origin/nmacchioni-patch-1 2025-11-03T16:58:02.1068118Z * [new branch] no_distributed_log_spew -> origin/no_distributed_log_spew 2025-11-03T16:58:02.1069912Z * [new branch] nofun-hack -> origin/nofun-hack 2025-11-03T16:58:02.1072374Z * [new branch] nullplay/fuse_matmul -> origin/nullplay/fuse_matmul 2025-11-03T16:58:02.1074120Z * [new branch] nullplay_fuse_matmul -> origin/nullplay_fuse_matmul 2025-11-03T16:58:02.1076875Z * [new branch] orig/release/1.10 -> origin/orig/release/1.10 2025-11-03T16:58:02.1078623Z * [new branch] orig/release/1.11 -> origin/orig/release/1.11 2025-11-03T16:58:02.1080385Z * [new branch] orig/release/1.12 -> origin/orig/release/1.12 2025-11-03T16:58:02.1082356Z * [new branch] orig/release/1.13 -> origin/orig/release/1.13 2025-11-03T16:58:02.1084420Z * [new branch] orig/release/1.6 -> origin/orig/release/1.6 2025-11-03T16:58:02.1086386Z * [new branch] orig/release/1.7 -> origin/orig/release/1.7 2025-11-03T16:58:02.1088084Z * [new branch] orig/release/1.8 -> origin/orig/release/1.8 2025-11-03T16:58:02.1089844Z * [new branch] orig/release/1.9 -> origin/orig/release/1.9 2025-11-03T16:58:02.1091640Z * [new branch] orig/release/2.0 -> origin/orig/release/2.0 2025-11-03T16:58:02.1093193Z * [new branch] orig/release/2.1 -> origin/orig/release/2.1 2025-11-03T16:58:02.1094935Z * [new branch] orig/release/2.2 -> origin/orig/release/2.2 2025-11-03T16:58:02.1096709Z * [new branch] orig/release/2.3 -> origin/orig/release/2.3 2025-11-03T16:58:02.1098412Z * [new branch] orig/release/2.4 -> origin/orig/release/2.4 2025-11-03T16:58:02.1100055Z * [new branch] orig/release/2.5 -> origin/orig/release/2.5 2025-11-03T16:58:02.1101815Z * [new branch] orig/release/2.6 -> origin/orig/release/2.6 2025-11-03T16:58:02.1103846Z * [new branch] orig/release/2.7 -> origin/orig/release/2.7 2025-11-03T16:58:02.1106002Z * [new branch] orig/release/2.8 -> origin/orig/release/2.8 2025-11-03T16:58:02.1107746Z * [new branch] orig/release/2.9 -> origin/orig/release/2.9 2025-11-03T16:58:02.1111551Z * [new branch] origin/gh/fxdawnn/1/base -> origin/origin/gh/fxdawnn/1/base 2025-11-03T16:58:02.1113744Z * [new branch] origin/gh/fxdawnn/1/orig -> origin/origin/gh/fxdawnn/1/orig 2025-11-03T16:58:02.1116517Z * [new branch] origin/gh/zpcore/14/orig -> origin/origin/gh/zpcore/14/orig 2025-11-03T16:58:02.1118292Z * [new branch] padded-tensor -> origin/padded-tensor 2025-11-03T16:58:02.1120576Z * [new branch] pca2 -> origin/pca2 2025-11-03T16:58:02.1122706Z * [new branch] per_channel_backup -> origin/per_channel_backup 2025-11-03T16:58:02.1124476Z * [new branch] perf_ops -> origin/perf_ops 2025-11-03T16:58:02.1126263Z * [new branch] perf_ops_2_9 -> origin/perf_ops_2_9 2025-11-03T16:58:02.1128191Z * [new branch] pianpwk-patch-1 -> origin/pianpwk-patch-1 2025-11-03T16:58:02.1130100Z * [new branch] pianpwk-patch-2 -> origin/pianpwk-patch-2 2025-11-03T16:58:02.1131906Z * [new branch] pianpwk-patch-3 -> origin/pianpwk-patch-3 2025-11-03T16:58:02.1134120Z * [new branch] pianpwk/__draft_debug_mode -> origin/pianpwk/__draft_debug_mode 2025-11-03T16:58:02.1135858Z * [new branch] pianpwk/_debug_mode_for_triton_draft -> origin/pianpwk/_debug_mode_for_triton_draft 2025-11-03T16:58:02.1137469Z * [new branch] pianpwk/_profile_w_dispatch_keys -> origin/pianpwk/_profile_w_dispatch_keys 2025-11-03T16:58:02.1139166Z * [new branch] pianpwk/_super_draft_debug_mode -> origin/pianpwk/_super_draft_debug_mode 2025-11-03T16:58:02.1140754Z * [new branch] pianpwk/anomaly_tb -> origin/pianpwk/anomaly_tb 2025-11-03T16:58:02.1143003Z * [new branch] pianpwk/auto_fx_annotate -> origin/pianpwk/auto_fx_annotate 2025-11-03T16:58:02.1145112Z * [new branch] pianpwk/backed_size_oblivious_export -> origin/pianpwk/backed_size_oblivious_export 2025-11-03T16:58:02.1146908Z * [new branch] pianpwk/base_view_shape_key -> origin/pianpwk/base_view_shape_key 2025-11-03T16:58:02.1148577Z * [new branch] pianpwk/bert_dynamic_perf -> origin/pianpwk/bert_dynamic_perf 2025-11-03T16:58:02.1150354Z * [new branch] pianpwk/debug_fwd_stack_traces -> origin/pianpwk/debug_fwd_stack_traces 2025-11-03T16:58:02.1152042Z * [new branch] pianpwk/debug_mode_hacks -> origin/pianpwk/debug_mode_hacks 2025-11-03T16:58:02.1154002Z * [new branch] pianpwk/debug_mode_hooks -> origin/pianpwk/debug_mode_hooks 2025-11-03T16:58:02.1155783Z * [new branch] pianpwk/debug_mode_inductor -> origin/pianpwk/debug_mode_inductor 2025-11-03T16:58:02.1157560Z * [new branch] pianpwk/debug_mode_opcall_refactor -> origin/pianpwk/debug_mode_opcall_refactor 2025-11-03T16:58:02.1159403Z * [new branch] pianpwk/debug_mode_show_ids -> origin/pianpwk/debug_mode_show_ids 2025-11-03T16:58:02.1161114Z * [new branch] pianpwk/debugmode_compile_tf -> origin/pianpwk/debugmode_compile_tf 2025-11-03T16:58:02.1162729Z * [new branch] pianpwk/debugmode_show_ids -> origin/pianpwk/debugmode_show_ids 2025-11-03T16:58:02.1164647Z * [new branch] pianpwk/dispatch_key_debugging_for_debug -> origin/pianpwk/dispatch_key_debugging_for_debug 2025-11-03T16:58:02.1166376Z * [new branch] pianpwk/draft_debug_mode_tfcompile -> origin/pianpwk/draft_debug_mode_tfcompile 2025-11-03T16:58:02.1168101Z * [new branch] pianpwk/draft_multikernel_nn -> origin/pianpwk/draft_multikernel_nn 2025-11-03T16:58:02.1169892Z * [new branch] pianpwk/draft_multikernel_status_10_5 -> origin/pianpwk/draft_multikernel_status_10_5 2025-11-03T16:58:02.1171626Z * [new branch] pianpwk/event_list_tree -> origin/pianpwk/event_list_tree 2025-11-03T16:58:02.1173455Z * [new branch] pianpwk/false_numel_refs -> origin/pianpwk/false_numel_refs 2025-11-03T16:58:02.1175136Z * [new branch] pianpwk/fx_export_annotate -> origin/pianpwk/fx_export_annotate 2025-11-03T16:58:02.1176807Z * [new branch] pianpwk/maybe_guard_rel -> origin/pianpwk/maybe_guard_rel 2025-11-03T16:58:02.1178622Z * [new branch] pianpwk/multikernel_hints_draft -> origin/pianpwk/multikernel_hints_draft 2025-11-03T16:58:02.1180371Z * [new branch] pianpwk/no_size_oblivious_slice_scat -> origin/pianpwk/no_size_oblivious_slice_scat 2025-11-03T16:58:02.1182233Z * [new branch] pianpwk/oblivious_reshape_view_better -> origin/pianpwk/oblivious_reshape_view_better 2025-11-03T16:58:02.1183954Z * [new branch] pianpwk/pre_forward_hook -> origin/pianpwk/pre_forward_hook 2025-11-03T16:58:02.1185674Z * [new branch] pianpwk/skip_python_keys_alternate -> origin/pianpwk/skip_python_keys_alternate 2025-11-03T16:58:02.1187434Z * [new branch] pianpwk/skip_python_keys_in_guards -> origin/pianpwk/skip_python_keys_in_guards 2025-11-03T16:58:02.1189157Z * [new branch] pianpwk/sym_tokens_draft -> origin/pianpwk/sym_tokens_draft 2025-11-03T16:58:02.1190982Z * [new branch] pianpwk/test_pointwise_guard_or_false -> origin/pianpwk/test_pointwise_guard_or_false 2025-11-03T16:58:02.1192802Z * [new branch] pianpwk/totally_draft_sym_wrap -> origin/pianpwk/totally_draft_sym_wrap 2025-11-03T16:58:02.1194441Z * [new branch] pianpwk/try_dumb_stuff -> origin/pianpwk/try_dumb_stuff 2025-11-03T16:58:02.1196590Z * [new branch] pianpwk/try_dumb_stuff_2 -> origin/pianpwk/try_dumb_stuff_2 2025-11-03T16:58:02.1198361Z * [new branch] pianpwk/user_symints -> origin/pianpwk/user_symints 2025-11-03T16:58:02.1200212Z * [new branch] pianpwk/wan21_reshape -> origin/pianpwk/wan21_reshape 2025-11-03T16:58:02.1202510Z * [new branch] piz/fall_back_missing_0716 -> origin/piz/fall_back_missing_0716 2025-11-03T16:58:02.1204128Z * [new branch] piz/fix_static_shard_method -> origin/piz/fix_static_shard_method 2025-11-03T16:58:02.1205905Z * [new branch] pool-separate -> origin/pool-separate 2025-11-03T16:58:02.1207693Z * [new branch] pr-156087 -> origin/pr-156087 2025-11-03T16:58:02.1210031Z * [new branch] pr/131860 -> origin/pr/131860 2025-11-03T16:58:02.1211855Z * [new branch] pr165329 -> origin/pr165329 2025-11-03T16:58:02.1213620Z * [new branch] pr165330 -> origin/pr165330 2025-11-03T16:58:02.1215386Z * [new branch] pr165564 -> origin/pr165564 2025-11-03T16:58:02.1217162Z * [new branch] pr165636 -> origin/pr165636 2025-11-03T16:58:02.1219074Z * [new branch] pr165637 -> origin/pr165637 2025-11-03T16:58:02.1220973Z * [new branch] pr165638 -> origin/pr165638 2025-11-03T16:58:02.1224213Z * [new branch] pr165639 -> origin/pr165639 2025-11-03T16:58:02.1226025Z * [new branch] pr165640 -> origin/pr165640 2025-11-03T16:58:02.1227802Z * [new branch] pr165642 -> origin/pr165642 2025-11-03T16:58:02.1229606Z * [new branch] pr165994 -> origin/pr165994 2025-11-03T16:58:02.1231363Z * [new branch] pr166102 -> origin/pr166102 2025-11-03T16:58:02.1233064Z * [new branch] pr166103 -> origin/pr166103 2025-11-03T16:58:02.1234891Z * [new branch] pr166149 -> origin/pr166149 2025-11-03T16:58:02.1236740Z * [new branch] pr166182 -> origin/pr166182 2025-11-03T16:58:02.1238587Z * [new branch] predispatch_to -> origin/predispatch_to 2025-11-03T16:58:02.1240409Z * [new branch] prepare-perf-baseline-number-2.8 -> origin/prepare-perf-baseline-number-2.8 2025-11-03T16:58:02.1242185Z * [new branch] prepare-perf-number-2.9 -> origin/prepare-perf-number-2.9 2025-11-03T16:58:02.1244160Z * [new branch] pt-opt-cuda3 -> origin/pt-opt-cuda3 2025-11-03T16:58:02.1246259Z * [new branch] python_compiled_autograd -> origin/python_compiled_autograd 2025-11-03T16:58:02.1249010Z * [new branch] qchip/export-D54134695 -> origin/qchip/export-D54134695 2025-11-03T16:58:02.1251254Z * [new branch] quint-bits -> origin/quint-bits 2025-11-03T16:58:02.1252946Z * [new branch] reland-accgrad-stream-warn -> origin/reland-accgrad-stream-warn 2025-11-03T16:58:02.1254722Z * [new branch] reland-fx-annotate -> origin/reland-fx-annotate 2025-11-03T16:58:02.1257136Z * [new branch] release/1.10 -> origin/release/1.10 2025-11-03T16:58:02.1258923Z * [new branch] release/1.11 -> origin/release/1.11 2025-11-03T16:58:02.1260731Z * [new branch] release/1.12 -> origin/release/1.12 2025-11-03T16:58:02.1262598Z * [new branch] release/1.13 -> origin/release/1.13 2025-11-03T16:58:02.1276557Z * [new branch] release/1.4 -> origin/release/1.4 2025-11-03T16:58:02.1276858Z * [new branch] release/1.4.1 -> origin/release/1.4.1 2025-11-03T16:58:02.1277024Z * [new branch] release/1.5 -> origin/release/1.5 2025-11-03T16:58:02.1277172Z * [new branch] release/1.6 -> origin/release/1.6 2025-11-03T16:58:02.1277456Z * [new branch] release/1.7 -> origin/release/1.7 2025-11-03T16:58:02.1277676Z * [new branch] release/1.8 -> origin/release/1.8 2025-11-03T16:58:02.1277901Z * [new branch] release/1.9 -> origin/release/1.9 2025-11-03T16:58:02.1278119Z * [new branch] release/2.0 -> origin/release/2.0 2025-11-03T16:58:02.1278325Z * [new branch] release/2.1 -> origin/release/2.1 2025-11-03T16:58:02.1279984Z * [new branch] release/2.2 -> origin/release/2.2 2025-11-03T16:58:02.1282080Z * [new branch] release/2.3 -> origin/release/2.3 2025-11-03T16:58:02.1284028Z * [new branch] release/2.4 -> origin/release/2.4 2025-11-03T16:58:02.1286280Z * [new branch] release/2.5 -> origin/release/2.5 2025-11-03T16:58:02.1288079Z * [new branch] release/2.6 -> origin/release/2.6 2025-11-03T16:58:02.1289957Z * [new branch] release/2.7 -> origin/release/2.7 2025-11-03T16:58:02.1291699Z * [new branch] release/2.8 -> origin/release/2.8 2025-11-03T16:58:02.1293576Z * [new branch] release/2.9 -> origin/release/2.9 2025-11-03T16:58:02.1295382Z * [new branch] release_notes -> origin/release_notes 2025-11-03T16:58:02.1297285Z * [new branch] remove-meta-files -> origin/remove-meta-files 2025-11-03T16:58:02.1299077Z * [new branch] remove_pyinterpreter -> origin/remove_pyinterpreter 2025-11-03T16:58:02.1300991Z * [new branch] replace-pytorch-labs-20250812-195836 -> origin/replace-pytorch-labs-20250812-195836 2025-11-03T16:58:02.1302843Z * [new branch] replace-pytorch-labs-20250812-200248 -> origin/replace-pytorch-labs-20250812-200248 2025-11-03T16:58:02.1304761Z * [new branch] replace-pytorch-labs-20250812-200324 -> origin/replace-pytorch-labs-20250812-200324 2025-11-03T16:58:02.1306565Z * [new branch] replace-pytorch-labs-20250812-204020 -> origin/replace-pytorch-labs-20250812-204020 2025-11-03T16:58:02.1309871Z * [new branch] revert-131069-gh/krzysztofjordan/1/head -> origin/revert-131069-gh/krzysztofjordan/1/head 2025-11-03T16:58:02.1312898Z * [new branch] revert-131469-gh/andrewor14/51/head -> origin/revert-131469-gh/andrewor14/51/head 2025-11-03T16:58:02.1316228Z * [new branch] revert-152361-gh/fadara01/1/head -> origin/revert-152361-gh/fadara01/1/head 2025-11-03T16:58:02.1319527Z * [new branch] revert-156870-gh/skarjala/3/head -> origin/revert-156870-gh/skarjala/3/head 2025-11-03T16:58:02.1321700Z * [new branch] revert-157914-cherry-pick-157503-by-pytorch_bot_bot_ -> origin/revert-157914-cherry-pick-157503-by-pytorch_bot_bot_ 2025-11-03T16:58:02.1325019Z * [new branch] revert-165144-gh/fadara01/2/head -> origin/revert-165144-gh/fadara01/2/head 2025-11-03T16:58:02.1326835Z * [new branch] revert-cpp -> origin/revert-cpp 2025-11-03T16:58:02.1328612Z * [new branch] revert-failed -> origin/revert-failed 2025-11-03T16:58:02.1330460Z * [new branch] revert_always_build_distributed -> origin/revert_always_build_distributed 2025-11-03T16:58:02.1332726Z * [new branch] rms_norm_patch -> origin/rms_norm_patch 2025-11-03T16:58:02.1335050Z * [new branch] ruisi/aot_eager_pass -> origin/ruisi/aot_eager_pass 2025-11-03T16:58:02.1336670Z * [new branch] ruisi/manual_bucket_pass -> origin/ruisi/manual_bucket_pass 2025-11-03T16:58:02.1338332Z * [new branch] ruisi/placement_trace -> origin/ruisi/placement_trace 2025-11-03T16:58:02.1341099Z * [new branch] ryanguo99/cleanup-dynamo-expected-failures -> origin/ryanguo99/cleanup-dynamo-expected-failures 2025-11-03T16:58:02.1342714Z * [new branch] ryanguo99/fix-closure-var -> origin/ryanguo99/fix-closure-var 2025-11-03T16:58:02.1344929Z * [new branch] rzou/faketensor_bench -> origin/rzou/faketensor_bench 2025-11-03T16:58:02.1346498Z * [new branch] rzou/njt -> origin/rzou/njt 2025-11-03T16:58:02.1348194Z * [new branch] rzou/pca -> origin/rzou/pca 2025-11-03T16:58:02.1349838Z * [new branch] rzou/realprop -> origin/rzou/realprop 2025-11-03T16:58:02.1351652Z * [new branch] samplevllm -> origin/samplevllm 2025-11-03T16:58:02.1354300Z * [new branch] sanchitintel/weird_thing_with_test_cpu_select_algorithm -> origin/sanchitintel/weird_thing_with_test_cpu_select_algorithm 2025-11-03T16:58:02.1356148Z * [new branch] sapling-pr-archive-SS-JIA -> origin/sapling-pr-archive-SS-JIA 2025-11-03T16:58:02.1358139Z * [new branch] sapling-pr-archive-tushar00jain -> origin/sapling-pr-archive-tushar00jain 2025-11-03T16:58:02.1359880Z * [new branch] save -> origin/save 2025-11-03T16:58:02.1361742Z * [new branch] save_github_env_rocm -> origin/save_github_env_rocm 2025-11-03T16:58:02.1363515Z * [new branch] sdpa-bs-zero -> origin/sdpa-bs-zero 2025-11-03T16:58:02.1365793Z * [new branch] sdym/2.5.1 -> origin/sdym/2.5.1 2025-11-03T16:58:02.1367684Z * [new branch] sekyondaMeta-dynamoconfig-fix -> origin/sekyondaMeta-dynamoconfig-fix 2025-11-03T16:58:02.1369948Z * [new branch] shengf/fx-xform-perf -> origin/shengf/fx-xform-perf 2025-11-03T16:58:02.1371779Z * [new branch] shoumikhin-patch-1 -> origin/shoumikhin-patch-1 2025-11-03T16:58:02.1373587Z * [new branch] solve-accuracy-fix -> origin/solve-accuracy-fix 2025-11-03T16:58:02.1375992Z * [new branch] soulitzer/reland-codev-grad-dtype -> origin/soulitzer/reland-codev-grad-dtype 2025-11-03T16:58:02.1377562Z * [new branch] soulitzer/stash-tls-ac -> origin/soulitzer/stash-tls-ac 2025-11-03T16:58:02.1379397Z * [new branch] stablize-compilation-time -> origin/stablize-compilation-time 2025-11-03T16:58:02.1381165Z * [new branch] starterTaskUpdate -> origin/starterTaskUpdate 2025-11-03T16:58:02.1383207Z * [new branch] suo -> origin/suo 2025-11-03T16:58:02.1384947Z * [new branch] sve-poc -> origin/sve-poc 2025-11-03T16:58:02.1386784Z * [new branch] switch-bn -> origin/switch-bn 2025-11-03T16:58:02.1388629Z * [new branch] sympy-bottleneck-repro -> origin/sympy-bottleneck-repro 2025-11-03T16:58:02.1390471Z * [new branch] tensordict_integration -> origin/tensordict_integration 2025-11-03T16:58:02.1392320Z * [new branch] test-move-conda-builds -> origin/test-move-conda-builds 2025-11-03T16:58:02.1394257Z * [new branch] test-myst-markdown-docstring -> origin/test-myst-markdown-docstring 2025-11-03T16:58:02.1395990Z * [new branch] test-old -> origin/test-old 2025-11-03T16:58:02.1397873Z * [new branch] test-vec-migration-internally -> origin/test-vec-migration-internally 2025-11-03T16:58:02.1400121Z * [new branch] test/bmm_heur -> origin/test/bmm_heur 2025-11-03T16:58:02.1401812Z * [new branch] test/inductor -> origin/test/inductor 2025-11-03T16:58:02.1404154Z * [new branch] tianren/customOp_autotune -> origin/tianren/customOp_autotune 2025-11-03T16:58:02.1405754Z * [new branch] tianren/customOp_autotune_fix -> origin/tianren/customOp_autotune_fix 2025-11-03T16:58:02.1407686Z * [new branch] tianren/customOp_enable_max_autotune -> origin/tianren/customOp_enable_max_autotune 2025-11-03T16:58:02.1409662Z * [new branch] tianren/customOp_fusion -> origin/tianren/customOp_fusion 2025-11-03T16:58:02.1411302Z * [new branch] tianren/flex_paged_attn_fix_temp -> origin/tianren/flex_paged_attn_fix_temp 2025-11-03T16:58:02.1413202Z * [new branch] tianren/remove_repeate -> origin/tianren/remove_repeate 2025-11-03T16:58:02.1415076Z * [new branch] tianren/test -> origin/tianren/test 2025-11-03T16:58:02.1416926Z * [new branch] tidy_performance_cyy -> origin/tidy_performance_cyy 2025-11-03T16:58:02.1418695Z * [new branch] tmp -> origin/tmp 2025-11-03T16:58:02.1420612Z * [new branch] torchtitan_ep -> origin/torchtitan_ep 2025-11-03T16:58:02.1423026Z * [new branch] trace_fsdp_torchtune_lora -> origin/trace_fsdp_torchtune_lora 2025-11-03T16:58:02.1425160Z * [new branch] traceable_fsdp_unit_tests -> origin/traceable_fsdp_unit_tests 2025-11-03T16:58:02.1426990Z * [new branch] tree_loop_vec_base -> origin/tree_loop_vec_base 2025-11-03T16:58:02.1428722Z * [new branch] triton_kernel -> origin/triton_kernel 2025-11-03T16:58:02.1430758Z * [new branch] trunk-tagging-multi-commits -> origin/trunk-tagging-multi-commits 2025-11-03T16:58:02.1432466Z * [new branch] tt_pkg_1908 -> origin/tt_pkg_1908 2025-11-03T16:58:02.1434290Z * [new branch] type_dec -> origin/type_dec 2025-11-03T16:58:02.1436129Z * [new branch] udate-sphinx-dependancies -> origin/udate-sphinx-dependancies 2025-11-03T16:58:02.1438592Z * [new branch] update-audio-commit-hash/17630256502-1803-1 -> origin/update-audio-commit-hash/17630256502-1803-1 2025-11-03T16:58:02.1440142Z * [new branch] update-audio-commit-hash/18392707270-1874-1 -> origin/update-audio-commit-hash/18392707270-1874-1 2025-11-03T16:58:02.1441967Z * [new branch] update-audio-commit-hash/18513819375-1881-1 -> origin/update-audio-commit-hash/18513819375-1881-1 2025-11-03T16:58:02.1443485Z * [new branch] update-audio-commit-hash/18607927372-1885-1 -> origin/update-audio-commit-hash/18607927372-1885-1 2025-11-03T16:58:02.1445381Z * [new branch] update-audio-commit-hash/18925946320-1900-1 -> origin/update-audio-commit-hash/18925946320-1900-1 2025-11-03T16:58:02.1448222Z * [new branch] update-executorch-commit-hash/15694981040-1626-1 -> origin/update-executorch-commit-hash/15694981040-1626-1 2025-11-03T16:58:02.1450474Z * [new branch] update-triton-commit-hash/13663274526-1487-2 -> origin/update-triton-commit-hash/13663274526-1487-2 2025-11-03T16:58:02.1452732Z * [new branch] update-vision-commit-hash/15336342773-1607-1 -> origin/update-vision-commit-hash/15336342773-1607-1 2025-11-03T16:58:02.1454368Z * [new branch] update-vision-commit-hash/18361653903-1869-1 -> origin/update-vision-commit-hash/18361653903-1869-1 2025-11-03T16:58:02.1455998Z * [new branch] update-vision-commit-hash/18513819375-1881-1 -> origin/update-vision-commit-hash/18513819375-1881-1 2025-11-03T16:58:02.1457543Z * [new branch] update-vision-commit-hash/18701484675-1891-1 -> origin/update-vision-commit-hash/18701484675-1891-1 2025-11-03T16:58:02.1459136Z * [new branch] update-vision-commit-hash/18860003792-1897-1 -> origin/update-vision-commit-hash/18860003792-1897-1 2025-11-03T16:58:02.1461083Z * [new branch] update-vision-commit-hash/18988459983-1903-1 -> origin/update-vision-commit-hash/18988459983-1903-1 2025-11-03T16:58:02.1463548Z * [new branch] update-vllm-commit-hash/18236802781-1857-1 -> origin/update-vllm-commit-hash/18236802781-1857-1 2025-11-03T16:58:02.1465427Z * [new branch] update-vllm-commit-hash/18451675449-1879-1 -> origin/update-vllm-commit-hash/18451675449-1879-1 2025-11-03T16:58:02.1467582Z * [new branch] update-xla-commit-hash/18273597034-206-1 -> origin/update-xla-commit-hash/18273597034-206-1 2025-11-03T16:58:02.1469185Z * [new branch] update-xla-commit-hash/18458620648-207-1 -> origin/update-xla-commit-hash/18458620648-207-1 2025-11-03T16:58:02.1470853Z * [new branch] update-xla-commit-hash/18645596657-208-1 -> origin/update-xla-commit-hash/18645596657-208-1 2025-11-03T16:58:02.1472484Z * [new branch] update-xla-commit-hash/19027297646-210-1 -> origin/update-xla-commit-hash/19027297646-210-1 2025-11-03T16:58:02.1474470Z * [new branch] update_docs_torch_multinomial_issue#125388 -> origin/update_docs_torch_multinomial_issue#125388 2025-11-03T16:58:02.1476184Z * [new branch] update_slow_tests_1722488736 -> origin/update_slow_tests_1722488736 2025-11-03T16:58:02.1477907Z * [new branch] update_slow_tests_1722879173 -> origin/update_slow_tests_1722879173 2025-11-03T16:58:02.1479617Z * [new branch] update_slow_tests_1759736444 -> origin/update_slow_tests_1759736444 2025-11-03T16:58:02.1481421Z * [new branch] update_slow_tests_1760341258 -> origin/update_slow_tests_1760341258 2025-11-03T16:58:02.1483159Z * [new branch] update_slow_tests_1760946059 -> origin/update_slow_tests_1760946059 2025-11-03T16:58:02.1484939Z * [new branch] update_slow_tests_1762155677 -> origin/update_slow_tests_1762155677 2025-11-03T16:58:02.1486736Z * [new branch] update_submodule_FBGEMM -> origin/update_submodule_FBGEMM 2025-11-03T16:58:02.1488471Z * [new branch] update_submodule_kineto -> origin/update_submodule_kineto 2025-11-03T16:58:02.1490298Z * [new branch] update_submodule_tensorpipe -> origin/update_submodule_tensorpipe 2025-11-03T16:58:02.1492199Z * [new branch] v0.1.2 -> origin/v0.1.2 2025-11-03T16:58:02.1494121Z * [new branch] v1.0.1 -> origin/v1.0.1 2025-11-03T16:58:02.1496040Z * [new branch] v1.0.3 -> origin/v1.0.3 2025-11-03T16:58:02.1497943Z * [new branch] v1.1.0 -> origin/v1.1.0 2025-11-03T16:58:02.1499887Z * [new branch] v1.2.0 -> origin/v1.2.0 2025-11-03T16:58:02.1502243Z * [new branch] v1.3.0 -> origin/v1.3.0 2025-11-03T16:58:02.1504154Z * [new branch] v1.3.1 -> origin/v1.3.1 2025-11-03T16:58:02.1505983Z * [new branch] validate_fn -> origin/validate_fn 2025-11-03T16:58:02.1507928Z * [new branch] validations_2.6 -> origin/validations_2.6 2025-11-03T16:58:02.1509775Z * [new branch] validations_2.8 -> origin/validations_2.8 2025-11-03T16:58:02.1511622Z * [new branch] varlen-api -> origin/varlen-api 2025-11-03T16:58:02.1513450Z * [new branch] varlen-api-backup -> origin/varlen-api-backup 2025-11-03T16:58:02.1515802Z * [new branch] viable/strict -> origin/viable/strict 2025-11-03T16:58:02.1518346Z * [new branch] vishal9-team/dtensor_parallelism_toy -> origin/vishal9-team/dtensor_parallelism_toy 2025-11-03T16:58:02.1520082Z * [new branch] vllmbuildci -> origin/vllmbuildci 2025-11-03T16:58:02.1522432Z * [new branch] vllmpin -> origin/vllmpin 2025-11-03T16:58:02.1525002Z * [new branch] wdvr/iss_145259 -> origin/wdvr/iss_145259 2025-11-03T16:58:02.1527697Z * [new branch] whc/stage2 -> origin/whc/stage2 2025-11-03T16:58:02.1529298Z * [new branch] whc/uneven -> origin/whc/uneven 2025-11-03T16:58:02.1531426Z * [new branch] whc/uneven-merge -> origin/whc/uneven-merge 2025-11-03T16:58:02.1532986Z * [new branch] win_12_8_build -> origin/win_12_8_build 2025-11-03T16:58:02.1534733Z * [new branch] win_ci -> origin/win_ci 2025-11-03T16:58:02.1536543Z * [new branch] win_warnings -> origin/win_warnings 2025-11-03T16:58:02.1538337Z * [new branch] windows_libtorch_free -> origin/windows_libtorch_free 2025-11-03T16:58:02.1540416Z * [new branch] windows_mmap -> origin/windows_mmap 2025-11-03T16:58:02.1542548Z * [new branch] xmfan-war -> origin/xmfan-war 2025-11-03T16:58:02.1545069Z * [new branch] xmfan/ca_0516 -> origin/xmfan/ca_0516 2025-11-03T16:58:02.1546673Z * [new branch] xmfan/ca_1051b93192 -> origin/xmfan/ca_1051b93192 2025-11-03T16:58:02.1548429Z * [new branch] xmfan/ca_1a722f62c248391fc4a542e8851a5559aa356ae8 -> origin/xmfan/ca_1a722f62c248391fc4a542e8851a5559aa356ae8 2025-11-03T16:58:02.1549722Z * [new branch] xmfan/ca_5a2be192d1 -> origin/xmfan/ca_5a2be192d1 2025-11-03T16:58:02.1551508Z * [new branch] xmfan/ca_9d59b516e9 -> origin/xmfan/ca_9d59b516e9 2025-11-03T16:58:02.1553405Z * [new branch] xmfan/ca_api -> origin/xmfan/ca_api 2025-11-03T16:58:02.1555116Z * [new branch] xmfan/ca_apr8 -> origin/xmfan/ca_apr8 2025-11-03T16:58:02.1557138Z * [new branch] xmfan/ca_base -> origin/xmfan/ca_base 2025-11-03T16:58:02.1559225Z * [new branch] xmfan/ca_dynamic -> origin/xmfan/ca_dynamic 2025-11-03T16:58:02.1560910Z * [new branch] xmfan/ca_fix_dyn -> origin/xmfan/ca_fix_dyn 2025-11-03T16:58:02.1562937Z * [new branch] xmfan/ca_fix_lowering -> origin/xmfan/ca_fix_lowering 2025-11-03T16:58:02.1564712Z * [new branch] xmfan/ca_fix_polyfills -> origin/xmfan/ca_fix_polyfills 2025-11-03T16:58:02.1566836Z * [new branch] xmfan/ca_jan3 -> origin/xmfan/ca_jan3 2025-11-03T16:58:02.1568526Z * [new branch] xmfan/ca_jun18 -> origin/xmfan/ca_jun18 2025-11-03T16:58:02.1570262Z * [new branch] xmfan/ca_jun24 -> origin/xmfan/ca_jun24 2025-11-03T16:58:02.1572217Z * [new branch] xmfan/ca_nested -> origin/xmfan/ca_nested 2025-11-03T16:58:02.1573831Z * [new branch] xmfan/ca_overhead -> origin/xmfan/ca_overhead 2025-11-03T16:58:02.1575596Z * [new branch] xmfan/ca_overhead_0eba7e5451 -> origin/xmfan/ca_overhead_0eba7e5451 2025-11-03T16:58:02.1577243Z * [new branch] xmfan/cacu_jun18 -> origin/xmfan/cacu_jun18 2025-11-03T16:58:02.1578975Z * [new branch] xmfan/cacu_jun19 -> origin/xmfan/cacu_jun19 2025-11-03T16:58:02.1580643Z * [new branch] xmfan/cacu_jun4 -> origin/xmfan/cacu_jun4 2025-11-03T16:58:02.1582783Z * [new branch] xmfan/cacu_may27 -> origin/xmfan/cacu_may27 2025-11-03T16:58:02.1585169Z * [new branch] xmfan/disable_duck_shape -> origin/xmfan/disable_duck_shape 2025-11-03T16:58:02.1586968Z * [new branch] xmfan/fca_cpp_node_passthrough -> origin/xmfan/fca_cpp_node_passthrough 2025-11-03T16:58:02.1588764Z * [new branch] xmfan/post_3945954741e2d37023c5d6954f9483008e0892f9 -> origin/xmfan/post_3945954741e2d37023c5d6954f9483008e0892f9 2025-11-03T16:58:02.1590529Z * [new branch] xmfan/pre_3945954741e2d37023c5d6954f9483008e0892f9 -> origin/xmfan/pre_3945954741e2d37023c5d6954f9483008e0892f9 2025-11-03T16:58:02.1592087Z * [new branch] xmfan/single_step -> origin/xmfan/single_step 2025-11-03T16:58:02.1593900Z * [new branch] xmfan/sth_0829 -> origin/xmfan/sth_0829 2025-11-03T16:58:02.1595829Z * [new branch] xmfan/test -> origin/xmfan/test 2025-11-03T16:58:02.1598167Z * [new branch] yguo/debug-0226-constexpr -> origin/yguo/debug-0226-constexpr 2025-11-03T16:58:02.1599768Z * [new branch] yguo/new_latest_changes -> origin/yguo/new_latest_changes 2025-11-03T16:58:02.1601396Z * [new branch] yguo/patch_constexpr_changes -> origin/yguo/patch_constexpr_changes 2025-11-03T16:58:02.1603611Z * [new branch] yihan_quantization -> origin/yihan_quantization 2025-11-03T16:58:02.1605839Z * [new branch] yiming/add_timm_models -> origin/yiming/add_timm_models 2025-11-03T16:58:02.1607422Z * [new branch] yiming/bootcamp -> origin/yiming/bootcamp 2025-11-03T16:58:02.1609304Z * [new branch] yiming/fix_aot_joint_graph_capture_test -> origin/yiming/fix_aot_joint_graph_capture_test 2025-11-03T16:58:02.1611054Z * [new branch] yiming/flex_attention_export_with_kwargs -> origin/yiming/flex_attention_export_with_kwargs 2025-11-03T16:58:02.1612974Z * [new branch] yiming/improve_custom_op_stack_trace -> origin/yiming/improve_custom_op_stack_trace 2025-11-03T16:58:02.1615201Z * [new branch] yiming/improve_sharding_error_msg -> origin/yiming/improve_sharding_error_msg 2025-11-03T16:58:02.1616963Z * [new branch] yiming/moe_dtensor_region_annotation -> origin/yiming/moe_dtensor_region_annotation 2025-11-03T16:58:02.1618650Z * [new branch] yiming/precompile_benchmark -> origin/yiming/precompile_benchmark 2025-11-03T16:58:02.1620654Z * [new branch] yiming/remove_blockmask_pytree_global_registration -> origin/yiming/remove_blockmask_pytree_global_registration 2025-11-03T16:58:02.1622669Z * [new branch] yolo-llama3 -> origin/yolo-llama3 2025-11-03T16:58:02.1625064Z * [new branch] yuxin/fix-save-memleak -> origin/yuxin/fix-save-memleak 2025-11-03T16:58:02.1627318Z * [new branch] zainr/canary-test -> origin/zainr/canary-test 2025-11-03T16:58:02.1629153Z * [new branch] zainr/cleanup-gh-runners -> origin/zainr/cleanup-gh-runners 2025-11-03T16:58:02.1630803Z * [new branch] zainr/pull-migration-c -> origin/zainr/pull-migration-c 2025-11-03T16:58:02.1632329Z * [new branch] zainr/test2 -> origin/zainr/test2 2025-11-03T16:58:02.1634149Z * [new branch] zainr/unstable -> origin/zainr/unstable 2025-11-03T16:58:02.1636069Z * [new branch] zasdfgbnm-patch-3 -> origin/zasdfgbnm-patch-3 2025-11-03T16:58:02.1637898Z * [new branch] zb2p -> origin/zb2p 2025-11-03T16:58:02.1639712Z * [new branch] zeros-and-scatter-part2 -> origin/zeros-and-scatter-part2 2025-11-03T16:58:02.1642525Z * [new branch] zhxchen17/aot_compile_fix_load_guard_manager -> origin/zhxchen17/aot_compile_fix_load_guard_manager 2025-11-03T16:58:02.1644721Z * [new branch] zhxchen17/ci/vllm_pin -> origin/zhxchen17/ci/vllm_pin 2025-11-03T16:58:02.1647070Z * [new branch] zhxchen17/precompile/closure_serde -> origin/zhxchen17/precompile/closure_serde 2025-11-03T16:58:02.1648690Z * [new branch] zhxchen17/precompile/default_args -> origin/zhxchen17/precompile/default_args 2025-11-03T16:58:02.1650471Z * [new branch] zhxchen17/precompile/export -> origin/zhxchen17/precompile/export 2025-11-03T16:58:02.1652257Z * [new branch] zhxchen17/precompile/export_gm -> origin/zhxchen17/precompile/export_gm 2025-11-03T16:58:02.1653888Z * [new branch] zhxchen17/precompile_vllm_fix_0 -> origin/zhxchen17/precompile_vllm_fix_0 2025-11-03T16:58:02.1656025Z * [new branch] zhxchen17/scratch/0 -> origin/zhxchen17/scratch/0 2025-11-03T16:58:02.1658440Z * [new branch] zhxhcen17/moodycamel -> origin/zhxhcen17/moodycamel 2025-11-03T16:58:02.1661048Z * [new branch] zxiiro/build-times -> origin/zxiiro/build-times 2025-11-03T16:58:02.1662813Z * [new branch] zxiiro/c7i-docs-build -> origin/zxiiro/c7i-docs-build 2025-11-03T16:58:02.1664500Z * [new branch] zxiiro/c7i-linux-4xlarge -> origin/zxiiro/c7i-linux-4xlarge 2025-11-03T16:58:02.1666188Z * [new branch] zxiiro/c7i-linux-build-yaml -> origin/zxiiro/c7i-linux-build-yaml 2025-11-03T16:58:02.1667853Z * [new branch] zxiiro/c7i-pull -> origin/zxiiro/c7i-pull 2025-11-03T16:58:02.1669571Z * [new branch] zxiiro/main -> origin/zxiiro/main 2025-11-03T16:58:02.1671291Z * [new branch] zxiiro/test-multicloud-arc -> origin/zxiiro/test-multicloud-arc 2025-11-03T16:58:02.1672748Z * [new tag] bc2caa7fdf006894eff7af936babde69ab5a40f8-huydhn-debug -> bc2caa7fdf006894eff7af936babde69ab5a40f8-huydhn-debug 2025-11-03T16:58:02.1674256Z * [new tag] ci/binaries/77164 -> ci/binaries/77164 2025-11-03T16:58:02.1675842Z * [new tag] ciflow/b200/115316 -> ciflow/b200/115316 2025-11-03T16:58:02.1676954Z * [new tag] ciflow/b200/157553 -> ciflow/b200/157553 2025-11-03T16:58:02.1678107Z * [new tag] ciflow/b200/157635 -> ciflow/b200/157635 2025-11-03T16:58:02.1679250Z * [new tag] ciflow/b200/160685 -> ciflow/b200/160685 2025-11-03T16:58:02.1680831Z * [new tag] ciflow/b200/161404 -> ciflow/b200/161404 2025-11-03T16:58:02.1681993Z * [new tag] ciflow/b200/163369 -> ciflow/b200/163369 2025-11-03T16:58:02.1683303Z * [new tag] ciflow/b200/164657 -> ciflow/b200/164657 2025-11-03T16:58:02.1684586Z * [new tag] ciflow/b200/164938 -> ciflow/b200/164938 2025-11-03T16:58:02.1685819Z * [new tag] ciflow/b200/165132 -> ciflow/b200/165132 2025-11-03T16:58:02.1687153Z * [new tag] ciflow/b200/165541 -> ciflow/b200/165541 2025-11-03T16:58:02.1688320Z * [new tag] ciflow/b200/165548 -> ciflow/b200/165548 2025-11-03T16:58:02.1689842Z * [new tag] ciflow/b200/165952 -> ciflow/b200/165952 2025-11-03T16:58:02.1691405Z * [new tag] ciflow/b200/166250 -> ciflow/b200/166250 2025-11-03T16:58:02.1692829Z * [new tag] ciflow/b200/166526 -> ciflow/b200/166526 2025-11-03T16:58:02.1694260Z * [new tag] ciflow/b200/166744 -> ciflow/b200/166744 2025-11-03T16:58:02.1695568Z * [new tag] ciflow/b200/166752 -> ciflow/b200/166752 2025-11-03T16:58:02.1696908Z * [new tag] ciflow/b200/166834 -> ciflow/b200/166834 2025-11-03T16:58:02.1698298Z * [new tag] ciflow/binaries/157432 -> ciflow/binaries/157432 2025-11-03T16:58:02.1699465Z * [new tag] ciflow/binaries/158104 -> ciflow/binaries/158104 2025-11-03T16:58:02.1700761Z * [new tag] ciflow/binaries/165922 -> ciflow/binaries/165922 2025-11-03T16:58:02.1702068Z * [new tag] ciflow/binaries/166044 -> ciflow/binaries/166044 2025-11-03T16:58:02.1703320Z * [new tag] ciflow/binaries/166621 -> ciflow/binaries/166621 2025-11-03T16:58:02.1704504Z * [new tag] ciflow/binaries/166764 -> ciflow/binaries/166764 2025-11-03T16:58:02.1705888Z * [new tag] ciflow/binaries_libtorch/157432 -> ciflow/binaries_libtorch/157432 2025-11-03T16:58:02.1707441Z * [new tag] ciflow/binaries_wheel/155731 -> ciflow/binaries_wheel/155731 2025-11-03T16:58:02.1708581Z * [new tag] ciflow/binaries_wheel/157432 -> ciflow/binaries_wheel/157432 2025-11-03T16:58:02.1709824Z * [new tag] ciflow/binaries_wheel/166380 -> ciflow/binaries_wheel/166380 2025-11-03T16:58:02.1711069Z * [new tag] ciflow/binaries_wheel/166407 -> ciflow/binaries_wheel/166407 2025-11-03T16:58:02.1712279Z * [new tag] ciflow/binaries_wheel/166829 -> ciflow/binaries_wheel/166829 2025-11-03T16:58:02.1713588Z * [new tag] ciflow/h100-distributed/166367 -> ciflow/h100-distributed/166367 2025-11-03T16:58:02.1715058Z * [new tag] ciflow/h100-symm-mem/151845 -> ciflow/h100-symm-mem/151845 2025-11-03T16:58:02.1716262Z * [new tag] ciflow/h100-symm-mem/157635 -> ciflow/h100-symm-mem/157635 2025-11-03T16:58:02.1717379Z * [new tag] ciflow/h100-symm-mem/163814 -> ciflow/h100-symm-mem/163814 2025-11-03T16:58:02.1718501Z * [new tag] ciflow/h100-symm-mem/163815 -> ciflow/h100-symm-mem/163815 2025-11-03T16:58:02.1719625Z * [new tag] ciflow/h100-symm-mem/165548 -> ciflow/h100-symm-mem/165548 2025-11-03T16:58:02.1720825Z * [new tag] ciflow/h100-symm-mem/166772 -> ciflow/h100-symm-mem/166772 2025-11-03T16:58:02.1724374Z * [new tag] ciflow/h100-symm-mem/166773 -> ciflow/h100-symm-mem/166773 2025-11-03T16:58:02.1725511Z * [new tag] ciflow/h100-symm-mem/166774 -> ciflow/h100-symm-mem/166774 2025-11-03T16:58:02.1726876Z * [new tag] ciflow/h100/115316 -> ciflow/h100/115316 2025-11-03T16:58:02.1728073Z * [new tag] ciflow/h100/157553 -> ciflow/h100/157553 2025-11-03T16:58:02.1729202Z * [new tag] ciflow/h100/157635 -> ciflow/h100/157635 2025-11-03T16:58:02.1730359Z * [new tag] ciflow/h100/160685 -> ciflow/h100/160685 2025-11-03T16:58:02.1731534Z * [new tag] ciflow/h100/161404 -> ciflow/h100/161404 2025-11-03T16:58:02.1732698Z * [new tag] ciflow/h100/163369 -> ciflow/h100/163369 2025-11-03T16:58:02.1733840Z * [new tag] ciflow/h100/164657 -> ciflow/h100/164657 2025-11-03T16:58:02.1735080Z * [new tag] ciflow/h100/165029 -> ciflow/h100/165029 2025-11-03T16:58:02.1736239Z * [new tag] ciflow/h100/165132 -> ciflow/h100/165132 2025-11-03T16:58:02.1737409Z * [new tag] ciflow/h100/165541 -> ciflow/h100/165541 2025-11-03T16:58:02.1738525Z * [new tag] ciflow/h100/165548 -> ciflow/h100/165548 2025-11-03T16:58:02.1740073Z * [new tag] ciflow/h100/165952 -> ciflow/h100/165952 2025-11-03T16:58:02.1741727Z * [new tag] ciflow/h100/166250 -> ciflow/h100/166250 2025-11-03T16:58:02.1743502Z * [new tag] ciflow/h100/166515 -> ciflow/h100/166515 2025-11-03T16:58:02.1744713Z * [new tag] ciflow/h100/166526 -> ciflow/h100/166526 2025-11-03T16:58:02.1746004Z * [new tag] ciflow/h100/166744 -> ciflow/h100/166744 2025-11-03T16:58:02.1747263Z * [new tag] ciflow/h100/166752 -> ciflow/h100/166752 2025-11-03T16:58:02.1748517Z * [new tag] ciflow/h100/166834 -> ciflow/h100/166834 2025-11-03T16:58:02.1750005Z * [new tag] ciflow/inductor-perf-compare/165029 -> ciflow/inductor-perf-compare/165029 2025-11-03T16:58:02.1751939Z * [new tag] ciflow/inductor-perf-test-nightly-rocm-mi300/162053 -> ciflow/inductor-perf-test-nightly-rocm-mi300/162053 2025-11-03T16:58:02.1753431Z * [new tag] ciflow/inductor-perf-test-nightly-x86-zen/162954 -> ciflow/inductor-perf-test-nightly-x86-zen/162954 2025-11-03T16:58:02.1754585Z * [new tag] ciflow/inductor-perf-test-nightly-x86-zen/164126 -> ciflow/inductor-perf-test-nightly-x86-zen/164126 2025-11-03T16:58:02.1755945Z * [new tag] ciflow/inductor-perf-test-nightly/162053 -> ciflow/inductor-perf-test-nightly/162053 2025-11-03T16:58:02.1757271Z * [new tag] ciflow/inductor-perf-test-nightly/165029 -> ciflow/inductor-perf-test-nightly/165029 2025-11-03T16:58:02.1759015Z * [new tag] ciflow/inductor-periodic/165029 -> ciflow/inductor-periodic/165029 2025-11-03T16:58:02.1760253Z * [new tag] ciflow/inductor-periodic/166743 -> ciflow/inductor-periodic/166743 2025-11-03T16:58:02.1761653Z * [new tag] ciflow/inductor-rocm/151845 -> ciflow/inductor-rocm/151845 2025-11-03T16:58:02.1762800Z * [new tag] ciflow/inductor-rocm/162053 -> ciflow/inductor-rocm/162053 2025-11-03T16:58:02.1763878Z * [new tag] ciflow/inductor-rocm/166515 -> ciflow/inductor-rocm/166515 2025-11-03T16:58:02.1765075Z * [new tag] ciflow/inductor-rocm/166743 -> ciflow/inductor-rocm/166743 2025-11-03T16:58:02.1766422Z * [new tag] ciflow/inductor-rocm/166870 -> ciflow/inductor-rocm/166870 2025-11-03T16:58:02.1767835Z * [new tag] ciflow/inductor/137400 -> ciflow/inductor/137400 2025-11-03T16:58:02.1768993Z * [new tag] ciflow/inductor/148180 -> ciflow/inductor/148180 2025-11-03T16:58:02.1770304Z * [new tag] ciflow/inductor/148294 -> ciflow/inductor/148294 2025-11-03T16:58:02.1771413Z * [new tag] ciflow/inductor/148328 -> ciflow/inductor/148328 2025-11-03T16:58:02.1772575Z * [new tag] ciflow/inductor/148484 -> ciflow/inductor/148484 2025-11-03T16:58:02.1773724Z * [new tag] ciflow/inductor/148492 -> ciflow/inductor/148492 2025-11-03T16:58:02.1774864Z * [new tag] ciflow/inductor/151845 -> ciflow/inductor/151845 2025-11-03T16:58:02.1775991Z * [new tag] ciflow/inductor/152624 -> ciflow/inductor/152624 2025-11-03T16:58:02.1777177Z * [new tag] ciflow/inductor/157635 -> ciflow/inductor/157635 2025-11-03T16:58:02.1778350Z * [new tag] ciflow/inductor/158104 -> ciflow/inductor/158104 2025-11-03T16:58:02.1779522Z * [new tag] ciflow/inductor/159523 -> ciflow/inductor/159523 2025-11-03T16:58:02.1780964Z * [new tag] ciflow/inductor/160174 -> ciflow/inductor/160174 2025-11-03T16:58:02.1782633Z * [new tag] ciflow/inductor/160685 -> ciflow/inductor/160685 2025-11-03T16:58:02.1784100Z * [new tag] ciflow/inductor/160686 -> ciflow/inductor/160686 2025-11-03T16:58:02.1785299Z * [new tag] ciflow/inductor/160687 -> ciflow/inductor/160687 2025-11-03T16:58:02.1786481Z * [new tag] ciflow/inductor/160688 -> ciflow/inductor/160688 2025-11-03T16:58:02.1787697Z * [new tag] ciflow/inductor/160706 -> ciflow/inductor/160706 2025-11-03T16:58:02.1788958Z * [new tag] ciflow/inductor/160729 -> ciflow/inductor/160729 2025-11-03T16:58:02.1790152Z * [new tag] ciflow/inductor/161158 -> ciflow/inductor/161158 2025-11-03T16:58:02.1791521Z * [new tag] ciflow/inductor/161246 -> ciflow/inductor/161246 2025-11-03T16:58:02.1792774Z * [new tag] ciflow/inductor/161404 -> ciflow/inductor/161404 2025-11-03T16:58:02.1794020Z * [new tag] ciflow/inductor/161495 -> ciflow/inductor/161495 2025-11-03T16:58:02.1795248Z * [new tag] ciflow/inductor/161512 -> ciflow/inductor/161512 2025-11-03T16:58:02.1796571Z * [new tag] ciflow/inductor/161938 -> ciflow/inductor/161938 2025-11-03T16:58:02.1797785Z * [new tag] ciflow/inductor/161939 -> ciflow/inductor/161939 2025-11-03T16:58:02.1799023Z * [new tag] ciflow/inductor/161940 -> ciflow/inductor/161940 2025-11-03T16:58:02.1800209Z * [new tag] ciflow/inductor/162053 -> ciflow/inductor/162053 2025-11-03T16:58:02.1801431Z * [new tag] ciflow/inductor/162469 -> ciflow/inductor/162469 2025-11-03T16:58:02.1802779Z * [new tag] ciflow/inductor/162709 -> ciflow/inductor/162709 2025-11-03T16:58:02.1803953Z * [new tag] ciflow/inductor/162795 -> ciflow/inductor/162795 2025-11-03T16:58:02.1805127Z * [new tag] ciflow/inductor/162954 -> ciflow/inductor/162954 2025-11-03T16:58:02.1806460Z * [new tag] ciflow/inductor/163249 -> ciflow/inductor/163249 2025-11-03T16:58:02.1807689Z * [new tag] ciflow/inductor/163335 -> ciflow/inductor/163335 2025-11-03T16:58:02.1808910Z * [new tag] ciflow/inductor/163369 -> ciflow/inductor/163369 2025-11-03T16:58:02.1810149Z * [new tag] ciflow/inductor/163490 -> ciflow/inductor/163490 2025-11-03T16:58:02.1811337Z * [new tag] ciflow/inductor/163503 -> ciflow/inductor/163503 2025-11-03T16:58:02.1812556Z * [new tag] ciflow/inductor/163714 -> ciflow/inductor/163714 2025-11-03T16:58:02.1813816Z * [new tag] ciflow/inductor/163936 -> ciflow/inductor/163936 2025-11-03T16:58:02.1815077Z * [new tag] ciflow/inductor/164202 -> ciflow/inductor/164202 2025-11-03T16:58:02.1816273Z * [new tag] ciflow/inductor/164384 -> ciflow/inductor/164384 2025-11-03T16:58:02.1817491Z * [new tag] ciflow/inductor/164657 -> ciflow/inductor/164657 2025-11-03T16:58:02.1818700Z * [new tag] ciflow/inductor/164938 -> ciflow/inductor/164938 2025-11-03T16:58:02.1819959Z * [new tag] ciflow/inductor/164979 -> ciflow/inductor/164979 2025-11-03T16:58:02.1821342Z * [new tag] ciflow/inductor/164980 -> ciflow/inductor/164980 2025-11-03T16:58:02.1822690Z * [new tag] ciflow/inductor/165005 -> ciflow/inductor/165005 2025-11-03T16:58:02.1823906Z * [new tag] ciflow/inductor/165029 -> ciflow/inductor/165029 2025-11-03T16:58:02.1825596Z * [new tag] ciflow/inductor/165092 -> ciflow/inductor/165092 2025-11-03T16:58:02.1826837Z * [new tag] ciflow/inductor/165132 -> ciflow/inductor/165132 2025-11-03T16:58:02.1828077Z * [new tag] ciflow/inductor/165197 -> ciflow/inductor/165197 2025-11-03T16:58:02.1829310Z * [new tag] ciflow/inductor/165274 -> ciflow/inductor/165274 2025-11-03T16:58:02.1830524Z * [new tag] ciflow/inductor/165283 -> ciflow/inductor/165283 2025-11-03T16:58:02.1831742Z * [new tag] ciflow/inductor/165284 -> ciflow/inductor/165284 2025-11-03T16:58:02.1833133Z * [new tag] ciflow/inductor/165367 -> ciflow/inductor/165367 2025-11-03T16:58:02.1834488Z * [new tag] ciflow/inductor/165423 -> ciflow/inductor/165423 2025-11-03T16:58:02.1835725Z * [new tag] ciflow/inductor/165487 -> ciflow/inductor/165487 2025-11-03T16:58:02.1837009Z * [new tag] ciflow/inductor/165541 -> ciflow/inductor/165541 2025-11-03T16:58:02.1838243Z * [new tag] ciflow/inductor/165597 -> ciflow/inductor/165597 2025-11-03T16:58:02.1839668Z * [new tag] ciflow/inductor/165686 -> ciflow/inductor/165686 2025-11-03T16:58:02.1841130Z * [new tag] ciflow/inductor/165790 -> ciflow/inductor/165790 2025-11-03T16:58:02.1842369Z * [new tag] ciflow/inductor/165824 -> ciflow/inductor/165824 2025-11-03T16:58:02.1843676Z * [new tag] ciflow/inductor/165856 -> ciflow/inductor/165856 2025-11-03T16:58:02.1844884Z * [new tag] ciflow/inductor/165885 -> ciflow/inductor/165885 2025-11-03T16:58:02.1846083Z * [new tag] ciflow/inductor/165952 -> ciflow/inductor/165952 2025-11-03T16:58:02.1847319Z * [new tag] ciflow/inductor/165953 -> ciflow/inductor/165953 2025-11-03T16:58:02.1848592Z * [new tag] ciflow/inductor/166071 -> ciflow/inductor/166071 2025-11-03T16:58:02.1850034Z * [new tag] ciflow/inductor/166083 -> ciflow/inductor/166083 2025-11-03T16:58:02.1851189Z * [new tag] ciflow/inductor/166170 -> ciflow/inductor/166170 2025-11-03T16:58:02.1852537Z * [new tag] ciflow/inductor/166198 -> ciflow/inductor/166198 2025-11-03T16:58:02.1853873Z * [new tag] ciflow/inductor/166225 -> ciflow/inductor/166225 2025-11-03T16:58:02.1855058Z * [new tag] ciflow/inductor/166236 -> ciflow/inductor/166236 2025-11-03T16:58:02.1856301Z * [new tag] ciflow/inductor/166250 -> ciflow/inductor/166250 2025-11-03T16:58:02.1857708Z * [new tag] ciflow/inductor/166254 -> ciflow/inductor/166254 2025-11-03T16:58:02.1858945Z * [new tag] ciflow/inductor/166276 -> ciflow/inductor/166276 2025-11-03T16:58:02.1860167Z * [new tag] ciflow/inductor/166294 -> ciflow/inductor/166294 2025-11-03T16:58:02.1861396Z * [new tag] ciflow/inductor/166300 -> ciflow/inductor/166300 2025-11-03T16:58:02.1862775Z * [new tag] ciflow/inductor/166320 -> ciflow/inductor/166320 2025-11-03T16:58:02.1864020Z * [new tag] ciflow/inductor/166321 -> ciflow/inductor/166321 2025-11-03T16:58:02.1865321Z * [new tag] ciflow/inductor/166337 -> ciflow/inductor/166337 2025-11-03T16:58:02.1866723Z * [new tag] ciflow/inductor/166341 -> ciflow/inductor/166341 2025-11-03T16:58:02.1867953Z * [new tag] ciflow/inductor/166354 -> ciflow/inductor/166354 2025-11-03T16:58:02.1869205Z * [new tag] ciflow/inductor/166355 -> ciflow/inductor/166355 2025-11-03T16:58:02.1870438Z * [new tag] ciflow/inductor/166361 -> ciflow/inductor/166361 2025-11-03T16:58:02.1871665Z * [new tag] ciflow/inductor/166366 -> ciflow/inductor/166366 2025-11-03T16:58:02.1872925Z * [new tag] ciflow/inductor/166367 -> ciflow/inductor/166367 2025-11-03T16:58:02.1874184Z * [new tag] ciflow/inductor/166370 -> ciflow/inductor/166370 2025-11-03T16:58:02.1875395Z * [new tag] ciflow/inductor/166371 -> ciflow/inductor/166371 2025-11-03T16:58:02.1876602Z * [new tag] ciflow/inductor/166372 -> ciflow/inductor/166372 2025-11-03T16:58:02.1878028Z * [new tag] ciflow/inductor/166377 -> ciflow/inductor/166377 2025-11-03T16:58:02.1879292Z * [new tag] ciflow/inductor/166400 -> ciflow/inductor/166400 2025-11-03T16:58:02.1880513Z * [new tag] ciflow/inductor/166402 -> ciflow/inductor/166402 2025-11-03T16:58:02.1881739Z * [new tag] ciflow/inductor/166417 -> ciflow/inductor/166417 2025-11-03T16:58:02.1882968Z * [new tag] ciflow/inductor/166433 -> ciflow/inductor/166433 2025-11-03T16:58:02.1884354Z * [new tag] ciflow/inductor/166442 -> ciflow/inductor/166442 2025-11-03T16:58:02.1885576Z * [new tag] ciflow/inductor/166467 -> ciflow/inductor/166467 2025-11-03T16:58:02.1886787Z * [new tag] ciflow/inductor/166479 -> ciflow/inductor/166479 2025-11-03T16:58:02.1888107Z * [new tag] ciflow/inductor/166482 -> ciflow/inductor/166482 2025-11-03T16:58:02.1889344Z * [new tag] ciflow/inductor/166483 -> ciflow/inductor/166483 2025-11-03T16:58:02.1890618Z * [new tag] ciflow/inductor/166494 -> ciflow/inductor/166494 2025-11-03T16:58:02.1891988Z * [new tag] ciflow/inductor/166498 -> ciflow/inductor/166498 2025-11-03T16:58:02.1893302Z * [new tag] ciflow/inductor/166499 -> ciflow/inductor/166499 2025-11-03T16:58:02.1894559Z * [new tag] ciflow/inductor/166515 -> ciflow/inductor/166515 2025-11-03T16:58:02.1895859Z * [new tag] ciflow/inductor/166536 -> ciflow/inductor/166536 2025-11-03T16:58:02.1897175Z * [new tag] ciflow/inductor/166540 -> ciflow/inductor/166540 2025-11-03T16:58:02.1898373Z * [new tag] ciflow/inductor/166541 -> ciflow/inductor/166541 2025-11-03T16:58:02.1899640Z * [new tag] ciflow/inductor/166545 -> ciflow/inductor/166545 2025-11-03T16:58:02.1900873Z * [new tag] ciflow/inductor/166581 -> ciflow/inductor/166581 2025-11-03T16:58:02.1902187Z * [new tag] ciflow/inductor/166584 -> ciflow/inductor/166584 2025-11-03T16:58:02.1903379Z * [new tag] ciflow/inductor/166593 -> ciflow/inductor/166593 2025-11-03T16:58:02.1904742Z * [new tag] ciflow/inductor/166600 -> ciflow/inductor/166600 2025-11-03T16:58:02.1905998Z * [new tag] ciflow/inductor/166608 -> ciflow/inductor/166608 2025-11-03T16:58:02.1907669Z * [new tag] ciflow/inductor/166610 -> ciflow/inductor/166610 2025-11-03T16:58:02.1908916Z * [new tag] ciflow/inductor/166629 -> ciflow/inductor/166629 2025-11-03T16:58:02.1910318Z * [new tag] ciflow/inductor/166640 -> ciflow/inductor/166640 2025-11-03T16:58:02.1911755Z * [new tag] ciflow/inductor/166648 -> ciflow/inductor/166648 2025-11-03T16:58:02.1913095Z * [new tag] ciflow/inductor/166658 -> ciflow/inductor/166658 2025-11-03T16:58:02.1914338Z * [new tag] ciflow/inductor/166659 -> ciflow/inductor/166659 2025-11-03T16:58:02.1915756Z * [new tag] ciflow/inductor/166662 -> ciflow/inductor/166662 2025-11-03T16:58:02.1916983Z * [new tag] ciflow/inductor/166663 -> ciflow/inductor/166663 2025-11-03T16:58:02.1918299Z * [new tag] ciflow/inductor/166664 -> ciflow/inductor/166664 2025-11-03T16:58:02.1919599Z * [new tag] ciflow/inductor/166669 -> ciflow/inductor/166669 2025-11-03T16:58:02.1920882Z * [new tag] ciflow/inductor/166673 -> ciflow/inductor/166673 2025-11-03T16:58:02.1922346Z * [new tag] ciflow/inductor/166674 -> ciflow/inductor/166674 2025-11-03T16:58:02.1923592Z * [new tag] ciflow/inductor/166676 -> ciflow/inductor/166676 2025-11-03T16:58:02.1924812Z * [new tag] ciflow/inductor/166679 -> ciflow/inductor/166679 2025-11-03T16:58:02.1926187Z * [new tag] ciflow/inductor/166701 -> ciflow/inductor/166701 2025-11-03T16:58:02.1927401Z * [new tag] ciflow/inductor/166720 -> ciflow/inductor/166720 2025-11-03T16:58:02.1928624Z * [new tag] ciflow/inductor/166738 -> ciflow/inductor/166738 2025-11-03T16:58:02.1929893Z * [new tag] ciflow/inductor/166740 -> ciflow/inductor/166740 2025-11-03T16:58:02.1931114Z * [new tag] ciflow/inductor/166741 -> ciflow/inductor/166741 2025-11-03T16:58:02.1932451Z * [new tag] ciflow/inductor/166745 -> ciflow/inductor/166745 2025-11-03T16:58:02.1933852Z * [new tag] ciflow/inductor/166747 -> ciflow/inductor/166747 2025-11-03T16:58:02.1935143Z * [new tag] ciflow/inductor/166757 -> ciflow/inductor/166757 2025-11-03T16:58:02.1936439Z * [new tag] ciflow/inductor/166762 -> ciflow/inductor/166762 2025-11-03T16:58:02.1937697Z * [new tag] ciflow/inductor/166775 -> ciflow/inductor/166775 2025-11-03T16:58:02.1938946Z * [new tag] ciflow/inductor/166776 -> ciflow/inductor/166776 2025-11-03T16:58:02.1940204Z * [new tag] ciflow/inductor/166777 -> ciflow/inductor/166777 2025-11-03T16:58:02.1941409Z * [new tag] ciflow/inductor/166788 -> ciflow/inductor/166788 2025-11-03T16:58:02.1942820Z * [new tag] ciflow/inductor/166791 -> ciflow/inductor/166791 2025-11-03T16:58:02.1944172Z * [new tag] ciflow/inductor/166793 -> ciflow/inductor/166793 2025-11-03T16:58:02.1945356Z * [new tag] ciflow/inductor/166803 -> ciflow/inductor/166803 2025-11-03T16:58:02.1946546Z * [new tag] ciflow/inductor/166808 -> ciflow/inductor/166808 2025-11-03T16:58:02.1947772Z * [new tag] ciflow/inductor/166822 -> ciflow/inductor/166822 2025-11-03T16:58:02.1949002Z * [new tag] ciflow/inductor/166834 -> ciflow/inductor/166834 2025-11-03T16:58:02.1950464Z * [new tag] ciflow/inductor/166839 -> ciflow/inductor/166839 2025-11-03T16:58:02.1951785Z * [new tag] ciflow/inductor/166846 -> ciflow/inductor/166846 2025-11-03T16:58:02.1953028Z * [new tag] ciflow/inductor/166855 -> ciflow/inductor/166855 2025-11-03T16:58:02.1954274Z * [new tag] ciflow/inductor/166867 -> ciflow/inductor/166867 2025-11-03T16:58:02.1955522Z * [new tag] ciflow/inductor/166868 -> ciflow/inductor/166868 2025-11-03T16:58:02.1956793Z * [new tag] ciflow/inductor/166869 -> ciflow/inductor/166869 2025-11-03T16:58:02.1958265Z * [new tag] ciflow/inductor/3b9a386 -> ciflow/inductor/3b9a386 2025-11-03T16:58:02.1959681Z * [new tag] ciflow/inductor/3d4b92b -> ciflow/inductor/3d4b92b 2025-11-03T16:58:02.1961116Z * [new tag] ciflow/inductor/d224ac7 -> ciflow/inductor/d224ac7 2025-11-03T16:58:02.1962597Z * [new tag] ciflow/linux-aarch64/166380 -> ciflow/linux-aarch64/166380 2025-11-03T16:58:02.1963742Z * [new tag] ciflow/linux-aarch64/166441 -> ciflow/linux-aarch64/166441 2025-11-03T16:58:02.1964898Z * [new tag] ciflow/linux-aarch64/166549 -> ciflow/linux-aarch64/166549 2025-11-03T16:58:02.1966061Z * [new tag] ciflow/linux-aarch64/166640 -> ciflow/linux-aarch64/166640 2025-11-03T16:58:02.1967317Z * [new tag] ciflow/linux-aarch64/166691 -> ciflow/linux-aarch64/166691 2025-11-03T16:58:02.1968472Z * [new tag] ciflow/linux-aarch64/166849 -> ciflow/linux-aarch64/166849 2025-11-03T16:58:02.1969864Z * [new tag] ciflow/mps/157553 -> ciflow/mps/157553 2025-11-03T16:58:02.1971033Z * [new tag] ciflow/mps/157554 -> ciflow/mps/157554 2025-11-03T16:58:02.1972579Z * [new tag] ciflow/mps/157635 -> ciflow/mps/157635 2025-11-03T16:58:02.1973694Z * [new tag] ciflow/mps/165952 -> ciflow/mps/165952 2025-11-03T16:58:02.1974868Z * [new tag] ciflow/mps/166254 -> ciflow/mps/166254 2025-11-03T16:58:02.1976023Z * [new tag] ciflow/mps/166273 -> ciflow/mps/166273 2025-11-03T16:58:02.1977354Z * [new tag] ciflow/mps/166396 -> ciflow/mps/166396 2025-11-03T16:58:02.1978500Z * [new tag] ciflow/mps/166615 -> ciflow/mps/166615 2025-11-03T16:58:02.1979773Z * [new tag] ciflow/mps/166712 -> ciflow/mps/166712 2025-11-03T16:58:02.1981206Z * [new tag] ciflow/nightly/158104 -> ciflow/nightly/158104 2025-11-03T16:58:02.1982514Z * [new tag] ciflow/nightly/166727 -> ciflow/nightly/166727 2025-11-03T16:58:02.1983943Z * [new tag] ciflow/op-benchmark/165915 -> ciflow/op-benchmark/165915 2025-11-03T16:58:02.1985167Z * [new tag] ciflow/op-benchmark/166331 -> ciflow/op-benchmark/166331 2025-11-03T16:58:02.1986285Z * [new tag] ciflow/op-benchmark/166640 -> ciflow/op-benchmark/166640 2025-11-03T16:58:02.1987447Z * [new tag] ciflow/op-benchmark/166652 -> ciflow/op-benchmark/166652 2025-11-03T16:58:02.1988598Z * [new tag] ciflow/op-benchmark/166731 -> ciflow/op-benchmark/166731 2025-11-03T16:58:02.1990151Z * [new tag] ciflow/periodic-rocm-mi200/166743 -> ciflow/periodic-rocm-mi200/166743 2025-11-03T16:58:02.1991675Z * [new tag] ciflow/periodic-rocm-mi300/166517 -> ciflow/periodic-rocm-mi300/166517 2025-11-03T16:58:02.1992803Z * [new tag] ciflow/periodic-rocm-mi300/166743 -> ciflow/periodic-rocm-mi300/166743 2025-11-03T16:58:02.1994384Z * [new tag] ciflow/periodic/054a2fd -> ciflow/periodic/054a2fd 2025-11-03T16:58:02.1995969Z * [new tag] ciflow/periodic/06e71c85583d9857f716b441eef9db40d3f27a39 -> ciflow/periodic/06e71c85583d9857f716b441eef9db40d3f27a39 2025-11-03T16:58:02.1997337Z * [new tag] ciflow/periodic/0e46a10aa7cf69354a0a0cbd268506b53a7cc882 -> ciflow/periodic/0e46a10aa7cf69354a0a0cbd268506b53a7cc882 2025-11-03T16:58:02.1998184Z * [new tag] ciflow/periodic/163490 -> ciflow/periodic/163490 2025-11-03T16:58:02.1999404Z * [new tag] ciflow/periodic/164202 -> ciflow/periodic/164202 2025-11-03T16:58:02.2000706Z * [new tag] ciflow/periodic/164938 -> ciflow/periodic/164938 2025-11-03T16:58:02.2001644Z * [new tag] ciflow/periodic/165885 -> ciflow/periodic/165885 2025-11-03T16:58:02.2003272Z * [new tag] ciflow/periodic/166517 -> ciflow/periodic/166517 2025-11-03T16:58:02.2005170Z * [new tag] ciflow/periodic/1e836bc769f9cfabc7659a172b8c2edee7c375d3 -> ciflow/periodic/1e836bc769f9cfabc7659a172b8c2edee7c375d3 2025-11-03T16:58:02.2006417Z * [new tag] ciflow/periodic/2a6d37d -> ciflow/periodic/2a6d37d 2025-11-03T16:58:02.2007715Z * [new tag] ciflow/periodic/317eeb8 -> ciflow/periodic/317eeb8 2025-11-03T16:58:02.2009045Z * [new tag] ciflow/periodic/3c32 -> ciflow/periodic/3c32 2025-11-03T16:58:02.2010502Z * [new tag] ciflow/periodic/3e98831 -> ciflow/periodic/3e98831 2025-11-03T16:58:02.2011986Z * [new tag] ciflow/periodic/687c15c0b3f01118536413d21efcd052838fa10d -> ciflow/periodic/687c15c0b3f01118536413d21efcd052838fa10d 2025-11-03T16:58:02.2013380Z * [new tag] ciflow/periodic/7379972cc0e8a4b4d88b4bea5c8be0aeffdb076a -> ciflow/periodic/7379972cc0e8a4b4d88b4bea5c8be0aeffdb076a 2025-11-03T16:58:02.2014667Z * [new tag] ciflow/periodic/94512-point -> ciflow/periodic/94512-point 2025-11-03T16:58:02.2016194Z * [new tag] ciflow/periodic/96b61844a722d2ae8b1a2dc283a6c0060b40782e -> ciflow/periodic/96b61844a722d2ae8b1a2dc283a6c0060b40782e 2025-11-03T16:58:02.2017578Z * [new tag] ciflow/periodic/be283297100ab86123e74b7a8372995d32b140c8 -> ciflow/periodic/be283297100ab86123e74b7a8372995d32b140c8 2025-11-03T16:58:02.2018952Z * [new tag] ciflow/periodic/bfc2050db975e589795cd3eceaed2e83bf89ad35 -> ciflow/periodic/bfc2050db975e589795cd3eceaed2e83bf89ad35 2025-11-03T16:58:02.2020353Z * [new tag] ciflow/periodic/c5701d0ab5f55b23243dd437f57e856970c45bb6 -> ciflow/periodic/c5701d0ab5f55b23243dd437f57e856970c45bb6 2025-11-03T16:58:02.2022002Z * [new tag] ciflow/periodic/cde81e92b95eee9af2879c9c75f7b03699ca72ad -> ciflow/periodic/cde81e92b95eee9af2879c9c75f7b03699ca72ad 2025-11-03T16:58:02.2023582Z * [new tag] ciflow/periodic/csl/test87519 -> ciflow/periodic/csl/test87519 2025-11-03T16:58:02.2024969Z * [new tag] ciflow/periodic/csltest88275 -> ciflow/periodic/csltest88275 2025-11-03T16:58:02.2026334Z * [new tag] ciflow/periodic/csltest88761 -> ciflow/periodic/csltest88761 2025-11-03T16:58:02.2027754Z * [new tag] ciflow/periodic/release_1.12 -> ciflow/periodic/release_1.12 2025-11-03T16:58:02.2029276Z * [new tag] ciflow/periodic/release_1.12.0 -> ciflow/periodic/release_1.12.0 2025-11-03T16:58:02.2030810Z * [new tag] ciflow/periodic/sha-ec5b83 -> ciflow/periodic/sha-ec5b83 2025-11-03T16:58:02.2032383Z * [new tag] ciflow/pull/166321 -> ciflow/pull/166321 2025-11-03T16:58:02.2034009Z * [new tag] ciflow/pull/166768 -> ciflow/pull/166768 2025-11-03T16:58:02.2035539Z * [new tag] ciflow/pull/18f425962656f009e151944d12f76e2f856b85df -> ciflow/pull/18f425962656f009e151944d12f76e2f856b85df 2025-11-03T16:58:02.2036809Z * [new tag] ciflow/pull/76780b1a3d208b142f067bdd655e7d4c5615efe1 -> ciflow/pull/76780b1a3d208b142f067bdd655e7d4c5615efe1 2025-11-03T16:58:02.2038441Z * [new tag] ciflow/pull/79aee77381b21d41c77148e5ff84c4b351aaf144 -> ciflow/pull/79aee77381b21d41c77148e5ff84c4b351aaf144 2025-11-03T16:58:02.2040153Z * [new tag] ciflow/pull/bc03d7c97462abfd45fc1b7b73d5c7acbbd2c185 -> ciflow/pull/bc03d7c97462abfd45fc1b7b73d5c7acbbd2c185 2025-11-03T16:58:02.2041532Z * [new tag] ciflow/pull/cee03634dab2dbf259d84bc389bdea80de0f80ac -> ciflow/pull/cee03634dab2dbf259d84bc389bdea80de0f80ac 2025-11-03T16:58:02.2043120Z * [new tag] ciflow/pull/d962bed15798961ffd10e64068f86a2cec411ffb -> ciflow/pull/d962bed15798961ffd10e64068f86a2cec411ffb 2025-11-03T16:58:02.2044580Z * [new tag] ciflow/pull/e471800dceb8a6592562fc4fc96a5e80bc494c0a -> ciflow/pull/e471800dceb8a6592562fc4fc96a5e80bc494c0a 2025-11-03T16:58:02.2045956Z * [new tag] ciflow/pull/f013e804c876a34cbc86a6a475597e3e29812883 -> ciflow/pull/f013e804c876a34cbc86a6a475597e3e29812883 2025-11-03T16:58:02.2047333Z * [new tag] ciflow/pull/f5cb9a4c68d9271c58ef4d3257210984b8e85099 -> ciflow/pull/f5cb9a4c68d9271c58ef4d3257210984b8e85099 2025-11-03T16:58:02.2048567Z * [new tag] ciflow/rocm-mi300/162053 -> ciflow/rocm-mi300/162053 2025-11-03T16:58:02.2049727Z * [new tag] ciflow/rocm-mi300/165548 -> ciflow/rocm-mi300/165548 2025-11-03T16:58:02.2051065Z * [new tag] ciflow/rocm-mi300/166317 -> ciflow/rocm-mi300/166317 2025-11-03T16:58:02.2052300Z * [new tag] ciflow/rocm-mi300/166743 -> ciflow/rocm-mi300/166743 2025-11-03T16:58:02.2053985Z * [new tag] ciflow/rocm-mi355/162053 -> ciflow/rocm-mi355/162053 2025-11-03T16:58:02.2055143Z * [new tag] ciflow/rocm-mi355/166743 -> ciflow/rocm-mi355/166743 2025-11-03T16:58:02.2056555Z * [new tag] ciflow/rocm/115316 -> ciflow/rocm/115316 2025-11-03T16:58:02.2057688Z * [new tag] ciflow/rocm/148492 -> ciflow/rocm/148492 2025-11-03T16:58:02.2058805Z * [new tag] ciflow/rocm/151845 -> ciflow/rocm/151845 2025-11-03T16:58:02.2059936Z * [new tag] ciflow/rocm/157553 -> ciflow/rocm/157553 2025-11-03T16:58:02.2061065Z * [new tag] ciflow/rocm/157635 -> ciflow/rocm/157635 2025-11-03T16:58:02.2062317Z * [new tag] ciflow/rocm/160685 -> ciflow/rocm/160685 2025-11-03T16:58:02.2063402Z * [new tag] ciflow/rocm/161404 -> ciflow/rocm/161404 2025-11-03T16:58:02.2064543Z * [new tag] ciflow/rocm/162053 -> ciflow/rocm/162053 2025-11-03T16:58:02.2065681Z * [new tag] ciflow/rocm/163369 -> ciflow/rocm/163369 2025-11-03T16:58:02.2066892Z * [new tag] ciflow/rocm/164657 -> ciflow/rocm/164657 2025-11-03T16:58:02.2068030Z * [new tag] ciflow/rocm/165132 -> ciflow/rocm/165132 2025-11-03T16:58:02.2069484Z * [new tag] ciflow/rocm/165541 -> ciflow/rocm/165541 2025-11-03T16:58:02.2071019Z * [new tag] ciflow/rocm/165548 -> ciflow/rocm/165548 2025-11-03T16:58:02.2072635Z * [new tag] ciflow/rocm/165952 -> ciflow/rocm/165952 2025-11-03T16:58:02.2074202Z * [new tag] ciflow/rocm/165997 -> ciflow/rocm/165997 2025-11-03T16:58:02.2075427Z * [new tag] ciflow/rocm/166250 -> ciflow/rocm/166250 2025-11-03T16:58:02.2076774Z * [new tag] ciflow/rocm/166317 -> ciflow/rocm/166317 2025-11-03T16:58:02.2077897Z * [new tag] ciflow/rocm/166377 -> ciflow/rocm/166377 2025-11-03T16:58:02.2079289Z * [new tag] ciflow/rocm/166391 -> ciflow/rocm/166391 2025-11-03T16:58:02.2080508Z * [new tag] ciflow/rocm/166515 -> ciflow/rocm/166515 2025-11-03T16:58:02.2081820Z * [new tag] ciflow/rocm/166517 -> ciflow/rocm/166517 2025-11-03T16:58:02.2083256Z * [new tag] ciflow/rocm/166526 -> ciflow/rocm/166526 2025-11-03T16:58:02.2084436Z * [new tag] ciflow/rocm/166743 -> ciflow/rocm/166743 2025-11-03T16:58:02.2085660Z * [new tag] ciflow/rocm/166744 -> ciflow/rocm/166744 2025-11-03T16:58:02.2086856Z * [new tag] ciflow/rocm/166752 -> ciflow/rocm/166752 2025-11-03T16:58:02.2088067Z * [new tag] ciflow/rocm/166764 -> ciflow/rocm/166764 2025-11-03T16:58:02.2089262Z * [new tag] ciflow/rocm/166834 -> ciflow/rocm/166834 2025-11-03T16:58:02.2090513Z * [new tag] ciflow/rocm/166870 -> ciflow/rocm/166870 2025-11-03T16:58:02.2092044Z * [new tag] ciflow/slow/01c7106 -> ciflow/slow/01c7106 2025-11-03T16:58:02.2093248Z * [new tag] ciflow/slow/0577043 -> ciflow/slow/0577043 2025-11-03T16:58:02.2094866Z * [new tag] ciflow/slow/0d5b74da0cab798fbfdb9caa53fad816999c8386-sdym -> ciflow/slow/0d5b74da0cab798fbfdb9caa53fad816999c8386-sdym 2025-11-03T16:58:02.2095952Z * [new tag] ciflow/slow/0e81104 -> ciflow/slow/0e81104 2025-11-03T16:58:02.2097105Z * [new tag] ciflow/slow/166844 -> ciflow/slow/166844 2025-11-03T16:58:02.2098415Z * [new tag] ciflow/slow/1732077 -> ciflow/slow/1732077 2025-11-03T16:58:02.2099748Z * [new tag] ciflow/slow/187eb7c -> ciflow/slow/187eb7c 2025-11-03T16:58:02.2101026Z * [new tag] ciflow/slow/1faef89 -> ciflow/slow/1faef89 2025-11-03T16:58:02.2102862Z * [new tag] ciflow/slow/3920ec1 -> ciflow/slow/3920ec1 2025-11-03T16:58:02.2104423Z * [new tag] ciflow/slow/3b7c6b2 -> ciflow/slow/3b7c6b2 2025-11-03T16:58:02.2105781Z * [new tag] ciflow/slow/59a3759 -> ciflow/slow/59a3759 2025-11-03T16:58:02.2107158Z * [new tag] ciflow/slow/70ef0bb -> ciflow/slow/70ef0bb 2025-11-03T16:58:02.2108528Z * [new tag] ciflow/slow/788ff06 -> ciflow/slow/788ff06 2025-11-03T16:58:02.2110184Z * [new tag] ciflow/slow/8751002215790a3a88750faa8f4366933e296693-sdym -> ciflow/slow/8751002215790a3a88750faa8f4366933e296693-sdym 2025-11-03T16:58:02.2111322Z * [new tag] ciflow/slow/9d85864 -> ciflow/slow/9d85864 2025-11-03T16:58:02.2112684Z * [new tag] ciflow/slow/9ffad5b -> ciflow/slow/9ffad5b 2025-11-03T16:58:02.2114014Z * [new tag] ciflow/slow/a206e8b -> ciflow/slow/a206e8b 2025-11-03T16:58:02.2115485Z * [new tag] ciflow/slow/a837609 -> ciflow/slow/a837609 2025-11-03T16:58:02.2116843Z * [new tag] ciflow/slow/af841f3 -> ciflow/slow/af841f3 2025-11-03T16:58:02.2118541Z * [new tag] ciflow/slow/da3aba1e46157c4df504b067477cdf2b3c96b194-sdym -> ciflow/slow/da3aba1e46157c4df504b067477cdf2b3c96b194-sdym 2025-11-03T16:58:02.2120127Z * [new tag] ciflow/trunk/04d6a6f3392a87af8baf620dc0a323ffbb188c12 -> ciflow/trunk/04d6a6f3392a87af8baf620dc0a323ffbb188c12 2025-11-03T16:58:02.2121770Z * [new tag] ciflow/trunk/0573747b6af273c6ac8cf9f1d81dc56066a57445 -> ciflow/trunk/0573747b6af273c6ac8cf9f1d81dc56066a57445 2025-11-03T16:58:02.2123926Z * [new tag] ciflow/trunk/137400 -> ciflow/trunk/137400 2025-11-03T16:58:02.2125238Z * [new tag] ciflow/trunk/148180 -> ciflow/trunk/148180 2025-11-03T16:58:02.2126111Z * [new tag] ciflow/trunk/148328 -> ciflow/trunk/148328 2025-11-03T16:58:02.2127814Z * [new tag] ciflow/trunk/148492 -> ciflow/trunk/148492 2025-11-03T16:58:02.2128918Z * [new tag] ciflow/trunk/151845 -> ciflow/trunk/151845 2025-11-03T16:58:02.2130375Z * [new tag] ciflow/trunk/152624 -> ciflow/trunk/152624 2025-11-03T16:58:02.2131896Z * [new tag] ciflow/trunk/155731 -> ciflow/trunk/155731 2025-11-03T16:58:02.2133137Z * [new tag] ciflow/trunk/156812 -> ciflow/trunk/156812 2025-11-03T16:58:02.2134520Z * [new tag] ciflow/trunk/157432 -> ciflow/trunk/157432 2025-11-03T16:58:02.2135748Z * [new tag] ciflow/trunk/158104 -> ciflow/trunk/158104 2025-11-03T16:58:02.2136982Z * [new tag] ciflow/trunk/160417 -> ciflow/trunk/160417 2025-11-03T16:58:02.2138379Z * [new tag] ciflow/trunk/161035 -> ciflow/trunk/161035 2025-11-03T16:58:02.2139726Z * [new tag] ciflow/trunk/161771 -> ciflow/trunk/161771 2025-11-03T16:58:02.2140927Z * [new tag] ciflow/trunk/162053 -> ciflow/trunk/162053 2025-11-03T16:58:02.2142383Z * [new tag] ciflow/trunk/162454 -> ciflow/trunk/162454 2025-11-03T16:58:02.2143741Z * [new tag] ciflow/trunk/162564 -> ciflow/trunk/162564 2025-11-03T16:58:02.2144969Z * [new tag] ciflow/trunk/162709 -> ciflow/trunk/162709 2025-11-03T16:58:02.2146203Z * [new tag] ciflow/trunk/162795 -> ciflow/trunk/162795 2025-11-03T16:58:02.2147409Z * [new tag] ciflow/trunk/163249 -> ciflow/trunk/163249 2025-11-03T16:58:02.2148640Z * [new tag] ciflow/trunk/163490 -> ciflow/trunk/163490 2025-11-03T16:58:02.2149843Z * [new tag] ciflow/trunk/164202 -> ciflow/trunk/164202 2025-11-03T16:58:02.2151067Z * [new tag] ciflow/trunk/164938 -> ciflow/trunk/164938 2025-11-03T16:58:02.2152263Z * [new tag] ciflow/trunk/164979 -> ciflow/trunk/164979 2025-11-03T16:58:02.2153475Z * [new tag] ciflow/trunk/164980 -> ciflow/trunk/164980 2025-11-03T16:58:02.2154692Z * [new tag] ciflow/trunk/165005 -> ciflow/trunk/165005 2025-11-03T16:58:02.2155923Z * [new tag] ciflow/trunk/165029 -> ciflow/trunk/165029 2025-11-03T16:58:02.2157328Z * [new tag] ciflow/trunk/165049 -> ciflow/trunk/165049 2025-11-03T16:58:02.2158719Z * [new tag] ciflow/trunk/165067 -> ciflow/trunk/165067 2025-11-03T16:58:02.2159932Z * [new tag] ciflow/trunk/165132 -> ciflow/trunk/165132 2025-11-03T16:58:02.2161359Z * [new tag] ciflow/trunk/165216 -> ciflow/trunk/165216 2025-11-03T16:58:02.2162556Z * [new tag] ciflow/trunk/165274 -> ciflow/trunk/165274 2025-11-03T16:58:02.2163893Z * [new tag] ciflow/trunk/165431 -> ciflow/trunk/165431 2025-11-03T16:58:02.2165079Z * [new tag] ciflow/trunk/165548 -> ciflow/trunk/165548 2025-11-03T16:58:02.2166309Z * [new tag] ciflow/trunk/165642 -> ciflow/trunk/165642 2025-11-03T16:58:02.2167511Z * [new tag] ciflow/trunk/165646 -> ciflow/trunk/165646 2025-11-03T16:58:02.2168739Z * [new tag] ciflow/trunk/165728 -> ciflow/trunk/165728 2025-11-03T16:58:02.2169951Z * [new tag] ciflow/trunk/165824 -> ciflow/trunk/165824 2025-11-03T16:58:02.2171167Z * [new tag] ciflow/trunk/165885 -> ciflow/trunk/165885 2025-11-03T16:58:02.2172377Z * [new tag] ciflow/trunk/165922 -> ciflow/trunk/165922 2025-11-03T16:58:02.2173664Z * [new tag] ciflow/trunk/165972 -> ciflow/trunk/165972 2025-11-03T16:58:02.2174845Z * [new tag] ciflow/trunk/166010 -> ciflow/trunk/166010 2025-11-03T16:58:02.2176011Z * [new tag] ciflow/trunk/166071 -> ciflow/trunk/166071 2025-11-03T16:58:02.2177353Z * [new tag] ciflow/trunk/166160 -> ciflow/trunk/166160 2025-11-03T16:58:02.2178616Z * [new tag] ciflow/trunk/166170 -> ciflow/trunk/166170 2025-11-03T16:58:02.2179922Z * [new tag] ciflow/trunk/166198 -> ciflow/trunk/166198 2025-11-03T16:58:02.2181117Z * [new tag] ciflow/trunk/166294 -> ciflow/trunk/166294 2025-11-03T16:58:02.2182471Z * [new tag] ciflow/trunk/166299 -> ciflow/trunk/166299 2025-11-03T16:58:02.2183666Z * [new tag] ciflow/trunk/166317 -> ciflow/trunk/166317 2025-11-03T16:58:02.2184895Z * [new tag] ciflow/trunk/166321 -> ciflow/trunk/166321 2025-11-03T16:58:02.2187514Z * [new tag] ciflow/trunk/166361 -> ciflow/trunk/166361 2025-11-03T16:58:02.2188065Z * [new tag] ciflow/trunk/166367 -> ciflow/trunk/166367 2025-11-03T16:58:02.2188839Z * [new tag] ciflow/trunk/166376 -> ciflow/trunk/166376 2025-11-03T16:58:02.2190004Z * [new tag] ciflow/trunk/166379 -> ciflow/trunk/166379 2025-11-03T16:58:02.2191109Z * [new tag] ciflow/trunk/166380 -> ciflow/trunk/166380 2025-11-03T16:58:02.2192330Z * [new tag] ciflow/trunk/166402 -> ciflow/trunk/166402 2025-11-03T16:58:02.2193488Z * [new tag] ciflow/trunk/166416 -> ciflow/trunk/166416 2025-11-03T16:58:02.2194748Z * [new tag] ciflow/trunk/166431 -> ciflow/trunk/166431 2025-11-03T16:58:02.2195962Z * [new tag] ciflow/trunk/166433 -> ciflow/trunk/166433 2025-11-03T16:58:02.2197255Z * [new tag] ciflow/trunk/166459 -> ciflow/trunk/166459 2025-11-03T16:58:02.2198493Z * [new tag] ciflow/trunk/166469 -> ciflow/trunk/166469 2025-11-03T16:58:02.2199887Z * [new tag] ciflow/trunk/166493 -> ciflow/trunk/166493 2025-11-03T16:58:02.2201062Z * [new tag] ciflow/trunk/166526 -> ciflow/trunk/166526 2025-11-03T16:58:02.2202354Z * [new tag] ciflow/trunk/166536 -> ciflow/trunk/166536 2025-11-03T16:58:02.2203574Z * [new tag] ciflow/trunk/166541 -> ciflow/trunk/166541 2025-11-03T16:58:02.2204811Z * [new tag] ciflow/trunk/166549 -> ciflow/trunk/166549 2025-11-03T16:58:02.2206224Z * [new tag] ciflow/trunk/166560 -> ciflow/trunk/166560 2025-11-03T16:58:02.2207481Z * [new tag] ciflow/trunk/166608 -> ciflow/trunk/166608 2025-11-03T16:58:02.2209152Z * [new tag] ciflow/trunk/166610 -> ciflow/trunk/166610 2025-11-03T16:58:02.2210502Z * [new tag] ciflow/trunk/166613 -> ciflow/trunk/166613 2025-11-03T16:58:02.2211689Z * [new tag] ciflow/trunk/166640 -> ciflow/trunk/166640 2025-11-03T16:58:02.2212897Z * [new tag] ciflow/trunk/166688 -> ciflow/trunk/166688 2025-11-03T16:58:02.2214093Z * [new tag] ciflow/trunk/166689 -> ciflow/trunk/166689 2025-11-03T16:58:02.2215471Z * [new tag] ciflow/trunk/166715 -> ciflow/trunk/166715 2025-11-03T16:58:02.2216641Z * [new tag] ciflow/trunk/166727 -> ciflow/trunk/166727 2025-11-03T16:58:02.2217843Z * [new tag] ciflow/trunk/166744 -> ciflow/trunk/166744 2025-11-03T16:58:02.2219211Z * [new tag] ciflow/trunk/166751 -> ciflow/trunk/166751 2025-11-03T16:58:02.2220691Z * [new tag] ciflow/trunk/166768 -> ciflow/trunk/166768 2025-11-03T16:58:02.2222091Z * [new tag] ciflow/trunk/166791 -> ciflow/trunk/166791 2025-11-03T16:58:02.2223235Z * [new tag] ciflow/trunk/166793 -> ciflow/trunk/166793 2025-11-03T16:58:02.2224477Z * [new tag] ciflow/trunk/166805 -> ciflow/trunk/166805 2025-11-03T16:58:02.2225708Z * [new tag] ciflow/trunk/166806 -> ciflow/trunk/166806 2025-11-03T16:58:02.2226918Z * [new tag] ciflow/trunk/166813 -> ciflow/trunk/166813 2025-11-03T16:58:02.2228122Z * [new tag] ciflow/trunk/166830 -> ciflow/trunk/166830 2025-11-03T16:58:02.2229425Z * [new tag] ciflow/trunk/166839 -> ciflow/trunk/166839 2025-11-03T16:58:02.2230905Z * [new tag] ciflow/trunk/166840 -> ciflow/trunk/166840 2025-11-03T16:58:02.2232119Z * [new tag] ciflow/trunk/166844 -> ciflow/trunk/166844 2025-11-03T16:58:02.2233391Z * [new tag] ciflow/trunk/166847 -> ciflow/trunk/166847 2025-11-03T16:58:02.2234566Z * [new tag] ciflow/trunk/166861 -> ciflow/trunk/166861 2025-11-03T16:58:02.2236099Z * [new tag] ciflow/trunk/18f425962656f009e151944d12f76e2f856b85df -> ciflow/trunk/18f425962656f009e151944d12f76e2f856b85df 2025-11-03T16:58:02.2237191Z * [new tag] ciflow/trunk/76780b1a3d208b142f067bdd655e7d4c5615efe1 -> ciflow/trunk/76780b1a3d208b142f067bdd655e7d4c5615efe1 2025-11-03T16:58:02.2238937Z * [new tag] ciflow/trunk/a19e92d43355489c8d33325321e688a6e6182267 -> ciflow/trunk/a19e92d43355489c8d33325321e688a6e6182267 2025-11-03T16:58:02.2240229Z * [new tag] ciflow/trunk/bc03d7c97462abfd45fc1b7b73d5c7acbbd2c185 -> ciflow/trunk/bc03d7c97462abfd45fc1b7b73d5c7acbbd2c185 2025-11-03T16:58:02.2241696Z * [new tag] ciflow/trunk/c3dc0c7089f3204c36961acc227a3ff503530094 -> ciflow/trunk/c3dc0c7089f3204c36961acc227a3ff503530094 2025-11-03T16:58:02.2242995Z * [new tag] ciflow/trunk/cee03634dab2dbf259d84bc389bdea80de0f80ac -> ciflow/trunk/cee03634dab2dbf259d84bc389bdea80de0f80ac 2025-11-03T16:58:02.2244053Z * [new tag] ciflow/trunk/d962bed15798961ffd10e64068f86a2cec411ffb -> ciflow/trunk/d962bed15798961ffd10e64068f86a2cec411ffb 2025-11-03T16:58:02.2245432Z * [new tag] ciflow/trunk/e471800dceb8a6592562fc4fc96a5e80bc494c0a -> ciflow/trunk/e471800dceb8a6592562fc4fc96a5e80bc494c0a 2025-11-03T16:58:02.2246718Z * [new tag] ciflow/trunk/f013e804c876a34cbc86a6a475597e3e29812883 -> ciflow/trunk/f013e804c876a34cbc86a6a475597e3e29812883 2025-11-03T16:58:02.2248218Z * [new tag] ciflow/unstable/123 -> ciflow/unstable/123 2025-11-03T16:58:02.2249621Z * [new tag] ciflow/vllm/165274 -> ciflow/vllm/165274 2025-11-03T16:58:02.2250772Z * [new tag] ciflow/vllm/166494 -> ciflow/vllm/166494 2025-11-03T16:58:02.2252233Z * [new tag] ciflow/win-arm64/158104 -> ciflow/win-arm64/158104 2025-11-03T16:58:02.2253582Z * [new tag] ciflow/xpu/156812 -> ciflow/xpu/156812 2025-11-03T16:58:02.2254704Z * [new tag] ciflow/xpu/161246 -> ciflow/xpu/161246 2025-11-03T16:58:02.2255861Z * [new tag] ciflow/xpu/162454 -> ciflow/xpu/162454 2025-11-03T16:58:02.2257128Z * [new tag] ciflow/xpu/162475 -> ciflow/xpu/162475 2025-11-03T16:58:02.2258273Z * [new tag] ciflow/xpu/162564 -> ciflow/xpu/162564 2025-11-03T16:58:02.2259395Z * [new tag] ciflow/xpu/165049 -> ciflow/xpu/165049 2025-11-03T16:58:02.2260599Z * [new tag] ciflow/xpu/165423 -> ciflow/xpu/165423 2025-11-03T16:58:02.2261799Z * [new tag] ciflow/xpu/166292 -> ciflow/xpu/166292 2025-11-03T16:58:02.2263033Z * [new tag] ciflow/xpu/166299 -> ciflow/xpu/166299 2025-11-03T16:58:02.2263901Z * [new tag] ciflow/xpu/166376 -> ciflow/xpu/166376 2025-11-03T16:58:02.2265250Z * [new tag] ciflow/xpu/166396 -> ciflow/xpu/166396 2025-11-03T16:58:02.2266645Z * [new tag] ciflow/xpu/166424 -> ciflow/xpu/166424 2025-11-03T16:58:02.2268289Z * [new tag] ciflow/xpu/166436 -> ciflow/xpu/166436 2025-11-03T16:58:02.2269724Z * [new tag] ciflow/xpu/166495 -> ciflow/xpu/166495 2025-11-03T16:58:02.2271137Z * [new tag] ciflow/xpu/166504 -> ciflow/xpu/166504 2025-11-03T16:58:02.2272334Z * [new tag] ciflow/xpu/166613 -> ciflow/xpu/166613 2025-11-03T16:58:02.2273724Z * [new tag] ciflow/xpu/166684 -> ciflow/xpu/166684 2025-11-03T16:58:02.2275023Z * [new tag] ciflow/xpu/166830 -> ciflow/xpu/166830 2025-11-03T16:58:02.2276411Z * [new tag] ciflow/xpu/166834 -> ciflow/xpu/166834 2025-11-03T16:58:02.2277556Z * [new tag] ciflow/xpu/166839 -> ciflow/xpu/166839 2025-11-03T16:58:02.2278749Z * [new tag] ciflow/xpu/166840 -> ciflow/xpu/166840 2025-11-03T16:58:02.2279964Z * [new tag] ciflow/xpu/166847 -> ciflow/xpu/166847 2025-11-03T16:58:02.2281228Z * [new tag] ciflow/xpu/166861 -> ciflow/xpu/166861 2025-11-03T16:58:02.2282404Z * [new tag] cslpull75 -> cslpull75 2025-11-03T16:58:02.2283612Z * [new tag] cslpull76 -> cslpull76 2025-11-03T16:58:02.2284792Z * [new tag] cslpull77 -> cslpull77 2025-11-03T16:58:02.2286091Z * [new tag] cslpull78 -> cslpull78 2025-11-03T16:58:02.2287538Z * [new tag] cslpull79 -> cslpull79 2025-11-03T16:58:02.2289030Z * [new tag] cslpull80 -> cslpull80 2025-11-03T16:58:02.2290301Z * [new tag] cslpull81 -> cslpull81 2025-11-03T16:58:02.2291617Z * [new tag] cslpull82 -> cslpull82 2025-11-03T16:58:02.2292965Z * [new tag] cslpull83 -> cslpull83 2025-11-03T16:58:02.2294204Z * [new tag] cslpull84 -> cslpull84 2025-11-03T16:58:02.2295451Z * [new tag] cslpull85 -> cslpull85 2025-11-03T16:58:02.2296785Z * [new tag] cslpull86 -> cslpull86 2025-11-03T16:58:02.2298047Z * [new tag] cslpull87 -> cslpull87 2025-11-03T16:58:02.2299368Z * [new tag] cslpull88 -> cslpull88 2025-11-03T16:58:02.2300675Z * [new tag] cslpull89 -> cslpull89 2025-11-03T16:58:02.2301954Z * [new tag] cslpull90 -> cslpull90 2025-11-03T16:58:02.2303722Z * [new tag] cslpull91 -> cslpull91 2025-11-03T16:58:02.2304924Z * [new tag] cslpull92 -> cslpull92 2025-11-03T16:58:02.2306178Z * [new tag] flight_5 -> flight_5 2025-11-03T16:58:02.2307688Z * [new tag] flight_5.1 -> flight_5.1 2025-11-03T16:58:02.2308991Z * [new tag] flight_5.2 -> flight_5.2 2025-11-03T16:58:02.2310318Z * [new tag] flight_5.3 -> flight_5.3 2025-11-03T16:58:02.2311569Z * [new tag] forpull1 -> forpull1 2025-11-03T16:58:02.2313189Z * [new tag] malfet/tag-2ef5611 -> malfet/tag-2ef5611 2025-11-03T16:58:02.2314518Z * [new tag] malfet/tag-317b1a0 -> malfet/tag-317b1a0 2025-11-03T16:58:02.2315936Z * [new tag] malfet/tag-ec6f767 -> malfet/tag-ec6f767 2025-11-03T16:58:02.2317222Z * [new tag] nightly-binary -> nightly-binary 2025-11-03T16:58:02.2318545Z * [new tag] sqzhang_flight4_plus -> sqzhang_flight4_plus 2025-11-03T16:58:02.2319901Z * [new tag] sqzhang_flight_3 -> sqzhang_flight_3 2025-11-03T16:58:02.2321801Z * [new tag] trunk/000f49551bfda0683d395e1fe67f53cfb785868f -> trunk/000f49551bfda0683d395e1fe67f53cfb785868f 2025-11-03T16:58:02.2324792Z * [new tag] trunk/009ea77234788b29910b1205679036eaae3ee3ac -> trunk/009ea77234788b29910b1205679036eaae3ee3ac 2025-11-03T16:58:02.2326471Z * [new tag] trunk/0187db88d463905e5f7dcdf7689519e2807e5a12 -> trunk/0187db88d463905e5f7dcdf7689519e2807e5a12 2025-11-03T16:58:02.2327969Z * [new tag] trunk/02095cc09db01f1fecae436ac1a6ecdf2cfca0db -> trunk/02095cc09db01f1fecae436ac1a6ecdf2cfca0db 2025-11-03T16:58:02.2329334Z * [new tag] trunk/030de07affa347a80fa9f11ab8e9234fc6d7fd4d -> trunk/030de07affa347a80fa9f11ab8e9234fc6d7fd4d 2025-11-03T16:58:02.2330785Z * [new tag] trunk/034e951b0cfb02d7b55327cd482e58cf2695dca0 -> trunk/034e951b0cfb02d7b55327cd482e58cf2695dca0 2025-11-03T16:58:02.2331935Z * [new tag] trunk/04d6a6f3392a87af8baf620dc0a323ffbb188c12 -> trunk/04d6a6f3392a87af8baf620dc0a323ffbb188c12 2025-11-03T16:58:02.2333068Z * [new tag] trunk/0573747b6af273c6ac8cf9f1d81dc56066a57445 -> trunk/0573747b6af273c6ac8cf9f1d81dc56066a57445 2025-11-03T16:58:02.2334486Z * [new tag] trunk/061fa73c97dcfcb39077ce8fc4042e7f22824b22 -> trunk/061fa73c97dcfcb39077ce8fc4042e7f22824b22 2025-11-03T16:58:02.2335923Z * [new tag] trunk/0674e0a0f14775f920296e9dfb8b61e4960bf99d -> trunk/0674e0a0f14775f920296e9dfb8b61e4960bf99d 2025-11-03T16:58:02.2337100Z * [new tag] trunk/06e71c85583d9857f716b441eef9db40d3f27a39 -> trunk/06e71c85583d9857f716b441eef9db40d3f27a39 2025-11-03T16:58:02.2338473Z * [new tag] trunk/08ae55021eafba28b304e78ef2da9a76741ce039 -> trunk/08ae55021eafba28b304e78ef2da9a76741ce039 2025-11-03T16:58:02.2339800Z * [new tag] trunk/08b0a8f11a00e0442bd3b3066617a501dca29876 -> trunk/08b0a8f11a00e0442bd3b3066617a501dca29876 2025-11-03T16:58:02.2341187Z * [new tag] trunk/08f4535378b9ab5c7ba74828bb20ad535e6de8eb -> trunk/08f4535378b9ab5c7ba74828bb20ad535e6de8eb 2025-11-03T16:58:02.2342632Z * [new tag] trunk/0918bf321c2cfaebb91c5f9e5f0ebe462962328e -> trunk/0918bf321c2cfaebb91c5f9e5f0ebe462962328e 2025-11-03T16:58:02.2344003Z * [new tag] trunk/0947765eb9208996f221dbcb088df800be3953d7 -> trunk/0947765eb9208996f221dbcb088df800be3953d7 2025-11-03T16:58:02.2345309Z * [new tag] trunk/0a3ac47c0a36d7ed619e6577294648a7c14b1ae1 -> trunk/0a3ac47c0a36d7ed619e6577294648a7c14b1ae1 2025-11-03T16:58:02.2346697Z * [new tag] trunk/0ae3e306217a67362371267bbba0ac3fa0855fdc -> trunk/0ae3e306217a67362371267bbba0ac3fa0855fdc 2025-11-03T16:58:02.2348029Z * [new tag] trunk/0d3a4f7155508656bd928527840ed155e8b1a88b -> trunk/0d3a4f7155508656bd928527840ed155e8b1a88b 2025-11-03T16:58:02.2349395Z * [new tag] trunk/0d4992c1702d9d14bc73c29d44029adb7e698eba -> trunk/0d4992c1702d9d14bc73c29d44029adb7e698eba 2025-11-03T16:58:02.2350700Z * [new tag] trunk/0d50e5d8d4132cb09508b3e35701e7b06433420e -> trunk/0d50e5d8d4132cb09508b3e35701e7b06433420e 2025-11-03T16:58:02.2352059Z * [new tag] trunk/0d81bb7f9cef1b735ff373a245aaa2bc037b31b0 -> trunk/0d81bb7f9cef1b735ff373a245aaa2bc037b31b0 2025-11-03T16:58:02.2353351Z * [new tag] trunk/0e19561e235372a3fbed81338fa4e674116c0448 -> trunk/0e19561e235372a3fbed81338fa4e674116c0448 2025-11-03T16:58:02.2354719Z * [new tag] trunk/0e46a10aa7cf69354a0a0cbd268506b53a7cc882 -> trunk/0e46a10aa7cf69354a0a0cbd268506b53a7cc882 2025-11-03T16:58:02.2355886Z * [new tag] trunk/0eacd934bc1c1bee002fcd084300d596d789d166 -> trunk/0eacd934bc1c1bee002fcd084300d596d789d166 2025-11-03T16:58:02.2357219Z * [new tag] trunk/0ec054982304473b51e6ded40c2dc3993d4036be -> trunk/0ec054982304473b51e6ded40c2dc3993d4036be 2025-11-03T16:58:02.2358463Z * [new tag] trunk/104b86861862ccc47f6df6528347a017bef813ad -> trunk/104b86861862ccc47f6df6528347a017bef813ad 2025-11-03T16:58:02.2359952Z * [new tag] trunk/108bb224f77842593009214ebf6258030b934642 -> trunk/108bb224f77842593009214ebf6258030b934642 2025-11-03T16:58:02.2361196Z * [new tag] trunk/110efe4df4759006ce5aad4ae1b87d93dcf33c3b -> trunk/110efe4df4759006ce5aad4ae1b87d93dcf33c3b 2025-11-03T16:58:02.2362541Z * [new tag] trunk/112960541529ab0d27bb8aa8b487ef194e048494 -> trunk/112960541529ab0d27bb8aa8b487ef194e048494 2025-11-03T16:58:02.2364147Z * [new tag] trunk/121235956bab7430fb8d080cee209607f8387ead -> trunk/121235956bab7430fb8d080cee209607f8387ead 2025-11-03T16:58:02.2365656Z * [new tag] trunk/12577064dddfc6f5daf66c5b5a73cb418a588f20 -> trunk/12577064dddfc6f5daf66c5b5a73cb418a588f20 2025-11-03T16:58:02.2366868Z * [new tag] trunk/12c12466b082834d23c1f512130da9280badfd28 -> trunk/12c12466b082834d23c1f512130da9280badfd28 2025-11-03T16:58:02.2368428Z * [new tag] trunk/13413b3b07cc72fa9c2671b2535f7e54c1b30ca2 -> trunk/13413b3b07cc72fa9c2671b2535f7e54c1b30ca2 2025-11-03T16:58:02.2369648Z * [new tag] trunk/13549e0e105291721f6f80ceb8f520ca272b971d -> trunk/13549e0e105291721f6f80ceb8f520ca272b971d 2025-11-03T16:58:02.2370954Z * [new tag] trunk/14102fb1f337d01ca61584011f5f657d14da5de6 -> trunk/14102fb1f337d01ca61584011f5f657d14da5de6 2025-11-03T16:58:02.2372647Z * [new tag] trunk/1425b40f29ab11ce21f7a98de9e35551213efe27 -> trunk/1425b40f29ab11ce21f7a98de9e35551213efe27 2025-11-03T16:58:02.2373776Z * [new tag] trunk/14d4a77495dc80da9cd7c5c29b21aaf4613ddc5b -> trunk/14d4a77495dc80da9cd7c5c29b21aaf4613ddc5b 2025-11-03T16:58:02.2375058Z * [new tag] trunk/160ab53dd57e67b3574763615cf8b33249e9afa5 -> trunk/160ab53dd57e67b3574763615cf8b33249e9afa5 2025-11-03T16:58:02.2376383Z * [new tag] trunk/16212f0d6bc1c4fb6ac0498badbe2fb74d5ff47b -> trunk/16212f0d6bc1c4fb6ac0498badbe2fb74d5ff47b 2025-11-03T16:58:02.2377895Z * [new tag] trunk/1656b253c5fcc34a53a9991303170892c6d90c1d -> trunk/1656b253c5fcc34a53a9991303170892c6d90c1d 2025-11-03T16:58:02.2379161Z * [new tag] trunk/173bcda436e9a00aa8215a414efc0c06cf4840a8 -> trunk/173bcda436e9a00aa8215a414efc0c06cf4840a8 2025-11-03T16:58:02.2380623Z * [new tag] trunk/1764f3a9c8eea4cf0aba853348446948597b7b78 -> trunk/1764f3a9c8eea4cf0aba853348446948597b7b78 2025-11-03T16:58:02.2382017Z * [new tag] trunk/17bdb232e1b3df3435fc6f54d3d713404ba96726 -> trunk/17bdb232e1b3df3435fc6f54d3d713404ba96726 2025-11-03T16:58:02.2383358Z * [new tag] trunk/17d5aa47674c5195a3d4919169fb6a334598e2a6 -> trunk/17d5aa47674c5195a3d4919169fb6a334598e2a6 2025-11-03T16:58:02.2384635Z * [new tag] trunk/181ee3bd42447b71a1a8435bf16c0877c4bc3ae7 -> trunk/181ee3bd42447b71a1a8435bf16c0877c4bc3ae7 2025-11-03T16:58:02.2385844Z * [new tag] trunk/18f425962656f009e151944d12f76e2f856b85df -> trunk/18f425962656f009e151944d12f76e2f856b85df 2025-11-03T16:58:02.2387432Z * [new tag] trunk/1abfa5f70b9aa0dcbd83dc40b092dda1cdcc7805 -> trunk/1abfa5f70b9aa0dcbd83dc40b092dda1cdcc7805 2025-11-03T16:58:02.2388613Z * [new tag] trunk/1aef88c72d3aef629b20e97a188c9dc4bab46a1a -> trunk/1aef88c72d3aef629b20e97a188c9dc4bab46a1a 2025-11-03T16:58:02.2389936Z * [new tag] trunk/1b655a87ef137d2cc9603a982532c5e033432daa -> trunk/1b655a87ef137d2cc9603a982532c5e033432daa 2025-11-03T16:58:02.2391474Z * [new tag] trunk/1c4ced2eafa470832f2f4e24487785c5b7b188c9 -> trunk/1c4ced2eafa470832f2f4e24487785c5b7b188c9 2025-11-03T16:58:02.2392716Z * [new tag] trunk/1dd6b76914397d466a7d9bfff551b6d8299fdce4 -> trunk/1dd6b76914397d466a7d9bfff551b6d8299fdce4 2025-11-03T16:58:02.2394049Z * [new tag] trunk/1dec8a67a8dc88736e9c476a3eaee301684094cd -> trunk/1dec8a67a8dc88736e9c476a3eaee301684094cd 2025-11-03T16:58:02.2395460Z * [new tag] trunk/1e3600b5287346b29a835ad67f8b33945e0ec698 -> trunk/1e3600b5287346b29a835ad67f8b33945e0ec698 2025-11-03T16:58:02.2397232Z * [new tag] trunk/1e836bc769f9cfabc7659a172b8c2edee7c375d3 -> trunk/1e836bc769f9cfabc7659a172b8c2edee7c375d3 2025-11-03T16:58:02.2398541Z * [new tag] trunk/1fa520ea654f5fc0b3c65ce6e056dd73442dd65d -> trunk/1fa520ea654f5fc0b3c65ce6e056dd73442dd65d 2025-11-03T16:58:02.2399857Z * [new tag] trunk/1fdef664a5dffc3bd3d9eba9d78ed458a5f9dbcb -> trunk/1fdef664a5dffc3bd3d9eba9d78ed458a5f9dbcb 2025-11-03T16:58:02.2401610Z * [new tag] trunk/20be077085daf739af76325d04f649dcd908316c -> trunk/20be077085daf739af76325d04f649dcd908316c 2025-11-03T16:58:02.2402916Z * [new tag] trunk/21b48f8dfa7685699df4c97c0ba373d5364230d9 -> trunk/21b48f8dfa7685699df4c97c0ba373d5364230d9 2025-11-03T16:58:02.2404481Z * [new tag] trunk/23669d02a67023ef3192f7dca0ac0547436cae50 -> trunk/23669d02a67023ef3192f7dca0ac0547436cae50 2025-11-03T16:58:02.2405750Z * [new tag] trunk/236ce736a1a84352b21b564e3e5b0be833d7c10c -> trunk/236ce736a1a84352b21b564e3e5b0be833d7c10c 2025-11-03T16:58:02.2406984Z * [new tag] trunk/239e7b541a94d16a33a4a48d3edc9a561777d86a -> trunk/239e7b541a94d16a33a4a48d3edc9a561777d86a 2025-11-03T16:58:02.2408454Z * [new tag] trunk/23b57a445c6394d46a6b46e32b7787663f40d13b -> trunk/23b57a445c6394d46a6b46e32b7787663f40d13b 2025-11-03T16:58:02.2410227Z * [new tag] trunk/24b6eb7727f7eef078f9366302f2aa3e1fef1db4 -> trunk/24b6eb7727f7eef078f9366302f2aa3e1fef1db4 2025-11-03T16:58:02.2411492Z * [new tag] trunk/24e94e021a5a96bf16397ea66243070a7be4aaef -> trunk/24e94e021a5a96bf16397ea66243070a7be4aaef 2025-11-03T16:58:02.2412952Z * [new tag] trunk/26534e9809eb2f7cd804fde5152cdd13dda2293f -> trunk/26534e9809eb2f7cd804fde5152cdd13dda2293f 2025-11-03T16:58:02.2414126Z * [new tag] trunk/267d0197bfca0232488d51dd1ff735d619adc2cf -> trunk/267d0197bfca0232488d51dd1ff735d619adc2cf 2025-11-03T16:58:02.2415681Z * [new tag] trunk/2699f5410b6802277db724f145c281814d6c9ac4 -> trunk/2699f5410b6802277db724f145c281814d6c9ac4 2025-11-03T16:58:02.2416934Z * [new tag] trunk/2829d48bd113043a98e81b4ab7fa0f3ca9ae0072 -> trunk/2829d48bd113043a98e81b4ab7fa0f3ca9ae0072 2025-11-03T16:58:02.2418579Z * [new tag] trunk/284716a691580cf0508a7c5a4f9f7306a32092ad -> trunk/284716a691580cf0508a7c5a4f9f7306a32092ad 2025-11-03T16:58:02.2419792Z * [new tag] trunk/298666631baa6634666ac4dd4599f1d52bafa718 -> trunk/298666631baa6634666ac4dd4599f1d52bafa718 2025-11-03T16:58:02.2421244Z * [new tag] trunk/2a058bfecf4f4a47927b4eaae9324c4999a8a794 -> trunk/2a058bfecf4f4a47927b4eaae9324c4999a8a794 2025-11-03T16:58:02.2422971Z * [new tag] trunk/2a5f87decf34b3d0ea7670238e2fd4620ed19e9f -> trunk/2a5f87decf34b3d0ea7670238e2fd4620ed19e9f 2025-11-03T16:58:02.2424190Z * [new tag] trunk/2c9f877fa71c90a1313deeda4b4486d37e72a8a9 -> trunk/2c9f877fa71c90a1313deeda4b4486d37e72a8a9 2025-11-03T16:58:02.2425468Z * [new tag] trunk/2ce894bb1d3e9698306b87091c900a72acac75ea -> trunk/2ce894bb1d3e9698306b87091c900a72acac75ea 2025-11-03T16:58:02.2426751Z * [new tag] trunk/2dc56456cbb2645c37c3176b7d9a263933bdec27 -> trunk/2dc56456cbb2645c37c3176b7d9a263933bdec27 2025-11-03T16:58:02.2428287Z * [new tag] trunk/2de4cf2102fdc35da00cc8070fdbcaf57198a576 -> trunk/2de4cf2102fdc35da00cc8070fdbcaf57198a576 2025-11-03T16:58:02.2429511Z * [new tag] trunk/2df2c316e2929519a5bad029936e9e37ed637b12 -> trunk/2df2c316e2929519a5bad029936e9e37ed637b12 2025-11-03T16:58:02.2431188Z * [new tag] trunk/30157d30f0e0db6a4fa7a2a3ea75e515645e36de -> trunk/30157d30f0e0db6a4fa7a2a3ea75e515645e36de 2025-11-03T16:58:02.2432451Z * [new tag] trunk/3041ede082d3b0f33bec0cef5062eabc5173b446 -> trunk/3041ede082d3b0f33bec0cef5062eabc5173b446 2025-11-03T16:58:02.2433935Z * [new tag] trunk/311ea0dec0c50f395e6dac7b3875e81ee243fceb -> trunk/311ea0dec0c50f395e6dac7b3875e81ee243fceb 2025-11-03T16:58:02.2435271Z * [new tag] trunk/31e42eb732d32e71c9954d6b2104cf1d129370d5 -> trunk/31e42eb732d32e71c9954d6b2104cf1d129370d5 2025-11-03T16:58:02.2436856Z * [new tag] trunk/32066772b3dee643b1657b8957f32b5ac8b1390a -> trunk/32066772b3dee643b1657b8957f32b5ac8b1390a 2025-11-03T16:58:02.2438072Z * [new tag] trunk/32920926f07e573083ecf81a40c898f47f4df631 -> trunk/32920926f07e573083ecf81a40c898f47f4df631 2025-11-03T16:58:02.2439504Z * [new tag] trunk/32fe4f681e2bfb2fdecf20027e29e1aeb6ad5297 -> trunk/32fe4f681e2bfb2fdecf20027e29e1aeb6ad5297 2025-11-03T16:58:02.2440843Z * [new tag] trunk/335b5c7d4bf3295d517902370142f007ca024cd0 -> trunk/335b5c7d4bf3295d517902370142f007ca024cd0 2025-11-03T16:58:02.2442475Z * [new tag] trunk/34d6ef7022be252ce18a9673f75eed1732dff22a -> trunk/34d6ef7022be252ce18a9673f75eed1732dff22a 2025-11-03T16:58:02.2443721Z * [new tag] trunk/35f3572fa483a8edb101d5765564e1ae274f3d45 -> trunk/35f3572fa483a8edb101d5765564e1ae274f3d45 2025-11-03T16:58:02.2445293Z * [new tag] trunk/365ed62f61dde7fc14dca2199692555e5d1abbe1 -> trunk/365ed62f61dde7fc14dca2199692555e5d1abbe1 2025-11-03T16:58:02.2446530Z * [new tag] trunk/369f2d6951d4c12571eb03b95d74ef2f21296847 -> trunk/369f2d6951d4c12571eb03b95d74ef2f21296847 2025-11-03T16:58:02.2447873Z * [new tag] trunk/36a48e7e6d2d01cd47f002140589a1ac2580ef3f -> trunk/36a48e7e6d2d01cd47f002140589a1ac2580ef3f 2025-11-03T16:58:02.2449356Z * [new tag] trunk/3895ce093f3e71b768d032a28a949a9f390c0d00 -> trunk/3895ce093f3e71b768d032a28a949a9f390c0d00 2025-11-03T16:58:02.2450635Z * [new tag] trunk/392acee68a5ca90279109df023583edaef73d50f -> trunk/392acee68a5ca90279109df023583edaef73d50f 2025-11-03T16:58:02.2452255Z * [new tag] trunk/398775a43e9808205f75c81d36f5087117d3f3f4 -> trunk/398775a43e9808205f75c81d36f5087117d3f3f4 2025-11-03T16:58:02.2453443Z * [new tag] trunk/398fdd32bbf7b5f8a2b67e637c8a2e46572d9f28 -> trunk/398fdd32bbf7b5f8a2b67e637c8a2e46572d9f28 2025-11-03T16:58:02.2454764Z * [new tag] trunk/39e5cdddf7e57881c52473d1288a66f0222527e1 -> trunk/39e5cdddf7e57881c52473d1288a66f0222527e1 2025-11-03T16:58:02.2456384Z * [new tag] trunk/3b4315940d371385674f957f810cb0f1ca0060bd -> trunk/3b4315940d371385674f957f810cb0f1ca0060bd 2025-11-03T16:58:02.2457662Z * [new tag] trunk/3b5d38a3bc55c32232cee8cf87e492c965f276fd -> trunk/3b5d38a3bc55c32232cee8cf87e492c965f276fd 2025-11-03T16:58:02.2459258Z * [new tag] trunk/3ca216ae172e35adde34a319a1a01faaf218e7c5 -> trunk/3ca216ae172e35adde34a319a1a01faaf218e7c5 2025-11-03T16:58:02.2460569Z * [new tag] trunk/3cc5949dc2c6e588d42cff36121819a8539902a7 -> trunk/3cc5949dc2c6e588d42cff36121819a8539902a7 2025-11-03T16:58:02.2462144Z * [new tag] trunk/3d4a2d8a93ad57e9fd1c4e5a03dd431dd0123bb3 -> trunk/3d4a2d8a93ad57e9fd1c4e5a03dd431dd0123bb3 2025-11-03T16:58:02.2463458Z * [new tag] trunk/3d4ca228beb851e2f1cc24d8300804d5884973c1 -> trunk/3d4ca228beb851e2f1cc24d8300804d5884973c1 2025-11-03T16:58:02.2464798Z * [new tag] trunk/3dc92d69ed40fd952244e54bbda0240928756654 -> trunk/3dc92d69ed40fd952244e54bbda0240928756654 2025-11-03T16:58:02.2466261Z * [new tag] trunk/3eddf049221fc04c2ac9d4af53c00305484ef325 -> trunk/3eddf049221fc04c2ac9d4af53c00305484ef325 2025-11-03T16:58:02.2467530Z * [new tag] trunk/3f1824742cac2ffb9a3afd90953c492c6c7f2f50 -> trunk/3f1824742cac2ffb9a3afd90953c492c6c7f2f50 2025-11-03T16:58:02.2468751Z * [new tag] trunk/3f5401020b3fd66238a5eda31576b898d086e14e -> trunk/3f5401020b3fd66238a5eda31576b898d086e14e 2025-11-03T16:58:02.2470236Z * [new tag] trunk/3f6538febd727b782e6e13cfd026a309fb14351d -> trunk/3f6538febd727b782e6e13cfd026a309fb14351d 2025-11-03T16:58:02.2471484Z * [new tag] trunk/3f69b4d9b46b1de627896c457bce1937e53d82e5 -> trunk/3f69b4d9b46b1de627896c457bce1937e53d82e5 2025-11-03T16:58:02.2472929Z * [new tag] trunk/401c2f965726fda8f78b08f3745e11008b38e7bf -> trunk/401c2f965726fda8f78b08f3745e11008b38e7bf 2025-11-03T16:58:02.2474083Z * [new tag] trunk/4295a9a1587cc62a3ef4ad9b1e552c78471d5664 -> trunk/4295a9a1587cc62a3ef4ad9b1e552c78471d5664 2025-11-03T16:58:02.2475503Z * [new tag] trunk/4316df857c9e7f301142eb54d06a85a43f8d617b -> trunk/4316df857c9e7f301142eb54d06a85a43f8d617b 2025-11-03T16:58:02.2476982Z * [new tag] trunk/43c30f607eeca0d3e9a26911d9c2131fc250eadd -> trunk/43c30f607eeca0d3e9a26911d9c2131fc250eadd 2025-11-03T16:58:02.2478299Z * [new tag] trunk/45c3f02d69b0a8447381fbc3f3f8731ab0d48ed1 -> trunk/45c3f02d69b0a8447381fbc3f3f8731ab0d48ed1 2025-11-03T16:58:02.2479628Z * [new tag] trunk/467c21ad9ae4133c20a3c098a0355e9ac20d48aa -> trunk/467c21ad9ae4133c20a3c098a0355e9ac20d48aa 2025-11-03T16:58:02.2481236Z * [new tag] trunk/46d17e887150c5de8a74314b6d9fcba1feca259b -> trunk/46d17e887150c5de8a74314b6d9fcba1feca259b 2025-11-03T16:58:02.2482485Z * [new tag] trunk/476b149a00cf52ceba2b2d8b6e14989d2bf540e2 -> trunk/476b149a00cf52ceba2b2d8b6e14989d2bf540e2 2025-11-03T16:58:02.2483695Z * [new tag] trunk/47ec1e9990b89bf9705dc463cad587bd357bc848 -> trunk/47ec1e9990b89bf9705dc463cad587bd357bc848 2025-11-03T16:58:02.2485052Z * [new tag] trunk/47f00243106918eec3ef9119272977fdf2864ad8 -> trunk/47f00243106918eec3ef9119272977fdf2864ad8 2025-11-03T16:58:02.2486925Z * [new tag] trunk/47f50cfd456313d8b46fcc7a1f6de477aa0a5aee -> trunk/47f50cfd456313d8b46fcc7a1f6de477aa0a5aee 2025-11-03T16:58:02.2488011Z * [new tag] trunk/483845a9c42267a6a139b7382e5ee340502376f8 -> trunk/483845a9c42267a6a139b7382e5ee340502376f8 2025-11-03T16:58:02.2489483Z * [new tag] trunk/48c3b71ecc245cb422ec9ab4c7a2078f7ef0b3df -> trunk/48c3b71ecc245cb422ec9ab4c7a2078f7ef0b3df 2025-11-03T16:58:02.2491105Z * [new tag] trunk/48e672d14994c912caccba0c7f2f5df54c0d86b3 -> trunk/48e672d14994c912caccba0c7f2f5df54c0d86b3 2025-11-03T16:58:02.2492391Z * [new tag] trunk/4a7bc1d522bed4f3f792b4641372ec3d06b79b5d -> trunk/4a7bc1d522bed4f3f792b4641372ec3d06b79b5d 2025-11-03T16:58:02.2493983Z * [new tag] trunk/4a7fefd7c783913b811f49e5d6242086758dca6c -> trunk/4a7fefd7c783913b811f49e5d6242086758dca6c 2025-11-03T16:58:02.2495256Z * [new tag] trunk/4a94591321a62b31658f9259680de87e4be9e50d -> trunk/4a94591321a62b31658f9259680de87e4be9e50d 2025-11-03T16:58:02.2496848Z * [new tag] trunk/4acc66f1192ab7743abcc50383aefc5447447f9d -> trunk/4acc66f1192ab7743abcc50383aefc5447447f9d 2025-11-03T16:58:02.2498088Z * [new tag] trunk/4c38887346b6a6185bb3e170621dc13d5f34e9ce -> trunk/4c38887346b6a6185bb3e170621dc13d5f34e9ce 2025-11-03T16:58:02.2499704Z * [new tag] trunk/4cc64d62346edb2e13320f20740f41ba41f64c70 -> trunk/4cc64d62346edb2e13320f20740f41ba41f64c70 2025-11-03T16:58:02.2500865Z * [new tag] trunk/4e6afa8c07b73e0b339e3baacbc73102748e61dd -> trunk/4e6afa8c07b73e0b339e3baacbc73102748e61dd 2025-11-03T16:58:02.2502667Z * [new tag] trunk/4e7232c5daf753e04e8f4189229e3c33888a33e5 -> trunk/4e7232c5daf753e04e8f4189229e3c33888a33e5 2025-11-03T16:58:02.2503861Z * [new tag] trunk/4e8ba37ce3320cac033019131b817b4aca29699b -> trunk/4e8ba37ce3320cac033019131b817b4aca29699b 2025-11-03T16:58:02.2505374Z * [new tag] trunk/4fada51ada8b3190971f2f47d97346057320866a -> trunk/4fada51ada8b3190971f2f47d97346057320866a 2025-11-03T16:58:02.2506614Z * [new tag] trunk/5016e7b2eb73589e35c7a8f0f0ea3bfba221f17d -> trunk/5016e7b2eb73589e35c7a8f0f0ea3bfba221f17d 2025-11-03T16:58:02.2508151Z * [new tag] trunk/51667435f50c025ca3655ff0eeb917b4ef0ffb78 -> trunk/51667435f50c025ca3655ff0eeb917b4ef0ffb78 2025-11-03T16:58:02.2509331Z * [new tag] trunk/52db60170dd192afa48233f03574a8d8bcd39d35 -> trunk/52db60170dd192afa48233f03574a8d8bcd39d35 2025-11-03T16:58:02.2510946Z * [new tag] trunk/544b443ea1d1a9b19e65f981168a01cb87a2d333 -> trunk/544b443ea1d1a9b19e65f981168a01cb87a2d333 2025-11-03T16:58:02.2512151Z * [new tag] trunk/551921d484d209e9a95d77262b1472bc1d18422d -> trunk/551921d484d209e9a95d77262b1472bc1d18422d 2025-11-03T16:58:02.2513544Z * [new tag] trunk/56838bad5f197d32c0ee77ce2eaa7a0cb462af68 -> trunk/56838bad5f197d32c0ee77ce2eaa7a0cb462af68 2025-11-03T16:58:02.2514887Z * [new tag] trunk/56a809aa0765ec2edea57ab37f92b1e78a214dcc -> trunk/56a809aa0765ec2edea57ab37f92b1e78a214dcc 2025-11-03T16:58:02.2516326Z * [new tag] trunk/56afad4eb300be2008d8b24581bcddcd62021036 -> trunk/56afad4eb300be2008d8b24581bcddcd62021036 2025-11-03T16:58:02.2517645Z * [new tag] trunk/572cc12b42f6edf76affd45ea9a23f4c8db5c7e1 -> trunk/572cc12b42f6edf76affd45ea9a23f4c8db5c7e1 2025-11-03T16:58:02.2519115Z * [new tag] trunk/5849eea129b47556630aac4276529afe287b91e1 -> trunk/5849eea129b47556630aac4276529afe287b91e1 2025-11-03T16:58:02.2521178Z * [new tag] trunk/59ddfb69a731970bf5762fbea43ff2987425d3ea -> trunk/59ddfb69a731970bf5762fbea43ff2987425d3ea 2025-11-03T16:58:02.2522761Z * [new tag] trunk/5a3930abbc19eac9a179455df82e206e69765ed2 -> trunk/5a3930abbc19eac9a179455df82e206e69765ed2 2025-11-03T16:58:02.2524005Z * [new tag] trunk/5bcfdae71da2f405a89bace6b09d616107cc3965 -> trunk/5bcfdae71da2f405a89bace6b09d616107cc3965 2025-11-03T16:58:02.2525533Z * [new tag] trunk/5cbdade914d6bc9f67827ef917974980a74e9372 -> trunk/5cbdade914d6bc9f67827ef917974980a74e9372 2025-11-03T16:58:02.2526779Z * [new tag] trunk/5cdbcb52334222d1a2667dabd6da993d1e9ddc93 -> trunk/5cdbcb52334222d1a2667dabd6da993d1e9ddc93 2025-11-03T16:58:02.2528514Z * [new tag] trunk/5d0b3e28dc18fac3c09edb24078998371f878f82 -> trunk/5d0b3e28dc18fac3c09edb24078998371f878f82 2025-11-03T16:58:02.2529810Z * [new tag] trunk/5d288bc3f73873887f681e15af83c5525e6a60bd -> trunk/5d288bc3f73873887f681e15af83c5525e6a60bd 2025-11-03T16:58:02.2531101Z * [new tag] trunk/5d6230779d920e9c449c22d22fd0f69fe7b73632 -> trunk/5d6230779d920e9c449c22d22fd0f69fe7b73632 2025-11-03T16:58:02.2532721Z * [new tag] trunk/5e05a0ae99c320bc570a3a42c5fcd6bbf2b2e9b4 -> trunk/5e05a0ae99c320bc570a3a42c5fcd6bbf2b2e9b4 2025-11-03T16:58:02.2533873Z * [new tag] trunk/5e7272b60a5d7203fd1d048ecac6f202526cac63 -> trunk/5e7272b60a5d7203fd1d048ecac6f202526cac63 2025-11-03T16:58:02.2535451Z * [new tag] trunk/5e769ff86780a7ffd561615dbf4b0defe80cfbb9 -> trunk/5e769ff86780a7ffd561615dbf4b0defe80cfbb9 2025-11-03T16:58:02.2536770Z * [new tag] trunk/5ebf74a655b3335fb079ee03f41228a77bcc630e -> trunk/5ebf74a655b3335fb079ee03f41228a77bcc630e 2025-11-03T16:58:02.2538331Z * [new tag] trunk/5fc2c7a2a11444c31338da9134452810f3ca8acd -> trunk/5fc2c7a2a11444c31338da9134452810f3ca8acd 2025-11-03T16:58:02.2539706Z * [new tag] trunk/5fd1d41e62ce4436c2bb08c7f4702df0021e2c62 -> trunk/5fd1d41e62ce4436c2bb08c7f4702df0021e2c62 2025-11-03T16:58:02.2541380Z * [new tag] trunk/60333de85dec845d6a0e651764c16a730a27828d -> trunk/60333de85dec845d6a0e651764c16a730a27828d 2025-11-03T16:58:02.2542688Z * [new tag] trunk/6096c0fc746dc99189510a1845fd45c667178c03 -> trunk/6096c0fc746dc99189510a1845fd45c667178c03 2025-11-03T16:58:02.2544307Z * [new tag] trunk/60bcb4ee8844e7bac7278ff662c7d5735a326758 -> trunk/60bcb4ee8844e7bac7278ff662c7d5735a326758 2025-11-03T16:58:02.2545634Z * [new tag] trunk/610c09f8f47585e811945e48abc895f88e09bda3 -> trunk/610c09f8f47585e811945e48abc895f88e09bda3 2025-11-03T16:58:02.2547102Z * [new tag] trunk/61bad3c1ea718311d7c256b05614d207d73276e8 -> trunk/61bad3c1ea718311d7c256b05614d207d73276e8 2025-11-03T16:58:02.2548430Z * [new tag] trunk/61bcc8d75ad065a40b2d7953046616d5979233b9 -> trunk/61bcc8d75ad065a40b2d7953046616d5979233b9 2025-11-03T16:58:02.2550250Z * [new tag] trunk/6268883f9c081b88b611ef01be3821503ad180ca -> trunk/6268883f9c081b88b611ef01be3821503ad180ca 2025-11-03T16:58:02.2551573Z * [new tag] trunk/629293f568229adaf60cdaf57fe69306d7782c2d -> trunk/629293f568229adaf60cdaf57fe69306d7782c2d 2025-11-03T16:58:02.2553146Z * [new tag] trunk/639a0b12395e72a8a220dc8443317f4ab3b3f1a0 -> trunk/639a0b12395e72a8a220dc8443317f4ab3b3f1a0 2025-11-03T16:58:02.2554416Z * [new tag] trunk/6530bc70fbc7d600ab0e600c02ecd9b341ce8237 -> trunk/6530bc70fbc7d600ab0e600c02ecd9b341ce8237 2025-11-03T16:58:02.2555904Z * [new tag] trunk/657f8c3e21bd8901dd8ce79ca9a54a45b27f604f -> trunk/657f8c3e21bd8901dd8ce79ca9a54a45b27f604f 2025-11-03T16:58:02.2557480Z * [new tag] trunk/65868156c613521e9635eb3f8eb3f404826ac1f9 -> trunk/65868156c613521e9635eb3f8eb3f404826ac1f9 2025-11-03T16:58:02.2558539Z * [new tag] trunk/687c15c0b3f01118536413d21efcd052838fa10d -> trunk/687c15c0b3f01118536413d21efcd052838fa10d 2025-11-03T16:58:02.2560145Z * [new tag] trunk/68b3984b77a19ed2b2afcbcae201e8b06e20f8bc -> trunk/68b3984b77a19ed2b2afcbcae201e8b06e20f8bc 2025-11-03T16:58:02.2561372Z * [new tag] trunk/694d205143b03cb26626cb86ecb399df6d5e3dd8 -> trunk/694d205143b03cb26626cb86ecb399df6d5e3dd8 2025-11-03T16:58:02.2563025Z * [new tag] trunk/694db5f54927697c9e914d35029f7e5bd9b85b96 -> trunk/694db5f54927697c9e914d35029f7e5bd9b85b96 2025-11-03T16:58:02.2564348Z * [new tag] trunk/695cb0d342f67db563fb57ce1a5f3261b073b790 -> trunk/695cb0d342f67db563fb57ce1a5f3261b073b790 2025-11-03T16:58:02.2565901Z * [new tag] trunk/69be99ee51bd0439fead4c83d1ed3eea9e1e073e -> trunk/69be99ee51bd0439fead4c83d1ed3eea9e1e073e 2025-11-03T16:58:02.2567138Z * [new tag] trunk/69fb3ebb5dbec0bd4cfd73cc9a6c870b7f604840 -> trunk/69fb3ebb5dbec0bd4cfd73cc9a6c870b7f604840 2025-11-03T16:58:02.2568624Z * [new tag] trunk/6a5a436624d08feff1de191e61eb4d4773f73bdb -> trunk/6a5a436624d08feff1de191e61eb4d4773f73bdb 2025-11-03T16:58:02.2570217Z * [new tag] trunk/6c7cad697287dff7e47545ae4b1df1e407185868 -> trunk/6c7cad697287dff7e47545ae4b1df1e407185868 2025-11-03T16:58:02.2571446Z * [new tag] trunk/6d5e651a5024acdbaf9633d9d1d9dc33c6a283c5 -> trunk/6d5e651a5024acdbaf9633d9d1d9dc33c6a283c5 2025-11-03T16:58:02.2572753Z * [new tag] trunk/6ecd6b23b6d2ab17da269ceebe9f944153f67083 -> trunk/6ecd6b23b6d2ab17da269ceebe9f944153f67083 2025-11-03T16:58:02.2574324Z * [new tag] trunk/7045aab143b6d89c2e7de11bc8cbf919d8fd78d0 -> trunk/7045aab143b6d89c2e7de11bc8cbf919d8fd78d0 2025-11-03T16:58:02.2575631Z * [new tag] trunk/70aeb491987401a380914a72c1b9d19e4c882429 -> trunk/70aeb491987401a380914a72c1b9d19e4c882429 2025-11-03T16:58:02.2576906Z * [new tag] trunk/7379972cc0e8a4b4d88b4bea5c8be0aeffdb076a -> trunk/7379972cc0e8a4b4d88b4bea5c8be0aeffdb076a 2025-11-03T16:58:02.2578951Z * [new tag] trunk/73da7a40b6dd0509b179f0ca92d9fa79973ee306 -> trunk/73da7a40b6dd0509b179f0ca92d9fa79973ee306 2025-11-03T16:58:02.2580125Z * [new tag] trunk/74336f8c77f1a906314c274e43239306c8077e90 -> trunk/74336f8c77f1a906314c274e43239306c8077e90 2025-11-03T16:58:02.2581689Z * [new tag] trunk/748162223768fb9f7daaca1f7f93d0bc37be4de9 -> trunk/748162223768fb9f7daaca1f7f93d0bc37be4de9 2025-11-03T16:58:02.2583048Z * [new tag] trunk/753d9bd8063356c0e31517ef9aabd814efd754c5 -> trunk/753d9bd8063356c0e31517ef9aabd814efd754c5 2025-11-03T16:58:02.2584741Z * [new tag] trunk/7563f61cc8a40a5ba21a498a2d98895b4eec3f39 -> trunk/7563f61cc8a40a5ba21a498a2d98895b4eec3f39 2025-11-03T16:58:02.2586083Z * [new tag] trunk/75f798e05b35cbdf1814673e745b9b7817bb29cd -> trunk/75f798e05b35cbdf1814673e745b9b7817bb29cd 2025-11-03T16:58:02.2587777Z * [new tag] trunk/764c54ecaeb8b0da486b009f4d4ef1fa280e2517 -> trunk/764c54ecaeb8b0da486b009f4d4ef1fa280e2517 2025-11-03T16:58:02.2588923Z * [new tag] trunk/76780b1a3d208b142f067bdd655e7d4c5615efe1 -> trunk/76780b1a3d208b142f067bdd655e7d4c5615efe1 2025-11-03T16:58:02.2590289Z * [new tag] trunk/7692fa09cdd47fdbec66efca9edf42a318e6c6e4 -> trunk/7692fa09cdd47fdbec66efca9edf42a318e6c6e4 2025-11-03T16:58:02.2591806Z * [new tag] trunk/76b2c37045e52540ec51e967aa7b6436a6b9b174 -> trunk/76b2c37045e52540ec51e967aa7b6436a6b9b174 2025-11-03T16:58:02.2593089Z * [new tag] trunk/76bb27e248750fc558ec4ea9050e09a6f171e9bc -> trunk/76bb27e248750fc558ec4ea9050e09a6f171e9bc 2025-11-03T16:58:02.2594666Z * [new tag] trunk/774abb018eb1d44c07fda9fa78549755ac7d4367 -> trunk/774abb018eb1d44c07fda9fa78549755ac7d4367 2025-11-03T16:58:02.2595916Z * [new tag] trunk/791ca80d3a2ca72a77b5dedc75bbcbb35b859cfc -> trunk/791ca80d3a2ca72a77b5dedc75bbcbb35b859cfc 2025-11-03T16:58:02.2597527Z * [new tag] trunk/797cd80b2670a51601f997f8c67387bd30440a36 -> trunk/797cd80b2670a51601f997f8c67387bd30440a36 2025-11-03T16:58:02.2598813Z * [new tag] trunk/79aa88cc5dcc55597013cbe36f0e00557b185803 -> trunk/79aa88cc5dcc55597013cbe36f0e00557b185803 2025-11-03T16:58:02.2600064Z * [new tag] trunk/79aee77381b21d41c77148e5ff84c4b351aaf144 -> trunk/79aee77381b21d41c77148e5ff84c4b351aaf144 2025-11-03T16:58:02.2601657Z * [new tag] trunk/7a0cd8ed092a62123fcc1f4c40de2669f46fe2b1 -> trunk/7a0cd8ed092a62123fcc1f4c40de2669f46fe2b1 2025-11-03T16:58:02.2602917Z * [new tag] trunk/7ae8aaf4c02e7cc14b2e3cce0c88a7333dff439c -> trunk/7ae8aaf4c02e7cc14b2e3cce0c88a7333dff439c 2025-11-03T16:58:02.2604401Z * [new tag] trunk/7c203b8420d26bd707c6052f93a4ba77dca2b2c0 -> trunk/7c203b8420d26bd707c6052f93a4ba77dca2b2c0 2025-11-03T16:58:02.2605718Z * [new tag] trunk/7ce723d21cc2a1a874f988f3459f0da834c75daa -> trunk/7ce723d21cc2a1a874f988f3459f0da834c75daa 2025-11-03T16:58:02.2607185Z * [new tag] trunk/7d16fcf2df9b240895056107a015735774cbe4b5 -> trunk/7d16fcf2df9b240895056107a015735774cbe4b5 2025-11-03T16:58:02.2608640Z * [new tag] trunk/7d39401fa07eb2f7782d8f234aa19f8b2abff481 -> trunk/7d39401fa07eb2f7782d8f234aa19f8b2abff481 2025-11-03T16:58:02.2610309Z * [new tag] trunk/7d67a41db4801632d5862e99c79b766284dfd307 -> trunk/7d67a41db4801632d5862e99c79b766284dfd307 2025-11-03T16:58:02.2611612Z * [new tag] trunk/7e3b9d105e3359e31fb252becb01d789a72dc8d9 -> trunk/7e3b9d105e3359e31fb252becb01d789a72dc8d9 2025-11-03T16:58:02.2613085Z * [new tag] trunk/7f7a28046b5927ddbd8ef95d00aa1ca76f773bb1 -> trunk/7f7a28046b5927ddbd8ef95d00aa1ca76f773bb1 2025-11-03T16:58:02.2614551Z * [new tag] trunk/80ba6e458ff2bd178da9ed62e27a61fe46e4252a -> trunk/80ba6e458ff2bd178da9ed62e27a61fe46e4252a 2025-11-03T16:58:02.2616265Z * [new tag] trunk/8101fd46d4619d0f9ef69e5829d1ee6f77289b72 -> trunk/8101fd46d4619d0f9ef69e5829d1ee6f77289b72 2025-11-03T16:58:02.2617266Z * [new tag] trunk/8110ce02a2830f562ed60607e30e2f29aed119dd -> trunk/8110ce02a2830f562ed60607e30e2f29aed119dd 2025-11-03T16:58:02.2618898Z * [new tag] trunk/81fa4a204cf268bc55e96137fa803ae4938472cd -> trunk/81fa4a204cf268bc55e96137fa803ae4938472cd 2025-11-03T16:58:02.2620148Z * [new tag] trunk/8209a0506b399b40a48716126ee22737aa4607c5 -> trunk/8209a0506b399b40a48716126ee22737aa4607c5 2025-11-03T16:58:02.2622000Z * [new tag] trunk/8221ee6db9fd319c249087ad9bb9abaa7d56e113 -> trunk/8221ee6db9fd319c249087ad9bb9abaa7d56e113 2025-11-03T16:58:02.2623356Z * [new tag] trunk/82d86bacf3c34760d6ab8a8a1eafa22d0ea259d9 -> trunk/82d86bacf3c34760d6ab8a8a1eafa22d0ea259d9 2025-11-03T16:58:02.2624928Z * [new tag] trunk/82fafb3304f8baffdb2d54578926d8bb4aba621c -> trunk/82fafb3304f8baffdb2d54578926d8bb4aba621c 2025-11-03T16:58:02.2626168Z * [new tag] trunk/82ff07c7884d478ddd5d638bebbb938e55c9bebf -> trunk/82ff07c7884d478ddd5d638bebbb938e55c9bebf 2025-11-03T16:58:02.2627619Z * [new tag] trunk/83cc38d9c1b4371d4395bda8b6861b98986f527b -> trunk/83cc38d9c1b4371d4395bda8b6861b98986f527b 2025-11-03T16:58:02.2628894Z * [new tag] trunk/840d63c12d255dac1ae3c5e442c6ea6eb87a7256 -> trunk/840d63c12d255dac1ae3c5e442c6ea6eb87a7256 2025-11-03T16:58:02.2630506Z * [new tag] trunk/8417981c96b7577f0228226ccf76d23842173144 -> trunk/8417981c96b7577f0228226ccf76d23842173144 2025-11-03T16:58:02.2631951Z * [new tag] trunk/845da9c8172b0e6e8b1e326359379dc6e9ddbddf -> trunk/845da9c8172b0e6e8b1e326359379dc6e9ddbddf 2025-11-03T16:58:02.2633543Z * [new tag] trunk/84776e13744db6d59b41a063bb8714e2bffe7a06 -> trunk/84776e13744db6d59b41a063bb8714e2bffe7a06 2025-11-03T16:58:02.2635162Z * [new tag] trunk/84a2715d341f068a26a281a252c3907bbe827d9b -> trunk/84a2715d341f068a26a281a252c3907bbe827d9b 2025-11-03T16:58:02.2636489Z * [new tag] trunk/84fe84850383fc11e7150681742bf53f11cd425e -> trunk/84fe84850383fc11e7150681742bf53f11cd425e 2025-11-03T16:58:02.2638025Z * [new tag] trunk/856a7a529872aa2d222e2491708cb458c682bb93 -> trunk/856a7a529872aa2d222e2491708cb458c682bb93 2025-11-03T16:58:02.2639291Z * [new tag] trunk/85a7c745aa5a2428a7cc640ef4b304296fe07c42 -> trunk/85a7c745aa5a2428a7cc640ef4b304296fe07c42 2025-11-03T16:58:02.2640959Z * [new tag] trunk/85b035ca9c230199d91ff86d362def01ad36dc67 -> trunk/85b035ca9c230199d91ff86d362def01ad36dc67 2025-11-03T16:58:02.2642314Z * [new tag] trunk/85b85f6c2c7e6ca98abf9d9a38410d4b5703d64a -> trunk/85b85f6c2c7e6ca98abf9d9a38410d4b5703d64a 2025-11-03T16:58:02.2643909Z * [new tag] trunk/877f126e353f537417c53bff073fc56d3faecd60 -> trunk/877f126e353f537417c53bff073fc56d3faecd60 2025-11-03T16:58:02.2645406Z * [new tag] trunk/8887a33ede05c3e77286447a8ebd82b458878943 -> trunk/8887a33ede05c3e77286447a8ebd82b458878943 2025-11-03T16:58:02.2646598Z * [new tag] trunk/895795f07cc0348ebe0b7e369d555b545d76eb9f -> trunk/895795f07cc0348ebe0b7e369d555b545d76eb9f 2025-11-03T16:58:02.2648143Z * [new tag] trunk/8aa087a29dcb5eba4cffd6a452e7f327abfd14b2 -> trunk/8aa087a29dcb5eba4cffd6a452e7f327abfd14b2 2025-11-03T16:58:02.2649308Z * [new tag] trunk/8af9ed08243222853cf8ebfc9177c5b1ea364d00 -> trunk/8af9ed08243222853cf8ebfc9177c5b1ea364d00 2025-11-03T16:58:02.2650867Z * [new tag] trunk/8b188647cfdc1355070ccd5aaa18a8060d4f67bf -> trunk/8b188647cfdc1355070ccd5aaa18a8060d4f67bf 2025-11-03T16:58:02.2652245Z * [new tag] trunk/8d4e48831e5abb6cbca8d97a2df6c47b770c303b -> trunk/8d4e48831e5abb6cbca8d97a2df6c47b770c303b 2025-11-03T16:58:02.2653788Z * [new tag] trunk/8d599045cf4102e451a9e8a9ff215d053ebbe0e8 -> trunk/8d599045cf4102e451a9e8a9ff215d053ebbe0e8 2025-11-03T16:58:02.2655220Z * [new tag] trunk/8e1e4ee8e0c43202085b47c9d2b05b2c0d13b368 -> trunk/8e1e4ee8e0c43202085b47c9d2b05b2c0d13b368 2025-11-03T16:58:02.2656553Z * [new tag] trunk/8f40a0c634a65246889b478b7f14da7a448174a9 -> trunk/8f40a0c634a65246889b478b7f14da7a448174a9 2025-11-03T16:58:02.2658124Z * [new tag] trunk/8f51556daa347041fc7ac06a832d2fdc62ca594d -> trunk/8f51556daa347041fc7ac06a832d2fdc62ca594d 2025-11-03T16:58:02.2659412Z * [new tag] trunk/904abfc2ca285910b7b1458779bd9c63619f2f7d -> trunk/904abfc2ca285910b7b1458779bd9c63619f2f7d 2025-11-03T16:58:02.2660835Z * [new tag] trunk/90519402c2006237f891289a0afdec804515aa73 -> trunk/90519402c2006237f891289a0afdec804515aa73 2025-11-03T16:58:02.2662487Z * [new tag] trunk/90b30ebf7ecd7b0da9deeb35b165de9e28e90aff -> trunk/90b30ebf7ecd7b0da9deeb35b165de9e28e90aff 2025-11-03T16:58:02.2663838Z * [new tag] trunk/90d7be35e980a39cb4044f8fefea33a6167e5733 -> trunk/90d7be35e980a39cb4044f8fefea33a6167e5733 2025-11-03T16:58:02.2665248Z * [new tag] trunk/9139368b64ebcc32c8c1190fd114c7f4733e7c6d -> trunk/9139368b64ebcc32c8c1190fd114c7f4733e7c6d 2025-11-03T16:58:02.2666813Z * [new tag] trunk/92381a5aa7c27b740d375dad2ff9ca7be3c1bc24 -> trunk/92381a5aa7c27b740d375dad2ff9ca7be3c1bc24 2025-11-03T16:58:02.2668135Z * [new tag] trunk/924482a6f63305b0f00760170ad513305700b79c -> trunk/924482a6f63305b0f00760170ad513305700b79c 2025-11-03T16:58:02.2669709Z * [new tag] trunk/9261a1fb128412201ef009d30844a2417364d73b -> trunk/9261a1fb128412201ef009d30844a2417364d73b 2025-11-03T16:58:02.2671549Z * [new tag] trunk/93a70c717a5636505da0b4ccfbca0e111c88c6a8 -> trunk/93a70c717a5636505da0b4ccfbca0e111c88c6a8 2025-11-03T16:58:02.2672832Z * [new tag] trunk/94eaeb9cb812744251cef11d147e93b00a633823 -> trunk/94eaeb9cb812744251cef11d147e93b00a633823 2025-11-03T16:58:02.2674415Z * [new tag] trunk/94f2657c4b534136aa8958bc35d44ceac5ccd60c -> trunk/94f2657c4b534136aa8958bc35d44ceac5ccd60c 2025-11-03T16:58:02.2675645Z * [new tag] trunk/9501405de686068c5d7d0e4eab21e3ef0b95cc73 -> trunk/9501405de686068c5d7d0e4eab21e3ef0b95cc73 2025-11-03T16:58:02.2677260Z * [new tag] trunk/95b55347730b071d0b79df5ddd477db9c7f91fe4 -> trunk/95b55347730b071d0b79df5ddd477db9c7f91fe4 2025-11-03T16:58:02.2678509Z * [new tag] trunk/96b61844a722d2ae8b1a2dc283a6c0060b40782e -> trunk/96b61844a722d2ae8b1a2dc283a6c0060b40782e 2025-11-03T16:58:02.2680171Z * [new tag] trunk/972030fe2ef0a7500afb7acb86a5f971c30536b9 -> trunk/972030fe2ef0a7500afb7acb86a5f971c30536b9 2025-11-03T16:58:02.2681761Z * [new tag] trunk/984e64b2cddecfe43501a7e01a1cce0c25473f54 -> trunk/984e64b2cddecfe43501a7e01a1cce0c25473f54 2025-11-03T16:58:02.2682995Z * [new tag] trunk/98d640bb113a0b6a39c590bbe36873faeb71ceb0 -> trunk/98d640bb113a0b6a39c590bbe36873faeb71ceb0 2025-11-03T16:58:02.2684519Z * [new tag] trunk/9901d4441862cc0be881f913e1a7f932f7cb0177 -> trunk/9901d4441862cc0be881f913e1a7f932f7cb0177 2025-11-03T16:58:02.2686018Z * [new tag] trunk/9940e894ea9c4374691ec49eba96db0a3159a8ae -> trunk/9940e894ea9c4374691ec49eba96db0a3159a8ae 2025-11-03T16:58:02.2687325Z * [new tag] trunk/9970fb97fff5fb5ab0f0f3acfa73558cc5c1618a -> trunk/9970fb97fff5fb5ab0f0f3acfa73558cc5c1618a 2025-11-03T16:58:02.2688940Z * [new tag] trunk/99b05d1b786dc1f2448c23617bbb80d441a5d64a -> trunk/99b05d1b786dc1f2448c23617bbb80d441a5d64a 2025-11-03T16:58:02.2690493Z * [new tag] trunk/99e07c39ec3e9cf4573da6e6737c3d0362bc6b13 -> trunk/99e07c39ec3e9cf4573da6e6737c3d0362bc6b13 2025-11-03T16:58:02.2691651Z * [new tag] trunk/9a91486e453f300804c7031d0e904b0e2af7d802 -> trunk/9a91486e453f300804c7031d0e904b0e2af7d802 2025-11-03T16:58:02.2693372Z * [new tag] trunk/9c22bbb2dce31b854e3387db77eaff501434f352 -> trunk/9c22bbb2dce31b854e3387db77eaff501434f352 2025-11-03T16:58:02.2694555Z * [new tag] trunk/9d6597b1e9f95f4477e88c5bee473c2850c2e63f -> trunk/9d6597b1e9f95f4477e88c5bee473c2850c2e63f 2025-11-03T16:58:02.2695879Z * [new tag] trunk/9ee1afbf66a6533c712956a3ba5e8f60c5875a74 -> trunk/9ee1afbf66a6533c712956a3ba5e8f60c5875a74 2025-11-03T16:58:02.2697382Z * [new tag] trunk/9f9dbe0a9aa5e5c985f36ad6cc736b3fb3506056 -> trunk/9f9dbe0a9aa5e5c985f36ad6cc736b3fb3506056 2025-11-03T16:58:02.2698749Z * [new tag] trunk/a04edcb27acc6e1dd0fc27233738e6dec718b109 -> trunk/a04edcb27acc6e1dd0fc27233738e6dec718b109 2025-11-03T16:58:02.2700053Z * [new tag] trunk/a076b4d7ac09b7be59915a088b348bcbd952c8bd -> trunk/a076b4d7ac09b7be59915a088b348bcbd952c8bd 2025-11-03T16:58:02.2701701Z * [new tag] trunk/a186aa8d6cc197054deba309605f4b75ca59899c -> trunk/a186aa8d6cc197054deba309605f4b75ca59899c 2025-11-03T16:58:02.2702969Z * [new tag] trunk/a19e92d43355489c8d33325321e688a6e6182267 -> trunk/a19e92d43355489c8d33325321e688a6e6182267 2025-11-03T16:58:02.2704371Z * [new tag] trunk/a1eb6b55387b6ae21cf9bcad76a471a2820fa1b0 -> trunk/a1eb6b55387b6ae21cf9bcad76a471a2820fa1b0 2025-11-03T16:58:02.2706015Z * [new tag] trunk/a25818cf7ee2c0ed5c862dff214dc46a30211671 -> trunk/a25818cf7ee2c0ed5c862dff214dc46a30211671 2025-11-03T16:58:02.2707350Z * [new tag] trunk/a2da69385abb4800732e698afacdddd78e9993aa -> trunk/a2da69385abb4800732e698afacdddd78e9993aa 2025-11-03T16:58:02.2708813Z * [new tag] trunk/a3fe1825aa4d95707875ffa8f15fd00c3b94977d -> trunk/a3fe1825aa4d95707875ffa8f15fd00c3b94977d 2025-11-03T16:58:02.2710150Z * [new tag] trunk/a4077b568f891f8cad948cc70cbb14043791d762 -> trunk/a4077b568f891f8cad948cc70cbb14043791d762 2025-11-03T16:58:02.2711435Z * [new tag] trunk/a4a0378e6bb8afd774e4e253304cf3bf9d7e188a -> trunk/a4a0378e6bb8afd774e4e253304cf3bf9d7e188a 2025-11-03T16:58:02.2712953Z * [new tag] trunk/a51f8772873bc91de8aad1ce865fe8d98a58aec7 -> trunk/a51f8772873bc91de8aad1ce865fe8d98a58aec7 2025-11-03T16:58:02.2714108Z * [new tag] trunk/a5335263d32b5be2b2647661334d81225c3cc3fc -> trunk/a5335263d32b5be2b2647661334d81225c3cc3fc 2025-11-03T16:58:02.2715685Z * [new tag] trunk/a553ea9ea4e71b66a58a1c3c37f18d67ffa4ecd3 -> trunk/a553ea9ea4e71b66a58a1c3c37f18d67ffa4ecd3 2025-11-03T16:58:02.2716836Z * [new tag] trunk/a5c3c08d1099f96548398a637463efb0cbe115ae -> trunk/a5c3c08d1099f96548398a637463efb0cbe115ae 2025-11-03T16:58:02.2718467Z * [new tag] trunk/a5f00077fc9623c762c755c926b7eee60eb0ec3f -> trunk/a5f00077fc9623c762c755c926b7eee60eb0ec3f 2025-11-03T16:58:02.2719804Z * [new tag] trunk/a663eb9c80e92190832b3dd616f138f938f51525 -> trunk/a663eb9c80e92190832b3dd616f138f938f51525 2025-11-03T16:58:02.2721591Z * [new tag] trunk/a6b1ef17173f56ba93ac97ff4384fa4060b5e41e -> trunk/a6b1ef17173f56ba93ac97ff4384fa4060b5e41e 2025-11-03T16:58:02.2724934Z * [new tag] trunk/a76b59cc45a17b728d9da6f937c8a9ab537c657f -> trunk/a76b59cc45a17b728d9da6f937c8a9ab537c657f 2025-11-03T16:58:02.2726487Z * [new tag] trunk/a77f5d9a005a789804fc4d5064036490db5f48fb -> trunk/a77f5d9a005a789804fc4d5064036490db5f48fb 2025-11-03T16:58:02.2727722Z * [new tag] trunk/a7fd0b40018da3c692f74257fc9b1d8cc5c410ff -> trunk/a7fd0b40018da3c692f74257fc9b1d8cc5c410ff 2025-11-03T16:58:02.2729316Z * [new tag] trunk/a988510c3396a668e92801c61b93f7b06219ad0e -> trunk/a988510c3396a668e92801c61b93f7b06219ad0e 2025-11-03T16:58:02.2730868Z * [new tag] trunk/a9b29caeae77480abaee14c89cff19f2c23921a1 -> trunk/a9b29caeae77480abaee14c89cff19f2c23921a1 2025-11-03T16:58:02.2732087Z * [new tag] trunk/aa9c96af041b26c9c55adac490f3449b98f27d06 -> trunk/aa9c96af041b26c9c55adac490f3449b98f27d06 2025-11-03T16:58:02.2733626Z * [new tag] trunk/aab27b051a5e24e71931ef0386f9084d51347211 -> trunk/aab27b051a5e24e71931ef0386f9084d51347211 2025-11-03T16:58:02.2734860Z * [new tag] trunk/ac841267a197e102e1800786b3f6bfcd068b3d85 -> trunk/ac841267a197e102e1800786b3f6bfcd068b3d85 2025-11-03T16:58:02.2736518Z * [new tag] trunk/acd936cc1a53b3c97f1265b6e40e2c09ba0a182d -> trunk/acd936cc1a53b3c97f1265b6e40e2c09ba0a182d 2025-11-03T16:58:02.2737776Z * [new tag] trunk/ad02bd13dfa017f69def846b265a566c4ec5cb3f -> trunk/ad02bd13dfa017f69def846b265a566c4ec5cb3f 2025-11-03T16:58:02.2739476Z * [new tag] trunk/ad3a56ab9857905cdc8f62fb635d578982797c73 -> trunk/ad3a56ab9857905cdc8f62fb635d578982797c73 2025-11-03T16:58:02.2741036Z * [new tag] trunk/ad559072db0cc68aad277380d20bacfdf0d944ac -> trunk/ad559072db0cc68aad277380d20bacfdf0d944ac 2025-11-03T16:58:02.2742326Z * [new tag] trunk/add37bacda45d7de852dedee652346f2857df6fc -> trunk/add37bacda45d7de852dedee652346f2857df6fc 2025-11-03T16:58:02.2743768Z * [new tag] trunk/adedf26e21d1ce835750c99464b3ffa0f6b38fd0 -> trunk/adedf26e21d1ce835750c99464b3ffa0f6b38fd0 2025-11-03T16:58:02.2745372Z * [new tag] trunk/ae038f871b3cb91c36df3a9d4b0ffb4f35d20482 -> trunk/ae038f871b3cb91c36df3a9d4b0ffb4f35d20482 2025-11-03T16:58:02.2746571Z * [new tag] trunk/afaaaa314cc9358a10e9b1986642d49c00773560 -> trunk/afaaaa314cc9358a10e9b1986642d49c00773560 2025-11-03T16:58:02.2747894Z * [new tag] trunk/b060e5c131e9cf43ba6e9b5bda2cb5789c61de47 -> trunk/b060e5c131e9cf43ba6e9b5bda2cb5789c61de47 2025-11-03T16:58:02.2749519Z * [new tag] trunk/b0831930edaa732dd73dc81223fc0a9f5fe6b956 -> trunk/b0831930edaa732dd73dc81223fc0a9f5fe6b956 2025-11-03T16:58:02.2750739Z * [new tag] trunk/b09fb481e0fb00dc475f218e9b14660161b40a94 -> trunk/b09fb481e0fb00dc475f218e9b14660161b40a94 2025-11-03T16:58:02.2752357Z * [new tag] trunk/b2a0f90501dd3a16a6ccaf4c49e1c10f6df4ce1d -> trunk/b2a0f90501dd3a16a6ccaf4c49e1c10f6df4ce1d 2025-11-03T16:58:02.2753570Z * [new tag] trunk/b33762bd2f5f1b115f049d83bf553b08d1ffc97f -> trunk/b33762bd2f5f1b115f049d83bf553b08d1ffc97f 2025-11-03T16:58:02.2755150Z * [new tag] trunk/b3861ac8e70a2334278c17f834880e2d45bbe7a9 -> trunk/b3861ac8e70a2334278c17f834880e2d45bbe7a9 2025-11-03T16:58:02.2756489Z * [new tag] trunk/b4403bfc62ca97eec554cdf815baab1fe93057d9 -> trunk/b4403bfc62ca97eec554cdf815baab1fe93057d9 2025-11-03T16:58:02.2758111Z * [new tag] trunk/b44423bbb43860c1e340cbebc9d101dc18031ecb -> trunk/b44423bbb43860c1e340cbebc9d101dc18031ecb 2025-11-03T16:58:02.2759335Z * [new tag] trunk/b470e59c382649557c49d3c31b8dec1b036eb329 -> trunk/b470e59c382649557c49d3c31b8dec1b036eb329 2025-11-03T16:58:02.2760830Z * [new tag] trunk/b5189e269e899a6ad9ddc4e2ce6f33029290af82 -> trunk/b5189e269e899a6ad9ddc4e2ce6f33029290af82 2025-11-03T16:58:02.2762417Z * [new tag] trunk/b71966f67b2b688f9d71f1602bd3c52d413dacef -> trunk/b71966f67b2b688f9d71f1602bd3c52d413dacef 2025-11-03T16:58:02.2763563Z * [new tag] trunk/b7d348a9074af70f34848298a4e7a7985b197742 -> trunk/b7d348a9074af70f34848298a4e7a7985b197742 2025-11-03T16:58:02.2765534Z * [new tag] trunk/b903018c26217edca67c9023b104027aea377b83 -> trunk/b903018c26217edca67c9023b104027aea377b83 2025-11-03T16:58:02.2766841Z * [new tag] trunk/b939de26d115d33d3cbc17dcf49dbdc1a84e9727 -> trunk/b939de26d115d33d3cbc17dcf49dbdc1a84e9727 2025-11-03T16:58:02.2768437Z * [new tag] trunk/b9bcb37f40e72dc4c2fc1d9281175409fe474194 -> trunk/b9bcb37f40e72dc4c2fc1d9281175409fe474194 2025-11-03T16:58:02.2769757Z * [new tag] trunk/ba71e9ca9ac21e6570b20353d0a67089ad32e442 -> trunk/ba71e9ca9ac21e6570b20353d0a67089ad32e442 2025-11-03T16:58:02.2771479Z * [new tag] trunk/bb54296258854897058e8f158afa8efedc4fcf18 -> trunk/bb54296258854897058e8f158afa8efedc4fcf18 2025-11-03T16:58:02.2772730Z * [new tag] trunk/bbb7d2270b6f647b97f8a9c1d4e69ddaf48ef661 -> trunk/bbb7d2270b6f647b97f8a9c1d4e69ddaf48ef661 2025-11-03T16:58:02.2774042Z * [new tag] trunk/bc03d7c97462abfd45fc1b7b73d5c7acbbd2c185 -> trunk/bc03d7c97462abfd45fc1b7b73d5c7acbbd2c185 2025-11-03T16:58:02.2775749Z * [new tag] trunk/bc5111cd8dc5167cbcbcb2c0f23e53e61a6e8ea5 -> trunk/bc5111cd8dc5167cbcbcb2c0f23e53e61a6e8ea5 2025-11-03T16:58:02.2776836Z * [new tag] trunk/be283297100ab86123e74b7a8372995d32b140c8 -> trunk/be283297100ab86123e74b7a8372995d32b140c8 2025-11-03T16:58:02.2778494Z * [new tag] trunk/bea89d60603ca33f1903e57db5c734a67535c26d -> trunk/bea89d60603ca33f1903e57db5c734a67535c26d 2025-11-03T16:58:02.2779820Z * [new tag] trunk/bebabd7fce29ea49b9269aeaa9fe3f34a3e1127e -> trunk/bebabd7fce29ea49b9269aeaa9fe3f34a3e1127e 2025-11-03T16:58:02.2781067Z * [new tag] trunk/bfb47ec50e09dc2921ba47fedf84a1e28ee991a7 -> trunk/bfb47ec50e09dc2921ba47fedf84a1e28ee991a7 2025-11-03T16:58:02.2782496Z * [new tag] trunk/bfc2050db975e589795cd3eceaed2e83bf89ad35 -> trunk/bfc2050db975e589795cd3eceaed2e83bf89ad35 2025-11-03T16:58:02.2783776Z * [new tag] trunk/c01636e1bc7338eecbcfd2c17fd3fa2cad836983 -> trunk/c01636e1bc7338eecbcfd2c17fd3fa2cad836983 2025-11-03T16:58:02.2785434Z * [new tag] trunk/c0bbda37e8d74734b64421762eb365e620f258c1 -> trunk/c0bbda37e8d74734b64421762eb365e620f258c1 2025-11-03T16:58:02.2786674Z * [new tag] trunk/c201a1cab1439bfcfaff7e4fdab0d6eeeb880a70 -> trunk/c201a1cab1439bfcfaff7e4fdab0d6eeeb880a70 2025-11-03T16:58:02.2788257Z * [new tag] trunk/c2e3cc7aedb2e7d89443225c7cccd08a0f8a3587 -> trunk/c2e3cc7aedb2e7d89443225c7cccd08a0f8a3587 2025-11-03T16:58:02.2789489Z * [new tag] trunk/c3047938a01c844c752e5462b05d180a04d438fd -> trunk/c3047938a01c844c752e5462b05d180a04d438fd 2025-11-03T16:58:02.2791072Z * [new tag] trunk/c37802a8c4388160faa65f8bc9c29d98b4d2fb3c -> trunk/c37802a8c4388160faa65f8bc9c29d98b4d2fb3c 2025-11-03T16:58:02.2792310Z * [new tag] trunk/c3b71d54995a16fd09c18a5da5d598a23388bbd6 -> trunk/c3b71d54995a16fd09c18a5da5d598a23388bbd6 2025-11-03T16:58:02.2793626Z * [new tag] trunk/c3d205d5986824e72df353c9bf112dc9b2a3612a -> trunk/c3d205d5986824e72df353c9bf112dc9b2a3612a 2025-11-03T16:58:02.2794917Z * [new tag] trunk/c3dc0c7089f3204c36961acc227a3ff503530094 -> trunk/c3dc0c7089f3204c36961acc227a3ff503530094 2025-11-03T16:58:02.2796540Z * [new tag] trunk/c54e2c5b4138cb85412a74189d1f7f771ad1116e -> trunk/c54e2c5b4138cb85412a74189d1f7f771ad1116e 2025-11-03T16:58:02.2797665Z * [new tag] trunk/c5701d0ab5f55b23243dd437f57e856970c45bb6 -> trunk/c5701d0ab5f55b23243dd437f57e856970c45bb6 2025-11-03T16:58:02.2811498Z * [new tag] trunk/c58d0ad85dfbf026d5f21ee739c139407dc20e54 -> trunk/c58d0ad85dfbf026d5f21ee739c139407dc20e54 2025-11-03T16:58:02.2812246Z * [new tag] trunk/c594950e867e420a6086190a88e4b8fbd687af91 -> trunk/c594950e867e420a6086190a88e4b8fbd687af91 2025-11-03T16:58:02.2812881Z * [new tag] trunk/c6a02eae5b77758ee7ee0b1dfbcf20a8341914a7 -> trunk/c6a02eae5b77758ee7ee0b1dfbcf20a8341914a7 2025-11-03T16:58:02.2813511Z * [new tag] trunk/c8adc08b3b19b838dcdc71f2055600f420e6a91b -> trunk/c8adc08b3b19b838dcdc71f2055600f420e6a91b 2025-11-03T16:58:02.2814124Z * [new tag] trunk/c9eabadc5e964678556a6bfcde455a5e11dc0785 -> trunk/c9eabadc5e964678556a6bfcde455a5e11dc0785 2025-11-03T16:58:02.2814729Z * [new tag] trunk/cb6966704cf70ebac6d2b36b65e9a75e11c29e71 -> trunk/cb6966704cf70ebac6d2b36b65e9a75e11c29e71 2025-11-03T16:58:02.2815327Z * [new tag] trunk/cde81e92b95eee9af2879c9c75f7b03699ca72ad -> trunk/cde81e92b95eee9af2879c9c75f7b03699ca72ad 2025-11-03T16:58:02.2816065Z * [new tag] trunk/cee03634dab2dbf259d84bc389bdea80de0f80ac -> trunk/cee03634dab2dbf259d84bc389bdea80de0f80ac 2025-11-03T16:58:02.2816687Z * [new tag] trunk/cf7756da385e03c169f7e6a5b5ff2c0e29b2814a -> trunk/cf7756da385e03c169f7e6a5b5ff2c0e29b2814a 2025-11-03T16:58:02.2817285Z * [new tag] trunk/cf9a834f39d6c3b769381cc1321bcb5c54132415 -> trunk/cf9a834f39d6c3b769381cc1321bcb5c54132415 2025-11-03T16:58:02.2817879Z * [new tag] trunk/d049ed2cb1619c44279cb716b8a1d94e4df3b372 -> trunk/d049ed2cb1619c44279cb716b8a1d94e4df3b372 2025-11-03T16:58:02.2818475Z * [new tag] trunk/d17790072381f0e76e177ae4093504a6908a5a4d -> trunk/d17790072381f0e76e177ae4093504a6908a5a4d 2025-11-03T16:58:02.2819044Z * [new tag] trunk/d1a6e006e090b1365e6653853982f175cca452e8 -> trunk/d1a6e006e090b1365e6653853982f175cca452e8 2025-11-03T16:58:02.2819640Z * [new tag] trunk/d2be06f673ff11e59b234156dccaad2dc22c7746 -> trunk/d2be06f673ff11e59b234156dccaad2dc22c7746 2025-11-03T16:58:02.2820404Z * [new tag] trunk/d2eff5d454ab2cb0a5ccdfb5eb6e7d6dcc75e097 -> trunk/d2eff5d454ab2cb0a5ccdfb5eb6e7d6dcc75e097 2025-11-03T16:58:02.2821418Z * [new tag] trunk/d3be06cbdca4b2a9cd8ae35926b3f20b1bed1fe2 -> trunk/d3be06cbdca4b2a9cd8ae35926b3f20b1bed1fe2 2025-11-03T16:58:02.2823575Z * [new tag] trunk/d3e511f07c0b1fc6df22347aec7f8e700bdb0d24 -> trunk/d3e511f07c0b1fc6df22347aec7f8e700bdb0d24 2025-11-03T16:58:02.2824807Z * [new tag] trunk/d401e4e70a1dd19733d3e3c84affc4884605efb4 -> trunk/d401e4e70a1dd19733d3e3c84affc4884605efb4 2025-11-03T16:58:02.2826114Z * [new tag] trunk/d46d8d6f54b15ded4f2483c7bde31be124281ab8 -> trunk/d46d8d6f54b15ded4f2483c7bde31be124281ab8 2025-11-03T16:58:02.2827388Z * [new tag] trunk/d6d6fa26f540c10c57ac80547a9475e9f4c201f2 -> trunk/d6d6fa26f540c10c57ac80547a9475e9f4c201f2 2025-11-03T16:58:02.2829014Z * [new tag] trunk/d7040e6d7515cea485824d2b810bea94e5958dea -> trunk/d7040e6d7515cea485824d2b810bea94e5958dea 2025-11-03T16:58:02.2830304Z * [new tag] trunk/d80ae738c9c1810b30a88233bdcd34cae16c5c6e -> trunk/d80ae738c9c1810b30a88233bdcd34cae16c5c6e 2025-11-03T16:58:02.2831828Z * [new tag] trunk/d8283a317a335c628c4541c6a60e102fdead57b7 -> trunk/d8283a317a335c628c4541c6a60e102fdead57b7 2025-11-03T16:58:02.2833109Z * [new tag] trunk/d9483d4c8d3f220ad505afd05606e068184896a4 -> trunk/d9483d4c8d3f220ad505afd05606e068184896a4 2025-11-03T16:58:02.2834425Z * [new tag] trunk/d962bed15798961ffd10e64068f86a2cec411ffb -> trunk/d962bed15798961ffd10e64068f86a2cec411ffb 2025-11-03T16:58:02.2835751Z * [new tag] trunk/d97144d31ef42e762ce65f467cbaf065edb241cc -> trunk/d97144d31ef42e762ce65f467cbaf065edb241cc 2025-11-03T16:58:02.2837378Z * [new tag] trunk/dc011d3203b24ebf1f78ab8dd64766205e64d8ec -> trunk/dc011d3203b24ebf1f78ab8dd64766205e64d8ec 2025-11-03T16:58:02.2838675Z * [new tag] trunk/dd1fe7c22f8eee3984a22663636b0d1403b079ba -> trunk/dd1fe7c22f8eee3984a22663636b0d1403b079ba 2025-11-03T16:58:02.2840226Z * [new tag] trunk/deb776319b12cad7fffb5cf9d8851a50e2b0e9ea -> trunk/deb776319b12cad7fffb5cf9d8851a50e2b0e9ea 2025-11-03T16:58:02.2841519Z * [new tag] trunk/defac66e39c60f867795346ded171366ed682ee3 -> trunk/defac66e39c60f867795346ded171366ed682ee3 2025-11-03T16:58:02.2843051Z * [new tag] trunk/df71b7072799c451a008cb36142dfdb1487f0d5e -> trunk/df71b7072799c451a008cb36142dfdb1487f0d5e 2025-11-03T16:58:02.2844255Z * [new tag] trunk/dfebdcab86acbaa0eaa996b47595e5f27a66492e -> trunk/dfebdcab86acbaa0eaa996b47595e5f27a66492e 2025-11-03T16:58:02.2845520Z * [new tag] trunk/e0604d3170f1123ebacbced66be400c7f162621f -> trunk/e0604d3170f1123ebacbced66be400c7f162621f 2025-11-03T16:58:02.2847188Z * [new tag] trunk/e0791fc11dc0024a828495985898b29120dcc4c1 -> trunk/e0791fc11dc0024a828495985898b29120dcc4c1 2025-11-03T16:58:02.2848402Z * [new tag] trunk/e0ca3049c0c7726747094f47f80cadcfad51320d -> trunk/e0ca3049c0c7726747094f47f80cadcfad51320d 2025-11-03T16:58:02.2849945Z * [new tag] trunk/e105a475756ae6dd45c8134bd4583afc57ee416e -> trunk/e105a475756ae6dd45c8134bd4583afc57ee416e 2025-11-03T16:58:02.2851284Z * [new tag] trunk/e137cd0a1046ead2ca9d631f427655adced606c2 -> trunk/e137cd0a1046ead2ca9d631f427655adced606c2 2025-11-03T16:58:02.2852888Z * [new tag] trunk/e1d011d6eb571cd98ec7c7ed8e8b518a5463ec97 -> trunk/e1d011d6eb571cd98ec7c7ed8e8b518a5463ec97 2025-11-03T16:58:02.2854085Z * [new tag] trunk/e214af6ae8e613485ac3dac3e2094cf53dfc3c1d -> trunk/e214af6ae8e613485ac3dac3e2094cf53dfc3c1d 2025-11-03T16:58:02.2855617Z * [new tag] trunk/e2dc32f4baf853e6171777aed530fc83bcdb35ae -> trunk/e2dc32f4baf853e6171777aed530fc83bcdb35ae 2025-11-03T16:58:02.2857478Z * [new tag] trunk/e380028a51c38de80b8888d027c63e290a773651 -> trunk/e380028a51c38de80b8888d027c63e290a773651 2025-11-03T16:58:02.2858883Z * [new tag] trunk/e3ae0594d16134632ff587c9ab400d4148c83e9f -> trunk/e3ae0594d16134632ff587c9ab400d4148c83e9f 2025-11-03T16:58:02.2860174Z * [new tag] trunk/e3e93c7107830c13f4139c3a62fda62c6b84bbf5 -> trunk/e3e93c7107830c13f4139c3a62fda62c6b84bbf5 2025-11-03T16:58:02.2861773Z * [new tag] trunk/e4043884c734a3947f4c8da0cfbba43fc6319f3d -> trunk/e4043884c734a3947f4c8da0cfbba43fc6319f3d 2025-11-03T16:58:02.2863066Z * [new tag] trunk/e471800dceb8a6592562fc4fc96a5e80bc494c0a -> trunk/e471800dceb8a6592562fc4fc96a5e80bc494c0a 2025-11-03T16:58:02.2864711Z * [new tag] trunk/e83be7042e809ef51c67f53d83ac45794447c73f -> trunk/e83be7042e809ef51c67f53d83ac45794447c73f 2025-11-03T16:58:02.2866040Z * [new tag] trunk/e8d887ae3fceb97283bc2d751621b9615524b4b7 -> trunk/e8d887ae3fceb97283bc2d751621b9615524b4b7 2025-11-03T16:58:02.2867598Z * [new tag] trunk/e8fadba28cc225782ddce1bf0ef8315ee1743761 -> trunk/e8fadba28cc225782ddce1bf0ef8315ee1743761 2025-11-03T16:58:02.2868949Z * [new tag] trunk/e95920e3e6f9d8218c6aed2b8b6ce9b838a49104 -> trunk/e95920e3e6f9d8218c6aed2b8b6ce9b838a49104 2025-11-03T16:58:02.2870274Z * [new tag] trunk/ea698e8bfcd2ec1820b3ba90b9c47416a33348de -> trunk/ea698e8bfcd2ec1820b3ba90b9c47416a33348de 2025-11-03T16:58:02.2871911Z * [new tag] trunk/eae701cad03feb87908107cb5246d9e13a9d426e -> trunk/eae701cad03feb87908107cb5246d9e13a9d426e 2025-11-03T16:58:02.2873140Z * [new tag] trunk/eb2bad5bb54df8a2563f2619819e9e689e0ba10e -> trunk/eb2bad5bb54df8a2563f2619819e9e689e0ba10e 2025-11-03T16:58:02.2874677Z * [new tag] trunk/ebb2b2e894a4ede8efc5f5fff068d4ac9972f77a -> trunk/ebb2b2e894a4ede8efc5f5fff068d4ac9972f77a 2025-11-03T16:58:02.2875997Z * [new tag] trunk/ee7434be822cf6e75b4566d8159f550ee233d8ae -> trunk/ee7434be822cf6e75b4566d8159f550ee233d8ae 2025-11-03T16:58:02.2877551Z * [new tag] trunk/ef8d97efcf05e1f3a05ad5ce3444e3af42481505 -> trunk/ef8d97efcf05e1f3a05ad5ce3444e3af42481505 2025-11-03T16:58:02.2878762Z * [new tag] trunk/f013e804c876a34cbc86a6a475597e3e29812883 -> trunk/f013e804c876a34cbc86a6a475597e3e29812883 2025-11-03T16:58:02.2880293Z * [new tag] trunk/f02708c2be66c7cf0edec52663de69f66ba4b1df -> trunk/f02708c2be66c7cf0edec52663de69f66ba4b1df 2025-11-03T16:58:02.2881589Z * [new tag] trunk/f0745ddb11ab66a8bb4de8031bd7e9dccba19379 -> trunk/f0745ddb11ab66a8bb4de8031bd7e9dccba19379 2025-11-03T16:58:02.2883139Z * [new tag] trunk/f167fd09fabf3f5be63cd31451f29d49796c14e8 -> trunk/f167fd09fabf3f5be63cd31451f29d49796c14e8 2025-11-03T16:58:02.2884548Z * [new tag] trunk/f1a34407152302812bf633360cd29e9fba8de71e -> trunk/f1a34407152302812bf633360cd29e9fba8de71e 2025-11-03T16:58:02.2885861Z * [new tag] trunk/f1af679270392c83e03808c8af5e2cbe3cdf16ce -> trunk/f1af679270392c83e03808c8af5e2cbe3cdf16ce 2025-11-03T16:58:02.2887410Z * [new tag] trunk/f1e4c42b6ef3d3cea08ab3babb693e3ce42cf08b -> trunk/f1e4c42b6ef3d3cea08ab3babb693e3ce42cf08b 2025-11-03T16:58:02.2888706Z * [new tag] trunk/f20bf778746a7662a9bdbd2fa49d326772ca90ed -> trunk/f20bf778746a7662a9bdbd2fa49d326772ca90ed 2025-11-03T16:58:02.2890217Z * [new tag] trunk/f2450798cd6f30b811f6d8256da0c271cad2285c -> trunk/f2450798cd6f30b811f6d8256da0c271cad2285c 2025-11-03T16:58:02.2891527Z * [new tag] trunk/f2c81635c85bf8d1c523d6b17f8016ad78ea02e2 -> trunk/f2c81635c85bf8d1c523d6b17f8016ad78ea02e2 2025-11-03T16:58:02.2893112Z * [new tag] trunk/f33abae6957b8ef56268ce22e660d3ed3111c7c3 -> trunk/f33abae6957b8ef56268ce22e660d3ed3111c7c3 2025-11-03T16:58:02.2894626Z * [new tag] trunk/f36f372acc28062e0988d84699c62689b0d89a6e -> trunk/f36f372acc28062e0988d84699c62689b0d89a6e 2025-11-03T16:58:02.2895912Z * [new tag] trunk/f452edd782fb7342e2a5740cea9b78135e714858 -> trunk/f452edd782fb7342e2a5740cea9b78135e714858 2025-11-03T16:58:02.2897462Z * [new tag] trunk/f4d05feb7a5f9191509e1e539ea210fcbe538b32 -> trunk/f4d05feb7a5f9191509e1e539ea210fcbe538b32 2025-11-03T16:58:02.2898790Z * [new tag] trunk/f5543e3741bbfba4579afc3b49d0720f2eee7e4d -> trunk/f5543e3741bbfba4579afc3b49d0720f2eee7e4d 2025-11-03T16:58:02.2900062Z * [new tag] trunk/f5cb9a4c68d9271c58ef4d3257210984b8e85099 -> trunk/f5cb9a4c68d9271c58ef4d3257210984b8e85099 2025-11-03T16:58:02.2901707Z * [new tag] trunk/f60751024ecac57840a2d25132d0c1562351e66e -> trunk/f60751024ecac57840a2d25132d0c1562351e66e 2025-11-03T16:58:02.2903286Z * [new tag] trunk/f6951cb8ea1ef96a33627734eb68b5dc7a04b6ed -> trunk/f6951cb8ea1ef96a33627734eb68b5dc7a04b6ed 2025-11-03T16:58:02.2904604Z * [new tag] trunk/f89a7e9fe80fb0f74af52bad892e1afafa57be9c -> trunk/f89a7e9fe80fb0f74af52bad892e1afafa57be9c 2025-11-03T16:58:02.2905943Z * [new tag] trunk/f8b4c002940d96ba2afc14ac4cf0b5277776bc02 -> trunk/f8b4c002940d96ba2afc14ac4cf0b5277776bc02 2025-11-03T16:58:02.2907455Z * [new tag] trunk/f911d64750efb22097eef5505552a7669a0c53de -> trunk/f911d64750efb22097eef5505552a7669a0c53de 2025-11-03T16:58:02.2908852Z * [new tag] trunk/f91899ca6c15d611c453591527e1be8b43e5d4ea -> trunk/f91899ca6c15d611c453591527e1be8b43e5d4ea 2025-11-03T16:58:02.2910447Z * [new tag] trunk/f93ea7dab190fa0953b7a4921b76993cadafc7c9 -> trunk/f93ea7dab190fa0953b7a4921b76993cadafc7c9 2025-11-03T16:58:02.2911880Z * [new tag] trunk/fa4cb918462d3cbe7cf48db7c1498824c1ca1cc4 -> trunk/fa4cb918462d3cbe7cf48db7c1498824c1ca1cc4 2025-11-03T16:58:02.2913089Z * [new tag] trunk/fa560e1158b496d1c37ca2a2443a0b7f5a08dad5 -> trunk/fa560e1158b496d1c37ca2a2443a0b7f5a08dad5 2025-11-03T16:58:02.2914656Z * [new tag] trunk/fa8e073a4e17f12e7e4bb0b26471179d2ca57471 -> trunk/fa8e073a4e17f12e7e4bb0b26471179d2ca57471 2025-11-03T16:58:02.2915960Z * [new tag] trunk/fb545fb0686c94cc017c15ca7e342996b49edb81 -> trunk/fb545fb0686c94cc017c15ca7e342996b49edb81 2025-11-03T16:58:02.2917505Z * [new tag] trunk/fc540cefd498f1001a5fc7a4f187080628b27839 -> trunk/fc540cefd498f1001a5fc7a4f187080628b27839 2025-11-03T16:58:02.2918788Z * [new tag] trunk/fc8ac1216c61d8736f45a58825abf58e7d51e088 -> trunk/fc8ac1216c61d8736f45a58825abf58e7d51e088 2025-11-03T16:58:02.2920406Z * [new tag] trunk/fcc10635660ceb619f78d06eaaa8b3c0bdd52ab8 -> trunk/fcc10635660ceb619f78d06eaaa8b3c0bdd52ab8 2025-11-03T16:58:02.2921766Z * [new tag] trunk/fcd5f8c352b5b75bd32e57fa044ec5df095032da -> trunk/fcd5f8c352b5b75bd32e57fa044ec5df095032da 2025-11-03T16:58:02.2923558Z * [new tag] trunk/fd5da81fdd76c0a78ef935d077a93b48b78fca3a -> trunk/fd5da81fdd76c0a78ef935d077a93b48b78fca3a 2025-11-03T16:58:02.2924753Z * [new tag] trunk/fd68d409ada709450ced3030bde89ec662a3f7b7 -> trunk/fd68d409ada709450ced3030bde89ec662a3f7b7 2025-11-03T16:58:02.2926066Z * [new tag] trunk/fea819ed080175f9a93205cc9074022bfa486325 -> trunk/fea819ed080175f9a93205cc9074022bfa486325 2025-11-03T16:58:02.2927559Z * [new tag] trunk/fee1ac927d66fce67f862311292dcc325b8def44 -> trunk/fee1ac927d66fce67f862311292dcc325b8def44 2025-11-03T16:58:02.2928918Z * [new tag] trunk/fee7624bd66e64f575c1396fcaaa54cabce49ca4 -> trunk/fee7624bd66e64f575c1396fcaaa54cabce49ca4 2025-11-03T16:58:02.2930438Z * [new tag] trunk/fefb546b9100614da1fa703e12651bc115059f02 -> trunk/fefb546b9100614da1fa703e12651bc115059f02 2025-11-03T16:58:02.2932054Z * [new tag] trunk/ff46d5a79bec09d9ae334d325f5f8cc89306b536 -> trunk/ff46d5a79bec09d9ae334d325f5f8cc89306b536 2025-11-03T16:58:02.2933463Z * [new tag] trunk/ffaa6578b7a641b1ac1fcf451f04cf5e1212c9fe -> trunk/ffaa6578b7a641b1ac1fcf451f04cf5e1212c9fe 2025-11-03T16:58:02.2934794Z * [new tag] v0.1.1 -> v0.1.1 2025-11-03T16:58:02.2936147Z * [new tag] v0.1.10 -> v0.1.10 2025-11-03T16:58:02.2937429Z * [new tag] v0.1.11 -> v0.1.11 2025-11-03T16:58:02.2938792Z * [new tag] v0.1.12 -> v0.1.12 2025-11-03T16:58:02.2940032Z * [new tag] v0.1.2 -> v0.1.2 2025-11-03T16:58:02.2941289Z * [new tag] v0.1.3 -> v0.1.3 2025-11-03T16:58:02.2942768Z * [new tag] v0.1.4 -> v0.1.4 2025-11-03T16:58:02.2944065Z * [new tag] v0.1.5 -> v0.1.5 2025-11-03T16:58:02.2945393Z * [new tag] v0.1.6 -> v0.1.6 2025-11-03T16:58:02.2946654Z * [new tag] v0.1.7 -> v0.1.7 2025-11-03T16:58:02.2947928Z * [new tag] v0.1.8 -> v0.1.8 2025-11-03T16:58:02.2949179Z * [new tag] v0.1.9 -> v0.1.9 2025-11-03T16:58:02.2950596Z * [new tag] v0.2.0 -> v0.2.0 2025-11-03T16:58:02.2951950Z * [new tag] v0.3.0 -> v0.3.0 2025-11-03T16:58:02.2953380Z * [new tag] v0.3.1 -> v0.3.1 2025-11-03T16:58:02.2954765Z * [new tag] v0.4.0 -> v0.4.0 2025-11-03T16:58:02.2956099Z * [new tag] v0.4.1 -> v0.4.1 2025-11-03T16:58:02.2957413Z * [new tag] v1.0.0 -> v1.0.0 2025-11-03T16:58:02.2958803Z * [new tag] v1.0.0a0 -> v1.0.0a0 2025-11-03T16:58:02.2960104Z * [new tag] v1.0.1 -> v1.0.1 2025-11-03T16:58:02.2961476Z * [new tag] v1.0rc0 -> v1.0rc0 2025-11-03T16:58:02.2962505Z * [new tag] v1.0rc1 -> v1.0rc1 2025-11-03T16:58:02.2963936Z * [new tag] v1.1.0 -> v1.1.0 2025-11-03T16:58:02.2965271Z * [new tag] v1.1.0a0 -> v1.1.0a0 2025-11-03T16:58:02.2966813Z * [new tag] v1.10.0 -> v1.10.0 2025-11-03T16:58:02.2968237Z * [new tag] v1.10.0-rc1 -> v1.10.0-rc1 2025-11-03T16:58:02.2969597Z * [new tag] v1.10.0-rc2 -> v1.10.0-rc2 2025-11-03T16:58:02.2970628Z * [new tag] v1.10.0-rc3 -> v1.10.0-rc3 2025-11-03T16:58:02.2972075Z * [new tag] v1.10.1 -> v1.10.1 2025-11-03T16:58:02.2973314Z * [new tag] v1.10.1-rc1 -> v1.10.1-rc1 2025-11-03T16:58:02.2974321Z * [new tag] v1.10.2 -> v1.10.2 2025-11-03T16:58:02.2975602Z * [new tag] v1.10.2-rc1 -> v1.10.2-rc1 2025-11-03T16:58:02.2976977Z * [new tag] v1.11.0 -> v1.11.0 2025-11-03T16:58:02.2978477Z * [new tag] v1.11.0-rc1 -> v1.11.0-rc1 2025-11-03T16:58:02.2979917Z * [new tag] v1.11.0-rc2 -> v1.11.0-rc2 2025-11-03T16:58:02.2981855Z * [new tag] v1.11.0-rc3 -> v1.11.0-rc3 2025-11-03T16:58:02.2983264Z * [new tag] v1.11.0-rc4 -> v1.11.0-rc4 2025-11-03T16:58:02.2984670Z * [new tag] v1.11.0-rc5 -> v1.11.0-rc5 2025-11-03T16:58:02.2985688Z * [new tag] v1.11.0-rc6 -> v1.11.0-rc6 2025-11-03T16:58:02.2986944Z * [new tag] v1.11.0-rc7 -> v1.11.0-rc7 2025-11-03T16:58:02.2988280Z * [new tag] v1.12.0 -> v1.12.0 2025-11-03T16:58:02.2989560Z * [new tag] v1.12.0-rc1 -> v1.12.0-rc1 2025-11-03T16:58:02.2990926Z * [new tag] v1.12.0-rc2 -> v1.12.0-rc2 2025-11-03T16:58:02.2992360Z * [new tag] v1.12.0-rc3 -> v1.12.0-rc3 2025-11-03T16:58:02.2993766Z * [new tag] v1.12.0-rc4 -> v1.12.0-rc4 2025-11-03T16:58:02.2995065Z * [new tag] v1.12.0-rc5 -> v1.12.0-rc5 2025-11-03T16:58:02.2996499Z * [new tag] v1.12.0-rc6 -> v1.12.0-rc6 2025-11-03T16:58:02.2997722Z * [new tag] v1.12.0-rc7 -> v1.12.0-rc7 2025-11-03T16:58:02.2998752Z * [new tag] v1.12.0-rc8 -> v1.12.0-rc8 2025-11-03T16:58:02.3000033Z * [new tag] v1.12.1 -> v1.12.1 2025-11-03T16:58:02.3001506Z * [new tag] v1.12.1-rc1 -> v1.12.1-rc1 2025-11-03T16:58:02.3002824Z * [new tag] v1.12.1-rc2 -> v1.12.1-rc2 2025-11-03T16:58:02.3004249Z * [new tag] v1.12.1-rc3 -> v1.12.1-rc3 2025-11-03T16:58:02.3005627Z * [new tag] v1.12.1-rc4 -> v1.12.1-rc4 2025-11-03T16:58:02.3006793Z * [new tag] v1.12.1-rc5 -> v1.12.1-rc5 2025-11-03T16:58:02.3008263Z * [new tag] v1.13.0 -> v1.13.0 2025-11-03T16:58:02.3009599Z * [new tag] v1.13.0-rc1 -> v1.13.0-rc1 2025-11-03T16:58:02.3010962Z * [new tag] v1.13.0-rc2 -> v1.13.0-rc2 2025-11-03T16:58:02.3012324Z * [new tag] v1.13.0-rc3 -> v1.13.0-rc3 2025-11-03T16:58:02.3013739Z * [new tag] v1.13.0-rc4 -> v1.13.0-rc4 2025-11-03T16:58:02.3014788Z * [new tag] v1.13.0-rc5 -> v1.13.0-rc5 2025-11-03T16:58:02.3016022Z * [new tag] v1.13.0-rc6 -> v1.13.0-rc6 2025-11-03T16:58:02.3017453Z * [new tag] v1.13.1 -> v1.13.1 2025-11-03T16:58:02.3018592Z * [new tag] v1.13.1-rc1 -> v1.13.1-rc1 2025-11-03T16:58:02.3019932Z * [new tag] v1.2.0 -> v1.2.0 2025-11-03T16:58:02.3021423Z * [new tag] v1.2.0a0 -> v1.2.0a0 2025-11-03T16:58:02.3022880Z * [new tag] v1.3.0 -> v1.3.0 2025-11-03T16:58:02.3024203Z * [new tag] v1.3.0a0 -> v1.3.0a0 2025-11-03T16:58:02.3025344Z * [new tag] v1.3.1 -> v1.3.1 2025-11-03T16:58:02.3026670Z * [new tag] v1.4.0 -> v1.4.0 2025-11-03T16:58:02.3028019Z * [new tag] v1.4.0a0 -> v1.4.0a0 2025-11-03T16:58:02.3029336Z * [new tag] v1.4.1 -> v1.4.1 2025-11-03T16:58:02.3030622Z * [new tag] v1.5.0 -> v1.5.0 2025-11-03T16:58:02.3031942Z * [new tag] v1.5.0-rc1 -> v1.5.0-rc1 2025-11-03T16:58:02.3033343Z * [new tag] v1.5.0-rc2 -> v1.5.0-rc2 2025-11-03T16:58:02.3034711Z * [new tag] v1.5.0-rc3 -> v1.5.0-rc3 2025-11-03T16:58:02.3035959Z * [new tag] v1.5.0-rc4 -> v1.5.0-rc4 2025-11-03T16:58:02.3037143Z * [new tag] v1.5.0-rc5 -> v1.5.0-rc5 2025-11-03T16:58:02.3038528Z * [new tag] v1.5.1 -> v1.5.1 2025-11-03T16:58:02.3039721Z * [new tag] v1.5.1-rc1 -> v1.5.1-rc1 2025-11-03T16:58:02.3040892Z * [new tag] v1.6.0 -> v1.6.0 2025-11-03T16:58:02.3042232Z * [new tag] v1.6.0-rc1 -> v1.6.0-rc1 2025-11-03T16:58:02.3043665Z * [new tag] v1.6.0-rc2 -> v1.6.0-rc2 2025-11-03T16:58:02.3045019Z * [new tag] v1.6.0-rc3 -> v1.6.0-rc3 2025-11-03T16:58:02.3046374Z * [new tag] v1.6.0-rc4 -> v1.6.0-rc4 2025-11-03T16:58:02.3047722Z * [new tag] v1.6.0-rc5 -> v1.6.0-rc5 2025-11-03T16:58:02.3049049Z * [new tag] v1.6.0-rc6 -> v1.6.0-rc6 2025-11-03T16:58:02.3050177Z * [new tag] v1.6.0-rc7 -> v1.6.0-rc7 2025-11-03T16:58:02.3051674Z * [new tag] v1.7.0 -> v1.7.0 2025-11-03T16:58:02.3053291Z * [new tag] v1.7.0-rc1 -> v1.7.0-rc1 2025-11-03T16:58:02.3054723Z * [new tag] v1.7.0-rc2 -> v1.7.0-rc2 2025-11-03T16:58:02.3056081Z * [new tag] v1.7.0-rc3 -> v1.7.0-rc3 2025-11-03T16:58:02.3057117Z * [new tag] v1.7.0-rc4 -> v1.7.0-rc4 2025-11-03T16:58:02.3058597Z * [new tag] v1.7.1 -> v1.7.1 2025-11-03T16:58:02.3060082Z * [new tag] v1.7.1-rc1 -> v1.7.1-rc1 2025-11-03T16:58:02.3061530Z * [new tag] v1.7.1-rc2 -> v1.7.1-rc2 2025-11-03T16:58:02.3063159Z * [new tag] v1.7.1-rc3 -> v1.7.1-rc3 2025-11-03T16:58:02.3064605Z * [new tag] v1.8.0 -> v1.8.0 2025-11-03T16:58:02.3065819Z * [new tag] v1.8.0-rc1 -> v1.8.0-rc1 2025-11-03T16:58:02.3067598Z * [new tag] v1.8.0-rc2 -> v1.8.0-rc2 2025-11-03T16:58:02.3068939Z * [new tag] v1.8.0-rc3 -> v1.8.0-rc3 2025-11-03T16:58:02.3070270Z * [new tag] v1.8.0-rc4 -> v1.8.0-rc4 2025-11-03T16:58:02.3071457Z * [new tag] v1.8.0-rc5 -> v1.8.0-rc5 2025-11-03T16:58:02.3072603Z * [new tag] v1.8.1 -> v1.8.1 2025-11-03T16:58:02.3073929Z * [new tag] v1.8.1-rc1 -> v1.8.1-rc1 2025-11-03T16:58:02.3075156Z * [new tag] v1.8.1-rc2 -> v1.8.1-rc2 2025-11-03T16:58:02.3076317Z * [new tag] v1.8.1-rc3 -> v1.8.1-rc3 2025-11-03T16:58:02.3078078Z * [new tag] v1.8.2 -> v1.8.2 2025-11-03T16:58:02.3079215Z * [new tag] v1.8.2-rc1 -> v1.8.2-rc1 2025-11-03T16:58:02.3080624Z * [new tag] v1.9.0 -> v1.9.0 2025-11-03T16:58:02.3081925Z * [new tag] v1.9.0-rc1 -> v1.9.0-rc1 2025-11-03T16:58:02.3083391Z * [new tag] v1.9.0-rc2 -> v1.9.0-rc2 2025-11-03T16:58:02.3084884Z * [new tag] v1.9.0-rc3 -> v1.9.0-rc3 2025-11-03T16:58:02.3085986Z * [new tag] v1.9.0-rc4 -> v1.9.0-rc4 2025-11-03T16:58:02.3087362Z * [new tag] v1.9.1 -> v1.9.1 2025-11-03T16:58:02.3088839Z * [new tag] v1.9.1-rc1 -> v1.9.1-rc1 2025-11-03T16:58:02.3089992Z * [new tag] v1.9.1-rc2 -> v1.9.1-rc2 2025-11-03T16:58:02.3091389Z * [new tag] v2.0.0 -> v2.0.0 2025-11-03T16:58:02.3092715Z * [new tag] v2.0.0-rc1 -> v2.0.0-rc1 2025-11-03T16:58:02.3094147Z * [new tag] v2.0.0-rc2 -> v2.0.0-rc2 2025-11-03T16:58:02.3095505Z * [new tag] v2.0.0-rc3 -> v2.0.0-rc3 2025-11-03T16:58:02.3096928Z * [new tag] v2.0.0-rc4 -> v2.0.0-rc4 2025-11-03T16:58:02.3098258Z * [new tag] v2.0.0-rc5 -> v2.0.0-rc5 2025-11-03T16:58:02.3099415Z * [new tag] v2.0.0-rc6 -> v2.0.0-rc6 2025-11-03T16:58:02.3100780Z * [new tag] v2.0.1 -> v2.0.1 2025-11-03T16:58:02.3102274Z * [new tag] v2.0.1-rc1 -> v2.0.1-rc1 2025-11-03T16:58:02.3103424Z * [new tag] v2.0.1-rc2 -> v2.0.1-rc2 2025-11-03T16:58:02.3104702Z * [new tag] v2.0.1-rc3 -> v2.0.1-rc3 2025-11-03T16:58:02.3105854Z * [new tag] v2.0.1-rc4 -> v2.0.1-rc4 2025-11-03T16:58:02.3107592Z * [new tag] v2.1.0 -> v2.1.0 2025-11-03T16:58:02.3108941Z * [new tag] v2.1.0-rc1 -> v2.1.0-rc1 2025-11-03T16:58:02.3110334Z * [new tag] v2.1.0-rc2 -> v2.1.0-rc2 2025-11-03T16:58:02.3111748Z * [new tag] v2.1.0-rc3 -> v2.1.0-rc3 2025-11-03T16:58:02.3113087Z * [new tag] v2.1.0-rc4 -> v2.1.0-rc4 2025-11-03T16:58:02.3114501Z * [new tag] v2.1.0-rc5 -> v2.1.0-rc5 2025-11-03T16:58:02.3115643Z * [new tag] v2.1.0-rc6 -> v2.1.0-rc6 2025-11-03T16:58:02.3116993Z * [new tag] v2.1.1 -> v2.1.1 2025-11-03T16:58:02.3118404Z * [new tag] v2.1.1-rc1 -> v2.1.1-rc1 2025-11-03T16:58:02.3119701Z * [new tag] v2.1.1-rc2 -> v2.1.1-rc2 2025-11-03T16:58:02.3121571Z * [new tag] v2.1.1-rc3 -> v2.1.1-rc3 2025-11-03T16:58:02.3123942Z * [new tag] v2.1.1-rc4 -> v2.1.1-rc4 2025-11-03T16:58:02.3125270Z * [new tag] v2.1.1-rc5 -> v2.1.1-rc5 2025-11-03T16:58:02.3126433Z * [new tag] v2.1.1-rc6 -> v2.1.1-rc6 2025-11-03T16:58:02.3127817Z * [new tag] v2.1.2 -> v2.1.2 2025-11-03T16:58:02.3129244Z * [new tag] v2.1.2-rc1 -> v2.1.2-rc1 2025-11-03T16:58:02.3130629Z * [new tag] v2.1.2-rc2 -> v2.1.2-rc2 2025-11-03T16:58:02.3131798Z * [new tag] v2.1.2-rc3 -> v2.1.2-rc3 2025-11-03T16:58:02.3133179Z * [new tag] v2.2.0 -> v2.2.0 2025-11-03T16:58:02.3134555Z * [new tag] v2.2.0-rc1 -> v2.2.0-rc1 2025-11-03T16:58:02.3135826Z * [new tag] v2.2.0-rc2 -> v2.2.0-rc2 2025-11-03T16:58:02.3137127Z * [new tag] v2.2.0-rc3 -> v2.2.0-rc3 2025-11-03T16:58:02.3138472Z * [new tag] v2.2.0-rc4 -> v2.2.0-rc4 2025-11-03T16:58:02.3139780Z * [new tag] v2.2.0-rc5 -> v2.2.0-rc5 2025-11-03T16:58:02.3141355Z * [new tag] v2.2.0-rc6 -> v2.2.0-rc6 2025-11-03T16:58:02.3142600Z * [new tag] v2.2.0-rc7 -> v2.2.0-rc7 2025-11-03T16:58:02.3143731Z * [new tag] v2.2.0-rc8 -> v2.2.0-rc8 2025-11-03T16:58:02.3145106Z * [new tag] v2.2.1 -> v2.2.1 2025-11-03T16:58:02.3146516Z * [new tag] v2.2.1-rc1 -> v2.2.1-rc1 2025-11-03T16:58:02.3147641Z * [new tag] v2.2.1-rc2 -> v2.2.1-rc2 2025-11-03T16:58:02.3148758Z * [new tag] v2.2.1-rc3 -> v2.2.1-rc3 2025-11-03T16:58:02.3149902Z * [new tag] v2.2.2 -> v2.2.2 2025-11-03T16:58:02.3151389Z * [new tag] v2.2.2-rc1 -> v2.2.2-rc1 2025-11-03T16:58:02.3152657Z * [new tag] v2.2.2-rc2 -> v2.2.2-rc2 2025-11-03T16:58:02.3154165Z * [new tag] v2.2.2-rc3 -> v2.2.2-rc3 2025-11-03T16:58:02.3155634Z * [new tag] v2.3.0 -> v2.3.0 2025-11-03T16:58:02.3156830Z * [new tag] v2.3.0-rc1 -> v2.3.0-rc1 2025-11-03T16:58:02.3158355Z * [new tag] v2.3.0-rc10 -> v2.3.0-rc10 2025-11-03T16:58:02.3159894Z * [new tag] v2.3.0-rc11 -> v2.3.0-rc11 2025-11-03T16:58:02.3161041Z * [new tag] v2.3.0-rc12 -> v2.3.0-rc12 2025-11-03T16:58:02.3162379Z * [new tag] v2.3.0-rc2 -> v2.3.0-rc2 2025-11-03T16:58:02.3163861Z * [new tag] v2.3.0-rc3 -> v2.3.0-rc3 2025-11-03T16:58:02.3165161Z * [new tag] v2.3.0-rc4 -> v2.3.0-rc4 2025-11-03T16:58:02.3166520Z * [new tag] v2.3.0-rc5 -> v2.3.0-rc5 2025-11-03T16:58:02.3167679Z * [new tag] v2.3.0-rc6 -> v2.3.0-rc6 2025-11-03T16:58:02.3169052Z * [new tag] v2.3.0-rc7 -> v2.3.0-rc7 2025-11-03T16:58:02.3170398Z * [new tag] v2.3.0-rc8 -> v2.3.0-rc8 2025-11-03T16:58:02.3171546Z * [new tag] v2.3.0-rc9 -> v2.3.0-rc9 2025-11-03T16:58:02.3172649Z * [new tag] v2.3.1 -> v2.3.1 2025-11-03T16:58:02.3173984Z * [new tag] v2.3.1-rc1 -> v2.3.1-rc1 2025-11-03T16:58:02.3175367Z * [new tag] v2.3.1-rc2 -> v2.3.1-rc2 2025-11-03T16:58:02.3176717Z * [new tag] v2.3.1-rc3 -> v2.3.1-rc3 2025-11-03T16:58:02.3178067Z * [new tag] v2.4.0 -> v2.4.0 2025-11-03T16:58:02.3179458Z * [new tag] v2.4.0-rc1 -> v2.4.0-rc1 2025-11-03T16:58:02.3180827Z * [new tag] v2.4.0-rc2 -> v2.4.0-rc2 2025-11-03T16:58:02.3182389Z * [new tag] v2.4.0-rc3 -> v2.4.0-rc3 2025-11-03T16:58:02.3183742Z * [new tag] v2.4.0-rc4 -> v2.4.0-rc4 2025-11-03T16:58:02.3185117Z * [new tag] v2.4.0-rc5 -> v2.4.0-rc5 2025-11-03T16:58:02.3186484Z * [new tag] v2.4.0-rc6 -> v2.4.0-rc6 2025-11-03T16:58:02.3187849Z * [new tag] v2.4.0-rc7 -> v2.4.0-rc7 2025-11-03T16:58:02.3189154Z * [new tag] v2.4.0-rc8 -> v2.4.0-rc8 2025-11-03T16:58:02.3190568Z * [new tag] v2.4.0-rc9 -> v2.4.0-rc9 2025-11-03T16:58:02.3191711Z * [new tag] v2.4.1 -> v2.4.1 2025-11-03T16:58:02.3193045Z * [new tag] v2.4.1-rc1 -> v2.4.1-rc1 2025-11-03T16:58:02.3194433Z * [new tag] v2.4.1-rc2 -> v2.4.1-rc2 2025-11-03T16:58:02.3196006Z * [new tag] v2.4.1-rc3 -> v2.4.1-rc3 2025-11-03T16:58:02.3197261Z * [new tag] v2.5.0 -> v2.5.0 2025-11-03T16:58:02.3198621Z * [new tag] v2.5.0-rc1 -> v2.5.0-rc1 2025-11-03T16:58:02.3199769Z * [new tag] v2.5.0-rc10 -> v2.5.0-rc10 2025-11-03T16:58:02.3201091Z * [new tag] v2.5.0-rc2 -> v2.5.0-rc2 2025-11-03T16:58:02.3202420Z * [new tag] v2.5.0-rc3 -> v2.5.0-rc3 2025-11-03T16:58:02.3203799Z * [new tag] v2.5.0-rc4 -> v2.5.0-rc4 2025-11-03T16:58:02.3205141Z * [new tag] v2.5.0-rc5 -> v2.5.0-rc5 2025-11-03T16:58:02.3206512Z * [new tag] v2.5.0-rc6 -> v2.5.0-rc6 2025-11-03T16:58:02.3207926Z * [new tag] v2.5.0-rc7 -> v2.5.0-rc7 2025-11-03T16:58:02.3209308Z * [new tag] v2.5.0-rc8 -> v2.5.0-rc8 2025-11-03T16:58:02.3210660Z * [new tag] v2.5.0-rc9 -> v2.5.0-rc9 2025-11-03T16:58:02.3211852Z * [new tag] v2.5.1 -> v2.5.1 2025-11-03T16:58:02.3212982Z * [new tag] v2.5.1-rc1 -> v2.5.1-rc1 2025-11-03T16:58:02.3214173Z * [new tag] v2.6.0 -> v2.6.0 2025-11-03T16:58:02.3215531Z * [new tag] v2.6.0-rc1 -> v2.6.0-rc1 2025-11-03T16:58:02.3216971Z * [new tag] v2.6.0-rc2 -> v2.6.0-rc2 2025-11-03T16:58:02.3218306Z * [new tag] v2.6.0-rc3 -> v2.6.0-rc3 2025-11-03T16:58:02.3219626Z * [new tag] v2.6.0-rc4 -> v2.6.0-rc4 2025-11-03T16:58:02.3221316Z * [new tag] v2.6.0-rc5 -> v2.6.0-rc5 2025-11-03T16:58:02.3223032Z * [new tag] v2.6.0-rc6 -> v2.6.0-rc6 2025-11-03T16:58:02.3224418Z * [new tag] v2.6.0-rc7 -> v2.6.0-rc7 2025-11-03T16:58:02.3225820Z * [new tag] v2.6.0-rc8 -> v2.6.0-rc8 2025-11-03T16:58:02.3227187Z * [new tag] v2.6.0-rc9 -> v2.6.0-rc9 2025-11-03T16:58:02.3228729Z * [new tag] v2.7.0 -> v2.7.0 2025-11-03T16:58:02.3230054Z * [new tag] v2.7.0-rc1 -> v2.7.0-rc1 2025-11-03T16:58:02.3231223Z * [new tag] v2.7.0-rc10 -> v2.7.0-rc10 2025-11-03T16:58:02.3232649Z * [new tag] v2.7.0-rc2 -> v2.7.0-rc2 2025-11-03T16:58:02.3234014Z * [new tag] v2.7.0-rc3 -> v2.7.0-rc3 2025-11-03T16:58:02.3235325Z * [new tag] v2.7.0-rc4 -> v2.7.0-rc4 2025-11-03T16:58:02.3236684Z * [new tag] v2.7.0-rc5 -> v2.7.0-rc5 2025-11-03T16:58:02.3238071Z * [new tag] v2.7.0-rc6 -> v2.7.0-rc6 2025-11-03T16:58:02.3239516Z * [new tag] v2.7.0-rc7 -> v2.7.0-rc7 2025-11-03T16:58:02.3241360Z * [new tag] v2.7.0-rc8 -> v2.7.0-rc8 2025-11-03T16:58:02.3242804Z * [new tag] v2.7.0-rc9 -> v2.7.0-rc9 2025-11-03T16:58:02.3243939Z * [new tag] v2.7.1 -> v2.7.1 2025-11-03T16:58:02.3245437Z * [new tag] v2.7.1-rc1 -> v2.7.1-rc1 2025-11-03T16:58:02.3246780Z * [new tag] v2.7.1-rc2 -> v2.7.1-rc2 2025-11-03T16:58:02.3248243Z * [new tag] v2.7.1-rc3 -> v2.7.1-rc3 2025-11-03T16:58:02.3249633Z * [new tag] v2.7.1-rc4 -> v2.7.1-rc4 2025-11-03T16:58:02.3251032Z * [new tag] v2.7.1-rc5 -> v2.7.1-rc5 2025-11-03T16:58:02.3252256Z * [new tag] v2.8.0 -> v2.8.0 2025-11-03T16:58:02.3253757Z * [new tag] v2.8.0-rc1 -> v2.8.0-rc1 2025-11-03T16:58:02.3255019Z * [new tag] v2.8.0-rc2 -> v2.8.0-rc2 2025-11-03T16:58:02.3256443Z * [new tag] v2.8.0-rc3 -> v2.8.0-rc3 2025-11-03T16:58:02.3257882Z * [new tag] v2.8.0-rc4 -> v2.8.0-rc4 2025-11-03T16:58:02.3259266Z * [new tag] v2.8.0-rc5 -> v2.8.0-rc5 2025-11-03T16:58:02.3260705Z * [new tag] v2.8.0-rc6 -> v2.8.0-rc6 2025-11-03T16:58:02.3262207Z * [new tag] v2.8.0-rc7 -> v2.8.0-rc7 2025-11-03T16:58:02.3263572Z * [new tag] v2.8.0-rc8 -> v2.8.0-rc8 2025-11-03T16:58:02.3264998Z * [new tag] v2.9.0 -> v2.9.0 2025-11-03T16:58:02.3266353Z * [new tag] v2.9.0-rc1 -> v2.9.0-rc1 2025-11-03T16:58:02.3267759Z * [new tag] v2.9.0-rc10 -> v2.9.0-rc10 2025-11-03T16:58:02.3268968Z * [new tag] v2.9.0-rc11 -> v2.9.0-rc11 2025-11-03T16:58:02.3270607Z * [new tag] v2.9.0-rc2 -> v2.9.0-rc2 2025-11-03T16:58:02.3271986Z * [new tag] v2.9.0-rc3 -> v2.9.0-rc3 2025-11-03T16:58:02.3273366Z * [new tag] v2.9.0-rc4 -> v2.9.0-rc4 2025-11-03T16:58:02.3274773Z * [new tag] v2.9.0-rc5 -> v2.9.0-rc5 2025-11-03T16:58:02.3276352Z * [new tag] v2.9.0-rc6 -> v2.9.0-rc6 2025-11-03T16:58:02.3277638Z * [new tag] v2.9.0-rc7 -> v2.9.0-rc7 2025-11-03T16:58:02.3279272Z * [new tag] v2.9.0-rc8 -> v2.9.0-rc8 2025-11-03T16:58:02.3280464Z * [new tag] v2.9.0-rc9 -> v2.9.0-rc9 2025-11-03T16:58:02.3282415Z * [new tag] viable/strict/1759343184 -> viable/strict/1759343184 2025-11-03T16:58:02.3283709Z * [new tag] viable/strict/1759346540 -> viable/strict/1759346540 2025-11-03T16:58:02.3284990Z * [new tag] viable/strict/1759348181 -> viable/strict/1759348181 2025-11-03T16:58:02.3286263Z * [new tag] viable/strict/1759350324 -> viable/strict/1759350324 2025-11-03T16:58:02.3287557Z * [new tag] viable/strict/1759351793 -> viable/strict/1759351793 2025-11-03T16:58:02.3288835Z * [new tag] viable/strict/1759353844 -> viable/strict/1759353844 2025-11-03T16:58:02.3290178Z * [new tag] viable/strict/1759355374 -> viable/strict/1759355374 2025-11-03T16:58:02.3291931Z * [new tag] viable/strict/1759357472 -> viable/strict/1759357472 2025-11-03T16:58:02.3293054Z * [new tag] viable/strict/1759361002 -> viable/strict/1759361002 2025-11-03T16:58:02.3294229Z * [new tag] viable/strict/1759362585 -> viable/strict/1759362585 2025-11-03T16:58:02.3295904Z * [new tag] viable/strict/1759365359 -> viable/strict/1759365359 2025-11-03T16:58:02.3297239Z * [new tag] viable/strict/1759370089 -> viable/strict/1759370089 2025-11-03T16:58:02.3298603Z * [new tag] viable/strict/1759377554 -> viable/strict/1759377554 2025-11-03T16:58:02.3299835Z * [new tag] viable/strict/1759379133 -> viable/strict/1759379133 2025-11-03T16:58:02.3301218Z * [new tag] viable/strict/1759389871 -> viable/strict/1759389871 2025-11-03T16:58:02.3302866Z * [new tag] viable/strict/1759393562 -> viable/strict/1759393562 2025-11-03T16:58:02.3304279Z * [new tag] viable/strict/1759395076 -> viable/strict/1759395076 2025-11-03T16:58:02.3305626Z * [new tag] viable/strict/1759398579 -> viable/strict/1759398579 2025-11-03T16:58:02.3307135Z * [new tag] viable/strict/1759404142 -> viable/strict/1759404142 2025-11-03T16:58:02.3308324Z * [new tag] viable/strict/1759405773 -> viable/strict/1759405773 2025-11-03T16:58:02.3309719Z * [new tag] viable/strict/1759408041 -> viable/strict/1759408041 2025-11-03T16:58:02.3311056Z * [new tag] viable/strict/1759411593 -> viable/strict/1759411593 2025-11-03T16:58:02.3312415Z * [new tag] viable/strict/1759427395 -> viable/strict/1759427395 2025-11-03T16:58:02.3313722Z * [new tag] viable/strict/1759434582 -> viable/strict/1759434582 2025-11-03T16:58:02.3315069Z * [new tag] viable/strict/1759436720 -> viable/strict/1759436720 2025-11-03T16:58:02.3316430Z * [new tag] viable/strict/1759440219 -> viable/strict/1759440219 2025-11-03T16:58:02.3317816Z * [new tag] viable/strict/1759441948 -> viable/strict/1759441948 2025-11-03T16:58:02.3319182Z * [new tag] viable/strict/1759443860 -> viable/strict/1759443860 2025-11-03T16:58:02.3320550Z * [new tag] viable/strict/1759445377 -> viable/strict/1759445377 2025-11-03T16:58:02.3322228Z * [new tag] viable/strict/1759447415 -> viable/strict/1759447415 2025-11-03T16:58:02.3323601Z * [new tag] viable/strict/1759451750 -> viable/strict/1759451750 2025-11-03T16:58:02.3324977Z * [new tag] viable/strict/1759453910 -> viable/strict/1759453910 2025-11-03T16:58:02.3326436Z * [new tag] viable/strict/1759456483 -> viable/strict/1759456483 2025-11-03T16:58:02.3327799Z * [new tag] viable/strict/1759459279 -> viable/strict/1759459279 2025-11-03T16:58:02.3329208Z * [new tag] viable/strict/1759460742 -> viable/strict/1759460742 2025-11-03T16:58:02.3330544Z * [new tag] viable/strict/1759462025 -> viable/strict/1759462025 2025-11-03T16:58:02.3331976Z * [new tag] viable/strict/1759469086 -> viable/strict/1759469086 2025-11-03T16:58:02.3333349Z * [new tag] viable/strict/1759470581 -> viable/strict/1759470581 2025-11-03T16:58:02.3334700Z * [new tag] viable/strict/1759472786 -> viable/strict/1759472786 2025-11-03T16:58:02.3336048Z * [new tag] viable/strict/1759476294 -> viable/strict/1759476294 2025-11-03T16:58:02.3337449Z * [new tag] viable/strict/1759479963 -> viable/strict/1759479963 2025-11-03T16:58:02.3338801Z * [new tag] viable/strict/1759492177 -> viable/strict/1759492177 2025-11-03T16:58:02.3340174Z * [new tag] viable/strict/1759519278 -> viable/strict/1759519278 2025-11-03T16:58:02.3341580Z * [new tag] viable/strict/1759524580 -> viable/strict/1759524580 2025-11-03T16:58:02.3342980Z * [new tag] viable/strict/1759528193 -> viable/strict/1759528193 2025-11-03T16:58:02.3344216Z * [new tag] viable/strict/1759533797 -> viable/strict/1759533797 2025-11-03T16:58:02.3345641Z * [new tag] viable/strict/1759542780 -> viable/strict/1759542780 2025-11-03T16:58:02.3347125Z * [new tag] viable/strict/1759549779 -> viable/strict/1759549779 2025-11-03T16:58:02.3348511Z * [new tag] viable/strict/1759555455 -> viable/strict/1759555455 2025-11-03T16:58:02.3349868Z * [new tag] viable/strict/1759559176 -> viable/strict/1759559176 2025-11-03T16:58:02.3351242Z * [new tag] viable/strict/1759560629 -> viable/strict/1759560629 2025-11-03T16:58:02.3352577Z * [new tag] viable/strict/1759569848 -> viable/strict/1759569848 2025-11-03T16:58:02.3354125Z * [new tag] viable/strict/1759571382 -> viable/strict/1759571382 2025-11-03T16:58:02.3355932Z * [new tag] viable/strict/1759573474 -> viable/strict/1759573474 2025-11-03T16:58:02.3357480Z * [new tag] viable/strict/1759618187 -> viable/strict/1759618187 2025-11-03T16:58:02.3358754Z * [new tag] viable/strict/1759626742 -> viable/strict/1759626742 2025-11-03T16:58:02.3360141Z * [new tag] viable/strict/1759632427 -> viable/strict/1759632427 2025-11-03T16:58:02.3361481Z * [new tag] viable/strict/1759634971 -> viable/strict/1759634971 2025-11-03T16:58:02.3362882Z * [new tag] viable/strict/1759661382 -> viable/strict/1759661382 2025-11-03T16:58:02.3364277Z * [new tag] viable/strict/1759663294 -> viable/strict/1759663294 2025-11-03T16:58:02.3365549Z * [new tag] viable/strict/1759708178 -> viable/strict/1759708178 2025-11-03T16:58:02.3366898Z * [new tag] viable/strict/1759715695 -> viable/strict/1759715695 2025-11-03T16:58:02.3368263Z * [new tag] viable/strict/1759728293 -> viable/strict/1759728293 2025-11-03T16:58:02.3369633Z * [new tag] viable/strict/1759735513 -> viable/strict/1759735513 2025-11-03T16:58:02.3371143Z * [new tag] viable/strict/1759739177 -> viable/strict/1759739177 2025-11-03T16:58:02.3372586Z * [new tag] viable/strict/1759758635 -> viable/strict/1759758635 2025-11-03T16:58:02.3374002Z * [new tag] viable/strict/1759765784 -> viable/strict/1759765784 2025-11-03T16:58:02.3375348Z * [new tag] viable/strict/1759767948 -> viable/strict/1759767948 2025-11-03T16:58:02.3376730Z * [new tag] viable/strict/1759771461 -> viable/strict/1759771461 2025-11-03T16:58:02.3378186Z * [new tag] viable/strict/1759776706 -> viable/strict/1759776706 2025-11-03T16:58:02.3379569Z * [new tag] viable/strict/1759782317 -> viable/strict/1759782317 2025-11-03T16:58:02.3380997Z * [new tag] viable/strict/1759783777 -> viable/strict/1759783777 2025-11-03T16:58:02.3382571Z * [new tag] viable/strict/1759785815 -> viable/strict/1759785815 2025-11-03T16:58:02.3383829Z * [new tag] viable/strict/1759789459 -> viable/strict/1759789459 2025-11-03T16:58:02.3385020Z * [new tag] viable/strict/1759790974 -> viable/strict/1759790974 2025-11-03T16:58:02.3386223Z * [new tag] viable/strict/1759794583 -> viable/strict/1759794583 2025-11-03T16:58:02.3387667Z * [new tag] viable/strict/1759797408 -> viable/strict/1759797408 2025-11-03T16:58:02.3389099Z * [new tag] viable/strict/1759799518 -> viable/strict/1759799518 2025-11-03T16:58:02.3390407Z * [new tag] viable/strict/1759804909 -> viable/strict/1759804909 2025-11-03T16:58:02.3391792Z * [new tag] viable/strict/1759807643 -> viable/strict/1759807643 2025-11-03T16:58:02.3392955Z * [new tag] viable/strict/1759809089 -> viable/strict/1759809089 2025-11-03T16:58:02.3394342Z * [new tag] viable/strict/1759811145 -> viable/strict/1759811145 2025-11-03T16:58:02.3395702Z * [new tag] viable/strict/1759812581 -> viable/strict/1759812581 2025-11-03T16:58:02.3397068Z * [new tag] viable/strict/1759814683 -> viable/strict/1759814683 2025-11-03T16:58:02.3398475Z * [new tag] viable/strict/1759821889 -> viable/strict/1759821889 2025-11-03T16:58:02.3399942Z * [new tag] viable/strict/1759823376 -> viable/strict/1759823376 2025-11-03T16:58:02.3401257Z * [new tag] viable/strict/1759827107 -> viable/strict/1759827107 2025-11-03T16:58:02.3402708Z * [new tag] viable/strict/1759830577 -> viable/strict/1759830577 2025-11-03T16:58:02.3404179Z * [new tag] viable/strict/1759832720 -> viable/strict/1759832720 2025-11-03T16:58:02.3405551Z * [new tag] viable/strict/1759842063 -> viable/strict/1759842063 2025-11-03T16:58:02.3406954Z * [new tag] viable/strict/1759847121 -> viable/strict/1759847121 2025-11-03T16:58:02.3408554Z * [new tag] viable/strict/1759850721 -> viable/strict/1759850721 2025-11-03T16:58:02.3409789Z * [new tag] viable/strict/1759857870 -> viable/strict/1759857870 2025-11-03T16:58:02.3411062Z * [new tag] viable/strict/1759863143 -> viable/strict/1759863143 2025-11-03T16:58:02.3412389Z * [new tag] viable/strict/1759875874 -> viable/strict/1759875874 2025-11-03T16:58:02.3413623Z * [new tag] viable/strict/1759877385 -> viable/strict/1759877385 2025-11-03T16:58:02.3415058Z * [new tag] viable/strict/1759883801 -> viable/strict/1759883801 2025-11-03T16:58:02.3416428Z * [new tag] viable/strict/1759885922 -> viable/strict/1759885922 2025-11-03T16:58:02.3417688Z * [new tag] viable/strict/1759888488 -> viable/strict/1759888488 2025-11-03T16:58:02.3419064Z * [new tag] viable/strict/1759895471 -> viable/strict/1759895471 2025-11-03T16:58:02.3420452Z * [new tag] viable/strict/1759904803 -> viable/strict/1759904803 2025-11-03T16:58:02.3422356Z * [new tag] viable/strict/1759908300 -> viable/strict/1759908300 2025-11-03T16:58:02.3423718Z * [new tag] viable/strict/1759915520 -> viable/strict/1759915520 2025-11-03T16:58:02.3425086Z * [new tag] viable/strict/1759916978 -> viable/strict/1759916978 2025-11-03T16:58:02.3426312Z * [new tag] viable/strict/1759930024 -> viable/strict/1759930024 2025-11-03T16:58:02.3427767Z * [new tag] viable/strict/1759948122 -> viable/strict/1759948122 2025-11-03T16:58:02.3429142Z * [new tag] viable/strict/1759952983 -> viable/strict/1759952983 2025-11-03T16:58:02.3430380Z * [new tag] viable/strict/1759955121 -> viable/strict/1759955121 2025-11-03T16:58:02.3431725Z * [new tag] viable/strict/1759962298 -> viable/strict/1759962298 2025-11-03T16:58:02.3432981Z * [new tag] viable/strict/1759965837 -> viable/strict/1759965837 2025-11-03T16:58:02.3434343Z * [new tag] viable/strict/1759970213 -> viable/strict/1759970213 2025-11-03T16:58:02.3435688Z * [new tag] viable/strict/1759974894 -> viable/strict/1759974894 2025-11-03T16:58:02.3437043Z * [new tag] viable/strict/1759977763 -> viable/strict/1759977763 2025-11-03T16:58:02.3438501Z * [new tag] viable/strict/1759979241 -> viable/strict/1759979241 2025-11-03T16:58:02.3439878Z * [new tag] viable/strict/1759985417 -> viable/strict/1759985417 2025-11-03T16:58:02.3441244Z * [new tag] viable/strict/1759987490 -> viable/strict/1759987490 2025-11-03T16:58:02.3442647Z * [new tag] viable/strict/1759996180 -> viable/strict/1759996180 2025-11-03T16:58:02.3444489Z * [new tag] viable/strict/1760065682 -> viable/strict/1760065682 2025-11-03T16:58:02.3445882Z * [new tag] viable/strict/1760066894 -> viable/strict/1760066894 2025-11-03T16:58:02.3447321Z * [new tag] viable/strict/1760070345 -> viable/strict/1760070345 2025-11-03T16:58:02.3448691Z * [new tag] viable/strict/1760089782 -> viable/strict/1760089782 2025-11-03T16:58:02.3450114Z * [new tag] viable/strict/1760091921 -> viable/strict/1760091921 2025-11-03T16:58:02.3451463Z * [new tag] viable/strict/1760127924 -> viable/strict/1760127924 2025-11-03T16:58:02.3452825Z * [new tag] viable/strict/1760129489 -> viable/strict/1760129489 2025-11-03T16:58:02.3454284Z * [new tag] viable/strict/1760132980 -> viable/strict/1760132980 2025-11-03T16:58:02.3455704Z * [new tag] viable/strict/1760135060 -> viable/strict/1760135060 2025-11-03T16:58:02.3457218Z * [new tag] viable/strict/1760215782 -> viable/strict/1760215782 2025-11-03T16:58:02.3458476Z * [new tag] viable/strict/1760273849 -> viable/strict/1760273849 2025-11-03T16:58:02.3459880Z * [new tag] viable/strict/1760275517 -> viable/strict/1760275517 2025-11-03T16:58:02.3461291Z * [new tag] viable/strict/1760276979 -> viable/strict/1760276979 2025-11-03T16:58:02.3462739Z * [new tag] viable/strict/1760279007 -> viable/strict/1760279007 2025-11-03T16:58:02.3463946Z * [new tag] viable/strict/1760286328 -> viable/strict/1760286328 2025-11-03T16:58:02.3465162Z * [new tag] viable/strict/1760493304 -> viable/strict/1760493304 2025-11-03T16:58:02.3466412Z * [new tag] viable/strict/1760496298 -> viable/strict/1760496298 2025-11-03T16:58:02.3467746Z * [new tag] viable/strict/1760518396 -> viable/strict/1760518396 2025-11-03T16:58:02.3469121Z * [new tag] viable/strict/1760534864 -> viable/strict/1760534864 2025-11-03T16:58:02.3470408Z * [new tag] viable/strict/1760549062 -> viable/strict/1760549062 2025-11-03T16:58:02.3471769Z * [new tag] viable/strict/1760552799 -> viable/strict/1760552799 2025-11-03T16:58:02.3473144Z * [new tag] viable/strict/1760554355 -> viable/strict/1760554355 2025-11-03T16:58:02.3474566Z * [new tag] viable/strict/1760556275 -> viable/strict/1760556275 2025-11-03T16:58:02.3476022Z * [new tag] viable/strict/1760564979 -> viable/strict/1760564979 2025-11-03T16:58:02.3477398Z * [new tag] viable/strict/1760567049 -> viable/strict/1760567049 2025-11-03T16:58:02.3479170Z * [new tag] viable/strict/1760568585 -> viable/strict/1760568585 2025-11-03T16:58:02.3480511Z * [new tag] viable/strict/1760570630 -> viable/strict/1760570630 2025-11-03T16:58:02.3481919Z * [new tag] viable/strict/1760572180 -> viable/strict/1760572180 2025-11-03T16:58:02.3483266Z * [new tag] viable/strict/1760575094 -> viable/strict/1760575094 2025-11-03T16:58:02.3484684Z * [new tag] viable/strict/1760579709 -> viable/strict/1760579709 2025-11-03T16:58:02.3486432Z * [new tag] viable/strict/1760582614 -> viable/strict/1760582614 2025-11-03T16:58:02.3487931Z * [new tag] viable/strict/1760586815 -> viable/strict/1760586815 2025-11-03T16:58:02.3489059Z * [new tag] viable/strict/1760588829 -> viable/strict/1760588829 2025-11-03T16:58:02.3490420Z * [new tag] viable/strict/1760590200 -> viable/strict/1760590200 2025-11-03T16:58:02.3491837Z * [new tag] viable/strict/1760592311 -> viable/strict/1760592311 2025-11-03T16:58:02.3493239Z * [new tag] viable/strict/1760619733 -> viable/strict/1760619733 2025-11-03T16:58:02.3494444Z * [new tag] viable/strict/1760628335 -> viable/strict/1760628335 2025-11-03T16:58:02.3495825Z * [new tag] viable/strict/1760635490 -> viable/strict/1760635490 2025-11-03T16:58:02.3497182Z * [new tag] viable/strict/1760640743 -> viable/strict/1760640743 2025-11-03T16:58:02.3498527Z * [new tag] viable/strict/1760642528 -> viable/strict/1760642528 2025-11-03T16:58:02.3499852Z * [new tag] viable/strict/1760646330 -> viable/strict/1760646330 2025-11-03T16:58:02.3501253Z * [new tag] viable/strict/1760666101 -> viable/strict/1760666101 2025-11-03T16:58:02.3502863Z * [new tag] viable/strict/1760668990 -> viable/strict/1760668990 2025-11-03T16:58:02.3504325Z * [new tag] viable/strict/1760670600 -> viable/strict/1760670600 2025-11-03T16:58:02.3505654Z * [new tag] viable/strict/1760671704 -> viable/strict/1760671704 2025-11-03T16:58:02.3507125Z * [new tag] viable/strict/1760673121 -> viable/strict/1760673121 2025-11-03T16:58:02.3508421Z * [new tag] viable/strict/1760675352 -> viable/strict/1760675352 2025-11-03T16:58:02.3509770Z * [new tag] viable/strict/1760696731 -> viable/strict/1760696731 2025-11-03T16:58:02.3512282Z * [new tag] viable/strict/1760723515 -> viable/strict/1760723515 2025-11-03T16:58:02.3513641Z * [new tag] viable/strict/1760727234 -> viable/strict/1760727234 2025-11-03T16:58:02.3515101Z * [new tag] viable/strict/1760730578 -> viable/strict/1760730578 2025-11-03T16:58:02.3516328Z * [new tag] viable/strict/1760732726 -> viable/strict/1760732726 2025-11-03T16:58:02.3517688Z * [new tag] viable/strict/1760734180 -> viable/strict/1760734180 2025-11-03T16:58:02.3519069Z * [new tag] viable/strict/1760736251 -> viable/strict/1760736251 2025-11-03T16:58:02.3520488Z * [new tag] viable/strict/1760737772 -> viable/strict/1760737772 2025-11-03T16:58:02.3522196Z * [new tag] viable/strict/1760758005 -> viable/strict/1760758005 2025-11-03T16:58:02.3523564Z * [new tag] viable/strict/1760761532 -> viable/strict/1760761532 2025-11-03T16:58:02.3524957Z * [new tag] viable/strict/1760802581 -> viable/strict/1760802581 2025-11-03T16:58:02.3526289Z * [new tag] viable/strict/1760827772 -> viable/strict/1760827772 2025-11-03T16:58:02.3527663Z * [new tag] viable/strict/1760834524 -> viable/strict/1760834524 2025-11-03T16:58:02.3528885Z * [new tag] viable/strict/1760845009 -> viable/strict/1760845009 2025-11-03T16:58:02.3530274Z * [new tag] viable/strict/1760876836 -> viable/strict/1760876836 2025-11-03T16:58:02.3531670Z * [new tag] viable/strict/1760880329 -> viable/strict/1760880329 2025-11-03T16:58:02.3533059Z * [new tag] viable/strict/1760888987 -> viable/strict/1760888987 2025-11-03T16:58:02.3534855Z * [new tag] viable/strict/1760912664 -> viable/strict/1760912664 2025-11-03T16:58:02.3536254Z * [new tag] viable/strict/1760925321 -> viable/strict/1760925321 2025-11-03T16:58:02.3537636Z * [new tag] viable/strict/1760931488 -> viable/strict/1760931488 2025-11-03T16:58:02.3539018Z * [new tag] viable/strict/1760932693 -> viable/strict/1760932693 2025-11-03T16:58:02.3540386Z * [new tag] viable/strict/1761004184 -> viable/strict/1761004184 2025-11-03T16:58:02.3541849Z * [new tag] viable/strict/1761014748 -> viable/strict/1761014748 2025-11-03T16:58:02.3543075Z * [new tag] viable/strict/1761017491 -> viable/strict/1761017491 2025-11-03T16:58:02.3544485Z * [new tag] viable/strict/1761018806 -> viable/strict/1761018806 2025-11-03T16:58:02.3545776Z * [new tag] viable/strict/1761020754 -> viable/strict/1761020754 2025-11-03T16:58:02.3547173Z * [new tag] viable/strict/1761024303 -> viable/strict/1761024303 2025-11-03T16:58:02.3548522Z * [new tag] viable/strict/1761029582 -> viable/strict/1761029582 2025-11-03T16:58:02.3549988Z * [new tag] viable/strict/1761031535 -> viable/strict/1761031535 2025-11-03T16:58:02.3551359Z * [new tag] viable/strict/1761035196 -> viable/strict/1761035196 2025-11-03T16:58:02.3552977Z * [new tag] viable/strict/1761045825 -> viable/strict/1761045825 2025-11-03T16:58:02.3554411Z * [new tag] viable/strict/1761054796 -> viable/strict/1761054796 2025-11-03T16:58:02.3555821Z * [new tag] viable/strict/1761060314 -> viable/strict/1761060314 2025-11-03T16:58:02.3557203Z * [new tag] viable/strict/1761071198 -> viable/strict/1761071198 2025-11-03T16:58:02.3558764Z * [new tag] viable/strict/1761074628 -> viable/strict/1761074628 2025-11-03T16:58:02.3560109Z * [new tag] viable/strict/1761078351 -> viable/strict/1761078351 2025-11-03T16:58:02.3561475Z * [new tag] viable/strict/1761079822 -> viable/strict/1761079822 2025-11-03T16:58:02.3562687Z * [new tag] viable/strict/1761081873 -> viable/strict/1761081873 2025-11-03T16:58:02.3564049Z * [new tag] viable/strict/1761083392 -> viable/strict/1761083392 2025-11-03T16:58:02.3565499Z * [new tag] viable/strict/1761085465 -> viable/strict/1761085465 2025-11-03T16:58:02.3566946Z * [new tag] viable/strict/1761089099 -> viable/strict/1761089099 2025-11-03T16:58:02.3568317Z * [new tag] viable/strict/1761095535 -> viable/strict/1761095535 2025-11-03T16:58:02.3569656Z * [new tag] viable/strict/1761098119 -> viable/strict/1761098119 2025-11-03T16:58:02.3571424Z * [new tag] viable/strict/1761101330 -> viable/strict/1761101330 2025-11-03T16:58:02.3572901Z * [new tag] viable/strict/1761114425 -> viable/strict/1761114425 2025-11-03T16:58:02.3574210Z * [new tag] viable/strict/1761116036 -> viable/strict/1761116036 2025-11-03T16:58:02.3575612Z * [new tag] viable/strict/1761119379 -> viable/strict/1761119379 2025-11-03T16:58:02.3577021Z * [new tag] viable/strict/1761121601 -> viable/strict/1761121601 2025-11-03T16:58:02.3578339Z * [new tag] viable/strict/1761123234 -> viable/strict/1761123234 2025-11-03T16:58:02.3579735Z * [new tag] viable/strict/1761126621 -> viable/strict/1761126621 2025-11-03T16:58:02.3581100Z * [new tag] viable/strict/1761132259 -> viable/strict/1761132259 2025-11-03T16:58:02.3582707Z * [new tag] viable/strict/1761146746 -> viable/strict/1761146746 2025-11-03T16:58:02.3584052Z * [new tag] viable/strict/1761164752 -> viable/strict/1761164752 2025-11-03T16:58:02.3585372Z * [new tag] viable/strict/1761166198 -> viable/strict/1761166198 2025-11-03T16:58:02.3586775Z * [new tag] viable/strict/1761175424 -> viable/strict/1761175424 2025-11-03T16:58:02.3588248Z * [new tag] viable/strict/1761176983 -> viable/strict/1761176983 2025-11-03T16:58:02.3589676Z * [new tag] viable/strict/1761179891 -> viable/strict/1761179891 2025-11-03T16:58:02.3590943Z * [new tag] viable/strict/1761181930 -> viable/strict/1761181930 2025-11-03T16:58:02.3592335Z * [new tag] viable/strict/1761184516 -> viable/strict/1761184516 2025-11-03T16:58:02.3593827Z * [new tag] viable/strict/1761190179 -> viable/strict/1761190179 2025-11-03T16:58:02.3595206Z * [new tag] viable/strict/1761193558 -> viable/strict/1761193558 2025-11-03T16:58:02.3596581Z * [new tag] viable/strict/1761207990 -> viable/strict/1761207990 2025-11-03T16:58:02.3597972Z * [new tag] viable/strict/1761229539 -> viable/strict/1761229539 2025-11-03T16:58:02.3599568Z * [new tag] viable/strict/1761244031 -> viable/strict/1761244031 2025-11-03T16:58:02.3600941Z * [new tag] viable/strict/1761248986 -> viable/strict/1761248986 2025-11-03T16:58:02.3602180Z * [new tag] viable/strict/1761259791 -> viable/strict/1761259791 2025-11-03T16:58:02.3603560Z * [new tag] viable/strict/1761266139 -> viable/strict/1761266139 2025-11-03T16:58:02.3605063Z * [new tag] viable/strict/1761268316 -> viable/strict/1761268316 2025-11-03T16:58:02.3606297Z * [new tag] viable/strict/1761273805 -> viable/strict/1761273805 2025-11-03T16:58:02.3607783Z * [new tag] viable/strict/1761275261 -> viable/strict/1761275261 2025-11-03T16:58:02.3609156Z * [new tag] viable/strict/1761277913 -> viable/strict/1761277913 2025-11-03T16:58:02.3610563Z * [new tag] viable/strict/1761290701 -> viable/strict/1761290701 2025-11-03T16:58:02.3611948Z * [new tag] viable/strict/1761294396 -> viable/strict/1761294396 2025-11-03T16:58:02.3613415Z * [new tag] viable/strict/1761303047 -> viable/strict/1761303047 2025-11-03T16:58:02.3614801Z * [new tag] viable/strict/1761335388 -> viable/strict/1761335388 2025-11-03T16:58:02.3616163Z * [new tag] viable/strict/1761337551 -> viable/strict/1761337551 2025-11-03T16:58:02.3617557Z * [new tag] viable/strict/1761339007 -> viable/strict/1761339007 2025-11-03T16:58:02.3618801Z * [new tag] viable/strict/1761341050 -> viable/strict/1761341050 2025-11-03T16:58:02.3620212Z * [new tag] viable/strict/1761346188 -> viable/strict/1761346188 2025-11-03T16:58:02.3621875Z * [new tag] viable/strict/1761349792 -> viable/strict/1761349792 2025-11-03T16:58:02.3623713Z * [new tag] viable/strict/1761352620 -> viable/strict/1761352620 2025-11-03T16:58:02.3625431Z * [new tag] viable/strict/1761354730 -> viable/strict/1761354730 2025-11-03T16:58:02.3626845Z * [new tag] viable/strict/1761357298 -> viable/strict/1761357298 2025-11-03T16:58:02.3628287Z * [new tag] viable/strict/1761360201 -> viable/strict/1761360201 2025-11-03T16:58:02.3629659Z * [new tag] viable/strict/1761361753 -> viable/strict/1761361753 2025-11-03T16:58:02.3631028Z * [new tag] viable/strict/1761364351 -> viable/strict/1761364351 2025-11-03T16:58:02.3632471Z * [new tag] viable/strict/1761366338 -> viable/strict/1761366338 2025-11-03T16:58:02.3633983Z * [new tag] viable/strict/1761367802 -> viable/strict/1761367802 2025-11-03T16:58:02.3635359Z * [new tag] viable/strict/1761369889 -> viable/strict/1761369889 2025-11-03T16:58:02.3636630Z * [new tag] viable/strict/1761371385 -> viable/strict/1761371385 2025-11-03T16:58:02.3638078Z * [new tag] viable/strict/1761373581 -> viable/strict/1761373581 2025-11-03T16:58:02.3639559Z * [new tag] viable/strict/1761375054 -> viable/strict/1761375054 2025-11-03T16:58:02.3641052Z * [new tag] viable/strict/1761421785 -> viable/strict/1761421785 2025-11-03T16:58:02.3642515Z * [new tag] viable/strict/1761434614 -> viable/strict/1761434614 2025-11-03T16:58:02.3644176Z * [new tag] viable/strict/1761439254 -> viable/strict/1761439254 2025-11-03T16:58:02.3645446Z * [new tag] viable/strict/1761454187 -> viable/strict/1761454187 2025-11-03T16:58:02.3646898Z * [new tag] viable/strict/1761459991 -> viable/strict/1761459991 2025-11-03T16:58:02.3648333Z * [new tag] viable/strict/1761470668 -> viable/strict/1761470668 2025-11-03T16:58:02.3650085Z * [new tag] viable/strict/1761472188 -> viable/strict/1761472188 2025-11-03T16:58:02.3651503Z * [new tag] viable/strict/1761503178 -> viable/strict/1761503178 2025-11-03T16:58:02.3652906Z * [new tag] viable/strict/1761517492 -> viable/strict/1761517492 2025-11-03T16:58:02.3654295Z * [new tag] viable/strict/1761518981 -> viable/strict/1761518981 2025-11-03T16:58:02.3655610Z * [new tag] viable/strict/1761533609 -> viable/strict/1761533609 2025-11-03T16:58:02.3656850Z * [new tag] viable/strict/1761546438 -> viable/strict/1761546438 2025-11-03T16:58:02.3658167Z * [new tag] viable/strict/1761548133 -> viable/strict/1761548133 2025-11-03T16:58:02.3659629Z * [new tag] viable/strict/1761555186 -> viable/strict/1761555186 2025-11-03T16:58:02.3660795Z * [new tag] viable/strict/1761557178 -> viable/strict/1761557178 2025-11-03T16:58:02.3662303Z * [new tag] viable/strict/1761560772 -> viable/strict/1761560772 2025-11-03T16:58:02.3663549Z * [new tag] viable/strict/1761562266 -> viable/strict/1761562266 2025-11-03T16:58:02.3664822Z * [new tag] viable/strict/1761564260 -> viable/strict/1761564260 2025-11-03T16:58:02.3666113Z * [new tag] viable/strict/1761568072 -> viable/strict/1761568072 2025-11-03T16:58:02.3667353Z * [new tag] viable/strict/1761571683 -> viable/strict/1761571683 2025-11-03T16:58:02.3668651Z * [new tag] viable/strict/1761580199 -> viable/strict/1761580199 2025-11-03T16:58:02.3669899Z * [new tag] viable/strict/1761587383 -> viable/strict/1761587383 2025-11-03T16:58:02.3671236Z * [new tag] viable/strict/1761591165 -> viable/strict/1761591165 2025-11-03T16:58:02.3672539Z * [new tag] viable/strict/1761594575 -> viable/strict/1761594575 2025-11-03T16:58:02.3673753Z * [new tag] viable/strict/1761596710 -> viable/strict/1761596710 2025-11-03T16:58:02.3675005Z * [new tag] viable/strict/1761598189 -> viable/strict/1761598189 2025-11-03T16:58:02.3676297Z * [new tag] viable/strict/1761600254 -> viable/strict/1761600254 2025-11-03T16:58:02.3677540Z * [new tag] viable/strict/1761603879 -> viable/strict/1761603879 2025-11-03T16:58:02.3678806Z * [new tag] viable/strict/1761605429 -> viable/strict/1761605429 2025-11-03T16:58:02.3680058Z * [new tag] viable/strict/1761607468 -> viable/strict/1761607468 2025-11-03T16:58:02.3681320Z * [new tag] viable/strict/1761608983 -> viable/strict/1761608983 2025-11-03T16:58:02.3682619Z * [new tag] viable/strict/1761611846 -> viable/strict/1761611846 2025-11-03T16:58:02.3683923Z * [new tag] viable/strict/1761613922 -> viable/strict/1761613922 2025-11-03T16:58:02.3685187Z * [new tag] viable/strict/1761616504 -> viable/strict/1761616504 2025-11-03T16:58:02.3686564Z * [new tag] viable/strict/1761619599 -> viable/strict/1761619599 2025-11-03T16:58:02.3687821Z * [new tag] viable/strict/1761686693 -> viable/strict/1761686693 2025-11-03T16:58:02.3689065Z * [new tag] viable/strict/1761688179 -> viable/strict/1761688179 2025-11-03T16:58:02.3690332Z * [new tag] viable/strict/1761691973 -> viable/strict/1761691973 2025-11-03T16:58:02.3691605Z * [new tag] viable/strict/1761693884 -> viable/strict/1761693884 2025-11-03T16:58:02.3692792Z * [new tag] viable/strict/1761695389 -> viable/strict/1761695389 2025-11-03T16:58:02.3694086Z * [new tag] viable/strict/1761698408 -> viable/strict/1761698408 2025-11-03T16:58:02.3695321Z * [new tag] viable/strict/1761702931 -> viable/strict/1761702931 2025-11-03T16:58:02.3696633Z * [new tag] viable/strict/1761706307 -> viable/strict/1761706307 2025-11-03T16:58:02.3697926Z * [new tag] viable/strict/1761709065 -> viable/strict/1761709065 2025-11-03T16:58:02.3699231Z * [new tag] viable/strict/1761710285 -> viable/strict/1761710285 2025-11-03T16:58:02.3700544Z * [new tag] viable/strict/1761711983 -> viable/strict/1761711983 2025-11-03T16:58:02.3701952Z * [new tag] viable/strict/1761713514 -> viable/strict/1761713514 2025-11-03T16:58:02.3703270Z * [new tag] viable/strict/1761715523 -> viable/strict/1761715523 2025-11-03T16:58:02.3704569Z * [new tag] viable/strict/1761727973 -> viable/strict/1761727973 2025-11-03T16:58:02.3706046Z * [new tag] viable/strict/1761751558 -> viable/strict/1761751558 2025-11-03T16:58:02.3707317Z * [new tag] viable/strict/1761755187 -> viable/strict/1761755187 2025-11-03T16:58:02.3708585Z * [new tag] viable/strict/1761756826 -> viable/strict/1761756826 2025-11-03T16:58:02.3709874Z * [new tag] viable/strict/1761769551 -> viable/strict/1761769551 2025-11-03T16:58:02.3711697Z * [new tag] viable/strict/1761771032 -> viable/strict/1761771032 2025-11-03T16:58:02.3713198Z * [new tag] viable/strict/1761773101 -> viable/strict/1761773101 2025-11-03T16:58:02.3714702Z * [new tag] viable/strict/1761781792 -> viable/strict/1761781792 2025-11-03T16:58:02.3715963Z * [new tag] viable/strict/1761784788 -> viable/strict/1761784788 2025-11-03T16:58:02.3717319Z * [new tag] viable/strict/1761786740 -> viable/strict/1761786740 2025-11-03T16:58:02.3718626Z * [new tag] viable/strict/1761789332 -> viable/strict/1761789332 2025-11-03T16:58:02.3719942Z * [new tag] viable/strict/1761792569 -> viable/strict/1761792569 2025-11-03T16:58:02.3721337Z * [new tag] viable/strict/1761795289 -> viable/strict/1761795289 2025-11-03T16:58:02.3722931Z * [new tag] viable/strict/1761798345 -> viable/strict/1761798345 2025-11-03T16:58:02.3724212Z * [new tag] viable/strict/1761799827 -> viable/strict/1761799827 2025-11-03T16:58:02.3725536Z * [new tag] viable/strict/1761805604 -> viable/strict/1761805604 2025-11-03T16:58:02.3726914Z * [new tag] viable/strict/1761807202 -> viable/strict/1761807202 2025-11-03T16:58:02.3728276Z * [new tag] viable/strict/1761809094 -> viable/strict/1761809094 2025-11-03T16:58:02.3729555Z * [new tag] viable/strict/1761810576 -> viable/strict/1761810576 2025-11-03T16:58:02.3730851Z * [new tag] viable/strict/1761812771 -> viable/strict/1761812771 2025-11-03T16:58:02.3732143Z * [new tag] viable/strict/1761814363 -> viable/strict/1761814363 2025-11-03T16:58:02.3733481Z * [new tag] viable/strict/1761857410 -> viable/strict/1761857410 2025-11-03T16:58:02.3734747Z * [new tag] viable/strict/1761860985 -> viable/strict/1761860985 2025-11-03T16:58:02.3736070Z * [new tag] viable/strict/1761863094 -> viable/strict/1761863094 2025-11-03T16:58:02.3737368Z * [new tag] viable/strict/1761864590 -> viable/strict/1761864590 2025-11-03T16:58:02.3738725Z * [new tag] viable/strict/1761866675 -> viable/strict/1761866675 2025-11-03T16:58:02.3740008Z * [new tag] viable/strict/1761868178 -> viable/strict/1761868178 2025-11-03T16:58:02.3741294Z * [new tag] viable/strict/1761871111 -> viable/strict/1761871111 2025-11-03T16:58:02.3742764Z * [new tag] viable/strict/1761873126 -> viable/strict/1761873126 2025-11-03T16:58:02.3744040Z * [new tag] viable/strict/1761875714 -> viable/strict/1761875714 2025-11-03T16:58:02.3745370Z * [new tag] viable/strict/1761878924 -> viable/strict/1761878924 2025-11-03T16:58:02.3746669Z * [new tag] viable/strict/1761881727 -> viable/strict/1761881727 2025-11-03T16:58:02.3747960Z * [new tag] viable/strict/1761882959 -> viable/strict/1761882959 2025-11-03T16:58:02.3749297Z * [new tag] viable/strict/1761886268 -> viable/strict/1761886268 2025-11-03T16:58:02.3750587Z * [new tag] viable/strict/1761893641 -> viable/strict/1761893641 2025-11-03T16:58:02.3751875Z * [new tag] viable/strict/1761931517 -> viable/strict/1761931517 2025-11-03T16:58:02.3753136Z * [new tag] viable/strict/1761933080 -> viable/strict/1761933080 2025-11-03T16:58:02.3754600Z * [new tag] viable/strict/1761935217 -> viable/strict/1761935217 2025-11-03T16:58:02.3755882Z * [new tag] viable/strict/1761938533 -> viable/strict/1761938533 2025-11-03T16:58:02.3757122Z * [new tag] viable/strict/1761940184 -> viable/strict/1761940184 2025-11-03T16:58:02.3758441Z * [new tag] viable/strict/1761942338 -> viable/strict/1761942338 2025-11-03T16:58:02.3759750Z * [new tag] viable/strict/1761946100 -> viable/strict/1761946100 2025-11-03T16:58:02.3761051Z * [new tag] viable/strict/1761947374 -> viable/strict/1761947374 2025-11-03T16:58:02.3762328Z * [new tag] viable/strict/1761950978 -> viable/strict/1761950978 2025-11-03T16:58:02.3763713Z * [new tag] viable/strict/1761957727 -> viable/strict/1761957727 2025-11-03T16:58:02.3765018Z * [new tag] viable/strict/1761959532 -> viable/strict/1761959532 2025-11-03T16:58:02.3766329Z * [new tag] viable/strict/1761965366 -> viable/strict/1761965366 2025-11-03T16:58:02.3767726Z * [new tag] viable/strict/1761968066 -> viable/strict/1761968066 2025-11-03T16:58:02.3769103Z * [new tag] viable/strict/1761969322 -> viable/strict/1761969322 2025-11-03T16:58:02.3770406Z * [new tag] viable/strict/1761974723 -> viable/strict/1761974723 2025-11-03T16:58:02.3771713Z * [new tag] viable/strict/1761981837 -> viable/strict/1761981837 2025-11-03T16:58:02.3772932Z * [new tag] viable/strict/1761985546 -> viable/strict/1761985546 2025-11-03T16:58:02.3774256Z * [new tag] viable/strict/1761987030 -> viable/strict/1761987030 2025-11-03T16:58:02.3775605Z * [new tag] viable/strict/1762003554 -> viable/strict/1762003554 2025-11-03T16:58:02.3776919Z * [new tag] viable/strict/1762021560 -> viable/strict/1762021560 2025-11-03T16:58:02.3778218Z * [new tag] viable/strict/1762032190 -> viable/strict/1762032190 2025-11-03T16:58:02.3779516Z * [new tag] viable/strict/1762040981 -> viable/strict/1762040981 2025-11-03T16:58:02.3780809Z * [new tag] viable/strict/1762048525 -> viable/strict/1762048525 2025-11-03T16:58:02.3782225Z * [new tag] viable/strict/1762104223 -> viable/strict/1762104223 2025-11-03T16:58:02.3783596Z * [new tag] viable/strict/1762105778 -> viable/strict/1762105778 2025-11-03T16:58:02.3784889Z * [new tag] viable/strict/1762115109 -> viable/strict/1762115109 2025-11-03T16:58:02.3786209Z * [new tag] viable/strict/1762125840 -> viable/strict/1762125840 2025-11-03T16:58:02.3787505Z * [new tag] viable/strict/1762127377 -> viable/strict/1762127377 2025-11-03T16:58:02.3788810Z * [new tag] viable/strict/1762134925 -> viable/strict/1762134925 2025-11-03T16:58:02.3790107Z * [new tag] viable/strict/1762138338 -> viable/strict/1762138338 2025-11-03T16:58:02.3791422Z * [new tag] viable/strict/1762148993 -> viable/strict/1762148993 2025-11-03T16:58:02.3792719Z * [new tag] viable/strict/1762152871 -> viable/strict/1762152871 2025-11-03T16:58:02.3794015Z * [new tag] viable/strict/1762156183 -> viable/strict/1762156183 2025-11-03T16:58:02.3795394Z * [new tag] viable/strict/1762163457 -> viable/strict/1762163457 2025-11-03T16:58:02.3797104Z * [new tag] viable/strict/1762165569 -> viable/strict/1762165569 2025-11-03T16:58:02.3798406Z * [new tag] viable/strict/1762169035 -> viable/strict/1762169035 2025-11-03T16:58:02.3799711Z * [new tag] viable/strict/1762174936 -> viable/strict/1762174936 2025-11-03T16:58:02.3801160Z * [new tag] whc_flight_1 -> whc_flight_1 2025-11-03T16:58:02.3802716Z * [new tag] whc_flight_2 -> whc_flight_2 2025-11-03T16:58:02.3804170Z * [new tag] whc_flight_4 -> whc_flight_4 2025-11-03T16:58:02.4843026Z [command]/usr/bin/git rev-parse --verify --quiet 3f6538febd727b782e6e13cfd026a309fb14351d^{object} 2025-11-03T16:58:02.4870462Z 3f6538febd727b782e6e13cfd026a309fb14351d 2025-11-03T16:58:02.4875352Z ##[endgroup] 2025-11-03T16:58:02.4875653Z ##[group]Determining the checkout info 2025-11-03T16:58:02.4876050Z ##[endgroup] 2025-11-03T16:58:02.4879875Z [command]/usr/bin/git sparse-checkout disable 2025-11-03T16:58:02.4918215Z [command]/usr/bin/git config --local --unset-all extensions.worktreeConfig 2025-11-03T16:58:02.4947136Z ##[group]Checking out the ref 2025-11-03T16:58:02.4952372Z [command]/usr/bin/git checkout --progress --force 3f6538febd727b782e6e13cfd026a309fb14351d 2025-11-03T16:58:03.5241952Z Updating files: 73% (14728/20004) 2025-11-03T16:58:03.5505825Z Updating files: 74% (14803/20004) 2025-11-03T16:58:03.5965911Z Updating files: 75% (15003/20004) 2025-11-03T16:58:03.6160586Z Updating files: 76% (15204/20004) 2025-11-03T16:58:03.6318693Z Updating files: 77% (15404/20004) 2025-11-03T16:58:03.6493625Z Updating files: 78% (15604/20004) 2025-11-03T16:58:03.6783844Z Updating files: 79% (15804/20004) 2025-11-03T16:58:03.7080718Z Updating files: 80% (16004/20004) 2025-11-03T16:58:03.7369787Z Updating files: 81% (16204/20004) 2025-11-03T16:58:03.7615354Z Updating files: 82% (16404/20004) 2025-11-03T16:58:03.7802659Z Updating files: 83% (16604/20004) 2025-11-03T16:58:03.7960158Z Updating files: 84% (16804/20004) 2025-11-03T16:58:03.8142916Z Updating files: 85% (17004/20004) 2025-11-03T16:58:03.8319175Z Updating files: 86% (17204/20004) 2025-11-03T16:58:03.8476751Z Updating files: 87% (17404/20004) 2025-11-03T16:58:03.8622662Z Updating files: 88% (17604/20004) 2025-11-03T16:58:03.8776198Z Updating files: 89% (17804/20004) 2025-11-03T16:58:03.8964416Z Updating files: 90% (18004/20004) 2025-11-03T16:58:03.9112402Z Updating files: 91% (18204/20004) 2025-11-03T16:58:03.9273526Z Updating files: 92% (18404/20004) 2025-11-03T16:58:03.9476954Z Updating files: 93% (18604/20004) 2025-11-03T16:58:03.9685060Z Updating files: 94% (18804/20004) 2025-11-03T16:58:03.9883167Z Updating files: 95% (19004/20004) 2025-11-03T16:58:04.0060772Z Updating files: 96% (19204/20004) 2025-11-03T16:58:04.0246091Z Updating files: 97% (19404/20004) 2025-11-03T16:58:04.0517078Z Updating files: 98% (19604/20004) 2025-11-03T16:58:04.0708785Z Updating files: 99% (19804/20004) 2025-11-03T16:58:04.0709076Z Updating files: 100% (20004/20004) 2025-11-03T16:58:04.0709366Z Updating files: 100% (20004/20004), done. 2025-11-03T16:58:04.0940431Z Note: switching to '3f6538febd727b782e6e13cfd026a309fb14351d'. 2025-11-03T16:58:04.0940706Z 2025-11-03T16:58:04.0940919Z You are in 'detached HEAD' state. You can look around, make experimental 2025-11-03T16:58:04.0941408Z changes and commit them, and you can discard any commits you make in this 2025-11-03T16:58:04.0941998Z state without impacting any branches by switching back to a branch. 2025-11-03T16:58:04.0942313Z 2025-11-03T16:58:04.0942502Z If you want to create a new branch to retain commits you create, you may 2025-11-03T16:58:04.0942953Z do so (now or later) by using -c with the switch command. Example: 2025-11-03T16:58:04.0943202Z 2025-11-03T16:58:04.0943317Z git switch -c 2025-11-03T16:58:04.0943496Z 2025-11-03T16:58:04.0943603Z Or undo this operation with: 2025-11-03T16:58:04.0943773Z 2025-11-03T16:58:04.0943854Z git switch - 2025-11-03T16:58:04.0943995Z 2025-11-03T16:58:04.0944216Z Turn off this advice by setting config variable advice.detachedHead to false 2025-11-03T16:58:04.0944540Z 2025-11-03T16:58:04.0947084Z HEAD is now at 3f6538febd7 Remove tools from BC linter (#166858) 2025-11-03T16:58:04.1070224Z ##[endgroup] 2025-11-03T16:58:04.1070665Z ##[group]Setting up auth for fetching submodules 2025-11-03T16:58:04.1078410Z [command]/usr/bin/git config --global http.https://github.com/.extraheader AUTHORIZATION: basic *** 2025-11-03T16:58:04.1130944Z [command]/usr/bin/git config --global --unset-all url.https://github.com/.insteadOf 2025-11-03T16:58:04.1162589Z [command]/usr/bin/git config --global --add url.https://github.com/.insteadOf git@github.com: 2025-11-03T16:58:04.1195104Z [command]/usr/bin/git config --global --add url.https://github.com/.insteadOf org-21003710@github.com: 2025-11-03T16:58:04.1222379Z ##[endgroup] 2025-11-03T16:58:04.1223182Z ##[group]Fetching submodules 2025-11-03T16:58:04.1226476Z [command]/usr/bin/git submodule sync --recursive 2025-11-03T16:58:04.1610406Z [command]/usr/bin/git -c protocol.version=2 submodule update --init --force --recursive 2025-11-03T16:58:04.1979037Z Submodule 'android/libs/fbjni' (https://github.com/facebookincubator/fbjni.git) registered for path 'android/libs/fbjni' 2025-11-03T16:58:04.1981991Z Submodule 'third_party/NNPACK_deps/FP16' (https://github.com/Maratyszcza/FP16.git) registered for path 'third_party/FP16' 2025-11-03T16:58:04.1985791Z Submodule 'third_party/NNPACK_deps/FXdiv' (https://github.com/Maratyszcza/FXdiv.git) registered for path 'third_party/FXdiv' 2025-11-03T16:58:04.1989918Z Submodule 'third_party/NNPACK' (https://github.com/Maratyszcza/NNPACK.git) registered for path 'third_party/NNPACK' 2025-11-03T16:58:04.1993970Z Submodule 'third_party/NVTX' (https://github.com/NVIDIA/NVTX.git) registered for path 'third_party/NVTX' 2025-11-03T16:58:04.1998379Z Submodule 'third_party/VulkanMemoryAllocator' (https://github.com/GPUOpen-LibrariesAndSDKs/VulkanMemoryAllocator.git) registered for path 'third_party/VulkanMemoryAllocator' 2025-11-03T16:58:04.2002325Z Submodule 'third_party/XNNPACK' (https://github.com/google/XNNPACK.git) registered for path 'third_party/XNNPACK' 2025-11-03T16:58:04.2006575Z Submodule 'third_party/aiter' (https://github.com/ROCm/aiter.git) registered for path 'third_party/aiter' 2025-11-03T16:58:04.2010996Z Submodule 'third_party/benchmark' (https://github.com/google/benchmark.git) registered for path 'third_party/benchmark' 2025-11-03T16:58:04.2015581Z Submodule 'third_party/composable_kernel' (https://github.com/ROCm/composable_kernel.git) registered for path 'third_party/composable_kernel' 2025-11-03T16:58:04.2020107Z Submodule 'third_party/cpp-httplib' (https://github.com/yhirose/cpp-httplib.git) registered for path 'third_party/cpp-httplib' 2025-11-03T16:58:04.2025122Z Submodule 'third_party/cpuinfo' (https://github.com/pytorch/cpuinfo.git) registered for path 'third_party/cpuinfo' 2025-11-03T16:58:04.2029815Z Submodule 'third_party/cudnn_frontend' (https://github.com/NVIDIA/cudnn-frontend.git) registered for path 'third_party/cudnn_frontend' 2025-11-03T16:58:04.2034584Z Submodule 'third_party/cutlass' (https://github.com/NVIDIA/cutlass.git) registered for path 'third_party/cutlass' 2025-11-03T16:58:04.2039451Z Submodule 'third_party/fbgemm' (https://github.com/pytorch/fbgemm) registered for path 'third_party/fbgemm' 2025-11-03T16:58:04.2045318Z Submodule 'third_party/flash-attention' (https://github.com/Dao-AILab/flash-attention.git) registered for path 'third_party/flash-attention' 2025-11-03T16:58:04.2053457Z Submodule 'third_party/flatbuffers' (https://github.com/google/flatbuffers.git) registered for path 'third_party/flatbuffers' 2025-11-03T16:58:04.2058471Z Submodule 'third_party/fmt' (https://github.com/fmtlib/fmt.git) registered for path 'third_party/fmt' 2025-11-03T16:58:04.2063902Z Submodule 'third_party/gemmlowp/gemmlowp' (https://github.com/google/gemmlowp.git) registered for path 'third_party/gemmlowp/gemmlowp' 2025-11-03T16:58:04.2069022Z Submodule 'third_party/gloo' (https://github.com/pytorch/gloo) registered for path 'third_party/gloo' 2025-11-03T16:58:04.2075397Z Submodule 'third_party/googletest' (https://github.com/google/googletest.git) registered for path 'third_party/googletest' 2025-11-03T16:58:04.2080885Z Submodule 'third_party/ideep' (https://github.com/intel/ideep) registered for path 'third_party/ideep' 2025-11-03T16:58:04.2086454Z Submodule 'third_party/ittapi' (https://github.com/intel/ittapi.git) registered for path 'third_party/ittapi' 2025-11-03T16:58:04.2091783Z Submodule 'third_party/kineto' (https://github.com/pytorch/kineto) registered for path 'third_party/kineto' 2025-11-03T16:58:04.2097433Z Submodule 'third_party/kleidiai' (https://github.com/ARM-software/kleidiai.git) registered for path 'third_party/kleidiai' 2025-11-03T16:58:04.2103237Z Submodule 'third_party/mimalloc' (https://github.com/microsoft/mimalloc.git) registered for path 'third_party/mimalloc' 2025-11-03T16:58:04.2109066Z Submodule 'third_party/nlohmann' (https://github.com/nlohmann/json.git) registered for path 'third_party/nlohmann' 2025-11-03T16:58:04.2114864Z Submodule 'third_party/onnx' (https://github.com/onnx/onnx.git) registered for path 'third_party/onnx' 2025-11-03T16:58:04.2120950Z Submodule 'third_party/opentelemetry-cpp' (https://github.com/open-telemetry/opentelemetry-cpp.git) registered for path 'third_party/opentelemetry-cpp' 2025-11-03T16:58:04.2127671Z Submodule 'third_party/pocketfft' (https://github.com/mreineck/pocketfft) registered for path 'third_party/pocketfft' 2025-11-03T16:58:04.2133885Z Submodule 'third_party/protobuf' (https://github.com/protocolbuffers/protobuf.git) registered for path 'third_party/protobuf' 2025-11-03T16:58:04.2140015Z Submodule 'third_party/NNPACK_deps/psimd' (https://github.com/Maratyszcza/psimd.git) registered for path 'third_party/psimd' 2025-11-03T16:58:04.2146671Z Submodule 'third_party/NNPACK_deps/pthreadpool' (https://github.com/Maratyszcza/pthreadpool.git) registered for path 'third_party/pthreadpool' 2025-11-03T16:58:04.2156119Z Submodule 'third_party/pybind11' (https://github.com/pybind/pybind11.git) registered for path 'third_party/pybind11' 2025-11-03T16:58:04.2162695Z Submodule 'third_party/python-peachpy' (https://github.com/malfet/PeachPy.git) registered for path 'third_party/python-peachpy' 2025-11-03T16:58:04.2168950Z Submodule 'third_party/sleef' (https://github.com/shibatch/sleef) registered for path 'third_party/sleef' 2025-11-03T16:58:04.2175581Z Submodule 'third_party/tensorpipe' (https://github.com/pytorch/tensorpipe.git) registered for path 'third_party/tensorpipe' 2025-11-03T16:58:04.2211451Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/android/libs/fbjni'... 2025-11-03T16:58:04.4933168Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/FXdiv'... 2025-11-03T16:58:04.4934155Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/FP16'... 2025-11-03T16:58:04.4941282Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/fmt'... 2025-11-03T16:58:07.1706827Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/NNPACK'... 2025-11-03T16:58:07.1707816Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/benchmark'... 2025-11-03T16:58:07.1708711Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/NVTX'... 2025-11-03T16:58:07.1709587Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/gloo'... 2025-11-03T16:58:07.1710450Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/cpuinfo'... 2025-11-03T16:58:07.1711390Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/gemmlowp/gemmlowp'... 2025-11-03T16:58:07.1712345Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/cpp-httplib'... 2025-11-03T16:58:07.1713290Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/flash-attention'... 2025-11-03T16:58:07.1714212Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/ideep'... 2025-11-03T16:58:07.1715066Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/ittapi'... 2025-11-03T16:58:07.1716279Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/cudnn_frontend'... 2025-11-03T16:58:07.1717972Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/kleidiai'... 2025-11-03T16:58:07.1719139Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/pocketfft'... 2025-11-03T16:58:07.1720083Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/googletest'... 2025-11-03T16:58:07.1720996Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/mimalloc'... 2025-11-03T16:58:07.1722064Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/psimd'... 2025-11-03T16:58:07.1723031Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/VulkanMemoryAllocator'... 2025-11-03T16:58:07.1724022Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/pthreadpool'... 2025-11-03T16:58:07.2425459Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/flatbuffers'... 2025-11-03T16:58:07.8477636Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/python-peachpy'... 2025-11-03T16:58:07.8479955Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/tensorpipe'... 2025-11-03T16:58:07.8481425Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/kineto'... 2025-11-03T16:58:07.9478968Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/opentelemetry-cpp'... 2025-11-03T16:58:19.4133249Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/sleef'... 2025-11-03T16:58:19.4134224Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/pybind11'... 2025-11-03T16:58:19.4135123Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/cutlass'... 2025-11-03T16:58:19.4135999Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/fbgemm'... 2025-11-03T16:58:19.4136854Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/onnx'... 2025-11-03T16:58:19.4137784Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/composable_kernel'... 2025-11-03T16:58:19.4138759Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/aiter'... 2025-11-03T16:58:19.4139641Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/nlohmann'... 2025-11-03T16:58:19.5134843Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/XNNPACK'... 2025-11-03T16:58:23.1234851Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/protobuf'... 2025-11-03T16:58:23.1434814Z Submodule path 'android/libs/fbjni': checked out '7e1e1fe3858c63c251c637ae41a20de425dde96f' 2025-11-03T16:58:23.1595453Z Submodule path 'third_party/FP16': checked out '4dfe081cf6bcd15db339cf2680b9281b8451eeb3' 2025-11-03T16:58:23.1731954Z Submodule path 'third_party/FXdiv': checked out 'b408327ac2a15ec3e43352421954f5b1967701d1' 2025-11-03T16:58:23.2059267Z Submodule path 'third_party/NNPACK': checked out 'c07e3a0400713d546e0dea2d5466dd22ea389c73' 2025-11-03T16:58:23.2936591Z Submodule path 'third_party/NVTX': checked out '2942f167cc30c5e3a44a2aecd5b0d9c07ff61a07' 2025-11-03T16:58:23.3497995Z Submodule path 'third_party/VulkanMemoryAllocator': checked out '1d8f600fd424278486eade7ed3e877c99f0846b1' 2025-11-03T16:58:24.2733125Z Submodule path 'third_party/XNNPACK': checked out '51a0103656eff6fc9bfd39a4597923c4b542c883' 2025-11-03T16:58:24.4594658Z Submodule path 'third_party/aiter': checked out '01aae101b9e5e94d6c16a9514c9fb8df99c93150' 2025-11-03T16:58:24.4619584Z Submodule '3rdparty/composable_kernel' (https://github.com/ROCm/composable_kernel.git) registered for path 'third_party/aiter/3rdparty/composable_kernel' 2025-11-03T16:58:24.4656011Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/aiter/3rdparty/composable_kernel'... 2025-11-03T16:58:27.8531796Z Submodule path 'third_party/aiter/3rdparty/composable_kernel': checked out 'cffe8fa2a442ac8e80dd236a1a5d24fe3d7e0cbf' 2025-11-03T16:58:27.8833089Z Submodule path 'third_party/benchmark': checked out '299e5928955cc62af9968370293b916f5130916f' 2025-11-03T16:58:28.3083029Z Submodule path 'third_party/composable_kernel': checked out '7fe50dc3da2069d6645d9deb8c017a876472a977' 2025-11-03T16:58:28.3631820Z Submodule path 'third_party/cpp-httplib': checked out '89c932f313c6437c38f2982869beacc89c2f2246' 2025-11-03T16:58:28.4681888Z Submodule path 'third_party/cpuinfo': checked out '5e3d2445e6a84d9599bee2bf78edbb4d80865e1d' 2025-11-03T16:58:28.5226828Z Submodule path 'third_party/cudnn_frontend': checked out '0b1577c8c83401237d601d0d0db5210506705396' 2025-11-03T16:58:29.2599833Z Submodule path 'third_party/cutlass': checked out 'f3fde58372d33e9a5650ba7b80fc48b3b49d40c8' 2025-11-03T16:58:29.4408829Z Submodule path 'third_party/fbgemm': checked out 'c0b988d39a9e47c794d699f29930ed4d7c7e13a4' 2025-11-03T16:58:29.4435092Z Submodule 'external/asmjit' (https://github.com/asmjit/asmjit.git) registered for path 'third_party/fbgemm/external/asmjit' 2025-11-03T16:58:29.4438716Z Submodule 'external/composable_kernel' (https://github.com/ROCm/composable_kernel.git) registered for path 'third_party/fbgemm/external/composable_kernel' 2025-11-03T16:58:29.4442899Z Submodule 'external/cpuinfo' (https://github.com/pytorch/cpuinfo) registered for path 'third_party/fbgemm/external/cpuinfo' 2025-11-03T16:58:29.4446490Z Submodule 'external/cutlass' (https://github.com/jwfromm/cutlass) registered for path 'third_party/fbgemm/external/cutlass' 2025-11-03T16:58:29.4450339Z Submodule 'external/googletest' (https://github.com/google/googletest) registered for path 'third_party/fbgemm/external/googletest' 2025-11-03T16:58:29.4454415Z Submodule 'external/hipify_torch' (https://github.com/ROCmSoftwarePlatform/hipify_torch.git) registered for path 'third_party/fbgemm/external/hipify_torch' 2025-11-03T16:58:29.4458362Z Submodule 'external/json' (https://github.com/nlohmann/json.git) registered for path 'third_party/fbgemm/external/json' 2025-11-03T16:58:29.4494123Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/fbgemm/external/asmjit'... 2025-11-03T16:58:30.6898175Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/fbgemm/external/hipify_torch'... 2025-11-03T16:58:30.6898948Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/fbgemm/external/cpuinfo'... 2025-11-03T16:58:30.6899797Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/fbgemm/external/googletest'... 2025-11-03T16:58:30.7899430Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/fbgemm/external/composable_kernel'... 2025-11-03T16:58:33.0747675Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/fbgemm/external/cutlass'... 2025-11-03T16:58:33.1748521Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/fbgemm/external/json'... 2025-11-03T16:58:36.0899431Z Submodule path 'third_party/fbgemm/external/asmjit': checked out 'a3199e8857792cd10b7589ff5d58343d2c9008ea' 2025-11-03T16:58:36.5150238Z Submodule path 'third_party/fbgemm/external/composable_kernel': checked out '7fe50dc3da2069d6645d9deb8c017a876472a977' 2025-11-03T16:58:36.6215258Z Submodule path 'third_party/fbgemm/external/cpuinfo': checked out '6543fec09b2f04ac4a666882998b534afc9c1349' 2025-11-03T16:58:37.3664468Z Submodule path 'third_party/fbgemm/external/cutlass': checked out '98125ce499b0fdf7ffbe0e3052f5b8709f4840f8' 2025-11-03T16:58:37.4937129Z Submodule path 'third_party/fbgemm/external/googletest': checked out '52eb8108c5bdec04579160ae17225d66034bd723' 2025-11-03T16:58:37.5107656Z Submodule path 'third_party/fbgemm/external/hipify_torch': checked out '63b6a7b541fa7f08f8475ca7d74054db36ff2691' 2025-11-03T16:58:37.6319503Z Submodule path 'third_party/fbgemm/external/json': checked out '9cca280a4d0ccf0c08f47a99aa71d1b0e52f8d03' 2025-11-03T16:58:37.7198189Z Submodule path 'third_party/flash-attention': checked out '979702c87a8713a8e0a5e9fee122b90d2ef13be5' 2025-11-03T16:58:37.7223666Z Submodule 'csrc/composable_kernel' (https://github.com/ROCm/composable_kernel.git) registered for path 'third_party/flash-attention/csrc/composable_kernel' 2025-11-03T16:58:37.7226731Z Submodule 'csrc/cutlass' (https://github.com/NVIDIA/cutlass.git) registered for path 'third_party/flash-attention/csrc/cutlass' 2025-11-03T16:58:37.7258955Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/flash-attention/csrc/composable_kernel'... 2025-11-03T16:58:40.6758810Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/flash-attention/csrc/cutlass'... 2025-11-03T16:58:40.9853552Z Submodule path 'third_party/flash-attention/csrc/composable_kernel': checked out '888317e698e9803c62bd38568abc9e05d7709f33' 2025-11-03T16:58:41.6416908Z Submodule path 'third_party/flash-attention/csrc/cutlass': checked out 'c506e16788cb08416a4a57e11a9067beeee29420' 2025-11-03T16:58:41.8131588Z Submodule path 'third_party/flatbuffers': checked out 'a2cd1ea3b6d3fee220106b5fed3f7ce8da9eb757' 2025-11-03T16:58:41.8464976Z Submodule path 'third_party/fmt': checked out 'e424e3f2e607da02742f73db84873b8084fc714c' 2025-11-03T16:58:41.8895799Z Submodule path 'third_party/gemmlowp/gemmlowp': checked out '3fb5c176c17c765a3492cd2f0321b0dab712f350' 2025-11-03T16:58:41.9226310Z Submodule path 'third_party/gloo': checked out '54cbae0d3a67fa890b4c3d9ee162b7860315e341' 2025-11-03T16:58:41.9734553Z Submodule path 'third_party/googletest': checked out '52eb8108c5bdec04579160ae17225d66034bd723' 2025-11-03T16:58:41.9899211Z Submodule path 'third_party/ideep': checked out '719d8e6cd7f7a0e01b155657526d693acf97c2b3' 2025-11-03T16:58:41.9920260Z Submodule 'mkl-dnn' (https://github.com/intel/mkl-dnn.git) registered for path 'third_party/ideep/mkl-dnn' 2025-11-03T16:58:41.9953233Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/ideep/mkl-dnn'... 2025-11-03T16:58:56.9814417Z Submodule path 'third_party/ideep/mkl-dnn': checked out '8d263e693366ef8db40acc569cc7d8edf644556d' 2025-11-03T16:58:57.0087002Z Submodule path 'third_party/ittapi': checked out 'dec1d23ca65ab069d225dfe40dea14f455170959' 2025-11-03T16:58:57.0983325Z Submodule path 'third_party/kineto': checked out '6fcbc53d33dd275c0aba1e5d7701d471b7f6eeb3' 2025-11-03T16:58:57.1006955Z Submodule 'libkineto/third_party/dynolog' (https://github.com/facebookincubator/dynolog.git) registered for path 'third_party/kineto/libkineto/third_party/dynolog' 2025-11-03T16:58:57.1008943Z Submodule 'libkineto/third_party/fmt' (https://github.com/fmtlib/fmt.git) registered for path 'third_party/kineto/libkineto/third_party/fmt' 2025-11-03T16:58:57.1013101Z Submodule 'libkineto/third_party/googletest' (https://github.com/google/googletest.git) registered for path 'third_party/kineto/libkineto/third_party/googletest' 2025-11-03T16:58:57.1046616Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/kineto/libkineto/third_party/dynolog'... 2025-11-03T16:58:57.7946617Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/kineto/libkineto/third_party/fmt'... 2025-11-03T16:58:58.2712405Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/kineto/libkineto/third_party/googletest'... 2025-11-03T16:58:58.3761578Z Submodule path 'third_party/kineto/libkineto/third_party/dynolog': checked out 'd2ffe0a4e3acace628db49974246b66fc3e85fb1' 2025-11-03T16:58:58.3784834Z Submodule 'third_party/DCGM' (https://github.com/NVIDIA/DCGM.git) registered for path 'third_party/kineto/libkineto/third_party/dynolog/third_party/DCGM' 2025-11-03T16:58:58.3788619Z Submodule 'third_party/cpr' (https://github.com/libcpr/cpr.git) registered for path 'third_party/kineto/libkineto/third_party/dynolog/third_party/cpr' 2025-11-03T16:58:58.3792891Z Submodule 'third_party/fmt' (https://github.com/fmtlib/fmt.git) registered for path 'third_party/kineto/libkineto/third_party/dynolog/third_party/fmt' 2025-11-03T16:58:58.3797017Z Submodule 'third_party/gflags' (https://github.com/gflags/gflags.git) registered for path 'third_party/kineto/libkineto/third_party/dynolog/third_party/gflags' 2025-11-03T16:58:58.3801541Z Submodule 'third_party/glog' (https://github.com/google/glog.git) registered for path 'third_party/kineto/libkineto/third_party/dynolog/third_party/glog' 2025-11-03T16:58:58.3805591Z Submodule 'third_party/googletest' (https://github.com/google/googletest.git) registered for path 'third_party/kineto/libkineto/third_party/dynolog/third_party/googletest' 2025-11-03T16:58:58.3810140Z Submodule 'third_party/json' (https://github.com/nlohmann/json.git) registered for path 'third_party/kineto/libkineto/third_party/dynolog/third_party/json' 2025-11-03T16:58:58.3814668Z Submodule 'third_party/pfs' (https://github.com/dtrugman/pfs.git) registered for path 'third_party/kineto/libkineto/third_party/dynolog/third_party/pfs' 2025-11-03T16:58:58.3818949Z Submodule 'third_party/prometheus-cpp' (https://github.com/jupp0r/prometheus-cpp.git) registered for path 'third_party/kineto/libkineto/third_party/dynolog/third_party/prometheus-cpp' 2025-11-03T16:58:58.3855550Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/kineto/libkineto/third_party/dynolog/third_party/DCGM'... 2025-11-03T16:59:00.2603930Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/kineto/libkineto/third_party/dynolog/third_party/prometheus-cpp'... 2025-11-03T16:59:00.2605058Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/kineto/libkineto/third_party/dynolog/third_party/pfs'... 2025-11-03T16:59:00.2606043Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/kineto/libkineto/third_party/dynolog/third_party/gflags'... 2025-11-03T16:59:00.2607042Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/kineto/libkineto/third_party/dynolog/third_party/cpr'... 2025-11-03T16:59:00.2608013Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/kineto/libkineto/third_party/dynolog/third_party/glog'... 2025-11-03T16:59:00.2609039Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/kineto/libkineto/third_party/dynolog/third_party/googletest'... 2025-11-03T16:59:00.2609935Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/kineto/libkineto/third_party/dynolog/third_party/fmt'... 2025-11-03T16:59:00.3606293Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/kineto/libkineto/third_party/dynolog/third_party/json'... 2025-11-03T16:59:04.9056086Z Submodule path 'third_party/kineto/libkineto/third_party/dynolog/third_party/DCGM': checked out 'ffde4e54bc7249a6039a5e6b45b395141e1217f9' 2025-11-03T16:59:04.9301794Z Submodule path 'third_party/kineto/libkineto/third_party/dynolog/third_party/cpr': checked out '871ed52d350214a034f6ef8a3b8f51c5ce1bd400' 2025-11-03T16:59:04.9720886Z Submodule path 'third_party/kineto/libkineto/third_party/dynolog/third_party/fmt': checked out 'cd4af11efc9c622896a3e4cb599fa28668ca3d05' 2025-11-03T16:59:04.9904219Z Submodule path 'third_party/kineto/libkineto/third_party/dynolog/third_party/gflags': checked out 'e171aa2d15ed9eb17054558e0b3a6a413bb01067' 2025-11-03T16:59:04.9926264Z Submodule 'doc' (https://github.com/gflags/gflags.git) registered for path 'third_party/kineto/libkineto/third_party/dynolog/third_party/gflags/doc' 2025-11-03T16:59:04.9959736Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/kineto/libkineto/third_party/dynolog/third_party/gflags/doc'... 2025-11-03T16:59:05.2771537Z Submodule path 'third_party/kineto/libkineto/third_party/dynolog/third_party/gflags/doc': checked out '8411df715cf522606e3b1aca386ddfc0b63d34b4' 2025-11-03T16:59:05.3005383Z Submodule path 'third_party/kineto/libkineto/third_party/dynolog/third_party/glog': checked out 'b33e3bad4c46c8a6345525fd822af355e5ef9446' 2025-11-03T16:59:05.3509045Z Submodule path 'third_party/kineto/libkineto/third_party/dynolog/third_party/googletest': checked out '52eb8108c5bdec04579160ae17225d66034bd723' 2025-11-03T16:59:05.4636775Z Submodule path 'third_party/kineto/libkineto/third_party/dynolog/third_party/json': checked out '4f8fba14066156b73f1189a2b8bd568bde5284c5' 2025-11-03T16:59:05.4854192Z Submodule path 'third_party/kineto/libkineto/third_party/dynolog/third_party/pfs': checked out 'f68a2fa8ea36c783bdd760371411fcb495aa3150' 2025-11-03T16:59:05.5098463Z Submodule path 'third_party/kineto/libkineto/third_party/dynolog/third_party/prometheus-cpp': checked out 'b1234816facfdda29845c46696a02998a4af115a' 2025-11-03T16:59:05.5120028Z Submodule 'civetweb' (https://github.com/civetweb/civetweb.git) registered for path 'third_party/kineto/libkineto/third_party/dynolog/third_party/prometheus-cpp/3rdparty/civetweb' 2025-11-03T16:59:05.5124204Z Submodule 'googletest' (https://github.com/google/googletest.git) registered for path 'third_party/kineto/libkineto/third_party/dynolog/third_party/prometheus-cpp/3rdparty/googletest' 2025-11-03T16:59:05.5159271Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/kineto/libkineto/third_party/dynolog/third_party/prometheus-cpp/3rdparty/civetweb'... 2025-11-03T16:59:07.5169582Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/kineto/libkineto/third_party/dynolog/third_party/prometheus-cpp/3rdparty/googletest'... 2025-11-03T16:59:07.7744829Z Submodule path 'third_party/kineto/libkineto/third_party/dynolog/third_party/prometheus-cpp/3rdparty/civetweb': checked out 'd7ba35bbb649209c66e582d5a0244ba988a15159' 2025-11-03T16:59:07.8277175Z Submodule path 'third_party/kineto/libkineto/third_party/dynolog/third_party/prometheus-cpp/3rdparty/googletest': checked out 'e2239ee6043f73722e7aa812a459f54a28552929' 2025-11-03T16:59:07.8635768Z Submodule path 'third_party/kineto/libkineto/third_party/fmt': checked out '40626af88bd7df9a5fb80be7b25ac85b122d6c21' 2025-11-03T16:59:07.9134177Z Submodule path 'third_party/kineto/libkineto/third_party/googletest': checked out '52eb8108c5bdec04579160ae17225d66034bd723' 2025-11-03T16:59:07.9788027Z Submodule path 'third_party/kleidiai': checked out 'd7770c89632329a9914ef1a90289917597639cbe' 2025-11-03T16:59:08.0242923Z Submodule path 'third_party/mimalloc': checked out 'fbd8b99c2b828428947d70fdc046bb55609be93e' 2025-11-03T16:59:08.1561607Z Submodule path 'third_party/nlohmann': checked out '55f93686c01528224f448c19128836e7df245f72' 2025-11-03T16:59:08.7369137Z Submodule path 'third_party/onnx': checked out 'e709452ef2bbc1d113faf678c24e6d3467696e83' 2025-11-03T16:59:08.7409012Z Submodule 'third_party/pybind11' (https://github.com/pybind/pybind11.git) registered for path 'third_party/onnx/third_party/pybind11' 2025-11-03T16:59:08.7444220Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/onnx/third_party/pybind11'... 2025-11-03T16:59:09.7553035Z Submodule path 'third_party/onnx/third_party/pybind11': checked out 'a2e59f0e7065404b44dfe92a28aca47ba1378dc4' 2025-11-03T16:59:09.8475762Z Submodule path 'third_party/opentelemetry-cpp': checked out 'a799f4aed9c94b765dcdaabaeab7d5e7e2310878' 2025-11-03T16:59:09.8500764Z Submodule 'third_party/benchmark' (https://github.com/google/benchmark) registered for path 'third_party/opentelemetry-cpp/third_party/benchmark' 2025-11-03T16:59:09.8504529Z Submodule 'third_party/googletest' (https://github.com/google/googletest) registered for path 'third_party/opentelemetry-cpp/third_party/googletest' 2025-11-03T16:59:09.8508252Z Submodule 'third_party/ms-gsl' (https://github.com/microsoft/GSL) registered for path 'third_party/opentelemetry-cpp/third_party/ms-gsl' 2025-11-03T16:59:09.8512106Z Submodule 'third_party/nlohmann-json' (https://github.com/nlohmann/json) registered for path 'third_party/opentelemetry-cpp/third_party/nlohmann-json' 2025-11-03T16:59:09.8516114Z Submodule 'third_party/opentelemetry-proto' (https://github.com/open-telemetry/opentelemetry-proto) registered for path 'third_party/opentelemetry-cpp/third_party/opentelemetry-proto' 2025-11-03T16:59:09.8520021Z Submodule 'third_party/opentracing-cpp' (https://github.com/opentracing/opentracing-cpp.git) registered for path 'third_party/opentelemetry-cpp/third_party/opentracing-cpp' 2025-11-03T16:59:09.8525107Z Submodule 'third_party/prometheus-cpp' (https://github.com/jupp0r/prometheus-cpp) registered for path 'third_party/opentelemetry-cpp/third_party/prometheus-cpp' 2025-11-03T16:59:09.8528653Z Submodule 'tools/vcpkg' (https://github.com/Microsoft/vcpkg) registered for path 'third_party/opentelemetry-cpp/tools/vcpkg' 2025-11-03T16:59:09.8565249Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/opentelemetry-cpp/third_party/benchmark'... 2025-11-03T16:59:10.3160315Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/opentelemetry-cpp/third_party/opentracing-cpp'... 2025-11-03T16:59:10.3161573Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/opentelemetry-cpp/third_party/opentelemetry-proto'... 2025-11-03T16:59:10.3163385Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/opentelemetry-cpp/third_party/prometheus-cpp'... 2025-11-03T16:59:10.3164571Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/opentelemetry-cpp/third_party/ms-gsl'... 2025-11-03T16:59:10.4161937Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/opentelemetry-cpp/third_party/googletest'... 2025-11-03T16:59:10.8876249Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/opentelemetry-cpp/third_party/nlohmann-json'... 2025-11-03T16:59:17.4423345Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/opentelemetry-cpp/tools/vcpkg'... 2025-11-03T16:59:17.6819852Z Submodule path 'third_party/opentelemetry-cpp/third_party/benchmark': checked out 'd572f4777349d43653b21d6c2fc63020ab326db2' 2025-11-03T16:59:17.7286026Z Submodule path 'third_party/opentelemetry-cpp/third_party/googletest': checked out 'b796f7d44681514f58a683a3a71ff17c94edb0c1' 2025-11-03T16:59:17.7483073Z Submodule path 'third_party/opentelemetry-cpp/third_party/ms-gsl': checked out '6f4529395c5b7c2d661812257cd6780c67e54afa' 2025-11-03T16:59:17.8724772Z Submodule path 'third_party/opentelemetry-cpp/third_party/nlohmann-json': checked out 'bc889afb4c5bf1c0d8ee29ef35eaaf4c8bef8a5d' 2025-11-03T16:59:17.8900942Z Submodule path 'third_party/opentelemetry-cpp/third_party/opentelemetry-proto': checked out '4ca4f0335c63cda7ab31ea7ed70d6553aee14dce' 2025-11-03T16:59:17.9106611Z Submodule path 'third_party/opentelemetry-cpp/third_party/opentracing-cpp': checked out '06b57f48ded1fa3bdd3d4346f6ef29e40e08eaf5' 2025-11-03T16:59:17.9327346Z Submodule path 'third_party/opentelemetry-cpp/third_party/prometheus-cpp': checked out 'c9ffcdda9086ffd9e1283ea7a0276d831f3c8a8d' 2025-11-03T16:59:17.9348736Z Submodule 'civetweb' (https://github.com/civetweb/civetweb.git) registered for path 'third_party/opentelemetry-cpp/third_party/prometheus-cpp/3rdparty/civetweb' 2025-11-03T16:59:17.9352571Z Submodule 'googletest' (https://github.com/google/googletest.git) registered for path 'third_party/opentelemetry-cpp/third_party/prometheus-cpp/3rdparty/googletest' 2025-11-03T16:59:17.9388558Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/opentelemetry-cpp/third_party/prometheus-cpp/3rdparty/civetweb'... 2025-11-03T16:59:19.9391032Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/opentelemetry-cpp/third_party/prometheus-cpp/3rdparty/googletest'... 2025-11-03T16:59:20.1935892Z Submodule path 'third_party/opentelemetry-cpp/third_party/prometheus-cpp/3rdparty/civetweb': checked out 'eefb26f82b233268fc98577d265352720d477ba4' 2025-11-03T16:59:20.2464523Z Submodule path 'third_party/opentelemetry-cpp/third_party/prometheus-cpp/3rdparty/googletest': checked out 'e2239ee6043f73722e7aa812a459f54a28552929' 2025-11-03T16:59:20.9403369Z Submodule path 'third_party/opentelemetry-cpp/tools/vcpkg': checked out '8eb57355a4ffb410a2e94c07b4dca2dffbee8e50' 2025-11-03T16:59:20.9553189Z Submodule path 'third_party/pocketfft': checked out '0fa0ef591e38c2758e3184c6c23e497b9f732ffa' 2025-11-03T16:59:21.2602848Z Submodule path 'third_party/protobuf': checked out 'd1eca4e4b421cd2997495c4b4e65cea6be4e9b8a' 2025-11-03T16:59:21.2629901Z Submodule 'third_party/benchmark' (https://github.com/google/benchmark.git) registered for path 'third_party/protobuf/third_party/benchmark' 2025-11-03T16:59:21.2633313Z Submodule 'third_party/googletest' (https://github.com/google/googletest.git) registered for path 'third_party/protobuf/third_party/googletest' 2025-11-03T16:59:21.2668532Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/protobuf/third_party/benchmark'... 2025-11-03T16:59:21.7892737Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/protobuf/third_party/googletest'... 2025-11-03T16:59:22.0550617Z Submodule path 'third_party/protobuf/third_party/benchmark': checked out '5b7683f49e1e9223cf9927b24f6fd3d6bd82e3f8' 2025-11-03T16:59:22.1299313Z Submodule path 'third_party/protobuf/third_party/googletest': checked out '5ec7f0c4a113e2f18ac2c6cc7df51ad6afc24081' 2025-11-03T16:59:22.1442482Z Submodule path 'third_party/psimd': checked out '072586a71b55b7f8c584153d223e95687148a900' 2025-11-03T16:59:22.1597029Z Submodule path 'third_party/pthreadpool': checked out '4fe0e1e183925bf8cfa6aae24237e724a96479b8' 2025-11-03T16:59:22.2079733Z Submodule path 'third_party/pybind11': checked out 'f5fbe867d2d26e4a0a9177a51f6e568868ad3dc8' 2025-11-03T16:59:22.2411265Z Submodule path 'third_party/python-peachpy': checked out 'f45429b087dd7d5bc78bb40dc7cf06425c252d67' 2025-11-03T16:59:22.2899285Z Submodule path 'third_party/sleef': checked out '5a1d179df9cf652951b59010a2d2075372d67f68' 2025-11-03T16:59:22.3244739Z Submodule path 'third_party/tensorpipe': checked out 'af0118d13e52f5a08841464a768e01a0bf3e3075' 2025-11-03T16:59:22.3268319Z Submodule 'third_party/googletest' (https://github.com/google/googletest.git) registered for path 'third_party/tensorpipe/third_party/googletest' 2025-11-03T16:59:22.3272172Z Submodule 'third_party/libnop' (https://github.com/google/libnop.git) registered for path 'third_party/tensorpipe/third_party/libnop' 2025-11-03T16:59:22.3275961Z Submodule 'third_party/libuv' (https://github.com/libuv/libuv.git) registered for path 'third_party/tensorpipe/third_party/libuv' 2025-11-03T16:59:22.3279549Z Submodule 'third_party/pybind11' (https://github.com/pybind/pybind11.git) registered for path 'third_party/tensorpipe/third_party/pybind11' 2025-11-03T16:59:22.3313990Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/tensorpipe/third_party/googletest'... 2025-11-03T16:59:23.2328016Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/tensorpipe/third_party/libnop'... 2025-11-03T16:59:23.3170610Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/tensorpipe/third_party/libuv'... 2025-11-03T16:59:23.5190765Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/tensorpipe/third_party/pybind11'... 2025-11-03T16:59:23.5790681Z Submodule path 'third_party/tensorpipe/third_party/googletest': checked out 'aee0f9d9b5b87796ee8a0ab26b7587ec30e8858e' 2025-11-03T16:59:23.5995296Z Submodule path 'third_party/tensorpipe/third_party/libnop': checked out '910b55815be16109f04f4180e9adee14fb4ce281' 2025-11-03T16:59:23.6793175Z Submodule path 'third_party/tensorpipe/third_party/libuv': checked out '5152db2cbfeb5582e9c27c5ea1dba2cd9e10759b' 2025-11-03T16:59:23.7140707Z Submodule path 'third_party/tensorpipe/third_party/pybind11': checked out 'a23996fce38ff6ccfbcdc09f1e63f2c4be5ea2ef' 2025-11-03T16:59:23.7162295Z Submodule 'tools/clang' (https://github.com/wjakob/clang-cindex-python3) registered for path 'third_party/tensorpipe/third_party/pybind11/tools/clang' 2025-11-03T16:59:23.7195780Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/tensorpipe/third_party/pybind11/tools/clang'... 2025-11-03T16:59:23.9576767Z Submodule path 'third_party/tensorpipe/third_party/pybind11/tools/clang': checked out '6a00cbc4a9b8e68b71caf7f774b3f9c753ae84d5' 2025-11-03T16:59:23.9625910Z [command]/usr/bin/git submodule foreach --recursive git config --local gc.auto 0 2025-11-03T16:59:23.9986634Z Entering 'android/libs/fbjni' 2025-11-03T16:59:24.0041720Z Entering 'third_party/FP16' 2025-11-03T16:59:24.0103411Z Entering 'third_party/FXdiv' 2025-11-03T16:59:24.0158776Z Entering 'third_party/NNPACK' 2025-11-03T16:59:24.0215934Z Entering 'third_party/NVTX' 2025-11-03T16:59:24.0271831Z Entering 'third_party/VulkanMemoryAllocator' 2025-11-03T16:59:24.0325006Z Entering 'third_party/XNNPACK' 2025-11-03T16:59:24.0391418Z Entering 'third_party/aiter' 2025-11-03T16:59:24.0444640Z Entering 'third_party/aiter/3rdparty/composable_kernel' 2025-11-03T16:59:24.0505646Z Entering 'third_party/benchmark' 2025-11-03T16:59:24.0559897Z Entering 'third_party/composable_kernel' 2025-11-03T16:59:24.0622836Z Entering 'third_party/cpp-httplib' 2025-11-03T16:59:24.0674071Z Entering 'third_party/cpuinfo' 2025-11-03T16:59:24.0729151Z Entering 'third_party/cudnn_frontend' 2025-11-03T16:59:24.0783363Z Entering 'third_party/cutlass' 2025-11-03T16:59:24.0845570Z Entering 'third_party/fbgemm' 2025-11-03T16:59:24.0902267Z Entering 'third_party/fbgemm/external/asmjit' 2025-11-03T16:59:24.0956593Z Entering 'third_party/fbgemm/external/composable_kernel' 2025-11-03T16:59:24.1018268Z Entering 'third_party/fbgemm/external/cpuinfo' 2025-11-03T16:59:24.1072332Z Entering 'third_party/fbgemm/external/cutlass' 2025-11-03T16:59:24.1132778Z Entering 'third_party/fbgemm/external/googletest' 2025-11-03T16:59:24.1186646Z Entering 'third_party/fbgemm/external/hipify_torch' 2025-11-03T16:59:24.1240392Z Entering 'third_party/fbgemm/external/json' 2025-11-03T16:59:24.1297149Z Entering 'third_party/flash-attention' 2025-11-03T16:59:24.1352315Z Entering 'third_party/flash-attention/csrc/composable_kernel' 2025-11-03T16:59:24.1411115Z Entering 'third_party/flash-attention/csrc/cutlass' 2025-11-03T16:59:24.1474057Z Entering 'third_party/flatbuffers' 2025-11-03T16:59:24.1528459Z Entering 'third_party/fmt' 2025-11-03T16:59:24.1580292Z Entering 'third_party/gemmlowp/gemmlowp' 2025-11-03T16:59:24.1636061Z Entering 'third_party/gloo' 2025-11-03T16:59:24.1691954Z Entering 'third_party/googletest' 2025-11-03T16:59:24.1744881Z Entering 'third_party/ideep' 2025-11-03T16:59:24.1797303Z Entering 'third_party/ideep/mkl-dnn' 2025-11-03T16:59:24.1859545Z Entering 'third_party/ittapi' 2025-11-03T16:59:24.1912793Z Entering 'third_party/kineto' 2025-11-03T16:59:24.1966513Z Entering 'third_party/kineto/libkineto/third_party/dynolog' 2025-11-03T16:59:24.2022197Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/DCGM' 2025-11-03T16:59:24.2076683Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/cpr' 2025-11-03T16:59:24.2133854Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/fmt' 2025-11-03T16:59:24.2188934Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/gflags' 2025-11-03T16:59:24.2242071Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/gflags/doc' 2025-11-03T16:59:24.2297338Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/glog' 2025-11-03T16:59:24.2353216Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/googletest' 2025-11-03T16:59:24.2406647Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/json' 2025-11-03T16:59:24.2463042Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/pfs' 2025-11-03T16:59:24.2518265Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/prometheus-cpp' 2025-11-03T16:59:24.2570672Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/prometheus-cpp/3rdparty/civetweb' 2025-11-03T16:59:24.2630137Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/prometheus-cpp/3rdparty/googletest' 2025-11-03T16:59:24.2690396Z Entering 'third_party/kineto/libkineto/third_party/fmt' 2025-11-03T16:59:24.2743075Z Entering 'third_party/kineto/libkineto/third_party/googletest' 2025-11-03T16:59:24.2799890Z Entering 'third_party/kleidiai' 2025-11-03T16:59:24.2854318Z Entering 'third_party/mimalloc' 2025-11-03T16:59:24.2907989Z Entering 'third_party/nlohmann' 2025-11-03T16:59:24.2964534Z Entering 'third_party/onnx' 2025-11-03T16:59:24.3038832Z Entering 'third_party/onnx/third_party/pybind11' 2025-11-03T16:59:24.3097074Z Entering 'third_party/opentelemetry-cpp' 2025-11-03T16:59:24.3150094Z Entering 'third_party/opentelemetry-cpp/third_party/benchmark' 2025-11-03T16:59:24.3201223Z Entering 'third_party/opentelemetry-cpp/third_party/googletest' 2025-11-03T16:59:24.3254406Z Entering 'third_party/opentelemetry-cpp/third_party/ms-gsl' 2025-11-03T16:59:24.3307701Z Entering 'third_party/opentelemetry-cpp/third_party/nlohmann-json' 2025-11-03T16:59:24.3362631Z Entering 'third_party/opentelemetry-cpp/third_party/opentelemetry-proto' 2025-11-03T16:59:24.3416780Z Entering 'third_party/opentelemetry-cpp/third_party/opentracing-cpp' 2025-11-03T16:59:24.3471429Z Entering 'third_party/opentelemetry-cpp/third_party/prometheus-cpp' 2025-11-03T16:59:24.3526662Z Entering 'third_party/opentelemetry-cpp/third_party/prometheus-cpp/3rdparty/civetweb' 2025-11-03T16:59:24.3582669Z Entering 'third_party/opentelemetry-cpp/third_party/prometheus-cpp/3rdparty/googletest' 2025-11-03T16:59:24.3641078Z Entering 'third_party/opentelemetry-cpp/tools/vcpkg' 2025-11-03T16:59:24.3712577Z Entering 'third_party/pocketfft' 2025-11-03T16:59:24.3766464Z Entering 'third_party/protobuf' 2025-11-03T16:59:24.3823616Z Entering 'third_party/protobuf/third_party/benchmark' 2025-11-03T16:59:24.3877024Z Entering 'third_party/protobuf/third_party/googletest' 2025-11-03T16:59:24.3934299Z Entering 'third_party/psimd' 2025-11-03T16:59:24.3988831Z Entering 'third_party/pthreadpool' 2025-11-03T16:59:24.4040162Z Entering 'third_party/pybind11' 2025-11-03T16:59:24.4093826Z Entering 'third_party/python-peachpy' 2025-11-03T16:59:24.4147908Z Entering 'third_party/sleef' 2025-11-03T16:59:24.4202035Z Entering 'third_party/tensorpipe' 2025-11-03T16:59:24.4252838Z Entering 'third_party/tensorpipe/third_party/googletest' 2025-11-03T16:59:24.4304778Z Entering 'third_party/tensorpipe/third_party/libnop' 2025-11-03T16:59:24.4357755Z Entering 'third_party/tensorpipe/third_party/libuv' 2025-11-03T16:59:24.4409915Z Entering 'third_party/tensorpipe/third_party/pybind11' 2025-11-03T16:59:24.4461399Z Entering 'third_party/tensorpipe/third_party/pybind11/tools/clang' 2025-11-03T16:59:24.4535427Z ##[endgroup] 2025-11-03T16:59:24.4535893Z ##[group]Persisting credentials for submodules 2025-11-03T16:59:24.4541400Z [command]/usr/bin/git submodule foreach --recursive sh -c "git config --local --name-only --get-regexp 'url\.https\:\/\/github\.com\/\.insteadOf' && git config --local --unset-all 'url.https://github.com/.insteadOf' || :" 2025-11-03T16:59:24.4906663Z Entering 'android/libs/fbjni' 2025-11-03T16:59:24.4984383Z Entering 'third_party/FP16' 2025-11-03T16:59:24.5057189Z Entering 'third_party/FXdiv' 2025-11-03T16:59:24.5129482Z Entering 'third_party/NNPACK' 2025-11-03T16:59:24.5206204Z Entering 'third_party/NVTX' 2025-11-03T16:59:24.5275145Z Entering 'third_party/VulkanMemoryAllocator' 2025-11-03T16:59:24.5346426Z Entering 'third_party/XNNPACK' 2025-11-03T16:59:24.5428092Z Entering 'third_party/aiter' 2025-11-03T16:59:24.5504812Z Entering 'third_party/aiter/3rdparty/composable_kernel' 2025-11-03T16:59:24.5583484Z Entering 'third_party/benchmark' 2025-11-03T16:59:24.5654161Z Entering 'third_party/composable_kernel' 2025-11-03T16:59:24.5731358Z Entering 'third_party/cpp-httplib' 2025-11-03T16:59:24.5803867Z Entering 'third_party/cpuinfo' 2025-11-03T16:59:24.5874071Z Entering 'third_party/cudnn_frontend' 2025-11-03T16:59:24.5943967Z Entering 'third_party/cutlass' 2025-11-03T16:59:24.6022060Z Entering 'third_party/fbgemm' 2025-11-03T16:59:24.6095149Z Entering 'third_party/fbgemm/external/asmjit' 2025-11-03T16:59:24.6163917Z Entering 'third_party/fbgemm/external/composable_kernel' 2025-11-03T16:59:24.6240571Z Entering 'third_party/fbgemm/external/cpuinfo' 2025-11-03T16:59:24.6311771Z Entering 'third_party/fbgemm/external/cutlass' 2025-11-03T16:59:24.6390250Z Entering 'third_party/fbgemm/external/googletest' 2025-11-03T16:59:24.6461775Z Entering 'third_party/fbgemm/external/hipify_torch' 2025-11-03T16:59:24.6531442Z Entering 'third_party/fbgemm/external/json' 2025-11-03T16:59:24.6607046Z Entering 'third_party/flash-attention' 2025-11-03T16:59:24.6679214Z Entering 'third_party/flash-attention/csrc/composable_kernel' 2025-11-03T16:59:24.6756201Z Entering 'third_party/flash-attention/csrc/cutlass' 2025-11-03T16:59:24.6835601Z Entering 'third_party/flatbuffers' 2025-11-03T16:59:24.6907363Z Entering 'third_party/fmt' 2025-11-03T16:59:24.6980950Z Entering 'third_party/gemmlowp/gemmlowp' 2025-11-03T16:59:24.7053951Z Entering 'third_party/gloo' 2025-11-03T16:59:24.7124792Z Entering 'third_party/googletest' 2025-11-03T16:59:24.7194016Z Entering 'third_party/ideep' 2025-11-03T16:59:24.7262759Z Entering 'third_party/ideep/mkl-dnn' 2025-11-03T16:59:24.7342855Z Entering 'third_party/ittapi' 2025-11-03T16:59:24.7418778Z Entering 'third_party/kineto' 2025-11-03T16:59:24.7493363Z Entering 'third_party/kineto/libkineto/third_party/dynolog' 2025-11-03T16:59:24.7562207Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/DCGM' 2025-11-03T16:59:24.7641304Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/cpr' 2025-11-03T16:59:24.7711443Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/fmt' 2025-11-03T16:59:24.7783046Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/gflags' 2025-11-03T16:59:24.7851904Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/gflags/doc' 2025-11-03T16:59:24.7927287Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/glog' 2025-11-03T16:59:24.8000382Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/googletest' 2025-11-03T16:59:24.8072469Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/json' 2025-11-03T16:59:24.8145819Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/pfs' 2025-11-03T16:59:24.8214701Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/prometheus-cpp' 2025-11-03T16:59:24.8283825Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/prometheus-cpp/3rdparty/civetweb' 2025-11-03T16:59:24.8355977Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/prometheus-cpp/3rdparty/googletest' 2025-11-03T16:59:24.8436402Z Entering 'third_party/kineto/libkineto/third_party/fmt' 2025-11-03T16:59:24.8507533Z Entering 'third_party/kineto/libkineto/third_party/googletest' 2025-11-03T16:59:24.8584674Z Entering 'third_party/kleidiai' 2025-11-03T16:59:24.8654756Z Entering 'third_party/mimalloc' 2025-11-03T16:59:24.8724831Z Entering 'third_party/nlohmann' 2025-11-03T16:59:24.8795344Z Entering 'third_party/onnx' 2025-11-03T16:59:24.8883915Z Entering 'third_party/onnx/third_party/pybind11' 2025-11-03T16:59:24.8958377Z Entering 'third_party/opentelemetry-cpp' 2025-11-03T16:59:24.9041814Z Entering 'third_party/opentelemetry-cpp/third_party/benchmark' 2025-11-03T16:59:24.9114254Z Entering 'third_party/opentelemetry-cpp/third_party/googletest' 2025-11-03T16:59:24.9184057Z Entering 'third_party/opentelemetry-cpp/third_party/ms-gsl' 2025-11-03T16:59:24.9253143Z Entering 'third_party/opentelemetry-cpp/third_party/nlohmann-json' 2025-11-03T16:59:24.9324501Z Entering 'third_party/opentelemetry-cpp/third_party/opentelemetry-proto' 2025-11-03T16:59:24.9395638Z Entering 'third_party/opentelemetry-cpp/third_party/opentracing-cpp' 2025-11-03T16:59:24.9464159Z Entering 'third_party/opentelemetry-cpp/third_party/prometheus-cpp' 2025-11-03T16:59:24.9532301Z Entering 'third_party/opentelemetry-cpp/third_party/prometheus-cpp/3rdparty/civetweb' 2025-11-03T16:59:24.9604313Z Entering 'third_party/opentelemetry-cpp/third_party/prometheus-cpp/3rdparty/googletest' 2025-11-03T16:59:24.9678571Z Entering 'third_party/opentelemetry-cpp/tools/vcpkg' 2025-11-03T16:59:24.9770435Z Entering 'third_party/pocketfft' 2025-11-03T16:59:24.9841732Z Entering 'third_party/protobuf' 2025-11-03T16:59:24.9915787Z Entering 'third_party/protobuf/third_party/benchmark' 2025-11-03T16:59:24.9984611Z Entering 'third_party/protobuf/third_party/googletest' 2025-11-03T16:59:25.0057215Z Entering 'third_party/psimd' 2025-11-03T16:59:25.0132567Z Entering 'third_party/pthreadpool' 2025-11-03T16:59:25.0203647Z Entering 'third_party/pybind11' 2025-11-03T16:59:25.0274160Z Entering 'third_party/python-peachpy' 2025-11-03T16:59:25.0344562Z Entering 'third_party/sleef' 2025-11-03T16:59:25.0415261Z Entering 'third_party/tensorpipe' 2025-11-03T16:59:25.0487236Z Entering 'third_party/tensorpipe/third_party/googletest' 2025-11-03T16:59:25.0559415Z Entering 'third_party/tensorpipe/third_party/libnop' 2025-11-03T16:59:25.0634648Z Entering 'third_party/tensorpipe/third_party/libuv' 2025-11-03T16:59:25.0704424Z Entering 'third_party/tensorpipe/third_party/pybind11' 2025-11-03T16:59:25.0772356Z Entering 'third_party/tensorpipe/third_party/pybind11/tools/clang' 2025-11-03T16:59:25.0866911Z [command]/usr/bin/git submodule foreach --recursive sh -c "git config --local 'http.https://github.com/.extraheader' 'AUTHORIZATION: basic ***' && git config --local --show-origin --name-only --get-regexp remote.origin.url" 2025-11-03T16:59:25.1237607Z Entering 'android/libs/fbjni' 2025-11-03T16:59:25.1302130Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/android/libs/fbjni/config remote.origin.url 2025-11-03T16:59:25.1323965Z Entering 'third_party/FP16' 2025-11-03T16:59:25.1388963Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/NNPACK_deps/FP16/config remote.origin.url 2025-11-03T16:59:25.1409496Z Entering 'third_party/FXdiv' 2025-11-03T16:59:25.1480595Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/NNPACK_deps/FXdiv/config remote.origin.url 2025-11-03T16:59:25.1502545Z Entering 'third_party/NNPACK' 2025-11-03T16:59:25.1568330Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/NNPACK/config remote.origin.url 2025-11-03T16:59:25.1590768Z Entering 'third_party/NVTX' 2025-11-03T16:59:25.1657534Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/NVTX/config remote.origin.url 2025-11-03T16:59:25.1681115Z Entering 'third_party/VulkanMemoryAllocator' 2025-11-03T16:59:25.1748717Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/VulkanMemoryAllocator/config remote.origin.url 2025-11-03T16:59:25.1768935Z Entering 'third_party/XNNPACK' 2025-11-03T16:59:25.1833791Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/XNNPACK/config remote.origin.url 2025-11-03T16:59:25.1871607Z Entering 'third_party/aiter' 2025-11-03T16:59:25.1938329Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/aiter/config remote.origin.url 2025-11-03T16:59:25.1960396Z Entering 'third_party/aiter/3rdparty/composable_kernel' 2025-11-03T16:59:25.2029192Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/aiter/modules/3rdparty/composable_kernel/config remote.origin.url 2025-11-03T16:59:25.2058115Z Entering 'third_party/benchmark' 2025-11-03T16:59:25.2122589Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/benchmark/config remote.origin.url 2025-11-03T16:59:25.2143798Z Entering 'third_party/composable_kernel' 2025-11-03T16:59:25.2209291Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/composable_kernel/config remote.origin.url 2025-11-03T16:59:25.2240012Z Entering 'third_party/cpp-httplib' 2025-11-03T16:59:25.2306309Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/cpp-httplib/config remote.origin.url 2025-11-03T16:59:25.2330456Z Entering 'third_party/cpuinfo' 2025-11-03T16:59:25.2395232Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/cpuinfo/config remote.origin.url 2025-11-03T16:59:25.2417908Z Entering 'third_party/cudnn_frontend' 2025-11-03T16:59:25.2484657Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/cudnn_frontend/config remote.origin.url 2025-11-03T16:59:25.2506311Z Entering 'third_party/cutlass' 2025-11-03T16:59:25.2569938Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/cutlass/config remote.origin.url 2025-11-03T16:59:25.2600187Z Entering 'third_party/fbgemm' 2025-11-03T16:59:25.2667234Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/fbgemm/config remote.origin.url 2025-11-03T16:59:25.2691198Z Entering 'third_party/fbgemm/external/asmjit' 2025-11-03T16:59:25.2757902Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/fbgemm/modules/external/asmjit/config remote.origin.url 2025-11-03T16:59:25.2780341Z Entering 'third_party/fbgemm/external/composable_kernel' 2025-11-03T16:59:25.2845943Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/fbgemm/modules/external/composable_kernel/config remote.origin.url 2025-11-03T16:59:25.2875780Z Entering 'third_party/fbgemm/external/cpuinfo' 2025-11-03T16:59:25.2943898Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/fbgemm/modules/external/cpuinfo/config remote.origin.url 2025-11-03T16:59:25.2965065Z Entering 'third_party/fbgemm/external/cutlass' 2025-11-03T16:59:25.3030406Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/fbgemm/modules/external/cutlass/config remote.origin.url 2025-11-03T16:59:25.3060890Z Entering 'third_party/fbgemm/external/googletest' 2025-11-03T16:59:25.3125901Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/fbgemm/modules/external/googletest/config remote.origin.url 2025-11-03T16:59:25.3148711Z Entering 'third_party/fbgemm/external/hipify_torch' 2025-11-03T16:59:25.3217882Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/fbgemm/modules/external/hipify_torch/config remote.origin.url 2025-11-03T16:59:25.3239508Z Entering 'third_party/fbgemm/external/json' 2025-11-03T16:59:25.3309843Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/fbgemm/modules/external/json/config remote.origin.url 2025-11-03T16:59:25.3335933Z Entering 'third_party/flash-attention' 2025-11-03T16:59:25.3407685Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/flash-attention/config remote.origin.url 2025-11-03T16:59:25.3430753Z Entering 'third_party/flash-attention/csrc/composable_kernel' 2025-11-03T16:59:25.3495487Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/flash-attention/modules/csrc/composable_kernel/config remote.origin.url 2025-11-03T16:59:25.3529004Z Entering 'third_party/flash-attention/csrc/cutlass' 2025-11-03T16:59:25.3596807Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/flash-attention/modules/csrc/cutlass/config remote.origin.url 2025-11-03T16:59:25.3629810Z Entering 'third_party/flatbuffers' 2025-11-03T16:59:25.3695460Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/flatbuffers/config remote.origin.url 2025-11-03T16:59:25.3721507Z Entering 'third_party/fmt' 2025-11-03T16:59:25.3787874Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/fmt/config remote.origin.url 2025-11-03T16:59:25.3810154Z Entering 'third_party/gemmlowp/gemmlowp' 2025-11-03T16:59:25.3878425Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/gemmlowp/gemmlowp/config remote.origin.url 2025-11-03T16:59:25.3899374Z Entering 'third_party/gloo' 2025-11-03T16:59:25.3976332Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/gloo/config remote.origin.url 2025-11-03T16:59:25.3999277Z Entering 'third_party/googletest' 2025-11-03T16:59:25.4066881Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/googletest/config remote.origin.url 2025-11-03T16:59:25.4089328Z Entering 'third_party/ideep' 2025-11-03T16:59:25.4162482Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/ideep/config remote.origin.url 2025-11-03T16:59:25.4182710Z Entering 'third_party/ideep/mkl-dnn' 2025-11-03T16:59:25.4246460Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/ideep/modules/mkl-dnn/config remote.origin.url 2025-11-03T16:59:25.4276812Z Entering 'third_party/ittapi' 2025-11-03T16:59:25.4340973Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/ittapi/config remote.origin.url 2025-11-03T16:59:25.4363046Z Entering 'third_party/kineto' 2025-11-03T16:59:25.4430582Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/kineto/config remote.origin.url 2025-11-03T16:59:25.4451497Z Entering 'third_party/kineto/libkineto/third_party/dynolog' 2025-11-03T16:59:25.4519164Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/kineto/modules/libkineto/third_party/dynolog/config remote.origin.url 2025-11-03T16:59:25.4540856Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/DCGM' 2025-11-03T16:59:25.4608359Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/kineto/modules/libkineto/third_party/dynolog/modules/third_party/DCGM/config remote.origin.url 2025-11-03T16:59:25.4632063Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/cpr' 2025-11-03T16:59:25.4698774Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/kineto/modules/libkineto/third_party/dynolog/modules/third_party/cpr/config remote.origin.url 2025-11-03T16:59:25.4721554Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/fmt' 2025-11-03T16:59:25.4788921Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/kineto/modules/libkineto/third_party/dynolog/modules/third_party/fmt/config remote.origin.url 2025-11-03T16:59:25.4809294Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/gflags' 2025-11-03T16:59:25.4876608Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/kineto/modules/libkineto/third_party/dynolog/modules/third_party/gflags/config remote.origin.url 2025-11-03T16:59:25.4897144Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/gflags/doc' 2025-11-03T16:59:25.4966295Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/kineto/modules/libkineto/third_party/dynolog/modules/third_party/gflags/modules/doc/config remote.origin.url 2025-11-03T16:59:25.4992243Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/glog' 2025-11-03T16:59:25.5057570Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/kineto/modules/libkineto/third_party/dynolog/modules/third_party/glog/config remote.origin.url 2025-11-03T16:59:25.5085275Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/googletest' 2025-11-03T16:59:25.5149170Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/kineto/modules/libkineto/third_party/dynolog/modules/third_party/googletest/config remote.origin.url 2025-11-03T16:59:25.5171174Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/json' 2025-11-03T16:59:25.5239247Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/kineto/modules/libkineto/third_party/dynolog/modules/third_party/json/config remote.origin.url 2025-11-03T16:59:25.5262485Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/pfs' 2025-11-03T16:59:25.5329241Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/kineto/modules/libkineto/third_party/dynolog/modules/third_party/pfs/config remote.origin.url 2025-11-03T16:59:25.5351578Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/prometheus-cpp' 2025-11-03T16:59:25.5418395Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/kineto/modules/libkineto/third_party/dynolog/modules/third_party/prometheus-cpp/config remote.origin.url 2025-11-03T16:59:25.5439846Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/prometheus-cpp/3rdparty/civetweb' 2025-11-03T16:59:25.5507671Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/kineto/modules/libkineto/third_party/dynolog/modules/third_party/prometheus-cpp/modules/civetweb/config remote.origin.url 2025-11-03T16:59:25.5536603Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/prometheus-cpp/3rdparty/googletest' 2025-11-03T16:59:25.5605582Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/kineto/modules/libkineto/third_party/dynolog/modules/third_party/prometheus-cpp/modules/googletest/config remote.origin.url 2025-11-03T16:59:25.5632280Z Entering 'third_party/kineto/libkineto/third_party/fmt' 2025-11-03T16:59:25.5697854Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/kineto/modules/libkineto/third_party/fmt/config remote.origin.url 2025-11-03T16:59:25.5721418Z Entering 'third_party/kineto/libkineto/third_party/googletest' 2025-11-03T16:59:25.5788203Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/kineto/modules/libkineto/third_party/googletest/config remote.origin.url 2025-11-03T16:59:25.5813788Z Entering 'third_party/kleidiai' 2025-11-03T16:59:25.5878074Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/kleidiai/config remote.origin.url 2025-11-03T16:59:25.5901384Z Entering 'third_party/mimalloc' 2025-11-03T16:59:25.5967457Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/mimalloc/config remote.origin.url 2025-11-03T16:59:25.5990643Z Entering 'third_party/nlohmann' 2025-11-03T16:59:25.6056125Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/nlohmann/config remote.origin.url 2025-11-03T16:59:25.6080278Z Entering 'third_party/onnx' 2025-11-03T16:59:25.6145564Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/onnx/config remote.origin.url 2025-11-03T16:59:25.6181885Z Entering 'third_party/onnx/third_party/pybind11' 2025-11-03T16:59:25.6249054Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/onnx/modules/third_party/pybind11/config remote.origin.url 2025-11-03T16:59:25.6276061Z Entering 'third_party/opentelemetry-cpp' 2025-11-03T16:59:25.6341757Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/opentelemetry-cpp/config remote.origin.url 2025-11-03T16:59:25.6363841Z Entering 'third_party/opentelemetry-cpp/third_party/benchmark' 2025-11-03T16:59:25.6428881Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/opentelemetry-cpp/modules/third_party/benchmark/config remote.origin.url 2025-11-03T16:59:25.6450847Z Entering 'third_party/opentelemetry-cpp/third_party/googletest' 2025-11-03T16:59:25.6518654Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/opentelemetry-cpp/modules/third_party/googletest/config remote.origin.url 2025-11-03T16:59:25.6541103Z Entering 'third_party/opentelemetry-cpp/third_party/ms-gsl' 2025-11-03T16:59:25.6605901Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/opentelemetry-cpp/modules/third_party/ms-gsl/config remote.origin.url 2025-11-03T16:59:25.6632916Z Entering 'third_party/opentelemetry-cpp/third_party/nlohmann-json' 2025-11-03T16:59:25.6697589Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/opentelemetry-cpp/modules/third_party/nlohmann-json/config remote.origin.url 2025-11-03T16:59:25.6720847Z Entering 'third_party/opentelemetry-cpp/third_party/opentelemetry-proto' 2025-11-03T16:59:25.6782543Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/opentelemetry-cpp/modules/third_party/opentelemetry-proto/config remote.origin.url 2025-11-03T16:59:25.6803850Z Entering 'third_party/opentelemetry-cpp/third_party/opentracing-cpp' 2025-11-03T16:59:25.6868986Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/opentelemetry-cpp/modules/third_party/opentracing-cpp/config remote.origin.url 2025-11-03T16:59:25.6890848Z Entering 'third_party/opentelemetry-cpp/third_party/prometheus-cpp' 2025-11-03T16:59:25.6957558Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/opentelemetry-cpp/modules/third_party/prometheus-cpp/config remote.origin.url 2025-11-03T16:59:25.6978859Z Entering 'third_party/opentelemetry-cpp/third_party/prometheus-cpp/3rdparty/civetweb' 2025-11-03T16:59:25.7046986Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/opentelemetry-cpp/modules/third_party/prometheus-cpp/modules/civetweb/config remote.origin.url 2025-11-03T16:59:25.7071636Z Entering 'third_party/opentelemetry-cpp/third_party/prometheus-cpp/3rdparty/googletest' 2025-11-03T16:59:25.7138982Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/opentelemetry-cpp/modules/third_party/prometheus-cpp/modules/googletest/config remote.origin.url 2025-11-03T16:59:25.7164054Z Entering 'third_party/opentelemetry-cpp/tools/vcpkg' 2025-11-03T16:59:25.7228876Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/opentelemetry-cpp/modules/tools/vcpkg/config remote.origin.url 2025-11-03T16:59:25.7269836Z Entering 'third_party/pocketfft' 2025-11-03T16:59:25.7335998Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/pocketfft/config remote.origin.url 2025-11-03T16:59:25.7356788Z Entering 'third_party/protobuf' 2025-11-03T16:59:25.7419096Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/protobuf/config remote.origin.url 2025-11-03T16:59:25.7443145Z Entering 'third_party/protobuf/third_party/benchmark' 2025-11-03T16:59:25.7508690Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/protobuf/modules/third_party/benchmark/config remote.origin.url 2025-11-03T16:59:25.7531344Z Entering 'third_party/protobuf/third_party/googletest' 2025-11-03T16:59:25.7596535Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/protobuf/modules/third_party/googletest/config remote.origin.url 2025-11-03T16:59:25.7621636Z Entering 'third_party/psimd' 2025-11-03T16:59:25.7685404Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/NNPACK_deps/psimd/config remote.origin.url 2025-11-03T16:59:25.7708422Z Entering 'third_party/pthreadpool' 2025-11-03T16:59:25.7777395Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/NNPACK_deps/pthreadpool/config remote.origin.url 2025-11-03T16:59:25.7805326Z Entering 'third_party/pybind11' 2025-11-03T16:59:25.7867551Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/pybind11/config remote.origin.url 2025-11-03T16:59:25.7890665Z Entering 'third_party/python-peachpy' 2025-11-03T16:59:25.7957872Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/python-peachpy/config remote.origin.url 2025-11-03T16:59:25.7980707Z Entering 'third_party/sleef' 2025-11-03T16:59:25.8050447Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/sleef/config remote.origin.url 2025-11-03T16:59:25.8072954Z Entering 'third_party/tensorpipe' 2025-11-03T16:59:25.8139121Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/tensorpipe/config remote.origin.url 2025-11-03T16:59:25.8160399Z Entering 'third_party/tensorpipe/third_party/googletest' 2025-11-03T16:59:25.8226811Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/tensorpipe/modules/third_party/googletest/config remote.origin.url 2025-11-03T16:59:25.8248779Z Entering 'third_party/tensorpipe/third_party/libnop' 2025-11-03T16:59:25.8320338Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/tensorpipe/modules/third_party/libnop/config remote.origin.url 2025-11-03T16:59:25.8341692Z Entering 'third_party/tensorpipe/third_party/libuv' 2025-11-03T16:59:25.8407346Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/tensorpipe/modules/third_party/libuv/config remote.origin.url 2025-11-03T16:59:25.8430424Z Entering 'third_party/tensorpipe/third_party/pybind11' 2025-11-03T16:59:25.8499881Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/tensorpipe/modules/third_party/pybind11/config remote.origin.url 2025-11-03T16:59:25.8519865Z Entering 'third_party/tensorpipe/third_party/pybind11/tools/clang' 2025-11-03T16:59:25.8587685Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/tensorpipe/modules/third_party/pybind11/modules/tools/clang/config remote.origin.url 2025-11-03T16:59:25.9254864Z [command]/usr/bin/git submodule foreach --recursive git config --local --add 'url.https://github.com/.insteadOf' 'git@github.com:' 2025-11-03T16:59:25.9625062Z Entering 'android/libs/fbjni' 2025-11-03T16:59:25.9686579Z Entering 'third_party/FP16' 2025-11-03T16:59:25.9750182Z Entering 'third_party/FXdiv' 2025-11-03T16:59:25.9805357Z Entering 'third_party/NNPACK' 2025-11-03T16:59:25.9859013Z Entering 'third_party/NVTX' 2025-11-03T16:59:25.9913572Z Entering 'third_party/VulkanMemoryAllocator' 2025-11-03T16:59:25.9965530Z Entering 'third_party/XNNPACK' 2025-11-03T16:59:26.0033097Z Entering 'third_party/aiter' 2025-11-03T16:59:26.0085191Z Entering 'third_party/aiter/3rdparty/composable_kernel' 2025-11-03T16:59:26.0148017Z Entering 'third_party/benchmark' 2025-11-03T16:59:26.0202358Z Entering 'third_party/composable_kernel' 2025-11-03T16:59:26.0262239Z Entering 'third_party/cpp-httplib' 2025-11-03T16:59:26.0319140Z Entering 'third_party/cpuinfo' 2025-11-03T16:59:26.0375363Z Entering 'third_party/cudnn_frontend' 2025-11-03T16:59:26.0435147Z Entering 'third_party/cutlass' 2025-11-03T16:59:26.0497663Z Entering 'third_party/fbgemm' 2025-11-03T16:59:26.0552004Z Entering 'third_party/fbgemm/external/asmjit' 2025-11-03T16:59:26.0604625Z Entering 'third_party/fbgemm/external/composable_kernel' 2025-11-03T16:59:26.0665988Z Entering 'third_party/fbgemm/external/cpuinfo' 2025-11-03T16:59:26.0719429Z Entering 'third_party/fbgemm/external/cutlass' 2025-11-03T16:59:26.0781270Z Entering 'third_party/fbgemm/external/googletest' 2025-11-03T16:59:26.0833838Z Entering 'third_party/fbgemm/external/hipify_torch' 2025-11-03T16:59:26.0884168Z Entering 'third_party/fbgemm/external/json' 2025-11-03T16:59:26.0940066Z Entering 'third_party/flash-attention' 2025-11-03T16:59:26.0994562Z Entering 'third_party/flash-attention/csrc/composable_kernel' 2025-11-03T16:59:26.1052389Z Entering 'third_party/flash-attention/csrc/cutlass' 2025-11-03T16:59:26.1115777Z Entering 'third_party/flatbuffers' 2025-11-03T16:59:26.1172274Z Entering 'third_party/fmt' 2025-11-03T16:59:26.1225731Z Entering 'third_party/gemmlowp/gemmlowp' 2025-11-03T16:59:26.1279254Z Entering 'third_party/gloo' 2025-11-03T16:59:26.1334073Z Entering 'third_party/googletest' 2025-11-03T16:59:26.1389968Z Entering 'third_party/ideep' 2025-11-03T16:59:26.1441446Z Entering 'third_party/ideep/mkl-dnn' 2025-11-03T16:59:26.1503135Z Entering 'third_party/ittapi' 2025-11-03T16:59:26.1561953Z Entering 'third_party/kineto' 2025-11-03T16:59:26.1615349Z Entering 'third_party/kineto/libkineto/third_party/dynolog' 2025-11-03T16:59:26.1668298Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/DCGM' 2025-11-03T16:59:26.1722358Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/cpr' 2025-11-03T16:59:26.1775787Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/fmt' 2025-11-03T16:59:26.1830847Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/gflags' 2025-11-03T16:59:26.1883094Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/gflags/doc' 2025-11-03T16:59:26.1939130Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/glog' 2025-11-03T16:59:26.1994423Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/googletest' 2025-11-03T16:59:26.2047649Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/json' 2025-11-03T16:59:26.2101979Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/pfs' 2025-11-03T16:59:26.2155297Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/prometheus-cpp' 2025-11-03T16:59:26.2205744Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/prometheus-cpp/3rdparty/civetweb' 2025-11-03T16:59:26.2265073Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/prometheus-cpp/3rdparty/googletest' 2025-11-03T16:59:26.2326313Z Entering 'third_party/kineto/libkineto/third_party/fmt' 2025-11-03T16:59:26.2378486Z Entering 'third_party/kineto/libkineto/third_party/googletest' 2025-11-03T16:59:26.2435938Z Entering 'third_party/kleidiai' 2025-11-03T16:59:26.2490995Z Entering 'third_party/mimalloc' 2025-11-03T16:59:26.2546149Z Entering 'third_party/nlohmann' 2025-11-03T16:59:26.2600201Z Entering 'third_party/onnx' 2025-11-03T16:59:26.2668113Z Entering 'third_party/onnx/third_party/pybind11' 2025-11-03T16:59:26.2726795Z Entering 'third_party/opentelemetry-cpp' 2025-11-03T16:59:26.2781126Z Entering 'third_party/opentelemetry-cpp/third_party/benchmark' 2025-11-03T16:59:26.2835124Z Entering 'third_party/opentelemetry-cpp/third_party/googletest' 2025-11-03T16:59:26.2886105Z Entering 'third_party/opentelemetry-cpp/third_party/ms-gsl' 2025-11-03T16:59:26.2939154Z Entering 'third_party/opentelemetry-cpp/third_party/nlohmann-json' 2025-11-03T16:59:26.2993861Z Entering 'third_party/opentelemetry-cpp/third_party/opentelemetry-proto' 2025-11-03T16:59:26.3047144Z Entering 'third_party/opentelemetry-cpp/third_party/opentracing-cpp' 2025-11-03T16:59:26.3101620Z Entering 'third_party/opentelemetry-cpp/third_party/prometheus-cpp' 2025-11-03T16:59:26.3153709Z Entering 'third_party/opentelemetry-cpp/third_party/prometheus-cpp/3rdparty/civetweb' 2025-11-03T16:59:26.3207513Z Entering 'third_party/opentelemetry-cpp/third_party/prometheus-cpp/3rdparty/googletest' 2025-11-03T16:59:26.3265431Z Entering 'third_party/opentelemetry-cpp/tools/vcpkg' 2025-11-03T16:59:26.3339551Z Entering 'third_party/pocketfft' 2025-11-03T16:59:26.3393667Z Entering 'third_party/protobuf' 2025-11-03T16:59:26.3448962Z Entering 'third_party/protobuf/third_party/benchmark' 2025-11-03T16:59:26.3502635Z Entering 'third_party/protobuf/third_party/googletest' 2025-11-03T16:59:26.3558745Z Entering 'third_party/psimd' 2025-11-03T16:59:26.3612326Z Entering 'third_party/pthreadpool' 2025-11-03T16:59:26.3668935Z Entering 'third_party/pybind11' 2025-11-03T16:59:26.3723816Z Entering 'third_party/python-peachpy' 2025-11-03T16:59:26.3778038Z Entering 'third_party/sleef' 2025-11-03T16:59:26.3833758Z Entering 'third_party/tensorpipe' 2025-11-03T16:59:26.3884048Z Entering 'third_party/tensorpipe/third_party/googletest' 2025-11-03T16:59:26.3938339Z Entering 'third_party/tensorpipe/third_party/libnop' 2025-11-03T16:59:26.3993031Z Entering 'third_party/tensorpipe/third_party/libuv' 2025-11-03T16:59:26.4046494Z Entering 'third_party/tensorpipe/third_party/pybind11' 2025-11-03T16:59:26.4097867Z Entering 'third_party/tensorpipe/third_party/pybind11/tools/clang' 2025-11-03T16:59:26.4171687Z [command]/usr/bin/git submodule foreach --recursive git config --local --add 'url.https://github.com/.insteadOf' 'org-21003710@github.com:' 2025-11-03T16:59:26.4530600Z Entering 'android/libs/fbjni' 2025-11-03T16:59:26.4585038Z Entering 'third_party/FP16' 2025-11-03T16:59:26.4638695Z Entering 'third_party/FXdiv' 2025-11-03T16:59:26.4693715Z Entering 'third_party/NNPACK' 2025-11-03T16:59:26.4749804Z Entering 'third_party/NVTX' 2025-11-03T16:59:26.4805177Z Entering 'third_party/VulkanMemoryAllocator' 2025-11-03T16:59:26.4859061Z Entering 'third_party/XNNPACK' 2025-11-03T16:59:26.4927295Z Entering 'third_party/aiter' 2025-11-03T16:59:26.4981568Z Entering 'third_party/aiter/3rdparty/composable_kernel' 2025-11-03T16:59:26.5045414Z Entering 'third_party/benchmark' 2025-11-03T16:59:26.5099472Z Entering 'third_party/composable_kernel' 2025-11-03T16:59:26.5159832Z Entering 'third_party/cpp-httplib' 2025-11-03T16:59:26.5222444Z Entering 'third_party/cpuinfo' 2025-11-03T16:59:26.5276090Z Entering 'third_party/cudnn_frontend' 2025-11-03T16:59:26.5330405Z Entering 'third_party/cutlass' 2025-11-03T16:59:26.5392554Z Entering 'third_party/fbgemm' 2025-11-03T16:59:26.5447765Z Entering 'third_party/fbgemm/external/asmjit' 2025-11-03T16:59:26.5503707Z Entering 'third_party/fbgemm/external/composable_kernel' 2025-11-03T16:59:26.5565298Z Entering 'third_party/fbgemm/external/cpuinfo' 2025-11-03T16:59:26.5619211Z Entering 'third_party/fbgemm/external/cutlass' 2025-11-03T16:59:26.5682029Z Entering 'third_party/fbgemm/external/googletest' 2025-11-03T16:59:26.5734033Z Entering 'third_party/fbgemm/external/hipify_torch' 2025-11-03T16:59:26.5784444Z Entering 'third_party/fbgemm/external/json' 2025-11-03T16:59:26.5839638Z Entering 'third_party/flash-attention' 2025-11-03T16:59:26.5893848Z Entering 'third_party/flash-attention/csrc/composable_kernel' 2025-11-03T16:59:26.5953029Z Entering 'third_party/flash-attention/csrc/cutlass' 2025-11-03T16:59:26.6016056Z Entering 'third_party/flatbuffers' 2025-11-03T16:59:26.6072765Z Entering 'third_party/fmt' 2025-11-03T16:59:26.6127262Z Entering 'third_party/gemmlowp/gemmlowp' 2025-11-03T16:59:26.6187120Z Entering 'third_party/gloo' 2025-11-03T16:59:26.6242295Z Entering 'third_party/googletest' 2025-11-03T16:59:26.6295677Z Entering 'third_party/ideep' 2025-11-03T16:59:26.6348517Z Entering 'third_party/ideep/mkl-dnn' 2025-11-03T16:59:26.6410219Z Entering 'third_party/ittapi' 2025-11-03T16:59:26.6472181Z Entering 'third_party/kineto' 2025-11-03T16:59:26.6525558Z Entering 'third_party/kineto/libkineto/third_party/dynolog' 2025-11-03T16:59:26.6578459Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/DCGM' 2025-11-03T16:59:26.6635939Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/cpr' 2025-11-03T16:59:26.6690478Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/fmt' 2025-11-03T16:59:26.6748909Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/gflags' 2025-11-03T16:59:26.6800474Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/gflags/doc' 2025-11-03T16:59:26.6857973Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/glog' 2025-11-03T16:59:26.6916772Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/googletest' 2025-11-03T16:59:26.6971262Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/json' 2025-11-03T16:59:26.7037459Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/pfs' 2025-11-03T16:59:26.7091369Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/prometheus-cpp' 2025-11-03T16:59:26.7145949Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/prometheus-cpp/3rdparty/civetweb' 2025-11-03T16:59:26.7203711Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/prometheus-cpp/3rdparty/googletest' 2025-11-03T16:59:26.7266103Z Entering 'third_party/kineto/libkineto/third_party/fmt' 2025-11-03T16:59:26.7318941Z Entering 'third_party/kineto/libkineto/third_party/googletest' 2025-11-03T16:59:26.7375549Z Entering 'third_party/kleidiai' 2025-11-03T16:59:26.7431488Z Entering 'third_party/mimalloc' 2025-11-03T16:59:26.7485131Z Entering 'third_party/nlohmann' 2025-11-03T16:59:26.7544256Z Entering 'third_party/onnx' 2025-11-03T16:59:26.7611040Z Entering 'third_party/onnx/third_party/pybind11' 2025-11-03T16:59:26.7674278Z Entering 'third_party/opentelemetry-cpp' 2025-11-03T16:59:26.7738462Z Entering 'third_party/opentelemetry-cpp/third_party/benchmark' 2025-11-03T16:59:26.7781810Z Entering 'third_party/opentelemetry-cpp/third_party/googletest' 2025-11-03T16:59:26.7834846Z Entering 'third_party/opentelemetry-cpp/third_party/ms-gsl' 2025-11-03T16:59:26.7887131Z Entering 'third_party/opentelemetry-cpp/third_party/nlohmann-json' 2025-11-03T16:59:26.7938705Z Entering 'third_party/opentelemetry-cpp/third_party/opentelemetry-proto' 2025-11-03T16:59:26.7991966Z Entering 'third_party/opentelemetry-cpp/third_party/opentracing-cpp' 2025-11-03T16:59:26.8046419Z Entering 'third_party/opentelemetry-cpp/third_party/prometheus-cpp' 2025-11-03T16:59:26.8097364Z Entering 'third_party/opentelemetry-cpp/third_party/prometheus-cpp/3rdparty/civetweb' 2025-11-03T16:59:26.8152717Z Entering 'third_party/opentelemetry-cpp/third_party/prometheus-cpp/3rdparty/googletest' 2025-11-03T16:59:26.8208405Z Entering 'third_party/opentelemetry-cpp/tools/vcpkg' 2025-11-03T16:59:26.8282035Z Entering 'third_party/pocketfft' 2025-11-03T16:59:26.8335222Z Entering 'third_party/protobuf' 2025-11-03T16:59:26.8392257Z Entering 'third_party/protobuf/third_party/benchmark' 2025-11-03T16:59:26.8443228Z Entering 'third_party/protobuf/third_party/googletest' 2025-11-03T16:59:26.8497231Z Entering 'third_party/psimd' 2025-11-03T16:59:26.8551333Z Entering 'third_party/pthreadpool' 2025-11-03T16:59:26.8604427Z Entering 'third_party/pybind11' 2025-11-03T16:59:26.8655990Z Entering 'third_party/python-peachpy' 2025-11-03T16:59:26.8709254Z Entering 'third_party/sleef' 2025-11-03T16:59:26.8763024Z Entering 'third_party/tensorpipe' 2025-11-03T16:59:26.8816039Z Entering 'third_party/tensorpipe/third_party/googletest' 2025-11-03T16:59:26.8871667Z Entering 'third_party/tensorpipe/third_party/libnop' 2025-11-03T16:59:26.8924350Z Entering 'third_party/tensorpipe/third_party/libuv' 2025-11-03T16:59:26.8976298Z Entering 'third_party/tensorpipe/third_party/pybind11' 2025-11-03T16:59:26.9034746Z Entering 'third_party/tensorpipe/third_party/pybind11/tools/clang' 2025-11-03T16:59:26.9108953Z ##[endgroup] 2025-11-03T16:59:26.9152865Z [command]/usr/bin/git log -1 --format=%H 2025-11-03T16:59:26.9177577Z 3f6538febd727b782e6e13cfd026a309fb14351d 2025-11-03T16:59:26.9287381Z ##[group]Run cd "${GITHUB_WORKSPACE}" 2025-11-03T16:59:26.9287808Z cd "${GITHUB_WORKSPACE}" 2025-11-03T16:59:26.9288177Z # Clean stale submodule dirs 2025-11-03T16:59:26.9288590Z if [ -z "${NO_SUDO}" ]; then 2025-11-03T16:59:26.9289046Z  sudo git submodule foreach --recursive git clean -ffdx 2025-11-03T16:59:26.9289521Z else 2025-11-03T16:59:26.9289885Z  git submodule foreach --recursive git clean -ffdx 2025-11-03T16:59:26.9290309Z fi 2025-11-03T16:59:26.9305381Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2025-11-03T16:59:26.9305856Z env: 2025-11-03T16:59:26.9306132Z GIT_DEFAULT_BRANCH: main 2025-11-03T16:59:26.9306453Z NO_SUDO: true 2025-11-03T16:59:26.9306707Z ##[endgroup] 2025-11-03T16:59:26.9699011Z Entering 'android/libs/fbjni' 2025-11-03T16:59:26.9743302Z Entering 'third_party/FP16' 2025-11-03T16:59:26.9782573Z Entering 'third_party/FXdiv' 2025-11-03T16:59:26.9821804Z Entering 'third_party/NNPACK' 2025-11-03T16:59:26.9868571Z Entering 'third_party/NVTX' 2025-11-03T16:59:26.9919201Z Entering 'third_party/VulkanMemoryAllocator' 2025-11-03T16:59:26.9962542Z Entering 'third_party/XNNPACK' 2025-11-03T16:59:27.0100506Z Entering 'third_party/aiter' 2025-11-03T16:59:27.0153898Z Entering 'third_party/aiter/3rdparty/composable_kernel' 2025-11-03T16:59:27.0292459Z Entering 'third_party/benchmark' 2025-11-03T16:59:27.0334784Z Entering 'third_party/composable_kernel' 2025-11-03T16:59:27.0479847Z Entering 'third_party/cpp-httplib' 2025-11-03T16:59:27.0523107Z Entering 'third_party/cpuinfo' 2025-11-03T16:59:27.0570624Z Entering 'third_party/cudnn_frontend' 2025-11-03T16:59:27.0622962Z Entering 'third_party/cutlass' 2025-11-03T16:59:27.0741682Z Entering 'third_party/fbgemm' 2025-11-03T16:59:27.0816600Z Entering 'third_party/fbgemm/external/asmjit' 2025-11-03T16:59:27.0859505Z Entering 'third_party/fbgemm/external/composable_kernel' 2025-11-03T16:59:27.1002159Z Entering 'third_party/fbgemm/external/cpuinfo' 2025-11-03T16:59:27.1050287Z Entering 'third_party/fbgemm/external/cutlass' 2025-11-03T16:59:27.1175325Z Entering 'third_party/fbgemm/external/googletest' 2025-11-03T16:59:27.1219070Z Entering 'third_party/fbgemm/external/hipify_torch' 2025-11-03T16:59:27.1261333Z Entering 'third_party/fbgemm/external/json' 2025-11-03T16:59:27.1317962Z Entering 'third_party/flash-attention' 2025-11-03T16:59:27.1366461Z Entering 'third_party/flash-attention/csrc/composable_kernel' 2025-11-03T16:59:27.1487242Z Entering 'third_party/flash-attention/csrc/cutlass' 2025-11-03T16:59:27.1602393Z Entering 'third_party/flatbuffers' 2025-11-03T16:59:27.1691819Z Entering 'third_party/fmt' 2025-11-03T16:59:27.1734413Z Entering 'third_party/gemmlowp/gemmlowp' 2025-11-03T16:59:27.1778753Z Entering 'third_party/gloo' 2025-11-03T16:59:27.1823131Z Entering 'third_party/googletest' 2025-11-03T16:59:27.1870031Z Entering 'third_party/ideep' 2025-11-03T16:59:27.1909333Z Entering 'third_party/ideep/mkl-dnn' 2025-11-03T16:59:27.2011398Z Entering 'third_party/ittapi' 2025-11-03T16:59:27.2060940Z Entering 'third_party/kineto' 2025-11-03T16:59:27.2105772Z Entering 'third_party/kineto/libkineto/third_party/dynolog' 2025-11-03T16:59:27.2153260Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/DCGM' 2025-11-03T16:59:27.2212565Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/cpr' 2025-11-03T16:59:27.2255160Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/fmt' 2025-11-03T16:59:27.2298293Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/gflags' 2025-11-03T16:59:27.2343426Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/gflags/doc' 2025-11-03T16:59:27.2385280Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/glog' 2025-11-03T16:59:27.2431905Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/googletest' 2025-11-03T16:59:27.2475524Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/json' 2025-11-03T16:59:27.2525365Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/pfs' 2025-11-03T16:59:27.2565319Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/prometheus-cpp' 2025-11-03T16:59:27.2604703Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/prometheus-cpp/3rdparty/civetweb' 2025-11-03T16:59:27.2665149Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/prometheus-cpp/3rdparty/googletest' 2025-11-03T16:59:27.2715688Z Entering 'third_party/kineto/libkineto/third_party/fmt' 2025-11-03T16:59:27.2758802Z Entering 'third_party/kineto/libkineto/third_party/googletest' 2025-11-03T16:59:27.2806722Z Entering 'third_party/kleidiai' 2025-11-03T16:59:27.2855372Z Entering 'third_party/mimalloc' 2025-11-03T16:59:27.2899683Z Entering 'third_party/nlohmann' 2025-11-03T16:59:27.2967017Z Entering 'third_party/onnx' 2025-11-03T16:59:27.3397628Z Entering 'third_party/onnx/third_party/pybind11' 2025-11-03T16:59:27.3448029Z Entering 'third_party/opentelemetry-cpp' 2025-11-03T16:59:27.3521856Z Entering 'third_party/opentelemetry-cpp/third_party/benchmark' 2025-11-03T16:59:27.3563760Z Entering 'third_party/opentelemetry-cpp/third_party/googletest' 2025-11-03T16:59:27.3604823Z Entering 'third_party/opentelemetry-cpp/third_party/ms-gsl' 2025-11-03T16:59:27.3644486Z Entering 'third_party/opentelemetry-cpp/third_party/nlohmann-json' 2025-11-03T16:59:27.3697183Z Entering 'third_party/opentelemetry-cpp/third_party/opentelemetry-proto' 2025-11-03T16:59:27.3745947Z Entering 'third_party/opentelemetry-cpp/third_party/opentracing-cpp' 2025-11-03T16:59:27.3789895Z Entering 'third_party/opentelemetry-cpp/third_party/prometheus-cpp' 2025-11-03T16:59:27.3832410Z Entering 'third_party/opentelemetry-cpp/third_party/prometheus-cpp/3rdparty/civetweb' 2025-11-03T16:59:27.3892557Z Entering 'third_party/opentelemetry-cpp/third_party/prometheus-cpp/3rdparty/googletest' 2025-11-03T16:59:27.3939270Z Entering 'third_party/opentelemetry-cpp/tools/vcpkg' 2025-11-03T16:59:27.4273684Z Entering 'third_party/pocketfft' 2025-11-03T16:59:27.4313454Z Entering 'third_party/protobuf' 2025-11-03T16:59:27.4412543Z Entering 'third_party/protobuf/third_party/benchmark' 2025-11-03T16:59:27.4452814Z Entering 'third_party/protobuf/third_party/googletest' 2025-11-03T16:59:27.4506885Z Entering 'third_party/psimd' 2025-11-03T16:59:27.4546473Z Entering 'third_party/pthreadpool' 2025-11-03T16:59:27.4584609Z Entering 'third_party/pybind11' 2025-11-03T16:59:27.4630316Z Entering 'third_party/python-peachpy' 2025-11-03T16:59:27.4672670Z Entering 'third_party/sleef' 2025-11-03T16:59:27.4721521Z Entering 'third_party/tensorpipe' 2025-11-03T16:59:27.4764966Z Entering 'third_party/tensorpipe/third_party/googletest' 2025-11-03T16:59:27.4805814Z Entering 'third_party/tensorpipe/third_party/libnop' 2025-11-03T16:59:27.4844095Z Entering 'third_party/tensorpipe/third_party/libuv' 2025-11-03T16:59:27.4892906Z Entering 'third_party/tensorpipe/third_party/pybind11' 2025-11-03T16:59:27.4933469Z Entering 'third_party/tensorpipe/third_party/pybind11/tools/clang' 2025-11-03T16:59:27.5080458Z Prepare all required actions 2025-11-03T16:59:27.5080915Z Getting action download info 2025-11-03T16:59:27.6468398Z ##[group]Run ./.github/actions/setup-linux 2025-11-03T16:59:27.6468630Z env: 2025-11-03T16:59:27.6468784Z GIT_DEFAULT_BRANCH: main 2025-11-03T16:59:27.6468968Z ##[endgroup] 2025-11-03T16:59:27.6506057Z ##[group]Run set -euo pipefail 2025-11-03T16:59:27.6506324Z set -euo pipefail 2025-11-03T16:59:27.6506544Z function get_ec2_metadata() { 2025-11-03T16:59:27.6506814Z  # Pulled from instance metadata endpoint for EC2 2025-11-03T16:59:27.6507281Z  # see https://docs.aws.amazon.com/AWSEC2/latest/UserGuide/instancedata-data-retrieval.html 2025-11-03T16:59:27.6507698Z  category=$1 2025-11-03T16:59:27.6507963Z  # If it is GCP runner (runner name contains gcp), do not run this 2025-11-03T16:59:27.6508273Z  runner_name_str=i-0aa7ffe8f5edfd321 2025-11-03T16:59:27.6508550Z  if [[ -f /.inarc ]]; then 2025-11-03T16:59:27.6508799Z  echo "ARC Runner, no info on ec2 metadata" 2025-11-03T16:59:27.6509080Z  elif [[ $runner_name_str == *"gcp"* ]]; then 2025-11-03T16:59:27.6509411Z  echo "Runner is from Google Cloud Platform, No info on ec2 metadata" 2025-11-03T16:59:27.6509722Z  else 2025-11-03T16:59:27.6510328Z  curl -H "X-aws-ec2-metadata-token: $(curl -s -X PUT "http://169.254.169.254/latest/api/token" -H "X-aws-ec2-metadata-token-ttl-seconds: 30")" -fsSL "http://169.254.169.254/latest/meta-data/${category}" 2025-11-03T16:59:27.6510948Z  fi 2025-11-03T16:59:27.6511100Z } 2025-11-03T16:59:27.6511288Z echo "ami-id: $(get_ec2_metadata ami-id)" 2025-11-03T16:59:27.6511591Z echo "instance-id: $(get_ec2_metadata instance-id)" 2025-11-03T16:59:27.6511953Z echo "instance-type: $(get_ec2_metadata instance-type)" 2025-11-03T16:59:27.6512240Z echo "system info $(uname -a)" 2025-11-03T16:59:27.6520113Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2025-11-03T16:59:27.6520406Z env: 2025-11-03T16:59:27.6520566Z GIT_DEFAULT_BRANCH: main 2025-11-03T16:59:27.6520762Z ##[endgroup] 2025-11-03T16:59:27.6677756Z ami-id: ami-08982f1c5bf93d976 2025-11-03T16:59:27.6787569Z instance-id: i-0aa7ffe8f5edfd321 2025-11-03T16:59:27.6900333Z instance-type: g6.4xlarge 2025-11-03T16:59:27.6915737Z system info Linux ip-10-0-2-76.ec2.internal 6.1.150-174.273.amzn2023.x86_64 #1 SMP PREEMPT_DYNAMIC Tue Sep 9 12:21:26 UTC 2025 x86_64 x86_64 x86_64 GNU/Linux 2025-11-03T16:59:27.6939333Z ##[group]Run if [ -f /usr/bin/nvidia-smi ]; then nvidia-smi; fi 2025-11-03T16:59:27.6939694Z if [ -f /usr/bin/nvidia-smi ]; then nvidia-smi; fi 2025-11-03T16:59:27.6947648Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2025-11-03T16:59:27.6947925Z env: 2025-11-03T16:59:27.6948095Z GIT_DEFAULT_BRANCH: main 2025-11-03T16:59:27.6948289Z ##[endgroup] 2025-11-03T16:59:29.1456298Z Mon Nov 3 16:59:29 2025 2025-11-03T16:59:29.1457006Z +-----------------------------------------------------------------------------------------+ 2025-11-03T16:59:29.1457624Z | NVIDIA-SMI 580.82.07 Driver Version: 580.82.07 CUDA Version: 13.0 | 2025-11-03T16:59:29.1458090Z +-----------------------------------------+------------------------+----------------------+ 2025-11-03T16:59:29.1458554Z | GPU Name Persistence-M | Bus-Id Disp.A | Volatile Uncorr. ECC | 2025-11-03T16:59:29.1459043Z | Fan Temp Perf Pwr:Usage/Cap | Memory-Usage | GPU-Util Compute M. | 2025-11-03T16:59:29.1459437Z | | | MIG M. | 2025-11-03T16:59:29.1459738Z |=========================================+========================+======================| 2025-11-03T16:59:29.1535511Z | 0 NVIDIA L4 Off | 00000000:35:00.0 Off | 0 | 2025-11-03T16:59:29.1536336Z | N/A 41C P0 29W / 72W | 0MiB / 23034MiB | 4% Default | 2025-11-03T16:59:29.1536890Z | | | N/A | 2025-11-03T16:59:29.1537387Z +-----------------------------------------+------------------------+----------------------+ 2025-11-03T16:59:29.1537794Z 2025-11-03T16:59:29.1537960Z +-----------------------------------------------------------------------------------------+ 2025-11-03T16:59:29.1538362Z | Processes: | 2025-11-03T16:59:29.1538765Z | GPU GI CI PID Type Process name GPU Memory | 2025-11-03T16:59:29.1539136Z | ID ID Usage | 2025-11-03T16:59:29.1539437Z |=========================================================================================| 2025-11-03T16:59:29.1540008Z | No running processes found | 2025-11-03T16:59:29.1540465Z +-----------------------------------------------------------------------------------------+ 2025-11-03T16:59:29.4797502Z ##[group]Run echo "IN_CONTAINER_RUNNER=$(if [ -f /.inarc ] || [ -f /.incontainer ]; then echo true ; else echo false; fi)" >> "$GITHUB_OUTPUT" 2025-11-03T16:59:29.4798177Z echo "IN_CONTAINER_RUNNER=$(if [ -f /.inarc ] || [ -f /.incontainer ]; then echo true ; else echo false; fi)" >> "$GITHUB_OUTPUT" 2025-11-03T16:59:29.4807450Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2025-11-03T16:59:29.4807730Z env: 2025-11-03T16:59:29.4807892Z GIT_DEFAULT_BRANCH: main 2025-11-03T16:59:29.4808076Z ##[endgroup] 2025-11-03T16:59:29.4877687Z ##[group]Run if systemctl is-active --quiet docker; then 2025-11-03T16:59:29.4878013Z if systemctl is-active --quiet docker; then 2025-11-03T16:59:29.4878308Z  echo "Docker daemon is running..."; 2025-11-03T16:59:29.4878561Z else 2025-11-03T16:59:29.4878816Z  echo "Starting docker daemon..." && sudo systemctl start docker; 2025-11-03T16:59:29.4879124Z fi 2025-11-03T16:59:29.4886706Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2025-11-03T16:59:29.4886981Z env: 2025-11-03T16:59:29.4887146Z GIT_DEFAULT_BRANCH: main 2025-11-03T16:59:29.4887341Z ##[endgroup] 2025-11-03T16:59:29.4998778Z Docker daemon is running... 2025-11-03T16:59:29.5037793Z ##[group]Run nick-fields/retry@v3.0.0 2025-11-03T16:59:29.5038013Z with: 2025-11-03T16:59:29.5038157Z shell: bash 2025-11-03T16:59:29.5038319Z timeout_minutes: 5 2025-11-03T16:59:29.5038493Z max_attempts: 3 2025-11-03T16:59:29.5038662Z retry_wait_seconds: 30 2025-11-03T16:59:29.5040263Z command: AWS_ACCOUNT_ID=$(aws sts get-caller-identity|grep Account|cut -f4 -d\") aws ecr get-login-password --region "$AWS_DEFAULT_REGION" | docker login --username AWS \ --password-stdin "$AWS_ACCOUNT_ID.dkr.ecr.$AWS_DEFAULT_REGION.amazonaws.com" # For LF Runners we need to make sure we also login to Meta's ECR docker registry too. META_AWS_ACCOUNT_ID=308535385114 if [ "$AWS_ACCOUNT_ID" != "$META_AWS_ACCOUNT_ID" ] ; then aws ecr get-login-password --region "$AWS_DEFAULT_REGION" | docker login --username AWS \ --password-stdin "$META_AWS_ACCOUNT_ID.dkr.ecr.$AWS_DEFAULT_REGION.amazonaws.com" fi 2025-11-03T16:59:29.5041867Z polling_interval_seconds: 1 2025-11-03T16:59:29.5042068Z warning_on_retry: true 2025-11-03T16:59:29.5042246Z continue_on_error: false 2025-11-03T16:59:29.5042423Z env: 2025-11-03T16:59:29.5042582Z GIT_DEFAULT_BRANCH: main 2025-11-03T16:59:29.5042766Z AWS_RETRY_MODE: standard 2025-11-03T16:59:29.5042941Z AWS_MAX_ATTEMPTS: 5 2025-11-03T16:59:29.5043128Z AWS_DEFAULT_REGION: us-east-1 2025-11-03T16:59:29.5043321Z ##[endgroup] 2025-11-03T16:59:30.5274456Z WARNING! Your password will be stored unencrypted in /home/ec2-user/.docker/config.json. 2025-11-03T16:59:30.5275644Z Configure a credential helper to remove this warning. See 2025-11-03T16:59:30.5276405Z https://docs.docker.com/engine/reference/commandline/login/#credentials-store 2025-11-03T16:59:30.5276977Z 2025-11-03T16:59:30.5277067Z Login Succeeded 2025-11-03T16:59:30.5825471Z Command completed after 1 attempt(s). 2025-11-03T16:59:30.5903219Z ##[group]Run env | grep '^GITHUB' >> "/tmp/github_env_${GITHUB_RUN_ID}" 2025-11-03T16:59:30.5903614Z env | grep '^GITHUB' >> "/tmp/github_env_${GITHUB_RUN_ID}" 2025-11-03T16:59:30.5903945Z env | grep '^CI' >> "/tmp/github_env_${GITHUB_RUN_ID}" 2025-11-03T16:59:30.5913436Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2025-11-03T16:59:30.5913713Z env: 2025-11-03T16:59:30.5913872Z GIT_DEFAULT_BRANCH: main 2025-11-03T16:59:30.5914056Z ##[endgroup] 2025-11-03T16:59:30.6015186Z ##[group]Run # ignore expansion of "docker ps -q" since it could be empty 2025-11-03T16:59:30.6015602Z # ignore expansion of "docker ps -q" since it could be empty 2025-11-03T16:59:30.6015922Z # shellcheck disable=SC2046 2025-11-03T16:59:30.6016174Z docker stop $(docker ps -q) || true 2025-11-03T16:59:30.6016428Z # Prune all of the docker images 2025-11-03T16:59:30.6016674Z docker system prune -af 2025-11-03T16:59:30.6024418Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2025-11-03T16:59:30.6024701Z env: 2025-11-03T16:59:30.6024868Z GIT_DEFAULT_BRANCH: main 2025-11-03T16:59:30.6025059Z ##[endgroup] 2025-11-03T16:59:30.6329043Z "docker stop" requires at least 1 argument. 2025-11-03T16:59:30.6329450Z See 'docker stop --help'. 2025-11-03T16:59:30.6329626Z 2025-11-03T16:59:30.6329785Z Usage: docker stop [OPTIONS] CONTAINER [CONTAINER...] 2025-11-03T16:59:30.6330034Z 2025-11-03T16:59:30.6330136Z Stop one or more running containers 2025-11-03T16:59:30.6685031Z Total reclaimed space: 0B 2025-11-03T16:59:30.6846487Z ##[group]Run pytorch/test-infra/.github/actions/calculate-docker-image@main 2025-11-03T16:59:30.6846860Z with: 2025-11-03T16:59:30.6847466Z docker-image-name: 308535385114.dkr.ecr.us-east-1.amazonaws.com/pytorch/ci-image:pytorch-linux-jammy-cuda12.8-cudnn9-py3.12-gcc11-vllm-30147fb242f88ab8f13e14b76ab186d189d8605c 2025-11-03T16:59:30.6848155Z use-custom-docker-registry: true 2025-11-03T16:59:30.6848377Z docker-build-dir: .ci/docker 2025-11-03T16:59:30.6848594Z docker-build-script: ./build.sh 2025-11-03T16:59:30.6848809Z working-directory: . 2025-11-03T16:59:30.6849065Z docker-registry: 308535385114.dkr.ecr.us-east-1.amazonaws.com 2025-11-03T16:59:30.6849347Z force-push: false 2025-11-03T16:59:30.6849514Z env: 2025-11-03T16:59:30.6849665Z GIT_DEFAULT_BRANCH: main 2025-11-03T16:59:30.6849846Z ##[endgroup] 2025-11-03T16:59:30.6876976Z ##[group]Run set -ex 2025-11-03T16:59:30.6877199Z set -ex 2025-11-03T16:59:30.6877363Z  2025-11-03T16:59:30.6877671Z # If the docker build directory or the build script doesn't exist, the action will 2025-11-03T16:59:30.6878140Z # gracefully return the docker image name as it is. Pulling docker image in Linux 2025-11-03T16:59:30.6878527Z # job could then download the pre-built image as usual 2025-11-03T16:59:30.6879016Z if [[ -d "${DOCKER_BUILD_DIR}" ]] && [[ -f "${DOCKER_BUILD_DIR}/${DOCKER_BUILD_SCRIPT}" ]] && [[ "${USE_CUSTOM_DOCKER_REGISTRY}" == "true" ]]; then 2025-11-03T16:59:30.6879456Z  echo "skip=false" >> "${GITHUB_OUTPUT}" 2025-11-03T16:59:30.6879684Z else 2025-11-03T16:59:30.6879886Z  echo "skip=true" >> "${GITHUB_OUTPUT}" 2025-11-03T16:59:30.6880192Z  echo "docker-image=${DOCKER_IMAGE_NAME}" >> "${GITHUB_OUTPUT}" 2025-11-03T16:59:30.6880478Z  2025-11-03T16:59:30.6880871Z  echo "Not using custom ECR registry. Either it was not requested or there is no Docker build script in the ${REPO_NAME} repo..." 2025-11-03T16:59:30.6881314Z  exit 0 2025-11-03T16:59:30.6881465Z fi 2025-11-03T16:59:30.6881614Z  2025-11-03T16:59:30.6881849Z if [[ "${DOCKER_IMAGE_NAME}" == *"${DOCKER_REGISTRY}/${REPO_NAME}"* ]]; then 2025-11-03T16:59:30.6882443Z  # The docker image name already includes the ECR prefix and tag, so we can just 2025-11-03T16:59:30.6882818Z  # use it as it is, but first let's extract the tag 2025-11-03T16:59:30.6883149Z  DOCKER_TAG=$(echo "${DOCKER_IMAGE_NAME}" | awk -F '[:,]' '{print $2}') 2025-11-03T16:59:30.6883506Z  echo "docker-tag=${DOCKER_TAG}" >> "${GITHUB_OUTPUT}" 2025-11-03T16:59:30.6883860Z  echo "docker-image=${DOCKER_IMAGE_NAME}" >> "${GITHUB_OUTPUT}" 2025-11-03T16:59:30.6884194Z else 2025-11-03T16:59:30.6884448Z  if [[ "${DOCKER_IMAGE_NAME}" == *:* ]]; then 2025-11-03T16:59:30.6884766Z  CUSTOM_TAG_PREFIX=${DOCKER_IMAGE_NAME#*:} 2025-11-03T16:59:30.6885100Z  DOCKER_IMAGE_NAME=${DOCKER_IMAGE_NAME%%:*} 2025-11-03T16:59:30.6885388Z  fi 2025-11-03T16:59:30.6885730Z  DOCKER_TAG=${CUSTOM_TAG_PREFIX:+${CUSTOM_TAG_PREFIX}-}$(git rev-parse HEAD:"${DOCKER_BUILD_DIR}") 2025-11-03T16:59:30.6886151Z  echo "docker-tag=${DOCKER_TAG}" >> "${GITHUB_OUTPUT}" 2025-11-03T16:59:30.6886603Z  echo "docker-image=${DOCKER_REGISTRY}/${REPO_NAME}/${DOCKER_IMAGE_NAME}:${DOCKER_TAG}" >> "${GITHUB_OUTPUT}" 2025-11-03T16:59:30.6887087Z  echo "custom-tag-prefix=${CUSTOM_TAG_PREFIX}" >> "${GITHUB_OUTPUT}" 2025-11-03T16:59:30.6887387Z fi 2025-11-03T16:59:30.6895906Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2025-11-03T16:59:30.6896181Z env: 2025-11-03T16:59:30.6896339Z GIT_DEFAULT_BRANCH: main 2025-11-03T16:59:30.6896529Z REPO_NAME: pytorch 2025-11-03T16:59:30.6897285Z DOCKER_IMAGE_NAME: 308535385114.dkr.ecr.us-east-1.amazonaws.com/pytorch/ci-image:pytorch-linux-jammy-cuda12.8-cudnn9-py3.12-gcc11-vllm-30147fb242f88ab8f13e14b76ab186d189d8605c 2025-11-03T16:59:30.6897947Z DOCKER_BUILD_DIR: .ci/docker 2025-11-03T16:59:30.6898158Z DOCKER_BUILD_SCRIPT: ./build.sh 2025-11-03T16:59:30.6898433Z DOCKER_REGISTRY: 308535385114.dkr.ecr.us-east-1.amazonaws.com 2025-11-03T16:59:30.6898740Z USE_CUSTOM_DOCKER_REGISTRY: true 2025-11-03T16:59:30.6898941Z CUSTOM_TAG_PREFIX: 2025-11-03T16:59:30.6899116Z ##[endgroup] 2025-11-03T16:59:30.6929678Z + [[ -d .ci/docker ]] 2025-11-03T16:59:30.6929982Z + [[ -f .ci/docker/./build.sh ]] 2025-11-03T16:59:30.6930249Z + [[ true == \t\r\u\e ]] 2025-11-03T16:59:30.6930485Z + echo skip=false 2025-11-03T16:59:30.6931507Z + [[ 308535385114.dkr.ecr.us-east-1.amazonaws.com/pytorch/ci-image:pytorch-linux-jammy-cuda12.8-cudnn9-py3.12-gcc11-vllm-30147fb242f88ab8f13e14b76ab186d189d8605c == *\3\0\8\5\3\5\3\8\5\1\1\4\.\d\k\r\.\e\c\r\.\u\s\-\e\a\s\t\-\1\.\a\m\a\z\o\n\a\w\s\.\c\o\m\/\p\y\t\o\r\c\h* ]] 2025-11-03T16:59:30.6938300Z ++ echo 308535385114.dkr.ecr.us-east-1.amazonaws.com/pytorch/ci-image:pytorch-linux-jammy-cuda12.8-cudnn9-py3.12-gcc11-vllm-30147fb242f88ab8f13e14b76ab186d189d8605c 2025-11-03T16:59:30.6939424Z ++ awk -F '[:,]' '{print $2}' 2025-11-03T16:59:30.6963452Z + DOCKER_TAG=pytorch-linux-jammy-cuda12.8-cudnn9-py3.12-gcc11-vllm-30147fb242f88ab8f13e14b76ab186d189d8605c 2025-11-03T16:59:30.6964822Z + echo docker-tag=pytorch-linux-jammy-cuda12.8-cudnn9-py3.12-gcc11-vllm-30147fb242f88ab8f13e14b76ab186d189d8605c 2025-11-03T16:59:30.6966132Z + echo docker-image=308535385114.dkr.ecr.us-east-1.amazonaws.com/pytorch/ci-image:pytorch-linux-jammy-cuda12.8-cudnn9-py3.12-gcc11-vllm-30147fb242f88ab8f13e14b76ab186d189d8605c 2025-11-03T16:59:30.6996824Z ##[group]Run set +e 2025-11-03T16:59:30.6997042Z set +e 2025-11-03T16:59:30.6997214Z set -x 2025-11-03T16:59:30.6997376Z  2025-11-03T16:59:30.6997522Z login() { 2025-11-03T16:59:30.6997879Z  aws ecr get-login-password --region us-east-1 | docker login -u AWS --password-stdin "$1" 2025-11-03T16:59:30.6998243Z } 2025-11-03T16:59:30.6998391Z  2025-11-03T16:59:30.6998544Z retry () { 2025-11-03T16:59:30.6998738Z  $* || (sleep 1 && $*) || (sleep 2 && $*) 2025-11-03T16:59:30.6998963Z } 2025-11-03T16:59:30.6999284Z  2025-11-03T16:59:30.6999447Z retry login "${DOCKER_REGISTRY}" 2025-11-03T16:59:30.6999670Z  2025-11-03T16:59:30.6999825Z START_TIME=$(date +%s) 2025-11-03T16:59:30.7000034Z # Wait up to 120 minutes 2025-11-03T16:59:30.7000300Z while [[ $(( $(date +%s) - 7200 )) -lt $START_TIME ]]; do 2025-11-03T16:59:30.7000644Z  # Check if image already exists, if it does then skip building it 2025-11-03T16:59:30.7000994Z  if docker manifest inspect "${DOCKER_IMAGE}"; then 2025-11-03T16:59:30.7001252Z  exit 0 2025-11-03T16:59:30.7001424Z  fi 2025-11-03T16:59:30.7001571Z  2025-11-03T16:59:30.7001852Z  # NB: This flag is used by Docker build workflow to push the image to ECR, so we can 2025-11-03T16:59:30.7002338Z  # use this to differentiate between the Docker build and regular build jobs. For the 2025-11-03T16:59:30.7002808Z  # latter, it will wait for the Docker images to become available before continuing 2025-11-03T16:59:30.7003176Z  if [ "${DOCKER_PUSH:-false}" == "true" ]; then 2025-11-03T16:59:30.7003457Z  # It's a Docker build job, let's build the image 2025-11-03T16:59:30.7003719Z  break 2025-11-03T16:59:30.7003887Z  else 2025-11-03T16:59:30.7004129Z  # It's a regular build job, wait for the image to become available 2025-11-03T16:59:30.7004425Z  sleep 300 2025-11-03T16:59:30.7004602Z  fi 2025-11-03T16:59:30.7004754Z done 2025-11-03T16:59:30.7004903Z  2025-11-03T16:59:30.7005293Z # NB: This part requires a full checkout. Otherwise, the merge base will 2025-11-03T16:59:30.7005709Z # be empty. The default action would be to continue rebuild the image 2025-11-03T16:59:30.7006072Z if [[ "$BASE_REVISION" = "$(git rev-parse HEAD)" ]]; then 2025-11-03T16:59:30.7006393Z  # if we're on the base branch then use the parent commit 2025-11-03T16:59:30.7006693Z  MERGE_BASE=$(git rev-parse HEAD~) 2025-11-03T16:59:30.7006905Z else 2025-11-03T16:59:30.7007133Z  # otherwise we're on a PR, so use the most recent base commit 2025-11-03T16:59:30.7007468Z  MERGE_BASE=$(git merge-base HEAD "$BASE_REVISION") 2025-11-03T16:59:30.7007733Z fi 2025-11-03T16:59:30.7007878Z  2025-11-03T16:59:30.7008044Z if [[ -z "${MERGE_BASE}" ]]; then 2025-11-03T16:59:30.7008290Z  echo "rebuild=true" >> "${GITHUB_OUTPUT}" 2025-11-03T16:59:30.7008517Z  2025-11-03T16:59:30.7008834Z  echo "Finding merge base only works with full checkout, please set fetch-depth to 0, continuing ..." 2025-11-03T16:59:30.7009216Z  exit 0 2025-11-03T16:59:30.7009384Z fi 2025-11-03T16:59:30.7009529Z  2025-11-03T16:59:30.7009738Z if ! git rev-parse "${MERGE_BASE}:${DOCKER_BUILD_DIR}"; then 2025-11-03T16:59:30.7010213Z  echo "Directory '${DOCKER_BUILD_DIR}' not found in commit $MERGE_BASE, you should rebase onto a more recent commit" 2025-11-03T16:59:30.7010615Z  exit 1 2025-11-03T16:59:30.7010770Z fi 2025-11-03T16:59:30.7010913Z  2025-11-03T16:59:30.7011161Z PREVIOUS_DOCKER_TAG=$(git rev-parse "${MERGE_BASE}:${DOCKER_BUILD_DIR}") 2025-11-03T16:59:30.7011620Z # If no image exists but the hash is the same as the previous hash then we should error out here 2025-11-03T16:59:30.7012026Z if [[ "${PREVIOUS_DOCKER_TAG}" == "${DOCKER_TAG}" ]]; then 2025-11-03T16:59:30.7012500Z  echo "WARNING: Something has gone wrong and the previous image isn't available for the merge-base of your branch" 2025-11-03T16:59:30.7013024Z  echo " Will re-build docker image to store in local cache, TTS may be longer" 2025-11-03T16:59:30.7013330Z fi 2025-11-03T16:59:30.7013474Z  2025-11-03T16:59:30.7013752Z echo "rebuild=true" >> "${GITHUB_OUTPUT}" 2025-11-03T16:59:30.7020941Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2025-11-03T16:59:30.7021595Z env: 2025-11-03T16:59:30.7021755Z GIT_DEFAULT_BRANCH: main 2025-11-03T16:59:30.7021968Z DOCKER_BUILD_DIR: .ci/docker 2025-11-03T16:59:30.7022220Z BASE_REVISION: 3f6538febd727b782e6e13cfd026a309fb14351d 2025-11-03T16:59:30.7022909Z DOCKER_IMAGE: 308535385114.dkr.ecr.us-east-1.amazonaws.com/pytorch/ci-image:pytorch-linux-jammy-cuda12.8-cudnn9-py3.12-gcc11-vllm-30147fb242f88ab8f13e14b76ab186d189d8605c 2025-11-03T16:59:30.7023747Z DOCKER_TAG: pytorch-linux-jammy-cuda12.8-cudnn9-py3.12-gcc11-vllm-30147fb242f88ab8f13e14b76ab186d189d8605c 2025-11-03T16:59:30.7024260Z DOCKER_REGISTRY: 308535385114.dkr.ecr.us-east-1.amazonaws.com 2025-11-03T16:59:30.7024544Z DOCKER_PUSH: 2025-11-03T16:59:30.7024709Z ##[endgroup] 2025-11-03T16:59:30.7051629Z + retry login 308535385114.dkr.ecr.us-east-1.amazonaws.com 2025-11-03T16:59:30.7054661Z + login 308535385114.dkr.ecr.us-east-1.amazonaws.com 2025-11-03T16:59:30.7055159Z + aws ecr get-login-password --region us-east-1 2025-11-03T16:59:30.7055897Z + docker login -u AWS --password-stdin 308535385114.dkr.ecr.us-east-1.amazonaws.com 2025-11-03T16:59:31.1611512Z WARNING! Your password will be stored unencrypted in /home/ec2-user/.docker/config.json. 2025-11-03T16:59:31.1612421Z Configure a credential helper to remove this warning. See 2025-11-03T16:59:31.1612949Z https://docs.docker.com/engine/reference/commandline/login/#credentials-store 2025-11-03T16:59:31.1613302Z 2025-11-03T16:59:31.1613747Z Login Succeeded 2025-11-03T16:59:31.1634009Z ++ date +%s 2025-11-03T16:59:31.1646412Z + START_TIME=1762189171 2025-11-03T16:59:31.1650516Z ++ date +%s 2025-11-03T16:59:31.1662711Z + [[ 1762181971 -lt 1762189171 ]] 2025-11-03T16:59:31.1664065Z + docker manifest inspect 308535385114.dkr.ecr.us-east-1.amazonaws.com/pytorch/ci-image:pytorch-linux-jammy-cuda12.8-cudnn9-py3.12-gcc11-vllm-30147fb242f88ab8f13e14b76ab186d189d8605c 2025-11-03T16:59:31.4047036Z { 2025-11-03T16:59:31.4047512Z "schemaVersion": 2, 2025-11-03T16:59:31.4048045Z "mediaType": "application/vnd.docker.distribution.manifest.v2+json", 2025-11-03T16:59:31.4048553Z "config": { 2025-11-03T16:59:31.4048944Z "mediaType": "application/vnd.docker.container.image.v1+json", 2025-11-03T16:59:31.4049403Z "size": 32985, 2025-11-03T16:59:31.4049868Z "digest": "sha256:123a0a8d2d8124dd2cca825f1276a192d04ee365eb7de45a4829163bae42bcc5" 2025-11-03T16:59:31.4050404Z }, 2025-11-03T16:59:31.4050624Z "layers": [ 2025-11-03T16:59:31.4050862Z { 2025-11-03T16:59:31.4051235Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-11-03T16:59:31.4051721Z "size": 30447990, 2025-11-03T16:59:31.4052225Z "digest": "sha256:828c1365039a657352c737a62d13e1932951b5658eb6bd9b9096ea9b73562453" 2025-11-03T16:59:31.4052770Z }, 2025-11-03T16:59:31.4052990Z { 2025-11-03T16:59:31.4053373Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-11-03T16:59:31.4053890Z "size": 1552, 2025-11-03T16:59:31.4054378Z "digest": "sha256:1c4ee0d43392f96acf52d2e8e2f6257611a1334051823ff62d69c216dbc7daca" 2025-11-03T16:59:31.4054933Z }, 2025-11-03T16:59:31.4055164Z { 2025-11-03T16:59:31.4055537Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-11-03T16:59:31.4055995Z "size": 313626101, 2025-11-03T16:59:31.4056453Z "digest": "sha256:728590afcb21193f16499a4744dada5e256b1413ed8ced7560c4eb4543953d2b" 2025-11-03T16:59:31.4056994Z }, 2025-11-03T16:59:31.4057282Z { 2025-11-03T16:59:31.4057816Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-11-03T16:59:31.4058381Z "size": 791, 2025-11-03T16:59:31.4059174Z "digest": "sha256:9a925ed696bfcc8ebda91a3c220252d69635e24adf4fc2835afd1a4aa9e601f6" 2025-11-03T16:59:31.4059831Z }, 2025-11-03T16:59:31.4060117Z { 2025-11-03T16:59:31.4060715Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-11-03T16:59:31.4091872Z "size": 106, 2025-11-03T16:59:31.4092785Z "digest": "sha256:e7ab1efdd1bd8cfda48e301193b8e1eb058d2c83851044818db8099ff597ce35" 2025-11-03T16:59:31.4093377Z }, 2025-11-03T16:59:31.4093579Z { 2025-11-03T16:59:31.4093841Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-11-03T16:59:31.4094154Z "size": 703, 2025-11-03T16:59:31.4094453Z "digest": "sha256:33a0c5e93563ea73282e573325b1527d301f618dbec76bb05c07228cf1c69230" 2025-11-03T16:59:31.4094795Z }, 2025-11-03T16:59:31.4094937Z { 2025-11-03T16:59:31.4095177Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-11-03T16:59:31.4095471Z "size": 1219, 2025-11-03T16:59:31.4095780Z "digest": "sha256:7401e108f2b86e7530e0d9f8fa1dea4ae5f79c68810edceb2a7abdd1630ee9d6" 2025-11-03T16:59:31.4096112Z }, 2025-11-03T16:59:31.4096250Z { 2025-11-03T16:59:31.4096478Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-11-03T16:59:31.4096790Z "size": 484, 2025-11-03T16:59:31.4097093Z "digest": "sha256:9d9c7be466f87846f48df125fcf7b25bc5d633cefbf0584ed645d3ab90c043bd" 2025-11-03T16:59:31.4097490Z + exit 0 2025-11-03T16:59:31.4097635Z }, 2025-11-03T16:59:31.4097774Z { 2025-11-03T16:59:31.4098008Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-11-03T16:59:31.4098309Z "size": 110363285, 2025-11-03T16:59:31.4098610Z "digest": "sha256:faa9a1190c2e97d043159143a313b91d39d5a6068aee31e2f2708a4eac04afa1" 2025-11-03T16:59:31.4098949Z }, 2025-11-03T16:59:31.4099090Z { 2025-11-03T16:59:31.4099326Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-11-03T16:59:31.4099617Z "size": 5132, 2025-11-03T16:59:31.4100122Z "digest": "sha256:7c2887fa8aa0685c83bbb51790f6890622425f1e8bfc2aa5ae9a6156505842a9" 2025-11-03T16:59:31.4100475Z }, 2025-11-03T16:59:31.4100615Z { 2025-11-03T16:59:31.4100840Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-11-03T16:59:31.4101130Z "size": 1755, 2025-11-03T16:59:31.4101514Z "digest": "sha256:95bb0abcaee9af0c3aafb8021c45ce9381c0a1e2c8b4c6a394a49b98107f8102" 2025-11-03T16:59:31.4101869Z }, 2025-11-03T16:59:31.4102007Z { 2025-11-03T16:59:31.4102245Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-11-03T16:59:31.4102541Z "size": 724, 2025-11-03T16:59:31.4102839Z "digest": "sha256:b3a429f4faee83e3d7e616cbcca1b458319d213977b4807ea863fc416abb2229" 2025-11-03T16:59:31.4103170Z }, 2025-11-03T16:59:31.4103308Z { 2025-11-03T16:59:31.4103546Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-11-03T16:59:31.4103840Z "size": 544, 2025-11-03T16:59:31.4104126Z "digest": "sha256:f4cb10b91a16ac44a669c09f1037854ceb9884ac4c6a118da74679c3f2629f24" 2025-11-03T16:59:31.4104470Z }, 2025-11-03T16:59:31.4104614Z { 2025-11-03T16:59:31.4104852Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-11-03T16:59:31.4105143Z "size": 3365931269, 2025-11-03T16:59:31.4105449Z "digest": "sha256:48618db2b1a6e401f30542bcf94fe34cdac711cdc8811c56800d9b013220cf85" 2025-11-03T16:59:31.4105785Z }, 2025-11-03T16:59:31.4105925Z { 2025-11-03T16:59:31.4106144Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-11-03T16:59:31.4106429Z "size": 32, 2025-11-03T16:59:31.4106716Z "digest": "sha256:4f4fb700ef54461cfa02571ae0db9a0dc1e0cdb5577484a6d75e68dc38e8acc1" 2025-11-03T16:59:31.4107046Z }, 2025-11-03T16:59:31.4107174Z { 2025-11-03T16:59:31.4107397Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-11-03T16:59:31.4107681Z "size": 380, 2025-11-03T16:59:31.4107976Z "digest": "sha256:2ec7ba999b860bc831c8e5b4e42a1087c8cfde1d02d8f2b3b0a1a297def702b1" 2025-11-03T16:59:31.4108303Z }, 2025-11-03T16:59:31.4108442Z { 2025-11-03T16:59:31.4108670Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-11-03T16:59:31.4108959Z "size": 236652, 2025-11-03T16:59:31.4109251Z "digest": "sha256:85d9004dcaf2a14dc2db8fb4caf12cc30e40747d20860a246e7c99fce1029e90" 2025-11-03T16:59:31.4109714Z }, 2025-11-03T16:59:31.4109849Z { 2025-11-03T16:59:31.4110074Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-11-03T16:59:31.4110355Z "size": 230, 2025-11-03T16:59:31.4110639Z "digest": "sha256:df8f8108c158cd5519b0ea50bd22522f354b19bf2388184e23faa3733b3fd6e3" 2025-11-03T16:59:31.4110966Z }, 2025-11-03T16:59:31.4111102Z { 2025-11-03T16:59:31.4111317Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-11-03T16:59:31.4111603Z "size": 3267656, 2025-11-03T16:59:31.4111893Z "digest": "sha256:329fb7ac2c9ea86ede71036735c0b2f3b2f388129235a33e002409205f4514f1" 2025-11-03T16:59:31.4112216Z }, 2025-11-03T16:59:31.4112344Z { 2025-11-03T16:59:31.4112583Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-11-03T16:59:31.4112873Z "size": 1478, 2025-11-03T16:59:31.4113155Z "digest": "sha256:397c5718353c9270969b965bfddacd659394a4189be78c230558a674cbde1a2c" 2025-11-03T16:59:31.4113466Z }, 2025-11-03T16:59:31.4113606Z { 2025-11-03T16:59:31.4113825Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-11-03T16:59:31.4114113Z "size": 481, 2025-11-03T16:59:31.4114381Z "digest": "sha256:099139c9d8b455920b3e71991a0af82a00315f2708e1b8398089dad4889d0c82" 2025-11-03T16:59:31.4114697Z }, 2025-11-03T16:59:31.4114841Z { 2025-11-03T16:59:31.4115074Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-11-03T16:59:31.4115353Z "size": 198, 2025-11-03T16:59:31.4115635Z "digest": "sha256:c094f15f3938d9123ed1f3a086212ccf36d3a0471de7dbfdbe14a185e16cb362" 2025-11-03T16:59:31.4115959Z }, 2025-11-03T16:59:31.4116095Z { 2025-11-03T16:59:31.4116405Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-11-03T16:59:31.4116700Z "size": 607, 2025-11-03T16:59:31.4116997Z "digest": "sha256:a99b38a3dcdd8fe005f4f28ca62e73a4c465ac67ea320f9cfee2fcae35797245" 2025-11-03T16:59:31.4117330Z }, 2025-11-03T16:59:31.4117458Z { 2025-11-03T16:59:31.4117683Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-11-03T16:59:31.4117970Z "size": 7871224158, 2025-11-03T16:59:31.4118284Z "digest": "sha256:bc183ac25ebc0a82c062b5aa9d512109c3349be9dada7361068edf5814ac9a76" 2025-11-03T16:59:31.4118611Z }, 2025-11-03T16:59:31.4118746Z { 2025-11-03T16:59:31.4118970Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-11-03T16:59:31.4119253Z "size": 830, 2025-11-03T16:59:31.4119521Z "digest": "sha256:1a6a81cb899b0f9a55045ca3f89823d22816308be9440327b64990ae8108de42" 2025-11-03T16:59:31.4119851Z }, 2025-11-03T16:59:31.4119988Z { 2025-11-03T16:59:31.4120212Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-11-03T16:59:31.4120494Z "size": 33451730, 2025-11-03T16:59:31.4120789Z "digest": "sha256:41ec23fc0915ee04c6a7f6142343834517dcca5c446aa66557f147eafc52de95" 2025-11-03T16:59:31.4121439Z }, 2025-11-03T16:59:31.4121586Z { 2025-11-03T16:59:31.4121814Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-11-03T16:59:31.4122109Z "size": 104, 2025-11-03T16:59:31.4122401Z "digest": "sha256:540c0b6aa5f3cd25baec56a97c4cad8bb1cfaee48f742276c41a24fa07ff619a" 2025-11-03T16:59:31.4122735Z }, 2025-11-03T16:59:31.4122866Z { 2025-11-03T16:59:31.4123086Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-11-03T16:59:31.4123371Z "size": 1496, 2025-11-03T16:59:31.4123654Z "digest": "sha256:5f740b394ea310b0c2f28c055d3ba824427d5b2b55a9226ac4039be54b839698" 2025-11-03T16:59:31.4123967Z }, 2025-11-03T16:59:31.4124104Z { 2025-11-03T16:59:31.4124325Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-11-03T16:59:31.4124625Z "size": 458410283, 2025-11-03T16:59:31.4124921Z "digest": "sha256:69c4fff3ff1a24255d59f1b56a0f00a70f25b3b1f627fe4d5c330d8c15c83b1b" 2025-11-03T16:59:31.4125247Z }, 2025-11-03T16:59:31.4125382Z { 2025-11-03T16:59:31.4125606Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-11-03T16:59:31.4126022Z "size": 163, 2025-11-03T16:59:31.4126302Z "digest": "sha256:ac00f880c769556a073909f55054fea6299ed9137a1bcf4478f23acfd73e2fdf" 2025-11-03T16:59:31.4126624Z }, 2025-11-03T16:59:31.4126757Z { 2025-11-03T16:59:31.4126976Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-11-03T16:59:31.4127258Z "size": 346, 2025-11-03T16:59:31.4127554Z "digest": "sha256:c084e8e1076b1f971aad4348e38c6204c20c0d3b768959f05a5f7e298c600d46" 2025-11-03T16:59:31.4127878Z }, 2025-11-03T16:59:31.4128004Z { 2025-11-03T16:59:31.4128232Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-11-03T16:59:31.4128521Z "size": 32, 2025-11-03T16:59:31.4128814Z "digest": "sha256:4f4fb700ef54461cfa02571ae0db9a0dc1e0cdb5577484a6d75e68dc38e8acc1" 2025-11-03T16:59:31.4129142Z }, 2025-11-03T16:59:31.4129282Z { 2025-11-03T16:59:31.4129514Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-11-03T16:59:31.4129811Z "size": 106, 2025-11-03T16:59:31.4130099Z "digest": "sha256:a300a654b9ecbe91618f11f1bf727a7e1df8e1819158bc0c647fa85654f25a9b" 2025-11-03T16:59:31.4130426Z }, 2025-11-03T16:59:31.4130563Z { 2025-11-03T16:59:31.4130782Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-11-03T16:59:31.4131060Z "size": 425, 2025-11-03T16:59:31.4131338Z "digest": "sha256:6345aec6eb298d06b276130fd8c432205e01f3090bf30bce281256fb484b6f4c" 2025-11-03T16:59:31.4131661Z }, 2025-11-03T16:59:31.4131796Z { 2025-11-03T16:59:31.4132014Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-11-03T16:59:31.4132303Z "size": 19309365, 2025-11-03T16:59:31.4132726Z "digest": "sha256:e7298879fae82b1621bb2ec3afaec102f5d3d19babf009f2af49f96fd1cefe6a" 2025-11-03T16:59:31.4133059Z }, 2025-11-03T16:59:31.4133185Z { 2025-11-03T16:59:31.4133411Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-11-03T16:59:31.4133687Z "size": 108, 2025-11-03T16:59:31.4133959Z "digest": "sha256:943364a6285e1bc14c6e7f9a03efb0b26a1d40da8d4ef89a6284ad066ad3aec7" 2025-11-03T16:59:31.4134285Z }, 2025-11-03T16:59:31.4134410Z { 2025-11-03T16:59:31.4134626Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-11-03T16:59:31.4134897Z "size": 639, 2025-11-03T16:59:31.4135175Z "digest": "sha256:14f0993d5468420d7d0c2454a7b63ebd7a0a1c8a27674982e3a4427f49ee1b4c" 2025-11-03T16:59:31.4135489Z }, 2025-11-03T16:59:31.4135613Z { 2025-11-03T16:59:31.4135823Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-11-03T16:59:31.4136102Z "size": 724, 2025-11-03T16:59:31.4136384Z "digest": "sha256:b3a429f4faee83e3d7e616cbcca1b458319d213977b4807ea863fc416abb2229" 2025-11-03T16:59:31.4136706Z }, 2025-11-03T16:59:31.4136830Z { 2025-11-03T16:59:31.4137058Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-11-03T16:59:31.4137336Z "size": 149, 2025-11-03T16:59:31.4137615Z "digest": "sha256:d6a558a7de5e9dd5633745ba04365d261f0c5ef255d68335a1e34d2393acabd6" 2025-11-03T16:59:31.4137931Z }, 2025-11-03T16:59:31.4138060Z { 2025-11-03T16:59:31.4138280Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-11-03T16:59:31.4138560Z "size": 135, 2025-11-03T16:59:31.4138831Z "digest": "sha256:05ee86267141af52f7acaa389053961b9b3af580b4cfa52635fb11c3cc9ac77a" 2025-11-03T16:59:31.4139152Z }, 2025-11-03T16:59:31.4139281Z { 2025-11-03T16:59:31.4139501Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-11-03T16:59:31.4139792Z "size": 140, 2025-11-03T16:59:31.4140077Z "digest": "sha256:f0ccf17afa340e506e78617ef552a5e1bcde2d3a8f329d0741997cabdd68b3df" 2025-11-03T16:59:31.4140404Z }, 2025-11-03T16:59:31.4140533Z { 2025-11-03T16:59:31.4140755Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-11-03T16:59:31.4141036Z "size": 32, 2025-11-03T16:59:31.4141319Z "digest": "sha256:4f4fb700ef54461cfa02571ae0db9a0dc1e0cdb5577484a6d75e68dc38e8acc1" 2025-11-03T16:59:31.4141824Z }, 2025-11-03T16:59:31.4141951Z { 2025-11-03T16:59:31.4142175Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-11-03T16:59:31.4142462Z "size": 221, 2025-11-03T16:59:31.4142752Z "digest": "sha256:a68d4d73adc3e7837c49e151841fa4d2ac9ad874ffae2ae9f6582d4da63ee803" 2025-11-03T16:59:31.4143093Z }, 2025-11-03T16:59:31.4143227Z { 2025-11-03T16:59:31.4143448Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-11-03T16:59:31.4143733Z "size": 255, 2025-11-03T16:59:31.4144009Z "digest": "sha256:1ef3a6930902936a8a4d2f964a465c0b4ceecec2fd03a78c0e5b4a4174a6dcc7" 2025-11-03T16:59:31.4144349Z }, 2025-11-03T16:59:31.4144486Z { 2025-11-03T16:59:31.4144712Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-11-03T16:59:31.4144992Z "size": 32, 2025-11-03T16:59:31.4145278Z "digest": "sha256:4f4fb700ef54461cfa02571ae0db9a0dc1e0cdb5577484a6d75e68dc38e8acc1" 2025-11-03T16:59:31.4145603Z }, 2025-11-03T16:59:31.4145734Z { 2025-11-03T16:59:31.4145955Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-11-03T16:59:31.4146240Z "size": 106, 2025-11-03T16:59:31.4146520Z "digest": "sha256:ba2105a84ce6992ef1d41f379ee51bbd5b34872691f4924d2ca3a46aee8ffe61" 2025-11-03T16:59:31.4146849Z }, 2025-11-03T16:59:31.4146978Z { 2025-11-03T16:59:31.4147197Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-11-03T16:59:31.4147483Z "size": 312468923, 2025-11-03T16:59:31.4147772Z "digest": "sha256:85c483845f299fec36365e6fb5d71ad97d8f0e1478233c63d51699e16212a287" 2025-11-03T16:59:31.4148085Z }, 2025-11-03T16:59:31.4148220Z { 2025-11-03T16:59:31.4148538Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-11-03T16:59:31.4148831Z "size": 3222059909, 2025-11-03T16:59:31.4149128Z "digest": "sha256:ce611413a46e5e0df2bb13c62f1fcaa11d5e2e474b803d0df2f5ef3daa50f9e9" 2025-11-03T16:59:31.4149458Z }, 2025-11-03T16:59:31.4149588Z { 2025-11-03T16:59:31.4149806Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-11-03T16:59:31.4150090Z "size": 128, 2025-11-03T16:59:31.4150371Z "digest": "sha256:2b12a35a9605735c61f0f93dbabfe19540a2849f46f257f8eb5bf2ff0e2f37e2" 2025-11-03T16:59:31.4150697Z }, 2025-11-03T16:59:31.4150829Z { 2025-11-03T16:59:31.4151058Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-11-03T16:59:31.4151351Z "size": 880, 2025-11-03T16:59:31.4151627Z "digest": "sha256:2a86c078c73d1d0224e1046a3e2ca94663ca04652ec9ac44700e7fd6a4d9fc9c" 2025-11-03T16:59:31.4151949Z }, 2025-11-03T16:59:31.4152075Z { 2025-11-03T16:59:31.4152299Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-11-03T16:59:31.4152588Z "size": 724, 2025-11-03T16:59:31.4152872Z "digest": "sha256:b3a429f4faee83e3d7e616cbcca1b458319d213977b4807ea863fc416abb2229" 2025-11-03T16:59:31.4153193Z }, 2025-11-03T16:59:31.4153326Z { 2025-11-03T16:59:31.4153545Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-11-03T16:59:31.4153832Z "size": 139, 2025-11-03T16:59:31.4154108Z "digest": "sha256:44e5c47c9ce4024f40e82cfb276df42771be0a6991ba260f37a9789f3caee04a" 2025-11-03T16:59:31.4154431Z }, 2025-11-03T16:59:31.4154564Z { 2025-11-03T16:59:31.4154792Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-11-03T16:59:31.4155077Z "size": 32, 2025-11-03T16:59:31.4155363Z "digest": "sha256:4f4fb700ef54461cfa02571ae0db9a0dc1e0cdb5577484a6d75e68dc38e8acc1" 2025-11-03T16:59:31.4155704Z }, 2025-11-03T16:59:31.4155842Z { 2025-11-03T16:59:31.4156061Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-11-03T16:59:31.4156351Z "size": 159, 2025-11-03T16:59:31.4156639Z "digest": "sha256:92d25d2ab4f36bf32d7e163eeaf2c24a582062777d0bc5231fbb32482184cfd7" 2025-11-03T16:59:31.4156965Z }, 2025-11-03T16:59:31.4157093Z { 2025-11-03T16:59:31.4157316Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-11-03T16:59:31.4157603Z "size": 1011, 2025-11-03T16:59:31.4157966Z "digest": "sha256:07c8d78a73e65a5453a1045bb24c0080b003ffc3168f56f16fea4c30f9581f5d" 2025-11-03T16:59:31.4158281Z }, 2025-11-03T16:59:31.4158414Z { 2025-11-03T16:59:31.4158633Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-11-03T16:59:31.4158918Z "size": 724, 2025-11-03T16:59:31.4159200Z "digest": "sha256:b3a429f4faee83e3d7e616cbcca1b458319d213977b4807ea863fc416abb2229" 2025-11-03T16:59:31.4159526Z }, 2025-11-03T16:59:31.4159659Z { 2025-11-03T16:59:31.4159875Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-11-03T16:59:31.4160152Z "size": 134, 2025-11-03T16:59:31.4160428Z "digest": "sha256:66372f032f924ed974c345878c60ae95400e3826d05cadb854b5b378c86fef10" 2025-11-03T16:59:31.4160754Z }, 2025-11-03T16:59:31.4160885Z { 2025-11-03T16:59:31.4161099Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-11-03T16:59:31.4161384Z "size": 32, 2025-11-03T16:59:31.4161663Z "digest": "sha256:4f4fb700ef54461cfa02571ae0db9a0dc1e0cdb5577484a6d75e68dc38e8acc1" 2025-11-03T16:59:31.4161992Z }, 2025-11-03T16:59:31.4162117Z { 2025-11-03T16:59:31.4162334Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-11-03T16:59:31.4162619Z "size": 157, 2025-11-03T16:59:31.4162896Z "digest": "sha256:60194b590b00d0238b1e90ea860d06fc775fe687151dc536f5d082458f0a386c" 2025-11-03T16:59:31.4163216Z }, 2025-11-03T16:59:31.4163344Z { 2025-11-03T16:59:31.4163563Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-11-03T16:59:31.4163849Z "size": 1371, 2025-11-03T16:59:31.4164143Z "digest": "sha256:b41ffa5a6a4588148bf64c1e6d509af9d49b9ff7f27f4f7be606e4c0a48f3a12" 2025-11-03T16:59:31.4164551Z }, 2025-11-03T16:59:31.4164686Z { 2025-11-03T16:59:31.4164914Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-11-03T16:59:31.4165201Z "size": 32, 2025-11-03T16:59:31.4165492Z "digest": "sha256:4f4fb700ef54461cfa02571ae0db9a0dc1e0cdb5577484a6d75e68dc38e8acc1" 2025-11-03T16:59:31.4165830Z }, 2025-11-03T16:59:31.4165967Z { 2025-11-03T16:59:31.4166187Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-11-03T16:59:31.4166464Z "size": 136, 2025-11-03T16:59:31.4166748Z "digest": "sha256:f165304e20915f2cdcc03f0e977c2db943a17feb53a13343371f12bae7fd74d3" 2025-11-03T16:59:31.4167075Z }, 2025-11-03T16:59:31.4167208Z { 2025-11-03T16:59:31.4167424Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-11-03T16:59:31.4167710Z "size": 528, 2025-11-03T16:59:31.4167994Z "digest": "sha256:16ec736149b0011270995bb0bab5befaafbc716cfdf9f280ff25f75a2072c90d" 2025-11-03T16:59:31.4168324Z }, 2025-11-03T16:59:31.4168457Z { 2025-11-03T16:59:31.4168686Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-11-03T16:59:31.4168980Z "size": 32, 2025-11-03T16:59:31.4169259Z "digest": "sha256:4f4fb700ef54461cfa02571ae0db9a0dc1e0cdb5577484a6d75e68dc38e8acc1" 2025-11-03T16:59:31.4169581Z }, 2025-11-03T16:59:31.4169710Z { 2025-11-03T16:59:31.4169927Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-11-03T16:59:31.4170210Z "size": 104, 2025-11-03T16:59:31.4170493Z "digest": "sha256:ea0ef0105e5f56368e0dbe0d4cea9d772ddf42bb0e3f6b5066fbe970c014a684" 2025-11-03T16:59:31.4170820Z }, 2025-11-03T16:59:31.4170950Z { 2025-11-03T16:59:31.4171168Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-11-03T16:59:31.4171444Z "size": 428, 2025-11-03T16:59:31.4171723Z "digest": "sha256:b294669af2b5786b9bc5601bffdc874e4673b0fcb879e1d77c17e674e7954b10" 2025-11-03T16:59:31.4172044Z }, 2025-11-03T16:59:31.4172172Z { 2025-11-03T16:59:31.4172387Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-11-03T16:59:31.4172671Z "size": 32, 2025-11-03T16:59:31.4172963Z "digest": "sha256:4f4fb700ef54461cfa02571ae0db9a0dc1e0cdb5577484a6d75e68dc38e8acc1" 2025-11-03T16:59:31.4173286Z }, 2025-11-03T16:59:31.4173509Z { 2025-11-03T16:59:31.4173729Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-11-03T16:59:31.4174016Z "size": 107, 2025-11-03T16:59:31.4174291Z "digest": "sha256:726179d41040afffbfeb80d353b6032e11543ff3d0337dc5c67a9238cebefc27" 2025-11-03T16:59:31.4174607Z }, 2025-11-03T16:59:31.4174739Z { 2025-11-03T16:59:31.4174960Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-11-03T16:59:31.4175253Z "size": 1896, 2025-11-03T16:59:31.4175530Z "digest": "sha256:375f6cf74921eec948c69077e8280fd9d9c2a6ccf83f8a33cccc413fc36edca8" 2025-11-03T16:59:31.4175857Z }, 2025-11-03T16:59:31.4175992Z { 2025-11-03T16:59:31.4176216Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-11-03T16:59:31.4176496Z "size": 245097719, 2025-11-03T16:59:31.4176801Z "digest": "sha256:0cc2718afe447c4d220e3db62787c33345e73283935f307a2033aeb8c875b6af" 2025-11-03T16:59:31.4177122Z }, 2025-11-03T16:59:31.4177254Z { 2025-11-03T16:59:31.4177474Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-11-03T16:59:31.4177764Z "size": 106, 2025-11-03T16:59:31.4178045Z "digest": "sha256:cc98ed46781d140520a78924755c1054fee09f9303a3611a63fea0ff97d5b478" 2025-11-03T16:59:31.4178363Z }, 2025-11-03T16:59:31.4178492Z { 2025-11-03T16:59:31.4178714Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-11-03T16:59:31.4179005Z "size": 165, 2025-11-03T16:59:31.4179283Z "digest": "sha256:83ecc1143684a6535b14ce79395dec432eca66768055598312e14a7d4522ad8b" 2025-11-03T16:59:31.4179609Z }, 2025-11-03T16:59:31.4179748Z { 2025-11-03T16:59:31.4179973Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-11-03T16:59:31.4180341Z "size": 7944, 2025-11-03T16:59:31.4180618Z "digest": "sha256:84d7e8a5d95e595a910304a466fdf855f4ffc34878881b1e1a3c9124207fdb5d" 2025-11-03T16:59:31.4180934Z }, 2025-11-03T16:59:31.4181063Z { 2025-11-03T16:59:31.4181278Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-11-03T16:59:31.4181649Z "size": 8070, 2025-11-03T16:59:31.4181924Z "digest": "sha256:9913742e4da406263abf521b5f54b52b61c82f5523433268948ae207565470e4" 2025-11-03T16:59:31.4182239Z }, 2025-11-03T16:59:31.4182367Z { 2025-11-03T16:59:31.4182579Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-11-03T16:59:31.4182860Z "size": 303, 2025-11-03T16:59:31.4183133Z "digest": "sha256:a5d33870f153d309e9c3150f196979fdaee17378d91b1ed8257bad66194c99c6" 2025-11-03T16:59:31.4183450Z }, 2025-11-03T16:59:31.4183574Z { 2025-11-03T16:59:31.4183791Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-11-03T16:59:31.4184076Z "size": 13360444, 2025-11-03T16:59:31.4184365Z "digest": "sha256:50b140b19ce99832006da1669a1feb8202de98a2960e50cbc1d2107212f6de30" 2025-11-03T16:59:31.4184684Z }, 2025-11-03T16:59:31.4184816Z { 2025-11-03T16:59:31.4185046Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-11-03T16:59:31.4185334Z "size": 108, 2025-11-03T16:59:31.4185616Z "digest": "sha256:e7227b2cf3c5ba811ee334ab6ae30e3894df5c07ace96da7bd9d79fa63569e97" 2025-11-03T16:59:31.4185953Z }, 2025-11-03T16:59:31.4186090Z { 2025-11-03T16:59:31.4186312Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-11-03T16:59:31.4186589Z "size": 54145699, 2025-11-03T16:59:31.4186893Z "digest": "sha256:4807408eff38f15afc1aad24567cabdde845bbc9b1f79a1d63509dac466a197b" 2025-11-03T16:59:31.4187216Z }, 2025-11-03T16:59:31.4187351Z { 2025-11-03T16:59:31.4187573Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-11-03T16:59:31.4187859Z "size": 32, 2025-11-03T16:59:31.4188143Z "digest": "sha256:4f4fb700ef54461cfa02571ae0db9a0dc1e0cdb5577484a6d75e68dc38e8acc1" 2025-11-03T16:59:31.4188468Z } 2025-11-03T16:59:31.4188600Z ] 2025-11-03T16:59:31.4188741Z } 2025-11-03T16:59:31.4223311Z ##[group]Run set -eux 2025-11-03T16:59:31.4223517Z set -eux 2025-11-03T16:59:31.4223947Z # It's ok if this steps fails, it would then be an anonymous user like what we used to have 2025-11-03T16:59:31.4224792Z aws secretsmanager get-secret-value --secret-id docker_hub_readonly_token | jq --raw-output '.SecretString' | jq -r .docker_hub_readonly_token | docker login --username pytorchbot --password-stdin || true 2025-11-03T16:59:31.4233106Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2025-11-03T16:59:31.4233378Z env: 2025-11-03T16:59:31.4233539Z GIT_DEFAULT_BRANCH: main 2025-11-03T16:59:31.4233744Z ##[endgroup] 2025-11-03T16:59:31.4265729Z + aws secretsmanager get-secret-value --secret-id docker_hub_readonly_token 2025-11-03T16:59:31.4266565Z + jq --raw-output .SecretString 2025-11-03T16:59:31.4268645Z + jq -r .docker_hub_readonly_token 2025-11-03T16:59:31.4269888Z + docker login --username pytorchbot --password-stdin 2025-11-03T16:59:31.9620733Z WARNING! Your password will be stored unencrypted in /home/ec2-user/.docker/config.json. 2025-11-03T16:59:31.9622785Z Configure a credential helper to remove this warning. See 2025-11-03T16:59:31.9624146Z https://docs.docker.com/engine/reference/commandline/login/#credentials-store 2025-11-03T16:59:31.9625062Z 2025-11-03T16:59:31.9625225Z Login Succeeded 2025-11-03T16:59:31.9710887Z ##[group]Run tag=${ECR_DOCKER_IMAGE##*:} 2025-11-03T16:59:31.9711198Z tag=${ECR_DOCKER_IMAGE##*:} 2025-11-03T16:59:31.9711498Z echo "docker pull ghcr.io/pytorch/ci-image:${tag/:/-}" 2025-11-03T16:59:31.9719700Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2025-11-03T16:59:31.9719984Z env: 2025-11-03T16:59:31.9720148Z GIT_DEFAULT_BRANCH: main 2025-11-03T16:59:31.9720776Z ECR_DOCKER_IMAGE: 308535385114.dkr.ecr.us-east-1.amazonaws.com/pytorch/ci-image:pytorch-linux-jammy-cuda12.8-cudnn9-py3.12-gcc11-vllm-30147fb242f88ab8f13e14b76ab186d189d8605c 2025-11-03T16:59:31.9721755Z ##[endgroup] 2025-11-03T16:59:31.9752610Z docker pull ghcr.io/pytorch/ci-image:pytorch-linux-jammy-cuda12.8-cudnn9-py3.12-gcc11-vllm-30147fb242f88ab8f13e14b76ab186d189d8605c 2025-11-03T16:59:31.9806198Z ##[group]Run pytorch/test-infra/.github/actions/pull-docker-image@main 2025-11-03T16:59:31.9806551Z with: 2025-11-03T16:59:31.9807135Z docker-image: 308535385114.dkr.ecr.us-east-1.amazonaws.com/pytorch/ci-image:pytorch-linux-jammy-cuda12.8-cudnn9-py3.12-gcc11-vllm-30147fb242f88ab8f13e14b76ab186d189d8605c 2025-11-03T16:59:31.9807826Z docker-registry: 308535385114.dkr.ecr.us-east-1.amazonaws.com 2025-11-03T16:59:31.9808095Z env: 2025-11-03T16:59:31.9808246Z GIT_DEFAULT_BRANCH: main 2025-11-03T16:59:31.9808432Z ##[endgroup] 2025-11-03T16:59:31.9831171Z ##[group]Run set -x 2025-11-03T16:59:31.9831375Z set -x 2025-11-03T16:59:31.9831544Z set +e 2025-11-03T16:59:31.9831708Z  2025-11-03T16:59:31.9831859Z login() { 2025-11-03T16:59:31.9832213Z  aws ecr get-login-password --region us-east-1 | docker login -u AWS --password-stdin "$1" 2025-11-03T16:59:31.9832578Z } 2025-11-03T16:59:31.9832751Z  2025-11-03T16:59:31.9832925Z retry () { 2025-11-03T16:59:31.9833137Z  $* || (sleep 1 && $*) || (sleep 2 && $*) 2025-11-03T16:59:31.9833366Z } 2025-11-03T16:59:31.9833520Z  2025-11-03T16:59:31.9833688Z retry login "${DOCKER_REGISTRY}" 2025-11-03T16:59:31.9833909Z  2025-11-03T16:59:31.9834263Z IMAGE_SIZE=$(docker manifest inspect "${DOCKER_IMAGE}" | jq '[.layers[].size, .config.size] | add / 1024 / 1024') 2025-11-03T16:59:31.9834756Z echo "Compressed size of image in MB: ${IMAGE_SIZE}" 2025-11-03T16:59:31.9835024Z  2025-11-03T16:59:31.9835179Z set -e 2025-11-03T16:59:31.9835432Z # ignore output since only exit code is used for conditional 2025-11-03T16:59:31.9835784Z # only pull docker image if it's not available locally 2025-11-03T16:59:31.9836178Z if ! docker inspect --type=image "${DOCKER_IMAGE}" >/dev/null 2>/dev/null; then 2025-11-03T16:59:31.9848086Z  retry docker pull "${DOCKER_IMAGE}" 2025-11-03T16:59:31.9848352Z fi 2025-11-03T16:59:31.9855542Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2025-11-03T16:59:31.9855816Z env: 2025-11-03T16:59:31.9855977Z GIT_DEFAULT_BRANCH: main 2025-11-03T16:59:31.9856595Z DOCKER_IMAGE: 308535385114.dkr.ecr.us-east-1.amazonaws.com/pytorch/ci-image:pytorch-linux-jammy-cuda12.8-cudnn9-py3.12-gcc11-vllm-30147fb242f88ab8f13e14b76ab186d189d8605c 2025-11-03T16:59:31.9857288Z DOCKER_REGISTRY: 308535385114.dkr.ecr.us-east-1.amazonaws.com 2025-11-03T16:59:31.9857570Z ##[endgroup] 2025-11-03T16:59:31.9883593Z + set +e 2025-11-03T16:59:31.9883918Z + retry login 308535385114.dkr.ecr.us-east-1.amazonaws.com 2025-11-03T16:59:31.9884316Z + login 308535385114.dkr.ecr.us-east-1.amazonaws.com 2025-11-03T16:59:31.9887159Z + aws ecr get-login-password --region us-east-1 2025-11-03T16:59:31.9888493Z + docker login -u AWS --password-stdin 308535385114.dkr.ecr.us-east-1.amazonaws.com 2025-11-03T16:59:32.4639657Z WARNING! Your password will be stored unencrypted in /home/ec2-user/.docker/config.json. 2025-11-03T16:59:32.4640789Z Configure a credential helper to remove this warning. See 2025-11-03T16:59:32.4641818Z https://docs.docker.com/engine/reference/commandline/login/#credentials-store 2025-11-03T16:59:32.4642676Z 2025-11-03T16:59:32.4643085Z Login Succeeded 2025-11-03T16:59:32.4669220Z ++ docker manifest inspect 308535385114.dkr.ecr.us-east-1.amazonaws.com/pytorch/ci-image:pytorch-linux-jammy-cuda12.8-cudnn9-py3.12-gcc11-vllm-30147fb242f88ab8f13e14b76ab186d189d8605c 2025-11-03T16:59:32.4670147Z ++ jq '[.layers[].size, .config.size] | add / 1024 / 1024' 2025-11-03T16:59:32.6875665Z + IMAGE_SIZE=15309.793606758118 2025-11-03T16:59:32.6876036Z + echo 'Compressed size of image in MB: 15309.793606758118' 2025-11-03T16:59:32.6876395Z + set -e 2025-11-03T16:59:32.6877502Z + docker inspect --type=image 308535385114.dkr.ecr.us-east-1.amazonaws.com/pytorch/ci-image:pytorch-linux-jammy-cuda12.8-cudnn9-py3.12-gcc11-vllm-30147fb242f88ab8f13e14b76ab186d189d8605c 2025-11-03T16:59:32.6878434Z Compressed size of image in MB: 15309.793606758118 2025-11-03T16:59:32.7003176Z + retry docker pull 308535385114.dkr.ecr.us-east-1.amazonaws.com/pytorch/ci-image:pytorch-linux-jammy-cuda12.8-cudnn9-py3.12-gcc11-vllm-30147fb242f88ab8f13e14b76ab186d189d8605c 2025-11-03T16:59:32.7004555Z + docker pull 308535385114.dkr.ecr.us-east-1.amazonaws.com/pytorch/ci-image:pytorch-linux-jammy-cuda12.8-cudnn9-py3.12-gcc11-vllm-30147fb242f88ab8f13e14b76ab186d189d8605c 2025-11-03T16:59:32.9494145Z pytorch-linux-jammy-cuda12.8-cudnn9-py3.12-gcc11-vllm-30147fb242f88ab8f13e14b76ab186d189d8605c: Pulling from pytorch/ci-image 2025-11-03T16:59:32.9498100Z 828c1365039a: Pulling fs layer 2025-11-03T16:59:32.9498930Z 1c4ee0d43392: Pulling fs layer 2025-11-03T16:59:32.9499701Z 728590afcb21: Pulling fs layer 2025-11-03T16:59:32.9500324Z 9a925ed696bf: Pulling fs layer 2025-11-03T16:59:32.9500611Z e7ab1efdd1bd: Pulling fs layer 2025-11-03T16:59:32.9500919Z 33a0c5e93563: Pulling fs layer 2025-11-03T16:59:32.9501149Z 7401e108f2b8: Pulling fs layer 2025-11-03T16:59:32.9501390Z 9d9c7be466f8: Pulling fs layer 2025-11-03T16:59:32.9501760Z faa9a1190c2e: Pulling fs layer 2025-11-03T16:59:32.9501995Z 7c2887fa8aa0: Pulling fs layer 2025-11-03T16:59:32.9502264Z 95bb0abcaee9: Pulling fs layer 2025-11-03T16:59:32.9502510Z b3a429f4faee: Pulling fs layer 2025-11-03T16:59:32.9502744Z f4cb10b91a16: Pulling fs layer 2025-11-03T16:59:32.9502983Z 48618db2b1a6: Pulling fs layer 2025-11-03T16:59:32.9503279Z 4f4fb700ef54: Pulling fs layer 2025-11-03T16:59:32.9503715Z 2ec7ba999b86: Pulling fs layer 2025-11-03T16:59:32.9504122Z 85d9004dcaf2: Pulling fs layer 2025-11-03T16:59:32.9504538Z df8f8108c158: Pulling fs layer 2025-11-03T16:59:32.9504959Z 329fb7ac2c9e: Pulling fs layer 2025-11-03T16:59:32.9505385Z 397c5718353c: Pulling fs layer 2025-11-03T16:59:32.9505795Z 099139c9d8b4: Pulling fs layer 2025-11-03T16:59:32.9506212Z c094f15f3938: Pulling fs layer 2025-11-03T16:59:32.9507026Z a99b38a3dcdd: Pulling fs layer 2025-11-03T16:59:32.9507322Z bc183ac25ebc: Pulling fs layer 2025-11-03T16:59:32.9507513Z 1a6a81cb899b: Pulling fs layer 2025-11-03T16:59:32.9507805Z 41ec23fc0915: Pulling fs layer 2025-11-03T16:59:32.9508051Z 540c0b6aa5f3: Pulling fs layer 2025-11-03T16:59:32.9508254Z 5f740b394ea3: Pulling fs layer 2025-11-03T16:59:32.9508444Z 69c4fff3ff1a: Pulling fs layer 2025-11-03T16:59:32.9508640Z ac00f880c769: Pulling fs layer 2025-11-03T16:59:32.9508832Z c084e8e1076b: Pulling fs layer 2025-11-03T16:59:32.9509059Z a300a654b9ec: Pulling fs layer 2025-11-03T16:59:32.9509251Z 6345aec6eb29: Pulling fs layer 2025-11-03T16:59:32.9509444Z e7298879fae8: Pulling fs layer 2025-11-03T16:59:32.9509632Z 943364a6285e: Pulling fs layer 2025-11-03T16:59:32.9509820Z 14f0993d5468: Pulling fs layer 2025-11-03T16:59:32.9510004Z d6a558a7de5e: Pulling fs layer 2025-11-03T16:59:32.9510194Z 05ee86267141: Pulling fs layer 2025-11-03T16:59:32.9510389Z f0ccf17afa34: Pulling fs layer 2025-11-03T16:59:32.9510587Z a68d4d73adc3: Pulling fs layer 2025-11-03T16:59:32.9510778Z 1ef3a6930902: Pulling fs layer 2025-11-03T16:59:32.9510966Z ba2105a84ce6: Pulling fs layer 2025-11-03T16:59:32.9511150Z 85c483845f29: Pulling fs layer 2025-11-03T16:59:32.9511336Z ce611413a46e: Pulling fs layer 2025-11-03T16:59:32.9511517Z 2b12a35a9605: Pulling fs layer 2025-11-03T16:59:32.9511702Z 2a86c078c73d: Pulling fs layer 2025-11-03T16:59:32.9511901Z 44e5c47c9ce4: Pulling fs layer 2025-11-03T16:59:32.9512087Z 92d25d2ab4f3: Pulling fs layer 2025-11-03T16:59:32.9512268Z 07c8d78a73e6: Pulling fs layer 2025-11-03T16:59:32.9512454Z 66372f032f92: Pulling fs layer 2025-11-03T16:59:32.9512640Z 60194b590b00: Pulling fs layer 2025-11-03T16:59:32.9512820Z 943364a6285e: Waiting 2025-11-03T16:59:32.9512992Z ba2105a84ce6: Waiting 2025-11-03T16:59:32.9513157Z 14f0993d5468: Waiting 2025-11-03T16:59:32.9513328Z b41ffa5a6a45: Pulling fs layer 2025-11-03T16:59:32.9513514Z f4cb10b91a16: Waiting 2025-11-03T16:59:32.9513672Z 85c483845f29: Waiting 2025-11-03T16:59:32.9513836Z 9a925ed696bf: Waiting 2025-11-03T16:59:32.9514207Z 4f4fb700ef54: Waiting 2025-11-03T16:59:32.9514377Z d6a558a7de5e: Waiting 2025-11-03T16:59:32.9514546Z 2ec7ba999b86: Waiting 2025-11-03T16:59:32.9514720Z f165304e2091: Pulling fs layer 2025-11-03T16:59:32.9514899Z 05ee86267141: Waiting 2025-11-03T16:59:32.9515065Z e7ab1efdd1bd: Waiting 2025-11-03T16:59:32.9515221Z 85d9004dcaf2: Waiting 2025-11-03T16:59:32.9515382Z f0ccf17afa34: Waiting 2025-11-03T16:59:32.9515544Z df8f8108c158: Waiting 2025-11-03T16:59:32.9515704Z a68d4d73adc3: Waiting 2025-11-03T16:59:32.9515856Z 48618db2b1a6: Waiting 2025-11-03T16:59:32.9516023Z 16ec736149b0: Pulling fs layer 2025-11-03T16:59:32.9516216Z ce611413a46e: Waiting 2025-11-03T16:59:32.9516476Z 9d9c7be466f8: Waiting 2025-11-03T16:59:32.9516647Z 2b12a35a9605: Waiting 2025-11-03T16:59:32.9516807Z 66372f032f92: Waiting 2025-11-03T16:59:32.9516966Z 2a86c078c73d: Waiting 2025-11-03T16:59:32.9517125Z 33a0c5e93563: Waiting 2025-11-03T16:59:32.9517281Z e7298879fae8: Waiting 2025-11-03T16:59:32.9517451Z 7401e108f2b8: Waiting 2025-11-03T16:59:32.9517620Z ea0ef0105e5f: Pulling fs layer 2025-11-03T16:59:32.9517806Z 95bb0abcaee9: Waiting 2025-11-03T16:59:32.9517968Z b3a429f4faee: Waiting 2025-11-03T16:59:32.9518134Z 7c2887fa8aa0: Waiting 2025-11-03T16:59:32.9518296Z 92d25d2ab4f3: Waiting 2025-11-03T16:59:32.9518461Z faa9a1190c2e: Waiting 2025-11-03T16:59:32.9518613Z 07c8d78a73e6: Waiting 2025-11-03T16:59:32.9518785Z 099139c9d8b4: Waiting 2025-11-03T16:59:32.9518956Z b294669af2b5: Pulling fs layer 2025-11-03T16:59:32.9519133Z c094f15f3938: Waiting 2025-11-03T16:59:32.9519282Z 60194b590b00: Waiting 2025-11-03T16:59:32.9519446Z 726179d41040: Pulling fs layer 2025-11-03T16:59:32.9519632Z a99b38a3dcdd: Waiting 2025-11-03T16:59:32.9519788Z f165304e2091: Waiting 2025-11-03T16:59:32.9519936Z 16ec736149b0: Waiting 2025-11-03T16:59:32.9520095Z ea0ef0105e5f: Waiting 2025-11-03T16:59:32.9520262Z 375f6cf74921: Pulling fs layer 2025-11-03T16:59:32.9520456Z b294669af2b5: Waiting 2025-11-03T16:59:32.9520733Z 540c0b6aa5f3: Waiting 2025-11-03T16:59:32.9520899Z 5f740b394ea3: Waiting 2025-11-03T16:59:32.9521245Z 69c4fff3ff1a: Waiting 2025-11-03T16:59:32.9521442Z 0cc2718afe44: Pulling fs layer 2025-11-03T16:59:32.9521644Z cc98ed46781d: Pulling fs layer 2025-11-03T16:59:32.9521835Z 83ecc1143684: Pulling fs layer 2025-11-03T16:59:32.9522029Z 84d7e8a5d95e: Pulling fs layer 2025-11-03T16:59:32.9522236Z 9913742e4da4: Pulling fs layer 2025-11-03T16:59:32.9522427Z bc183ac25ebc: Waiting 2025-11-03T16:59:32.9522597Z 6345aec6eb29: Waiting 2025-11-03T16:59:32.9522766Z a5d33870f153: Pulling fs layer 2025-11-03T16:59:32.9522942Z 726179d41040: Waiting 2025-11-03T16:59:32.9523109Z 50b140b19ce9: Pulling fs layer 2025-11-03T16:59:32.9523292Z 397c5718353c: Waiting 2025-11-03T16:59:32.9523457Z ac00f880c769: Waiting 2025-11-03T16:59:32.9523617Z e7227b2cf3c5: Pulling fs layer 2025-11-03T16:59:32.9523797Z 9913742e4da4: Waiting 2025-11-03T16:59:32.9523963Z 4807408eff38: Pulling fs layer 2025-11-03T16:59:32.9524138Z a300a654b9ec: Waiting 2025-11-03T16:59:32.9524305Z 84d7e8a5d95e: Waiting 2025-11-03T16:59:32.9524465Z 0cc2718afe44: Waiting 2025-11-03T16:59:32.9524625Z 375f6cf74921: Waiting 2025-11-03T16:59:32.9524787Z cc98ed46781d: Waiting 2025-11-03T16:59:32.9524955Z a5d33870f153: Waiting 2025-11-03T16:59:32.9525114Z 83ecc1143684: Waiting 2025-11-03T16:59:32.9525277Z 4807408eff38: Waiting 2025-11-03T16:59:32.9525429Z 50b140b19ce9: Waiting 2025-11-03T16:59:32.9525587Z b41ffa5a6a45: Waiting 2025-11-03T16:59:32.9525746Z 44e5c47c9ce4: Waiting 2025-11-03T16:59:32.9525904Z e7227b2cf3c5: Waiting 2025-11-03T16:59:32.9526055Z c084e8e1076b: Waiting 2025-11-03T16:59:32.9526225Z 1a6a81cb899b: Waiting 2025-11-03T16:59:32.9526386Z 1ef3a6930902: Waiting 2025-11-03T16:59:32.9526544Z 41ec23fc0915: Waiting 2025-11-03T16:59:32.9526698Z 329fb7ac2c9e: Waiting 2025-11-03T16:59:33.0384769Z 1c4ee0d43392: Download complete 2025-11-03T16:59:33.1063495Z 9a925ed696bf: Verifying Checksum 2025-11-03T16:59:33.1063773Z 9a925ed696bf: Download complete 2025-11-03T16:59:33.1892425Z e7ab1efdd1bd: Verifying Checksum 2025-11-03T16:59:33.1892777Z e7ab1efdd1bd: Download complete 2025-11-03T16:59:33.2774844Z 33a0c5e93563: Verifying Checksum 2025-11-03T16:59:33.2775178Z 33a0c5e93563: Download complete 2025-11-03T16:59:33.2925405Z 828c1365039a: Verifying Checksum 2025-11-03T16:59:33.2925724Z 828c1365039a: Download complete 2025-11-03T16:59:33.3803262Z 9d9c7be466f8: Verifying Checksum 2025-11-03T16:59:33.3803602Z 9d9c7be466f8: Download complete 2025-11-03T16:59:33.3847852Z 7401e108f2b8: Verifying Checksum 2025-11-03T16:59:33.3848201Z 7401e108f2b8: Download complete 2025-11-03T16:59:33.4678326Z 7c2887fa8aa0: Verifying Checksum 2025-11-03T16:59:33.4678625Z 7c2887fa8aa0: Download complete 2025-11-03T16:59:33.5449574Z 95bb0abcaee9: Download complete 2025-11-03T16:59:33.6111165Z b3a429f4faee: Download complete 2025-11-03T16:59:33.6796565Z f4cb10b91a16: Verifying Checksum 2025-11-03T16:59:33.6796869Z f4cb10b91a16: Download complete 2025-11-03T16:59:34.1961786Z 828c1365039a: Pull complete 2025-11-03T16:59:34.2177792Z 1c4ee0d43392: Pull complete 2025-11-03T16:59:34.5288988Z faa9a1190c2e: Verifying Checksum 2025-11-03T16:59:34.5289292Z faa9a1190c2e: Download complete 2025-11-03T16:59:34.5365203Z 4f4fb700ef54: Verifying Checksum 2025-11-03T16:59:34.5365594Z 4f4fb700ef54: Download complete 2025-11-03T16:59:34.6221553Z 2ec7ba999b86: Verifying Checksum 2025-11-03T16:59:34.6221817Z 2ec7ba999b86: Download complete 2025-11-03T16:59:34.6936136Z 85d9004dcaf2: Verifying Checksum 2025-11-03T16:59:34.6936658Z 85d9004dcaf2: Download complete 2025-11-03T16:59:34.7514075Z df8f8108c158: Verifying Checksum 2025-11-03T16:59:34.7514395Z df8f8108c158: Download complete 2025-11-03T16:59:34.8399152Z 329fb7ac2c9e: Verifying Checksum 2025-11-03T16:59:34.8399482Z 329fb7ac2c9e: Download complete 2025-11-03T16:59:34.9251658Z 397c5718353c: Verifying Checksum 2025-11-03T16:59:34.9252000Z 397c5718353c: Download complete 2025-11-03T16:59:35.0191991Z 099139c9d8b4: Download complete 2025-11-03T16:59:35.1034056Z c094f15f3938: Verifying Checksum 2025-11-03T16:59:35.1034617Z c094f15f3938: Download complete 2025-11-03T16:59:35.1772595Z a99b38a3dcdd: Verifying Checksum 2025-11-03T16:59:35.1773031Z a99b38a3dcdd: Download complete 2025-11-03T16:59:36.1538195Z 728590afcb21: Verifying Checksum 2025-11-03T16:59:36.1538629Z 728590afcb21: Download complete 2025-11-03T16:59:36.2284721Z 1a6a81cb899b: Verifying Checksum 2025-11-03T16:59:36.2285116Z 1a6a81cb899b: Download complete 2025-11-03T16:59:36.6346369Z 41ec23fc0915: Verifying Checksum 2025-11-03T16:59:36.7094568Z 540c0b6aa5f3: Verifying Checksum 2025-11-03T16:59:36.8035621Z 540c0b6aa5f3: Download complete 2025-11-03T16:59:36.8035991Z 5f740b394ea3: Verifying Checksum 2025-11-03T16:59:36.8036282Z 5f740b394ea3: Download complete 2025-11-03T16:59:41.4574053Z 69c4fff3ff1a: Verifying Checksum 2025-11-03T16:59:41.4574362Z 69c4fff3ff1a: Download complete 2025-11-03T16:59:41.5476825Z ac00f880c769: Verifying Checksum 2025-11-03T16:59:41.5477256Z ac00f880c769: Download complete 2025-11-03T16:59:41.6474037Z c084e8e1076b: Verifying Checksum 2025-11-03T16:59:41.6474628Z c084e8e1076b: Download complete 2025-11-03T16:59:41.7071977Z a300a654b9ec: Download complete 2025-11-03T16:59:41.7984573Z 6345aec6eb29: Download complete 2025-11-03T16:59:42.0536785Z e7298879fae8: Verifying Checksum 2025-11-03T16:59:42.0537247Z e7298879fae8: Download complete 2025-11-03T16:59:42.1193080Z 943364a6285e: Verifying Checksum 2025-11-03T16:59:42.1193367Z 943364a6285e: Download complete 2025-11-03T16:59:42.1868229Z 14f0993d5468: Verifying Checksum 2025-11-03T16:59:42.1868591Z 14f0993d5468: Download complete 2025-11-03T16:59:42.2483255Z d6a558a7de5e: Verifying Checksum 2025-11-03T16:59:42.2483614Z d6a558a7de5e: Download complete 2025-11-03T16:59:42.3297985Z 05ee86267141: Verifying Checksum 2025-11-03T16:59:42.3298381Z 05ee86267141: Download complete 2025-11-03T16:59:42.4118240Z f0ccf17afa34: Download complete 2025-11-03T16:59:42.5253767Z a68d4d73adc3: Verifying Checksum 2025-11-03T16:59:42.5254609Z a68d4d73adc3: Download complete 2025-11-03T16:59:42.6081259Z 1ef3a6930902: Verifying Checksum 2025-11-03T16:59:42.6081654Z 1ef3a6930902: Download complete 2025-11-03T16:59:42.7190905Z ba2105a84ce6: Verifying Checksum 2025-11-03T16:59:42.7191197Z ba2105a84ce6: Download complete 2025-11-03T16:59:43.2571109Z 728590afcb21: Pull complete 2025-11-03T16:59:43.2837784Z 9a925ed696bf: Pull complete 2025-11-03T16:59:43.3068846Z e7ab1efdd1bd: Pull complete 2025-11-03T16:59:43.3350494Z 33a0c5e93563: Pull complete 2025-11-03T16:59:43.3634371Z 7401e108f2b8: Pull complete 2025-11-03T16:59:43.3883866Z 9d9c7be466f8: Pull complete 2025-11-03T16:59:45.3333074Z faa9a1190c2e: Pull complete 2025-11-03T16:59:45.5326361Z 7c2887fa8aa0: Pull complete 2025-11-03T16:59:45.7011330Z 95bb0abcaee9: Pull complete 2025-11-03T16:59:45.8008374Z b3a429f4faee: Pull complete 2025-11-03T16:59:45.8963760Z f4cb10b91a16: Pull complete 2025-11-03T16:59:45.8972845Z 85c483845f29: Verifying Checksum 2025-11-03T16:59:45.8973179Z 85c483845f29: Download complete 2025-11-03T17:00:08.6315975Z 48618db2b1a6: Verifying Checksum 2025-11-03T17:00:08.6316321Z 48618db2b1a6: Download complete 2025-11-03T17:00:08.7393956Z 2b12a35a9605: Verifying Checksum 2025-11-03T17:00:08.7394368Z 2b12a35a9605: Download complete 2025-11-03T17:00:08.8470553Z 2a86c078c73d: Verifying Checksum 2025-11-03T17:00:08.8470990Z 2a86c078c73d: Download complete 2025-11-03T17:00:08.9308516Z 44e5c47c9ce4: Verifying Checksum 2025-11-03T17:00:08.9308894Z 44e5c47c9ce4: Download complete 2025-11-03T17:00:08.9955528Z 92d25d2ab4f3: Verifying Checksum 2025-11-03T17:00:08.9955908Z 92d25d2ab4f3: Download complete 2025-11-03T17:00:09.0813765Z 07c8d78a73e6: Verifying Checksum 2025-11-03T17:00:09.0814272Z 07c8d78a73e6: Download complete 2025-11-03T17:00:09.1535693Z 66372f032f92: Verifying Checksum 2025-11-03T17:00:09.1536068Z 66372f032f92: Download complete 2025-11-03T17:00:09.2243449Z 60194b590b00: Download complete 2025-11-03T17:00:09.3005816Z b41ffa5a6a45: Download complete 2025-11-03T17:00:09.4015122Z f165304e2091: Verifying Checksum 2025-11-03T17:00:09.4015771Z f165304e2091: Download complete 2025-11-03T17:00:09.4939953Z 16ec736149b0: Download complete 2025-11-03T17:00:09.5859241Z ea0ef0105e5f: Verifying Checksum 2025-11-03T17:00:09.5859624Z ea0ef0105e5f: Download complete 2025-11-03T17:00:09.6549452Z b294669af2b5: Verifying Checksum 2025-11-03T17:00:09.6550368Z b294669af2b5: Download complete 2025-11-03T17:00:09.7514411Z 726179d41040: Verifying Checksum 2025-11-03T17:00:09.7514716Z 726179d41040: Download complete 2025-11-03T17:00:09.8749569Z 375f6cf74921: Verifying Checksum 2025-11-03T17:00:09.8749902Z 375f6cf74921: Download complete 2025-11-03T17:00:12.9398420Z 0cc2718afe44: Verifying Checksum 2025-11-03T17:00:12.9398831Z 0cc2718afe44: Download complete 2025-11-03T17:00:13.0117546Z cc98ed46781d: Verifying Checksum 2025-11-03T17:00:13.0118091Z cc98ed46781d: Download complete 2025-11-03T17:00:13.0894759Z 83ecc1143684: Verifying Checksum 2025-11-03T17:00:13.0895302Z 83ecc1143684: Download complete 2025-11-03T17:00:13.1708661Z 84d7e8a5d95e: Verifying Checksum 2025-11-03T17:00:13.1709135Z 84d7e8a5d95e: Download complete 2025-11-03T17:00:13.2675671Z 9913742e4da4: Verifying Checksum 2025-11-03T17:00:13.2676185Z 9913742e4da4: Download complete 2025-11-03T17:00:13.3437102Z a5d33870f153: Verifying Checksum 2025-11-03T17:00:13.3437418Z a5d33870f153: Download complete 2025-11-03T17:00:13.5735831Z 50b140b19ce9: Verifying Checksum 2025-11-03T17:00:13.5736247Z 50b140b19ce9: Download complete 2025-11-03T17:00:13.6683713Z e7227b2cf3c5: Verifying Checksum 2025-11-03T17:00:13.6684046Z e7227b2cf3c5: Download complete 2025-11-03T17:00:14.3592742Z 4807408eff38: Verifying Checksum 2025-11-03T17:00:14.3593167Z 4807408eff38: Download complete 2025-11-03T17:00:20.9918209Z ce611413a46e: Verifying Checksum 2025-11-03T17:00:20.9918577Z ce611413a46e: Download complete 2025-11-03T17:01:00.9111908Z bc183ac25ebc: Verifying Checksum 2025-11-03T17:01:00.9112297Z bc183ac25ebc: Download complete 2025-11-03T17:01:32.8126541Z 48618db2b1a6: Pull complete 2025-11-03T17:01:33.0719292Z 4f4fb700ef54: Pull complete 2025-11-03T17:01:33.3445114Z 2ec7ba999b86: Pull complete 2025-11-03T17:01:33.5546263Z 85d9004dcaf2: Pull complete 2025-11-03T17:01:33.6565075Z df8f8108c158: Pull complete 2025-11-03T17:01:33.8005877Z 329fb7ac2c9e: Pull complete 2025-11-03T17:01:33.9725176Z 397c5718353c: Pull complete 2025-11-03T17:01:34.2985630Z 099139c9d8b4: Pull complete 2025-11-03T17:01:34.5764365Z c094f15f3938: Pull complete 2025-11-03T17:01:34.8569490Z a99b38a3dcdd: Pull complete 2025-11-03T17:03:08.4606121Z bc183ac25ebc: Pull complete 2025-11-03T17:03:08.5933969Z 1a6a81cb899b: Pull complete 2025-11-03T17:03:09.1806155Z 41ec23fc0915: Pull complete 2025-11-03T17:03:09.3091541Z 540c0b6aa5f3: Pull complete 2025-11-03T17:03:09.3410509Z 5f740b394ea3: Pull complete 2025-11-03T17:03:16.0128857Z 69c4fff3ff1a: Pull complete 2025-11-03T17:03:16.2362044Z ac00f880c769: Pull complete 2025-11-03T17:03:16.4614245Z c084e8e1076b: Pull complete 2025-11-03T17:03:16.8997950Z a300a654b9ec: Pull complete 2025-11-03T17:03:17.1259451Z 6345aec6eb29: Pull complete 2025-11-03T17:03:17.4867404Z e7298879fae8: Pull complete 2025-11-03T17:03:17.7115972Z 943364a6285e: Pull complete 2025-11-03T17:03:17.9246946Z 14f0993d5468: Pull complete 2025-11-03T17:03:18.3656554Z d6a558a7de5e: Pull complete 2025-11-03T17:03:18.5903340Z 05ee86267141: Pull complete 2025-11-03T17:03:18.8077583Z f0ccf17afa34: Pull complete 2025-11-03T17:03:19.2489838Z a68d4d73adc3: Pull complete 2025-11-03T17:03:19.4903349Z 1ef3a6930902: Pull complete 2025-11-03T17:03:19.9410035Z ba2105a84ce6: Pull complete 2025-11-03T17:03:21.0346071Z 85c483845f29: Pull complete 2025-11-03T17:04:05.3063700Z ce611413a46e: Pull complete 2025-11-03T17:04:05.3773230Z 2b12a35a9605: Pull complete 2025-11-03T17:04:05.4797506Z 2a86c078c73d: Pull complete 2025-11-03T17:04:05.7352528Z 44e5c47c9ce4: Pull complete 2025-11-03T17:04:06.1073580Z 92d25d2ab4f3: Pull complete 2025-11-03T17:04:06.2768176Z 07c8d78a73e6: Pull complete 2025-11-03T17:04:06.6438387Z 66372f032f92: Pull complete 2025-11-03T17:04:07.0169366Z 60194b590b00: Pull complete 2025-11-03T17:04:07.2278241Z b41ffa5a6a45: Pull complete 2025-11-03T17:04:07.5835461Z f165304e2091: Pull complete 2025-11-03T17:04:07.8085284Z 16ec736149b0: Pull complete 2025-11-03T17:04:08.2056355Z ea0ef0105e5f: Pull complete 2025-11-03T17:04:08.4201892Z b294669af2b5: Pull complete 2025-11-03T17:04:08.7922636Z 726179d41040: Pull complete 2025-11-03T17:04:08.9968619Z 375f6cf74921: Pull complete 2025-11-03T17:04:14.6998597Z 0cc2718afe44: Pull complete 2025-11-03T17:04:14.9082887Z cc98ed46781d: Pull complete 2025-11-03T17:04:15.1292641Z 83ecc1143684: Pull complete 2025-11-03T17:04:15.3380285Z 84d7e8a5d95e: Pull complete 2025-11-03T17:04:15.5644297Z 9913742e4da4: Pull complete 2025-11-03T17:04:15.7790407Z a5d33870f153: Pull complete 2025-11-03T17:04:17.2168561Z 50b140b19ce9: Pull complete 2025-11-03T17:04:17.4374841Z e7227b2cf3c5: Pull complete 2025-11-03T17:04:18.6927014Z 4807408eff38: Pull complete 2025-11-03T17:04:18.9339341Z Digest: sha256:655b37e47c7ab0fab00339d6162ef718600c6b4e76f1601297e6a07026860230 2025-11-03T17:04:18.9608961Z Status: Downloaded newer image for 308535385114.dkr.ecr.us-east-1.amazonaws.com/pytorch/ci-image:pytorch-linux-jammy-cuda12.8-cudnn9-py3.12-gcc11-vllm-30147fb242f88ab8f13e14b76ab186d189d8605c 2025-11-03T17:04:18.9773312Z 308535385114.dkr.ecr.us-east-1.amazonaws.com/pytorch/ci-image:pytorch-linux-jammy-cuda12.8-cudnn9-py3.12-gcc11-vllm-30147fb242f88ab8f13e14b76ab186d189d8605c 2025-11-03T17:04:18.9832509Z ##[group]Run echo "IN_CONTAINER_RUNNER=$(if [ -f /.inarc ] || [ -f /.incontainer ]; then echo true ; else echo false; fi)" >> "$GITHUB_OUTPUT" 2025-11-03T17:04:18.9833239Z echo "IN_CONTAINER_RUNNER=$(if [ -f /.inarc ] || [ -f /.incontainer ]; then echo true ; else echo false; fi)" >> "$GITHUB_OUTPUT" 2025-11-03T17:04:18.9843787Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2025-11-03T17:04:18.9844071Z env: 2025-11-03T17:04:18.9844240Z GIT_DEFAULT_BRANCH: main 2025-11-03T17:04:18.9844434Z ##[endgroup] 2025-11-03T17:04:19.0019371Z ##[group]Run pytorch/test-infra/.github/actions/setup-nvidia@main 2025-11-03T17:04:19.0019783Z with: 2025-11-03T17:04:19.0019988Z driver-version: 580.82.07 2025-11-03T17:04:19.0020223Z env: 2025-11-03T17:04:19.0020416Z GIT_DEFAULT_BRANCH: main 2025-11-03T17:04:19.0020644Z ##[endgroup] 2025-11-03T17:04:19.0129583Z ##[group]Run nick-fields/retry@3e91a01664abd3c5cd539100d10d33b9c5b68482 2025-11-03T17:04:19.0129890Z with: 2025-11-03T17:04:19.0130048Z timeout_minutes: 10 2025-11-03T17:04:19.0130226Z max_attempts: 3 2025-11-03T17:04:19.0149867Z command: # Is it disgusting to have a full shell script here in this github action? Sure # But is it the best way to make it so that this action relies on nothing else? Absolutely set -eou pipefail DISTRIBUTION=$(. /etc/os-release;echo $ID$VERSION_ID) DRIVER_FN="NVIDIA-Linux-x86_64-${DRIVER_VERSION}.run" install_nvidia_docker2_amzn2() { ( set -x # Needed for yum-config-manager sudo yum install -y yum-utils if [[ "${DISTRIBUTION}" == "amzn2023" ]] ; then YUM_REPO_URL="https://nvidia.github.io/libnvidia-container/stable/rpm/nvidia-container-toolkit.repo" else # Amazon Linux 2 YUM_REPO_URL="https://nvidia.github.io/nvidia-docker/${DISTRIBUTION}/nvidia-docker.repo" fi sudo yum-config-manager --add-repo "${YUM_REPO_URL}" sudo yum install -y \ nvidia-container-toolkit-1.17.8 \ libnvidia-container-tools-1.17.8 \ libnvidia-container1-1.17.8 \ nvidia-container-toolkit-base-1.17.8 sudo systemctl restart docker ) } install_nvidia_docker2_ubuntu20() { ( set -x # Install nvidia-driver package if not installed status="$(dpkg-query -W --showformat='${db:Status-Status}' nvidia-docker2 2>&1)" if [ ! $? = 0 ] || [ ! "$status" = installed ]; then sudo apt-get install -y nvidia-container-toolkit-1.17.8 sudo systemctl restart docker fi ) } pre_install_nvidia_driver_amzn2() { ( # Purge any nvidia driver installed from RHEL repo sudo yum remove -y nvidia-driver-latest-dkms ) } install_nvidia_driver_common() { ( # Try to gather more information about the runner and its existing NVIDIA driver if any echo "Before installing NVIDIA driver" lspci lsmod modinfo nvidia || true HAS_NVIDIA_DRIVER=0 # Check if NVIDIA driver has already been installed if [ -x "$(command -v nvidia-smi)" ]; then set +e # The driver exists, check its version next. Also check only the first GPU if there are more than one of them # so that the same driver version is not print over multiple lines INSTALLED_DRIVER_VERSION=$(nvidia-smi --query-gpu=driver_version --format=csv,noheader --id=0) NVIDIA_SMI_STATUS=$? if [ "$NVIDIA_SMI_STATUS" -ne 0 ] && [ "$NVIDIA_SMI_STATUS" -ne 14 ]; then echo "Failed to get NVIDIA driver version ($INSTALLED_DRIVER_VERSION). Continuing" elif [ "$INSTALLED_DRIVER_VERSION" != "$DRIVER_VERSION" ]; then echo "NVIDIA driver ($INSTALLED_DRIVER_VERSION) has been installed, but we expect to have $DRIVER_VERSION instead. Continuing" # Turn off persistent mode so that the installation script can unload the kernel module sudo killall nvidia-persistenced || true else HAS_NVIDIA_DRIVER=1 echo "NVIDIA driver ($INSTALLED_DRIVER_VERSION) has already been installed. Skipping NVIDIA driver installation" fi set -e fi if [ "$HAS_NVIDIA_DRIVER" -eq 0 ]; then # CAUTION: this may need to be updated in future if [ "${DISTRIBUTION}" != ubuntu20.04 ]; then sudo yum groupinstall -y "Development Tools" # ensure our kernel install is the same as our underlying kernel, # groupinstall "Development Tools" has a habit of mismatching kernel headers sudo yum install -y "kernel-devel-uname-r == $(uname -r)" sudo modprobe backlight fi sudo curl -fsL -o /tmp/nvidia_driver "https://s3.amazonaws.com/ossci-linux/nvidia_driver/$DRIVER_FN" set +e sudo /bin/bash /tmp/nvidia_driver -s --no-drm NVIDIA_INSTALLATION_STATUS=$? RESET_GPU=0 if [ "$NVIDIA_INSTALLATION_STATUS" -ne 0 ]; then sudo cat /var/log/nvidia-installer.log # Fail to install NVIDIA driver, try to reset the GPU RESET_GPU=1 elif [ -x "$(command -v nvidia-smi)" ]; then # Check again if nvidia-smi works even if the driver installation completes successfully INSTALLED_DRIVER_VERSION=$(nvidia-smi --query-gpu=driver_version --format=csv,noheader --id=0) NVIDIA_SMI_STATUS=$? if [ "$NVIDIA_SMI_STATUS" -ne 0 ] && [ "$NVIDIA_SMI_STATUS" -ne 14 ]; then RESET_GPU=1 fi fi if [ "$RESET_GPU" -eq 1 ]; then NVIDIA_DEVICES=$(lspci -D | grep -i NVIDIA | cut -d' ' -f1) # The GPU can get stuck in a failure state if somehow the test crashs the GPU microcode. When this # happens, we'll try to reset all NVIDIA devices https://github.com/pytorch/pytorch/issues/88388 for PCI_ID in $NVIDIA_DEVICES; do DEVICE_ENABLED=$(cat /sys/bus/pci/devices/$PCI_ID/enable) echo "Reseting $PCI_ID (enabled state: $DEVICE_ENABLED)" # This requires sudo permission of course echo "1" | sudo tee /sys/bus/pci/devices/$PCI_ID/reset sleep 1 done fi sudo rm -fv /tmp/nvidia_driver set -e fi ) } post_install_nvidia_driver_common() { ( sudo modprobe nvidia || true echo "After installing NVIDIA driver" lspci lsmod modinfo nvidia || true ( set +e nvidia-smi # NB: Annoyingly, nvidia-smi command returns successfully with return code 0 even in # the case where the driver has already crashed as it still can get the driver version # and some basic information like the bus ID. However, the rest of the information # would be missing (ERR!), for example: # # +-----------------------------------------------------------------------------+ # | NVIDIA-SMI 525.89.02 Driver Version: 525.89.02 CUDA Version: 12.0 | # |-------------------------------+----------------------+----------------------+ # | GPU Name Persistence-M| Bus-Id Disp.A | Volatile Uncorr. ECC | # | Fan Temp Perf Pwr:Usage/Cap| Memory-Usage | GPU-Util Compute M. | # | | | MIG M. | # |===============================+======================+======================| # | 0 ERR! Off | 00000000:00:1E.0 Off | ERR! | # |ERR! ERR! ERR! ERR! / ERR! | 4184MiB / 23028MiB | ERR! Default | # | | | ERR! | # +-------------------------------+----------------------+----------------------+ # # +-----------------------------------------------------------------------------+ # | Processes: | # | GPU GI CI PID Type Process name GPU Memory | # | ID ID Usage | # |=============================================================================| # +-----------------------------------------------------------------------------+ # # This should be reported as a failure instead as it will guarantee to fail when # Docker tries to run with --gpus all # # So, the correct check here is to query one of the missing piece of info like # GPU name, so that the command can fail accordingly nvidia-smi --query-gpu=gpu_name --format=csv,noheader --id=0 NVIDIA_SMI_STATUS=$? # Allowable exit statuses for nvidia-smi, see: https://github.com/NVIDIA/gpu-operator/issues/285 if [ "$NVIDIA_SMI_STATUS" -eq 0 ] || [ "$NVIDIA_SMI_STATUS" -eq 14 ]; then echo "INFO: Ignoring allowed status ${NVIDIA_SMI_STATUS}" else echo "ERROR: nvidia-smi exited with unresolved status ${NVIDIA_SMI_STATUS}" exit ${NVIDIA_SMI_STATUS} fi set -e ) ) } install_nvidia_driver_amzn2() { ( set -x pre_install_nvidia_driver_amzn2 install_nvidia_driver_common post_install_nvidia_driver_common ) } install_nvidia_driver_ubuntu20() { ( set -x install_nvidia_driver_common post_install_nvidia_driver_common ) } echo "== Installing nvidia driver ${DRIVER_FN} ==" case "${DISTRIBUTION}" in amzn*) install_nvidia_driver_amzn2 ;; ubuntu20.04) install_nvidia_driver_ubuntu20 ;; *) echo "ERROR: Unknown distribution ${DISTRIBUTION}" exit 1 ;; esac # Install container toolkit based on distribution echo "== Installing nvidia container toolkit for ${DISTRIBUTION} ==" case "${DISTRIBUTION}" in amzn*) install_nvidia_docker2_amzn2 ;; ubuntu20.04) install_nvidia_docker2_ubuntu20 ;; *) echo "ERROR: Unknown distribution ${DISTRIBUTION}" exit 1 ;; esac echo "GPU_FLAG=--gpus all -e NVIDIA_DRIVER_CAPABILITIES=all" >> "${GITHUB_ENV}" # Fix https://github.com/NVIDIA/nvidia-docker/issues/1648 on runners with # more than one GPUs. This just needs to be run once. The command fails # on subsequent runs and complains that the mode is already on, but that's # ok sudo nvidia-persistenced || true # This should show persistence mode ON nvidia-smi # check if the container-toolkit is correctly installed and CUDA is available inside a container docker run --rm -t --gpus=all public.ecr.aws/docker/library/python:3.13 nvidia-smi 2025-11-03T17:04:19.0169385Z retry_wait_seconds: 10 2025-11-03T17:04:19.0169590Z polling_interval_seconds: 1 2025-11-03T17:04:19.0169794Z warning_on_retry: true 2025-11-03T17:04:19.0169979Z continue_on_error: false 2025-11-03T17:04:19.0170162Z env: 2025-11-03T17:04:19.0170317Z GIT_DEFAULT_BRANCH: main 2025-11-03T17:04:19.0170505Z DRIVER_VERSION: 580.82.07 2025-11-03T17:04:19.0170698Z ##[endgroup] 2025-11-03T17:04:19.1247281Z == Installing nvidia driver NVIDIA-Linux-x86_64-580.82.07.run == 2025-11-03T17:04:19.1248434Z + pre_install_nvidia_driver_amzn2 2025-11-03T17:04:19.1250546Z + sudo yum remove -y nvidia-driver-latest-dkms 2025-11-03T17:04:19.8789443Z No match for argument: nvidia-driver-latest-dkms 2025-11-03T17:04:19.8790373Z No packages marked for removal. 2025-11-03T17:04:19.8848131Z Dependencies resolved. 2025-11-03T17:04:19.8856792Z Nothing to do. 2025-11-03T17:04:19.8857529Z Complete! 2025-11-03T17:04:19.9685339Z + install_nvidia_driver_common 2025-11-03T17:04:19.9691342Z + echo 'Before installing NVIDIA driver' 2025-11-03T17:04:19.9693029Z Before installing NVIDIA driver 2025-11-03T17:04:19.9695057Z + lspci 2025-11-03T17:04:20.0496383Z 00:00.0 Host bridge: Intel Corporation 440FX - 82441FX PMC [Natoma] 2025-11-03T17:04:20.0496874Z 00:01.0 ISA bridge: Intel Corporation 82371SB PIIX3 ISA [Natoma/Triton II] 2025-11-03T17:04:20.0497409Z 00:01.3 Non-VGA unclassified device: Intel Corporation 82371AB/EB/MB PIIX4 ACPI (rev 08) 2025-11-03T17:04:20.0498203Z 00:03.0 VGA compatible controller: Amazon.com, Inc. Device 1111 2025-11-03T17:04:20.0498730Z 00:04.0 Non-Volatile memory controller: Amazon.com, Inc. NVMe EBS Controller 2025-11-03T17:04:20.0499160Z 01:00.0 PCI bridge: Amazon.com, Inc. Device 0200 2025-11-03T17:04:20.0499493Z 02:00.0 PCI bridge: Amazon.com, Inc. Device 0200 2025-11-03T17:04:20.0499815Z 03:00.0 PCI bridge: Amazon.com, Inc. Device 0200 2025-11-03T17:04:20.0500121Z 03:00.1 PCI bridge: Amazon.com, Inc. Device 0200 2025-11-03T17:04:20.0500424Z 03:00.2 PCI bridge: Amazon.com, Inc. Device 0200 2025-11-03T17:04:20.0500719Z 03:00.3 PCI bridge: Amazon.com, Inc. Device 0200 2025-11-03T17:04:20.0501016Z 03:00.4 PCI bridge: Amazon.com, Inc. Device 0200 2025-11-03T17:04:20.0501315Z 03:00.5 PCI bridge: Amazon.com, Inc. Device 0200 2025-11-03T17:04:20.0501679Z 03:00.6 PCI bridge: Amazon.com, Inc. Device 0200 2025-11-03T17:04:20.0501916Z 03:00.7 PCI bridge: Amazon.com, Inc. Device 0200 2025-11-03T17:04:20.0502172Z 03:01.0 PCI bridge: Amazon.com, Inc. Device 0200 2025-11-03T17:04:20.0502423Z 03:01.1 PCI bridge: Amazon.com, Inc. Device 0200 2025-11-03T17:04:20.0502670Z 03:01.2 PCI bridge: Amazon.com, Inc. Device 0200 2025-11-03T17:04:20.0502907Z 03:01.3 PCI bridge: Amazon.com, Inc. Device 0200 2025-11-03T17:04:20.0503149Z 03:01.4 PCI bridge: Amazon.com, Inc. Device 0200 2025-11-03T17:04:20.0503390Z 03:01.5 PCI bridge: Amazon.com, Inc. Device 0200 2025-11-03T17:04:20.0503632Z 03:01.6 PCI bridge: Amazon.com, Inc. Device 0200 2025-11-03T17:04:20.0503872Z 03:01.7 PCI bridge: Amazon.com, Inc. Device 0200 2025-11-03T17:04:20.0504106Z 03:02.0 PCI bridge: Amazon.com, Inc. Device 0200 2025-11-03T17:04:20.0504349Z 03:02.1 PCI bridge: Amazon.com, Inc. Device 0200 2025-11-03T17:04:20.0504594Z 03:02.2 PCI bridge: Amazon.com, Inc. Device 0200 2025-11-03T17:04:20.0504837Z 03:02.3 PCI bridge: Amazon.com, Inc. Device 0200 2025-11-03T17:04:20.0505074Z 03:02.4 PCI bridge: Amazon.com, Inc. Device 0200 2025-11-03T17:04:20.0505314Z 03:02.5 PCI bridge: Amazon.com, Inc. Device 0200 2025-11-03T17:04:20.0505574Z 03:02.6 PCI bridge: Amazon.com, Inc. Device 0200 2025-11-03T17:04:20.0506777Z 03:02.7 PCI bridge: Amazon.com, Inc. Device 0200 2025-11-03T17:04:20.0507033Z 03:03.0 PCI bridge: Amazon.com, Inc. Device 0200 2025-11-03T17:04:20.0507277Z 03:03.1 PCI bridge: Amazon.com, Inc. Device 0200 2025-11-03T17:04:20.0507521Z 03:03.2 PCI bridge: Amazon.com, Inc. Device 0200 2025-11-03T17:04:20.0507764Z 03:03.3 PCI bridge: Amazon.com, Inc. Device 0200 2025-11-03T17:04:20.0508001Z 03:03.4 PCI bridge: Amazon.com, Inc. Device 0200 2025-11-03T17:04:20.0508240Z 03:03.5 PCI bridge: Amazon.com, Inc. Device 0200 2025-11-03T17:04:20.0508482Z 03:03.6 PCI bridge: Amazon.com, Inc. Device 0200 2025-11-03T17:04:20.0508726Z 03:03.7 PCI bridge: Amazon.com, Inc. Device 0200 2025-11-03T17:04:20.0508982Z 24:00.0 PCI bridge: Amazon.com, Inc. Device 0200 2025-11-03T17:04:20.0509238Z 25:00.0 PCI bridge: Amazon.com, Inc. Device 0200 2025-11-03T17:04:20.0509484Z 26:00.0 PCI bridge: Amazon.com, Inc. Device 0200 2025-11-03T17:04:20.0509743Z 26:00.1 PCI bridge: Amazon.com, Inc. Device 0200 2025-11-03T17:04:20.0509988Z 26:00.2 PCI bridge: Amazon.com, Inc. Device 0200 2025-11-03T17:04:20.0510225Z 26:00.3 PCI bridge: Amazon.com, Inc. Device 0200 2025-11-03T17:04:20.0510466Z 26:00.4 PCI bridge: Amazon.com, Inc. Device 0200 2025-11-03T17:04:20.0510706Z 26:00.5 PCI bridge: Amazon.com, Inc. Device 0200 2025-11-03T17:04:20.0510946Z 26:00.6 PCI bridge: Amazon.com, Inc. Device 0200 2025-11-03T17:04:20.0511182Z 26:00.7 PCI bridge: Amazon.com, Inc. Device 0200 2025-11-03T17:04:20.0511421Z 26:01.0 PCI bridge: Amazon.com, Inc. Device 0200 2025-11-03T17:04:20.0511741Z 27:00.0 Ethernet controller: Amazon.com, Inc. Elastic Network Adapter (ENA) 2025-11-03T17:04:20.0512068Z 30:00.0 PCI bridge: Amazon.com, Inc. Device 0200 2025-11-03T17:04:20.0512306Z 31:00.0 PCI bridge: Amazon.com, Inc. Device 0200 2025-11-03T17:04:20.0512549Z 32:00.0 PCI bridge: Amazon.com, Inc. Device 0200 2025-11-03T17:04:20.0512869Z 33:00.0 Non-Volatile memory controller: Amazon.com, Inc. NVMe SSD Controller 2025-11-03T17:04:20.0513282Z 34:00.0 PCI bridge: Amazon.com, Inc. Device 0200 2025-11-03T17:04:20.0513571Z 35:00.0 3D controller: NVIDIA Corporation AD104GL [L4] (rev a1) 2025-11-03T17:04:20.0513828Z + lsmod 2025-11-03T17:04:20.0546609Z Module Size Used by 2025-11-03T17:04:20.0546938Z nvidia_uvm 1925120 0 2025-11-03T17:04:20.0547211Z nvidia 14286848 1 nvidia_uvm 2025-11-03T17:04:20.0547494Z drm 602112 1 nvidia 2025-11-03T17:04:20.0547777Z drm_panel_orientation_quirks 32768 1 drm 2025-11-03T17:04:20.0548076Z backlight 24576 1 drm 2025-11-03T17:04:20.0548349Z i2c_core 110592 2 nvidia,drm 2025-11-03T17:04:20.0548627Z xt_conntrack 16384 1 2025-11-03T17:04:20.0548861Z nft_chain_nat 16384 3 2025-11-03T17:04:20.0549101Z xt_MASQUERADE 20480 1 2025-11-03T17:04:20.0549374Z nf_nat 57344 2 nft_chain_nat,xt_MASQUERADE 2025-11-03T17:04:20.0549847Z nf_conntrack_netlink 57344 0 2025-11-03T17:04:20.0550330Z nf_conntrack 184320 4 xt_conntrack,nf_nat,nf_conntrack_netlink,xt_MASQUERADE 2025-11-03T17:04:20.0550756Z nf_defrag_ipv6 24576 1 nf_conntrack 2025-11-03T17:04:20.0551047Z nf_defrag_ipv4 16384 1 nf_conntrack 2025-11-03T17:04:20.0551334Z xfrm_user 57344 1 2025-11-03T17:04:20.0551580Z xfrm_algo 16384 1 xfrm_user 2025-11-03T17:04:20.0551785Z xt_addrtype 16384 2 2025-11-03T17:04:20.0551977Z nft_compat 20480 4 2025-11-03T17:04:20.0552204Z nf_tables 311296 57 nft_compat,nft_chain_nat 2025-11-03T17:04:20.0552516Z nfnetlink 20480 4 nft_compat,nf_conntrack_netlink,nf_tables 2025-11-03T17:04:20.0552787Z br_netfilter 36864 0 2025-11-03T17:04:20.0552989Z bridge 323584 1 br_netfilter 2025-11-03T17:04:20.0553216Z stp 16384 1 bridge 2025-11-03T17:04:20.0553433Z llc 16384 2 bridge,stp 2025-11-03T17:04:20.0553718Z overlay 167936 0 2025-11-03T17:04:20.0554168Z tls 139264 0 2025-11-03T17:04:20.0554356Z nls_ascii 16384 1 2025-11-03T17:04:20.0554544Z nls_cp437 20480 1 2025-11-03T17:04:20.0554727Z vfat 24576 1 2025-11-03T17:04:20.0554911Z fat 86016 1 vfat 2025-11-03T17:04:20.0555101Z ena 184320 0 2025-11-03T17:04:20.0555276Z i8042 45056 0 2025-11-03T17:04:20.0555452Z serio 28672 3 i8042 2025-11-03T17:04:20.0555655Z ghash_clmulni_intel 16384 0 2025-11-03T17:04:20.0555840Z button 24576 0 2025-11-03T17:04:20.0556018Z sunrpc 700416 1 2025-11-03T17:04:20.0556194Z sch_fq_codel 20480 9 2025-11-03T17:04:20.0556388Z dm_mod 188416 0 2025-11-03T17:04:20.0556565Z fuse 184320 1 2025-11-03T17:04:20.0556738Z loop 36864 0 2025-11-03T17:04:20.0556907Z configfs 57344 1 2025-11-03T17:04:20.0557089Z dmi_sysfs 20480 0 2025-11-03T17:04:20.0557272Z crc32_pclmul 16384 0 2025-11-03T17:04:20.0557455Z crc32c_intel 24576 0 2025-11-03T17:04:20.0557627Z efivarfs 24576 1 2025-11-03T17:04:20.0557804Z + modinfo nvidia 2025-11-03T17:04:20.0567582Z filename: /lib/modules/6.1.150-174.273.amzn2023.x86_64/kernel/drivers/video/nvidia.ko 2025-11-03T17:04:20.0568010Z import_ns: DMA_BUF 2025-11-03T17:04:20.0568242Z alias: char-major-195-* 2025-11-03T17:04:20.0568493Z version: 580.82.07 2025-11-03T17:04:20.0568719Z supported: external 2025-11-03T17:04:20.0568952Z license: Dual MIT/GPL 2025-11-03T17:04:20.0569213Z firmware: nvidia/580.82.07/gsp_tu10x.bin 2025-11-03T17:04:20.0569530Z firmware: nvidia/580.82.07/gsp_ga10x.bin 2025-11-03T17:04:20.0569825Z srcversion: BA7240A71DCF7DC6FE88C1D 2025-11-03T17:04:20.0570135Z alias: of:N*T*Cnvidia,tegra264-displayC* 2025-11-03T17:04:20.0570672Z alias: of:N*T*Cnvidia,tegra264-display 2025-11-03T17:04:20.0571009Z alias: of:N*T*Cnvidia,tegra234-displayC* 2025-11-03T17:04:20.0571333Z alias: of:N*T*Cnvidia,tegra234-display 2025-11-03T17:04:20.0571640Z alias: pci:v000010DEd*sv*sd*bc06sc80i00* 2025-11-03T17:04:20.0571931Z alias: pci:v000010DEd*sv*sd*bc03sc02i00* 2025-11-03T17:04:20.0572189Z alias: pci:v000010DEd*sv*sd*bc03sc00i00* 2025-11-03T17:04:20.0572415Z depends: i2c-core,drm 2025-11-03T17:04:20.0572609Z retpoline: Y 2025-11-03T17:04:20.0572775Z name: nvidia 2025-11-03T17:04:20.0573040Z vermagic: 6.1.150-174.273.amzn2023.x86_64 SMP preempt mod_unload modversions 2025-11-03T17:04:20.0573407Z parm: NvSwitchRegDwords:NvSwitch regkey (charp) 2025-11-03T17:04:20.0573753Z parm: NvSwitchBlacklist:NvSwitchBlacklist=uuid[,uuid...] (charp) 2025-11-03T17:04:20.0574073Z parm: NVreg_ResmanDebugLevel:int 2025-11-03T17:04:20.0574304Z parm: NVreg_RmLogonRC:int 2025-11-03T17:04:20.0574529Z parm: NVreg_ModifyDeviceFiles:int 2025-11-03T17:04:20.0574762Z parm: NVreg_DeviceFileUID:int 2025-11-03T17:04:20.0574988Z parm: NVreg_DeviceFileGID:int 2025-11-03T17:04:20.0575209Z parm: NVreg_DeviceFileMode:int 2025-11-03T17:04:20.0575473Z parm: NVreg_InitializeSystemMemoryAllocations:int 2025-11-03T17:04:20.0575758Z parm: NVreg_UsePageAttributeTable:int 2025-11-03T17:04:20.0576000Z parm: NVreg_EnablePCIeGen3:int 2025-11-03T17:04:20.0576219Z parm: NVreg_EnableMSI:int 2025-11-03T17:04:20.0576435Z parm: NVreg_EnableStreamMemOPs:int 2025-11-03T17:04:20.0576700Z parm: NVreg_RestrictProfilingToAdminUsers:int 2025-11-03T17:04:20.0576993Z parm: NVreg_PreserveVideoMemoryAllocations:int 2025-11-03T17:04:20.0577275Z parm: NVreg_EnableS0ixPowerManagement:int 2025-11-03T17:04:20.0577570Z parm: NVreg_S0ixPowerManagementVideoMemoryThreshold:int 2025-11-03T17:04:20.0577877Z parm: NVreg_DynamicPowerManagement:int 2025-11-03T17:04:20.0578275Z parm: NVreg_DynamicPowerManagementVideoMemoryThreshold:int 2025-11-03T17:04:20.0578576Z parm: NVreg_EnableGpuFirmware:int 2025-11-03T17:04:20.0578813Z parm: NVreg_EnableGpuFirmwareLogs:int 2025-11-03T17:04:20.0579082Z parm: NVreg_OpenRmEnableUnsupportedGpus:int 2025-11-03T17:04:20.0579356Z parm: NVreg_EnableUserNUMAManagement:int 2025-11-03T17:04:20.0579604Z parm: NVreg_MemoryPoolSize:int 2025-11-03T17:04:20.0579838Z parm: NVreg_KMallocHeapMaxSize:int 2025-11-03T17:04:20.0580071Z parm: NVreg_VMallocHeapMaxSize:int 2025-11-03T17:04:20.0580303Z parm: NVreg_IgnoreMMIOCheck:int 2025-11-03T17:04:20.0580530Z parm: NVreg_NvLinkDisable:int 2025-11-03T17:04:20.0580799Z parm: NVreg_EnablePCIERelaxedOrderingMode:int 2025-11-03T17:04:20.0581056Z parm: NVreg_RegisterPCIDriver:int 2025-11-03T17:04:20.0581316Z parm: NVreg_RegisterPlatformDeviceDriver:int 2025-11-03T17:04:20.0581741Z parm: NVreg_EnableResizableBar:int 2025-11-03T17:04:20.0581994Z parm: NVreg_EnableDbgBreakpoint:int 2025-11-03T17:04:20.0582279Z parm: NVreg_EnableNonblockingOpen:int 2025-11-03T17:04:20.0582540Z parm: NVreg_CoherentGPUMemoryMode:charp 2025-11-03T17:04:20.0582791Z parm: NVreg_RegistryDwords:charp 2025-11-03T17:04:20.0583040Z parm: NVreg_RegistryDwordsPerDevice:charp 2025-11-03T17:04:20.0583274Z parm: NVreg_RmMsg:charp 2025-11-03T17:04:20.0583492Z parm: NVreg_GpuBlacklist:charp 2025-11-03T17:04:20.0583734Z parm: NVreg_TemporaryFilePath:charp 2025-11-03T17:04:20.0583973Z parm: NVreg_ExcludedGpus:charp 2025-11-03T17:04:20.0584199Z parm: NVreg_DmaRemapPeerMmio:int 2025-11-03T17:04:20.0584441Z parm: NVreg_RmNvlinkBandwidth:charp 2025-11-03T17:04:20.0584700Z parm: NVreg_RmNvlinkBandwidthLinkCount:int 2025-11-03T17:04:20.0585092Z parm: NVreg_ImexChannelCount:int 2025-11-03T17:04:20.0585334Z parm: NVreg_CreateImexChannel0:int 2025-11-03T17:04:20.0585591Z parm: NVreg_GrdmaPciTopoCheckOverride:int 2025-11-03T17:04:20.0585842Z parm: rm_firmware_active:charp 2025-11-03T17:04:20.0586061Z + HAS_NVIDIA_DRIVER=0 2025-11-03T17:04:20.0586246Z ++ command -v nvidia-smi 2025-11-03T17:04:20.0586441Z + '[' -x /usr/bin/nvidia-smi ']' 2025-11-03T17:04:20.0586633Z + set +e 2025-11-03T17:04:20.0586871Z ++ nvidia-smi --query-gpu=driver_version --format=csv,noheader --id=0 2025-11-03T17:04:21.6963598Z + INSTALLED_DRIVER_VERSION=580.82.07 2025-11-03T17:04:21.6964133Z + NVIDIA_SMI_STATUS=0 2025-11-03T17:04:21.6965293Z + '[' 0 -ne 0 ']' 2025-11-03T17:04:21.6965866Z + '[' 580.82.07 '!=' 580.82.07 ']' 2025-11-03T17:04:21.6966474Z + HAS_NVIDIA_DRIVER=1 2025-11-03T17:04:21.6967211Z + echo 'NVIDIA driver (580.82.07) has already been installed. Skipping NVIDIA driver installation' 2025-11-03T17:04:21.6967927Z + set -e 2025-11-03T17:04:21.6968239Z + '[' 1 -eq 0 ']' 2025-11-03T17:04:21.6968820Z NVIDIA driver (580.82.07) has already been installed. Skipping NVIDIA driver installation 2025-11-03T17:04:21.6969473Z + post_install_nvidia_driver_common 2025-11-03T17:04:21.6973227Z + sudo modprobe nvidia 2025-11-03T17:04:21.8117608Z + echo 'After installing NVIDIA driver' 2025-11-03T17:04:21.8118050Z + lspci 2025-11-03T17:04:21.8118329Z After installing NVIDIA driver 2025-11-03T17:04:21.8292813Z 00:00.0 Host bridge: Intel Corporation 440FX - 82441FX PMC [Natoma] 2025-11-03T17:04:21.8293338Z 00:01.0 ISA bridge: Intel Corporation 82371SB PIIX3 ISA [Natoma/Triton II] 2025-11-03T17:04:21.8293873Z 00:01.3 Non-VGA unclassified device: Intel Corporation 82371AB/EB/MB PIIX4 ACPI (rev 08) 2025-11-03T17:04:21.8294378Z 00:03.0 VGA compatible controller: Amazon.com, Inc. Device 1111 2025-11-03T17:04:21.8294850Z 00:04.0 Non-Volatile memory controller: Amazon.com, Inc. NVMe EBS Controller 2025-11-03T17:04:21.8295306Z 01:00.0 PCI bridge: Amazon.com, Inc. Device 0200 2025-11-03T17:04:21.8295948Z 02:00.0 PCI bridge: Amazon.com, Inc. Device 0200 2025-11-03T17:04:21.8296274Z 03:00.0 PCI bridge: Amazon.com, Inc. Device 0200 2025-11-03T17:04:21.8296599Z 03:00.1 PCI bridge: Amazon.com, Inc. Device 0200 2025-11-03T17:04:21.8296900Z 03:00.2 PCI bridge: Amazon.com, Inc. Device 0200 2025-11-03T17:04:21.8297200Z 03:00.3 PCI bridge: Amazon.com, Inc. Device 0200 2025-11-03T17:04:21.8297500Z 03:00.4 PCI bridge: Amazon.com, Inc. Device 0200 2025-11-03T17:04:21.8297796Z 03:00.5 PCI bridge: Amazon.com, Inc. Device 0200 2025-11-03T17:04:21.8298090Z 03:00.6 PCI bridge: Amazon.com, Inc. Device 0200 2025-11-03T17:04:21.8298387Z 03:00.7 PCI bridge: Amazon.com, Inc. Device 0200 2025-11-03T17:04:21.8298683Z 03:01.0 PCI bridge: Amazon.com, Inc. Device 0200 2025-11-03T17:04:21.8298999Z 03:01.1 PCI bridge: Amazon.com, Inc. Device 0200 2025-11-03T17:04:21.8299300Z 03:01.2 PCI bridge: Amazon.com, Inc. Device 0200 2025-11-03T17:04:21.8299616Z 03:01.3 PCI bridge: Amazon.com, Inc. Device 0200 2025-11-03T17:04:21.8299927Z 03:01.4 PCI bridge: Amazon.com, Inc. Device 0200 2025-11-03T17:04:21.8300235Z 03:01.5 PCI bridge: Amazon.com, Inc. Device 0200 2025-11-03T17:04:21.8300529Z 03:01.6 PCI bridge: Amazon.com, Inc. Device 0200 2025-11-03T17:04:21.8300828Z 03:01.7 PCI bridge: Amazon.com, Inc. Device 0200 2025-11-03T17:04:21.8301129Z 03:02.0 PCI bridge: Amazon.com, Inc. Device 0200 2025-11-03T17:04:21.8301431Z 03:02.1 PCI bridge: Amazon.com, Inc. Device 0200 2025-11-03T17:04:21.8301856Z 03:02.2 PCI bridge: Amazon.com, Inc. Device 0200 2025-11-03T17:04:21.8302158Z 03:02.3 PCI bridge: Amazon.com, Inc. Device 0200 2025-11-03T17:04:21.8302457Z 03:02.4 PCI bridge: Amazon.com, Inc. Device 0200 2025-11-03T17:04:21.8302751Z 03:02.5 PCI bridge: Amazon.com, Inc. Device 0200 2025-11-03T17:04:21.8303056Z 03:02.6 PCI bridge: Amazon.com, Inc. Device 0200 2025-11-03T17:04:21.8303361Z 03:02.7 PCI bridge: Amazon.com, Inc. Device 0200 2025-11-03T17:04:21.8303604Z 03:03.0 PCI bridge: Amazon.com, Inc. Device 0200 2025-11-03T17:04:21.8304047Z 03:03.1 PCI bridge: Amazon.com, Inc. Device 0200 2025-11-03T17:04:21.8304317Z 03:03.2 PCI bridge: Amazon.com, Inc. Device 0200 2025-11-03T17:04:21.8304563Z 03:03.3 PCI bridge: Amazon.com, Inc. Device 0200 2025-11-03T17:04:21.8304803Z 03:03.4 PCI bridge: Amazon.com, Inc. Device 0200 2025-11-03T17:04:21.8305043Z 03:03.5 PCI bridge: Amazon.com, Inc. Device 0200 2025-11-03T17:04:21.8305280Z 03:03.6 PCI bridge: Amazon.com, Inc. Device 0200 2025-11-03T17:04:21.8305518Z 03:03.7 PCI bridge: Amazon.com, Inc. Device 0200 2025-11-03T17:04:21.8305758Z 24:00.0 PCI bridge: Amazon.com, Inc. Device 0200 2025-11-03T17:04:21.8305998Z 25:00.0 PCI bridge: Amazon.com, Inc. Device 0200 2025-11-03T17:04:21.8306231Z 26:00.0 PCI bridge: Amazon.com, Inc. Device 0200 2025-11-03T17:04:21.8306471Z 26:00.1 PCI bridge: Amazon.com, Inc. Device 0200 2025-11-03T17:04:21.8306729Z 26:00.2 PCI bridge: Amazon.com, Inc. Device 0200 2025-11-03T17:04:21.8306977Z 26:00.3 PCI bridge: Amazon.com, Inc. Device 0200 2025-11-03T17:04:21.8307219Z 26:00.4 PCI bridge: Amazon.com, Inc. Device 0200 2025-11-03T17:04:21.8307461Z 26:00.5 PCI bridge: Amazon.com, Inc. Device 0200 2025-11-03T17:04:21.8307703Z 26:00.6 PCI bridge: Amazon.com, Inc. Device 0200 2025-11-03T17:04:21.8307954Z 26:00.7 PCI bridge: Amazon.com, Inc. Device 0200 2025-11-03T17:04:21.8308197Z 26:01.0 PCI bridge: Amazon.com, Inc. Device 0200 2025-11-03T17:04:21.8308511Z 27:00.0 Ethernet controller: Amazon.com, Inc. Elastic Network Adapter (ENA) 2025-11-03T17:04:21.8308843Z 30:00.0 PCI bridge: Amazon.com, Inc. Device 0200 2025-11-03T17:04:21.8309089Z 31:00.0 PCI bridge: Amazon.com, Inc. Device 0200 2025-11-03T17:04:21.8309331Z 32:00.0 PCI bridge: Amazon.com, Inc. Device 0200 2025-11-03T17:04:21.8309644Z 33:00.0 Non-Volatile memory controller: Amazon.com, Inc. NVMe SSD Controller 2025-11-03T17:04:21.8309967Z 34:00.0 PCI bridge: Amazon.com, Inc. Device 0200 2025-11-03T17:04:21.8310251Z 35:00.0 3D controller: NVIDIA Corporation AD104GL [L4] (rev a1) 2025-11-03T17:04:21.8310511Z + lsmod 2025-11-03T17:04:21.8331863Z Module Size Used by 2025-11-03T17:04:21.8332361Z nvidia_uvm 1925120 0 2025-11-03T17:04:21.8332624Z nvidia 14286848 1 nvidia_uvm 2025-11-03T17:04:21.8332904Z drm 602112 1 nvidia 2025-11-03T17:04:21.8333207Z drm_panel_orientation_quirks 32768 1 drm 2025-11-03T17:04:21.8333500Z backlight 24576 1 drm 2025-11-03T17:04:21.8333776Z i2c_core 110592 2 nvidia,drm 2025-11-03T17:04:21.8334048Z xt_conntrack 16384 1 2025-11-03T17:04:21.8334295Z nft_chain_nat 16384 3 2025-11-03T17:04:21.8334522Z xt_MASQUERADE 20480 1 2025-11-03T17:04:21.8334807Z nf_nat 57344 2 nft_chain_nat,xt_MASQUERADE 2025-11-03T17:04:21.8335115Z nf_conntrack_netlink 57344 0 2025-11-03T17:04:21.8335607Z nf_conntrack 184320 4 xt_conntrack,nf_nat,nf_conntrack_netlink,xt_MASQUERADE 2025-11-03T17:04:21.8336098Z nf_defrag_ipv6 24576 1 nf_conntrack 2025-11-03T17:04:21.8336529Z nf_defrag_ipv4 16384 1 nf_conntrack 2025-11-03T17:04:21.8336800Z xfrm_user 57344 1 2025-11-03T17:04:21.8337139Z xfrm_algo 16384 1 xfrm_user 2025-11-03T17:04:21.8337439Z xt_addrtype 16384 2 2025-11-03T17:04:21.8337665Z nft_compat 20480 4 2025-11-03T17:04:21.8337942Z nf_tables 311296 57 nft_compat,nft_chain_nat 2025-11-03T17:04:21.8338327Z nfnetlink 20480 4 nft_compat,nf_conntrack_netlink,nf_tables 2025-11-03T17:04:21.8338674Z br_netfilter 36864 0 2025-11-03T17:04:21.8338913Z bridge 323584 1 br_netfilter 2025-11-03T17:04:21.8339180Z stp 16384 1 bridge 2025-11-03T17:04:21.8339443Z llc 16384 2 bridge,stp 2025-11-03T17:04:21.8339710Z overlay 167936 0 2025-11-03T17:04:21.8339926Z tls 139264 0 2025-11-03T17:04:21.8340141Z nls_ascii 16384 1 2025-11-03T17:04:21.8340359Z nls_cp437 20480 1 2025-11-03T17:04:21.8340756Z vfat 24576 1 2025-11-03T17:04:21.8340987Z fat 86016 1 vfat 2025-11-03T17:04:21.8341228Z ena 184320 0 2025-11-03T17:04:21.8341556Z i8042 45056 0 2025-11-03T17:04:21.8341796Z serio 28672 3 i8042 2025-11-03T17:04:21.8342040Z ghash_clmulni_intel 16384 0 2025-11-03T17:04:21.8342285Z button 24576 0 2025-11-03T17:04:21.8342508Z sunrpc 700416 1 2025-11-03T17:04:21.8342730Z sch_fq_codel 20480 9 2025-11-03T17:04:21.8342944Z dm_mod 188416 0 2025-11-03T17:04:21.8343143Z fuse 184320 1 2025-11-03T17:04:21.8343320Z loop 36864 0 2025-11-03T17:04:21.8343498Z configfs 57344 1 2025-11-03T17:04:21.8343671Z dmi_sysfs 20480 0 2025-11-03T17:04:21.8343862Z crc32_pclmul 16384 0 2025-11-03T17:04:21.8344050Z crc32c_intel 24576 0 2025-11-03T17:04:21.8344232Z efivarfs 24576 1 2025-11-03T17:04:21.8344413Z + modinfo nvidia 2025-11-03T17:04:21.8351443Z filename: /lib/modules/6.1.150-174.273.amzn2023.x86_64/kernel/drivers/video/nvidia.ko 2025-11-03T17:04:21.8351807Z import_ns: DMA_BUF 2025-11-03T17:04:21.8352011Z alias: char-major-195-* 2025-11-03T17:04:21.8352217Z version: 580.82.07 2025-11-03T17:04:21.8352409Z supported: external 2025-11-03T17:04:21.8352600Z license: Dual MIT/GPL 2025-11-03T17:04:21.8352837Z firmware: nvidia/580.82.07/gsp_tu10x.bin 2025-11-03T17:04:21.8353201Z firmware: nvidia/580.82.07/gsp_ga10x.bin 2025-11-03T17:04:21.8353482Z srcversion: BA7240A71DCF7DC6FE88C1D 2025-11-03T17:04:21.8353812Z alias: of:N*T*Cnvidia,tegra264-displayC* 2025-11-03T17:04:21.8354146Z alias: of:N*T*Cnvidia,tegra264-display 2025-11-03T17:04:21.8354468Z alias: of:N*T*Cnvidia,tegra234-displayC* 2025-11-03T17:04:21.8354775Z alias: of:N*T*Cnvidia,tegra234-display 2025-11-03T17:04:21.8355091Z alias: pci:v000010DEd*sv*sd*bc06sc80i00* 2025-11-03T17:04:21.8355600Z alias: pci:v000010DEd*sv*sd*bc03sc02i00* 2025-11-03T17:04:21.8355968Z alias: pci:v000010DEd*sv*sd*bc03sc00i00* 2025-11-03T17:04:21.8356238Z depends: i2c-core,drm 2025-11-03T17:04:21.8356475Z retpoline: Y 2025-11-03T17:04:21.8356677Z name: nvidia 2025-11-03T17:04:21.8357023Z vermagic: 6.1.150-174.273.amzn2023.x86_64 SMP preempt mod_unload modversions 2025-11-03T17:04:21.8357464Z parm: NvSwitchRegDwords:NvSwitch regkey (charp) 2025-11-03T17:04:21.8357881Z parm: NvSwitchBlacklist:NvSwitchBlacklist=uuid[,uuid...] (charp) 2025-11-03T17:04:21.8358266Z parm: NVreg_ResmanDebugLevel:int 2025-11-03T17:04:21.8358550Z parm: NVreg_RmLogonRC:int 2025-11-03T17:04:21.8358823Z parm: NVreg_ModifyDeviceFiles:int 2025-11-03T17:04:21.8359111Z parm: NVreg_DeviceFileUID:int 2025-11-03T17:04:21.8359385Z parm: NVreg_DeviceFileGID:int 2025-11-03T17:04:21.8359663Z parm: NVreg_DeviceFileMode:int 2025-11-03T17:04:21.8360044Z parm: NVreg_InitializeSystemMemoryAllocations:int 2025-11-03T17:04:21.8360484Z parm: NVreg_UsePageAttributeTable:int 2025-11-03T17:04:21.8360784Z parm: NVreg_EnablePCIeGen3:int 2025-11-03T17:04:21.8361056Z parm: NVreg_EnableMSI:int 2025-11-03T17:04:21.8361328Z parm: NVreg_EnableStreamMemOPs:int 2025-11-03T17:04:21.8361658Z parm: NVreg_RestrictProfilingToAdminUsers:int 2025-11-03T17:04:21.8362017Z parm: NVreg_PreserveVideoMemoryAllocations:int 2025-11-03T17:04:21.8362363Z parm: NVreg_EnableS0ixPowerManagement:int 2025-11-03T17:04:21.8362741Z parm: NVreg_S0ixPowerManagementVideoMemoryThreshold:int 2025-11-03T17:04:21.8363060Z parm: NVreg_DynamicPowerManagement:int 2025-11-03T17:04:21.8363366Z parm: NVreg_DynamicPowerManagementVideoMemoryThreshold:int 2025-11-03T17:04:21.8363794Z parm: NVreg_EnableGpuFirmware:int 2025-11-03T17:04:21.8364062Z parm: NVreg_EnableGpuFirmwareLogs:int 2025-11-03T17:04:21.8364340Z parm: NVreg_OpenRmEnableUnsupportedGpus:int 2025-11-03T17:04:21.8364614Z parm: NVreg_EnableUserNUMAManagement:int 2025-11-03T17:04:21.8364862Z parm: NVreg_MemoryPoolSize:int 2025-11-03T17:04:21.8365099Z parm: NVreg_KMallocHeapMaxSize:int 2025-11-03T17:04:21.8365450Z parm: NVreg_VMallocHeapMaxSize:int 2025-11-03T17:04:21.8365696Z parm: NVreg_IgnoreMMIOCheck:int 2025-11-03T17:04:21.8365929Z parm: NVreg_NvLinkDisable:int 2025-11-03T17:04:21.8366180Z parm: NVreg_EnablePCIERelaxedOrderingMode:int 2025-11-03T17:04:21.8366436Z parm: NVreg_RegisterPCIDriver:int 2025-11-03T17:04:21.8366697Z parm: NVreg_RegisterPlatformDeviceDriver:int 2025-11-03T17:04:21.8366961Z parm: NVreg_EnableResizableBar:int 2025-11-03T17:04:21.8367208Z parm: NVreg_EnableDbgBreakpoint:int 2025-11-03T17:04:21.8367468Z parm: NVreg_EnableNonblockingOpen:int 2025-11-03T17:04:21.8367734Z parm: NVreg_CoherentGPUMemoryMode:charp 2025-11-03T17:04:21.8367984Z parm: NVreg_RegistryDwords:charp 2025-11-03T17:04:21.8368236Z parm: NVreg_RegistryDwordsPerDevice:charp 2025-11-03T17:04:21.8368479Z parm: NVreg_RmMsg:charp 2025-11-03T17:04:21.8368686Z parm: NVreg_GpuBlacklist:charp 2025-11-03T17:04:21.8368925Z parm: NVreg_TemporaryFilePath:charp 2025-11-03T17:04:21.8369165Z parm: NVreg_ExcludedGpus:charp 2025-11-03T17:04:21.8369389Z parm: NVreg_DmaRemapPeerMmio:int 2025-11-03T17:04:21.8369617Z parm: NVreg_RmNvlinkBandwidth:charp 2025-11-03T17:04:21.8369873Z parm: NVreg_RmNvlinkBandwidthLinkCount:int 2025-11-03T17:04:21.8370137Z parm: NVreg_ImexChannelCount:int 2025-11-03T17:04:21.8370368Z parm: NVreg_CreateImexChannel0:int 2025-11-03T17:04:21.8370609Z parm: NVreg_GrdmaPciTopoCheckOverride:int 2025-11-03T17:04:21.8370851Z parm: rm_firmware_active:charp 2025-11-03T17:04:21.8371146Z + set +e 2025-11-03T17:04:21.8371287Z + nvidia-smi 2025-11-03T17:04:23.2764288Z Mon Nov 3 17:04:23 2025 2025-11-03T17:04:23.2764698Z +-----------------------------------------------------------------------------------------+ 2025-11-03T17:04:23.2765189Z | NVIDIA-SMI 580.82.07 Driver Version: 580.82.07 CUDA Version: 13.0 | 2025-11-03T17:04:23.2765646Z +-----------------------------------------+------------------------+----------------------+ 2025-11-03T17:04:23.2766108Z | GPU Name Persistence-M | Bus-Id Disp.A | Volatile Uncorr. ECC | 2025-11-03T17:04:23.2766623Z | Fan Temp Perf Pwr:Usage/Cap | Memory-Usage | GPU-Util Compute M. | 2025-11-03T17:04:23.2767155Z | | | MIG M. | 2025-11-03T17:04:23.2767482Z |=========================================+========================+======================| 2025-11-03T17:04:23.2834477Z | 0 NVIDIA L4 Off | 00000000:35:00.0 Off | 0 | 2025-11-03T17:04:23.2834941Z | N/A 40C P0 29W / 72W | 0MiB / 23034MiB | 4% Default | 2025-11-03T17:04:23.2835303Z | | | N/A | 2025-11-03T17:04:23.2835747Z +-----------------------------------------+------------------------+----------------------+ 2025-11-03T17:04:23.2836188Z 2025-11-03T17:04:23.2836357Z +-----------------------------------------------------------------------------------------+ 2025-11-03T17:04:23.2836752Z | Processes: | 2025-11-03T17:04:23.2837147Z | GPU GI CI PID Type Process name GPU Memory | 2025-11-03T17:04:23.2837820Z | ID ID Usage | 2025-11-03T17:04:23.2838162Z |=========================================================================================| 2025-11-03T17:04:23.2839428Z | No running processes found | 2025-11-03T17:04:23.2839897Z +-----------------------------------------------------------------------------------------+ 2025-11-03T17:04:23.6092401Z + nvidia-smi --query-gpu=gpu_name --format=csv,noheader --id=0 2025-11-03T17:04:25.0444431Z NVIDIA L4 2025-11-03T17:04:25.2264174Z + NVIDIA_SMI_STATUS=0 2025-11-03T17:04:25.2264485Z + '[' 0 -eq 0 ']' 2025-11-03T17:04:25.2264732Z + echo 'INFO: Ignoring allowed status 0' 2025-11-03T17:04:25.2265014Z + set -e 2025-11-03T17:04:25.2265210Z INFO: Ignoring allowed status 0 2025-11-03T17:04:25.2272895Z == Installing nvidia container toolkit for amzn2023 == 2025-11-03T17:04:25.2276487Z + sudo yum install -y yum-utils 2025-11-03T17:04:25.6492730Z Last metadata expiration check: 0:20:00 ago on Mon Nov 3 16:44:25 2025. 2025-11-03T17:04:25.6708836Z Package dnf-utils-4.3.0-13.amzn2023.0.5.noarch is already installed. 2025-11-03T17:04:25.7112516Z Dependencies resolved. 2025-11-03T17:04:25.7350522Z Nothing to do. 2025-11-03T17:04:25.7351119Z Complete! 2025-11-03T17:04:25.9892343Z + [[ amzn2023 == \a\m\z\n\2\0\2\3 ]] 2025-11-03T17:04:25.9892866Z + YUM_REPO_URL=https://nvidia.github.io/libnvidia-container/stable/rpm/nvidia-container-toolkit.repo 2025-11-03T17:04:25.9893580Z + sudo yum-config-manager --add-repo https://nvidia.github.io/libnvidia-container/stable/rpm/nvidia-container-toolkit.repo 2025-11-03T17:04:26.2784931Z Adding repo from: https://nvidia.github.io/libnvidia-container/stable/rpm/nvidia-container-toolkit.repo 2025-11-03T17:04:26.3282316Z + sudo yum install -y nvidia-container-toolkit-1.17.8 libnvidia-container-tools-1.17.8 libnvidia-container1-1.17.8 nvidia-container-toolkit-base-1.17.8 2025-11-03T17:04:26.8766173Z nvidia-container-toolkit 23 kB/s | 833 B 00:00 2025-11-03T17:04:26.9446152Z Dependencies resolved. 2025-11-03T17:04:26.9676322Z ================================================================================ 2025-11-03T17:04:26.9676760Z Package Arch Version Repository Size 2025-11-03T17:04:26.9677133Z ================================================================================ 2025-11-03T17:04:26.9677434Z Downgrading: 2025-11-03T17:04:26.9677783Z libnvidia-container-tools x86_64 1.17.8-1 nvidia-container-toolkit 40 k 2025-11-03T17:04:26.9678301Z libnvidia-container1 x86_64 1.17.8-1 nvidia-container-toolkit 1.0 M 2025-11-03T17:04:26.9678811Z nvidia-container-toolkit x86_64 1.17.8-1 nvidia-container-toolkit 1.2 M 2025-11-03T17:04:26.9679345Z nvidia-container-toolkit-base x86_64 1.17.8-1 nvidia-container-toolkit 5.8 M 2025-11-03T17:04:26.9679670Z 2025-11-03T17:04:26.9679755Z Transaction Summary 2025-11-03T17:04:26.9679989Z ================================================================================ 2025-11-03T17:04:26.9680298Z Downgrade 4 Packages 2025-11-03T17:04:26.9680450Z 2025-11-03T17:04:26.9680562Z Total download size: 8.0 M 2025-11-03T17:04:26.9682132Z Downloading Packages: 2025-11-03T17:04:27.0023965Z (1/4): libnvidia-container-tools-1.17.8-1.x86_6 1.2 MB/s | 40 kB 00:00 2025-11-03T17:04:27.0199353Z (2/4): libnvidia-container1-1.17.8-1.x86_64.rpm 19 MB/s | 1.0 MB 00:00 2025-11-03T17:04:27.0279851Z (3/4): nvidia-container-toolkit-1.17.8-1.x86_64 21 MB/s | 1.2 MB 00:00 2025-11-03T17:04:27.0566770Z (4/4): nvidia-container-toolkit-base-1.17.8-1.x 107 MB/s | 5.8 MB 00:00 2025-11-03T17:04:27.0579415Z -------------------------------------------------------------------------------- 2025-11-03T17:04:27.0582779Z Total 89 MB/s | 8.0 MB 00:00 2025-11-03T17:04:27.0585530Z Running transaction check 2025-11-03T17:04:27.0705977Z Transaction check succeeded. 2025-11-03T17:04:27.0706752Z Running transaction test 2025-11-03T17:04:27.1071283Z Transaction test succeeded. 2025-11-03T17:04:27.1074008Z Running transaction 2025-11-03T17:04:27.5895080Z Preparing : 1/1 2025-11-03T17:04:27.7145089Z Downgrading : nvidia-container-toolkit-base-1.17.8-1.x86_64 1/8 2025-11-03T17:04:27.7434018Z Downgrading : libnvidia-container1-1.17.8-1.x86_64 2/8 2025-11-03T17:04:27.8169080Z Running scriptlet: libnvidia-container1-1.17.8-1.x86_64 2/8 2025-11-03T17:04:27.9314069Z Downgrading : libnvidia-container-tools-1.17.8-1.x86_64 3/8 2025-11-03T17:04:27.9610388Z Downgrading : nvidia-container-toolkit-1.17.8-1.x86_64 4/8 2025-11-03T17:04:27.9738736Z Running scriptlet: nvidia-container-toolkit-1.17.8-1.x86_64 4/8 2025-11-03T17:04:27.9739669Z Cleanup : nvidia-container-toolkit-1.18.0-1.x86_64 5/8 2025-11-03T17:04:28.0069140Z Running scriptlet: nvidia-container-toolkit-1.18.0-1.x86_64 5/8 2025-11-03T17:04:28.0382454Z Cleanup : libnvidia-container-tools-1.18.0-1.x86_64 6/8 2025-11-03T17:04:28.0383455Z Cleanup : libnvidia-container1-1.18.0-1.x86_64 7/8 2025-11-03T17:04:28.1359555Z Running scriptlet: libnvidia-container1-1.18.0-1.x86_64 7/8 2025-11-03T17:04:28.1795840Z Cleanup : nvidia-container-toolkit-base-1.18.0-1.x86_64 8/8 2025-11-03T17:04:28.2256345Z Running scriptlet: nvidia-container-toolkit-1.17.8-1.x86_64 8/8 2025-11-03T17:05:19.8087572Z Running scriptlet: nvidia-container-toolkit-base-1.18.0-1.x86_64 8/8 2025-11-03T17:05:19.8088529Z Verifying : libnvidia-container-tools-1.17.8-1.x86_64 1/8 2025-11-03T17:05:19.8089462Z Verifying : libnvidia-container-tools-1.18.0-1.x86_64 2/8 2025-11-03T17:05:19.8090098Z Verifying : libnvidia-container1-1.17.8-1.x86_64 3/8 2025-11-03T17:05:19.8090990Z Verifying : libnvidia-container1-1.18.0-1.x86_64 4/8 2025-11-03T17:05:19.8091395Z Verifying : nvidia-container-toolkit-1.17.8-1.x86_64 5/8 2025-11-03T17:05:19.8091801Z Verifying : nvidia-container-toolkit-1.18.0-1.x86_64 6/8 2025-11-03T17:05:19.8092202Z Verifying : nvidia-container-toolkit-base-1.17.8-1.x86_64 7/8 2025-11-03T17:05:19.9418871Z Verifying : nvidia-container-toolkit-base-1.18.0-1.x86_64 8/8================================================================================ 2025-11-03T17:05:19.9419791Z WARNING: 2025-11-03T17:05:19.9420093Z A newer release of "Amazon Linux" is available. 2025-11-03T17:05:19.9420376Z 2025-11-03T17:05:19.9420528Z Available Versions: 2025-11-03T17:05:19.9420789Z 2025-11-03T17:05:19.9420937Z Version 2023.9.20250929: 2025-11-03T17:05:19.9421719Z Run the following command to upgrade to 2023.9.20250929: 2025-11-03T17:05:19.9421992Z 2025-11-03T17:05:19.9422161Z dnf upgrade --releasever=2023.9.20250929 2025-11-03T17:05:19.9422484Z 2025-11-03T17:05:19.9422578Z Release notes: 2025-11-03T17:05:19.9422990Z https://docs.aws.amazon.com/linux/al2023/release-notes/relnotes-2023.9.20250929.html 2025-11-03T17:05:19.9423370Z 2025-11-03T17:05:19.9423452Z Version 2023.9.20251014: 2025-11-03T17:05:19.9423745Z Run the following command to upgrade to 2023.9.20251014: 2025-11-03T17:05:19.9423980Z 2025-11-03T17:05:19.9424095Z dnf upgrade --releasever=2023.9.20251014 2025-11-03T17:05:19.9424294Z 2025-11-03T17:05:19.9424370Z Release notes: 2025-11-03T17:05:19.9424744Z https://docs.aws.amazon.com/linux/al2023/release-notes/relnotes-2023.9.20251014.html 2025-11-03T17:05:19.9425103Z 2025-11-03T17:05:19.9425185Z Version 2023.9.20251020: 2025-11-03T17:05:19.9425470Z Run the following command to upgrade to 2023.9.20251020: 2025-11-03T17:05:19.9425701Z 2025-11-03T17:05:19.9426124Z dnf upgrade --releasever=2023.9.20251020 2025-11-03T17:05:19.9426337Z 2025-11-03T17:05:19.9426416Z Release notes: 2025-11-03T17:05:19.9426785Z https://docs.aws.amazon.com/linux/al2023/release-notes/relnotes-2023.9.20251020.html 2025-11-03T17:05:19.9427140Z 2025-11-03T17:05:19.9427227Z Version 2023.9.20251027: 2025-11-03T17:05:19.9427507Z Run the following command to upgrade to 2023.9.20251027: 2025-11-03T17:05:19.9427742Z 2025-11-03T17:05:19.9427853Z dnf upgrade --releasever=2023.9.20251027 2025-11-03T17:05:19.9428046Z 2025-11-03T17:05:19.9428121Z Release notes: 2025-11-03T17:05:19.9428489Z https://docs.aws.amazon.com/linux/al2023/release-notes/relnotes-2023.9.20251027.html 2025-11-03T17:05:19.9428835Z 2025-11-03T17:05:19.9428938Z ================================================================================ 2025-11-03T17:05:19.9886702Z 2025-11-03T17:05:19.9886825Z 2025-11-03T17:05:19.9886914Z Downgraded: 2025-11-03T17:05:19.9887284Z libnvidia-container-tools-1.17.8-1.x86_64 2025-11-03T17:05:19.9887830Z libnvidia-container1-1.17.8-1.x86_64 2025-11-03T17:05:19.9888343Z nvidia-container-toolkit-1.17.8-1.x86_64 2025-11-03T17:05:19.9888868Z nvidia-container-toolkit-base-1.17.8-1.x86_64 2025-11-03T17:05:19.9889182Z 2025-11-03T17:05:19.9889260Z Complete! 2025-11-03T17:05:20.0350517Z + sudo systemctl restart docker 2025-11-03T17:05:27.3548528Z Mon Nov 3 17:05:27 2025 2025-11-03T17:05:27.3548955Z +-----------------------------------------------------------------------------------------+ 2025-11-03T17:05:27.3549465Z | NVIDIA-SMI 580.82.07 Driver Version: 580.82.07 CUDA Version: 13.0 | 2025-11-03T17:05:27.3549935Z +-----------------------------------------+------------------------+----------------------+ 2025-11-03T17:05:27.3550447Z | GPU Name Persistence-M | Bus-Id Disp.A | Volatile Uncorr. ECC | 2025-11-03T17:05:27.3551291Z | Fan Temp Perf Pwr:Usage/Cap | Memory-Usage | GPU-Util Compute M. | 2025-11-03T17:05:27.3551704Z | | | MIG M. | 2025-11-03T17:05:27.3552003Z |=========================================+========================+======================| 2025-11-03T17:05:27.3620612Z | 0 NVIDIA L4 On | 00000000:35:00.0 Off | 0 | 2025-11-03T17:05:27.3621453Z | N/A 40C P0 29W / 72W | 0MiB / 23034MiB | 4% Default | 2025-11-03T17:05:27.3622048Z | | | N/A | 2025-11-03T17:05:27.3622559Z +-----------------------------------------+------------------------+----------------------+ 2025-11-03T17:05:27.3622972Z 2025-11-03T17:05:27.3623196Z +-----------------------------------------------------------------------------------------+ 2025-11-03T17:05:27.3623634Z | Processes: | 2025-11-03T17:05:27.3624047Z | GPU GI CI PID Type Process name GPU Memory | 2025-11-03T17:05:27.3624424Z | ID ID Usage | 2025-11-03T17:05:27.3624734Z |=========================================================================================| 2025-11-03T17:05:27.3625647Z | No running processes found | 2025-11-03T17:05:27.3626072Z +-----------------------------------------------------------------------------------------+ 2025-11-03T17:05:27.5218765Z Unable to find image 'public.ecr.aws/docker/library/python:3.13' locally 2025-11-03T17:05:27.7724386Z 3.13: Pulling from docker/library/python 2025-11-03T17:05:27.8740714Z 795dbedde24d: Pulling fs layer 2025-11-03T17:05:27.8740995Z 89d573bf42b3: Pulling fs layer 2025-11-03T17:05:27.8741465Z 26dfe2fac1c4: Pulling fs layer 2025-11-03T17:05:27.8741825Z 79d5bd8a8d26: Pulling fs layer 2025-11-03T17:05:27.8742031Z d53b4b1b9ebd: Pulling fs layer 2025-11-03T17:05:27.8742234Z 73ea62972d18: Pulling fs layer 2025-11-03T17:05:27.8742430Z eb23eeea1475: Pulling fs layer 2025-11-03T17:05:27.8742618Z 79d5bd8a8d26: Waiting 2025-11-03T17:05:27.8742798Z d53b4b1b9ebd: Waiting 2025-11-03T17:05:27.8742983Z 73ea62972d18: Waiting 2025-11-03T17:05:27.8743146Z eb23eeea1475: Waiting 2025-11-03T17:05:28.0604826Z 89d573bf42b3: Verifying Checksum 2025-11-03T17:05:28.0605206Z 89d573bf42b3: Download complete 2025-11-03T17:05:28.0731160Z 795dbedde24d: Verifying Checksum 2025-11-03T17:05:28.0731534Z 795dbedde24d: Download complete 2025-11-03T17:05:28.0914405Z 26dfe2fac1c4: Verifying Checksum 2025-11-03T17:05:28.0914748Z 26dfe2fac1c4: Download complete 2025-11-03T17:05:28.1536997Z d53b4b1b9ebd: Verifying Checksum 2025-11-03T17:05:28.1537517Z d53b4b1b9ebd: Download complete 2025-11-03T17:05:28.1846895Z eb23eeea1475: Verifying Checksum 2025-11-03T17:05:28.1847213Z eb23eeea1475: Download complete 2025-11-03T17:05:28.2270274Z 73ea62972d18: Verifying Checksum 2025-11-03T17:05:28.2270575Z 73ea62972d18: Download complete 2025-11-03T17:05:29.1090595Z 79d5bd8a8d26: Verifying Checksum 2025-11-03T17:05:29.1091134Z 79d5bd8a8d26: Download complete 2025-11-03T17:05:29.4174394Z 795dbedde24d: Pull complete 2025-11-03T17:05:29.9610302Z 89d573bf42b3: Pull complete 2025-11-03T17:05:31.7793097Z 26dfe2fac1c4: Pull complete 2025-11-03T17:05:36.9657311Z 79d5bd8a8d26: Pull complete 2025-11-03T17:05:37.3430104Z d53b4b1b9ebd: Pull complete 2025-11-03T17:05:37.9790165Z 73ea62972d18: Pull complete 2025-11-03T17:05:38.0031120Z eb23eeea1475: Pull complete 2025-11-03T17:05:38.0167307Z Digest: sha256:12513c633252a28bcfee85839aa384e1af322f11275779c6645076c6cd0cfe52 2025-11-03T17:05:38.0208064Z Status: Downloaded newer image for public.ecr.aws/docker/library/python:3.13 2025-11-03T17:05:45.2257208Z Mon Nov 3 17:05:45 2025 2025-11-03T17:05:45.2257670Z +-----------------------------------------------------------------------------------------+ 2025-11-03T17:05:45.2258504Z | NVIDIA-SMI 580.82.07 Driver Version: 580.82.07 CUDA Version: 13.0 | 2025-11-03T17:05:45.2258961Z +-----------------------------------------+------------------------+----------------------+ 2025-11-03T17:05:45.2259420Z | GPU Name Persistence-M | Bus-Id Disp.A | Volatile Uncorr. ECC | 2025-11-03T17:05:45.2259927Z | Fan Temp Perf Pwr:Usage/Cap | Memory-Usage | GPU-Util Compute M. | 2025-11-03T17:05:45.2260339Z | | | MIG M. | 2025-11-03T17:05:45.2260646Z |=========================================+========================+======================| 2025-11-03T17:05:45.2373917Z | 0 NVIDIA L4 On | 00000000:35:00.0 Off | 0 | 2025-11-03T17:05:45.2374381Z | N/A 39C P8 13W / 72W | 0MiB / 23034MiB | 0% Default | 2025-11-03T17:05:45.2374763Z | | | N/A | 2025-11-03T17:05:45.2375128Z +-----------------------------------------+------------------------+----------------------+ 2025-11-03T17:05:45.2377409Z 2025-11-03T17:05:45.2377609Z +-----------------------------------------------------------------------------------------+ 2025-11-03T17:05:45.2378022Z | Processes: | 2025-11-03T17:05:45.2378444Z | GPU GI CI PID Type Process name GPU Memory | 2025-11-03T17:05:45.2378826Z | ID ID Usage | 2025-11-03T17:05:45.2379135Z |=========================================================================================| 2025-11-03T17:05:45.2382765Z | No running processes found | 2025-11-03T17:05:45.2383171Z +-----------------------------------------------------------------------------------------+ 2025-11-03T17:05:46.1658149Z Command completed after 1 attempt(s). 2025-11-03T17:05:46.1743391Z Prepare all required actions 2025-11-03T17:05:46.1767510Z ##[group]Run ./.github/actions/get-workflow-job-id 2025-11-03T17:05:46.1767758Z with: 2025-11-03T17:05:46.1768327Z github-token: *** 2025-11-03T17:05:46.1768512Z env: 2025-11-03T17:05:46.1768662Z GIT_DEFAULT_BRANCH: main 2025-11-03T17:05:46.1768906Z GPU_FLAG: --gpus all -e NVIDIA_DRIVER_CAPABILITIES=all 2025-11-03T17:05:46.1769163Z ##[endgroup] 2025-11-03T17:05:46.1796859Z ##[group]Run set -eux 2025-11-03T17:05:46.1797046Z set -eux 2025-11-03T17:05:46.1797362Z python3 .github/scripts/get_workflow_job_id.py "${GITHUB_RUN_ID}" "${RUNNER_NAME}" 2025-11-03T17:05:46.1810028Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2025-11-03T17:05:46.1810299Z env: 2025-11-03T17:05:46.1810458Z GIT_DEFAULT_BRANCH: main 2025-11-03T17:05:46.1810719Z GPU_FLAG: --gpus all -e NVIDIA_DRIVER_CAPABILITIES=all 2025-11-03T17:05:46.1811113Z GITHUB_TOKEN: *** 2025-11-03T17:05:46.1811295Z ##[endgroup] 2025-11-03T17:05:46.1847077Z + python3 .github/scripts/get_workflow_job_id.py 19040285420 i-0aa7ffe8f5edfd321 2025-11-03T17:05:46.7914490Z Setting output job-id=54383051168 2025-11-03T17:05:46.7915089Z Setting output job-name=ci-vllm-test / test (vllm_basic_models_test, 1, 1, linux.g6.4xlarge.experimental.nvidia.gpu) 2025-11-03T17:05:46.8030271Z ##[group]Run python3 -m pip install psutil==5.9.8 dataclasses_json==0.6.7 nvidia-ml-py==11.525.84 2025-11-03T17:05:46.8030839Z python3 -m pip install psutil==5.9.8 dataclasses_json==0.6.7 nvidia-ml-py==11.525.84 2025-11-03T17:05:46.8031520Z python3 -m tools.stats.monitor --log-interval "$MONITOR_LOG_INTERVAL" --data-collect-interval "$MONITOR_DATA_COLLECT_INTERVAL" > usage_log.txt 2>&1 & 2025-11-03T17:05:46.8032117Z echo "monitor-script-pid=${!}" >> "${GITHUB_OUTPUT}" 2025-11-03T17:05:46.8041566Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2025-11-03T17:05:46.8041841Z env: 2025-11-03T17:05:46.8041998Z GIT_DEFAULT_BRANCH: main 2025-11-03T17:05:46.8042235Z GPU_FLAG: --gpus all -e NVIDIA_DRIVER_CAPABILITIES=all 2025-11-03T17:05:46.8042474Z JOB_ID: 54383051168 2025-11-03T17:05:46.8042815Z JOB_NAME: ci-vllm-test / test (vllm_basic_models_test, 1, 1, linux.g6.4xlarge.experimental.nvidia.gpu) 2025-11-03T17:05:46.8043217Z WORKFLOW_NAME: vllm-test 2025-11-03T17:05:46.8043409Z WORKFLOW_RUN_ID: 19040285420 2025-11-03T17:05:46.8043598Z MONITOR_LOG_INTERVAL: 5 2025-11-03T17:05:46.8043790Z MONITOR_DATA_COLLECT_INTERVAL: 1 2025-11-03T17:05:46.8043991Z ##[endgroup] 2025-11-03T17:05:47.0603769Z Defaulting to user installation because normal site-packages is not writeable 2025-11-03T17:05:47.3951915Z Collecting psutil==5.9.8 2025-11-03T17:05:47.4134751Z Downloading psutil-5.9.8-cp36-abi3-manylinux_2_12_x86_64.manylinux2010_x86_64.manylinux_2_17_x86_64.manylinux2014_x86_64.whl (288 kB) 2025-11-03T17:05:47.4830044Z Collecting dataclasses_json==0.6.7 2025-11-03T17:05:47.4875446Z Downloading dataclasses_json-0.6.7-py3-none-any.whl (28 kB) 2025-11-03T17:05:47.5144349Z Collecting nvidia-ml-py==11.525.84 2025-11-03T17:05:47.5183034Z Downloading nvidia_ml_py-11.525.84-py3-none-any.whl (34 kB) 2025-11-03T17:05:47.6276866Z Collecting marshmallow<4.0.0,>=3.18.0 2025-11-03T17:05:47.6316918Z Downloading marshmallow-3.26.1-py3-none-any.whl (50 kB) 2025-11-03T17:05:47.6546766Z Collecting typing-inspect<1,>=0.4.0 2025-11-03T17:05:47.6585415Z Downloading typing_inspect-0.9.0-py3-none-any.whl (8.8 kB) 2025-11-03T17:05:47.7081298Z Collecting packaging>=17.0 2025-11-03T17:05:47.7120047Z Downloading packaging-25.0-py3-none-any.whl (66 kB) 2025-11-03T17:05:47.7357446Z Collecting mypy-extensions>=0.3.0 2025-11-03T17:05:47.7396756Z Downloading mypy_extensions-1.1.0-py3-none-any.whl (5.0 kB) 2025-11-03T17:05:47.7829770Z Collecting typing-extensions>=3.7.4 2025-11-03T17:05:47.7869520Z Downloading typing_extensions-4.15.0-py3-none-any.whl (44 kB) 2025-11-03T17:05:47.8722138Z Installing collected packages: typing-extensions, packaging, mypy-extensions, typing-inspect, marshmallow, psutil, nvidia-ml-py, dataclasses-json 2025-11-03T17:05:48.1190157Z Successfully installed dataclasses-json-0.6.7 marshmallow-3.26.1 mypy-extensions-1.1.0 nvidia-ml-py-11.525.84 packaging-25.0 psutil-5.9.8 typing-extensions-4.15.0 typing-inspect-0.9.0 2025-11-03T17:05:48.2764955Z Prepare all required actions 2025-11-03T17:05:48.2765307Z Getting action download info 2025-11-03T17:05:48.4882944Z Download action repository 'seemethere/download-artifact-s3@v4' (SHA:1da556a7aa0a088e3153970611f6c432d58e80e6) 2025-11-03T17:05:48.7265462Z Download action repository 'actions/download-artifact@v4' (SHA:d3f86a106a0bac45b974a628896c90dbdf5c8093) 2025-11-03T17:05:49.0823587Z ##[group]Run ./.github/actions/download-build-artifacts 2025-11-03T17:05:49.0823869Z with: 2025-11-03T17:05:49.0824056Z name: linux-jammy-cuda12.8-py3.12-gcc11 2025-11-03T17:05:49.0824306Z s3-bucket: gha-artifacts 2025-11-03T17:05:49.0824521Z env: 2025-11-03T17:05:49.0824707Z GIT_DEFAULT_BRANCH: main 2025-11-03T17:05:49.0824990Z GPU_FLAG: --gpus all -e NVIDIA_DRIVER_CAPABILITIES=all 2025-11-03T17:05:49.0825298Z ##[endgroup] 2025-11-03T17:05:49.0854876Z ##[group]Run seemethere/download-artifact-s3@v4 2025-11-03T17:05:49.0855192Z with: 2025-11-03T17:05:49.0855427Z name: linux-jammy-cuda12.8-py3.12-gcc11 2025-11-03T17:05:49.0855664Z s3-bucket: gha-artifacts 2025-11-03T17:05:49.0855862Z region: us-east-1 2025-11-03T17:05:49.0856021Z env: 2025-11-03T17:05:49.0856184Z GIT_DEFAULT_BRANCH: main 2025-11-03T17:05:49.0856420Z GPU_FLAG: --gpus all -e NVIDIA_DRIVER_CAPABILITIES=all 2025-11-03T17:05:49.0856673Z ##[endgroup] 2025-11-03T17:05:49.4908766Z (node:61200) NOTE: We are formalizing our plans to enter AWS SDK for JavaScript (v2) into maintenance mode in 2023. 2025-11-03T17:05:49.4909244Z 2025-11-03T17:05:49.4909436Z Please migrate your code to use AWS SDK for JavaScript (v3). 2025-11-03T17:05:49.4910302Z For more information, check the migration guide at https://a.co/7PzMCcy 2025-11-03T17:05:49.4910832Z (Use `node --trace-warnings ...` to show where the warning was created) 2025-11-03T17:05:49.7752003Z Found 1 objects with prefix pytorch/pytorch/19040285420/linux-jammy-cuda12.8-py3.12-gcc11/ 2025-11-03T17:05:49.7752771Z Starting download (1/1): /home/ec2-user/actions-runner/_work/pytorch/pytorch/artifacts.zip 2025-11-03T17:06:15.0528444Z Finished download (1/1): /home/ec2-user/actions-runner/_work/pytorch/pytorch/artifacts.zip 2025-11-03T17:06:15.0533874Z Artifact download has finished successfully 2025-11-03T17:06:15.0808500Z ##[group]Run unzip -o artifacts.zip 2025-11-03T17:06:15.0808751Z unzip -o artifacts.zip 2025-11-03T17:06:15.0817479Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2025-11-03T17:06:15.0817779Z env: 2025-11-03T17:06:15.0817951Z GIT_DEFAULT_BRANCH: main 2025-11-03T17:06:15.0818194Z GPU_FLAG: --gpus all -e NVIDIA_DRIVER_CAPABILITIES=all 2025-11-03T17:06:15.0818465Z ##[endgroup] 2025-11-03T17:06:15.1231350Z Archive: artifacts.zip 2025-11-03T17:06:15.1232561Z creating: dist/ 2025-11-03T17:06:17.6680470Z inflating: dist/torch-2.10.0a0+git3f6538f-cp312-cp312-linux_x86_64.whl 2025-11-03T17:06:17.6682567Z creating: dist/vision/ 2025-11-03T17:06:17.6808972Z inflating: dist/vision/torchvision-0.25.0a0+cfbc5c2-cp312-cp312-linux_x86_64.whl 2025-11-03T17:06:17.6809479Z creating: dist/audio/ 2025-11-03T17:06:17.6882246Z inflating: dist/audio/torchaudio-2.10.0a0+3b0e7a6-cp312-cp312-linux_x86_64.whl 2025-11-03T17:06:17.6998106Z inflating: dist/.ninja_log 2025-11-03T17:06:17.6998407Z creating: dist/external/ 2025-11-03T17:06:17.6998849Z creating: dist/external/vllm/ 2025-11-03T17:06:17.6999196Z creating: dist/external/vllm/wheels/ 2025-11-03T17:06:17.7000181Z inflating: dist/external/vllm/wheels/build_summary.txt 2025-11-03T17:06:17.7000617Z creating: dist/external/vllm/wheels/flashinfer-python/ 2025-11-03T17:06:18.3147027Z inflating: dist/external/vllm/wheels/flashinfer-python/flashinfer_python-0.2.14.post1-cp39-abi3-linux_x86_64.whl 2025-11-03T17:06:18.3147693Z creating: dist/external/vllm/wheels/vllm/ 2025-11-03T17:06:21.0098450Z inflating: dist/external/vllm/wheels/vllm/vllm-0.11.0rc2.dev389+ge51928192.d20251103-cp38-abi3-linux_x86_64.whl 2025-11-03T17:06:21.0099524Z creating: dist/external/vllm/wheels/xformers/ 2025-11-03T17:06:22.5923830Z inflating: dist/external/vllm/wheels/xformers/xformers-0.0.33+5d4b92a5.d20251103-cp39-abi3-linux_x86_64.whl 2025-11-03T17:06:22.5924481Z creating: build/custom_test_artifacts/ 2025-11-03T17:06:22.5924853Z creating: build/custom_test_artifacts/custom-op-build/ 2025-11-03T17:06:22.5925447Z creating: build/custom_test_artifacts/custom-op-build/CMakeFiles/ 2025-11-03T17:06:22.5926294Z creating: build/custom_test_artifacts/custom-op-build/CMakeFiles/pkgRedirects/ 2025-11-03T17:06:22.5933918Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/CMakeConfigureLog.yaml 2025-11-03T17:06:22.5934566Z creating: build/custom_test_artifacts/custom-op-build/CMakeFiles/3.31.6/ 2025-11-03T17:06:22.5935462Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/3.31.6/CMakeSystem.cmake 2025-11-03T17:06:22.5936076Z creating: build/custom_test_artifacts/custom-op-build/CMakeFiles/3.31.6/CompilerIdC/ 2025-11-03T17:06:22.5936806Z creating: build/custom_test_artifacts/custom-op-build/CMakeFiles/3.31.6/CompilerIdC/tmp/ 2025-11-03T17:06:22.5939734Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/3.31.6/CompilerIdC/CMakeCCompilerId.c 2025-11-03T17:06:22.5941588Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/3.31.6/CompilerIdC/a.out 2025-11-03T17:06:22.5942754Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/3.31.6/CMakeCCompiler.cmake 2025-11-03T17:06:22.5943419Z creating: build/custom_test_artifacts/custom-op-build/CMakeFiles/3.31.6/CompilerIdCXX/ 2025-11-03T17:06:22.5944372Z creating: build/custom_test_artifacts/custom-op-build/CMakeFiles/3.31.6/CompilerIdCXX/tmp/ 2025-11-03T17:06:22.5947086Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/3.31.6/CompilerIdCXX/CMakeCXXCompilerId.cpp 2025-11-03T17:06:22.5948682Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/3.31.6/CompilerIdCXX/a.out 2025-11-03T17:06:22.5950083Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/3.31.6/CMakeCXXCompiler.cmake 2025-11-03T17:06:22.5952161Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/3.31.6/CMakeDetermineCompilerABI_C.bin 2025-11-03T17:06:22.5954247Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/3.31.6/CMakeDetermineCompilerABI_CXX.bin 2025-11-03T17:06:22.5954852Z creating: build/custom_test_artifacts/custom-op-build/CMakeFiles/3.31.6/CompilerIdCUDA/ 2025-11-03T17:06:22.5955501Z creating: build/custom_test_artifacts/custom-op-build/CMakeFiles/3.31.6/CompilerIdCUDA/tmp/ 2025-11-03T17:06:22.6007821Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/3.31.6/CompilerIdCUDA/tmp/CMakeCUDACompilerId.cpp4.ii 2025-11-03T17:06:22.6061134Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/3.31.6/CompilerIdCUDA/tmp/CMakeCUDACompilerId.cudafe1.cpp 2025-11-03T17:06:22.6062165Z extracting: build/custom_test_artifacts/custom-op-build/CMakeFiles/3.31.6/CompilerIdCUDA/tmp/CMakeCUDACompilerId.module_id 2025-11-03T17:06:22.6118590Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/3.31.6/CompilerIdCUDA/tmp/CMakeCUDACompilerId.cpp1.ii 2025-11-03T17:06:22.6119631Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/3.31.6/CompilerIdCUDA/tmp/CMakeCUDACompilerId.cudafe1.c 2025-11-03T17:06:22.6121409Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/3.31.6/CompilerIdCUDA/tmp/CMakeCUDACompilerId.cudafe1.gpu 2025-11-03T17:06:22.6122536Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/3.31.6/CompilerIdCUDA/tmp/CMakeCUDACompilerId.cudafe1.stub.c 2025-11-03T17:06:22.6123555Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/3.31.6/CompilerIdCUDA/tmp/CMakeCUDACompilerId.ptx 2025-11-03T17:06:22.6124691Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/3.31.6/CompilerIdCUDA/tmp/CMakeCUDACompilerId.sm_52.cubin 2025-11-03T17:06:22.6126105Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/3.31.6/CompilerIdCUDA/tmp/CMakeCUDACompilerId.fatbin 2025-11-03T17:06:22.6127197Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/3.31.6/CompilerIdCUDA/tmp/CMakeCUDACompilerId.fatbin.c 2025-11-03T17:06:22.6128987Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/3.31.6/CompilerIdCUDA/tmp/CMakeCUDACompilerId.o 2025-11-03T17:06:22.6130022Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/3.31.6/CompilerIdCUDA/tmp/a_dlink.sm_52.cubin 2025-11-03T17:06:22.6131045Z extracting: build/custom_test_artifacts/custom-op-build/CMakeFiles/3.31.6/CompilerIdCUDA/tmp/a_dlink.reg.c 2025-11-03T17:06:22.6132192Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/3.31.6/CompilerIdCUDA/tmp/a_dlink.fatbin 2025-11-03T17:06:22.6133438Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/3.31.6/CompilerIdCUDA/tmp/a_dlink.fatbin.c 2025-11-03T17:06:22.6134734Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/3.31.6/CompilerIdCUDA/tmp/a_dlink.o 2025-11-03T17:06:22.6137566Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/3.31.6/CompilerIdCUDA/CMakeCUDACompilerId.cu 2025-11-03T17:06:22.6202389Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/3.31.6/CompilerIdCUDA/a.out 2025-11-03T17:06:22.6203486Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/3.31.6/CMakeCUDACompiler.cmake 2025-11-03T17:06:22.6268645Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/3.31.6/CMakeDetermineCompilerABI_CUDA.bin 2025-11-03T17:06:22.6269552Z creating: build/custom_test_artifacts/custom-op-build/CMakeFiles/CMakeScratch/ 2025-11-03T17:06:22.6270105Z creating: build/custom_test_artifacts/custom-op-build/CMakeFiles/CMakeTmp/ 2025-11-03T17:06:22.6270693Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/cmake.check_cache 2025-11-03T17:06:22.6271284Z creating: build/custom_test_artifacts/custom-op-build/CMakeFiles/custom_ops.dir/ 2025-11-03T17:06:22.6272371Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/custom_ops.dir/compiler_depend.ts 2025-11-03T17:06:22.6273242Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/custom_ops.dir/compiler_depend.make 2025-11-03T17:06:22.6274221Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/custom_ops.dir/depend.make 2025-11-03T17:06:22.6275281Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/custom_ops.dir/link.txt 2025-11-03T17:06:22.6276323Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/custom_ops.dir/cmake_clean.cmake 2025-11-03T17:06:22.6277888Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/custom_ops.dir/build.make 2025-11-03T17:06:22.6278760Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/custom_ops.dir/DependInfo.cmake 2025-11-03T17:06:22.6279774Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/custom_ops.dir/flags.make 2025-11-03T17:06:22.6280926Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/custom_ops.dir/progress.make 2025-11-03T17:06:22.6299840Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/custom_ops.dir/op.cpp.o.d 2025-11-03T17:06:22.6473912Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/custom_ops.dir/op.cpp.o 2025-11-03T17:06:22.6474568Z creating: build/custom_test_artifacts/custom-op-build/CMakeFiles/test_custom_ops.dir/ 2025-11-03T17:06:22.6475423Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/test_custom_ops.dir/compiler_depend.ts 2025-11-03T17:06:22.6476448Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/test_custom_ops.dir/compiler_depend.make 2025-11-03T17:06:22.6477378Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/test_custom_ops.dir/depend.make 2025-11-03T17:06:22.6478606Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/test_custom_ops.dir/link.txt 2025-11-03T17:06:22.6479551Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/test_custom_ops.dir/cmake_clean.cmake 2025-11-03T17:06:22.6480855Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/test_custom_ops.dir/build.make 2025-11-03T17:06:22.6481948Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/test_custom_ops.dir/DependInfo.cmake 2025-11-03T17:06:22.6482880Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/test_custom_ops.dir/flags.make 2025-11-03T17:06:22.6484066Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/test_custom_ops.dir/progress.make 2025-11-03T17:06:22.6503203Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/test_custom_ops.dir/test_custom_ops.cpp.o.d 2025-11-03T17:06:22.6573845Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/test_custom_ops.dir/test_custom_ops.cpp.o 2025-11-03T17:06:22.6574936Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/CMakeDirectoryInformation.cmake 2025-11-03T17:06:22.6575881Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/TargetDirectories.txt 2025-11-03T17:06:22.6576787Z extracting: build/custom_test_artifacts/custom-op-build/CMakeFiles/progress.marks 2025-11-03T17:06:22.6578163Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/Makefile2 2025-11-03T17:06:22.6580238Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/Makefile.cmake 2025-11-03T17:06:22.6581027Z inflating: build/custom_test_artifacts/custom-op-build/detect_cuda_version.cc 2025-11-03T17:06:22.6583963Z inflating: build/custom_test_artifacts/custom-op-build/CMakeCache.txt 2025-11-03T17:06:22.6585089Z inflating: build/custom_test_artifacts/custom-op-build/Makefile 2025-11-03T17:06:22.6586181Z inflating: build/custom_test_artifacts/custom-op-build/cmake_install.cmake 2025-11-03T17:06:22.6736683Z inflating: build/custom_test_artifacts/custom-op-build/libcustom_ops.so 2025-11-03T17:06:22.6785532Z inflating: build/custom_test_artifacts/custom-op-build/test_custom_ops 2025-11-03T17:06:22.6786007Z creating: build/custom_test_artifacts/jit-hook-build/ 2025-11-03T17:06:22.6786428Z creating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/ 2025-11-03T17:06:22.6787095Z creating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/pkgRedirects/ 2025-11-03T17:06:22.6794696Z inflating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/CMakeConfigureLog.yaml 2025-11-03T17:06:22.6795323Z creating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/3.31.6/ 2025-11-03T17:06:22.6796128Z inflating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/3.31.6/CMakeSystem.cmake 2025-11-03T17:06:22.6796742Z creating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/3.31.6/CompilerIdC/ 2025-11-03T17:06:22.6797410Z creating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/3.31.6/CompilerIdC/tmp/ 2025-11-03T17:06:22.6800326Z inflating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/3.31.6/CompilerIdC/CMakeCCompilerId.c 2025-11-03T17:06:22.6801981Z inflating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/3.31.6/CompilerIdC/a.out 2025-11-03T17:06:22.6803828Z inflating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/3.31.6/CMakeCCompiler.cmake 2025-11-03T17:06:22.6804370Z creating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/3.31.6/CompilerIdCXX/ 2025-11-03T17:06:22.6804976Z creating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/3.31.6/CompilerIdCXX/tmp/ 2025-11-03T17:06:22.6807907Z inflating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/3.31.6/CompilerIdCXX/CMakeCXXCompilerId.cpp 2025-11-03T17:06:22.6809734Z inflating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/3.31.6/CompilerIdCXX/a.out 2025-11-03T17:06:22.6811302Z inflating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/3.31.6/CMakeCXXCompiler.cmake 2025-11-03T17:06:22.6813115Z inflating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/3.31.6/CMakeDetermineCompilerABI_C.bin 2025-11-03T17:06:22.6815152Z inflating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/3.31.6/CMakeDetermineCompilerABI_CXX.bin 2025-11-03T17:06:22.6816091Z creating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/3.31.6/CompilerIdCUDA/ 2025-11-03T17:06:22.6816711Z creating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/3.31.6/CompilerIdCUDA/tmp/ 2025-11-03T17:06:22.6869212Z inflating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/3.31.6/CompilerIdCUDA/tmp/CMakeCUDACompilerId.cpp4.ii 2025-11-03T17:06:22.6921992Z inflating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/3.31.6/CompilerIdCUDA/tmp/CMakeCUDACompilerId.cudafe1.cpp 2025-11-03T17:06:22.6923028Z extracting: build/custom_test_artifacts/jit-hook-build/CMakeFiles/3.31.6/CompilerIdCUDA/tmp/CMakeCUDACompilerId.module_id 2025-11-03T17:06:22.6979515Z inflating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/3.31.6/CompilerIdCUDA/tmp/CMakeCUDACompilerId.cpp1.ii 2025-11-03T17:06:22.6980459Z inflating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/3.31.6/CompilerIdCUDA/tmp/CMakeCUDACompilerId.cudafe1.c 2025-11-03T17:06:22.6982069Z inflating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/3.31.6/CompilerIdCUDA/tmp/CMakeCUDACompilerId.cudafe1.gpu 2025-11-03T17:06:22.6983203Z inflating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/3.31.6/CompilerIdCUDA/tmp/CMakeCUDACompilerId.cudafe1.stub.c 2025-11-03T17:06:22.6984340Z inflating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/3.31.6/CompilerIdCUDA/tmp/CMakeCUDACompilerId.ptx 2025-11-03T17:06:22.6985466Z inflating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/3.31.6/CompilerIdCUDA/tmp/CMakeCUDACompilerId.sm_52.cubin 2025-11-03T17:06:22.6986606Z inflating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/3.31.6/CompilerIdCUDA/tmp/CMakeCUDACompilerId.fatbin 2025-11-03T17:06:22.6987920Z inflating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/3.31.6/CompilerIdCUDA/tmp/CMakeCUDACompilerId.fatbin.c 2025-11-03T17:06:22.6989477Z inflating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/3.31.6/CompilerIdCUDA/tmp/CMakeCUDACompilerId.o 2025-11-03T17:06:22.6990558Z inflating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/3.31.6/CompilerIdCUDA/tmp/a_dlink.sm_52.cubin 2025-11-03T17:06:22.6991565Z extracting: build/custom_test_artifacts/jit-hook-build/CMakeFiles/3.31.6/CompilerIdCUDA/tmp/a_dlink.reg.c 2025-11-03T17:06:22.6992724Z inflating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/3.31.6/CompilerIdCUDA/tmp/a_dlink.fatbin 2025-11-03T17:06:22.6993983Z inflating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/3.31.6/CompilerIdCUDA/tmp/a_dlink.fatbin.c 2025-11-03T17:06:22.6995268Z inflating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/3.31.6/CompilerIdCUDA/tmp/a_dlink.o 2025-11-03T17:06:22.6998139Z inflating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/3.31.6/CompilerIdCUDA/CMakeCUDACompilerId.cu 2025-11-03T17:06:22.7062943Z inflating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/3.31.6/CompilerIdCUDA/a.out 2025-11-03T17:06:22.7064097Z inflating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/3.31.6/CMakeCUDACompiler.cmake 2025-11-03T17:06:22.7129224Z inflating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/3.31.6/CMakeDetermineCompilerABI_CUDA.bin 2025-11-03T17:06:22.7129956Z creating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/CMakeScratch/ 2025-11-03T17:06:22.7130494Z creating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/CMakeTmp/ 2025-11-03T17:06:22.7131268Z inflating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/cmake.check_cache 2025-11-03T17:06:22.7131854Z creating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/test_jit_hooks.dir/ 2025-11-03T17:06:22.7133114Z inflating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/test_jit_hooks.dir/compiler_depend.ts 2025-11-03T17:06:22.7133961Z inflating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/test_jit_hooks.dir/compiler_depend.make 2025-11-03T17:06:22.7134895Z inflating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/test_jit_hooks.dir/depend.make 2025-11-03T17:06:22.7135948Z inflating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/test_jit_hooks.dir/link.txt 2025-11-03T17:06:22.7136977Z inflating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/test_jit_hooks.dir/cmake_clean.cmake 2025-11-03T17:06:22.7138560Z inflating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/test_jit_hooks.dir/build.make 2025-11-03T17:06:22.7139447Z inflating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/test_jit_hooks.dir/DependInfo.cmake 2025-11-03T17:06:22.7140472Z inflating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/test_jit_hooks.dir/flags.make 2025-11-03T17:06:22.7141692Z inflating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/test_jit_hooks.dir/progress.make 2025-11-03T17:06:22.7160784Z inflating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/test_jit_hooks.dir/test_jit_hooks.cpp.o.d 2025-11-03T17:06:22.7215755Z inflating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/test_jit_hooks.dir/test_jit_hooks.cpp.o 2025-11-03T17:06:22.7216814Z inflating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/CMakeDirectoryInformation.cmake 2025-11-03T17:06:22.7218271Z inflating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/TargetDirectories.txt 2025-11-03T17:06:22.7219244Z extracting: build/custom_test_artifacts/jit-hook-build/CMakeFiles/progress.marks 2025-11-03T17:06:22.7220524Z inflating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/Makefile2 2025-11-03T17:06:22.7222868Z inflating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/Makefile.cmake 2025-11-03T17:06:22.7223498Z inflating: build/custom_test_artifacts/jit-hook-build/detect_cuda_version.cc 2025-11-03T17:06:22.7226477Z inflating: build/custom_test_artifacts/jit-hook-build/CMakeCache.txt 2025-11-03T17:06:22.7227616Z inflating: build/custom_test_artifacts/jit-hook-build/Makefile 2025-11-03T17:06:22.7228725Z inflating: build/custom_test_artifacts/jit-hook-build/cmake_install.cmake 2025-11-03T17:06:22.7262438Z inflating: build/custom_test_artifacts/jit-hook-build/test_jit_hooks 2025-11-03T17:06:22.7262945Z creating: build/custom_test_artifacts/custom-backend-build/ 2025-11-03T17:06:22.7263536Z creating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/ 2025-11-03T17:06:22.7264166Z creating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/pkgRedirects/ 2025-11-03T17:06:22.7271768Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/CMakeConfigureLog.yaml 2025-11-03T17:06:22.7272428Z creating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/3.31.6/ 2025-11-03T17:06:22.7273157Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/3.31.6/CMakeSystem.cmake 2025-11-03T17:06:22.7273818Z creating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/3.31.6/CompilerIdC/ 2025-11-03T17:06:22.7274459Z creating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/3.31.6/CompilerIdC/tmp/ 2025-11-03T17:06:22.7277816Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/3.31.6/CompilerIdC/CMakeCCompilerId.c 2025-11-03T17:06:22.7279453Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/3.31.6/CompilerIdC/a.out 2025-11-03T17:06:22.7280639Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/3.31.6/CMakeCCompiler.cmake 2025-11-03T17:06:22.7281313Z creating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/3.31.6/CompilerIdCXX/ 2025-11-03T17:06:22.7282162Z creating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/3.31.6/CompilerIdCXX/tmp/ 2025-11-03T17:06:22.7284697Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/3.31.6/CompilerIdCXX/CMakeCXXCompilerId.cpp 2025-11-03T17:06:22.7286448Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/3.31.6/CompilerIdCXX/a.out 2025-11-03T17:06:22.7287682Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/3.31.6/CMakeCXXCompiler.cmake 2025-11-03T17:06:22.7289718Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/3.31.6/CMakeDetermineCompilerABI_C.bin 2025-11-03T17:06:22.7291805Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/3.31.6/CMakeDetermineCompilerABI_CXX.bin 2025-11-03T17:06:22.7292429Z creating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/3.31.6/CompilerIdCUDA/ 2025-11-03T17:06:22.7293095Z creating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/3.31.6/CompilerIdCUDA/tmp/ 2025-11-03T17:06:22.7345603Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/3.31.6/CompilerIdCUDA/tmp/CMakeCUDACompilerId.cpp4.ii 2025-11-03T17:06:22.7398356Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/3.31.6/CompilerIdCUDA/tmp/CMakeCUDACompilerId.cudafe1.cpp 2025-11-03T17:06:22.7399369Z extracting: build/custom_test_artifacts/custom-backend-build/CMakeFiles/3.31.6/CompilerIdCUDA/tmp/CMakeCUDACompilerId.module_id 2025-11-03T17:06:22.7455720Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/3.31.6/CompilerIdCUDA/tmp/CMakeCUDACompilerId.cpp1.ii 2025-11-03T17:06:22.7456989Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/3.31.6/CompilerIdCUDA/tmp/CMakeCUDACompilerId.cudafe1.c 2025-11-03T17:06:22.7458333Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/3.31.6/CompilerIdCUDA/tmp/CMakeCUDACompilerId.cudafe1.gpu 2025-11-03T17:06:22.7459493Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/3.31.6/CompilerIdCUDA/tmp/CMakeCUDACompilerId.cudafe1.stub.c 2025-11-03T17:06:22.7460590Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/3.31.6/CompilerIdCUDA/tmp/CMakeCUDACompilerId.ptx 2025-11-03T17:06:22.7461757Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/3.31.6/CompilerIdCUDA/tmp/CMakeCUDACompilerId.sm_52.cubin 2025-11-03T17:06:22.7462895Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/3.31.6/CompilerIdCUDA/tmp/CMakeCUDACompilerId.fatbin 2025-11-03T17:06:22.7464167Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/3.31.6/CompilerIdCUDA/tmp/CMakeCUDACompilerId.fatbin.c 2025-11-03T17:06:22.7466052Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/3.31.6/CompilerIdCUDA/tmp/CMakeCUDACompilerId.o 2025-11-03T17:06:22.7466982Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/3.31.6/CompilerIdCUDA/tmp/a_dlink.sm_52.cubin 2025-11-03T17:06:22.7468040Z extracting: build/custom_test_artifacts/custom-backend-build/CMakeFiles/3.31.6/CompilerIdCUDA/tmp/a_dlink.reg.c 2025-11-03T17:06:22.7469188Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/3.31.6/CompilerIdCUDA/tmp/a_dlink.fatbin 2025-11-03T17:06:22.7470461Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/3.31.6/CompilerIdCUDA/tmp/a_dlink.fatbin.c 2025-11-03T17:06:22.7471745Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/3.31.6/CompilerIdCUDA/tmp/a_dlink.o 2025-11-03T17:06:22.7474601Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/3.31.6/CompilerIdCUDA/CMakeCUDACompilerId.cu 2025-11-03T17:06:22.7539369Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/3.31.6/CompilerIdCUDA/a.out 2025-11-03T17:06:22.7540444Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/3.31.6/CMakeCUDACompiler.cmake 2025-11-03T17:06:22.7605457Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/3.31.6/CMakeDetermineCompilerABI_CUDA.bin 2025-11-03T17:06:22.7606236Z creating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/CMakeScratch/ 2025-11-03T17:06:22.7606807Z creating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/CMakeTmp/ 2025-11-03T17:06:22.7607410Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/cmake.check_cache 2025-11-03T17:06:22.7608037Z creating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/custom_backend.dir/ 2025-11-03T17:06:22.7609396Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/custom_backend.dir/compiler_depend.ts 2025-11-03T17:06:22.7610430Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/custom_backend.dir/compiler_depend.make 2025-11-03T17:06:22.7611352Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/custom_backend.dir/depend.make 2025-11-03T17:06:22.7612333Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/custom_backend.dir/link.txt 2025-11-03T17:06:22.7613338Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/custom_backend.dir/cmake_clean.cmake 2025-11-03T17:06:22.7614668Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/custom_backend.dir/build.make 2025-11-03T17:06:22.7615732Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/custom_backend.dir/DependInfo.cmake 2025-11-03T17:06:22.7616795Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/custom_backend.dir/flags.make 2025-11-03T17:06:22.7617995Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/custom_backend.dir/progress.make 2025-11-03T17:06:22.7622810Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/custom_backend.dir/custom_backend.cpp.o.d 2025-11-03T17:06:22.7726945Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/custom_backend.dir/custom_backend.cpp.o 2025-11-03T17:06:22.7727729Z creating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/test_custom_backend.dir/ 2025-11-03T17:06:22.7728528Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/test_custom_backend.dir/compiler_depend.ts 2025-11-03T17:06:22.7729463Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/test_custom_backend.dir/compiler_depend.make 2025-11-03T17:06:22.7730373Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/test_custom_backend.dir/depend.make 2025-11-03T17:06:22.7731482Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/test_custom_backend.dir/link.txt 2025-11-03T17:06:22.7732486Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/test_custom_backend.dir/cmake_clean.cmake 2025-11-03T17:06:22.7734463Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/test_custom_backend.dir/build.make 2025-11-03T17:06:22.7735370Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/test_custom_backend.dir/DependInfo.cmake 2025-11-03T17:06:22.7736353Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/test_custom_backend.dir/flags.make 2025-11-03T17:06:22.7737524Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/test_custom_backend.dir/progress.make 2025-11-03T17:06:22.7756684Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/test_custom_backend.dir/test_custom_backend.cpp.o.d 2025-11-03T17:06:22.7804294Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/test_custom_backend.dir/test_custom_backend.cpp.o 2025-11-03T17:06:22.7805327Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/CMakeDirectoryInformation.cmake 2025-11-03T17:06:22.7806275Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/TargetDirectories.txt 2025-11-03T17:06:22.7807419Z extracting: build/custom_test_artifacts/custom-backend-build/CMakeFiles/progress.marks 2025-11-03T17:06:22.7808623Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/Makefile2 2025-11-03T17:06:22.7810711Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/Makefile.cmake 2025-11-03T17:06:22.7811409Z inflating: build/custom_test_artifacts/custom-backend-build/detect_cuda_version.cc 2025-11-03T17:06:22.7814330Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeCache.txt 2025-11-03T17:06:22.7815747Z inflating: build/custom_test_artifacts/custom-backend-build/Makefile 2025-11-03T17:06:22.7816741Z inflating: build/custom_test_artifacts/custom-backend-build/cmake_install.cmake 2025-11-03T17:06:22.7904497Z inflating: build/custom_test_artifacts/custom-backend-build/libcustom_backend.so 2025-11-03T17:06:22.7938935Z inflating: build/custom_test_artifacts/custom-backend-build/test_custom_backend 2025-11-03T17:06:22.7939359Z creating: build/lib/ 2025-11-03T17:06:22.8011457Z inflating: build/lib/libprotobuf-lite.a 2025-11-03T17:06:22.8020325Z inflating: build/lib/libpthreadpool.a 2025-11-03T17:06:22.8405418Z inflating: build/lib/libprotobuf.a 2025-11-03T17:06:22.8412962Z inflating: build/lib/libcpuinfo.a 2025-11-03T17:06:22.8844325Z inflating: build/lib/libprotoc.a 2025-11-03T17:06:22.8851639Z inflating: build/lib/libcpuinfo_internals.a 2025-11-03T17:06:22.8852889Z inflating: build/lib/libclog.a 2025-11-03T17:06:22.8855644Z inflating: build/lib/libnnpack_reference_layers.a 2025-11-03T17:06:22.8872785Z inflating: build/lib/libpytorch_qnnpack.a 2025-11-03T17:06:22.8888620Z inflating: build/lib/libnnpack.a 2025-11-03T17:06:22.9050586Z inflating: build/lib/libmicrokernels-prod.a 2025-11-03T17:06:22.9808798Z inflating: build/lib/libmicrokernels-all.a 2025-11-03T17:06:22.9870113Z inflating: build/lib/libgtest.a 2025-11-03T17:06:22.9885420Z inflating: build/lib/libgmock.a 2025-11-03T17:06:22.9886496Z inflating: build/lib/libgtest_main.a 2025-11-03T17:06:22.9952777Z inflating: build/lib/libbenchmark.a 2025-11-03T17:06:22.9953746Z inflating: build/lib/libgmock_main.a 2025-11-03T17:06:22.9955227Z inflating: build/lib/libbenchmark_main.a 2025-11-03T17:06:23.0034171Z inflating: build/lib/libXNNPACK.a 2025-11-03T17:06:23.0035258Z inflating: build/lib/libjitprofiling.a 2025-11-03T17:06:23.0042901Z inflating: build/lib/libittnotify.a 2025-11-03T17:06:23.0100218Z inflating: build/lib/libasmjit.a 2025-11-03T17:06:23.1126474Z inflating: build/lib/libfbgemm.a 2025-11-03T17:06:23.1153615Z inflating: build/lib/libtensorpipe_uv.a 2025-11-03T17:06:23.1633001Z inflating: build/lib/libtensorpipe.a 2025-11-03T17:06:23.1848360Z inflating: build/lib/libtensorpipe_cuda.a 2025-11-03T17:06:23.1965658Z inflating: build/lib/libgloo.a 2025-11-03T17:06:23.2007184Z inflating: build/lib/libonnx_proto.a 2025-11-03T17:06:23.2630528Z inflating: build/lib/libonnx.a 2025-11-03T17:06:23.3015455Z inflating: build/lib/libgloo_cuda.a 2025-11-03T17:06:23.3032787Z inflating: build/lib/libfmt.a 2025-11-03T17:06:24.1892625Z inflating: build/lib/libdnnl.a 2025-11-03T17:06:24.2307122Z inflating: build/lib/libkineto.a 2025-11-03T17:06:24.2308888Z inflating: build/lib/libtorch_global_deps.so 2025-11-03T17:06:24.2414556Z inflating: build/lib/libc10.so 2025-11-03T17:06:24.2416721Z inflating: build/lib/libcaffe2_nvrtc.so 2025-11-03T17:06:24.2459928Z inflating: build/lib/libc10_cuda.so 2025-11-03T17:06:26.9260673Z inflating: build/lib/libtorch_cpu.so 2025-11-03T17:06:26.9946607Z inflating: build/lib/libtorch_nvshmem.so 2025-11-03T17:06:31.0539118Z inflating: build/lib/libtorch_cuda.so 2025-11-03T17:06:31.0543183Z inflating: build/lib/libshm.so 2025-11-03T17:06:31.0545878Z inflating: build/lib/libc10d_cuda_test.so 2025-11-03T17:06:31.0547280Z inflating: build/lib/libtorch.so 2025-11-03T17:06:31.0592595Z inflating: build/lib/libtorch_cuda_linalg.so 2025-11-03T17:06:31.0653953Z inflating: build/lib/libtorchbind_test.so 2025-11-03T17:06:31.0670050Z inflating: build/lib/libjitbackend_test.so 2025-11-03T17:06:31.0691066Z inflating: build/lib/libbackend_with_compiler.so 2025-11-03T17:06:31.0714050Z inflating: build/lib/libaoti_custom_ops.so 2025-11-03T17:06:31.2784533Z inflating: build/lib/libtorch_python.so 2025-11-03T17:06:31.2815338Z inflating: build/lib/libnnapi_backend.so 2025-11-03T17:06:31.2815651Z creating: build/bin/ 2025-11-03T17:06:31.3210443Z inflating: build/bin/protoc-3.13.0.0 2025-11-03T17:06:31.3606240Z inflating: build/bin/protoc 2025-11-03T17:06:31.3656434Z inflating: build/bin/c10_Device_test 2025-11-03T17:06:31.3703703Z inflating: build/bin/c10_StreamGuard_test 2025-11-03T17:06:31.3758521Z inflating: build/bin/c10_SymInt_test 2025-11-03T17:06:31.3810532Z inflating: build/bin/c10_InlineDeviceGuard_test 2025-11-03T17:06:31.3863946Z inflating: build/bin/c10_Metaprogramming_test 2025-11-03T17:06:31.3917533Z inflating: build/bin/c10_InlineStreamGuard_test 2025-11-03T17:06:31.3970955Z inflating: build/bin/c10_ThreadLocal_test 2025-11-03T17:06:31.4018923Z inflating: build/bin/c10_Semaphore_test 2025-11-03T17:06:31.4069919Z inflating: build/bin/c10_Bitset_test 2025-11-03T17:06:31.4121958Z inflating: build/bin/c10_Scalar_test 2025-11-03T17:06:31.4171059Z inflating: build/bin/c10_TypeList_test 2025-11-03T17:06:31.4228029Z inflating: build/bin/c10_DispatchKeySet_test 2025-11-03T17:06:31.4279503Z inflating: build/bin/c10_AllocatorConfig_test 2025-11-03T17:06:31.4327504Z inflating: build/bin/c10_DeadlockDetection_test 2025-11-03T17:06:31.4375700Z inflating: build/bin/c10_Synchronized_test 2025-11-03T17:06:31.4425616Z inflating: build/bin/c10_TypeIndex_test 2025-11-03T17:06:31.4479183Z inflating: build/bin/c10_LeftRight_test 2025-11-03T17:06:31.4533030Z inflating: build/bin/c10_SizesAndStrides_test 2025-11-03T17:06:31.4580310Z inflating: build/bin/c10_ConstexprCrc_test 2025-11-03T17:06:31.4631407Z inflating: build/bin/c10_IntrusiveList_test 2025-11-03T17:06:31.4681033Z inflating: build/bin/c10_accumulate_test 2025-11-03T17:06:31.4735891Z inflating: build/bin/c10_Enumerate_test 2025-11-03T17:06:31.4783927Z inflating: build/bin/c10_CompileTimeFunctionPointer_test 2025-11-03T17:06:31.4832479Z inflating: build/bin/c10_Half_test 2025-11-03T17:06:31.4883607Z inflating: build/bin/c10_NetworkFlow_test 2025-11-03T17:06:31.4937875Z inflating: build/bin/c10_complex_math_test 2025-11-03T17:06:31.4986420Z inflating: build/bin/c10_flags_test 2025-11-03T17:06:31.5033956Z inflating: build/bin/c10_TypeTraits_test 2025-11-03T17:06:31.5084378Z inflating: build/bin/c10_exception_test 2025-11-03T17:06:31.5135640Z inflating: build/bin/c10_registry_test 2025-11-03T17:06:31.5194461Z inflating: build/bin/c10_ordered_preserving_dict_test 2025-11-03T17:06:31.5248896Z inflating: build/bin/c10_logging_test 2025-11-03T17:06:31.5297451Z inflating: build/bin/c10_generic_math_test 2025-11-03T17:06:31.5346574Z inflating: build/bin/c10_tempfile_test 2025-11-03T17:06:31.5400700Z inflating: build/bin/c10_string_util_test 2025-11-03T17:06:31.5448100Z inflating: build/bin/c10_string_view_test 2025-11-03T17:06:31.5497737Z inflating: build/bin/c10_DeviceGuard_test 2025-11-03T17:06:31.5545816Z inflating: build/bin/c10_ArrayRef_test 2025-11-03T17:06:31.5594535Z inflating: build/bin/c10_bit_cast_test 2025-11-03T17:06:31.5661118Z inflating: build/bin/c10_cow_test 2025-11-03T17:06:31.5802541Z inflating: build/bin/c10_small_vector_test 2025-11-03T17:06:31.5851647Z inflating: build/bin/c10_irange_test 2025-11-03T17:06:31.5901094Z inflating: build/bin/c10_ssize_test 2025-11-03T17:06:31.5952581Z inflating: build/bin/c10_lazy_test 2025-11-03T17:06:31.6000475Z inflating: build/bin/c10_error_test 2025-11-03T17:06:31.6054257Z inflating: build/bin/c10_bfloat16_test 2025-11-03T17:06:31.6106957Z inflating: build/bin/c10_complex_test 2025-11-03T17:06:31.6258524Z inflating: build/bin/c10_intrusive_ptr_test 2025-11-03T17:06:31.6300699Z inflating: build/bin/c10_intrusive_ptr_benchmark 2025-11-03T17:06:31.6371987Z inflating: build/bin/c10_optional_test 2025-11-03T17:06:31.6426615Z inflating: build/bin/c10_typeid_test 2025-11-03T17:06:31.6474353Z inflating: build/bin/c10_cuda_CUDATest 2025-11-03T17:06:31.6522743Z inflating: build/bin/test_vec_half_AVX2 2025-11-03T17:06:31.6596265Z inflating: build/bin/test_aoti_abi_check 2025-11-03T17:06:31.6645145Z inflating: build/bin/test_vec_half_AVX512 2025-11-03T17:06:31.6692708Z inflating: build/bin/test_vec_half_DEFAULT 2025-11-03T17:06:31.7219613Z inflating: build/bin/vec_test_all_types_DEFAULT 2025-11-03T17:06:31.7762880Z inflating: build/bin/vec_test_all_types_AVX512 2025-11-03T17:06:31.8310516Z inflating: build/bin/vec_test_all_types_AVX2 2025-11-03T17:06:31.8361617Z inflating: build/bin/c10_cuda_CUDAAssertionsTest_catches_thread_and_block_and_device 2025-11-03T17:06:31.8412142Z inflating: build/bin/c10_cuda_CUDAAssertionsTest_multiple_writes_from_same_block 2025-11-03T17:06:31.8463284Z inflating: build/bin/c10_cuda_CUDAAssertionsTest_multiple_writes_from_multiple_blocks 2025-11-03T17:06:31.8513393Z inflating: build/bin/c10_cuda_CUDAAssertionsTest_from_2_processes 2025-11-03T17:06:31.8564358Z inflating: build/bin/c10_cuda_CUDAAssertionsTest_1_var_test 2025-11-03T17:06:31.8615010Z inflating: build/bin/c10_cuda_CUDAAssertionsTest_multiple_writes_from_blocks_and_threads 2025-11-03T17:06:31.8666144Z inflating: build/bin/c10_cuda_CUDAAssertionsTest_catches_stream 2025-11-03T17:06:31.8716810Z inflating: build/bin/BackoffTest 2025-11-03T17:06:31.8768977Z inflating: build/bin/FileStoreTest 2025-11-03T17:06:31.8824003Z inflating: build/bin/TCPStoreTest 2025-11-03T17:06:31.8875553Z inflating: build/bin/HashStoreTest 2025-11-03T17:06:31.8887912Z inflating: build/bin/ProcessGroupMPITest 2025-11-03T17:06:31.8891385Z inflating: build/bin/torch_shm_manager 2025-11-03T17:06:31.8894347Z inflating: build/bin/example_allreduce 2025-11-03T17:06:31.8948974Z inflating: build/bin/ProcessGroupGlooAsyncTest 2025-11-03T17:06:31.9009528Z inflating: build/bin/ProcessGroupNCCLTest 2025-11-03T17:06:31.9068362Z inflating: build/bin/ProcessGroupNCCLErrorsTest 2025-11-03T17:06:31.9131489Z inflating: build/bin/ProcessGroupGlooTest 2025-11-03T17:06:31.9192509Z inflating: build/bin/basic 2025-11-03T17:06:31.9245641Z inflating: build/bin/broadcast_test 2025-11-03T17:06:31.9299858Z inflating: build/bin/NamedTensor_test 2025-11-03T17:06:31.9350594Z inflating: build/bin/Dimname_test 2025-11-03T17:06:31.9401676Z inflating: build/bin/cpu_profiling_allocator_test 2025-11-03T17:06:31.9455321Z inflating: build/bin/half_test 2025-11-03T17:06:31.9512537Z inflating: build/bin/apply_utils_test 2025-11-03T17:06:31.9564406Z inflating: build/bin/math_kernel_test 2025-11-03T17:06:31.9654034Z inflating: build/bin/ivalue_test 2025-11-03T17:06:31.9708182Z inflating: build/bin/native_test 2025-11-03T17:06:31.9770404Z inflating: build/bin/MaybeOwned_test 2025-11-03T17:06:31.9834433Z inflating: build/bin/pow_test 2025-11-03T17:06:31.9903987Z inflating: build/bin/Dict_test 2025-11-03T17:06:31.9955968Z inflating: build/bin/memory_format_test 2025-11-03T17:06:32.0010161Z inflating: build/bin/scalar_tensor_test 2025-11-03T17:06:32.0059845Z inflating: build/bin/operators_test 2025-11-03T17:06:32.0116118Z inflating: build/bin/atest 2025-11-03T17:06:32.0164837Z inflating: build/bin/thread_init_test 2025-11-03T17:06:32.0250816Z inflating: build/bin/cpu_rng_test 2025-11-03T17:06:32.0299852Z inflating: build/bin/cpu_allocator_test 2025-11-03T17:06:32.0356008Z inflating: build/bin/cpu_generator_test 2025-11-03T17:06:32.0404037Z inflating: build/bin/verify_api_visibility 2025-11-03T17:06:32.0452079Z inflating: build/bin/reduce_ops_test 2025-11-03T17:06:32.0504044Z inflating: build/bin/memory_overlapping_test 2025-11-03T17:06:32.0578278Z inflating: build/bin/tensor_iterator_test 2025-11-03T17:06:32.0628610Z inflating: build/bin/stride_properties_test 2025-11-03T17:06:32.0677923Z inflating: build/bin/StorageUtils_test 2025-11-03T17:06:32.0732534Z inflating: build/bin/extension_backend_test 2025-11-03T17:06:32.0843191Z inflating: build/bin/kernel_function_legacy_test 2025-11-03T17:06:32.0891152Z inflating: build/bin/lazy_tensor_test 2025-11-03T17:06:32.0942675Z inflating: build/bin/mobile_memory_cleanup 2025-11-03T17:06:32.0997487Z inflating: build/bin/quantized_test 2025-11-03T17:06:32.1085779Z inflating: build/bin/make_boxed_from_unboxed_functor_test 2025-11-03T17:06:32.1136065Z inflating: build/bin/packedtensoraccessor_test 2025-11-03T17:06:32.1184775Z inflating: build/bin/operator_name_test 2025-11-03T17:06:32.1232350Z inflating: build/bin/op_allowlist_test 2025-11-03T17:06:32.1288449Z inflating: build/bin/scalar_test 2025-11-03T17:06:32.1356827Z inflating: build/bin/cuda_complex_math_test 2025-11-03T17:06:32.1412900Z inflating: build/bin/type_test 2025-11-03T17:06:32.1469603Z inflating: build/bin/IListRef_test 2025-11-03T17:06:32.1518762Z inflating: build/bin/reportMemoryUsage_test 2025-11-03T17:06:32.1568364Z inflating: build/bin/weakref_test 2025-11-03T17:06:32.1619217Z inflating: build/bin/cuda_apply_test 2025-11-03T17:06:32.1669763Z inflating: build/bin/undefined_tensor_test 2025-11-03T17:06:32.1732142Z inflating: build/bin/cuda_cub_test 2025-11-03T17:06:32.1781213Z inflating: build/bin/CppSignature_test 2025-11-03T17:06:32.1834208Z inflating: build/bin/type_ptr_test 2025-11-03T17:06:32.1922726Z inflating: build/bin/kernel_function_test 2025-11-03T17:06:32.1972465Z inflating: build/bin/cuda_dlconvertor_test 2025-11-03T17:06:32.2025029Z inflating: build/bin/test_parallel 2025-11-03T17:06:32.2072841Z inflating: build/bin/cuda_device_test 2025-11-03T17:06:32.2167260Z inflating: build/bin/kernel_lambda_test 2025-11-03T17:06:32.2216253Z inflating: build/bin/dlconvertor_test 2025-11-03T17:06:32.2279773Z inflating: build/bin/inline_container_test 2025-11-03T17:06:32.2338441Z inflating: build/bin/cuda_atomic_ops_test 2025-11-03T17:06:32.2389424Z inflating: build/bin/cuda_vectorized_test 2025-11-03T17:06:32.2441286Z inflating: build/bin/cuda_allocator_test 2025-11-03T17:06:32.2491059Z inflating: build/bin/xla_tensor_test 2025-11-03T17:06:32.2548478Z inflating: build/bin/cuda_complex_test 2025-11-03T17:06:32.2611465Z inflating: build/bin/KernelFunction_test 2025-11-03T17:06:32.2661037Z inflating: build/bin/wrapdim_test 2025-11-03T17:06:32.2716126Z inflating: build/bin/cuda_generator_test 2025-11-03T17:06:32.2766769Z inflating: build/bin/cuda_reportMemoryUsage_test 2025-11-03T17:06:32.2824477Z inflating: build/bin/kernel_stackbased_test 2025-11-03T17:06:32.2882673Z inflating: build/bin/cuda_stream_test 2025-11-03T17:06:32.2931179Z inflating: build/bin/cuda_half_test 2025-11-03T17:06:32.2978960Z inflating: build/bin/cuda_exchange_device_test 2025-11-03T17:06:32.3032212Z inflating: build/bin/backend_fallback_test 2025-11-03T17:06:32.3080229Z inflating: build/bin/cuda_cudnn_test 2025-11-03T17:06:32.3130189Z inflating: build/bin/cuda_integer_divider_test 2025-11-03T17:06:32.3209909Z inflating: build/bin/cuda_distributions_test 2025-11-03T17:06:32.3325850Z inflating: build/bin/kernel_lambda_legacy_test 2025-11-03T17:06:32.3373889Z inflating: build/bin/cuda_optional_test 2025-11-03T17:06:32.3427906Z inflating: build/bin/cuda_caching_host_allocator_test 2025-11-03T17:06:32.3707166Z inflating: build/bin/op_registration_test 2025-11-03T17:06:32.3805927Z inflating: build/bin/List_test 2025-11-03T17:06:32.3872069Z inflating: build/bin/legacy_vmap_test 2025-11-03T17:06:32.3920460Z inflating: build/bin/cuda_allocatorTraceTracker_test 2025-11-03T17:06:32.3984843Z inflating: build/bin/test_cpp_rpc 2025-11-03T17:06:32.4037908Z inflating: build/bin/test_dist_autograd 2025-11-03T17:06:32.4040357Z inflating: build/bin/parallel_benchmark 2025-11-03T17:06:32.5073047Z inflating: build/bin/test_api 2025-11-03T17:06:32.5124093Z inflating: build/bin/cuda_packedtensoraccessor_test 2025-11-03T17:06:32.5435931Z inflating: build/bin/test_lazy 2025-11-03T17:06:32.6415469Z inflating: build/bin/test_jit 2025-11-03T17:06:32.6415975Z creating: .additional_ci_files/ 2025-11-03T17:06:32.6465588Z inflating: .additional_ci_files/test-times.json 2025-11-03T17:06:32.6650154Z inflating: .additional_ci_files/test-class-times.json 2025-11-03T17:06:32.6697308Z ##[group]Run rm artifacts.zip 2025-11-03T17:06:32.6697549Z rm artifacts.zip 2025-11-03T17:06:32.6706895Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2025-11-03T17:06:32.6707174Z env: 2025-11-03T17:06:32.6707340Z GIT_DEFAULT_BRANCH: main 2025-11-03T17:06:32.6707593Z GPU_FLAG: --gpus all -e NVIDIA_DRIVER_CAPABILITIES=all 2025-11-03T17:06:32.6707848Z ##[endgroup] 2025-11-03T17:06:32.8664529Z ##[group]Run df -H 2025-11-03T17:06:32.8664713Z df -H 2025-11-03T17:06:32.8672109Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2025-11-03T17:06:32.8672389Z env: 2025-11-03T17:06:32.8672548Z GIT_DEFAULT_BRANCH: main 2025-11-03T17:06:32.8672792Z GPU_FLAG: --gpus all -e NVIDIA_DRIVER_CAPABILITIES=all 2025-11-03T17:06:32.8673230Z ##[endgroup] 2025-11-03T17:06:32.8723719Z Filesystem Size Used Avail Use% Mounted on 2025-11-03T17:06:32.8724088Z devtmpfs 4.2M 0 4.2M 0% /dev 2025-11-03T17:06:32.8724409Z tmpfs 33G 0 33G 0% /dev/shm 2025-11-03T17:06:32.8724710Z tmpfs 13G 779k 13G 1% /run 2025-11-03T17:06:32.8724979Z /dev/nvme0n1p1 161G 56G 106G 35% / 2025-11-03T17:06:32.8725289Z tmpfs 33G 13k 33G 1% /tmp 2025-11-03T17:06:32.8725617Z /dev/nvme0n1p128 11M 1.4M 9.2M 13% /boot/efi 2025-11-03T17:06:32.8726127Z tmpfs 6.5G 0 6.5G 0% /run/user/0 2025-11-03T17:06:32.8756999Z Prepare all required actions 2025-11-03T17:06:32.8757469Z Getting action download info 2025-11-03T17:06:33.0246912Z ##[group]Run ./.github/actions/download-td-artifacts 2025-11-03T17:06:33.0247164Z with: 2025-11-03T17:06:33.0247315Z env: 2025-11-03T17:06:33.0247472Z GIT_DEFAULT_BRANCH: main 2025-11-03T17:06:33.0247726Z GPU_FLAG: --gpus all -e NVIDIA_DRIVER_CAPABILITIES=all 2025-11-03T17:06:33.0247996Z ##[endgroup] 2025-11-03T17:06:33.0283848Z ##[group]Run seemethere/download-artifact-s3@v4 2025-11-03T17:06:33.0284092Z with: 2025-11-03T17:06:33.0284243Z name: td_results 2025-11-03T17:06:33.0284420Z s3-bucket: gha-artifacts 2025-11-03T17:06:33.0284613Z region: us-east-1 2025-11-03T17:06:33.0284771Z env: 2025-11-03T17:06:33.0284915Z GIT_DEFAULT_BRANCH: main 2025-11-03T17:06:33.0285153Z GPU_FLAG: --gpus all -e NVIDIA_DRIVER_CAPABILITIES=all 2025-11-03T17:06:33.0285460Z ##[endgroup] 2025-11-03T17:06:33.5301555Z (node:61234) NOTE: We are formalizing our plans to enter AWS SDK for JavaScript (v2) into maintenance mode in 2023. 2025-11-03T17:06:33.5302310Z 2025-11-03T17:06:33.5302606Z Please migrate your code to use AWS SDK for JavaScript (v3). 2025-11-03T17:06:33.5303145Z For more information, check the migration guide at https://a.co/7PzMCcy 2025-11-03T17:06:33.5303672Z (Use `node --trace-warnings ...` to show where the warning was created) 2025-11-03T17:06:33.6178764Z Found 0 objects with prefix pytorch/pytorch/19040285420/td_results/ 2025-11-03T17:06:33.6184520Z Artifact download has finished successfully 2025-11-03T17:06:33.6434399Z ##[group]Run mkdir -p .additional_ci_files 2025-11-03T17:06:33.6434685Z mkdir -p .additional_ci_files 2025-11-03T17:06:33.6435023Z mv td_results.json .additional_ci_files/td_results.json || true 2025-11-03T17:06:33.6442935Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2025-11-03T17:06:33.6443202Z env: 2025-11-03T17:06:33.6443364Z GIT_DEFAULT_BRANCH: main 2025-11-03T17:06:33.6443622Z GPU_FLAG: --gpus all -e NVIDIA_DRIVER_CAPABILITIES=all 2025-11-03T17:06:33.6443876Z ##[endgroup] 2025-11-03T17:06:33.6505556Z mv: cannot stat 'td_results.json': No such file or directory 2025-11-03T17:06:33.6555772Z ##[group]Run .github/scripts/parse_ref.py 2025-11-03T17:06:33.6556091Z .github/scripts/parse_ref.py 2025-11-03T17:06:33.6563175Z shell: /usr/bin/bash -e {0} 2025-11-03T17:06:33.6563395Z env: 2025-11-03T17:06:33.6563549Z GIT_DEFAULT_BRANCH: main 2025-11-03T17:06:33.6563806Z GPU_FLAG: --gpus all -e NVIDIA_DRIVER_CAPABILITIES=all 2025-11-03T17:06:33.6564062Z ##[endgroup] 2025-11-03T17:06:33.6786238Z Setting output branch=main 2025-11-03T17:06:33.6883795Z Prepare all required actions 2025-11-03T17:06:33.6884137Z Getting action download info 2025-11-03T17:06:33.8452257Z ##[group]Run ./.github/actions/filter-test-configs 2025-11-03T17:06:33.8452590Z with: 2025-11-03T17:06:33.8452987Z github-token: *** 2025-11-03T17:06:33.8457698Z test-matrix: {"include": [{"config": "vllm_basic_correctness_test", "shard": 1, "num_shards": 1, "runner": "linux.g6.4xlarge.experimental.nvidia.gpu"}, {"config": "vllm_basic_models_test", "shard": 1, "num_shards": 1, "runner": "linux.g6.4xlarge.experimental.nvidia.gpu"}, {"config": "vllm_entrypoints_test", "shard": 1, "num_shards": 1, "runner": "linux.g6.4xlarge.experimental.nvidia.gpu"}, {"config": "vllm_regression_test", "shard": 1, "num_shards": 1, "runner": "linux.g6.4xlarge.experimental.nvidia.gpu"}, {"config": "vllm_multi_model_processor_test", "shard": 1, "num_shards": 1, "runner": "linux.g6.4xlarge.experimental.nvidia.gpu"}, {"config": "vllm_pytorch_compilation_unit_tests", "shard": 1, "num_shards": 1, "runner": "linux.g6.4xlarge.experimental.nvidia.gpu"}, {"config": "vllm_lora_28_failure_test", "shard": 1, "num_shards": 1, "runner": "linux.g6.4xlarge.experimental.nvidia.gpu"}, {"config": "vllm_multi_model_test_28_failure_test", "shard": 1, "num_shards": 1, "runner": "linux.g6.4xlarge.experimental.nvidia.gpu"}, {"config": "vllm_language_model_test_extended_generation_28_failure_test", "shard": 1, "num_shards": 1, "runner": "linux.g6.4xlarge.experimental.nvidia.gpu"}, {"config": "vllm_distributed_test_2_gpu_28_failure_test", "shard": 1, "num_shards": 1, "runner": "linux.g6.4xlarge.experimental.nvidia.gpu"}, {"config": "vllm_lora_test", "shard": 0, "num_shards": 4, "runner": "linux.g6.4xlarge.experimental.nvidia.gpu"}, {"config": "vllm_lora_test", "shard": 1, "num_shards": 4, "runner": "linux.g6.4xlarge.experimental.nvidia.gpu"}, {"config": "vllm_lora_test", "shard": 2, "num_shards": 4, "runner": "linux.g6.4xlarge.experimental.nvidia.gpu"}, {"config": "vllm_lora_test", "shard": 3, "num_shards": 4, "runner": "linux.g6.4xlarge.experimental.nvidia.gpu"}, {"config": "vllm_lora_tp_test_distributed", "shard": 1, "num_shards": 1, "runner": "linux.g6.12xlarge.nvidia.gpu"}, {"config": "vllm_distributed_test_28_failure_test", "shard": 1, "num_shards": 1, "runner": "linux.g6.12xlarge.nvidia.gpu"}]} 2025-11-03T17:06:33.8463068Z job-name: ci-vllm-test / test (vllm_basic_models_test, 1, 1, linux.g6.4xlarge.experimental.nvidia.gpu) 2025-11-03T17:06:33.8463583Z env: 2025-11-03T17:06:33.8463746Z GIT_DEFAULT_BRANCH: main 2025-11-03T17:06:33.8463991Z GPU_FLAG: --gpus all -e NVIDIA_DRIVER_CAPABILITIES=all 2025-11-03T17:06:33.8464246Z ##[endgroup] 2025-11-03T17:06:33.8530147Z ##[group]Run nick-fields/retry@v3.0.0 2025-11-03T17:06:33.8530366Z with: 2025-11-03T17:06:33.8530521Z shell: bash 2025-11-03T17:06:33.8530681Z timeout_minutes: 10 2025-11-03T17:06:33.8530855Z max_attempts: 5 2025-11-03T17:06:33.8531025Z retry_wait_seconds: 30 2025-11-03T17:06:33.8531597Z command: set -eux # PyYAML 6.0 doesn't work with MacOS x86 anymore # This must run on Python-3.7 (AmazonLinux2) so can't use request=3.32.2 python3 -m pip install requests==2.27.1 pyyaml==6.0.2 2025-11-03T17:06:33.8532186Z polling_interval_seconds: 1 2025-11-03T17:06:33.8532385Z warning_on_retry: true 2025-11-03T17:06:33.8532569Z continue_on_error: false 2025-11-03T17:06:33.8532749Z env: 2025-11-03T17:06:33.8532894Z GIT_DEFAULT_BRANCH: main 2025-11-03T17:06:33.8533139Z GPU_FLAG: --gpus all -e NVIDIA_DRIVER_CAPABILITIES=all 2025-11-03T17:06:33.8533508Z GITHUB_TOKEN: *** 2025-11-03T17:06:33.8533679Z ##[endgroup] 2025-11-03T17:06:33.9453456Z + python3 -m pip install requests==2.27.1 pyyaml==6.0.2 2025-11-03T17:06:34.1633912Z Defaulting to user installation because normal site-packages is not writeable 2025-11-03T17:06:34.2644312Z Collecting requests==2.27.1 2025-11-03T17:06:34.2813401Z Downloading requests-2.27.1-py2.py3-none-any.whl (63 kB) 2025-11-03T17:06:34.4409026Z Collecting pyyaml==6.0.2 2025-11-03T17:06:34.4448322Z Downloading PyYAML-6.0.2-cp39-cp39-manylinux_2_17_x86_64.manylinux2014_x86_64.whl (737 kB) 2025-11-03T17:06:34.8192252Z Collecting charset-normalizer~=2.0.0 2025-11-03T17:06:34.8234269Z Downloading charset_normalizer-2.0.12-py3-none-any.whl (39 kB) 2025-11-03T17:06:34.8704285Z Collecting certifi>=2017.4.17 2025-11-03T17:06:34.8744148Z Downloading certifi-2025.10.5-py3-none-any.whl (163 kB) 2025-11-03T17:06:34.8806390Z Requirement already satisfied: urllib3<1.27,>=1.21.1 in /usr/lib/python3.9/site-packages (from requests==2.27.1) (1.25.10) 2025-11-03T17:06:34.8813973Z Requirement already satisfied: idna<4,>=2.5 in /usr/lib/python3.9/site-packages (from requests==2.27.1) (2.10) 2025-11-03T17:06:34.9539011Z Installing collected packages: charset-normalizer, certifi, requests, pyyaml 2025-11-03T17:06:35.0652267Z Successfully installed certifi-2025.10.5 charset-normalizer-2.0.12 pyyaml-6.0.2 requests-2.27.1 2025-11-03T17:06:35.9217472Z Command completed after 1 attempt(s). 2025-11-03T17:06:35.9283815Z ##[group]Run set -x 2025-11-03T17:06:35.9284046Z set -x 2025-11-03T17:06:35.9284214Z  2025-11-03T17:06:35.9284490Z # Use relative path here as this could be checked out anywhere, not necessarily 2025-11-03T17:06:35.9295139Z # in runner workspace 2025-11-03T17:06:35.9295703Z python3 "${GITHUB_ACTION_PATH}/../../scripts/parse_ref.py" 2025-11-03T17:06:35.9304049Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2025-11-03T17:06:35.9304327Z env: 2025-11-03T17:06:35.9304491Z GIT_DEFAULT_BRANCH: main 2025-11-03T17:06:35.9304738Z GPU_FLAG: --gpus all -e NVIDIA_DRIVER_CAPABILITIES=all 2025-11-03T17:06:35.9304997Z ##[endgroup] 2025-11-03T17:06:35.9334089Z + python3 /home/ec2-user/actions-runner/_work/pytorch/pytorch/./.github/actions/filter-test-configs/../../scripts/parse_ref.py 2025-11-03T17:06:35.9505178Z Setting output branch=main 2025-11-03T17:06:35.9561941Z ##[group]Run echo "Workflow: ${GITHUB_WORKFLOW}" 2025-11-03T17:06:35.9562273Z echo "Workflow: ${GITHUB_WORKFLOW}" 2025-11-03T17:06:35.9562519Z echo "Job name: ${JOB_NAME}" 2025-11-03T17:06:35.9562731Z  2025-11-03T17:06:35.9563012Z # Use relative path here as this could be checked out anywhere, not necessarily 2025-11-03T17:06:35.9563354Z # in runner workspace 2025-11-03T17:06:35.9563675Z python3 "${GITHUB_ACTION_PATH}/../../scripts/filter_test_configs.py" \ 2025-11-03T17:06:35.9564036Z  --workflow "${GITHUB_WORKFLOW}" \ 2025-11-03T17:06:35.9564275Z  --job-name "${JOB_NAME}" \ 2025-11-03T17:06:35.9569288Z  --test-matrix "{"include": [{"config": "vllm_basic_correctness_test", "shard": 1, "num_shards": 1, "runner": "linux.g6.4xlarge.experimental.nvidia.gpu"}, {"config": "vllm_basic_models_test", "shard": 1, "num_shards": 1, "runner": "linux.g6.4xlarge.experimental.nvidia.gpu"}, {"config": "vllm_entrypoints_test", "shard": 1, "num_shards": 1, "runner": "linux.g6.4xlarge.experimental.nvidia.gpu"}, {"config": "vllm_regression_test", "shard": 1, "num_shards": 1, "runner": "linux.g6.4xlarge.experimental.nvidia.gpu"}, {"config": "vllm_multi_model_processor_test", "shard": 1, "num_shards": 1, "runner": "linux.g6.4xlarge.experimental.nvidia.gpu"}, {"config": "vllm_pytorch_compilation_unit_tests", "shard": 1, "num_shards": 1, "runner": "linux.g6.4xlarge.experimental.nvidia.gpu"}, {"config": "vllm_lora_28_failure_test", "shard": 1, "num_shards": 1, "runner": "linux.g6.4xlarge.experimental.nvidia.gpu"}, {"config": "vllm_multi_model_test_28_failure_test", "shard": 1, "num_shards": 1, "runner": "linux.g6.4xlarge.experimental.nvidia.gpu"}, {"config": "vllm_language_model_test_extended_generation_28_failure_test", "shard": 1, "num_shards": 1, "runner": "linux.g6.4xlarge.experimental.nvidia.gpu"}, {"config": "vllm_distributed_test_2_gpu_28_failure_test", "shard": 1, "num_shards": 1, "runner": "linux.g6.4xlarge.experimental.nvidia.gpu"}, {"config": "vllm_lora_test", "shard": 0, "num_shards": 4, "runner": "linux.g6.4xlarge.experimental.nvidia.gpu"}, {"config": "vllm_lora_test", "shard": 1, "num_shards": 4, "runner": "linux.g6.4xlarge.experimental.nvidia.gpu"}, {"config": "vllm_lora_test", "shard": 2, "num_shards": 4, "runner": "linux.g6.4xlarge.experimental.nvidia.gpu"}, {"config": "vllm_lora_test", "shard": 3, "num_shards": 4, "runner": "linux.g6.4xlarge.experimental.nvidia.gpu"}, {"config": "vllm_lora_tp_test_distributed", "shard": 1, "num_shards": 1, "runner": "linux.g6.12xlarge.nvidia.gpu"}, {"config": "vllm_distributed_test_28_failure_test", "shard": 1, "num_shards": 1, "runner": "linux.g6.12xlarge.nvidia.gpu"}]}" \ 2025-11-03T17:06:35.9574703Z  --selected-test-configs "" \ 2025-11-03T17:06:35.9574987Z  --pr-number "${PR_NUMBER}" \ 2025-11-03T17:06:35.9575218Z  --tag "${TAG}" \ 2025-11-03T17:06:35.9575576Z  --event-name "${EVENT_NAME}" \ 2025-11-03T17:06:35.9575807Z  --schedule "${SCHEDULE}" \ 2025-11-03T17:06:35.9576032Z  --branch "${HEAD_BRANCH}" 2025-11-03T17:06:35.9583351Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2025-11-03T17:06:35.9583636Z env: 2025-11-03T17:06:35.9583792Z GIT_DEFAULT_BRANCH: main 2025-11-03T17:06:35.9584039Z GPU_FLAG: --gpus all -e NVIDIA_DRIVER_CAPABILITIES=all 2025-11-03T17:06:35.9584577Z GITHUB_TOKEN: *** 2025-11-03T17:06:35.9584945Z JOB_NAME: ci-vllm-test / test (vllm_basic_models_test, 1, 1, linux.g6.4xlarge.experimental.nvidia.gpu) 2025-11-03T17:06:35.9585345Z PR_NUMBER: 2025-11-03T17:06:35.9585500Z TAG: 2025-11-03T17:06:35.9585651Z EVENT_NAME: push 2025-11-03T17:06:35.9585819Z SCHEDULE: 2025-11-03T17:06:35.9585979Z HEAD_BRANCH: main 2025-11-03T17:06:35.9586153Z ##[endgroup] 2025-11-03T17:06:35.9612168Z Workflow: vllm-test 2025-11-03T17:06:35.9612558Z Job name: ci-vllm-test / test (vllm_basic_models_test, 1, 1, linux.g6.4xlarge.experimental.nvidia.gpu) 2025-11-03T17:06:36.1556744Z Setting output keep-going=True 2025-11-03T17:06:36.1557112Z Setting output ci-verbose-test-logs=False 2025-11-03T17:06:36.1557476Z Setting output ci-test-showlocals=False 2025-11-03T17:06:36.1557798Z Setting output ci-no-test-timeout=False 2025-11-03T17:06:36.1558101Z Setting output ci-no-td=False 2025-11-03T17:06:36.1558378Z Setting output ci-td-distributed=False 2025-11-03T17:06:36.1558663Z Setting output is-unstable=False 2025-11-03T17:06:36.1558938Z Setting output reenabled-issues= 2025-11-03T17:06:36.1565029Z Setting output test-matrix={"include": [{"config": "vllm_basic_correctness_test", "shard": 1, "num_shards": 1, "runner": "linux.g6.4xlarge.experimental.nvidia.gpu"}, {"config": "vllm_basic_models_test", "shard": 1, "num_shards": 1, "runner": "linux.g6.4xlarge.experimental.nvidia.gpu"}, {"config": "vllm_entrypoints_test", "shard": 1, "num_shards": 1, "runner": "linux.g6.4xlarge.experimental.nvidia.gpu"}, {"config": "vllm_regression_test", "shard": 1, "num_shards": 1, "runner": "linux.g6.4xlarge.experimental.nvidia.gpu"}, {"config": "vllm_multi_model_processor_test", "shard": 1, "num_shards": 1, "runner": "linux.g6.4xlarge.experimental.nvidia.gpu"}, {"config": "vllm_pytorch_compilation_unit_tests", "shard": 1, "num_shards": 1, "runner": "linux.g6.4xlarge.experimental.nvidia.gpu"}, {"config": "vllm_lora_28_failure_test", "shard": 1, "num_shards": 1, "runner": "linux.g6.4xlarge.experimental.nvidia.gpu"}, {"config": "vllm_multi_model_test_28_failure_test", "shard": 1, "num_shards": 1, "runner": "linux.g6.4xlarge.experimental.nvidia.gpu"}, {"config": "vllm_language_model_test_extended_generation_28_failure_test", "shard": 1, "num_shards": 1, "runner": "linux.g6.4xlarge.experimental.nvidia.gpu"}, {"config": "vllm_distributed_test_2_gpu_28_failure_test", "shard": 1, "num_shards": 1, "runner": "linux.g6.4xlarge.experimental.nvidia.gpu"}, {"config": "vllm_lora_test", "shard": 0, "num_shards": 4, "runner": "linux.g6.4xlarge.experimental.nvidia.gpu"}, {"config": "vllm_lora_test", "shard": 1, "num_shards": 4, "runner": "linux.g6.4xlarge.experimental.nvidia.gpu"}, {"config": "vllm_lora_test", "shard": 2, "num_shards": 4, "runner": "linux.g6.4xlarge.experimental.nvidia.gpu"}, {"config": "vllm_lora_test", "shard": 3, "num_shards": 4, "runner": "linux.g6.4xlarge.experimental.nvidia.gpu"}, {"config": "vllm_lora_tp_test_distributed", "shard": 1, "num_shards": 1, "runner": "linux.g6.12xlarge.nvidia.gpu"}, {"config": "vllm_distributed_test_28_failure_test", "shard": 1, "num_shards": 1, "runner": "linux.g6.12xlarge.nvidia.gpu"}]} 2025-11-03T17:06:36.1570162Z Setting output is-test-matrix-empty=False 2025-11-03T17:06:36.1683981Z ##[group]Run echo "Filtered matrix:" 2025-11-03T17:06:36.1684241Z echo "Filtered matrix:" 2025-11-03T17:06:36.1688940Z echo "{"include": [{"config": "vllm_basic_correctness_test", "shard": 1, "num_shards": 1, "runner": "linux.g6.4xlarge.experimental.nvidia.gpu"}, {"config": "vllm_basic_models_test", "shard": 1, "num_shards": 1, "runner": "linux.g6.4xlarge.experimental.nvidia.gpu"}, {"config": "vllm_entrypoints_test", "shard": 1, "num_shards": 1, "runner": "linux.g6.4xlarge.experimental.nvidia.gpu"}, {"config": "vllm_regression_test", "shard": 1, "num_shards": 1, "runner": "linux.g6.4xlarge.experimental.nvidia.gpu"}, {"config": "vllm_multi_model_processor_test", "shard": 1, "num_shards": 1, "runner": "linux.g6.4xlarge.experimental.nvidia.gpu"}, {"config": "vllm_pytorch_compilation_unit_tests", "shard": 1, "num_shards": 1, "runner": "linux.g6.4xlarge.experimental.nvidia.gpu"}, {"config": "vllm_lora_28_failure_test", "shard": 1, "num_shards": 1, "runner": "linux.g6.4xlarge.experimental.nvidia.gpu"}, {"config": "vllm_multi_model_test_28_failure_test", "shard": 1, "num_shards": 1, "runner": "linux.g6.4xlarge.experimental.nvidia.gpu"}, {"config": "vllm_language_model_test_extended_generation_28_failure_test", "shard": 1, "num_shards": 1, "runner": "linux.g6.4xlarge.experimental.nvidia.gpu"}, {"config": "vllm_distributed_test_2_gpu_28_failure_test", "shard": 1, "num_shards": 1, "runner": "linux.g6.4xlarge.experimental.nvidia.gpu"}, {"config": "vllm_lora_test", "shard": 0, "num_shards": 4, "runner": "linux.g6.4xlarge.experimental.nvidia.gpu"}, {"config": "vllm_lora_test", "shard": 1, "num_shards": 4, "runner": "linux.g6.4xlarge.experimental.nvidia.gpu"}, {"config": "vllm_lora_test", "shard": 2, "num_shards": 4, "runner": "linux.g6.4xlarge.experimental.nvidia.gpu"}, {"config": "vllm_lora_test", "shard": 3, "num_shards": 4, "runner": "linux.g6.4xlarge.experimental.nvidia.gpu"}, {"config": "vllm_lora_tp_test_distributed", "shard": 1, "num_shards": 1, "runner": "linux.g6.12xlarge.nvidia.gpu"}, {"config": "vllm_distributed_test_28_failure_test", "shard": 1, "num_shards": 1, "runner": "linux.g6.12xlarge.nvidia.gpu"}]}" 2025-11-03T17:06:36.1693942Z  2025-11-03T17:06:36.1694091Z echo 2025-11-03T17:06:36.1694288Z echo "Is the current job unstable? False" 2025-11-03T17:06:36.1694525Z  2025-11-03T17:06:36.1694662Z echo 2025-11-03T17:06:36.1694863Z echo "Is keep-going label set? True" 2025-11-03T17:06:36.1695083Z  2025-11-03T17:06:36.1695220Z echo 2025-11-03T17:06:36.1695407Z echo "Reenabled issues? " 2025-11-03T17:06:36.1703235Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2025-11-03T17:06:36.1703516Z env: 2025-11-03T17:06:36.1703673Z GIT_DEFAULT_BRANCH: main 2025-11-03T17:06:36.1703919Z GPU_FLAG: --gpus all -e NVIDIA_DRIVER_CAPABILITIES=all 2025-11-03T17:06:36.1704173Z ##[endgroup] 2025-11-03T17:06:36.1731786Z Filtered matrix: 2025-11-03T17:06:36.1738104Z {include: [{config: vllm_basic_correctness_test, shard: 1, num_shards: 1, runner: linux.g6.4xlarge.experimental.nvidia.gpu}, {config: vllm_basic_models_test, shard: 1, num_shards: 1, runner: linux.g6.4xlarge.experimental.nvidia.gpu}, {config: vllm_entrypoints_test, shard: 1, num_shards: 1, runner: linux.g6.4xlarge.experimental.nvidia.gpu}, {config: vllm_regression_test, shard: 1, num_shards: 1, runner: linux.g6.4xlarge.experimental.nvidia.gpu}, {config: vllm_multi_model_processor_test, shard: 1, num_shards: 1, runner: linux.g6.4xlarge.experimental.nvidia.gpu}, {config: vllm_pytorch_compilation_unit_tests, shard: 1, num_shards: 1, runner: linux.g6.4xlarge.experimental.nvidia.gpu}, {config: vllm_lora_28_failure_test, shard: 1, num_shards: 1, runner: linux.g6.4xlarge.experimental.nvidia.gpu}, {config: vllm_multi_model_test_28_failure_test, shard: 1, num_shards: 1, runner: linux.g6.4xlarge.experimental.nvidia.gpu}, {config: vllm_language_model_test_extended_generation_28_failure_test, shard: 1, num_shards: 1, runner: linux.g6.4xlarge.experimental.nvidia.gpu}, {config: vllm_distributed_test_2_gpu_28_failure_test, shard: 1, num_shards: 1, runner: linux.g6.4xlarge.experimental.nvidia.gpu}, {config: vllm_lora_test, shard: 0, num_shards: 4, runner: linux.g6.4xlarge.experimental.nvidia.gpu}, {config: vllm_lora_test, shard: 1, num_shards: 4, runner: linux.g6.4xlarge.experimental.nvidia.gpu}, {config: vllm_lora_test, shard: 2, num_shards: 4, runner: linux.g6.4xlarge.experimental.nvidia.gpu}, {config: vllm_lora_test, shard: 3, num_shards: 4, runner: linux.g6.4xlarge.experimental.nvidia.gpu}, {config: vllm_lora_tp_test_distributed, shard: 1, num_shards: 1, runner: linux.g6.12xlarge.nvidia.gpu}, {config: vllm_distributed_test_28_failure_test, shard: 1, num_shards: 1, runner: linux.g6.12xlarge.nvidia.gpu}]} 2025-11-03T17:06:36.1743046Z 2025-11-03T17:06:36.1743152Z Is the current job unstable? False 2025-11-03T17:06:36.1743302Z 2025-11-03T17:06:36.1743384Z Is keep-going label set? True 2025-11-03T17:06:36.1743525Z 2025-11-03T17:06:36.1743591Z Reenabled issues? 2025-11-03T17:06:36.1780929Z ##[group]Run echo "timeout=$((JOB_TIMEOUT-30))" >> "${GITHUB_OUTPUT}" 2025-11-03T17:06:36.1781351Z echo "timeout=$((JOB_TIMEOUT-30))" >> "${GITHUB_OUTPUT}" 2025-11-03T17:06:36.1788424Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2025-11-03T17:06:36.1788705Z env: 2025-11-03T17:06:36.1788875Z GIT_DEFAULT_BRANCH: main 2025-11-03T17:06:36.1789120Z GPU_FLAG: --gpus all -e NVIDIA_DRIVER_CAPABILITIES=all 2025-11-03T17:06:36.1789374Z JOB_TIMEOUT: 240 2025-11-03T17:06:36.1789543Z ##[endgroup] 2025-11-03T17:06:36.1858858Z ##[group]Run env | grep '^GITHUB' >> "/tmp/github_env_${GITHUB_RUN_ID}" 2025-11-03T17:06:36.1859269Z env | grep '^GITHUB' >> "/tmp/github_env_${GITHUB_RUN_ID}" 2025-11-03T17:06:36.1859587Z env | grep '^CI' >> "/tmp/github_env_${GITHUB_RUN_ID}" 2025-11-03T17:06:36.1866688Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2025-11-03T17:06:36.1866964Z env: 2025-11-03T17:06:36.1867123Z GIT_DEFAULT_BRANCH: main 2025-11-03T17:06:36.1867366Z GPU_FLAG: --gpus all -e NVIDIA_DRIVER_CAPABILITIES=all 2025-11-03T17:06:36.1867618Z ##[endgroup] 2025-11-03T17:06:36.1980407Z ##[group]Run set -x 2025-11-03T17:06:36.1980828Z set -x 2025-11-03T17:06:36.1981005Z  2025-11-03T17:06:36.1981201Z if [[ $TEST_CONFIG == 'multigpu' ]]; then 2025-11-03T17:06:36.1981699Z  TEST_COMMAND=.ci/pytorch/multigpu-test.sh 2025-11-03T17:06:36.1982117Z elif [[ $BUILD_ENVIRONMENT == *onnx* ]]; then 2025-11-03T17:06:36.1982386Z  TEST_COMMAND=.ci/onnx/test.sh 2025-11-03T17:06:36.1982598Z else 2025-11-03T17:06:36.1982784Z  TEST_COMMAND=.ci/pytorch/test.sh 2025-11-03T17:06:36.1983006Z fi 2025-11-03T17:06:36.1983150Z  2025-11-03T17:06:36.1983338Z # Leaving 1GB for the runner and other things 2025-11-03T17:06:36.1983753Z TOTAL_AVAILABLE_MEMORY_IN_GB=$(awk '/MemTotal/ { printf "%.3f \n", $2/1024/1024 - 1 }' /proc/meminfo) 2025-11-03T17:06:36.1984406Z # https://docs.docker.com/engine/containers/resource_constraints/#--memory-swap-details, the 3GB swap 2025-11-03T17:06:36.1984907Z # comes from https://github.com/pytorch/test-infra/pull/6058 2025-11-03T17:06:36.1985289Z TOTAL_MEMORY_WITH_SWAP=$(("${TOTAL_AVAILABLE_MEMORY_IN_GB%.*}" + 3)) 2025-11-03T17:06:36.1985598Z  2025-11-03T17:06:36.1985793Z if [[ ${BUILD_ENVIRONMENT} == *"s390x"* ]]; then 2025-11-03T17:06:36.1986035Z  SHM_OPTS= 2025-11-03T17:06:36.1986209Z  JENKINS_USER= 2025-11-03T17:06:36.1986472Z  # ensure that docker container cleanly exits in 12 hours 2025-11-03T17:06:36.1986805Z  # if for some reason cleanup action doesn't stop container 2025-11-03T17:06:36.1987081Z  # when job is cancelled 2025-11-03T17:06:36.1987299Z  DOCKER_SHELL_CMD="sleep 12h" 2025-11-03T17:06:36.1987500Z else 2025-11-03T17:06:36.1987677Z  SHM_OPTS="--shm-size=${SHM_SIZE}" 2025-11-03T17:06:36.1987910Z  JENKINS_USER="--user jenkins" 2025-11-03T17:06:36.1988133Z  DOCKER_SHELL_CMD= 2025-11-03T17:06:36.1988312Z fi 2025-11-03T17:06:36.1988456Z  2025-11-03T17:06:36.1988691Z # detached container should get cleaned up by teardown_ec2_linux 2025-11-03T17:06:36.1989242Z # TODO: Stop building test binaries as part of the build phase 2025-11-03T17:06:36.1989677Z # Used for GPU_FLAG, SHM_OPTS, JENKINS_USER and DOCKER_SHELL_CMD since that doesn't play nice 2025-11-03T17:06:36.1990056Z # shellcheck disable=SC2086,SC2090 2025-11-03T17:06:36.1990298Z container_name=$(docker run \ 2025-11-03T17:06:36.1990521Z  ${GPU_FLAG:-} \ 2025-11-03T17:06:36.1990747Z  ${SCCACHE_SERVER_PORT_DOCKER_FLAG:-} \ 2025-11-03T17:06:36.1990988Z  -e BUILD_ENVIRONMENT \ 2025-11-03T17:06:36.1991198Z  -e PR_NUMBER \ 2025-11-03T17:06:36.1991399Z  -e GITHUB_ACTIONS \ 2025-11-03T17:06:36.1991606Z  -e GITHUB_REPOSITORY \ 2025-11-03T17:06:36.1991814Z  -e GITHUB_WORKFLOW \ 2025-11-03T17:06:36.1992023Z  -e GITHUB_JOB \ 2025-11-03T17:06:36.1992220Z  -e GITHUB_RUN_ID \ 2025-11-03T17:06:36.1992425Z  -e GITHUB_RUN_NUMBER \ 2025-11-03T17:06:36.1992643Z  -e GITHUB_RUN_ATTEMPT \ 2025-11-03T17:06:36.1992852Z  -e JOB_ID \ 2025-11-03T17:06:36.1993032Z  -e JOB_NAME \ 2025-11-03T17:06:36.1993214Z  -e BASE_SHA \ 2025-11-03T17:06:36.1993385Z  -e BRANCH \ 2025-11-03T17:06:36.1993560Z  -e SHA1 \ 2025-11-03T17:06:36.1993752Z  -e AWS_DEFAULT_REGION \ 2025-11-03T17:06:36.1993962Z  -e IN_WHEEL_TEST \ 2025-11-03T17:06:36.1994151Z  -e SHARD_NUMBER \ 2025-11-03T17:06:36.1994345Z  -e TEST_CONFIG \ 2025-11-03T17:06:36.1994542Z  -e NUM_TEST_SHARDS \ 2025-11-03T17:06:36.1994749Z  -e REENABLED_ISSUES \ 2025-11-03T17:06:36.1994959Z  -e CONTINUE_THROUGH_ERROR \ 2025-11-03T17:06:36.1995292Z  -e VERBOSE_TEST_LOGS \ 2025-11-03T17:06:36.1995519Z  -e TEST_SHOWLOCALS \ 2025-11-03T17:06:36.1995719Z  -e NO_TEST_TIMEOUT \ 2025-11-03T17:06:36.1995908Z  -e NO_TD \ 2025-11-03T17:06:36.1996084Z  -e TD_DISTRIBUTED \ 2025-11-03T17:06:36.1996280Z  -e PR_LABELS \ 2025-11-03T17:06:36.1996495Z  -e MAX_JOBS="$(nproc --ignore=2)" \ 2025-11-03T17:06:36.1996725Z  -e SCCACHE_BUCKET \ 2025-11-03T17:06:36.1996924Z  -e SCCACHE_REGION \ 2025-11-03T17:06:36.1997114Z  -e XLA_CUDA \ 2025-11-03T17:06:36.1997315Z  -e XLA_CLANG_CACHE_S3_BUCKET_NAME \ 2025-11-03T17:06:36.1997567Z  -e PYTORCH_TEST_CUDA_MEM_LEAK_CHECK \ 2025-11-03T17:06:36.1997820Z  -e PYTORCH_TEST_RERUN_DISABLED_TESTS \ 2025-11-03T17:06:36.1998076Z  -e SKIP_SCCACHE_INITIALIZATION=1 \ 2025-11-03T17:06:36.1998317Z  -e HUGGING_FACE_HUB_TOKEN \ 2025-11-03T17:06:36.1998551Z  -e VLLM_TEST_HUGGING_FACE_TOKEN \ 2025-11-03T17:06:36.1998794Z  -e SCRIBE_GRAPHQL_ACCESS_TOKEN \ 2025-11-03T17:06:36.1999020Z  -e DASHBOARD_TAG \ 2025-11-03T17:06:36.1999228Z  -e ARTIFACTS_FILE_SUFFIX \ 2025-11-03T17:06:36.1999505Z  --memory="${TOTAL_AVAILABLE_MEMORY_IN_GB%.*}g" \ 2025-11-03T17:06:36.1999799Z  --memory-swap="${TOTAL_MEMORY_WITH_SWAP}g" \ 2025-11-03T17:06:36.2000089Z  --env-file="/tmp/github_env_${GITHUB_RUN_ID}" \ 2025-11-03T17:06:36.2000377Z  --security-opt seccomp=unconfined \ 2025-11-03T17:06:36.2000622Z  --cap-add=SYS_PTRACE \ 2025-11-03T17:06:36.2000833Z  --ipc=host \ 2025-11-03T17:06:36.2001014Z  ${SHM_OPTS} \ 2025-11-03T17:06:36.2001192Z  --tty \ 2025-11-03T17:06:36.2001363Z  --detach \ 2025-11-03T17:06:36.2001551Z  --name="${container_name}" \ 2025-11-03T17:06:36.2001768Z  ${JENKINS_USER} \ 2025-11-03T17:06:36.2002021Z  -v "${GITHUB_WORKSPACE}:/var/lib/jenkins/workspace" \ 2025-11-03T17:06:36.2002302Z  -w /var/lib/jenkins/workspace \ 2025-11-03T17:06:36.2002525Z  "${DOCKER_IMAGE}" \ 2025-11-03T17:06:36.2002819Z  ${DOCKER_SHELL_CMD} 2025-11-03T17:06:36.2003001Z ) 2025-11-03T17:06:36.2003237Z echo "DOCKER_CONTAINER_ID=${container_name}" >> "${GITHUB_ENV}" 2025-11-03T17:06:36.2003524Z  2025-11-03T17:06:36.2003719Z if [[ ${BUILD_ENVIRONMENT} == *"s390x"* ]]; then 2025-11-03T17:06:36.2004137Z  docker exec -t "${container_name}" sh -c "python3 -m pip install -r .ci/docker/requirements-ci.txt" 2025-11-03T17:06:36.2004507Z fi 2025-11-03T17:06:36.2004653Z  2025-11-03T17:06:36.2004997Z docker exec -t "${container_name}" sh -c "python3 -m pip install $(echo dist/*.whl)[opt-einsum] && ${TEST_COMMAND}" 2025-11-03T17:06:36.2012079Z shell: /usr/bin/bash -e {0} 2025-11-03T17:06:36.2012294Z env: 2025-11-03T17:06:36.2012464Z GIT_DEFAULT_BRANCH: main 2025-11-03T17:06:36.2012712Z GPU_FLAG: --gpus all -e NVIDIA_DRIVER_CAPABILITIES=all 2025-11-03T17:06:36.2013025Z BUILD_ENVIRONMENT: linux-jammy-cuda12.8-py3.12-gcc11 2025-11-03T17:06:36.2013291Z PR_NUMBER: 2025-11-03T17:06:36.2013482Z GITHUB_REPOSITORY: pytorch/pytorch 2025-11-03T17:06:36.2013705Z GITHUB_WORKFLOW: vllm-test 2025-11-03T17:06:36.2013890Z GITHUB_JOB: test 2025-11-03T17:06:36.2014057Z GITHUB_RUN_ID: 19040285420 2025-11-03T17:06:36.2014257Z GITHUB_RUN_NUMBER: 3325 2025-11-03T17:06:36.2014443Z GITHUB_RUN_ATTEMPT: 1 2025-11-03T17:06:36.2014613Z JOB_ID: 54383051168 2025-11-03T17:06:36.2014953Z JOB_NAME: ci-vllm-test / test (vllm_basic_models_test, 1, 1, linux.g6.4xlarge.experimental.nvidia.gpu) 2025-11-03T17:06:36.2015324Z BRANCH: main 2025-11-03T17:06:36.2015517Z SHA1: 3f6538febd727b782e6e13cfd026a309fb14351d 2025-11-03T17:06:36.2015780Z BASE_SHA: 3f6538febd727b782e6e13cfd026a309fb14351d 2025-11-03T17:06:36.2016144Z TEST_CONFIG: vllm_basic_models_test 2025-11-03T17:06:36.2016376Z SHARD_NUMBER: 1 2025-11-03T17:06:36.2016544Z NUM_TEST_SHARDS: 1 2025-11-03T17:06:36.2016704Z EXTRA_FLAGS: 2025-11-03T17:06:36.2016872Z OP_BENCHMARK_TESTS: 2025-11-03T17:06:36.2017046Z REENABLED_ISSUES: 2025-11-03T17:06:36.2017223Z CONTINUE_THROUGH_ERROR: True 2025-11-03T17:06:36.2017415Z VERBOSE_TEST_LOGS: False 2025-11-03T17:06:36.2017601Z TEST_SHOWLOCALS: False 2025-11-03T17:06:36.2017783Z NO_TEST_TIMEOUT: False 2025-11-03T17:06:36.2017953Z NO_TD: False 2025-11-03T17:06:36.2018109Z TD_DISTRIBUTED: False 2025-11-03T17:06:36.2018336Z SCCACHE_BUCKET: ossci-compiler-cache-circleci-v2 2025-11-03T17:06:36.2018591Z SCCACHE_REGION: us-east-1 2025-11-03T17:06:36.2018776Z SHM_SIZE: 2g 2025-11-03T17:06:36.2019349Z DOCKER_IMAGE: 308535385114.dkr.ecr.us-east-1.amazonaws.com/pytorch/ci-image:pytorch-linux-jammy-cuda12.8-cudnn9-py3.12-gcc11-vllm-30147fb242f88ab8f13e14b76ab186d189d8605c 2025-11-03T17:06:36.2019955Z XLA_CUDA: 2025-11-03T17:06:36.2020213Z XLA_CLANG_CACHE_S3_BUCKET_NAME: ossci-compiler-clang-cache-circleci-xla 2025-11-03T17:06:36.2020536Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK: 0 2025-11-03T17:06:36.2020768Z PYTORCH_TEST_RERUN_DISABLED_TESTS: 0 2025-11-03T17:06:36.2020984Z DASHBOARD_TAG: 2025-11-03T17:06:36.2021674Z VLLM_TEST_HUGGING_FACE_TOKEN: *** 2025-11-03T17:06:36.2022002Z HUGGING_FACE_HUB_TOKEN: *** 2025-11-03T17:06:36.2022313Z SCRIBE_GRAPHQL_ACCESS_TOKEN: *** 2025-11-03T17:06:36.2022744Z ARTIFACTS_FILE_SUFFIX: test-vllm_basic_models_test-1-1-linux.g6.4xlarge.experimental.nvidia.gpu_54383051168 2025-11-03T17:06:36.2023156Z ##[endgroup] 2025-11-03T17:06:36.2049319Z + [[ vllm_basic_models_test == \m\u\l\t\i\g\p\u ]] 2025-11-03T17:06:36.2049711Z + [[ linux-jammy-cuda12.8-py3.12-gcc11 == *onnx* ]] 2025-11-03T17:06:36.2050041Z + TEST_COMMAND=.ci/pytorch/test.sh 2025-11-03T17:06:36.2052945Z ++ awk '/MemTotal/ { printf "%.3f \n", $2/1024/1024 - 1 }' /proc/meminfo 2025-11-03T17:06:36.2081396Z + TOTAL_AVAILABLE_MEMORY_IN_GB='59.453 ' 2025-11-03T17:06:36.2081952Z + TOTAL_MEMORY_WITH_SWAP=62 2025-11-03T17:06:36.2082488Z + [[ linux-jammy-cuda12.8-py3.12-gcc11 == *\s\3\9\0\x* ]] 2025-11-03T17:06:36.2083326Z + SHM_OPTS=--shm-size=2g 2025-11-03T17:06:36.2083729Z + JENKINS_USER='--user jenkins' 2025-11-03T17:06:36.2084088Z + DOCKER_SHELL_CMD= 2025-11-03T17:06:36.2091890Z +++ nproc --ignore=2 2025-11-03T17:06:36.2125553Z ++ docker run --gpus all -e NVIDIA_DRIVER_CAPABILITIES=all -e BUILD_ENVIRONMENT -e PR_NUMBER -e GITHUB_ACTIONS -e GITHUB_REPOSITORY -e GITHUB_WORKFLOW -e GITHUB_JOB -e GITHUB_RUN_ID -e GITHUB_RUN_NUMBER -e GITHUB_RUN_ATTEMPT -e JOB_ID -e JOB_NAME -e BASE_SHA -e BRANCH -e SHA1 -e AWS_DEFAULT_REGION -e IN_WHEEL_TEST -e SHARD_NUMBER -e TEST_CONFIG -e NUM_TEST_SHARDS -e REENABLED_ISSUES -e CONTINUE_THROUGH_ERROR -e VERBOSE_TEST_LOGS -e TEST_SHOWLOCALS -e NO_TEST_TIMEOUT -e NO_TD -e TD_DISTRIBUTED -e PR_LABELS -e MAX_JOBS=14 -e SCCACHE_BUCKET -e SCCACHE_REGION -e XLA_CUDA -e XLA_CLANG_CACHE_S3_BUCKET_NAME -e PYTORCH_TEST_CUDA_MEM_LEAK_CHECK -e PYTORCH_TEST_RERUN_DISABLED_TESTS -e SKIP_SCCACHE_INITIALIZATION=1 -e HUGGING_FACE_HUB_TOKEN -e VLLM_TEST_HUGGING_FACE_TOKEN -e SCRIBE_GRAPHQL_ACCESS_TOKEN -e DASHBOARD_TAG -e ARTIFACTS_FILE_SUFFIX --memory=59g --memory-swap=62g --env-file=/tmp/github_env_19040285420 --security-opt seccomp=unconfined --cap-add=SYS_PTRACE --ipc=host --shm-size=2g --tty --detach --name= --user jenkins -v /home/ec2-user/actions-runner/_work/pytorch/pytorch:/var/lib/jenkins/workspace -w /var/lib/jenkins/workspace 308535385114.dkr.ecr.us-east-1.amazonaws.com/pytorch/ci-image:pytorch-linux-jammy-cuda12.8-cudnn9-py3.12-gcc11-vllm-30147fb242f88ab8f13e14b76ab186d189d8605c 2025-11-03T17:06:58.0393465Z + container_name=c5e792ba8f7ad58ea91feb6e358bcdc693306479a78a475d1e0ca6fca41e9d73 2025-11-03T17:06:58.0394175Z + echo DOCKER_CONTAINER_ID=c5e792ba8f7ad58ea91feb6e358bcdc693306479a78a475d1e0ca6fca41e9d73 2025-11-03T17:06:58.0395075Z + [[ linux-jammy-cuda12.8-py3.12-gcc11 == *\s\3\9\0\x* ]] 2025-11-03T17:06:58.0402364Z ++ echo dist/torch-2.10.0a0+git3f6538f-cp312-cp312-linux_x86_64.whl 2025-11-03T17:06:58.0404886Z + docker exec -t c5e792ba8f7ad58ea91feb6e358bcdc693306479a78a475d1e0ca6fca41e9d73 sh -c 'python3 -m pip install dist/torch-2.10.0a0+git3f6538f-cp312-cp312-linux_x86_64.whl[opt-einsum] && .ci/pytorch/test.sh' 2025-11-03T17:06:58.5739223Z Processing ./dist/torch-2.10.0a0+git3f6538f-cp312-cp312-linux_x86_64.whl (from torch==2.10.0a0+git3f6538f) 2025-11-03T17:06:58.9689492Z Requirement already satisfied: filelock in /opt/conda/envs/py_3.12/lib/python3.12/site-packages (from torch==2.10.0a0+git3f6538f->torch==2.10.0a0+git3f6538f) (3.18.0) 2025-11-03T17:06:58.9691374Z Requirement already satisfied: typing-extensions>=4.10.0 in /opt/conda/envs/py_3.12/lib/python3.12/site-packages (from torch==2.10.0a0+git3f6538f->torch==2.10.0a0+git3f6538f) (4.12.2) 2025-11-03T17:06:58.9695918Z Requirement already satisfied: setuptools in /opt/conda/envs/py_3.12/lib/python3.12/site-packages (from torch==2.10.0a0+git3f6538f->torch==2.10.0a0+git3f6538f) (78.1.1) 2025-11-03T17:06:58.9699108Z Requirement already satisfied: sympy>=1.13.3 in /opt/conda/envs/py_3.12/lib/python3.12/site-packages (from torch==2.10.0a0+git3f6538f->torch==2.10.0a0+git3f6538f) (1.13.3) 2025-11-03T17:06:58.9702497Z Requirement already satisfied: networkx>=2.5.1 in /opt/conda/envs/py_3.12/lib/python3.12/site-packages (from torch==2.10.0a0+git3f6538f->torch==2.10.0a0+git3f6538f) (2.8.8) 2025-11-03T17:06:58.9705778Z Requirement already satisfied: jinja2 in /opt/conda/envs/py_3.12/lib/python3.12/site-packages (from torch==2.10.0a0+git3f6538f->torch==2.10.0a0+git3f6538f) (3.1.6) 2025-11-03T17:06:58.9708998Z Requirement already satisfied: fsspec>=0.8.5 in /opt/conda/envs/py_3.12/lib/python3.12/site-packages (from torch==2.10.0a0+git3f6538f->torch==2.10.0a0+git3f6538f) (2025.9.0) 2025-11-03T17:06:58.9720710Z Requirement already satisfied: opt-einsum>=3.3 in /opt/conda/envs/py_3.12/lib/python3.12/site-packages (from torch==2.10.0a0+git3f6538f->torch==2.10.0a0+git3f6538f) (3.3.0) 2025-11-03T17:06:58.9820336Z Requirement already satisfied: numpy>=1.7 in /opt/conda/envs/py_3.12/lib/python3.12/site-packages (from opt-einsum>=3.3->torch==2.10.0a0+git3f6538f->torch==2.10.0a0+git3f6538f) (1.26.2) 2025-11-03T17:06:58.9858988Z Requirement already satisfied: mpmath<1.4,>=1.1.0 in /opt/conda/envs/py_3.12/lib/python3.12/site-packages (from sympy>=1.13.3->torch==2.10.0a0+git3f6538f->torch==2.10.0a0+git3f6538f) (1.3.0) 2025-11-03T17:06:58.9902266Z Requirement already satisfied: MarkupSafe>=2.0 in /opt/conda/envs/py_3.12/lib/python3.12/site-packages (from jinja2->torch==2.10.0a0+git3f6538f->torch==2.10.0a0+git3f6538f) (3.0.3) 2025-11-03T17:06:59.1886771Z Installing collected packages: torch 2025-11-03T17:07:11.2121242Z Successfully installed torch-2.10.0a0+git3f6538f 2025-11-03T17:07:11.2766338Z + export TERM=vt100 2025-11-03T17:07:11.2766638Z + TERM=vt100 2025-11-03T17:07:11.2768651Z ++ dirname .ci/pytorch/test.sh 2025-11-03T17:07:11.2778775Z + source .ci/pytorch/common.sh 2025-11-03T17:07:11.2782620Z +++ dirname .ci/pytorch/common.sh 2025-11-03T17:07:11.2790649Z ++ source .ci/pytorch/common_utils.sh 2025-11-03T17:07:11.2791864Z +++ declare -f -t trap_add 2025-11-03T17:07:11.2796141Z ++ set -ex -o pipefail 2025-11-03T17:07:11.2796545Z ++ [[ linux-jammy-cuda12.8-py3.12-gcc11 == *rocm* ]] 2025-11-03T17:07:11.2796889Z ++ BUILD_TEST_LIBTORCH=0 2025-11-03T17:07:11.2799870Z ++ dirname .ci/pytorch/test.sh 2025-11-03T17:07:11.2808613Z + source .ci/pytorch/common-build.sh 2025-11-03T17:07:11.2810266Z ++ [[ linux-jammy-cuda12.8-py3.12-gcc11 != *win-* ]] 2025-11-03T17:07:11.2816534Z ++++ dirname .ci/pytorch/common-build.sh 2025-11-03T17:07:11.2825668Z +++ cd .ci/pytorch 2025-11-03T17:07:11.2826031Z +++ pwd -P 2025-11-03T17:07:11.2904899Z ++ script_dir=/var/lib/jenkins/workspace/.ci/pytorch 2025-11-03T17:07:11.2905401Z ++ [[ linux-jammy-cuda12.8-py3.12-gcc11 == *-pch* ]] 2025-11-03T17:07:11.2905705Z ++ which sccache 2025-11-03T17:07:11.2974257Z ++ [[ -z ossci-compiler-cache-circleci-v2 ]] 2025-11-03T17:07:11.2975007Z ++ sccache --stop-server 2025-11-03T17:07:11.3004583Z ++ true 2025-11-03T17:07:11.3004945Z ++ rm -f /var/lib/jenkins/sccache_error.log 2025-11-03T17:07:11.3016572Z ++ trap_add sccache_epilogue EXIT 2025-11-03T17:07:11.3016979Z ++ trap_add_cmd=sccache_epilogue 2025-11-03T17:07:11.3017222Z ++ shift 2025-11-03T17:07:11.3017421Z ++ for trap_add_name in "$@" 2025-11-03T17:07:11.3024467Z ++++ trap -p EXIT 2025-11-03T17:07:11.3027163Z +++ eval 'extract_trap_cmd ' 2025-11-03T17:07:11.3027610Z ++++ extract_trap_cmd 2025-11-03T17:07:11.3027858Z ++++ printf '%s\n' '' 2025-11-03T17:07:11.3028107Z +++ printf '%s\n' sccache_epilogue 2025-11-03T17:07:11.3030675Z ++ trap -- ' 2025-11-03T17:07:11.3031102Z sccache_epilogue' EXIT 2025-11-03T17:07:11.3031376Z ++ [[ -n 1 ]] 2025-11-03T17:07:11.3031877Z ++ echo 'Skipping sccache server initialization, setting environment variables' 2025-11-03T17:07:11.3032521Z Skipping sccache server initialization, setting environment variables 2025-11-03T17:07:11.3032885Z ++ export SCCACHE_IDLE_TIMEOUT=0 2025-11-03T17:07:11.3033262Z ++ SCCACHE_IDLE_TIMEOUT=0 2025-11-03T17:07:11.3033626Z ++ export SCCACHE_ERROR_LOG=/var/lib/jenkins/sccache_error.log 2025-11-03T17:07:11.3033983Z ++ SCCACHE_ERROR_LOG=/var/lib/jenkins/sccache_error.log 2025-11-03T17:07:11.3040214Z ++ export RUST_LOG=sccache::server=error 2025-11-03T17:07:11.3040502Z ++ RUST_LOG=sccache::server=error 2025-11-03T17:07:11.3040723Z ++ sccache --zero-stats 2025-11-03T17:07:11.6435430Z Statistics zeroed. 2025-11-03T17:07:11.6442935Z ++ which ccache 2025-11-03T17:07:11.6502484Z + [[ linux-jammy-cuda12.8-py3.12-gcc11 != *rocm* ]] 2025-11-03T17:07:11.6503511Z + [[ linux-jammy-cuda12.8-py3.12-gcc11 != *s390x* ]] 2025-11-03T17:07:11.6504305Z + [[ -d /var/lib/jenkins/workspace ]] 2025-11-03T17:07:11.6509360Z ++ stat -c %u /var/lib/jenkins/workspace 2025-11-03T17:07:11.6526713Z + WORKSPACE_ORIGINAL_OWNER_ID=1000 2025-11-03T17:07:11.6527107Z + trap_add cleanup_workspace EXIT 2025-11-03T17:07:11.6527387Z + trap_add_cmd=cleanup_workspace 2025-11-03T17:07:11.6527661Z + shift 2025-11-03T17:07:11.6527860Z + for trap_add_name in "$@" 2025-11-03T17:07:11.6534752Z +++ trap -p EXIT 2025-11-03T17:07:11.6537406Z ++ eval 'extract_trap_cmd trap -- '\'' 2025-11-03T17:07:11.6538249Z sccache_epilogue'\'' EXIT' 2025-11-03T17:07:11.6538666Z +++ extract_trap_cmd trap -- ' 2025-11-03T17:07:11.6539086Z sccache_epilogue' EXIT 2025-11-03T17:07:11.6539443Z +++ printf '%s\n' ' 2025-11-03T17:07:11.6539737Z sccache_epilogue' 2025-11-03T17:07:11.6539970Z ++ printf '%s\n' cleanup_workspace 2025-11-03T17:07:11.6540341Z + trap -- ' 2025-11-03T17:07:11.6540661Z sccache_epilogue 2025-11-03T17:07:11.6541016Z cleanup_workspace' EXIT 2025-11-03T17:07:11.6541549Z + sudo chown -R jenkins /var/lib/jenkins/workspace 2025-11-03T17:07:12.6214284Z + git config --global --add safe.directory /var/lib/jenkins/workspace 2025-11-03T17:07:12.6235788Z + [[ linux-jammy-cuda12.8-py3.12-gcc11 == *cuda* ]] 2025-11-03T17:07:12.6239680Z ++ python -c 'import os;import numba.cuda; print(os.path.dirname(numba.cuda.__file__))' 2025-11-03T17:07:13.1121838Z + NUMBA_CUDA_DIR=/opt/conda/envs/py_3.12/lib/python3.12/site-packages/numba/cuda 2025-11-03T17:07:13.1122466Z + '[' -n /opt/conda/envs/py_3.12/lib/python3.12/site-packages/numba/cuda ']' 2025-11-03T17:07:13.1133447Z +++ realpath .ci/pytorch/test.sh 2025-11-03T17:07:13.1140532Z ++ dirname /var/lib/jenkins/workspace/.ci/pytorch/test.sh 2025-11-03T17:07:13.1337962Z + NUMBA_PATCH=/var/lib/jenkins/workspace/.ci/pytorch/numba-cuda-13.patch 2025-11-03T17:07:13.1338543Z + pushd /opt/conda/envs/py_3.12/lib/python3.12/site-packages/numba/cuda 2025-11-03T17:07:13.1339106Z /opt/conda/envs/py_3.12/lib/python3.12/site-packages/numba/cuda ~/workspace 2025-11-03T17:07:13.1339531Z + patch -p4 2025-11-03T17:07:13.1355265Z patching file cudadrv/driver.py 2025-11-03T17:07:13.1359706Z + popd 2025-11-03T17:07:13.1359936Z ~/workspace 2025-11-03T17:07:13.1360166Z + echo 'Environment variables:' 2025-11-03T17:07:13.1360442Z Environment variables: 2025-11-03T17:07:13.1360656Z + env 2025-11-03T17:07:13.1370864Z GITHUB_WORKSPACE=/home/ec2-user/actions-runner/_work/pytorch/pytorch 2025-11-03T17:07:13.1371446Z CONTINUE_THROUGH_ERROR=True 2025-11-03T17:07:13.1371886Z BUILD_ENVIRONMENT=linux-jammy-cuda12.8-py3.12-gcc11 2025-11-03T17:07:13.1372625Z VLLM_TEST_HUGGING_FACE_TOKEN=*** 2025-11-03T17:07:13.1372936Z HOSTNAME=c5e792ba8f7a 2025-11-03T17:07:13.1373382Z GITHUB_PATH=/home/ec2-user/actions-runner/_work/_temp/_runner_file_commands/add_path_edfee93f-f196-45b5-89e0-86041d1a5f42 2025-11-03T17:07:13.1373947Z GITHUB_ACTION=__run_3 2025-11-03T17:07:13.1374133Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=0 2025-11-03T17:07:13.1374353Z GITHUB_RUN_NUMBER=3325 2025-11-03T17:07:13.1374545Z TEST_CONFIG=vllm_basic_models_test 2025-11-03T17:07:13.1374767Z GITHUB_REPOSITORY_OWNER_ID=21003710 2025-11-03T17:07:13.1374995Z TORCH_NVCC_FLAGS=-Xfatbin -compress-all 2025-11-03T17:07:13.1375306Z SCCACHE_IDLE_TIMEOUT=0 2025-11-03T17:07:13.1375670Z SCRIBE_GRAPHQL_ACCESS_TOKEN=*** 2025-11-03T17:07:13.1375912Z GITHUB_TRIGGERING_ACTOR=pytorchmergebot 2025-11-03T17:07:13.1376144Z GITHUB_REF_TYPE=branch 2025-11-03T17:07:13.1376371Z BASE_SHA=3f6538febd727b782e6e13cfd026a309fb14351d 2025-11-03T17:07:13.1376614Z XLA_CUDA= 2025-11-03T17:07:13.1376777Z NCCL_LIB_DIR=/usr/local/cuda/lib64/ 2025-11-03T17:07:13.1377068Z HUGGING_FACE_HUB_TOKEN=*** 2025-11-03T17:07:13.1377442Z *** 2025-11-03T17:07:13.1377612Z GITHUB_REPOSITORY_ID=65600975 2025-11-03T17:07:13.1377815Z GITHUB_ACTIONS=true 2025-11-03T17:07:13.1377997Z NVIDIA_DRIVER_CAPABILITIES=all 2025-11-03T17:07:13.1378228Z SCCACHE_ERROR_LOG=/var/lib/jenkins/sccache_error.log 2025-11-03T17:07:13.1378502Z SHA1=3f6538febd727b782e6e13cfd026a309fb14351d 2025-11-03T17:07:13.1378762Z GITHUB_SHA=3f6538febd727b782e6e13cfd026a309fb14351d 2025-11-03T17:07:13.1379119Z GITHUB_WORKFLOW_REF=pytorch/pytorch/.github/workflows/vllm.yml@refs/heads/main 2025-11-03T17:07:13.1379438Z UCC_HOME=/usr 2025-11-03T17:07:13.1379598Z VERBOSE_TEST_LOGS=False 2025-11-03T17:07:13.1379781Z GITHUB_REF=refs/heads/main 2025-11-03T17:07:13.1379968Z SHARD_NUMBER=1 2025-11-03T17:07:13.1380128Z GITHUB_REF_PROTECTED=true 2025-11-03T17:07:13.1380315Z HOME=/var/lib/jenkins 2025-11-03T17:07:13.1380733Z GITHUB_API_URL=https://api.github.com 2025-11-03T17:07:13.1380975Z PYTORCH_TEST_RERUN_DISABLED_TESTS=0 2025-11-03T17:07:13.1381213Z UCX_COMMIT=7836b165abdbe468a2f607e7254011c07d788152 2025-11-03T17:07:13.1381450Z USE_SYSTEM_NCCL=1 2025-11-03T17:07:13.1381763Z NUM_TEST_SHARDS=1 2025-11-03T17:07:13.1381923Z UCX_HOME=/usr 2025-11-03T17:07:13.1382322Z GITHUB_STATE=/home/ec2-user/actions-runner/_work/_temp/_runner_file_commands/save_state_edfee93f-f196-45b5-89e0-86041d1a5f42 2025-11-03T17:07:13.1382934Z JOB_NAME=ci-vllm-test / test (vllm_basic_models_test, 1, 1, linux.g6.4xlarge.experimental.nvidia.gpu) 2025-11-03T17:07:13.1383670Z GITHUB_ENV=/home/ec2-user/actions-runner/_work/_temp/_runner_file_commands/set_env_edfee93f-f196-45b5-89e0-86041d1a5f42 2025-11-03T17:07:13.1384240Z GITHUB_EVENT_PATH=/home/ec2-user/actions-runner/_work/_temp/_github_workflow/event.json 2025-11-03T17:07:13.1384605Z GITHUB_EVENT_NAME=push 2025-11-03T17:07:13.1384777Z DASHBOARD_TAG= 2025-11-03T17:07:13.1384943Z GITHUB_RUN_ID=19040285420 2025-11-03T17:07:13.1385137Z INSTALLED_OPENBLAS= 2025-11-03T17:07:13.1385561Z GITHUB_STEP_SUMMARY=/home/ec2-user/actions-runner/_work/_temp/_runner_file_commands/step_summary_edfee93f-f196-45b5-89e0-86041d1a5f42 2025-11-03T17:07:13.1386032Z GITHUB_ACTOR=pytorchmergebot 2025-11-03T17:07:13.1386222Z PR_NUMBER= 2025-11-03T17:07:13.1386384Z DESIRED_CUDA=12.8.1 2025-11-03T17:07:13.1386557Z GITHUB_RUN_ATTEMPT=1 2025-11-03T17:07:13.1386729Z ANACONDA_PYTHON_VERSION=3.12 2025-11-03T17:07:13.1386965Z GITHUB_GRAPHQL_URL=https://api.github.com/graphql 2025-11-03T17:07:13.1387203Z TERM=vt100 2025-11-03T17:07:13.1387353Z INSTALLED_VISION=yes 2025-11-03T17:07:13.1387511Z BRANCH=main 2025-11-03T17:07:13.1387669Z SCCACHE_REGION=us-east-1 2025-11-03T17:07:13.1387864Z OPENSSL_ROOT_DIR=/opt/openssl 2025-11-03T17:07:13.1388059Z BUILD_AOT_INDUCTOR_TEST= 2025-11-03T17:07:13.1388356Z CUDA_PATH=/usr/local/cuda 2025-11-03T17:07:13.1388725Z GITHUB_ACTION_PATH=/home/ec2-user/actions-runner/_work/pytorch/pytorch/./.github/actions/setup-linux 2025-11-03T17:07:13.1389131Z GITHUB_SERVER_URL=https://github.com 2025-11-03T17:07:13.1389376Z UCC_COMMIT=430e241bf5d38cbc73fc7a6b89155397232e3f96 2025-11-03T17:07:13.1389606Z REENABLED_ISSUES= 2025-11-03T17:07:13.1389761Z DOCS= 2025-11-03T17:07:13.1389901Z SHLVL=1 2025-11-03T17:07:13.1390042Z MAX_JOBS=14 2025-11-03T17:07:13.1390190Z GITHUB_ACTOR_ID=97764156 2025-11-03T17:07:13.1390435Z GITHUB_WORKFLOW_SHA=3f6538febd727b782e6e13cfd026a309fb14351d 2025-11-03T17:07:13.1390702Z GITHUB_REF_NAME=main 2025-11-03T17:07:13.1390966Z XLA_CLANG_CACHE_S3_BUCKET_NAME=ossci-compiler-clang-cache-circleci-xla 2025-11-03T17:07:13.1391268Z GITHUB_JOB=test 2025-11-03T17:07:13.1391429Z NO_TEST_TIMEOUT=False 2025-11-03T17:07:13.1391604Z TD_DISTRIBUTED=False 2025-11-03T17:07:13.1391791Z GITHUB_REPOSITORY=pytorch/pytorch 2025-11-03T17:07:13.1392004Z GITHUB_RETENTION_DAYS=90 2025-11-03T17:07:13.1392188Z OPENSSL_DIR=/opt/openssl 2025-11-03T17:07:13.1392372Z GITHUB_ACTION_REPOSITORY= 2025-11-03T17:07:13.1392909Z PATH=/opt/cache/bin:/usr/local/nvidia/bin:/usr/local/cuda/bin:/opt/conda/envs/py_3.12/bin:/opt/conda/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin 2025-11-03T17:07:13.1393454Z GITHUB_BASE_REF= 2025-11-03T17:07:13.1393614Z INSTALLED_ACL= 2025-11-03T17:07:13.1393977Z ARTIFACTS_FILE_SUFFIX=test-vllm_basic_models_test-1-1-linux.g6.4xlarge.experimental.nvidia.gpu_54383051168 2025-11-03T17:07:13.1394396Z CI=true 2025-11-03T17:07:13.1394553Z GITHUB_REPOSITORY_OWNER=pytorch 2025-11-03T17:07:13.1394806Z RUST_LOG=sccache::server=error 2025-11-03T17:07:13.1394999Z JOB_ID=54383051168 2025-11-03T17:07:13.1395165Z GITHUB_HEAD_REF= 2025-11-03T17:07:13.1395321Z GITHUB_ACTION_REF= 2025-11-03T17:07:13.1395530Z SCCACHE_BUCKET=ossci-compiler-cache-circleci-v2 2025-11-03T17:07:13.1395781Z TEST_SHOWLOCALS=False 2025-11-03T17:07:13.1395968Z GITHUB_WORKFLOW=vllm-test 2025-11-03T17:07:13.1396170Z DEBIAN_FRONTEND=noninteractive 2025-11-03T17:07:13.1396627Z GITHUB_OUTPUT=/home/ec2-user/actions-runner/_work/_temp/_runner_file_commands/set_output_edfee93f-f196-45b5-89e0-86041d1a5f42 2025-11-03T17:07:13.1397169Z NO_TD=False 2025-11-03T17:07:13.1397337Z SKIP_SCCACHE_INITIALIZATION=1 2025-11-03T17:07:13.1397551Z NCCL_INCLUDE_DIR=/usr/local/cuda/include/ 2025-11-03T17:07:13.1397766Z _=/usr/bin/env 2025-11-03T17:07:13.1398017Z OLDPWD=/opt/conda/envs/py_3.12/lib/python3.12/site-packages/numba/cuda 2025-11-03T17:07:13.1398385Z ++ python -c 'import site; print(site.getsitepackages()[0])' 2025-11-03T17:07:13.1500665Z + TORCH_INSTALL_DIR=/opt/conda/envs/py_3.12/lib/python3.12/site-packages/torch 2025-11-03T17:07:13.1501716Z + TORCH_BIN_DIR=/opt/conda/envs/py_3.12/lib/python3.12/site-packages/torch/bin 2025-11-03T17:07:13.1502361Z + TORCH_LIB_DIR=/opt/conda/envs/py_3.12/lib/python3.12/site-packages/torch/lib 2025-11-03T17:07:13.1502878Z + TORCH_TEST_DIR=/opt/conda/envs/py_3.12/lib/python3.12/site-packages/torch/test 2025-11-03T17:07:13.1503267Z + BUILD_DIR=build 2025-11-03T17:07:13.1503483Z + BUILD_RENAMED_DIR=build_renamed 2025-11-03T17:07:13.1503765Z + BUILD_BIN_DIR=build/bin 2025-11-03T17:07:13.1503992Z + SHARD_NUMBER=1 2025-11-03T17:07:13.1504198Z + NUM_TEST_SHARDS=1 2025-11-03T17:07:13.1504415Z + export TORCH_SERIALIZATION_DEBUG=1 2025-11-03T17:07:13.1504688Z + TORCH_SERIALIZATION_DEBUG=1 2025-11-03T17:07:13.1504937Z + export VALGRIND=ON 2025-11-03T17:07:13.1505108Z + VALGRIND=ON 2025-11-03T17:07:13.1505306Z + [[ linux-jammy-cuda12.8-py3.12-gcc11 == *clang9* ]] 2025-11-03T17:07:13.1505592Z + [[ linux-jammy-cuda12.8-py3.12-gcc11 == *xpu* ]] 2025-11-03T17:07:13.1505829Z + detect_cuda_arch 2025-11-03T17:07:13.1506026Z + [[ linux-jammy-cuda12.8-py3.12-gcc11 == *cuda* ]] 2025-11-03T17:07:13.1506269Z + command -v nvidia-smi 2025-11-03T17:07:13.1506453Z /usr/bin/nvidia-smi 2025-11-03T17:07:13.1511757Z ++ nvidia-smi --query-gpu=compute_cap --format=csv 2025-11-03T17:07:13.1512890Z ++ tail -n 1 2025-11-03T17:07:13.1760392Z + TORCH_CUDA_ARCH_LIST=8.9 2025-11-03T17:07:13.1760729Z + export TORCH_CUDA_ARCH_LIST 2025-11-03T17:07:13.1761057Z + [[ linux-jammy-cuda12.8-py3.12-gcc11 == *s390x* ]] 2025-11-03T17:07:13.1761384Z + [[ 0 == \1 ]] 2025-11-03T17:07:13.1761606Z + [[ True == \1 ]] 2025-11-03T17:07:13.1761863Z + [[ linux-jammy-cuda12.8-py3.12-gcc11 != *bazel* ]] 2025-11-03T17:07:13.1765299Z ++ realpath build/custom_test_artifacts 2025-11-03T17:07:13.2003526Z + CUSTOM_TEST_ARTIFACT_BUILD_DIR=/var/lib/jenkins/workspace/build/custom_test_artifacts 2025-11-03T17:07:13.2003984Z + [[ -n '' ]] 2025-11-03T17:07:13.2004214Z + echo 'Environment variables' 2025-11-03T17:07:13.2004486Z Environment variables 2025-11-03T17:07:13.2004735Z + env 2025-11-03T17:07:13.2172078Z GITHUB_WORKSPACE=/home/ec2-user/actions-runner/_work/pytorch/pytorch 2025-11-03T17:07:13.2172637Z CONTINUE_THROUGH_ERROR=True 2025-11-03T17:07:13.2173043Z BUILD_ENVIRONMENT=linux-jammy-cuda12.8-py3.12-gcc11 2025-11-03T17:07:13.2173807Z VLLM_TEST_HUGGING_FACE_TOKEN=*** 2025-11-03T17:07:13.2174109Z HOSTNAME=c5e792ba8f7a 2025-11-03T17:07:13.2174536Z GITHUB_PATH=/home/ec2-user/actions-runner/_work/_temp/_runner_file_commands/add_path_edfee93f-f196-45b5-89e0-86041d1a5f42 2025-11-03T17:07:13.2175000Z GITHUB_ACTION=__run_3 2025-11-03T17:07:13.2175190Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=0 2025-11-03T17:07:13.2175513Z GITHUB_RUN_NUMBER=3325 2025-11-03T17:07:13.2175770Z TEST_CONFIG=vllm_basic_models_test 2025-11-03T17:07:13.2175990Z GITHUB_REPOSITORY_OWNER_ID=21003710 2025-11-03T17:07:13.2176220Z TORCH_NVCC_FLAGS=-Xfatbin -compress-all 2025-11-03T17:07:13.2176441Z SCCACHE_IDLE_TIMEOUT=0 2025-11-03T17:07:13.2176735Z SCRIBE_GRAPHQL_ACCESS_TOKEN=*** 2025-11-03T17:07:13.2176966Z GITHUB_TRIGGERING_ACTOR=pytorchmergebot 2025-11-03T17:07:13.2177188Z GITHUB_REF_TYPE=branch 2025-11-03T17:07:13.2177373Z TORCH_CUDA_ARCH_LIST=8.9 2025-11-03T17:07:13.2177601Z BASE_SHA=3f6538febd727b782e6e13cfd026a309fb14351d 2025-11-03T17:07:13.2177843Z XLA_CUDA= 2025-11-03T17:07:13.2178007Z NCCL_LIB_DIR=/usr/local/cuda/lib64/ 2025-11-03T17:07:13.2178404Z HUGGING_FACE_HUB_TOKEN=*** 2025-11-03T17:07:13.2178657Z *** 2025-11-03T17:07:13.2178837Z GITHUB_REPOSITORY_ID=65600975 2025-11-03T17:07:13.2179278Z GITHUB_ACTIONS=true 2025-11-03T17:07:13.2179462Z NVIDIA_DRIVER_CAPABILITIES=all 2025-11-03T17:07:13.2179696Z SCCACHE_ERROR_LOG=/var/lib/jenkins/sccache_error.log 2025-11-03T17:07:13.2179972Z SHA1=3f6538febd727b782e6e13cfd026a309fb14351d 2025-11-03T17:07:13.2180234Z GITHUB_SHA=3f6538febd727b782e6e13cfd026a309fb14351d 2025-11-03T17:07:13.2180587Z GITHUB_WORKFLOW_REF=pytorch/pytorch/.github/workflows/vllm.yml@refs/heads/main 2025-11-03T17:07:13.2181006Z UCC_HOME=/usr 2025-11-03T17:07:13.2181299Z TORCH_SERIALIZATION_DEBUG=1 2025-11-03T17:07:13.2181769Z VERBOSE_TEST_LOGS=False 2025-11-03T17:07:13.2182075Z GITHUB_REF=refs/heads/main 2025-11-03T17:07:13.2182370Z SHARD_NUMBER=1 2025-11-03T17:07:13.2182642Z GITHUB_REF_PROTECTED=true 2025-11-03T17:07:13.2182835Z HOME=/var/lib/jenkins 2025-11-03T17:07:13.2183161Z GITHUB_API_URL=https://api.github.com 2025-11-03T17:07:13.2183453Z PYTORCH_TEST_RERUN_DISABLED_TESTS=0 2025-11-03T17:07:13.2183778Z UCX_COMMIT=7836b165abdbe468a2f607e7254011c07d788152 2025-11-03T17:07:13.2184028Z USE_SYSTEM_NCCL=1 2025-11-03T17:07:13.2184192Z NUM_TEST_SHARDS=1 2025-11-03T17:07:13.2184350Z UCX_HOME=/usr 2025-11-03T17:07:13.2184754Z GITHUB_STATE=/home/ec2-user/actions-runner/_work/_temp/_runner_file_commands/save_state_edfee93f-f196-45b5-89e0-86041d1a5f42 2025-11-03T17:07:13.2185369Z JOB_NAME=ci-vllm-test / test (vllm_basic_models_test, 1, 1, linux.g6.4xlarge.experimental.nvidia.gpu) 2025-11-03T17:07:13.2185983Z GITHUB_ENV=/home/ec2-user/actions-runner/_work/_temp/_runner_file_commands/set_env_edfee93f-f196-45b5-89e0-86041d1a5f42 2025-11-03T17:07:13.2186544Z GITHUB_EVENT_PATH=/home/ec2-user/actions-runner/_work/_temp/_github_workflow/event.json 2025-11-03T17:07:13.2186901Z GITHUB_EVENT_NAME=push 2025-11-03T17:07:13.2187079Z DASHBOARD_TAG= 2025-11-03T17:07:13.2187248Z GITHUB_RUN_ID=19040285420 2025-11-03T17:07:13.2187618Z INSTALLED_OPENBLAS= 2025-11-03T17:07:13.2188056Z GITHUB_STEP_SUMMARY=/home/ec2-user/actions-runner/_work/_temp/_runner_file_commands/step_summary_edfee93f-f196-45b5-89e0-86041d1a5f42 2025-11-03T17:07:13.2188548Z GITHUB_ACTOR=pytorchmergebot 2025-11-03T17:07:13.2188743Z PR_NUMBER= 2025-11-03T17:07:13.2188898Z DESIRED_CUDA=12.8.1 2025-11-03T17:07:13.2189059Z GITHUB_RUN_ATTEMPT=1 2025-11-03T17:07:13.2189244Z VALGRIND=ON 2025-11-03T17:07:13.2189408Z ANACONDA_PYTHON_VERSION=3.12 2025-11-03T17:07:13.2189647Z GITHUB_GRAPHQL_URL=https://api.github.com/graphql 2025-11-03T17:07:13.2189886Z TERM=vt100 2025-11-03T17:07:13.2190043Z INSTALLED_VISION=yes 2025-11-03T17:07:13.2190213Z BRANCH=main 2025-11-03T17:07:13.2190372Z SCCACHE_REGION=us-east-1 2025-11-03T17:07:13.2190564Z OPENSSL_ROOT_DIR=/opt/openssl 2025-11-03T17:07:13.2190766Z BUILD_AOT_INDUCTOR_TEST= 2025-11-03T17:07:13.2190948Z CUDA_PATH=/usr/local/cuda 2025-11-03T17:07:13.2191324Z GITHUB_ACTION_PATH=/home/ec2-user/actions-runner/_work/pytorch/pytorch/./.github/actions/setup-linux 2025-11-03T17:07:13.2191834Z GITHUB_SERVER_URL=https://github.com 2025-11-03T17:07:13.2192309Z UCC_COMMIT=430e241bf5d38cbc73fc7a6b89155397232e3f96 2025-11-03T17:07:13.2192559Z REENABLED_ISSUES= 2025-11-03T17:07:13.2192718Z DOCS= 2025-11-03T17:07:13.2192855Z SHLVL=1 2025-11-03T17:07:13.2192995Z MAX_JOBS=14 2025-11-03T17:07:13.2193154Z GITHUB_ACTOR_ID=97764156 2025-11-03T17:07:13.2193398Z GITHUB_WORKFLOW_SHA=3f6538febd727b782e6e13cfd026a309fb14351d 2025-11-03T17:07:13.2193660Z GITHUB_REF_NAME=main 2025-11-03T17:07:13.2193926Z XLA_CLANG_CACHE_S3_BUCKET_NAME=ossci-compiler-clang-cache-circleci-xla 2025-11-03T17:07:13.2194230Z GITHUB_JOB=test 2025-11-03T17:07:13.2194393Z NO_TEST_TIMEOUT=False 2025-11-03T17:07:13.2194561Z TD_DISTRIBUTED=False 2025-11-03T17:07:13.2194746Z GITHUB_REPOSITORY=pytorch/pytorch 2025-11-03T17:07:13.2194974Z GITHUB_RETENTION_DAYS=90 2025-11-03T17:07:13.2195156Z OPENSSL_DIR=/opt/openssl 2025-11-03T17:07:13.2195342Z GITHUB_ACTION_REPOSITORY= 2025-11-03T17:07:13.2195884Z PATH=/opt/cache/bin:/usr/local/nvidia/bin:/usr/local/cuda/bin:/opt/conda/envs/py_3.12/bin:/opt/conda/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin 2025-11-03T17:07:13.2196560Z GITHUB_BASE_REF= 2025-11-03T17:07:13.2196726Z INSTALLED_ACL= 2025-11-03T17:07:13.2197083Z ARTIFACTS_FILE_SUFFIX=test-vllm_basic_models_test-1-1-linux.g6.4xlarge.experimental.nvidia.gpu_54383051168 2025-11-03T17:07:13.2197486Z CI=true 2025-11-03T17:07:13.2197662Z GITHUB_REPOSITORY_OWNER=pytorch 2025-11-03T17:07:13.2197909Z RUST_LOG=sccache::server=error 2025-11-03T17:07:13.2198093Z JOB_ID=54383051168 2025-11-03T17:07:13.2198256Z GITHUB_HEAD_REF= 2025-11-03T17:07:13.2198421Z GITHUB_ACTION_REF= 2025-11-03T17:07:13.2198627Z SCCACHE_BUCKET=ossci-compiler-cache-circleci-v2 2025-11-03T17:07:13.2198871Z TEST_SHOWLOCALS=False 2025-11-03T17:07:13.2199052Z GITHUB_WORKFLOW=vllm-test 2025-11-03T17:07:13.2199247Z DEBIAN_FRONTEND=noninteractive 2025-11-03T17:07:13.2199701Z GITHUB_OUTPUT=/home/ec2-user/actions-runner/_work/_temp/_runner_file_commands/set_output_edfee93f-f196-45b5-89e0-86041d1a5f42 2025-11-03T17:07:13.2200138Z NO_TD=False 2025-11-03T17:07:13.2200302Z SKIP_SCCACHE_INITIALIZATION=1 2025-11-03T17:07:13.2200528Z NCCL_INCLUDE_DIR=/usr/local/cuda/include/ 2025-11-03T17:07:13.2200842Z OLDPWD=/opt/conda/envs/py_3.12/lib/python3.12/site-packages/numba/cuda 2025-11-03T17:07:13.2201131Z _=/usr/bin/env 2025-11-03T17:07:13.2201301Z + echo 'Testing pytorch' 2025-11-03T17:07:13.2201489Z Testing pytorch 2025-11-03T17:07:13.2201667Z + export LANG=C.UTF-8 2025-11-03T17:07:13.2201839Z + LANG=C.UTF-8 2025-11-03T17:07:13.2202006Z + PR_NUMBER= 2025-11-03T17:07:13.2202190Z + [[ vllm_basic_models_test == \d\e\f\a\u\l\t ]] 2025-11-03T17:07:13.2202458Z + [[ vllm_basic_models_test == \d\i\s\t\r\i\b\u\t\e\d ]] 2025-11-03T17:07:13.2202710Z + [[ vllm_basic_models_test == \s\l\o\w ]] 2025-11-03T17:07:13.2202992Z + [[ linux-jammy-cuda12.8-py3.12-gcc11 == *slow-gradcheck* ]] 2025-11-03T17:07:13.2203389Z + [[ linux-jammy-cuda12.8-py3.12-gcc11 == *cuda* ]] 2025-11-03T17:07:13.2203662Z + export PYTORCH_TESTING_DEVICE_ONLY_FOR=cuda 2025-11-03T17:07:13.2203900Z + PYTORCH_TESTING_DEVICE_ONLY_FOR=cuda 2025-11-03T17:07:13.2204141Z + [[ vllm_basic_models_test == *crossref* ]] 2025-11-03T17:07:13.2204394Z + [[ linux-jammy-cuda12.8-py3.12-gcc11 == *rocm* ]] 2025-11-03T17:07:13.2204662Z + [[ linux-jammy-cuda12.8-py3.12-gcc11 == *xpu* ]] 2025-11-03T17:07:13.2204930Z + [[ linux-jammy-cuda12.8-py3.12-gcc11 != *-bazel-* ]] 2025-11-03T17:07:13.2205176Z + pip_install ninja==1.10.2 2025-11-03T17:07:13.2205429Z + pip_install_pkg='python3 -m pip install --progress-bar off' 2025-11-03T17:07:13.2205865Z + python3 -m pip install --progress-bar off ninja==1.10.2 2025-11-03T17:07:13.6983491Z Collecting ninja==1.10.2 2025-11-03T17:07:13.7150724Z Downloading ninja-1.10.2-py2.py3-none-manylinux_2_5_x86_64.manylinux1_x86_64.whl.metadata (5.0 kB) 2025-11-03T17:07:13.7260909Z Downloading ninja-1.10.2-py2.py3-none-manylinux_2_5_x86_64.manylinux1_x86_64.whl (108 kB) 2025-11-03T17:07:13.9227662Z Installing collected packages: ninja 2025-11-03T17:07:13.9228027Z Attempting uninstall: ninja 2025-11-03T17:07:13.9256851Z Found existing installation: ninja 1.11.1.4 2025-11-03T17:07:13.9281685Z Uninstalling ninja-1.11.1.4: 2025-11-03T17:07:13.9417621Z Successfully uninstalled ninja-1.11.1.4 2025-11-03T17:07:13.9896367Z Successfully installed ninja-1.10.2 2025-11-03T17:07:14.0357474Z + export PATH=/var/lib/jenkins/.local/bin:/opt/cache/bin:/usr/local/nvidia/bin:/usr/local/cuda/bin:/opt/conda/envs/py_3.12/bin:/opt/conda/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin 2025-11-03T17:07:14.0358920Z + PATH=/var/lib/jenkins/.local/bin:/opt/cache/bin:/usr/local/nvidia/bin:/usr/local/cuda/bin:/opt/conda/envs/py_3.12/bin:/opt/conda/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin 2025-11-03T17:07:14.0359774Z + [[ linux-jammy-cuda12.8-py3.12-gcc11 == *aarch64* ]] 2025-11-03T17:07:14.0360161Z + [[ linux-jammy-cuda12.8-py3.12-gcc11 == *asan* ]] 2025-11-03T17:07:14.0360512Z + [[ linux-jammy-cuda12.8-py3.12-gcc11 == *-debug* ]] 2025-11-03T17:07:14.0361055Z + [[ linux-jammy-cuda12.8-py3.12-gcc11 != *-bazel-* ]] 2025-11-03T17:07:14.0362089Z + echo 'We are not in debug mode: linux-jammy-cuda12.8-py3.12-gcc11. Expect the assertion to pass' 2025-11-03T17:07:14.0362711Z We are not in debug mode: linux-jammy-cuda12.8-py3.12-gcc11. Expect the assertion to pass 2025-11-03T17:07:14.0363134Z + cd test 2025-11-03T17:07:14.0363444Z + python -c 'import torch; torch._C._crash_if_debug_asserts_fail(424242)' 2025-11-03T17:07:15.5118739Z + [[ vllm_basic_models_test == \n\o\g\p\u\_\N\O\_\A\V\X\2 ]] 2025-11-03T17:07:15.5119224Z + [[ vllm_basic_models_test == \n\o\g\p\u\_\A\V\X\5\1\2 ]] 2025-11-03T17:07:15.5119651Z + [[ vllm_basic_models_test == \l\e\g\a\c\y\_\n\v\i\d\i\a\_\d\r\i\v\e\r ]] 2025-11-03T17:07:15.5122486Z + DYNAMO_BENCHMARK_FLAGS=() 2025-11-03T17:07:15.5122907Z + [[ vllm_basic_models_test == *pr_time_benchmarks* ]] 2025-11-03T17:07:15.5123306Z + [[ vllm_basic_models_test == *dynamo_eager* ]] 2025-11-03T17:07:15.5123664Z + [[ vllm_basic_models_test == *aot_eager* ]] 2025-11-03T17:07:15.5123969Z + [[ vllm_basic_models_test == *aot_inductor* ]] 2025-11-03T17:07:15.5124321Z + [[ vllm_basic_models_test == *max_autotune_inductor* ]] 2025-11-03T17:07:15.5124651Z + [[ vllm_basic_models_test == *inductor* ]] 2025-11-03T17:07:15.5124944Z + [[ vllm_basic_models_test == *dynamic* ]] 2025-11-03T17:07:15.5125227Z + [[ vllm_basic_models_test == *cpu* ]] 2025-11-03T17:07:15.5125497Z + [[ vllm_basic_models_test == *xpu* ]] 2025-11-03T17:07:15.5125791Z + DYNAMO_BENCHMARK_FLAGS+=(--device cuda) 2025-11-03T17:07:15.5232615Z + [[ linux-jammy-cuda12.8-py3.12-gcc11 == *libtorch* ]] 2025-11-03T17:07:15.5233017Z + [[ linux-jammy-cuda12.8-py3.12-gcc11 == *-bazel-* ]] 2025-11-03T17:07:15.5236383Z + cd test 2025-11-03T17:07:15.5237381Z + python -c 'import torch; print(torch.__config__.show())' 2025-11-03T17:07:17.0131803Z PyTorch built with: 2025-11-03T17:07:17.0132098Z - GCC 11.4 2025-11-03T17:07:17.0132679Z - C++ Version: 201703 2025-11-03T17:07:17.0133250Z - Intel(R) oneAPI Math Kernel Library Version 2024.2-Product Build 20240605 for Intel(R) 64 architecture applications 2025-11-03T17:07:17.0133924Z - Intel(R) MKL-DNN v3.7.1 (Git Hash 8d263e693366ef8db40acc569cc7d8edf644556d) 2025-11-03T17:07:17.0134320Z - OpenMP 201511 (a.k.a. OpenMP 4.5) 2025-11-03T17:07:17.0134628Z - LAPACK is enabled (usually provided by MKL) 2025-11-03T17:07:17.0134921Z - NNPACK is enabled 2025-11-03T17:07:17.0135165Z - CPU capability usage: AVX2 2025-11-03T17:07:17.0135408Z - CUDA Runtime 12.8 2025-11-03T17:07:17.0135946Z - NVCC architecture flags: -gencode;arch=compute_80,code=sm_80;-gencode;arch=compute_89,code=sm_89;-gencode;arch=compute_90,code=sm_90 2025-11-03T17:07:17.0136508Z - CuDNN 90.8 2025-11-03T17:07:17.0140489Z - Build settings: BLAS_INFO=mkl, BUILD_TYPE=Release, COMMIT_SHA=3f6538febd727b782e6e13cfd026a309fb14351d, CUDA_VERSION=12.8, CUDNN_VERSION=9.8.0, CXX_COMPILER=/opt/cache/bin/c++, CXX_FLAGS= -fvisibility-inlines-hidden -DUSE_PTHREADPOOL -DNDEBUG -DUSE_KINETO -DLIBKINETO_NOROCTRACER -DLIBKINETO_NOXPUPTI=ON -DUSE_FBGEMM -DUSE_PYTORCH_QNNPACK -DUSE_XNNPACK -DSYMBOLICATE_MOBILE_DEBUG_HANDLE -O2 -fPIC -DC10_NODEPRECATED -Wall -Wextra -Werror=return-type -Werror=non-virtual-dtor -Werror=range-loop-construct -Werror=bool-operation -Wnarrowing -Wno-missing-field-initializers -Wno-unknown-pragmas -Wno-unused-parameter -Wno-strict-overflow -Wno-strict-aliasing -Wno-stringop-overflow -Wsuggest-override -Wno-psabi -Wno-error=old-style-cast -faligned-new -Werror -Wno-maybe-uninitialized -fno-math-errno -fno-trapping-math -Werror=format -Wno-stringop-overflow, FORCE_FALLBACK_CUDA_MPI=1, LAPACK_INFO=mkl, PERF_WITH_AVX=1, PERF_WITH_AVX2=1, TORCH_VERSION=2.10.0, USE_CUDA=ON, USE_CUDNN=ON, USE_CUSPARSELT=ON, USE_GFLAGS=OFF, USE_GLOG=OFF, USE_GLOO=ON, USE_MKL=ON, USE_MKLDNN=ON, USE_MPI=ON, USE_NCCL=ON, USE_NNPACK=ON, USE_OPENMP=ON, USE_ROCM=OFF, USE_ROCM_KERNEL_ASSERT=OFF, USE_XCCL=OFF, USE_XPU=OFF, 2025-11-03T17:07:17.0144025Z 2025-11-03T17:07:17.2732703Z + cd test 2025-11-03T17:07:17.2733341Z + python -c 'import torch; print(torch.__config__.parallel_info())' 2025-11-03T17:07:18.4736555Z ATen/Parallel: 2025-11-03T17:07:18.4736897Z at::get_num_threads() : 8 2025-11-03T17:07:18.4737178Z at::get_num_interop_threads() : 8 2025-11-03T17:07:18.4737457Z OpenMP 201511 (a.k.a. OpenMP 4.5) 2025-11-03T17:07:18.4737735Z omp_get_max_threads() : 8 2025-11-03T17:07:18.4738257Z Intel(R) oneAPI Math Kernel Library Version 2024.2-Product Build 20240605 for Intel(R) 64 architecture applications 2025-11-03T17:07:18.4738794Z mkl_get_max_threads() : 8 2025-11-03T17:07:18.4739134Z Intel(R) MKL-DNN v3.7.1 (Git Hash 8d263e693366ef8db40acc569cc7d8edf644556d) 2025-11-03T17:07:18.4739530Z std::thread::hardware_concurrency() : 16 2025-11-03T17:07:18.4739810Z Environment variables: 2025-11-03T17:07:18.4740038Z OMP_NUM_THREADS : [not set] 2025-11-03T17:07:18.4740272Z MKL_NUM_THREADS : [not set] 2025-11-03T17:07:18.4740543Z ATen parallel backend: OpenMP 2025-11-03T17:07:18.4740711Z 2025-11-03T17:07:18.6991402Z + [[ vllm_basic_models_test == *numpy_2* ]] 2025-11-03T17:07:18.6991849Z + [[ linux-jammy-cuda12.8-py3.12-gcc11 == *aarch64* ]] 2025-11-03T17:07:18.6992239Z + [[ vllm_basic_models_test == *backward* ]] 2025-11-03T17:07:18.6992536Z + [[ vllm_basic_models_test == *xla* ]] 2025-11-03T17:07:18.6992813Z + [[ vllm_basic_models_test == *vllm* ]] 2025-11-03T17:07:18.6993127Z + echo 'vLLM CI uses TORCH_CUDA_ARCH_LIST: 8.9' 2025-11-03T17:07:18.6993437Z vLLM CI uses TORCH_CUDA_ARCH_LIST: 8.9 2025-11-03T17:07:18.6995445Z + cd .ci/lumen_cli 2025-11-03T17:07:18.6995705Z + python -m pip install -e . 2025-11-03T17:07:18.9852909Z Obtaining file:///var/lib/jenkins/workspace/.ci/lumen_cli 2025-11-03T17:07:19.7204620Z Installing build dependencies ... [?25l- \ done 2025-11-03T17:07:19.9314076Z [?25h Checking if build backend supports build_editable ... [?25ldone 2025-11-03T17:07:20.1916057Z [?25h Getting requirements to build editable ... [?25l- done 2025-11-03T17:07:20.3530435Z [?25h Preparing editable metadata (pyproject.toml) ... [?25l- done 2025-11-03T17:07:20.4107222Z [?25hCollecting pyyaml==6.0.2 (from lumen-ci==0.1.0) 2025-11-03T17:07:20.4277385Z Downloading PyYAML-6.0.2-cp312-cp312-manylinux_2_17_x86_64.manylinux2014_x86_64.whl.metadata (2.1 kB) 2025-11-03T17:07:20.4568565Z Collecting GitPython==3.1.45 (from lumen-ci==0.1.0) 2025-11-03T17:07:20.4600713Z Downloading gitpython-3.1.45-py3-none-any.whl.metadata (13 kB) 2025-11-03T17:07:20.4861338Z Collecting docker==7.1.0 (from lumen-ci==0.1.0) 2025-11-03T17:07:20.4915662Z Downloading docker-7.1.0-py3-none-any.whl.metadata (3.8 kB) 2025-11-03T17:07:20.4995554Z Requirement already satisfied: pytest==7.3.2 in /opt/conda/envs/py_3.12/lib/python3.12/site-packages (from lumen-ci==0.1.0) (7.3.2) 2025-11-03T17:07:20.6855135Z Collecting uv==0.9.6 (from lumen-ci==0.1.0) 2025-11-03T17:07:20.6905222Z Downloading uv-0.9.6-py3-none-manylinux_2_17_x86_64.manylinux2014_x86_64.whl.metadata (11 kB) 2025-11-03T17:07:20.6990176Z Requirement already satisfied: requests>=2.26.0 in /opt/conda/envs/py_3.12/lib/python3.12/site-packages (from docker==7.1.0->lumen-ci==0.1.0) (2.32.5) 2025-11-03T17:07:20.6993107Z Requirement already satisfied: urllib3>=1.26.0 in /opt/conda/envs/py_3.12/lib/python3.12/site-packages (from docker==7.1.0->lumen-ci==0.1.0) (2.5.0) 2025-11-03T17:07:20.7115979Z Collecting gitdb<5,>=4.0.1 (from GitPython==3.1.45->lumen-ci==0.1.0) 2025-11-03T17:07:20.7152951Z Downloading gitdb-4.0.12-py3-none-any.whl.metadata (1.2 kB) 2025-11-03T17:07:20.7232406Z Requirement already satisfied: iniconfig in /opt/conda/envs/py_3.12/lib/python3.12/site-packages (from pytest==7.3.2->lumen-ci==0.1.0) (2.3.0) 2025-11-03T17:07:20.7235179Z Requirement already satisfied: packaging in /opt/conda/envs/py_3.12/lib/python3.12/site-packages (from pytest==7.3.2->lumen-ci==0.1.0) (23.1) 2025-11-03T17:07:20.7239889Z Requirement already satisfied: pluggy<2.0,>=0.12 in /opt/conda/envs/py_3.12/lib/python3.12/site-packages (from pytest==7.3.2->lumen-ci==0.1.0) (1.6.0) 2025-11-03T17:07:20.7377225Z Collecting smmap<6,>=3.0.1 (from gitdb<5,>=4.0.1->GitPython==3.1.45->lumen-ci==0.1.0) 2025-11-03T17:07:20.7415368Z Downloading smmap-5.0.2-py3-none-any.whl.metadata (4.3 kB) 2025-11-03T17:07:20.7503687Z Requirement already satisfied: charset_normalizer<4,>=2 in /opt/conda/envs/py_3.12/lib/python3.12/site-packages (from requests>=2.26.0->docker==7.1.0->lumen-ci==0.1.0) (3.4.4) 2025-11-03T17:07:20.7508048Z Requirement already satisfied: idna<4,>=2.5 in /opt/conda/envs/py_3.12/lib/python3.12/site-packages (from requests>=2.26.0->docker==7.1.0->lumen-ci==0.1.0) (3.11) 2025-11-03T17:07:20.7513508Z Requirement already satisfied: certifi>=2017.4.17 in /opt/conda/envs/py_3.12/lib/python3.12/site-packages (from requests>=2.26.0->docker==7.1.0->lumen-ci==0.1.0) (2025.10.5) 2025-11-03T17:07:20.7626319Z Downloading docker-7.1.0-py3-none-any.whl (147 kB) 2025-11-03T17:07:20.7755163Z Downloading gitpython-3.1.45-py3-none-any.whl (208 kB) 2025-11-03T17:07:20.7885023Z Downloading PyYAML-6.0.2-cp312-cp312-manylinux_2_17_x86_64.manylinux2014_x86_64.whl (767 kB) 2025-11-03T17:07:20.8013553Z [?25l ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 0.0/767.5 kB ? eta -:--:-- 2025-11-03T17:07:20.8014270Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 767.5/767.5 kB 51.9 MB/s 0:00:00 2025-11-03T17:07:20.8054855Z [?25hDownloading uv-0.9.6-py3-none-manylinux_2_17_x86_64.manylinux2014_x86_64.whl (21.4 MB) 2025-11-03T17:07:21.0079514Z [?25l ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 0.0/21.4 MB ? eta -:--:-- 2025-11-03T17:07:21.0123532Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╸ 21.2/21.4 MB 134.4 MB/s eta 0:00:01 2025-11-03T17:07:21.0124199Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 21.4/21.4 MB 103.5 MB/s 0:00:00 2025-11-03T17:07:21.0162793Z [?25hDownloading gitdb-4.0.12-py3-none-any.whl (62 kB) 2025-11-03T17:07:21.0236824Z Downloading smmap-5.0.2-py3-none-any.whl (24 kB) 2025-11-03T17:07:21.0480306Z Building wheels for collected packages: lumen-ci 2025-11-03T17:07:21.2277363Z Building editable for lumen-ci (pyproject.toml) ... [?25l- done 2025-11-03T17:07:21.2283652Z [?25h Created wheel for lumen-ci: filename=lumen_ci-0.1.0-0.editable-py3-none-any.whl size=2707 sha256=7a42ce4f8904a49b64161b63d5e83cc61fce27b4997d8277ff99ec9687488cc1 2025-11-03T17:07:21.2284839Z Stored in directory: /tmp/pip-ephem-wheel-cache-v4f2_xcb/wheels/6d/29/6c/82214c85262c5bdb1d615291039f4ec9f974034103e7a303fa 2025-11-03T17:07:21.2298411Z Successfully built lumen-ci 2025-11-03T17:07:21.3939095Z Installing collected packages: uv, smmap, pyyaml, gitdb, docker, GitPython, lumen-ci 2025-11-03T17:07:21.5620029Z [?25l 2025-11-03T17:07:21.6367687Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 0/7 [uv] 2025-11-03T17:07:21.6368160Z  Attempting uninstall: pyyaml 2025-11-03T17:07:21.6419364Z ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 0/7 [uv] 2025-11-03T17:07:21.6419960Z  Found existing installation: PyYAML 6.0.3 2025-11-03T17:07:21.6454742Z ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 0/7 [uv] 2025-11-03T17:07:21.6455316Z  Uninstalling PyYAML-6.0.3: 2025-11-03T17:07:21.6597391Z ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 0/7 [uv] 2025-11-03T17:07:21.6597956Z  Successfully uninstalled PyYAML-6.0.3 2025-11-03T17:07:21.7314196Z ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 0/7 [uv] 2025-11-03T17:07:21.9014318Z  ━━━━━━━━━━━━━━━━━╺━━━━━━━━━━━━━━━━━━━━━━ 3/7 [gitdb] 2025-11-03T17:07:21.9112166Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━╸━━━━━━━━━━━ 5/7 [GitPython] 2025-11-03T17:07:21.9112756Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 7/7 [lumen-ci] 2025-11-03T17:07:21.9113161Z [?25h 2025-11-03T17:07:21.9209048Z Successfully installed GitPython-3.1.45 docker-7.1.0 gitdb-4.0.12 lumen-ci-0.1.0 pyyaml-6.0.2 smmap-5.0.2 uv-0.9.6 2025-11-03T17:07:22.0080117Z + python -m cli.run test external vllm --test-plan vllm_basic_models_test --shard-id 1 --num-shards 1 2025-11-03T17:07:22.2055083Z 2025-11-03 17:07:22,204 [INFO] cli.lib.core.vllm.vllm_test: Display VllmTestParameters VllmTestParameters(torch_whls_path=PosixPath('/var/lib/jenkins/workspace/dist'), vllm_whls_path=PosixPath('/var/lib/jenkins/workspace/dist/external/vllm/wheels'), torch_cuda_arch_list='8.9', cleaning_script=PosixPath('/var/lib/jenkins/workspace/.github/ci_configs/vllm/use_existing_torch.py')) 2025-11-03T17:07:22.2056761Z 2025-11-03 17:07:22,205 [INFO] cli.lib.common.git_helper: Cloning vllm to vllm 2025-11-03T17:07:22.4212134Z 2025-11-03 17:07:22,420 [INFO] cli.lib.common.git_helper: Progress: 5% - remote: Counting objects: 5% (2/39) 2025-11-03T17:07:22.4213649Z 2025-11-03 17:07:22,420 [INFO] cli.lib.common.git_helper: Progress: 10% - remote: Counting objects: 10% (4/39) 2025-11-03T17:07:22.4214325Z 2025-11-03 17:07:22,420 [INFO] cli.lib.common.git_helper: Progress: 15% - remote: Counting objects: 15% (6/39) 2025-11-03T17:07:22.4215144Z 2025-11-03 17:07:22,420 [INFO] cli.lib.common.git_helper: Progress: 20% - remote: Counting objects: 20% (8/39) 2025-11-03T17:07:22.4215780Z 2025-11-03 17:07:22,420 [INFO] cli.lib.common.git_helper: Progress: 25% - remote: Counting objects: 25% (10/39) 2025-11-03T17:07:22.4216302Z 2025-11-03 17:07:22,421 [INFO] cli.lib.common.git_helper: Progress: 30% - remote: Counting objects: 30% (12/39) 2025-11-03T17:07:22.4216986Z 2025-11-03 17:07:22,421 [INFO] cli.lib.common.git_helper: Progress: 35% - remote: Counting objects: 35% (14/39) 2025-11-03T17:07:22.4217714Z 2025-11-03 17:07:22,421 [INFO] cli.lib.common.git_helper: Progress: 100% - remote: Counting objects: 100% (39/39) 2025-11-03T17:07:22.4218363Z 2025-11-03 17:07:22,421 [INFO] cli.lib.common.git_helper: Progress: 5% - remote: Compressing objects: 5% (2/39) 2025-11-03T17:07:22.4235977Z 2025-11-03 17:07:22,423 [INFO] cli.lib.common.git_helper: Progress: 10% - remote: Compressing objects: 10% (4/39) 2025-11-03T17:07:22.4240980Z 2025-11-03 17:07:22,423 [INFO] cli.lib.common.git_helper: Progress: 15% - remote: Compressing objects: 15% (6/39) 2025-11-03T17:07:22.4254137Z 2025-11-03 17:07:22,425 [INFO] cli.lib.common.git_helper: Progress: 20% - remote: Compressing objects: 20% (8/39) 2025-11-03T17:07:22.4254965Z 2025-11-03 17:07:22,425 [INFO] cli.lib.common.git_helper: Progress: 25% - remote: Compressing objects: 25% (10/39) 2025-11-03T17:07:22.4256264Z 2025-11-03 17:07:22,425 [INFO] cli.lib.common.git_helper: Progress: 30% - remote: Compressing objects: 30% (12/39) 2025-11-03T17:07:22.4256939Z 2025-11-03 17:07:22,425 [INFO] cli.lib.common.git_helper: Progress: 35% - remote: Compressing objects: 35% (14/39) 2025-11-03T17:07:22.4267583Z 2025-11-03 17:07:22,426 [INFO] cli.lib.common.git_helper: Progress: 100% - remote: Compressing objects: 100% (39/39) 2025-11-03T17:07:22.4481020Z 2025-11-03 17:07:22,447 [INFO] cli.lib.common.git_helper: Progress: 0% - Receiving objects: 0% (1/136572) 2025-11-03T17:07:22.5460609Z 2025-11-03 17:07:22,545 [INFO] cli.lib.common.git_helper: Progress: 5% - Receiving objects: 5% (6829/136572) 2025-11-03T17:07:22.7800801Z 2025-11-03 17:07:22,779 [INFO] cli.lib.common.git_helper: Progress: 10% - Receiving objects: 10% (13658/136572) 2025-11-03T17:07:22.9860473Z 2025-11-03 17:07:22,985 [INFO] cli.lib.common.git_helper: Progress: 15% - Receiving objects: 15% (20486/136572), 27.89 MiB | 55.77 MiB/s 2025-11-03T17:07:23.1318250Z 2025-11-03 17:07:23,131 [INFO] cli.lib.common.git_helper: Progress: 20% - Receiving objects: 20% (27315/136572), 27.89 MiB | 55.77 MiB/s 2025-11-03T17:07:23.2509054Z 2025-11-03 17:07:23,250 [INFO] cli.lib.common.git_helper: Progress: 25% - Receiving objects: 25% (34143/136572), 27.89 MiB | 55.77 MiB/s 2025-11-03T17:07:23.3850430Z 2025-11-03 17:07:23,384 [INFO] cli.lib.common.git_helper: Progress: 30% - Receiving objects: 30% (40972/136572), 27.89 MiB | 55.77 MiB/s 2025-11-03T17:07:23.5123656Z 2025-11-03 17:07:23,512 [INFO] cli.lib.common.git_helper: Progress: 35% - Receiving objects: 35% (47801/136572), 55.91 MiB | 55.90 MiB/s 2025-11-03T17:07:23.6272863Z 2025-11-03 17:07:23,626 [INFO] cli.lib.common.git_helper: Progress: 40% - Receiving objects: 40% (54629/136572), 55.91 MiB | 55.90 MiB/s 2025-11-03T17:07:23.7245626Z 2025-11-03 17:07:23,724 [INFO] cli.lib.common.git_helper: Progress: 45% - Receiving objects: 45% (61458/136572), 55.91 MiB | 55.90 MiB/s 2025-11-03T17:07:23.8172631Z 2025-11-03 17:07:23,816 [INFO] cli.lib.common.git_helper: Progress: 50% - Receiving objects: 50% (68286/136572), 55.91 MiB | 55.90 MiB/s 2025-11-03T17:07:23.8608980Z 2025-11-03 17:07:23,860 [INFO] cli.lib.common.git_helper: Progress: 55% - Receiving objects: 55% (75115/136572), 55.91 MiB | 55.90 MiB/s 2025-11-03T17:07:23.9061818Z 2025-11-03 17:07:23,905 [INFO] cli.lib.common.git_helper: Progress: 60% - Receiving objects: 60% (81944/136572), 55.91 MiB | 55.90 MiB/s 2025-11-03T17:07:23.9415942Z 2025-11-03 17:07:23,941 [INFO] cli.lib.common.git_helper: Progress: 65% - Receiving objects: 65% (88772/136572), 86.06 MiB | 57.37 MiB/s 2025-11-03T17:07:23.9819509Z 2025-11-03 17:07:23,981 [INFO] cli.lib.common.git_helper: Progress: 70% - Receiving objects: 70% (95601/136572), 86.06 MiB | 57.37 MiB/s 2025-11-03T17:07:24.0419162Z 2025-11-03 17:07:24,041 [INFO] cli.lib.common.git_helper: Progress: 75% - Receiving objects: 75% (102429/136572), 86.06 MiB | 57.37 MiB/s 2025-11-03T17:07:24.1040906Z 2025-11-03 17:07:24,103 [INFO] cli.lib.common.git_helper: Progress: 80% - Receiving objects: 80% (109258/136572), 86.06 MiB | 57.37 MiB/s 2025-11-03T17:07:24.1656850Z 2025-11-03 17:07:24,165 [INFO] cli.lib.common.git_helper: Progress: 85% - Receiving objects: 85% (116087/136572), 86.06 MiB | 57.37 MiB/s 2025-11-03T17:07:24.1982134Z 2025-11-03 17:07:24,197 [INFO] cli.lib.common.git_helper: Progress: 90% - Receiving objects: 90% (122915/136572), 86.06 MiB | 57.37 MiB/s 2025-11-03T17:07:24.2456626Z 2025-11-03 17:07:24,245 [INFO] cli.lib.common.git_helper: Progress: 95% - Receiving objects: 95% (129744/136572), 86.06 MiB | 57.37 MiB/s 2025-11-03T17:07:24.2790596Z 2025-11-03 17:07:24,278 [INFO] cli.lib.common.git_helper: Progress: 100% - Receiving objects: 100% (136572/136572), 86.06 MiB | 57.37 MiB/s 2025-11-03T17:07:24.2966877Z 2025-11-03 17:07:24,296 [INFO] cli.lib.common.git_helper: Resolving deltas: 0% (0/108697) 2025-11-03T17:07:24.3293469Z 2025-11-03 17:07:24,328 [INFO] cli.lib.common.git_helper: Progress: 5% - Resolving deltas: 5% (5435/108697) 2025-11-03T17:07:24.3812305Z 2025-11-03 17:07:24,380 [INFO] cli.lib.common.git_helper: Progress: 10% - Resolving deltas: 10% (10870/108697) 2025-11-03T17:07:24.4319006Z 2025-11-03 17:07:24,431 [INFO] cli.lib.common.git_helper: Progress: 15% - Resolving deltas: 15% (16305/108697) 2025-11-03T17:07:24.4704975Z 2025-11-03 17:07:24,470 [INFO] cli.lib.common.git_helper: Progress: 20% - Resolving deltas: 20% (21740/108697) 2025-11-03T17:07:24.5183395Z 2025-11-03 17:07:24,518 [INFO] cli.lib.common.git_helper: Progress: 25% - Resolving deltas: 25% (27175/108697) 2025-11-03T17:07:24.5657038Z 2025-11-03 17:07:24,565 [INFO] cli.lib.common.git_helper: Progress: 30% - Resolving deltas: 30% (32610/108697) 2025-11-03T17:07:24.6038708Z 2025-11-03 17:07:24,603 [INFO] cli.lib.common.git_helper: Progress: 35% - Resolving deltas: 35% (38044/108697) 2025-11-03T17:07:24.6348259Z 2025-11-03 17:07:24,634 [INFO] cli.lib.common.git_helper: Progress: 40% - Resolving deltas: 40% (43479/108697) 2025-11-03T17:07:24.6692501Z 2025-11-03 17:07:24,668 [INFO] cli.lib.common.git_helper: Progress: 45% - Resolving deltas: 45% (48914/108697) 2025-11-03T17:07:24.7009296Z 2025-11-03 17:07:24,700 [INFO] cli.lib.common.git_helper: Progress: 50% - Resolving deltas: 50% (54349/108697) 2025-11-03T17:07:24.7335397Z 2025-11-03 17:07:24,733 [INFO] cli.lib.common.git_helper: Progress: 55% - Resolving deltas: 55% (59784/108697) 2025-11-03T17:07:24.7577406Z 2025-11-03 17:07:24,757 [INFO] cli.lib.common.git_helper: Progress: 60% - Resolving deltas: 60% (65219/108697) 2025-11-03T17:07:24.7762023Z 2025-11-03 17:07:24,775 [INFO] cli.lib.common.git_helper: Progress: 65% - Resolving deltas: 65% (70654/108697) 2025-11-03T17:07:24.8002431Z 2025-11-03 17:07:24,799 [INFO] cli.lib.common.git_helper: Progress: 70% - Resolving deltas: 70% (76088/108697) 2025-11-03T17:07:24.8303980Z 2025-11-03 17:07:24,830 [INFO] cli.lib.common.git_helper: Progress: 75% - Resolving deltas: 75% (81523/108697) 2025-11-03T17:07:24.8527115Z 2025-11-03 17:07:24,852 [INFO] cli.lib.common.git_helper: Progress: 80% - Resolving deltas: 80% (86958/108697) 2025-11-03T17:07:24.8749964Z 2025-11-03 17:07:24,874 [INFO] cli.lib.common.git_helper: Progress: 85% - Resolving deltas: 85% (92393/108697) 2025-11-03T17:07:24.9007179Z 2025-11-03 17:07:24,900 [INFO] cli.lib.common.git_helper: Progress: 90% - Resolving deltas: 90% (97828/108697) 2025-11-03T17:07:24.9193172Z 2025-11-03 17:07:24,919 [INFO] cli.lib.common.git_helper: Progress: 95% - Resolving deltas: 95% (103263/108697) 2025-11-03T17:07:24.9314586Z 2025-11-03 17:07:24,931 [INFO] cli.lib.common.git_helper: Progress: 100% - Resolving deltas: 100% (108697/108697) 2025-11-03T17:07:25.7108035Z 2025-11-03 17:07:25,710 [INFO] cli.lib.common.git_helper: Checking out pinned vllm commit e5192819208c4d68194844b7dfafbc00020d0dea 2025-11-03T17:07:26.0654323Z 2025-11-03 17:07:26,064 [INFO] cli.lib.common.git_helper: Successfully cloned vllm 2025-11-03T17:07:26.1032862Z 2025-11-03 17:07:26,102 [INFO] cli.lib.core.vllm.vllm_test: Running vllm test with inputs: VllmTestParameters(torch_whls_path=PosixPath('/var/lib/jenkins/workspace/dist'), vllm_whls_path=PosixPath('/var/lib/jenkins/workspace/dist/external/vllm/wheels'), torch_cuda_arch_list='8.9', cleaning_script=PosixPath('/var/lib/jenkins/workspace/.github/ci_configs/vllm/use_existing_torch.py')) 2025-11-03T17:07:26.1054597Z 2025-11-03 17:07:26,105 [INFO] cli.lib.common.pip_helper: torch already exist with version: 2.10.0a0+git3f6538f 2025-11-03T17:07:26.1057085Z 2025-11-03 17:07:26,105 [INFO] cli.lib.common.pip_helper: Installing /var/lib/jenkins/workspace/dist/vision/torchvision-0.25.0a0+cfbc5c2-cp312-cp312-linux_x86_64.whl... 2025-11-03T17:07:26.1058648Z 2025-11-03 17:07:26,105 [INFO] cli.lib.common.pip_helper: pip installing packages: /opt/conda/envs/py_3.12/bin/python -m pip install /var/lib/jenkins/workspace/dist/vision/torchvision-0.25.0a0+cfbc5c2-cp312-cp312-linux_x86_64.whl 2025-11-03T17:07:26.1060174Z 2025-11-03 17:07:26,105 [INFO] cli.lib.common.utils: [cmd] /opt/conda/envs/py_3.12/bin/python -m pip install /var/lib/jenkins/workspace/dist/vision/torchvision-0.25.0a0+cfbc5c2-cp312-cp312-linux_x86_64.whl 2025-11-03T17:07:26.3964709Z Processing /var/lib/jenkins/workspace/dist/vision/torchvision-0.25.0a0+cfbc5c2-cp312-cp312-linux_x86_64.whl 2025-11-03T17:07:26.4040330Z Requirement already satisfied: numpy in /opt/conda/envs/py_3.12/lib/python3.12/site-packages (from torchvision==0.25.0a0+cfbc5c2) (1.26.2) 2025-11-03T17:07:26.4043043Z Requirement already satisfied: torch in /opt/conda/envs/py_3.12/lib/python3.12/site-packages (from torchvision==0.25.0a0+cfbc5c2) (2.10.0a0+git3f6538f) 2025-11-03T17:07:26.4046546Z Requirement already satisfied: pillow!=8.3.*,>=5.3.0 in /opt/conda/envs/py_3.12/lib/python3.12/site-packages (from torchvision==0.25.0a0+cfbc5c2) (11.0.0) 2025-11-03T17:07:26.4098745Z Requirement already satisfied: filelock in /opt/conda/envs/py_3.12/lib/python3.12/site-packages (from torch->torchvision==0.25.0a0+cfbc5c2) (3.18.0) 2025-11-03T17:07:26.4102820Z Requirement already satisfied: typing-extensions>=4.10.0 in /opt/conda/envs/py_3.12/lib/python3.12/site-packages (from torch->torchvision==0.25.0a0+cfbc5c2) (4.12.2) 2025-11-03T17:07:26.4107576Z Requirement already satisfied: setuptools in /opt/conda/envs/py_3.12/lib/python3.12/site-packages (from torch->torchvision==0.25.0a0+cfbc5c2) (78.1.1) 2025-11-03T17:07:26.4111398Z Requirement already satisfied: sympy>=1.13.3 in /opt/conda/envs/py_3.12/lib/python3.12/site-packages (from torch->torchvision==0.25.0a0+cfbc5c2) (1.13.3) 2025-11-03T17:07:26.4119611Z Requirement already satisfied: networkx>=2.5.1 in /opt/conda/envs/py_3.12/lib/python3.12/site-packages (from torch->torchvision==0.25.0a0+cfbc5c2) (2.8.8) 2025-11-03T17:07:26.4122955Z Requirement already satisfied: jinja2 in /opt/conda/envs/py_3.12/lib/python3.12/site-packages (from torch->torchvision==0.25.0a0+cfbc5c2) (3.1.6) 2025-11-03T17:07:26.4126783Z Requirement already satisfied: fsspec>=0.8.5 in /opt/conda/envs/py_3.12/lib/python3.12/site-packages (from torch->torchvision==0.25.0a0+cfbc5c2) (2025.9.0) 2025-11-03T17:07:26.4230829Z Requirement already satisfied: mpmath<1.4,>=1.1.0 in /opt/conda/envs/py_3.12/lib/python3.12/site-packages (from sympy>=1.13.3->torch->torchvision==0.25.0a0+cfbc5c2) (1.3.0) 2025-11-03T17:07:26.4272166Z Requirement already satisfied: MarkupSafe>=2.0 in /opt/conda/envs/py_3.12/lib/python3.12/site-packages (from jinja2->torch->torchvision==0.25.0a0+cfbc5c2) (3.0.3) 2025-11-03T17:07:26.5917897Z Installing collected packages: torchvision 2025-11-03T17:07:27.0262627Z Successfully installed torchvision-0.25.0a0+cfbc5c2 2025-11-03T17:07:27.0727473Z 2025-11-03 17:07:27,072 [INFO] cli.lib.common.pip_helper: Installing /var/lib/jenkins/workspace/dist/audio/torchaudio-2.10.0a0+3b0e7a6-cp312-cp312-linux_x86_64.whl... 2025-11-03T17:07:27.0728930Z 2025-11-03 17:07:27,072 [INFO] cli.lib.common.pip_helper: pip installing packages: /opt/conda/envs/py_3.12/bin/python -m pip install /var/lib/jenkins/workspace/dist/audio/torchaudio-2.10.0a0+3b0e7a6-cp312-cp312-linux_x86_64.whl 2025-11-03T17:07:27.0730427Z 2025-11-03 17:07:27,072 [INFO] cli.lib.common.utils: [cmd] /opt/conda/envs/py_3.12/bin/python -m pip install /var/lib/jenkins/workspace/dist/audio/torchaudio-2.10.0a0+3b0e7a6-cp312-cp312-linux_x86_64.whl 2025-11-03T17:07:27.3645725Z Processing /var/lib/jenkins/workspace/dist/audio/torchaudio-2.10.0a0+3b0e7a6-cp312-cp312-linux_x86_64.whl 2025-11-03T17:07:27.3701774Z Requirement already satisfied: torch in /opt/conda/envs/py_3.12/lib/python3.12/site-packages (from torchaudio==2.10.0a0+3b0e7a6) (2.10.0a0+git3f6538f) 2025-11-03T17:07:27.3711035Z Requirement already satisfied: filelock in /opt/conda/envs/py_3.12/lib/python3.12/site-packages (from torch->torchaudio==2.10.0a0+3b0e7a6) (3.18.0) 2025-11-03T17:07:27.3715586Z Requirement already satisfied: typing-extensions>=4.10.0 in /opt/conda/envs/py_3.12/lib/python3.12/site-packages (from torch->torchaudio==2.10.0a0+3b0e7a6) (4.12.2) 2025-11-03T17:07:27.3721308Z Requirement already satisfied: setuptools in /opt/conda/envs/py_3.12/lib/python3.12/site-packages (from torch->torchaudio==2.10.0a0+3b0e7a6) (78.1.1) 2025-11-03T17:07:27.3725895Z Requirement already satisfied: sympy>=1.13.3 in /opt/conda/envs/py_3.12/lib/python3.12/site-packages (from torch->torchaudio==2.10.0a0+3b0e7a6) (1.13.3) 2025-11-03T17:07:27.3729773Z Requirement already satisfied: networkx>=2.5.1 in /opt/conda/envs/py_3.12/lib/python3.12/site-packages (from torch->torchaudio==2.10.0a0+3b0e7a6) (2.8.8) 2025-11-03T17:07:27.3732983Z Requirement already satisfied: jinja2 in /opt/conda/envs/py_3.12/lib/python3.12/site-packages (from torch->torchaudio==2.10.0a0+3b0e7a6) (3.1.6) 2025-11-03T17:07:27.3737285Z Requirement already satisfied: fsspec>=0.8.5 in /opt/conda/envs/py_3.12/lib/python3.12/site-packages (from torch->torchaudio==2.10.0a0+3b0e7a6) (2025.9.0) 2025-11-03T17:07:27.3844537Z Requirement already satisfied: mpmath<1.4,>=1.1.0 in /opt/conda/envs/py_3.12/lib/python3.12/site-packages (from sympy>=1.13.3->torch->torchaudio==2.10.0a0+3b0e7a6) (1.3.0) 2025-11-03T17:07:27.3885027Z Requirement already satisfied: MarkupSafe>=2.0 in /opt/conda/envs/py_3.12/lib/python3.12/site-packages (from jinja2->torch->torchaudio==2.10.0a0+3b0e7a6) (3.0.3) 2025-11-03T17:07:27.5758635Z Installing collected packages: torchaudio 2025-11-03T17:07:27.7260555Z Successfully installed torchaudio-2.10.0a0+3b0e7a6 2025-11-03T17:07:27.7673423Z 2025-11-03 17:07:27,766 [INFO] cli.lib.core.vllm.vllm_test: Done. Installed torch and other torch-related wheels 2025-11-03T17:07:27.7674051Z 2025-11-03 17:07:27,767 [INFO] cli.lib.core.vllm.vllm_test: Installing vllm wheels 2025-11-03T17:07:27.7676987Z 2025-11-03 17:07:27,767 [INFO] cli.lib.common.pip_helper: Installing /var/lib/jenkins/workspace/dist/external/vllm/wheels/xformers/xformers-0.0.33+5d4b92a5.d20251103-cp39-abi3-linux_x86_64.whl... 2025-11-03T17:07:27.7678380Z 2025-11-03 17:07:27,767 [INFO] cli.lib.common.pip_helper: pip installing packages: /opt/conda/envs/py_3.12/bin/python -m pip install /var/lib/jenkins/workspace/dist/external/vllm/wheels/xformers/xformers-0.0.33+5d4b92a5.d20251103-cp39-abi3-linux_x86_64.whl 2025-11-03T17:07:27.7679757Z 2025-11-03 17:07:27,767 [INFO] cli.lib.common.utils: [cmd] /opt/conda/envs/py_3.12/bin/python -m pip install /var/lib/jenkins/workspace/dist/external/vllm/wheels/xformers/xformers-0.0.33+5d4b92a5.d20251103-cp39-abi3-linux_x86_64.whl 2025-11-03T17:07:28.0602061Z Processing /var/lib/jenkins/workspace/dist/external/vllm/wheels/xformers/xformers-0.0.33+5d4b92a5.d20251103-cp39-abi3-linux_x86_64.whl 2025-11-03T17:07:28.2498314Z Requirement already satisfied: torch>=2.8 in /opt/conda/envs/py_3.12/lib/python3.12/site-packages (from xformers==0.0.33+5d4b92a5.d20251103) (2.10.0a0+git3f6538f) 2025-11-03T17:07:28.2499500Z Requirement already satisfied: numpy in /opt/conda/envs/py_3.12/lib/python3.12/site-packages (from xformers==0.0.33+5d4b92a5.d20251103) (1.26.2) 2025-11-03T17:07:28.2509905Z Requirement already satisfied: filelock in /opt/conda/envs/py_3.12/lib/python3.12/site-packages (from torch>=2.8->xformers==0.0.33+5d4b92a5.d20251103) (3.18.0) 2025-11-03T17:07:28.2513915Z Requirement already satisfied: typing-extensions>=4.10.0 in /opt/conda/envs/py_3.12/lib/python3.12/site-packages (from torch>=2.8->xformers==0.0.33+5d4b92a5.d20251103) (4.12.2) 2025-11-03T17:07:28.2519410Z Requirement already satisfied: setuptools in /opt/conda/envs/py_3.12/lib/python3.12/site-packages (from torch>=2.8->xformers==0.0.33+5d4b92a5.d20251103) (78.1.1) 2025-11-03T17:07:28.2523664Z Requirement already satisfied: sympy>=1.13.3 in /opt/conda/envs/py_3.12/lib/python3.12/site-packages (from torch>=2.8->xformers==0.0.33+5d4b92a5.d20251103) (1.13.3) 2025-11-03T17:07:28.2528290Z Requirement already satisfied: networkx>=2.5.1 in /opt/conda/envs/py_3.12/lib/python3.12/site-packages (from torch>=2.8->xformers==0.0.33+5d4b92a5.d20251103) (2.8.8) 2025-11-03T17:07:28.2531422Z Requirement already satisfied: jinja2 in /opt/conda/envs/py_3.12/lib/python3.12/site-packages (from torch>=2.8->xformers==0.0.33+5d4b92a5.d20251103) (3.1.6) 2025-11-03T17:07:28.2535406Z Requirement already satisfied: fsspec>=0.8.5 in /opt/conda/envs/py_3.12/lib/python3.12/site-packages (from torch>=2.8->xformers==0.0.33+5d4b92a5.d20251103) (2025.9.0) 2025-11-03T17:07:28.2644071Z Requirement already satisfied: mpmath<1.4,>=1.1.0 in /opt/conda/envs/py_3.12/lib/python3.12/site-packages (from sympy>=1.13.3->torch>=2.8->xformers==0.0.33+5d4b92a5.d20251103) (1.3.0) 2025-11-03T17:07:28.2684721Z Requirement already satisfied: MarkupSafe>=2.0 in /opt/conda/envs/py_3.12/lib/python3.12/site-packages (from jinja2->torch>=2.8->xformers==0.0.33+5d4b92a5.d20251103) (3.0.3) 2025-11-03T17:07:28.4406083Z Installing collected packages: xformers 2025-11-03T17:07:32.0967538Z Successfully installed xformers-0.0.33+5d4b92a5.d20251103 2025-11-03T17:07:32.1409661Z 2025-11-03 17:07:32,140 [INFO] cli.lib.common.pip_helper: Installing /var/lib/jenkins/workspace/dist/external/vllm/wheels/vllm/vllm-0.11.0rc2.dev389+ge51928192.d20251103-cp38-abi3-linux_x86_64.whl... 2025-11-03T17:07:32.1411635Z 2025-11-03 17:07:32,140 [INFO] cli.lib.common.pip_helper: pip installing packages: /opt/conda/envs/py_3.12/bin/python -m pip install /var/lib/jenkins/workspace/dist/external/vllm/wheels/vllm/vllm-0.11.0rc2.dev389+ge51928192.d20251103-cp38-abi3-linux_x86_64.whl 2025-11-03T17:07:32.1413157Z 2025-11-03 17:07:32,140 [INFO] cli.lib.common.utils: [cmd] /opt/conda/envs/py_3.12/bin/python -m pip install /var/lib/jenkins/workspace/dist/external/vllm/wheels/vllm/vllm-0.11.0rc2.dev389+ge51928192.d20251103-cp38-abi3-linux_x86_64.whl 2025-11-03T17:07:32.4315074Z Processing /var/lib/jenkins/workspace/dist/external/vllm/wheels/vllm/vllm-0.11.0rc2.dev389+ge51928192.d20251103-cp38-abi3-linux_x86_64.whl 2025-11-03T17:07:33.0307842Z Collecting regex (from vllm==0.11.0rc2.dev389+ge51928192.d20251103) 2025-11-03T17:07:33.0556818Z Downloading regex-2025.10.23-cp312-cp312-manylinux2014_x86_64.manylinux_2_17_x86_64.manylinux_2_28_x86_64.whl.metadata (40 kB) 2025-11-03T17:07:33.0624548Z Requirement already satisfied: cachetools in /opt/conda/envs/py_3.12/lib/python3.12/site-packages (from vllm==0.11.0rc2.dev389+ge51928192.d20251103) (6.2.1) 2025-11-03T17:07:33.0627293Z Requirement already satisfied: psutil in /opt/conda/envs/py_3.12/lib/python3.12/site-packages (from vllm==0.11.0rc2.dev389+ge51928192.d20251103) (7.1.2) 2025-11-03T17:07:33.0988242Z Collecting sentencepiece (from vllm==0.11.0rc2.dev389+ge51928192.d20251103) 2025-11-03T17:07:33.1035025Z Downloading sentencepiece-0.2.1-cp312-cp312-manylinux_2_27_x86_64.manylinux_2_28_x86_64.whl.metadata (10 kB) 2025-11-03T17:07:33.1083674Z Requirement already satisfied: numpy in /opt/conda/envs/py_3.12/lib/python3.12/site-packages (from vllm==0.11.0rc2.dev389+ge51928192.d20251103) (1.26.2) 2025-11-03T17:07:33.1087037Z Requirement already satisfied: requests>=2.26.0 in /opt/conda/envs/py_3.12/lib/python3.12/site-packages (from vllm==0.11.0rc2.dev389+ge51928192.d20251103) (2.32.5) 2025-11-03T17:07:33.1089790Z Requirement already satisfied: tqdm in /opt/conda/envs/py_3.12/lib/python3.12/site-packages (from vllm==0.11.0rc2.dev389+ge51928192.d20251103) (4.67.1) 2025-11-03T17:07:33.1603522Z Collecting blake3 (from vllm==0.11.0rc2.dev389+ge51928192.d20251103) 2025-11-03T17:07:33.1647807Z Downloading blake3-1.0.8-cp312-cp312-manylinux_2_17_x86_64.manylinux2014_x86_64.whl.metadata (4.6 kB) 2025-11-03T17:07:33.1760211Z Collecting py-cpuinfo (from vllm==0.11.0rc2.dev389+ge51928192.d20251103) 2025-11-03T17:07:33.1796144Z Downloading py_cpuinfo-9.0.0-py3-none-any.whl.metadata (794 bytes) 2025-11-03T17:07:33.2114706Z Collecting transformers>=4.55.2 (from vllm==0.11.0rc2.dev389+ge51928192.d20251103) 2025-11-03T17:07:33.2152806Z Downloading transformers-4.57.1-py3-none-any.whl.metadata (43 kB) 2025-11-03T17:07:33.4672623Z Collecting tokenizers>=0.21.1 (from vllm==0.11.0rc2.dev389+ge51928192.d20251103) 2025-11-03T17:07:33.4708423Z Downloading tokenizers-0.22.1-cp39-abi3-manylinux_2_17_x86_64.manylinux2014_x86_64.whl.metadata (6.8 kB) 2025-11-03T17:07:33.4766813Z Requirement already satisfied: protobuf in /opt/conda/envs/py_3.12/lib/python3.12/site-packages (from vllm==0.11.0rc2.dev389+ge51928192.d20251103) (5.29.5) 2025-11-03T17:07:33.5269393Z Collecting fastapi>=0.115.0 (from fastapi[standard]>=0.115.0->vllm==0.11.0rc2.dev389+ge51928192.d20251103) 2025-11-03T17:07:33.5307119Z Downloading fastapi-0.121.0-py3-none-any.whl.metadata (28 kB) 2025-11-03T17:07:33.5397827Z Requirement already satisfied: aiohttp in /opt/conda/envs/py_3.12/lib/python3.12/site-packages (from vllm==0.11.0rc2.dev389+ge51928192.d20251103) (3.13.2) 2025-11-03T17:07:33.5804342Z Collecting openai>=1.99.1 (from vllm==0.11.0rc2.dev389+ge51928192.d20251103) 2025-11-03T17:07:33.5838535Z Downloading openai-2.6.1-py3-none-any.whl.metadata (29 kB) 2025-11-03T17:07:33.6756632Z Collecting pydantic>=2.12.0 (from vllm==0.11.0rc2.dev389+ge51928192.d20251103) 2025-11-03T17:07:33.6797752Z Downloading pydantic-2.12.3-py3-none-any.whl.metadata (87 kB) 2025-11-03T17:07:33.6984937Z Collecting prometheus_client>=0.18.0 (from vllm==0.11.0rc2.dev389+ge51928192.d20251103) 2025-11-03T17:07:33.7026266Z Downloading prometheus_client-0.23.1-py3-none-any.whl.metadata (1.9 kB) 2025-11-03T17:07:33.7073190Z Requirement already satisfied: pillow in /opt/conda/envs/py_3.12/lib/python3.12/site-packages (from vllm==0.11.0rc2.dev389+ge51928192.d20251103) (11.0.0) 2025-11-03T17:07:33.7190650Z Collecting prometheus-fastapi-instrumentator>=7.0.0 (from vllm==0.11.0rc2.dev389+ge51928192.d20251103) 2025-11-03T17:07:33.7232838Z Downloading prometheus_fastapi_instrumentator-7.1.0-py3-none-any.whl.metadata (13 kB) 2025-11-03T17:07:33.7560376Z Collecting tiktoken>=0.6.0 (from vllm==0.11.0rc2.dev389+ge51928192.d20251103) 2025-11-03T17:07:33.7594865Z Downloading tiktoken-0.12.0-cp312-cp312-manylinux_2_28_x86_64.whl.metadata (6.7 kB) 2025-11-03T17:07:33.7797555Z Collecting lm-format-enforcer==0.11.3 (from vllm==0.11.0rc2.dev389+ge51928192.d20251103) 2025-11-03T17:07:33.7857728Z Downloading lm_format_enforcer-0.11.3-py3-none-any.whl.metadata (17 kB) 2025-11-03T17:07:33.8253158Z Collecting llguidance<0.8.0,>=0.7.11 (from vllm==0.11.0rc2.dev389+ge51928192.d20251103) 2025-11-03T17:07:33.8318994Z Downloading llguidance-0.7.30-cp39-abi3-manylinux_2_17_x86_64.manylinux2014_x86_64.whl.metadata (10 kB) 2025-11-03T17:07:33.8602809Z Collecting outlines_core==0.2.11 (from vllm==0.11.0rc2.dev389+ge51928192.d20251103) 2025-11-03T17:07:33.8660020Z Downloading outlines_core-0.2.11-cp312-cp312-manylinux_2_17_x86_64.manylinux2014_x86_64.whl.metadata (5.8 kB) 2025-11-03T17:07:33.8967222Z Collecting diskcache==5.6.3 (from vllm==0.11.0rc2.dev389+ge51928192.d20251103) 2025-11-03T17:07:33.8999715Z Downloading diskcache-5.6.3-py3-none-any.whl.metadata (20 kB) 2025-11-03T17:07:33.9125492Z Collecting lark==1.2.2 (from vllm==0.11.0rc2.dev389+ge51928192.d20251103) 2025-11-03T17:07:33.9171950Z Downloading lark-1.2.2-py3-none-any.whl.metadata (1.8 kB) 2025-11-03T17:07:33.9492669Z Collecting xgrammar==0.1.25 (from vllm==0.11.0rc2.dev389+ge51928192.d20251103) 2025-11-03T17:07:33.9531564Z Downloading xgrammar-0.1.25-cp312-cp312-manylinux_2_17_x86_64.manylinux2014_x86_64.whl.metadata (6.5 kB) 2025-11-03T17:07:33.9585168Z Requirement already satisfied: typing_extensions>=4.10 in /opt/conda/envs/py_3.12/lib/python3.12/site-packages (from vllm==0.11.0rc2.dev389+ge51928192.d20251103) (4.12.2) 2025-11-03T17:07:33.9587833Z Requirement already satisfied: filelock>=3.16.1 in /opt/conda/envs/py_3.12/lib/python3.12/site-packages (from vllm==0.11.0rc2.dev389+ge51928192.d20251103) (3.18.0) 2025-11-03T17:07:33.9680220Z Collecting partial-json-parser (from vllm==0.11.0rc2.dev389+ge51928192.d20251103) 2025-11-03T17:07:33.9725187Z Downloading partial_json_parser-0.2.1.1.post6-py3-none-any.whl.metadata (6.1 kB) 2025-11-03T17:07:34.1565705Z Collecting pyzmq>=25.0.0 (from vllm==0.11.0rc2.dev389+ge51928192.d20251103) 2025-11-03T17:07:34.1608857Z Downloading pyzmq-27.1.0-cp312-abi3-manylinux_2_26_x86_64.manylinux_2_28_x86_64.whl.metadata (6.0 kB) 2025-11-03T17:07:34.2056307Z Collecting msgspec (from vllm==0.11.0rc2.dev389+ge51928192.d20251103) 2025-11-03T17:07:34.2113630Z Downloading msgspec-0.19.0-cp312-cp312-manylinux_2_17_x86_64.manylinux2014_x86_64.whl.metadata (6.9 kB) 2025-11-03T17:07:34.2273110Z Collecting gguf>=0.13.0 (from vllm==0.11.0rc2.dev389+ge51928192.d20251103) 2025-11-03T17:07:34.2310746Z Downloading gguf-0.17.1-py3-none-any.whl.metadata (4.3 kB) 2025-11-03T17:07:34.2710864Z Collecting mistral_common>=1.8.5 (from mistral_common[audio,image]>=1.8.5->vllm==0.11.0rc2.dev389+ge51928192.d20251103) 2025-11-03T17:07:34.2775120Z Downloading mistral_common-1.8.5-py3-none-any.whl.metadata (5.1 kB) 2025-11-03T17:07:34.3303501Z Collecting opencv-python-headless>=4.11.0 (from vllm==0.11.0rc2.dev389+ge51928192.d20251103) 2025-11-03T17:07:34.3338789Z Downloading opencv_python_headless-4.12.0.88-cp37-abi3-manylinux2014_x86_64.manylinux_2_17_x86_64.whl.metadata (19 kB) 2025-11-03T17:07:34.3392202Z Requirement already satisfied: pyyaml in /opt/conda/envs/py_3.12/lib/python3.12/site-packages (from vllm==0.11.0rc2.dev389+ge51928192.d20251103) (6.0.2) 2025-11-03T17:07:34.3397043Z Requirement already satisfied: six>=1.16.0 in /opt/conda/envs/py_3.12/lib/python3.12/site-packages (from vllm==0.11.0rc2.dev389+ge51928192.d20251103) (1.17.0) 2025-11-03T17:07:34.3402321Z Requirement already satisfied: setuptools<80,>=77.0.3 in /opt/conda/envs/py_3.12/lib/python3.12/site-packages (from vllm==0.11.0rc2.dev389+ge51928192.d20251103) (78.1.1) 2025-11-03T17:07:34.3747184Z Collecting einops (from vllm==0.11.0rc2.dev389+ge51928192.d20251103) 2025-11-03T17:07:34.3815454Z Downloading einops-0.8.1-py3-none-any.whl.metadata (13 kB) 2025-11-03T17:07:34.4031469Z Collecting compressed-tensors==0.11.0 (from vllm==0.11.0rc2.dev389+ge51928192.d20251103) 2025-11-03T17:07:34.4081615Z Downloading compressed_tensors-0.11.0-py3-none-any.whl.metadata (7.0 kB) 2025-11-03T17:07:34.4237107Z Collecting depyf==0.19.0 (from vllm==0.11.0rc2.dev389+ge51928192.d20251103) 2025-11-03T17:07:34.4318903Z Downloading depyf-0.19.0-py3-none-any.whl.metadata (7.3 kB) 2025-11-03T17:07:34.4502057Z Collecting cloudpickle (from vllm==0.11.0rc2.dev389+ge51928192.d20251103) 2025-11-03T17:07:34.4537470Z Downloading cloudpickle-3.1.2-py3-none-any.whl.metadata (7.1 kB) 2025-11-03T17:07:34.5168269Z Collecting watchfiles (from vllm==0.11.0rc2.dev389+ge51928192.d20251103) 2025-11-03T17:07:34.5228130Z Downloading watchfiles-1.1.1-cp312-cp312-manylinux_2_17_x86_64.manylinux2014_x86_64.whl.metadata (4.9 kB) 2025-11-03T17:07:34.5377699Z Collecting python-json-logger (from vllm==0.11.0rc2.dev389+ge51928192.d20251103) 2025-11-03T17:07:34.5415497Z Downloading python_json_logger-4.0.0-py3-none-any.whl.metadata (4.0 kB) 2025-11-03T17:07:34.5468512Z Requirement already satisfied: scipy in /opt/conda/envs/py_3.12/lib/python3.12/site-packages (from vllm==0.11.0rc2.dev389+ge51928192.d20251103) (1.14.1) 2025-11-03T17:07:34.5470816Z Requirement already satisfied: ninja in /opt/conda/envs/py_3.12/lib/python3.12/site-packages (from vllm==0.11.0rc2.dev389+ge51928192.d20251103) (1.10.2) 2025-11-03T17:07:34.6149967Z Collecting pybase64 (from vllm==0.11.0rc2.dev389+ge51928192.d20251103) 2025-11-03T17:07:34.6198817Z Downloading pybase64-1.4.2-cp312-cp312-manylinux1_x86_64.manylinux2014_x86_64.manylinux_2_17_x86_64.manylinux_2_5_x86_64.whl.metadata (8.7 kB) 2025-11-03T17:07:34.6572134Z Collecting cbor2 (from vllm==0.11.0rc2.dev389+ge51928192.d20251103) 2025-11-03T17:07:34.6617807Z Downloading cbor2-5.7.1-cp312-cp312-manylinux2014_x86_64.manylinux_2_17_x86_64.manylinux_2_28_x86_64.whl.metadata (5.4 kB) 2025-11-03T17:07:34.7035053Z Collecting setproctitle (from vllm==0.11.0rc2.dev389+ge51928192.d20251103) 2025-11-03T17:07:34.7078028Z Downloading setproctitle-1.3.7-cp312-cp312-manylinux1_x86_64.manylinux_2_28_x86_64.manylinux_2_5_x86_64.whl.metadata (10 kB) 2025-11-03T17:07:34.7274736Z Collecting openai-harmony>=0.0.3 (from vllm==0.11.0rc2.dev389+ge51928192.d20251103) 2025-11-03T17:07:34.7376312Z Downloading openai_harmony-0.0.4-cp38-abi3-manylinux_2_17_x86_64.manylinux2014_x86_64.whl.metadata (8.0 kB) 2025-11-03T17:07:34.8112955Z Collecting numba==0.61.2 (from vllm==0.11.0rc2.dev389+ge51928192.d20251103) 2025-11-03T17:07:34.8178734Z Downloading numba-0.61.2-cp312-cp312-manylinux2014_x86_64.manylinux_2_17_x86_64.whl.metadata (2.8 kB) 2025-11-03T17:07:34.9071176Z Collecting ray>=2.48.0 (from ray[cgraph]>=2.48.0->vllm==0.11.0rc2.dev389+ge51928192.d20251103) 2025-11-03T17:07:34.9119024Z Downloading ray-2.51.1-cp312-cp312-manylinux2014_x86_64.whl.metadata (21 kB) 2025-11-03T17:07:34.9713505Z Collecting flashinfer-python==0.4.0 (from vllm==0.11.0rc2.dev389+ge51928192.d20251103) 2025-11-03T17:07:34.9779459Z Downloading flashinfer_python-0.4.0.tar.gz (4.5 MB) 2025-11-03T17:07:35.0122123Z [?25l ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 0.0/4.5 MB ? eta -:--:-- 2025-11-03T17:07:35.0122802Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 4.5/4.5 MB 135.8 MB/s 0:00:00 2025-11-03T17:07:36.6589666Z [?25h Installing build dependencies ... [?25l- \ | / done 2025-11-03T17:07:37.2381415Z [?25h Getting requirements to build wheel ... [?25l- done 2025-11-03T17:07:38.0695553Z [?25h Preparing metadata (pyproject.toml) ... [?25l- done 2025-11-03T17:07:38.0722709Z [?25hRequirement already satisfied: torch>=1.7.0 in /opt/conda/envs/py_3.12/lib/python3.12/site-packages (from compressed-tensors==0.11.0->vllm==0.11.0rc2.dev389+ge51928192.d20251103) (2.10.0a0+git3f6538f) 2025-11-03T17:07:38.1528949Z Collecting frozendict (from compressed-tensors==0.11.0->vllm==0.11.0rc2.dev389+ge51928192.d20251103) 2025-11-03T17:07:38.1576646Z Downloading frozendict-2.4.6-py312-none-any.whl.metadata (23 kB) 2025-11-03T17:07:38.1721535Z Collecting astor (from depyf==0.19.0->vllm==0.11.0rc2.dev389+ge51928192.d20251103) 2025-11-03T17:07:38.1763017Z Downloading astor-0.8.1-py2.py3-none-any.whl.metadata (4.2 kB) 2025-11-03T17:07:38.1802975Z Requirement already satisfied: dill in /opt/conda/envs/py_3.12/lib/python3.12/site-packages (from depyf==0.19.0->vllm==0.11.0rc2.dev389+ge51928192.d20251103) (0.3.7) 2025-11-03T17:07:38.2035523Z Collecting apache-tvm-ffi==0.1.0b15 (from flashinfer-python==0.4.0->vllm==0.11.0rc2.dev389+ge51928192.d20251103) 2025-11-03T17:07:38.2045985Z Using cached apache_tvm_ffi-0.1.0b15-cp312-abi3-manylinux_2_24_x86_64.manylinux_2_28_x86_64.whl.metadata (2.0 kB) 2025-11-03T17:07:38.2065247Z Requirement already satisfied: click in /opt/conda/envs/py_3.12/lib/python3.12/site-packages (from flashinfer-python==0.4.0->vllm==0.11.0rc2.dev389+ge51928192.d20251103) (8.3.0) 2025-11-03T17:07:38.2186962Z Collecting nvidia-cudnn-frontend>=1.13.0 (from flashinfer-python==0.4.0->vllm==0.11.0rc2.dev389+ge51928192.d20251103) 2025-11-03T17:07:38.2239503Z Downloading nvidia_cudnn_frontend-1.15.0-cp312-cp312-manylinux_2_27_x86_64.manylinux_2_28_x86_64.whl.metadata (6.7 kB) 2025-11-03T17:07:38.2354737Z Collecting nvidia-cutlass-dsl>=4.2.1 (from flashinfer-python==0.4.0->vllm==0.11.0rc2.dev389+ge51928192.d20251103) 2025-11-03T17:07:38.2398724Z Downloading nvidia_cutlass_dsl-4.2.1-cp312-cp312-manylinux_2_28_x86_64.whl.metadata (2.5 kB) 2025-11-03T17:07:38.2549744Z Collecting nvidia-ml-py (from flashinfer-python==0.4.0->vllm==0.11.0rc2.dev389+ge51928192.d20251103) 2025-11-03T17:07:38.2590850Z Downloading nvidia_ml_py-13.580.82-py3-none-any.whl.metadata (9.6 kB) 2025-11-03T17:07:38.2753538Z Collecting packaging>=24.2 (from flashinfer-python==0.4.0->vllm==0.11.0rc2.dev389+ge51928192.d20251103) 2025-11-03T17:07:38.2763702Z Using cached packaging-25.0-py3-none-any.whl.metadata (3.3 kB) 2025-11-03T17:07:38.2778935Z Requirement already satisfied: tabulate in /opt/conda/envs/py_3.12/lib/python3.12/site-packages (from flashinfer-python==0.4.0->vllm==0.11.0rc2.dev389+ge51928192.d20251103) (0.9.0) 2025-11-03T17:07:38.2887082Z Collecting interegular>=0.3.2 (from lm-format-enforcer==0.11.3->vllm==0.11.0rc2.dev389+ge51928192.d20251103) 2025-11-03T17:07:38.3058904Z Downloading interegular-0.3.3-py37-none-any.whl.metadata (3.0 kB) 2025-11-03T17:07:38.3621593Z Collecting llvmlite<0.45,>=0.44.0dev0 (from numba==0.61.2->vllm==0.11.0rc2.dev389+ge51928192.d20251103) 2025-11-03T17:07:38.3741541Z Downloading llvmlite-0.44.0-cp312-cp312-manylinux_2_17_x86_64.manylinux2014_x86_64.whl.metadata (5.0 kB) 2025-11-03T17:07:38.3895115Z Requirement already satisfied: triton in /opt/conda/envs/py_3.12/lib/python3.12/site-packages (from xgrammar==0.1.25->vllm==0.11.0rc2.dev389+ge51928192.d20251103) (3.5.0+git7416ffcb) 2025-11-03T17:07:38.4208926Z Collecting starlette<0.50.0,>=0.40.0 (from fastapi>=0.115.0->fastapi[standard]>=0.115.0->vllm==0.11.0rc2.dev389+ge51928192.d20251103) 2025-11-03T17:07:38.4245819Z Downloading starlette-0.49.3-py3-none-any.whl.metadata (6.4 kB) 2025-11-03T17:07:38.4497932Z Collecting annotated-doc>=0.0.2 (from fastapi>=0.115.0->fastapi[standard]>=0.115.0->vllm==0.11.0rc2.dev389+ge51928192.d20251103) 2025-11-03T17:07:38.4529976Z Downloading annotated_doc-0.0.3-py3-none-any.whl.metadata (6.6 kB) 2025-11-03T17:07:38.4652598Z Collecting annotated-types>=0.6.0 (from pydantic>=2.12.0->vllm==0.11.0rc2.dev389+ge51928192.d20251103) 2025-11-03T17:07:38.4686695Z Downloading annotated_types-0.7.0-py3-none-any.whl.metadata (15 kB) 2025-11-03T17:07:39.0373864Z Collecting pydantic-core==2.41.4 (from pydantic>=2.12.0->vllm==0.11.0rc2.dev389+ge51928192.d20251103) 2025-11-03T17:07:39.0418406Z Downloading pydantic_core-2.41.4-cp312-cp312-manylinux_2_17_x86_64.manylinux2014_x86_64.whl.metadata (7.3 kB) 2025-11-03T17:07:39.0590105Z Collecting typing_extensions>=4.10 (from vllm==0.11.0rc2.dev389+ge51928192.d20251103) 2025-11-03T17:07:39.0600742Z Using cached typing_extensions-4.15.0-py3-none-any.whl.metadata (3.3 kB) 2025-11-03T17:07:39.0690631Z Collecting typing-inspection>=0.4.2 (from pydantic>=2.12.0->vllm==0.11.0rc2.dev389+ge51928192.d20251103) 2025-11-03T17:07:39.0726637Z Downloading typing_inspection-0.4.2-py3-none-any.whl.metadata (2.6 kB) 2025-11-03T17:07:39.0938916Z Collecting anyio<5,>=3.6.2 (from starlette<0.50.0,>=0.40.0->fastapi>=0.115.0->fastapi[standard]>=0.115.0->vllm==0.11.0rc2.dev389+ge51928192.d20251103) 2025-11-03T17:07:39.0973910Z Downloading anyio-4.11.0-py3-none-any.whl.metadata (4.1 kB) 2025-11-03T17:07:39.1036870Z Requirement already satisfied: idna>=2.8 in /opt/conda/envs/py_3.12/lib/python3.12/site-packages (from anyio<5,>=3.6.2->starlette<0.50.0,>=0.40.0->fastapi>=0.115.0->fastapi[standard]>=0.115.0->vllm==0.11.0rc2.dev389+ge51928192.d20251103) (3.11) 2025-11-03T17:07:39.1108886Z Collecting sniffio>=1.1 (from anyio<5,>=3.6.2->starlette<0.50.0,>=0.40.0->fastapi>=0.115.0->fastapi[standard]>=0.115.0->vllm==0.11.0rc2.dev389+ge51928192.d20251103) 2025-11-03T17:07:39.1143142Z Downloading sniffio-1.3.1-py3-none-any.whl.metadata (3.9 kB) 2025-11-03T17:07:39.1301018Z Collecting fastapi-cli>=0.0.8 (from fastapi-cli[standard]>=0.0.8; extra == "standard"->fastapi[standard]>=0.115.0->vllm==0.11.0rc2.dev389+ge51928192.d20251103) 2025-11-03T17:07:39.1345570Z Downloading fastapi_cli-0.0.14-py3-none-any.whl.metadata (6.4 kB) 2025-11-03T17:07:39.1551424Z Collecting httpx<1.0.0,>=0.23.0 (from fastapi[standard]>=0.115.0->vllm==0.11.0rc2.dev389+ge51928192.d20251103) 2025-11-03T17:07:39.1590428Z Downloading httpx-0.28.1-py3-none-any.whl.metadata (7.1 kB) 2025-11-03T17:07:39.1645206Z Requirement already satisfied: jinja2>=3.1.5 in /opt/conda/envs/py_3.12/lib/python3.12/site-packages (from fastapi[standard]>=0.115.0->vllm==0.11.0rc2.dev389+ge51928192.d20251103) (3.1.6) 2025-11-03T17:07:39.1743266Z Collecting python-multipart>=0.0.18 (from fastapi[standard]>=0.115.0->vllm==0.11.0rc2.dev389+ge51928192.d20251103) 2025-11-03T17:07:39.1777577Z Downloading python_multipart-0.0.20-py3-none-any.whl.metadata (1.8 kB) 2025-11-03T17:07:39.1914516Z Collecting email-validator>=2.0.0 (from fastapi[standard]>=0.115.0->vllm==0.11.0rc2.dev389+ge51928192.d20251103) 2025-11-03T17:07:39.1945332Z Downloading email_validator-2.3.0-py3-none-any.whl.metadata (26 kB) 2025-11-03T17:07:39.2213137Z Collecting uvicorn>=0.12.0 (from uvicorn[standard]>=0.12.0; extra == "standard"->fastapi[standard]>=0.115.0->vllm==0.11.0rc2.dev389+ge51928192.d20251103) 2025-11-03T17:07:39.2251500Z Downloading uvicorn-0.38.0-py3-none-any.whl.metadata (6.8 kB) 2025-11-03T17:07:39.2324117Z Requirement already satisfied: certifi in /opt/conda/envs/py_3.12/lib/python3.12/site-packages (from httpx<1.0.0,>=0.23.0->fastapi[standard]>=0.115.0->vllm==0.11.0rc2.dev389+ge51928192.d20251103) (2025.10.5) 2025-11-03T17:07:39.2476706Z Collecting httpcore==1.* (from httpx<1.0.0,>=0.23.0->fastapi[standard]>=0.115.0->vllm==0.11.0rc2.dev389+ge51928192.d20251103) 2025-11-03T17:07:39.2502819Z Downloading httpcore-1.0.9-py3-none-any.whl.metadata (21 kB) 2025-11-03T17:07:39.2688118Z Collecting h11>=0.16 (from httpcore==1.*->httpx<1.0.0,>=0.23.0->fastapi[standard]>=0.115.0->vllm==0.11.0rc2.dev389+ge51928192.d20251103) 2025-11-03T17:07:39.2723798Z Downloading h11-0.16.0-py3-none-any.whl.metadata (8.3 kB) 2025-11-03T17:07:39.2879885Z Collecting dnspython>=2.0.0 (from email-validator>=2.0.0->fastapi[standard]>=0.115.0->vllm==0.11.0rc2.dev389+ge51928192.d20251103) 2025-11-03T17:07:39.2910357Z Downloading dnspython-2.8.0-py3-none-any.whl.metadata (5.7 kB) 2025-11-03T17:07:39.3147563Z Collecting typer>=0.15.1 (from fastapi-cli>=0.0.8->fastapi-cli[standard]>=0.0.8; extra == "standard"->fastapi[standard]>=0.115.0->vllm==0.11.0rc2.dev389+ge51928192.d20251103) 2025-11-03T17:07:39.3185056Z Downloading typer-0.20.0-py3-none-any.whl.metadata (16 kB) 2025-11-03T17:07:39.3465826Z Collecting rich-toolkit>=0.14.8 (from fastapi-cli>=0.0.8->fastapi-cli[standard]>=0.0.8; extra == "standard"->fastapi[standard]>=0.115.0->vllm==0.11.0rc2.dev389+ge51928192.d20251103) 2025-11-03T17:07:39.3504870Z Downloading rich_toolkit-0.15.1-py3-none-any.whl.metadata (1.0 kB) 2025-11-03T17:07:39.3646602Z Collecting fastapi-cloud-cli>=0.1.1 (from fastapi-cli[standard]>=0.0.8; extra == "standard"->fastapi[standard]>=0.115.0->vllm==0.11.0rc2.dev389+ge51928192.d20251103) 2025-11-03T17:07:39.3693205Z Downloading fastapi_cloud_cli-0.3.1-py3-none-any.whl.metadata (3.2 kB) 2025-11-03T17:07:39.4554108Z Collecting rignore>=0.5.1 (from fastapi-cloud-cli>=0.1.1->fastapi-cli[standard]>=0.0.8; extra == "standard"->fastapi[standard]>=0.115.0->vllm==0.11.0rc2.dev389+ge51928192.d20251103) 2025-11-03T17:07:39.4604140Z Downloading rignore-0.7.4-cp312-cp312-manylinux_2_17_x86_64.manylinux2014_x86_64.whl.metadata (4.2 kB) 2025-11-03T17:07:39.5249353Z Collecting sentry-sdk>=2.20.0 (from fastapi-cloud-cli>=0.1.1->fastapi-cli[standard]>=0.0.8; extra == "standard"->fastapi[standard]>=0.115.0->vllm==0.11.0rc2.dev389+ge51928192.d20251103) 2025-11-03T17:07:39.5293125Z Downloading sentry_sdk-2.43.0-py2.py3-none-any.whl.metadata (10 kB) 2025-11-03T17:07:39.5529157Z Requirement already satisfied: MarkupSafe>=2.0 in /opt/conda/envs/py_3.12/lib/python3.12/site-packages (from jinja2>=3.1.5->fastapi[standard]>=0.115.0->vllm==0.11.0rc2.dev389+ge51928192.d20251103) (3.0.3) 2025-11-03T17:07:39.5858534Z Collecting jsonschema>=4.21.1 (from mistral_common>=1.8.5->mistral_common[audio,image]>=1.8.5->vllm==0.11.0rc2.dev389+ge51928192.d20251103) 2025-11-03T17:07:39.5904035Z Downloading jsonschema-4.25.1-py3-none-any.whl.metadata (7.6 kB) 2025-11-03T17:07:39.6106947Z Collecting pydantic-extra-types>=2.10.5 (from pydantic-extra-types[pycountry]>=2.10.5->mistral_common>=1.8.5->mistral_common[audio,image]>=1.8.5->vllm==0.11.0rc2.dev389+ge51928192.d20251103) 2025-11-03T17:07:39.6150030Z Downloading pydantic_extra_types-2.10.6-py3-none-any.whl.metadata (4.0 kB) 2025-11-03T17:07:39.6234932Z Requirement already satisfied: attrs>=22.2.0 in /opt/conda/envs/py_3.12/lib/python3.12/site-packages (from jsonschema>=4.21.1->mistral_common>=1.8.5->mistral_common[audio,image]>=1.8.5->vllm==0.11.0rc2.dev389+ge51928192.d20251103) (25.4.0) 2025-11-03T17:07:39.6348178Z Collecting jsonschema-specifications>=2023.03.6 (from jsonschema>=4.21.1->mistral_common>=1.8.5->mistral_common[audio,image]>=1.8.5->vllm==0.11.0rc2.dev389+ge51928192.d20251103) 2025-11-03T17:07:39.6390946Z Downloading jsonschema_specifications-2025.9.1-py3-none-any.whl.metadata (2.9 kB) 2025-11-03T17:07:39.6613037Z Collecting referencing>=0.28.4 (from jsonschema>=4.21.1->mistral_common>=1.8.5->mistral_common[audio,image]>=1.8.5->vllm==0.11.0rc2.dev389+ge51928192.d20251103) 2025-11-03T17:07:39.6653583Z Downloading referencing-0.37.0-py3-none-any.whl.metadata (2.8 kB) 2025-11-03T17:07:39.9351720Z Collecting rpds-py>=0.7.1 (from jsonschema>=4.21.1->mistral_common>=1.8.5->mistral_common[audio,image]>=1.8.5->vllm==0.11.0rc2.dev389+ge51928192.d20251103) 2025-11-03T17:07:39.9395685Z Downloading rpds_py-0.28.0-cp312-cp312-manylinux_2_17_x86_64.manylinux2014_x86_64.whl.metadata (4.1 kB) 2025-11-03T17:07:39.9788960Z Collecting cuda-python>=12.8 (from nvidia-cutlass-dsl>=4.2.1->flashinfer-python==0.4.0->vllm==0.11.0rc2.dev389+ge51928192.d20251103) 2025-11-03T17:07:39.9834364Z Downloading cuda_python-13.0.3-py3-none-any.whl.metadata (4.7 kB) 2025-11-03T17:07:40.0031116Z Collecting cuda-bindings~=13.0.3 (from cuda-python>=12.8->nvidia-cutlass-dsl>=4.2.1->flashinfer-python==0.4.0->vllm==0.11.0rc2.dev389+ge51928192.d20251103) 2025-11-03T17:07:40.0082753Z Downloading cuda_bindings-13.0.3-cp312-cp312-manylinux_2_24_x86_64.manylinux_2_28_x86_64.whl.metadata (2.3 kB) 2025-11-03T17:07:40.0134296Z Requirement already satisfied: cuda-pathfinder~=1.1 in /opt/conda/envs/py_3.12/lib/python3.12/site-packages (from cuda-python>=12.8->nvidia-cutlass-dsl>=4.2.1->flashinfer-python==0.4.0->vllm==0.11.0rc2.dev389+ge51928192.d20251103) (1.3.2) 2025-11-03T17:07:40.0208916Z Requirement already satisfied: distro<2,>=1.7.0 in /opt/conda/envs/py_3.12/lib/python3.12/site-packages (from openai>=1.99.1->vllm==0.11.0rc2.dev389+ge51928192.d20251103) (1.9.0) 2025-11-03T17:07:40.0729159Z Collecting jiter<1,>=0.10.0 (from openai>=1.99.1->vllm==0.11.0rc2.dev389+ge51928192.d20251103) 2025-11-03T17:07:40.0771951Z Downloading jiter-0.11.1-cp312-cp312-manylinux_2_17_x86_64.manylinux2014_x86_64.whl.metadata (5.2 kB) 2025-11-03T17:07:40.2512030Z Collecting numpy (from vllm==0.11.0rc2.dev389+ge51928192.d20251103) 2025-11-03T17:07:40.2592642Z Downloading numpy-2.2.6-cp312-cp312-manylinux_2_17_x86_64.manylinux2014_x86_64.whl.metadata (62 kB) 2025-11-03T17:07:40.3176579Z Collecting pycountry>=23 (from pydantic-extra-types[pycountry]>=2.10.5->mistral_common>=1.8.5->mistral_common[audio,image]>=1.8.5->vllm==0.11.0rc2.dev389+ge51928192.d20251103) 2025-11-03T17:07:40.3213327Z Downloading pycountry-24.6.1-py3-none-any.whl.metadata (12 kB) 2025-11-03T17:07:40.3508543Z Collecting click (from flashinfer-python==0.4.0->vllm==0.11.0rc2.dev389+ge51928192.d20251103) 2025-11-03T17:07:40.3543264Z Downloading click-8.2.1-py3-none-any.whl.metadata (2.5 kB) 2025-11-03T17:07:40.3589555Z Requirement already satisfied: msgpack<2.0.0,>=1.0.0 in /opt/conda/envs/py_3.12/lib/python3.12/site-packages (from ray>=2.48.0->ray[cgraph]>=2.48.0->vllm==0.11.0rc2.dev389+ge51928192.d20251103) (1.1.2) 2025-11-03T17:07:40.3822372Z Collecting cupy-cuda12x (from ray[cgraph]>=2.48.0->vllm==0.11.0rc2.dev389+ge51928192.d20251103) 2025-11-03T17:07:40.3865198Z Downloading cupy_cuda12x-13.6.0-cp312-cp312-manylinux2014_x86_64.whl.metadata (2.4 kB) 2025-11-03T17:07:40.4063244Z Requirement already satisfied: charset_normalizer<4,>=2 in /opt/conda/envs/py_3.12/lib/python3.12/site-packages (from requests>=2.26.0->vllm==0.11.0rc2.dev389+ge51928192.d20251103) (3.4.4) 2025-11-03T17:07:40.4069327Z Requirement already satisfied: urllib3<3,>=1.21.1 in /opt/conda/envs/py_3.12/lib/python3.12/site-packages (from requests>=2.26.0->vllm==0.11.0rc2.dev389+ge51928192.d20251103) (2.5.0) 2025-11-03T17:07:40.4463012Z Collecting rich>=13.7.1 (from rich-toolkit>=0.14.8->fastapi-cli>=0.0.8->fastapi-cli[standard]>=0.0.8; extra == "standard"->fastapi[standard]>=0.115.0->vllm==0.11.0rc2.dev389+ge51928192.d20251103) 2025-11-03T17:07:40.4497672Z Downloading rich-14.2.0-py3-none-any.whl.metadata (18 kB) 2025-11-03T17:07:40.4695565Z Collecting markdown-it-py>=2.2.0 (from rich>=13.7.1->rich-toolkit>=0.14.8->fastapi-cli>=0.0.8->fastapi-cli[standard]>=0.0.8; extra == "standard"->fastapi[standard]>=0.115.0->vllm==0.11.0rc2.dev389+ge51928192.d20251103) 2025-11-03T17:07:40.4728636Z Downloading markdown_it_py-4.0.0-py3-none-any.whl.metadata (7.3 kB) 2025-11-03T17:07:40.4800736Z Requirement already satisfied: pygments<3.0.0,>=2.13.0 in /opt/conda/envs/py_3.12/lib/python3.12/site-packages (from rich>=13.7.1->rich-toolkit>=0.14.8->fastapi-cli>=0.0.8->fastapi-cli[standard]>=0.0.8; extra == "standard"->fastapi[standard]>=0.115.0->vllm==0.11.0rc2.dev389+ge51928192.d20251103) (2.15.0) 2025-11-03T17:07:40.4897601Z Collecting mdurl~=0.1 (from markdown-it-py>=2.2.0->rich>=13.7.1->rich-toolkit>=0.14.8->fastapi-cli>=0.0.8->fastapi-cli[standard]>=0.0.8; extra == "standard"->fastapi[standard]>=0.115.0->vllm==0.11.0rc2.dev389+ge51928192.d20251103) 2025-11-03T17:07:40.4931911Z Downloading mdurl-0.1.2-py3-none-any.whl.metadata (1.6 kB) 2025-11-03T17:07:40.5539730Z Collecting huggingface-hub<2.0,>=0.16.4 (from tokenizers>=0.21.1->vllm==0.11.0rc2.dev389+ge51928192.d20251103) 2025-11-03T17:07:40.5574939Z Downloading huggingface_hub-1.0.1-py3-none-any.whl.metadata (13 kB) 2025-11-03T17:07:40.5731640Z Requirement already satisfied: fsspec>=2023.5.0 in /opt/conda/envs/py_3.12/lib/python3.12/site-packages (from huggingface-hub<2.0,>=0.16.4->tokenizers>=0.21.1->vllm==0.11.0rc2.dev389+ge51928192.d20251103) (2025.9.0) 2025-11-03T17:07:40.5857830Z Collecting shellingham (from huggingface-hub<2.0,>=0.16.4->tokenizers>=0.21.1->vllm==0.11.0rc2.dev389+ge51928192.d20251103) 2025-11-03T17:07:40.5893705Z Downloading shellingham-1.5.4-py2.py3-none-any.whl.metadata (3.5 kB) 2025-11-03T17:07:40.6051509Z Collecting typer-slim (from huggingface-hub<2.0,>=0.16.4->tokenizers>=0.21.1->vllm==0.11.0rc2.dev389+ge51928192.d20251103) 2025-11-03T17:07:40.6088220Z Downloading typer_slim-0.20.0-py3-none-any.whl.metadata (16 kB) 2025-11-03T17:07:40.6431046Z Collecting hf-xet<2.0.0,>=1.2.0 (from huggingface-hub<2.0,>=0.16.4->tokenizers>=0.21.1->vllm==0.11.0rc2.dev389+ge51928192.d20251103) 2025-11-03T17:07:40.6472780Z Downloading hf_xet-1.2.0-cp37-abi3-manylinux_2_17_x86_64.manylinux2014_x86_64.whl.metadata (4.9 kB) 2025-11-03T17:07:40.6641809Z Requirement already satisfied: sympy>=1.13.3 in /opt/conda/envs/py_3.12/lib/python3.12/site-packages (from torch>=1.7.0->compressed-tensors==0.11.0->vllm==0.11.0rc2.dev389+ge51928192.d20251103) (1.13.3) 2025-11-03T17:07:40.6645901Z Requirement already satisfied: networkx>=2.5.1 in /opt/conda/envs/py_3.12/lib/python3.12/site-packages (from torch>=1.7.0->compressed-tensors==0.11.0->vllm==0.11.0rc2.dev389+ge51928192.d20251103) (2.8.8) 2025-11-03T17:07:40.6700873Z Requirement already satisfied: mpmath<1.4,>=1.1.0 in /opt/conda/envs/py_3.12/lib/python3.12/site-packages (from sympy>=1.13.3->torch>=1.7.0->compressed-tensors==0.11.0->vllm==0.11.0rc2.dev389+ge51928192.d20251103) (1.3.0) 2025-11-03T17:07:40.6843813Z Collecting huggingface-hub<2.0,>=0.16.4 (from tokenizers>=0.21.1->vllm==0.11.0rc2.dev389+ge51928192.d20251103) 2025-11-03T17:07:40.6881425Z Downloading huggingface_hub-0.36.0-py3-none-any.whl.metadata (14 kB) 2025-11-03T17:07:40.8054791Z Collecting safetensors>=0.4.3 (from transformers>=4.55.2->vllm==0.11.0rc2.dev389+ge51928192.d20251103) 2025-11-03T17:07:40.8092269Z Downloading safetensors-0.6.2-cp38-abi3-manylinux_2_17_x86_64.manylinux2014_x86_64.whl.metadata (4.1 kB) 2025-11-03T17:07:40.8801291Z Collecting httptools>=0.6.3 (from uvicorn[standard]>=0.12.0; extra == "standard"->fastapi[standard]>=0.115.0->vllm==0.11.0rc2.dev389+ge51928192.d20251103) 2025-11-03T17:07:40.8841653Z Downloading httptools-0.7.1-cp312-cp312-manylinux1_x86_64.manylinux_2_28_x86_64.manylinux_2_5_x86_64.whl.metadata (3.5 kB) 2025-11-03T17:07:40.9424999Z Collecting python-dotenv>=0.13 (from uvicorn[standard]>=0.12.0; extra == "standard"->fastapi[standard]>=0.115.0->vllm==0.11.0rc2.dev389+ge51928192.d20251103) 2025-11-03T17:07:40.9463509Z Downloading python_dotenv-1.2.1-py3-none-any.whl.metadata (25 kB) 2025-11-03T17:07:40.9840011Z Collecting uvloop>=0.15.1 (from uvicorn[standard]>=0.12.0; extra == "standard"->fastapi[standard]>=0.115.0->vllm==0.11.0rc2.dev389+ge51928192.d20251103) 2025-11-03T17:07:40.9878181Z Downloading uvloop-0.22.1-cp312-cp312-manylinux2014_x86_64.manylinux_2_17_x86_64.manylinux_2_28_x86_64.whl.metadata (4.9 kB) 2025-11-03T17:07:41.0644933Z Collecting websockets>=10.4 (from uvicorn[standard]>=0.12.0; extra == "standard"->fastapi[standard]>=0.115.0->vllm==0.11.0rc2.dev389+ge51928192.d20251103) 2025-11-03T17:07:41.0677767Z Downloading websockets-15.0.1-cp312-cp312-manylinux_2_5_x86_64.manylinux1_x86_64.manylinux_2_17_x86_64.manylinux2014_x86_64.whl.metadata (6.8 kB) 2025-11-03T17:07:41.0890055Z Requirement already satisfied: aiohappyeyeballs>=2.5.0 in /opt/conda/envs/py_3.12/lib/python3.12/site-packages (from aiohttp->vllm==0.11.0rc2.dev389+ge51928192.d20251103) (2.6.1) 2025-11-03T17:07:41.0893632Z Requirement already satisfied: aiosignal>=1.4.0 in /opt/conda/envs/py_3.12/lib/python3.12/site-packages (from aiohttp->vllm==0.11.0rc2.dev389+ge51928192.d20251103) (1.4.0) 2025-11-03T17:07:41.0900039Z Requirement already satisfied: frozenlist>=1.1.1 in /opt/conda/envs/py_3.12/lib/python3.12/site-packages (from aiohttp->vllm==0.11.0rc2.dev389+ge51928192.d20251103) (1.8.0) 2025-11-03T17:07:41.0904105Z Requirement already satisfied: multidict<7.0,>=4.5 in /opt/conda/envs/py_3.12/lib/python3.12/site-packages (from aiohttp->vllm==0.11.0rc2.dev389+ge51928192.d20251103) (6.7.0) 2025-11-03T17:07:41.0907715Z Requirement already satisfied: propcache>=0.2.0 in /opt/conda/envs/py_3.12/lib/python3.12/site-packages (from aiohttp->vllm==0.11.0rc2.dev389+ge51928192.d20251103) (0.4.1) 2025-11-03T17:07:41.0911423Z Requirement already satisfied: yarl<2.0,>=1.17.0 in /opt/conda/envs/py_3.12/lib/python3.12/site-packages (from aiohttp->vllm==0.11.0rc2.dev389+ge51928192.d20251103) (1.22.0) 2025-11-03T17:07:41.1465093Z Collecting fastrlock>=0.5 (from cupy-cuda12x->ray[cgraph]>=2.48.0->vllm==0.11.0rc2.dev389+ge51928192.d20251103) 2025-11-03T17:07:41.1519766Z Downloading fastrlock-0.8.3-cp312-cp312-manylinux_2_5_x86_64.manylinux1_x86_64.manylinux_2_28_x86_64.whl.metadata (7.7 kB) 2025-11-03T17:07:41.1703091Z Requirement already satisfied: soundfile>=0.12.1 in /opt/conda/envs/py_3.12/lib/python3.12/site-packages (from mistral_common>=1.8.5->mistral_common[audio,image]>=1.8.5->vllm==0.11.0rc2.dev389+ge51928192.d20251103) (0.13.1) 2025-11-03T17:07:41.1723965Z Requirement already satisfied: cffi>=1.0 in /opt/conda/envs/py_3.12/lib/python3.12/site-packages (from soundfile>=0.12.1->mistral_common>=1.8.5->mistral_common[audio,image]>=1.8.5->vllm==0.11.0rc2.dev389+ge51928192.d20251103) (2.0.0) 2025-11-03T17:07:41.1744360Z Requirement already satisfied: pycparser in /opt/conda/envs/py_3.12/lib/python3.12/site-packages (from cffi>=1.0->soundfile>=0.12.1->mistral_common>=1.8.5->mistral_common[audio,image]>=1.8.5->vllm==0.11.0rc2.dev389+ge51928192.d20251103) (2.23) 2025-11-03T17:07:41.1770665Z Requirement already satisfied: soxr>=0.5.0 in /opt/conda/envs/py_3.12/lib/python3.12/site-packages (from mistral_common>=1.8.5->mistral_common[audio,image]>=1.8.5->vllm==0.11.0rc2.dev389+ge51928192.d20251103) (1.0.0) 2025-11-03T17:07:41.2334509Z Downloading compressed_tensors-0.11.0-py3-none-any.whl (179 kB) 2025-11-03T17:07:41.2424126Z Downloading depyf-0.19.0-py3-none-any.whl (39 kB) 2025-11-03T17:07:41.2493943Z Downloading diskcache-5.6.3-py3-none-any.whl (45 kB) 2025-11-03T17:07:41.2536189Z Using cached apache_tvm_ffi-0.1.0b15-cp312-abi3-manylinux_2_24_x86_64.manylinux_2_28_x86_64.whl (1.6 MB) 2025-11-03T17:07:41.2582248Z Downloading lark-1.2.2-py3-none-any.whl (111 kB) 2025-11-03T17:07:41.2681677Z Downloading lm_format_enforcer-0.11.3-py3-none-any.whl (45 kB) 2025-11-03T17:07:41.2792697Z Downloading numba-0.61.2-cp312-cp312-manylinux2014_x86_64.manylinux_2_17_x86_64.whl (3.9 MB) 2025-11-03T17:07:41.3020903Z [?25l ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 0.0/3.9 MB ? eta -:--:-- 2025-11-03T17:07:41.3021816Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 3.9/3.9 MB 180.2 MB/s 0:00:00 2025-11-03T17:07:41.3122251Z [?25hDownloading outlines_core-0.2.11-cp312-cp312-manylinux_2_17_x86_64.manylinux2014_x86_64.whl (2.3 MB) 2025-11-03T17:07:41.3441461Z [?25l ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 0.0/2.3 MB ? eta -:--:-- 2025-11-03T17:07:41.3442092Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 2.3/2.3 MB 69.1 MB/s 0:00:00 2025-11-03T17:07:41.3494611Z [?25hDownloading xgrammar-0.1.25-cp312-cp312-manylinux_2_17_x86_64.manylinux2014_x86_64.whl (8.7 MB) 2025-11-03T17:07:41.4060097Z [?25l ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 0.0/8.7 MB ? eta -:--:-- 2025-11-03T17:07:41.4061050Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 8.7/8.7 MB 156.1 MB/s 0:00:00 2025-11-03T17:07:41.4119498Z [?25hDownloading llguidance-0.7.30-cp39-abi3-manylinux_2_17_x86_64.manylinux2014_x86_64.whl (15.0 MB) 2025-11-03T17:07:41.5173094Z [?25l ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 0.0/15.0 MB ? eta -:--:-- 2025-11-03T17:07:41.5173776Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 15.0/15.0 MB 143.6 MB/s 0:00:00 2025-11-03T17:07:41.5243847Z [?25hDownloading llvmlite-0.44.0-cp312-cp312-manylinux_2_17_x86_64.manylinux2014_x86_64.whl (42.4 MB) 2025-11-03T17:07:41.7271051Z [?25l ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 0.0/42.4 MB ? eta -:--:-- 2025-11-03T17:07:41.8633087Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╸━━ 39.8/42.4 MB 200.0 MB/s eta 0:00:01 2025-11-03T17:07:41.8633822Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 42.4/42.4 MB 125.1 MB/s 0:00:00 2025-11-03T17:07:41.8672789Z [?25hDownloading fastapi-0.121.0-py3-none-any.whl (109 kB) 2025-11-03T17:07:41.8756720Z Downloading pydantic-2.12.3-py3-none-any.whl (462 kB) 2025-11-03T17:07:41.8863641Z Downloading pydantic_core-2.41.4-cp312-cp312-manylinux_2_17_x86_64.manylinux2014_x86_64.whl (2.1 MB) 2025-11-03T17:07:41.9070751Z [?25l ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 0.0/2.1 MB ? eta -:--:-- 2025-11-03T17:07:41.9071416Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 2.1/2.1 MB 100.4 MB/s 0:00:00 2025-11-03T17:07:41.9115679Z [?25hDownloading starlette-0.49.3-py3-none-any.whl (74 kB) 2025-11-03T17:07:41.9183425Z Downloading anyio-4.11.0-py3-none-any.whl (109 kB) 2025-11-03T17:07:41.9282867Z Downloading annotated_doc-0.0.3-py3-none-any.whl (5.5 kB) 2025-11-03T17:07:41.9343941Z Downloading annotated_types-0.7.0-py3-none-any.whl (13 kB) 2025-11-03T17:07:41.9426708Z Downloading httpx-0.28.1-py3-none-any.whl (73 kB) 2025-11-03T17:07:41.9488206Z Downloading httpcore-1.0.9-py3-none-any.whl (78 kB) 2025-11-03T17:07:41.9563754Z Downloading email_validator-2.3.0-py3-none-any.whl (35 kB) 2025-11-03T17:07:41.9642170Z Downloading dnspython-2.8.0-py3-none-any.whl (331 kB) 2025-11-03T17:07:41.9741824Z Downloading fastapi_cli-0.0.14-py3-none-any.whl (11 kB) 2025-11-03T17:07:41.9803736Z Downloading fastapi_cloud_cli-0.3.1-py3-none-any.whl (19 kB) 2025-11-03T17:07:41.9901204Z Downloading gguf-0.17.1-py3-none-any.whl (96 kB) 2025-11-03T17:07:41.9990128Z Downloading h11-0.16.0-py3-none-any.whl (37 kB) 2025-11-03T17:07:42.0074651Z Downloading interegular-0.3.3-py37-none-any.whl (23 kB) 2025-11-03T17:07:42.0174710Z Downloading mistral_common-1.8.5-py3-none-any.whl (6.5 MB) 2025-11-03T17:07:42.0646321Z [?25l ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 0.0/6.5 MB ? eta -:--:-- 2025-11-03T17:07:42.0647406Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 6.5/6.5 MB 140.6 MB/s 0:00:00 2025-11-03T17:07:42.0677374Z [?25hDownloading jsonschema-4.25.1-py3-none-any.whl (90 kB) 2025-11-03T17:07:42.0742333Z Downloading jsonschema_specifications-2025.9.1-py3-none-any.whl (18 kB) 2025-11-03T17:07:42.0830673Z Downloading nvidia_cudnn_frontend-1.15.0-cp312-cp312-manylinux_2_27_x86_64.manylinux_2_28_x86_64.whl (1.9 MB) 2025-11-03T17:07:42.1019521Z [?25l ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 0.0/1.9 MB ? eta -:--:-- 2025-11-03T17:07:42.1020195Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 1.9/1.9 MB 97.2 MB/s 0:00:00 2025-11-03T17:07:42.1097847Z [?25hDownloading nvidia_cutlass_dsl-4.2.1-cp312-cp312-manylinux_2_28_x86_64.whl (62.2 MB) 2025-11-03T17:07:42.3122054Z [?25l ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 0.0/62.2 MB ? eta -:--:-- 2025-11-03T17:07:42.5138731Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╸━━━━━━━━━━ 46.1/62.2 MB 233.1 MB/s eta 0:00:01 2025-11-03T17:07:42.6066038Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╸ 62.1/62.2 MB 233.9 MB/s eta 0:00:01 2025-11-03T17:07:42.6066997Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 62.2/62.2 MB 125.3 MB/s 0:00:00 2025-11-03T17:07:42.6108391Z [?25hDownloading cuda_python-13.0.3-py3-none-any.whl (7.6 kB) 2025-11-03T17:07:42.6253022Z Downloading cuda_bindings-13.0.3-cp312-cp312-manylinux_2_24_x86_64.manylinux_2_28_x86_64.whl (12.1 MB) 2025-11-03T17:07:42.7495744Z [?25l ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 0.0/12.1 MB ? eta -:--:-- 2025-11-03T17:07:42.7496440Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 12.1/12.1 MB 97.7 MB/s 0:00:00 2025-11-03T17:07:42.7531559Z [?25hDownloading openai-2.6.1-py3-none-any.whl (1.0 MB) 2025-11-03T17:07:42.7779105Z [?25l ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 0.0/1.0 MB ? eta -:--:-- 2025-11-03T17:07:42.7779746Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 1.0/1.0 MB 34.9 MB/s 0:00:00 2025-11-03T17:07:42.7816451Z [?25hDownloading jiter-0.11.1-cp312-cp312-manylinux_2_17_x86_64.manylinux2014_x86_64.whl (358 kB) 2025-11-03T17:07:42.7954700Z Using cached typing_extensions-4.15.0-py3-none-any.whl (44 kB) 2025-11-03T17:07:42.8026143Z Downloading openai_harmony-0.0.4-cp38-abi3-manylinux_2_17_x86_64.manylinux2014_x86_64.whl (3.0 MB) 2025-11-03T17:07:42.8302691Z [?25l ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 0.0/3.0 MB ? eta -:--:-- 2025-11-03T17:07:42.8303333Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 3.0/3.0 MB 110.0 MB/s 0:00:00 2025-11-03T17:07:42.8364803Z [?25hDownloading opencv_python_headless-4.12.0.88-cp37-abi3-manylinux2014_x86_64.manylinux_2_17_x86_64.whl (54.0 MB) 2025-11-03T17:07:43.0386657Z [?25l ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 0.0/54.0 MB ? eta -:--:-- 2025-11-03T17:07:43.2073088Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╸━━━━━━ 45.6/54.0 MB 227.5 MB/s eta 0:00:01 2025-11-03T17:07:43.2073973Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 54.0/54.0 MB 146.0 MB/s 0:00:00 2025-11-03T17:07:43.2165681Z [?25hDownloading numpy-2.2.6-cp312-cp312-manylinux_2_17_x86_64.manylinux2014_x86_64.whl (16.5 MB) 2025-11-03T17:07:43.3388493Z [?25l ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 0.0/16.5 MB ? eta -:--:-- 2025-11-03T17:07:43.3389167Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 16.5/16.5 MB 136.0 MB/s 0:00:00 2025-11-03T17:07:43.3398462Z [?25hUsing cached packaging-25.0-py3-none-any.whl (66 kB) 2025-11-03T17:07:43.3448049Z Downloading prometheus_client-0.23.1-py3-none-any.whl (61 kB) 2025-11-03T17:07:43.3529283Z Downloading prometheus_fastapi_instrumentator-7.1.0-py3-none-any.whl (19 kB) 2025-11-03T17:07:43.3595597Z Downloading pydantic_extra_types-2.10.6-py3-none-any.whl (40 kB) 2025-11-03T17:07:43.3671612Z Downloading pycountry-24.6.1-py3-none-any.whl (6.3 MB) 2025-11-03T17:07:43.3978075Z [?25l ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 0.0/6.3 MB ? eta -:--:-- 2025-11-03T17:07:43.3978734Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 6.3/6.3 MB 216.0 MB/s 0:00:00 2025-11-03T17:07:43.4013731Z [?25hDownloading python_multipart-0.0.20-py3-none-any.whl (24 kB) 2025-11-03T17:07:43.4084520Z Downloading pyzmq-27.1.0-cp312-abi3-manylinux_2_26_x86_64.manylinux_2_28_x86_64.whl (840 kB) 2025-11-03T17:07:43.4197214Z [?25l ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 0.0/841.0 kB ? eta -:--:-- 2025-11-03T17:07:43.4197906Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 841.0/841.0 kB 69.2 MB/s 0:00:00 2025-11-03T17:07:43.4285471Z [?25hDownloading ray-2.51.1-cp312-cp312-manylinux2014_x86_64.whl (71.4 MB) 2025-11-03T17:07:43.6311897Z [?25l ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 0.0/71.4 MB ? eta -:--:-- 2025-11-03T17:07:43.8323696Z  ━━━━━━━━━━━━━━━━━━━━━━━━╸━━━━━━━━━━━━━━━ 44.0/71.4 MB 222.6 MB/s eta 0:00:01 2025-11-03T17:07:43.9708660Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╸ 71.3/71.4 MB 187.3 MB/s eta 0:00:01 2025-11-03T17:07:43.9709397Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 71.4/71.4 MB 131.7 MB/s 0:00:00 2025-11-03T17:07:43.9745921Z [?25hDownloading click-8.2.1-py3-none-any.whl (102 kB) 2025-11-03T17:07:43.9816334Z Downloading referencing-0.37.0-py3-none-any.whl (26 kB) 2025-11-03T17:07:43.9886126Z Downloading rich_toolkit-0.15.1-py3-none-any.whl (29 kB) 2025-11-03T17:07:43.9956014Z Downloading rich-14.2.0-py3-none-any.whl (243 kB) 2025-11-03T17:07:44.0036770Z Downloading markdown_it_py-4.0.0-py3-none-any.whl (87 kB) 2025-11-03T17:07:44.0110227Z Downloading mdurl-0.1.2-py3-none-any.whl (10.0 kB) 2025-11-03T17:07:44.0205055Z Downloading rignore-0.7.4-cp312-cp312-manylinux_2_17_x86_64.manylinux2014_x86_64.whl (959 kB) 2025-11-03T17:07:44.0300477Z [?25l ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 0.0/959.8 kB ? eta -:--:-- 2025-11-03T17:07:44.0301147Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 959.8/959.8 kB 100.8 MB/s 0:00:00 2025-11-03T17:07:44.0334850Z [?25hDownloading rpds_py-0.28.0-cp312-cp312-manylinux_2_17_x86_64.manylinux2014_x86_64.whl (386 kB) 2025-11-03T17:07:44.0444975Z Downloading sentry_sdk-2.43.0-py2.py3-none-any.whl (400 kB) 2025-11-03T17:07:44.0528849Z Downloading sniffio-1.3.1-py3-none-any.whl (10 kB) 2025-11-03T17:07:44.0594627Z Downloading tiktoken-0.12.0-cp312-cp312-manylinux_2_28_x86_64.whl (1.2 MB) 2025-11-03T17:07:44.0723824Z [?25l ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 0.0/1.2 MB ? eta -:--:-- 2025-11-03T17:07:44.0724333Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 1.2/1.2 MB 84.8 MB/s 0:00:00 2025-11-03T17:07:44.0758919Z [?25hDownloading regex-2025.10.23-cp312-cp312-manylinux2014_x86_64.manylinux_2_17_x86_64.manylinux_2_28_x86_64.whl (803 kB) 2025-11-03T17:07:44.0845436Z [?25l ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 0.0/803.4 kB ? eta -:--:-- 2025-11-03T17:07:44.0846348Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 803.4/803.4 kB 90.3 MB/s 0:00:00 2025-11-03T17:07:44.0879381Z [?25hDownloading tokenizers-0.22.1-cp39-abi3-manylinux_2_17_x86_64.manylinux2014_x86_64.whl (3.3 MB) 2025-11-03T17:07:44.1061779Z [?25l ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 0.0/3.3 MB ? eta -:--:-- 2025-11-03T17:07:44.1062433Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 3.3/3.3 MB 193.5 MB/s 0:00:00 2025-11-03T17:07:44.1100337Z [?25hDownloading hf_xet-1.2.0-cp37-abi3-manylinux_2_17_x86_64.manylinux2014_x86_64.whl (3.3 MB) 2025-11-03T17:07:44.1359926Z [?25l ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 0.0/3.3 MB ? eta -:--:-- 2025-11-03T17:07:44.1360570Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 3.3/3.3 MB 130.6 MB/s 0:00:00 2025-11-03T17:07:44.1395435Z [?25hDownloading transformers-4.57.1-py3-none-any.whl (12.0 MB) 2025-11-03T17:07:44.1926906Z [?25l ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 0.0/12.0 MB ? eta -:--:-- 2025-11-03T17:07:44.1927569Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 12.0/12.0 MB 236.4 MB/s 0:00:00 2025-11-03T17:07:44.1965306Z [?25hDownloading huggingface_hub-0.36.0-py3-none-any.whl (566 kB) 2025-11-03T17:07:44.2050634Z [?25l ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 0.0/566.1 kB ? eta -:--:-- 2025-11-03T17:07:44.2051304Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 566.1/566.1 kB 50.4 MB/s 0:00:00 2025-11-03T17:07:44.2086088Z [?25hDownloading safetensors-0.6.2-cp38-abi3-manylinux_2_17_x86_64.manylinux2014_x86_64.whl (485 kB) 2025-11-03T17:07:44.2183248Z Downloading typer-0.20.0-py3-none-any.whl (47 kB) 2025-11-03T17:07:44.2254833Z Downloading shellingham-1.5.4-py2.py3-none-any.whl (9.8 kB) 2025-11-03T17:07:44.2324195Z Downloading typing_inspection-0.4.2-py3-none-any.whl (14 kB) 2025-11-03T17:07:44.2385212Z Downloading uvicorn-0.38.0-py3-none-any.whl (68 kB) 2025-11-03T17:07:44.2475964Z Downloading httptools-0.7.1-cp312-cp312-manylinux1_x86_64.manylinux_2_28_x86_64.manylinux_2_5_x86_64.whl (517 kB) 2025-11-03T17:07:44.2561956Z Downloading python_dotenv-1.2.1-py3-none-any.whl (21 kB) 2025-11-03T17:07:44.2653868Z Downloading uvloop-0.22.1-cp312-cp312-manylinux2014_x86_64.manylinux_2_17_x86_64.manylinux_2_28_x86_64.whl (4.4 MB) 2025-11-03T17:07:44.2947153Z [?25l ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 0.0/4.4 MB ? eta -:--:-- 2025-11-03T17:07:44.2947795Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 4.4/4.4 MB 155.1 MB/s 0:00:00 2025-11-03T17:07:44.3013152Z [?25hDownloading watchfiles-1.1.1-cp312-cp312-manylinux_2_17_x86_64.manylinux2014_x86_64.whl (456 kB) 2025-11-03T17:07:44.3104180Z Downloading websockets-15.0.1-cp312-cp312-manylinux_2_5_x86_64.manylinux1_x86_64.manylinux_2_17_x86_64.manylinux2014_x86_64.whl (182 kB) 2025-11-03T17:07:44.3172510Z Downloading astor-0.8.1-py2.py3-none-any.whl (27 kB) 2025-11-03T17:07:44.3266219Z Downloading blake3-1.0.8-cp312-cp312-manylinux_2_17_x86_64.manylinux2014_x86_64.whl (388 kB) 2025-11-03T17:07:44.3366999Z Downloading cbor2-5.7.1-cp312-cp312-manylinux2014_x86_64.manylinux_2_17_x86_64.manylinux_2_28_x86_64.whl (285 kB) 2025-11-03T17:07:44.3443566Z Downloading cloudpickle-3.1.2-py3-none-any.whl (22 kB) 2025-11-03T17:07:44.3547206Z Downloading cupy_cuda12x-13.6.0-cp312-cp312-manylinux2014_x86_64.whl (112.9 MB) 2025-11-03T17:07:44.5577051Z [?25l ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 0.0/112.9 MB ? eta -:--:-- 2025-11-03T17:07:44.8444228Z  ━━━━━━━━━━━━━━╸━━━━━━━━━━━━━━━━━━━━━━━━━ 41.9/112.9 MB 209.5 MB/s eta 0:00:01 2025-11-03T17:07:45.0460227Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╺━━━━━━━━━ 86.0/112.9 MB 214.0 MB/s eta 0:00:01 2025-11-03T17:07:45.1798546Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╸ 112.7/112.9 MB 191.3 MB/s eta 0:00:01 2025-11-03T17:07:45.1799837Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 112.9/112.9 MB 137.0 MB/s 0:00:00 2025-11-03T17:07:45.1856296Z [?25hDownloading fastrlock-0.8.3-cp312-cp312-manylinux_2_5_x86_64.manylinux1_x86_64.manylinux_2_28_x86_64.whl (53 kB) 2025-11-03T17:07:45.1936956Z Downloading einops-0.8.1-py3-none-any.whl (64 kB) 2025-11-03T17:07:45.2009692Z Downloading frozendict-2.4.6-py312-none-any.whl (16 kB) 2025-11-03T17:07:45.2074297Z Downloading msgspec-0.19.0-cp312-cp312-manylinux_2_17_x86_64.manylinux2014_x86_64.whl (213 kB) 2025-11-03T17:07:45.2153886Z Downloading nvidia_ml_py-13.580.82-py3-none-any.whl (49 kB) 2025-11-03T17:07:45.2242938Z Downloading partial_json_parser-0.2.1.1.post6-py3-none-any.whl (10 kB) 2025-11-03T17:07:45.2304759Z Downloading py_cpuinfo-9.0.0-py3-none-any.whl (22 kB) 2025-11-03T17:07:45.2389892Z Downloading pybase64-1.4.2-cp312-cp312-manylinux1_x86_64.manylinux2014_x86_64.manylinux_2_17_x86_64.manylinux_2_5_x86_64.whl (71 kB) 2025-11-03T17:07:45.2453320Z Downloading python_json_logger-4.0.0-py3-none-any.whl (15 kB) 2025-11-03T17:07:45.2538676Z Downloading sentencepiece-0.2.1-cp312-cp312-manylinux_2_27_x86_64.manylinux_2_28_x86_64.whl (1.4 MB) 2025-11-03T17:07:45.2653148Z [?25l ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 0.0/1.4 MB ? eta -:--:-- 2025-11-03T17:07:45.2653806Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 1.4/1.4 MB 129.3 MB/s 0:00:00 2025-11-03T17:07:45.2703044Z [?25hDownloading setproctitle-1.3.7-cp312-cp312-manylinux1_x86_64.manylinux_2_28_x86_64.manylinux_2_5_x86_64.whl (32 kB) 2025-11-03T17:07:45.6576264Z Building wheels for collected packages: flashinfer-python 2025-11-03T17:07:48.3086129Z Building wheel for flashinfer-python (pyproject.toml) ... [?25l- \ | / - \ | / - \ | / - \ | / done 2025-11-03T17:07:48.3137736Z [?25h Created wheel for flashinfer-python: filename=flashinfer_python-0.4.0-py3-none-any.whl size=6764886 sha256=01b7c72e4e3a110b5ba9f57d807d23c648733e392b5f09a9048e9833ccd46dc6 2025-11-03T17:07:48.3138873Z Stored in directory: /var/lib/jenkins/.cache/pip/wheels/86/5d/0d/2752ded4190ef42ea316eafbbf63e32fd02342fe0764a0dfc6 2025-11-03T17:07:48.3220762Z Successfully built flashinfer-python 2025-11-03T17:07:48.6294127Z Installing collected packages: py-cpuinfo, nvidia-ml-py, fastrlock, websockets, uvloop, typing_extensions, sniffio, shellingham, setproctitle, sentry-sdk, sentencepiece, safetensors, rpds-py, rignore, regex, pyzmq, python-multipart, python-json-logger, python-dotenv, pycountry, pybase64, prometheus_client, partial-json-parser, packaging, outlines_core, nvidia-cudnn-frontend, numpy, msgspec, mdurl, llvmlite, llguidance, lark, jiter, interegular, httptools, hf-xet, h11, frozendict, einops, dnspython, diskcache, cuda-bindings, cloudpickle, click, cbor2, blake3, astor, annotated-types, annotated-doc, uvicorn, typing-inspection, tiktoken, referencing, pydantic-core, opencv-python-headless, numba, markdown-it-py, huggingface-hub, httpcore, gguf, email-validator, depyf, cupy-cuda12x, cuda-python, apache-tvm-ffi, anyio, watchfiles, tokenizers, starlette, rich, pydantic, nvidia-cutlass-dsl, jsonschema-specifications, httpx, typer, transformers, rich-toolkit, pydantic-extra-types, prometheus-fastapi-instrumentator, openai-harmony, openai, lm-format-enforcer, jsonschema, flashinfer-python, fastapi, xgrammar, ray, fastapi-cloud-cli, fastapi-cli, compressed-tensors, mistral_common, vllm 2025-11-03T17:07:48.7971543Z [?25l 2025-11-03T17:07:48.8498623Z  ━╸━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━  4/92 [uvloop] 2025-11-03T17:07:48.8499104Z  Attempting uninstall: typing_extensions 2025-11-03T17:07:48.8555237Z ━╸━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━  4/92 [uvloop] 2025-11-03T17:07:48.8555771Z  Found existing installation: typing_extensions 4.12.2 2025-11-03T17:07:48.8575500Z ━╸━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━  4/92 [uvloop] 2025-11-03T17:07:48.8576292Z  Uninstalling typing_extensions-4.12.2: 2025-11-03T17:07:48.9646592Z ━╸━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━  4/92 [uvloop] 2025-11-03T17:07:49.1329699Z  ━━╺━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━  5/92 [typing_extensions] 2025-11-03T17:07:49.3003359Z  ━━╺━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━  5/92 [typing_extensions] 2025-11-03T17:07:49.3361439Z  ━━╺━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━  5/92 [typing_extensions] 2025-11-03T17:07:49.3362028Z  Successfully uninstalled typing_extensions-4.12.2 2025-11-03T17:07:49.4685934Z ━━╺━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━  5/92 [typing_extensions] 2025-11-03T17:07:49.6401069Z  ━━━╸━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━  9/92 [sentry-sdk] 2025-11-03T17:07:49.8078007Z  ━━━━╺━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 10/92 [sentencepiece] 2025-11-03T17:07:49.9753858Z  ━━━━━━╸━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 15/92 [pyzmq] 2025-11-03T17:07:50.1429877Z  ━━━━━━━━╺━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 19/92 [pycountry] 2025-11-03T17:07:50.1461400Z  ━━━━━━━━━╸━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 22/92 [partial-json-parser] 2025-11-03T17:07:50.1462062Z  Attempting uninstall: packaging 2025-11-03T17:07:50.1507674Z ━━━━━━━━━╸━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 22/92 [partial-json-parser] 2025-11-03T17:07:50.1508335Z  Found existing installation: packaging 23.1 2025-11-03T17:07:50.1540542Z ━━━━━━━━━╸━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 22/92 [partial-json-parser] 2025-11-03T17:07:50.1541077Z  Uninstalling packaging-23.1: 2025-11-03T17:07:50.1642187Z ━━━━━━━━━╸━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 22/92 [partial-json-parser] 2025-11-03T17:07:50.1642797Z  Successfully uninstalled packaging-23.1 2025-11-03T17:07:50.2772741Z ━━━━━━━━━╸━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 22/92 [partial-json-parser] 2025-11-03T17:07:50.2773250Z  Attempting uninstall: numpy 2025-11-03T17:07:50.2853287Z ━━━━━━━━━╸━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 22/92 [partial-json-parser] 2025-11-03T17:07:50.2853835Z  Found existing installation: numpy 1.26.2 2025-11-03T17:07:50.3103947Z ━━━━━━━━━╸━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 22/92 [partial-json-parser] 2025-11-03T17:07:50.3405324Z  ━━━━━━━━━━━╺━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 26/92 [numpy] 2025-11-03T17:07:50.3405907Z  Uninstalling numpy-1.26.2: 2025-11-03T17:07:50.4780318Z ━━━━━━━━━━━╺━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 26/92 [numpy] 2025-11-03T17:07:50.6455328Z  ━━━━━━━━━━━╺━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 26/92 [numpy] 2025-11-03T17:07:50.6865300Z  ━━━━━━━━━━━╺━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 26/92 [numpy] 2025-11-03T17:07:50.6865826Z  Successfully uninstalled numpy-1.26.2 2025-11-03T17:07:50.8130795Z ━━━━━━━━━━━╺━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 26/92 [numpy] 2025-11-03T17:07:50.9807952Z  ━━━━━━━━━━━╺━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 26/92 [numpy] 2025-11-03T17:07:51.1570639Z  ━━━━━━━━━━━╺━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 26/92 [numpy] 2025-11-03T17:07:51.3300746Z  ━━━━━━━━━━━╺━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 26/92 [numpy] 2025-11-03T17:07:51.4977019Z  ━━━━━━━━━━━╺━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 26/92 [numpy] 2025-11-03T17:07:51.6695891Z  ━━━━━━━━━━━╺━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 26/92 [numpy] 2025-11-03T17:07:51.8371512Z  ━━━━━━━━━━━╺━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 26/92 [numpy] 2025-11-03T17:07:52.0066988Z  ━━━━━━━━━━━╺━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 26/92 [numpy] 2025-11-03T17:07:52.1742796Z  ━━━━━━━━━━━╺━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 26/92 [numpy] 2025-11-03T17:07:52.2302608Z  ━━━━━━━━━━━╺━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 26/92 [numpy] 2025-11-03T17:07:52.2303085Z  Attempting uninstall: llvmlite 2025-11-03T17:07:52.2343386Z ━━━━━━━━━━━╺━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 26/92 [numpy] 2025-11-03T17:07:52.2343934Z  Found existing installation: llvmlite 0.43.0 2025-11-03T17:07:52.2394200Z ━━━━━━━━━━━╺━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 26/92 [numpy] 2025-11-03T17:07:52.2394665Z  Uninstalling llvmlite-0.43.0: 2025-11-03T17:07:52.3248011Z ━━━━━━━━━━━╺━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 26/92 [numpy] 2025-11-03T17:07:52.3248659Z  Successfully uninstalled llvmlite-0.43.0 2025-11-03T17:07:52.3417245Z ━━━━━━━━━━━╺━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 26/92 [numpy] 2025-11-03T17:07:52.5094052Z  ━━━━━━━━━━━━╸━━━━━━━━━━━━━━━━━━━━━━━━━━━ 29/92 [llvmlite] 2025-11-03T17:07:52.6770829Z  ━━━━━━━━━━━━╸━━━━━━━━━━━━━━━━━━━━━━━━━━━ 29/92 [llvmlite] 2025-11-03T17:07:52.8458273Z  ━━━━━━━━━━━━╸━━━━━━━━━━━━━━━━━━━━━━━━━━━ 29/92 [llvmlite] 2025-11-03T17:07:53.0133613Z  ━━━━━━━━━━━━╸━━━━━━━━━━━━━━━━━━━━━━━━━━━ 29/92 [llvmlite] 2025-11-03T17:07:53.0951703Z  ━━━━━━━━━━━━━╺━━━━━━━━━━━━━━━━━━━━━━━━━━ 30/92 [llguidance] 2025-11-03T17:07:53.0952469Z  Attempting uninstall: lark 2025-11-03T17:07:53.0991467Z ━━━━━━━━━━━━━╺━━━━━━━━━━━━━━━━━━━━━━━━━━ 30/92 [llguidance] 2025-11-03T17:07:53.0992119Z  Found existing installation: lark 0.12.0 2025-11-03T17:07:53.1048635Z ━━━━━━━━━━━━━╺━━━━━━━━━━━━━━━━━━━━━━━━━━ 30/92 [llguidance] 2025-11-03T17:07:53.1049065Z  Uninstalling lark-0.12.0: 2025-11-03T17:07:53.1291633Z ━━━━━━━━━━━━━╺━━━━━━━━━━━━━━━━━━━━━━━━━━ 30/92 [llguidance] 2025-11-03T17:07:53.1292129Z  Successfully uninstalled lark-0.12.0 2025-11-03T17:07:53.1810004Z ━━━━━━━━━━━━━╺━━━━━━━━━━━━━━━━━━━━━━━━━━ 30/92 [llguidance] 2025-11-03T17:07:53.3500404Z  ━━━━━━━━━━━━━╺━━━━━━━━━━━━━━━━━━━━━━━━━━ 31/92 [lark] 2025-11-03T17:07:53.5179657Z  ━━━━━━━━━━━━━━━━╸━━━━━━━━━━━━━━━━━━━━━━━ 38/92 [einops] 2025-11-03T17:07:53.5708199Z  ━━━━━━━━━━━━━━━━╸━━━━━━━━━━━━━━━━━━━━━━━ 39/92 [dnspython] 2025-11-03T17:07:53.5708703Z  Attempting uninstall: cuda-bindings 2025-11-03T17:07:53.5746750Z ━━━━━━━━━━━━━━━━╸━━━━━━━━━━━━━━━━━━━━━━━ 39/92 [dnspython] 2025-11-03T17:07:53.5747363Z  Found existing installation: cuda-bindings 12.9.4 2025-11-03T17:07:53.5811205Z ━━━━━━━━━━━━━━━━╸━━━━━━━━━━━━━━━━━━━━━━━ 39/92 [dnspython] 2025-11-03T17:07:53.5811666Z  Uninstalling cuda-bindings-12.9.4: 2025-11-03T17:07:53.6854342Z ━━━━━━━━━━━━━━━━╸━━━━━━━━━━━━━━━━━━━━━━━ 39/92 [dnspython] 2025-11-03T17:07:53.8530850Z  ━━━━━━━━━━━━━━━━━╸━━━━━━━━━━━━━━━━━━━━━━ 41/92 [cuda-bindings] 2025-11-03T17:07:54.0208496Z  ━━━━━━━━━━━━━━━━━╸━━━━━━━━━━━━━━━━━━━━━━ 41/92 [cuda-bindings] 2025-11-03T17:07:54.1017171Z  ━━━━━━━━━━━━━━━━━╸━━━━━━━━━━━━━━━━━━━━━━ 41/92 [cuda-bindings] 2025-11-03T17:07:54.1017706Z  Successfully uninstalled cuda-bindings-12.9.4 2025-11-03T17:07:54.1884506Z ━━━━━━━━━━━━━━━━━╸━━━━━━━━━━━━━━━━━━━━━━ 41/92 [cuda-bindings] 2025-11-03T17:07:54.3560822Z  ━━━━━━━━━━━━━━━━━╸━━━━━━━━━━━━━━━━━━━━━━ 41/92 [cuda-bindings] 2025-11-03T17:07:54.3703483Z  ━━━━━━━━━━━━━━━━━╸━━━━━━━━━━━━━━━━━━━━━━ 41/92 [cuda-bindings] 2025-11-03T17:07:54.3703926Z  Attempting uninstall: click 2025-11-03T17:07:54.3736728Z ━━━━━━━━━━━━━━━━━╸━━━━━━━━━━━━━━━━━━━━━━ 41/92 [cuda-bindings] 2025-11-03T17:07:54.3737267Z  Found existing installation: click 8.3.0 2025-11-03T17:07:54.3772001Z ━━━━━━━━━━━━━━━━━╸━━━━━━━━━━━━━━━━━━━━━━ 41/92 [cuda-bindings] 2025-11-03T17:07:54.3772486Z  Uninstalling click-8.3.0: 2025-11-03T17:07:54.3891546Z ━━━━━━━━━━━━━━━━━╸━━━━━━━━━━━━━━━━━━━━━━ 41/92 [cuda-bindings] 2025-11-03T17:07:54.3892048Z  Successfully uninstalled click-8.3.0 2025-11-03T17:07:54.5234722Z ━━━━━━━━━━━━━━━━━╸━━━━━━━━━━━━━━━━━━━━━━ 41/92 [cuda-bindings] 2025-11-03T17:07:54.6911554Z  ━━━━━━━━━━━━━━━━━━━━━╺━━━━━━━━━━━━━━━━━━ 49/92 [uvicorn] 2025-11-03T17:07:54.8586334Z  ━━━━━━━━━━━━━━━━━━━━━━━╺━━━━━━━━━━━━━━━━ 54/92 [opencv-python-headless] 2025-11-03T17:07:55.0261415Z  ━━━━━━━━━━━━━━━━━━━━━━━╺━━━━━━━━━━━━━━━━ 54/92 [opencv-python-headless] 2025-11-03T17:07:55.1936632Z  ━━━━━━━━━━━━━━━━━━━━━━━╺━━━━━━━━━━━━━━━━ 54/92 [opencv-python-headless] 2025-11-03T17:07:55.2516113Z  ━━━━━━━━━━━━━━━━━━━━━━━╺━━━━━━━━━━━━━━━━ 54/92 [opencv-python-headless] 2025-11-03T17:07:55.2516663Z  Attempting uninstall: numba 2025-11-03T17:07:55.2561242Z ━━━━━━━━━━━━━━━━━━━━━━━╺━━━━━━━━━━━━━━━━ 54/92 [opencv-python-headless] 2025-11-03T17:07:55.2561806Z  Found existing installation: numba 0.60.0 2025-11-03T17:07:55.3306338Z ━━━━━━━━━━━━━━━━━━━━━━━╺━━━━━━━━━━━━━━━━ 54/92 [opencv-python-headless] 2025-11-03T17:07:55.3306848Z  Uninstalling numba-0.60.0: 2025-11-03T17:07:55.3610695Z ━━━━━━━━━━━━━━━━━━━━━━━╺━━━━━━━━━━━━━━━━ 54/92 [opencv-python-headless] 2025-11-03T17:07:55.5290627Z  ━━━━━━━━━━━━━━━━━━━━━━━╸━━━━━━━━━━━━━━━━ 55/92 [numba] 2025-11-03T17:07:55.6965291Z  ━━━━━━━━━━━━━━━━━━━━━━━╸━━━━━━━━━━━━━━━━ 55/92 [numba] 2025-11-03T17:07:55.7235232Z  ━━━━━━━━━━━━━━━━━━━━━━━╸━━━━━━━━━━━━━━━━ 55/92 [numba] 2025-11-03T17:07:55.7236099Z  Successfully uninstalled numba-0.60.0 2025-11-03T17:07:55.8640543Z ━━━━━━━━━━━━━━━━━━━━━━━╸━━━━━━━━━━━━━━━━ 55/92 [numba] 2025-11-03T17:07:56.0427512Z  ━━━━━━━━━━━━━━━━━━━━━━━╸━━━━━━━━━━━━━━━━ 55/92 [numba] 2025-11-03T17:07:56.2102935Z  ━━━━━━━━━━━━━━━━━━━━━━━╸━━━━━━━━━━━━━━━━ 55/92 [numba] 2025-11-03T17:07:56.3784120Z  ━━━━━━━━━━━━━━━━━━━━━━━╸━━━━━━━━━━━━━━━━ 55/92 [numba] 2025-11-03T17:07:56.5463400Z  ━━━━━━━━━━━━━━━━━━━━━━━╸━━━━━━━━━━━━━━━━ 55/92 [numba] 2025-11-03T17:07:56.7148730Z  ━━━━━━━━━━━━━━━━━━━━━━━╸━━━━━━━━━━━━━━━━ 55/92 [numba] 2025-11-03T17:07:56.8825717Z  ━━━━━━━━━━━━━━━━━━━━━━━╸━━━━━━━━━━━━━━━━ 55/92 [numba] 2025-11-03T17:07:57.0594232Z  ━━━━━━━━━━━━━━━━━━━━━━━╸━━━━━━━━━━━━━━━━ 55/92 [numba] 2025-11-03T17:07:57.2270528Z  ━━━━━━━━━━━━━━━━━━━━━━━╸━━━━━━━━━━━━━━━━ 55/92 [numba] 2025-11-03T17:07:57.3946625Z  ━━━━━━━━━━━━━━━━━━━━━━━╸━━━━━━━━━━━━━━━━ 55/92 [numba] 2025-11-03T17:07:57.5623004Z  ━━━━━━━━━━━━━━━━━━━━━━━╸━━━━━━━━━━━━━━━━ 55/92 [numba] 2025-11-03T17:07:57.7312468Z  ━━━━━━━━━━━━━━━━━━━━━━━╸━━━━━━━━━━━━━━━━ 55/92 [numba] 2025-11-03T17:07:57.8994192Z  ━━━━━━━━━━━━━━━━━━━━━━━━╸━━━━━━━━━━━━━━━ 57/92 [huggingface-hub] 2025-11-03T17:07:58.0672374Z  ━━━━━━━━━━━━━━━━━━━━━━━━╸━━━━━━━━━━━━━━━ 57/92 [huggingface-hub] 2025-11-03T17:07:58.2348978Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━╸━━━━━━━━━━━━━ 61/92 [depyf] 2025-11-03T17:07:58.4025033Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━╸━━━━━━━━━━━━━ 62/92 [cupy-cuda12x] 2025-11-03T17:07:58.5701681Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━╸━━━━━━━━━━━━━ 62/92 [cupy-cuda12x] 2025-11-03T17:07:58.7377120Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━╸━━━━━━━━━━━━━ 62/92 [cupy-cuda12x] 2025-11-03T17:07:58.9053587Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━╸━━━━━━━━━━━━━ 62/92 [cupy-cuda12x] 2025-11-03T17:07:59.0728499Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━╸━━━━━━━━━━━━━ 62/92 [cupy-cuda12x] 2025-11-03T17:07:59.2404005Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━╸━━━━━━━━━━━━━ 62/92 [cupy-cuda12x] 2025-11-03T17:07:59.4079876Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━╸━━━━━━━━━━━━━ 62/92 [cupy-cuda12x] 2025-11-03T17:07:59.5766187Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━╸━━━━━━━━━━━━━ 62/92 [cupy-cuda12x] 2025-11-03T17:07:59.7441574Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━╸━━━━━━━━━━━━━ 62/92 [cupy-cuda12x] 2025-11-03T17:07:59.9117624Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━╸━━━━━━━━━━━━━ 62/92 [cupy-cuda12x] 2025-11-03T17:08:00.0793592Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━╸━━━━━━━━━━━━ 64/92 [apache-tvm-ffi] 2025-11-03T17:08:00.2485534Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╺━━━━━━━━━━ 67/92 [tokenizers] 2025-11-03T17:08:00.4176198Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╺━━━━━━━━━ 69/92 [rich] 2025-11-03T17:08:00.5852014Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╺━━━━━━━━━ 70/92 [pydantic] 2025-11-03T17:08:00.7527109Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╸━━━━━━━━━ 71/92 [nvidia-cutlass-dsl] 2025-11-03T17:08:00.9201983Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╸━━━━━━━━━ 71/92 [nvidia-cutlass-dsl] 2025-11-03T17:08:01.0877241Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╸━━━━━━━━━ 71/92 [nvidia-cutlass-dsl] 2025-11-03T17:08:01.2622888Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╸━━━━━━━━━ 71/92 [nvidia-cutlass-dsl] 2025-11-03T17:08:01.4299337Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╸━━━━━━━━━ 71/92 [nvidia-cutlass-dsl] 2025-11-03T17:08:01.5978896Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╸━━━━━━━━━ 71/92 [nvidia-cutlass-dsl] 2025-11-03T17:08:01.7653901Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╺━━━━━━━━ 72/92 [jsonschema-specifications] 2025-11-03T17:08:01.9330026Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╸━━━━━━━ 75/92 [transformers] 2025-11-03T17:08:02.1005285Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╸━━━━━━━ 75/92 [transformers] 2025-11-03T17:08:02.2680606Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╸━━━━━━━ 75/92 [transformers] 2025-11-03T17:08:02.4359577Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╸━━━━━━━ 75/92 [transformers] 2025-11-03T17:08:02.6034776Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╸━━━━━━━ 75/92 [transformers] 2025-11-03T17:08:02.7728547Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╸━━━━━━━ 75/92 [transformers] 2025-11-03T17:08:02.9406458Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╸━━━━━━━ 75/92 [transformers] 2025-11-03T17:08:03.1111908Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╸━━━━━━━ 75/92 [transformers] 2025-11-03T17:08:03.2831030Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╸━━━━━━━ 75/92 [transformers] 2025-11-03T17:08:03.4506237Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╸━━━━━━━ 75/92 [transformers] 2025-11-03T17:08:03.6182906Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╸━━━━━━━ 75/92 [transformers] 2025-11-03T17:08:03.7889373Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╸━━━━━━━ 75/92 [transformers] 2025-11-03T17:08:03.9597896Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╸━━━━━━━ 75/92 [transformers] 2025-11-03T17:08:04.1299842Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╸━━━━━━━ 75/92 [transformers] 2025-11-03T17:08:04.3008166Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╸━━━━━━━ 75/92 [transformers] 2025-11-03T17:08:04.4689299Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╸━━━━━━━ 75/92 [transformers] 2025-11-03T17:08:04.6373592Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╸━━━━━━━ 75/92 [transformers] 2025-11-03T17:08:04.8057242Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╸━━━━━━━ 75/92 [transformers] 2025-11-03T17:08:04.9744527Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╸━━━━━━━ 75/92 [transformers] 2025-11-03T17:08:05.1476845Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╸━━━━━━━ 75/92 [transformers] 2025-11-03T17:08:05.3224540Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╸━━━━━━━ 75/92 [transformers] 2025-11-03T17:08:05.4899706Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╸━━━━━━━ 75/92 [transformers] 2025-11-03T17:08:05.6580879Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╸━━━━━━━ 75/92 [transformers] 2025-11-03T17:08:05.8309846Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╸━━━━━━━ 75/92 [transformers] 2025-11-03T17:08:06.0009529Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╸━━━━━━━ 75/92 [transformers] 2025-11-03T17:08:06.1724828Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╸━━━━━━━ 75/92 [transformers] 2025-11-03T17:08:06.3400922Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╸━━━━━━━ 75/92 [transformers] 2025-11-03T17:08:06.5076178Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╸━━━━━━━ 75/92 [transformers] 2025-11-03T17:08:06.6772126Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╸━━━━━━━ 75/92 [transformers] 2025-11-03T17:08:06.8457666Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╸━━━━━━━ 75/92 [transformers] 2025-11-03T17:08:07.0177818Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╸━━━━━━━ 75/92 [transformers] 2025-11-03T17:08:07.1949968Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╸━━━━━━━ 75/92 [transformers] 2025-11-03T17:08:07.3626810Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╸━━━━━━━ 75/92 [transformers] 2025-11-03T17:08:07.5301802Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╸━━━━━━━ 75/92 [transformers] 2025-11-03T17:08:07.6984684Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╺━━━━━━ 77/92 [pydantic-extra-types] 2025-11-03T17:08:07.8657698Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╺━━━━━ 79/92 [openai-harmony] 2025-11-03T17:08:08.0335104Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╸━━━━━ 80/92 [openai] 2025-11-03T17:08:08.2011753Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╸━━━━━ 80/92 [openai] 2025-11-03T17:08:08.3692833Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╸━━━━━ 80/92 [openai] 2025-11-03T17:08:08.5366992Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╸━━━━━ 80/92 [openai] 2025-11-03T17:08:08.7116939Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╸━━━━━ 80/92 [openai] 2025-11-03T17:08:08.8793068Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╸━━━━ 82/92 [jsonschema] 2025-11-03T17:08:09.0467371Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╺━━━ 83/92 [flashinfer-python] 2025-11-03T17:08:09.2214987Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╺━━━ 83/92 [flashinfer-python] 2025-11-03T17:08:09.3935585Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╺━━━ 83/92 [flashinfer-python] 2025-11-03T17:08:09.5613453Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╺━━━ 83/92 [flashinfer-python] 2025-11-03T17:08:09.7295185Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╺━━━ 83/92 [flashinfer-python] 2025-11-03T17:08:09.8966581Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╺━━━ 83/92 [flashinfer-python] 2025-11-03T17:08:10.0648855Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╺━━━ 83/92 [flashinfer-python] 2025-11-03T17:08:10.2326536Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╸━━━ 84/92 [fastapi] 2025-11-03T17:08:10.4002752Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╸━━━ 85/92 [xgrammar] 2025-11-03T17:08:10.5677031Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╺━━ 86/92 [ray] 2025-11-03T17:08:10.7351965Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╺━━ 86/92 [ray] 2025-11-03T17:08:10.9027846Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╺━━ 86/92 [ray] 2025-11-03T17:08:11.0702795Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╺━━ 86/92 [ray] 2025-11-03T17:08:11.2395680Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╺━━ 86/92 [ray] 2025-11-03T17:08:11.4097258Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╺━━ 86/92 [ray] 2025-11-03T17:08:11.5785346Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╺━━ 86/92 [ray] 2025-11-03T17:08:11.7471049Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╺━━ 86/92 [ray] 2025-11-03T17:08:11.9168163Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╺━━ 86/92 [ray] 2025-11-03T17:08:12.0847973Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╺━━ 86/92 [ray] 2025-11-03T17:08:12.2522921Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╺━━ 86/92 [ray] 2025-11-03T17:08:12.4200528Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╺━━ 86/92 [ray] 2025-11-03T17:08:12.5876375Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╺━━ 86/92 [ray] 2025-11-03T17:08:12.7551827Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╺━━ 86/92 [ray] 2025-11-03T17:08:12.9228629Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╺━━ 86/92 [ray] 2025-11-03T17:08:13.0916089Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╺━━ 86/92 [ray] 2025-11-03T17:08:13.2591501Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╺━━ 86/92 [ray] 2025-11-03T17:08:13.4273863Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╺━━ 86/92 [ray] 2025-11-03T17:08:13.5948614Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╺━━ 86/92 [ray] 2025-11-03T17:08:13.7638325Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╺━━ 86/92 [ray] 2025-11-03T17:08:13.9317691Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╺━━ 86/92 [ray] 2025-11-03T17:08:14.0996230Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╺━━ 86/92 [ray] 2025-11-03T17:08:14.2671509Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╺ 90/92 [mistral_common] 2025-11-03T17:08:14.4348045Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╸ 91/92 [vllm] 2025-11-03T17:08:14.6024165Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╸ 91/92 [vllm] 2025-11-03T17:08:14.7699732Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╸ 91/92 [vllm] 2025-11-03T17:08:14.9376050Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╸ 91/92 [vllm] 2025-11-03T17:08:15.1051808Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╸ 91/92 [vllm] 2025-11-03T17:08:15.2728152Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╸ 91/92 [vllm] 2025-11-03T17:08:15.4404515Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╸ 91/92 [vllm] 2025-11-03T17:08:15.6081290Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╸ 91/92 [vllm] 2025-11-03T17:08:15.7757915Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╸ 91/92 [vllm] 2025-11-03T17:08:15.9435206Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╸ 91/92 [vllm] 2025-11-03T17:08:16.1110000Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╸ 91/92 [vllm] 2025-11-03T17:08:16.2785053Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╸ 91/92 [vllm] 2025-11-03T17:08:16.4460236Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╸ 91/92 [vllm] 2025-11-03T17:08:16.6138551Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╸ 91/92 [vllm] 2025-11-03T17:08:16.7812218Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╸ 91/92 [vllm] 2025-11-03T17:08:16.9486833Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╸ 91/92 [vllm] 2025-11-03T17:08:17.1161293Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╸ 91/92 [vllm] 2025-11-03T17:08:17.2837897Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╸ 91/92 [vllm] 2025-11-03T17:08:17.4513750Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╸ 91/92 [vllm] 2025-11-03T17:08:17.6190650Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╸ 91/92 [vllm] 2025-11-03T17:08:17.7867348Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╸ 91/92 [vllm] 2025-11-03T17:08:17.9543084Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╸ 91/92 [vllm] 2025-11-03T17:08:18.1219345Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╸ 91/92 [vllm] 2025-11-03T17:08:18.2894323Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╸ 91/92 [vllm] 2025-11-03T17:08:18.4568735Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╸ 91/92 [vllm] 2025-11-03T17:08:18.6244139Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╸ 91/92 [vllm] 2025-11-03T17:08:18.7919117Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╸ 91/92 [vllm] 2025-11-03T17:08:18.9593359Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╸ 91/92 [vllm] 2025-11-03T17:08:19.1268597Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╸ 91/92 [vllm] 2025-11-03T17:08:19.2942463Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╸ 91/92 [vllm] 2025-11-03T17:08:19.4618191Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╸ 91/92 [vllm] 2025-11-03T17:08:19.6293684Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╸ 91/92 [vllm] 2025-11-03T17:08:19.7974768Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╸ 91/92 [vllm] 2025-11-03T17:08:19.9664802Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╸ 91/92 [vllm] 2025-11-03T17:08:20.1375183Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╸ 91/92 [vllm] 2025-11-03T17:08:20.3106031Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╸ 91/92 [vllm] 2025-11-03T17:08:20.4793940Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╸ 91/92 [vllm] 2025-11-03T17:08:20.6515441Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╸ 91/92 [vllm] 2025-11-03T17:08:20.8191580Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╸ 91/92 [vllm] 2025-11-03T17:08:20.9870442Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╸ 91/92 [vllm] 2025-11-03T17:08:21.0966516Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╸ 91/92 [vllm] 2025-11-03T17:08:21.0967097Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 92/92 [vllm] 2025-11-03T17:08:21.0967448Z [?25h 2025-11-03T17:08:21.1117950Z Successfully installed annotated-doc-0.0.3 annotated-types-0.7.0 anyio-4.11.0 apache-tvm-ffi-0.1.0b15 astor-0.8.1 blake3-1.0.8 cbor2-5.7.1 click-8.2.1 cloudpickle-3.1.2 compressed-tensors-0.11.0 cuda-bindings-13.0.3 cuda-python-13.0.3 cupy-cuda12x-13.6.0 depyf-0.19.0 diskcache-5.6.3 dnspython-2.8.0 einops-0.8.1 email-validator-2.3.0 fastapi-0.121.0 fastapi-cli-0.0.14 fastapi-cloud-cli-0.3.1 fastrlock-0.8.3 flashinfer-python-0.4.0 frozendict-2.4.6 gguf-0.17.1 h11-0.16.0 hf-xet-1.2.0 httpcore-1.0.9 httptools-0.7.1 httpx-0.28.1 huggingface-hub-0.36.0 interegular-0.3.3 jiter-0.11.1 jsonschema-4.25.1 jsonschema-specifications-2025.9.1 lark-1.2.2 llguidance-0.7.30 llvmlite-0.44.0 lm-format-enforcer-0.11.3 markdown-it-py-4.0.0 mdurl-0.1.2 mistral_common-1.8.5 msgspec-0.19.0 numba-0.61.2 numpy-2.2.6 nvidia-cudnn-frontend-1.15.0 nvidia-cutlass-dsl-4.2.1 nvidia-ml-py-13.580.82 openai-2.6.1 openai-harmony-0.0.4 opencv-python-headless-4.12.0.88 outlines_core-0.2.11 packaging-25.0 partial-json-parser-0.2.1.1.post6 prometheus-fastapi-instrumentator-7.1.0 prometheus_client-0.23.1 py-cpuinfo-9.0.0 pybase64-1.4.2 pycountry-24.6.1 pydantic-2.12.3 pydantic-core-2.41.4 pydantic-extra-types-2.10.6 python-dotenv-1.2.1 python-json-logger-4.0.0 python-multipart-0.0.20 pyzmq-27.1.0 ray-2.51.1 referencing-0.37.0 regex-2025.10.23 rich-14.2.0 rich-toolkit-0.15.1 rignore-0.7.4 rpds-py-0.28.0 safetensors-0.6.2 sentencepiece-0.2.1 sentry-sdk-2.43.0 setproctitle-1.3.7 shellingham-1.5.4 sniffio-1.3.1 starlette-0.49.3 tiktoken-0.12.0 tokenizers-0.22.1 transformers-4.57.1 typer-0.20.0 typing-inspection-0.4.2 typing_extensions-4.15.0 uvicorn-0.38.0 uvloop-0.22.1 vllm-0.11.0rc2.dev389+ge51928192.d20251103 watchfiles-1.1.1 websockets-15.0.1 xgrammar-0.1.25 2025-11-03T17:08:21.9835165Z 2025-11-03 17:08:21,982 [INFO] cli.lib.common.pip_helper: Installing /var/lib/jenkins/workspace/dist/external/vllm/wheels/flashinfer-python/flashinfer_python-0.2.14.post1-cp39-abi3-linux_x86_64.whl... 2025-11-03T17:08:21.9837002Z 2025-11-03 17:08:21,983 [INFO] cli.lib.common.pip_helper: pip installing packages: /opt/conda/envs/py_3.12/bin/python -m pip install /var/lib/jenkins/workspace/dist/external/vllm/wheels/flashinfer-python/flashinfer_python-0.2.14.post1-cp39-abi3-linux_x86_64.whl 2025-11-03T17:08:21.9838793Z 2025-11-03 17:08:21,983 [INFO] cli.lib.common.utils: [cmd] /opt/conda/envs/py_3.12/bin/python -m pip install /var/lib/jenkins/workspace/dist/external/vllm/wheels/flashinfer-python/flashinfer_python-0.2.14.post1-cp39-abi3-linux_x86_64.whl 2025-11-03T17:08:22.2783372Z Processing /var/lib/jenkins/workspace/dist/external/vllm/wheels/flashinfer-python/flashinfer_python-0.2.14.post1-cp39-abi3-linux_x86_64.whl 2025-11-03T17:08:22.3531353Z Requirement already satisfied: numpy in /opt/conda/envs/py_3.12/lib/python3.12/site-packages (from flashinfer-python==0.2.14.post1) (2.2.6) 2025-11-03T17:08:22.3533216Z Requirement already satisfied: torch in /opt/conda/envs/py_3.12/lib/python3.12/site-packages (from flashinfer-python==0.2.14.post1) (2.10.0a0+git3f6538f) 2025-11-03T17:08:22.3535902Z Requirement already satisfied: ninja in /opt/conda/envs/py_3.12/lib/python3.12/site-packages (from flashinfer-python==0.2.14.post1) (1.10.2) 2025-11-03T17:08:22.3538499Z Requirement already satisfied: requests in /opt/conda/envs/py_3.12/lib/python3.12/site-packages (from flashinfer-python==0.2.14.post1) (2.32.5) 2025-11-03T17:08:22.4077544Z Collecting cuda-python<=12.9 (from flashinfer-python==0.2.14.post1) 2025-11-03T17:08:22.4269381Z Downloading cuda_python-12.9.0-py3-none-any.whl.metadata (4.6 kB) 2025-11-03T17:08:22.4436890Z Collecting pynvml (from flashinfer-python==0.2.14.post1) 2025-11-03T17:08:22.4469692Z Downloading pynvml-13.0.1-py3-none-any.whl.metadata (5.6 kB) 2025-11-03T17:08:22.4537879Z Requirement already satisfied: einops in /opt/conda/envs/py_3.12/lib/python3.12/site-packages (from flashinfer-python==0.2.14.post1) (0.8.1) 2025-11-03T17:08:22.4541733Z Requirement already satisfied: packaging>=24.2 in /opt/conda/envs/py_3.12/lib/python3.12/site-packages (from flashinfer-python==0.2.14.post1) (25.0) 2025-11-03T17:08:22.4545050Z Requirement already satisfied: nvidia-cudnn-frontend>=1.13.0 in /opt/conda/envs/py_3.12/lib/python3.12/site-packages (from flashinfer-python==0.2.14.post1) (1.15.0) 2025-11-03T17:08:22.4708438Z Collecting cuda-bindings~=12.9.0 (from cuda-python<=12.9->flashinfer-python==0.2.14.post1) 2025-11-03T17:08:22.4719148Z Using cached cuda_bindings-12.9.4-cp312-cp312-manylinux_2_24_x86_64.manylinux_2_28_x86_64.whl.metadata (2.6 kB) 2025-11-03T17:08:22.4747552Z Requirement already satisfied: cuda-pathfinder~=1.1 in /opt/conda/envs/py_3.12/lib/python3.12/site-packages (from cuda-bindings~=12.9.0->cuda-python<=12.9->flashinfer-python==0.2.14.post1) (1.3.2) 2025-11-03T17:08:22.4761765Z Requirement already satisfied: filelock in /opt/conda/envs/py_3.12/lib/python3.12/site-packages (from torch->flashinfer-python==0.2.14.post1) (3.18.0) 2025-11-03T17:08:22.4765721Z Requirement already satisfied: typing-extensions>=4.10.0 in /opt/conda/envs/py_3.12/lib/python3.12/site-packages (from torch->flashinfer-python==0.2.14.post1) (4.15.0) 2025-11-03T17:08:22.4770313Z Requirement already satisfied: setuptools in /opt/conda/envs/py_3.12/lib/python3.12/site-packages (from torch->flashinfer-python==0.2.14.post1) (78.1.1) 2025-11-03T17:08:22.4773875Z Requirement already satisfied: sympy>=1.13.3 in /opt/conda/envs/py_3.12/lib/python3.12/site-packages (from torch->flashinfer-python==0.2.14.post1) (1.13.3) 2025-11-03T17:08:22.4777371Z Requirement already satisfied: networkx>=2.5.1 in /opt/conda/envs/py_3.12/lib/python3.12/site-packages (from torch->flashinfer-python==0.2.14.post1) (2.8.8) 2025-11-03T17:08:22.4780013Z Requirement already satisfied: jinja2 in /opt/conda/envs/py_3.12/lib/python3.12/site-packages (from torch->flashinfer-python==0.2.14.post1) (3.1.6) 2025-11-03T17:08:22.4783506Z Requirement already satisfied: fsspec>=0.8.5 in /opt/conda/envs/py_3.12/lib/python3.12/site-packages (from torch->flashinfer-python==0.2.14.post1) (2025.9.0) 2025-11-03T17:08:22.4897054Z Requirement already satisfied: mpmath<1.4,>=1.1.0 in /opt/conda/envs/py_3.12/lib/python3.12/site-packages (from sympy>=1.13.3->torch->flashinfer-python==0.2.14.post1) (1.3.0) 2025-11-03T17:08:22.4945577Z Requirement already satisfied: MarkupSafe>=2.0 in /opt/conda/envs/py_3.12/lib/python3.12/site-packages (from jinja2->torch->flashinfer-python==0.2.14.post1) (3.0.3) 2025-11-03T17:08:22.4995022Z Requirement already satisfied: nvidia-ml-py>=12.0.0 in /opt/conda/envs/py_3.12/lib/python3.12/site-packages (from pynvml->flashinfer-python==0.2.14.post1) (13.580.82) 2025-11-03T17:08:22.5010449Z Requirement already satisfied: charset_normalizer<4,>=2 in /opt/conda/envs/py_3.12/lib/python3.12/site-packages (from requests->flashinfer-python==0.2.14.post1) (3.4.4) 2025-11-03T17:08:22.5014445Z Requirement already satisfied: idna<4,>=2.5 in /opt/conda/envs/py_3.12/lib/python3.12/site-packages (from requests->flashinfer-python==0.2.14.post1) (3.11) 2025-11-03T17:08:22.5018243Z Requirement already satisfied: urllib3<3,>=1.21.1 in /opt/conda/envs/py_3.12/lib/python3.12/site-packages (from requests->flashinfer-python==0.2.14.post1) (2.5.0) 2025-11-03T17:08:22.5021983Z Requirement already satisfied: certifi>=2017.4.17 in /opt/conda/envs/py_3.12/lib/python3.12/site-packages (from requests->flashinfer-python==0.2.14.post1) (2025.10.5) 2025-11-03T17:08:22.5165608Z Downloading cuda_python-12.9.0-py3-none-any.whl (7.5 kB) 2025-11-03T17:08:22.5228913Z Using cached cuda_bindings-12.9.4-cp312-cp312-manylinux_2_24_x86_64.manylinux_2_28_x86_64.whl (12.2 MB) 2025-11-03T17:08:22.5327779Z Downloading pynvml-13.0.1-py3-none-any.whl (28 kB) 2025-11-03T17:08:22.8436236Z Installing collected packages: pynvml, cuda-bindings, cuda-python, flashinfer-python 2025-11-03T17:08:22.8727544Z [?25l 2025-11-03T17:08:22.8727878Z  Attempting uninstall: cuda-bindings 2025-11-03T17:08:22.8796863Z 2025-11-03T17:08:22.8797154Z  Found existing installation: cuda-bindings 13.0.3 2025-11-03T17:08:22.8842815Z 2025-11-03T17:08:22.8843039Z  Uninstalling cuda-bindings-13.0.3: 2025-11-03T17:08:22.8857195Z 2025-11-03T17:08:22.8857457Z  Successfully uninstalled cuda-bindings-13.0.3 2025-11-03T17:08:23.0120347Z 2025-11-03T17:08:23.1609570Z  ━━━━━━━━━━╺━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 1/4 [cuda-bindings] 2025-11-03T17:08:23.1610964Z  Attempting uninstall: cuda-python 2025-11-03T17:08:23.1685917Z ━━━━━━━━━━╺━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 1/4 [cuda-bindings] 2025-11-03T17:08:23.1686482Z  Found existing installation: cuda-python 13.0.3 2025-11-03T17:08:23.1704084Z ━━━━━━━━━━╺━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 1/4 [cuda-bindings] 2025-11-03T17:08:23.1704560Z  Uninstalling cuda-python-13.0.3: 2025-11-03T17:08:23.1712522Z ━━━━━━━━━━╺━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 1/4 [cuda-bindings] 2025-11-03T17:08:23.1713323Z  Successfully uninstalled cuda-python-13.0.3 2025-11-03T17:08:23.1797768Z ━━━━━━━━━━╺━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 1/4 [cuda-bindings] 2025-11-03T17:08:23.1811017Z  ━━━━━━━━━━━━━━━━━━━━╺━━━━━━━━━━━━━━━━━━━ 2/4 [cuda-python] 2025-11-03T17:08:23.1811486Z  Attempting uninstall: flashinfer-python 2025-11-03T17:08:23.1889724Z ━━━━━━━━━━━━━━━━━━━━╺━━━━━━━━━━━━━━━━━━━ 2/4 [cuda-python] 2025-11-03T17:08:23.1890238Z  Found existing installation: flashinfer-python 0.4.0 2025-11-03T17:08:23.2611076Z ━━━━━━━━━━━━━━━━━━━━╺━━━━━━━━━━━━━━━━━━━ 2/4 [cuda-python] 2025-11-03T17:08:23.2611809Z  Uninstalling flashinfer-python-0.4.0: 2025-11-03T17:08:23.3473881Z ━━━━━━━━━━━━━━━━━━━━╺━━━━━━━━━━━━━━━━━━━ 2/4 [cuda-python] 2025-11-03T17:08:23.5148749Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╺━━━━━━━━━ 3/4 [flashinfer-python] 2025-11-03T17:08:23.6824637Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╺━━━━━━━━━ 3/4 [flashinfer-python] 2025-11-03T17:08:23.8196407Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╺━━━━━━━━━ 3/4 [flashinfer-python] 2025-11-03T17:08:23.8197154Z  Successfully uninstalled flashinfer-python-0.4.0 2025-11-03T17:08:23.8499886Z ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╺━━━━━━━━━ 3/4 [flashinfer-python] 2025-11-03T17:08:24.0176906Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╺━━━━━━━━━ 3/4 [flashinfer-python] 2025-11-03T17:08:24.1854069Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╺━━━━━━━━━ 3/4 [flashinfer-python] 2025-11-03T17:08:24.3529670Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╺━━━━━━━━━ 3/4 [flashinfer-python] 2025-11-03T17:08:24.5206832Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╺━━━━━━━━━ 3/4 [flashinfer-python] 2025-11-03T17:08:24.6883544Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╺━━━━━━━━━ 3/4 [flashinfer-python] 2025-11-03T17:08:24.8559833Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╺━━━━━━━━━ 3/4 [flashinfer-python] 2025-11-03T17:08:25.0236030Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╺━━━━━━━━━ 3/4 [flashinfer-python] 2025-11-03T17:08:25.1916463Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╺━━━━━━━━━ 3/4 [flashinfer-python] 2025-11-03T17:08:25.3597046Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╺━━━━━━━━━ 3/4 [flashinfer-python] 2025-11-03T17:08:25.5277447Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╺━━━━━━━━━ 3/4 [flashinfer-python] 2025-11-03T17:08:25.6399433Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╺━━━━━━━━━ 3/4 [flashinfer-python] 2025-11-03T17:08:25.6400459Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 4/4 [flashinfer-python] 2025-11-03T17:08:25.6400831Z [?25h 2025-11-03T17:08:25.6551338Z ERROR: pip's dependency resolver does not currently take into account all the packages that are installed. This behaviour is the source of the following dependency conflicts. 2025-11-03T17:08:25.6552958Z vllm 0.11.0rc2.dev389+ge51928192.d20251103 requires flashinfer-python==0.4.0, but you have flashinfer-python 0.2.14.post1 which is incompatible. 2025-11-03T17:08:25.6554351Z Successfully installed cuda-bindings-12.9.4 cuda-python-12.9.0 flashinfer-python-0.2.14.post1 pynvml-13.0.1 2025-11-03T17:08:25.7213519Z 2025-11-03 17:08:25,720 [INFO] cli.lib.core.vllm.vllm_test: Done. Installed vllm wheels 2025-11-03T17:08:25.7214279Z 2025-11-03 17:08:25,721 [INFO] cli.lib.common.pip_helper: pip installing packages: /opt/conda/envs/py_3.12/bin/python -m uv pip install -e tests/vllm_test_utils 2025-11-03T17:08:25.7215044Z 2025-11-03 17:08:25,721 [INFO] cli.lib.common.utils: [cmd] /opt/conda/envs/py_3.12/bin/python -m uv pip install -e tests/vllm_test_utils 2025-11-03T17:08:25.8304011Z Using Python 3.12.12 environment at: /opt/conda/envs/py_3.12 2025-11-03T17:08:26.6900467Z Resolved 1 package in 850ms 2025-11-03T17:08:26.6903617Z Building vllm-test-utils @ file:///var/lib/jenkins/workspace/vllm/tests/vllm_test_utils 2025-11-03T17:08:26.8681293Z Built vllm-test-utils @ file:///var/lib/jenkins/workspace/vllm/tests/vllm_test_utils 2025-11-03T17:08:26.8711838Z Prepared 1 package in 181ms 2025-11-03T17:08:26.8816686Z Installed 1 package in 10ms 2025-11-03T17:08:26.8820211Z + vllm-test-utils==0.1 (from file:///var/lib/jenkins/workspace/vllm/tests/vllm_test_utils) 2025-11-03T17:08:26.9050982Z 2025-11-03 17:08:26,904 [INFO] cli.lib.common.pip_helper: pip installing packages: /opt/conda/envs/py_3.12/bin/python -m uv pip install hf_transfer 2025-11-03T17:08:26.9052003Z 2025-11-03 17:08:26,904 [INFO] cli.lib.common.utils: [cmd] /opt/conda/envs/py_3.12/bin/python -m uv pip install hf_transfer 2025-11-03T17:08:26.9857296Z Using Python 3.12.12 environment at: /opt/conda/envs/py_3.12 2025-11-03T17:08:27.0168487Z Resolved 1 package in 24ms 2025-11-03T17:08:27.0226083Z Downloading hf-transfer (3.4MiB) 2025-11-03T17:08:27.0495861Z Downloading hf-transfer 2025-11-03T17:08:27.0498711Z Prepared 1 package in 32ms 2025-11-03T17:08:27.0549640Z Installed 1 package in 4ms 2025-11-03T17:08:27.0550260Z + hf-transfer==0.1.9 2025-11-03T17:08:27.0580919Z 2025-11-03 17:08:27,057 [INFO] cli.lib.common.utils: [cmd] /opt/conda/envs/py_3.12/bin/python use_existing_torch.py 2025-11-03T17:08:27.0981897Z >>> cleaning requirements/cpu.txt 2025-11-03T17:08:27.0982228Z removed: 2025-11-03T17:08:27.0987880Z <<< done cleaning requirements/cpu.txt 2025-11-03T17:08:27.0996456Z 2025-11-03T17:08:27.0996581Z >>> cleaning requirements/dev.txt 2025-11-03T17:08:27.0996850Z <<< done cleaning requirements/dev.txt 2025-11-03T17:08:27.0997009Z 2025-11-03T17:08:27.0997104Z >>> cleaning requirements/kv_connectors.txt 2025-11-03T17:08:27.0997392Z <<< done cleaning requirements/kv_connectors.txt 2025-11-03T17:08:27.0997576Z 2025-11-03T17:08:27.0997659Z >>> cleaning requirements/lint.txt 2025-11-03T17:08:27.0997895Z <<< done cleaning requirements/lint.txt 2025-11-03T17:08:27.0998050Z 2025-11-03T17:08:27.0998135Z >>> cleaning requirements/tpu.txt 2025-11-03T17:08:27.0998338Z removed: 2025-11-03T17:08:27.1002287Z <<< done cleaning requirements/tpu.txt 2025-11-03T17:08:27.1002556Z 2025-11-03T17:08:27.1002650Z >>> cleaning requirements/build.txt 2025-11-03T17:08:27.1002871Z removed: 2025-11-03T17:08:27.1012316Z <<< done cleaning requirements/build.txt 2025-11-03T17:08:27.1012673Z 2025-11-03T17:08:27.1012828Z >>> cleaning requirements/common.txt 2025-11-03T17:08:27.1013288Z <<< done cleaning requirements/common.txt 2025-11-03T17:08:27.1013493Z 2025-11-03T17:08:27.1013605Z >>> cleaning requirements/cpu-build.txt 2025-11-03T17:08:27.1013892Z removed: 2025-11-03T17:08:27.1019010Z <<< done cleaning requirements/cpu-build.txt 2025-11-03T17:08:27.1019369Z 2025-11-03T17:08:27.1019532Z >>> cleaning requirements/cuda.txt 2025-11-03T17:08:27.1019793Z removed: 2025-11-03T17:08:27.1025673Z <<< done cleaning requirements/cuda.txt 2025-11-03T17:08:27.1025844Z 2025-11-03T17:08:27.1025967Z >>> cleaning requirements/docs.txt 2025-11-03T17:08:27.1026338Z removed: 2025-11-03T17:08:27.1026984Z <<< done cleaning requirements/docs.txt 2025-11-03T17:08:27.1027500Z 2025-11-03T17:08:27.1027692Z >>> cleaning requirements/nightly_torch_test.txt 2025-11-03T17:08:27.1028028Z <<< done cleaning requirements/nightly_torch_test.txt 2025-11-03T17:08:27.1028316Z 2025-11-03T17:08:27.1028476Z >>> cleaning requirements/rocm-build.txt 2025-11-03T17:08:27.1028732Z removed: 2025-11-03T17:08:27.1035190Z <<< done cleaning requirements/rocm-build.txt 2025-11-03T17:08:27.1035451Z 2025-11-03T17:08:27.1035551Z >>> cleaning requirements/rocm-test.txt 2025-11-03T17:08:27.1035821Z <<< done cleaning requirements/rocm-test.txt 2025-11-03T17:08:27.1035986Z 2025-11-03T17:08:27.1036067Z >>> cleaning requirements/rocm.txt 2025-11-03T17:08:27.1036297Z <<< done cleaning requirements/rocm.txt 2025-11-03T17:08:27.1036451Z 2025-11-03T17:08:27.1036524Z >>> cleaning requirements/test.txt 2025-11-03T17:08:27.1037929Z removed: 2025-11-03T17:08:27.1043058Z <<< done cleaning requirements/test.txt 2025-11-03T17:08:27.1043291Z 2025-11-03T17:08:27.1043383Z >>> cleaning requirements/xpu.txt 2025-11-03T17:08:27.1043612Z removed: 2025-11-03T17:08:27.1050043Z <<< done cleaning requirements/xpu.txt 2025-11-03T17:08:27.1050323Z 2025-11-03T17:08:27.1050411Z >>> cleaning pyproject.toml 2025-11-03T17:08:27.1050622Z removed: 2025-11-03T17:08:27.1061026Z <<< done cleaning pyproject.toml 2025-11-03T17:08:27.1061209Z 2025-11-03T17:08:27.1118245Z 2025-11-03 17:08:27,111 [INFO] cli.lib.common.pip_helper: pip installing packages: /opt/conda/envs/py_3.12/bin/python -m uv pip install -r requirements/common.txt 2025-11-03T17:08:27.1119215Z 2025-11-03 17:08:27,111 [INFO] cli.lib.common.utils: [cmd] /opt/conda/envs/py_3.12/bin/python -m uv pip install -r requirements/common.txt 2025-11-03T17:08:27.1630320Z Using Python 3.12.12 environment at: /opt/conda/envs/py_3.12 2025-11-03T17:08:27.1794861Z Audited 50 packages in 21ms 2025-11-03T17:08:27.1815346Z 2025-11-03 17:08:27,181 [INFO] cli.lib.common.pip_helper: pip installing packages: /opt/conda/envs/py_3.12/bin/python -m uv pip install -r requirements/build.txt 2025-11-03T17:08:27.1816296Z 2025-11-03 17:08:27,181 [INFO] cli.lib.common.utils: [cmd] /opt/conda/envs/py_3.12/bin/python -m uv pip install -r requirements/build.txt 2025-11-03T17:08:27.2283814Z Using Python 3.12.12 environment at: /opt/conda/envs/py_3.12 2025-11-03T17:08:27.2952477Z Resolved 11 packages in 59ms 2025-11-03T17:08:27.3058752Z Prepared 1 package in 10ms 2025-11-03T17:08:27.3125203Z Installed 1 package in 6ms 2025-11-03T17:08:27.3125640Z + setuptools-scm==9.2.2 2025-11-03T17:08:27.3173450Z 2025-11-03 17:08:27,317 [INFO] cli.lib.core.vllm.vllm_test: generate test.txt from requirements/test.in with local torch whls 2025-11-03T17:08:27.6769637Z 2025-11-03 17:08:27,676 [INFO] cli.lib.core.vllm.vllm_test: [INFO] Updated requirements/test.in 2025-11-03T17:08:27.6773664Z 2025-11-03 17:08:27,677 [INFO] cli.lib.common.utils: [cmd] /opt/conda/envs/py_3.12/bin/python -m uv pip compile requirements/test.in -o test.txt --index-strategy unsafe-best-match --constraint snapshot_constraint.txt --torch-backend cu128 2025-11-03T17:08:27.8056524Z Updating https://github.com/IBM/terratorch.git (1.1.rc3) 2025-11-03T17:08:29.5145431Z Updated https://github.com/IBM/terratorch.git (07184fcf91a1324f831ff521dd238d97fe350e3e) 2025-11-03T17:08:32.3023872Z Resolved 329 packages in 4.57s 2025-11-03T17:08:32.3024485Z # This file was autogenerated by uv via the following command: 2025-11-03T17:08:32.3046078Z # uv pip compile requirements/test.in -o test.txt --index-strategy unsafe-best-match --constraint snapshot_constraint.txt --torch-backend cu128 2025-11-03T17:08:32.3108248Z absl-py==2.1.0 2025-11-03T17:08:32.3108776Z  # via 2025-11-03T17:08:32.3109047Z # -c snapshot_constraint.txt 2025-11-03T17:08:32.3109601Z # rouge-score 2025-11-03T17:08:32.3110020Z accelerate==1.0.1 2025-11-03T17:08:32.3110768Z  # via 2025-11-03T17:08:32.3111057Z # -c snapshot_constraint.txt 2025-11-03T17:08:32.3111511Z # lm-eval 2025-11-03T17:08:32.3111946Z # peft 2025-11-03T17:08:32.3112312Z aenum==3.1.16 2025-11-03T17:08:32.3112648Z  # via 2025-11-03T17:08:32.3112954Z # -c snapshot_constraint.txt 2025-11-03T17:08:32.3113476Z # lightly 2025-11-03T17:08:32.3113862Z affine==2.4.0 2025-11-03T17:08:32.3114169Z  # via 2025-11-03T17:08:32.3114379Z # -c snapshot_constraint.txt 2025-11-03T17:08:32.3114686Z # rasterio 2025-11-03T17:08:32.3114907Z aiohappyeyeballs==2.6.1 2025-11-03T17:08:32.3115209Z  # via 2025-11-03T17:08:32.3115464Z # -c snapshot_constraint.txt 2025-11-03T17:08:32.3115752Z # aiohttp 2025-11-03T17:08:32.3115955Z aiohttp==3.13.0 2025-11-03T17:08:32.3116167Z  # via 2025-11-03T17:08:32.3116372Z # -c snapshot_constraint.txt 2025-11-03T17:08:32.3116626Z # aiohttp-cors 2025-11-03T17:08:32.3116896Z # datasets 2025-11-03T17:08:32.3117231Z # fsspec 2025-11-03T17:08:32.3117546Z # gpt-oss 2025-11-03T17:08:32.3117865Z # lm-eval 2025-11-03T17:08:32.3118247Z # ray 2025-11-03T17:08:32.3118599Z aiohttp-cors==0.8.1 2025-11-03T17:08:32.3119021Z  # via 2025-11-03T17:08:32.3119391Z # -c snapshot_constraint.txt 2025-11-03T17:08:32.3119780Z # ray 2025-11-03T17:08:32.3119948Z aiosignal==1.4.0 2025-11-03T17:08:32.3120228Z  # via 2025-11-03T17:08:32.3120515Z # -c snapshot_constraint.txt 2025-11-03T17:08:32.3121281Z # aiohttp 2025-11-03T17:08:32.3121588Z albucore==0.0.16 2025-11-03T17:08:32.3121925Z  # via 2025-11-03T17:08:32.3122165Z # -c snapshot_constraint.txt 2025-11-03T17:08:32.3122537Z # terratorch 2025-11-03T17:08:32.3122861Z albumentations==1.4.6 2025-11-03T17:08:32.3123138Z  # via 2025-11-03T17:08:32.3123314Z # -c snapshot_constraint.txt 2025-11-03T17:08:32.3123550Z # terratorch 2025-11-03T17:08:32.3123719Z alembic==1.16.4 2025-11-03T17:08:32.3123893Z  # via 2025-11-03T17:08:32.3124057Z # -c snapshot_constraint.txt 2025-11-03T17:08:32.3124531Z # mlflow 2025-11-03T17:08:32.3124701Z annotated-types==0.7.0 2025-11-03T17:08:32.3124897Z  # via 2025-11-03T17:08:32.3125052Z # -c snapshot_constraint.txt 2025-11-03T17:08:32.3125345Z # pydantic 2025-11-03T17:08:32.3125537Z antlr4-python3-runtime==4.9.3 2025-11-03T17:08:32.3125753Z  # via 2025-11-03T17:08:32.3125913Z # -c snapshot_constraint.txt 2025-11-03T17:08:32.3126103Z # hydra-core 2025-11-03T17:08:32.3126294Z # omegaconf 2025-11-03T17:08:32.3126459Z anyio==4.6.2.post1 2025-11-03T17:08:32.3126628Z  # via 2025-11-03T17:08:32.3126788Z # -c snapshot_constraint.txt 2025-11-03T17:08:32.3126979Z # httpx 2025-11-03T17:08:32.3127158Z # starlette 2025-11-03T17:08:32.3127320Z argcomplete==3.5.1 2025-11-03T17:08:32.3127512Z  # via 2025-11-03T17:08:32.3127683Z # -c snapshot_constraint.txt 2025-11-03T17:08:32.3128062Z # datamodel-code-generator 2025-11-03T17:08:32.3128280Z arrow==1.3.0 2025-11-03T17:08:32.3128455Z  # via 2025-11-03T17:08:32.3128614Z # -c snapshot_constraint.txt 2025-11-03T17:08:32.3128846Z # isoduration 2025-11-03T17:08:32.3129014Z attrs==24.2.0 2025-11-03T17:08:32.3129197Z  # via 2025-11-03T17:08:32.3129357Z # -c snapshot_constraint.txt 2025-11-03T17:08:32.3129555Z # aiohttp 2025-11-03T17:08:32.3129697Z # fiona 2025-11-03T17:08:32.3129968Z # hypothesis 2025-11-03T17:08:32.3130127Z # jsonlines 2025-11-03T17:08:32.3130281Z # jsonschema 2025-11-03T17:08:32.3130435Z # pytest-subtests 2025-11-03T17:08:32.3130604Z # rasterio 2025-11-03T17:08:32.3130799Z # referencing 2025-11-03T17:08:32.3130986Z audioread==3.0.1 2025-11-03T17:08:32.3131160Z  # via 2025-11-03T17:08:32.3131459Z # -c snapshot_constraint.txt 2025-11-03T17:08:32.3131695Z # librosa 2025-11-03T17:08:32.3131850Z backoff==2.2.1 2025-11-03T17:08:32.3132022Z  # via 2025-11-03T17:08:32.3132201Z # -c snapshot_constraint.txt 2025-11-03T17:08:32.3132429Z # -r requirements/test.in 2025-11-03T17:08:32.3132648Z # schemathesis 2025-11-03T17:08:32.3132826Z bitsandbytes==0.46.1 2025-11-03T17:08:32.3133009Z  # via 2025-11-03T17:08:32.3133166Z # -c snapshot_constraint.txt 2025-11-03T17:08:32.3133360Z # -r requirements/test.in 2025-11-03T17:08:32.3133571Z # lightning 2025-11-03T17:08:32.3133737Z black==24.10.0 2025-11-03T17:08:32.3133958Z  # via 2025-11-03T17:08:32.3134234Z # -c snapshot_constraint.txt 2025-11-03T17:08:32.3134714Z # datamodel-code-generator 2025-11-03T17:08:32.3135047Z blinker==1.9.0 2025-11-03T17:08:32.3135314Z  # via 2025-11-03T17:08:32.3135574Z # -c snapshot_constraint.txt 2025-11-03T17:08:32.3135994Z # flask 2025-11-03T17:08:32.3136280Z blobfile==3.0.0 2025-11-03T17:08:32.3136624Z  # via 2025-11-03T17:08:32.3136926Z # -c snapshot_constraint.txt 2025-11-03T17:08:32.3137539Z # -r requirements/test.in 2025-11-03T17:08:32.3137916Z bm25s==0.2.13 2025-11-03T17:08:32.3138251Z  # via 2025-11-03T17:08:32.3138552Z # -c snapshot_constraint.txt 2025-11-03T17:08:32.3138958Z # mteb 2025-11-03T17:08:32.3139226Z boto3==1.35.57 2025-11-03T17:08:32.3139508Z  # via 2025-11-03T17:08:32.3139764Z # -c snapshot_constraint.txt 2025-11-03T17:08:32.3140061Z # runai-model-streamer-s3 2025-11-03T17:08:32.3140570Z # tensorizer 2025-11-03T17:08:32.3140861Z botocore==1.35.57 2025-11-03T17:08:32.3141052Z  # via 2025-11-03T17:08:32.3141217Z # -c snapshot_constraint.txt 2025-11-03T17:08:32.3141413Z # boto3 2025-11-03T17:08:32.3141736Z # s3transfer 2025-11-03T17:08:32.3141927Z bounded-pool-executor==0.0.3 2025-11-03T17:08:32.3142165Z  # via 2025-11-03T17:08:32.3142333Z # -c snapshot_constraint.txt 2025-11-03T17:08:32.3142557Z # pqdm 2025-11-03T17:08:32.3142803Z buildkite-test-collector==0.1.9 2025-11-03T17:08:32.3143294Z  # via 2025-11-03T17:08:32.3143564Z # -c snapshot_constraint.txt 2025-11-03T17:08:32.3143983Z # -r requirements/test.in 2025-11-03T17:08:32.3144367Z cachetools==5.5.2 2025-11-03T17:08:32.3144634Z  # via 2025-11-03T17:08:32.3144889Z # -c snapshot_constraint.txt 2025-11-03T17:08:32.3145157Z # google-auth 2025-11-03T17:08:32.3145381Z # mlflow-skinny 2025-11-03T17:08:32.3145568Z certifi==2024.8.30 2025-11-03T17:08:32.3145854Z  # via 2025-11-03T17:08:32.3146138Z # -c snapshot_constraint.txt 2025-11-03T17:08:32.3146409Z # fiona 2025-11-03T17:08:32.3146678Z # httpcore 2025-11-03T17:08:32.3146889Z # httpx 2025-11-03T17:08:32.3147066Z # lightly 2025-11-03T17:08:32.3147286Z # pyogrio 2025-11-03T17:08:32.3147522Z # pyproj 2025-11-03T17:08:32.3147706Z # rasterio 2025-11-03T17:08:32.3148052Z # requests 2025-11-03T17:08:32.3148334Z cffi==1.17.1 2025-11-03T17:08:32.3148646Z  # via 2025-11-03T17:08:32.3148931Z # -c snapshot_constraint.txt 2025-11-03T17:08:32.3149289Z # soundfile 2025-11-03T17:08:32.3149564Z chardet==5.2.0 2025-11-03T17:08:32.3149884Z  # via 2025-11-03T17:08:32.3150133Z # -c snapshot_constraint.txt 2025-11-03T17:08:32.3150414Z # mbstrdecoder 2025-11-03T17:08:32.3150754Z charset-normalizer==3.4.0 2025-11-03T17:08:32.3151123Z  # via 2025-11-03T17:08:32.3151415Z # -c snapshot_constraint.txt 2025-11-03T17:08:32.3151765Z # requests 2025-11-03T17:08:32.3152045Z chz==0.3.0 2025-11-03T17:08:32.3152318Z  # via 2025-11-03T17:08:32.3152570Z # -c snapshot_constraint.txt 2025-11-03T17:08:32.3152967Z # gpt-oss 2025-11-03T17:08:32.3153187Z click==8.1.7 2025-11-03T17:08:32.3153443Z  # via 2025-11-03T17:08:32.3153810Z # -c snapshot_constraint.txt 2025-11-03T17:08:32.3154142Z # black 2025-11-03T17:08:32.3154323Z # click-plugins 2025-11-03T17:08:32.3154495Z # cligj 2025-11-03T17:08:32.3154632Z # fiona 2025-11-03T17:08:32.3154775Z # flask 2025-11-03T17:08:32.3154915Z # jiwer 2025-11-03T17:08:32.3155096Z # mlflow-skinny 2025-11-03T17:08:32.3155326Z # nltk 2025-11-03T17:08:32.3155474Z # rasterio 2025-11-03T17:08:32.3155625Z # ray 2025-11-03T17:08:32.3155776Z # schemathesis 2025-11-03T17:08:32.3155986Z # typer 2025-11-03T17:08:32.3156237Z # uvicorn 2025-11-03T17:08:32.3156496Z click-plugins==1.1.1.2 2025-11-03T17:08:32.3156696Z  # via 2025-11-03T17:08:32.3156915Z # -c snapshot_constraint.txt 2025-11-03T17:08:32.3157231Z # fiona 2025-11-03T17:08:32.3157561Z # rasterio 2025-11-03T17:08:32.3157767Z cligj==0.7.2 2025-11-03T17:08:32.3158026Z  # via 2025-11-03T17:08:32.3158304Z # -c snapshot_constraint.txt 2025-11-03T17:08:32.3158507Z # fiona 2025-11-03T17:08:32.3158684Z # rasterio 2025-11-03T17:08:32.3158878Z cloudpickle==3.1.1 2025-11-03T17:08:32.3159142Z  # via 2025-11-03T17:08:32.3159399Z # -c snapshot_constraint.txt 2025-11-03T17:08:32.3159634Z # mlflow-skinny 2025-11-03T17:08:32.3159813Z colorama==0.4.6 2025-11-03T17:08:32.3160000Z  # via 2025-11-03T17:08:32.3160286Z # -c snapshot_constraint.txt 2025-11-03T17:08:32.3160525Z # sacrebleu 2025-11-03T17:08:32.3160683Z # schemathesis 2025-11-03T17:08:32.3160904Z # tqdm-multiprocess 2025-11-03T17:08:32.3161163Z colorful==0.5.6 2025-11-03T17:08:32.3161378Z  # via 2025-11-03T17:08:32.3161539Z # -c snapshot_constraint.txt 2025-11-03T17:08:32.3161757Z # ray 2025-11-03T17:08:32.3161962Z contourpy==1.3.0 2025-11-03T17:08:32.3162211Z  # via 2025-11-03T17:08:32.3162377Z # -c snapshot_constraint.txt 2025-11-03T17:08:32.3162614Z # matplotlib 2025-11-03T17:08:32.3162843Z coverage==7.10.6 2025-11-03T17:08:32.3163075Z  # via 2025-11-03T17:08:32.3163318Z # -c snapshot_constraint.txt 2025-11-03T17:08:32.3163689Z # pytest-cov 2025-11-03T17:08:32.3163859Z cramjam==2.9.0 2025-11-03T17:08:32.3164123Z  # via 2025-11-03T17:08:32.3164337Z # -c snapshot_constraint.txt 2025-11-03T17:08:32.3164573Z # fastparquet 2025-11-03T17:08:32.3164752Z cupy-cuda12x==13.6.0 2025-11-03T17:08:32.3165038Z  # via 2025-11-03T17:08:32.3165266Z # -c snapshot_constraint.txt 2025-11-03T17:08:32.3165493Z # ray 2025-11-03T17:08:32.3165639Z cycler==0.12.1 2025-11-03T17:08:32.3165810Z  # via 2025-11-03T17:08:32.3166042Z # -c snapshot_constraint.txt 2025-11-03T17:08:32.3166418Z # matplotlib 2025-11-03T17:08:32.3166674Z databricks-sdk==0.59.0 2025-11-03T17:08:32.3166982Z  # via 2025-11-03T17:08:32.3167260Z # -c snapshot_constraint.txt 2025-11-03T17:08:32.3167664Z # mlflow-skinny 2025-11-03T17:08:32.3167987Z datamodel-code-generator==0.26.3 2025-11-03T17:08:32.3168354Z  # via 2025-11-03T17:08:32.3168625Z # -c snapshot_constraint.txt 2025-11-03T17:08:32.3169036Z # -r requirements/test.in 2025-11-03T17:08:32.3169375Z dataproperty==1.0.1 2025-11-03T17:08:32.3169670Z  # via 2025-11-03T17:08:32.3169939Z # -c snapshot_constraint.txt 2025-11-03T17:08:32.3170260Z # pytablewriter 2025-11-03T17:08:32.3170581Z # tabledata 2025-11-03T17:08:32.3170858Z datasets==3.0.2 2025-11-03T17:08:32.3171145Z  # via 2025-11-03T17:08:32.3171411Z # -c snapshot_constraint.txt 2025-11-03T17:08:32.3171724Z # evaluate 2025-11-03T17:08:32.3171976Z # lm-eval 2025-11-03T17:08:32.3172268Z # mteb 2025-11-03T17:08:32.3172526Z decorator==5.1.1 2025-11-03T17:08:32.3172819Z  # via 2025-11-03T17:08:32.3173083Z # -c snapshot_constraint.txt 2025-11-03T17:08:32.3173614Z # librosa 2025-11-03T17:08:32.3173884Z decord==0.6.0 2025-11-03T17:08:32.3174162Z  # via 2025-11-03T17:08:32.3174429Z # -c snapshot_constraint.txt 2025-11-03T17:08:32.3174843Z # -r requirements/test.in 2025-11-03T17:08:32.3175168Z dill==0.3.8 2025-11-03T17:08:32.3175454Z  # via 2025-11-03T17:08:32.3175728Z # -c snapshot_constraint.txt 2025-11-03T17:08:32.3176045Z # datasets 2025-11-03T17:08:32.3176283Z # evaluate 2025-11-03T17:08:32.3176528Z # lm-eval 2025-11-03T17:08:32.3176846Z # multiprocess 2025-11-03T17:08:32.3177137Z distlib==0.3.9 2025-11-03T17:08:32.3177423Z  # via 2025-11-03T17:08:32.3177692Z # -c snapshot_constraint.txt 2025-11-03T17:08:32.3178082Z # virtualenv 2025-11-03T17:08:32.3178264Z dnspython==2.7.0 2025-11-03T17:08:32.3178439Z  # via 2025-11-03T17:08:32.3178597Z # -c snapshot_constraint.txt 2025-11-03T17:08:32.3178839Z # email-validator 2025-11-03T17:08:32.3179029Z docker==7.1.0 2025-11-03T17:08:32.3179191Z  # via 2025-11-03T17:08:32.3179356Z # -c snapshot_constraint.txt 2025-11-03T17:08:32.3179546Z # gpt-oss 2025-11-03T17:08:32.3179726Z # mlflow 2025-11-03T17:08:32.3179878Z docopt==0.6.2 2025-11-03T17:08:32.3180047Z  # via 2025-11-03T17:08:32.3180205Z # -c snapshot_constraint.txt 2025-11-03T17:08:32.3180419Z # num2words 2025-11-03T17:08:32.3180596Z docstring-parser==0.17.0 2025-11-03T17:08:32.3180794Z  # via 2025-11-03T17:08:32.3180966Z # -c snapshot_constraint.txt 2025-11-03T17:08:32.3181190Z # jsonargparse 2025-11-03T17:08:32.3181383Z efficientnet-pytorch==0.7.1 2025-11-03T17:08:32.3181779Z  # via segmentation-models-pytorch 2025-11-03T17:08:32.3182012Z einops==0.8.1 2025-11-03T17:08:32.3182179Z  # via 2025-11-03T17:08:32.3182339Z # -c snapshot_constraint.txt 2025-11-03T17:08:32.3182545Z # -r requirements/test.in 2025-11-03T17:08:32.3182729Z # encodec 2025-11-03T17:08:32.3182877Z # terratorch 2025-11-03T17:08:32.3183034Z # torchgeo 2025-11-03T17:08:32.3183200Z # vector-quantize-pytorch 2025-11-03T17:08:32.3183571Z # vocos 2025-11-03T17:08:32.3183738Z einx==0.3.0 2025-11-03T17:08:32.3183912Z  # via 2025-11-03T17:08:32.3184079Z # -c snapshot_constraint.txt 2025-11-03T17:08:32.3184326Z # vector-quantize-pytorch 2025-11-03T17:08:32.3184542Z email-validator==2.2.0 2025-11-03T17:08:32.3184737Z  # via 2025-11-03T17:08:32.3184902Z # -c snapshot_constraint.txt 2025-11-03T17:08:32.3185121Z # pydantic 2025-11-03T17:08:32.3185288Z encodec==0.1.1 2025-11-03T17:08:32.3185462Z  # via 2025-11-03T17:08:32.3185623Z # -c snapshot_constraint.txt 2025-11-03T17:08:32.3185831Z # vocos 2025-11-03T17:08:32.3186001Z eval-type-backport==0.2.2 2025-11-03T17:08:32.3186201Z  # via 2025-11-03T17:08:32.3186359Z # -c snapshot_constraint.txt 2025-11-03T17:08:32.3186574Z # mteb 2025-11-03T17:08:32.3186740Z evaluate==0.4.3 2025-11-03T17:08:32.3186910Z  # via 2025-11-03T17:08:32.3187067Z # -c snapshot_constraint.txt 2025-11-03T17:08:32.3187283Z # lm-eval 2025-11-03T17:08:32.3187444Z fastapi==0.116.1 2025-11-03T17:08:32.3187617Z  # via 2025-11-03T17:08:32.3187769Z # -c snapshot_constraint.txt 2025-11-03T17:08:32.3187957Z # gpt-oss 2025-11-03T17:08:32.3188145Z # mlflow-skinny 2025-11-03T17:08:32.3188328Z fastparquet==2024.11.0 2025-11-03T17:08:32.3188508Z  # via 2025-11-03T17:08:32.3188665Z # -c snapshot_constraint.txt 2025-11-03T17:08:32.3188887Z # genai-perf 2025-11-03T17:08:32.3189062Z fastrlock==0.8.2 2025-11-03T17:08:32.3189228Z  # via 2025-11-03T17:08:32.3189387Z # -c snapshot_constraint.txt 2025-11-03T17:08:32.3189613Z # cupy-cuda12x 2025-11-03T17:08:32.3189804Z fastsafetensors==0.1.10 2025-11-03T17:08:32.3189994Z  # via 2025-11-03T17:08:32.3190258Z # -c snapshot_constraint.txt 2025-11-03T17:08:32.3190505Z # -r requirements/test.in 2025-11-03T17:08:32.3190712Z filelock==3.16.1 2025-11-03T17:08:32.3190887Z  # via 2025-11-03T17:08:32.3191044Z # -c snapshot_constraint.txt 2025-11-03T17:08:32.3191235Z # blobfile 2025-11-03T17:08:32.3191384Z # datasets 2025-11-03T17:08:32.3191532Z # huggingface-hub 2025-11-03T17:08:32.3191696Z # ray 2025-11-03T17:08:32.3191836Z # torch 2025-11-03T17:08:32.3191983Z # transformers 2025-11-03T17:08:32.3192171Z # virtualenv 2025-11-03T17:08:32.3192334Z fiona==1.10.1 2025-11-03T17:08:32.3192503Z  # via 2025-11-03T17:08:32.3192662Z # -c snapshot_constraint.txt 2025-11-03T17:08:32.3192875Z # torchgeo 2025-11-03T17:08:32.3193039Z flask==3.1.1 2025-11-03T17:08:32.3193208Z  # via 2025-11-03T17:08:32.3193364Z # -c snapshot_constraint.txt 2025-11-03T17:08:32.3193595Z # mlflow 2025-11-03T17:08:32.3193761Z fonttools==4.55.0 2025-11-03T17:08:32.3193938Z  # via 2025-11-03T17:08:32.3194091Z # -c snapshot_constraint.txt 2025-11-03T17:08:32.3194315Z # matplotlib 2025-11-03T17:08:32.3194481Z fqdn==1.5.1 2025-11-03T17:08:32.3194644Z  # via 2025-11-03T17:08:32.3194796Z # -c snapshot_constraint.txt 2025-11-03T17:08:32.3195017Z # jsonschema 2025-11-03T17:08:32.3195182Z frozendict==2.4.6 2025-11-03T17:08:32.3195357Z  # via 2025-11-03T17:08:32.3195507Z # -c snapshot_constraint.txt 2025-11-03T17:08:32.3195719Z # einx 2025-11-03T17:08:32.3195868Z frozenlist==1.5.0 2025-11-03T17:08:32.3196039Z  # via 2025-11-03T17:08:32.3196186Z # -c snapshot_constraint.txt 2025-11-03T17:08:32.3196373Z # aiohttp 2025-11-03T17:08:32.3196567Z # aiosignal 2025-11-03T17:08:32.3196729Z fsspec==2024.9.0 2025-11-03T17:08:32.3196893Z  # via 2025-11-03T17:08:32.3197054Z # -c snapshot_constraint.txt 2025-11-03T17:08:32.3197242Z # datasets 2025-11-03T17:08:32.3197394Z # evaluate 2025-11-03T17:08:32.3197538Z # fastparquet 2025-11-03T17:08:32.3197702Z # huggingface-hub 2025-11-03T17:08:32.3197964Z # lightning 2025-11-03T17:08:32.3198121Z # pytorch-lightning 2025-11-03T17:08:32.3198327Z # torch 2025-11-03T17:08:32.3198480Z ftfy==6.3.1 2025-11-03T17:08:32.3198642Z  # via 2025-11-03T17:08:32.3198796Z # -c snapshot_constraint.txt 2025-11-03T17:08:32.3199032Z # open-clip-torch 2025-11-03T17:08:32.3199224Z genai-perf==0.0.8 2025-11-03T17:08:32.3199401Z  # via 2025-11-03T17:08:32.3199552Z # -c snapshot_constraint.txt 2025-11-03T17:08:32.3199788Z # -r requirements/test.in 2025-11-03T17:08:32.3199986Z genson==1.3.0 2025-11-03T17:08:32.3200157Z  # via 2025-11-03T17:08:32.3200309Z # -c snapshot_constraint.txt 2025-11-03T17:08:32.3200575Z # datamodel-code-generator 2025-11-03T17:08:32.3200787Z geopandas==1.0.1 2025-11-03T17:08:32.3200968Z  # via 2025-11-03T17:08:32.3201121Z # -c snapshot_constraint.txt 2025-11-03T17:08:32.3201345Z # terratorch 2025-11-03T17:08:32.3201510Z gitdb==4.0.12 2025-11-03T17:08:32.3201679Z  # via 2025-11-03T17:08:32.3201832Z # -c snapshot_constraint.txt 2025-11-03T17:08:32.3202053Z # gitpython 2025-11-03T17:08:32.3202231Z gitpython==3.1.44 2025-11-03T17:08:32.3202410Z  # via 2025-11-03T17:08:32.3202559Z # -c snapshot_constraint.txt 2025-11-03T17:08:32.3202789Z # mlflow-skinny 2025-11-03T17:08:32.3202977Z google-api-core==2.24.2 2025-11-03T17:08:32.3203165Z  # via 2025-11-03T17:08:32.3203318Z # -c snapshot_constraint.txt 2025-11-03T17:08:32.3203519Z # google-cloud-core 2025-11-03T17:08:32.3203704Z # google-cloud-storage 2025-11-03T17:08:32.3203919Z # opencensus 2025-11-03T17:08:32.3204080Z google-auth==2.40.2 2025-11-03T17:08:32.3204256Z  # via 2025-11-03T17:08:32.3204415Z # -c snapshot_constraint.txt 2025-11-03T17:08:32.3204688Z # databricks-sdk 2025-11-03T17:08:32.3204862Z # google-api-core 2025-11-03T17:08:32.3205032Z # google-cloud-core 2025-11-03T17:08:32.3205217Z # google-cloud-storage 2025-11-03T17:08:32.3205453Z # runai-model-streamer-gcs 2025-11-03T17:08:32.3205674Z google-cloud-core==2.4.3 2025-11-03T17:08:32.3205871Z  # via 2025-11-03T17:08:32.3206032Z # -c snapshot_constraint.txt 2025-11-03T17:08:32.3206265Z # google-cloud-storage 2025-11-03T17:08:32.3206484Z google-cloud-storage==3.4.0 2025-11-03T17:08:32.3206694Z  # via 2025-11-03T17:08:32.3206851Z # -c snapshot_constraint.txt 2025-11-03T17:08:32.3207091Z # runai-model-streamer-gcs 2025-11-03T17:08:32.3207300Z google-crc32c==1.7.1 2025-11-03T17:08:32.3207484Z  # via 2025-11-03T17:08:32.3207646Z # -c snapshot_constraint.txt 2025-11-03T17:08:32.3207840Z # google-cloud-storage 2025-11-03T17:08:32.3208072Z # google-resumable-media 2025-11-03T17:08:32.3208296Z google-resumable-media==2.7.2 2025-11-03T17:08:32.3208509Z  # via 2025-11-03T17:08:32.3208660Z # -c snapshot_constraint.txt 2025-11-03T17:08:32.3208902Z # google-cloud-storage 2025-11-03T17:08:32.3209118Z googleapis-common-protos==1.70.0 2025-11-03T17:08:32.3209333Z  # via 2025-11-03T17:08:32.3209483Z # -c snapshot_constraint.txt 2025-11-03T17:08:32.3209716Z # google-api-core 2025-11-03T17:08:32.3209902Z gpt-oss==0.0.8 2025-11-03T17:08:32.3210085Z  # via 2025-11-03T17:08:32.3210240Z # -c snapshot_constraint.txt 2025-11-03T17:08:32.3210482Z # -r requirements/test.in 2025-11-03T17:08:32.3210684Z graphene==3.4.3 2025-11-03T17:08:32.3210862Z  # via 2025-11-03T17:08:32.3211015Z # -c snapshot_constraint.txt 2025-11-03T17:08:32.3211230Z # mlflow 2025-11-03T17:08:32.3211395Z graphql-core==3.2.6 2025-11-03T17:08:32.3211573Z  # via 2025-11-03T17:08:32.3211723Z # -c snapshot_constraint.txt 2025-11-03T17:08:32.3211910Z # graphene 2025-11-03T17:08:32.3212068Z # graphql-relay 2025-11-03T17:08:32.3212278Z # hypothesis-graphql 2025-11-03T17:08:32.3212561Z graphql-relay==3.2.0 2025-11-03T17:08:32.3212746Z  # via 2025-11-03T17:08:32.3212900Z # -c snapshot_constraint.txt 2025-11-03T17:08:32.3213111Z # graphene 2025-11-03T17:08:32.3213416Z greenlet==3.2.3 2025-11-03T17:08:32.3213588Z  # via 2025-11-03T17:08:32.3213744Z # -c snapshot_constraint.txt 2025-11-03T17:08:32.3213966Z # sqlalchemy 2025-11-03T17:08:32.3214128Z grpcio==1.71.0 2025-11-03T17:08:32.3214293Z  # via 2025-11-03T17:08:32.3214463Z # -c snapshot_constraint.txt 2025-11-03T17:08:32.3214672Z # ray 2025-11-03T17:08:32.3214826Z gunicorn==23.0.0 2025-11-03T17:08:32.3214999Z  # via 2025-11-03T17:08:32.3215157Z # -c snapshot_constraint.txt 2025-11-03T17:08:32.3215365Z # mlflow 2025-11-03T17:08:32.3215517Z h11==0.14.0 2025-11-03T17:08:32.3215682Z  # via 2025-11-03T17:08:32.3215839Z # -c snapshot_constraint.txt 2025-11-03T17:08:32.3216020Z # httpcore 2025-11-03T17:08:32.3216202Z # uvicorn 2025-11-03T17:08:32.3216362Z h5py==3.13.0 2025-11-03T17:08:32.3216532Z  # via 2025-11-03T17:08:32.3216683Z # -c snapshot_constraint.txt 2025-11-03T17:08:32.3216904Z # terratorch 2025-11-03T17:08:32.3217083Z harfile==0.3.0 2025-11-03T17:08:32.3217256Z  # via 2025-11-03T17:08:32.3217406Z # -c snapshot_constraint.txt 2025-11-03T17:08:32.3217631Z # schemathesis 2025-11-03T17:08:32.3217812Z hf-xet==1.1.7 2025-11-03T17:08:32.3217980Z  # via 2025-11-03T17:08:32.3218145Z # -c snapshot_constraint.txt 2025-11-03T17:08:32.3218384Z # huggingface-hub 2025-11-03T17:08:32.3218570Z hiredis==3.0.0 2025-11-03T17:08:32.3218739Z  # via 2025-11-03T17:08:32.3218891Z # -c snapshot_constraint.txt 2025-11-03T17:08:32.3219199Z # tensorizer 2025-11-03T17:08:32.3219387Z html2text==2025.4.15 2025-11-03T17:08:32.3219566Z  # via 2025-11-03T17:08:32.3219723Z # -c snapshot_constraint.txt 2025-11-03T17:08:32.3219946Z # gpt-oss 2025-11-03T17:08:32.3220109Z httpcore==1.0.6 2025-11-03T17:08:32.3220274Z  # via 2025-11-03T17:08:32.3220426Z # -c snapshot_constraint.txt 2025-11-03T17:08:32.3220642Z # httpx 2025-11-03T17:08:32.3220796Z httpx==0.27.2 2025-11-03T17:08:32.3220956Z  # via 2025-11-03T17:08:32.3221573Z # -c snapshot_constraint.txt 2025-11-03T17:08:32.3221792Z # -r requirements/test.in 2025-11-03T17:08:32.3222030Z # schemathesis 2025-11-03T17:08:32.3222209Z huggingface-hub==0.34.3 2025-11-03T17:08:32.3222397Z  # via 2025-11-03T17:08:32.3222559Z # -c snapshot_constraint.txt 2025-11-03T17:08:32.3222752Z # accelerate 2025-11-03T17:08:32.3222901Z # datasets 2025-11-03T17:08:32.3223049Z # evaluate 2025-11-03T17:08:32.3223217Z # open-clip-torch 2025-11-03T17:08:32.3223384Z # peft 2025-11-03T17:08:32.3223547Z # segmentation-models-pytorch 2025-11-03T17:08:32.3223767Z # sentence-transformers 2025-11-03T17:08:32.3223960Z # terratorch 2025-11-03T17:08:32.3224112Z # timm 2025-11-03T17:08:32.3224252Z # tokenizers 2025-11-03T17:08:32.3224407Z # transformers 2025-11-03T17:08:32.3224604Z # vocos 2025-11-03T17:08:32.3224758Z humanize==4.11.0 2025-11-03T17:08:32.3224934Z  # via 2025-11-03T17:08:32.3225095Z # -c snapshot_constraint.txt 2025-11-03T17:08:32.3225339Z # runai-model-streamer 2025-11-03T17:08:32.3225536Z hydra-core==1.3.2 2025-11-03T17:08:32.3225720Z  # via 2025-11-03T17:08:32.3225887Z # -c snapshot_constraint.txt 2025-11-03T17:08:32.3226086Z # lightly 2025-11-03T17:08:32.3226264Z # lightning 2025-11-03T17:08:32.3226439Z hypothesis==6.131.0 2025-11-03T17:08:32.3226624Z  # via 2025-11-03T17:08:32.3226791Z # -c snapshot_constraint.txt 2025-11-03T17:08:32.3226995Z # hypothesis-graphql 2025-11-03T17:08:32.3227194Z # hypothesis-jsonschema 2025-11-03T17:08:32.3227425Z # schemathesis 2025-11-03T17:08:32.3227794Z hypothesis-graphql==0.11.1 2025-11-03T17:08:32.3227997Z  # via 2025-11-03T17:08:32.3228162Z # -c snapshot_constraint.txt 2025-11-03T17:08:32.3228393Z # schemathesis 2025-11-03T17:08:32.3228582Z hypothesis-jsonschema==0.23.1 2025-11-03T17:08:32.3228788Z  # via 2025-11-03T17:08:32.3228951Z # -c snapshot_constraint.txt 2025-11-03T17:08:32.3229178Z # schemathesis 2025-11-03T17:08:32.3229346Z idna==3.10 2025-11-03T17:08:32.3229504Z  # via 2025-11-03T17:08:32.3229666Z # -c snapshot_constraint.txt 2025-11-03T17:08:32.3229852Z # anyio 2025-11-03T17:08:32.3230003Z # email-validator 2025-11-03T17:08:32.3230162Z # httpx 2025-11-03T17:08:32.3230308Z # jsonschema 2025-11-03T17:08:32.3230464Z # requests 2025-11-03T17:08:32.3230652Z # yarl 2025-11-03T17:08:32.3230811Z imageio==2.37.0 2025-11-03T17:08:32.3230986Z  # via 2025-11-03T17:08:32.3231146Z # -c snapshot_constraint.txt 2025-11-03T17:08:32.3231372Z # scikit-image 2025-11-03T17:08:32.3231557Z importlib-metadata==8.7.0 2025-11-03T17:08:32.3231751Z  # via 2025-11-03T17:08:32.3231909Z # -c snapshot_constraint.txt 2025-11-03T17:08:32.3232099Z # mlflow-skinny 2025-11-03T17:08:32.3232307Z # opentelemetry-api 2025-11-03T17:08:32.3232511Z importlib-resources==6.5.2 2025-11-03T17:08:32.3232712Z  # via 2025-11-03T17:08:32.3232862Z # -c snapshot_constraint.txt 2025-11-03T17:08:32.3233094Z # typeshed-client 2025-11-03T17:08:32.3233279Z inflect==5.6.2 2025-11-03T17:08:32.3233464Z  # via 2025-11-03T17:08:32.3233617Z # -c snapshot_constraint.txt 2025-11-03T17:08:32.3233870Z # datamodel-code-generator 2025-11-03T17:08:32.3234080Z iniconfig==2.0.0 2025-11-03T17:08:32.3234393Z  # via 2025-11-03T17:08:32.3234553Z # -c snapshot_constraint.txt 2025-11-03T17:08:32.3234775Z # pytest 2025-11-03T17:08:32.3234939Z isoduration==20.11.0 2025-11-03T17:08:32.3235148Z  # via 2025-11-03T17:08:32.3235299Z # -c snapshot_constraint.txt 2025-11-03T17:08:32.3235525Z # jsonschema 2025-11-03T17:08:32.3235697Z isort==5.13.2 2025-11-03T17:08:32.3235869Z  # via 2025-11-03T17:08:32.3236024Z # -c snapshot_constraint.txt 2025-11-03T17:08:32.3236271Z # datamodel-code-generator 2025-11-03T17:08:32.3236485Z itsdangerous==2.2.0 2025-11-03T17:08:32.3236666Z  # via 2025-11-03T17:08:32.3236818Z # -c snapshot_constraint.txt 2025-11-03T17:08:32.3237039Z # flask 2025-11-03T17:08:32.3237198Z jinja2==3.1.6 2025-11-03T17:08:32.3237370Z  # via 2025-11-03T17:08:32.3237522Z # -c snapshot_constraint.txt 2025-11-03T17:08:32.3237730Z # datamodel-code-generator 2025-11-03T17:08:32.3237920Z # flask 2025-11-03T17:08:32.3238061Z # mlflow 2025-11-03T17:08:32.3238241Z # torch 2025-11-03T17:08:32.3238392Z jiwer==3.0.5 2025-11-03T17:08:32.3238556Z  # via 2025-11-03T17:08:32.3238712Z # -c snapshot_constraint.txt 2025-11-03T17:08:32.3238951Z # -r requirements/test.in 2025-11-03T17:08:32.3239154Z jmespath==1.0.1 2025-11-03T17:08:32.3239331Z  # via 2025-11-03T17:08:32.3239490Z # -c snapshot_constraint.txt 2025-11-03T17:08:32.3239677Z # boto3 2025-11-03T17:08:32.3239850Z # botocore 2025-11-03T17:08:32.3240012Z joblib==1.4.2 2025-11-03T17:08:32.3240182Z  # via 2025-11-03T17:08:32.3240341Z # -c snapshot_constraint.txt 2025-11-03T17:08:32.3240527Z # librosa 2025-11-03T17:08:32.3240674Z # nltk 2025-11-03T17:08:32.3240849Z # scikit-learn 2025-11-03T17:08:32.3241031Z jsonargparse==4.35.0 2025-11-03T17:08:32.3241218Z  # via 2025-11-03T17:08:32.3241376Z # -c snapshot_constraint.txt 2025-11-03T17:08:32.3241564Z # lightning 2025-11-03T17:08:32.3241750Z # terratorch 2025-11-03T17:08:32.3241922Z jsonlines==4.0.0 2025-11-03T17:08:32.3242101Z  # via 2025-11-03T17:08:32.3243014Z # -c snapshot_constraint.txt 2025-11-03T17:08:32.3243243Z # lm-eval 2025-11-03T17:08:32.3243407Z jsonpointer==3.0.0 2025-11-03T17:08:32.3243581Z  # via 2025-11-03T17:08:32.3243740Z # -c snapshot_constraint.txt 2025-11-03T17:08:32.3243965Z # jsonschema 2025-11-03T17:08:32.3244136Z jsonschema==4.23.0 2025-11-03T17:08:32.3244307Z  # via 2025-11-03T17:08:32.3244478Z # -c snapshot_constraint.txt 2025-11-03T17:08:32.3244684Z # hypothesis-jsonschema 2025-11-03T17:08:32.3244879Z # mistral-common 2025-11-03T17:08:32.3245040Z # ray 2025-11-03T17:08:32.3245226Z # schemathesis 2025-11-03T17:08:32.3245427Z jsonschema-specifications==2024.10.1 2025-11-03T17:08:32.3245659Z  # via 2025-11-03T17:08:32.3245820Z # -c snapshot_constraint.txt 2025-11-03T17:08:32.3246044Z # jsonschema 2025-11-03T17:08:32.3246216Z junit-xml==1.9 2025-11-03T17:08:32.3246388Z  # via 2025-11-03T17:08:32.3246545Z # -c snapshot_constraint.txt 2025-11-03T17:08:32.3246788Z # schemathesis 2025-11-03T17:08:32.3246967Z kaleido==0.2.1 2025-11-03T17:08:32.3247136Z  # via 2025-11-03T17:08:32.3247289Z # -c snapshot_constraint.txt 2025-11-03T17:08:32.3247512Z # genai-perf 2025-11-03T17:08:32.3247680Z kiwisolver==1.4.7 2025-11-03T17:08:32.3247857Z  # via 2025-11-03T17:08:32.3248011Z # -c snapshot_constraint.txt 2025-11-03T17:08:32.3248235Z # matplotlib 2025-11-03T17:08:32.3248396Z kornia==0.8.1 2025-11-03T17:08:32.3248560Z  # via 2025-11-03T17:08:32.3248723Z # -c snapshot_constraint.txt 2025-11-03T17:08:32.3248946Z # torchgeo 2025-11-03T17:08:32.3249114Z kornia-rs==0.1.9 2025-11-03T17:08:32.3249283Z  # via 2025-11-03T17:08:32.3249528Z # -c snapshot_constraint.txt 2025-11-03T17:08:32.3249755Z # kornia 2025-11-03T17:08:32.3249915Z lazy-loader==0.4 2025-11-03T17:08:32.3250086Z  # via 2025-11-03T17:08:32.3250247Z # -c snapshot_constraint.txt 2025-11-03T17:08:32.3250436Z # librosa 2025-11-03T17:08:32.3250619Z # scikit-image 2025-11-03T17:08:32.3250784Z libnacl==2.1.0 2025-11-03T17:08:32.3250957Z  # via 2025-11-03T17:08:32.3251115Z # -c snapshot_constraint.txt 2025-11-03T17:08:32.3251336Z # tensorizer 2025-11-03T17:08:32.3251512Z librosa==0.10.2.post1 2025-11-03T17:08:32.3251699Z  # via 2025-11-03T17:08:32.3251856Z # -c snapshot_constraint.txt 2025-11-03T17:08:32.3252094Z # -r requirements/test.in 2025-11-03T17:08:32.3252290Z lightly==1.5.20 2025-11-03T17:08:32.3252463Z  # via 2025-11-03T17:08:32.3252622Z # -c snapshot_constraint.txt 2025-11-03T17:08:32.3252814Z # terratorch 2025-11-03T17:08:32.3252993Z # torchgeo 2025-11-03T17:08:32.3253164Z lightly-utils==0.0.2 2025-11-03T17:08:32.3253358Z  # via 2025-11-03T17:08:32.3253516Z # -c snapshot_constraint.txt 2025-11-03T17:08:32.3264258Z # lightly 2025-11-03T17:08:32.3264518Z lightning==2.5.1.post0 2025-11-03T17:08:32.3264773Z  # via 2025-11-03T17:08:32.3264968Z # -c snapshot_constraint.txt 2025-11-03T17:08:32.3265185Z # terratorch 2025-11-03T17:08:32.3265386Z # torchgeo 2025-11-03T17:08:32.3265571Z lightning-utilities==0.14.3 2025-11-03T17:08:32.3265795Z  # via 2025-11-03T17:08:32.3265974Z # -c snapshot_constraint.txt 2025-11-03T17:08:32.3266180Z # lightning 2025-11-03T17:08:32.3266347Z # pytorch-lightning 2025-11-03T17:08:32.3266575Z # torchmetrics 2025-11-03T17:08:32.3266760Z llvmlite==0.44.0 2025-11-03T17:08:32.3266942Z  # via 2025-11-03T17:08:32.3267112Z # -c snapshot_constraint.txt 2025-11-03T17:08:32.3267339Z # numba 2025-11-03T17:08:32.3267752Z lm-eval @ git+https://github.com/EleutherAI/lm-evaluation-harness.git@206b7722158f58c35b7ffcd53b035fdbdda5126d 2025-11-03T17:08:32.3268194Z  # via 2025-11-03T17:08:32.3268363Z # -c snapshot_constraint.txt 2025-11-03T17:08:32.3268746Z # -r requirements/test.in 2025-11-03T17:08:32.3268952Z lxml==5.3.0 2025-11-03T17:08:32.3269125Z  # via 2025-11-03T17:08:32.3269287Z # -c snapshot_constraint.txt 2025-11-03T17:08:32.3269482Z # blobfile 2025-11-03T17:08:32.3269650Z # gpt-oss 2025-11-03T17:08:32.3269836Z # sacrebleu 2025-11-03T17:08:32.3270005Z mako==1.3.10 2025-11-03T17:08:32.3270178Z  # via 2025-11-03T17:08:32.3270348Z # -c snapshot_constraint.txt 2025-11-03T17:08:32.3270575Z # alembic 2025-11-03T17:08:32.3270738Z markdown==3.8.2 2025-11-03T17:08:32.3270927Z  # via 2025-11-03T17:08:32.3271090Z # -c snapshot_constraint.txt 2025-11-03T17:08:32.3271309Z # mlflow 2025-11-03T17:08:32.3271474Z markdown-it-py==3.0.0 2025-11-03T17:08:32.3271676Z  # via 2025-11-03T17:08:32.3271839Z # -c snapshot_constraint.txt 2025-11-03T17:08:32.3272062Z # rich 2025-11-03T17:08:32.3272220Z markupsafe==3.0.1 2025-11-03T17:08:32.3272404Z  # via 2025-11-03T17:08:32.3272568Z # -c snapshot_constraint.txt 2025-11-03T17:08:32.3272763Z # flask 2025-11-03T17:08:32.3272903Z # jinja2 2025-11-03T17:08:32.3273047Z # mako 2025-11-03T17:08:32.3273231Z # werkzeug 2025-11-03T17:08:32.3273399Z matplotlib==3.9.2 2025-11-03T17:08:32.3273576Z  # via 2025-11-03T17:08:32.3273742Z # -c snapshot_constraint.txt 2025-11-03T17:08:32.3273960Z # -r requirements/test.in 2025-11-03T17:08:32.3274147Z # lightning 2025-11-03T17:08:32.3274300Z # mlflow 2025-11-03T17:08:32.3274452Z # pycocotools 2025-11-03T17:08:32.3274638Z # torchgeo 2025-11-03T17:08:32.3274799Z mbstrdecoder==1.1.3 2025-11-03T17:08:32.3274988Z  # via 2025-11-03T17:08:32.3275229Z # -c snapshot_constraint.txt 2025-11-03T17:08:32.3275431Z # dataproperty 2025-11-03T17:08:32.3275590Z # pytablewriter 2025-11-03T17:08:32.3275780Z # typepy 2025-11-03T17:08:32.3275948Z mdurl==0.1.2 2025-11-03T17:08:32.3276121Z  # via 2025-11-03T17:08:32.3276275Z # -c snapshot_constraint.txt 2025-11-03T17:08:32.3276510Z # markdown-it-py 2025-11-03T17:08:32.3276706Z mistral-common==1.8.5 2025-11-03T17:08:32.3276898Z  # via 2025-11-03T17:08:32.3277059Z # -c snapshot_constraint.txt 2025-11-03T17:08:32.3277310Z # -r requirements/test.in 2025-11-03T17:08:32.3277512Z mlflow==2.22.0 2025-11-03T17:08:32.3277684Z  # via 2025-11-03T17:08:32.3277837Z # -c snapshot_constraint.txt 2025-11-03T17:08:32.3278064Z # terratorch 2025-11-03T17:08:32.3278242Z mlflow-skinny==2.22.0 2025-11-03T17:08:32.3278425Z  # via 2025-11-03T17:08:32.3278584Z # -c snapshot_constraint.txt 2025-11-03T17:08:32.3278802Z # mlflow 2025-11-03T17:08:32.3278972Z more-itertools==10.5.0 2025-11-03T17:08:32.3279165Z  # via 2025-11-03T17:08:32.3279320Z # -c snapshot_constraint.txt 2025-11-03T17:08:32.3279541Z # lm-eval 2025-11-03T17:08:32.3279708Z mpmath==1.3.0 2025-11-03T17:08:32.3279875Z  # via 2025-11-03T17:08:32.3280037Z # -c snapshot_constraint.txt 2025-11-03T17:08:32.3280255Z # sympy 2025-11-03T17:08:32.3280415Z msgpack==1.1.0 2025-11-03T17:08:32.3280582Z  # via 2025-11-03T17:08:32.3280740Z # -c snapshot_constraint.txt 2025-11-03T17:08:32.3280927Z # librosa 2025-11-03T17:08:32.3281102Z # ray 2025-11-03T17:08:32.3281254Z mteb==1.38.11 2025-11-03T17:08:32.3281424Z  # via 2025-11-03T17:08:32.3281599Z # -c snapshot_constraint.txt 2025-11-03T17:08:32.3281845Z # -r requirements/test.in 2025-11-03T17:08:32.3282044Z multidict==6.1.0 2025-11-03T17:08:32.3282222Z  # via 2025-11-03T17:08:32.3282378Z # -c snapshot_constraint.txt 2025-11-03T17:08:32.3282574Z # aiohttp 2025-11-03T17:08:32.3282742Z # yarl 2025-11-03T17:08:32.3282910Z multiprocess==0.70.16 2025-11-03T17:08:32.3283095Z  # via 2025-11-03T17:08:32.3283354Z # -c snapshot_constraint.txt 2025-11-03T17:08:32.3283541Z # datasets 2025-11-03T17:08:32.3283723Z # evaluate 2025-11-03T17:08:32.3283881Z munch==4.0.0 2025-11-03T17:08:32.3284048Z  # via 2025-11-03T17:08:32.3284202Z # -c snapshot_constraint.txt 2025-11-03T17:08:32.3284441Z # pretrainedmodels 2025-11-03T17:08:32.3284642Z mypy-extensions==1.0.0 2025-11-03T17:08:32.3284834Z  # via 2025-11-03T17:08:32.3284988Z # -c snapshot_constraint.txt 2025-11-03T17:08:32.3285202Z # black 2025-11-03T17:08:32.3285360Z networkx==3.2.1 2025-11-03T17:08:32.3285521Z  # via 2025-11-03T17:08:32.3285676Z # -c snapshot_constraint.txt 2025-11-03T17:08:32.3285872Z # scikit-image 2025-11-03T17:08:32.3286055Z # torch 2025-11-03T17:08:32.3286205Z nltk==3.9.1 2025-11-03T17:08:32.3286376Z  # via 2025-11-03T17:08:32.3286536Z # -c snapshot_constraint.txt 2025-11-03T17:08:32.3286762Z # rouge-score 2025-11-03T17:08:32.3286950Z num2words==0.5.14 2025-11-03T17:08:32.3287130Z  # via 2025-11-03T17:08:32.3287290Z # -c snapshot_constraint.txt 2025-11-03T17:08:32.3287532Z # -r requirements/test.in 2025-11-03T17:08:32.3287727Z numba==0.61.2 2025-11-03T17:08:32.3287897Z  # via 2025-11-03T17:08:32.3288056Z # -c snapshot_constraint.txt 2025-11-03T17:08:32.3288262Z # -r requirements/test.in 2025-11-03T17:08:32.3288475Z # librosa 2025-11-03T17:08:32.3288638Z numexpr==2.10.1 2025-11-03T17:08:32.3288808Z  # via 2025-11-03T17:08:32.3288971Z # -c snapshot_constraint.txt 2025-11-03T17:08:32.3289207Z # lm-eval 2025-11-03T17:08:32.3289370Z numpy==1.26.4 2025-11-03T17:08:32.3289541Z  # via 2025-11-03T17:08:32.3289703Z # -c snapshot_constraint.txt 2025-11-03T17:08:32.3289981Z # -r requirements/test.in 2025-11-03T17:08:32.3290171Z # accelerate 2025-11-03T17:08:32.3290325Z # albucore 2025-11-03T17:08:32.3290483Z # albumentations 2025-11-03T17:08:32.3290651Z # bitsandbytes 2025-11-03T17:08:32.3290804Z # bm25s 2025-11-03T17:08:32.3290951Z # contourpy 2025-11-03T17:08:32.3291103Z # cupy-cuda12x 2025-11-03T17:08:32.3291264Z # datasets 2025-11-03T17:08:32.3291423Z # decord 2025-11-03T17:08:32.3291567Z # einx 2025-11-03T17:08:32.3291707Z # encodec 2025-11-03T17:08:32.3291855Z # evaluate 2025-11-03T17:08:32.3292003Z # fastparquet 2025-11-03T17:08:32.3292165Z # genai-perf 2025-11-03T17:08:32.3292315Z # geopandas 2025-11-03T17:08:32.3292466Z # h5py 2025-11-03T17:08:32.3292610Z # imageio 2025-11-03T17:08:32.3292754Z # librosa 2025-11-03T17:08:32.3292894Z # lightly 2025-11-03T17:08:32.3293045Z # lightly-utils 2025-11-03T17:08:32.3293212Z # matplotlib 2025-11-03T17:08:32.3293386Z # mistral-common 2025-11-03T17:08:32.3293544Z # mlflow 2025-11-03T17:08:32.3293688Z # mteb 2025-11-03T17:08:32.3293830Z # numba 2025-11-03T17:08:32.3293973Z # numexpr 2025-11-03T17:08:32.3294137Z # opencv-python-headless 2025-11-03T17:08:32.3294326Z # pandas 2025-11-03T17:08:32.3294471Z # patsy 2025-11-03T17:08:32.3294606Z # peft 2025-11-03T17:08:32.3294752Z # pycocotools 2025-11-03T17:08:32.3294909Z # pyogrio 2025-11-03T17:08:32.3295055Z # rasterio 2025-11-03T17:08:32.3295200Z # rioxarray 2025-11-03T17:08:32.3295356Z # rouge-score 2025-11-03T17:08:32.3295526Z # runai-model-streamer 2025-11-03T17:08:32.3295715Z # sacrebleu 2025-11-03T17:08:32.3295867Z # scikit-image 2025-11-03T17:08:32.3296035Z # scikit-learn 2025-11-03T17:08:32.3296197Z # scipy 2025-11-03T17:08:32.3296375Z # segmentation-models-pytorch 2025-11-03T17:08:32.3296592Z # shapely 2025-11-03T17:08:32.3296743Z # soxr 2025-11-03T17:08:32.3296895Z # statsmodels 2025-11-03T17:08:32.3297054Z # tensorboardx 2025-11-03T17:08:32.3297208Z # tensorizer 2025-11-03T17:08:32.3297466Z # tifffile 2025-11-03T17:08:32.3297616Z # torchgeo 2025-11-03T17:08:32.3297769Z # torchmetrics 2025-11-03T17:08:32.3297946Z # torchvision 2025-11-03T17:08:32.3298102Z # transformers 2025-11-03T17:08:32.3298263Z # tritonclient 2025-11-03T17:08:32.3298415Z # vocos 2025-11-03T17:08:32.3298597Z # xarray 2025-11-03T17:08:32.3298747Z omegaconf==2.3.0 2025-11-03T17:08:32.3298928Z  # via 2025-11-03T17:08:32.3299093Z # -c snapshot_constraint.txt 2025-11-03T17:08:32.3299295Z # hydra-core 2025-11-03T17:08:32.3299476Z # lightning 2025-11-03T17:08:32.3299655Z open-clip-torch==2.32.0 2025-11-03T17:08:32.3299892Z  # via -r requirements/test.in 2025-11-03T17:08:32.3300123Z openai-harmony==0.0.4 2025-11-03T17:08:32.3300318Z  # via 2025-11-03T17:08:32.3300488Z # -c snapshot_constraint.txt 2025-11-03T17:08:32.3300719Z # gpt-oss 2025-11-03T17:08:32.3300898Z opencensus==0.11.4 2025-11-03T17:08:32.3301088Z  # via 2025-11-03T17:08:32.3301261Z # -c snapshot_constraint.txt 2025-11-03T17:08:32.3301479Z # ray 2025-11-03T17:08:32.3301753Z opencensus-context==0.1.3 2025-11-03T17:08:32.3301956Z  # via 2025-11-03T17:08:32.3302114Z # -c snapshot_constraint.txt 2025-11-03T17:08:32.3302333Z # opencensus 2025-11-03T17:08:32.3302528Z opencv-python-headless==4.11.0.86 2025-11-03T17:08:32.3302746Z  # via 2025-11-03T17:08:32.3302910Z # -c snapshot_constraint.txt 2025-11-03T17:08:32.3303113Z # -r requirements/test.in 2025-11-03T17:08:32.3303302Z # albucore 2025-11-03T17:08:32.3303454Z # albumentations 2025-11-03T17:08:32.3303669Z # mistral-common 2025-11-03T17:08:32.3303868Z opentelemetry-api==1.35.0 2025-11-03T17:08:32.3304065Z  # via 2025-11-03T17:08:32.3304306Z # -c snapshot_constraint.txt 2025-11-03T17:08:32.3304510Z # mlflow-skinny 2025-11-03T17:08:32.3304709Z # opentelemetry-exporter-prometheus 2025-11-03T17:08:32.3304946Z # opentelemetry-sdk 2025-11-03T17:08:32.3305202Z # opentelemetry-semantic-conventions 2025-11-03T17:08:32.3305472Z opentelemetry-exporter-prometheus==0.56b0 2025-11-03T17:08:32.3305728Z  # via 2025-11-03T17:08:32.3305898Z # -c snapshot_constraint.txt 2025-11-03T17:08:32.3306110Z # ray 2025-11-03T17:08:32.3306280Z opentelemetry-proto==1.36.0 2025-11-03T17:08:32.3306489Z  # via 2025-11-03T17:08:32.3306646Z # -c snapshot_constraint.txt 2025-11-03T17:08:32.3306857Z # ray 2025-11-03T17:08:32.3307023Z opentelemetry-sdk==1.35.0 2025-11-03T17:08:32.3307223Z  # via 2025-11-03T17:08:32.3307380Z # -c snapshot_constraint.txt 2025-11-03T17:08:32.3307578Z # mlflow-skinny 2025-11-03T17:08:32.3307777Z # opentelemetry-exporter-prometheus 2025-11-03T17:08:32.3308026Z # ray 2025-11-03T17:08:32.3308211Z opentelemetry-semantic-conventions==0.56b0 2025-11-03T17:08:32.3308457Z  # via 2025-11-03T17:08:32.3308628Z # -c snapshot_constraint.txt 2025-11-03T17:08:32.3308865Z # opentelemetry-sdk 2025-11-03T17:08:32.3309054Z packaging==24.2 2025-11-03T17:08:32.3309232Z  # via 2025-11-03T17:08:32.3309398Z # -c snapshot_constraint.txt 2025-11-03T17:08:32.3309593Z # accelerate 2025-11-03T17:08:32.3309745Z # black 2025-11-03T17:08:32.3309917Z # datamodel-code-generator 2025-11-03T17:08:32.3310114Z # datasets 2025-11-03T17:08:32.3310265Z # evaluate 2025-11-03T17:08:32.3310414Z # fastparquet 2025-11-03T17:08:32.3310577Z # geopandas 2025-11-03T17:08:32.3310727Z # gunicorn 2025-11-03T17:08:32.3310885Z # huggingface-hub 2025-11-03T17:08:32.3311049Z # hydra-core 2025-11-03T17:08:32.3311202Z # kornia 2025-11-03T17:08:32.3311354Z # lazy-loader 2025-11-03T17:08:32.3311509Z # lightning 2025-11-03T17:08:32.3311667Z # lightning-utilities 2025-11-03T17:08:32.3311862Z # matplotlib 2025-11-03T17:08:32.3312104Z # mlflow-skinny 2025-11-03T17:08:32.3312270Z # peft 2025-11-03T17:08:32.3312407Z # plotly 2025-11-03T17:08:32.3312561Z # pooch 2025-11-03T17:08:32.3312708Z # pyogrio 2025-11-03T17:08:32.3312851Z # pytest 2025-11-03T17:08:32.3313011Z # pytest-rerunfailures 2025-11-03T17:08:32.3313208Z # pytorch-lightning 2025-11-03T17:08:32.3313384Z # ray 2025-11-03T17:08:32.3313526Z # rioxarray 2025-11-03T17:08:32.3313682Z # scikit-image 2025-11-03T17:08:32.3313846Z # statsmodels 2025-11-03T17:08:32.3314004Z # tensorboardx 2025-11-03T17:08:32.3314154Z # torchmetrics 2025-11-03T17:08:32.3314316Z # transformers 2025-11-03T17:08:32.3314480Z # typepy 2025-11-03T17:08:32.3314661Z # xarray 2025-11-03T17:08:32.3314813Z pandas==2.2.3 2025-11-03T17:08:32.3314995Z  # via 2025-11-03T17:08:32.3315170Z # -c snapshot_constraint.txt 2025-11-03T17:08:32.3315365Z # datasets 2025-11-03T17:08:32.3315513Z # evaluate 2025-11-03T17:08:32.3315661Z # fastparquet 2025-11-03T17:08:32.3315819Z # genai-perf 2025-11-03T17:08:32.3315974Z # geopandas 2025-11-03T17:08:32.3316121Z # mlflow 2025-11-03T17:08:32.3316271Z # statsmodels 2025-11-03T17:08:32.3316424Z # torchgeo 2025-11-03T17:08:32.3316598Z # xarray 2025-11-03T17:08:32.3316758Z pathspec==0.12.1 2025-11-03T17:08:32.3316935Z  # via 2025-11-03T17:08:32.3317098Z # -c snapshot_constraint.txt 2025-11-03T17:08:32.3317313Z # black 2025-11-03T17:08:32.3317474Z pathvalidate==3.2.1 2025-11-03T17:08:32.3317650Z  # via 2025-11-03T17:08:32.3317812Z # -c snapshot_constraint.txt 2025-11-03T17:08:32.3318039Z # pytablewriter 2025-11-03T17:08:32.3318222Z patsy==1.0.1 2025-11-03T17:08:32.3318394Z  # via 2025-11-03T17:08:32.3318648Z # -c snapshot_constraint.txt 2025-11-03T17:08:32.3318885Z # statsmodels 2025-11-03T17:08:32.3319062Z peft==0.16.0 2025-11-03T17:08:32.3319235Z  # via 2025-11-03T17:08:32.3319394Z # -c snapshot_constraint.txt 2025-11-03T17:08:32.3319593Z # -r requirements/test.in 2025-11-03T17:08:32.3319821Z # lm-eval 2025-11-03T17:08:32.3319984Z pillow==10.4.0 2025-11-03T17:08:32.3320156Z  # via 2025-11-03T17:08:32.3320310Z # -c snapshot_constraint.txt 2025-11-03T17:08:32.3320505Z # genai-perf 2025-11-03T17:08:32.3320660Z # imageio 2025-11-03T17:08:32.3320813Z # lightly-utils 2025-11-03T17:08:32.3320972Z # matplotlib 2025-11-03T17:08:32.3321427Z # mistral-common 2025-11-03T17:08:32.3321613Z # scikit-image 2025-11-03T17:08:32.3321795Z # segmentation-models-pytorch 2025-11-03T17:08:32.3322015Z # sentence-transformers 2025-11-03T17:08:32.3322205Z # torchgeo 2025-11-03T17:08:32.3322415Z # torchvision 2025-11-03T17:08:32.3322592Z platformdirs==4.3.6 2025-11-03T17:08:32.3322786Z  # via 2025-11-03T17:08:32.3322954Z # -c snapshot_constraint.txt 2025-11-03T17:08:32.3323152Z # black 2025-11-03T17:08:32.3323295Z # pooch 2025-11-03T17:08:32.3323477Z # virtualenv 2025-11-03T17:08:32.3323650Z plotly==5.24.1 2025-11-03T17:08:32.3323822Z  # via 2025-11-03T17:08:32.3323978Z # -c snapshot_constraint.txt 2025-11-03T17:08:32.3324204Z # genai-perf 2025-11-03T17:08:32.3324373Z pluggy==1.5.0 2025-11-03T17:08:32.3324544Z  # via 2025-11-03T17:08:32.3324697Z # -c snapshot_constraint.txt 2025-11-03T17:08:32.3324889Z # pytest 2025-11-03T17:08:32.3325070Z # pytest-cov 2025-11-03T17:08:32.3325239Z polars==1.29.0 2025-11-03T17:08:32.3325404Z  # via 2025-11-03T17:08:32.3325563Z # -c snapshot_constraint.txt 2025-11-03T17:08:32.3325795Z # mteb 2025-11-03T17:08:32.3325955Z pooch==1.8.2 2025-11-03T17:08:32.3326121Z  # via 2025-11-03T17:08:32.3326283Z # -c snapshot_constraint.txt 2025-11-03T17:08:32.3326502Z # librosa 2025-11-03T17:08:32.3326807Z portalocker==2.10.1 2025-11-03T17:08:32.3326994Z  # via 2025-11-03T17:08:32.3327161Z # -c snapshot_constraint.txt 2025-11-03T17:08:32.3327381Z # sacrebleu 2025-11-03T17:08:32.3327542Z pqdm==0.2.0 2025-11-03T17:08:32.3327715Z  # via 2025-11-03T17:08:32.3327873Z # -c snapshot_constraint.txt 2025-11-03T17:08:32.3328123Z # -r requirements/test.in 2025-11-03T17:08:32.3328337Z pretrainedmodels==0.7.4 2025-11-03T17:08:32.3328528Z  # via 2025-11-03T17:08:32.3328687Z # -c snapshot_constraint.txt 2025-11-03T17:08:32.3328947Z # segmentation-models-pytorch 2025-11-03T17:08:32.3329174Z prometheus-client==0.22.0 2025-11-03T17:08:32.3329374Z  # via 2025-11-03T17:08:32.3329530Z # -c snapshot_constraint.txt 2025-11-03T17:08:32.3329756Z # opentelemetry-exporter-prometheus 2025-11-03T17:08:32.3329999Z # ray 2025-11-03T17:08:32.3330160Z propcache==0.2.0 2025-11-03T17:08:32.3330335Z  # via 2025-11-03T17:08:32.3330498Z # -c snapshot_constraint.txt 2025-11-03T17:08:32.3330576Z # aiohttp 2025-11-03T17:08:32.3330660Z # yarl 2025-11-03T17:08:32.3330735Z proto-plus==1.26.1 2025-11-03T17:08:32.3330815Z  # via 2025-11-03T17:08:32.3330888Z # -c snapshot_constraint.txt 2025-11-03T17:08:32.3330997Z # google-api-core 2025-11-03T17:08:32.3331062Z protobuf==5.28.3 2025-11-03T17:08:32.3331142Z  # via 2025-11-03T17:08:32.3331219Z # -c snapshot_constraint.txt 2025-11-03T17:08:32.3331283Z # google-api-core 2025-11-03T17:08:32.3331369Z # googleapis-common-protos 2025-11-03T17:08:32.3331434Z # mlflow-skinny 2025-11-03T17:08:32.3331506Z # opentelemetry-proto 2025-11-03T17:08:32.3331574Z # proto-plus 2025-11-03T17:08:32.3331634Z # ray 2025-11-03T17:08:32.3331820Z # tensorboardx 2025-11-03T17:08:32.3331921Z # tensorizer 2025-11-03T17:08:32.3331983Z psutil==6.1.0 2025-11-03T17:08:32.3332070Z  # via 2025-11-03T17:08:32.3332160Z # -c snapshot_constraint.txt 2025-11-03T17:08:32.3332221Z # accelerate 2025-11-03T17:08:32.3332287Z # peft 2025-11-03T17:08:32.3332377Z # tensorizer 2025-11-03T17:08:32.3332441Z py==1.11.0 2025-11-03T17:08:32.3332519Z  # via 2025-11-03T17:08:32.3332591Z # -c snapshot_constraint.txt 2025-11-03T17:08:32.3332696Z # pytest-forked 2025-11-03T17:08:32.3332757Z py-spy==0.4.0 2025-11-03T17:08:32.3332842Z  # via 2025-11-03T17:08:32.3332914Z # -c snapshot_constraint.txt 2025-11-03T17:08:32.3332995Z # ray 2025-11-03T17:08:32.3333062Z pyarrow==18.0.0 2025-11-03T17:08:32.3333141Z  # via 2025-11-03T17:08:32.3333212Z # -c snapshot_constraint.txt 2025-11-03T17:08:32.3333274Z # datasets 2025-11-03T17:08:32.3333332Z # genai-perf 2025-11-03T17:08:32.3333427Z # mlflow 2025-11-03T17:08:32.3333488Z pyasn1==0.6.1 2025-11-03T17:08:32.3333568Z  # via 2025-11-03T17:08:32.3333645Z # -c snapshot_constraint.txt 2025-11-03T17:08:32.3333712Z # pyasn1-modules 2025-11-03T17:08:32.3333797Z # rsa 2025-11-03T17:08:32.3333864Z pyasn1-modules==0.4.2 2025-11-03T17:08:32.3333941Z  # via 2025-11-03T17:08:32.3334017Z # -c snapshot_constraint.txt 2025-11-03T17:08:32.3334108Z # google-auth 2025-11-03T17:08:32.3334170Z pybind11==2.13.6 2025-11-03T17:08:32.3334253Z  # via 2025-11-03T17:08:32.3334328Z # -c snapshot_constraint.txt 2025-11-03T17:08:32.3334430Z # lm-eval 2025-11-03T17:08:32.3334496Z pycocotools==2.0.8 2025-11-03T17:08:32.3334576Z  # via 2025-11-03T17:08:32.3334652Z # -c snapshot_constraint.txt 2025-11-03T17:08:32.3334742Z # terratorch 2025-11-03T17:08:32.3334810Z pycountry==24.6.1 2025-11-03T17:08:32.3334890Z  # via 2025-11-03T17:08:32.3334965Z # -c snapshot_constraint.txt 2025-11-03T17:08:32.3335080Z # pydantic-extra-types 2025-11-03T17:08:32.3335145Z pycparser==2.22 2025-11-03T17:08:32.3335314Z  # via 2025-11-03T17:08:32.3335394Z # -c snapshot_constraint.txt 2025-11-03T17:08:32.3335476Z # cffi 2025-11-03T17:08:32.3335550Z pycryptodomex==3.22.0 2025-11-03T17:08:32.3335629Z  # via 2025-11-03T17:08:32.3335702Z # -c snapshot_constraint.txt 2025-11-03T17:08:32.3335794Z # blobfile 2025-11-03T17:08:32.3335857Z pydantic==2.12.0 2025-11-03T17:08:32.3335941Z  # via 2025-11-03T17:08:32.3336010Z # -c snapshot_constraint.txt 2025-11-03T17:08:32.3336081Z # -r requirements/test.in 2025-11-03T17:08:32.3336150Z # albumentations 2025-11-03T17:08:32.3336226Z # datamodel-code-generator 2025-11-03T17:08:32.3336286Z # fastapi 2025-11-03T17:08:32.3336351Z # gpt-oss 2025-11-03T17:08:32.3336408Z # lightly 2025-11-03T17:08:32.3336492Z # mistral-common 2025-11-03T17:08:32.3336557Z # mlflow-skinny 2025-11-03T17:08:32.3336615Z # mteb 2025-11-03T17:08:32.3336687Z # openai-harmony 2025-11-03T17:08:32.3336764Z # pydantic-extra-types 2025-11-03T17:08:32.3336854Z # ray 2025-11-03T17:08:32.3336922Z pydantic-core==2.41.1 2025-11-03T17:08:32.3337001Z  # via 2025-11-03T17:08:32.3337079Z # -c snapshot_constraint.txt 2025-11-03T17:08:32.3337167Z # pydantic 2025-11-03T17:08:32.3337243Z pydantic-extra-types==2.10.5 2025-11-03T17:08:32.3337328Z  # via 2025-11-03T17:08:32.3337400Z # -c snapshot_constraint.txt 2025-11-03T17:08:32.3337501Z # mistral-common 2025-11-03T17:08:32.3337565Z pygments==2.18.0 2025-11-03T17:08:32.3337644Z  # via 2025-11-03T17:08:32.3337725Z # -c snapshot_constraint.txt 2025-11-03T17:08:32.3337806Z # rich 2025-11-03T17:08:32.3337875Z pyogrio==0.11.0 2025-11-03T17:08:32.3337953Z  # via 2025-11-03T17:08:32.3338126Z # -c snapshot_constraint.txt 2025-11-03T17:08:32.3338229Z # geopandas 2025-11-03T17:08:32.3338295Z pyparsing==3.2.0 2025-11-03T17:08:32.3338375Z  # via 2025-11-03T17:08:32.3338460Z # -c snapshot_constraint.txt 2025-11-03T17:08:32.3338522Z # matplotlib 2025-11-03T17:08:32.3338616Z # rasterio 2025-11-03T17:08:32.3338678Z pyproj==3.7.1 2025-11-03T17:08:32.3338758Z  # via 2025-11-03T17:08:32.3338836Z # -c snapshot_constraint.txt 2025-11-03T17:08:32.3338897Z # geopandas 2025-11-03T17:08:32.3338955Z # rioxarray 2025-11-03T17:08:32.3339048Z # torchgeo 2025-11-03T17:08:32.3339115Z pyrate-limiter==3.7.0 2025-11-03T17:08:32.3339200Z  # via 2025-11-03T17:08:32.3339271Z # -c snapshot_constraint.txt 2025-11-03T17:08:32.3339365Z # schemathesis 2025-11-03T17:08:32.3339435Z pystemmer==3.0.0 2025-11-03T17:08:32.3339515Z  # via 2025-11-03T17:08:32.3339606Z # -c snapshot_constraint.txt 2025-11-03T17:08:32.3339691Z # mteb 2025-11-03T17:08:32.3339757Z pytablewriter==1.2.0 2025-11-03T17:08:32.3339841Z  # via 2025-11-03T17:08:32.3339912Z # -c snapshot_constraint.txt 2025-11-03T17:08:32.3340001Z # lm-eval 2025-11-03T17:08:32.3340068Z pytest==8.3.5 2025-11-03T17:08:32.3340146Z  # via 2025-11-03T17:08:32.3340222Z # -c snapshot_constraint.txt 2025-11-03T17:08:32.3340290Z # -r requirements/test.in 2025-11-03T17:08:32.3340364Z # buildkite-test-collector 2025-11-03T17:08:32.3340430Z # genai-perf 2025-11-03T17:08:32.3340494Z # pytest-asyncio 2025-11-03T17:08:32.3340557Z # pytest-cov 2025-11-03T17:08:32.3340620Z # pytest-forked 2025-11-03T17:08:32.3340680Z # pytest-mock 2025-11-03T17:08:32.3340755Z # pytest-rerunfailures 2025-11-03T17:08:32.3340817Z # pytest-shard 2025-11-03T17:08:32.3340883Z # pytest-subtests 2025-11-03T17:08:32.3340952Z # pytest-timeout 2025-11-03T17:08:32.3341013Z # schemathesis 2025-11-03T17:08:32.3341114Z # terratorch 2025-11-03T17:08:32.3341180Z pytest-asyncio==0.24.0 2025-11-03T17:08:32.3341259Z  # via 2025-11-03T17:08:32.3341335Z # -c snapshot_constraint.txt 2025-11-03T17:08:32.3341617Z # -r requirements/test.in 2025-11-03T17:08:32.3341689Z pytest-cov==6.3.0 2025-11-03T17:08:32.3341772Z  # via 2025-11-03T17:08:32.3341844Z # -c snapshot_constraint.txt 2025-11-03T17:08:32.3341954Z # -r requirements/test.in 2025-11-03T17:08:32.3342019Z pytest-forked==1.6.0 2025-11-03T17:08:32.3342099Z  # via 2025-11-03T17:08:32.3342175Z # -c snapshot_constraint.txt 2025-11-03T17:08:32.3342280Z # -r requirements/test.in 2025-11-03T17:08:32.3342349Z pytest-mock==3.14.0 2025-11-03T17:08:32.3342427Z  # via 2025-11-03T17:08:32.3342497Z # -c snapshot_constraint.txt 2025-11-03T17:08:32.3342593Z # genai-perf 2025-11-03T17:08:32.3342668Z pytest-rerunfailures==14.0 2025-11-03T17:08:32.3342753Z  # via 2025-11-03T17:08:32.3342829Z # -c snapshot_constraint.txt 2025-11-03T17:08:32.3342933Z # -r requirements/test.in 2025-11-03T17:08:32.3343003Z pytest-shard==0.1.2 2025-11-03T17:08:32.3343097Z  # via 2025-11-03T17:08:32.3343177Z # -c snapshot_constraint.txt 2025-11-03T17:08:32.3343283Z # -r requirements/test.in 2025-11-03T17:08:32.3343353Z pytest-subtests==0.14.1 2025-11-03T17:08:32.3343438Z  # via 2025-11-03T17:08:32.3343509Z # -c snapshot_constraint.txt 2025-11-03T17:08:32.3343603Z # schemathesis 2025-11-03T17:08:32.3343676Z pytest-timeout==2.3.1 2025-11-03T17:08:32.3343755Z  # via 2025-11-03T17:08:32.3343831Z # -c snapshot_constraint.txt 2025-11-03T17:08:32.3343935Z # -r requirements/test.in 2025-11-03T17:08:32.3343999Z python-box==7.3.2 2025-11-03T17:08:32.3344085Z  # via 2025-11-03T17:08:32.3344156Z # -c snapshot_constraint.txt 2025-11-03T17:08:32.3344251Z # terratorch 2025-11-03T17:08:32.3344452Z python-dateutil==2.9.0.post0 2025-11-03T17:08:32.3344539Z  # via 2025-11-03T17:08:32.3344618Z # -c snapshot_constraint.txt 2025-11-03T17:08:32.3344675Z # arrow 2025-11-03T17:08:32.3344738Z # botocore 2025-11-03T17:08:32.3344805Z # graphene 2025-11-03T17:08:32.3344861Z # lightly 2025-11-03T17:08:32.3344927Z # matplotlib 2025-11-03T17:08:32.3344986Z # pandas 2025-11-03T17:08:32.3345073Z # typepy 2025-11-03T17:08:32.3345149Z python-rapidjson==1.20 2025-11-03T17:08:32.3345228Z  # via 2025-11-03T17:08:32.3345308Z # -c snapshot_constraint.txt 2025-11-03T17:08:32.3345402Z # tritonclient 2025-11-03T17:08:32.3345474Z pytorch-lightning==2.5.5 2025-11-03T17:08:32.3345566Z  # via 2025-11-03T17:08:32.3345626Z # lightly 2025-11-03T17:08:32.3345717Z # lightning 2025-11-03T17:08:32.3345798Z pytrec-eval-terrier==0.5.7 2025-11-03T17:08:32.3345878Z  # via 2025-11-03T17:08:32.3345958Z # -c snapshot_constraint.txt 2025-11-03T17:08:32.3346040Z # mteb 2025-11-03T17:08:32.3346100Z pytz==2024.2 2025-11-03T17:08:32.3346185Z  # via 2025-11-03T17:08:32.3346261Z # -c snapshot_constraint.txt 2025-11-03T17:08:32.3346324Z # pandas 2025-11-03T17:08:32.3346418Z # typepy 2025-11-03T17:08:32.3346480Z pyyaml==6.0.2 2025-11-03T17:08:32.3346565Z  # via 2025-11-03T17:08:32.3346637Z # -c snapshot_constraint.txt 2025-11-03T17:08:32.3346697Z # accelerate 2025-11-03T17:08:32.3346768Z # albumentations 2025-11-03T17:08:32.3346843Z # datamodel-code-generator 2025-11-03T17:08:32.3346908Z # datasets 2025-11-03T17:08:32.3346968Z # genai-perf 2025-11-03T17:08:32.3347032Z # huggingface-hub 2025-11-03T17:08:32.3347100Z # jsonargparse 2025-11-03T17:08:32.3347158Z # lightning 2025-11-03T17:08:32.3347222Z # mlflow-skinny 2025-11-03T17:08:32.3347286Z # omegaconf 2025-11-03T17:08:32.3347343Z # peft 2025-11-03T17:08:32.3347419Z # pytorch-lightning 2025-11-03T17:08:32.3347487Z # ray 2025-11-03T17:08:32.3347548Z # responses 2025-11-03T17:08:32.3347614Z # schemathesis 2025-11-03T17:08:32.3347756Z # timm 2025-11-03T17:08:32.3347833Z # transformers 2025-11-03T17:08:32.3347924Z # vocos 2025-11-03T17:08:32.3347987Z rapidfuzz==3.12.1 2025-11-03T17:08:32.3348070Z  # via 2025-11-03T17:08:32.3348145Z # -c snapshot_constraint.txt 2025-11-03T17:08:32.3348226Z # jiwer 2025-11-03T17:08:32.3348294Z rasterio==1.4.3 2025-11-03T17:08:32.3348371Z  # via 2025-11-03T17:08:32.3348447Z # -c snapshot_constraint.txt 2025-11-03T17:08:32.3348506Z # rioxarray 2025-11-03T17:08:32.3348565Z # terratorch 2025-11-03T17:08:32.3348659Z # torchgeo 2025-11-03T17:08:32.3348717Z ray==2.48.0 2025-11-03T17:08:32.3348799Z  # via 2025-11-03T17:08:32.3348872Z # -c snapshot_constraint.txt 2025-11-03T17:08:32.3348975Z # -r requirements/test.in 2025-11-03T17:08:32.3349048Z redis==5.2.0 2025-11-03T17:08:32.3349137Z  # via 2025-11-03T17:08:32.3349209Z # -c snapshot_constraint.txt 2025-11-03T17:08:32.3349308Z # tensorizer 2025-11-03T17:08:32.3349371Z referencing==0.35.1 2025-11-03T17:08:32.3349455Z  # via 2025-11-03T17:08:32.3349528Z # -c snapshot_constraint.txt 2025-11-03T17:08:32.3349586Z # jsonschema 2025-11-03T17:08:32.3349710Z # jsonschema-specifications 2025-11-03T17:08:32.3349770Z regex==2024.9.11 2025-11-03T17:08:32.3349853Z  # via 2025-11-03T17:08:32.3349925Z # -c snapshot_constraint.txt 2025-11-03T17:08:32.3349982Z # nltk 2025-11-03T17:08:32.3350055Z # open-clip-torch 2025-11-03T17:08:32.3350114Z # sacrebleu 2025-11-03T17:08:32.3350171Z # tiktoken 2025-11-03T17:08:32.3350269Z # transformers 2025-11-03T17:08:32.3350329Z requests==2.32.3 2025-11-03T17:08:32.3350412Z  # via 2025-11-03T17:08:32.3350482Z # -c snapshot_constraint.txt 2025-11-03T17:08:32.3350641Z # buildkite-test-collector 2025-11-03T17:08:32.3350714Z # databricks-sdk 2025-11-03T17:08:32.3350772Z # datasets 2025-11-03T17:08:32.3350833Z # docker 2025-11-03T17:08:32.3350897Z # evaluate 2025-11-03T17:08:32.3350965Z # google-api-core 2025-11-03T17:08:32.3351041Z # google-cloud-storage 2025-11-03T17:08:32.3351101Z # gpt-oss 2025-11-03T17:08:32.3351166Z # huggingface-hub 2025-11-03T17:08:32.3351229Z # lightly 2025-11-03T17:08:32.3351286Z # lm-eval 2025-11-03T17:08:32.3351356Z # mistral-common 2025-11-03T17:08:32.3351417Z # mlflow-skinny 2025-11-03T17:08:32.3351475Z # mteb 2025-11-03T17:08:32.3351542Z # pooch 2025-11-03T17:08:32.3351609Z # ray 2025-11-03T17:08:32.3351670Z # responses 2025-11-03T17:08:32.3351736Z # schemathesis 2025-11-03T17:08:32.3351808Z # starlette-testclient 2025-11-03T17:08:32.3351871Z # tiktoken 2025-11-03T17:08:32.3351974Z # transformers 2025-11-03T17:08:32.3352038Z responses==0.25.3 2025-11-03T17:08:32.3352122Z  # via 2025-11-03T17:08:32.3352195Z # -c snapshot_constraint.txt 2025-11-03T17:08:32.3352294Z # genai-perf 2025-11-03T17:08:32.3352366Z rfc3339-validator==0.1.4 2025-11-03T17:08:32.3352446Z  # via 2025-11-03T17:08:32.3352524Z # -c snapshot_constraint.txt 2025-11-03T17:08:32.3352615Z # jsonschema 2025-11-03T17:08:32.3352676Z rfc3987==1.3.8 2025-11-03T17:08:32.3352762Z  # via 2025-11-03T17:08:32.3352834Z # -c snapshot_constraint.txt 2025-11-03T17:08:32.3352930Z # jsonschema 2025-11-03T17:08:32.3352993Z rich==13.9.4 2025-11-03T17:08:32.3353072Z  # via 2025-11-03T17:08:32.3353150Z # -c snapshot_constraint.txt 2025-11-03T17:08:32.3353212Z # genai-perf 2025-11-03T17:08:32.3353277Z # lightning 2025-11-03T17:08:32.3353334Z # mteb 2025-11-03T17:08:32.3353419Z # typer 2025-11-03T17:08:32.3353491Z rioxarray==0.19.0 2025-11-03T17:08:32.3353572Z  # via 2025-11-03T17:08:32.3353644Z # -c snapshot_constraint.txt 2025-11-03T17:08:32.3353740Z # terratorch 2025-11-03T17:08:32.3353893Z rouge-score==0.1.2 2025-11-03T17:08:32.3353979Z  # via 2025-11-03T17:08:32.3354051Z # -c snapshot_constraint.txt 2025-11-03T17:08:32.3354136Z # lm-eval 2025-11-03T17:08:32.3354204Z rpds-py==0.20.1 2025-11-03T17:08:32.3354282Z  # via 2025-11-03T17:08:32.3354360Z # -c snapshot_constraint.txt 2025-11-03T17:08:32.3354419Z # jsonschema 2025-11-03T17:08:32.3354509Z # referencing 2025-11-03T17:08:32.3354571Z rsa==4.9.1 2025-11-03T17:08:32.3354651Z  # via 2025-11-03T17:08:32.3354723Z # -c snapshot_constraint.txt 2025-11-03T17:08:32.3354819Z # google-auth 2025-11-03T17:08:32.3354890Z rtree==1.4.0 2025-11-03T17:08:32.3354974Z  # via 2025-11-03T17:08:32.3355047Z # -c snapshot_constraint.txt 2025-11-03T17:08:32.3355133Z # torchgeo 2025-11-03T17:08:32.3355215Z runai-model-streamer==0.14.0 2025-11-03T17:08:32.3355293Z  # via 2025-11-03T17:08:32.3355363Z # -c snapshot_constraint.txt 2025-11-03T17:08:32.3355476Z # -r requirements/test.in 2025-11-03T17:08:32.3355552Z runai-model-streamer-gcs==0.14.0 2025-11-03T17:08:32.3355637Z  # via 2025-11-03T17:08:32.3355707Z # -c snapshot_constraint.txt 2025-11-03T17:08:32.3355822Z # runai-model-streamer 2025-11-03T17:08:32.3355905Z runai-model-streamer-s3==0.14.0 2025-11-03T17:08:32.3355985Z  # via 2025-11-03T17:08:32.3356060Z # -c snapshot_constraint.txt 2025-11-03T17:08:32.3356163Z # runai-model-streamer 2025-11-03T17:08:32.3356225Z s3transfer==0.10.3 2025-11-03T17:08:32.3356310Z  # via 2025-11-03T17:08:32.3356382Z # -c snapshot_constraint.txt 2025-11-03T17:08:32.3356471Z # boto3 2025-11-03T17:08:32.3356533Z sacrebleu==2.4.3 2025-11-03T17:08:32.3356611Z  # via 2025-11-03T17:08:32.3356772Z # -c snapshot_constraint.txt 2025-11-03T17:08:32.3356864Z # lm-eval 2025-11-03T17:08:32.3356927Z safetensors==0.4.5 2025-11-03T17:08:32.3357013Z  # via 2025-11-03T17:08:32.3357087Z # -c snapshot_constraint.txt 2025-11-03T17:08:32.3357152Z # accelerate 2025-11-03T17:08:32.3357219Z # open-clip-torch 2025-11-03T17:08:32.3357279Z # peft 2025-11-03T17:08:32.3357343Z # timm 2025-11-03T17:08:32.3357437Z # transformers 2025-11-03T17:08:32.3357508Z schemathesis==3.39.15 2025-11-03T17:08:32.3357586Z  # via 2025-11-03T17:08:32.3357658Z # -c snapshot_constraint.txt 2025-11-03T17:08:32.3357768Z # -r requirements/test.in 2025-11-03T17:08:32.3357833Z scikit-image==0.25.2 2025-11-03T17:08:32.3357913Z  # via 2025-11-03T17:08:32.3357991Z # -c snapshot_constraint.txt 2025-11-03T17:08:32.3358089Z # albumentations 2025-11-03T17:08:32.3358158Z scikit-learn==1.5.2 2025-11-03T17:08:32.3358238Z  # via 2025-11-03T17:08:32.3358313Z # -c snapshot_constraint.txt 2025-11-03T17:08:32.3358381Z # albumentations 2025-11-03T17:08:32.3358439Z # librosa 2025-11-03T17:08:32.3358506Z # lm-eval 2025-11-03T17:08:32.3358562Z # mlflow 2025-11-03T17:08:32.3358621Z # mteb 2025-11-03T17:08:32.3358733Z # sentence-transformers 2025-11-03T17:08:32.3358793Z scipy==1.13.1 2025-11-03T17:08:32.3358872Z  # via 2025-11-03T17:08:32.3358959Z # -c snapshot_constraint.txt 2025-11-03T17:08:32.3359025Z # albumentations 2025-11-03T17:08:32.3359086Z # bm25s 2025-11-03T17:08:32.3359143Z # librosa 2025-11-03T17:08:32.3359202Z # mlflow 2025-11-03T17:08:32.3359268Z # mteb 2025-11-03T17:08:32.3359333Z # scikit-image 2025-11-03T17:08:32.3359396Z # scikit-learn 2025-11-03T17:08:32.3359475Z # sentence-transformers 2025-11-03T17:08:32.3359537Z # statsmodels 2025-11-03T17:08:32.3359629Z # vocos 2025-11-03T17:08:32.3359715Z segmentation-models-pytorch==0.4.0 2025-11-03T17:08:32.3359794Z  # via 2025-11-03T17:08:32.3359860Z # terratorch 2025-11-03T17:08:32.3359948Z # torchgeo 2025-11-03T17:08:32.3360121Z sentence-transformers==3.2.1 2025-11-03T17:08:32.3360203Z  # via 2025-11-03T17:08:32.3360275Z # -c snapshot_constraint.txt 2025-11-03T17:08:32.3360352Z # -r requirements/test.in 2025-11-03T17:08:32.3360434Z # mteb 2025-11-03T17:08:32.3360499Z setuptools==77.0.3 2025-11-03T17:08:32.3360583Z  # via 2025-11-03T17:08:32.3360655Z # -c snapshot_constraint.txt 2025-11-03T17:08:32.3360731Z # lightning-utilities 2025-11-03T17:08:32.3360796Z # pytablewriter 2025-11-03T17:08:32.3360884Z # torch 2025-11-03T17:08:32.3360954Z shapely==2.1.1 2025-11-03T17:08:32.3361033Z  # via 2025-11-03T17:08:32.3361112Z # -c snapshot_constraint.txt 2025-11-03T17:08:32.3361171Z # geopandas 2025-11-03T17:08:32.3361258Z # torchgeo 2025-11-03T17:08:32.3361333Z shellingham==1.5.4 2025-11-03T17:08:32.3361412Z  # via 2025-11-03T17:08:32.3361483Z # -c snapshot_constraint.txt 2025-11-03T17:08:32.3361572Z # typer 2025-11-03T17:08:32.3361634Z six==1.16.0 2025-11-03T17:08:32.3361719Z  # via 2025-11-03T17:08:32.3361792Z # -c snapshot_constraint.txt 2025-11-03T17:08:32.3361852Z # junit-xml 2025-11-03T17:08:32.3361917Z # lightly 2025-11-03T17:08:32.3361976Z # opencensus 2025-11-03T17:08:32.3362046Z # python-dateutil 2025-11-03T17:08:32.3362121Z # rfc3339-validator 2025-11-03T17:08:32.3362183Z # rouge-score 2025-11-03T17:08:32.3362314Z # segmentation-models-pytorch 2025-11-03T17:08:32.3362376Z smart-open==7.1.0 2025-11-03T17:08:32.3362456Z  # via 2025-11-03T17:08:32.3362530Z # -c snapshot_constraint.txt 2025-11-03T17:08:32.3362611Z # ray 2025-11-03T17:08:32.3362675Z smmap==5.0.2 2025-11-03T17:08:32.3362755Z  # via 2025-11-03T17:08:32.3362919Z # -c snapshot_constraint.txt 2025-11-03T17:08:32.3363011Z # gitdb 2025-11-03T17:08:32.3363083Z sniffio==1.3.1 2025-11-03T17:08:32.3363169Z  # via 2025-11-03T17:08:32.3363246Z # -c snapshot_constraint.txt 2025-11-03T17:08:32.3363303Z # anyio 2025-11-03T17:08:32.3363394Z # httpx 2025-11-03T17:08:32.3363463Z sortedcontainers==2.4.0 2025-11-03T17:08:32.3363542Z  # via 2025-11-03T17:08:32.3363666Z # -c snapshot_constraint.txt 2025-11-03T17:08:32.3363837Z # hypothesis 2025-11-03T17:08:32.3363912Z soundfile==0.12.1 2025-11-03T17:08:32.3363996Z  # via 2025-11-03T17:08:32.3364071Z # -c snapshot_constraint.txt 2025-11-03T17:08:32.3364149Z # -r requirements/test.in 2025-11-03T17:08:32.3364207Z # librosa 2025-11-03T17:08:32.3364310Z # mistral-common 2025-11-03T17:08:32.3364373Z soxr==0.5.0.post1 2025-11-03T17:08:32.3364452Z  # via 2025-11-03T17:08:32.3364530Z # -c snapshot_constraint.txt 2025-11-03T17:08:32.3364591Z # librosa 2025-11-03T17:08:32.3364689Z # mistral-common 2025-11-03T17:08:32.3364769Z sqlalchemy==2.0.41 2025-11-03T17:08:32.3364851Z  # via 2025-11-03T17:08:32.3364934Z # -c snapshot_constraint.txt 2025-11-03T17:08:32.3364992Z # alembic 2025-11-03T17:08:32.3365078Z # mlflow 2025-11-03T17:08:32.3365147Z sqlitedict==2.1.0 2025-11-03T17:08:32.3365226Z  # via 2025-11-03T17:08:32.3365302Z # -c snapshot_constraint.txt 2025-11-03T17:08:32.3365388Z # lm-eval 2025-11-03T17:08:32.3365451Z sqlparse==0.5.3 2025-11-03T17:08:32.3365534Z  # via 2025-11-03T17:08:32.3365605Z # -c snapshot_constraint.txt 2025-11-03T17:08:32.3365700Z # mlflow-skinny 2025-11-03T17:08:32.3365769Z starlette==0.46.2 2025-11-03T17:08:32.3365847Z  # via 2025-11-03T17:08:32.3365922Z # -c snapshot_constraint.txt 2025-11-03T17:08:32.3365980Z # fastapi 2025-11-03T17:08:32.3366042Z # schemathesis 2025-11-03T17:08:32.3366158Z # starlette-testclient 2025-11-03T17:08:32.3366232Z starlette-testclient==0.4.1 2025-11-03T17:08:32.3366318Z  # via 2025-11-03T17:08:32.3366389Z # -c snapshot_constraint.txt 2025-11-03T17:08:32.3366585Z # schemathesis 2025-11-03T17:08:32.3366655Z statsmodels==0.14.4 2025-11-03T17:08:32.3366736Z  # via 2025-11-03T17:08:32.3366814Z # -c snapshot_constraint.txt 2025-11-03T17:08:32.3366910Z # genai-perf 2025-11-03T17:08:32.3366973Z structlog==25.4.0 2025-11-03T17:08:32.3367059Z  # via 2025-11-03T17:08:32.3367132Z # -c snapshot_constraint.txt 2025-11-03T17:08:32.3367219Z # gpt-oss 2025-11-03T17:08:32.3367287Z sympy==1.13.3 2025-11-03T17:08:32.3367366Z  # via 2025-11-03T17:08:32.3367447Z # -c snapshot_constraint.txt 2025-11-03T17:08:32.3367505Z # einx 2025-11-03T17:08:32.3367589Z # torch 2025-11-03T17:08:32.3367657Z tabledata==1.3.3 2025-11-03T17:08:32.3367739Z  # via 2025-11-03T17:08:32.3367814Z # -c snapshot_constraint.txt 2025-11-03T17:08:32.3367919Z # pytablewriter 2025-11-03T17:08:32.3367982Z tabulate==0.9.0 2025-11-03T17:08:32.3368079Z  # via 2025-11-03T17:08:32.3368158Z # -c snapshot_constraint.txt 2025-11-03T17:08:32.3368248Z # sacrebleu 2025-11-03T17:08:32.3368313Z tblib==3.1.0 2025-11-03T17:08:32.3368393Z  # via 2025-11-03T17:08:32.3368464Z # -c snapshot_constraint.txt 2025-11-03T17:08:32.3368576Z # -r requirements/test.in 2025-11-03T17:08:32.3368639Z tcolorpy==0.1.6 2025-11-03T17:08:32.3368724Z  # via 2025-11-03T17:08:32.3368797Z # -c snapshot_constraint.txt 2025-11-03T17:08:32.3368893Z # pytablewriter 2025-11-03T17:08:32.3368960Z tenacity==9.1.2 2025-11-03T17:08:32.3369078Z  # via 2025-11-03T17:08:32.3369207Z # -c snapshot_constraint.txt 2025-11-03T17:08:32.3369269Z # gpt-oss 2025-11-03T17:08:32.3369326Z # lm-eval 2025-11-03T17:08:32.3369423Z # plotly 2025-11-03T17:08:32.3369580Z tensorboardx==2.6.4 2025-11-03T17:08:32.3369665Z  # via 2025-11-03T17:08:32.3369744Z # -c snapshot_constraint.txt 2025-11-03T17:08:32.3369835Z # lightning 2025-11-03T17:08:32.3369908Z tensorizer==2.10.1 2025-11-03T17:08:32.3369986Z  # via 2025-11-03T17:08:32.3370057Z # -c snapshot_constraint.txt 2025-11-03T17:08:32.3370170Z # -r requirements/test.in 2025-11-03T17:08:32.3370236Z termcolor==3.1.0 2025-11-03T17:08:32.3370319Z  # via 2025-11-03T17:08:32.3370389Z # -c snapshot_constraint.txt 2025-11-03T17:08:32.3370475Z # gpt-oss 2025-11-03T17:08:32.3370730Z terratorch @ git+https://github.com/IBM/terratorch.git@07184fcf91a1324f831ff521dd238d97fe350e3e 2025-11-03T17:08:32.3370852Z  # via -r requirements/test.in 2025-11-03T17:08:32.3370923Z threadpoolctl==3.5.0 2025-11-03T17:08:32.3371001Z  # via 2025-11-03T17:08:32.3371085Z # -c snapshot_constraint.txt 2025-11-03T17:08:32.3371187Z # scikit-learn 2025-11-03T17:08:32.3371252Z tifffile==2025.3.30 2025-11-03T17:08:32.3371332Z  # via 2025-11-03T17:08:32.3371410Z # -c snapshot_constraint.txt 2025-11-03T17:08:32.3371474Z # scikit-image 2025-11-03T17:08:32.3371571Z # terratorch 2025-11-03T17:08:32.3371634Z tiktoken==0.12.0 2025-11-03T17:08:32.3371713Z  # via 2025-11-03T17:08:32.3371791Z # -c snapshot_constraint.txt 2025-11-03T17:08:32.3371850Z # gpt-oss 2025-11-03T17:08:32.3371914Z # lm-eval 2025-11-03T17:08:32.3372009Z # mistral-common 2025-11-03T17:08:32.3372069Z timm==1.0.17 2025-11-03T17:08:32.3372152Z  # via 2025-11-03T17:08:32.3372223Z # -c snapshot_constraint.txt 2025-11-03T17:08:32.3372293Z # -r requirements/test.in 2025-11-03T17:08:32.3372365Z # open-clip-torch 2025-11-03T17:08:32.3372446Z # segmentation-models-pytorch 2025-11-03T17:08:32.3372512Z # terratorch 2025-11-03T17:08:32.3372600Z # torchgeo 2025-11-03T17:08:32.3372666Z tokenizers==0.22.0 2025-11-03T17:08:32.3372751Z  # via 2025-11-03T17:08:32.3372822Z # -c snapshot_constraint.txt 2025-11-03T17:08:32.3372898Z # -r requirements/test.in 2025-11-03T17:08:32.3373125Z # transformers 2025-11-03T17:08:32.3373186Z tomli==2.2.1 2025-11-03T17:08:32.3373271Z  # via 2025-11-03T17:08:32.3373343Z # -c snapshot_constraint.txt 2025-11-03T17:08:32.3373437Z # schemathesis 2025-11-03T17:08:32.3373505Z tomli-w==1.2.0 2025-11-03T17:08:32.3373584Z  # via 2025-11-03T17:08:32.3373661Z # -c snapshot_constraint.txt 2025-11-03T17:08:32.3373753Z # schemathesis 2025-11-03T17:08:32.3374229Z torch @ file:///var/lib/jenkins/workspace/dist/torch-2.10.0a0%2Bgit3f6538f-cp312-cp312-linux_x86_64.whl#sha256=288b4779bb4efa62cf5264c0eca9fdb02025e7a7fb44d42e3f512a68d8e60da0 2025-11-03T17:08:32.3374316Z  # via 2025-11-03T17:08:32.3374385Z # -r requirements/test.in 2025-11-03T17:08:32.3374451Z # accelerate 2025-11-03T17:08:32.3374515Z # bitsandbytes 2025-11-03T17:08:32.3374589Z # efficientnet-pytorch 2025-11-03T17:08:32.3374662Z # encodec 2025-11-03T17:08:32.3374734Z # fastsafetensors 2025-11-03T17:08:32.3374800Z # kornia 2025-11-03T17:08:32.3374867Z # lightly 2025-11-03T17:08:32.3374928Z # lightning 2025-11-03T17:08:32.3374993Z # lm-eval 2025-11-03T17:08:32.3375051Z # mteb 2025-11-03T17:08:32.3375119Z # open-clip-torch 2025-11-03T17:08:32.3375183Z # peft 2025-11-03T17:08:32.3375250Z # pretrainedmodels 2025-11-03T17:08:32.3375325Z # pytorch-lightning 2025-11-03T17:08:32.3375396Z # runai-model-streamer 2025-11-03T17:08:32.3375476Z # segmentation-models-pytorch 2025-11-03T17:08:32.3375556Z # sentence-transformers 2025-11-03T17:08:32.3375617Z # tensorizer 2025-11-03T17:08:32.3375675Z # terratorch 2025-11-03T17:08:32.3375737Z # timm 2025-11-03T17:08:32.3375797Z # torchaudio 2025-11-03T17:08:32.3375944Z # torchgeo 2025-11-03T17:08:32.3376014Z # torchmetrics 2025-11-03T17:08:32.3376075Z # torchvision 2025-11-03T17:08:32.3376156Z # vector-quantize-pytorch 2025-11-03T17:08:32.3376247Z # vocos 2025-11-03T17:08:32.3376764Z torchaudio @ file:///var/lib/jenkins/workspace/dist/audio/torchaudio-2.10.0a0%2B3b0e7a6-cp312-cp312-linux_x86_64.whl#sha256=70d7f19c17cbc4bb012226c17e5f5380b3bd6bda732a45f329984580e0bc17a6 2025-11-03T17:08:32.3376844Z  # via 2025-11-03T17:08:32.3376915Z # -r requirements/test.in 2025-11-03T17:08:32.3376979Z # encodec 2025-11-03T17:08:32.3377074Z # vocos 2025-11-03T17:08:32.3377143Z torchgeo==0.7.0 2025-11-03T17:08:32.3377243Z  # via terratorch 2025-11-03T17:08:32.3377307Z torchmetrics==1.8.2 2025-11-03T17:08:32.3377390Z  # via 2025-11-03T17:08:32.3377449Z # lightning 2025-11-03T17:08:32.3377518Z # pytorch-lightning 2025-11-03T17:08:32.3377585Z # terratorch 2025-11-03T17:08:32.3377675Z # torchgeo 2025-11-03T17:08:32.3378216Z torchvision @ file:///var/lib/jenkins/workspace/dist/vision/torchvision-0.25.0a0%2Bcfbc5c2-cp312-cp312-linux_x86_64.whl#sha256=7fa43396b4b9ce1b239050abea02e7fbabadf1cbdf180883becb1f6530983b15 2025-11-03T17:08:32.3378300Z  # via 2025-11-03T17:08:32.3378369Z # -r requirements/test.in 2025-11-03T17:08:32.3378431Z # lightly 2025-11-03T17:08:32.3378496Z # open-clip-torch 2025-11-03T17:08:32.3378567Z # pretrainedmodels 2025-11-03T17:08:32.3378644Z # segmentation-models-pytorch 2025-11-03T17:08:32.3378704Z # terratorch 2025-11-03T17:08:32.3378765Z # timm 2025-11-03T17:08:32.3378852Z # torchgeo 2025-11-03T17:08:32.3378916Z tqdm==4.66.6 2025-11-03T17:08:32.3378994Z  # via 2025-11-03T17:08:32.3379066Z # -c snapshot_constraint.txt 2025-11-03T17:08:32.3379129Z # datasets 2025-11-03T17:08:32.3379187Z # evaluate 2025-11-03T17:08:32.3379253Z # huggingface-hub 2025-11-03T17:08:32.3379320Z # lightly 2025-11-03T17:08:32.3379378Z # lightning 2025-11-03T17:08:32.3379440Z # lm-eval 2025-11-03T17:08:32.3379496Z # mteb 2025-11-03T17:08:32.3379639Z # nltk 2025-11-03T17:08:32.3379711Z # open-clip-torch 2025-11-03T17:08:32.3379768Z # peft 2025-11-03T17:08:32.3379825Z # pqdm 2025-11-03T17:08:32.3379900Z # pretrainedmodels 2025-11-03T17:08:32.3379966Z # pytorch-lightning 2025-11-03T17:08:32.3380050Z # segmentation-models-pytorch 2025-11-03T17:08:32.3380121Z # sentence-transformers 2025-11-03T17:08:32.3380189Z # tqdm-multiprocess 2025-11-03T17:08:32.3380292Z # transformers 2025-11-03T17:08:32.3380363Z tqdm-multiprocess==0.0.11 2025-11-03T17:08:32.3380448Z  # via 2025-11-03T17:08:32.3380522Z # -c snapshot_constraint.txt 2025-11-03T17:08:32.3380618Z # lm-eval 2025-11-03T17:08:32.3380691Z transformers==4.56.2 2025-11-03T17:08:32.3380770Z  # via 2025-11-03T17:08:32.3380847Z # -c snapshot_constraint.txt 2025-11-03T17:08:32.3380922Z # -r requirements/test.in 2025-11-03T17:08:32.3380984Z # genai-perf 2025-11-03T17:08:32.3381047Z # lm-eval 2025-11-03T17:08:32.3381111Z # peft 2025-11-03T17:08:32.3381184Z # sentence-transformers 2025-11-03T17:08:32.3381318Z # transformers-stream-generator 2025-11-03T17:08:32.3381402Z transformers-stream-generator==0.0.5 2025-11-03T17:08:32.3381489Z  # via 2025-11-03T17:08:32.3381652Z # -c snapshot_constraint.txt 2025-11-03T17:08:32.3381764Z # -r requirements/test.in 2025-11-03T17:08:32.3381835Z tritonclient==2.51.0 2025-11-03T17:08:32.3381917Z  # via 2025-11-03T17:08:32.3381995Z # -c snapshot_constraint.txt 2025-11-03T17:08:32.3382067Z # -r requirements/test.in 2025-11-03T17:08:32.3382160Z # genai-perf 2025-11-03T17:08:32.3382227Z typepy==1.3.2 2025-11-03T17:08:32.3382306Z  # via 2025-11-03T17:08:32.3382378Z # -c snapshot_constraint.txt 2025-11-03T17:08:32.3382530Z # dataproperty 2025-11-03T17:08:32.3382732Z # pytablewriter 2025-11-03T17:08:32.3382834Z # tabledata 2025-11-03T17:08:32.3382895Z typer==0.15.2 2025-11-03T17:08:32.3382981Z  # via 2025-11-03T17:08:32.3383064Z # -c snapshot_constraint.txt 2025-11-03T17:08:32.3383164Z # fastsafetensors 2025-11-03T17:08:32.3383256Z types-python-dateutil==2.9.0.20241206 2025-11-03T17:08:32.3383337Z  # via 2025-11-03T17:08:32.3383412Z # -c snapshot_constraint.txt 2025-11-03T17:08:32.3383500Z # arrow 2025-11-03T17:08:32.3383570Z typeshed-client==2.8.2 2025-11-03T17:08:32.3383648Z  # via 2025-11-03T17:08:32.3383729Z # -c snapshot_constraint.txt 2025-11-03T17:08:32.3383823Z # jsonargparse 2025-11-03T17:08:32.3383900Z typing-extensions==4.15.0 2025-11-03T17:08:32.3383980Z  # via 2025-11-03T17:08:32.3384053Z # -c snapshot_constraint.txt 2025-11-03T17:08:32.3384120Z # aiosignal 2025-11-03T17:08:32.3384187Z # albumentations 2025-11-03T17:08:32.3384252Z # alembic 2025-11-03T17:08:32.3384309Z # chz 2025-11-03T17:08:32.3384367Z # fastapi 2025-11-03T17:08:32.3384455Z # graphene 2025-11-03T17:08:32.3384572Z # huggingface-hub 2025-11-03T17:08:32.3384659Z # librosa 2025-11-03T17:08:32.3384726Z # lightning 2025-11-03T17:08:32.3384800Z # lightning-utilities 2025-11-03T17:08:32.3384867Z # mistral-common 2025-11-03T17:08:32.3384930Z # mlflow-skinny 2025-11-03T17:08:32.3384987Z # mteb 2025-11-03T17:08:32.3385062Z # opentelemetry-api 2025-11-03T17:08:32.3385131Z # opentelemetry-sdk 2025-11-03T17:08:32.3385234Z # opentelemetry-semantic-conventions 2025-11-03T17:08:32.3385292Z # pqdm 2025-11-03T17:08:32.3385351Z # pydantic 2025-11-03T17:08:32.3385418Z # pydantic-core 2025-11-03T17:08:32.3385490Z # pydantic-extra-types 2025-11-03T17:08:32.3385567Z # pytorch-lightning 2025-11-03T17:08:32.3385634Z # sqlalchemy 2025-11-03T17:08:32.3385696Z # torch 2025-11-03T17:08:32.3385759Z # torchgeo 2025-11-03T17:08:32.3385816Z # typer 2025-11-03T17:08:32.3385880Z # typeshed-client 2025-11-03T17:08:32.3386122Z # typing-inspection 2025-11-03T17:08:32.3386193Z typing-inspection==0.4.2 2025-11-03T17:08:32.3386270Z  # via 2025-11-03T17:08:32.3386347Z # -c snapshot_constraint.txt 2025-11-03T17:08:32.3386437Z # pydantic 2025-11-03T17:08:32.3386501Z tzdata==2024.2 2025-11-03T17:08:32.3386580Z  # via 2025-11-03T17:08:32.3386653Z # -c snapshot_constraint.txt 2025-11-03T17:08:32.3386744Z # pandas 2025-11-03T17:08:32.3386809Z uri-template==1.3.0 2025-11-03T17:08:32.3386892Z  # via 2025-11-03T17:08:32.3386964Z # -c snapshot_constraint.txt 2025-11-03T17:08:32.3387054Z # jsonschema 2025-11-03T17:08:32.3387119Z urllib3==2.2.3 2025-11-03T17:08:32.3387197Z  # via 2025-11-03T17:08:32.3387268Z # -c snapshot_constraint.txt 2025-11-03T17:08:32.3387334Z # blobfile 2025-11-03T17:08:32.3387393Z # botocore 2025-11-03T17:08:32.3387454Z # docker 2025-11-03T17:08:32.3387512Z # lightly 2025-11-03T17:08:32.3387573Z # requests 2025-11-03T17:08:32.3387639Z # responses 2025-11-03T17:08:32.3387732Z # tritonclient 2025-11-03T17:08:32.3387803Z uvicorn==0.35.0 2025-11-03T17:08:32.3387880Z  # via 2025-11-03T17:08:32.3387952Z # -c snapshot_constraint.txt 2025-11-03T17:08:32.3388017Z # gpt-oss 2025-11-03T17:08:32.3388112Z # mlflow-skinny 2025-11-03T17:08:32.3388191Z vector-quantize-pytorch==1.24.2 2025-11-03T17:08:32.3388319Z  # via -r requirements/test.in 2025-11-03T17:08:32.3388385Z virtualenv==20.31.2 2025-11-03T17:08:32.3388471Z  # via 2025-11-03T17:08:32.3388543Z # -c snapshot_constraint.txt 2025-11-03T17:08:32.3388624Z # ray 2025-11-03T17:08:32.3388699Z vocos==0.1.0 2025-11-03T17:08:32.3388780Z  # via 2025-11-03T17:08:32.3388935Z # -c snapshot_constraint.txt 2025-11-03T17:08:32.3389045Z # -r requirements/test.in 2025-11-03T17:08:32.3389107Z wcwidth==0.2.13 2025-11-03T17:08:32.3389195Z  # via 2025-11-03T17:08:32.3389268Z # -c snapshot_constraint.txt 2025-11-03T17:08:32.3389349Z # ftfy 2025-11-03T17:08:32.3389418Z webcolors==24.11.1 2025-11-03T17:08:32.3389496Z  # via 2025-11-03T17:08:32.3389574Z # -c snapshot_constraint.txt 2025-11-03T17:08:32.3389674Z # jsonschema 2025-11-03T17:08:32.3389738Z werkzeug==3.1.3 2025-11-03T17:08:32.3389825Z  # via 2025-11-03T17:08:32.3389897Z # -c snapshot_constraint.txt 2025-11-03T17:08:32.3389959Z # flask 2025-11-03T17:08:32.3390052Z # schemathesis 2025-11-03T17:08:32.3390115Z word2number==1.1 2025-11-03T17:08:32.3390208Z  # via 2025-11-03T17:08:32.3390281Z # -c snapshot_constraint.txt 2025-11-03T17:08:32.3390368Z # lm-eval 2025-11-03T17:08:32.3390434Z wrapt==1.17.2 2025-11-03T17:08:32.3390519Z  # via 2025-11-03T17:08:32.3390600Z # -c snapshot_constraint.txt 2025-11-03T17:08:32.3390693Z # smart-open 2025-11-03T17:08:32.3390757Z xarray==2025.7.1 2025-11-03T17:08:32.3390844Z  # via 2025-11-03T17:08:32.3390927Z # -c snapshot_constraint.txt 2025-11-03T17:08:32.3391026Z # rioxarray 2025-11-03T17:08:32.3391088Z xxhash==3.5.0 2025-11-03T17:08:32.3391167Z  # via 2025-11-03T17:08:32.3391246Z # -c snapshot_constraint.txt 2025-11-03T17:08:32.3391304Z # datasets 2025-11-03T17:08:32.3391393Z # evaluate 2025-11-03T17:08:32.3391463Z yarl==1.17.1 2025-11-03T17:08:32.3391542Z  # via 2025-11-03T17:08:32.3391620Z # -c snapshot_constraint.txt 2025-11-03T17:08:32.3391682Z # aiohttp 2025-11-03T17:08:32.3391776Z # schemathesis 2025-11-03T17:08:32.3391848Z zipp==3.23.0 2025-11-03T17:08:32.3391928Z  # via 2025-11-03T17:08:32.3392002Z # -c snapshot_constraint.txt 2025-11-03T17:08:32.3399096Z # importlib-metadata 2025-11-03T17:08:32.3399208Z zstandard==0.23.0 2025-11-03T17:08:32.3399335Z  # via 2025-11-03T17:08:32.3399425Z # -c snapshot_constraint.txt 2025-11-03T17:08:32.3399653Z # lm-eval 2025-11-03T17:08:32.4506946Z 2025-11-03 17:08:32,450 [INFO] cli.lib.common.pip_helper: pip installing packages: /opt/conda/envs/py_3.12/bin/python -m uv pip install -r test.txt 2025-11-03T17:08:32.4508248Z 2025-11-03 17:08:32,450 [INFO] cli.lib.common.utils: [cmd] /opt/conda/envs/py_3.12/bin/python -m uv pip install -r test.txt 2025-11-03T17:08:32.4996156Z Using Python 3.12.12 environment at: /opt/conda/envs/py_3.12 2025-11-03T17:08:32.6517544Z Resolved 329 packages in 145ms 2025-11-03T17:08:32.6766720Z Updating https://github.com/EleutherAI/lm-evaluation-harness.git (206b7722158f58c35b7ffcd53b035fdbdda5126d) 2025-11-03T17:08:32.7554760Z Building terratorch @ git+https://github.com/IBM/terratorch.git@07184fcf91a1324f831ff521dd238d97fe350e3e 2025-11-03T17:08:32.7966832Z Downloading networkx (1.6MiB) 2025-11-03T17:08:32.8071794Z Downloading zstandard (5.2MiB) 2025-11-03T17:08:32.8296881Z Downloading decord (13.0MiB) 2025-11-03T17:08:32.8297796Z Downloading virtualenv (5.8MiB) 2025-11-03T17:08:32.8300138Z Downloading sqlalchemy (3.2MiB) 2025-11-03T17:08:32.8302872Z Downloading numpy (17.1MiB) 2025-11-03T17:08:32.8305517Z Downloading opencv-python-headless (47.7MiB) 2025-11-03T17:08:32.8307897Z Downloading runai-model-streamer-gcs (4.2MiB) 2025-11-03T17:08:32.8310283Z Downloading scikit-learn (12.3MiB) 2025-11-03T17:08:32.8312405Z Downloading pandas (12.1MiB) 2025-11-03T17:08:32.8314936Z Downloading plotly (18.2MiB) 2025-11-03T17:08:32.8317132Z Downloading pillow (4.3MiB) 2025-11-03T17:08:32.8387129Z Downloading scikit-image (14.3MiB) 2025-11-03T17:08:32.8390070Z Downloading kaleido (76.2MiB) 2025-11-03T17:08:32.8392740Z Downloading tritonclient (13.3MiB) 2025-11-03T17:08:32.8395440Z Downloading botocore (12.1MiB) 2025-11-03T17:08:32.8398082Z Downloading grpcio (5.6MiB) 2025-11-03T17:08:32.8399976Z Downloading hf-xet (3.0MiB) 2025-11-03T17:08:32.8402343Z Downloading h5py (4.7MiB) 2025-11-03T17:08:32.8406344Z Downloading runai-model-streamer-s3 (5.6MiB) 2025-11-03T17:08:32.8407059Z Downloading kornia-rs (2.6MiB) 2025-11-03T17:08:32.8409530Z Downloading py-spy (2.6MiB) 2025-11-03T17:08:32.8411626Z Downloading matplotlib (7.9MiB) 2025-11-03T17:08:32.8413472Z Downloading mlflow-skinny (6.0MiB) 2025-11-03T17:08:32.8415803Z Downloading fiona (16.4MiB) 2025-11-03T17:08:32.8418016Z Downloading statsmodels (10.2MiB) 2025-11-03T17:08:32.8420370Z Downloading fastparquet (1.7MiB) 2025-11-03T17:08:32.8424531Z Downloading black (1.7MiB) 2025-11-03T17:08:32.8426043Z Downloading tokenizers (3.2MiB) 2025-11-03T17:08:32.8428715Z Downloading python-box (4.1MiB) 2025-11-03T17:08:32.8430786Z Downloading pydantic-core (2.0MiB) 2025-11-03T17:08:32.8433506Z Downloading python-rapidjson (1.6MiB) 2025-11-03T17:08:32.8435393Z Downloading transformers (11.1MiB) 2025-11-03T17:08:32.8437887Z Downloading cramjam (2.2MiB) 2025-11-03T17:08:32.8439911Z Downloading mteb (1.5MiB) 2025-11-03T17:08:32.8443775Z Downloading pyproj (9.1MiB) 2025-11-03T17:08:32.8444939Z Downloading fonttools (4.7MiB) 2025-11-03T17:08:32.8447571Z Downloading ray (66.9MiB) 2025-11-03T17:08:32.8449940Z Downloading pyarrow (38.2MiB) 2025-11-03T17:08:32.8453080Z Downloading pycryptodomex (2.2MiB) 2025-11-03T17:08:32.8455481Z Downloading scipy (36.4MiB) 2025-11-03T17:08:32.8459888Z Downloading polars (33.2MiB) 2025-11-03T17:08:32.8461791Z Downloading rapidfuzz (3.0MiB) 2025-11-03T17:08:32.8464570Z Downloading pyogrio (26.4MiB) 2025-11-03T17:08:32.8467396Z Downloading timm (2.4MiB) 2025-11-03T17:08:32.8471156Z Downloading bitsandbytes (69.5MiB) 2025-11-03T17:08:32.8473976Z Downloading rasterio (21.2MiB) 2025-11-03T17:08:32.8476122Z Downloading aiohttp (1.7MiB) 2025-11-03T17:08:32.8478578Z Downloading mlflow (27.7MiB) 2025-11-03T17:08:32.8492711Z Downloading shapely (3.0MiB) 2025-11-03T17:08:33.6384570Z Downloading python-rapidjson 2025-11-03T17:08:33.6575925Z Downloading open-clip-torch (1.5MiB) 2025-11-03T17:08:33.7126009Z Downloading fastparquet 2025-11-03T17:08:33.7238849Z Downloading nltk (1.4MiB) 2025-11-03T17:08:33.7399780Z Downloading aiohttp 2025-11-03T17:08:33.7575765Z Downloading kiwisolver (1.4MiB) 2025-11-03T17:08:33.8176822Z Downloading black 2025-11-03T17:08:33.8312663Z Downloading pydantic-core 2025-11-03T17:08:33.8394243Z Downloading fastsafetensors (1.4MiB) 2025-11-03T17:08:33.8478090Z Downloading xarray (1.3MiB) 2025-11-03T17:08:33.9466659Z Downloading cramjam 2025-11-03T17:08:33.9618047Z Downloading setuptools (1.2MiB) 2025-11-03T17:08:33.9637845Z Downloading pycryptodomex 2025-11-03T17:08:33.9778074Z Downloading pygments (1.1MiB) 2025-11-03T17:08:34.0599272Z Downloading kornia-rs 2025-11-03T17:08:34.0721927Z Downloading soundfile (1.1MiB) 2025-11-03T17:08:34.0913331Z Built terratorch @ git+https://github.com/IBM/terratorch.git@07184fcf91a1324f831ff521dd238d97fe350e3e 2025-11-03T17:08:34.1603128Z Downloading timm 2025-11-03T17:08:34.1726402Z Downloading kornia (1.0MiB) 2025-11-03T17:08:34.1730825Z Downloading py-spy 2025-11-03T17:08:34.2367195Z Downloading networkx 2025-11-03T17:08:34.3135905Z Downloading hf-xet 2025-11-03T17:08:34.3247506Z Downloading rapidfuzz 2025-11-03T17:08:34.3990306Z Downloading kiwisolver 2025-11-03T17:08:34.3991474Z Downloading tokenizers 2025-11-03T17:08:34.4213378Z Downloading sqlalchemy 2025-11-03T17:08:34.4250221Z Downloading shapely 2025-11-03T17:08:34.4458725Z Downloading xarray 2025-11-03T17:08:34.5084679Z Downloading fastsafetensors 2025-11-03T17:08:34.5543041Z Downloading open-clip-torch 2025-11-03T17:08:34.5965153Z Downloading soundfile 2025-11-03T17:08:34.6288968Z Downloading nltk 2025-11-03T17:08:34.6452673Z Downloading pygments 2025-11-03T17:08:34.7562616Z Downloading python-box 2025-11-03T17:08:34.8403035Z Downloading runai-model-streamer-gcs 2025-11-03T17:08:34.9378655Z Downloading setuptools 2025-11-03T17:08:34.9671545Z Downloading pillow 2025-11-03T17:08:35.0168430Z Downloading mteb 2025-11-03T17:08:35.0293989Z Downloading fonttools 2025-11-03T17:08:35.0803335Z Downloading h5py 2025-11-03T17:08:35.0901330Z Downloading kornia 2025-11-03T17:08:35.1995648Z Downloading zstandard 2025-11-03T17:08:35.5611912Z Downloading runai-model-streamer-s3 2025-11-03T17:08:35.6002584Z Downloading grpcio 2025-11-03T17:08:35.6568483Z Downloading virtualenv 2025-11-03T17:08:36.0044797Z Building encodec==0.1.1 2025-11-03T17:08:36.0073667Z Building antlr4-python3-runtime==4.9.3 2025-11-03T17:08:36.0122182Z Building sqlitedict==2.1.0 2025-11-03T17:08:36.0122726Z Building docopt==0.6.2 2025-11-03T17:08:36.0125733Z Building pretrainedmodels==0.7.4 2025-11-03T17:08:36.0161811Z Building efficientnet-pytorch==0.7.1 2025-11-03T17:08:36.0181118Z Building rouge-score==0.1.2 2025-11-03T17:08:36.0182378Z Building transformers-stream-generator==0.0.5 2025-11-03T17:08:36.0184555Z Building word2number==1.1 2025-11-03T17:08:36.2394652Z Downloading mlflow-skinny 2025-11-03T17:08:36.4590672Z Downloading matplotlib 2025-11-03T17:08:36.5944745Z Built word2number==1.1 2025-11-03T17:08:36.5951205Z Downloading pyproj 2025-11-03T17:08:36.6136023Z Built rouge-score==0.1.2 2025-11-03T17:08:36.6490294Z Built docopt==0.6.2 2025-11-03T17:08:36.6698476Z Built efficientnet-pytorch==0.7.1 2025-11-03T17:08:36.6818660Z Built sqlitedict==2.1.0 2025-11-03T17:08:36.7072238Z Built encodec==0.1.1 2025-11-03T17:08:36.7112281Z Built antlr4-python3-runtime==4.9.3 2025-11-03T17:08:36.7717270Z Built transformers-stream-generator==0.0.5 2025-11-03T17:08:36.7908864Z Built pretrainedmodels==0.7.4 2025-11-03T17:08:36.8356821Z Updated https://github.com/EleutherAI/lm-evaluation-harness.git (206b7722158f58c35b7ffcd53b035fdbdda5126d) 2025-11-03T17:08:36.8364973Z Building lm-eval @ git+https://github.com/EleutherAI/lm-evaluation-harness.git@206b7722158f58c35b7ffcd53b035fdbdda5126d 2025-11-03T17:08:36.9197482Z Downloading statsmodels 2025-11-03T17:08:37.1753122Z Downloading pandas 2025-11-03T17:08:37.1868819Z Downloading scikit-learn 2025-11-03T17:08:37.2665095Z Downloading decord 2025-11-03T17:08:37.3504900Z Downloading botocore 2025-11-03T17:08:37.3582152Z Downloading tritonclient 2025-11-03T17:08:37.4041099Z Downloading transformers 2025-11-03T17:08:37.4871064Z Downloading scikit-image 2025-11-03T17:08:37.6108330Z Downloading fiona 2025-11-03T17:08:37.6467318Z Downloading numpy 2025-11-03T17:08:38.0197282Z Downloading rasterio 2025-11-03T17:08:38.2279777Z Downloading pyogrio 2025-11-03T17:08:38.5516878Z Downloading polars 2025-11-03T17:08:38.6074949Z Downloading mlflow 2025-11-03T17:08:39.2755922Z Downloading opencv-python-headless 2025-11-03T17:08:39.4973018Z Downloading scipy 2025-11-03T17:08:39.5051942Z Downloading bitsandbytes 2025-11-03T17:08:39.5263907Z Downloading pyarrow 2025-11-03T17:08:39.7105948Z Downloading kaleido 2025-11-03T17:08:39.7983885Z Downloading ray 2025-11-03T17:08:41.3000387Z Downloading plotly 2025-11-03T17:08:51.3114532Z Built lm-eval @ git+https://github.com/EleutherAI/lm-evaluation-harness.git@206b7722158f58c35b7ffcd53b035fdbdda5126d 2025-11-03T17:08:51.5500950Z Prepared 290 packages in 18.88s 2025-11-03T17:08:54.3843578Z Uninstalled 94 packages in 2.83s 2025-11-03T17:08:55.4140406Z Installed 290 packages in 1.02s 2025-11-03T17:08:55.4144572Z - absl-py==2.3.1 2025-11-03T17:08:55.4145203Z + absl-py==2.1.0 2025-11-03T17:08:55.4145977Z + accelerate==1.0.1 2025-11-03T17:08:55.4146464Z + aenum==3.1.16 2025-11-03T17:08:55.4147084Z + affine==2.4.0 2025-11-03T17:08:55.4148232Z - aiohttp==3.13.2 2025-11-03T17:08:55.4148907Z + aiohttp==3.13.0 2025-11-03T17:08:55.4150056Z + aiohttp-cors==0.8.1 2025-11-03T17:08:55.4151197Z + albucore==0.0.16 2025-11-03T17:08:55.4152378Z + albumentations==1.4.6 2025-11-03T17:08:55.4153237Z + alembic==1.16.4 2025-11-03T17:08:55.4154653Z + antlr4-python3-runtime==4.9.3 2025-11-03T17:08:55.4155291Z - anyio==4.11.0 2025-11-03T17:08:55.4157272Z + anyio==4.6.2.post1 2025-11-03T17:08:55.4157755Z + argcomplete==3.5.1 2025-11-03T17:08:55.4159688Z + arrow==1.3.0 2025-11-03T17:08:55.4160671Z - attrs==25.4.0 2025-11-03T17:08:55.4161134Z + attrs==24.2.0 2025-11-03T17:08:55.4161790Z - audioread==3.1.0 2025-11-03T17:08:55.4162849Z + audioread==3.0.1 2025-11-03T17:08:55.4163746Z + backoff==2.2.1 2025-11-03T17:08:55.4164985Z + bitsandbytes==0.46.1 2025-11-03T17:08:55.4165468Z + black==24.10.0 2025-11-03T17:08:55.4166570Z + blinker==1.9.0 2025-11-03T17:08:55.4167466Z + blobfile==3.0.0 2025-11-03T17:08:55.4169336Z + bm25s==0.2.13 2025-11-03T17:08:55.4169797Z - boto3==1.35.42 2025-11-03T17:08:55.4170088Z + boto3==1.35.57 2025-11-03T17:08:55.4171092Z - botocore==1.35.99 2025-11-03T17:08:55.4171708Z + botocore==1.35.57 2025-11-03T17:08:55.4172674Z + bounded-pool-executor==0.0.3 2025-11-03T17:08:55.4173828Z + buildkite-test-collector==0.1.9 2025-11-03T17:08:55.4174508Z - cachetools==6.2.1 2025-11-03T17:08:55.4176022Z + cachetools==5.5.2 2025-11-03T17:08:55.4176662Z - certifi==2025.10.5 2025-11-03T17:08:55.4177724Z + certifi==2024.8.30 2025-11-03T17:08:55.4179576Z - cffi==2.0.0 2025-11-03T17:08:55.4180079Z + cffi==1.17.1 2025-11-03T17:08:55.4180667Z + chardet==5.2.0 2025-11-03T17:08:55.4181798Z - charset-normalizer==3.4.4 2025-11-03T17:08:55.4182407Z + charset-normalizer==3.4.0 2025-11-03T17:08:55.4182938Z + chz==0.3.0 2025-11-03T17:08:55.4184083Z - click==8.2.1 2025-11-03T17:08:55.4184680Z + click==8.1.7 2025-11-03T17:08:55.4185534Z + click-plugins==1.1.1.2 2025-11-03T17:08:55.4185955Z + cligj==0.7.2 2025-11-03T17:08:55.4186396Z - cloudpickle==3.1.2 2025-11-03T17:08:55.4187299Z + cloudpickle==3.1.1 2025-11-03T17:08:55.4188961Z + colorful==0.5.6 2025-11-03T17:08:55.4189730Z + contourpy==1.3.0 2025-11-03T17:08:55.4190130Z + coverage==7.10.6 2025-11-03T17:08:55.4190726Z + cramjam==2.9.0 2025-11-03T17:08:55.4191450Z + cycler==0.12.1 2025-11-03T17:08:55.4192375Z + databricks-sdk==0.59.0 2025-11-03T17:08:55.4192893Z + datamodel-code-generator==0.26.3 2025-11-03T17:08:55.4193717Z + dataproperty==1.0.1 2025-11-03T17:08:55.4194787Z + datasets==3.0.2 2025-11-03T17:08:55.4195920Z - decorator==5.2.1 2025-11-03T17:08:55.4196381Z + decorator==5.1.1 2025-11-03T17:08:55.4197299Z + decord==0.6.0 2025-11-03T17:08:55.4198639Z - dill==0.3.7 2025-11-03T17:08:55.4199059Z + dill==0.3.8 2025-11-03T17:08:55.4200070Z + distlib==0.3.9 2025-11-03T17:08:55.4200777Z - dnspython==2.8.0 2025-11-03T17:08:55.4202120Z + dnspython==2.7.0 2025-11-03T17:08:55.4202808Z + docopt==0.6.2 2025-11-03T17:08:55.4203804Z + docstring-parser==0.17.0 2025-11-03T17:08:55.4204858Z + efficientnet-pytorch==0.7.1 2025-11-03T17:08:55.4205446Z + einx==0.3.0 2025-11-03T17:08:55.4206686Z - email-validator==2.3.0 2025-11-03T17:08:55.4207836Z + email-validator==2.2.0 2025-11-03T17:08:55.4208712Z + encodec==0.1.1 2025-11-03T17:08:55.4209972Z + eval-type-backport==0.2.2 2025-11-03T17:08:55.4210754Z + evaluate==0.4.3 2025-11-03T17:08:55.4211671Z - fastapi==0.121.0 2025-11-03T17:08:55.4212641Z + fastapi==0.116.1 2025-11-03T17:08:55.4213712Z + fastparquet==2024.11.0 2025-11-03T17:08:55.4214792Z - fastrlock==0.8.3 2025-11-03T17:08:55.4215460Z + fastrlock==0.8.2 2025-11-03T17:08:55.4216801Z + fastsafetensors==0.1.10 2025-11-03T17:08:55.4217397Z - filelock==3.18.0 2025-11-03T17:08:55.4218274Z + filelock==3.16.1 2025-11-03T17:08:55.4219141Z + fiona==1.10.1 2025-11-03T17:08:55.4219808Z + flask==3.1.1 2025-11-03T17:08:55.4220618Z + fonttools==4.55.0 2025-11-03T17:08:55.4222187Z + fqdn==1.5.1 2025-11-03T17:08:55.4223180Z - frozenlist==1.8.0 2025-11-03T17:08:55.4223914Z + frozenlist==1.5.0 2025-11-03T17:08:55.4224689Z - fsspec==2025.9.0 2025-11-03T17:08:55.4225590Z + fsspec==2024.9.0 2025-11-03T17:08:55.4226262Z + ftfy==6.3.1 2025-11-03T17:08:55.4227465Z + genai-perf==0.0.8 2025-11-03T17:08:55.4228639Z + genson==1.3.0 2025-11-03T17:08:55.4229559Z + geopandas==1.0.1 2025-11-03T17:08:55.4230810Z - gitpython==3.1.45 2025-11-03T17:08:55.4231650Z + gitpython==3.1.44 2025-11-03T17:08:55.4232862Z + google-api-core==2.24.2 2025-11-03T17:08:55.4233968Z - google-auth==2.42.1 2025-11-03T17:08:55.4235663Z + google-auth==2.40.2 2025-11-03T17:08:55.4236404Z + google-cloud-core==2.4.3 2025-11-03T17:08:55.4237625Z + google-cloud-storage==3.4.0 2025-11-03T17:08:55.4239197Z + google-crc32c==1.7.1 2025-11-03T17:08:55.4239900Z + google-resumable-media==2.7.2 2025-11-03T17:08:55.4241112Z + googleapis-common-protos==1.70.0 2025-11-03T17:08:55.4242530Z + gpt-oss==0.0.8 2025-11-03T17:08:55.4243706Z + graphene==3.4.3 2025-11-03T17:08:55.4244633Z + graphql-core==3.2.6 2025-11-03T17:08:55.4245840Z + graphql-relay==3.2.0 2025-11-03T17:08:55.4246867Z + greenlet==3.2.3 2025-11-03T17:08:55.4247568Z - grpcio==1.76.0 2025-11-03T17:08:55.4248897Z + grpcio==1.71.0 2025-11-03T17:08:55.4249831Z + gunicorn==23.0.0 2025-11-03T17:08:55.4251121Z - h11==0.16.0 2025-11-03T17:08:55.4251948Z + h11==0.14.0 2025-11-03T17:08:55.4253154Z + h5py==3.13.0 2025-11-03T17:08:55.4253878Z + harfile==0.3.0 2025-11-03T17:08:55.4255509Z - hf-xet==1.2.0 2025-11-03T17:08:55.4256468Z + hf-xet==1.1.7 2025-11-03T17:08:55.4257713Z + hiredis==3.0.0 2025-11-03T17:08:55.4258875Z + html2text==2025.4.15 2025-11-03T17:08:55.4259789Z - httpcore==1.0.9 2025-11-03T17:08:55.4260671Z + httpcore==1.0.6 2025-11-03T17:08:55.4261795Z - httpx==0.28.1 2025-11-03T17:08:55.4262343Z + httpx==0.27.2 2025-11-03T17:08:55.4263141Z - huggingface-hub==0.36.0 2025-11-03T17:08:55.4264395Z + huggingface-hub==0.34.3 2025-11-03T17:08:55.4265020Z + humanize==4.11.0 2025-11-03T17:08:55.4266481Z + hydra-core==1.3.2 2025-11-03T17:08:55.4266974Z - hypothesis==6.56.4 2025-11-03T17:08:55.4267909Z + hypothesis==6.131.0 2025-11-03T17:08:55.4269075Z + hypothesis-graphql==0.11.1 2025-11-03T17:08:55.4269827Z + hypothesis-jsonschema==0.23.1 2025-11-03T17:08:55.4270566Z - idna==3.11 2025-11-03T17:08:55.4271547Z + idna==3.10 2025-11-03T17:08:55.4272395Z + importlib-metadata==8.7.0 2025-11-03T17:08:55.4273338Z + importlib-resources==6.5.2 2025-11-03T17:08:55.4274204Z + inflect==5.6.2 2025-11-03T17:08:55.4275144Z - iniconfig==2.3.0 2025-11-03T17:08:55.4275794Z + iniconfig==2.0.0 2025-11-03T17:08:55.4277046Z + isoduration==20.11.0 2025-11-03T17:08:55.4278139Z + isort==5.13.2 2025-11-03T17:08:55.4278934Z + itsdangerous==2.2.0 2025-11-03T17:08:55.4279864Z + jiwer==3.0.5 2025-11-03T17:08:55.4281032Z - joblib==1.5.2 2025-11-03T17:08:55.4282014Z + joblib==1.4.2 2025-11-03T17:08:55.4282846Z + jsonargparse==4.35.0 2025-11-03T17:08:55.4283951Z + jsonlines==4.0.0 2025-11-03T17:08:55.4285292Z + jsonpointer==3.0.0 2025-11-03T17:08:55.4286115Z - jsonschema==4.25.1 2025-11-03T17:08:55.4287400Z + jsonschema==4.23.0 2025-11-03T17:08:55.4290407Z - jsonschema-specifications==2025.9.1 2025-11-03T17:08:55.4291144Z + jsonschema-specifications==2024.10.1 2025-11-03T17:08:55.4291747Z + junit-xml==1.9 2025-11-03T17:08:55.4292260Z + kaleido==0.2.1 2025-11-03T17:08:55.4292777Z + kiwisolver==1.4.7 2025-11-03T17:08:55.4293271Z + kornia==0.8.1 2025-11-03T17:08:55.4293951Z + kornia-rs==0.1.9 2025-11-03T17:08:55.4294445Z + libnacl==2.1.0 2025-11-03T17:08:55.4294942Z - librosa==0.10.2 2025-11-03T17:08:55.4295551Z + librosa==0.10.2.post1 2025-11-03T17:08:55.4296581Z + lightly==1.5.20 2025-11-03T17:08:55.4297229Z + lightly-utils==0.0.2 2025-11-03T17:08:55.4298194Z + lightning==2.5.1.post0 2025-11-03T17:08:55.4298949Z + lightning-utilities==0.14.3 2025-11-03T17:08:55.4300549Z + lm-eval==0.4.9.1 (from git+https://github.com/EleutherAI/lm-evaluation-harness.git@206b7722158f58c35b7ffcd53b035fdbdda5126d) 2025-11-03T17:08:55.4301603Z + mako==1.3.10 2025-11-03T17:08:55.4302104Z - markdown==3.9 2025-11-03T17:08:55.4303081Z + markdown==3.8.2 2025-11-03T17:08:55.4303595Z - markdown-it-py==4.0.0 2025-11-03T17:08:55.4304581Z + markdown-it-py==3.0.0 2025-11-03T17:08:55.4305162Z - markupsafe==3.0.3 2025-11-03T17:08:55.4306184Z + markupsafe==3.0.1 2025-11-03T17:08:55.4307088Z + matplotlib==3.9.2 2025-11-03T17:08:55.4308140Z + mbstrdecoder==1.1.3 2025-11-03T17:08:55.4309255Z + mlflow==2.22.0 2025-11-03T17:08:55.4309777Z + mlflow-skinny==2.22.0 2025-11-03T17:08:55.4310586Z + more-itertools==10.5.0 2025-11-03T17:08:55.4311431Z - msgpack==1.1.2 2025-11-03T17:08:55.4312300Z + msgpack==1.1.0 2025-11-03T17:08:55.4313135Z + mteb==1.38.11 2025-11-03T17:08:55.4314202Z - multidict==6.7.0 2025-11-03T17:08:55.4314710Z + multidict==6.1.0 2025-11-03T17:08:55.4315851Z + multiprocess==0.70.16 2025-11-03T17:08:55.4316672Z + munch==4.0.0 2025-11-03T17:08:55.4317431Z - mypy-extensions==1.1.0 2025-11-03T17:08:55.4318603Z + mypy-extensions==1.0.0 2025-11-03T17:08:55.4319441Z - networkx==2.8.8 2025-11-03T17:08:55.4319898Z + networkx==3.2.1 2025-11-03T17:08:55.4321632Z + nltk==3.9.1 2025-11-03T17:08:55.4321945Z + num2words==0.5.14 2025-11-03T17:08:55.4323061Z + numexpr==2.10.1 2025-11-03T17:08:55.4323952Z - numpy==2.2.6 2025-11-03T17:08:55.4325050Z + numpy==1.26.4 2025-11-03T17:08:55.4325968Z + omegaconf==2.3.0 2025-11-03T17:08:55.4326377Z + open-clip-torch==2.32.0 2025-11-03T17:08:55.4327681Z + opencensus==0.11.4 2025-11-03T17:08:55.4328089Z + opencensus-context==0.1.3 2025-11-03T17:08:55.4329232Z - opencv-python-headless==4.12.0.88 2025-11-03T17:08:55.4330190Z + opencv-python-headless==4.11.0.86 2025-11-03T17:08:55.4331597Z + opentelemetry-api==1.35.0 2025-11-03T17:08:55.4332365Z + opentelemetry-exporter-prometheus==0.56b0 2025-11-03T17:08:55.4333204Z + opentelemetry-proto==1.36.0 2025-11-03T17:08:55.4334617Z + opentelemetry-sdk==1.35.0 2025-11-03T17:08:55.4335311Z + opentelemetry-semantic-conventions==0.56b0 2025-11-03T17:08:55.4335955Z - packaging==25.0 2025-11-03T17:08:55.4336710Z + packaging==24.2 2025-11-03T17:08:55.4337847Z - pandas==2.0.3 2025-11-03T17:08:55.4338517Z + pandas==2.2.3 2025-11-03T17:08:55.4339176Z + pathvalidate==3.2.1 2025-11-03T17:08:55.4340771Z + patsy==1.0.1 2025-11-03T17:08:55.4341319Z + peft==0.16.0 2025-11-03T17:08:55.4342003Z - pillow==11.0.0 2025-11-03T17:08:55.4343063Z + pillow==10.4.0 2025-11-03T17:08:55.4343939Z - platformdirs==4.5.0 2025-11-03T17:08:55.4345215Z + platformdirs==4.3.6 2025-11-03T17:08:55.4345734Z + plotly==5.24.1 2025-11-03T17:08:55.4346656Z - pluggy==1.6.0 2025-11-03T17:08:55.4347923Z + pluggy==1.5.0 2025-11-03T17:08:55.4348530Z + polars==1.29.0 2025-11-03T17:08:55.4349541Z + portalocker==2.10.1 2025-11-03T17:08:55.4350486Z + pqdm==0.2.0 2025-11-03T17:08:55.4351757Z + pretrainedmodels==0.7.4 2025-11-03T17:08:55.4352547Z - prometheus-client==0.23.1 2025-11-03T17:08:55.4353655Z + prometheus-client==0.22.0 2025-11-03T17:08:55.4354605Z - propcache==0.4.1 2025-11-03T17:08:55.4355162Z + propcache==0.2.0 2025-11-03T17:08:55.4356754Z + proto-plus==1.26.1 2025-11-03T17:08:55.4357244Z - protobuf==5.29.5 2025-11-03T17:08:55.4358206Z + protobuf==5.28.3 2025-11-03T17:08:55.4359024Z - psutil==7.1.2 2025-11-03T17:08:55.4359639Z + psutil==6.1.0 2025-11-03T17:08:55.4360553Z + py==1.11.0 2025-11-03T17:08:55.4361613Z + py-spy==0.4.0 2025-11-03T17:08:55.4362706Z + pyarrow==18.0.0 2025-11-03T17:08:55.4363754Z + pybind11==2.13.6 2025-11-03T17:08:55.4364587Z + pycocotools==2.0.8 2025-11-03T17:08:55.4365553Z - pycparser==2.23 2025-11-03T17:08:55.4366371Z + pycparser==2.22 2025-11-03T17:08:55.4367262Z + pycryptodomex==3.22.0 2025-11-03T17:08:55.4368428Z - pydantic==2.12.3 2025-11-03T17:08:55.4369298Z + pydantic==2.12.0 2025-11-03T17:08:55.4370101Z - pydantic-core==2.41.4 2025-11-03T17:08:55.4371055Z + pydantic-core==2.41.1 2025-11-03T17:08:55.4372268Z - pydantic-extra-types==2.10.6 2025-11-03T17:08:55.4372893Z + pydantic-extra-types==2.10.5 2025-11-03T17:08:55.4374164Z - pygments==2.15.0 2025-11-03T17:08:55.4374656Z + pygments==2.18.0 2025-11-03T17:08:55.4375646Z + pyogrio==0.11.0 2025-11-03T17:08:55.4376506Z + pyparsing==3.2.0 2025-11-03T17:08:55.4377436Z + pyproj==3.7.1 2025-11-03T17:08:55.4378561Z + pyrate-limiter==3.7.0 2025-11-03T17:08:55.4379381Z + pystemmer==3.0.0 2025-11-03T17:08:55.4380451Z + pytablewriter==1.2.0 2025-11-03T17:08:55.4381083Z - pytest==7.3.2 2025-11-03T17:08:55.4382155Z + pytest==8.3.5 2025-11-03T17:08:55.4382936Z + pytest-asyncio==0.24.0 2025-11-03T17:08:55.4383586Z + pytest-cov==6.3.0 2025-11-03T17:08:55.4384791Z + pytest-forked==1.6.0 2025-11-03T17:08:55.4386236Z + pytest-mock==3.14.0 2025-11-03T17:08:55.4386947Z + pytest-shard==0.1.2 2025-11-03T17:08:55.4387886Z - pytest-subtests==0.13.1 2025-11-03T17:08:55.4388801Z + pytest-subtests==0.14.1 2025-11-03T17:08:55.4389972Z + pytest-timeout==2.3.1 2025-11-03T17:08:55.4390574Z + python-box==7.3.2 2025-11-03T17:08:55.4391328Z + python-rapidjson==1.20 2025-11-03T17:08:55.4392323Z + pytorch-lightning==2.5.5 2025-11-03T17:08:55.4393067Z + pytrec-eval-terrier==0.5.7 2025-11-03T17:08:55.4393698Z - pytz==2025.2 2025-11-03T17:08:55.4395291Z + pytz==2024.2 2025-11-03T17:08:55.4395703Z + rapidfuzz==3.12.1 2025-11-03T17:08:55.4396275Z + rasterio==1.4.3 2025-11-03T17:08:55.4396828Z - ray==2.51.1 2025-11-03T17:08:55.4397892Z + ray==2.48.0 2025-11-03T17:08:55.4398613Z - redis==7.0.1 2025-11-03T17:08:55.4399753Z + redis==5.2.0 2025-11-03T17:08:55.4400708Z - referencing==0.37.0 2025-11-03T17:08:55.4401802Z + referencing==0.35.1 2025-11-03T17:08:55.4402791Z - regex==2025.10.23 2025-11-03T17:08:55.4403285Z + regex==2024.9.11 2025-11-03T17:08:55.4404024Z - requests==2.32.5 2025-11-03T17:08:55.4405215Z + requests==2.32.3 2025-11-03T17:08:55.4405780Z + responses==0.25.3 2025-11-03T17:08:55.4406901Z + rfc3339-validator==0.1.4 2025-11-03T17:08:55.4407659Z + rfc3987==1.3.8 2025-11-03T17:08:55.4408206Z - rich==14.2.0 2025-11-03T17:08:55.4410117Z + rich==13.9.4 2025-11-03T17:08:55.4410639Z + rioxarray==0.19.0 2025-11-03T17:08:55.4411832Z + rouge-score==0.1.2 2025-11-03T17:08:55.4412998Z - rpds-py==0.28.0 2025-11-03T17:08:55.4413642Z + rpds-py==0.20.1 2025-11-03T17:08:55.4414366Z + rtree==1.4.0 2025-11-03T17:08:55.4415518Z + runai-model-streamer==0.14.0 2025-11-03T17:08:55.4416230Z + runai-model-streamer-gcs==0.14.0 2025-11-03T17:08:55.4417855Z + runai-model-streamer-s3==0.14.0 2025-11-03T17:08:55.4418506Z - s3transfer==0.10.4 2025-11-03T17:08:55.4419022Z + s3transfer==0.10.3 2025-11-03T17:08:55.4420110Z + sacrebleu==2.4.3 2025-11-03T17:08:55.4420616Z - safetensors==0.6.2 2025-11-03T17:08:55.4422137Z + safetensors==0.4.5 2025-11-03T17:08:55.4422934Z + schemathesis==3.39.15 2025-11-03T17:08:55.4423980Z - scikit-image==0.22.0 2025-11-03T17:08:55.4424972Z + scikit-image==0.25.2 2025-11-03T17:08:55.4426002Z - scikit-learn==1.7.2 2025-11-03T17:08:55.4427102Z + scikit-learn==1.5.2 2025-11-03T17:08:55.4428260Z - scipy==1.14.1 2025-11-03T17:08:55.4429156Z + scipy==1.13.1 2025-11-03T17:08:55.4430155Z + segmentation-models-pytorch==0.4.0 2025-11-03T17:08:55.4430776Z + sentence-transformers==3.2.1 2025-11-03T17:08:55.4432185Z - setuptools==78.1.1 2025-11-03T17:08:55.4433206Z + setuptools==77.0.3 2025-11-03T17:08:55.4434243Z + shapely==2.1.1 2025-11-03T17:08:55.4435211Z - six==1.17.0 2025-11-03T17:08:55.4436388Z + six==1.16.0 2025-11-03T17:08:55.4437136Z + smart-open==7.1.0 2025-11-03T17:08:55.4437862Z - soundfile==0.13.1 2025-11-03T17:08:55.4439389Z + soundfile==0.12.1 2025-11-03T17:08:55.4440182Z - soxr==1.0.0 2025-11-03T17:08:55.4441425Z + soxr==0.5.0.post1 2025-11-03T17:08:55.4442072Z + sqlalchemy==2.0.41 2025-11-03T17:08:55.4443140Z + sqlitedict==2.1.0 2025-11-03T17:08:55.4444221Z + sqlparse==0.5.3 2025-11-03T17:08:55.4444945Z - starlette==0.49.3 2025-11-03T17:08:55.4446009Z + starlette==0.46.2 2025-11-03T17:08:55.4446862Z + starlette-testclient==0.4.1 2025-11-03T17:08:55.4447964Z + statsmodels==0.14.4 2025-11-03T17:08:55.4448627Z + structlog==25.4.0 2025-11-03T17:08:55.4450115Z + tabledata==1.3.3 2025-11-03T17:08:55.4450960Z + tblib==3.1.0 2025-11-03T17:08:55.4451870Z + tcolorpy==0.1.6 2025-11-03T17:08:55.4452782Z + tenacity==9.1.2 2025-11-03T17:08:55.4453517Z + tensorboardx==2.6.4 2025-11-03T17:08:55.4454616Z + tensorizer==2.10.1 2025-11-03T17:08:55.4455688Z + termcolor==3.1.0 2025-11-03T17:08:55.4456798Z + terratorch==1.0.2 (from git+https://github.com/IBM/terratorch.git@07184fcf91a1324f831ff521dd238d97fe350e3e) 2025-11-03T17:08:55.4457574Z - threadpoolctl==3.6.0 2025-11-03T17:08:55.4458314Z + threadpoolctl==3.5.0 2025-11-03T17:08:55.4459306Z - tifffile==2025.10.16 2025-11-03T17:08:55.4460158Z + tifffile==2025.3.30 2025-11-03T17:08:55.4460850Z + timm==1.0.17 2025-11-03T17:08:55.4462231Z - tokenizers==0.22.1 2025-11-03T17:08:55.4462934Z + tokenizers==0.22.0 2025-11-03T17:08:55.4463525Z + tomli==2.2.1 2025-11-03T17:08:55.4464576Z + tomli-w==1.2.0 2025-11-03T17:08:55.4466547Z - torch==2.10.0a0+git3f6538f (from file:///var/lib/jenkins/workspace/dist/torch-2.10.0a0%2Bgit3f6538f-cp312-cp312-linux_x86_64.whl) 2025-11-03T17:08:55.4468556Z + torch==2.10.0a0+git3f6538f (from file:///var/lib/jenkins/workspace/dist/torch-2.10.0a0+git3f6538f-cp312-cp312-linux_x86_64.whl#sha256=288b4779bb4efa62cf5264c0eca9fdb02025e7a7fb44d42e3f512a68d8e60da0) 2025-11-03T17:08:55.4470242Z - torchaudio==2.10.0a0+3b0e7a6 (from file:///var/lib/jenkins/workspace/dist/audio/torchaudio-2.10.0a0%2B3b0e7a6-cp312-cp312-linux_x86_64.whl) 2025-11-03T17:08:55.4472003Z + torchaudio==2.10.0a0+3b0e7a6 (from file:///var/lib/jenkins/workspace/dist/audio/torchaudio-2.10.0a0+3b0e7a6-cp312-cp312-linux_x86_64.whl#sha256=70d7f19c17cbc4bb012226c17e5f5380b3bd6bda732a45f329984580e0bc17a6) 2025-11-03T17:08:55.4473137Z + torchgeo==0.7.0 2025-11-03T17:08:55.4473603Z + torchmetrics==1.8.2 2025-11-03T17:08:55.4474711Z - torchvision==0.25.0a0+cfbc5c2 (from file:///var/lib/jenkins/workspace/dist/vision/torchvision-0.25.0a0%2Bcfbc5c2-cp312-cp312-linux_x86_64.whl) 2025-11-03T17:08:55.4476505Z + torchvision==0.25.0a0+cfbc5c2 (from file:///var/lib/jenkins/workspace/dist/vision/torchvision-0.25.0a0+cfbc5c2-cp312-cp312-linux_x86_64.whl#sha256=7fa43396b4b9ce1b239050abea02e7fbabadf1cbdf180883becb1f6530983b15) 2025-11-03T17:08:55.4477677Z - tqdm==4.67.1 2025-11-03T17:08:55.4478057Z + tqdm==4.66.6 2025-11-03T17:08:55.4478483Z + tqdm-multiprocess==0.0.11 2025-11-03T17:08:55.4478993Z - transformers==4.57.1 2025-11-03T17:08:55.4479525Z + transformers==4.56.2 2025-11-03T17:08:55.4480042Z + transformers-stream-generator==0.0.5 2025-11-03T17:08:55.4480530Z + tritonclient==2.51.0 2025-11-03T17:08:55.4480933Z + typepy==1.3.2 2025-11-03T17:08:55.4481293Z - typer==0.20.0 2025-11-03T17:08:55.4481874Z + typer==0.15.2 2025-11-03T17:08:55.4482684Z + types-python-dateutil==2.9.0.20241206 2025-11-03T17:08:55.4483908Z + typeshed-client==2.8.2 2025-11-03T17:08:55.4484447Z - tzdata==2025.2 2025-11-03T17:08:55.4485345Z + tzdata==2024.2 2025-11-03T17:08:55.4486396Z + uri-template==1.3.0 2025-11-03T17:08:55.4487149Z - urllib3==2.5.0 2025-11-03T17:08:55.4487608Z + urllib3==2.2.3 2025-11-03T17:08:55.4488669Z - uvicorn==0.38.0 2025-11-03T17:08:55.4489047Z + uvicorn==0.35.0 2025-11-03T17:08:55.4489940Z + vector-quantize-pytorch==1.24.2 2025-11-03T17:08:55.4491204Z + virtualenv==20.31.2 2025-11-03T17:08:55.4491902Z + vocos==0.1.0 2025-11-03T17:08:55.4492778Z + wcwidth==0.2.13 2025-11-03T17:08:55.4493450Z + webcolors==24.11.1 2025-11-03T17:08:55.4493900Z + word2number==1.1 2025-11-03T17:08:55.4494765Z - wrapt==2.0.0 2025-11-03T17:08:55.4495511Z + wrapt==1.17.2 2025-11-03T17:08:55.4496682Z + xarray==2025.7.1 2025-11-03T17:08:55.4497276Z + xxhash==3.5.0 2025-11-03T17:08:55.4498297Z - yarl==1.22.0 2025-11-03T17:08:55.4498690Z + yarl==1.17.1 2025-11-03T17:08:55.4499329Z + zipp==3.23.0 2025-11-03T17:08:55.4500691Z + zstandard==0.23.0 2025-11-03T17:08:55.6204029Z 2025-11-03 17:08:55,619 [INFO] cli.lib.core.vllm.vllm_test: Done. installed requirements for test dependencies 2025-11-03T17:08:55.6204685Z 2025-11-03 17:08:55,620 [INFO] cli.lib.core.vllm.vllm_test: Double check installed packages 2025-11-03T17:08:55.6252863Z 2025-11-03 17:08:55,625 [INFO] cli.lib.common.pip_helper: torch already exist with version: 2.10.0a0+git3f6538f 2025-11-03T17:08:55.6256227Z 2025-11-03 17:08:55,625 [INFO] cli.lib.common.pip_helper: xformers already exist with version: 0.0.33+5d4b92a5.d20251103 2025-11-03T17:08:55.6260085Z 2025-11-03 17:08:55,625 [INFO] cli.lib.common.pip_helper: torchvision already exist with version: 0.25.0a0+cfbc5c2 2025-11-03T17:08:55.6264408Z 2025-11-03 17:08:55,626 [INFO] cli.lib.common.pip_helper: torchaudio already exist with version: 2.10.0a0+3b0e7a6 2025-11-03T17:08:55.6269811Z 2025-11-03 17:08:55,626 [INFO] cli.lib.common.pip_helper: vllm already exist with version: 0.11.0rc2.dev389+ge51928192.d20251103 2025-11-03T17:08:55.6270697Z 2025-11-03 17:08:55,626 [INFO] cli.lib.core.vllm.vllm_test: Done. checked installed packages 2025-11-03T17:08:55.6271414Z 2025-11-03 17:08:55,626 [INFO] cli.lib.core.vllm.lib: run vllm tests..... 2025-11-03T17:08:55.6272065Z 2025-11-03 17:08:55,626 [INFO] cli.lib.core.vllm.lib: Running tests: Basic models test 2025-11-03T17:08:55.6272648Z 2025-11-03 17:08:55,627 [INFO] cli.lib.core.vllm.lib: Running step: pytest -v -s models/test_transformers.py 2025-11-03T17:08:55.6273268Z 2025-11-03 17:08:55,627 [INFO] cli.lib.common.utils: [shell] pytest -v -s models/test_transformers.py 2025-11-03T17:09:06.0389657Z INFO 11-03 17:09:06 [__init__.py:224] Automatically detected platform cuda. 2025-11-03T17:09:08.6338494Z /opt/conda/envs/py_3.12/lib/python3.12/site-packages/pytest_asyncio/plugin.py:208: PytestDeprecationWarning: The configuration option "asyncio_default_fixture_loop_scope" is unset. 2025-11-03T17:09:08.6341298Z The event loop scope for asynchronous fixtures will default to the fixture caching scope. Future versions of pytest-asyncio will default the loop scope for asynchronous fixtures to function scope. Set the default fixture loop scope explicitly in order to avoid unexpected behavior in the future. Valid fixture loop scopes are: "function", "class", "module", "package", "session" 2025-11-03T17:09:08.6342774Z 2025-11-03T17:09:08.6343018Z warnings.warn(PytestDeprecationWarning(_DEFAULT_FIXTURE_LOOP_SCOPE_UNSET)) 2025-11-03T17:09:08.6408339Z ============================= test session starts ============================== 2025-11-03T17:09:08.6408916Z platform linux -- Python 3.12.12, pytest-8.3.5, pluggy-1.5.0 -- /opt/conda/envs/py_3.12/bin/python 2025-11-03T17:09:08.6521416Z cachedir: .pytest_cache 2025-11-03T17:09:08.6522169Z hypothesis profile 'ci' -> database=None, deadline=None, print_blob=True, derandomize=True, suppress_health_check=[HealthCheck.too_slow] 2025-11-03T17:09:08.6522813Z rootdir: /var/lib/jenkins/workspace/vllm 2025-11-03T17:09:08.6523107Z configfile: pyproject.toml 2025-11-03T17:09:08.6524111Z plugins: cpp-2.3.0, flakefinder-1.1.0, rerunfailures-14.0, xdist-3.3.1, xdoctest-1.3.0, asyncio-0.24.0, cov-6.3.0, timeout-2.3.1, hydra-core-1.3.2, mock-3.14.0, schemathesis-3.39.15, buildkite-test-collector-0.1.9, subtests-0.14.1, hypothesis-6.131.0, forked-1.6.0, shard-0.1.2, anyio-4.6.2.post1 2025-11-03T17:09:08.6525177Z asyncio: mode=Mode.STRICT, default_loop_scope=None 2025-11-03T17:09:11.4001421Z collecting ... WARNING 11-03 17:09:11 [interface.py:514] Current platform cuda does not have '_pytestfixturefunction' attribute. 2025-11-03T17:09:11.8621718Z WARNING 11-03 17:09:11 [interface.py:514] Current platform cuda does not have '__test__' attribute. 2025-11-03T17:09:11.8622505Z WARNING 11-03 17:09:11 [interface.py:514] Current platform cuda does not have '__bases__' attribute. 2025-11-03T17:09:11.8623006Z WARNING 11-03 17:09:11 [interface.py:514] Current platform cuda does not have '__test__' attribute. 2025-11-03T17:09:11.8623529Z WARNING 11-03 17:09:11 [interface.py:514] Current platform cuda does not have '_schemathesis_test' attribute. 2025-11-03T17:09:11.8651866Z  2025-11-03T17:09:11.8655789Z collecting 12 items  2025-11-03T17:09:11.8656532Z collected 12 items  2025-11-03T17:09:11.8662743Z Running 12 items in this shard: tests/models/test_transformers.py::test_models[meta-llama/Llama-3.2-1B-Instruct-transformers], tests/models/test_transformers.py::test_models[hmellor/Ilama-3.2-1B-auto], tests/models/test_transformers.py::test_models[allenai/OLMoE-1B-7B-0924-transformers], tests/models/test_transformers.py::test_hybrid_attention, tests/models/test_transformers.py::test_distributed, tests/models/test_transformers.py::test_quantization[5-32-TheBloke/TinyLlama-1.1B-Chat-v0.3-AWQ-quantization_kwargs0], tests/models/test_transformers.py::test_quantization[5-32-TheBloke/TinyLlama-1.1B-Chat-v0.3-GPTQ-quantization_kwargs1], tests/models/test_transformers.py::test_quantization[5-32-meta-llama/Llama-3.2-1B-Instruct-quantization_kwargs2], tests/models/test_transformers.py::test_embed_loading[Qwen/Qwen3-Embedding-0.6B], tests/models/test_transformers.py::test_embed_loading[meta-llama/Llama-3.2-1B-Instruct], tests/models/test_transformers.py::test_pooling[TransformersEmbeddingModel], tests/models/test_transformers.py::test_pooling[TransformersForSequenceClassification] 2025-11-03T17:09:11.8668767Z 2025-11-03T17:09:11.8920306Z models/test_transformers.py::test_models[meta-llama/Llama-3.2-1B-Instruct-transformers] INFO 11-03 17:09:11 [utils.py:239] non-default args: {'trust_remote_code': True, 'seed': 0, 'max_model_len': 1024, 'block_size': 16, 'disable_log_stats': True, 'enable_chunked_prefill': False, 'compilation_config': {'level': None, 'debug_dump_path': None, 'cache_dir': '', 'backend': '', 'custom_ops': [], 'splitting_ops': None, 'use_inductor': True, 'compile_sizes': None, 'inductor_compile_config': {'enable_auto_functionalized_v2': False}, 'inductor_passes': {}, 'cudagraph_mode': None, 'use_cudagraph': True, 'cudagraph_num_of_warmups': 0, 'cudagraph_capture_sizes': [4], 'cudagraph_copy_inputs': False, 'full_cuda_graph': False, 'use_inductor_graph_partition': False, 'pass_config': {}, 'max_capture_size': None, 'local_cache_dir': None}, 'model_impl': 'transformers', 'model': 'meta-llama/Llama-3.2-1B-Instruct'} 2025-11-03T17:09:11.9957516Z 2025-11-03T17:09:11.9958269Z config.json: 0% 0.00/877 [00:00, 'use_cudagraph': True, 'cudagraph_num_of_warmups': 1, 'cudagraph_capture_sizes': [4], 'cudagraph_copy_inputs': False, 'full_cuda_graph': True, 'use_inductor_graph_partition': False, 'pass_config': {}, 'max_capture_size': 4, 'local_cache_dir': None} 2025-11-03T17:09:20.9805668Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:09:20.9833591Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:09:20.9841608Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:09:20.9849450Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:09:20.9856875Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:09:20.9864941Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:09:20.9867722Z (EngineCore_DP0 pid=768) INFO 11-03 17:09:20 [parallel_state.py:1231] rank 0 in world size 1 is assigned as DP rank 0, PP rank 0, TP rank 0, EP rank 0 2025-11-03T17:09:21.3466181Z (EngineCore_DP0 pid=768) INFO 11-03 17:09:21 [topk_topp_sampler.py:59] Using FlashInfer for top-p & top-k sampling. 2025-11-03T17:09:21.3961185Z (EngineCore_DP0 pid=768) INFO 11-03 17:09:21 [gpu_model_runner.py:2840] Starting to load model meta-llama/Llama-3.2-1B-Instruct... 2025-11-03T17:09:21.6744429Z (EngineCore_DP0 pid=768) INFO 11-03 17:09:21 [gpu_model_runner.py:2870] Loading model from scratch... 2025-11-03T17:09:21.6746004Z (EngineCore_DP0 pid=768) INFO 11-03 17:09:21 [transformers.py:493] Using Transformers backend. 2025-11-03T17:09:21.6749500Z (EngineCore_DP0 pid=768) `torch_dtype` is deprecated! Use `dtype` instead! 2025-11-03T17:09:21.8770449Z (EngineCore_DP0 pid=768) INFO 11-03 17:09:21 [cuda.py:404] Using Flash Attention backend on V1 engine. 2025-11-03T17:09:22.0206290Z (EngineCore_DP0 pid=768) INFO 11-03 17:09:22 [weight_utils.py:419] Using model weights format ['*.safetensors'] 2025-11-03T17:09:22.1191664Z (EngineCore_DP0 pid=768) 2025-11-03T17:09:23.2416358Z model.safetensors: 0% 0.00/2.47G [00:00, 'use_cudagraph': True, 'cudagraph_num_of_warmups': 1, 'cudagraph_capture_sizes': [4], 'cudagraph_copy_inputs': False, 'full_cuda_graph': True, 'use_inductor_graph_partition': False, 'pass_config': {}, 'max_capture_size': 4, 'local_cache_dir': None} 2025-11-03T17:10:20.8458841Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:10:20.8488280Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:10:20.8496006Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:10:20.8503824Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:10:20.8510942Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:10:20.8517883Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:10:20.8520450Z (EngineCore_DP0 pid=1114) INFO 11-03 17:10:20 [parallel_state.py:1231] rank 0 in world size 1 is assigned as DP rank 0, PP rank 0, TP rank 0, EP rank 0 2025-11-03T17:10:21.0809161Z (EngineCore_DP0 pid=1114) INFO 11-03 17:10:21 [topk_topp_sampler.py:59] Using FlashInfer for top-p & top-k sampling. 2025-11-03T17:10:21.2951459Z (EngineCore_DP0 pid=1114) WARNING 11-03 17:10:21 [utils.py:188] TransformersForCausalLM has no vLLM implementation, falling back to Transformers implementation. Some features may not be supported and performance may not be optimal. 2025-11-03T17:10:21.2953456Z (EngineCore_DP0 pid=1114) INFO 11-03 17:10:21 [gpu_model_runner.py:2840] Starting to load model hmellor/Ilama-3.2-1B... 2025-11-03T17:10:21.5196754Z (EngineCore_DP0 pid=1114) INFO 11-03 17:10:21 [gpu_model_runner.py:2870] Loading model from scratch... 2025-11-03T17:10:21.5197817Z (EngineCore_DP0 pid=1114) INFO 11-03 17:10:21 [transformers.py:493] Using Transformers backend. 2025-11-03T17:10:21.5472255Z (EngineCore_DP0 pid=1114) `torch_dtype` is deprecated! Use `dtype` instead! 2025-11-03T17:10:21.7317810Z (EngineCore_DP0 pid=1114) INFO 11-03 17:10:21 [cuda.py:404] Using Flash Attention backend on V1 engine. 2025-11-03T17:10:21.8531649Z (EngineCore_DP0 pid=1114) INFO 11-03 17:10:21 [weight_utils.py:419] Using model weights format ['*.safetensors'] 2025-11-03T17:10:22.1746396Z (EngineCore_DP0 pid=1114) 2025-11-03T17:10:23.5023750Z model.safetensors: 0% 0.00/4.94G [00:00, 'use_cudagraph': False, 'cudagraph_num_of_warmups': 0, 'cudagraph_capture_sizes': [4], 'cudagraph_copy_inputs': False, 'full_cuda_graph': False, 'use_inductor_graph_partition': False, 'pass_config': {}, 'max_capture_size': 4, 'local_cache_dir': None} 2025-11-03T17:11:28.3802019Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:11:28.3832372Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:11:28.3839667Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:11:28.3847179Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:11:28.3854959Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:11:28.3862530Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:11:28.3864964Z (EngineCore_DP0 pid=1504) INFO 11-03 17:11:28 [parallel_state.py:1231] rank 0 in world size 1 is assigned as DP rank 0, PP rank 0, TP rank 0, EP rank 0 2025-11-03T17:11:28.6185142Z (EngineCore_DP0 pid=1504) INFO 11-03 17:11:28 [topk_topp_sampler.py:59] Using FlashInfer for top-p & top-k sampling. 2025-11-03T17:11:28.6734780Z (EngineCore_DP0 pid=1504) INFO 11-03 17:11:28 [gpu_model_runner.py:2840] Starting to load model hmellor/tiny-random-Gemma2ForCausalLM... 2025-11-03T17:11:28.9081516Z (EngineCore_DP0 pid=1504) INFO 11-03 17:11:28 [gpu_model_runner.py:2870] Loading model from scratch... 2025-11-03T17:11:28.9082398Z (EngineCore_DP0 pid=1504) INFO 11-03 17:11:28 [transformers.py:493] Using Transformers backend. 2025-11-03T17:11:28.9084935Z (EngineCore_DP0 pid=1504) `torch_dtype` is deprecated! Use `dtype` instead! 2025-11-03T17:11:29.0096801Z (EngineCore_DP0 pid=1504) INFO 11-03 17:11:29 [cuda.py:404] Using Flash Attention backend on V1 engine. 2025-11-03T17:11:29.0741876Z (EngineCore_DP0 pid=1504) INFO 11-03 17:11:29 [weight_utils.py:419] Using model weights format ['*.safetensors'] 2025-11-03T17:11:29.2006367Z (EngineCore_DP0 pid=1504) 2025-11-03T17:11:29.3333109Z model.safetensors: 0% 0.00/16.9M [00:00, 'use_cudagraph': False, 'cudagraph_num_of_warmups': 0, 'cudagraph_capture_sizes': [4], 'cudagraph_copy_inputs': False, 'full_cuda_graph': False, 'use_inductor_graph_partition': False, 'pass_config': {}, 'max_capture_size': 4, 'local_cache_dir': None} 2025-11-03T17:11:53.3437272Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:11:53.3470019Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:11:53.3477350Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:11:53.3485237Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:11:53.3493576Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:11:53.3500272Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:11:53.3503053Z (EngineCore_DP0 pid=1800) INFO 11-03 17:11:53 [parallel_state.py:1231] rank 0 in world size 1 is assigned as DP rank 0, PP rank 0, TP rank 0, EP rank 0 2025-11-03T17:11:53.5806445Z (EngineCore_DP0 pid=1800) INFO 11-03 17:11:53 [topk_topp_sampler.py:59] Using FlashInfer for top-p & top-k sampling. 2025-11-03T17:11:53.6110402Z (EngineCore_DP0 pid=1800) INFO 11-03 17:11:53 [gpu_model_runner.py:2840] Starting to load model hmellor/tiny-random-Gemma2ForCausalLM... 2025-11-03T17:11:53.8367783Z (EngineCore_DP0 pid=1800) INFO 11-03 17:11:53 [gpu_model_runner.py:2870] Loading model from scratch... 2025-11-03T17:11:53.9012546Z (EngineCore_DP0 pid=1800) INFO 11-03 17:11:53 [cuda.py:404] Using Flash Attention backend on V1 engine. 2025-11-03T17:11:53.9726471Z (EngineCore_DP0 pid=1800) INFO 11-03 17:11:53 [weight_utils.py:419] Using model weights format ['*.safetensors'] 2025-11-03T17:11:54.0560163Z (EngineCore_DP0 pid=1800) INFO 11-03 17:11:54 [weight_utils.py:480] No model.safetensors.index.json found in remote. 2025-11-03T17:11:54.0565643Z (EngineCore_DP0 pid=1800) 2025-11-03T17:11:54.0654025Z Loading safetensors checkpoint shards: 0% 0/1 [00:00, 'use_cudagraph': False, 'cudagraph_num_of_warmups': 0, 'cudagraph_capture_sizes': [4], 'cudagraph_copy_inputs': False, 'full_cuda_graph': False, 'use_inductor_graph_partition': False, 'pass_config': {}, 'max_capture_size': 4, 'local_cache_dir': None} 2025-11-03T17:12:17.3407938Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:12:17.3438046Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:12:17.3445220Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:12:17.3452665Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:12:17.3459854Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:12:17.3467142Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:12:17.3469703Z (EngineCore_DP0 pid=2046) INFO 11-03 17:12:17 [parallel_state.py:1231] rank 0 in world size 1 is assigned as DP rank 0, PP rank 0, TP rank 0, EP rank 0 2025-11-03T17:12:17.5809047Z (EngineCore_DP0 pid=2046) INFO 11-03 17:12:17 [topk_topp_sampler.py:59] Using FlashInfer for top-p & top-k sampling. 2025-11-03T17:12:17.6274281Z (EngineCore_DP0 pid=2046) INFO 11-03 17:12:17 [gpu_model_runner.py:2840] Starting to load model TheBloke/TinyLlama-1.1B-Chat-v0.3-AWQ... 2025-11-03T17:12:17.8506967Z (EngineCore_DP0 pid=2046) INFO 11-03 17:12:17 [gpu_model_runner.py:2870] Loading model from scratch... 2025-11-03T17:12:17.9151107Z (EngineCore_DP0 pid=2046) INFO 11-03 17:12:17 [cuda.py:404] Using Flash Attention backend on V1 engine. 2025-11-03T17:12:18.0372207Z (EngineCore_DP0 pid=2046) INFO 11-03 17:12:18 [weight_utils.py:419] Using model weights format ['*.safetensors'] 2025-11-03T17:12:18.1154534Z (EngineCore_DP0 pid=2046) 2025-11-03T17:12:18.9394973Z model.safetensors: 0% 0.00/766M [00:00, 'use_cudagraph': False, 'cudagraph_num_of_warmups': 0, 'cudagraph_capture_sizes': [4], 'cudagraph_copy_inputs': False, 'full_cuda_graph': False, 'use_inductor_graph_partition': False, 'pass_config': {}, 'max_capture_size': 4, 'local_cache_dir': None} 2025-11-03T17:12:31.5226252Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:12:31.5255875Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:12:31.5263492Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:12:31.5271429Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:12:31.5279315Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:12:31.5286782Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:12:31.5289602Z (EngineCore_DP0 pid=2206) INFO 11-03 17:12:31 [parallel_state.py:1231] rank 0 in world size 1 is assigned as DP rank 0, PP rank 0, TP rank 0, EP rank 0 2025-11-03T17:12:31.7613844Z (EngineCore_DP0 pid=2206) INFO 11-03 17:12:31 [topk_topp_sampler.py:59] Using FlashInfer for top-p & top-k sampling. 2025-11-03T17:12:31.8325259Z (EngineCore_DP0 pid=2206) INFO 11-03 17:12:31 [gpu_model_runner.py:2840] Starting to load model TheBloke/TinyLlama-1.1B-Chat-v0.3-AWQ... 2025-11-03T17:12:32.0652105Z (EngineCore_DP0 pid=2206) INFO 11-03 17:12:32 [gpu_model_runner.py:2870] Loading model from scratch... 2025-11-03T17:12:32.0653408Z (EngineCore_DP0 pid=2206) INFO 11-03 17:12:32 [transformers.py:493] Using Transformers backend. 2025-11-03T17:12:32.0655877Z (EngineCore_DP0 pid=2206) `torch_dtype` is deprecated! Use `dtype` instead! 2025-11-03T17:12:32.2450579Z (EngineCore_DP0 pid=2206) INFO 11-03 17:12:32 [cuda.py:404] Using Flash Attention backend on V1 engine. 2025-11-03T17:12:32.3210290Z (EngineCore_DP0 pid=2206) INFO 11-03 17:12:32 [weight_utils.py:419] Using model weights format ['*.safetensors'] 2025-11-03T17:12:32.3801656Z (EngineCore_DP0 pid=2206) INFO 11-03 17:12:32 [weight_utils.py:480] No model.safetensors.index.json found in remote. 2025-11-03T17:12:32.3807682Z (EngineCore_DP0 pid=2206) 2025-11-03T17:12:32.5402717Z Loading safetensors checkpoint shards: 0% 0/1 [00:00, 'use_cudagraph': False, 'cudagraph_num_of_warmups': 0, 'cudagraph_capture_sizes': [4], 'cudagraph_copy_inputs': False, 'full_cuda_graph': False, 'use_inductor_graph_partition': False, 'pass_config': {}, 'max_capture_size': 4, 'local_cache_dir': None} 2025-11-03T17:12:45.6804618Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:12:45.6833920Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:12:45.6840855Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:12:45.6848184Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:12:45.6855207Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:12:45.6862641Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:12:45.6865341Z (EngineCore_DP0 pid=2337) INFO 11-03 17:12:45 [parallel_state.py:1231] rank 0 in world size 1 is assigned as DP rank 0, PP rank 0, TP rank 0, EP rank 0 2025-11-03T17:12:45.9188669Z (EngineCore_DP0 pid=2337) INFO 11-03 17:12:45 [topk_topp_sampler.py:59] Using FlashInfer for top-p & top-k sampling. 2025-11-03T17:12:45.9644610Z (EngineCore_DP0 pid=2337) INFO 11-03 17:12:45 [gpu_model_runner.py:2840] Starting to load model TheBloke/TinyLlama-1.1B-Chat-v0.3-GPTQ... 2025-11-03T17:12:46.1938148Z (EngineCore_DP0 pid=2337) INFO 11-03 17:12:46 [gpu_model_runner.py:2870] Loading model from scratch... 2025-11-03T17:12:46.1951434Z (EngineCore_DP0 pid=2337) INFO 11-03 17:12:46 [gptq_marlin.py:357] Using MarlinLinearKernel for GPTQMarlinLinearMethod 2025-11-03T17:12:46.2587132Z (EngineCore_DP0 pid=2337) INFO 11-03 17:12:46 [cuda.py:404] Using Flash Attention backend on V1 engine. 2025-11-03T17:12:46.4044042Z (EngineCore_DP0 pid=2337) INFO 11-03 17:12:46 [weight_utils.py:419] Using model weights format ['*.safetensors'] 2025-11-03T17:12:46.4841371Z (EngineCore_DP0 pid=2337) 2025-11-03T17:12:48.2392609Z model.safetensors: 0% 0.00/768M [00:00, 'use_cudagraph': False, 'cudagraph_num_of_warmups': 0, 'cudagraph_capture_sizes': [4], 'cudagraph_copy_inputs': False, 'full_cuda_graph': False, 'use_inductor_graph_partition': False, 'pass_config': {}, 'max_capture_size': 4, 'local_cache_dir': None} 2025-11-03T17:13:01.5664988Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:13:01.5694948Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:13:01.5702577Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:13:01.5711635Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:13:01.5719127Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:13:01.5726336Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:13:01.5728909Z (EngineCore_DP0 pid=2527) INFO 11-03 17:13:01 [parallel_state.py:1231] rank 0 in world size 1 is assigned as DP rank 0, PP rank 0, TP rank 0, EP rank 0 2025-11-03T17:13:01.8053171Z (EngineCore_DP0 pid=2527) INFO 11-03 17:13:01 [topk_topp_sampler.py:59] Using FlashInfer for top-p & top-k sampling. 2025-11-03T17:13:01.8768381Z (EngineCore_DP0 pid=2527) INFO 11-03 17:13:01 [gpu_model_runner.py:2840] Starting to load model TheBloke/TinyLlama-1.1B-Chat-v0.3-GPTQ... 2025-11-03T17:13:02.1072073Z (EngineCore_DP0 pid=2527) INFO 11-03 17:13:02 [gpu_model_runner.py:2870] Loading model from scratch... 2025-11-03T17:13:02.1073310Z (EngineCore_DP0 pid=2527) INFO 11-03 17:13:02 [transformers.py:493] Using Transformers backend. 2025-11-03T17:13:02.1075606Z (EngineCore_DP0 pid=2527) `torch_dtype` is deprecated! Use `dtype` instead! 2025-11-03T17:13:02.2399251Z (EngineCore_DP0 pid=2527) INFO 11-03 17:13:02 [gptq_marlin.py:357] Using MarlinLinearKernel for GPTQMarlinLinearMethod 2025-11-03T17:13:02.3238550Z (EngineCore_DP0 pid=2527) INFO 11-03 17:13:02 [cuda.py:404] Using Flash Attention backend on V1 engine. 2025-11-03T17:13:02.4010445Z (EngineCore_DP0 pid=2527) INFO 11-03 17:13:02 [weight_utils.py:419] Using model weights format ['*.safetensors'] 2025-11-03T17:13:02.4642247Z (EngineCore_DP0 pid=2527) INFO 11-03 17:13:02 [weight_utils.py:480] No model.safetensors.index.json found in remote. 2025-11-03T17:13:02.4648165Z (EngineCore_DP0 pid=2527) 2025-11-03T17:13:02.6465622Z Loading safetensors checkpoint shards: 0% 0/1 [00:00, 'use_cudagraph': False, 'cudagraph_num_of_warmups': 0, 'cudagraph_capture_sizes': [4], 'cudagraph_copy_inputs': False, 'full_cuda_graph': False, 'use_inductor_graph_partition': False, 'pass_config': {}, 'max_capture_size': 4, 'local_cache_dir': None} 2025-11-03T17:13:15.2929964Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:13:15.2959770Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:13:15.2966432Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:13:15.2974112Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:13:15.2980959Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:13:15.2988280Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:13:15.2991027Z (EngineCore_DP0 pid=2655) INFO 11-03 17:13:15 [parallel_state.py:1231] rank 0 in world size 1 is assigned as DP rank 0, PP rank 0, TP rank 0, EP rank 0 2025-11-03T17:13:15.5284243Z (EngineCore_DP0 pid=2655) INFO 11-03 17:13:15 [topk_topp_sampler.py:59] Using FlashInfer for top-p & top-k sampling. 2025-11-03T17:13:15.5739255Z (EngineCore_DP0 pid=2655) INFO 11-03 17:13:15 [gpu_model_runner.py:2840] Starting to load model meta-llama/Llama-3.2-1B-Instruct... 2025-11-03T17:13:15.8028264Z (EngineCore_DP0 pid=2655) INFO 11-03 17:13:15 [gpu_model_runner.py:2870] Loading model from scratch... 2025-11-03T17:13:16.0184975Z (EngineCore_DP0 pid=2655) INFO 11-03 17:13:16 [cuda.py:404] Using Flash Attention backend on V1 engine. 2025-11-03T17:13:16.0414243Z (EngineCore_DP0 pid=2655) INFO 11-03 17:13:16 [bitsandbytes_loader.py:791] Loading weights with BitsAndBytes quantization. May take a while ... 2025-11-03T17:13:16.1655752Z (EngineCore_DP0 pid=2655) INFO 11-03 17:13:16 [weight_utils.py:419] Using model weights format ['*.safetensors'] 2025-11-03T17:13:16.2797154Z (EngineCore_DP0 pid=2655) INFO 11-03 17:13:16 [weight_utils.py:480] No model.safetensors.index.json found in remote. 2025-11-03T17:13:16.2804736Z (EngineCore_DP0 pid=2655) 2025-11-03T17:13:16.6641911Z Loading safetensors checkpoint shards: 0% 0/1 [00:00, 'use_cudagraph': False, 'cudagraph_num_of_warmups': 0, 'cudagraph_capture_sizes': [4], 'cudagraph_copy_inputs': False, 'full_cuda_graph': False, 'use_inductor_graph_partition': False, 'pass_config': {}, 'max_capture_size': 4, 'local_cache_dir': None} 2025-11-03T17:13:29.5771249Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:13:29.5801050Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:13:29.5807525Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:13:29.5814927Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:13:29.5822029Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:13:29.5829505Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:13:29.5831887Z (EngineCore_DP0 pid=2783) INFO 11-03 17:13:29 [parallel_state.py:1231] rank 0 in world size 1 is assigned as DP rank 0, PP rank 0, TP rank 0, EP rank 0 2025-11-03T17:13:29.8201319Z (EngineCore_DP0 pid=2783) INFO 11-03 17:13:29 [topk_topp_sampler.py:59] Using FlashInfer for top-p & top-k sampling. 2025-11-03T17:13:29.8907384Z (EngineCore_DP0 pid=2783) INFO 11-03 17:13:29 [gpu_model_runner.py:2840] Starting to load model meta-llama/Llama-3.2-1B-Instruct... 2025-11-03T17:13:30.1258808Z (EngineCore_DP0 pid=2783) INFO 11-03 17:13:30 [gpu_model_runner.py:2870] Loading model from scratch... 2025-11-03T17:13:30.1260101Z (EngineCore_DP0 pid=2783) INFO 11-03 17:13:30 [transformers.py:493] Using Transformers backend. 2025-11-03T17:13:30.1262751Z (EngineCore_DP0 pid=2783) `torch_dtype` is deprecated! Use `dtype` instead! 2025-11-03T17:13:30.3819431Z (EngineCore_DP0 pid=2783) INFO 11-03 17:13:30 [cuda.py:404] Using Flash Attention backend on V1 engine. 2025-11-03T17:13:30.3894813Z (EngineCore_DP0 pid=2783) INFO 11-03 17:13:30 [bitsandbytes_loader.py:791] Loading weights with BitsAndBytes quantization. May take a while ... 2025-11-03T17:13:30.4984020Z (EngineCore_DP0 pid=2783) INFO 11-03 17:13:30 [weight_utils.py:419] Using model weights format ['*.safetensors'] 2025-11-03T17:13:30.5591795Z (EngineCore_DP0 pid=2783) INFO 11-03 17:13:30 [weight_utils.py:480] No model.safetensors.index.json found in remote. 2025-11-03T17:13:30.5601254Z (EngineCore_DP0 pid=2783) 2025-11-03T17:13:30.9495671Z Loading safetensors checkpoint shards: 0% 0/1 [00:00, 'use_cudagraph': False, 'cudagraph_num_of_warmups': 0, 'cudagraph_capture_sizes': [4], 'cudagraph_copy_inputs': False, 'full_cuda_graph': False, 'use_inductor_graph_partition': False, 'pass_config': {}, 'max_capture_size': 4, 'local_cache_dir': None} 2025-11-03T17:13:51.5119194Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:13:51.5148636Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:13:51.5155461Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:13:51.5162845Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:13:51.5169794Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:13:51.5176727Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:13:51.5179488Z (EngineCore_DP0 pid=2965) INFO 11-03 17:13:51 [parallel_state.py:1231] rank 0 in world size 1 is assigned as DP rank 0, PP rank 0, TP rank 0, EP rank 0 2025-11-03T17:13:51.7489723Z (EngineCore_DP0 pid=2965) INFO 11-03 17:13:51 [topk_topp_sampler.py:59] Using FlashInfer for top-p & top-k sampling. 2025-11-03T17:13:51.7853458Z (EngineCore_DP0 pid=2965) INFO 11-03 17:13:51 [gpu_model_runner.py:2840] Starting to load model Qwen/Qwen3-Embedding-0.6B... 2025-11-03T17:13:52.0124648Z (EngineCore_DP0 pid=2965) INFO 11-03 17:13:52 [gpu_model_runner.py:2870] Loading model from scratch... 2025-11-03T17:13:52.0125887Z (EngineCore_DP0 pid=2965) INFO 11-03 17:13:52 [transformers.py:493] Using Transformers backend. 2025-11-03T17:13:52.0128690Z (EngineCore_DP0 pid=2965) `torch_dtype` is deprecated! Use `dtype` instead! 2025-11-03T17:13:52.2285809Z (EngineCore_DP0 pid=2965) INFO 11-03 17:13:52 [cuda.py:404] Using Flash Attention backend on V1 engine. 2025-11-03T17:13:52.3020842Z (EngineCore_DP0 pid=2965) INFO 11-03 17:13:52 [weight_utils.py:419] Using model weights format ['*.safetensors'] 2025-11-03T17:13:52.3937945Z (EngineCore_DP0 pid=2965) 2025-11-03T17:13:53.1532265Z model.safetensors: 0% 0.00/1.19G [00:00, 'use_cudagraph': False, 'cudagraph_num_of_warmups': 0, 'cudagraph_capture_sizes': [4], 'cudagraph_copy_inputs': False, 'full_cuda_graph': False, 'use_inductor_graph_partition': False, 'pass_config': {}, 'max_capture_size': 4, 'local_cache_dir': None} 2025-11-03T17:14:05.1539026Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:14:05.1568438Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:14:05.1575499Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:14:05.1583091Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:14:05.1590154Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:14:05.1597544Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:14:05.1599868Z (EngineCore_DP0 pid=3125) INFO 11-03 17:14:05 [parallel_state.py:1231] rank 0 in world size 1 is assigned as DP rank 0, PP rank 0, TP rank 0, EP rank 0 2025-11-03T17:14:05.3878885Z (EngineCore_DP0 pid=3125) INFO 11-03 17:14:05 [topk_topp_sampler.py:59] Using FlashInfer for top-p & top-k sampling. 2025-11-03T17:14:05.4244578Z (EngineCore_DP0 pid=3125) INFO 11-03 17:14:05 [gpu_model_runner.py:2840] Starting to load model meta-llama/Llama-3.2-1B-Instruct... 2025-11-03T17:14:05.6484691Z (EngineCore_DP0 pid=3125) INFO 11-03 17:14:05 [gpu_model_runner.py:2870] Loading model from scratch... 2025-11-03T17:14:05.6485534Z (EngineCore_DP0 pid=3125) INFO 11-03 17:14:05 [transformers.py:493] Using Transformers backend. 2025-11-03T17:14:05.6487803Z (EngineCore_DP0 pid=3125) `torch_dtype` is deprecated! Use `dtype` instead! 2025-11-03T17:14:05.8331469Z (EngineCore_DP0 pid=3125) INFO 11-03 17:14:05 [cuda.py:404] Using Flash Attention backend on V1 engine. 2025-11-03T17:14:05.9868065Z (EngineCore_DP0 pid=3125) INFO 11-03 17:14:05 [weight_utils.py:419] Using model weights format ['*.safetensors'] 2025-11-03T17:14:06.0512402Z (EngineCore_DP0 pid=3125) INFO 11-03 17:14:06 [weight_utils.py:480] No model.safetensors.index.json found in remote. 2025-11-03T17:14:06.0518055Z (EngineCore_DP0 pid=3125) 2025-11-03T17:14:06.3747259Z Loading safetensors checkpoint shards: 0% 0/1 [00:00:488 2025-11-03T17:14:10.3071076Z :488: DeprecationWarning: builtin type SwigPyPacked has no __module__ attribute 2025-11-03T17:14:10.3071520Z 2025-11-03T17:14:10.3071665Z :488 2025-11-03T17:14:10.3072225Z :488: DeprecationWarning: builtin type SwigPyObject has no __module__ attribute 2025-11-03T17:14:10.3072568Z 2025-11-03T17:14:10.3072828Z ../../../../../../opt/conda/envs/py_3.12/lib/python3.12/site-packages/schemathesis/generation/coverage.py:305 2025-11-03T17:14:10.3074589Z /opt/conda/envs/py_3.12/lib/python3.12/site-packages/schemathesis/generation/coverage.py:305: DeprecationWarning: jsonschema.exceptions.RefResolutionError is deprecated as of version 4.18.0. If you wish to catch potential reference resolution errors, directly catch referencing.exceptions.Unresolvable. 2025-11-03T17:14:10.3075818Z ref_error: type[Exception] = jsonschema.RefResolutionError, 2025-11-03T17:14:10.3076033Z 2025-11-03T17:14:10.3076286Z tests/models/test_transformers.py::test_models[meta-llama/Llama-3.2-1B-Instruct-transformers] 2025-11-03T17:14:10.3077187Z /opt/conda/envs/py_3.12/lib/python3.12/multiprocessing/popen_fork.py:66: DeprecationWarning: This process (pid=659) is multi-threaded, use of fork() may lead to deadlocks in the child. 2025-11-03T17:14:10.3077807Z self.pid = os.fork() 2025-11-03T17:14:10.3077939Z 2025-11-03T17:14:10.3078179Z tests/models/test_transformers.py::test_models[meta-llama/Llama-3.2-1B-Instruct-transformers] 2025-11-03T17:14:10.3078715Z /var/lib/jenkins/workspace/vllm/tests/models/test_transformers.py:54: UserWarning: Test0: 2025-11-03T17:14:10.3079331Z Matched tokens: [85, 4178, 44, 374, 6319, 311, 387, 7701] 2025-11-03T17:14:10.3080286Z ref: 'vLLM is designed to be highly parallelizable, allowing for efficient inference and serving of large-scale language models. It leverages a combination of techniques, including' {15638: -1.381659984588623, 69311: -1.381659984588623, 63174: -2.256659984588623, 34440: -2.756659984588623, 11297: -2.881659984588623} 2025-11-03T17:14:10.3082516Z test: 'vLLM is designed to be highly scalable, flexible, and adaptable to various use cases, including but not limited to, language translation, text summarization,' {69311: Logprob(logprob=-1.3481978178024292, rank=1, decoded_token=' scalable'), 15638: Logprob(logprob=-1.4731978178024292, rank=2, decoded_token=' parallel'), 63174: Logprob(logprob=-2.2231979370117188, rank=3, decoded_token=' customizable'), 34440: Logprob(logprob=-2.7231979370117188, rank=4, decoded_token=' optimized'), 11297: Logprob(logprob=-2.8481979370117188, rank=5, decoded_token=' efficient')} 2025-11-03T17:14:10.3084048Z check_logprobs_close( 2025-11-03T17:14:10.3084170Z 2025-11-03T17:14:10.3084414Z tests/models/test_transformers.py::test_models[meta-llama/Llama-3.2-1B-Instruct-transformers] 2025-11-03T17:14:10.3084955Z /var/lib/jenkins/workspace/vllm/tests/models/test_transformers.py:54: UserWarning: Test1: 2025-11-03T17:14:10.3085434Z Matched tokens: [] 2025-11-03T17:14:10.3086191Z ref: '**Answer:**\n1. **1950s:** The first AI programs were developed, including the Logical Theorist, which was a program that could reason' {334: -1.832015037536621, 8586: -1.832015037536621, 791: -2.082015037536621, 567: -2.082015037536621, 9470: -2.332015037536621} 2025-11-03T17:14:10.3088152Z test: 'Here are the major milestones in the development of artificial intelligence from 1950 to 2020:\n\n**1950s:**\n\n* Alan Turing publishes "Comput' {8586: Logprob(logprob=-1.7877479791641235, rank=1, decoded_token='Here'), 334: Logprob(logprob=-1.9127479791641235, rank=2, decoded_token='**'), 567: Logprob(logprob=-2.037747859954834, rank=3, decoded_token='##'), 791: Logprob(logprob=-2.162747859954834, rank=4, decoded_token='The'), 9470: Logprob(logprob=-2.287747859954834, rank=5, decoded_token='Art')} 2025-11-03T17:14:10.3089457Z check_logprobs_close( 2025-11-03T17:14:10.3089588Z 2025-11-03T17:14:10.3089830Z tests/models/test_transformers.py::test_models[meta-llama/Llama-3.2-1B-Instruct-transformers] 2025-11-03T17:14:10.3090371Z /var/lib/jenkins/workspace/vllm/tests/models/test_transformers.py:54: UserWarning: Test3: 2025-11-03T17:14:10.3090750Z Matched tokens: [32, 30828, 4009, 374, 264] 2025-11-03T17:14:10.3091587Z ref: 'A neural network is a complex system of interconnected nodes or "neurons" that process information. The basic components of a neural network include:\n1. **Art' {6485: -1.63894784450531, 955: -1.76394784450531, 6500: -2.0139479637145996, 55580: -2.3889479637145996, 1646: -2.7639479637145996} 2025-11-03T17:14:10.3093688Z test: 'A neural network is a type of machine learning model that is inspired by the structure and function of the human brain. It is composed of layers of interconnected nodes or' {955: Logprob(logprob=-1.7134931087493896, rank=2, decoded_token=' type'), 6485: Logprob(logprob=-1.7134931087493896, rank=1, decoded_token=' complex'), 6500: Logprob(logprob=-2.0884931087493896, rank=3, decoded_token=' computer'), 55580: Logprob(logprob=-2.3384931087493896, rank=4, decoded_token=' computational'), 1646: Logprob(logprob=-2.7134931087493896, rank=5, decoded_token=' model')} 2025-11-03T17:14:10.3095172Z check_logprobs_close( 2025-11-03T17:14:10.3095302Z 2025-11-03T17:14:10.3095534Z tests/models/test_transformers.py::test_models[meta-llama/Llama-3.2-1B-Instruct-transformers] 2025-11-03T17:14:10.3096070Z /var/lib/jenkins/workspace/vllm/tests/models/test_transformers.py:54: UserWarning: Test4: 2025-11-03T17:14:10.3096557Z Matched tokens: [2170, 279, 7160, 743, 927, 279, 3363, 11, 264, 47766, 12585, 7086, 1901, 1955, 14980] 2025-11-03T17:14:10.3097529Z ref: 'As the sun set over the city, a lone robot named Zeta stood on the rooftop of a high-rise building, gazing out at the twinkling lights' {389: -1.337048888206482, 304: -1.462048888206482, 520: -2.4620490074157715, 7636: -2.4620490074157715, 16615: -2.8370490074157715} 2025-11-03T17:14:10.3099345Z test: 'As the sun set over the city, a lone robot named Zeta stood in the empty lot where he had spent countless hours before. He had been built to' {304: Logprob(logprob=-1.4122865200042725, rank=2, decoded_token=' in'), 389: Logprob(logprob=-1.4122865200042725, rank=1, decoded_token=' on'), 520: Logprob(logprob=-2.4122865200042725, rank=3, decoded_token=' at'), 7636: Logprob(logprob=-2.5372865200042725, rank=4, decoded_token=' alone'), 16615: Logprob(logprob=-2.7872865200042725, rank=5, decoded_token=' tall')} 2025-11-03T17:14:10.3100636Z check_logprobs_close( 2025-11-03T17:14:10.3100762Z 2025-11-03T17:14:10.3100989Z tests/models/test_transformers.py::test_models[meta-llama/Llama-3.2-1B-Instruct-transformers] 2025-11-03T17:14:10.3101643Z /var/lib/jenkins/workspace/vllm/tests/models/test_transformers.py:54: UserWarning: Test5: 2025-11-03T17:14:10.3102251Z Matched tokens: [791, 20562, 12, 777, 28522, 706, 1047, 264, 28254, 5536, 389, 3728, 7100, 14726, 323, 3938, 2626, 4211, 13, 578, 28522, 706, 69627, 8312, 27271, 11, 9057, 24716] 2025-11-03T17:14:10.3103402Z ref: 'The COVID-19 pandemic has had a profound impact on global economic structures and future business models. The pandemic has disrupted supply chains, caused widespread lockdowns, and' {51235: -0.9368098974227905, 2683: -1.0618098974227905, 7100: -2.18681001663208, 61000: -3.06181001663208, 2626: -3.56181001663208} 2025-11-03T17:14:10.3105717Z test: 'The COVID-19 pandemic has had a profound impact on global economic structures and future business models. The pandemic has disrupted supply chains, caused widespread job losses, and' {2683: Logprob(logprob=-0.9869983196258545, rank=1, decoded_token=' job'), 51235: Logprob(logprob=-0.9869983196258545, rank=2, decoded_token=' lockdown'), 7100: Logprob(logprob=-2.2369983196258545, rank=3, decoded_token=' economic'), 61000: Logprob(logprob=-3.1119983196258545, rank=4, decoded_token=' closures'), 2626: Logprob(logprob=-3.4869983196258545, rank=5, decoded_token=' business')} 2025-11-03T17:14:10.3107418Z check_logprobs_close( 2025-11-03T17:14:10.3107542Z 2025-11-03T17:14:10.3107728Z tests/models/test_transformers.py::test_models[hmellor/Ilama-3.2-1B-auto] 2025-11-03T17:14:10.3108205Z /var/lib/jenkins/workspace/vllm/tests/models/test_transformers.py:54: UserWarning: Test2: 2025-11-03T17:14:10.3108693Z Matched tokens: [28474, 323, 13168, 21075, 11478, 449, 3823, 11478, 304, 3878, 315, 8863, 2038, 13] 2025-11-03T17:14:10.3109822Z ref: 'Compare and contrast artificial intelligence with human intelligence in terms of processing information. Artificial intelligence is a branch of computer science that deals with the design of intelligent machines. It' {59294: -2.893242359161377, 83017: -2.893242359161377, 763: -3.018242359161377, 66379: -3.143242359161377, 2650: -3.143242359161377} 2025-11-03T17:14:10.3112186Z test: 'Compare and contrast artificial intelligence with human intelligence in terms of processing information. Explain how the two differ.\nArtificial intelligence (AI) is a branch of computer science' {83017: Logprob(logprob=-2.7976644039154053, rank=1, decoded_token=' Explain'), 59294: Logprob(logprob=-2.9226644039154053, rank=2, decoded_token=' Artificial'), 66379: Logprob(logprob=-3.0476644039154053, rank=3, decoded_token=' Discuss'), 763: Logprob(logprob=-3.0476644039154053, rank=4, decoded_token=' In'), 2650: Logprob(logprob=-3.1726644039154053, rank=5, decoded_token=' How')} 2025-11-03T17:14:10.3113747Z check_logprobs_close( 2025-11-03T17:14:10.3113884Z 2025-11-03T17:14:10.3114199Z tests/models/test_transformers.py::test_quantization[5-32-TheBloke/TinyLlama-1.1B-Chat-v0.3-AWQ-quantization_kwargs0] 2025-11-03T17:14:10.3114810Z /var/lib/jenkins/workspace/vllm/tests/models/test_transformers.py:179: UserWarning: Test6: 2025-11-03T17:14:10.3115424Z Matched tokens: [5618, 338, 278, 16375, 26002, 310, 278, 2598, 29874, 29420, 20413, 29973, 13, 1576, 2598, 29874, 29420] 2025-11-03T17:14:10.3116947Z transformers: 'What is the cultural significance of the Mona Lisa painting?\nThe Mona Lisa painting is a famous painting by Leonardo da Vinci, which was created' {20413: Logprob(logprob=-0.8314066529273987, rank=1, decoded_token='painting'), 338: Logprob(logprob=-0.8392191529273987, rank=2, decoded_token='is'), 29892: Logprob(logprob=-3.057969093322754, rank=3, decoded_token=','), 756: Logprob(logprob=-3.761094093322754, rank=4, decoded_token='has'), 313: Logprob(logprob=-4.229844093322754, rank=5, decoded_token='(')} 2025-11-03T17:14:10.3119424Z vllm: 'What is the cultural significance of the Mona Lisa painting?\nThe Mona Lisa is a famous painting by Leonardo da Vinci, which was created for' {338: Logprob(logprob=-0.8350272178649902, rank=1, decoded_token='is'), 20413: Logprob(logprob=-0.8350272178649902, rank=2, decoded_token='painting'), 29892: Logprob(logprob=-3.0615897178649902, rank=3, decoded_token=','), 756: Logprob(logprob=-3.7647147178649902, rank=4, decoded_token='has'), 313: Logprob(logprob=-4.23346471786499, rank=5, decoded_token='(')} 2025-11-03T17:14:10.3120734Z check_logprobs_close( 2025-11-03T17:14:10.3120852Z 2025-11-03T17:14:10.3121265Z -- Docs: https://docs.pytest.org/en/stable/how-to/capture-warnings.html 2025-11-03T17:14:10.3121903Z ============ 8 passed, 4 skipped, 12 warnings in 301.67s (0:05:01) ============= 2025-11-03T17:14:10.6030286Z sys:1: DeprecationWarning: builtin type swigvarlink has no __module__ attribute 2025-11-03T17:14:11.9233346Z 2025-11-03 17:14:11,922 [INFO] cli.lib.core.vllm.lib: Finish running step: pytest -v -s models/test_transformers.py 2025-11-03T17:14:11.9233973Z 2025-11-03 17:14:11,922 [INFO] cli.lib.core.vllm.lib: Running step: pytest -v -s models/test_registry.py 2025-11-03T17:14:11.9234491Z 2025-11-03 17:14:11,922 [INFO] cli.lib.common.utils: [shell] pytest -v -s models/test_registry.py 2025-11-03T17:14:16.5363299Z INFO 11-03 17:14:16 [__init__.py:224] Automatically detected platform cuda. 2025-11-03T17:14:18.1116788Z /opt/conda/envs/py_3.12/lib/python3.12/site-packages/pytest_asyncio/plugin.py:208: PytestDeprecationWarning: The configuration option "asyncio_default_fixture_loop_scope" is unset. 2025-11-03T17:14:18.1119234Z The event loop scope for asynchronous fixtures will default to the fixture caching scope. Future versions of pytest-asyncio will default the loop scope for asynchronous fixtures to function scope. Set the default fixture loop scope explicitly in order to avoid unexpected behavior in the future. Valid fixture loop scopes are: "function", "class", "module", "package", "session" 2025-11-03T17:14:18.1120868Z 2025-11-03T17:14:18.1121408Z warnings.warn(PytestDeprecationWarning(_DEFAULT_FIXTURE_LOOP_SCOPE_UNSET)) 2025-11-03T17:14:18.1180759Z ============================= test session starts ============================== 2025-11-03T17:14:18.1181426Z platform linux -- Python 3.12.12, pytest-8.3.5, pluggy-1.5.0 -- /opt/conda/envs/py_3.12/bin/python 2025-11-03T17:14:18.1285197Z cachedir: .pytest_cache 2025-11-03T17:14:18.1285792Z hypothesis profile 'ci' -> database=None, deadline=None, print_blob=True, derandomize=True, suppress_health_check=[HealthCheck.too_slow] 2025-11-03T17:14:18.1286407Z rootdir: /var/lib/jenkins/workspace/vllm 2025-11-03T17:14:18.1286705Z configfile: pyproject.toml 2025-11-03T17:14:18.1287738Z plugins: cpp-2.3.0, flakefinder-1.1.0, rerunfailures-14.0, xdist-3.3.1, xdoctest-1.3.0, asyncio-0.24.0, cov-6.3.0, timeout-2.3.1, hydra-core-1.3.2, mock-3.14.0, schemathesis-3.39.15, buildkite-test-collector-0.1.9, subtests-0.14.1, hypothesis-6.131.0, forked-1.6.0, shard-0.1.2, anyio-4.6.2.post1 2025-11-03T17:14:18.1288806Z asyncio: mode=Mode.STRICT, default_loop_scope=None 2025-11-03T17:14:19.4432341Z collecting ... WARNING 11-03 17:14:19 [interface.py:514] Current platform cuda does not have '_pytestfixturefunction' attribute. 2025-11-03T17:14:19.7462316Z WARNING 11-03 17:14:19 [interface.py:514] Current platform cuda does not have '__test__' attribute. 2025-11-03T17:14:19.7463005Z WARNING 11-03 17:14:19 [interface.py:514] Current platform cuda does not have '__bases__' attribute. 2025-11-03T17:14:19.7463614Z WARNING 11-03 17:14:19 [interface.py:514] Current platform cuda does not have '__test__' attribute. 2025-11-03T17:14:19.7464248Z WARNING 11-03 17:14:19 [interface.py:514] Current platform cuda does not have '_schemathesis_test' attribute. 2025-11-03T17:14:19.7528740Z  2025-11-03T17:14:19.7553268Z collecting 229 items  2025-11-03T17:14:19.7554034Z collected 229 items  2025-11-03T17:14:19.7608314Z Running 229 items in this shard: tests/models/test_registry.py::test_registry_imports[ApertusForCausalLM], tests/models/test_registry.py::test_registry_imports[AquilaModel], tests/models/test_registry.py::test_registry_imports[AquilaForCausalLM], tests/models/test_registry.py::test_registry_imports[ArceeForCausalLM], tests/models/test_registry.py::test_registry_imports[ArcticForCausalLM], tests/models/test_registry.py::test_registry_imports[MiniMaxForCausalLM], tests/models/test_registry.py::test_registry_imports[MiniMaxText01ForCausalLM], tests/models/test_registry.py::test_registry_imports[MiniMaxM1ForCausalLM], tests/models/test_registry.py::test_registry_imports[BaiChuanForCausalLM], tests/models/test_registry.py::test_registry_imports[BaichuanForCausalLM], tests/models/test_registry.py::test_registry_imports[BailingMoeForCausalLM], tests/models/test_registry.py::test_registry_imports[BailingMoeV2ForCausalLM], tests/models/test_registry.py::test_registry_imports[BambaForCausalLM], tests/models/test_registry.py::test_registry_imports[BloomForCausalLM], tests/models/test_registry.py::test_registry_imports[ChatGLMModel], tests/models/test_registry.py::test_registry_imports[ChatGLMForConditionalGeneration], tests/models/test_registry.py::test_registry_imports[CohereForCausalLM], tests/models/test_registry.py::test_registry_imports[Cohere2ForCausalLM], tests/models/test_registry.py::test_registry_imports[CwmForCausalLM], tests/models/test_registry.py::test_registry_imports[DbrxForCausalLM], tests/models/test_registry.py::test_registry_imports[DeciLMForCausalLM], tests/models/test_registry.py::test_registry_imports[DeepseekForCausalLM], tests/models/test_registry.py::test_registry_imports[DeepseekV2ForCausalLM], tests/models/test_registry.py::test_registry_imports[DeepseekV3ForCausalLM], tests/models/test_registry.py::test_registry_imports[DeepseekV32ForCausalLM], tests/models/test_registry.py::test_registry_imports[Dots1ForCausalLM], tests/models/test_registry.py::test_registry_imports[Ernie4_5ForCausalLM], tests/models/test_registry.py::test_registry_imports[Ernie4_5_MoeForCausalLM], tests/models/test_registry.py::test_registry_imports[ExaoneForCausalLM], tests/models/test_registry.py::test_registry_imports[Exaone4ForCausalLM], tests/models/test_registry.py::test_registry_imports[FalconForCausalLM], tests/models/test_registry.py::test_registry_imports[Fairseq2LlamaForCausalLM], tests/models/test_registry.py::test_registry_imports[GemmaForCausalLM], tests/models/test_registry.py::test_registry_imports[Gemma2ForCausalLM], tests/models/test_registry.py::test_registry_imports[Gemma3ForCausalLM], tests/models/test_registry.py::test_registry_imports[Gemma3nForCausalLM], tests/models/test_registry.py::test_registry_imports[Qwen3NextForCausalLM], tests/models/test_registry.py::test_registry_imports[GlmForCausalLM], tests/models/test_registry.py::test_registry_imports[Glm4ForCausalLM], tests/models/test_registry.py::test_registry_imports[Glm4MoeForCausalLM], tests/models/test_registry.py::test_registry_imports[GptOssForCausalLM], tests/models/test_registry.py::test_registry_imports[GPT2LMHeadModel], tests/models/test_registry.py::test_registry_imports[GPTBigCodeForCausalLM], tests/models/test_registry.py::test_registry_imports[GPTJForCausalLM], tests/models/test_registry.py::test_registry_imports[GPTNeoXForCausalLM], tests/models/test_registry.py::test_registry_imports[GraniteForCausalLM], tests/models/test_registry.py::test_registry_imports[GraniteMoeForCausalLM], tests/models/test_registry.py::test_registry_imports[GraniteMoeHybridForCausalLM], tests/models/test_registry.py::test_registry_imports[GraniteMoeSharedForCausalLM], tests/models/test_registry.py::test_registry_imports[GritLM], tests/models/test_registry.py::test_registry_imports[Grok1ModelForCausalLM], tests/models/test_registry.py::test_registry_imports[HunYuanMoEV1ForCausalLM], tests/models/test_registry.py::test_registry_imports[HunYuanDenseV1ForCausalLM], tests/models/test_registry.py::test_registry_imports[HCXVisionForCausalLM], tests/models/test_registry.py::test_registry_imports[InternLMForCausalLM], tests/models/test_registry.py::test_registry_imports[InternLM2ForCausalLM], tests/models/test_registry.py::test_registry_imports[InternLM2VEForCausalLM], tests/models/test_registry.py::test_registry_imports[InternLM3ForCausalLM], tests/models/test_registry.py::test_registry_imports[JAISLMHeadModel], tests/models/test_registry.py::test_registry_imports[JambaForCausalLM], tests/models/test_registry.py::test_registry_imports[Lfm2ForCausalLM], tests/models/test_registry.py::test_registry_imports[Lfm2MoeForCausalLM], tests/models/test_registry.py::test_registry_imports[LlamaForCausalLM], tests/models/test_registry.py::test_registry_imports[Llama4ForCausalLM], tests/models/test_registry.py::test_registry_imports[LLaMAForCausalLM], tests/models/test_registry.py::test_registry_imports[LongcatFlashForCausalLM], tests/models/test_registry.py::test_registry_imports[MambaForCausalLM], tests/models/test_registry.py::test_registry_imports[FalconMambaForCausalLM], tests/models/test_registry.py::test_registry_imports[FalconH1ForCausalLM], tests/models/test_registry.py::test_registry_imports[Mamba2ForCausalLM], tests/models/test_registry.py::test_registry_imports[MiniCPMForCausalLM], tests/models/test_registry.py::test_registry_imports[MiniCPM3ForCausalLM], tests/models/test_registry.py::test_registry_imports[MistralForCausalLM], tests/models/test_registry.py::test_registry_imports[MixtralForCausalLM], tests/models/test_registry.py::test_registry_imports[MptForCausalLM], tests/models/test_registry.py::test_registry_imports[MPTForCausalLM], tests/models/test_registry.py::test_registry_imports[MiMoForCausalLM], tests/models/test_registry.py::test_registry_imports[NemotronForCausalLM], tests/models/test_registry.py::test_registry_imports[NemotronHForCausalLM], tests/models/test_registry.py::test_registry_imports[OlmoForCausalLM], tests/models/test_registry.py::test_registry_imports[Olmo2ForCausalLM], tests/models/test_registry.py::test_registry_imports[Olmo3ForCausalLM], tests/models/test_registry.py::test_registry_imports[OlmoeForCausalLM], tests/models/test_registry.py::test_registry_imports[OPTForCausalLM], tests/models/test_registry.py::test_registry_imports[OrionForCausalLM], tests/models/test_registry.py::test_registry_imports[PersimmonForCausalLM], tests/models/test_registry.py::test_registry_imports[PhiForCausalLM], tests/models/test_registry.py::test_registry_imports[Phi3ForCausalLM], tests/models/test_registry.py::test_registry_imports[PhiMoEForCausalLM], tests/models/test_registry.py::test_registry_imports[Plamo2ForCausalLM], tests/models/test_registry.py::test_registry_imports[QWenLMHeadModel], tests/models/test_registry.py::test_registry_imports[Qwen2ForCausalLM], tests/models/test_registry.py::test_registry_imports[Qwen2MoeForCausalLM], tests/models/test_registry.py::test_registry_imports[Qwen3ForCausalLM], tests/models/test_registry.py::test_registry_imports[Qwen3MoeForCausalLM], tests/models/test_registry.py::test_registry_imports[RWForCausalLM], tests/models/test_registry.py::test_registry_imports[SeedOssForCausalLM], tests/models/test_registry.py::test_registry_imports[Step3TextForCausalLM], tests/models/test_registry.py::test_registry_imports[StableLMEpochForCausalLM], tests/models/test_registry.py::test_registry_imports[StableLmForCausalLM], tests/models/test_registry.py::test_registry_imports[Starcoder2ForCausalLM], tests/models/test_registry.py::test_registry_imports[SolarForCausalLM], tests/models/test_registry.py::test_registry_imports[TeleChat2ForCausalLM], tests/models/test_registry.py::test_registry_imports[TeleFLMForCausalLM], tests/models/test_registry.py::test_registry_imports[XverseForCausalLM], tests/models/test_registry.py::test_registry_imports[Zamba2ForCausalLM], tests/models/test_registry.py::test_registry_imports[BertModel], tests/models/test_registry.py::test_registry_imports[Gemma2Model], tests/models/test_registry.py::test_registry_imports[Gemma3TextModel], tests/models/test_registry.py::test_registry_imports[GPT2ForSequenceClassification], tests/models/test_registry.py::test_registry_imports[GteModel], tests/models/test_registry.py::test_registry_imports[GteNewModel], tests/models/test_registry.py::test_registry_imports[InternLM2ForRewardModel], tests/models/test_registry.py::test_registry_imports[JambaForSequenceClassification], tests/models/test_registry.py::test_registry_imports[LlamaModel], tests/models/test_registry.py::test_registry_imports[MistralModel], tests/models/test_registry.py::test_registry_imports[ModernBertModel], tests/models/test_registry.py::test_registry_imports[NomicBertModel], tests/models/test_registry.py::test_registry_imports[Qwen2Model], tests/models/test_registry.py::test_registry_imports[Qwen2ForRewardModel], tests/models/test_registry.py::test_registry_imports[Qwen2ForProcessRewardModel], tests/models/test_registry.py::test_registry_imports[RobertaForMaskedLM], tests/models/test_registry.py::test_registry_imports[RobertaModel], tests/models/test_registry.py::test_registry_imports[XLMRobertaModel], tests/models/test_registry.py::test_registry_imports[CLIPModel], tests/models/test_registry.py::test_registry_imports[LlavaNextForConditionalGeneration], tests/models/test_registry.py::test_registry_imports[Phi3VForCausalLM], tests/models/test_registry.py::test_registry_imports[Qwen2VLForConditionalGeneration], tests/models/test_registry.py::test_registry_imports[PrithviGeoSpatialMAE], tests/models/test_registry.py::test_registry_imports[Terratorch], tests/models/test_registry.py::test_registry_imports[BertForSequenceClassification], tests/models/test_registry.py::test_registry_imports[BertForTokenClassification], tests/models/test_registry.py::test_registry_imports[GteNewForSequenceClassification], tests/models/test_registry.py::test_registry_imports[ModernBertForSequenceClassification], tests/models/test_registry.py::test_registry_imports[ModernBertForTokenClassification], tests/models/test_registry.py::test_registry_imports[RobertaForSequenceClassification], tests/models/test_registry.py::test_registry_imports[XLMRobertaForSequenceClassification], tests/models/test_registry.py::test_registry_imports[JinaVLForRanking], tests/models/test_registry.py::test_registry_imports[AriaForConditionalGeneration], tests/models/test_registry.py::test_registry_imports[AyaVisionForConditionalGeneration], tests/models/test_registry.py::test_registry_imports[Blip2ForConditionalGeneration], tests/models/test_registry.py::test_registry_imports[ChameleonForConditionalGeneration], tests/models/test_registry.py::test_registry_imports[Cohere2VisionForConditionalGeneration], tests/models/test_registry.py::test_registry_imports[DeepseekVLV2ForCausalLM], tests/models/test_registry.py::test_registry_imports[DotsOCRForCausalLM], tests/models/test_registry.py::test_registry_imports[Ernie4_5_VLMoeForConditionalGeneration], tests/models/test_registry.py::test_registry_imports[FuyuForCausalLM], tests/models/test_registry.py::test_registry_imports[Gemma3ForConditionalGeneration], tests/models/test_registry.py::test_registry_imports[Gemma3nForConditionalGeneration], tests/models/test_registry.py::test_registry_imports[GLM4VForCausalLM], tests/models/test_registry.py::test_registry_imports[Glm4vForConditionalGeneration], tests/models/test_registry.py::test_registry_imports[Glm4vMoeForConditionalGeneration], tests/models/test_registry.py::test_registry_imports[GraniteSpeechForConditionalGeneration], tests/models/test_registry.py::test_registry_imports[H2OVLChatModel], tests/models/test_registry.py::test_registry_imports[InternVLChatModel], tests/models/test_registry.py::test_registry_imports[NemotronH_Nano_VL_V2], tests/models/test_registry.py::test_registry_imports[InternS1ForConditionalGeneration], tests/models/test_registry.py::test_registry_imports[InternVLForConditionalGeneration], tests/models/test_registry.py::test_registry_imports[Idefics3ForConditionalGeneration], tests/models/test_registry.py::test_registry_imports[SmolVLMForConditionalGeneration], tests/models/test_registry.py::test_registry_imports[KeyeForConditionalGeneration], tests/models/test_registry.py::test_registry_imports[KeyeVL1_5ForConditionalGeneration], tests/models/test_registry.py::test_registry_imports[RForConditionalGeneration], tests/models/test_registry.py::test_registry_imports[KimiVLForConditionalGeneration], tests/models/test_registry.py::test_registry_imports[Llama_Nemotron_Nano_VL], tests/models/test_registry.py::test_registry_imports[Llama4ForConditionalGeneration], tests/models/test_registry.py::test_registry_imports[LlavaForConditionalGeneration], tests/models/test_registry.py::test_registry_imports[LlavaNextVideoForConditionalGeneration], tests/models/test_registry.py::test_registry_imports[LlavaOnevisionForConditionalGeneration], tests/models/test_registry.py::test_registry_imports[MantisForConditionalGeneration], tests/models/test_registry.py::test_registry_imports[MiDashengLMModel], tests/models/test_registry.py::test_registry_imports[MiniMaxVL01ForConditionalGeneration], tests/models/test_registry.py::test_registry_imports[MiniCPMO], tests/models/test_registry.py::test_registry_imports[MiniCPMV], tests/models/test_registry.py::test_registry_imports[Mistral3ForConditionalGeneration], tests/models/test_registry.py::test_registry_imports[MolmoForCausalLM], tests/models/test_registry.py::test_registry_imports[NVLM_D], tests/models/test_registry.py::test_registry_imports[Ovis], tests/models/test_registry.py::test_registry_imports[Ovis2_5], tests/models/test_registry.py::test_registry_imports[PaliGemmaForConditionalGeneration], tests/models/test_registry.py::test_registry_imports[Phi4MMForCausalLM], tests/models/test_registry.py::test_registry_imports[Phi4MultimodalForCausalLM], tests/models/test_registry.py::test_registry_imports[PixtralForConditionalGeneration], tests/models/test_registry.py::test_registry_imports[QwenVLForConditionalGeneration], tests/models/test_registry.py::test_registry_imports[Qwen2_5_VLForConditionalGeneration], tests/models/test_registry.py::test_registry_imports[Qwen2AudioForConditionalGeneration], tests/models/test_registry.py::test_registry_imports[Qwen2_5OmniModel], tests/models/test_registry.py::test_registry_imports[Qwen2_5OmniForConditionalGeneration], tests/models/test_registry.py::test_registry_imports[Qwen3VLForConditionalGeneration], tests/models/test_registry.py::test_registry_imports[Qwen3VLMoeForConditionalGeneration], tests/models/test_registry.py::test_registry_imports[SkyworkR1VChatModel], tests/models/test_registry.py::test_registry_imports[Step3VLForConditionalGeneration], tests/models/test_registry.py::test_registry_imports[TarsierForConditionalGeneration], tests/models/test_registry.py::test_registry_imports[Tarsier2ForConditionalGeneration], tests/models/test_registry.py::test_registry_imports[UltravoxModel], tests/models/test_registry.py::test_registry_imports[VoxtralForConditionalGeneration], tests/models/test_registry.py::test_registry_imports[WhisperForConditionalGeneration], tests/models/test_registry.py::test_registry_imports[MiMoMTPModel], tests/models/test_registry.py::test_registry_imports[EagleLlamaForCausalLM], tests/models/test_registry.py::test_registry_imports[EagleLlama4ForCausalLM], tests/models/test_registry.py::test_registry_imports[EagleMiniCPMForCausalLM], tests/models/test_registry.py::test_registry_imports[Eagle3LlamaForCausalLM], tests/models/test_registry.py::test_registry_imports[LlamaForCausalLMEagle3], tests/models/test_registry.py::test_registry_imports[Eagle3Qwen2_5vlForCausalLM], tests/models/test_registry.py::test_registry_imports[EagleDeepSeekMTPModel], tests/models/test_registry.py::test_registry_imports[DeepSeekMTPModel], tests/models/test_registry.py::test_registry_imports[ErnieMTPModel], tests/models/test_registry.py::test_registry_imports[LongCatFlashMTPModel], tests/models/test_registry.py::test_registry_imports[Glm4MoeMTPModel], tests/models/test_registry.py::test_registry_imports[MedusaModel], tests/models/test_registry.py::test_registry_imports[Qwen3NextMTP], tests/models/test_registry.py::test_registry_imports[SmolLM3ForCausalLM], tests/models/test_registry.py::test_registry_imports[Emu3ForConditionalGeneration], tests/models/test_registry.py::test_registry_imports[TransformersForCausalLM], tests/models/test_registry.py::test_registry_imports[TransformersForMultimodalLM], tests/models/test_registry.py::test_registry_imports[TransformersMoEForCausalLM], tests/models/test_registry.py::test_registry_imports[TransformersMoEForMultimodalLM], tests/models/test_registry.py::test_registry_imports[TransformersEmbeddingModel], tests/models/test_registry.py::test_registry_imports[TransformersForSequenceClassification], tests/models/test_registry.py::test_registry_imports[TransformersMoEForSequenceClassification], tests/models/test_registry.py::test_registry_imports[TransformersMoEEmbeddingModel], tests/models/test_registry.py::test_registry_model_property[LlamaForCausalLM-False-False-False], tests/models/test_registry.py::test_registry_model_property[LlavaForConditionalGeneration-True-True-False], tests/models/test_registry.py::test_registry_model_property[BertForSequenceClassification-False-False-True], tests/models/test_registry.py::test_registry_model_property[RobertaForSequenceClassification-False-False-True], tests/models/test_registry.py::test_registry_model_property[XLMRobertaForSequenceClassification-False-False-True], tests/models/test_registry.py::test_registry_is_pp[DeepseekV2ForCausalLM-True-False], tests/models/test_registry.py::test_registry_is_pp[Qwen2VLForConditionalGeneration-True-True], tests/models/test_registry.py::test_hf_registry_coverage 2025-11-03T17:14:19.7659564Z 2025-11-03T17:14:20.0290243Z models/test_registry.py::test_registry_imports[ApertusForCausalLM] PASSED 2025-11-03T17:14:20.2625168Z models/test_registry.py::test_registry_imports[AquilaModel] PASSED 2025-11-03T17:14:20.4922677Z models/test_registry.py::test_registry_imports[AquilaForCausalLM] PASSED 2025-11-03T17:14:20.7246046Z models/test_registry.py::test_registry_imports[ArceeForCausalLM] PASSED 2025-11-03T17:14:20.9897771Z models/test_registry.py::test_registry_imports[ArcticForCausalLM] PASSED 2025-11-03T17:14:21.2351320Z models/test_registry.py::test_registry_imports[MiniMaxForCausalLM] PASSED 2025-11-03T17:14:21.4646524Z models/test_registry.py::test_registry_imports[MiniMaxText01ForCausalLM] PASSED 2025-11-03T17:14:21.6988783Z models/test_registry.py::test_registry_imports[MiniMaxM1ForCausalLM] PASSED 2025-11-03T17:14:21.9316264Z models/test_registry.py::test_registry_imports[BaiChuanForCausalLM] PASSED 2025-11-03T17:14:22.1626813Z models/test_registry.py::test_registry_imports[BaichuanForCausalLM] PASSED 2025-11-03T17:14:22.3962906Z models/test_registry.py::test_registry_imports[BailingMoeForCausalLM] PASSED 2025-11-03T17:14:22.6271865Z models/test_registry.py::test_registry_imports[BailingMoeV2ForCausalLM] PASSED 2025-11-03T17:14:22.9010592Z models/test_registry.py::test_registry_imports[BambaForCausalLM] PASSED 2025-11-03T17:14:23.1446750Z models/test_registry.py::test_registry_imports[BloomForCausalLM] PASSED 2025-11-03T17:14:23.3751597Z models/test_registry.py::test_registry_imports[ChatGLMModel] SKIPPED 2025-11-03T17:14:23.6083457Z models/test_registry.py::test_registry_imports[ChatGLMForConditionalGeneration] PASSED 2025-11-03T17:14:23.8509576Z models/test_registry.py::test_registry_imports[CohereForCausalLM] PASSED 2025-11-03T17:14:24.0830236Z models/test_registry.py::test_registry_imports[Cohere2ForCausalLM] PASSED 2025-11-03T17:14:24.3151807Z models/test_registry.py::test_registry_imports[CwmForCausalLM] PASSED 2025-11-03T17:14:24.5519835Z models/test_registry.py::test_registry_imports[DbrxForCausalLM] PASSED 2025-11-03T17:14:24.7900229Z models/test_registry.py::test_registry_imports[DeciLMForCausalLM] PASSED 2025-11-03T17:14:25.0300937Z models/test_registry.py::test_registry_imports[DeepseekForCausalLM] PASSED 2025-11-03T17:14:25.2734968Z models/test_registry.py::test_registry_imports[DeepseekV2ForCausalLM] PASSED 2025-11-03T17:14:25.5089884Z models/test_registry.py::test_registry_imports[DeepseekV3ForCausalLM] PASSED 2025-11-03T17:14:25.7431198Z models/test_registry.py::test_registry_imports[DeepseekV32ForCausalLM] PASSED 2025-11-03T17:14:25.9987571Z models/test_registry.py::test_registry_imports[Dots1ForCausalLM] PASSED 2025-11-03T17:14:26.2444024Z models/test_registry.py::test_registry_imports[Ernie4_5ForCausalLM] PASSED 2025-11-03T17:14:26.4932639Z models/test_registry.py::test_registry_imports[Ernie4_5_MoeForCausalLM] PASSED 2025-11-03T17:14:26.7366390Z models/test_registry.py::test_registry_imports[ExaoneForCausalLM] PASSED 2025-11-03T17:14:26.9892549Z models/test_registry.py::test_registry_imports[Exaone4ForCausalLM] PASSED 2025-11-03T17:14:27.2436145Z models/test_registry.py::test_registry_imports[FalconForCausalLM] PASSED 2025-11-03T17:14:27.4813699Z models/test_registry.py::test_registry_imports[Fairseq2LlamaForCausalLM] PASSED 2025-11-03T17:14:27.7351349Z models/test_registry.py::test_registry_imports[GemmaForCausalLM] PASSED 2025-11-03T17:14:27.9773890Z models/test_registry.py::test_registry_imports[Gemma2ForCausalLM] PASSED 2025-11-03T17:14:28.2336726Z models/test_registry.py::test_registry_imports[Gemma3ForCausalLM] PASSED 2025-11-03T17:14:30.2326102Z models/test_registry.py::test_registry_imports[Gemma3nForCausalLM] PASSED 2025-11-03T17:14:30.4790196Z models/test_registry.py::test_registry_imports[Qwen3NextForCausalLM] SKIPPED 2025-11-03T17:14:30.7273556Z models/test_registry.py::test_registry_imports[GlmForCausalLM] PASSED 2025-11-03T17:14:30.9880393Z models/test_registry.py::test_registry_imports[Glm4ForCausalLM] PASSED 2025-11-03T17:14:31.2496559Z models/test_registry.py::test_registry_imports[Glm4MoeForCausalLM] PASSED 2025-11-03T17:14:31.5137960Z models/test_registry.py::test_registry_imports[GptOssForCausalLM] PASSED 2025-11-03T17:14:31.7836274Z models/test_registry.py::test_registry_imports[GPT2LMHeadModel] PASSED 2025-11-03T17:14:32.0435559Z models/test_registry.py::test_registry_imports[GPTBigCodeForCausalLM] PASSED 2025-11-03T17:14:32.3070509Z models/test_registry.py::test_registry_imports[GPTJForCausalLM] PASSED 2025-11-03T17:14:32.5813756Z models/test_registry.py::test_registry_imports[GPTNeoXForCausalLM] PASSED 2025-11-03T17:14:32.8440220Z models/test_registry.py::test_registry_imports[GraniteForCausalLM] PASSED 2025-11-03T17:14:33.1159361Z models/test_registry.py::test_registry_imports[GraniteMoeForCausalLM] PASSED 2025-11-03T17:14:33.3926369Z models/test_registry.py::test_registry_imports[GraniteMoeHybridForCausalLM] PASSED 2025-11-03T17:14:33.6393961Z models/test_registry.py::test_registry_imports[GraniteMoeSharedForCausalLM] PASSED 2025-11-03T17:14:33.9010847Z models/test_registry.py::test_registry_imports[GritLM] PASSED 2025-11-03T17:14:34.1733826Z models/test_registry.py::test_registry_imports[Grok1ModelForCausalLM] PASSED 2025-11-03T17:14:34.4348887Z models/test_registry.py::test_registry_imports[HunYuanMoEV1ForCausalLM] PASSED 2025-11-03T17:14:34.6824298Z models/test_registry.py::test_registry_imports[HunYuanDenseV1ForCausalLM] PASSED 2025-11-03T17:14:35.0024627Z models/test_registry.py::test_registry_imports[HCXVisionForCausalLM] PASSED 2025-11-03T17:14:35.2517544Z models/test_registry.py::test_registry_imports[InternLMForCausalLM] PASSED 2025-11-03T17:14:35.5174570Z models/test_registry.py::test_registry_imports[InternLM2ForCausalLM] PASSED 2025-11-03T17:14:35.7824728Z models/test_registry.py::test_registry_imports[InternLM2VEForCausalLM] PASSED 2025-11-03T17:14:36.0325262Z models/test_registry.py::test_registry_imports[InternLM3ForCausalLM] PASSED 2025-11-03T17:14:36.2949505Z models/test_registry.py::test_registry_imports[JAISLMHeadModel] PASSED 2025-11-03T17:14:36.5674809Z models/test_registry.py::test_registry_imports[JambaForCausalLM] PASSED 2025-11-03T17:14:36.8342269Z models/test_registry.py::test_registry_imports[Lfm2ForCausalLM] PASSED 2025-11-03T17:14:37.0817809Z models/test_registry.py::test_registry_imports[Lfm2MoeForCausalLM] SKIPPED 2025-11-03T17:14:37.3303454Z models/test_registry.py::test_registry_imports[LlamaForCausalLM] PASSED 2025-11-03T17:14:37.5952439Z models/test_registry.py::test_registry_imports[Llama4ForCausalLM] PASSED 2025-11-03T17:14:37.8428408Z models/test_registry.py::test_registry_imports[LLaMAForCausalLM] PASSED 2025-11-03T17:14:38.1061754Z models/test_registry.py::test_registry_imports[LongcatFlashForCausalLM] PASSED 2025-11-03T17:14:38.3703096Z models/test_registry.py::test_registry_imports[MambaForCausalLM] PASSED 2025-11-03T17:14:38.6185595Z models/test_registry.py::test_registry_imports[FalconMambaForCausalLM] PASSED 2025-11-03T17:14:38.8866723Z models/test_registry.py::test_registry_imports[FalconH1ForCausalLM] PASSED 2025-11-03T17:14:39.1478567Z models/test_registry.py::test_registry_imports[Mamba2ForCausalLM] PASSED 2025-11-03T17:14:39.4062084Z models/test_registry.py::test_registry_imports[MiniCPMForCausalLM] PASSED 2025-11-03T17:14:39.6684882Z models/test_registry.py::test_registry_imports[MiniCPM3ForCausalLM] PASSED 2025-11-03T17:14:39.9175925Z models/test_registry.py::test_registry_imports[MistralForCausalLM] PASSED 2025-11-03T17:14:40.1855207Z models/test_registry.py::test_registry_imports[MixtralForCausalLM] PASSED 2025-11-03T17:14:40.4515448Z models/test_registry.py::test_registry_imports[MptForCausalLM] PASSED 2025-11-03T17:14:40.7012244Z models/test_registry.py::test_registry_imports[MPTForCausalLM] PASSED 2025-11-03T17:14:40.9683572Z models/test_registry.py::test_registry_imports[MiMoForCausalLM] PASSED 2025-11-03T17:14:41.2265001Z models/test_registry.py::test_registry_imports[NemotronForCausalLM] PASSED 2025-11-03T17:14:41.4928510Z models/test_registry.py::test_registry_imports[NemotronHForCausalLM] PASSED 2025-11-03T17:14:41.7550587Z models/test_registry.py::test_registry_imports[OlmoForCausalLM] PASSED 2025-11-03T17:14:42.0228498Z models/test_registry.py::test_registry_imports[Olmo2ForCausalLM] PASSED 2025-11-03T17:14:42.2711110Z models/test_registry.py::test_registry_imports[Olmo3ForCausalLM] PASSED 2025-11-03T17:14:42.5240960Z models/test_registry.py::test_registry_imports[OlmoeForCausalLM] PASSED 2025-11-03T17:14:42.7763989Z models/test_registry.py::test_registry_imports[OPTForCausalLM] PASSED 2025-11-03T17:14:43.0255659Z models/test_registry.py::test_registry_imports[OrionForCausalLM] PASSED 2025-11-03T17:14:43.2768650Z models/test_registry.py::test_registry_imports[PersimmonForCausalLM] PASSED 2025-11-03T17:14:43.5302146Z models/test_registry.py::test_registry_imports[PhiForCausalLM] PASSED 2025-11-03T17:14:43.7793537Z models/test_registry.py::test_registry_imports[Phi3ForCausalLM] PASSED 2025-11-03T17:14:44.0298665Z models/test_registry.py::test_registry_imports[PhiMoEForCausalLM] PASSED 2025-11-03T17:14:44.2782581Z models/test_registry.py::test_registry_imports[Plamo2ForCausalLM] SKIPPED 2025-11-03T17:14:44.5307439Z models/test_registry.py::test_registry_imports[QWenLMHeadModel] SKIPPED 2025-11-03T17:14:44.7796250Z models/test_registry.py::test_registry_imports[Qwen2ForCausalLM] PASSED 2025-11-03T17:14:45.0326795Z models/test_registry.py::test_registry_imports[Qwen2MoeForCausalLM] PASSED 2025-11-03T17:14:45.2837719Z models/test_registry.py::test_registry_imports[Qwen3ForCausalLM] PASSED 2025-11-03T17:14:45.5379561Z models/test_registry.py::test_registry_imports[Qwen3MoeForCausalLM] PASSED 2025-11-03T17:14:45.7873513Z models/test_registry.py::test_registry_imports[RWForCausalLM] PASSED 2025-11-03T17:14:46.0378527Z models/test_registry.py::test_registry_imports[SeedOssForCausalLM] PASSED 2025-11-03T17:14:46.2878336Z models/test_registry.py::test_registry_imports[Step3TextForCausalLM] PASSED 2025-11-03T17:14:46.5446269Z models/test_registry.py::test_registry_imports[StableLMEpochForCausalLM] PASSED 2025-11-03T17:14:46.7939829Z models/test_registry.py::test_registry_imports[StableLmForCausalLM] PASSED 2025-11-03T17:14:47.0467593Z models/test_registry.py::test_registry_imports[Starcoder2ForCausalLM] PASSED 2025-11-03T17:14:47.2972497Z models/test_registry.py::test_registry_imports[SolarForCausalLM] PASSED 2025-11-03T17:14:47.5498317Z models/test_registry.py::test_registry_imports[TeleChat2ForCausalLM] PASSED 2025-11-03T17:14:47.8008424Z models/test_registry.py::test_registry_imports[TeleFLMForCausalLM] PASSED 2025-11-03T17:14:48.0501745Z models/test_registry.py::test_registry_imports[XverseForCausalLM] PASSED 2025-11-03T17:14:48.3041846Z models/test_registry.py::test_registry_imports[Zamba2ForCausalLM] PASSED 2025-11-03T17:14:48.5613067Z models/test_registry.py::test_registry_imports[BertModel] PASSED 2025-11-03T17:14:48.8121871Z models/test_registry.py::test_registry_imports[Gemma2Model] PASSED 2025-11-03T17:14:49.0624007Z models/test_registry.py::test_registry_imports[Gemma3TextModel] PASSED 2025-11-03T17:14:49.0630374Z models/test_registry.py::test_registry_imports[GPT2ForSequenceClassification] WARNING 11-03 17:14:49 [interfaces_base.py:74] The model () is missing the `get_input_embeddings` method. 2025-11-03T17:14:49.0633924Z WARNING 11-03 17:14:49 [interfaces_base.py:74] The model () is missing the `get_input_embeddings` method. 2025-11-03T17:14:49.0634954Z WARNING 11-03 17:14:49 [interfaces_base.py:74] The model () is missing the `get_input_embeddings` method. 2025-11-03T17:14:49.3124801Z PASSED 2025-11-03T17:14:49.5641305Z models/test_registry.py::test_registry_imports[GteModel] PASSED 2025-11-03T17:14:49.8148677Z models/test_registry.py::test_registry_imports[GteNewModel] PASSED 2025-11-03T17:14:50.0640101Z models/test_registry.py::test_registry_imports[InternLM2ForRewardModel] PASSED 2025-11-03T17:14:50.3134404Z models/test_registry.py::test_registry_imports[JambaForSequenceClassification] PASSED 2025-11-03T17:14:50.5626380Z models/test_registry.py::test_registry_imports[LlamaModel] PASSED 2025-11-03T17:14:50.8123715Z models/test_registry.py::test_registry_imports[MistralModel] PASSED 2025-11-03T17:14:51.0666440Z models/test_registry.py::test_registry_imports[ModernBertModel] PASSED 2025-11-03T17:14:51.3156161Z models/test_registry.py::test_registry_imports[NomicBertModel] PASSED 2025-11-03T17:14:51.5651499Z models/test_registry.py::test_registry_imports[Qwen2Model] PASSED 2025-11-03T17:14:51.8150235Z models/test_registry.py::test_registry_imports[Qwen2ForRewardModel] SKIPPED 2025-11-03T17:14:52.0650629Z models/test_registry.py::test_registry_imports[Qwen2ForProcessRewardModel] SKIPPED 2025-11-03T17:14:52.3263721Z models/test_registry.py::test_registry_imports[RobertaForMaskedLM] PASSED 2025-11-03T17:14:52.5787506Z models/test_registry.py::test_registry_imports[RobertaModel] PASSED 2025-11-03T17:14:52.8284724Z models/test_registry.py::test_registry_imports[XLMRobertaModel] PASSED 2025-11-03T17:14:53.0778257Z models/test_registry.py::test_registry_imports[CLIPModel] PASSED 2025-11-03T17:14:53.9387951Z models/test_registry.py::test_registry_imports[LlavaNextForConditionalGeneration] PASSED 2025-11-03T17:14:54.1959242Z models/test_registry.py::test_registry_imports[Phi3VForCausalLM] SKIPPED 2025-11-03T17:14:54.4669344Z models/test_registry.py::test_registry_imports[Qwen2VLForConditionalGeneration] PASSED 2025-11-03T17:15:01.8373645Z models/test_registry.py::test_registry_imports[PrithviGeoSpatialMAE] PASSED 2025-11-03T17:15:02.2335019Z models/test_registry.py::test_registry_imports[Terratorch] PASSED 2025-11-03T17:15:02.6246000Z models/test_registry.py::test_registry_imports[BertForSequenceClassification] PASSED 2025-11-03T17:15:03.0167852Z models/test_registry.py::test_registry_imports[BertForTokenClassification] PASSED 2025-11-03T17:15:03.4029060Z models/test_registry.py::test_registry_imports[GteNewForSequenceClassification] PASSED 2025-11-03T17:15:03.7892965Z models/test_registry.py::test_registry_imports[ModernBertForSequenceClassification] PASSED 2025-11-03T17:15:04.1764083Z models/test_registry.py::test_registry_imports[ModernBertForTokenClassification] PASSED 2025-11-03T17:15:04.5663282Z models/test_registry.py::test_registry_imports[RobertaForSequenceClassification] PASSED 2025-11-03T17:15:04.9546139Z models/test_registry.py::test_registry_imports[XLMRobertaForSequenceClassification] PASSED 2025-11-03T17:15:05.3441965Z models/test_registry.py::test_registry_imports[JinaVLForRanking] PASSED 2025-11-03T17:15:05.7582630Z models/test_registry.py::test_registry_imports[AriaForConditionalGeneration] PASSED 2025-11-03T17:15:06.1611398Z models/test_registry.py::test_registry_imports[AyaVisionForConditionalGeneration] PASSED 2025-11-03T17:15:06.5657496Z models/test_registry.py::test_registry_imports[Blip2ForConditionalGeneration] PASSED 2025-11-03T17:15:06.9624571Z models/test_registry.py::test_registry_imports[ChameleonForConditionalGeneration] PASSED 2025-11-03T17:15:07.3654381Z models/test_registry.py::test_registry_imports[Cohere2VisionForConditionalGeneration] PASSED 2025-11-03T17:15:07.7556666Z models/test_registry.py::test_registry_imports[DeepseekVLV2ForCausalLM] SKIPPED 2025-11-03T17:15:08.1593277Z models/test_registry.py::test_registry_imports[DotsOCRForCausalLM] PASSED 2025-11-03T17:15:08.5574206Z models/test_registry.py::test_registry_imports[Ernie4_5_VLMoeForConditionalGeneration] PASSED 2025-11-03T17:15:08.9637963Z models/test_registry.py::test_registry_imports[FuyuForCausalLM] PASSED 2025-11-03T17:15:09.3676314Z models/test_registry.py::test_registry_imports[Gemma3ForConditionalGeneration] PASSED 2025-11-03T17:15:09.8036727Z models/test_registry.py::test_registry_imports[Gemma3nForConditionalGeneration] PASSED 2025-11-03T17:15:10.1990120Z models/test_registry.py::test_registry_imports[GLM4VForCausalLM] PASSED 2025-11-03T17:15:10.6067405Z models/test_registry.py::test_registry_imports[Glm4vForConditionalGeneration] PASSED 2025-11-03T17:15:10.9981593Z models/test_registry.py::test_registry_imports[Glm4vMoeForConditionalGeneration] PASSED 2025-11-03T17:15:11.3997122Z models/test_registry.py::test_registry_imports[GraniteSpeechForConditionalGeneration] PASSED 2025-11-03T17:15:11.7914965Z models/test_registry.py::test_registry_imports[H2OVLChatModel] SKIPPED 2025-11-03T17:15:12.1885083Z models/test_registry.py::test_registry_imports[InternVLChatModel] PASSED 2025-11-03T17:15:12.5873779Z models/test_registry.py::test_registry_imports[NemotronH_Nano_VL_V2] PASSED 2025-11-03T17:15:12.9983404Z models/test_registry.py::test_registry_imports[InternS1ForConditionalGeneration] PASSED 2025-11-03T17:15:13.3960100Z models/test_registry.py::test_registry_imports[InternVLForConditionalGeneration] PASSED 2025-11-03T17:15:13.8021276Z models/test_registry.py::test_registry_imports[Idefics3ForConditionalGeneration] PASSED 2025-11-03T17:15:14.2977488Z models/test_registry.py::test_registry_imports[SmolVLMForConditionalGeneration] PASSED 2025-11-03T17:15:14.6983427Z models/test_registry.py::test_registry_imports[KeyeForConditionalGeneration] PASSED 2025-11-03T17:15:15.0964414Z models/test_registry.py::test_registry_imports[KeyeVL1_5ForConditionalGeneration] PASSED 2025-11-03T17:15:15.5186904Z models/test_registry.py::test_registry_imports[RForConditionalGeneration] PASSED 2025-11-03T17:15:15.9235752Z models/test_registry.py::test_registry_imports[KimiVLForConditionalGeneration] PASSED 2025-11-03T17:15:16.3233568Z models/test_registry.py::test_registry_imports[Llama_Nemotron_Nano_VL] PASSED 2025-11-03T17:15:16.7286974Z models/test_registry.py::test_registry_imports[Llama4ForConditionalGeneration] PASSED 2025-11-03T17:15:17.1268596Z models/test_registry.py::test_registry_imports[LlavaForConditionalGeneration] PASSED 2025-11-03T17:15:17.5323809Z models/test_registry.py::test_registry_imports[LlavaNextVideoForConditionalGeneration] PASSED 2025-11-03T17:15:17.9291653Z models/test_registry.py::test_registry_imports[LlavaOnevisionForConditionalGeneration] PASSED 2025-11-03T17:15:18.3253717Z models/test_registry.py::test_registry_imports[MantisForConditionalGeneration] SKIPPED 2025-11-03T17:15:18.8974373Z models/test_registry.py::test_registry_imports[MiDashengLMModel] PASSED 2025-11-03T17:15:19.2979376Z models/test_registry.py::test_registry_imports[MiniMaxVL01ForConditionalGeneration] PASSED 2025-11-03T17:15:19.7047707Z models/test_registry.py::test_registry_imports[MiniCPMO] PASSED 2025-11-03T17:15:20.1070316Z models/test_registry.py::test_registry_imports[MiniCPMV] PASSED 2025-11-03T17:15:20.5116800Z models/test_registry.py::test_registry_imports[Mistral3ForConditionalGeneration] PASSED 2025-11-03T17:15:20.9150110Z models/test_registry.py::test_registry_imports[MolmoForCausalLM] SKIPPED 2025-11-03T17:15:21.3116357Z models/test_registry.py::test_registry_imports[NVLM_D] PASSED 2025-11-03T17:15:21.7145223Z models/test_registry.py::test_registry_imports[Ovis] SKIPPED (`trans...) 2025-11-03T17:15:22.1281932Z models/test_registry.py::test_registry_imports[Ovis2_5] PASSED 2025-11-03T17:15:22.5366164Z models/test_registry.py::test_registry_imports[PaliGemmaForConditionalGeneration] PASSED 2025-11-03T17:15:22.9423241Z models/test_registry.py::test_registry_imports[Phi4MMForCausalLM] PASSED 2025-11-03T17:15:23.3721375Z models/test_registry.py::test_registry_imports[Phi4MultimodalForCausalLM] PASSED 2025-11-03T17:15:23.7742142Z models/test_registry.py::test_registry_imports[PixtralForConditionalGeneration] PASSED 2025-11-03T17:15:24.1794867Z models/test_registry.py::test_registry_imports[QwenVLForConditionalGeneration] PASSED 2025-11-03T17:15:24.5823831Z models/test_registry.py::test_registry_imports[Qwen2_5_VLForConditionalGeneration] PASSED 2025-11-03T17:15:25.0033816Z models/test_registry.py::test_registry_imports[Qwen2AudioForConditionalGeneration] PASSED 2025-11-03T17:15:25.4511435Z models/test_registry.py::test_registry_imports[Qwen2_5OmniModel] PASSED 2025-11-03T17:15:25.8538337Z models/test_registry.py::test_registry_imports[Qwen2_5OmniForConditionalGeneration] PASSED 2025-11-03T17:15:26.2536791Z models/test_registry.py::test_registry_imports[Qwen3VLForConditionalGeneration] SKIPPED 2025-11-03T17:15:26.6526515Z models/test_registry.py::test_registry_imports[Qwen3VLMoeForConditionalGeneration] SKIPPED 2025-11-03T17:15:27.0577172Z models/test_registry.py::test_registry_imports[SkyworkR1VChatModel] PASSED 2025-11-03T17:15:27.4700453Z models/test_registry.py::test_registry_imports[Step3VLForConditionalGeneration] PASSED 2025-11-03T17:15:27.8767880Z models/test_registry.py::test_registry_imports[TarsierForConditionalGeneration] PASSED 2025-11-03T17:15:28.2821442Z models/test_registry.py::test_registry_imports[Tarsier2ForConditionalGeneration] PASSED 2025-11-03T17:15:28.6921819Z models/test_registry.py::test_registry_imports[UltravoxModel] PASSED 2025-11-03T17:15:29.1408166Z models/test_registry.py::test_registry_imports[VoxtralForConditionalGeneration] PASSED 2025-11-03T17:15:29.5517261Z models/test_registry.py::test_registry_imports[WhisperForConditionalGeneration] PASSED 2025-11-03T17:15:29.9655888Z models/test_registry.py::test_registry_imports[MiMoMTPModel] PASSED 2025-11-03T17:15:30.3705406Z models/test_registry.py::test_registry_imports[EagleLlamaForCausalLM] PASSED 2025-11-03T17:15:30.7728539Z models/test_registry.py::test_registry_imports[EagleLlama4ForCausalLM] PASSED 2025-11-03T17:15:31.1783204Z models/test_registry.py::test_registry_imports[EagleMiniCPMForCausalLM] PASSED 2025-11-03T17:15:31.5864793Z models/test_registry.py::test_registry_imports[Eagle3LlamaForCausalLM] PASSED 2025-11-03T17:15:31.9989649Z models/test_registry.py::test_registry_imports[LlamaForCausalLMEagle3] PASSED 2025-11-03T17:15:32.4122836Z models/test_registry.py::test_registry_imports[Eagle3Qwen2_5vlForCausalLM] PASSED 2025-11-03T17:15:32.8245470Z models/test_registry.py::test_registry_imports[EagleDeepSeekMTPModel] PASSED 2025-11-03T17:15:33.2348257Z models/test_registry.py::test_registry_imports[DeepSeekMTPModel] PASSED 2025-11-03T17:15:33.6454893Z models/test_registry.py::test_registry_imports[ErnieMTPModel] PASSED 2025-11-03T17:15:34.0596027Z models/test_registry.py::test_registry_imports[LongCatFlashMTPModel] PASSED 2025-11-03T17:15:34.4713609Z models/test_registry.py::test_registry_imports[Glm4MoeMTPModel] PASSED 2025-11-03T17:15:34.8807490Z models/test_registry.py::test_registry_imports[MedusaModel] PASSED 2025-11-03T17:15:35.2909291Z models/test_registry.py::test_registry_imports[Qwen3NextMTP] SKIPPED 2025-11-03T17:15:35.6984658Z models/test_registry.py::test_registry_imports[SmolLM3ForCausalLM] PASSED 2025-11-03T17:15:36.1041182Z models/test_registry.py::test_registry_imports[Emu3ForConditionalGeneration] PASSED 2025-11-03T17:15:36.5100692Z models/test_registry.py::test_registry_imports[TransformersForCausalLM] PASSED 2025-11-03T17:15:36.9153707Z models/test_registry.py::test_registry_imports[TransformersForMultimodalLM] PASSED 2025-11-03T17:15:37.3241066Z models/test_registry.py::test_registry_imports[TransformersMoEForCausalLM] SKIPPED 2025-11-03T17:15:37.7248798Z models/test_registry.py::test_registry_imports[TransformersMoEForMultimodalLM] SKIPPED 2025-11-03T17:15:38.1361251Z models/test_registry.py::test_registry_imports[TransformersEmbeddingModel] SKIPPED 2025-11-03T17:15:38.5446379Z models/test_registry.py::test_registry_imports[TransformersForSequenceClassification] SKIPPED 2025-11-03T17:15:38.9523608Z models/test_registry.py::test_registry_imports[TransformersMoEForSequenceClassification] SKIPPED 2025-11-03T17:15:39.3621874Z models/test_registry.py::test_registry_imports[TransformersMoEEmbeddingModel] SKIPPED 2025-11-03T17:15:39.3862083Z models/test_registry.py::test_registry_model_property[LlamaForCausalLM-False-False-False] Fork a new process to run a test 3303 2025-11-03T17:15:39.3873319Z Fork a new process to run a test 0 2025-11-03T17:15:39.8901923Z PASSED 2025-11-03T17:15:39.9135081Z models/test_registry.py::test_registry_model_property[LlavaForConditionalGeneration-True-True-False] Fork a new process to run a test 3304 2025-11-03T17:15:39.9148470Z Fork a new process to run a test 0 2025-11-03T17:15:47.1783005Z PASSED 2025-11-03T17:15:47.2015804Z models/test_registry.py::test_registry_model_property[BertForSequenceClassification-False-False-True] Fork a new process to run a test 3356 2025-11-03T17:15:47.2029301Z Fork a new process to run a test 0 2025-11-03T17:15:53.9551089Z PASSED 2025-11-03T17:15:53.9784651Z models/test_registry.py::test_registry_model_property[RobertaForSequenceClassification-False-False-True] Fork a new process to run a test 3408 2025-11-03T17:15:53.9796694Z Fork a new process to run a test 0 2025-11-03T17:16:00.8154377Z PASSED 2025-11-03T17:16:00.8385857Z models/test_registry.py::test_registry_model_property[XLMRobertaForSequenceClassification-False-False-True] Fork a new process to run a test 3460 2025-11-03T17:16:00.8398350Z Fork a new process to run a test 0 2025-11-03T17:16:01.3406062Z PASSED 2025-11-03T17:16:01.3639015Z models/test_registry.py::test_registry_is_pp[DeepseekV2ForCausalLM-True-False] Fork a new process to run a test 3461 2025-11-03T17:16:01.3651910Z Fork a new process to run a test 0 2025-11-03T17:16:08.1071085Z PASSED 2025-11-03T17:16:08.1302059Z models/test_registry.py::test_registry_is_pp[Qwen2VLForConditionalGeneration-True-True] Fork a new process to run a test 3512 2025-11-03T17:16:08.1315247Z Fork a new process to run a test 0 2025-11-03T17:16:14.8387090Z PASSED 2025-11-03T17:16:15.2481417Z models/test_registry.py::test_hf_registry_coverage PASSED 2025-11-03T17:16:15.2482718Z 2025-11-03T17:16:15.2483638Z =============================== warnings summary =============================== 2025-11-03T17:16:15.2484502Z ../../../../../../opt/conda/envs/py_3.12/lib/python3.12/site-packages/torch/cuda/__init__.py:65 2025-11-03T17:16:15.2486238Z /opt/conda/envs/py_3.12/lib/python3.12/site-packages/torch/cuda/__init__.py:65: FutureWarning: The pynvml package is deprecated. Please install nvidia-ml-py instead. If you did not install pynvml directly, please report this to the maintainers of the package that installed pynvml for you. 2025-11-03T17:16:15.2487570Z import pynvml # type: ignore[import] 2025-11-03T17:16:15.2487814Z 2025-11-03T17:16:15.2487929Z :488 2025-11-03T17:16:15.2488726Z :488: DeprecationWarning: builtin type SwigPyPacked has no __module__ attribute 2025-11-03T17:16:15.2489309Z 2025-11-03T17:16:15.2489479Z :488 2025-11-03T17:16:15.2490167Z :488: DeprecationWarning: builtin type SwigPyObject has no __module__ attribute 2025-11-03T17:16:15.2490727Z 2025-11-03T17:16:15.2490992Z ../../../../../../opt/conda/envs/py_3.12/lib/python3.12/site-packages/schemathesis/generation/coverage.py:305 2025-11-03T17:16:15.2492151Z /opt/conda/envs/py_3.12/lib/python3.12/site-packages/schemathesis/generation/coverage.py:305: DeprecationWarning: jsonschema.exceptions.RefResolutionError is deprecated as of version 4.18.0. If you wish to catch potential reference resolution errors, directly catch referencing.exceptions.Unresolvable. 2025-11-03T17:16:15.2493215Z ref_error: type[Exception] = jsonschema.RefResolutionError, 2025-11-03T17:16:15.2493429Z 2025-11-03T17:16:15.2493665Z tests/models/test_registry.py::test_registry_imports[LlavaNextForConditionalGeneration] 2025-11-03T17:16:15.2494161Z tests/models/test_registry.py::test_registry_imports[LlavaNextForConditionalGeneration] 2025-11-03T17:16:15.2494853Z tests/models/test_registry.py::test_registry_imports[LlavaNextForConditionalGeneration] 2025-11-03T17:16:15.2495349Z tests/models/test_registry.py::test_registry_imports[LlavaNextForConditionalGeneration] 2025-11-03T17:16:15.2496273Z /opt/conda/envs/py_3.12/lib/python3.12/site-packages/triton/runtime/autotuner.py:99: DeprecationWarning: warmup, rep, and use_cuda_graph parameters are deprecated. See https://github.com/triton-lang/triton/pull/4496 for details. 2025-11-03T17:16:15.2497365Z warnings.warn(("warmup, rep, and use_cuda_graph parameters are deprecated. See " 2025-11-03T17:16:15.2497615Z 2025-11-03T17:16:15.2497829Z tests/models/test_registry.py::test_registry_imports[LlavaNextForConditionalGeneration] 2025-11-03T17:16:15.2498772Z /opt/conda/envs/py_3.12/lib/python3.12/site-packages/torch/library.py:356: UserWarning: Warning only once for all operators, other operators may also be overridden. 2025-11-03T17:16:15.2499488Z Overriding a previously registered kernel for the same operator and the same dispatch key 2025-11-03T17:16:15.2500466Z operator: xformers_python::_fused_allgather_and_linear_impl(Tensor scattered_input, Tensor[] weights, str process_group_name, Tensor(a3!)[] gathered_outputs, SymInt timeout_s, bool _wait, bool _memcpy, Tensor scale_scattered_input, Tensor?[] scales_weights) -> () 2025-11-03T17:16:15.2501557Z registered at /opt/conda/envs/py_3.12/lib/python3.12/site-packages/torch/_library/custom_ops.py:922 2025-11-03T17:16:15.2501963Z dispatch key: ADInplaceOrView 2025-11-03T17:16:15.2502192Z previous kernel: no debug info 2025-11-03T17:16:15.2502939Z new kernel: registered at /opt/conda/envs/py_3.12/lib/python3.12/site-packages/torch/_library/custom_ops.py:922 (Triggered internally at /var/lib/jenkins/workspace/aten/src/ATen/core/dispatch/OperatorEntry.cpp:208.) 2025-11-03T17:16:15.2503654Z self.m.impl( 2025-11-03T17:16:15.2503764Z 2025-11-03T17:16:15.2503950Z tests/models/test_registry.py::test_registry_imports[PrithviGeoSpatialMAE] 2025-11-03T17:16:15.2504724Z /opt/conda/envs/py_3.12/lib/python3.12/site-packages/pretrainedmodels/models/dpn.py:255: SyntaxWarning: "is" with 'str' literal. Did you mean "=="? 2025-11-03T17:16:15.2505273Z if block_type is 'proj': 2025-11-03T17:16:15.2505410Z 2025-11-03T17:16:15.2505598Z tests/models/test_registry.py::test_registry_imports[PrithviGeoSpatialMAE] 2025-11-03T17:16:15.2506237Z /opt/conda/envs/py_3.12/lib/python3.12/site-packages/pretrainedmodels/models/dpn.py:258: SyntaxWarning: "is" with 'str' literal. Did you mean "=="? 2025-11-03T17:16:15.2506778Z elif block_type is 'down': 2025-11-03T17:16:15.2506909Z 2025-11-03T17:16:15.2507078Z tests/models/test_registry.py::test_registry_imports[PrithviGeoSpatialMAE] 2025-11-03T17:16:15.2507706Z /opt/conda/envs/py_3.12/lib/python3.12/site-packages/pretrainedmodels/models/dpn.py:262: SyntaxWarning: "is" with 'str' literal. Did you mean "=="? 2025-11-03T17:16:15.2508231Z assert block_type is 'normal' 2025-11-03T17:16:15.2508366Z 2025-11-03T17:16:15.2508537Z tests/models/test_registry.py::test_registry_imports[PrithviGeoSpatialMAE] 2025-11-03T17:16:15.2509549Z /opt/conda/envs/py_3.12/lib/python3.12/site-packages/geopandas/_compat.py:7: DeprecationWarning: The 'shapely.geos' module is deprecated, and will be removed in a future version. All attributes of 'shapely.geos' are available directly from the top-level 'shapely' namespace (since shapely 2.0.0). 2025-11-03T17:16:15.2510465Z import shapely.geos 2025-11-03T17:16:15.2510578Z 2025-11-03T17:16:15.2510752Z tests/models/test_registry.py::test_registry_imports[PrithviGeoSpatialMAE] 2025-11-03T17:16:15.2512710Z /opt/conda/envs/py_3.12/lib/python3.12/site-packages/torch/__init__.py:1628: UserWarning: Please use the new API settings to control TF32 behavior, such as torch.backends.cudnn.conv.fp32_precision = 'tf32' or torch.backends.cuda.matmul.fp32_precision = 'ieee'. Old settings, e.g, torch.backends.cuda.matmul.allow_tf32 = True, torch.backends.cudnn.allow_tf32 = True, allowTF32CuDNN() and allowTF32CuBLAS() will be deprecated after Pytorch 2.9. Please see https://pytorch.org/docs/main/notes/cuda.html#tensorfloat-32-tf32-on-ampere-and-later-devices (Triggered internally at /var/lib/jenkins/workspace/aten/src/ATen/Context.cpp:45.) 2025-11-03T17:16:15.2514662Z _C._set_float32_matmul_precision(precision) 2025-11-03T17:16:15.2514833Z 2025-11-03T17:16:15.2515074Z tests/models/test_registry.py::test_registry_model_property[LlamaForCausalLM-False-False-False] 2025-11-03T17:16:15.2515660Z tests/models/test_registry.py::test_registry_model_property[LlavaForConditionalGeneration-True-True-False] 2025-11-03T17:16:15.2516287Z tests/models/test_registry.py::test_registry_model_property[BertForSequenceClassification-False-False-True] 2025-11-03T17:16:15.2516924Z tests/models/test_registry.py::test_registry_model_property[RobertaForSequenceClassification-False-False-True] 2025-11-03T17:16:15.2517592Z tests/models/test_registry.py::test_registry_model_property[XLMRobertaForSequenceClassification-False-False-True] 2025-11-03T17:16:15.2518178Z tests/models/test_registry.py::test_registry_is_pp[DeepseekV2ForCausalLM-True-False] 2025-11-03T17:16:15.2518699Z tests/models/test_registry.py::test_registry_is_pp[Qwen2VLForConditionalGeneration-True-True] 2025-11-03T17:16:15.2519444Z /var/lib/jenkins/workspace/vllm/tests/utils.py:872: DeprecationWarning: This process (pid=3246) is multi-threaded, use of fork() may lead to deadlocks in the child. 2025-11-03T17:16:15.2520003Z pid = os.fork() 2025-11-03T17:16:15.2520116Z 2025-11-03T17:16:15.2520287Z -- Docs: https://docs.pytest.org/en/stable/how-to/capture-warnings.html 2025-11-03T17:16:15.2520927Z =========== 207 passed, 22 skipped, 21 warnings in 117.13s (0:01:57) =========== 2025-11-03T17:16:15.6799619Z sys:1: DeprecationWarning: builtin type swigvarlink has no __module__ attribute 2025-11-03T17:16:17.5211951Z 2025-11-03 17:16:17,520 [INFO] cli.lib.core.vllm.lib: Finish running step: pytest -v -s models/test_registry.py 2025-11-03T17:16:17.5213067Z 2025-11-03 17:16:17,520 [INFO] cli.lib.core.vllm.lib: Running step: pytest -v -s models/test_utils.py 2025-11-03T17:16:17.5213719Z 2025-11-03 17:16:17,520 [INFO] cli.lib.common.utils: [shell] pytest -v -s models/test_utils.py 2025-11-03T17:16:22.0687407Z INFO 11-03 17:16:22 [__init__.py:224] Automatically detected platform cuda. 2025-11-03T17:16:23.5564867Z /opt/conda/envs/py_3.12/lib/python3.12/site-packages/pytest_asyncio/plugin.py:208: PytestDeprecationWarning: The configuration option "asyncio_default_fixture_loop_scope" is unset. 2025-11-03T17:16:23.5566866Z The event loop scope for asynchronous fixtures will default to the fixture caching scope. Future versions of pytest-asyncio will default the loop scope for asynchronous fixtures to function scope. Set the default fixture loop scope explicitly in order to avoid unexpected behavior in the future. Valid fixture loop scopes are: "function", "class", "module", "package", "session" 2025-11-03T17:16:23.5568313Z 2025-11-03T17:16:23.5568585Z warnings.warn(PytestDeprecationWarning(_DEFAULT_FIXTURE_LOOP_SCOPE_UNSET)) 2025-11-03T17:16:23.5628529Z ============================= test session starts ============================== 2025-11-03T17:16:23.5629123Z platform linux -- Python 3.12.12, pytest-8.3.5, pluggy-1.5.0 -- /opt/conda/envs/py_3.12/bin/python 2025-11-03T17:16:23.5733241Z cachedir: .pytest_cache 2025-11-03T17:16:23.5734066Z hypothesis profile 'ci' -> database=None, deadline=None, print_blob=True, derandomize=True, suppress_health_check=[HealthCheck.too_slow] 2025-11-03T17:16:23.5734699Z rootdir: /var/lib/jenkins/workspace/vllm 2025-11-03T17:16:23.5734990Z configfile: pyproject.toml 2025-11-03T17:16:23.5736016Z plugins: cpp-2.3.0, flakefinder-1.1.0, rerunfailures-14.0, xdist-3.3.1, xdoctest-1.3.0, asyncio-0.24.0, cov-6.3.0, timeout-2.3.1, hydra-core-1.3.2, mock-3.14.0, schemathesis-3.39.15, buildkite-test-collector-0.1.9, subtests-0.14.1, hypothesis-6.131.0, forked-1.6.0, shard-0.1.2, anyio-4.6.2.post1 2025-11-03T17:16:23.5737100Z asyncio: mode=Mode.STRICT, default_loop_scope=None 2025-11-03T17:16:24.9676719Z collecting ...  2025-11-03T17:16:24.9677332Z collecting 4 items  2025-11-03T17:16:24.9678303Z collected 4 items  2025-11-03T17:16:24.9679533Z Running 4 items in this shard: tests/models/test_utils.py::test_module_with_batchnorm_can_load, tests/models/test_utils.py::test_module_with_child_containing_batchnorm_can_autoload, tests/models/test_utils.py::test_module_skip_prefix, tests/models/test_utils.py::test_module_skip_substr 2025-11-03T17:16:24.9685555Z 2025-11-03T17:16:25.1949336Z models/test_utils.py::test_module_with_batchnorm_can_load PASSED 2025-11-03T17:16:25.4102111Z models/test_utils.py::test_module_with_child_containing_batchnorm_can_autoload PASSED 2025-11-03T17:16:25.6294161Z models/test_utils.py::test_module_skip_prefix PASSED 2025-11-03T17:16:25.8449802Z models/test_utils.py::test_module_skip_substr PASSED 2025-11-03T17:16:25.8450118Z 2025-11-03T17:16:25.8450444Z =============================== warnings summary =============================== 2025-11-03T17:16:25.8451214Z ../../../../../../opt/conda/envs/py_3.12/lib/python3.12/site-packages/torch/cuda/__init__.py:65 2025-11-03T17:16:25.8452772Z /opt/conda/envs/py_3.12/lib/python3.12/site-packages/torch/cuda/__init__.py:65: FutureWarning: The pynvml package is deprecated. Please install nvidia-ml-py instead. If you did not install pynvml directly, please report this to the maintainers of the package that installed pynvml for you. 2025-11-03T17:16:25.8454128Z import pynvml # type: ignore[import] 2025-11-03T17:16:25.8454333Z 2025-11-03T17:16:25.8454451Z :488 2025-11-03T17:16:25.8454969Z :488: DeprecationWarning: builtin type SwigPyPacked has no __module__ attribute 2025-11-03T17:16:25.8455533Z 2025-11-03T17:16:25.8455702Z :488 2025-11-03T17:16:25.8456552Z :488: DeprecationWarning: builtin type SwigPyObject has no __module__ attribute 2025-11-03T17:16:25.8457012Z 2025-11-03T17:16:25.8457345Z ../../../../../../opt/conda/envs/py_3.12/lib/python3.12/site-packages/schemathesis/generation/coverage.py:305 2025-11-03T17:16:25.8459065Z /opt/conda/envs/py_3.12/lib/python3.12/site-packages/schemathesis/generation/coverage.py:305: DeprecationWarning: jsonschema.exceptions.RefResolutionError is deprecated as of version 4.18.0. If you wish to catch potential reference resolution errors, directly catch referencing.exceptions.Unresolvable. 2025-11-03T17:16:25.8460162Z ref_error: type[Exception] = jsonschema.RefResolutionError, 2025-11-03T17:16:25.8460378Z 2025-11-03T17:16:25.8460563Z -- Docs: https://docs.pytest.org/en/stable/how-to/capture-warnings.html 2025-11-03T17:16:25.8461137Z ======================== 4 passed, 4 warnings in 2.28s ========================= 2025-11-03T17:16:26.0738563Z sys:1: DeprecationWarning: builtin type swigvarlink has no __module__ attribute 2025-11-03T17:16:27.0009626Z 2025-11-03 17:16:27,000 [INFO] cli.lib.core.vllm.lib: Finish running step: pytest -v -s models/test_utils.py 2025-11-03T17:16:27.0010367Z 2025-11-03 17:16:27,000 [INFO] cli.lib.core.vllm.lib: Running step: pytest -v -s models/test_vision.py 2025-11-03T17:16:27.0010967Z 2025-11-03 17:16:27,000 [INFO] cli.lib.common.utils: [shell] pytest -v -s models/test_vision.py 2025-11-03T17:16:31.5759113Z INFO 11-03 17:16:31 [__init__.py:224] Automatically detected platform cuda. 2025-11-03T17:16:33.0756167Z /opt/conda/envs/py_3.12/lib/python3.12/site-packages/pytest_asyncio/plugin.py:208: PytestDeprecationWarning: The configuration option "asyncio_default_fixture_loop_scope" is unset. 2025-11-03T17:16:33.0758323Z The event loop scope for asynchronous fixtures will default to the fixture caching scope. Future versions of pytest-asyncio will default the loop scope for asynchronous fixtures to function scope. Set the default fixture loop scope explicitly in order to avoid unexpected behavior in the future. Valid fixture loop scopes are: "function", "class", "module", "package", "session" 2025-11-03T17:16:33.0760068Z 2025-11-03T17:16:33.0760317Z warnings.warn(PytestDeprecationWarning(_DEFAULT_FIXTURE_LOOP_SCOPE_UNSET)) 2025-11-03T17:16:33.0819942Z ============================= test session starts ============================== 2025-11-03T17:16:33.0820528Z platform linux -- Python 3.12.12, pytest-8.3.5, pluggy-1.5.0 -- /opt/conda/envs/py_3.12/bin/python 2025-11-03T17:16:33.0925690Z cachedir: .pytest_cache 2025-11-03T17:16:33.0926372Z hypothesis profile 'ci' -> database=None, deadline=None, print_blob=True, derandomize=True, suppress_health_check=[HealthCheck.too_slow] 2025-11-03T17:16:33.0926999Z rootdir: /var/lib/jenkins/workspace/vllm 2025-11-03T17:16:33.0927305Z configfile: pyproject.toml 2025-11-03T17:16:33.0928347Z plugins: cpp-2.3.0, flakefinder-1.1.0, rerunfailures-14.0, xdist-3.3.1, xdoctest-1.3.0, asyncio-0.24.0, cov-6.3.0, timeout-2.3.1, hydra-core-1.3.2, mock-3.14.0, schemathesis-3.39.15, buildkite-test-collector-0.1.9, subtests-0.14.1, hypothesis-6.131.0, forked-1.6.0, shard-0.1.2, anyio-4.6.2.post1 2025-11-03T17:16:33.0929444Z asyncio: mode=Mode.STRICT, default_loop_scope=None 2025-11-03T17:16:34.3836981Z collecting ... WARNING 11-03 17:16:34 [interface.py:514] Current platform cuda does not have '_pytestfixturefunction' attribute. 2025-11-03T17:16:34.6867414Z WARNING 11-03 17:16:34 [interface.py:514] Current platform cuda does not have '__test__' attribute. 2025-11-03T17:16:34.6868105Z WARNING 11-03 17:16:34 [interface.py:514] Current platform cuda does not have '__bases__' attribute. 2025-11-03T17:16:34.6868711Z WARNING 11-03 17:16:34 [interface.py:514] Current platform cuda does not have '__test__' attribute. 2025-11-03T17:16:34.6869347Z WARNING 11-03 17:16:34 [interface.py:514] Current platform cuda does not have '_schemathesis_test' attribute. 2025-11-03T17:16:34.6901092Z  2025-11-03T17:16:34.6906432Z collecting 19 items  2025-11-03T17:16:34.6906991Z collected 19 items  2025-11-03T17:16:34.6912852Z Running 19 items in this shard: tests/models/test_vision.py::test_resolve_visual_encoder_outputs[select_layers0-10-10-expected_features0], tests/models/test_vision.py::test_resolve_visual_encoder_outputs[select_layers1-10-10-expected_features1], tests/models/test_vision.py::test_resolve_visual_encoder_outputs[select_layers2-10-20-expected_features2], tests/models/test_vision.py::test_resolve_visual_encoder_outputs[select_layers3-10-20-expected_features3], tests/models/test_vision.py::test_run_dp_sharded_vision_model[1], tests/models/test_vision.py::test_run_dp_sharded_vision_model[4], tests/models/test_vision.py::test_run_dp_sharded_vision_model[5], tests/models/test_vision.py::test_get_load_balance_assignment_cases[sizes0-2-expected_shuffle_indices0-expected_gpu_sample_counts0-expected_grouped_sizes_per_gpu0-empty input], tests/models/test_vision.py::test_get_load_balance_assignment_cases[sizes1-4-expected_shuffle_indices1-expected_gpu_sample_counts1-expected_grouped_sizes_per_gpu1-fewer samples than GPUs], tests/models/test_vision.py::test_get_load_balance_assignment_cases[sizes2-1-expected_shuffle_indices2-expected_gpu_sample_counts2-expected_grouped_sizes_per_gpu2-single GPU], tests/models/test_vision.py::test_get_load_balance_assignment_cases[sizes3-2-expected_shuffle_indices3-expected_gpu_sample_counts3-expected_grouped_sizes_per_gpu3-balanced assignment], tests/models/test_vision.py::test_get_load_balance_assignment_cases[sizes4-2-expected_shuffle_indices4-expected_gpu_sample_counts4-expected_grouped_sizes_per_gpu4-unbalanced sizes], tests/models/test_vision.py::test_run_dp_sharded_mrope_vision_model[1], tests/models/test_vision.py::test_run_dp_sharded_mrope_vision_model[3], tests/models/test_vision.py::test_run_dp_sharded_mrope_vision_model[5], tests/models/test_vision.py::test_run_dp_sharded_mrope_vision_model_empty_input, tests/models/test_vision.py::test_run_dp_sharded_mrope_vision_model_uneven_load, tests/models/test_vision.py::test_simple_mrope_vision_model_spatial_merge[2], tests/models/test_vision.py::test_simple_mrope_vision_model_spatial_merge[4] 2025-11-03T17:16:34.6919230Z 2025-11-03T17:16:34.9250409Z models/test_vision.py::test_resolve_visual_encoder_outputs[select_layers0-10-10-expected_features0] PASSED 2025-11-03T17:16:35.1539249Z models/test_vision.py::test_resolve_visual_encoder_outputs[select_layers1-10-10-expected_features1] PASSED 2025-11-03T17:16:35.3802273Z models/test_vision.py::test_resolve_visual_encoder_outputs[select_layers2-10-20-expected_features2] PASSED 2025-11-03T17:16:35.6062711Z models/test_vision.py::test_resolve_visual_encoder_outputs[select_layers3-10-20-expected_features3] PASSED 2025-11-03T17:16:35.6071730Z models/test_vision.py::test_run_dp_sharded_vision_model[1] SKIPPED (...) 2025-11-03T17:16:35.6079493Z models/test_vision.py::test_run_dp_sharded_vision_model[4] SKIPPED (...) 2025-11-03T17:16:35.6087321Z models/test_vision.py::test_run_dp_sharded_vision_model[5] SKIPPED (...) 2025-11-03T17:16:35.8347908Z models/test_vision.py::test_get_load_balance_assignment_cases[sizes0-2-expected_shuffle_indices0-expected_gpu_sample_counts0-expected_grouped_sizes_per_gpu0-empty input] PASSED 2025-11-03T17:16:36.0618175Z models/test_vision.py::test_get_load_balance_assignment_cases[sizes1-4-expected_shuffle_indices1-expected_gpu_sample_counts1-expected_grouped_sizes_per_gpu1-fewer samples than GPUs] PASSED 2025-11-03T17:16:36.2891043Z models/test_vision.py::test_get_load_balance_assignment_cases[sizes2-1-expected_shuffle_indices2-expected_gpu_sample_counts2-expected_grouped_sizes_per_gpu2-single GPU] PASSED 2025-11-03T17:16:36.5157743Z models/test_vision.py::test_get_load_balance_assignment_cases[sizes3-2-expected_shuffle_indices3-expected_gpu_sample_counts3-expected_grouped_sizes_per_gpu3-balanced assignment] PASSED 2025-11-03T17:16:36.7426566Z models/test_vision.py::test_get_load_balance_assignment_cases[sizes4-2-expected_shuffle_indices4-expected_gpu_sample_counts4-expected_grouped_sizes_per_gpu4-unbalanced sizes] PASSED 2025-11-03T17:16:36.7433606Z models/test_vision.py::test_run_dp_sharded_mrope_vision_model[1] SKIPPED 2025-11-03T17:16:36.7441777Z models/test_vision.py::test_run_dp_sharded_mrope_vision_model[3] SKIPPED 2025-11-03T17:16:36.7449372Z models/test_vision.py::test_run_dp_sharded_mrope_vision_model[5] SKIPPED 2025-11-03T17:16:36.7456729Z models/test_vision.py::test_run_dp_sharded_mrope_vision_model_empty_input SKIPPED 2025-11-03T17:16:36.7464404Z models/test_vision.py::test_run_dp_sharded_mrope_vision_model_uneven_load SKIPPED 2025-11-03T17:16:37.2896105Z models/test_vision.py::test_simple_mrope_vision_model_spatial_merge[2] PASSED 2025-11-03T17:16:37.5190880Z models/test_vision.py::test_simple_mrope_vision_model_spatial_merge[4] PASSED 2025-11-03T17:16:37.5191360Z 2025-11-03T17:16:37.5191702Z =============================== warnings summary =============================== 2025-11-03T17:16:37.5192421Z ../../../../../../opt/conda/envs/py_3.12/lib/python3.12/site-packages/torch/cuda/__init__.py:65 2025-11-03T17:16:37.5193856Z /opt/conda/envs/py_3.12/lib/python3.12/site-packages/torch/cuda/__init__.py:65: FutureWarning: The pynvml package is deprecated. Please install nvidia-ml-py instead. If you did not install pynvml directly, please report this to the maintainers of the package that installed pynvml for you. 2025-11-03T17:16:37.5195172Z import pynvml # type: ignore[import] 2025-11-03T17:16:37.5195374Z 2025-11-03T17:16:37.5195492Z :488 2025-11-03T17:16:37.5196006Z :488: DeprecationWarning: builtin type SwigPyPacked has no __module__ attribute 2025-11-03T17:16:37.5196433Z 2025-11-03T17:16:37.5196542Z :488 2025-11-03T17:16:37.5197120Z :488: DeprecationWarning: builtin type SwigPyObject has no __module__ attribute 2025-11-03T17:16:37.5197886Z 2025-11-03T17:16:37.5198241Z ../../../../../../opt/conda/envs/py_3.12/lib/python3.12/site-packages/schemathesis/generation/coverage.py:305 2025-11-03T17:16:37.5199726Z /opt/conda/envs/py_3.12/lib/python3.12/site-packages/schemathesis/generation/coverage.py:305: DeprecationWarning: jsonschema.exceptions.RefResolutionError is deprecated as of version 4.18.0. If you wish to catch potential reference resolution errors, directly catch referencing.exceptions.Unresolvable. 2025-11-03T17:16:37.5200813Z ref_error: type[Exception] = jsonschema.RefResolutionError, 2025-11-03T17:16:37.5201033Z 2025-11-03T17:16:37.5201223Z -- Docs: https://docs.pytest.org/en/stable/how-to/capture-warnings.html 2025-11-03T17:16:37.5201840Z ================== 11 passed, 8 skipped, 4 warnings in 4.44s =================== 2025-11-03T17:16:37.7620959Z sys:1: DeprecationWarning: builtin type swigvarlink has no __module__ attribute 2025-11-03T17:16:38.8514492Z 2025-11-03 17:16:38,850 [INFO] cli.lib.core.vllm.lib: Finish running step: pytest -v -s models/test_vision.py 2025-11-03T17:16:38.8515247Z 2025-11-03 17:16:38,851 [INFO] cli.lib.core.vllm.lib: Running step: pytest -v -s models/test_initialization.py 2025-11-03T17:16:38.8515954Z 2025-11-03 17:16:38,851 [INFO] cli.lib.common.utils: [shell] pytest -v -s models/test_initialization.py 2025-11-03T17:16:43.3929573Z INFO 11-03 17:16:43 [__init__.py:224] Automatically detected platform cuda. 2025-11-03T17:16:44.8892124Z /opt/conda/envs/py_3.12/lib/python3.12/site-packages/pytest_asyncio/plugin.py:208: PytestDeprecationWarning: The configuration option "asyncio_default_fixture_loop_scope" is unset. 2025-11-03T17:16:44.8894505Z The event loop scope for asynchronous fixtures will default to the fixture caching scope. Future versions of pytest-asyncio will default the loop scope for asynchronous fixtures to function scope. Set the default fixture loop scope explicitly in order to avoid unexpected behavior in the future. Valid fixture loop scopes are: "function", "class", "module", "package", "session" 2025-11-03T17:16:44.8895687Z 2025-11-03T17:16:44.8895885Z warnings.warn(PytestDeprecationWarning(_DEFAULT_FIXTURE_LOOP_SCOPE_UNSET)) 2025-11-03T17:16:44.8956462Z ============================= test session starts ============================== 2025-11-03T17:16:44.8957038Z platform linux -- Python 3.12.12, pytest-8.3.5, pluggy-1.5.0 -- /opt/conda/envs/py_3.12/bin/python 2025-11-03T17:16:44.9061230Z cachedir: .pytest_cache 2025-11-03T17:16:44.9062083Z hypothesis profile 'ci' -> database=None, deadline=None, print_blob=True, derandomize=True, suppress_health_check=[HealthCheck.too_slow] 2025-11-03T17:16:44.9062725Z rootdir: /var/lib/jenkins/workspace/vllm 2025-11-03T17:16:44.9063034Z configfile: pyproject.toml 2025-11-03T17:16:44.9064075Z plugins: cpp-2.3.0, flakefinder-1.1.0, rerunfailures-14.0, xdist-3.3.1, xdoctest-1.3.0, asyncio-0.24.0, cov-6.3.0, timeout-2.3.1, hydra-core-1.3.2, mock-3.14.0, schemathesis-3.39.15, buildkite-test-collector-0.1.9, subtests-0.14.1, hypothesis-6.131.0, forked-1.6.0, shard-0.1.2, anyio-4.6.2.post1 2025-11-03T17:16:44.9065125Z asyncio: mode=Mode.STRICT, default_loop_scope=None 2025-11-03T17:16:46.5124811Z collecting ...  2025-11-03T17:16:46.5147036Z collecting 225 items  2025-11-03T17:16:46.5147624Z collected 225 items  2025-11-03T17:16:46.5211883Z Running 225 items in this shard: tests/models/test_initialization.py::test_can_initialize_small_subset[LlavaForConditionalGeneration], tests/models/test_initialization.py::test_can_initialize_small_subset[Llama4ForConditionalGeneration], tests/models/test_initialization.py::test_can_initialize_small_subset[BertForSequenceClassification], tests/models/test_initialization.py::test_can_initialize_small_subset[Gemma3nForCausalLM], tests/models/test_initialization.py::test_can_initialize_small_subset[JinaVLForRanking], tests/models/test_initialization.py::test_can_initialize_small_subset[InternVLChatModel], tests/models/test_initialization.py::test_can_initialize_small_subset[InternLM2ForRewardModel], tests/models/test_initialization.py::test_can_initialize_small_subset[TransformersForMultimodalLM], tests/models/test_initialization.py::test_can_initialize_small_subset[PrithviGeoSpatialMAE], tests/models/test_initialization.py::test_can_initialize_small_subset[UltravoxModel], tests/models/test_initialization.py::test_can_initialize_small_subset[DeepSeekMTPModel], tests/models/test_initialization.py::test_can_initialize_small_subset[XLMRobertaModel], tests/models/test_initialization.py::test_can_initialize_large_subset[CohereForCausalLM], tests/models/test_initialization.py::test_can_initialize_large_subset[EagleLlama4ForCausalLM], tests/models/test_initialization.py::test_can_initialize_large_subset[InternLM2VEForCausalLM], tests/models/test_initialization.py::test_can_initialize_large_subset[JAISLMHeadModel], tests/models/test_initialization.py::test_can_initialize_large_subset[DeepseekV3ForCausalLM], tests/models/test_initialization.py::test_can_initialize_large_subset[MptForCausalLM], tests/models/test_initialization.py::test_can_initialize_large_subset[LlavaNextVideoForConditionalGeneration], tests/models/test_initialization.py::test_can_initialize_large_subset[Ovis], tests/models/test_initialization.py::test_can_initialize_large_subset[KeyeVL1_5ForConditionalGeneration], tests/models/test_initialization.py::test_can_initialize_large_subset[Gemma3nForConditionalGeneration], tests/models/test_initialization.py::test_can_initialize_large_subset[KeyeForConditionalGeneration], tests/models/test_initialization.py::test_can_initialize_large_subset[SmolVLMForConditionalGeneration], tests/models/test_initialization.py::test_can_initialize_large_subset[HCXVisionForCausalLM], tests/models/test_initialization.py::test_can_initialize_large_subset[Blip2ForConditionalGeneration], tests/models/test_initialization.py::test_can_initialize_large_subset[InternLM2ForCausalLM], tests/models/test_initialization.py::test_can_initialize_large_subset[EagleLlamaForCausalLM], tests/models/test_initialization.py::test_can_initialize_large_subset[MiDashengLMModel], tests/models/test_initialization.py::test_can_initialize_large_subset[QWenLMHeadModel], tests/models/test_initialization.py::test_can_initialize_large_subset[ApertusForCausalLM], tests/models/test_initialization.py::test_can_initialize_large_subset[LlavaOnevisionForConditionalGeneration], tests/models/test_initialization.py::test_can_initialize_large_subset[OrionForCausalLM], tests/models/test_initialization.py::test_can_initialize_large_subset[Qwen2Model], tests/models/test_initialization.py::test_can_initialize_large_subset[Idefics3ForConditionalGeneration], tests/models/test_initialization.py::test_can_initialize_large_subset[MiniMaxForCausalLM], tests/models/test_initialization.py::test_can_initialize_large_subset[GlmForCausalLM], tests/models/test_initialization.py::test_can_initialize_large_subset[RobertaModel], tests/models/test_initialization.py::test_can_initialize_large_subset[Gemma2ForCausalLM], tests/models/test_initialization.py::test_can_initialize_large_subset[ArceeForCausalLM], tests/models/test_initialization.py::test_can_initialize_large_subset[Ernie4_5ForCausalLM], tests/models/test_initialization.py::test_can_initialize_large_subset[Lfm2MoeForCausalLM], tests/models/test_initialization.py::test_can_initialize_large_subset[Llama4ForCausalLM], tests/models/test_initialization.py::test_can_initialize_large_subset[ChameleonForConditionalGeneration], tests/models/test_initialization.py::test_can_initialize_large_subset[Mistral3ForConditionalGeneration], tests/models/test_initialization.py::test_can_initialize_large_subset[Qwen3MoeForCausalLM], tests/models/test_initialization.py::test_can_initialize_large_subset[MiMoForCausalLM], tests/models/test_initialization.py::test_can_initialize_large_subset[GteModel], tests/models/test_initialization.py::test_can_initialize_large_subset[PixtralForConditionalGeneration], tests/models/test_initialization.py::test_can_initialize_large_subset[Exaone4ForCausalLM], tests/models/test_initialization.py::test_can_initialize_large_subset[GPT2LMHeadModel], tests/models/test_initialization.py::test_can_initialize_large_subset[MantisForConditionalGeneration], tests/models/test_initialization.py::test_can_initialize_large_subset[BailingMoeV2ForCausalLM], tests/models/test_initialization.py::test_can_initialize_large_subset[SkyworkR1VChatModel], tests/models/test_initialization.py::test_can_initialize_large_subset[GLM4VForCausalLM], tests/models/test_initialization.py::test_can_initialize_large_subset[ChatGLMModel], tests/models/test_initialization.py::test_can_initialize_large_subset[Qwen2_5OmniModel], tests/models/test_initialization.py::test_can_initialize_large_subset[Eagle3Qwen2_5vlForCausalLM], tests/models/test_initialization.py::test_can_initialize_large_subset[GteNewForSequenceClassification], tests/models/test_initialization.py::test_can_initialize_large_subset[AquilaForCausalLM], tests/models/test_initialization.py::test_can_initialize_large_subset[MiniMaxText01ForCausalLM], tests/models/test_initialization.py::test_can_initialize_large_subset[Step3TextForCausalLM], tests/models/test_initialization.py::test_can_initialize_large_subset[Glm4vForConditionalGeneration], tests/models/test_initialization.py::test_can_initialize_large_subset[LlamaModel], tests/models/test_initialization.py::test_can_initialize_large_subset[KimiVLForConditionalGeneration], tests/models/test_initialization.py::test_can_initialize_large_subset[TransformersMoEForSequenceClassification], tests/models/test_initialization.py::test_can_initialize_large_subset[FalconMambaForCausalLM], tests/models/test_initialization.py::test_can_initialize_large_subset[RWForCausalLM], tests/models/test_initialization.py::test_can_initialize_large_subset[WhisperForConditionalGeneration], tests/models/test_initialization.py::test_can_initialize_large_subset[DeciLMForCausalLM], tests/models/test_initialization.py::test_can_initialize_large_subset[BailingMoeForCausalLM], tests/models/test_initialization.py::test_can_initialize_large_subset[LlamaForCausalLMEagle3], tests/models/test_initialization.py::test_can_initialize_large_subset[DeepseekV2ForCausalLM], tests/models/test_initialization.py::test_can_initialize_large_subset[TransformersForSequenceClassification], tests/models/test_initialization.py::test_can_initialize_large_subset[OPTForCausalLM], tests/models/test_initialization.py::test_can_initialize_large_subset[MiniCPMV], tests/models/test_initialization.py::test_can_initialize_large_subset[MiniCPMForCausalLM], tests/models/test_initialization.py::test_can_initialize_large_subset[ModernBertModel], tests/models/test_initialization.py::test_can_initialize_large_subset[Mamba2ForCausalLM], tests/models/test_initialization.py::test_can_initialize_large_subset[MiniMaxVL01ForConditionalGeneration], tests/models/test_initialization.py::test_can_initialize_large_subset[Gemma3ForConditionalGeneration], tests/models/test_initialization.py::test_can_initialize_large_subset[GraniteMoeHybridForCausalLM], tests/models/test_initialization.py::test_can_initialize_large_subset[Grok1ModelForCausalLM], tests/models/test_initialization.py::test_can_initialize_large_subset[GteNewModel], tests/models/test_initialization.py::test_can_initialize_large_subset[XverseForCausalLM], tests/models/test_initialization.py::test_can_initialize_large_subset[MixtralForCausalLM], tests/models/test_initialization.py::test_can_initialize_large_subset[NVLM_D], tests/models/test_initialization.py::test_can_initialize_large_subset[GraniteMoeSharedForCausalLM], tests/models/test_initialization.py::test_can_initialize_large_subset[InternVLForConditionalGeneration], tests/models/test_initialization.py::test_can_initialize_large_subset[NemotronHForCausalLM], tests/models/test_initialization.py::test_can_initialize_large_subset[NemotronForCausalLM], tests/models/test_initialization.py::test_can_initialize_large_subset[ExaoneForCausalLM], tests/models/test_initialization.py::test_can_initialize_large_subset[TarsierForConditionalGeneration], tests/models/test_initialization.py::test_can_initialize_large_subset[MistralForCausalLM], tests/models/test_initialization.py::test_can_initialize_large_subset[ModernBertForSequenceClassification], tests/models/test_initialization.py::test_can_initialize_large_subset[Qwen2ForCausalLM], tests/models/test_initialization.py::test_can_initialize_large_subset[MPTForCausalLM], tests/models/test_initialization.py::test_can_initialize_large_subset[Starcoder2ForCausalLM], tests/models/test_initialization.py::test_can_initialize_large_subset[Qwen2_5_VLForConditionalGeneration], tests/models/test_initialization.py::test_can_initialize_large_subset[DeepseekV32ForCausalLM], tests/models/test_initialization.py::test_can_initialize_large_subset[PhiMoEForCausalLM], tests/models/test_initialization.py::test_can_initialize_large_subset[Qwen2MoeForCausalLM], tests/models/test_initialization.py::test_can_initialize_large_subset[DotsOCRForCausalLM], tests/models/test_initialization.py::test_can_initialize_large_subset[Qwen3NextMTP], tests/models/test_initialization.py::test_can_initialize_large_subset[BaichuanForCausalLM], tests/models/test_initialization.py::test_can_initialize_large_subset[Qwen2ForRewardModel], tests/models/test_initialization.py::test_can_initialize_large_subset[MiniCPMO], tests/models/test_initialization.py::test_can_initialize_large_subset[DeepseekForCausalLM], tests/models/test_initialization.py::test_can_initialize_large_subset[LlamaForCausalLM], tests/models/test_initialization.py::test_can_initialize_large_subset[Ernie4_5_MoeForCausalLM], tests/models/test_initialization.py::test_can_initialize_large_subset[MistralModel], tests/models/test_initialization.py::test_can_initialize_large_subset[GritLM], tests/models/test_initialization.py::test_can_initialize_large_subset[MiniCPM3ForCausalLM], tests/models/test_initialization.py::test_can_initialize_large_subset[SmolLM3ForCausalLM], tests/models/test_initialization.py::test_can_initialize_large_subset[BambaForCausalLM], tests/models/test_initialization.py::test_can_initialize_large_subset[BertForTokenClassification], tests/models/test_initialization.py::test_can_initialize_large_subset[Ovis2_5], tests/models/test_initialization.py::test_can_initialize_large_subset[EagleMiniCPMForCausalLM], tests/models/test_initialization.py::test_can_initialize_large_subset[Tarsier2ForConditionalGeneration], tests/models/test_initialization.py::test_can_initialize_large_subset[Glm4vMoeForConditionalGeneration], tests/models/test_initialization.py::test_can_initialize_large_subset[MiniMaxM1ForCausalLM], tests/models/test_initialization.py::test_can_initialize_large_subset[GptOssForCausalLM], tests/models/test_initialization.py::test_can_initialize_large_subset[Eagle3LlamaForCausalLM], tests/models/test_initialization.py::test_can_initialize_large_subset[Cohere2ForCausalLM], tests/models/test_initialization.py::test_can_initialize_large_subset[GPT2ForSequenceClassification], tests/models/test_initialization.py::test_can_initialize_large_subset[Qwen3VLForConditionalGeneration], tests/models/test_initialization.py::test_can_initialize_large_subset[DbrxForCausalLM], tests/models/test_initialization.py::test_can_initialize_large_subset[AriaForConditionalGeneration], tests/models/test_initialization.py::test_can_initialize_large_subset[LlavaNextForConditionalGeneration], tests/models/test_initialization.py::test_can_initialize_large_subset[ChatGLMForConditionalGeneration], tests/models/test_initialization.py::test_can_initialize_large_subset[ArcticForCausalLM], tests/models/test_initialization.py::test_can_initialize_large_subset[Lfm2ForCausalLM], tests/models/test_initialization.py::test_can_initialize_large_subset[H2OVLChatModel], tests/models/test_initialization.py::test_can_initialize_large_subset[SeedOssForCausalLM], tests/models/test_initialization.py::test_can_initialize_large_subset[MiMoMTPModel], tests/models/test_initialization.py::test_can_initialize_large_subset[OlmoForCausalLM], tests/models/test_initialization.py::test_can_initialize_large_subset[QwenVLForConditionalGeneration], tests/models/test_initialization.py::test_can_initialize_large_subset[Glm4MoeForCausalLM], tests/models/test_initialization.py::test_can_initialize_large_subset[StableLmForCausalLM], tests/models/test_initialization.py::test_can_initialize_large_subset[Phi3VForCausalLM], tests/models/test_initialization.py::test_can_initialize_large_subset[TransformersMoEEmbeddingModel], tests/models/test_initialization.py::test_can_initialize_large_subset[Gemma3ForCausalLM], tests/models/test_initialization.py::test_can_initialize_large_subset[CwmForCausalLM], tests/models/test_initialization.py::test_can_initialize_large_subset[GPTNeoXForCausalLM], tests/models/test_initialization.py::test_can_initialize_large_subset[ModernBertForTokenClassification], tests/models/test_initialization.py::test_can_initialize_large_subset[FalconForCausalLM], tests/models/test_initialization.py::test_can_initialize_large_subset[Phi4MMForCausalLM], tests/models/test_initialization.py::test_can_initialize_large_subset[TeleFLMForCausalLM], tests/models/test_initialization.py::test_can_initialize_large_subset[Cohere2VisionForConditionalGeneration], tests/models/test_initialization.py::test_can_initialize_large_subset[Olmo2ForCausalLM], tests/models/test_initialization.py::test_can_initialize_large_subset[OlmoeForCausalLM], tests/models/test_initialization.py::test_can_initialize_large_subset[MolmoForCausalLM], tests/models/test_initialization.py::test_can_initialize_large_subset[MedusaModel], tests/models/test_initialization.py::test_can_initialize_large_subset[GraniteForCausalLM], tests/models/test_initialization.py::test_can_initialize_large_subset[Olmo3ForCausalLM], tests/models/test_initialization.py::test_can_initialize_large_subset[StableLMEpochForCausalLM], tests/models/test_initialization.py::test_can_initialize_large_subset[HunYuanMoEV1ForCausalLM], tests/models/test_initialization.py::test_can_initialize_large_subset[AyaVisionForConditionalGeneration], tests/models/test_initialization.py::test_can_initialize_large_subset[LongCatFlashMTPModel], tests/models/test_initialization.py::test_can_initialize_large_subset[Qwen3NextForCausalLM], tests/models/test_initialization.py::test_can_initialize_large_subset[Llama_Nemotron_Nano_VL], tests/models/test_initialization.py::test_can_initialize_large_subset[NemotronH_Nano_VL_V2], tests/models/test_initialization.py::test_can_initialize_large_subset[InternS1ForConditionalGeneration], tests/models/test_initialization.py::test_can_initialize_large_subset[Phi4MultimodalForCausalLM], tests/models/test_initialization.py::test_can_initialize_large_subset[Fairseq2LlamaForCausalLM], tests/models/test_initialization.py::test_can_initialize_large_subset[HunYuanDenseV1ForCausalLM], tests/models/test_initialization.py::test_can_initialize_large_subset[DeepseekVLV2ForCausalLM], tests/models/test_initialization.py::test_can_initialize_large_subset[FalconH1ForCausalLM], tests/models/test_initialization.py::test_can_initialize_large_subset[Emu3ForConditionalGeneration], tests/models/test_initialization.py::test_can_initialize_large_subset[BloomForCausalLM], tests/models/test_initialization.py::test_can_initialize_large_subset[LLaMAForCausalLM], tests/models/test_initialization.py::test_can_initialize_large_subset[PersimmonForCausalLM], tests/models/test_initialization.py::test_can_initialize_large_subset[Terratorch], tests/models/test_initialization.py::test_can_initialize_large_subset[Qwen2AudioForConditionalGeneration], tests/models/test_initialization.py::test_can_initialize_large_subset[RobertaForSequenceClassification], tests/models/test_initialization.py::test_can_initialize_large_subset[FuyuForCausalLM], tests/models/test_initialization.py::test_can_initialize_large_subset[BertModel], tests/models/test_initialization.py::test_can_initialize_large_subset[VoxtralForConditionalGeneration], tests/models/test_initialization.py::test_can_initialize_large_subset[TeleChat2ForCausalLM], tests/models/test_initialization.py::test_can_initialize_large_subset[Qwen3VLMoeForConditionalGeneration], tests/models/test_initialization.py::test_can_initialize_large_subset[Dots1ForCausalLM], tests/models/test_initialization.py::test_can_initialize_large_subset[EagleDeepSeekMTPModel], tests/models/test_initialization.py::test_can_initialize_large_subset[Zamba2ForCausalLM], tests/models/test_initialization.py::test_can_initialize_large_subset[CLIPModel], tests/models/test_initialization.py::test_can_initialize_large_subset[ErnieMTPModel], tests/models/test_initialization.py::test_can_initialize_large_subset[TransformersMoEForCausalLM], tests/models/test_initialization.py::test_can_initialize_large_subset[GraniteSpeechForConditionalGeneration], tests/models/test_initialization.py::test_can_initialize_large_subset[Qwen2_5OmniForConditionalGeneration], tests/models/test_initialization.py::test_can_initialize_large_subset[RForConditionalGeneration], tests/models/test_initialization.py::test_can_initialize_large_subset[Qwen3ForCausalLM], tests/models/test_initialization.py::test_can_initialize_large_subset[NomicBertModel], tests/models/test_initialization.py::test_can_initialize_large_subset[Qwen2VLForConditionalGeneration], tests/models/test_initialization.py::test_can_initialize_large_subset[JambaForSequenceClassification], tests/models/test_initialization.py::test_can_initialize_large_subset[Glm4MoeMTPModel], tests/models/test_initialization.py::test_can_initialize_large_subset[MambaForCausalLM], tests/models/test_initialization.py::test_can_initialize_large_subset[InternLMForCausalLM], tests/models/test_initialization.py::test_can_initialize_large_subset[XLMRobertaForSequenceClassification], tests/models/test_initialization.py::test_can_initialize_large_subset[Phi3ForCausalLM], tests/models/test_initialization.py::test_can_initialize_large_subset[SolarForCausalLM], tests/models/test_initialization.py::test_can_initialize_large_subset[Gemma2Model], tests/models/test_initialization.py::test_can_initialize_large_subset[Step3VLForConditionalGeneration], tests/models/test_initialization.py::test_can_initialize_large_subset[Gemma3TextModel], tests/models/test_initialization.py::test_can_initialize_large_subset[TransformersEmbeddingModel], tests/models/test_initialization.py::test_can_initialize_large_subset[AquilaModel], tests/models/test_initialization.py::test_can_initialize_large_subset[LongcatFlashForCausalLM], tests/models/test_initialization.py::test_can_initialize_large_subset[InternLM3ForCausalLM], tests/models/test_initialization.py::test_can_initialize_large_subset[JambaForCausalLM], tests/models/test_initialization.py::test_can_initialize_large_subset[PhiForCausalLM], tests/models/test_initialization.py::test_can_initialize_large_subset[GemmaForCausalLM], tests/models/test_initialization.py::test_can_initialize_large_subset[TransformersForCausalLM], tests/models/test_initialization.py::test_can_initialize_large_subset[Glm4ForCausalLM], tests/models/test_initialization.py::test_can_initialize_large_subset[TransformersMoEForMultimodalLM], tests/models/test_initialization.py::test_can_initialize_large_subset[GraniteMoeForCausalLM], tests/models/test_initialization.py::test_can_initialize_large_subset[Ernie4_5_VLMoeForConditionalGeneration], tests/models/test_initialization.py::test_can_initialize_large_subset[Plamo2ForCausalLM], tests/models/test_initialization.py::test_can_initialize_large_subset[Qwen2ForProcessRewardModel], tests/models/test_initialization.py::test_can_initialize_large_subset[RobertaForMaskedLM], tests/models/test_initialization.py::test_can_initialize_large_subset[GPTBigCodeForCausalLM], tests/models/test_initialization.py::test_can_initialize_large_subset[PaliGemmaForConditionalGeneration], tests/models/test_initialization.py::test_can_initialize_large_subset[GPTJForCausalLM], tests/models/test_initialization.py::test_can_initialize_large_subset[BaiChuanForCausalLM], tests/models/test_initialization.py::test_implicit_converted_models[GemmaForSequenceClassification], tests/models/test_initialization.py::test_implicit_converted_models[LlamaForSequenceClassification], tests/models/test_initialization.py::test_implicit_converted_models[Qwen2ForSequenceClassification], tests/models/test_initialization.py::test_implicit_converted_models[Qwen3ForSequenceClassification] 2025-11-03T17:16:46.5273897Z 2025-11-03T17:16:46.5310174Z models/test_initialization.py::test_can_initialize_small_subset[LlavaForConditionalGeneration] Fork a new process to run a test 3730 2025-11-03T17:16:46.5320266Z Fork a new process to run a test 0 2025-11-03T17:16:46.5594707Z INFO 11-03 17:16:46 [utils.py:239] non-default args: {'load_format': 'dummy', 'gpu_memory_utilization': 0.8, 'disable_log_stats': True, 'hf_overrides': functools.partial(, model_arch='LlavaForConditionalGeneration', exist_overrides={}, use_original_num_layers=False), 'model_impl': 'vllm', 'model': 'llava-hf/llava-1.5-7b-hf'} 2025-11-03T17:16:46.6430582Z 2025-11-03T17:16:46.6432087Z config.json: 0% 0.00/950 [00:00, 'use_cudagraph': True, 'cudagraph_num_of_warmups': 1, 'cudagraph_capture_sizes': [512, 504, 496, 488, 480, 472, 464, 456, 448, 440, 432, 424, 416, 408, 400, 392, 384, 376, 368, 360, 352, 344, 336, 328, 320, 312, 304, 296, 288, 280, 272, 264, 256, 248, 240, 232, 224, 216, 208, 200, 192, 184, 176, 168, 160, 152, 144, 136, 128, 120, 112, 104, 96, 88, 80, 72, 64, 56, 48, 40, 32, 24, 16, 8, 4, 2, 1], 'cudagraph_copy_inputs': False, 'full_cuda_graph': True, 'use_inductor_graph_partition': False, 'pass_config': {}, 'max_capture_size': 512, 'local_cache_dir': None} 2025-11-03T17:16:55.1382081Z (EngineCore_DP0 pid=3759) /opt/conda/envs/py_3.12/lib/python3.12/site-packages/torch/library.py:356: UserWarning: Warning only once for all operators, other operators may also be overridden. 2025-11-03T17:16:55.1383304Z (EngineCore_DP0 pid=3759) Overriding a previously registered kernel for the same operator and the same dispatch key 2025-11-03T17:16:55.1387005Z (EngineCore_DP0 pid=3759) operator: xformers_python::_fused_allgather_and_linear_impl(Tensor scattered_input, Tensor[] weights, str process_group_name, Tensor(a3!)[] gathered_outputs, SymInt timeout_s, bool _wait, bool _memcpy, Tensor scale_scattered_input, Tensor?[] scales_weights) -> () 2025-11-03T17:16:55.1388308Z (EngineCore_DP0 pid=3759) registered at /opt/conda/envs/py_3.12/lib/python3.12/site-packages/torch/_library/custom_ops.py:922 2025-11-03T17:16:55.1388932Z (EngineCore_DP0 pid=3759) dispatch key: ADInplaceOrView 2025-11-03T17:16:55.1389360Z (EngineCore_DP0 pid=3759) previous kernel: no debug info 2025-11-03T17:16:55.1390405Z (EngineCore_DP0 pid=3759) new kernel: registered at /opt/conda/envs/py_3.12/lib/python3.12/site-packages/torch/_library/custom_ops.py:922 (Triggered internally at /var/lib/jenkins/workspace/aten/src/ATen/core/dispatch/OperatorEntry.cpp:208.) 2025-11-03T17:16:55.1391282Z (EngineCore_DP0 pid=3759) self.m.impl( 2025-11-03T17:16:55.5664468Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:16:55.5693005Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:16:55.5700034Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:16:55.5707593Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:16:55.5714929Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:16:55.5722780Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:16:55.5725061Z (EngineCore_DP0 pid=3759) INFO 11-03 17:16:55 [parallel_state.py:1231] rank 0 in world size 1 is assigned as DP rank 0, PP rank 0, TP rank 0, EP rank 0 2025-11-03T17:16:55.8021696Z (EngineCore_DP0 pid=3759) INFO 11-03 17:16:55 [topk_topp_sampler.py:59] Using FlashInfer for top-p & top-k sampling. 2025-11-03T17:16:55.9291953Z (EngineCore_DP0 pid=3759) 2025-11-03T17:16:55.9292403Z processor_config.json: 0% 0.00/173 [00:00, model_arch='Llama4ForConditionalGeneration', exist_overrides={}, use_original_num_layers=False), 'model_impl': 'vllm', 'model': 'meta-llama/Llama-4-Scout-17B-16E-Instruct'} 2025-11-03T17:17:06.0762091Z 2025-11-03T17:17:06.0763073Z config.json: 0% 0.00/2.18k [00:00, 'use_cudagraph': True, 'cudagraph_num_of_warmups': 1, 'cudagraph_capture_sizes': [512, 504, 496, 488, 480, 472, 464, 456, 448, 440, 432, 424, 416, 408, 400, 392, 384, 376, 368, 360, 352, 344, 336, 328, 320, 312, 304, 296, 288, 280, 272, 264, 256, 248, 240, 232, 224, 216, 208, 200, 192, 184, 176, 168, 160, 152, 144, 136, 128, 120, 112, 104, 96, 88, 80, 72, 64, 56, 48, 40, 32, 24, 16, 8, 4, 2, 1], 'cudagraph_copy_inputs': False, 'full_cuda_graph': True, 'use_inductor_graph_partition': False, 'pass_config': {}, 'max_capture_size': 512, 'local_cache_dir': None} 2025-11-03T17:17:16.1778160Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:17:16.1806607Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:17:16.1814147Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:17:16.1822065Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:17:16.1830296Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:17:16.1838017Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:17:16.1840745Z (EngineCore_DP0 pid=4100) INFO 11-03 17:17:16 [parallel_state.py:1231] rank 0 in world size 1 is assigned as DP rank 0, PP rank 0, TP rank 0, EP rank 0 2025-11-03T17:17:16.5292731Z (EngineCore_DP0 pid=4100) INFO 11-03 17:17:16 [topk_topp_sampler.py:59] Using FlashInfer for top-p & top-k sampling. 2025-11-03T17:17:16.7336013Z (EngineCore_DP0 pid=4100) 2025-11-03T17:17:16.7336453Z processor_config.json: 0% 0.00/128 [00:00, model_arch='BertForSequenceClassification', exist_overrides={}, use_original_num_layers=False), 'model_impl': 'vllm', 'model': 'cross-encoder/ms-marco-MiniLM-L-6-v2'} 2025-11-03T17:17:26.4838544Z 2025-11-03T17:17:26.4838882Z config.json: 0% 0.00/794 [00:00, 'use_cudagraph': True, 'cudagraph_num_of_warmups': 1, 'cudagraph_capture_sizes': [512, 504, 496, 488, 480, 472, 464, 456, 448, 440, 432, 424, 416, 408, 400, 392, 384, 376, 368, 360, 352, 344, 336, 328, 320, 312, 304, 296, 288, 280, 272, 264, 256, 248, 240, 232, 224, 216, 208, 200, 192, 184, 176, 168, 160, 152, 144, 136, 128, 120, 112, 104, 96, 88, 80, 72, 64, 56, 48, 40, 32, 24, 16, 8, 4, 2, 1], 'cudagraph_copy_inputs': False, 'full_cuda_graph': False, 'use_inductor_graph_partition': False, 'pass_config': {}, 'max_capture_size': 512, 'local_cache_dir': None} 2025-11-03T17:17:28.4725262Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:17:28.4756458Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:17:28.4764890Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:17:28.4773834Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:17:28.4782119Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:17:28.4790859Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:17:28.4793531Z (EngineCore_DP0 pid=4190) INFO 11-03 17:17:28 [parallel_state.py:1231] rank 0 in world size 1 is assigned as DP rank 0, PP rank 0, TP rank 0, EP rank 0 2025-11-03T17:17:28.8174435Z (EngineCore_DP0 pid=4190) INFO 11-03 17:17:28 [topk_topp_sampler.py:59] Using FlashInfer for top-p & top-k sampling. 2025-11-03T17:17:28.8276155Z (EngineCore_DP0 pid=4190) INFO 11-03 17:17:28 [gpu_model_runner.py:2840] Starting to load model cross-encoder/ms-marco-MiniLM-L-6-v2... 2025-11-03T17:17:29.0791062Z (EngineCore_DP0 pid=4190) INFO 11-03 17:17:29 [gpu_model_runner.py:2870] Loading model from scratch... 2025-11-03T17:17:29.0836435Z (EngineCore_DP0 pid=4190) INFO 11-03 17:17:29 [cuda.py:404] Using Flash Attention backend on V1 engine. 2025-11-03T17:17:29.6603307Z (EngineCore_DP0 pid=4190) INFO 11-03 17:17:29 [gpu_model_runner.py:2902] Model loading took 0.0263 GiB and 0.066128 seconds 2025-11-03T17:17:29.7986599Z (EngineCore_DP0 pid=4190) INFO 11-03 17:17:29 [core.py:142] Disabling chunked prefill for model without KVCache 2025-11-03T17:17:29.8013748Z (EngineCore_DP0 pid=4190) INFO 11-03 17:17:29 [vllm.py:433] Only "last" pooling supports chunked prefill and prefix caching; disabling both. 2025-11-03T17:17:30.0668049Z (EngineCore_DP0 pid=4190) INFO 11-03 17:17:30 [gc_utils.py:40] GC Debug Config. enabled:False,top_objects:-1 2025-11-03T17:17:30.0681812Z INFO 11-03 17:17:30 [llm.py:337] Supported tasks: ['score', 'classify', 'encode'] 2025-11-03T17:17:30.4849142Z PASSED 2025-11-03T17:17:30.4992106Z models/test_initialization.py::test_can_initialize_small_subset[Gemma3nForCausalLM] Fork a new process to run a test 4256 2025-11-03T17:17:30.5001896Z Fork a new process to run a test 0 2025-11-03T17:17:30.5278759Z INFO 11-03 17:17:30 [utils.py:239] non-default args: {'load_format': 'dummy', 'gpu_memory_utilization': 0.8, 'disable_log_stats': True, 'hf_overrides': functools.partial(, model_arch='Gemma3nForCausalLM', exist_overrides={}, use_original_num_layers=False), 'model_impl': 'vllm', 'model': 'google/gemma-3n-E2B-it'} 2025-11-03T17:17:33.5727767Z 2025-11-03T17:17:33.5728220Z config.json: 0% 0.00/4.25k [00:00, 'use_cudagraph': True, 'cudagraph_num_of_warmups': 1, 'cudagraph_capture_sizes': [512, 504, 496, 488, 480, 472, 464, 456, 448, 440, 432, 424, 416, 408, 400, 392, 384, 376, 368, 360, 352, 344, 336, 328, 320, 312, 304, 296, 288, 280, 272, 264, 256, 248, 240, 232, 224, 216, 208, 200, 192, 184, 176, 168, 160, 152, 144, 136, 128, 120, 112, 104, 96, 88, 80, 72, 64, 56, 48, 40, 32, 24, 16, 8, 4, 2, 1], 'cudagraph_copy_inputs': False, 'full_cuda_graph': True, 'use_inductor_graph_partition': False, 'pass_config': {}, 'max_capture_size': 512, 'local_cache_dir': None} 2025-11-03T17:17:45.4025249Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:17:45.4053875Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:17:45.4061308Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:17:45.4069724Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:17:45.4077632Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:17:45.4085502Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:17:45.4088270Z (EngineCore_DP0 pid=4335) INFO 11-03 17:17:45 [parallel_state.py:1231] rank 0 in world size 1 is assigned as DP rank 0, PP rank 0, TP rank 0, EP rank 0 2025-11-03T17:17:45.7751936Z (EngineCore_DP0 pid=4335) INFO 11-03 17:17:45 [topk_topp_sampler.py:59] Using FlashInfer for top-p & top-k sampling. 2025-11-03T17:17:45.9735434Z (EngineCore_DP0 pid=4335) 2025-11-03T17:17:45.9737356Z processor_config.json: 0% 0.00/98.0 [00:00, model_arch='JinaVLForRanking', exist_overrides={}, use_original_num_layers=False), 'model_impl': 'vllm', 'model': 'jinaai/jina-reranker-m0'} 2025-11-03T17:18:01.3562341Z 2025-11-03T17:18:01.3564952Z config.json: 0.00B [00:00, ?B/s] 2025-11-03T17:18:01.3565268Z config.json: 1.16kB [00:00, 5.45MB/s] 2025-11-03T17:18:01.5075547Z 2025-11-03T17:18:01.5076036Z preprocessor_config.json: 0% 0.00/316 [00:00, 'use_cudagraph': True, 'cudagraph_num_of_warmups': 1, 'cudagraph_capture_sizes': [512, 504, 496, 488, 480, 472, 464, 456, 448, 440, 432, 424, 416, 408, 400, 392, 384, 376, 368, 360, 352, 344, 336, 328, 320, 312, 304, 296, 288, 280, 272, 264, 256, 248, 240, 232, 224, 216, 208, 200, 192, 184, 176, 168, 160, 152, 144, 136, 128, 120, 112, 104, 96, 88, 80, 72, 64, 56, 48, 40, 32, 24, 16, 8, 4, 2, 1], 'cudagraph_copy_inputs': False, 'full_cuda_graph': False, 'use_inductor_graph_partition': False, 'pass_config': {}, 'max_capture_size': 512, 'local_cache_dir': None} 2025-11-03T17:18:10.3759121Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:18:10.3787967Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:18:10.3795472Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:18:10.3803378Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:18:10.3811269Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:18:10.3819182Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:18:10.3822957Z (EngineCore_DP0 pid=4512) INFO 11-03 17:18:10 [parallel_state.py:1231] rank 0 in world size 1 is assigned as DP rank 0, PP rank 0, TP rank 0, EP rank 0 2025-11-03T17:18:10.7126210Z (EngineCore_DP0 pid=4512) INFO 11-03 17:18:10 [topk_topp_sampler.py:59] Using FlashInfer for top-p & top-k sampling. 2025-11-03T17:18:12.2655366Z (EngineCore_DP0 pid=4512) 2025-11-03T17:18:12.2656476Z chat_template.json: 0.00B [00:00, ?B/s] 2025-11-03T17:18:12.2656891Z chat_template.json: 1.05kB [00:00, 3.28MB/s] 2025-11-03T17:18:12.8279437Z (EngineCore_DP0 pid=4512) INFO 11-03 17:18:12 [gpu_model_runner.py:2840] Starting to load model jinaai/jina-reranker-m0... 2025-11-03T17:18:13.1695318Z (EngineCore_DP0 pid=4512) INFO 11-03 17:18:13 [gpu_model_runner.py:2870] Loading model from scratch... 2025-11-03T17:18:13.2397118Z (EngineCore_DP0 pid=4512) INFO 11-03 17:18:13 [cuda.py:404] Using Flash Attention backend on V1 engine. 2025-11-03T17:18:13.8111571Z (EngineCore_DP0 pid=4512) INFO 11-03 17:18:13 [gpu_model_runner.py:2902] Model loading took 1.8029 GiB and 0.136192 seconds 2025-11-03T17:18:13.8115296Z (EngineCore_DP0 pid=4512) INFO 11-03 17:18:13 [kv_cache_utils.py:1199] GPU KV cache size: 10,485,760 tokens 2025-11-03T17:18:13.8116705Z (EngineCore_DP0 pid=4512) INFO 11-03 17:18:13 [kv_cache_utils.py:1204] Maximum concurrency for 32,768 tokens per request: 320.00x 2025-11-03T17:18:15.4197560Z (EngineCore_DP0 pid=4512) INFO 11-03 17:18:15 [gc_utils.py:40] GC Debug Config. enabled:False,top_objects:-1 2025-11-03T17:18:15.4210891Z INFO 11-03 17:18:15 [llm.py:337] Supported tasks: ['score', 'classify'] 2025-11-03T17:18:15.9072312Z PASSED 2025-11-03T17:18:15.9215618Z models/test_initialization.py::test_can_initialize_small_subset[InternVLChatModel] Fork a new process to run a test 4578 2025-11-03T17:18:15.9226486Z Fork a new process to run a test 0 2025-11-03T17:18:15.9498459Z INFO 11-03 17:18:15 [utils.py:239] non-default args: {'trust_remote_code': True, 'load_format': 'dummy', 'gpu_memory_utilization': 0.8, 'disable_log_stats': True, 'hf_overrides': functools.partial(, model_arch='InternVLChatModel', exist_overrides={}, use_original_num_layers=False), 'model_impl': 'vllm', 'model': 'OpenGVLab/InternVL2-1B'} 2025-11-03T17:18:16.0267372Z 2025-11-03T17:18:16.0272708Z config.json: 0.00B [00:00, ?B/s] 2025-11-03T17:18:16.0273418Z config.json: 3.73kB [00:00, 8.95MB/s] 2025-11-03T17:18:16.1034837Z 2025-11-03T17:18:16.1036441Z configuration_internvl_chat.py: 0.00B [00:00, ?B/s] 2025-11-03T17:18:16.1037168Z configuration_internvl_chat.py: 4.04kB [00:00, 13.9MB/s] 2025-11-03T17:18:16.1457687Z 2025-11-03T17:18:16.1459272Z configuration_intern_vit.py: 0.00B [00:00, ?B/s] 2025-11-03T17:18:16.1459918Z configuration_intern_vit.py: 5.55kB [00:00, 29.2MB/s] 2025-11-03T17:18:16.1480955Z A new version of the following files was downloaded from https://huggingface.co/OpenGVLab/InternVL2-1B: 2025-11-03T17:18:16.1481826Z - configuration_intern_vit.py 2025-11-03T17:18:16.1482746Z . Make sure to double-check they do not contain any added malicious code. To avoid downloading new versions of the code file, you can pin a revision. 2025-11-03T17:18:16.1484188Z A new version of the following files was downloaded from https://huggingface.co/OpenGVLab/InternVL2-1B: 2025-11-03T17:18:16.1485008Z - configuration_internvl_chat.py 2025-11-03T17:18:16.1485288Z - configuration_intern_vit.py 2025-11-03T17:18:16.1485896Z . Make sure to double-check they do not contain any added malicious code. To avoid downloading new versions of the code file, you can pin a revision. 2025-11-03T17:18:16.3836964Z 2025-11-03T17:18:16.3837471Z preprocessor_config.json: 0% 0.00/287 [00:00, 'use_cudagraph': True, 'cudagraph_num_of_warmups': 1, 'cudagraph_capture_sizes': [512, 504, 496, 488, 480, 472, 464, 456, 448, 440, 432, 424, 416, 408, 400, 392, 384, 376, 368, 360, 352, 344, 336, 328, 320, 312, 304, 296, 288, 280, 272, 264, 256, 248, 240, 232, 224, 216, 208, 200, 192, 184, 176, 168, 160, 152, 144, 136, 128, 120, 112, 104, 96, 88, 80, 72, 64, 56, 48, 40, 32, 24, 16, 8, 4, 2, 1], 'cudagraph_copy_inputs': False, 'full_cuda_graph': True, 'use_inductor_graph_partition': False, 'pass_config': {}, 'max_capture_size': 512, 'local_cache_dir': None} 2025-11-03T17:18:24.5233980Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:18:24.5263214Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:18:24.5270844Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:18:24.5278737Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:18:24.5286673Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:18:24.5294442Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:18:24.5297431Z (EngineCore_DP0 pid=4636) INFO 11-03 17:18:24 [parallel_state.py:1231] rank 0 in world size 1 is assigned as DP rank 0, PP rank 0, TP rank 0, EP rank 0 2025-11-03T17:18:24.8722344Z (EngineCore_DP0 pid=4636) INFO 11-03 17:18:24 [topk_topp_sampler.py:59] Using FlashInfer for top-p & top-k sampling. 2025-11-03T17:18:24.9330369Z (EngineCore_DP0 pid=4636) WARNING 11-03 17:18:24 [__init__.py:2040] The following intended overrides are not keyword args and will be dropped: {'truncation'} 2025-11-03T17:18:24.9359527Z (EngineCore_DP0 pid=4636) WARNING 11-03 17:18:24 [__init__.py:2040] The following intended overrides are not keyword args and will be dropped: {'truncation'} 2025-11-03T17:18:25.0111202Z (EngineCore_DP0 pid=4636) INFO 11-03 17:18:25 [gpu_model_runner.py:2840] Starting to load model OpenGVLab/InternVL2-1B... 2025-11-03T17:18:25.2687328Z (EngineCore_DP0 pid=4636) INFO 11-03 17:18:25 [gpu_model_runner.py:2870] Loading model from scratch... 2025-11-03T17:18:25.2731392Z (EngineCore_DP0 pid=4636) INFO 11-03 17:18:25 [layer.py:497] MultiHeadAttention attn_backend: _Backend.FLASH_ATTN, use_upstream_fa: False 2025-11-03T17:18:25.3348768Z (EngineCore_DP0 pid=4636) INFO 11-03 17:18:25 [cuda.py:404] Using Flash Attention backend on V1 engine. 2025-11-03T17:18:25.9188998Z (EngineCore_DP0 pid=4636) INFO 11-03 17:18:25 [gpu_model_runner.py:2902] Model loading took 0.5767 GiB and 0.126955 seconds 2025-11-03T17:18:25.9194489Z (EngineCore_DP0 pid=4636) INFO 11-03 17:18:25 [kv_cache_utils.py:1199] GPU KV cache size: 20,971,520 tokens 2025-11-03T17:18:25.9195597Z (EngineCore_DP0 pid=4636) INFO 11-03 17:18:25 [kv_cache_utils.py:1204] Maximum concurrency for 32,768 tokens per request: 640.00x 2025-11-03T17:18:26.3424785Z (EngineCore_DP0 pid=4636) WARNING 11-03 17:18:26 [__init__.py:2040] The following intended overrides are not keyword args and will be dropped: {'truncation'} 2025-11-03T17:18:28.5480942Z (EngineCore_DP0 pid=4636) INFO 11-03 17:18:28 [gc_utils.py:40] GC Debug Config. enabled:False,top_objects:-1 2025-11-03T17:18:28.5495135Z INFO 11-03 17:18:28 [llm.py:337] Supported tasks: ['generate'] 2025-11-03T17:18:29.0208512Z PASSED 2025-11-03T17:18:29.0351381Z models/test_initialization.py::test_can_initialize_small_subset[InternLM2ForRewardModel] Fork a new process to run a test 4718 2025-11-03T17:18:29.0361377Z Fork a new process to run a test 0 2025-11-03T17:18:29.0641489Z INFO 11-03 17:18:29 [utils.py:239] non-default args: {'trust_remote_code': True, 'load_format': 'dummy', 'gpu_memory_utilization': 0.8, 'disable_log_stats': True, 'hf_overrides': functools.partial(, model_arch='InternLM2ForRewardModel', exist_overrides={}, use_original_num_layers=False), 'model_impl': 'vllm', 'model': 'internlm/internlm2-1_8b-reward'} 2025-11-03T17:18:29.1401392Z 2025-11-03T17:18:29.1403105Z config.json: 0% 0.00/813 [00:00, 'use_cudagraph': True, 'cudagraph_num_of_warmups': 1, 'cudagraph_capture_sizes': [512, 504, 496, 488, 480, 472, 464, 456, 448, 440, 432, 424, 416, 408, 400, 392, 384, 376, 368, 360, 352, 344, 336, 328, 320, 312, 304, 296, 288, 280, 272, 264, 256, 248, 240, 232, 224, 216, 208, 200, 192, 184, 176, 168, 160, 152, 144, 136, 128, 120, 112, 104, 96, 88, 80, 72, 64, 56, 48, 40, 32, 24, 16, 8, 4, 2, 1], 'cudagraph_copy_inputs': False, 'full_cuda_graph': False, 'use_inductor_graph_partition': False, 'pass_config': {}, 'max_capture_size': 512, 'local_cache_dir': None} 2025-11-03T17:18:38.0856274Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:18:38.0884321Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:18:38.0892558Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:18:38.0901183Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:18:38.0909871Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:18:38.0918120Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:18:38.0920935Z (EngineCore_DP0 pid=4797) INFO 11-03 17:18:38 [parallel_state.py:1231] rank 0 in world size 1 is assigned as DP rank 0, PP rank 0, TP rank 0, EP rank 0 2025-11-03T17:18:38.4294264Z (EngineCore_DP0 pid=4797) INFO 11-03 17:18:38 [topk_topp_sampler.py:59] Using FlashInfer for top-p & top-k sampling. 2025-11-03T17:18:38.5097914Z (EngineCore_DP0 pid=4797) INFO 11-03 17:18:38 [gpu_model_runner.py:2840] Starting to load model internlm/internlm2-1_8b-reward... 2025-11-03T17:18:38.7634340Z (EngineCore_DP0 pid=4797) INFO 11-03 17:18:38 [gpu_model_runner.py:2870] Loading model from scratch... 2025-11-03T17:18:38.8281252Z (EngineCore_DP0 pid=4797) INFO 11-03 17:18:38 [cuda.py:404] Using Flash Attention backend on V1 engine. 2025-11-03T17:18:39.3920880Z (EngineCore_DP0 pid=4797) INFO 11-03 17:18:39 [gpu_model_runner.py:2902] Model loading took 0.4785 GiB and 0.128110 seconds 2025-11-03T17:18:39.3924727Z (EngineCore_DP0 pid=4797) INFO 11-03 17:18:39 [kv_cache_utils.py:1199] GPU KV cache size: 2,621,440 tokens 2025-11-03T17:18:39.3926165Z (EngineCore_DP0 pid=4797) INFO 11-03 17:18:39 [kv_cache_utils.py:1204] Maximum concurrency for 32,768 tokens per request: 80.00x 2025-11-03T17:18:40.5001267Z (EngineCore_DP0 pid=4797) INFO 11-03 17:18:40 [vllm.py:433] Only "last" pooling supports chunked prefill and prefix caching; disabling both. 2025-11-03T17:18:40.7743391Z (EngineCore_DP0 pid=4797) INFO 11-03 17:18:40 [gc_utils.py:40] GC Debug Config. enabled:False,top_objects:-1 2025-11-03T17:18:40.7756301Z INFO 11-03 17:18:40 [llm.py:337] Supported tasks: ['encode'] 2025-11-03T17:18:41.2446028Z PASSED 2025-11-03T17:18:41.2588542Z models/test_initialization.py::test_can_initialize_small_subset[TransformersForMultimodalLM] Fork a new process to run a test 4863 2025-11-03T17:18:41.2598363Z Fork a new process to run a test 0 2025-11-03T17:18:41.2877776Z INFO 11-03 17:18:41 [utils.py:239] non-default args: {'load_format': 'dummy', 'gpu_memory_utilization': 0.8, 'disable_log_stats': True, 'hf_overrides': functools.partial(, model_arch='TransformersForMultimodalLM', exist_overrides={}, use_original_num_layers=False), 'model_impl': 'transformers', 'model': 'BAAI/Emu3-Chat-hf'} 2025-11-03T17:18:41.4574008Z 2025-11-03T17:18:41.4806292Z config.json: 0.00B [00:00, ?B/s] 2025-11-03T17:18:41.4806799Z config.json: 6.55MB [00:00, 283MB/s] 2025-11-03T17:18:41.9338142Z 2025-11-03T17:18:41.9338798Z preprocessor_config.json: 0% 0.00/611 [00:00, 'use_cudagraph': True, 'cudagraph_num_of_warmups': 1, 'cudagraph_capture_sizes': [512, 504, 496, 488, 480, 472, 464, 456, 448, 440, 432, 424, 416, 408, 400, 392, 384, 376, 368, 360, 352, 344, 336, 328, 320, 312, 304, 296, 288, 280, 272, 264, 256, 248, 240, 232, 224, 216, 208, 200, 192, 184, 176, 168, 160, 152, 144, 136, 128, 120, 112, 104, 96, 88, 80, 72, 64, 56, 48, 40, 32, 24, 16, 8, 4, 2, 1], 'cudagraph_copy_inputs': False, 'full_cuda_graph': True, 'use_inductor_graph_partition': False, 'pass_config': {}, 'max_capture_size': 512, 'local_cache_dir': None} 2025-11-03T17:19:11.1514310Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:19:11.1542225Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:19:11.1549161Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:19:11.1557870Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:19:11.1566688Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:19:11.1574519Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:19:11.1577306Z (EngineCore_DP0 pid=4943) INFO 11-03 17:19:11 [parallel_state.py:1231] rank 0 in world size 1 is assigned as DP rank 0, PP rank 0, TP rank 0, EP rank 0 2025-11-03T17:19:11.4918466Z (EngineCore_DP0 pid=4943) INFO 11-03 17:19:11 [topk_topp_sampler.py:59] Using FlashInfer for top-p & top-k sampling. 2025-11-03T17:19:11.6369106Z (EngineCore_DP0 pid=4943) 2025-11-03T17:19:11.6370254Z processor_config.json: 0% 0.00/69.0 [00:00, model_arch='PrithviGeoSpatialMAE', exist_overrides={}, use_original_num_layers=False), 'enforce_eager': True, 'model_impl': 'vllm', 'model': 'ibm-nasa-geospatial/Prithvi-EO-2.0-300M-TL-Sen1Floods11'} 2025-11-03T17:19:57.2272323Z 2025-11-03T17:19:57.2275404Z config.json: 0.00B [00:00, ?B/s] 2025-11-03T17:19:57.2275784Z config.json: 3.45kB [00:00, 9.62MB/s] 2025-11-03T17:20:07.8851720Z INFO 11-03 17:20:07 [model.py:915] Resolved `--runner auto` to `--runner pooling`. Pass the value explicitly to silence this message. 2025-11-03T17:20:07.8852778Z INFO 11-03 17:20:07 [model.py:653] Resolved architecture: Terratorch 2025-11-03T17:20:07.8853213Z `torch_dtype` is deprecated! Use `dtype` instead! 2025-11-03T17:20:07.9337998Z ERROR 11-03 17:20:07 [config.py:298] Error retrieving safetensors: 'ibm-nasa-geospatial/Prithvi-EO-2.0-300M-TL-Sen1Floods11' is not a safetensors repo. Couldn't find 'model.safetensors.index.json' or 'model.safetensors' files., retrying 1 of 2 2025-11-03T17:20:10.0025706Z ERROR 11-03 17:20:10 [config.py:296] Error retrieving safetensors: 'ibm-nasa-geospatial/Prithvi-EO-2.0-300M-TL-Sen1Floods11' is not a safetensors repo. Couldn't find 'model.safetensors.index.json' or 'model.safetensors' files. 2025-11-03T17:20:10.0028071Z INFO 11-03 17:20:10 [model.py:1939] Downcasting torch.float32 to torch.float16. 2025-11-03T17:20:10.0029785Z WARNING 11-03 17:20:10 [model.py:2042] The model's config.json does not contain any of the following keys to determine the original maximum length of the model: ['max_position_embeddings', 'n_positions', 'max_seq_len', 'seq_length', 'model_max_length', 'max_target_positions', 'max_sequence_length', 'max_seq_length', 'seq_len']. Assuming the model's maximum length is 2048. 2025-11-03T17:20:10.0031080Z INFO 11-03 17:20:10 [model.py:1714] Using max model len 2048 2025-11-03T17:20:10.0346105Z INFO 11-03 17:20:10 [arg_utils.py:1725] (Disabling) chunked prefill by default 2025-11-03T17:20:10.0346591Z INFO 11-03 17:20:10 [arg_utils.py:1728] (Disabling) prefix caching by default 2025-11-03T17:20:10.1852256Z INFO 11-03 17:20:10 [vllm.py:354] Cudagraph is disabled under eager mode 2025-11-03T17:20:10.1852906Z INFO 11-03 17:20:10 [vllm.py:433] Only "last" pooling supports chunked prefill and prefix caching; disabling both. 2025-11-03T17:20:13.7913418Z WARNING 11-03 17:20:13 [__init__.py:2879] We must use the `spawn` multiprocessing start method. Overriding VLLM_WORKER_MULTIPROC_METHOD to 'spawn'. See https://docs.vllm.ai/en/latest/usage/troubleshooting.html#python-multiprocessing for more information. Reasons: CUDA is initialized 2025-11-03T17:20:14.3385204Z /opt/conda/envs/py_3.12/lib/python3.12/site-packages/torch/cuda/__init__.py:65: FutureWarning: The pynvml package is deprecated. Please install nvidia-ml-py instead. If you did not install pynvml directly, please report this to the maintainers of the package that installed pynvml for you. 2025-11-03T17:20:14.3386447Z import pynvml # type: ignore[import] 2025-11-03T17:20:16.5932798Z INFO 11-03 17:20:16 [__init__.py:224] Automatically detected platform cuda. 2025-11-03T17:20:19.1225442Z (EngineCore_DP0 pid=5071) INFO 11-03 17:20:19 [core.py:727] Waiting for init message from front-end. 2025-11-03T17:20:19.1483435Z (EngineCore_DP0 pid=5071) INFO 11-03 17:20:19 [core.py:94] Initializing a V1 LLM engine (v0.11.0rc2.dev389+ge51928192.d20251103) with config: model='ibm-nasa-geospatial/Prithvi-EO-2.0-300M-TL-Sen1Floods11', speculative_config=None, tokenizer='ibm-nasa-geospatial/Prithvi-EO-2.0-300M-TL-Sen1Floods11', skip_tokenizer_init=True, tokenizer_mode=auto, revision=None, tokenizer_revision=None, trust_remote_code=False, dtype=torch.float16, max_seq_len=2048, download_dir=None, load_format=dummy, tensor_parallel_size=1, pipeline_parallel_size=1, data_parallel_size=1, disable_custom_all_reduce=False, quantization=None, enforce_eager=True, kv_cache_dtype=auto, device_config=cuda, structured_outputs_config=StructuredOutputsConfig(backend='auto', disable_fallback=False, disable_any_whitespace=False, disable_additional_properties=False, reasoning_parser=''), observability_config=ObservabilityConfig(show_hidden_metrics_for_version=None, otlp_traces_endpoint=None, collect_detailed_traces=None), seed=0, served_model_name=ibm-nasa-geospatial/Prithvi-EO-2.0-300M-TL-Sen1Floods11, enable_prefix_caching=False, chunked_prefill_enabled=False, pooler_config=PoolerConfig(pooling_type='All', normalize=None, dimensions=None, enable_chunked_processing=None, max_embed_len=None, activation=None, logit_bias=None, softmax=None, step_tag_id=None, returned_token_ids=None), compilation_config={'level': 0, 'debug_dump_path': None, 'cache_dir': '', 'backend': '', 'custom_ops': [], 'splitting_ops': None, 'use_inductor': True, 'compile_sizes': [], 'inductor_compile_config': {'enable_auto_functionalized_v2': False}, 'inductor_passes': {}, 'cudagraph_mode': , 'use_cudagraph': False, 'cudagraph_num_of_warmups': 0, 'cudagraph_capture_sizes': [], 'cudagraph_copy_inputs': False, 'full_cuda_graph': False, 'use_inductor_graph_partition': False, 'pass_config': {}, 'max_capture_size': 0, 'local_cache_dir': None} 2025-11-03T17:20:22.8715947Z (EngineCore_DP0 pid=5071) /opt/conda/envs/py_3.12/lib/python3.12/site-packages/torch/library.py:356: UserWarning: Warning only once for all operators, other operators may also be overridden. 2025-11-03T17:20:22.8717262Z (EngineCore_DP0 pid=5071) Overriding a previously registered kernel for the same operator and the same dispatch key 2025-11-03T17:20:22.8719038Z (EngineCore_DP0 pid=5071) operator: xformers_python::_fused_allgather_and_linear_impl(Tensor scattered_input, Tensor[] weights, str process_group_name, Tensor(a3!)[] gathered_outputs, SymInt timeout_s, bool _wait, bool _memcpy, Tensor scale_scattered_input, Tensor?[] scales_weights) -> () 2025-11-03T17:20:22.8720586Z (EngineCore_DP0 pid=5071) registered at /opt/conda/envs/py_3.12/lib/python3.12/site-packages/torch/_library/custom_ops.py:922 2025-11-03T17:20:22.8721562Z (EngineCore_DP0 pid=5071) dispatch key: ADInplaceOrView 2025-11-03T17:20:22.8722081Z (EngineCore_DP0 pid=5071) previous kernel: no debug info 2025-11-03T17:20:22.8723737Z (EngineCore_DP0 pid=5071) new kernel: registered at /opt/conda/envs/py_3.12/lib/python3.12/site-packages/torch/_library/custom_ops.py:922 (Triggered internally at /var/lib/jenkins/workspace/aten/src/ATen/core/dispatch/OperatorEntry.cpp:208.) 2025-11-03T17:20:22.8724754Z (EngineCore_DP0 pid=5071) self.m.impl( 2025-11-03T17:20:22.9002693Z (EngineCore_DP0 pid=5071) /opt/conda/envs/py_3.12/lib/python3.12/site-packages/torch/__init__.py:1628: UserWarning: Please use the new API settings to control TF32 behavior, such as torch.backends.cudnn.conv.fp32_precision = 'tf32' or torch.backends.cuda.matmul.fp32_precision = 'ieee'. Old settings, e.g, torch.backends.cuda.matmul.allow_tf32 = True, torch.backends.cudnn.allow_tf32 = True, allowTF32CuDNN() and allowTF32CuBLAS() will be deprecated after Pytorch 2.9. Please see https://pytorch.org/docs/main/notes/cuda.html#tensorfloat-32-tf32-on-ampere-and-later-devices (Triggered internally at /var/lib/jenkins/workspace/aten/src/ATen/Context.cpp:45.) 2025-11-03T17:20:22.9005205Z (EngineCore_DP0 pid=5071) _C._set_float32_matmul_precision(precision) 2025-11-03T17:20:23.1618162Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:20:23.1645362Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:20:23.1652581Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:20:23.1660290Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:20:23.1667773Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:20:23.1675146Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:20:23.1677664Z (EngineCore_DP0 pid=5071) INFO 11-03 17:20:23 [parallel_state.py:1231] rank 0 in world size 1 is assigned as DP rank 0, PP rank 0, TP rank 0, EP rank 0 2025-11-03T17:20:23.4814930Z (EngineCore_DP0 pid=5071) INFO 11-03 17:20:23 [topk_topp_sampler.py:59] Using FlashInfer for top-p & top-k sampling. 2025-11-03T17:20:23.4906820Z (EngineCore_DP0 pid=5071) INFO 11-03 17:20:23 [gpu_model_runner.py:2840] Starting to load model ibm-nasa-geospatial/Prithvi-EO-2.0-300M-TL-Sen1Floods11... 2025-11-03T17:20:23.7980233Z (EngineCore_DP0 pid=5071) INFO 11-03 17:20:23 [gpu_model_runner.py:2870] Loading model from scratch... 2025-11-03T17:20:23.8880992Z (EngineCore_DP0 pid=5071) 2025-11-03T17:20:23.8881417Z config.json: 0% 0.00/776 [00:00, model_arch='UltravoxModel', exist_overrides={}, use_original_num_layers=False), 'model_impl': 'vllm', 'model': 'fixie-ai/ultravox-v0_5-llama-3_2-1b'} 2025-11-03T17:20:29.3855303Z 2025-11-03T17:20:29.3857042Z config.json: 0.00B [00:00, ?B/s] 2025-11-03T17:20:29.6412223Z config.json: 2.15kB [00:00, 10.6MB/s] 2025-11-03T17:20:29.6412559Z 2025-11-03T17:20:29.6412794Z preprocessor_config.json: 0% 0.00/339 [00:00, 'use_cudagraph': True, 'cudagraph_num_of_warmups': 1, 'cudagraph_capture_sizes': [512, 504, 496, 488, 480, 472, 464, 456, 448, 440, 432, 424, 416, 408, 400, 392, 384, 376, 368, 360, 352, 344, 336, 328, 320, 312, 304, 296, 288, 280, 272, 264, 256, 248, 240, 232, 224, 216, 208, 200, 192, 184, 176, 168, 160, 152, 144, 136, 128, 120, 112, 104, 96, 88, 80, 72, 64, 56, 48, 40, 32, 24, 16, 8, 4, 2, 1], 'cudagraph_copy_inputs': False, 'full_cuda_graph': True, 'use_inductor_graph_partition': False, 'pass_config': {}, 'max_capture_size': 512, 'local_cache_dir': None} 2025-11-03T17:20:38.5317221Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:20:38.5346841Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:20:38.5353768Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:20:38.5361717Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:20:38.5369373Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:20:38.5377348Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:20:38.5380262Z (EngineCore_DP0 pid=5303) INFO 11-03 17:20:38 [parallel_state.py:1231] rank 0 in world size 1 is assigned as DP rank 0, PP rank 0, TP rank 0, EP rank 0 2025-11-03T17:20:38.8707645Z (EngineCore_DP0 pid=5303) INFO 11-03 17:20:38 [topk_topp_sampler.py:59] Using FlashInfer for top-p & top-k sampling. 2025-11-03T17:20:39.0036918Z (EngineCore_DP0 pid=5303) 2025-11-03T17:20:39.0037321Z processor_config.json: 0% 0.00/270 [00:00, model_arch='DeepSeekMTPModel', exist_overrides={}, use_original_num_layers=False), 'speculative_config': {'model': 'luccafong/deepseek_mtp_draft_random', 'num_speculative_tokens': 1}, 'model_impl': 'vllm', 'model': 'luccafong/deepseek_mtp_main_random'} 2025-11-03T17:20:50.4656862Z 2025-11-03T17:20:50.4659100Z config.json: 0.00B [00:00, ?B/s] 2025-11-03T17:20:50.4659511Z config.json: 1.95kB [00:00, 11.4MB/s] 2025-11-03T17:20:50.5175148Z INFO 11-03 17:20:50 [config.py:408] Replacing legacy 'type' key with 'rope_type' 2025-11-03T17:20:56.9165111Z INFO 11-03 17:20:56 [model.py:653] Resolved architecture: DeepseekV3ForCausalLM 2025-11-03T17:20:56.9166683Z `torch_dtype` is deprecated! Use `dtype` instead! 2025-11-03T17:20:56.9415522Z INFO 11-03 17:20:56 [model.py:1714] Using max model len 163840 2025-11-03T17:20:57.0364931Z The argument `trust_remote_code` is to be used with Auto classes. It has no effect here and is ignored. 2025-11-03T17:20:57.2766832Z 2025-11-03T17:20:57.2769174Z config.json: 0.00B [00:00, ?B/s] 2025-11-03T17:20:57.2769538Z config.json: 1.95kB [00:00, 5.90MB/s] 2025-11-03T17:20:57.3074948Z INFO 11-03 17:20:57 [config.py:408] Replacing legacy 'type' key with 'rope_type' 2025-11-03T17:21:03.6395246Z INFO 11-03 17:21:03 [model.py:653] Resolved architecture: DeepSeekMTPModel 2025-11-03T17:21:03.6396321Z INFO 11-03 17:21:03 [model.py:1714] Using max model len 163840 2025-11-03T17:21:03.6403371Z INFO 11-03 17:21:03 [scheduler.py:225] Chunked prefill is enabled with max_num_batched_tokens=8192. 2025-11-03T17:21:03.7628333Z 2025-11-03T17:21:03.7629952Z tokenizer_config.json: 0.00B [00:00, ?B/s] 2025-11-03T17:21:03.7630325Z tokenizer_config.json: 3.13kB [00:00, 18.8MB/s] 2025-11-03T17:21:03.9023803Z 2025-11-03T17:21:03.9338460Z tokenizer.json: 0.00B [00:00, ?B/s] 2025-11-03T17:21:03.9338936Z tokenizer.json: 7.85MB [00:00, 248MB/s] 2025-11-03T17:21:04.5023908Z INFO 11-03 17:21:04 [config.py:408] Replacing legacy 'type' key with 'rope_type' 2025-11-03T17:21:04.5246589Z (EngineCore_DP0 pid=5493) INFO 11-03 17:21:04 [core.py:727] Waiting for init message from front-end. 2025-11-03T17:21:04.5288442Z (EngineCore_DP0 pid=5493) INFO 11-03 17:21:04 [core.py:94] Initializing a V1 LLM engine (v0.11.0rc2.dev389+ge51928192.d20251103) with config: model='luccafong/deepseek_mtp_main_random', speculative_config=SpeculativeConfig(method='mtp', model='luccafong/deepseek_mtp_draft_random', num_spec_tokens=1), tokenizer='luccafong/deepseek_mtp_main_random', skip_tokenizer_init=False, tokenizer_mode=auto, revision=None, tokenizer_revision=None, trust_remote_code=True, dtype=torch.bfloat16, max_seq_len=163840, download_dir=None, load_format=dummy, tensor_parallel_size=1, pipeline_parallel_size=1, data_parallel_size=1, disable_custom_all_reduce=False, quantization=fp8, enforce_eager=False, kv_cache_dtype=auto, device_config=cuda, structured_outputs_config=StructuredOutputsConfig(backend='auto', disable_fallback=False, disable_any_whitespace=False, disable_additional_properties=False, reasoning_parser=''), observability_config=ObservabilityConfig(show_hidden_metrics_for_version=None, otlp_traces_endpoint=None, collect_detailed_traces=None), seed=0, served_model_name=luccafong/deepseek_mtp_main_random, enable_prefix_caching=True, chunked_prefill_enabled=True, pooler_config=None, compilation_config={'level': 3, 'debug_dump_path': None, 'cache_dir': '', 'backend': '', 'custom_ops': ['+quant_fp8'], 'splitting_ops': ['vllm.unified_attention', 'vllm.unified_attention_with_output', 'vllm.unified_mla_attention', 'vllm.unified_mla_attention_with_output', 'vllm.mamba_mixer2', 'vllm.mamba_mixer', 'vllm.short_conv', 'vllm.linear_attention', 'vllm.plamo2_mamba_mixer', 'vllm.gdn_attention', 'vllm.sparse_attn_indexer'], 'use_inductor': True, 'compile_sizes': [], 'inductor_compile_config': {'enable_auto_functionalized_v2': False}, 'inductor_passes': {}, 'cudagraph_mode': , 'use_cudagraph': True, 'cudagraph_num_of_warmups': 1, 'cudagraph_capture_sizes': [512, 504, 496, 488, 480, 472, 464, 456, 448, 440, 432, 424, 416, 408, 400, 392, 384, 376, 368, 360, 352, 344, 336, 328, 320, 312, 304, 296, 288, 280, 272, 264, 256, 248, 240, 232, 224, 216, 208, 200, 192, 184, 176, 168, 160, 152, 144, 136, 128, 120, 112, 104, 96, 88, 80, 72, 64, 56, 48, 40, 32, 24, 16, 8, 4, 2, 1], 'cudagraph_copy_inputs': False, 'full_cuda_graph': True, 'use_inductor_graph_partition': False, 'pass_config': {}, 'max_capture_size': 512, 'local_cache_dir': None} 2025-11-03T17:21:05.2583230Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:21:05.2610891Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:21:05.2619009Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:21:05.2627294Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:21:05.2635371Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:21:05.2643359Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:21:05.2646168Z (EngineCore_DP0 pid=5493) INFO 11-03 17:21:05 [parallel_state.py:1231] rank 0 in world size 1 is assigned as DP rank 0, PP rank 0, TP rank 0, EP rank 0 2025-11-03T17:21:05.6099874Z (EngineCore_DP0 pid=5493) INFO 11-03 17:21:05 [topk_topp_sampler.py:59] Using FlashInfer for top-p & top-k sampling. 2025-11-03T17:21:05.6246928Z (EngineCore_DP0 pid=5493) WARNING 11-03 17:21:05 [__init__.py:194] min_p, logit_bias, and min_tokens parameters won't currently work with speculative decoding enabled. 2025-11-03T17:21:05.6882874Z (EngineCore_DP0 pid=5493) INFO 11-03 17:21:05 [gpu_model_runner.py:2840] Starting to load model luccafong/deepseek_mtp_main_random... 2025-11-03T17:21:05.9595762Z (EngineCore_DP0 pid=5493) INFO 11-03 17:21:05 [gpu_model_runner.py:2870] Loading model from scratch... 2025-11-03T17:21:06.0384905Z (EngineCore_DP0 pid=5493) INFO 11-03 17:21:06 [cuda.py:328] Using Triton MLA backend on V1 engine. 2025-11-03T17:21:06.1442746Z (EngineCore_DP0 pid=5493) WARNING 11-03 17:21:06 [fp8.py:150] DeepGEMM backend requested but not available. 2025-11-03T17:21:06.1443665Z (EngineCore_DP0 pid=5493) INFO 11-03 17:21:06 [fp8.py:165] Using Triton backend for FP8 MoE 2025-11-03T17:21:06.2802436Z (EngineCore_DP0 pid=5493) WARNING 11-03 17:21:06 [fp8_utils.py:785] Using default W8A8 Block FP8 kernel config. Performance might be sub-optimal! Config file not found at /opt/conda/envs/py_3.12/lib/python3.12/site-packages/vllm/model_executor/layers/quantization/utils/configs/N=8192,K=512,device_name=NVIDIA_L4,dtype=fp8_w8a8,block_shape=[128,128].json 2025-11-03T17:21:07.3368937Z (EngineCore_DP0 pid=5493) INFO 11-03 17:21:07 [gpu_model_runner.py:2879] Loading drafter model... 2025-11-03T17:21:07.3467315Z (EngineCore_DP0 pid=5493) INFO 11-03 17:21:07 [eagle.py:1005] Assuming the EAGLE head shares the same vocab embedding with the target model. 2025-11-03T17:21:07.3468423Z (EngineCore_DP0 pid=5493) INFO 11-03 17:21:07 [eagle.py:1027] Loading EAGLE LM head weights from the target model. 2025-11-03T17:21:07.8768698Z (EngineCore_DP0 pid=5493) INFO 11-03 17:21:07 [gpu_model_runner.py:2902] Model loading took 2.8305 GiB and 1.387459 seconds 2025-11-03T17:21:07.8772527Z (EngineCore_DP0 pid=5493) INFO 11-03 17:21:07 [kv_cache_utils.py:1199] GPU KV cache size: 4,660,336 tokens 2025-11-03T17:21:07.8773471Z (EngineCore_DP0 pid=5493) INFO 11-03 17:21:07 [kv_cache_utils.py:1204] Maximum concurrency for 163,840 tokens per request: 28.44x 2025-11-03T17:21:09.0049419Z (EngineCore_DP0 pid=5493) INFO 11-03 17:21:09 [gc_utils.py:40] GC Debug Config. enabled:False,top_objects:-1 2025-11-03T17:21:09.0063294Z INFO 11-03 17:21:09 [llm.py:337] Supported tasks: ['generate'] 2025-11-03T17:21:09.4944904Z PASSED 2025-11-03T17:21:09.5088320Z models/test_initialization.py::test_can_initialize_small_subset[XLMRobertaModel] Fork a new process to run a test 5595 2025-11-03T17:21:09.5098000Z Fork a new process to run a test 0 2025-11-03T17:21:09.5375189Z INFO 11-03 17:21:09 [utils.py:239] non-default args: {'load_format': 'dummy', 'gpu_memory_utilization': 0.8, 'disable_log_stats': True, 'hf_overrides': functools.partial(, model_arch='XLMRobertaModel', exist_overrides={}, use_original_num_layers=False), 'model_impl': 'vllm', 'model': 'intfloat/multilingual-e5-small'} 2025-11-03T17:21:09.6660607Z 2025-11-03T17:21:09.6665077Z config.json: 0% 0.00/655 [00:00, 'use_cudagraph': True, 'cudagraph_num_of_warmups': 1, 'cudagraph_capture_sizes': [512, 504, 496, 488, 480, 472, 464, 456, 448, 440, 432, 424, 416, 408, 400, 392, 384, 376, 368, 360, 352, 344, 336, 328, 320, 312, 304, 296, 288, 280, 272, 264, 256, 248, 240, 232, 224, 216, 208, 200, 192, 184, 176, 168, 160, 152, 144, 136, 128, 120, 112, 104, 96, 88, 80, 72, 64, 56, 48, 40, 32, 24, 16, 8, 4, 2, 1], 'cudagraph_copy_inputs': False, 'full_cuda_graph': False, 'use_inductor_graph_partition': False, 'pass_config': {}, 'max_capture_size': 512, 'local_cache_dir': None} 2025-11-03T17:21:19.0493371Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:21:19.0526107Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:21:19.0533392Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:21:19.0542006Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:21:19.0550083Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:21:19.0557745Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:21:19.0560420Z (EngineCore_DP0 pid=5675) INFO 11-03 17:21:19 [parallel_state.py:1231] rank 0 in world size 1 is assigned as DP rank 0, PP rank 0, TP rank 0, EP rank 0 2025-11-03T17:21:19.3971544Z (EngineCore_DP0 pid=5675) INFO 11-03 17:21:19 [topk_topp_sampler.py:59] Using FlashInfer for top-p & top-k sampling. 2025-11-03T17:21:19.4078199Z (EngineCore_DP0 pid=5675) INFO 11-03 17:21:19 [gpu_model_runner.py:2840] Starting to load model intfloat/multilingual-e5-small... 2025-11-03T17:21:19.6591805Z (EngineCore_DP0 pid=5675) INFO 11-03 17:21:19 [gpu_model_runner.py:2870] Loading model from scratch... 2025-11-03T17:21:19.6637323Z (EngineCore_DP0 pid=5675) INFO 11-03 17:21:19 [cuda.py:404] Using Flash Attention backend on V1 engine. 2025-11-03T17:21:20.2338548Z (EngineCore_DP0 pid=5675) INFO 11-03 17:21:20 [gpu_model_runner.py:2902] Model loading took 0.1834 GiB and 0.065302 seconds 2025-11-03T17:21:21.3236746Z (EngineCore_DP0 pid=5675) INFO 11-03 17:21:21 [core.py:142] Disabling chunked prefill for model without KVCache 2025-11-03T17:21:21.3288377Z (EngineCore_DP0 pid=5675) INFO 11-03 17:21:21 [vllm.py:433] Only "last" pooling supports chunked prefill and prefix caching; disabling both. 2025-11-03T17:21:21.5904428Z (EngineCore_DP0 pid=5675) INFO 11-03 17:21:21 [gc_utils.py:40] GC Debug Config. enabled:False,top_objects:-1 2025-11-03T17:21:21.5918975Z INFO 11-03 17:21:21 [llm.py:337] Supported tasks: ['embed', 'encode'] 2025-11-03T17:21:22.0671034Z PASSED 2025-11-03T17:21:22.0814597Z models/test_initialization.py::test_can_initialize_large_subset[CohereForCausalLM] Fork a new process to run a test 5741 2025-11-03T17:21:22.0825117Z Fork a new process to run a test 0 2025-11-03T17:21:22.1109724Z INFO 11-03 17:21:22 [utils.py:239] non-default args: {'trust_remote_code': True, 'load_format': 'dummy', 'gpu_memory_utilization': 0.8, 'disable_log_stats': True, 'hf_overrides': functools.partial(, model_arch='CohereForCausalLM', exist_overrides={}, use_original_num_layers=False), 'model_impl': 'vllm', 'model': 'CohereForAI/c4ai-command-r-v01'} 2025-11-03T17:21:22.3316727Z 2025-11-03T17:21:22.3317870Z config.json: 0% 0.00/765 [00:00, 'use_cudagraph': True, 'cudagraph_num_of_warmups': 1, 'cudagraph_capture_sizes': [512, 504, 496, 488, 480, 472, 464, 456, 448, 440, 432, 424, 416, 408, 400, 392, 384, 376, 368, 360, 352, 344, 336, 328, 320, 312, 304, 296, 288, 280, 272, 264, 256, 248, 240, 232, 224, 216, 208, 200, 192, 184, 176, 168, 160, 152, 144, 136, 128, 120, 112, 104, 96, 88, 80, 72, 64, 56, 48, 40, 32, 24, 16, 8, 4, 2, 1], 'cudagraph_copy_inputs': False, 'full_cuda_graph': True, 'use_inductor_graph_partition': False, 'pass_config': {}, 'max_capture_size': 512, 'local_cache_dir': None} 2025-11-03T17:21:31.6356646Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:21:31.6384411Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:21:31.6391700Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:21:31.6400275Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:21:31.6408571Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:21:31.6416848Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:21:31.6419954Z (EngineCore_DP0 pid=5819) INFO 11-03 17:21:31 [parallel_state.py:1231] rank 0 in world size 1 is assigned as DP rank 0, PP rank 0, TP rank 0, EP rank 0 2025-11-03T17:21:31.9794851Z (EngineCore_DP0 pid=5819) INFO 11-03 17:21:31 [topk_topp_sampler.py:59] Using FlashInfer for top-p & top-k sampling. 2025-11-03T17:21:32.1106786Z (EngineCore_DP0 pid=5819) INFO 11-03 17:21:32 [gpu_model_runner.py:2840] Starting to load model CohereForAI/c4ai-command-r-v01... 2025-11-03T17:21:32.3696271Z (EngineCore_DP0 pid=5819) INFO 11-03 17:21:32 [gpu_model_runner.py:2870] Loading model from scratch... 2025-11-03T17:21:32.4395060Z (EngineCore_DP0 pid=5819) INFO 11-03 17:21:32 [cuda.py:404] Using Flash Attention backend on V1 engine. 2025-11-03T17:21:33.0066618Z (EngineCore_DP0 pid=5819) INFO 11-03 17:21:33 [gpu_model_runner.py:2902] Model loading took 5.4688 GiB and 0.118111 seconds 2025-11-03T17:21:33.0071081Z (EngineCore_DP0 pid=5819) INFO 11-03 17:21:33 [kv_cache_utils.py:1199] GPU KV cache size: 327,680 tokens 2025-11-03T17:21:33.0072274Z (EngineCore_DP0 pid=5819) INFO 11-03 17:21:33 [kv_cache_utils.py:1204] Maximum concurrency for 131,072 tokens per request: 2.50x 2025-11-03T17:21:34.0431162Z (EngineCore_DP0 pid=5819) INFO 11-03 17:21:34 [gc_utils.py:40] GC Debug Config. enabled:False,top_objects:-1 2025-11-03T17:21:34.0444648Z INFO 11-03 17:21:34 [llm.py:337] Supported tasks: ['generate'] 2025-11-03T17:21:34.5460074Z PASSED 2025-11-03T17:21:34.5604474Z models/test_initialization.py::test_can_initialize_large_subset[EagleLlama4ForCausalLM] Fork a new process to run a test 5885 2025-11-03T17:21:34.5614545Z Fork a new process to run a test 0 2025-11-03T17:21:34.5888865Z INFO 11-03 17:21:34 [utils.py:239] non-default args: {'tokenizer': 'meta-llama/Llama-4-Scout-17B-16E-Instruct', 'trust_remote_code': True, 'load_format': 'dummy', 'gpu_memory_utilization': 0.8, 'disable_log_stats': True, 'hf_overrides': functools.partial(, model_arch='EagleLlama4ForCausalLM', exist_overrides={}, use_original_num_layers=False), 'speculative_config': {'model': 'morgendave/EAGLE-Llama-4-Scout-17B-16E-Instruct', 'num_speculative_tokens': 1}, 'model_impl': 'vllm', 'model': 'morgendave/EAGLE-Llama-4-Scout-17B-16E-Instruct'} 2025-11-03T17:21:34.6794510Z 2025-11-03T17:21:34.6796891Z config.json: 0.00B [00:00, ?B/s] 2025-11-03T17:21:34.6797293Z config.json: 1.51kB [00:00, 7.72MB/s] 2025-11-03T17:21:34.7885083Z 2025-11-03T17:21:34.7886210Z preprocessor_config.json: 0% 0.00/636 [00:00, 'use_cudagraph': True, 'cudagraph_num_of_warmups': 1, 'cudagraph_capture_sizes': [512, 504, 496, 488, 480, 472, 464, 456, 448, 440, 432, 424, 416, 408, 400, 392, 384, 376, 368, 360, 352, 344, 336, 328, 320, 312, 304, 296, 288, 280, 272, 264, 256, 248, 240, 232, 224, 216, 208, 200, 192, 184, 176, 168, 160, 152, 144, 136, 128, 120, 112, 104, 96, 88, 80, 72, 64, 56, 48, 40, 32, 24, 16, 8, 4, 2, 1], 'cudagraph_copy_inputs': False, 'full_cuda_graph': True, 'use_inductor_graph_partition': False, 'pass_config': {}, 'max_capture_size': 512, 'local_cache_dir': None} 2025-11-03T17:21:43.0412258Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:21:43.0440828Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:21:43.0448497Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:21:43.0455923Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:21:43.0463861Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:21:43.0471851Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:21:43.0475172Z (EngineCore_DP0 pid=5943) INFO 11-03 17:21:43 [parallel_state.py:1231] rank 0 in world size 1 is assigned as DP rank 0, PP rank 0, TP rank 0, EP rank 0 2025-11-03T17:21:43.3828017Z (EngineCore_DP0 pid=5943) INFO 11-03 17:21:43 [topk_topp_sampler.py:59] Using FlashInfer for top-p & top-k sampling. 2025-11-03T17:21:43.4013844Z (EngineCore_DP0 pid=5943) WARNING 11-03 17:21:43 [__init__.py:194] min_p, logit_bias, and min_tokens parameters won't currently work with speculative decoding enabled. 2025-11-03T17:21:43.5119814Z (EngineCore_DP0 pid=5943) INFO 11-03 17:21:43 [gpu_model_runner.py:2840] Starting to load model morgendave/EAGLE-Llama-4-Scout-17B-16E-Instruct... 2025-11-03T17:21:43.7652866Z (EngineCore_DP0 pid=5943) INFO 11-03 17:21:43 [gpu_model_runner.py:2870] Loading model from scratch... 2025-11-03T17:21:43.9033622Z (EngineCore_DP0 pid=5943) INFO 11-03 17:21:43 [cuda.py:404] Using Flash Attention backend on V1 engine. 2025-11-03T17:21:43.9641956Z (EngineCore_DP0 pid=5943) INFO 11-03 17:21:43 [gpu_model_runner.py:2879] Loading drafter model... 2025-11-03T17:21:43.9895024Z (EngineCore_DP0 pid=5943) INFO 11-03 17:21:43 [eagle.py:1005] Assuming the EAGLE head shares the same vocab embedding with the target model. 2025-11-03T17:21:43.9896027Z (EngineCore_DP0 pid=5943) INFO 11-03 17:21:43 [eagle.py:1027] Loading EAGLE LM head weights from the target model. 2025-11-03T17:21:44.4961853Z (EngineCore_DP0 pid=5943) INFO 11-03 17:21:44 [gpu_model_runner.py:2902] Model loading took 6.3595 GiB and 0.224551 seconds 2025-11-03T17:21:44.4967266Z (EngineCore_DP0 pid=5943) INFO 11-03 17:21:44 [kv_cache_utils.py:1199] GPU KV cache size: 655,360 tokens 2025-11-03T17:21:44.4968357Z (EngineCore_DP0 pid=5943) INFO 11-03 17:21:44 [kv_cache_utils.py:1204] Maximum concurrency for 262,144 tokens per request: 2.50x 2025-11-03T17:21:45.7062464Z (EngineCore_DP0 pid=5943) INFO 11-03 17:21:45 [gc_utils.py:40] GC Debug Config. enabled:False,top_objects:-1 2025-11-03T17:21:45.7073255Z INFO 11-03 17:21:45 [llm.py:337] Supported tasks: ['generate'] 2025-11-03T17:21:46.2429963Z PASSED 2025-11-03T17:21:46.2572605Z models/test_initialization.py::test_can_initialize_large_subset[InternLM2VEForCausalLM] Fork a new process to run a test 6009 2025-11-03T17:21:46.2582607Z Fork a new process to run a test 0 2025-11-03T17:21:46.2858247Z INFO 11-03 17:21:46 [utils.py:239] non-default args: {'trust_remote_code': True, 'load_format': 'dummy', 'gpu_memory_utilization': 0.8, 'disable_log_stats': True, 'hf_overrides': functools.partial(, model_arch='InternLM2VEForCausalLM', exist_overrides={}, use_original_num_layers=False), 'model_impl': 'vllm', 'model': 'OpenGVLab/Mono-InternVL-2B'} 2025-11-03T17:21:46.3925581Z 2025-11-03T17:21:46.3926917Z config.json: 0.00B [00:00, ?B/s] 2025-11-03T17:21:46.3927263Z config.json: 3.75kB [00:00, 23.5MB/s] 2025-11-03T17:21:46.4730423Z 2025-11-03T17:21:46.4732059Z configuration_internvl_chat.py: 0.00B [00:00, ?B/s] 2025-11-03T17:21:46.4732640Z configuration_internvl_chat.py: 4.44kB [00:00, 32.3MB/s] 2025-11-03T17:21:46.5287098Z 2025-11-03T17:21:46.5289905Z configuration_internlm2.py: 0.00B [00:00, ?B/s] 2025-11-03T17:21:46.5290460Z configuration_internlm2.py: 7.00kB [00:00, 15.2MB/s] 2025-11-03T17:21:46.5312784Z A new version of the following files was downloaded from https://huggingface.co/OpenGVLab/Mono-InternVL-2B: 2025-11-03T17:21:46.5313438Z - configuration_internlm2.py 2025-11-03T17:21:46.5314107Z . Make sure to double-check they do not contain any added malicious code. To avoid downloading new versions of the code file, you can pin a revision. 2025-11-03T17:21:46.5642210Z 2025-11-03T17:21:46.5644051Z configuration_intern_patch.py: 0.00B [00:00, ?B/s] 2025-11-03T17:21:46.5644589Z configuration_intern_patch.py: 4.40kB [00:00, 31.8MB/s] 2025-11-03T17:21:46.5662873Z A new version of the following files was downloaded from https://huggingface.co/OpenGVLab/Mono-InternVL-2B: 2025-11-03T17:21:46.5663832Z - configuration_intern_patch.py 2025-11-03T17:21:46.5664492Z . Make sure to double-check they do not contain any added malicious code. To avoid downloading new versions of the code file, you can pin a revision. 2025-11-03T17:21:46.5665380Z A new version of the following files was downloaded from https://huggingface.co/OpenGVLab/Mono-InternVL-2B: 2025-11-03T17:21:46.5665904Z - configuration_internvl_chat.py 2025-11-03T17:21:46.5666183Z - configuration_internlm2.py 2025-11-03T17:21:46.5666432Z - configuration_intern_patch.py 2025-11-03T17:21:46.5667052Z . Make sure to double-check they do not contain any added malicious code. To avoid downloading new versions of the code file, you can pin a revision. 2025-11-03T17:21:46.7813842Z INFO 11-03 17:21:46 [model.py:653] Resolved architecture: InternVLChatModel 2025-11-03T17:21:46.7814334Z `torch_dtype` is deprecated! Use `dtype` instead! 2025-11-03T17:21:46.8054710Z INFO 11-03 17:21:46 [model.py:1714] Using max model len 32768 2025-11-03T17:21:46.8056677Z The argument `trust_remote_code` is to be used with Auto classes. It has no effect here and is ignored. 2025-11-03T17:21:46.8633261Z INFO 11-03 17:21:46 [scheduler.py:225] Chunked prefill is enabled with max_num_batched_tokens=8192. 2025-11-03T17:21:46.9060981Z 2025-11-03T17:21:46.9063066Z tokenizer_config.json: 0.00B [00:00, ?B/s] 2025-11-03T17:21:46.9063925Z tokenizer_config.json: 4.00kB [00:00, 23.3MB/s] 2025-11-03T17:21:46.9418341Z 2025-11-03T17:21:46.9420646Z tokenization_internlm2.py: 0.00B [00:00, ?B/s] 2025-11-03T17:21:46.9421340Z tokenization_internlm2.py: 8.79kB [00:00, 21.5MB/s] 2025-11-03T17:21:46.9470796Z A new version of the following files was downloaded from https://huggingface.co/OpenGVLab/Mono-InternVL-2B: 2025-11-03T17:21:46.9471455Z - tokenization_internlm2.py 2025-11-03T17:21:46.9472123Z . Make sure to double-check they do not contain any added malicious code. To avoid downloading new versions of the code file, you can pin a revision. 2025-11-03T17:21:47.0650688Z 2025-11-03T17:21:47.3242737Z ./tokenizer.model: 0% 0.00/1.48M [00:00, 'use_cudagraph': True, 'cudagraph_num_of_warmups': 1, 'cudagraph_capture_sizes': [512, 504, 496, 488, 480, 472, 464, 456, 448, 440, 432, 424, 416, 408, 400, 392, 384, 376, 368, 360, 352, 344, 336, 328, 320, 312, 304, 296, 288, 280, 272, 264, 256, 248, 240, 232, 224, 216, 208, 200, 192, 184, 176, 168, 160, 152, 144, 136, 128, 120, 112, 104, 96, 88, 80, 72, 64, 56, 48, 40, 32, 24, 16, 8, 4, 2, 1], 'cudagraph_copy_inputs': False, 'full_cuda_graph': True, 'use_inductor_graph_partition': False, 'pass_config': {}, 'max_capture_size': 512, 'local_cache_dir': None} 2025-11-03T17:21:48.8405193Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:21:48.8435751Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:21:48.8443211Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:21:48.8450943Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:21:48.8457919Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:21:48.8465932Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:21:48.8468602Z (EngineCore_DP0 pid=6037) INFO 11-03 17:21:48 [parallel_state.py:1231] rank 0 in world size 1 is assigned as DP rank 0, PP rank 0, TP rank 0, EP rank 0 2025-11-03T17:21:49.1877619Z (EngineCore_DP0 pid=6037) INFO 11-03 17:21:49 [topk_topp_sampler.py:59] Using FlashInfer for top-p & top-k sampling. 2025-11-03T17:21:49.2891023Z (EngineCore_DP0 pid=6037) WARNING 11-03 17:21:49 [__init__.py:2040] The following intended overrides are not keyword args and will be dropped: {'truncation'} 2025-11-03T17:21:49.2899039Z (EngineCore_DP0 pid=6037) WARNING 11-03 17:21:49 [__init__.py:2040] The following intended overrides are not keyword args and will be dropped: {'truncation'} 2025-11-03T17:21:49.4678309Z (EngineCore_DP0 pid=6037) INFO 11-03 17:21:49 [gpu_model_runner.py:2840] Starting to load model OpenGVLab/Mono-InternVL-2B... 2025-11-03T17:21:49.7268663Z (EngineCore_DP0 pid=6037) INFO 11-03 17:21:49 [gpu_model_runner.py:2870] Loading model from scratch... 2025-11-03T17:21:49.7926983Z (EngineCore_DP0 pid=6037) INFO 11-03 17:21:49 [cuda.py:404] Using Flash Attention backend on V1 engine. 2025-11-03T17:21:50.3724498Z (EngineCore_DP0 pid=6037) INFO 11-03 17:21:50 [gpu_model_runner.py:2902] Model loading took 0.9523 GiB and 0.125821 seconds 2025-11-03T17:21:50.3729216Z (EngineCore_DP0 pid=6037) INFO 11-03 17:21:50 [kv_cache_utils.py:1199] GPU KV cache size: 2,621,440 tokens 2025-11-03T17:21:50.3729976Z (EngineCore_DP0 pid=6037) INFO 11-03 17:21:50 [kv_cache_utils.py:1204] Maximum concurrency for 32,768 tokens per request: 80.00x 2025-11-03T17:21:50.6556562Z (EngineCore_DP0 pid=6037) WARNING 11-03 17:21:50 [tokenizer.py:255] Using a slow tokenizer. This might cause a significant slowdown. Consider using a fast tokenizer instead. 2025-11-03T17:21:50.7518595Z (EngineCore_DP0 pid=6037) WARNING 11-03 17:21:50 [__init__.py:2040] The following intended overrides are not keyword args and will be dropped: {'truncation'} 2025-11-03T17:21:51.3012888Z (EngineCore_DP0 pid=6037) INFO 11-03 17:21:51 [gc_utils.py:40] GC Debug Config. enabled:False,top_objects:-1 2025-11-03T17:21:51.3027449Z INFO 11-03 17:21:51 [llm.py:337] Supported tasks: ['generate'] 2025-11-03T17:21:51.7602244Z PASSED 2025-11-03T17:21:51.7743612Z models/test_initialization.py::test_can_initialize_large_subset[JAISLMHeadModel] Fork a new process to run a test 6103 2025-11-03T17:21:51.7754004Z Fork a new process to run a test 0 2025-11-03T17:21:51.8028346Z INFO 11-03 17:21:51 [utils.py:239] non-default args: {'load_format': 'dummy', 'gpu_memory_utilization': 0.8, 'disable_log_stats': True, 'hf_overrides': functools.partial(, model_arch='JAISLMHeadModel', exist_overrides={}, use_original_num_layers=False), 'model_impl': 'vllm', 'model': 'inceptionai/jais-13b-chat'} 2025-11-03T17:21:51.9754693Z 2025-11-03T17:21:51.9758681Z config.json: 0% 0.00/1.26k [00:00, 'use_cudagraph': True, 'cudagraph_num_of_warmups': 1, 'cudagraph_capture_sizes': [512, 504, 496, 488, 480, 472, 464, 456, 448, 440, 432, 424, 416, 408, 400, 392, 384, 376, 368, 360, 352, 344, 336, 328, 320, 312, 304, 296, 288, 280, 272, 264, 256, 248, 240, 232, 224, 216, 208, 200, 192, 184, 176, 168, 160, 152, 144, 136, 128, 120, 112, 104, 96, 88, 80, 72, 64, 56, 48, 40, 32, 24, 16, 8, 4, 2, 1], 'cudagraph_copy_inputs': False, 'full_cuda_graph': True, 'use_inductor_graph_partition': False, 'pass_config': {}, 'max_capture_size': 512, 'local_cache_dir': None} 2025-11-03T17:22:00.3845422Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:22:00.3873754Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:22:00.3881478Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:22:00.3889220Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:22:00.3897025Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:22:00.3905145Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:22:00.3908298Z (EngineCore_DP0 pid=6161) INFO 11-03 17:22:00 [parallel_state.py:1231] rank 0 in world size 1 is assigned as DP rank 0, PP rank 0, TP rank 0, EP rank 0 2025-11-03T17:22:00.7257539Z (EngineCore_DP0 pid=6161) INFO 11-03 17:22:00 [topk_topp_sampler.py:59] Using FlashInfer for top-p & top-k sampling. 2025-11-03T17:22:00.8234572Z (EngineCore_DP0 pid=6161) INFO 11-03 17:22:00 [gpu_model_runner.py:2840] Starting to load model inceptionai/jais-13b-chat... 2025-11-03T17:22:01.0745936Z (EngineCore_DP0 pid=6161) INFO 11-03 17:22:01 [gpu_model_runner.py:2870] Loading model from scratch... 2025-11-03T17:22:01.0770833Z (EngineCore_DP0 pid=6161) INFO 11-03 17:22:01 [cuda.py:404] Using Flash Attention backend on V1 engine. 2025-11-03T17:22:01.6345852Z (EngineCore_DP0 pid=6161) INFO 11-03 17:22:01 [gpu_model_runner.py:2902] Model loading took 1.3986 GiB and 0.061890 seconds 2025-11-03T17:22:01.6350269Z (EngineCore_DP0 pid=6161) INFO 11-03 17:22:01 [kv_cache_utils.py:1199] GPU KV cache size: 524,288 tokens 2025-11-03T17:22:01.6351384Z (EngineCore_DP0 pid=6161) INFO 11-03 17:22:01 [kv_cache_utils.py:1204] Maximum concurrency for 2,048 tokens per request: 256.00x 2025-11-03T17:22:02.1838126Z (EngineCore_DP0 pid=6161) INFO 11-03 17:22:02 [gc_utils.py:40] GC Debug Config. enabled:False,top_objects:-1 2025-11-03T17:22:02.1853730Z INFO 11-03 17:22:02 [llm.py:337] Supported tasks: ['generate'] 2025-11-03T17:22:02.6384680Z PASSED 2025-11-03T17:22:02.6526921Z models/test_initialization.py::test_can_initialize_large_subset[DeepseekV3ForCausalLM] Fork a new process to run a test 6227 2025-11-03T17:22:02.6536656Z Fork a new process to run a test 0 2025-11-03T17:22:02.6808322Z INFO 11-03 17:22:02 [utils.py:239] non-default args: {'trust_remote_code': True, 'load_format': 'dummy', 'gpu_memory_utilization': 0.8, 'disable_log_stats': True, 'hf_overrides': functools.partial(, model_arch='DeepseekV3ForCausalLM', exist_overrides={}, use_original_num_layers=False), 'model_impl': 'vllm', 'model': 'deepseek-ai/DeepSeek-V3'} 2025-11-03T17:22:02.7784277Z 2025-11-03T17:22:02.7786484Z config.json: 0.00B [00:00, ?B/s] 2025-11-03T17:22:02.7786944Z config.json: 1.66kB [00:00, 8.63MB/s] 2025-11-03T17:22:02.8236704Z INFO 11-03 17:22:02 [config.py:408] Replacing legacy 'type' key with 'rope_type' 2025-11-03T17:22:02.9427736Z INFO 11-03 17:22:02 [model.py:653] Resolved architecture: DeepseekV3ForCausalLM 2025-11-03T17:22:02.9428243Z `torch_dtype` is deprecated! Use `dtype` instead! 2025-11-03T17:22:02.9678009Z INFO 11-03 17:22:02 [model.py:1714] Using max model len 163840 2025-11-03T17:22:03.0631029Z The argument `trust_remote_code` is to be used with Auto classes. It has no effect here and is ignored. 2025-11-03T17:22:03.2313317Z INFO 11-03 17:22:03 [scheduler.py:225] Chunked prefill is enabled with max_num_batched_tokens=8192. 2025-11-03T17:22:03.2801266Z 2025-11-03T17:22:03.2802668Z tokenizer_config.json: 0.00B [00:00, ?B/s] 2025-11-03T17:22:03.2803114Z tokenizer_config.json: 3.13kB [00:00, 25.2MB/s] 2025-11-03T17:22:03.3981452Z 2025-11-03T17:22:03.4252589Z tokenizer.json: 0.00B [00:00, ?B/s] 2025-11-03T17:22:03.4253029Z tokenizer.json: 7.85MB [00:00, 288MB/s] 2025-11-03T17:22:04.1048562Z INFO 11-03 17:22:04 [config.py:408] Replacing legacy 'type' key with 'rope_type' 2025-11-03T17:22:04.1271643Z (EngineCore_DP0 pid=6235) INFO 11-03 17:22:04 [core.py:727] Waiting for init message from front-end. 2025-11-03T17:22:04.1308813Z (EngineCore_DP0 pid=6235) INFO 11-03 17:22:04 [core.py:94] Initializing a V1 LLM engine (v0.11.0rc2.dev389+ge51928192.d20251103) with config: model='deepseek-ai/DeepSeek-V3', speculative_config=None, tokenizer='deepseek-ai/DeepSeek-V3', skip_tokenizer_init=False, tokenizer_mode=auto, revision=None, tokenizer_revision=None, trust_remote_code=True, dtype=torch.bfloat16, max_seq_len=163840, download_dir=None, load_format=dummy, tensor_parallel_size=1, pipeline_parallel_size=1, data_parallel_size=1, disable_custom_all_reduce=False, quantization=fp8, enforce_eager=False, kv_cache_dtype=auto, device_config=cuda, structured_outputs_config=StructuredOutputsConfig(backend='auto', disable_fallback=False, disable_any_whitespace=False, disable_additional_properties=False, reasoning_parser=''), observability_config=ObservabilityConfig(show_hidden_metrics_for_version=None, otlp_traces_endpoint=None, collect_detailed_traces=None), seed=0, served_model_name=deepseek-ai/DeepSeek-V3, enable_prefix_caching=True, chunked_prefill_enabled=True, pooler_config=None, compilation_config={'level': 3, 'debug_dump_path': None, 'cache_dir': '', 'backend': '', 'custom_ops': ['+quant_fp8'], 'splitting_ops': ['vllm.unified_attention', 'vllm.unified_attention_with_output', 'vllm.unified_mla_attention', 'vllm.unified_mla_attention_with_output', 'vllm.mamba_mixer2', 'vllm.mamba_mixer', 'vllm.short_conv', 'vllm.linear_attention', 'vllm.plamo2_mamba_mixer', 'vllm.gdn_attention', 'vllm.sparse_attn_indexer'], 'use_inductor': True, 'compile_sizes': [], 'inductor_compile_config': {'enable_auto_functionalized_v2': False}, 'inductor_passes': {}, 'cudagraph_mode': , 'use_cudagraph': True, 'cudagraph_num_of_warmups': 1, 'cudagraph_capture_sizes': [512, 504, 496, 488, 480, 472, 464, 456, 448, 440, 432, 424, 416, 408, 400, 392, 384, 376, 368, 360, 352, 344, 336, 328, 320, 312, 304, 296, 288, 280, 272, 264, 256, 248, 240, 232, 224, 216, 208, 200, 192, 184, 176, 168, 160, 152, 144, 136, 128, 120, 112, 104, 96, 88, 80, 72, 64, 56, 48, 40, 32, 24, 16, 8, 4, 2, 1], 'cudagraph_copy_inputs': False, 'full_cuda_graph': True, 'use_inductor_graph_partition': False, 'pass_config': {}, 'max_capture_size': 512, 'local_cache_dir': None} 2025-11-03T17:22:04.8185806Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:22:04.8214180Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:22:04.8222229Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:22:04.8231256Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:22:04.8239175Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:22:04.8247742Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:22:04.8250920Z (EngineCore_DP0 pid=6235) INFO 11-03 17:22:04 [parallel_state.py:1231] rank 0 in world size 1 is assigned as DP rank 0, PP rank 0, TP rank 0, EP rank 0 2025-11-03T17:22:05.1708630Z (EngineCore_DP0 pid=6235) INFO 11-03 17:22:05 [topk_topp_sampler.py:59] Using FlashInfer for top-p & top-k sampling. 2025-11-03T17:22:05.2917109Z (EngineCore_DP0 pid=6235) INFO 11-03 17:22:05 [gpu_model_runner.py:2840] Starting to load model deepseek-ai/DeepSeek-V3... 2025-11-03T17:22:05.5551505Z (EngineCore_DP0 pid=6235) INFO 11-03 17:22:05 [gpu_model_runner.py:2870] Loading model from scratch... 2025-11-03T17:22:05.6446889Z (EngineCore_DP0 pid=6235) INFO 11-03 17:22:05 [cuda.py:328] Using Triton MLA backend on V1 engine. 2025-11-03T17:22:05.6601970Z (EngineCore_DP0 pid=6235) WARNING 11-03 17:22:05 [fp8.py:150] DeepGEMM backend requested but not available. 2025-11-03T17:22:05.6603071Z (EngineCore_DP0 pid=6235) INFO 11-03 17:22:05 [fp8.py:165] Using Triton backend for FP8 MoE 2025-11-03T17:22:05.7824174Z (EngineCore_DP0 pid=6235) WARNING 11-03 17:22:05 [fp8_utils.py:785] Using default W8A8 Block FP8 kernel config. Performance might be sub-optimal! Config file not found at /opt/conda/envs/py_3.12/lib/python3.12/site-packages/vllm/model_executor/layers/quantization/utils/configs/N=32768,K=512,device_name=NVIDIA_L4,dtype=fp8_w8a8,block_shape=[128,128].json 2025-11-03T17:22:06.8229039Z (EngineCore_DP0 pid=6235) INFO 11-03 17:22:06 [gpu_model_runner.py:2902] Model loading took 4.3759 GiB and 0.727321 seconds 2025-11-03T17:22:06.8230039Z (EngineCore_DP0 pid=6235) INFO 11-03 17:22:06 [kv_cache_utils.py:1199] GPU KV cache size: 9,320,672 tokens 2025-11-03T17:22:06.8230764Z (EngineCore_DP0 pid=6235) INFO 11-03 17:22:06 [kv_cache_utils.py:1204] Maximum concurrency for 163,840 tokens per request: 56.89x 2025-11-03T17:22:08.3944523Z (EngineCore_DP0 pid=6235) INFO 11-03 17:22:08 [gc_utils.py:40] GC Debug Config. enabled:False,top_objects:-1 2025-11-03T17:22:08.3957797Z INFO 11-03 17:22:08 [llm.py:337] Supported tasks: ['generate'] 2025-11-03T17:22:08.8961057Z PASSED 2025-11-03T17:22:08.9103794Z models/test_initialization.py::test_can_initialize_large_subset[MptForCausalLM] Fork a new process to run a test 6305 2025-11-03T17:22:08.9114668Z Fork a new process to run a test 0 2025-11-03T17:22:08.9116479Z Model is not available online 2025-11-03T17:22:09.2071770Z PASSED 2025-11-03T17:22:09.2219451Z models/test_initialization.py::test_can_initialize_large_subset[LlavaNextVideoForConditionalGeneration] Fork a new process to run a test 6306 2025-11-03T17:22:09.2230209Z Fork a new process to run a test 0 2025-11-03T17:22:09.2503980Z INFO 11-03 17:22:09 [utils.py:239] non-default args: {'load_format': 'dummy', 'gpu_memory_utilization': 0.8, 'disable_log_stats': True, 'hf_overrides': functools.partial(, model_arch='LlavaNextVideoForConditionalGeneration', exist_overrides={}, use_original_num_layers=False), 'model_impl': 'vllm', 'model': 'llava-hf/LLaVA-NeXT-Video-7B-hf'} 2025-11-03T17:22:09.3345176Z 2025-11-03T17:22:09.3346990Z config.json: 0.00B [00:00, ?B/s] 2025-11-03T17:22:09.3347291Z config.json: 1.41kB [00:00, 8.38MB/s] 2025-11-03T17:22:09.4423706Z 2025-11-03T17:22:09.4426247Z preprocessor_config.json: 0% 0.00/741 [00:00, 'use_cudagraph': True, 'cudagraph_num_of_warmups': 1, 'cudagraph_capture_sizes': [512, 504, 496, 488, 480, 472, 464, 456, 448, 440, 432, 424, 416, 408, 400, 392, 384, 376, 368, 360, 352, 344, 336, 328, 320, 312, 304, 296, 288, 280, 272, 264, 256, 248, 240, 232, 224, 216, 208, 200, 192, 184, 176, 168, 160, 152, 144, 136, 128, 120, 112, 104, 96, 88, 80, 72, 64, 56, 48, 40, 32, 24, 16, 8, 4, 2, 1], 'cudagraph_copy_inputs': False, 'full_cuda_graph': True, 'use_inductor_graph_partition': False, 'pass_config': {}, 'max_capture_size': 512, 'local_cache_dir': None} 2025-11-03T17:22:24.6839399Z (EngineCore_DP0 pid=6386) /opt/conda/envs/py_3.12/lib/python3.12/site-packages/torch/library.py:356: UserWarning: Warning only once for all operators, other operators may also be overridden. 2025-11-03T17:22:24.6840679Z (EngineCore_DP0 pid=6386) Overriding a previously registered kernel for the same operator and the same dispatch key 2025-11-03T17:22:24.6842452Z (EngineCore_DP0 pid=6386) operator: xformers_python::_fused_allgather_and_linear_impl(Tensor scattered_input, Tensor[] weights, str process_group_name, Tensor(a3!)[] gathered_outputs, SymInt timeout_s, bool _wait, bool _memcpy, Tensor scale_scattered_input, Tensor?[] scales_weights) -> () 2025-11-03T17:22:24.6843901Z (EngineCore_DP0 pid=6386) registered at /opt/conda/envs/py_3.12/lib/python3.12/site-packages/torch/_library/custom_ops.py:922 2025-11-03T17:22:24.6844511Z (EngineCore_DP0 pid=6386) dispatch key: ADInplaceOrView 2025-11-03T17:22:24.6844935Z (EngineCore_DP0 pid=6386) previous kernel: no debug info 2025-11-03T17:22:24.6845963Z (EngineCore_DP0 pid=6386) new kernel: registered at /opt/conda/envs/py_3.12/lib/python3.12/site-packages/torch/_library/custom_ops.py:922 (Triggered internally at /var/lib/jenkins/workspace/aten/src/ATen/core/dispatch/OperatorEntry.cpp:208.) 2025-11-03T17:22:24.6846865Z (EngineCore_DP0 pid=6386) self.m.impl( 2025-11-03T17:22:25.2262028Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:22:25.2289572Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:22:25.2296805Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:22:25.2304087Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:22:25.2311155Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:22:25.2317989Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:22:25.2320736Z (EngineCore_DP0 pid=6386) INFO 11-03 17:22:25 [parallel_state.py:1231] rank 0 in world size 1 is assigned as DP rank 0, PP rank 0, TP rank 0, EP rank 0 2025-11-03T17:22:25.4635135Z (EngineCore_DP0 pid=6386) INFO 11-03 17:22:25 [topk_topp_sampler.py:59] Using FlashInfer for top-p & top-k sampling. 2025-11-03T17:22:25.6212514Z (EngineCore_DP0 pid=6386) 2025-11-03T17:22:25.6212968Z processor_config.json: 0% 0.00/209 [00:00, model_arch='KeyeVL1_5ForConditionalGeneration', exist_overrides={}, use_original_num_layers=False), 'model_impl': 'vllm', 'model': 'Kwai-Keye/Keye-VL-1_5-8B'} 2025-11-03T17:22:37.3336167Z 2025-11-03T17:22:37.3338183Z config.json: 0.00B [00:00, ?B/s] 2025-11-03T17:22:37.3338559Z config.json: 2.03kB [00:00, 10.7MB/s] 2025-11-03T17:22:37.4017738Z 2025-11-03T17:22:37.4020794Z configuration_keye_vl_1_5.py: 0.00B [00:00, ?B/s] 2025-11-03T17:22:37.4021523Z configuration_keye_vl_1_5.py: 12.9kB [00:00, 27.8MB/s] 2025-11-03T17:22:37.4122370Z A new version of the following files was downloaded from https://huggingface.co/Kwai-Keye/Keye-VL-1_5-8B: 2025-11-03T17:22:37.4122933Z - configuration_keye_vl_1_5.py 2025-11-03T17:22:37.4123563Z . Make sure to double-check they do not contain any added malicious code. To avoid downloading new versions of the code file, you can pin a revision. 2025-11-03T17:22:37.6359348Z 2025-11-03T17:22:37.6360475Z preprocessor_config.json: 0% 0.00/458 [00:00, 'use_cudagraph': True, 'cudagraph_num_of_warmups': 1, 'cudagraph_capture_sizes': [512, 504, 496, 488, 480, 472, 464, 456, 448, 440, 432, 424, 416, 408, 400, 392, 384, 376, 368, 360, 352, 344, 336, 328, 320, 312, 304, 296, 288, 280, 272, 264, 256, 248, 240, 232, 224, 216, 208, 200, 192, 184, 176, 168, 160, 152, 144, 136, 128, 120, 112, 104, 96, 88, 80, 72, 64, 56, 48, 40, 32, 24, 16, 8, 4, 2, 1], 'cudagraph_copy_inputs': False, 'full_cuda_graph': True, 'use_inductor_graph_partition': False, 'pass_config': {}, 'max_capture_size': 512, 'local_cache_dir': None} 2025-11-03T17:22:46.3623256Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:22:46.3653367Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:22:46.3660804Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:22:46.3668711Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:22:46.3676691Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:22:46.3684803Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:22:46.3687409Z (EngineCore_DP0 pid=6727) INFO 11-03 17:22:46 [parallel_state.py:1231] rank 0 in world size 1 is assigned as DP rank 0, PP rank 0, TP rank 0, EP rank 0 2025-11-03T17:22:46.7132978Z (EngineCore_DP0 pid=6727) INFO 11-03 17:22:46 [topk_topp_sampler.py:59] Using FlashInfer for top-p & top-k sampling. 2025-11-03T17:22:46.8411671Z (EngineCore_DP0 pid=6727) 2025-11-03T17:22:46.8413710Z processor_config.json: 0% 0.00/132 [00:00, model_arch='Gemma3nForConditionalGeneration', exist_overrides={}, use_original_num_layers=False), 'model_impl': 'vllm', 'model': 'google/gemma-3n-E2B-it'} 2025-11-03T17:22:52.2806830Z INFO 11-03 17:22:52 [model.py:653] Resolved architecture: Gemma3nForConditionalGeneration 2025-11-03T17:22:52.2807729Z `torch_dtype` is deprecated! Use `dtype` instead! 2025-11-03T17:22:52.3045199Z INFO 11-03 17:22:52 [model.py:1714] Using max model len 32768 2025-11-03T17:22:52.5162447Z INFO 11-03 17:22:52 [scheduler.py:225] Chunked prefill is enabled with max_num_batched_tokens=8192. 2025-11-03T17:22:55.6609270Z (EngineCore_DP0 pid=6800) INFO 11-03 17:22:55 [core.py:727] Waiting for init message from front-end. 2025-11-03T17:22:55.6643676Z (EngineCore_DP0 pid=6800) INFO 11-03 17:22:55 [core.py:94] Initializing a V1 LLM engine (v0.11.0rc2.dev389+ge51928192.d20251103) with config: model='google/gemma-3n-E2B-it', speculative_config=None, tokenizer='google/gemma-3n-E2B-it', skip_tokenizer_init=False, tokenizer_mode=auto, revision=None, tokenizer_revision=None, trust_remote_code=False, dtype=torch.bfloat16, max_seq_len=32768, download_dir=None, load_format=dummy, tensor_parallel_size=1, pipeline_parallel_size=1, data_parallel_size=1, disable_custom_all_reduce=False, quantization=None, enforce_eager=False, kv_cache_dtype=auto, device_config=cuda, structured_outputs_config=StructuredOutputsConfig(backend='auto', disable_fallback=False, disable_any_whitespace=False, disable_additional_properties=False, reasoning_parser=''), observability_config=ObservabilityConfig(show_hidden_metrics_for_version=None, otlp_traces_endpoint=None, collect_detailed_traces=None), seed=0, served_model_name=google/gemma-3n-E2B-it, enable_prefix_caching=True, chunked_prefill_enabled=True, pooler_config=None, compilation_config={'level': 3, 'debug_dump_path': None, 'cache_dir': '', 'backend': '', 'custom_ops': [], 'splitting_ops': ['vllm.unified_attention', 'vllm.unified_attention_with_output', 'vllm.unified_mla_attention', 'vllm.unified_mla_attention_with_output', 'vllm.mamba_mixer2', 'vllm.mamba_mixer', 'vllm.short_conv', 'vllm.linear_attention', 'vllm.plamo2_mamba_mixer', 'vllm.gdn_attention', 'vllm.sparse_attn_indexer'], 'use_inductor': True, 'compile_sizes': [], 'inductor_compile_config': {'enable_auto_functionalized_v2': False}, 'inductor_passes': {}, 'cudagraph_mode': , 'use_cudagraph': True, 'cudagraph_num_of_warmups': 1, 'cudagraph_capture_sizes': [512, 504, 496, 488, 480, 472, 464, 456, 448, 440, 432, 424, 416, 408, 400, 392, 384, 376, 368, 360, 352, 344, 336, 328, 320, 312, 304, 296, 288, 280, 272, 264, 256, 248, 240, 232, 224, 216, 208, 200, 192, 184, 176, 168, 160, 152, 144, 136, 128, 120, 112, 104, 96, 88, 80, 72, 64, 56, 48, 40, 32, 24, 16, 8, 4, 2, 1], 'cudagraph_copy_inputs': False, 'full_cuda_graph': True, 'use_inductor_graph_partition': False, 'pass_config': {}, 'max_capture_size': 512, 'local_cache_dir': None} 2025-11-03T17:22:56.3880171Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:22:56.3909306Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:22:56.3917285Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:22:56.3925705Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:22:56.3933486Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:22:56.3941649Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:22:56.3944552Z (EngineCore_DP0 pid=6800) INFO 11-03 17:22:56 [parallel_state.py:1231] rank 0 in world size 1 is assigned as DP rank 0, PP rank 0, TP rank 0, EP rank 0 2025-11-03T17:22:56.7513364Z (EngineCore_DP0 pid=6800) INFO 11-03 17:22:56 [topk_topp_sampler.py:59] Using FlashInfer for top-p & top-k sampling. 2025-11-03T17:23:02.9835432Z (EngineCore_DP0 pid=6800) INFO 11-03 17:23:02 [gpu_model_runner.py:2840] Starting to load model google/gemma-3n-E2B-it... 2025-11-03T17:23:03.4157889Z (EngineCore_DP0 pid=6800) INFO 11-03 17:23:03 [gpu_model_runner.py:2870] Loading model from scratch... 2025-11-03T17:23:03.6365540Z (EngineCore_DP0 pid=6800) INFO 11-03 17:23:03 [cuda.py:404] Using Flash Attention backend on V1 engine. 2025-11-03T17:23:04.3041898Z (EngineCore_DP0 pid=6800) INFO 11-03 17:23:04 [gpu_model_runner.py:2902] Model loading took 3.6358 GiB and 0.322470 seconds 2025-11-03T17:23:04.3047241Z (EngineCore_DP0 pid=6800) INFO 11-03 17:23:04 [kv_cache_utils.py:1199] GPU KV cache size: 2,621,440 tokens 2025-11-03T17:23:04.3048317Z (EngineCore_DP0 pid=6800) INFO 11-03 17:23:04 [kv_cache_utils.py:1204] Maximum concurrency for 32,768 tokens per request: 300.62x 2025-11-03T17:23:06.1338002Z (EngineCore_DP0 pid=6800) INFO 11-03 17:23:06 [gc_utils.py:40] GC Debug Config. enabled:False,top_objects:-1 2025-11-03T17:23:06.1353451Z INFO 11-03 17:23:06 [llm.py:337] Supported tasks: ['generate', 'transcription'] 2025-11-03T17:23:06.6659498Z PASSED 2025-11-03T17:23:06.6803055Z models/test_initialization.py::test_can_initialize_large_subset[KeyeForConditionalGeneration] Fork a new process to run a test 6897 2025-11-03T17:23:06.6814033Z Fork a new process to run a test 0 2025-11-03T17:23:06.7091916Z INFO 11-03 17:23:06 [utils.py:239] non-default args: {'trust_remote_code': True, 'load_format': 'dummy', 'gpu_memory_utilization': 0.8, 'disable_log_stats': True, 'hf_overrides': functools.partial(, model_arch='KeyeForConditionalGeneration', exist_overrides={}, use_original_num_layers=False), 'model_impl': 'vllm', 'model': 'Kwai-Keye/Keye-VL-8B-Preview'} 2025-11-03T17:23:06.7801743Z 2025-11-03T17:23:06.7804118Z config.json: 0.00B [00:00, ?B/s] 2025-11-03T17:23:06.7804449Z config.json: 1.87kB [00:00, 9.44MB/s] 2025-11-03T17:23:06.8600812Z 2025-11-03T17:23:06.8603245Z configuration_keye.py: 0.00B [00:00, ?B/s] 2025-11-03T17:23:06.8603621Z configuration_keye.py: 12.3kB [00:00, 69.5MB/s] 2025-11-03T17:23:06.8696971Z A new version of the following files was downloaded from https://huggingface.co/Kwai-Keye/Keye-VL-8B-Preview: 2025-11-03T17:23:06.8697523Z - configuration_keye.py 2025-11-03T17:23:06.8698142Z . Make sure to double-check they do not contain any added malicious code. To avoid downloading new versions of the code file, you can pin a revision. 2025-11-03T17:23:07.1214802Z 2025-11-03T17:23:07.1215464Z preprocessor_config.json: 0% 0.00/380 [00:00, 'use_cudagraph': True, 'cudagraph_num_of_warmups': 1, 'cudagraph_capture_sizes': [512, 504, 496, 488, 480, 472, 464, 456, 448, 440, 432, 424, 416, 408, 400, 392, 384, 376, 368, 360, 352, 344, 336, 328, 320, 312, 304, 296, 288, 280, 272, 264, 256, 248, 240, 232, 224, 216, 208, 200, 192, 184, 176, 168, 160, 152, 144, 136, 128, 120, 112, 104, 96, 88, 80, 72, 64, 56, 48, 40, 32, 24, 16, 8, 4, 2, 1], 'cudagraph_copy_inputs': False, 'full_cuda_graph': True, 'use_inductor_graph_partition': False, 'pass_config': {}, 'max_capture_size': 512, 'local_cache_dir': None} 2025-11-03T17:23:15.8207184Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:23:15.8236766Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:23:15.8243972Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:23:15.8251946Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:23:15.8259561Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:23:15.8270110Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:23:15.8274495Z (EngineCore_DP0 pid=6975) INFO 11-03 17:23:15 [parallel_state.py:1231] rank 0 in world size 1 is assigned as DP rank 0, PP rank 0, TP rank 0, EP rank 0 2025-11-03T17:23:16.1607228Z (EngineCore_DP0 pid=6975) INFO 11-03 17:23:16 [topk_topp_sampler.py:59] Using FlashInfer for top-p & top-k sampling. 2025-11-03T17:23:16.2902930Z (EngineCore_DP0 pid=6975) 2025-11-03T17:23:16.2904789Z processor_config.json: 0% 0.00/115 [00:00, model_arch='SmolVLMForConditionalGeneration', exist_overrides={}, use_original_num_layers=False), 'model_impl': 'vllm', 'model': 'HuggingFaceTB/SmolVLM2-2.2B-Instruct'} 2025-11-03T17:23:21.6541913Z 2025-11-03T17:23:21.6544132Z config.json: 0.00B [00:00, ?B/s] 2025-11-03T17:23:21.6544424Z config.json: 3.64kB [00:00, 20.4MB/s] 2025-11-03T17:23:21.7749480Z 2025-11-03T17:23:21.7751259Z preprocessor_config.json: 0% 0.00/599 [00:00, 'use_cudagraph': True, 'cudagraph_num_of_warmups': 1, 'cudagraph_capture_sizes': [512, 504, 496, 488, 480, 472, 464, 456, 448, 440, 432, 424, 416, 408, 400, 392, 384, 376, 368, 360, 352, 344, 336, 328, 320, 312, 304, 296, 288, 280, 272, 264, 256, 248, 240, 232, 224, 216, 208, 200, 192, 184, 176, 168, 160, 152, 144, 136, 128, 120, 112, 104, 96, 88, 80, 72, 64, 56, 48, 40, 32, 24, 16, 8, 4, 2, 1], 'cudagraph_copy_inputs': False, 'full_cuda_graph': True, 'use_inductor_graph_partition': False, 'pass_config': {}, 'max_capture_size': 512, 'local_cache_dir': None} 2025-11-03T17:23:30.0179246Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:23:30.0208528Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:23:30.0216372Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:23:30.0224751Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:23:30.0232908Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:23:30.0240311Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:23:30.0243219Z (EngineCore_DP0 pid=7099) INFO 11-03 17:23:30 [parallel_state.py:1231] rank 0 in world size 1 is assigned as DP rank 0, PP rank 0, TP rank 0, EP rank 0 2025-11-03T17:23:30.3599092Z (EngineCore_DP0 pid=7099) INFO 11-03 17:23:30 [topk_topp_sampler.py:59] Using FlashInfer for top-p & top-k sampling. 2025-11-03T17:23:30.4876700Z (EngineCore_DP0 pid=7099) 2025-11-03T17:23:30.4878073Z processor_config.json: 0% 0.00/67.0 [00:00, model_arch='HCXVisionForCausalLM', exist_overrides={}, use_original_num_layers=False), 'model_impl': 'vllm', 'model': 'naver-hyperclovax/HyperCLOVAX-SEED-Vision-Instruct-3B'} 2025-11-03T17:23:33.7488177Z 2025-11-03T17:23:33.7490077Z config.json: 0.00B [00:00, ?B/s] 2025-11-03T17:23:33.7490432Z config.json: 5.51kB [00:00, 26.2MB/s] 2025-11-03T17:23:33.8274301Z 2025-11-03T17:23:33.8276528Z configuration_hyperclovax.py: 0.00B [00:00, ?B/s] 2025-11-03T17:23:33.8277301Z configuration_hyperclovax.py: 2.44kB [00:00, 13.0MB/s] 2025-11-03T17:23:33.8369570Z A new version of the following files was downloaded from https://huggingface.co/naver-hyperclovax/HyperCLOVAX-SEED-Vision-Instruct-3B: 2025-11-03T17:23:33.8370257Z - configuration_hyperclovax.py 2025-11-03T17:23:33.8370890Z . Make sure to double-check they do not contain any added malicious code. To avoid downloading new versions of the code file, you can pin a revision. 2025-11-03T17:23:34.0604878Z 2025-11-03T17:23:34.0605390Z preprocessor_config.json: 0.00B [00:00, ?B/s] 2025-11-03T17:23:34.0605799Z preprocessor_config.json: 1.76kB [00:00, 19.1MB/s] 2025-11-03T17:23:40.5299661Z INFO 11-03 17:23:40 [model.py:653] Resolved architecture: HCXVisionForCausalLM 2025-11-03T17:23:40.5300165Z `torch_dtype` is deprecated! Use `dtype` instead! 2025-11-03T17:23:40.5548523Z INFO 11-03 17:23:40 [model.py:1714] Using max model len 131072 2025-11-03T17:23:40.5551216Z The argument `trust_remote_code` is to be used with Auto classes. It has no effect here and is ignored. 2025-11-03T17:23:40.6331669Z INFO 11-03 17:23:40 [scheduler.py:225] Chunked prefill is enabled with max_num_batched_tokens=8192. 2025-11-03T17:23:40.6661138Z 2025-11-03T17:23:40.6663268Z tokenizer_config.json: 0.00B [00:00, ?B/s] 2025-11-03T17:23:40.6663632Z tokenizer_config.json: 11.7kB [00:00, 43.3MB/s] 2025-11-03T17:23:40.7367351Z 2025-11-03T17:23:40.7485153Z vocab.json: 0.00B [00:00, ?B/s] 2025-11-03T17:23:40.7485450Z vocab.json: 1.86MB [00:00, 156MB/s] 2025-11-03T17:23:40.7817745Z 2025-11-03T17:23:40.7871439Z merges.txt: 0.00B [00:00, ?B/s] 2025-11-03T17:23:40.7871718Z merges.txt: 1.08MB [00:00, 196MB/s] 2025-11-03T17:23:40.8227659Z 2025-11-03T17:23:40.8461201Z tokenizer.json: 0.00B [00:00, ?B/s] 2025-11-03T17:23:40.8461637Z tokenizer.json: 8.03MB [00:00, 341MB/s] 2025-11-03T17:23:40.8912393Z 2025-11-03T17:23:40.8913334Z added_tokens.json: 0% 0.00/925 [00:00, 'use_cudagraph': True, 'cudagraph_num_of_warmups': 1, 'cudagraph_capture_sizes': [512, 504, 496, 488, 480, 472, 464, 456, 448, 440, 432, 424, 416, 408, 400, 392, 384, 376, 368, 360, 352, 344, 336, 328, 320, 312, 304, 296, 288, 280, 272, 264, 256, 248, 240, 232, 224, 216, 208, 200, 192, 184, 176, 168, 160, 152, 144, 136, 128, 120, 112, 104, 96, 88, 80, 72, 64, 56, 48, 40, 32, 24, 16, 8, 4, 2, 1], 'cudagraph_copy_inputs': False, 'full_cuda_graph': True, 'use_inductor_graph_partition': False, 'pass_config': {}, 'max_capture_size': 512, 'local_cache_dir': None} 2025-11-03T17:23:42.7326613Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:23:42.7354664Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:23:42.7362464Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:23:42.7370452Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:23:42.7378392Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:23:42.7386691Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:23:42.7389420Z (EngineCore_DP0 pid=7240) INFO 11-03 17:23:42 [parallel_state.py:1231] rank 0 in world size 1 is assigned as DP rank 0, PP rank 0, TP rank 0, EP rank 0 2025-11-03T17:23:43.0907211Z (EngineCore_DP0 pid=7240) INFO 11-03 17:23:43 [topk_topp_sampler.py:59] Using FlashInfer for top-p & top-k sampling. 2025-11-03T17:23:43.5112917Z (EngineCore_DP0 pid=7240) 2025-11-03T17:23:43.5113368Z processor_config.json: 0% 0.00/120 [00:00, model_arch='Blip2ForConditionalGeneration', exist_overrides={}, use_original_num_layers=False), 'model_impl': 'vllm', 'model': 'Salesforce/blip2-opt-2.7b'} 2025-11-03T17:23:48.5649001Z 2025-11-03T17:23:48.5651364Z config.json: 0.00B [00:00, ?B/s] 2025-11-03T17:23:48.5651681Z config.json: 1.03kB [00:00, 4.95MB/s] 2025-11-03T17:23:48.6646833Z 2025-11-03T17:23:48.6647818Z preprocessor_config.json: 0% 0.00/432 [00:00, 'use_cudagraph': True, 'cudagraph_num_of_warmups': 1, 'cudagraph_capture_sizes': [512, 504, 496, 488, 480, 472, 464, 456, 448, 440, 432, 424, 416, 408, 400, 392, 384, 376, 368, 360, 352, 344, 336, 328, 320, 312, 304, 296, 288, 280, 272, 264, 256, 248, 240, 232, 224, 216, 208, 200, 192, 184, 176, 168, 160, 152, 144, 136, 128, 120, 112, 104, 96, 88, 80, 72, 64, 56, 48, 40, 32, 24, 16, 8, 4, 2, 1], 'cudagraph_copy_inputs': False, 'full_cuda_graph': True, 'use_inductor_graph_partition': False, 'pass_config': {}, 'max_capture_size': 512, 'local_cache_dir': None} 2025-11-03T17:23:56.4747965Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:23:56.4776202Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:23:56.4784271Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:23:56.4792338Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:23:56.4800387Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:23:56.4808385Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:23:56.4811373Z (EngineCore_DP0 pid=7380) INFO 11-03 17:23:56 [parallel_state.py:1231] rank 0 in world size 1 is assigned as DP rank 0, PP rank 0, TP rank 0, EP rank 0 2025-11-03T17:23:56.8209241Z (EngineCore_DP0 pid=7380) INFO 11-03 17:23:56 [topk_topp_sampler.py:59] Using FlashInfer for top-p & top-k sampling. 2025-11-03T17:23:56.9298992Z (EngineCore_DP0 pid=7380) 2025-11-03T17:23:56.9299397Z processor_config.json: 0% 0.00/68.0 [00:00, model_arch='InternLM2ForCausalLM', exist_overrides={}, use_original_num_layers=False), 'model_impl': 'vllm', 'model': 'internlm/internlm2-chat-7b'} 2025-11-03T17:24:00.0801758Z 2025-11-03T17:24:00.0802974Z config.json: 0% 0.00/916 [00:00, 'use_cudagraph': True, 'cudagraph_num_of_warmups': 1, 'cudagraph_capture_sizes': [512, 504, 496, 488, 480, 472, 464, 456, 448, 440, 432, 424, 416, 408, 400, 392, 384, 376, 368, 360, 352, 344, 336, 328, 320, 312, 304, 296, 288, 280, 272, 264, 256, 248, 240, 232, 224, 216, 208, 200, 192, 184, 176, 168, 160, 152, 144, 136, 128, 120, 112, 104, 96, 88, 80, 72, 64, 56, 48, 40, 32, 24, 16, 8, 4, 2, 1], 'cudagraph_copy_inputs': False, 'full_cuda_graph': True, 'use_inductor_graph_partition': False, 'pass_config': {}, 'max_capture_size': 512, 'local_cache_dir': None} 2025-11-03T17:24:08.9955949Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:24:08.9983457Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:24:08.9990833Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:24:08.9998545Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:24:09.0006374Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:24:09.0013844Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:24:09.0016502Z (EngineCore_DP0 pid=7541) INFO 11-03 17:24:09 [parallel_state.py:1231] rank 0 in world size 1 is assigned as DP rank 0, PP rank 0, TP rank 0, EP rank 0 2025-11-03T17:24:09.3356896Z (EngineCore_DP0 pid=7541) INFO 11-03 17:24:09 [topk_topp_sampler.py:59] Using FlashInfer for top-p & top-k sampling. 2025-11-03T17:24:09.4140263Z (EngineCore_DP0 pid=7541) INFO 11-03 17:24:09 [gpu_model_runner.py:2840] Starting to load model internlm/internlm2-chat-7b... 2025-11-03T17:24:09.6707055Z (EngineCore_DP0 pid=7541) INFO 11-03 17:24:09 [gpu_model_runner.py:2870] Loading model from scratch... 2025-11-03T17:24:09.7362210Z (EngineCore_DP0 pid=7541) INFO 11-03 17:24:09 [cuda.py:404] Using Flash Attention backend on V1 engine. 2025-11-03T17:24:10.3129442Z (EngineCore_DP0 pid=7541) INFO 11-03 17:24:10 [gpu_model_runner.py:2902] Model loading took 1.8360 GiB and 0.125680 seconds 2025-11-03T17:24:10.3133903Z (EngineCore_DP0 pid=7541) INFO 11-03 17:24:10 [kv_cache_utils.py:1199] GPU KV cache size: 2,621,440 tokens 2025-11-03T17:24:10.3134622Z (EngineCore_DP0 pid=7541) INFO 11-03 17:24:10 [kv_cache_utils.py:1204] Maximum concurrency for 65,536 tokens per request: 40.00x 2025-11-03T17:24:11.7357925Z (EngineCore_DP0 pid=7541) INFO 11-03 17:24:11 [gc_utils.py:40] GC Debug Config. enabled:False,top_objects:-1 2025-11-03T17:24:11.7371097Z INFO 11-03 17:24:11 [llm.py:337] Supported tasks: ['generate'] 2025-11-03T17:24:12.2026574Z PASSED 2025-11-03T17:24:12.2168407Z models/test_initialization.py::test_can_initialize_large_subset[EagleLlamaForCausalLM] Fork a new process to run a test 7607 2025-11-03T17:24:12.2178999Z Fork a new process to run a test 0 2025-11-03T17:24:12.2454540Z INFO 11-03 17:24:12 [utils.py:239] non-default args: {'tokenizer': 'meta-llama/Meta-Llama-3-8B-Instruct', 'trust_remote_code': True, 'load_format': 'dummy', 'gpu_memory_utilization': 0.8, 'disable_log_stats': True, 'hf_overrides': functools.partial(, model_arch='EagleLlamaForCausalLM', exist_overrides={}, use_original_num_layers=False), 'speculative_config': {'model': 'yuhuili/EAGLE-LLaMA3-Instruct-8B', 'num_speculative_tokens': 1}, 'model_impl': 'vllm', 'model': 'meta-llama/Meta-Llama-3-8B-Instruct'} 2025-11-03T17:24:12.3435944Z 2025-11-03T17:24:12.3437096Z config.json: 0% 0.00/654 [00:00, 'use_cudagraph': True, 'cudagraph_num_of_warmups': 1, 'cudagraph_capture_sizes': [512, 504, 496, 488, 480, 472, 464, 456, 448, 440, 432, 424, 416, 408, 400, 392, 384, 376, 368, 360, 352, 344, 336, 328, 320, 312, 304, 296, 288, 280, 272, 264, 256, 248, 240, 232, 224, 216, 208, 200, 192, 184, 176, 168, 160, 152, 144, 136, 128, 120, 112, 104, 96, 88, 80, 72, 64, 56, 48, 40, 32, 24, 16, 8, 4, 2, 1], 'cudagraph_copy_inputs': False, 'full_cuda_graph': True, 'use_inductor_graph_partition': False, 'pass_config': {}, 'max_capture_size': 512, 'local_cache_dir': None} 2025-11-03T17:24:14.6683165Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:24:14.6710976Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:24:14.6718250Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:24:14.6726465Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:24:14.6734293Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:24:14.6742626Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:24:14.6744795Z (EngineCore_DP0 pid=7615) INFO 11-03 17:24:14 [parallel_state.py:1231] rank 0 in world size 1 is assigned as DP rank 0, PP rank 0, TP rank 0, EP rank 0 2025-11-03T17:24:15.0040533Z (EngineCore_DP0 pid=7615) INFO 11-03 17:24:15 [topk_topp_sampler.py:59] Using FlashInfer for top-p & top-k sampling. 2025-11-03T17:24:15.0191660Z (EngineCore_DP0 pid=7615) WARNING 11-03 17:24:15 [__init__.py:194] min_p, logit_bias, and min_tokens parameters won't currently work with speculative decoding enabled. 2025-11-03T17:24:15.0573737Z (EngineCore_DP0 pid=7615) INFO 11-03 17:24:15 [gpu_model_runner.py:2840] Starting to load model meta-llama/Meta-Llama-3-8B-Instruct... 2025-11-03T17:24:15.3095691Z (EngineCore_DP0 pid=7615) INFO 11-03 17:24:15 [gpu_model_runner.py:2870] Loading model from scratch... 2025-11-03T17:24:15.3641165Z (EngineCore_DP0 pid=7615) INFO 11-03 17:24:15 [cuda.py:404] Using Flash Attention backend on V1 engine. 2025-11-03T17:24:15.4233307Z (EngineCore_DP0 pid=7615) INFO 11-03 17:24:15 [gpu_model_runner.py:2879] Loading drafter model... 2025-11-03T17:24:15.4355271Z (EngineCore_DP0 pid=7615) INFO 11-03 17:24:15 [eagle.py:1005] Assuming the EAGLE head shares the same vocab embedding with the target model. 2025-11-03T17:24:15.4356250Z (EngineCore_DP0 pid=7615) INFO 11-03 17:24:15 [eagle.py:1027] Loading EAGLE LM head weights from the target model. 2025-11-03T17:24:15.9350522Z (EngineCore_DP0 pid=7615) INFO 11-03 17:24:15 [gpu_model_runner.py:2902] Model loading took 2.8345 GiB and 0.126325 seconds 2025-11-03T17:24:15.9354477Z (EngineCore_DP0 pid=7615) INFO 11-03 17:24:15 [kv_cache_utils.py:1199] GPU KV cache size: 1,310,720 tokens 2025-11-03T17:24:15.9355902Z (EngineCore_DP0 pid=7615) INFO 11-03 17:24:15 [kv_cache_utils.py:1204] Maximum concurrency for 8,192 tokens per request: 160.00x 2025-11-03T17:24:16.6383483Z (EngineCore_DP0 pid=7615) INFO 11-03 17:24:16 [gc_utils.py:40] GC Debug Config. enabled:False,top_objects:-1 2025-11-03T17:24:16.6404186Z INFO 11-03 17:24:16 [llm.py:337] Supported tasks: ['generate'] 2025-11-03T17:24:17.1012517Z PASSED 2025-11-03T17:24:17.1156702Z models/test_initialization.py::test_can_initialize_large_subset[MiDashengLMModel] Fork a new process to run a test 7681 2025-11-03T17:24:17.1166924Z Fork a new process to run a test 0 2025-11-03T17:24:17.1443880Z INFO 11-03 17:24:17 [utils.py:239] non-default args: {'trust_remote_code': True, 'load_format': 'dummy', 'gpu_memory_utilization': 0.8, 'disable_log_stats': True, 'hf_overrides': functools.partial(, model_arch='MiDashengLMModel', exist_overrides={}, use_original_num_layers=False), 'model_impl': 'vllm', 'model': 'mispeech/midashenglm-7b'} 2025-11-03T17:24:17.2945813Z 2025-11-03T17:24:17.2946800Z config.json: 0.00B [00:00, ?B/s] 2025-11-03T17:24:17.2947227Z config.json: 1.65kB [00:00, 11.0MB/s] 2025-11-03T17:24:17.3705670Z Unrecognized keys in `rope_scaling` for 'rope_type'='default': {'mrope_section'} 2025-11-03T17:24:17.4864898Z 2025-11-03T17:24:17.4866041Z preprocessor_config.json: 0% 0.00/349 [00:00, 'use_cudagraph': True, 'cudagraph_num_of_warmups': 1, 'cudagraph_capture_sizes': [512, 504, 496, 488, 480, 472, 464, 456, 448, 440, 432, 424, 416, 408, 400, 392, 384, 376, 368, 360, 352, 344, 336, 328, 320, 312, 304, 296, 288, 280, 272, 264, 256, 248, 240, 232, 224, 216, 208, 200, 192, 184, 176, 168, 160, 152, 144, 136, 128, 120, 112, 104, 96, 88, 80, 72, 64, 56, 48, 40, 32, 24, 16, 8, 4, 2, 1], 'cudagraph_copy_inputs': False, 'full_cuda_graph': True, 'use_inductor_graph_partition': False, 'pass_config': {}, 'max_capture_size': 512, 'local_cache_dir': None} 2025-11-03T17:24:26.7817640Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:24:26.7845713Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:24:26.7853128Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:24:26.7860460Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:24:26.7869069Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:24:26.7876151Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:24:26.7878912Z (EngineCore_DP0 pid=7760) INFO 11-03 17:24:26 [parallel_state.py:1231] rank 0 in world size 1 is assigned as DP rank 0, PP rank 0, TP rank 0, EP rank 0 2025-11-03T17:24:27.1301279Z (EngineCore_DP0 pid=7760) INFO 11-03 17:24:27 [topk_topp_sampler.py:59] Using FlashInfer for top-p & top-k sampling. 2025-11-03T17:24:27.2853416Z (EngineCore_DP0 pid=7760) 2025-11-03T17:24:27.2854020Z processor_config.json: 0% 0.00/268 [00:00, model_arch='ApertusForCausalLM', exist_overrides={}, use_original_num_layers=False), 'model_impl': 'vllm', 'model': 'swiss-ai/Apertus-8B-Instruct-2509'} 2025-11-03T17:24:34.0422145Z 2025-11-03T17:24:34.0423699Z config.json: 0% 0.00/901 [00:00, 'use_cudagraph': True, 'cudagraph_num_of_warmups': 1, 'cudagraph_capture_sizes': [512, 504, 496, 488, 480, 472, 464, 456, 448, 440, 432, 424, 416, 408, 400, 392, 384, 376, 368, 360, 352, 344, 336, 328, 320, 312, 304, 296, 288, 280, 272, 264, 256, 248, 240, 232, 224, 216, 208, 200, 192, 184, 176, 168, 160, 152, 144, 136, 128, 120, 112, 104, 96, 88, 80, 72, 64, 56, 48, 40, 32, 24, 16, 8, 4, 2, 1], 'cudagraph_copy_inputs': False, 'full_cuda_graph': True, 'use_inductor_graph_partition': False, 'pass_config': {}, 'max_capture_size': 512, 'local_cache_dir': None} 2025-11-03T17:24:42.5084533Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:24:42.5112391Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:24:42.5119363Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:24:42.5127702Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:24:42.5135354Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:24:42.5143692Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:24:42.5146665Z (EngineCore_DP0 pid=7921) INFO 11-03 17:24:42 [parallel_state.py:1231] rank 0 in world size 1 is assigned as DP rank 0, PP rank 0, TP rank 0, EP rank 0 2025-11-03T17:24:42.8440976Z (EngineCore_DP0 pid=7921) INFO 11-03 17:24:42 [topk_topp_sampler.py:59] Using FlashInfer for top-p & top-k sampling. 2025-11-03T17:24:42.9520112Z (EngineCore_DP0 pid=7921) INFO 11-03 17:24:42 [gpu_model_runner.py:2840] Starting to load model swiss-ai/Apertus-8B-Instruct-2509... 2025-11-03T17:24:43.2044375Z (EngineCore_DP0 pid=7921) INFO 11-03 17:24:43 [gpu_model_runner.py:2870] Loading model from scratch... 2025-11-03T17:24:43.3004481Z (EngineCore_DP0 pid=7921) INFO 11-03 17:24:43 [cuda.py:404] Using Flash Attention backend on V1 engine. 2025-11-03T17:24:43.3257460Z (EngineCore_DP0 pid=7921) WARNING 11-03 17:24:43 [activation.py:420] CUDA-fused xIELU not available (No module named 'xielu') – falling back to a Python version. 2025-11-03T17:24:43.3258123Z 2025-11-03T17:24:43.3258750Z (EngineCore_DP0 pid=7921) WARNING 11-03 17:24:43 [activation.py:420] For CUDA xIELU (experimental), `pip install git+https://github.com/nickjbrowning/XIELU` 2025-11-03T17:24:43.8822559Z (EngineCore_DP0 pid=7921) INFO 11-03 17:24:43 [gpu_model_runner.py:2902] Model loading took 2.4219 GiB and 0.179758 seconds 2025-11-03T17:24:43.8827019Z (EngineCore_DP0 pid=7921) INFO 11-03 17:24:43 [kv_cache_utils.py:1199] GPU KV cache size: 2,621,440 tokens 2025-11-03T17:24:43.8828279Z (EngineCore_DP0 pid=7921) INFO 11-03 17:24:43 [kv_cache_utils.py:1204] Maximum concurrency for 65,536 tokens per request: 40.00x 2025-11-03T17:24:44.8994555Z (EngineCore_DP0 pid=7921) INFO 11-03 17:24:44 [gc_utils.py:40] GC Debug Config. enabled:False,top_objects:-1 2025-11-03T17:24:44.9008622Z INFO 11-03 17:24:44 [llm.py:337] Supported tasks: ['generate'] 2025-11-03T17:24:45.3860138Z PASSED 2025-11-03T17:24:45.4003709Z models/test_initialization.py::test_can_initialize_large_subset[LlavaOnevisionForConditionalGeneration] Fork a new process to run a test 7987 2025-11-03T17:24:45.4014163Z Fork a new process to run a test 0 2025-11-03T17:24:45.4294562Z INFO 11-03 17:24:45 [utils.py:239] non-default args: {'load_format': 'dummy', 'gpu_memory_utilization': 0.8, 'disable_log_stats': True, 'hf_overrides': functools.partial(, model_arch='LlavaOnevisionForConditionalGeneration', exist_overrides={}, use_original_num_layers=False), 'model_impl': 'vllm', 'model': 'llava-hf/llava-onevision-qwen2-0.5b-ov-hf'} 2025-11-03T17:24:45.5194327Z 2025-11-03T17:24:45.5196693Z config.json: 0.00B [00:00, ?B/s] 2025-11-03T17:24:45.5197126Z config.json: 2.59kB [00:00, 12.0MB/s] 2025-11-03T17:24:45.6625766Z 2025-11-03T17:24:45.6626953Z preprocessor_config.json: 0.00B [00:00, ?B/s] 2025-11-03T17:24:45.6627432Z preprocessor_config.json: 1.73kB [00:00, 18.1MB/s] 2025-11-03T17:24:52.3659875Z INFO 11-03 17:24:52 [model.py:653] Resolved architecture: LlavaOnevisionForConditionalGeneration 2025-11-03T17:24:52.3660678Z `torch_dtype` is deprecated! Use `dtype` instead! 2025-11-03T17:24:52.3914514Z INFO 11-03 17:24:52 [model.py:1714] Using max model len 32768 2025-11-03T17:24:52.5583168Z INFO 11-03 17:24:52 [scheduler.py:225] Chunked prefill is enabled with max_num_batched_tokens=8192. 2025-11-03T17:24:52.6111531Z 2025-11-03T17:24:52.6112951Z tokenizer_config.json: 0.00B [00:00, ?B/s] 2025-11-03T17:24:52.6113524Z tokenizer_config.json: 1.80kB [00:00, 13.1MB/s] 2025-11-03T17:24:52.6936899Z 2025-11-03T17:24:52.7104364Z vocab.json: 0.00B [00:00, ?B/s] 2025-11-03T17:24:52.7104858Z vocab.json: 2.78MB [00:00, 164MB/s] 2025-11-03T17:24:52.7447363Z 2025-11-03T17:24:52.7532462Z merges.txt: 0.00B [00:00, ?B/s] 2025-11-03T17:24:52.7532925Z merges.txt: 1.67MB [00:00, 196MB/s] 2025-11-03T17:24:52.7951715Z 2025-11-03T17:24:52.8190104Z tokenizer.json: 0.00B [00:00, ?B/s] 2025-11-03T17:24:52.8190616Z tokenizer.json: 7.03MB [00:00, 295MB/s] 2025-11-03T17:24:52.8548598Z 2025-11-03T17:24:52.8550780Z added_tokens.json: 0% 0.00/122 [00:00, 'use_cudagraph': True, 'cudagraph_num_of_warmups': 1, 'cudagraph_capture_sizes': [512, 504, 496, 488, 480, 472, 464, 456, 448, 440, 432, 424, 416, 408, 400, 392, 384, 376, 368, 360, 352, 344, 336, 328, 320, 312, 304, 296, 288, 280, 272, 264, 256, 248, 240, 232, 224, 216, 208, 200, 192, 184, 176, 168, 160, 152, 144, 136, 128, 120, 112, 104, 96, 88, 80, 72, 64, 56, 48, 40, 32, 24, 16, 8, 4, 2, 1], 'cudagraph_copy_inputs': False, 'full_cuda_graph': True, 'use_inductor_graph_partition': False, 'pass_config': {}, 'max_capture_size': 512, 'local_cache_dir': None} 2025-11-03T17:25:01.0362305Z (EngineCore_DP0 pid=8047) /opt/conda/envs/py_3.12/lib/python3.12/site-packages/torch/library.py:356: UserWarning: Warning only once for all operators, other operators may also be overridden. 2025-11-03T17:25:01.0363625Z (EngineCore_DP0 pid=8047) Overriding a previously registered kernel for the same operator and the same dispatch key 2025-11-03T17:25:01.0365409Z (EngineCore_DP0 pid=8047) operator: xformers_python::_fused_allgather_and_linear_impl(Tensor scattered_input, Tensor[] weights, str process_group_name, Tensor(a3!)[] gathered_outputs, SymInt timeout_s, bool _wait, bool _memcpy, Tensor scale_scattered_input, Tensor?[] scales_weights) -> () 2025-11-03T17:25:01.0366931Z (EngineCore_DP0 pid=8047) registered at /opt/conda/envs/py_3.12/lib/python3.12/site-packages/torch/_library/custom_ops.py:922 2025-11-03T17:25:01.0367677Z (EngineCore_DP0 pid=8047) dispatch key: ADInplaceOrView 2025-11-03T17:25:01.0368203Z (EngineCore_DP0 pid=8047) previous kernel: no debug info 2025-11-03T17:25:01.0369503Z (EngineCore_DP0 pid=8047) new kernel: registered at /opt/conda/envs/py_3.12/lib/python3.12/site-packages/torch/_library/custom_ops.py:922 (Triggered internally at /var/lib/jenkins/workspace/aten/src/ATen/core/dispatch/OperatorEntry.cpp:208.) 2025-11-03T17:25:01.0370607Z (EngineCore_DP0 pid=8047) self.m.impl( 2025-11-03T17:25:01.6759168Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:25:01.6789424Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:25:01.6796411Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:25:01.6803572Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:25:01.6811135Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:25:01.6817216Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:25:01.6820063Z (EngineCore_DP0 pid=8047) INFO 11-03 17:25:01 [parallel_state.py:1231] rank 0 in world size 1 is assigned as DP rank 0, PP rank 0, TP rank 0, EP rank 0 2025-11-03T17:25:01.9138652Z (EngineCore_DP0 pid=8047) INFO 11-03 17:25:01 [topk_topp_sampler.py:59] Using FlashInfer for top-p & top-k sampling. 2025-11-03T17:25:02.0340376Z (EngineCore_DP0 pid=8047) 2025-11-03T17:25:02.0340977Z processor_config.json: 0% 0.00/178 [00:00, model_arch='OrionForCausalLM', exist_overrides={}, use_original_num_layers=False), 'model_impl': 'vllm', 'model': 'OrionStarAI/Orion-14B-Chat'} 2025-11-03T17:25:16.6688940Z 2025-11-03T17:25:16.6689562Z config.json: 0% 0.00/771 [00:00, 'use_cudagraph': True, 'cudagraph_num_of_warmups': 1, 'cudagraph_capture_sizes': [512, 504, 496, 488, 480, 472, 464, 456, 448, 440, 432, 424, 416, 408, 400, 392, 384, 376, 368, 360, 352, 344, 336, 328, 320, 312, 304, 296, 288, 280, 272, 264, 256, 248, 240, 232, 224, 216, 208, 200, 192, 184, 176, 168, 160, 152, 144, 136, 128, 120, 112, 104, 96, 88, 80, 72, 64, 56, 48, 40, 32, 24, 16, 8, 4, 2, 1], 'cudagraph_copy_inputs': False, 'full_cuda_graph': True, 'use_inductor_graph_partition': False, 'pass_config': {}, 'max_capture_size': 512, 'local_cache_dir': None} 2025-11-03T17:25:25.2205893Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:25:25.2234016Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:25:25.2242401Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:25:25.2250101Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:25:25.2257822Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:25:25.2266478Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:25:25.2269295Z (EngineCore_DP0 pid=8388) INFO 11-03 17:25:25 [parallel_state.py:1231] rank 0 in world size 1 is assigned as DP rank 0, PP rank 0, TP rank 0, EP rank 0 2025-11-03T17:25:25.5617217Z (EngineCore_DP0 pid=8388) INFO 11-03 17:25:25 [topk_topp_sampler.py:59] Using FlashInfer for top-p & top-k sampling. 2025-11-03T17:25:25.6609257Z (EngineCore_DP0 pid=8388) INFO 11-03 17:25:25 [gpu_model_runner.py:2840] Starting to load model OrionStarAI/Orion-14B-Chat... 2025-11-03T17:25:25.9172750Z (EngineCore_DP0 pid=8388) INFO 11-03 17:25:25 [gpu_model_runner.py:2870] Loading model from scratch... 2025-11-03T17:25:25.9824404Z (EngineCore_DP0 pid=8388) INFO 11-03 17:25:25 [cuda.py:404] Using Flash Attention backend on V1 engine. 2025-11-03T17:25:26.5503419Z (EngineCore_DP0 pid=8388) INFO 11-03 17:25:26 [gpu_model_runner.py:2902] Model loading took 2.2496 GiB and 0.124945 seconds 2025-11-03T17:25:26.5508585Z (EngineCore_DP0 pid=8388) INFO 11-03 17:25:26 [kv_cache_utils.py:1199] GPU KV cache size: 524,288 tokens 2025-11-03T17:25:26.5509552Z (EngineCore_DP0 pid=8388) INFO 11-03 17:25:26 [kv_cache_utils.py:1204] Maximum concurrency for 4,096 tokens per request: 128.00x 2025-11-03T17:25:26.8407558Z (EngineCore_DP0 pid=8388) WARNING 11-03 17:25:26 [tokenizer.py:255] Using a slow tokenizer. This might cause a significant slowdown. Consider using a fast tokenizer instead. 2025-11-03T17:25:27.2838014Z (EngineCore_DP0 pid=8388) INFO 11-03 17:25:27 [gc_utils.py:40] GC Debug Config. enabled:False,top_objects:-1 2025-11-03T17:25:27.2854829Z INFO 11-03 17:25:27 [llm.py:337] Supported tasks: ['generate'] 2025-11-03T17:25:27.7606067Z PASSED 2025-11-03T17:25:27.7750358Z models/test_initialization.py::test_can_initialize_large_subset[Qwen2Model] Fork a new process to run a test 8454 2025-11-03T17:25:27.7760269Z Fork a new process to run a test 0 2025-11-03T17:25:27.8038671Z INFO 11-03 17:25:27 [utils.py:239] non-default args: {'load_format': 'dummy', 'gpu_memory_utilization': 0.8, 'disable_log_stats': True, 'hf_overrides': functools.partial(, model_arch='Qwen2Model', exist_overrides={}, use_original_num_layers=False), 'model_impl': 'vllm', 'model': 'ssmits/Qwen2-7B-Instruct-embed-base'} 2025-11-03T17:25:27.8948954Z 2025-11-03T17:25:27.8950163Z config.json: 0% 0.00/706 [00:00, 'use_cudagraph': True, 'cudagraph_num_of_warmups': 1, 'cudagraph_capture_sizes': [512, 504, 496, 488, 480, 472, 464, 456, 448, 440, 432, 424, 416, 408, 400, 392, 384, 376, 368, 360, 352, 344, 336, 328, 320, 312, 304, 296, 288, 280, 272, 264, 256, 248, 240, 232, 224, 216, 208, 200, 192, 184, 176, 168, 160, 152, 144, 136, 128, 120, 112, 104, 96, 88, 80, 72, 64, 56, 48, 40, 32, 24, 16, 8, 4, 2, 1], 'cudagraph_copy_inputs': False, 'full_cuda_graph': True, 'use_inductor_graph_partition': False, 'pass_config': {}, 'max_capture_size': 512, 'local_cache_dir': None} 2025-11-03T17:25:35.8869703Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:25:35.8897445Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:25:35.8905091Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:25:35.8912923Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:25:35.8920576Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:25:35.8928568Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:25:35.8931132Z (EngineCore_DP0 pid=8512) INFO 11-03 17:25:35 [parallel_state.py:1231] rank 0 in world size 1 is assigned as DP rank 0, PP rank 0, TP rank 0, EP rank 0 2025-11-03T17:25:36.2245645Z (EngineCore_DP0 pid=8512) INFO 11-03 17:25:36 [topk_topp_sampler.py:59] Using FlashInfer for top-p & top-k sampling. 2025-11-03T17:25:36.2934559Z (EngineCore_DP0 pid=8512) INFO 11-03 17:25:36 [gpu_model_runner.py:2840] Starting to load model ssmits/Qwen2-7B-Instruct-embed-base... 2025-11-03T17:25:36.5477394Z (EngineCore_DP0 pid=8512) INFO 11-03 17:25:36 [gpu_model_runner.py:2870] Loading model from scratch... 2025-11-03T17:25:36.6129634Z (EngineCore_DP0 pid=8512) INFO 11-03 17:25:36 [cuda.py:404] Using Flash Attention backend on V1 engine. 2025-11-03T17:25:37.1760810Z (EngineCore_DP0 pid=8512) INFO 11-03 17:25:37 [gpu_model_runner.py:2902] Model loading took 2.4751 GiB and 0.124926 seconds 2025-11-03T17:25:37.1765298Z (EngineCore_DP0 pid=8512) INFO 11-03 17:25:37 [kv_cache_utils.py:1199] GPU KV cache size: 5,242,880 tokens 2025-11-03T17:25:37.1766727Z (EngineCore_DP0 pid=8512) INFO 11-03 17:25:37 [kv_cache_utils.py:1204] Maximum concurrency for 32,768 tokens per request: 160.00x 2025-11-03T17:25:38.3027290Z (EngineCore_DP0 pid=8512) INFO 11-03 17:25:38 [gc_utils.py:40] GC Debug Config. enabled:False,top_objects:-1 2025-11-03T17:25:38.3043734Z INFO 11-03 17:25:38 [llm.py:337] Supported tasks: ['generate'] 2025-11-03T17:25:38.7673346Z PASSED 2025-11-03T17:25:38.7816970Z models/test_initialization.py::test_can_initialize_large_subset[Idefics3ForConditionalGeneration] Fork a new process to run a test 8578 2025-11-03T17:25:38.7827527Z Fork a new process to run a test 0 2025-11-03T17:25:38.8099112Z INFO 11-03 17:25:38 [utils.py:239] non-default args: {'load_format': 'dummy', 'gpu_memory_utilization': 0.8, 'disable_log_stats': True, 'hf_overrides': functools.partial(, model_arch='Idefics3ForConditionalGeneration', exist_overrides={}, use_original_num_layers=False), 'model_impl': 'vllm', 'model': 'HuggingFaceM4/Idefics3-8B-Llama3'} 2025-11-03T17:25:38.8989440Z 2025-11-03T17:25:38.8992463Z config.json: 0.00B [00:00, ?B/s] 2025-11-03T17:25:38.8992727Z config.json: 7.18kB [00:00, 24.9MB/s] 2025-11-03T17:25:39.0377466Z 2025-11-03T17:25:39.0378988Z preprocessor_config.json: 0% 0.00/435 [00:00, 'use_cudagraph': True, 'cudagraph_num_of_warmups': 1, 'cudagraph_capture_sizes': [512, 504, 496, 488, 480, 472, 464, 456, 448, 440, 432, 424, 416, 408, 400, 392, 384, 376, 368, 360, 352, 344, 336, 328, 320, 312, 304, 296, 288, 280, 272, 264, 256, 248, 240, 232, 224, 216, 208, 200, 192, 184, 176, 168, 160, 152, 144, 136, 128, 120, 112, 104, 96, 88, 80, 72, 64, 56, 48, 40, 32, 24, 16, 8, 4, 2, 1], 'cudagraph_copy_inputs': False, 'full_cuda_graph': True, 'use_inductor_graph_partition': False, 'pass_config': {}, 'max_capture_size': 512, 'local_cache_dir': None} 2025-11-03T17:25:47.4250364Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:25:47.4279708Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:25:47.4287019Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:25:47.4294682Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:25:47.4302325Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:25:47.4310109Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:25:47.4313536Z (EngineCore_DP0 pid=8636) INFO 11-03 17:25:47 [parallel_state.py:1231] rank 0 in world size 1 is assigned as DP rank 0, PP rank 0, TP rank 0, EP rank 0 2025-11-03T17:25:47.7621415Z (EngineCore_DP0 pid=8636) INFO 11-03 17:25:47 [topk_topp_sampler.py:59] Using FlashInfer for top-p & top-k sampling. 2025-11-03T17:25:49.5029644Z (EngineCore_DP0 pid=8636) 2025-11-03T17:25:49.5030944Z chat_template.json: 0% 0.00/434 [00:00, model_arch='MiniMaxForCausalLM', exist_overrides={}, use_original_num_layers=False), 'model_impl': 'vllm', 'model': 'MiniMaxAI/MiniMax-Text-01-hf'} 2025-11-03T17:25:52.8996149Z 2025-11-03T17:25:52.8998032Z config.json: 0.00B [00:00, ?B/s] 2025-11-03T17:25:52.8998495Z config.json: 3.00kB [00:00, 17.6MB/s] 2025-11-03T17:25:59.2691261Z INFO 11-03 17:25:59 [model.py:653] Resolved architecture: MiniMaxForCausalLM 2025-11-03T17:25:59.2691758Z `torch_dtype` is deprecated! Use `dtype` instead! 2025-11-03T17:25:59.4095676Z 2025-11-03T17:25:59.4116303Z model.safetensors.index.json: 0.00B [00:00, ?B/s] 2025-11-03T17:25:59.4116790Z model.safetensors.index.json: 823kB [00:00, 401MB/s] 2025-11-03T17:25:59.4377189Z 2025-11-03T17:25:59.5505869Z Parse safetensors files: 0% 0/413 [00:00= mamba page size. 2025-11-03T17:26:01.2287090Z 2025-11-03T17:26:01.2288776Z tokenizer_config.json: 0.00B [00:00, ?B/s] 2025-11-03T17:26:01.2289125Z tokenizer_config.json: 1.39kB [00:00, 7.15MB/s] 2025-11-03T17:26:01.3375596Z 2025-11-03T17:26:01.3580080Z vocab.json: 0.00B [00:00, ?B/s] 2025-11-03T17:26:01.3580422Z vocab.json: 4.71MB [00:00, 228MB/s] 2025-11-03T17:26:01.4136796Z 2025-11-03T17:26:01.4260622Z merges.txt: 0.00B [00:00, ?B/s] 2025-11-03T17:26:01.4261002Z merges.txt: 2.41MB [00:00, 194MB/s] 2025-11-03T17:26:01.5135835Z 2025-11-03T17:26:01.5497540Z tokenizer.json: 0.00B [00:00, ?B/s] 2025-11-03T17:26:01.5497886Z tokenizer.json: 9.72MB [00:00, 268MB/s] 2025-11-03T17:26:02.4433028Z (EngineCore_DP0 pid=8796) INFO 11-03 17:26:02 [core.py:727] Waiting for init message from front-end. 2025-11-03T17:26:02.4472071Z (EngineCore_DP0 pid=8796) INFO 11-03 17:26:02 [core.py:94] Initializing a V1 LLM engine (v0.11.0rc2.dev389+ge51928192.d20251103) with config: model='MiniMaxAI/MiniMax-Text-01-hf', speculative_config=None, tokenizer='MiniMaxAI/MiniMax-Text-01-hf', skip_tokenizer_init=False, tokenizer_mode=auto, revision=None, tokenizer_revision=None, trust_remote_code=False, dtype=torch.bfloat16, max_seq_len=10240000, download_dir=None, load_format=dummy, tensor_parallel_size=1, pipeline_parallel_size=1, data_parallel_size=1, disable_custom_all_reduce=False, quantization=None, enforce_eager=False, kv_cache_dtype=auto, device_config=cuda, structured_outputs_config=StructuredOutputsConfig(backend='auto', disable_fallback=False, disable_any_whitespace=False, disable_additional_properties=False, reasoning_parser=''), observability_config=ObservabilityConfig(show_hidden_metrics_for_version=None, otlp_traces_endpoint=None, collect_detailed_traces=None), seed=0, served_model_name=MiniMaxAI/MiniMax-Text-01-hf, enable_prefix_caching=False, chunked_prefill_enabled=True, pooler_config=None, compilation_config={'level': 3, 'debug_dump_path': None, 'cache_dir': '', 'backend': '', 'custom_ops': [], 'splitting_ops': ['vllm.unified_attention', 'vllm.unified_attention_with_output', 'vllm.unified_mla_attention', 'vllm.unified_mla_attention_with_output', 'vllm.mamba_mixer2', 'vllm.mamba_mixer', 'vllm.short_conv', 'vllm.linear_attention', 'vllm.plamo2_mamba_mixer', 'vllm.gdn_attention', 'vllm.sparse_attn_indexer'], 'use_inductor': True, 'compile_sizes': [], 'inductor_compile_config': {'enable_auto_functionalized_v2': False}, 'inductor_passes': {}, 'cudagraph_mode': , 'use_cudagraph': True, 'cudagraph_num_of_warmups': 1, 'cudagraph_capture_sizes': [512, 504, 496, 488, 480, 472, 464, 456, 448, 440, 432, 424, 416, 408, 400, 392, 384, 376, 368, 360, 352, 344, 336, 328, 320, 312, 304, 296, 288, 280, 272, 264, 256, 248, 240, 232, 224, 216, 208, 200, 192, 184, 176, 168, 160, 152, 144, 136, 128, 120, 112, 104, 96, 88, 80, 72, 64, 56, 48, 40, 32, 24, 16, 8, 4, 2, 1], 'cudagraph_copy_inputs': False, 'full_cuda_graph': True, 'use_inductor_graph_partition': False, 'pass_config': {}, 'max_capture_size': 512, 'local_cache_dir': None} 2025-11-03T17:26:03.1571153Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:26:03.1600566Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:26:03.1609130Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:26:03.1618637Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:26:03.1627575Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:26:03.1635992Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:26:03.1638717Z (EngineCore_DP0 pid=8796) INFO 11-03 17:26:03 [parallel_state.py:1231] rank 0 in world size 1 is assigned as DP rank 0, PP rank 0, TP rank 0, EP rank 0 2025-11-03T17:26:03.5038306Z (EngineCore_DP0 pid=8796) INFO 11-03 17:26:03 [topk_topp_sampler.py:59] Using FlashInfer for top-p & top-k sampling. 2025-11-03T17:26:04.9704490Z (EngineCore_DP0 pid=8796) INFO 11-03 17:26:04 [gpu_model_runner.py:2840] Starting to load model MiniMaxAI/MiniMax-Text-01-hf... 2025-11-03T17:26:05.2295815Z (EngineCore_DP0 pid=8796) INFO 11-03 17:26:05 [gpu_model_runner.py:2870] Loading model from scratch... 2025-11-03T17:26:05.9650788Z (EngineCore_DP0 pid=8796) INFO 11-03 17:26:05 [gpu_model_runner.py:2902] Model loading took 7.9713 GiB and 0.111534 seconds 2025-11-03T17:26:05.9656055Z (EngineCore_DP0 pid=8796) INFO 11-03 17:26:05 [kv_cache_utils.py:1199] GPU KV cache size: 52,428,800,000 tokens 2025-11-03T17:26:05.9656801Z (EngineCore_DP0 pid=8796) INFO 11-03 17:26:05 [kv_cache_utils.py:1204] Maximum concurrency for 10,240,000 tokens per request: 5120.00x 2025-11-03T17:26:06.8626925Z (EngineCore_DP0 pid=8796) INFO 11-03 17:26:06 [gc_utils.py:40] GC Debug Config. enabled:False,top_objects:-1 2025-11-03T17:26:06.8644530Z INFO 11-03 17:26:06 [llm.py:337] Supported tasks: ['generate'] 2025-11-03T17:26:08.3049937Z PASSED 2025-11-03T17:26:08.3192682Z models/test_initialization.py::test_can_initialize_large_subset[GlmForCausalLM] Fork a new process to run a test 8862 2025-11-03T17:26:08.3203277Z Fork a new process to run a test 0 2025-11-03T17:26:08.3482924Z INFO 11-03 17:26:08 [utils.py:239] non-default args: {'load_format': 'dummy', 'gpu_memory_utilization': 0.8, 'disable_log_stats': True, 'hf_overrides': functools.partial(, model_arch='GlmForCausalLM', exist_overrides={}, use_original_num_layers=False), 'model_impl': 'vllm', 'model': 'zai-org/glm-4-9b-chat-hf'} 2025-11-03T17:26:08.4307987Z 2025-11-03T17:26:08.4309483Z config.json: 0% 0.00/661 [00:00, 'use_cudagraph': True, 'cudagraph_num_of_warmups': 1, 'cudagraph_capture_sizes': [512, 504, 496, 488, 480, 472, 464, 456, 448, 440, 432, 424, 416, 408, 400, 392, 384, 376, 368, 360, 352, 344, 336, 328, 320, 312, 304, 296, 288, 280, 272, 264, 256, 248, 240, 232, 224, 216, 208, 200, 192, 184, 176, 168, 160, 152, 144, 136, 128, 120, 112, 104, 96, 88, 80, 72, 64, 56, 48, 40, 32, 24, 16, 8, 4, 2, 1], 'cudagraph_copy_inputs': False, 'full_cuda_graph': True, 'use_inductor_graph_partition': False, 'pass_config': {}, 'max_capture_size': 512, 'local_cache_dir': None} 2025-11-03T17:26:17.3393845Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:26:17.3421844Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:26:17.3429145Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:26:17.3436719Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:26:17.3444209Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:26:17.3452339Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:26:17.3454985Z (EngineCore_DP0 pid=8940) INFO 11-03 17:26:17 [parallel_state.py:1231] rank 0 in world size 1 is assigned as DP rank 0, PP rank 0, TP rank 0, EP rank 0 2025-11-03T17:26:17.6786576Z (EngineCore_DP0 pid=8940) INFO 11-03 17:26:17 [topk_topp_sampler.py:59] Using FlashInfer for top-p & top-k sampling. 2025-11-03T17:26:17.7786895Z (EngineCore_DP0 pid=8940) INFO 11-03 17:26:17 [gpu_model_runner.py:2840] Starting to load model zai-org/glm-4-9b-chat-hf... 2025-11-03T17:26:18.0350433Z (EngineCore_DP0 pid=8940) INFO 11-03 17:26:18 [gpu_model_runner.py:2870] Loading model from scratch... 2025-11-03T17:26:18.1195973Z (EngineCore_DP0 pid=8940) INFO 11-03 17:26:18 [cuda.py:404] Using Flash Attention backend on V1 engine. 2025-11-03T17:26:18.6850280Z (EngineCore_DP0 pid=8940) INFO 11-03 17:26:18 [gpu_model_runner.py:2902] Model loading took 2.7090 GiB and 0.144728 seconds 2025-11-03T17:26:18.6855407Z (EngineCore_DP0 pid=8940) INFO 11-03 17:26:18 [kv_cache_utils.py:1199] GPU KV cache size: 10,485,760 tokens 2025-11-03T17:26:18.6856769Z (EngineCore_DP0 pid=8940) INFO 11-03 17:26:18 [kv_cache_utils.py:1204] Maximum concurrency for 131,072 tokens per request: 80.00x 2025-11-03T17:26:20.5912106Z (EngineCore_DP0 pid=8940) INFO 11-03 17:26:20 [gc_utils.py:40] GC Debug Config. enabled:False,top_objects:-1 2025-11-03T17:26:20.5926157Z INFO 11-03 17:26:20 [llm.py:337] Supported tasks: ['generate'] 2025-11-03T17:26:21.1079929Z PASSED 2025-11-03T17:26:21.1222444Z models/test_initialization.py::test_can_initialize_large_subset[RobertaModel] Fork a new process to run a test 9006 2025-11-03T17:26:21.1233730Z Fork a new process to run a test 0 2025-11-03T17:26:21.1517938Z INFO 11-03 17:26:21 [utils.py:239] non-default args: {'load_format': 'dummy', 'gpu_memory_utilization': 0.8, 'disable_log_stats': True, 'hf_overrides': functools.partial(, model_arch='RobertaModel', exist_overrides={}, use_original_num_layers=False), 'model_impl': 'vllm', 'model': 'sentence-transformers/stsb-roberta-base-v2'} 2025-11-03T17:26:21.2537303Z 2025-11-03T17:26:21.2538492Z config.json: 0% 0.00/675 [00:00, 'use_cudagraph': True, 'cudagraph_num_of_warmups': 1, 'cudagraph_capture_sizes': [72, 64, 56, 48, 40, 32, 24, 16, 8, 4, 2, 1], 'cudagraph_copy_inputs': False, 'full_cuda_graph': False, 'use_inductor_graph_partition': False, 'pass_config': {}, 'max_capture_size': 72, 'local_cache_dir': None} 2025-11-03T17:26:29.4637601Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:26:29.4668155Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:26:29.4675941Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:26:29.4683369Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:26:29.4690766Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:26:29.4698506Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:26:29.4701222Z (EngineCore_DP0 pid=9065) INFO 11-03 17:26:29 [parallel_state.py:1231] rank 0 in world size 1 is assigned as DP rank 0, PP rank 0, TP rank 0, EP rank 0 2025-11-03T17:26:29.8138135Z (EngineCore_DP0 pid=9065) INFO 11-03 17:26:29 [topk_topp_sampler.py:59] Using FlashInfer for top-p & top-k sampling. 2025-11-03T17:26:29.8271428Z (EngineCore_DP0 pid=9065) INFO 11-03 17:26:29 [gpu_model_runner.py:2840] Starting to load model sentence-transformers/stsb-roberta-base-v2... 2025-11-03T17:26:30.0764779Z (EngineCore_DP0 pid=9065) INFO 11-03 17:26:30 [gpu_model_runner.py:2870] Loading model from scratch... 2025-11-03T17:26:30.0820233Z (EngineCore_DP0 pid=9065) INFO 11-03 17:26:30 [cuda.py:404] Using Flash Attention backend on V1 engine. 2025-11-03T17:26:30.6455085Z (EngineCore_DP0 pid=9065) INFO 11-03 17:26:30 [gpu_model_runner.py:2902] Model loading took 0.0862 GiB and 0.065899 seconds 2025-11-03T17:26:30.8374973Z (EngineCore_DP0 pid=9065) INFO 11-03 17:26:30 [core.py:142] Disabling chunked prefill for model without KVCache 2025-11-03T17:26:30.8403478Z (EngineCore_DP0 pid=9065) INFO 11-03 17:26:30 [vllm.py:433] Only "last" pooling supports chunked prefill and prefix caching; disabling both. 2025-11-03T17:26:31.1010869Z (EngineCore_DP0 pid=9065) INFO 11-03 17:26:31 [gc_utils.py:40] GC Debug Config. enabled:False,top_objects:-1 2025-11-03T17:26:31.1025694Z INFO 11-03 17:26:31 [llm.py:337] Supported tasks: ['embed', 'encode'] 2025-11-03T17:26:31.5275408Z PASSED 2025-11-03T17:26:31.5415577Z models/test_initialization.py::test_can_initialize_large_subset[Gemma2ForCausalLM] Fork a new process to run a test 9131 2025-11-03T17:26:31.5425505Z Fork a new process to run a test 0 2025-11-03T17:26:31.5704420Z INFO 11-03 17:26:31 [utils.py:239] non-default args: {'load_format': 'dummy', 'gpu_memory_utilization': 0.8, 'disable_log_stats': True, 'hf_overrides': functools.partial(, model_arch='Gemma2ForCausalLM', exist_overrides={}, use_original_num_layers=False), 'model_impl': 'vllm', 'model': 'google/gemma-2-9b'} 2025-11-03T17:26:31.6961218Z 2025-11-03T17:26:31.6961877Z config.json: 0% 0.00/856 [00:00, 'use_cudagraph': True, 'cudagraph_num_of_warmups': 1, 'cudagraph_capture_sizes': [512, 504, 496, 488, 480, 472, 464, 456, 448, 440, 432, 424, 416, 408, 400, 392, 384, 376, 368, 360, 352, 344, 336, 328, 320, 312, 304, 296, 288, 280, 272, 264, 256, 248, 240, 232, 224, 216, 208, 200, 192, 184, 176, 168, 160, 152, 144, 136, 128, 120, 112, 104, 96, 88, 80, 72, 64, 56, 48, 40, 32, 24, 16, 8, 4, 2, 1], 'cudagraph_copy_inputs': False, 'full_cuda_graph': True, 'use_inductor_graph_partition': False, 'pass_config': {}, 'max_capture_size': 512, 'local_cache_dir': None} 2025-11-03T17:26:35.2137161Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:26:35.2164018Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:26:35.2171527Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:26:35.2179610Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:26:35.2188012Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:26:35.2195735Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:26:35.2198353Z (EngineCore_DP0 pid=9162) INFO 11-03 17:26:35 [parallel_state.py:1231] rank 0 in world size 1 is assigned as DP rank 0, PP rank 0, TP rank 0, EP rank 0 2025-11-03T17:26:35.5564486Z (EngineCore_DP0 pid=9162) INFO 11-03 17:26:35 [topk_topp_sampler.py:59] Using FlashInfer for top-p & top-k sampling. 2025-11-03T17:26:35.6350149Z (EngineCore_DP0 pid=9162) INFO 11-03 17:26:35 [gpu_model_runner.py:2840] Starting to load model google/gemma-2-9b... 2025-11-03T17:26:35.8850407Z (EngineCore_DP0 pid=9162) INFO 11-03 17:26:35 [gpu_model_runner.py:2870] Loading model from scratch... 2025-11-03T17:26:35.9562001Z (EngineCore_DP0 pid=9162) INFO 11-03 17:26:35 [cuda.py:404] Using Flash Attention backend on V1 engine. 2025-11-03T17:26:36.5209348Z (EngineCore_DP0 pid=9162) INFO 11-03 17:26:36 [gpu_model_runner.py:2902] Model loading took 2.0821 GiB and 0.130141 seconds 2025-11-03T17:26:36.5213898Z (EngineCore_DP0 pid=9162) INFO 11-03 17:26:36 [kv_cache_utils.py:1199] GPU KV cache size: 1,310,720 tokens 2025-11-03T17:26:36.5214959Z (EngineCore_DP0 pid=9162) INFO 11-03 17:26:36 [kv_cache_utils.py:1204] Maximum concurrency for 8,192 tokens per request: 159.69x 2025-11-03T17:26:37.7103372Z (EngineCore_DP0 pid=9162) INFO 11-03 17:26:37 [gc_utils.py:40] GC Debug Config. enabled:False,top_objects:-1 2025-11-03T17:26:37.7118219Z INFO 11-03 17:26:37 [llm.py:337] Supported tasks: ['generate'] 2025-11-03T17:26:38.1904447Z PASSED 2025-11-03T17:26:38.2045126Z models/test_initialization.py::test_can_initialize_large_subset[ArceeForCausalLM] Fork a new process to run a test 9228 2025-11-03T17:26:38.2055887Z Fork a new process to run a test 0 2025-11-03T17:26:38.2333319Z INFO 11-03 17:26:38 [utils.py:239] non-default args: {'load_format': 'dummy', 'gpu_memory_utilization': 0.8, 'disable_log_stats': True, 'hf_overrides': functools.partial(, model_arch='ArceeForCausalLM', exist_overrides={}, use_original_num_layers=False), 'model_impl': 'vllm', 'model': 'arcee-ai/AFM-4.5B-Base'} 2025-11-03T17:26:38.3192166Z 2025-11-03T17:26:38.3194111Z config.json: 0% 0.00/843 [00:00, 'use_cudagraph': True, 'cudagraph_num_of_warmups': 1, 'cudagraph_capture_sizes': [512, 504, 496, 488, 480, 472, 464, 456, 448, 440, 432, 424, 416, 408, 400, 392, 384, 376, 368, 360, 352, 344, 336, 328, 320, 312, 304, 296, 288, 280, 272, 264, 256, 248, 240, 232, 224, 216, 208, 200, 192, 184, 176, 168, 160, 152, 144, 136, 128, 120, 112, 104, 96, 88, 80, 72, 64, 56, 48, 40, 32, 24, 16, 8, 4, 2, 1], 'cudagraph_copy_inputs': False, 'full_cuda_graph': True, 'use_inductor_graph_partition': False, 'pass_config': {}, 'max_capture_size': 512, 'local_cache_dir': None} 2025-11-03T17:26:47.1693773Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:26:47.1722488Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:26:47.1729202Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:26:47.1737081Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:26:47.1744786Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:26:47.1752192Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:26:47.1754880Z (EngineCore_DP0 pid=9306) INFO 11-03 17:26:47 [parallel_state.py:1231] rank 0 in world size 1 is assigned as DP rank 0, PP rank 0, TP rank 0, EP rank 0 2025-11-03T17:26:47.5114994Z (EngineCore_DP0 pid=9306) INFO 11-03 17:26:47 [topk_topp_sampler.py:59] Using FlashInfer for top-p & top-k sampling. 2025-11-03T17:26:47.5993974Z (EngineCore_DP0 pid=9306) INFO 11-03 17:26:47 [gpu_model_runner.py:2840] Starting to load model arcee-ai/AFM-4.5B-Base... 2025-11-03T17:26:47.8588535Z (EngineCore_DP0 pid=9306) INFO 11-03 17:26:47 [gpu_model_runner.py:2870] Loading model from scratch... 2025-11-03T17:26:47.9630903Z (EngineCore_DP0 pid=9306) INFO 11-03 17:26:47 [cuda.py:404] Using Flash Attention backend on V1 engine. 2025-11-03T17:26:48.5339300Z (EngineCore_DP0 pid=9306) INFO 11-03 17:26:48 [gpu_model_runner.py:2902] Model loading took 1.4478 GiB and 0.162902 seconds 2025-11-03T17:26:48.5344699Z (EngineCore_DP0 pid=9306) INFO 11-03 17:26:48 [kv_cache_utils.py:1199] GPU KV cache size: 5,242,880 tokens 2025-11-03T17:26:48.5345661Z (EngineCore_DP0 pid=9306) INFO 11-03 17:26:48 [kv_cache_utils.py:1204] Maximum concurrency for 81,920 tokens per request: 64.00x 2025-11-03T17:26:49.9226746Z (EngineCore_DP0 pid=9306) INFO 11-03 17:26:49 [gc_utils.py:40] GC Debug Config. enabled:False,top_objects:-1 2025-11-03T17:26:49.9241398Z INFO 11-03 17:26:49 [llm.py:337] Supported tasks: ['generate'] 2025-11-03T17:26:50.4218894Z PASSED 2025-11-03T17:26:50.4359322Z models/test_initialization.py::test_can_initialize_large_subset[Ernie4_5ForCausalLM] Fork a new process to run a test 9372 2025-11-03T17:26:50.4369324Z Fork a new process to run a test 0 2025-11-03T17:26:50.4650297Z INFO 11-03 17:26:50 [utils.py:239] non-default args: {'load_format': 'dummy', 'gpu_memory_utilization': 0.8, 'disable_log_stats': True, 'hf_overrides': functools.partial(, model_arch='Ernie4_5ForCausalLM', exist_overrides={}, use_original_num_layers=False), 'model_impl': 'vllm', 'model': 'baidu/ERNIE-4.5-0.3B-PT'} 2025-11-03T17:26:50.5526002Z 2025-11-03T17:26:50.5527286Z config.json: 0% 0.00/633 [00:00, 'use_cudagraph': True, 'cudagraph_num_of_warmups': 1, 'cudagraph_capture_sizes': [512, 504, 496, 488, 480, 472, 464, 456, 448, 440, 432, 424, 416, 408, 400, 392, 384, 376, 368, 360, 352, 344, 336, 328, 320, 312, 304, 296, 288, 280, 272, 264, 256, 248, 240, 232, 224, 216, 208, 200, 192, 184, 176, 168, 160, 152, 144, 136, 128, 120, 112, 104, 96, 88, 80, 72, 64, 56, 48, 40, 32, 24, 16, 8, 4, 2, 1], 'cudagraph_copy_inputs': False, 'full_cuda_graph': True, 'use_inductor_graph_partition': False, 'pass_config': {}, 'max_capture_size': 512, 'local_cache_dir': None} 2025-11-03T17:26:59.8875258Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:26:59.8902548Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:26:59.8909481Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:26:59.8918104Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:26:59.8926157Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:26:59.8934497Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:26:59.8937551Z (EngineCore_DP0 pid=9451) INFO 11-03 17:26:59 [parallel_state.py:1231] rank 0 in world size 1 is assigned as DP rank 0, PP rank 0, TP rank 0, EP rank 0 2025-11-03T17:27:00.2321962Z (EngineCore_DP0 pid=9451) INFO 11-03 17:27:00 [topk_topp_sampler.py:59] Using FlashInfer for top-p & top-k sampling. 2025-11-03T17:27:00.2929447Z (EngineCore_DP0 pid=9451) INFO 11-03 17:27:00 [gpu_model_runner.py:2840] Starting to load model baidu/ERNIE-4.5-0.3B-PT... 2025-11-03T17:27:00.5527063Z (EngineCore_DP0 pid=9451) INFO 11-03 17:27:00 [gpu_model_runner.py:2870] Loading model from scratch... 2025-11-03T17:27:00.6201449Z (EngineCore_DP0 pid=9451) INFO 11-03 17:27:00 [cuda.py:404] Using Flash Attention backend on V1 engine. 2025-11-03T17:27:01.1811439Z (EngineCore_DP0 pid=9451) INFO 11-03 17:27:01 [gpu_model_runner.py:2902] Model loading took 0.2549 GiB and 0.127743 seconds 2025-11-03T17:27:01.1815924Z (EngineCore_DP0 pid=9451) INFO 11-03 17:27:01 [kv_cache_utils.py:1199] GPU KV cache size: 10,485,760 tokens 2025-11-03T17:27:01.1816877Z (EngineCore_DP0 pid=9451) INFO 11-03 17:27:01 [kv_cache_utils.py:1204] Maximum concurrency for 131,072 tokens per request: 80.00x 2025-11-03T17:27:03.8079310Z (EngineCore_DP0 pid=9451) INFO 11-03 17:27:03 [gc_utils.py:40] GC Debug Config. enabled:False,top_objects:-1 2025-11-03T17:27:03.8093579Z INFO 11-03 17:27:03 [llm.py:337] Supported tasks: ['generate'] 2025-11-03T17:27:04.2780106Z PASSED 2025-11-03T17:27:04.2922080Z models/test_initialization.py::test_can_initialize_large_subset[Lfm2MoeForCausalLM] Fork a new process to run a test 9517 2025-11-03T17:27:04.2932411Z Fork a new process to run a test 0 2025-11-03T17:27:04.2936250Z `transformers==4.56.2` installed, but `transformers>=4.58` is required to run this model. 2025-11-03T17:27:04.5909127Z PASSED 2025-11-03T17:27:04.6050119Z models/test_initialization.py::test_can_initialize_large_subset[Llama4ForCausalLM] Fork a new process to run a test 9518 2025-11-03T17:27:04.6060551Z Fork a new process to run a test 0 2025-11-03T17:27:04.6062301Z Model is not available online 2025-11-03T17:27:04.8942061Z PASSED 2025-11-03T17:27:04.9083921Z models/test_initialization.py::test_can_initialize_large_subset[ChameleonForConditionalGeneration] Fork a new process to run a test 9519 2025-11-03T17:27:04.9093573Z Fork a new process to run a test 0 2025-11-03T17:27:04.9374071Z INFO 11-03 17:27:04 [utils.py:239] non-default args: {'load_format': 'dummy', 'gpu_memory_utilization': 0.8, 'disable_log_stats': True, 'hf_overrides': functools.partial(, model_arch='ChameleonForConditionalGeneration', exist_overrides={}, use_original_num_layers=False), 'model_impl': 'vllm', 'model': 'facebook/chameleon-7b'} 2025-11-03T17:27:05.0625156Z 2025-11-03T17:27:05.1227698Z config.json: 0% 0.00/1.71M [00:00, 'use_cudagraph': True, 'cudagraph_num_of_warmups': 1, 'cudagraph_capture_sizes': [512, 504, 496, 488, 480, 472, 464, 456, 448, 440, 432, 424, 416, 408, 400, 392, 384, 376, 368, 360, 352, 344, 336, 328, 320, 312, 304, 296, 288, 280, 272, 264, 256, 248, 240, 232, 224, 216, 208, 200, 192, 184, 176, 168, 160, 152, 144, 136, 128, 120, 112, 104, 96, 88, 80, 72, 64, 56, 48, 40, 32, 24, 16, 8, 4, 2, 1], 'cudagraph_copy_inputs': False, 'full_cuda_graph': True, 'use_inductor_graph_partition': False, 'pass_config': {}, 'max_capture_size': 512, 'local_cache_dir': None} 2025-11-03T17:27:17.8652146Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:27:17.8680169Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:27:17.8687711Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:27:17.8695683Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:27:17.8703511Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:27:17.8711548Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:27:17.8714311Z (EngineCore_DP0 pid=9577) INFO 11-03 17:27:17 [parallel_state.py:1231] rank 0 in world size 1 is assigned as DP rank 0, PP rank 0, TP rank 0, EP rank 0 2025-11-03T17:27:18.2110626Z (EngineCore_DP0 pid=9577) INFO 11-03 17:27:18 [topk_topp_sampler.py:59] Using FlashInfer for top-p & top-k sampling. 2025-11-03T17:27:18.4693399Z (EngineCore_DP0 pid=9577) 2025-11-03T17:27:18.4694375Z processor_config.json: 0% 0.00/102 [00:00, model_arch='Mistral3ForConditionalGeneration', exist_overrides={}, use_original_num_layers=False), 'model_impl': 'vllm', 'model': 'mistralai/Mistral-Small-3.1-24B-Instruct-2503'} 2025-11-03T17:27:27.2236324Z 2025-11-03T17:27:27.2237802Z config.json: 0% 0.00/1.18k [00:00, 'use_cudagraph': True, 'cudagraph_num_of_warmups': 1, 'cudagraph_capture_sizes': [512, 504, 496, 488, 480, 472, 464, 456, 448, 440, 432, 424, 416, 408, 400, 392, 384, 376, 368, 360, 352, 344, 336, 328, 320, 312, 304, 296, 288, 280, 272, 264, 256, 248, 240, 232, 224, 216, 208, 200, 192, 184, 176, 168, 160, 152, 144, 136, 128, 120, 112, 104, 96, 88, 80, 72, 64, 56, 48, 40, 32, 24, 16, 8, 4, 2, 1], 'cudagraph_copy_inputs': False, 'full_cuda_graph': True, 'use_inductor_graph_partition': False, 'pass_config': {}, 'max_capture_size': 512, 'local_cache_dir': None} 2025-11-03T17:27:43.8139485Z (EngineCore_DP0 pid=9738) /opt/conda/envs/py_3.12/lib/python3.12/site-packages/torch/library.py:356: UserWarning: Warning only once for all operators, other operators may also be overridden. 2025-11-03T17:27:43.8140801Z (EngineCore_DP0 pid=9738) Overriding a previously registered kernel for the same operator and the same dispatch key 2025-11-03T17:27:43.8142636Z (EngineCore_DP0 pid=9738) operator: xformers_python::_fused_allgather_and_linear_impl(Tensor scattered_input, Tensor[] weights, str process_group_name, Tensor(a3!)[] gathered_outputs, SymInt timeout_s, bool _wait, bool _memcpy, Tensor scale_scattered_input, Tensor?[] scales_weights) -> () 2025-11-03T17:27:43.8144193Z (EngineCore_DP0 pid=9738) registered at /opt/conda/envs/py_3.12/lib/python3.12/site-packages/torch/_library/custom_ops.py:922 2025-11-03T17:27:43.8144826Z (EngineCore_DP0 pid=9738) dispatch key: ADInplaceOrView 2025-11-03T17:27:43.8145575Z (EngineCore_DP0 pid=9738) previous kernel: no debug info 2025-11-03T17:27:43.8146657Z (EngineCore_DP0 pid=9738) new kernel: registered at /opt/conda/envs/py_3.12/lib/python3.12/site-packages/torch/_library/custom_ops.py:922 (Triggered internally at /var/lib/jenkins/workspace/aten/src/ATen/core/dispatch/OperatorEntry.cpp:208.) 2025-11-03T17:27:43.8147548Z (EngineCore_DP0 pid=9738) self.m.impl( 2025-11-03T17:27:43.8311866Z (EngineCore_DP0 pid=9738) /opt/conda/envs/py_3.12/lib/python3.12/site-packages/vllm/transformers_utils/tokenizer.py:271: FutureWarning: It is strongly recommended to run mistral models with `--tokenizer-mode "mistral"` to ensure correct encoding and decoding. 2025-11-03T17:27:43.8313119Z (EngineCore_DP0 pid=9738) return cached_get_tokenizer( 2025-11-03T17:27:44.6562358Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:27:44.6590965Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:27:44.6597638Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:27:44.6605120Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:27:44.6612213Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:27:44.6619408Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:27:44.6622280Z (EngineCore_DP0 pid=9738) INFO 11-03 17:27:44 [parallel_state.py:1231] rank 0 in world size 1 is assigned as DP rank 0, PP rank 0, TP rank 0, EP rank 0 2025-11-03T17:27:44.9018890Z (EngineCore_DP0 pid=9738) INFO 11-03 17:27:44 [topk_topp_sampler.py:59] Using FlashInfer for top-p & top-k sampling. 2025-11-03T17:27:45.1044960Z (EngineCore_DP0 pid=9738) 2025-11-03T17:27:45.1045381Z processor_config.json: 0% 0.00/189 [00:00, model_arch='Qwen3MoeForCausalLM', exist_overrides={}, use_original_num_layers=False), 'model_impl': 'vllm', 'model': 'Qwen/Qwen3-30B-A3B'} 2025-11-03T17:27:58.5258324Z 2025-11-03T17:27:58.5259486Z config.json: 0% 0.00/963 [00:00, 'use_cudagraph': True, 'cudagraph_num_of_warmups': 1, 'cudagraph_capture_sizes': [512, 504, 496, 488, 480, 472, 464, 456, 448, 440, 432, 424, 416, 408, 400, 392, 384, 376, 368, 360, 352, 344, 336, 328, 320, 312, 304, 296, 288, 280, 272, 264, 256, 248, 240, 232, 224, 216, 208, 200, 192, 184, 176, 168, 160, 152, 144, 136, 128, 120, 112, 104, 96, 88, 80, 72, 64, 56, 48, 40, 32, 24, 16, 8, 4, 2, 1], 'cudagraph_copy_inputs': False, 'full_cuda_graph': True, 'use_inductor_graph_partition': False, 'pass_config': {}, 'max_capture_size': 512, 'local_cache_dir': None} 2025-11-03T17:28:06.9709685Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:28:06.9737121Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:28:06.9744863Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:28:06.9753106Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:28:06.9761173Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:28:06.9769480Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:28:06.9772597Z (EngineCore_DP0 pid=10104) INFO 11-03 17:28:06 [parallel_state.py:1231] rank 0 in world size 1 is assigned as DP rank 0, PP rank 0, TP rank 0, EP rank 0 2025-11-03T17:28:07.3064339Z (EngineCore_DP0 pid=10104) INFO 11-03 17:28:07 [topk_topp_sampler.py:59] Using FlashInfer for top-p & top-k sampling. 2025-11-03T17:28:07.3707430Z (EngineCore_DP0 pid=10104) INFO 11-03 17:28:07 [gpu_model_runner.py:2840] Starting to load model Qwen/Qwen3-30B-A3B... 2025-11-03T17:28:07.6267041Z (EngineCore_DP0 pid=10104) INFO 11-03 17:28:07 [gpu_model_runner.py:2870] Loading model from scratch... 2025-11-03T17:28:07.6972749Z (EngineCore_DP0 pid=10104) INFO 11-03 17:28:07 [cuda.py:404] Using Flash Attention backend on V1 engine. 2025-11-03T17:28:08.2552036Z (EngineCore_DP0 pid=10104) INFO 11-03 17:28:08 [gpu_model_runner.py:2902] Model loading took 1.2227 GiB and 0.129003 seconds 2025-11-03T17:28:08.2555879Z (EngineCore_DP0 pid=10104) INFO 11-03 17:28:08 [kv_cache_utils.py:1199] GPU KV cache size: 5,242,880 tokens 2025-11-03T17:28:08.2556625Z (EngineCore_DP0 pid=10104) INFO 11-03 17:28:08 [kv_cache_utils.py:1204] Maximum concurrency for 40,960 tokens per request: 128.00x 2025-11-03T17:28:09.3941891Z (EngineCore_DP0 pid=10104) INFO 11-03 17:28:09 [gc_utils.py:40] GC Debug Config. enabled:False,top_objects:-1 2025-11-03T17:28:09.3956354Z INFO 11-03 17:28:09 [llm.py:337] Supported tasks: ['generate'] 2025-11-03T17:28:09.8652655Z PASSED 2025-11-03T17:28:09.8794141Z models/test_initialization.py::test_can_initialize_large_subset[MiMoForCausalLM] Fork a new process to run a test 10170 2025-11-03T17:28:09.8804310Z Fork a new process to run a test 0 2025-11-03T17:28:09.9076727Z INFO 11-03 17:28:09 [utils.py:239] non-default args: {'trust_remote_code': True, 'load_format': 'dummy', 'gpu_memory_utilization': 0.8, 'disable_log_stats': True, 'hf_overrides': functools.partial(, model_arch='MiMoForCausalLM', exist_overrides={}, use_original_num_layers=False), 'model_impl': 'vllm', 'model': 'XiaomiMiMo/MiMo-7B-RL'} 2025-11-03T17:28:10.0210655Z 2025-11-03T17:28:10.0212722Z config.json: 0.00B [00:00, ?B/s] 2025-11-03T17:28:10.0213057Z config.json: 1.04kB [00:00, 5.02MB/s] 2025-11-03T17:28:10.0829021Z 2025-11-03T17:28:10.0829856Z configuration_mimo.py: 0% 0.00/376 [00:00, 'use_cudagraph': True, 'cudagraph_num_of_warmups': 1, 'cudagraph_capture_sizes': [512, 504, 496, 488, 480, 472, 464, 456, 448, 440, 432, 424, 416, 408, 400, 392, 384, 376, 368, 360, 352, 344, 336, 328, 320, 312, 304, 296, 288, 280, 272, 264, 256, 248, 240, 232, 224, 216, 208, 200, 192, 184, 176, 168, 160, 152, 144, 136, 128, 120, 112, 104, 96, 88, 80, 72, 64, 56, 48, 40, 32, 24, 16, 8, 4, 2, 1], 'cudagraph_copy_inputs': False, 'full_cuda_graph': True, 'use_inductor_graph_partition': False, 'pass_config': {}, 'max_capture_size': 512, 'local_cache_dir': None} 2025-11-03T17:28:18.2386162Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:28:18.2413657Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:28:18.2421825Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:28:18.2429611Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:28:18.2437159Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:28:18.2444876Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:28:18.2447908Z (EngineCore_DP0 pid=10228) INFO 11-03 17:28:18 [parallel_state.py:1231] rank 0 in world size 1 is assigned as DP rank 0, PP rank 0, TP rank 0, EP rank 0 2025-11-03T17:28:18.5861224Z (EngineCore_DP0 pid=10228) INFO 11-03 17:28:18 [topk_topp_sampler.py:59] Using FlashInfer for top-p & top-k sampling. 2025-11-03T17:28:18.6588863Z (EngineCore_DP0 pid=10228) INFO 11-03 17:28:18 [gpu_model_runner.py:2840] Starting to load model XiaomiMiMo/MiMo-7B-RL... 2025-11-03T17:28:18.9194154Z (EngineCore_DP0 pid=10228) INFO 11-03 17:28:18 [gpu_model_runner.py:2870] Loading model from scratch... 2025-11-03T17:28:18.9863927Z (EngineCore_DP0 pid=10228) INFO 11-03 17:28:18 [cuda.py:404] Using Flash Attention backend on V1 engine. 2025-11-03T17:28:19.5494329Z (EngineCore_DP0 pid=10228) INFO 11-03 17:28:19 [gpu_model_runner.py:2902] Model loading took 2.6543 GiB and 0.127732 seconds 2025-11-03T17:28:19.5498785Z (EngineCore_DP0 pid=10228) INFO 11-03 17:28:19 [kv_cache_utils.py:1199] GPU KV cache size: 2,621,440 tokens 2025-11-03T17:28:19.5499708Z (EngineCore_DP0 pid=10228) INFO 11-03 17:28:19 [kv_cache_utils.py:1204] Maximum concurrency for 32,768 tokens per request: 80.00x 2025-11-03T17:28:20.2456265Z (EngineCore_DP0 pid=10228) INFO 11-03 17:28:20 [gc_utils.py:40] GC Debug Config. enabled:False,top_objects:-1 2025-11-03T17:28:20.2477960Z INFO 11-03 17:28:20 [llm.py:337] Supported tasks: ['generate'] 2025-11-03T17:28:20.7155611Z PASSED 2025-11-03T17:28:20.7296281Z models/test_initialization.py::test_can_initialize_large_subset[GteModel] Fork a new process to run a test 10294 2025-11-03T17:28:20.7306071Z Fork a new process to run a test 0 2025-11-03T17:28:20.7588931Z INFO 11-03 17:28:20 [utils.py:239] non-default args: {'trust_remote_code': True, 'load_format': 'dummy', 'gpu_memory_utilization': 0.8, 'disable_log_stats': True, 'hf_overrides': functools.partial(, model_arch='GteModel', exist_overrides={}, use_original_num_layers=False), 'model_impl': 'vllm', 'model': 'Snowflake/snowflake-arctic-embed-m-v2.0'} 2025-11-03T17:28:20.8426590Z 2025-11-03T17:28:20.8428241Z config.json: 0% 0.00/971 [00:00, 'use_cudagraph': True, 'cudagraph_num_of_warmups': 1, 'cudagraph_capture_sizes': [512, 504, 496, 488, 480, 472, 464, 456, 448, 440, 432, 424, 416, 408, 400, 392, 384, 376, 368, 360, 352, 344, 336, 328, 320, 312, 304, 296, 288, 280, 272, 264, 256, 248, 240, 232, 224, 216, 208, 200, 192, 184, 176, 168, 160, 152, 144, 136, 128, 120, 112, 104, 96, 88, 80, 72, 64, 56, 48, 40, 32, 24, 16, 8, 4, 2, 1], 'cudagraph_copy_inputs': False, 'full_cuda_graph': False, 'use_inductor_graph_partition': False, 'pass_config': {}, 'max_capture_size': 512, 'local_cache_dir': None} 2025-11-03T17:28:29.9017862Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:28:29.9046485Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:28:29.9053272Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:28:29.9061009Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:28:29.9068687Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:28:29.9076528Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:28:29.9079182Z (EngineCore_DP0 pid=10375) INFO 11-03 17:28:29 [parallel_state.py:1231] rank 0 in world size 1 is assigned as DP rank 0, PP rank 0, TP rank 0, EP rank 0 2025-11-03T17:28:30.2437501Z (EngineCore_DP0 pid=10375) INFO 11-03 17:28:30 [topk_topp_sampler.py:59] Using FlashInfer for top-p & top-k sampling. 2025-11-03T17:28:30.2750322Z (EngineCore_DP0 pid=10375) INFO 11-03 17:28:30 [gpu_model_runner.py:2840] Starting to load model Snowflake/snowflake-arctic-embed-m-v2.0... 2025-11-03T17:28:30.5288372Z (EngineCore_DP0 pid=10375) INFO 11-03 17:28:30 [gpu_model_runner.py:2870] Loading model from scratch... 2025-11-03T17:28:30.5939287Z (EngineCore_DP0 pid=10375) INFO 11-03 17:28:30 [cuda.py:404] Using Flash Attention backend on V1 engine. 2025-11-03T17:28:31.1675959Z (EngineCore_DP0 pid=10375) INFO 11-03 17:28:31 [gpu_model_runner.py:2902] Model loading took 0.3769 GiB and 0.128631 seconds 2025-11-03T17:28:31.9465885Z (EngineCore_DP0 pid=10375) INFO 11-03 17:28:31 [core.py:142] Disabling chunked prefill for model without KVCache 2025-11-03T17:28:31.9948096Z (EngineCore_DP0 pid=10375) INFO 11-03 17:28:31 [vllm.py:433] Only "last" pooling supports chunked prefill and prefix caching; disabling both. 2025-11-03T17:28:32.2501986Z (EngineCore_DP0 pid=10375) INFO 11-03 17:28:32 [gc_utils.py:40] GC Debug Config. enabled:False,top_objects:-1 2025-11-03T17:28:32.2518027Z INFO 11-03 17:28:32 [llm.py:337] Supported tasks: ['embed', 'encode'] 2025-11-03T17:28:32.7546689Z PASSED 2025-11-03T17:28:32.7687558Z models/test_initialization.py::test_can_initialize_large_subset[PixtralForConditionalGeneration] Fork a new process to run a test 10441 2025-11-03T17:28:32.7698863Z Fork a new process to run a test 0 2025-11-03T17:28:32.7973494Z INFO 11-03 17:28:32 [utils.py:239] non-default args: {'tokenizer_mode': 'mistral', 'load_format': 'dummy', 'gpu_memory_utilization': 0.8, 'disable_log_stats': True, 'hf_overrides': functools.partial(, model_arch='PixtralForConditionalGeneration', exist_overrides={}, use_original_num_layers=False), 'model_impl': 'vllm', 'model': 'mistralai/Pixtral-12B-2409'} 2025-11-03T17:28:32.8774287Z 2025-11-03T17:28:32.8775382Z params.json: 0% 0.00/552 [00:00, 'use_cudagraph': True, 'cudagraph_num_of_warmups': 1, 'cudagraph_capture_sizes': [512, 504, 496, 488, 480, 472, 464, 456, 448, 440, 432, 424, 416, 408, 400, 392, 384, 376, 368, 360, 352, 344, 336, 328, 320, 312, 304, 296, 288, 280, 272, 264, 256, 248, 240, 232, 224, 216, 208, 200, 192, 184, 176, 168, 160, 152, 144, 136, 128, 120, 112, 104, 96, 88, 80, 72, 64, 56, 48, 40, 32, 24, 16, 8, 4, 2, 1], 'cudagraph_copy_inputs': False, 'full_cuda_graph': True, 'use_inductor_graph_partition': False, 'pass_config': {}, 'max_capture_size': 512, 'local_cache_dir': None} 2025-11-03T17:28:51.6047943Z (EngineCore_DP0 pid=10522) /opt/conda/envs/py_3.12/lib/python3.12/site-packages/torch/library.py:356: UserWarning: Warning only once for all operators, other operators may also be overridden. 2025-11-03T17:28:51.6049221Z (EngineCore_DP0 pid=10522) Overriding a previously registered kernel for the same operator and the same dispatch key 2025-11-03T17:28:51.6051375Z (EngineCore_DP0 pid=10522) operator: xformers_python::_fused_allgather_and_linear_impl(Tensor scattered_input, Tensor[] weights, str process_group_name, Tensor(a3!)[] gathered_outputs, SymInt timeout_s, bool _wait, bool _memcpy, Tensor scale_scattered_input, Tensor?[] scales_weights) -> () 2025-11-03T17:28:51.6052955Z (EngineCore_DP0 pid=10522) registered at /opt/conda/envs/py_3.12/lib/python3.12/site-packages/torch/_library/custom_ops.py:922 2025-11-03T17:28:51.6053707Z (EngineCore_DP0 pid=10522) dispatch key: ADInplaceOrView 2025-11-03T17:28:51.6054220Z (EngineCore_DP0 pid=10522) previous kernel: no debug info 2025-11-03T17:28:51.6055502Z (EngineCore_DP0 pid=10522) new kernel: registered at /opt/conda/envs/py_3.12/lib/python3.12/site-packages/torch/_library/custom_ops.py:922 (Triggered internally at /var/lib/jenkins/workspace/aten/src/ATen/core/dispatch/OperatorEntry.cpp:208.) 2025-11-03T17:28:51.6056610Z (EngineCore_DP0 pid=10522) self.m.impl( 2025-11-03T17:28:52.7159121Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:28:52.7187619Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:28:52.7194386Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:28:52.7201644Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:28:52.7208523Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:28:52.7215593Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:28:52.7218140Z (EngineCore_DP0 pid=10522) INFO 11-03 17:28:52 [parallel_state.py:1231] rank 0 in world size 1 is assigned as DP rank 0, PP rank 0, TP rank 0, EP rank 0 2025-11-03T17:28:52.9571439Z (EngineCore_DP0 pid=10522) INFO 11-03 17:28:52 [topk_topp_sampler.py:59] Using FlashInfer for top-p & top-k sampling. 2025-11-03T17:28:53.1674955Z (EngineCore_DP0 pid=10522) INFO 11-03 17:28:53 [gpu_model_runner.py:2840] Starting to load model mistralai/Pixtral-12B-2409... 2025-11-03T17:28:53.3975174Z (EngineCore_DP0 pid=10522) INFO 11-03 17:28:53 [gpu_model_runner.py:2870] Loading model from scratch... 2025-11-03T17:28:53.4547425Z (EngineCore_DP0 pid=10522) INFO 11-03 17:28:53 [cuda.py:404] Using Flash Attention backend on V1 engine. 2025-11-03T17:28:53.9729651Z (EngineCore_DP0 pid=10522) INFO 11-03 17:28:53 [gpu_model_runner.py:2902] Model loading took 3.1304 GiB and 0.111675 seconds 2025-11-03T17:28:54.2065574Z (EngineCore_DP0 pid=10522) INFO 11-03 17:28:54 [gpu_model_runner.py:3647] Encoder cache will be initialized with a budget of 8192 tokens, and profiled with 1 image items of the maximum feature size. 2025-11-03T17:28:54.2565266Z (EngineCore_DP0 pid=10522) WARNING 11-03 17:28:54 [processing.py:1091] PixtralProcessorAdapter did not return `BatchFeature`. Make sure to match the behaviour of `ProcessorMixin` when implementing custom processors. 2025-11-03T17:28:55.7541547Z (EngineCore_DP0 pid=10522) INFO 11-03 17:28:55 [backends.py:594] Using cache directory: /var/lib/jenkins/.cache/vllm/torch_compile_cache/c56678db3c/rank_0_0/backbone for vLLM's torch.compile 2025-11-03T17:28:55.7544581Z (EngineCore_DP0 pid=10522) INFO 11-03 17:28:55 [backends.py:608] Dynamo bytecode transform time: 1.25 s 2025-11-03T17:28:56.3779255Z (EngineCore_DP0 pid=10522) [rank0]:W1103 17:28:56.376000 10522 site-packages/torch/_inductor/utils.py:1665] [0/0] Not enough SMs to use max_autotune_gemm mode 2025-11-03T17:28:57.6969836Z (EngineCore_DP0 pid=10522) INFO 11-03 17:28:57 [backends.py:214] Cache the graph for dynamic shape for later use 2025-11-03T17:28:58.6432047Z (EngineCore_DP0 pid=10522) INFO 11-03 17:28:58 [backends.py:241] Compiling a graph for dynamic shape takes 2.70 s 2025-11-03T17:28:59.2042243Z (EngineCore_DP0 pid=10522) INFO 11-03 17:28:59 [monitor.py:33] torch.compile takes 3.95 s in total 2025-11-03T17:29:00.2358113Z (EngineCore_DP0 pid=10522) INFO 11-03 17:29:00 [gpu_worker.py:314] Available KV cache memory: 13.48 GiB 2025-11-03T17:29:00.5126774Z (EngineCore_DP0 pid=10522) INFO 11-03 17:29:00 [kv_cache_utils.py:1199] GPU KV cache size: 3,534,544 tokens 2025-11-03T17:29:00.5127696Z (EngineCore_DP0 pid=10522) INFO 11-03 17:29:00 [kv_cache_utils.py:1204] Maximum concurrency for 128,000 tokens per request: 27.61x 2025-11-03T17:29:00.7870436Z (EngineCore_DP0 pid=10522) 2025-11-03T17:29:00.8889412Z Capturing CUDA graphs (mixed prefill-decode, PIECEWISE): 0% 0/67 [00:00, model_arch='Exaone4ForCausalLM', exist_overrides={}, use_original_num_layers=False), 'model_impl': 'vllm', 'model': 'LGAI-EXAONE/EXAONE-4.0-32B'} 2025-11-03T17:29:04.2888477Z 2025-11-03T17:29:04.2890759Z config.json: 0.00B [00:00, ?B/s] 2025-11-03T17:29:04.2891076Z config.json: 2.42kB [00:00, 13.4MB/s] 2025-11-03T17:29:10.7392058Z INFO 11-03 17:29:10 [model.py:653] Resolved architecture: Exaone4ForCausalLM 2025-11-03T17:29:10.7392583Z `torch_dtype` is deprecated! Use `dtype` instead! 2025-11-03T17:29:10.7642851Z INFO 11-03 17:29:10 [model.py:1714] Using max model len 131072 2025-11-03T17:29:10.9355189Z INFO 11-03 17:29:10 [scheduler.py:225] Chunked prefill is enabled with max_num_batched_tokens=8192. 2025-11-03T17:29:10.9882972Z 2025-11-03T17:29:10.9886975Z tokenizer_config.json: 0.00B [00:00, ?B/s] 2025-11-03T17:29:10.9887354Z tokenizer_config.json: 70.3kB [00:00, 169MB/s] 2025-11-03T17:29:11.0857338Z 2025-11-03T17:29:11.1004653Z vocab.json: 0.00B [00:00, ?B/s] 2025-11-03T17:29:11.1004959Z vocab.json: 1.93MB [00:00, 132MB/s] 2025-11-03T17:29:11.1462946Z 2025-11-03T17:29:11.1521310Z merges.txt: 0.00B [00:00, ?B/s] 2025-11-03T17:29:11.1521921Z merges.txt: 1.22MB [00:00, 211MB/s] 2025-11-03T17:29:11.2220930Z 2025-11-03T17:29:11.2468048Z tokenizer.json: 0.00B [00:00, ?B/s] 2025-11-03T17:29:11.2468360Z tokenizer.json: 7.91MB [00:00, 319MB/s] 2025-11-03T17:29:11.3268102Z 2025-11-03T17:29:11.3271128Z special_tokens_map.json: 0.00B [00:00, ?B/s] 2025-11-03T17:29:11.3271492Z special_tokens_map.json: 6.70kB [00:00, 26.4MB/s] 2025-11-03T17:29:11.3690246Z 2025-11-03T17:29:11.3693564Z chat_template.jinja: 0.00B [00:00, ?B/s] 2025-11-03T17:29:11.6712050Z chat_template.jinja: 5.49kB [00:00, 23.9MB/s] 2025-11-03T17:29:11.6712318Z 2025-11-03T17:29:11.6712457Z generation_config.json: 0% 0.00/170 [00:00, 'use_cudagraph': True, 'cudagraph_num_of_warmups': 1, 'cudagraph_capture_sizes': [512, 504, 496, 488, 480, 472, 464, 456, 448, 440, 432, 424, 416, 408, 400, 392, 384, 376, 368, 360, 352, 344, 336, 328, 320, 312, 304, 296, 288, 280, 272, 264, 256, 248, 240, 232, 224, 216, 208, 200, 192, 184, 176, 168, 160, 152, 144, 136, 128, 120, 112, 104, 96, 88, 80, 72, 64, 56, 48, 40, 32, 24, 16, 8, 4, 2, 1], 'cudagraph_copy_inputs': False, 'full_cuda_graph': True, 'use_inductor_graph_partition': False, 'pass_config': {}, 'max_capture_size': 512, 'local_cache_dir': None} 2025-11-03T17:29:12.4387120Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:29:12.4415080Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:29:12.4423036Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:29:12.4431122Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:29:12.4439231Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:29:12.4446918Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:29:12.4450124Z (EngineCore_DP0 pid=10849) INFO 11-03 17:29:12 [parallel_state.py:1231] rank 0 in world size 1 is assigned as DP rank 0, PP rank 0, TP rank 0, EP rank 0 2025-11-03T17:29:12.7941217Z (EngineCore_DP0 pid=10849) INFO 11-03 17:29:12 [topk_topp_sampler.py:59] Using FlashInfer for top-p & top-k sampling. 2025-11-03T17:29:12.9069571Z (EngineCore_DP0 pid=10849) INFO 11-03 17:29:12 [gpu_model_runner.py:2840] Starting to load model LGAI-EXAONE/EXAONE-4.0-32B... 2025-11-03T17:29:13.1717289Z (EngineCore_DP0 pid=10849) INFO 11-03 17:29:13 [gpu_model_runner.py:2870] Loading model from scratch... 2025-11-03T17:29:13.2665607Z (EngineCore_DP0 pid=10849) INFO 11-03 17:29:13 [cuda.py:404] Using Flash Attention backend on V1 engine. 2025-11-03T17:29:13.8450644Z (EngineCore_DP0 pid=10849) INFO 11-03 17:29:13 [gpu_model_runner.py:2902] Model loading took 2.8868 GiB and 0.154019 seconds 2025-11-03T17:29:13.8455857Z (EngineCore_DP0 pid=10849) INFO 11-03 17:29:13 [kv_cache_utils.py:1199] GPU KV cache size: 2,621,440 tokens 2025-11-03T17:29:13.8456901Z (EngineCore_DP0 pid=10849) INFO 11-03 17:29:13 [kv_cache_utils.py:1204] Maximum concurrency for 131,072 tokens per request: 213.06x 2025-11-03T17:29:14.5149969Z (EngineCore_DP0 pid=10849) INFO 11-03 17:29:14 [gc_utils.py:40] GC Debug Config. enabled:False,top_objects:-1 2025-11-03T17:29:14.5172966Z INFO 11-03 17:29:14 [llm.py:337] Supported tasks: ['generate'] 2025-11-03T17:29:15.0132902Z PASSED 2025-11-03T17:29:15.0274764Z models/test_initialization.py::test_can_initialize_large_subset[GPT2LMHeadModel] Fork a new process to run a test 10915 2025-11-03T17:29:15.0285399Z Fork a new process to run a test 0 2025-11-03T17:29:15.0565489Z INFO 11-03 17:29:15 [utils.py:239] non-default args: {'load_format': 'dummy', 'gpu_memory_utilization': 0.8, 'disable_log_stats': True, 'hf_overrides': functools.partial(, model_arch='GPT2LMHeadModel', exist_overrides={}, use_original_num_layers=False), 'model_impl': 'vllm', 'model': 'openai-community/gpt2'} 2025-11-03T17:29:15.1432842Z 2025-11-03T17:29:15.1434317Z config.json: 0% 0.00/665 [00:00, 'use_cudagraph': True, 'cudagraph_num_of_warmups': 1, 'cudagraph_capture_sizes': [512, 504, 496, 488, 480, 472, 464, 456, 448, 440, 432, 424, 416, 408, 400, 392, 384, 376, 368, 360, 352, 344, 336, 328, 320, 312, 304, 296, 288, 280, 272, 264, 256, 248, 240, 232, 224, 216, 208, 200, 192, 184, 176, 168, 160, 152, 144, 136, 128, 120, 112, 104, 96, 88, 80, 72, 64, 56, 48, 40, 32, 24, 16, 8, 4, 2, 1], 'cudagraph_copy_inputs': False, 'full_cuda_graph': True, 'use_inductor_graph_partition': False, 'pass_config': {}, 'max_capture_size': 512, 'local_cache_dir': None} 2025-11-03T17:29:23.0839031Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:29:23.0867034Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:29:23.0874525Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:29:23.0882694Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:29:23.0890519Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:29:23.0898369Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:29:23.0901244Z (EngineCore_DP0 pid=10974) INFO 11-03 17:29:23 [parallel_state.py:1231] rank 0 in world size 1 is assigned as DP rank 0, PP rank 0, TP rank 0, EP rank 0 2025-11-03T17:29:23.4240138Z (EngineCore_DP0 pid=10974) INFO 11-03 17:29:23 [topk_topp_sampler.py:59] Using FlashInfer for top-p & top-k sampling. 2025-11-03T17:29:23.4661599Z (EngineCore_DP0 pid=10974) INFO 11-03 17:29:23 [gpu_model_runner.py:2840] Starting to load model openai-community/gpt2... 2025-11-03T17:29:23.7193978Z (EngineCore_DP0 pid=10974) INFO 11-03 17:29:23 [gpu_model_runner.py:2870] Loading model from scratch... 2025-11-03T17:29:23.7233272Z (EngineCore_DP0 pid=10974) INFO 11-03 17:29:23 [cuda.py:404] Using Flash Attention backend on V1 engine. 2025-11-03T17:29:24.2905003Z (EngineCore_DP0 pid=10974) INFO 11-03 17:29:24 [gpu_model_runner.py:2902] Model loading took 0.0869 GiB and 0.062916 seconds 2025-11-03T17:29:24.2909969Z (EngineCore_DP0 pid=10974) INFO 11-03 17:29:24 [kv_cache_utils.py:1199] GPU KV cache size: 3,495,248 tokens 2025-11-03T17:29:24.2910913Z (EngineCore_DP0 pid=10974) INFO 11-03 17:29:24 [kv_cache_utils.py:1204] Maximum concurrency for 1,024 tokens per request: 3413.33x 2025-11-03T17:29:25.1327763Z (EngineCore_DP0 pid=10974) INFO 11-03 17:29:25 [gc_utils.py:40] GC Debug Config. enabled:False,top_objects:-1 2025-11-03T17:29:25.1342164Z INFO 11-03 17:29:25 [llm.py:337] Supported tasks: ['generate'] 2025-11-03T17:29:25.5601208Z PASSED 2025-11-03T17:29:25.5743210Z models/test_initialization.py::test_can_initialize_large_subset[MantisForConditionalGeneration] Fork a new process to run a test 11040 2025-11-03T17:29:25.5754221Z Fork a new process to run a test 0 2025-11-03T17:29:25.5757936Z `transformers==4.56.2` installed, but `transformers<=4.48` is required to run this model. Reason: HF model is not compatible. 2025-11-03T17:29:25.8658746Z PASSED 2025-11-03T17:29:25.8802369Z models/test_initialization.py::test_can_initialize_large_subset[BailingMoeV2ForCausalLM] Fork a new process to run a test 11041 2025-11-03T17:29:25.8813085Z Fork a new process to run a test 0 2025-11-03T17:29:25.9095813Z INFO 11-03 17:29:25 [utils.py:239] non-default args: {'trust_remote_code': True, 'load_format': 'dummy', 'gpu_memory_utilization': 0.8, 'disable_log_stats': True, 'hf_overrides': functools.partial(, model_arch='BailingMoeV2ForCausalLM', exist_overrides={}, use_original_num_layers=False), 'model_impl': 'vllm', 'model': 'inclusionAI/Ling-mini-2.0'} 2025-11-03T17:29:26.0204343Z 2025-11-03T17:29:26.0206355Z config.json: 0.00B [00:00, ?B/s] 2025-11-03T17:29:26.0206668Z config.json: 1.52kB [00:00, 7.61MB/s] 2025-11-03T17:29:26.0943701Z 2025-11-03T17:29:26.0945669Z configuration_bailing_moe_v2.py: 0.00B [00:00, ?B/s] 2025-11-03T17:29:26.0946110Z configuration_bailing_moe_v2.py: 3.16kB [00:00, 11.9MB/s] 2025-11-03T17:29:26.1039216Z A new version of the following files was downloaded from https://huggingface.co/inclusionAI/Ling-mini-2.0: 2025-11-03T17:29:26.1039767Z - configuration_bailing_moe_v2.py 2025-11-03T17:29:26.1040425Z . Make sure to double-check they do not contain any added malicious code. To avoid downloading new versions of the code file, you can pin a revision. 2025-11-03T17:29:26.1378577Z You are using a model of type bailing_moe to instantiate a model of type . This is not supported for all configurations of models and can yield errors. 2025-11-03T17:29:32.6379453Z INFO 11-03 17:29:32 [model.py:653] Resolved architecture: BailingMoeV2ForCausalLM 2025-11-03T17:29:32.6380347Z `torch_dtype` is deprecated! Use `dtype` instead! 2025-11-03T17:29:32.6636922Z INFO 11-03 17:29:32 [model.py:1714] Using max model len 32768 2025-11-03T17:29:32.6639201Z The argument `trust_remote_code` is to be used with Auto classes. It has no effect here and is ignored. 2025-11-03T17:29:32.7432369Z INFO 11-03 17:29:32 [scheduler.py:225] Chunked prefill is enabled with max_num_batched_tokens=8192. 2025-11-03T17:29:32.8105694Z 2025-11-03T17:29:32.8108087Z tokenizer_config.json: 0.00B [00:00, ?B/s] 2025-11-03T17:29:32.8108634Z tokenizer_config.json: 4.59kB [00:00, 16.2MB/s] 2025-11-03T17:29:32.8929619Z 2025-11-03T17:29:32.9246048Z tokenizer.json: 0.00B [00:00, ?B/s] 2025-11-03T17:29:32.9246502Z tokenizer.json: 7.66MB [00:00, 243MB/s] 2025-11-03T17:29:33.0197154Z 2025-11-03T17:29:33.0197487Z special_tokens_map.json: 0% 0.00/152 [00:00, 'use_cudagraph': True, 'cudagraph_num_of_warmups': 1, 'cudagraph_capture_sizes': [512, 504, 496, 488, 480, 472, 464, 456, 448, 440, 432, 424, 416, 408, 400, 392, 384, 376, 368, 360, 352, 344, 336, 328, 320, 312, 304, 296, 288, 280, 272, 264, 256, 248, 240, 232, 224, 216, 208, 200, 192, 184, 176, 168, 160, 152, 144, 136, 128, 120, 112, 104, 96, 88, 80, 72, 64, 56, 48, 40, 32, 24, 16, 8, 4, 2, 1], 'cudagraph_copy_inputs': False, 'full_cuda_graph': True, 'use_inductor_graph_partition': False, 'pass_config': {}, 'max_capture_size': 512, 'local_cache_dir': None} 2025-11-03T17:29:34.3394427Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:29:34.3422460Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:29:34.3430328Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:29:34.3438143Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:29:34.3446107Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:29:34.3454159Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:29:34.3456939Z (EngineCore_DP0 pid=11099) INFO 11-03 17:29:34 [parallel_state.py:1231] rank 0 in world size 1 is assigned as DP rank 0, PP rank 0, TP rank 0, EP rank 0 2025-11-03T17:29:34.6881886Z (EngineCore_DP0 pid=11099) INFO 11-03 17:29:34 [topk_topp_sampler.py:59] Using FlashInfer for top-p & top-k sampling. 2025-11-03T17:29:34.7501210Z (EngineCore_DP0 pid=11099) INFO 11-03 17:29:34 [gpu_model_runner.py:2840] Starting to load model inclusionAI/Ling-mini-2.0... 2025-11-03T17:29:35.0102500Z (EngineCore_DP0 pid=11099) INFO 11-03 17:29:35 [gpu_model_runner.py:2870] Loading model from scratch... 2025-11-03T17:29:35.0769339Z (EngineCore_DP0 pid=11099) INFO 11-03 17:29:35 [cuda.py:404] Using Flash Attention backend on V1 engine. 2025-11-03T17:29:35.6603581Z (EngineCore_DP0 pid=11099) INFO 11-03 17:29:35 [gpu_model_runner.py:2902] Model loading took 1.3224 GiB and 0.128014 seconds 2025-11-03T17:29:35.6608295Z (EngineCore_DP0 pid=11099) INFO 11-03 17:29:35 [kv_cache_utils.py:1199] GPU KV cache size: 5,242,880 tokens 2025-11-03T17:29:35.6609199Z (EngineCore_DP0 pid=11099) INFO 11-03 17:29:35 [kv_cache_utils.py:1204] Maximum concurrency for 32,768 tokens per request: 160.00x 2025-11-03T17:29:36.8746588Z (EngineCore_DP0 pid=11099) INFO 11-03 17:29:36 [gc_utils.py:40] GC Debug Config. enabled:False,top_objects:-1 2025-11-03T17:29:36.8761282Z INFO 11-03 17:29:36 [llm.py:337] Supported tasks: ['generate'] 2025-11-03T17:29:37.3358639Z PASSED 2025-11-03T17:29:37.3502592Z models/test_initialization.py::test_can_initialize_large_subset[SkyworkR1VChatModel] Fork a new process to run a test 11165 2025-11-03T17:29:37.3512865Z Fork a new process to run a test 0 2025-11-03T17:29:37.3792331Z INFO 11-03 17:29:37 [utils.py:239] non-default args: {'trust_remote_code': True, 'load_format': 'dummy', 'gpu_memory_utilization': 0.8, 'disable_log_stats': True, 'hf_overrides': functools.partial(, model_arch='SkyworkR1VChatModel', exist_overrides={}, use_original_num_layers=False), 'model_impl': 'vllm', 'model': 'Skywork/Skywork-R1V-38B'} 2025-11-03T17:29:37.5058392Z 2025-11-03T17:29:37.5061244Z config.json: 0.00B [00:00, ?B/s] 2025-11-03T17:29:37.5061675Z config.json: 3.71kB [00:00, 14.0MB/s] 2025-11-03T17:29:37.5960715Z 2025-11-03T17:29:37.5961966Z configuration_skywork_chat.py: 0.00B [00:00, ?B/s] 2025-11-03T17:29:37.5962381Z configuration_skywork_chat.py: 3.77kB [00:00, 32.5MB/s] 2025-11-03T17:29:37.6397938Z 2025-11-03T17:29:37.6399846Z configuration_skywork_lm2.py: 0.00B [00:00, ?B/s] 2025-11-03T17:29:37.6400241Z configuration_skywork_lm2.py: 6.40kB [00:00, 45.7MB/s] 2025-11-03T17:29:37.6422423Z A new version of the following files was downloaded from https://huggingface.co/Skywork/Skywork-R1V-38B: 2025-11-03T17:29:37.6422968Z - configuration_skywork_lm2.py 2025-11-03T17:29:37.6423598Z . Make sure to double-check they do not contain any added malicious code. To avoid downloading new versions of the code file, you can pin a revision. 2025-11-03T17:29:37.6897650Z 2025-11-03T17:29:37.6899349Z configuration_skywork_vit.py: 0.00B [00:00, ?B/s] 2025-11-03T17:29:37.6899736Z configuration_skywork_vit.py: 4.52kB [00:00, 18.3MB/s] 2025-11-03T17:29:37.6918860Z A new version of the following files was downloaded from https://huggingface.co/Skywork/Skywork-R1V-38B: 2025-11-03T17:29:37.6919438Z - configuration_skywork_vit.py 2025-11-03T17:29:37.6920108Z . Make sure to double-check they do not contain any added malicious code. To avoid downloading new versions of the code file, you can pin a revision. 2025-11-03T17:29:37.6920990Z A new version of the following files was downloaded from https://huggingface.co/Skywork/Skywork-R1V-38B: 2025-11-03T17:29:37.6922008Z - configuration_skywork_chat.py 2025-11-03T17:29:37.6922294Z - configuration_skywork_lm2.py 2025-11-03T17:29:37.6922549Z - configuration_skywork_vit.py 2025-11-03T17:29:37.6923132Z . Make sure to double-check they do not contain any added malicious code. To avoid downloading new versions of the code file, you can pin a revision. 2025-11-03T17:29:38.0726465Z 2025-11-03T17:29:38.0728497Z preprocessor_config.json: 0% 0.00/287 [00:00, 'use_cudagraph': True, 'cudagraph_num_of_warmups': 1, 'cudagraph_capture_sizes': [512, 504, 496, 488, 480, 472, 464, 456, 448, 440, 432, 424, 416, 408, 400, 392, 384, 376, 368, 360, 352, 344, 336, 328, 320, 312, 304, 296, 288, 280, 272, 264, 256, 248, 240, 232, 224, 216, 208, 200, 192, 184, 176, 168, 160, 152, 144, 136, 128, 120, 112, 104, 96, 88, 80, 72, 64, 56, 48, 40, 32, 24, 16, 8, 4, 2, 1], 'cudagraph_copy_inputs': False, 'full_cuda_graph': True, 'use_inductor_graph_partition': False, 'pass_config': {}, 'max_capture_size': 512, 'local_cache_dir': None} 2025-11-03T17:29:46.4014194Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:29:46.4041763Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:29:46.4048993Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:29:46.4057097Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:29:46.4065184Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:29:46.4073036Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:29:46.4075925Z (EngineCore_DP0 pid=11223) INFO 11-03 17:29:46 [parallel_state.py:1231] rank 0 in world size 1 is assigned as DP rank 0, PP rank 0, TP rank 0, EP rank 0 2025-11-03T17:29:46.7549927Z (EngineCore_DP0 pid=11223) INFO 11-03 17:29:46 [topk_topp_sampler.py:59] Using FlashInfer for top-p & top-k sampling. 2025-11-03T17:29:46.8790487Z (EngineCore_DP0 pid=11223) WARNING 11-03 17:29:46 [__init__.py:2040] The following intended overrides are not keyword args and will be dropped: {'truncation'} 2025-11-03T17:29:46.8810161Z (EngineCore_DP0 pid=11223) WARNING 11-03 17:29:46 [__init__.py:2040] The following intended overrides are not keyword args and will be dropped: {'truncation'} 2025-11-03T17:29:46.9070697Z (EngineCore_DP0 pid=11223) INFO 11-03 17:29:46 [gpu_model_runner.py:2840] Starting to load model Skywork/Skywork-R1V-38B... 2025-11-03T17:29:47.1653888Z (EngineCore_DP0 pid=11223) INFO 11-03 17:29:47 [gpu_model_runner.py:2870] Loading model from scratch... 2025-11-03T17:29:47.1701228Z (EngineCore_DP0 pid=11223) INFO 11-03 17:29:47 [layer.py:497] MultiHeadAttention attn_backend: _Backend.FLASH_ATTN, use_upstream_fa: False 2025-11-03T17:29:47.2284589Z (EngineCore_DP0 pid=11223) INFO 11-03 17:29:47 [cuda.py:404] Using Flash Attention backend on V1 engine. 2025-11-03T17:29:47.7969348Z (EngineCore_DP0 pid=11223) INFO 11-03 17:29:47 [gpu_model_runner.py:2902] Model loading took 4.2528 GiB and 0.123869 seconds 2025-11-03T17:29:47.7973926Z (EngineCore_DP0 pid=11223) INFO 11-03 17:29:47 [kv_cache_utils.py:1199] GPU KV cache size: 2,621,440 tokens 2025-11-03T17:29:47.7974952Z (EngineCore_DP0 pid=11223) INFO 11-03 17:29:47 [kv_cache_utils.py:1204] Maximum concurrency for 131,072 tokens per request: 20.00x 2025-11-03T17:29:48.2459929Z (EngineCore_DP0 pid=11223) WARNING 11-03 17:29:48 [__init__.py:2040] The following intended overrides are not keyword args and will be dropped: {'truncation'} 2025-11-03T17:29:48.6188971Z (EngineCore_DP0 pid=11223) INFO 11-03 17:29:48 [gc_utils.py:40] GC Debug Config. enabled:False,top_objects:-1 2025-11-03T17:29:48.6205276Z INFO 11-03 17:29:48 [llm.py:337] Supported tasks: ['generate'] 2025-11-03T17:29:49.1228584Z PASSED 2025-11-03T17:29:49.1370289Z models/test_initialization.py::test_can_initialize_large_subset[GLM4VForCausalLM] Fork a new process to run a test 11305 2025-11-03T17:29:49.1381098Z Fork a new process to run a test 0 2025-11-03T17:29:49.1660460Z INFO 11-03 17:29:49 [utils.py:239] non-default args: {'trust_remote_code': True, 'load_format': 'dummy', 'gpu_memory_utilization': 0.8, 'disable_log_stats': True, 'hf_overrides': functools.partial(, model_arch='GLM4VForCausalLM', exist_overrides={'architectures': ['GLM4VForCausalLM']}, use_original_num_layers=False), 'model_impl': 'vllm', 'model': 'zai-org/glm-4v-9b'} 2025-11-03T17:29:49.3446952Z 2025-11-03T17:29:49.3449259Z config.json: 0.00B [00:00, ?B/s] 2025-11-03T17:29:49.3449559Z config.json: 1.77kB [00:00, 9.24MB/s] 2025-11-03T17:29:55.9626651Z INFO 11-03 17:29:55 [model.py:653] Resolved architecture: GLM4VForCausalLM 2025-11-03T17:29:55.9627137Z `torch_dtype` is deprecated! Use `dtype` instead! 2025-11-03T17:29:55.9875948Z INFO 11-03 17:29:55 [model.py:1714] Using max model len 8192 2025-11-03T17:29:55.9878446Z The argument `trust_remote_code` is to be used with Auto classes. It has no effect here and is ignored. 2025-11-03T17:29:56.1565314Z INFO 11-03 17:29:56 [scheduler.py:225] Chunked prefill is enabled with max_num_batched_tokens=8192. 2025-11-03T17:29:56.2165522Z 2025-11-03T17:29:56.2166941Z tokenizer_config.json: 0.00B [00:00, ?B/s] 2025-11-03T17:29:56.2167302Z tokenizer_config.json: 3.22kB [00:00, 25.2MB/s] 2025-11-03T17:29:56.2487650Z 2025-11-03T17:29:56.2492888Z tokenization_chatglm.py: 0.00B [00:00, ?B/s] 2025-11-03T17:29:56.2493766Z tokenization_chatglm.py: 17.7kB [00:00, 40.2MB/s] 2025-11-03T17:29:56.2635584Z A new version of the following files was downloaded from https://huggingface.co/zai-org/glm-4v-9b: 2025-11-03T17:29:56.2636103Z - tokenization_chatglm.py 2025-11-03T17:29:56.2636740Z . Make sure to double-check they do not contain any added malicious code. To avoid downloading new versions of the code file, you can pin a revision. 2025-11-03T17:29:56.3708598Z 2025-11-03T17:29:56.6196778Z tokenizer.model: 0% 0.00/2.62M [00:00, 'use_cudagraph': True, 'cudagraph_num_of_warmups': 1, 'cudagraph_capture_sizes': [512, 504, 496, 488, 480, 472, 464, 456, 448, 440, 432, 424, 416, 408, 400, 392, 384, 376, 368, 360, 352, 344, 336, 328, 320, 312, 304, 296, 288, 280, 272, 264, 256, 248, 240, 232, 224, 216, 208, 200, 192, 184, 176, 168, 160, 152, 144, 136, 128, 120, 112, 104, 96, 88, 80, 72, 64, 56, 48, 40, 32, 24, 16, 8, 4, 2, 1], 'cudagraph_copy_inputs': False, 'full_cuda_graph': True, 'use_inductor_graph_partition': False, 'pass_config': {}, 'max_capture_size': 512, 'local_cache_dir': None} 2025-11-03T17:29:58.3417332Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:29:58.3446573Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:29:58.3453842Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:29:58.3461767Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:29:58.3469506Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:29:58.3477601Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:29:58.3480350Z (EngineCore_DP0 pid=11383) INFO 11-03 17:29:58 [parallel_state.py:1231] rank 0 in world size 1 is assigned as DP rank 0, PP rank 0, TP rank 0, EP rank 0 2025-11-03T17:29:58.6959307Z (EngineCore_DP0 pid=11383) INFO 11-03 17:29:58 [topk_topp_sampler.py:59] Using FlashInfer for top-p & top-k sampling. 2025-11-03T17:29:58.7739967Z (EngineCore_DP0 pid=11383) WARNING 11-03 17:29:58 [__init__.py:2040] The following intended overrides are not keyword args and will be dropped: {'truncation'} 2025-11-03T17:29:58.7746629Z (EngineCore_DP0 pid=11383) WARNING 11-03 17:29:58 [__init__.py:2040] The following intended overrides are not keyword args and will be dropped: {'truncation'} 2025-11-03T17:29:58.7865804Z (EngineCore_DP0 pid=11383) INFO 11-03 17:29:58 [gpu_model_runner.py:2840] Starting to load model zai-org/glm-4v-9b... 2025-11-03T17:29:59.0514187Z (EngineCore_DP0 pid=11383) INFO 11-03 17:29:59 [gpu_model_runner.py:2870] Loading model from scratch... 2025-11-03T17:29:59.1091999Z (EngineCore_DP0 pid=11383) INFO 11-03 17:29:59 [cuda.py:404] Using Flash Attention backend on V1 engine. 2025-11-03T17:29:59.1796853Z (EngineCore_DP0 pid=11383) INFO 11-03 17:29:59 [layer.py:497] MultiHeadAttention attn_backend: _Backend.XFORMERS, use_upstream_fa: False 2025-11-03T17:29:59.7029404Z (EngineCore_DP0 pid=11383) INFO 11-03 17:29:59 [gpu_model_runner.py:2902] Model loading took 3.2448 GiB and 0.132990 seconds 2025-11-03T17:29:59.7034286Z (EngineCore_DP0 pid=11383) INFO 11-03 17:29:59 [kv_cache_utils.py:1199] GPU KV cache size: 10,485,760 tokens 2025-11-03T17:29:59.7035177Z (EngineCore_DP0 pid=11383) INFO 11-03 17:29:59 [kv_cache_utils.py:1204] Maximum concurrency for 8,192 tokens per request: 1280.00x 2025-11-03T17:30:00.0975291Z (EngineCore_DP0 pid=11383) WARNING 11-03 17:30:00 [tokenizer.py:255] Using a slow tokenizer. This might cause a significant slowdown. Consider using a fast tokenizer instead. 2025-11-03T17:30:00.1263300Z (EngineCore_DP0 pid=11383) WARNING 11-03 17:30:00 [__init__.py:2040] The following intended overrides are not keyword args and will be dropped: {'truncation'} 2025-11-03T17:30:01.3265803Z (EngineCore_DP0 pid=11383) INFO 11-03 17:30:01 [gc_utils.py:40] GC Debug Config. enabled:False,top_objects:-1 2025-11-03T17:30:01.3280335Z INFO 11-03 17:30:01 [llm.py:337] Supported tasks: ['generate'] 2025-11-03T17:30:01.8042853Z PASSED 2025-11-03T17:30:01.8184249Z models/test_initialization.py::test_can_initialize_large_subset[ChatGLMModel] Fork a new process to run a test 11449 2025-11-03T17:30:01.8195261Z Fork a new process to run a test 0 2025-11-03T17:30:01.8198983Z `transformers==4.56.2` installed, but `transformers<=4.48` is required to run this model. 2025-11-03T17:30:02.1121747Z PASSED 2025-11-03T17:30:02.1262635Z models/test_initialization.py::test_can_initialize_large_subset[Qwen2_5OmniModel] Fork a new process to run a test 11450 2025-11-03T17:30:02.1274004Z Fork a new process to run a test 0 2025-11-03T17:30:02.1551792Z INFO 11-03 17:30:02 [utils.py:239] non-default args: {'load_format': 'dummy', 'gpu_memory_utilization': 0.8, 'disable_log_stats': True, 'hf_overrides': functools.partial(, model_arch='Qwen2_5OmniModel', exist_overrides={}, use_original_num_layers=False), 'model_impl': 'vllm', 'model': 'Qwen/Qwen2.5-Omni-3B'} 2025-11-03T17:30:02.2263672Z 2025-11-03T17:30:02.2266281Z config.json: 0.00B [00:00, ?B/s] 2025-11-03T17:30:02.2266634Z config.json: 13.2kB [00:00, 51.2MB/s] 2025-11-03T17:30:02.2577674Z Unrecognized keys in `rope_scaling` for 'rope_type'='default': {'mrope_section'} 2025-11-03T17:30:02.3243338Z 2025-11-03T17:30:02.3245037Z preprocessor_config.json: 0% 0.00/667 [00:00, 'use_cudagraph': True, 'cudagraph_num_of_warmups': 1, 'cudagraph_capture_sizes': [512, 504, 496, 488, 480, 472, 464, 456, 448, 440, 432, 424, 416, 408, 400, 392, 384, 376, 368, 360, 352, 344, 336, 328, 320, 312, 304, 296, 288, 280, 272, 264, 256, 248, 240, 232, 224, 216, 208, 200, 192, 184, 176, 168, 160, 152, 144, 136, 128, 120, 112, 104, 96, 88, 80, 72, 64, 56, 48, 40, 32, 24, 16, 8, 4, 2, 1], 'cudagraph_copy_inputs': False, 'full_cuda_graph': True, 'use_inductor_graph_partition': False, 'pass_config': {}, 'max_capture_size': 512, 'local_cache_dir': None} 2025-11-03T17:30:11.1577837Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:30:11.1616182Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:30:11.1624140Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:30:11.1632239Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:30:11.1639578Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:30:11.1647379Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:30:11.1650280Z (EngineCore_DP0 pid=11528) INFO 11-03 17:30:11 [parallel_state.py:1231] rank 0 in world size 1 is assigned as DP rank 0, PP rank 0, TP rank 0, EP rank 0 2025-11-03T17:30:11.5085895Z (EngineCore_DP0 pid=11528) INFO 11-03 17:30:11 [topk_topp_sampler.py:59] Using FlashInfer for top-p & top-k sampling. 2025-11-03T17:30:13.3451825Z (EngineCore_DP0 pid=11528) 2025-11-03T17:30:13.3452683Z chat_template.json: 0.00B [00:00, ?B/s] 2025-11-03T17:30:13.3452988Z chat_template.json: 1.31kB [00:00, 6.66MB/s] 2025-11-03T17:30:13.9181410Z (EngineCore_DP0 pid=11528) INFO 11-03 17:30:13 [gpu_model_runner.py:2840] Starting to load model Qwen/Qwen2.5-Omni-3B... 2025-11-03T17:30:14.2562713Z (EngineCore_DP0 pid=11528) INFO 11-03 17:30:14 [gpu_model_runner.py:2870] Loading model from scratch... 2025-11-03T17:30:14.2564035Z (EngineCore_DP0 pid=11528) WARNING 11-03 17:30:14 [qwen2_5_omni_thinker.py:865] flash_attn is not available, the model may not yield the exactly same result as the transformers implementation in the audio tower part. 2025-11-03T17:30:14.3782042Z (EngineCore_DP0 pid=11528) INFO 11-03 17:30:14 [cuda.py:404] Using Flash Attention backend on V1 engine. 2025-11-03T17:30:14.9715805Z (EngineCore_DP0 pid=11528) INFO 11-03 17:30:14 [gpu_model_runner.py:2902] Model loading took 3.7809 GiB and 0.194744 seconds 2025-11-03T17:30:14.9720030Z (EngineCore_DP0 pid=11528) INFO 11-03 17:30:14 [kv_cache_utils.py:1199] GPU KV cache size: 10,485,760 tokens 2025-11-03T17:30:14.9720961Z (EngineCore_DP0 pid=11528) INFO 11-03 17:30:14 [kv_cache_utils.py:1204] Maximum concurrency for 32,768 tokens per request: 320.00x 2025-11-03T17:30:16.5885938Z (EngineCore_DP0 pid=11528) INFO 11-03 17:30:16 [gc_utils.py:40] GC Debug Config. enabled:False,top_objects:-1 2025-11-03T17:30:16.5902025Z INFO 11-03 17:30:16 [llm.py:337] Supported tasks: ['generate'] 2025-11-03T17:30:17.0861319Z PASSED 2025-11-03T17:30:17.1003512Z models/test_initialization.py::test_can_initialize_large_subset[Eagle3Qwen2_5vlForCausalLM] Fork a new process to run a test 11594 2025-11-03T17:30:17.1014177Z Fork a new process to run a test 0 2025-11-03T17:30:17.1294582Z INFO 11-03 17:30:17 [utils.py:239] non-default args: {'load_format': 'dummy', 'gpu_memory_utilization': 0.8, 'disable_log_stats': True, 'hf_overrides': functools.partial(, model_arch='Eagle3Qwen2_5vlForCausalLM', exist_overrides={}, use_original_num_layers=False), 'speculative_config': {'model': 'Rayzl/qwen2.5-vl-7b-eagle3-sgl', 'num_speculative_tokens': 1}, 'model_impl': 'vllm', 'model': 'Qwen/Qwen2.5-VL-7B-Instruct'} 2025-11-03T17:30:17.2104326Z 2025-11-03T17:30:17.2106473Z config.json: 0.00B [00:00, ?B/s] 2025-11-03T17:30:17.2106936Z config.json: 1.37kB [00:00, 6.49MB/s] 2025-11-03T17:30:17.3138114Z 2025-11-03T17:30:17.3139144Z preprocessor_config.json: 0% 0.00/350 [00:00, 'use_cudagraph': True, 'cudagraph_num_of_warmups': 1, 'cudagraph_capture_sizes': [512, 504, 496, 488, 480, 472, 464, 456, 448, 440, 432, 424, 416, 408, 400, 392, 384, 376, 368, 360, 352, 344, 336, 328, 320, 312, 304, 296, 288, 280, 272, 264, 256, 248, 240, 232, 224, 216, 208, 200, 192, 184, 176, 168, 160, 152, 144, 136, 128, 120, 112, 104, 96, 88, 80, 72, 64, 56, 48, 40, 32, 24, 16, 8, 4, 2, 1], 'cudagraph_copy_inputs': False, 'full_cuda_graph': True, 'use_inductor_graph_partition': False, 'pass_config': {}, 'max_capture_size': 512, 'local_cache_dir': None} 2025-11-03T17:30:32.2255163Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:30:32.2284546Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:30:32.2292369Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:30:32.2299744Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:30:32.2307765Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:30:32.2315421Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:30:32.2317918Z (EngineCore_DP0 pid=11702) INFO 11-03 17:30:32 [parallel_state.py:1231] rank 0 in world size 1 is assigned as DP rank 0, PP rank 0, TP rank 0, EP rank 0 2025-11-03T17:30:32.5839529Z (EngineCore_DP0 pid=11702) INFO 11-03 17:30:32 [topk_topp_sampler.py:59] Using FlashInfer for top-p & top-k sampling. 2025-11-03T17:30:32.5986326Z (EngineCore_DP0 pid=11702) WARNING 11-03 17:30:32 [__init__.py:194] min_p, logit_bias, and min_tokens parameters won't currently work with speculative decoding enabled. 2025-11-03T17:30:34.1743824Z (EngineCore_DP0 pid=11702) 2025-11-03T17:30:34.1745441Z chat_template.json: 0.00B [00:00, ?B/s] 2025-11-03T17:30:34.1745771Z chat_template.json: 1.05kB [00:00, 3.33MB/s] 2025-11-03T17:30:34.7231457Z (EngineCore_DP0 pid=11702) INFO 11-03 17:30:34 [gpu_model_runner.py:2840] Starting to load model Qwen/Qwen2.5-VL-7B-Instruct... 2025-11-03T17:30:35.0565677Z (EngineCore_DP0 pid=11702) INFO 11-03 17:30:35 [gpu_model_runner.py:2870] Loading model from scratch... 2025-11-03T17:30:35.1220595Z (EngineCore_DP0 pid=11702) INFO 11-03 17:30:35 [cuda.py:404] Using Flash Attention backend on V1 engine. 2025-11-03T17:30:35.1882125Z (EngineCore_DP0 pid=11702) INFO 11-03 17:30:35 [gpu_model_runner.py:2879] Loading drafter model... 2025-11-03T17:30:35.2014741Z (EngineCore_DP0 pid=11702) WARNING 11-03 17:30:35 [vllm.py:821] `torch.compile` is turned on, but the model Qwen/Qwen2.5-VL-7B-Instruct does not support it. Please open an issue on GitHub if you want it to be supported. 2025-11-03T17:30:35.2272102Z (EngineCore_DP0 pid=11702) INFO 11-03 17:30:35 [eagle.py:1005] Assuming the EAGLE head shares the same vocab embedding with the target model. 2025-11-03T17:30:35.2273471Z (EngineCore_DP0 pid=11702) INFO 11-03 17:30:35 [eagle.py:1043] The EAGLE head's lm_head will be loaded separately from the target model. 2025-11-03T17:30:35.7477447Z (EngineCore_DP0 pid=11702) INFO 11-03 17:30:35 [gpu_model_runner.py:2902] Model loading took 4.6155 GiB and 0.170996 seconds 2025-11-03T17:30:35.7482076Z (EngineCore_DP0 pid=11702) INFO 11-03 17:30:35 [kv_cache_utils.py:1199] GPU KV cache size: 2,621,440 tokens 2025-11-03T17:30:35.7483030Z (EngineCore_DP0 pid=11702) INFO 11-03 17:30:35 [kv_cache_utils.py:1204] Maximum concurrency for 128,000 tokens per request: 20.48x 2025-11-03T17:30:36.4536462Z (EngineCore_DP0 pid=11702) INFO 11-03 17:30:36 [gc_utils.py:40] GC Debug Config. enabled:False,top_objects:-1 2025-11-03T17:30:36.4556761Z INFO 11-03 17:30:36 [llm.py:337] Supported tasks: ['generate'] 2025-11-03T17:30:36.9534233Z PASSED 2025-11-03T17:30:36.9675965Z models/test_initialization.py::test_can_initialize_large_subset[GteNewForSequenceClassification] Fork a new process to run a test 11768 2025-11-03T17:30:36.9686669Z Fork a new process to run a test 0 2025-11-03T17:30:36.9966814Z INFO 11-03 17:30:36 [utils.py:239] non-default args: {'trust_remote_code': True, 'load_format': 'dummy', 'gpu_memory_utilization': 0.8, 'disable_log_stats': True, 'hf_overrides': functools.partial(, model_arch='GteNewForSequenceClassification', exist_overrides={'architectures': ['GteNewForSequenceClassification']}, use_original_num_layers=False), 'model_impl': 'vllm', 'model': 'Alibaba-NLP/gte-multilingual-reranker-base'} 2025-11-03T17:30:37.0921650Z 2025-11-03T17:30:37.0923778Z config.json: 0.00B [00:00, ?B/s] 2025-11-03T17:30:37.0924089Z config.json: 1.49kB [00:00, 7.10MB/s] 2025-11-03T17:30:37.1821878Z 2025-11-03T17:30:37.1824753Z configuration.py: 0.00B [00:00, ?B/s] 2025-11-03T17:30:37.1825196Z configuration.py: 7.13kB [00:00, 26.1MB/s] 2025-11-03T17:30:37.1919935Z A new version of the following files was downloaded from https://huggingface.co/Alibaba-NLP/new-impl: 2025-11-03T17:30:37.1920464Z - configuration.py 2025-11-03T17:30:37.1921412Z . Make sure to double-check they do not contain any added malicious code. To avoid downloading new versions of the code file, you can pin a revision. 2025-11-03T17:30:37.2274557Z INFO 11-03 17:30:37 [config.py:408] Replacing legacy 'type' key with 'rope_type' 2025-11-03T17:30:43.7388609Z INFO 11-03 17:30:43 [model.py:915] Resolved `--runner auto` to `--runner pooling`. Pass the value explicitly to silence this message. 2025-11-03T17:30:43.7389457Z INFO 11-03 17:30:43 [model.py:653] Resolved architecture: GteNewForSequenceClassification 2025-11-03T17:30:43.7390074Z `torch_dtype` is deprecated! Use `dtype` instead! 2025-11-03T17:30:43.7644038Z INFO 11-03 17:30:43 [model.py:1714] Using max model len 65536 2025-11-03T17:30:43.7645961Z The argument `trust_remote_code` is to be used with Auto classes. It has no effect here and is ignored. 2025-11-03T17:30:43.8272393Z INFO 11-03 17:30:43 [arg_utils.py:1725] (Disabling) chunked prefill by default 2025-11-03T17:30:43.8272939Z INFO 11-03 17:30:43 [arg_utils.py:1728] (Disabling) prefix caching by default 2025-11-03T17:30:43.8541600Z INFO 11-03 17:30:43 [vllm.py:433] Only "last" pooling supports chunked prefill and prefix caching; disabling both. 2025-11-03T17:30:44.0058372Z 2025-11-03T17:30:44.0059402Z tokenizer_config.json: 0.00B [00:00, ?B/s] 2025-11-03T17:30:44.0059896Z tokenizer_config.json: 1.34kB [00:00, 10.9MB/s] 2025-11-03T17:30:44.1381895Z 2025-11-03T17:30:44.4201384Z tokenizer.json: 0% 0.00/17.1M [00:00, 'use_cudagraph': True, 'cudagraph_num_of_warmups': 1, 'cudagraph_capture_sizes': [512, 504, 496, 488, 480, 472, 464, 456, 448, 440, 432, 424, 416, 408, 400, 392, 384, 376, 368, 360, 352, 344, 336, 328, 320, 312, 304, 296, 288, 280, 272, 264, 256, 248, 240, 232, 224, 216, 208, 200, 192, 184, 176, 168, 160, 152, 144, 136, 128, 120, 112, 104, 96, 88, 80, 72, 64, 56, 48, 40, 32, 24, 16, 8, 4, 2, 1], 'cudagraph_copy_inputs': False, 'full_cuda_graph': False, 'use_inductor_graph_partition': False, 'pass_config': {}, 'max_capture_size': 512, 'local_cache_dir': None} 2025-11-03T17:30:46.2717668Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:30:46.2745860Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:30:46.2753314Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:30:46.2761867Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:30:46.2770099Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:30:46.2778966Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:30:46.2781938Z (EngineCore_DP0 pid=11847) INFO 11-03 17:30:46 [parallel_state.py:1231] rank 0 in world size 1 is assigned as DP rank 0, PP rank 0, TP rank 0, EP rank 0 2025-11-03T17:30:46.6102674Z (EngineCore_DP0 pid=11847) INFO 11-03 17:30:46 [topk_topp_sampler.py:59] Using FlashInfer for top-p & top-k sampling. 2025-11-03T17:30:46.7060084Z (EngineCore_DP0 pid=11847) INFO 11-03 17:30:46 [gpu_model_runner.py:2840] Starting to load model Alibaba-NLP/gte-multilingual-reranker-base... 2025-11-03T17:30:46.9637740Z (EngineCore_DP0 pid=11847) INFO 11-03 17:30:46 [gpu_model_runner.py:2870] Loading model from scratch... 2025-11-03T17:30:47.0291384Z (EngineCore_DP0 pid=11847) INFO 11-03 17:30:47 [cuda.py:404] Using Flash Attention backend on V1 engine. 2025-11-03T17:30:47.5993800Z (EngineCore_DP0 pid=11847) INFO 11-03 17:30:47 [gpu_model_runner.py:2902] Model loading took 0.3780 GiB and 0.129733 seconds 2025-11-03T17:30:48.4243443Z (EngineCore_DP0 pid=11847) INFO 11-03 17:30:48 [core.py:142] Disabling chunked prefill for model without KVCache 2025-11-03T17:30:48.4634628Z (EngineCore_DP0 pid=11847) INFO 11-03 17:30:48 [vllm.py:433] Only "last" pooling supports chunked prefill and prefix caching; disabling both. 2025-11-03T17:30:48.7180554Z (EngineCore_DP0 pid=11847) INFO 11-03 17:30:48 [gc_utils.py:40] GC Debug Config. enabled:False,top_objects:-1 2025-11-03T17:30:48.7198216Z INFO 11-03 17:30:48 [llm.py:337] Supported tasks: ['score', 'classify', 'encode'] 2025-11-03T17:30:49.2517679Z PASSED 2025-11-03T17:30:49.2659439Z models/test_initialization.py::test_can_initialize_large_subset[AquilaForCausalLM] Fork a new process to run a test 11913 2025-11-03T17:30:49.2670196Z Fork a new process to run a test 0 2025-11-03T17:30:49.2948874Z INFO 11-03 17:30:49 [utils.py:239] non-default args: {'trust_remote_code': True, 'load_format': 'dummy', 'gpu_memory_utilization': 0.8, 'disable_log_stats': True, 'hf_overrides': functools.partial(, model_arch='AquilaForCausalLM', exist_overrides={}, use_original_num_layers=False), 'model_impl': 'vllm', 'model': 'BAAI/AquilaChat2-7B'} 2025-11-03T17:30:49.4742558Z 2025-11-03T17:30:49.4744335Z config.json: 0% 0.00/678 [00:00, 'use_cudagraph': True, 'cudagraph_num_of_warmups': 1, 'cudagraph_capture_sizes': [512, 504, 496, 488, 480, 472, 464, 456, 448, 440, 432, 424, 416, 408, 400, 392, 384, 376, 368, 360, 352, 344, 336, 328, 320, 312, 304, 296, 288, 280, 272, 264, 256, 248, 240, 232, 224, 216, 208, 200, 192, 184, 176, 168, 160, 152, 144, 136, 128, 120, 112, 104, 96, 88, 80, 72, 64, 56, 48, 40, 32, 24, 16, 8, 4, 2, 1], 'cudagraph_copy_inputs': False, 'full_cuda_graph': True, 'use_inductor_graph_partition': False, 'pass_config': {}, 'max_capture_size': 512, 'local_cache_dir': None} 2025-11-03T17:30:51.7951151Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:30:51.7978912Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:30:51.7987115Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:30:51.7994843Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:30:51.8002511Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:30:51.8010367Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:30:51.8013384Z (EngineCore_DP0 pid=11921) INFO 11-03 17:30:51 [parallel_state.py:1231] rank 0 in world size 1 is assigned as DP rank 0, PP rank 0, TP rank 0, EP rank 0 2025-11-03T17:30:52.1397648Z (EngineCore_DP0 pid=11921) INFO 11-03 17:30:52 [topk_topp_sampler.py:59] Using FlashInfer for top-p & top-k sampling. 2025-11-03T17:30:52.2067270Z (EngineCore_DP0 pid=11921) INFO 11-03 17:30:52 [gpu_model_runner.py:2840] Starting to load model BAAI/AquilaChat2-7B... 2025-11-03T17:30:52.4606910Z (EngineCore_DP0 pid=11921) INFO 11-03 17:30:52 [gpu_model_runner.py:2870] Loading model from scratch... 2025-11-03T17:30:52.5274128Z (EngineCore_DP0 pid=11921) INFO 11-03 17:30:52 [cuda.py:404] Using Flash Attention backend on V1 engine. 2025-11-03T17:30:53.0989172Z (EngineCore_DP0 pid=11921) INFO 11-03 17:30:53 [gpu_model_runner.py:2902] Model loading took 1.9048 GiB and 0.125995 seconds 2025-11-03T17:30:53.0994538Z (EngineCore_DP0 pid=11921) INFO 11-03 17:30:53 [kv_cache_utils.py:1199] GPU KV cache size: 655,360 tokens 2025-11-03T17:30:53.0995484Z (EngineCore_DP0 pid=11921) INFO 11-03 17:30:53 [kv_cache_utils.py:1204] Maximum concurrency for 2,048 tokens per request: 320.00x 2025-11-03T17:30:53.9558672Z (EngineCore_DP0 pid=11921) INFO 11-03 17:30:53 [gc_utils.py:40] GC Debug Config. enabled:False,top_objects:-1 2025-11-03T17:30:53.9572727Z INFO 11-03 17:30:53 [llm.py:337] Supported tasks: ['generate'] 2025-11-03T17:30:54.4070253Z PASSED 2025-11-03T17:30:54.4211578Z models/test_initialization.py::test_can_initialize_large_subset[MiniMaxText01ForCausalLM] Fork a new process to run a test 11987 2025-11-03T17:30:54.4222488Z Fork a new process to run a test 0 2025-11-03T17:30:54.4503739Z INFO 11-03 17:30:54 [utils.py:239] non-default args: {'trust_remote_code': True, 'load_format': 'dummy', 'gpu_memory_utilization': 0.8, 'disable_log_stats': True, 'revision': 'a59aa9cbc53b9fb8742ca4e9e1531b9802b6fdc3', 'hf_overrides': functools.partial(, model_arch='MiniMaxText01ForCausalLM', exist_overrides={}, use_original_num_layers=False), 'model_impl': 'vllm', 'model': 'MiniMaxAI/MiniMax-Text-01'} 2025-11-03T17:30:54.7181606Z 2025-11-03T17:30:54.7184185Z config.json: 0.00B [00:00, ?B/s] 2025-11-03T17:30:54.7184624Z config.json: 1.89kB [00:00, 8.87MB/s] 2025-11-03T17:30:54.7780569Z 2025-11-03T17:30:54.7782563Z configuration_minimax_text_01.py: 0.00B [00:00, ?B/s] 2025-11-03T17:30:54.7783151Z configuration_minimax_text_01.py: 7.36kB [00:00, 28.6MB/s] 2025-11-03T17:30:54.7891387Z You are using a model of type minimax_text_01 to instantiate a model of type MiniMaxText01. This is not supported for all configurations of models and can yield errors. 2025-11-03T17:30:55.0141376Z INFO 11-03 17:30:55 [model.py:653] Resolved architecture: MiniMaxText01ForCausalLM 2025-11-03T17:30:55.0142086Z `torch_dtype` is deprecated! Use `dtype` instead! 2025-11-03T17:30:55.1688155Z 2025-11-03T17:30:55.1708478Z model.safetensors.index.json: 0.00B [00:00, ?B/s] 2025-11-03T17:30:55.1709030Z model.safetensors.index.json: 823kB [00:00, 409MB/s] 2025-11-03T17:30:55.1955978Z 2025-11-03T17:30:55.3767503Z Parse safetensors files: 0% 0/413 [00:00= mamba page size. 2025-11-03T17:30:57.9495445Z 2025-11-03T17:30:57.9496849Z tokenizer_config.json: 0.00B [00:00, ?B/s] 2025-11-03T17:30:57.9497211Z tokenizer_config.json: 1.39kB [00:00, 11.4MB/s] 2025-11-03T17:30:58.4518221Z 2025-11-03T17:30:58.4718671Z vocab.json: 0.00B [00:00, ?B/s] 2025-11-03T17:30:58.4719027Z vocab.json: 4.71MB [00:00, 233MB/s] 2025-11-03T17:30:58.8537797Z 2025-11-03T17:30:58.8657469Z merges.txt: 0.00B [00:00, ?B/s] 2025-11-03T17:30:58.8657883Z merges.txt: 2.41MB [00:00, 198MB/s] 2025-11-03T17:30:58.9469706Z 2025-11-03T17:30:58.9828868Z tokenizer.json: 0.00B [00:00, ?B/s] 2025-11-03T17:30:58.9829325Z tokenizer.json: 9.72MB [00:00, 270MB/s] 2025-11-03T17:30:59.7558949Z You are using a model of type minimax_text_01 to instantiate a model of type MiniMaxText01. This is not supported for all configurations of models and can yield errors. 2025-11-03T17:30:59.7794756Z (EngineCore_DP0 pid=12015) INFO 11-03 17:30:59 [core.py:727] Waiting for init message from front-end. 2025-11-03T17:30:59.7833749Z (EngineCore_DP0 pid=12015) INFO 11-03 17:30:59 [core.py:94] Initializing a V1 LLM engine (v0.11.0rc2.dev389+ge51928192.d20251103) with config: model='MiniMaxAI/MiniMax-Text-01', speculative_config=None, tokenizer='MiniMaxAI/MiniMax-Text-01', skip_tokenizer_init=False, tokenizer_mode=auto, revision=a59aa9cbc53b9fb8742ca4e9e1531b9802b6fdc3, tokenizer_revision=a59aa9cbc53b9fb8742ca4e9e1531b9802b6fdc3, trust_remote_code=True, dtype=torch.bfloat16, max_seq_len=10240000, download_dir=None, load_format=dummy, tensor_parallel_size=1, pipeline_parallel_size=1, data_parallel_size=1, disable_custom_all_reduce=False, quantization=None, enforce_eager=False, kv_cache_dtype=auto, device_config=cuda, structured_outputs_config=StructuredOutputsConfig(backend='auto', disable_fallback=False, disable_any_whitespace=False, disable_additional_properties=False, reasoning_parser=''), observability_config=ObservabilityConfig(show_hidden_metrics_for_version=None, otlp_traces_endpoint=None, collect_detailed_traces=None), seed=0, served_model_name=MiniMaxAI/MiniMax-Text-01, enable_prefix_caching=False, chunked_prefill_enabled=True, pooler_config=None, compilation_config={'level': 3, 'debug_dump_path': None, 'cache_dir': '', 'backend': '', 'custom_ops': [], 'splitting_ops': ['vllm.unified_attention', 'vllm.unified_attention_with_output', 'vllm.unified_mla_attention', 'vllm.unified_mla_attention_with_output', 'vllm.mamba_mixer2', 'vllm.mamba_mixer', 'vllm.short_conv', 'vllm.linear_attention', 'vllm.plamo2_mamba_mixer', 'vllm.gdn_attention', 'vllm.sparse_attn_indexer'], 'use_inductor': True, 'compile_sizes': [], 'inductor_compile_config': {'enable_auto_functionalized_v2': False}, 'inductor_passes': {}, 'cudagraph_mode': , 'use_cudagraph': True, 'cudagraph_num_of_warmups': 1, 'cudagraph_capture_sizes': [512, 504, 496, 488, 480, 472, 464, 456, 448, 440, 432, 424, 416, 408, 400, 392, 384, 376, 368, 360, 352, 344, 336, 328, 320, 312, 304, 296, 288, 280, 272, 264, 256, 248, 240, 232, 224, 216, 208, 200, 192, 184, 176, 168, 160, 152, 144, 136, 128, 120, 112, 104, 96, 88, 80, 72, 64, 56, 48, 40, 32, 24, 16, 8, 4, 2, 1], 'cudagraph_copy_inputs': False, 'full_cuda_graph': True, 'use_inductor_graph_partition': False, 'pass_config': {}, 'max_capture_size': 512, 'local_cache_dir': None} 2025-11-03T17:31:00.4767630Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:31:00.4797273Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:31:00.4805655Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:31:00.4815214Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:31:00.4823488Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:31:00.4832573Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:31:00.4835188Z (EngineCore_DP0 pid=12015) INFO 11-03 17:31:00 [parallel_state.py:1231] rank 0 in world size 1 is assigned as DP rank 0, PP rank 0, TP rank 0, EP rank 0 2025-11-03T17:31:00.8292611Z (EngineCore_DP0 pid=12015) INFO 11-03 17:31:00 [topk_topp_sampler.py:59] Using FlashInfer for top-p & top-k sampling. 2025-11-03T17:31:01.8029102Z (EngineCore_DP0 pid=12015) INFO 11-03 17:31:01 [gpu_model_runner.py:2840] Starting to load model MiniMaxAI/MiniMax-Text-01... 2025-11-03T17:31:02.0659109Z (EngineCore_DP0 pid=12015) INFO 11-03 17:31:02 [gpu_model_runner.py:2870] Loading model from scratch... 2025-11-03T17:31:02.6977202Z (EngineCore_DP0 pid=12015) INFO 11-03 17:31:02 [gpu_model_runner.py:2902] Model loading took 7.9713 GiB and 0.111803 seconds 2025-11-03T17:31:02.6983215Z (EngineCore_DP0 pid=12015) INFO 11-03 17:31:02 [kv_cache_utils.py:1199] GPU KV cache size: 52,428,800,000 tokens 2025-11-03T17:31:02.6984354Z (EngineCore_DP0 pid=12015) INFO 11-03 17:31:02 [kv_cache_utils.py:1204] Maximum concurrency for 10,240,000 tokens per request: 5120.00x 2025-11-03T17:31:03.5750635Z (EngineCore_DP0 pid=12015) INFO 11-03 17:31:03 [gc_utils.py:40] GC Debug Config. enabled:False,top_objects:-1 2025-11-03T17:31:03.5765773Z INFO 11-03 17:31:03 [llm.py:337] Supported tasks: ['generate'] 2025-11-03T17:31:04.9801892Z PASSED 2025-11-03T17:31:04.9943180Z models/test_initialization.py::test_can_initialize_large_subset[Step3TextForCausalLM] Fork a new process to run a test 12081 2025-11-03T17:31:04.9954153Z Fork a new process to run a test 0 2025-11-03T17:31:05.0236200Z INFO 11-03 17:31:05 [utils.py:239] non-default args: {'trust_remote_code': True, 'load_format': 'dummy', 'gpu_memory_utilization': 0.8, 'disable_log_stats': True, 'hf_overrides': functools.partial(, model_arch='Step3TextForCausalLM', exist_overrides={}, use_original_num_layers=False), 'model_impl': 'vllm', 'model': 'stepfun-ai/step3'} 2025-11-03T17:31:05.1885161Z 2025-11-03T17:31:05.1887096Z config.json: 0.00B [00:00, ?B/s] 2025-11-03T17:31:05.1887423Z config.json: 1.54kB [00:00, 8.82MB/s] 2025-11-03T17:31:11.6024524Z INFO 11-03 17:31:11 [model.py:653] Resolved architecture: Step3VLForConditionalGeneration 2025-11-03T17:31:11.6025088Z `torch_dtype` is deprecated! Use `dtype` instead! 2025-11-03T17:31:11.6274230Z INFO 11-03 17:31:11 [model.py:1714] Using max model len 65536 2025-11-03T17:31:11.6276623Z The argument `trust_remote_code` is to be used with Auto classes. It has no effect here and is ignored. 2025-11-03T17:31:11.8165878Z INFO 11-03 17:31:11 [scheduler.py:225] Chunked prefill is enabled with max_num_batched_tokens=8192. 2025-11-03T17:31:11.8590808Z 2025-11-03T17:31:11.8592398Z tokenizer_config.json: 0.00B [00:00, ?B/s] 2025-11-03T17:31:11.8592801Z tokenizer_config.json: 3.94kB [00:00, 28.5MB/s] 2025-11-03T17:31:11.9697040Z 2025-11-03T17:31:12.0007102Z tokenizer.json: 0.00B [00:00, ?B/s] 2025-11-03T17:31:12.0007484Z tokenizer.json: 7.85MB [00:00, 252MB/s] 2025-11-03T17:31:12.5339289Z 2025-11-03T17:31:12.5340200Z generation_config.json: 0% 0.00/158 [00:00, 'use_cudagraph': True, 'cudagraph_num_of_warmups': 1, 'cudagraph_capture_sizes': [512, 504, 496, 488, 480, 472, 464, 456, 448, 440, 432, 424, 416, 408, 400, 392, 384, 376, 368, 360, 352, 344, 336, 328, 320, 312, 304, 296, 288, 280, 272, 264, 256, 248, 240, 232, 224, 216, 208, 200, 192, 184, 176, 168, 160, 152, 144, 136, 128, 120, 112, 104, 96, 88, 80, 72, 64, 56, 48, 40, 32, 24, 16, 8, 4, 2, 1], 'cudagraph_copy_inputs': False, 'full_cuda_graph': True, 'use_inductor_graph_partition': False, 'pass_config': {}, 'max_capture_size': 512, 'local_cache_dir': None} 2025-11-03T17:31:13.7222418Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:31:13.7249616Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:31:13.7257052Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:31:13.7265116Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:31:13.7273426Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:31:13.7281084Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:31:13.7283972Z (EngineCore_DP0 pid=12139) INFO 11-03 17:31:13 [parallel_state.py:1231] rank 0 in world size 1 is assigned as DP rank 0, PP rank 0, TP rank 0, EP rank 0 2025-11-03T17:31:14.0667614Z (EngineCore_DP0 pid=12139) INFO 11-03 17:31:14 [topk_topp_sampler.py:59] Using FlashInfer for top-p & top-k sampling. 2025-11-03T17:31:14.1750378Z (EngineCore_DP0 pid=12139) INFO 11-03 17:31:14 [gpu_model_runner.py:2840] Starting to load model stepfun-ai/step3... 2025-11-03T17:31:14.4369123Z (EngineCore_DP0 pid=12139) INFO 11-03 17:31:14 [gpu_model_runner.py:2870] Loading model from scratch... 2025-11-03T17:31:14.4545497Z (EngineCore_DP0 pid=12139) INFO 11-03 17:31:14 [layer.py:497] MultiHeadAttention attn_backend: _Backend.XFORMERS, use_upstream_fa: False 2025-11-03T17:31:14.5219196Z (EngineCore_DP0 pid=12139) INFO 11-03 17:31:14 [cuda.py:404] Using Flash Attention backend on V1 engine. 2025-11-03T17:31:15.0984257Z (EngineCore_DP0 pid=12139) INFO 11-03 17:31:15 [gpu_model_runner.py:2902] Model loading took 5.3915 GiB and 0.144693 seconds 2025-11-03T17:31:15.0988793Z (EngineCore_DP0 pid=12139) INFO 11-03 17:31:15 [kv_cache_utils.py:1199] GPU KV cache size: 10,485,760 tokens 2025-11-03T17:31:17.2763998Z (EngineCore_DP0 pid=12139) INFO 11-03 17:31:15 [kv_cache_utils.py:1204] Maximum concurrency for 65,536 tokens per request: 160.00x 2025-11-03T17:31:17.2765113Z (EngineCore_DP0 pid=12139) INFO 11-03 17:31:17 [gc_utils.py:40] GC Debug Config. enabled:False,top_objects:-1 2025-11-03T17:31:17.2778125Z INFO 11-03 17:31:17 [llm.py:337] Supported tasks: ['generate'] 2025-11-03T17:31:17.7778795Z PASSED 2025-11-03T17:31:17.7919531Z models/test_initialization.py::test_can_initialize_large_subset[Glm4vForConditionalGeneration] Fork a new process to run a test 12205 2025-11-03T17:31:17.7929944Z Fork a new process to run a test 0 2025-11-03T17:31:17.8204571Z INFO 11-03 17:31:17 [utils.py:239] non-default args: {'load_format': 'dummy', 'gpu_memory_utilization': 0.8, 'disable_log_stats': True, 'hf_overrides': functools.partial(, model_arch='Glm4vForConditionalGeneration', exist_overrides={}, use_original_num_layers=False), 'model_impl': 'vllm', 'model': 'zai-org/GLM-4.1V-9B-Thinking'} 2025-11-03T17:31:17.9088036Z 2025-11-03T17:31:17.9089990Z config.json: 0.00B [00:00, ?B/s] 2025-11-03T17:31:17.9090296Z config.json: 1.52kB [00:00, 8.84MB/s] 2025-11-03T17:31:18.0255116Z 2025-11-03T17:31:18.0256148Z preprocessor_config.json: 0% 0.00/364 [00:00, 'use_cudagraph': True, 'cudagraph_num_of_warmups': 1, 'cudagraph_capture_sizes': [512, 504, 496, 488, 480, 472, 464, 456, 448, 440, 432, 424, 416, 408, 400, 392, 384, 376, 368, 360, 352, 344, 336, 328, 320, 312, 304, 296, 288, 280, 272, 264, 256, 248, 240, 232, 224, 216, 208, 200, 192, 184, 176, 168, 160, 152, 144, 136, 128, 120, 112, 104, 96, 88, 80, 72, 64, 56, 48, 40, 32, 24, 16, 8, 4, 2, 1], 'cudagraph_copy_inputs': False, 'full_cuda_graph': True, 'use_inductor_graph_partition': False, 'pass_config': {}, 'max_capture_size': 512, 'local_cache_dir': None} 2025-11-03T17:31:27.8160430Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:31:27.8189815Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:31:27.8197707Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:31:27.8205728Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:31:27.8213633Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:31:27.8222091Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:31:27.8225392Z (EngineCore_DP0 pid=12288) INFO 11-03 17:31:27 [parallel_state.py:1231] rank 0 in world size 1 is assigned as DP rank 0, PP rank 0, TP rank 0, EP rank 0 2025-11-03T17:31:28.1610501Z (EngineCore_DP0 pid=12288) INFO 11-03 17:31:28 [topk_topp_sampler.py:59] Using FlashInfer for top-p & top-k sampling. 2025-11-03T17:31:28.4425071Z (EngineCore_DP0 pid=12288) Using a slow image processor as `use_fast` is unset and a slow processor was saved with this model. `use_fast=True` will be the default behavior in v4.52, even if the model was saved with a slow processor. This will result in minor differences in outputs. You'll still be able to use a slow processor with `use_fast=False`. 2025-11-03T17:31:29.6959068Z (EngineCore_DP0 pid=12288) 2025-11-03T17:31:29.6959544Z video_preprocessor_config.json: 0% 0.00/365 [00:00, model_arch='KimiVLForConditionalGeneration', exist_overrides={}, use_original_num_layers=False), 'model_impl': 'vllm', 'model': 'moonshotai/Kimi-VL-A3B-Instruct'} 2025-11-03T17:31:55.4583483Z 2025-11-03T17:31:55.4585976Z config.json: 0.00B [00:00, ?B/s] 2025-11-03T17:31:55.4586264Z config.json: 2.00kB [00:00, 9.17MB/s] 2025-11-03T17:31:55.5721878Z 2025-11-03T17:31:55.5724933Z preprocessor_config.json: 0% 0.00/401 [00:00, 'use_cudagraph': True, 'cudagraph_num_of_warmups': 1, 'cudagraph_capture_sizes': [512, 504, 496, 488, 480, 472, 464, 456, 448, 440, 432, 424, 416, 408, 400, 392, 384, 376, 368, 360, 352, 344, 336, 328, 320, 312, 304, 296, 288, 280, 272, 264, 256, 248, 240, 232, 224, 216, 208, 200, 192, 184, 176, 168, 160, 152, 144, 136, 128, 120, 112, 104, 96, 88, 80, 72, 64, 56, 48, 40, 32, 24, 16, 8, 4, 2, 1], 'cudagraph_copy_inputs': False, 'full_cuda_graph': True, 'use_inductor_graph_partition': False, 'pass_config': {}, 'max_capture_size': 512, 'local_cache_dir': None} 2025-11-03T17:32:04.4714077Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:32:04.4742473Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:32:04.4749066Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:32:04.4757105Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:32:04.4764742Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:32:04.4776095Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:32:04.4780726Z (EngineCore_DP0 pid=12449) INFO 11-03 17:32:04 [parallel_state.py:1231] rank 0 in world size 1 is assigned as DP rank 0, PP rank 0, TP rank 0, EP rank 0 2025-11-03T17:32:04.8187016Z (EngineCore_DP0 pid=12449) INFO 11-03 17:32:04 [topk_topp_sampler.py:59] Using FlashInfer for top-p & top-k sampling. 2025-11-03T17:32:05.4766853Z (EngineCore_DP0 pid=12449) 2025-11-03T17:32:05.4767387Z configuration_kimi_vl.py: 0.00B [00:00, ?B/s] 2025-11-03T17:32:05.4767801Z configuration_kimi_vl.py: 13.2kB [00:00, 63.0MB/s] 2025-11-03T17:32:05.8959980Z (EngineCore_DP0 pid=12449) 2025-11-03T17:32:05.8961882Z processing_kimi_vl.py: 0.00B [00:00, ?B/s] 2025-11-03T17:32:05.8962225Z processing_kimi_vl.py: 7.54kB [00:00, 24.4MB/s] 2025-11-03T17:32:06.0279629Z (EngineCore_DP0 pid=12449) 2025-11-03T17:32:06.0280502Z image_processing_kimi_vl.py: 0.00B [00:00, ?B/s] 2025-11-03T17:32:06.0280871Z image_processing_kimi_vl.py: 4.68kB [00:00, 15.0MB/s] 2025-11-03T17:32:13.6042231Z (EngineCore_DP0 pid=12449) INFO 11-03 17:32:13 [gpu_model_runner.py:2840] Starting to load model moonshotai/Kimi-VL-A3B-Instruct... 2025-11-03T17:32:13.8625356Z (EngineCore_DP0 pid=12449) INFO 11-03 17:32:13 [gpu_model_runner.py:2870] Loading model from scratch... 2025-11-03T17:32:13.8951680Z (EngineCore_DP0 pid=12449) INFO 11-03 17:32:13 [cuda.py:328] Using Triton MLA backend on V1 engine. 2025-11-03T17:32:14.4862191Z (EngineCore_DP0 pid=12449) INFO 11-03 17:32:14 [gpu_model_runner.py:2902] Model loading took 1.4523 GiB and 0.111748 seconds 2025-11-03T17:32:14.4866741Z (EngineCore_DP0 pid=12449) INFO 11-03 17:32:14 [kv_cache_utils.py:1199] GPU KV cache size: 9,320,672 tokens 2025-11-03T17:32:14.4867651Z (EngineCore_DP0 pid=12449) INFO 11-03 17:32:14 [kv_cache_utils.py:1204] Maximum concurrency for 131,072 tokens per request: 71.11x 2025-11-03T17:32:14.9871249Z (EngineCore_DP0 pid=12449) WARNING 11-03 17:32:14 [tokenizer.py:255] Using a slow tokenizer. This might cause a significant slowdown. Consider using a fast tokenizer instead. 2025-11-03T17:32:16.1783158Z (EngineCore_DP0 pid=12449) INFO 11-03 17:32:16 [gc_utils.py:40] GC Debug Config. enabled:False,top_objects:-1 2025-11-03T17:32:16.1798023Z INFO 11-03 17:32:16 [llm.py:337] Supported tasks: ['generate'] 2025-11-03T17:32:16.6773770Z PASSED 2025-11-03T17:32:16.6913620Z models/test_initialization.py::test_can_initialize_large_subset[TransformersMoEForSequenceClassification] Fork a new process to run a test 12515 2025-11-03T17:32:16.6924946Z Fork a new process to run a test 0 2025-11-03T17:32:16.6928845Z `transformers==4.56.2` installed, but `transformers>=4.57.0.dev0` is required to run this model. 2025-11-03T17:32:16.9836354Z PASSED 2025-11-03T17:32:16.9977178Z models/test_initialization.py::test_can_initialize_large_subset[FalconMambaForCausalLM] Fork a new process to run a test 12516 2025-11-03T17:32:16.9987406Z Fork a new process to run a test 0 2025-11-03T17:32:17.0259659Z INFO 11-03 17:32:17 [utils.py:239] non-default args: {'load_format': 'dummy', 'gpu_memory_utilization': 0.8, 'disable_log_stats': True, 'hf_overrides': functools.partial(, model_arch='FalconMambaForCausalLM', exist_overrides={}, use_original_num_layers=False), 'model_impl': 'vllm', 'model': 'tiiuae/falcon-mamba-7b-instruct'} 2025-11-03T17:32:17.1016626Z 2025-11-03T17:32:17.1018631Z config.json: 0% 0.00/870 [00:00, 'use_cudagraph': True, 'cudagraph_num_of_warmups': 1, 'cudagraph_capture_sizes': [512, 504, 496, 488, 480, 472, 464, 456, 448, 440, 432, 424, 416, 408, 400, 392, 384, 376, 368, 360, 352, 344, 336, 328, 320, 312, 304, 296, 288, 280, 272, 264, 256, 248, 240, 232, 224, 216, 208, 200, 192, 184, 176, 168, 160, 152, 144, 136, 128, 120, 112, 104, 96, 88, 80, 72, 64, 56, 48, 40, 32, 24, 16, 8, 4, 2, 1], 'cudagraph_copy_inputs': False, 'full_cuda_graph': True, 'use_inductor_graph_partition': False, 'pass_config': {}, 'max_capture_size': 512, 'local_cache_dir': None} 2025-11-03T17:32:24.8733569Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:32:24.8761700Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:32:24.8769364Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:32:24.8777420Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:32:24.8785174Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:32:24.8792813Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:32:24.8795569Z (EngineCore_DP0 pid=12574) INFO 11-03 17:32:24 [parallel_state.py:1231] rank 0 in world size 1 is assigned as DP rank 0, PP rank 0, TP rank 0, EP rank 0 2025-11-03T17:32:25.2157280Z (EngineCore_DP0 pid=12574) INFO 11-03 17:32:25 [topk_topp_sampler.py:59] Using FlashInfer for top-p & top-k sampling. 2025-11-03T17:32:25.4155655Z (EngineCore_DP0 pid=12574) INFO 11-03 17:32:25 [gpu_model_runner.py:2840] Starting to load model tiiuae/falcon-mamba-7b-instruct... 2025-11-03T17:32:25.6741677Z (EngineCore_DP0 pid=12574) INFO 11-03 17:32:25 [gpu_model_runner.py:2870] Loading model from scratch... 2025-11-03T17:32:26.2529129Z (EngineCore_DP0 pid=12574) INFO 11-03 17:32:26 [gpu_model_runner.py:2902] Model loading took 1.1886 GiB and 0.061591 seconds 2025-11-03T17:32:26.2534401Z (EngineCore_DP0 pid=12574) INFO 11-03 17:32:26 [kv_cache_utils.py:1199] GPU KV cache size: 70,639,616 tokens 2025-11-03T17:32:26.2535344Z (EngineCore_DP0 pid=12574) INFO 11-03 17:32:26 [kv_cache_utils.py:1204] Maximum concurrency for 2,048 tokens per request: 34492.00x 2025-11-03T17:32:26.7109542Z (EngineCore_DP0 pid=12574) INFO 11-03 17:32:26 [gc_utils.py:40] GC Debug Config. enabled:False,top_objects:-1 2025-11-03T17:32:26.7123471Z INFO 11-03 17:32:26 [llm.py:337] Supported tasks: ['generate'] 2025-11-03T17:32:27.1557644Z PASSED 2025-11-03T17:32:27.1697943Z models/test_initialization.py::test_can_initialize_large_subset[RWForCausalLM] Fork a new process to run a test 12640 2025-11-03T17:32:27.1709590Z Fork a new process to run a test 0 2025-11-03T17:32:27.1983655Z INFO 11-03 17:32:27 [utils.py:239] non-default args: {'load_format': 'dummy', 'gpu_memory_utilization': 0.8, 'disable_log_stats': True, 'hf_overrides': functools.partial(, model_arch='RWForCausalLM', exist_overrides={}, use_original_num_layers=False), 'model_impl': 'vllm', 'model': 'tiiuae/falcon-40b'} 2025-11-03T17:32:27.3459566Z 2025-11-03T17:32:27.3461309Z config.json: 0.00B [00:00, ?B/s] 2025-11-03T17:32:27.3461727Z config.json: 1.05kB [00:00, 5.61MB/s] 2025-11-03T17:32:33.8105873Z INFO 11-03 17:32:33 [model.py:653] Resolved architecture: FalconForCausalLM 2025-11-03T17:32:33.8106361Z `torch_dtype` is deprecated! Use `dtype` instead! 2025-11-03T17:32:33.8355996Z INFO 11-03 17:32:33 [model.py:1714] Using max model len 2048 2025-11-03T17:32:34.0093140Z INFO 11-03 17:32:34 [scheduler.py:225] Chunked prefill is enabled with max_num_batched_tokens=8192. 2025-11-03T17:32:34.0508536Z 2025-11-03T17:32:34.0509234Z tokenizer_config.json: 0% 0.00/242 [00:00, 'use_cudagraph': True, 'cudagraph_num_of_warmups': 1, 'cudagraph_capture_sizes': [512, 504, 496, 488, 480, 472, 464, 456, 448, 440, 432, 424, 416, 408, 400, 392, 384, 376, 368, 360, 352, 344, 336, 328, 320, 312, 304, 296, 288, 280, 272, 264, 256, 248, 240, 232, 224, 216, 208, 200, 192, 184, 176, 168, 160, 152, 144, 136, 128, 120, 112, 104, 96, 88, 80, 72, 64, 56, 48, 40, 32, 24, 16, 8, 4, 2, 1], 'cudagraph_copy_inputs': False, 'full_cuda_graph': True, 'use_inductor_graph_partition': False, 'pass_config': {}, 'max_capture_size': 512, 'local_cache_dir': None} 2025-11-03T17:32:35.2400371Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:32:35.2440402Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:32:35.2450852Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:32:35.2460623Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:32:35.2470911Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:32:35.2481527Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:32:35.2484975Z (EngineCore_DP0 pid=12698) INFO 11-03 17:32:35 [parallel_state.py:1231] rank 0 in world size 1 is assigned as DP rank 0, PP rank 0, TP rank 0, EP rank 0 2025-11-03T17:32:35.6005361Z (EngineCore_DP0 pid=12698) INFO 11-03 17:32:35 [topk_topp_sampler.py:59] Using FlashInfer for top-p & top-k sampling. 2025-11-03T17:32:35.7124192Z (EngineCore_DP0 pid=12698) INFO 11-03 17:32:35 [gpu_model_runner.py:2840] Starting to load model tiiuae/falcon-40b... 2025-11-03T17:32:35.9763000Z (EngineCore_DP0 pid=12698) INFO 11-03 17:32:35 [gpu_model_runner.py:2870] Loading model from scratch... 2025-11-03T17:32:36.0427782Z (EngineCore_DP0 pid=12698) INFO 11-03 17:32:36 [cuda.py:404] Using Flash Attention backend on V1 engine. 2025-11-03T17:32:36.6273286Z (EngineCore_DP0 pid=12698) INFO 11-03 17:32:36 [gpu_model_runner.py:2902] Model loading took 2.2582 GiB and 0.126250 seconds 2025-11-03T17:32:36.6277942Z (EngineCore_DP0 pid=12698) INFO 11-03 17:32:36 [kv_cache_utils.py:1199] GPU KV cache size: 5,242,880 tokens 2025-11-03T17:32:36.6278851Z (EngineCore_DP0 pid=12698) INFO 11-03 17:32:36 [kv_cache_utils.py:1204] Maximum concurrency for 2,048 tokens per request: 2560.00x 2025-11-03T17:32:37.5760698Z (EngineCore_DP0 pid=12698) INFO 11-03 17:32:37 [gc_utils.py:40] GC Debug Config. enabled:False,top_objects:-1 2025-11-03T17:32:37.5777102Z INFO 11-03 17:32:37 [llm.py:337] Supported tasks: ['generate'] 2025-11-03T17:32:38.0438414Z PASSED 2025-11-03T17:32:38.0578951Z models/test_initialization.py::test_can_initialize_large_subset[WhisperForConditionalGeneration] Fork a new process to run a test 12764 2025-11-03T17:32:38.0590373Z Fork a new process to run a test 0 2025-11-03T17:32:38.0872693Z INFO 11-03 17:32:38 [utils.py:239] non-default args: {'load_format': 'dummy', 'gpu_memory_utilization': 0.8, 'disable_log_stats': True, 'hf_overrides': functools.partial(, model_arch='WhisperForConditionalGeneration', exist_overrides={}, use_original_num_layers=False), 'model_impl': 'vllm', 'model': 'openai/whisper-large-v3'} 2025-11-03T17:32:38.1636152Z 2025-11-03T17:32:38.1638258Z config.json: 0.00B [00:00, ?B/s] 2025-11-03T17:32:38.1638560Z config.json: 1.27kB [00:00, 6.99MB/s] 2025-11-03T17:32:38.2660624Z 2025-11-03T17:32:38.2661848Z preprocessor_config.json: 0% 0.00/340 [00:00, 'use_cudagraph': True, 'cudagraph_num_of_warmups': 1, 'cudagraph_capture_sizes': [512, 504, 496, 488, 480, 472, 464, 456, 448, 440, 432, 424, 416, 408, 400, 392, 384, 376, 368, 360, 352, 344, 336, 328, 320, 312, 304, 296, 288, 280, 272, 264, 256, 248, 240, 232, 224, 216, 208, 200, 192, 184, 176, 168, 160, 152, 144, 136, 128, 120, 112, 104, 96, 88, 80, 72, 64, 56, 48, 40, 32, 24, 16, 8, 4, 2, 1], 'cudagraph_copy_inputs': False, 'full_cuda_graph': False, 'use_inductor_graph_partition': False, 'pass_config': {}, 'max_capture_size': 512, 'local_cache_dir': None} 2025-11-03T17:32:57.9761053Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:32:57.9788574Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:32:57.9795662Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:32:57.9802726Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:32:57.9809705Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:32:57.9816734Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:32:57.9819359Z (EngineCore_DP0 pid=12846) INFO 11-03 17:32:57 [parallel_state.py:1231] rank 0 in world size 1 is assigned as DP rank 0, PP rank 0, TP rank 0, EP rank 0 2025-11-03T17:32:58.2124466Z (EngineCore_DP0 pid=12846) INFO 11-03 17:32:58 [topk_topp_sampler.py:59] Using FlashInfer for top-p & top-k sampling. 2025-11-03T17:32:59.7131443Z (EngineCore_DP0 pid=12846) INFO 11-03 17:32:59 [gpu_model_runner.py:2840] Starting to load model openai/whisper-large-v3... 2025-11-03T17:32:59.9618202Z (EngineCore_DP0 pid=12846) INFO 11-03 17:32:59 [gpu_model_runner.py:2870] Loading model from scratch... 2025-11-03T17:32:59.9648141Z (EngineCore_DP0 pid=12846) INFO 11-03 17:32:59 [layer.py:497] MultiHeadAttention attn_backend: _Backend.FLASH_ATTN, use_upstream_fa: False 2025-11-03T17:33:00.0223854Z (EngineCore_DP0 pid=12846) INFO 11-03 17:33:00 [cuda.py:404] Using Flash Attention backend on V1 engine. 2025-11-03T17:33:00.0985759Z (EngineCore_DP0 pid=12846) WARNING 11-03 17:33:00 [vllm.py:821] `torch.compile` is turned on, but the model openai/whisper-large-v3 does not support it. Please open an issue on GitHub if you want it to be supported. 2025-11-03T17:33:00.5731520Z (EngineCore_DP0 pid=12846) INFO 11-03 17:33:00 [gpu_model_runner.py:2902] Model loading took 2.8763 GiB and 0.157345 seconds 2025-11-03T17:33:00.7977395Z (EngineCore_DP0 pid=12846) INFO 11-03 17:33:00 [gpu_model_runner.py:3647] Encoder cache will be initialized with a budget of 1500 tokens, and profiled with 1 audio items of the maximum feature size. 2025-11-03T17:33:00.7996141Z (EngineCore_DP0 pid=12846) WARNING 11-03 17:33:00 [processing.py:1091] WhisperProcessor did not return `BatchFeature`. Make sure to match the behaviour of `ProcessorMixin` when implementing custom processors. 2025-11-03T17:33:01.8888601Z (EngineCore_DP0 pid=12846) INFO 11-03 17:33:01 [gpu_worker.py:314] Available KV cache memory: 14.45 GiB 2025-11-03T17:33:02.1122922Z (EngineCore_DP0 pid=12846) INFO 11-03 17:33:02 [kv_cache_utils.py:1199] GPU KV cache size: 47,328 tokens 2025-11-03T17:33:02.1123757Z (EngineCore_DP0 pid=12846) INFO 11-03 17:33:02 [kv_cache_utils.py:1204] Maximum concurrency for 448 tokens per request: 48.50x 2025-11-03T17:33:02.3489085Z (EngineCore_DP0 pid=12846) 2025-11-03T17:33:02.4798967Z Capturing CUDA graphs (mixed prefill-decode, PIECEWISE): 0% 0/67 [00:00, model_arch='DeciLMForCausalLM', exist_overrides={}, use_original_num_layers=False), 'model_impl': 'vllm', 'model': 'nvidia/Llama-3_3-Nemotron-Super-49B-v1'} 2025-11-03T17:33:06.0332584Z 2025-11-03T17:33:06.0336220Z config.json: 0.00B [00:00, ?B/s] 2025-11-03T17:33:06.0336507Z config.json: 37.1kB [00:00, 104MB/s] 2025-11-03T17:33:06.1103432Z 2025-11-03T17:33:06.1106064Z configuration_decilm.py: 0.00B [00:00, ?B/s] 2025-11-03T17:33:06.1106617Z configuration_decilm.py: 2.57kB [00:00, 12.3MB/s] 2025-11-03T17:33:06.1649041Z 2025-11-03T17:33:06.1658535Z (…)nsformers_4_44_2__modeling_rope_utils.py: 0.00B [00:00, ?B/s] 2025-11-03T17:33:06.1659087Z (…)nsformers_4_44_2__modeling_rope_utils.py: 28.1kB [00:00, 143MB/s] 2025-11-03T17:33:06.1737452Z A new version of the following files was downloaded from https://huggingface.co/nvidia/Llama-3_3-Nemotron-Super-49B-v1: 2025-11-03T17:33:06.1738067Z - transformers_4_44_2__modeling_rope_utils.py 2025-11-03T17:33:06.1738747Z . Make sure to double-check they do not contain any added malicious code. To avoid downloading new versions of the code file, you can pin a revision. 2025-11-03T17:33:06.2173307Z 2025-11-03T17:33:06.2175759Z block_config.py: 0.00B [00:00, ?B/s] 2025-11-03T17:33:06.2176217Z block_config.py: 4.35kB [00:00, 22.2MB/s] 2025-11-03T17:33:06.2206987Z A new version of the following files was downloaded from https://huggingface.co/nvidia/Llama-3_3-Nemotron-Super-49B-v1: 2025-11-03T17:33:06.2207564Z - block_config.py 2025-11-03T17:33:06.2208147Z . Make sure to double-check they do not contain any added malicious code. To avoid downloading new versions of the code file, you can pin a revision. 2025-11-03T17:33:06.2549990Z 2025-11-03T17:33:06.2553404Z (…)nsformers_4_44_2__configuration_llama.py: 0.00B [00:00, ?B/s] 2025-11-03T17:33:06.2553933Z (…)nsformers_4_44_2__configuration_llama.py: 11.1kB [00:00, 38.4MB/s] 2025-11-03T17:33:06.2575003Z A new version of the following files was downloaded from https://huggingface.co/nvidia/Llama-3_3-Nemotron-Super-49B-v1: 2025-11-03T17:33:06.2575604Z - transformers_4_44_2__configuration_llama.py 2025-11-03T17:33:06.2576251Z . Make sure to double-check they do not contain any added malicious code. To avoid downloading new versions of the code file, you can pin a revision. 2025-11-03T17:33:06.2577141Z A new version of the following files was downloaded from https://huggingface.co/nvidia/Llama-3_3-Nemotron-Super-49B-v1: 2025-11-03T17:33:06.2577868Z - configuration_decilm.py 2025-11-03T17:33:06.2578162Z - transformers_4_44_2__modeling_rope_utils.py 2025-11-03T17:33:06.2578448Z - block_config.py 2025-11-03T17:33:06.2578681Z - transformers_4_44_2__configuration_llama.py 2025-11-03T17:33:06.2579311Z . Make sure to double-check they do not contain any added malicious code. To avoid downloading new versions of the code file, you can pin a revision. 2025-11-03T17:33:12.8683407Z INFO 11-03 17:33:12 [model.py:653] Resolved architecture: DeciLMForCausalLM 2025-11-03T17:33:12.8683910Z `torch_dtype` is deprecated! Use `dtype` instead! 2025-11-03T17:33:12.8934676Z INFO 11-03 17:33:12 [model.py:1714] Using max model len 131072 2025-11-03T17:33:12.8936179Z The argument `trust_remote_code` is to be used with Auto classes. It has no effect here and is ignored. 2025-11-03T17:33:12.9465919Z INFO 11-03 17:33:12 [scheduler.py:225] Chunked prefill is enabled with max_num_batched_tokens=8192. 2025-11-03T17:33:12.9878672Z 2025-11-03T17:33:12.9882273Z tokenizer_config.json: 0.00B [00:00, ?B/s] 2025-11-03T17:33:12.9882630Z tokenizer_config.json: 51.3kB [00:00, 147MB/s] 2025-11-03T17:33:13.0728107Z 2025-11-03T17:33:13.4022018Z tokenizer.json: 0% 0.00/17.2M [00:00, 'use_cudagraph': True, 'cudagraph_num_of_warmups': 1, 'cudagraph_capture_sizes': [512, 504, 496, 488, 480, 472, 464, 456, 448, 440, 432, 424, 416, 408, 400, 392, 384, 376, 368, 360, 352, 344, 336, 328, 320, 312, 304, 296, 288, 280, 272, 264, 256, 248, 240, 232, 224, 216, 208, 200, 192, 184, 176, 168, 160, 152, 144, 136, 128, 120, 112, 104, 96, 88, 80, 72, 64, 56, 48, 40, 32, 24, 16, 8, 4, 2, 1], 'cudagraph_copy_inputs': False, 'full_cuda_graph': True, 'use_inductor_graph_partition': False, 'pass_config': {}, 'max_capture_size': 512, 'local_cache_dir': None} 2025-11-03T17:33:14.9092073Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:33:14.9120145Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:33:14.9127608Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:33:14.9135239Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:33:14.9142981Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:33:14.9150249Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:33:14.9153208Z (EngineCore_DP0 pid=13059) INFO 11-03 17:33:14 [parallel_state.py:1231] rank 0 in world size 1 is assigned as DP rank 0, PP rank 0, TP rank 0, EP rank 0 2025-11-03T17:33:15.2584666Z (EngineCore_DP0 pid=13059) INFO 11-03 17:33:15 [topk_topp_sampler.py:59] Using FlashInfer for top-p & top-k sampling. 2025-11-03T17:33:15.3920591Z (EngineCore_DP0 pid=13059) INFO 11-03 17:33:15 [gpu_model_runner.py:2840] Starting to load model nvidia/Llama-3_3-Nemotron-Super-49B-v1... 2025-11-03T17:33:15.6543467Z (EngineCore_DP0 pid=13059) INFO 11-03 17:33:15 [gpu_model_runner.py:2870] Loading model from scratch... 2025-11-03T17:33:15.7627844Z (EngineCore_DP0 pid=13059) INFO 11-03 17:33:15 [cuda.py:404] Using Flash Attention backend on V1 engine. 2025-11-03T17:33:16.3347442Z (EngineCore_DP0 pid=13059) INFO 11-03 17:33:16 [gpu_model_runner.py:2902] Model loading took 4.8829 GiB and 0.168441 seconds 2025-11-03T17:33:16.3351829Z (EngineCore_DP0 pid=13059) INFO 11-03 17:33:16 [kv_cache_utils.py:1199] GPU KV cache size: 2,621,440 tokens 2025-11-03T17:33:16.3352801Z (EngineCore_DP0 pid=13059) INFO 11-03 17:33:16 [kv_cache_utils.py:1204] Maximum concurrency for 131,072 tokens per request: 20.00x 2025-11-03T17:33:17.2496225Z (EngineCore_DP0 pid=13059) INFO 11-03 17:33:17 [gc_utils.py:40] GC Debug Config. enabled:False,top_objects:-1 2025-11-03T17:33:17.2511615Z INFO 11-03 17:33:17 [llm.py:337] Supported tasks: ['generate'] 2025-11-03T17:33:17.7658448Z PASSED 2025-11-03T17:33:17.7798400Z models/test_initialization.py::test_can_initialize_large_subset[BailingMoeForCausalLM] Fork a new process to run a test 13125 2025-11-03T17:33:17.7809016Z Fork a new process to run a test 0 2025-11-03T17:33:17.8089713Z INFO 11-03 17:33:17 [utils.py:239] non-default args: {'trust_remote_code': True, 'load_format': 'dummy', 'gpu_memory_utilization': 0.8, 'disable_log_stats': True, 'hf_overrides': functools.partial(, model_arch='BailingMoeForCausalLM', exist_overrides={}, use_original_num_layers=False), 'model_impl': 'vllm', 'model': 'inclusionAI/Ling-lite-1.5'} 2025-11-03T17:33:17.9418872Z 2025-11-03T17:33:17.9420900Z config.json: 0.00B [00:00, ?B/s] 2025-11-03T17:33:17.9421679Z config.json: 1.28kB [00:00, 6.67MB/s] 2025-11-03T17:33:18.0204122Z 2025-11-03T17:33:18.0206966Z configuration_bailing_moe.py: 0.00B [00:00, ?B/s] 2025-11-03T17:33:18.0207532Z configuration_bailing_moe.py: 2.86kB [00:00, 12.9MB/s] 2025-11-03T17:33:18.0296010Z A new version of the following files was downloaded from https://huggingface.co/inclusionAI/Ling-lite-1.5: 2025-11-03T17:33:18.0296578Z - configuration_bailing_moe.py 2025-11-03T17:33:18.0297219Z . Make sure to double-check they do not contain any added malicious code. To avoid downloading new versions of the code file, you can pin a revision. 2025-11-03T17:33:24.4883740Z INFO 11-03 17:33:24 [model.py:653] Resolved architecture: BailingMoeForCausalLM 2025-11-03T17:33:24.4884321Z `torch_dtype` is deprecated! Use `dtype` instead! 2025-11-03T17:33:24.5140851Z INFO 11-03 17:33:24 [model.py:1714] Using max model len 32768 2025-11-03T17:33:24.5142335Z The argument `trust_remote_code` is to be used with Auto classes. It has no effect here and is ignored. 2025-11-03T17:33:24.5672913Z INFO 11-03 17:33:24 [scheduler.py:225] Chunked prefill is enabled with max_num_batched_tokens=8192. 2025-11-03T17:33:24.6055884Z 2025-11-03T17:33:24.6059589Z tokenizer_config.json: 0.00B [00:00, ?B/s] 2025-11-03T17:33:24.6059976Z tokenizer_config.json: 51.1kB [00:00, 168MB/s] 2025-11-03T17:33:24.7592507Z 2025-11-03T17:33:24.7946556Z tokenizer.json: 0.00B [00:00, ?B/s] 2025-11-03T17:33:24.7946933Z tokenizer.json: 9.75MB [00:00, 275MB/s] 2025-11-03T17:33:24.8963757Z 2025-11-03T17:33:24.8964359Z special_tokens_map.json: 0% 0.00/747 [00:00, 'use_cudagraph': True, 'cudagraph_num_of_warmups': 1, 'cudagraph_capture_sizes': [512, 504, 496, 488, 480, 472, 464, 456, 448, 440, 432, 424, 416, 408, 400, 392, 384, 376, 368, 360, 352, 344, 336, 328, 320, 312, 304, 296, 288, 280, 272, 264, 256, 248, 240, 232, 224, 216, 208, 200, 192, 184, 176, 168, 160, 152, 144, 136, 128, 120, 112, 104, 96, 88, 80, 72, 64, 56, 48, 40, 32, 24, 16, 8, 4, 2, 1], 'cudagraph_copy_inputs': False, 'full_cuda_graph': True, 'use_inductor_graph_partition': False, 'pass_config': {}, 'max_capture_size': 512, 'local_cache_dir': None} 2025-11-03T17:33:26.1693252Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:33:26.1721549Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:33:26.1729882Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:33:26.1738004Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:33:26.1746456Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:33:26.1754853Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:33:26.1757798Z (EngineCore_DP0 pid=13183) INFO 11-03 17:33:26 [parallel_state.py:1231] rank 0 in world size 1 is assigned as DP rank 0, PP rank 0, TP rank 0, EP rank 0 2025-11-03T17:33:26.5077930Z (EngineCore_DP0 pid=13183) INFO 11-03 17:33:26 [topk_topp_sampler.py:59] Using FlashInfer for top-p & top-k sampling. 2025-11-03T17:33:26.5696066Z (EngineCore_DP0 pid=13183) INFO 11-03 17:33:26 [gpu_model_runner.py:2840] Starting to load model inclusionAI/Ling-lite-1.5... 2025-11-03T17:33:26.8217578Z (EngineCore_DP0 pid=13183) INFO 11-03 17:33:26 [gpu_model_runner.py:2870] Loading model from scratch... 2025-11-03T17:33:26.8894565Z (EngineCore_DP0 pid=13183) INFO 11-03 17:33:26 [cuda.py:404] Using Flash Attention backend on V1 engine. 2025-11-03T17:33:27.4374105Z (EngineCore_DP0 pid=13183) INFO 11-03 17:33:27 [gpu_model_runner.py:2902] Model loading took 1.0567 GiB and 0.127207 seconds 2025-11-03T17:33:27.4379174Z (EngineCore_DP0 pid=13183) INFO 11-03 17:33:27 [kv_cache_utils.py:1199] GPU KV cache size: 5,242,880 tokens 2025-11-03T17:33:27.4380127Z (EngineCore_DP0 pid=13183) INFO 11-03 17:33:27 [kv_cache_utils.py:1204] Maximum concurrency for 32,768 tokens per request: 426.11x 2025-11-03T17:33:28.6139123Z (EngineCore_DP0 pid=13183) INFO 11-03 17:33:28 [gc_utils.py:40] GC Debug Config. enabled:False,top_objects:-1 2025-11-03T17:33:28.6156232Z INFO 11-03 17:33:28 [llm.py:337] Supported tasks: ['generate'] 2025-11-03T17:33:29.0801079Z PASSED 2025-11-03T17:33:29.0940457Z models/test_initialization.py::test_can_initialize_large_subset[LlamaForCausalLMEagle3] Fork a new process to run a test 13249 2025-11-03T17:33:29.0951987Z Fork a new process to run a test 0 2025-11-03T17:33:29.1230991Z INFO 11-03 17:33:29 [utils.py:239] non-default args: {'tokenizer': 'Qwen/Qwen3-8B', 'trust_remote_code': True, 'load_format': 'dummy', 'gpu_memory_utilization': 0.8, 'disable_log_stats': True, 'hf_overrides': functools.partial(, model_arch='LlamaForCausalLMEagle3', exist_overrides={}, use_original_num_layers=True), 'speculative_config': {'model': 'AngelSlim/Qwen3-8B_eagle3', 'num_speculative_tokens': 1}, 'model_impl': 'vllm', 'model': 'Qwen/Qwen3-8B'} 2025-11-03T17:33:29.2069358Z 2025-11-03T17:33:29.2070349Z config.json: 0% 0.00/728 [00:00, 'use_cudagraph': True, 'cudagraph_num_of_warmups': 1, 'cudagraph_capture_sizes': [512, 504, 496, 488, 480, 472, 464, 456, 448, 440, 432, 424, 416, 408, 400, 392, 384, 376, 368, 360, 352, 344, 336, 328, 320, 312, 304, 296, 288, 280, 272, 264, 256, 248, 240, 232, 224, 216, 208, 200, 192, 184, 176, 168, 160, 152, 144, 136, 128, 120, 112, 104, 96, 88, 80, 72, 64, 56, 48, 40, 32, 24, 16, 8, 4, 2, 1], 'cudagraph_copy_inputs': False, 'full_cuda_graph': True, 'use_inductor_graph_partition': False, 'pass_config': {}, 'max_capture_size': 512, 'local_cache_dir': None} 2025-11-03T17:33:37.6908082Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:33:37.6936379Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:33:37.6943897Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:33:37.6951558Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:33:37.6959063Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:33:37.6966918Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:33:37.6969886Z (EngineCore_DP0 pid=13327) INFO 11-03 17:33:37 [parallel_state.py:1231] rank 0 in world size 1 is assigned as DP rank 0, PP rank 0, TP rank 0, EP rank 0 2025-11-03T17:33:38.0316600Z (EngineCore_DP0 pid=13327) INFO 11-03 17:33:38 [topk_topp_sampler.py:59] Using FlashInfer for top-p & top-k sampling. 2025-11-03T17:33:38.0512866Z (EngineCore_DP0 pid=13327) WARNING 11-03 17:33:38 [__init__.py:194] min_p, logit_bias, and min_tokens parameters won't currently work with speculative decoding enabled. 2025-11-03T17:33:38.1064666Z (EngineCore_DP0 pid=13327) INFO 11-03 17:33:38 [gpu_model_runner.py:2840] Starting to load model Qwen/Qwen3-8B... 2025-11-03T17:33:38.3596007Z (EngineCore_DP0 pid=13327) INFO 11-03 17:33:38 [gpu_model_runner.py:2870] Loading model from scratch... 2025-11-03T17:33:38.4197491Z (EngineCore_DP0 pid=13327) INFO 11-03 17:33:38 [cuda.py:404] Using Flash Attention backend on V1 engine. 2025-11-03T17:33:38.5485411Z (EngineCore_DP0 pid=13327) INFO 11-03 17:33:38 [gpu_model_runner.py:2879] Loading drafter model... 2025-11-03T17:33:38.6160430Z (EngineCore_DP0 pid=13327) INFO 11-03 17:33:38 [eagle.py:1005] Assuming the EAGLE head shares the same vocab embedding with the target model. 2025-11-03T17:33:38.6161529Z (EngineCore_DP0 pid=13327) INFO 11-03 17:33:38 [eagle.py:1043] The EAGLE head's lm_head will be loaded separately from the target model. 2025-11-03T17:33:39.1254028Z (EngineCore_DP0 pid=13327) INFO 11-03 17:33:39 [gpu_model_runner.py:2902] Model loading took 16.0127 GiB and 0.256846 seconds 2025-11-03T17:33:39.1262651Z (EngineCore_DP0 pid=13327) INFO 11-03 17:33:39 [kv_cache_utils.py:1199] GPU KV cache size: 70,848 tokens 2025-11-03T17:33:39.1263750Z (EngineCore_DP0 pid=13327) INFO 11-03 17:33:39 [kv_cache_utils.py:1204] Maximum concurrency for 40,960 tokens per request: 1.73x 2025-11-03T17:33:39.8416442Z (EngineCore_DP0 pid=13327) INFO 11-03 17:33:39 [gc_utils.py:40] GC Debug Config. enabled:False,top_objects:-1 2025-11-03T17:33:39.8432721Z INFO 11-03 17:33:39 [llm.py:337] Supported tasks: ['generate'] 2025-11-03T17:33:40.3273848Z PASSED 2025-11-03T17:33:40.3412571Z models/test_initialization.py::test_can_initialize_large_subset[DeepseekV2ForCausalLM] Fork a new process to run a test 13393 2025-11-03T17:33:40.3424427Z Fork a new process to run a test 0 2025-11-03T17:33:40.3701077Z INFO 11-03 17:33:40 [utils.py:239] non-default args: {'trust_remote_code': True, 'load_format': 'dummy', 'gpu_memory_utilization': 0.8, 'disable_log_stats': True, 'hf_overrides': functools.partial(, model_arch='DeepseekV2ForCausalLM', exist_overrides={}, use_original_num_layers=False), 'model_impl': 'vllm', 'model': 'deepseek-ai/DeepSeek-V2-Lite-Chat'} 2025-11-03T17:33:40.4556871Z 2025-11-03T17:33:40.4559241Z config.json: 0.00B [00:00, ?B/s] 2025-11-03T17:33:40.4559554Z config.json: 1.52kB [00:00, 7.13MB/s] 2025-11-03T17:33:40.5217283Z 2025-11-03T17:33:40.5219522Z configuration_deepseek.py: 0.00B [00:00, ?B/s] 2025-11-03T17:33:40.5219912Z configuration_deepseek.py: 10.3kB [00:00, 55.9MB/s] 2025-11-03T17:33:40.5315306Z A new version of the following files was downloaded from https://huggingface.co/deepseek-ai/DeepSeek-V2-Lite-Chat: 2025-11-03T17:33:40.5315889Z - configuration_deepseek.py 2025-11-03T17:33:40.5316511Z . Make sure to double-check they do not contain any added malicious code. To avoid downloading new versions of the code file, you can pin a revision. 2025-11-03T17:33:40.5622618Z INFO 11-03 17:33:40 [config.py:408] Replacing legacy 'type' key with 'rope_type' 2025-11-03T17:33:40.7373838Z INFO 11-03 17:33:40 [model.py:653] Resolved architecture: DeepseekV2ForCausalLM 2025-11-03T17:33:40.7374401Z `torch_dtype` is deprecated! Use `dtype` instead! 2025-11-03T17:33:40.7619510Z INFO 11-03 17:33:40 [model.py:1714] Using max model len 163840 2025-11-03T17:33:40.7621955Z The argument `trust_remote_code` is to be used with Auto classes. It has no effect here and is ignored. 2025-11-03T17:33:40.8114542Z INFO 11-03 17:33:40 [scheduler.py:225] Chunked prefill is enabled with max_num_batched_tokens=8192. 2025-11-03T17:33:40.8529130Z 2025-11-03T17:33:40.8530691Z tokenizer_config.json: 0.00B [00:00, ?B/s] 2025-11-03T17:33:40.8531068Z tokenizer_config.json: 1.28kB [00:00, 12.6MB/s] 2025-11-03T17:33:41.0598001Z 2025-11-03T17:33:41.0791619Z tokenizer.json: 0.00B [00:00, ?B/s] 2025-11-03T17:33:41.0791954Z tokenizer.json: 4.61MB [00:00, 239MB/s] 2025-11-03T17:33:41.5529240Z 2025-11-03T17:33:41.5529864Z generation_config.json: 0% 0.00/181 [00:00, 'use_cudagraph': True, 'cudagraph_num_of_warmups': 1, 'cudagraph_capture_sizes': [512, 504, 496, 488, 480, 472, 464, 456, 448, 440, 432, 424, 416, 408, 400, 392, 384, 376, 368, 360, 352, 344, 336, 328, 320, 312, 304, 296, 288, 280, 272, 264, 256, 248, 240, 232, 224, 216, 208, 200, 192, 184, 176, 168, 160, 152, 144, 136, 128, 120, 112, 104, 96, 88, 80, 72, 64, 56, 48, 40, 32, 24, 16, 8, 4, 2, 1], 'cudagraph_copy_inputs': False, 'full_cuda_graph': True, 'use_inductor_graph_partition': False, 'pass_config': {}, 'max_capture_size': 512, 'local_cache_dir': None} 2025-11-03T17:33:42.2457097Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:33:42.2484798Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:33:42.2492498Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:33:42.2500188Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:33:42.2508401Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:33:42.2516558Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:33:42.2519544Z (EngineCore_DP0 pid=13401) INFO 11-03 17:33:42 [parallel_state.py:1231] rank 0 in world size 1 is assigned as DP rank 0, PP rank 0, TP rank 0, EP rank 0 2025-11-03T17:33:42.5903479Z (EngineCore_DP0 pid=13401) INFO 11-03 17:33:42 [topk_topp_sampler.py:59] Using FlashInfer for top-p & top-k sampling. 2025-11-03T17:33:42.6633841Z (EngineCore_DP0 pid=13401) INFO 11-03 17:33:42 [gpu_model_runner.py:2840] Starting to load model deepseek-ai/DeepSeek-V2-Lite-Chat... 2025-11-03T17:33:42.9158193Z (EngineCore_DP0 pid=13401) INFO 11-03 17:33:42 [gpu_model_runner.py:2870] Loading model from scratch... 2025-11-03T17:33:43.0047941Z (EngineCore_DP0 pid=13401) INFO 11-03 17:33:43 [cuda.py:328] Using Triton MLA backend on V1 engine. 2025-11-03T17:33:43.5811343Z (EngineCore_DP0 pid=13401) INFO 11-03 17:33:43 [gpu_model_runner.py:2902] Model loading took 0.8919 GiB and 0.161985 seconds 2025-11-03T17:33:43.5815291Z (EngineCore_DP0 pid=13401) INFO 11-03 17:33:43 [kv_cache_utils.py:1199] GPU KV cache size: 9,320,672 tokens 2025-11-03T17:33:43.5816652Z (EngineCore_DP0 pid=13401) INFO 11-03 17:33:43 [kv_cache_utils.py:1204] Maximum concurrency for 163,840 tokens per request: 56.89x 2025-11-03T17:33:45.0074089Z (EngineCore_DP0 pid=13401) INFO 11-03 17:33:45 [gc_utils.py:40] GC Debug Config. enabled:False,top_objects:-1 2025-11-03T17:33:45.0087498Z INFO 11-03 17:33:45 [llm.py:337] Supported tasks: ['generate'] 2025-11-03T17:33:45.4808497Z PASSED 2025-11-03T17:33:45.4948486Z models/test_initialization.py::test_can_initialize_large_subset[TransformersForSequenceClassification] Fork a new process to run a test 13467 2025-11-03T17:33:45.4959356Z Fork a new process to run a test 0 2025-11-03T17:33:45.4962974Z `transformers==4.56.2` installed, but `transformers>=4.57.0.dev0` is required to run this model. 2025-11-03T17:33:45.8013923Z PASSED 2025-11-03T17:33:45.8152274Z models/test_initialization.py::test_can_initialize_large_subset[OPTForCausalLM] Fork a new process to run a test 13468 2025-11-03T17:33:45.8162016Z Fork a new process to run a test 0 2025-11-03T17:33:45.8441130Z INFO 11-03 17:33:45 [utils.py:239] non-default args: {'load_format': 'dummy', 'gpu_memory_utilization': 0.8, 'disable_log_stats': True, 'hf_overrides': functools.partial(, model_arch='OPTForCausalLM', exist_overrides={}, use_original_num_layers=False), 'model_impl': 'vllm', 'model': 'facebook/opt-125m'} 2025-11-03T17:33:45.9149939Z 2025-11-03T17:33:45.9151571Z config.json: 0% 0.00/651 [00:00, 'use_cudagraph': True, 'cudagraph_num_of_warmups': 1, 'cudagraph_capture_sizes': [512, 504, 496, 488, 480, 472, 464, 456, 448, 440, 432, 424, 416, 408, 400, 392, 384, 376, 368, 360, 352, 344, 336, 328, 320, 312, 304, 296, 288, 280, 272, 264, 256, 248, 240, 232, 224, 216, 208, 200, 192, 184, 176, 168, 160, 152, 144, 136, 128, 120, 112, 104, 96, 88, 80, 72, 64, 56, 48, 40, 32, 24, 16, 8, 4, 2, 1], 'cudagraph_copy_inputs': False, 'full_cuda_graph': True, 'use_inductor_graph_partition': False, 'pass_config': {}, 'max_capture_size': 512, 'local_cache_dir': None} 2025-11-03T17:33:53.6715182Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:33:53.6743422Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:33:53.6751659Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:33:53.6759731Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:33:53.6767719Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:33:53.6775285Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:33:53.6778302Z (EngineCore_DP0 pid=13526) INFO 11-03 17:33:53 [parallel_state.py:1231] rank 0 in world size 1 is assigned as DP rank 0, PP rank 0, TP rank 0, EP rank 0 2025-11-03T17:33:54.0237643Z (EngineCore_DP0 pid=13526) INFO 11-03 17:33:54 [topk_topp_sampler.py:59] Using FlashInfer for top-p & top-k sampling. 2025-11-03T17:33:54.0646474Z (EngineCore_DP0 pid=13526) INFO 11-03 17:33:54 [gpu_model_runner.py:2840] Starting to load model facebook/opt-125m... 2025-11-03T17:33:54.3236302Z (EngineCore_DP0 pid=13526) INFO 11-03 17:33:54 [gpu_model_runner.py:2870] Loading model from scratch... 2025-11-03T17:33:54.3276077Z (EngineCore_DP0 pid=13526) INFO 11-03 17:33:54 [cuda.py:404] Using Flash Attention backend on V1 engine. 2025-11-03T17:33:54.9014818Z (EngineCore_DP0 pid=13526) INFO 11-03 17:33:54 [gpu_model_runner.py:2902] Model loading took 0.0884 GiB and 0.062597 seconds 2025-11-03T17:33:54.9018680Z (EngineCore_DP0 pid=13526) INFO 11-03 17:33:54 [kv_cache_utils.py:1199] GPU KV cache size: 3,495,248 tokens 2025-11-03T17:33:54.9019743Z (EngineCore_DP0 pid=13526) INFO 11-03 17:33:54 [kv_cache_utils.py:1204] Maximum concurrency for 2,048 tokens per request: 1706.66x 2025-11-03T17:33:55.7784980Z (EngineCore_DP0 pid=13526) INFO 11-03 17:33:55 [gc_utils.py:40] GC Debug Config. enabled:False,top_objects:-1 2025-11-03T17:33:55.7800235Z INFO 11-03 17:33:55 [llm.py:337] Supported tasks: ['generate'] 2025-11-03T17:33:56.2032920Z PASSED 2025-11-03T17:33:56.2173448Z models/test_initialization.py::test_can_initialize_large_subset[MiniCPMV] Fork a new process to run a test 13592 2025-11-03T17:33:56.2184751Z Fork a new process to run a test 0 2025-11-03T17:33:56.2459046Z INFO 11-03 17:33:56 [utils.py:239] non-default args: {'trust_remote_code': True, 'load_format': 'dummy', 'gpu_memory_utilization': 0.8, 'disable_log_stats': True, 'hf_overrides': functools.partial(, model_arch='MiniCPMV', exist_overrides={}, use_original_num_layers=False), 'model_impl': 'vllm', 'model': 'openbmb/MiniCPM-Llama3-V-2_5'} 2025-11-03T17:33:56.3457222Z 2025-11-03T17:33:56.3459621Z config.json: 0.00B [00:00, ?B/s] 2025-11-03T17:33:56.3459953Z config.json: 1.39kB [00:00, 7.59MB/s] 2025-11-03T17:33:56.4201837Z 2025-11-03T17:33:56.4203670Z configuration_minicpm.py: 0.00B [00:00, ?B/s] 2025-11-03T17:33:56.4204064Z configuration_minicpm.py: 4.06kB [00:00, 34.2MB/s] 2025-11-03T17:33:56.4299710Z A new version of the following files was downloaded from https://huggingface.co/openbmb/MiniCPM-Llama3-V-2_5: 2025-11-03T17:33:56.4300246Z - configuration_minicpm.py 2025-11-03T17:33:56.4300884Z . Make sure to double-check they do not contain any added malicious code. To avoid downloading new versions of the code file, you can pin a revision. 2025-11-03T17:33:56.9994501Z 2025-11-03T17:33:56.9994996Z preprocessor_config.json: 0% 0.00/599 [00:00, 'use_cudagraph': True, 'cudagraph_num_of_warmups': 1, 'cudagraph_capture_sizes': [512, 504, 496, 488, 480, 472, 464, 456, 448, 440, 432, 424, 416, 408, 400, 392, 384, 376, 368, 360, 352, 344, 336, 328, 320, 312, 304, 296, 288, 280, 272, 264, 256, 248, 240, 232, 224, 216, 208, 200, 192, 184, 176, 168, 160, 152, 144, 136, 128, 120, 112, 104, 96, 88, 80, 72, 64, 56, 48, 40, 32, 24, 16, 8, 4, 2, 1], 'cudagraph_copy_inputs': False, 'full_cuda_graph': True, 'use_inductor_graph_partition': False, 'pass_config': {}, 'max_capture_size': 512, 'local_cache_dir': None} 2025-11-03T17:34:05.2557603Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:34:05.2587096Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:34:05.2594820Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:34:05.2602881Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:34:05.2610652Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:34:05.2618385Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:34:05.2621289Z (EngineCore_DP0 pid=13650) INFO 11-03 17:34:05 [parallel_state.py:1231] rank 0 in world size 1 is assigned as DP rank 0, PP rank 0, TP rank 0, EP rank 0 2025-11-03T17:34:05.6022735Z (EngineCore_DP0 pid=13650) INFO 11-03 17:34:05 [topk_topp_sampler.py:59] Using FlashInfer for top-p & top-k sampling. 2025-11-03T17:34:05.8573057Z (EngineCore_DP0 pid=13650) 2025-11-03T17:34:05.8574295Z processing_minicpmv.py: 0.00B [00:00, ?B/s] 2025-11-03T17:34:05.8574862Z processing_minicpmv.py: 11.9kB [00:00, 43.1MB/s] 2025-11-03T17:34:05.9049390Z (EngineCore_DP0 pid=13650) 2025-11-03T17:34:05.9051422Z image_processing_minicpmv.py: 0.00B [00:00, ?B/s] 2025-11-03T17:34:05.9052080Z image_processing_minicpmv.py: 15.5kB [00:00, 86.9MB/s] 2025-11-03T17:34:05.9885270Z (EngineCore_DP0 pid=13650) Using a slow image processor as `use_fast` is unset and a slow processor was saved with this model. `use_fast=True` will be the default behavior in v4.52, even if the model was saved with a slow processor. This will result in minor differences in outputs. You'll still be able to use a slow processor with `use_fast=False`. 2025-11-03T17:34:08.4681134Z (EngineCore_DP0 pid=13650) INFO 11-03 17:34:08 [gpu_model_runner.py:2840] Starting to load model openbmb/MiniCPM-Llama3-V-2_5... 2025-11-03T17:34:08.7897196Z (EngineCore_DP0 pid=13650) INFO 11-03 17:34:08 [gpu_model_runner.py:2870] Loading model from scratch... 2025-11-03T17:34:08.8139147Z (EngineCore_DP0 pid=13650) INFO 11-03 17:34:08 [cuda.py:404] Using Flash Attention backend on V1 engine. 2025-11-03T17:34:08.8897943Z (EngineCore_DP0 pid=13650) INFO 11-03 17:34:08 [layer.py:497] MultiHeadAttention attn_backend: _Backend.XFORMERS, use_upstream_fa: False 2025-11-03T17:34:09.4822962Z (EngineCore_DP0 pid=13650) INFO 11-03 17:34:09 [gpu_model_runner.py:2902] Model loading took 2.6087 GiB and 0.177520 seconds 2025-11-03T17:34:09.4827975Z (EngineCore_DP0 pid=13650) INFO 11-03 17:34:09 [kv_cache_utils.py:1199] GPU KV cache size: 2,621,440 tokens 2025-11-03T17:34:09.4829302Z (EngineCore_DP0 pid=13650) INFO 11-03 17:34:09 [kv_cache_utils.py:1204] Maximum concurrency for 8,192 tokens per request: 320.00x 2025-11-03T17:34:10.2956072Z (EngineCore_DP0 pid=13650) INFO 11-03 17:34:10 [gc_utils.py:40] GC Debug Config. enabled:False,top_objects:-1 2025-11-03T17:34:10.2972031Z INFO 11-03 17:34:10 [llm.py:337] Supported tasks: ['generate'] 2025-11-03T17:34:10.7769100Z PASSED 2025-11-03T17:34:10.7909404Z models/test_initialization.py::test_can_initialize_large_subset[MiniCPMForCausalLM] Fork a new process to run a test 13732 2025-11-03T17:34:10.7921529Z Fork a new process to run a test 0 2025-11-03T17:34:10.8206763Z INFO 11-03 17:34:10 [utils.py:239] non-default args: {'trust_remote_code': True, 'load_format': 'dummy', 'gpu_memory_utilization': 0.8, 'disable_log_stats': True, 'hf_overrides': functools.partial(, model_arch='MiniCPMForCausalLM', exist_overrides={}, use_original_num_layers=False), 'model_impl': 'vllm', 'model': 'openbmb/MiniCPM-2B-sft-bf16'} 2025-11-03T17:34:10.9492383Z 2025-11-03T17:34:10.9494894Z config.json: 0.00B [00:00, ?B/s] 2025-11-03T17:34:10.9495219Z config.json: 1.01kB [00:00, 4.65MB/s] 2025-11-03T17:34:11.0403738Z 2025-11-03T17:34:11.0406044Z configuration_minicpm.py: 0.00B [00:00, ?B/s] 2025-11-03T17:34:11.0406429Z configuration_minicpm.py: 9.77kB [00:00, 51.0MB/s] 2025-11-03T17:34:11.0506210Z A new version of the following files was downloaded from https://huggingface.co/openbmb/MiniCPM-2B-sft-bf16: 2025-11-03T17:34:11.0506754Z - configuration_minicpm.py 2025-11-03T17:34:11.0507403Z . Make sure to double-check they do not contain any added malicious code. To avoid downloading new versions of the code file, you can pin a revision. 2025-11-03T17:34:17.4960960Z INFO 11-03 17:34:17 [model.py:653] Resolved architecture: MiniCPMForCausalLM 2025-11-03T17:34:17.4961458Z `torch_dtype` is deprecated! Use `dtype` instead! 2025-11-03T17:34:17.5215021Z INFO 11-03 17:34:17 [model.py:1714] Using max model len 4096 2025-11-03T17:34:17.5217085Z The argument `trust_remote_code` is to be used with Auto classes. It has no effect here and is ignored. 2025-11-03T17:34:17.5775486Z INFO 11-03 17:34:17 [scheduler.py:225] Chunked prefill is enabled with max_num_batched_tokens=8192. 2025-11-03T17:34:17.6617012Z 2025-11-03T17:34:17.6618773Z tokenizer_config.json: 0.00B [00:00, ?B/s] 2025-11-03T17:34:17.6619234Z tokenizer_config.json: 1.12kB [00:00, 3.95MB/s] 2025-11-03T17:34:17.8129566Z 2025-11-03T17:34:18.1038076Z tokenizer.model: 0% 0.00/1.99M [00:00, 'use_cudagraph': True, 'cudagraph_num_of_warmups': 1, 'cudagraph_capture_sizes': [512, 504, 496, 488, 480, 472, 464, 456, 448, 440, 432, 424, 416, 408, 400, 392, 384, 376, 368, 360, 352, 344, 336, 328, 320, 312, 304, 296, 288, 280, 272, 264, 256, 248, 240, 232, 224, 216, 208, 200, 192, 184, 176, 168, 160, 152, 144, 136, 128, 120, 112, 104, 96, 88, 80, 72, 64, 56, 48, 40, 32, 24, 16, 8, 4, 2, 1], 'cudagraph_copy_inputs': False, 'full_cuda_graph': True, 'use_inductor_graph_partition': False, 'pass_config': {}, 'max_capture_size': 512, 'local_cache_dir': None} 2025-11-03T17:34:19.3560440Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:34:19.3587737Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:34:19.3594783Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:34:19.3602816Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:34:19.3610124Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:34:19.3618231Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:34:19.3620828Z (EngineCore_DP0 pid=13810) INFO 11-03 17:34:19 [parallel_state.py:1231] rank 0 in world size 1 is assigned as DP rank 0, PP rank 0, TP rank 0, EP rank 0 2025-11-03T17:34:19.7082098Z (EngineCore_DP0 pid=13810) INFO 11-03 17:34:19 [topk_topp_sampler.py:59] Using FlashInfer for top-p & top-k sampling. 2025-11-03T17:34:19.7753613Z (EngineCore_DP0 pid=13810) INFO 11-03 17:34:19 [gpu_model_runner.py:2840] Starting to load model openbmb/MiniCPM-2B-sft-bf16... 2025-11-03T17:34:20.0359995Z (EngineCore_DP0 pid=13810) INFO 11-03 17:34:20 [gpu_model_runner.py:2870] Loading model from scratch... 2025-11-03T17:34:20.1038123Z (EngineCore_DP0 pid=13810) INFO 11-03 17:34:20 [cuda.py:404] Using Flash Attention backend on V1 engine. 2025-11-03T17:34:20.6803034Z (EngineCore_DP0 pid=13810) INFO 11-03 17:34:20 [gpu_model_runner.py:2902] Model loading took 0.6422 GiB and 0.128081 seconds 2025-11-03T17:34:20.6807591Z (EngineCore_DP0 pid=13810) INFO 11-03 17:34:20 [kv_cache_utils.py:1199] GPU KV cache size: 1,165,072 tokens 2025-11-03T17:34:20.6808537Z (EngineCore_DP0 pid=13810) INFO 11-03 17:34:20 [kv_cache_utils.py:1204] Maximum concurrency for 4,096 tokens per request: 284.44x 2025-11-03T17:34:21.4196748Z (EngineCore_DP0 pid=13810) INFO 11-03 17:34:21 [gc_utils.py:40] GC Debug Config. enabled:False,top_objects:-1 2025-11-03T17:34:21.4220653Z INFO 11-03 17:34:21 [llm.py:337] Supported tasks: ['generate'] 2025-11-03T17:34:21.8784040Z PASSED 2025-11-03T17:34:21.8925716Z models/test_initialization.py::test_can_initialize_large_subset[ModernBertModel] Fork a new process to run a test 13876 2025-11-03T17:34:21.8935928Z Fork a new process to run a test 0 2025-11-03T17:34:21.9216754Z INFO 11-03 17:34:21 [utils.py:239] non-default args: {'trust_remote_code': True, 'load_format': 'dummy', 'gpu_memory_utilization': 0.8, 'disable_log_stats': True, 'hf_overrides': functools.partial(, model_arch='ModernBertModel', exist_overrides={}, use_original_num_layers=False), 'model_impl': 'vllm', 'model': 'Alibaba-NLP/gte-modernbert-base'} 2025-11-03T17:34:22.0166096Z 2025-11-03T17:34:22.0167940Z config.json: 0.00B [00:00, ?B/s] 2025-11-03T17:34:22.0168331Z config.json: 1.18kB [00:00, 7.07MB/s] 2025-11-03T17:34:28.4999164Z 2025-11-03T17:34:28.4999599Z modules.json: 0% 0.00/229 [00:00, 'use_cudagraph': True, 'cudagraph_num_of_warmups': 1, 'cudagraph_capture_sizes': [512, 504, 496, 488, 480, 472, 464, 456, 448, 440, 432, 424, 416, 408, 400, 392, 384, 376, 368, 360, 352, 344, 336, 328, 320, 312, 304, 296, 288, 280, 272, 264, 256, 248, 240, 232, 224, 216, 208, 200, 192, 184, 176, 168, 160, 152, 144, 136, 128, 120, 112, 104, 96, 88, 80, 72, 64, 56, 48, 40, 32, 24, 16, 8, 4, 2, 1], 'cudagraph_copy_inputs': False, 'full_cuda_graph': False, 'use_inductor_graph_partition': False, 'pass_config': {}, 'max_capture_size': 512, 'local_cache_dir': None} 2025-11-03T17:34:30.0744310Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:34:30.0775357Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:34:30.0782750Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:34:30.0790504Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:34:30.0798302Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:34:30.0805882Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:34:30.0808649Z (EngineCore_DP0 pid=13935) INFO 11-03 17:34:30 [parallel_state.py:1231] rank 0 in world size 1 is assigned as DP rank 0, PP rank 0, TP rank 0, EP rank 0 2025-11-03T17:34:30.4201565Z (EngineCore_DP0 pid=13935) INFO 11-03 17:34:30 [topk_topp_sampler.py:59] Using FlashInfer for top-p & top-k sampling. 2025-11-03T17:34:30.4409587Z (EngineCore_DP0 pid=13935) INFO 11-03 17:34:30 [gpu_model_runner.py:2840] Starting to load model Alibaba-NLP/gte-modernbert-base... 2025-11-03T17:34:30.6963127Z (EngineCore_DP0 pid=13935) INFO 11-03 17:34:30 [gpu_model_runner.py:2870] Loading model from scratch... 2025-11-03T17:34:30.7642795Z (EngineCore_DP0 pid=13935) INFO 11-03 17:34:30 [cuda.py:404] Using Flash Attention backend on V1 engine. 2025-11-03T17:34:31.3392188Z (EngineCore_DP0 pid=13935) INFO 11-03 17:34:31 [gpu_model_runner.py:2902] Model loading took 0.0826 GiB and 0.128071 seconds 2025-11-03T17:34:31.5273113Z (EngineCore_DP0 pid=13935) INFO 11-03 17:34:31 [core.py:142] Disabling chunked prefill for model without KVCache 2025-11-03T17:34:31.5299760Z (EngineCore_DP0 pid=13935) INFO 11-03 17:34:31 [vllm.py:433] Only "last" pooling supports chunked prefill and prefix caching; disabling both. 2025-11-03T17:34:31.7895942Z (EngineCore_DP0 pid=13935) INFO 11-03 17:34:31 [gc_utils.py:40] GC Debug Config. enabled:False,top_objects:-1 2025-11-03T17:34:31.7909885Z INFO 11-03 17:34:31 [llm.py:337] Supported tasks: ['embed', 'encode'] 2025-11-03T17:34:32.2325555Z PASSED 2025-11-03T17:34:32.2467272Z models/test_initialization.py::test_can_initialize_large_subset[Mamba2ForCausalLM] Fork a new process to run a test 14001 2025-11-03T17:34:32.2478096Z Fork a new process to run a test 0 2025-11-03T17:34:32.2758002Z INFO 11-03 17:34:32 [utils.py:239] non-default args: {'load_format': 'dummy', 'gpu_memory_utilization': 0.8, 'disable_log_stats': True, 'hf_overrides': functools.partial(, model_arch='Mamba2ForCausalLM', exist_overrides={}, use_original_num_layers=False), 'model_impl': 'vllm', 'model': 'mistralai/Mamba-Codestral-7B-v0.1'} 2025-11-03T17:34:32.3893393Z 2025-11-03T17:34:32.3895603Z config.json: 0.00B [00:00, ?B/s] 2025-11-03T17:34:32.3895918Z config.json: 1.01kB [00:00, 5.62MB/s] 2025-11-03T17:34:38.7742310Z INFO 11-03 17:34:38 [model.py:653] Resolved architecture: Mamba2ForCausalLM 2025-11-03T17:34:38.7742850Z `torch_dtype` is deprecated! Use `dtype` instead! 2025-11-03T17:34:38.7994435Z WARNING 11-03 17:34:38 [model.py:2042] The model's config.json does not contain any of the following keys to determine the original maximum length of the model: ['max_position_embeddings', 'n_positions', 'max_seq_len', 'seq_length', 'model_max_length', 'max_target_positions', 'max_sequence_length', 'max_seq_length', 'seq_len']. Assuming the model's maximum length is 2048. 2025-11-03T17:34:38.7995786Z INFO 11-03 17:34:38 [model.py:1714] Using max model len 2048 2025-11-03T17:34:38.9680955Z INFO 11-03 17:34:38 [scheduler.py:225] Chunked prefill is enabled with max_num_batched_tokens=8192. 2025-11-03T17:34:38.9681960Z INFO 11-03 17:34:38 [config.py:300] Warning: Prefix caching is currently enabled. Its support for Mamba2 layers is experimental. Please report any issues you may observe. 2025-11-03T17:34:38.9682855Z INFO 11-03 17:34:38 [config.py:313] Disabling cascade attention since it is not supported for hybrid models. 2025-11-03T17:34:39.0093810Z 2025-11-03T17:34:39.0100170Z tokenizer_config.json: 0.00B [00:00, ?B/s] 2025-11-03T17:34:39.0100567Z tokenizer_config.json: 137kB [00:00, 197MB/s] 2025-11-03T17:34:39.1208471Z 2025-11-03T17:34:39.3358268Z tokenizer.model: 0% 0.00/588k [00:00, 'use_cudagraph': True, 'cudagraph_num_of_warmups': 1, 'cudagraph_capture_sizes': [512, 504, 496, 488, 480, 472, 464, 456, 448, 440, 432, 424, 416, 408, 400, 392, 384, 376, 368, 360, 352, 344, 336, 328, 320, 312, 304, 296, 288, 280, 272, 264, 256, 248, 240, 232, 224, 216, 208, 200, 192, 184, 176, 168, 160, 152, 144, 136, 128, 120, 112, 104, 96, 88, 80, 72, 64, 56, 48, 40, 32, 24, 16, 8, 4, 2, 1], 'cudagraph_copy_inputs': False, 'full_cuda_graph': True, 'use_inductor_graph_partition': False, 'pass_config': {}, 'max_capture_size': 512, 'local_cache_dir': None} 2025-11-03T17:34:40.5012773Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:34:40.5040815Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:34:40.5048114Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:34:40.5055978Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:34:40.5064083Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:34:40.5071824Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:34:40.5074854Z (EngineCore_DP0 pid=14079) INFO 11-03 17:34:40 [parallel_state.py:1231] rank 0 in world size 1 is assigned as DP rank 0, PP rank 0, TP rank 0, EP rank 0 2025-11-03T17:34:40.8456153Z (EngineCore_DP0 pid=14079) INFO 11-03 17:34:40 [topk_topp_sampler.py:59] Using FlashInfer for top-p & top-k sampling. 2025-11-03T17:34:41.0485797Z (EngineCore_DP0 pid=14079) INFO 11-03 17:34:41 [gpu_model_runner.py:2840] Starting to load model mistralai/Mamba-Codestral-7B-v0.1... 2025-11-03T17:34:41.3117283Z (EngineCore_DP0 pid=14079) INFO 11-03 17:34:41 [gpu_model_runner.py:2870] Loading model from scratch... 2025-11-03T17:34:41.8818617Z (EngineCore_DP0 pid=14079) INFO 11-03 17:34:41 [gpu_model_runner.py:2902] Model loading took 0.7052 GiB and 0.060704 seconds 2025-11-03T17:34:41.8824341Z (EngineCore_DP0 pid=14079) INFO 11-03 17:34:41 [kv_cache_utils.py:1199] GPU KV cache size: 10,186,752 tokens 2025-11-03T17:34:41.8825320Z (EngineCore_DP0 pid=14079) INFO 11-03 17:34:41 [kv_cache_utils.py:1204] Maximum concurrency for 2,048 tokens per request: 4974.00x 2025-11-03T17:34:42.5610273Z (EngineCore_DP0 pid=14079) INFO 11-03 17:34:42 [gc_utils.py:40] GC Debug Config. enabled:False,top_objects:-1 2025-11-03T17:34:42.5634645Z INFO 11-03 17:34:42 [llm.py:337] Supported tasks: ['generate'] 2025-11-03T17:34:43.0022637Z PASSED 2025-11-03T17:34:43.0166207Z models/test_initialization.py::test_can_initialize_large_subset[MiniMaxVL01ForConditionalGeneration] Fork a new process to run a test 14145 2025-11-03T17:34:43.0176648Z Fork a new process to run a test 0 2025-11-03T17:34:43.3045078Z PASSED 2025-11-03T17:34:43.3186245Z models/test_initialization.py::test_can_initialize_large_subset[Gemma3ForConditionalGeneration] Fork a new process to run a test 14146 2025-11-03T17:34:43.3197060Z Fork a new process to run a test 0 2025-11-03T17:34:43.3480275Z INFO 11-03 17:34:43 [utils.py:239] non-default args: {'load_format': 'dummy', 'gpu_memory_utilization': 0.8, 'disable_log_stats': True, 'hf_overrides': functools.partial(, model_arch='Gemma3ForConditionalGeneration', exist_overrides={}, use_original_num_layers=False), 'model_impl': 'vllm', 'model': 'google/gemma-3-4b-it'} 2025-11-03T17:34:43.4536907Z 2025-11-03T17:34:43.4538407Z config.json: 0% 0.00/855 [00:00, 'use_cudagraph': True, 'cudagraph_num_of_warmups': 1, 'cudagraph_capture_sizes': [512, 504, 496, 488, 480, 472, 464, 456, 448, 440, 432, 424, 416, 408, 400, 392, 384, 376, 368, 360, 352, 344, 336, 328, 320, 312, 304, 296, 288, 280, 272, 264, 256, 248, 240, 232, 224, 216, 208, 200, 192, 184, 176, 168, 160, 152, 144, 136, 128, 120, 112, 104, 96, 88, 80, 72, 64, 56, 48, 40, 32, 24, 16, 8, 4, 2, 1], 'cudagraph_copy_inputs': False, 'full_cuda_graph': True, 'use_inductor_graph_partition': False, 'pass_config': {}, 'max_capture_size': 512, 'local_cache_dir': None} 2025-11-03T17:34:55.0675731Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:34:55.0705151Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:34:55.0712597Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:34:55.0720343Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:34:55.0728634Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:34:55.0736485Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:34:55.0739113Z (EngineCore_DP0 pid=14225) INFO 11-03 17:34:55 [parallel_state.py:1231] rank 0 in world size 1 is assigned as DP rank 0, PP rank 0, TP rank 0, EP rank 0 2025-11-03T17:34:55.4089038Z (EngineCore_DP0 pid=14225) INFO 11-03 17:34:55 [topk_topp_sampler.py:59] Using FlashInfer for top-p & top-k sampling. 2025-11-03T17:34:55.6062280Z (EngineCore_DP0 pid=14225) 2025-11-03T17:34:55.6063065Z processor_config.json: 0% 0.00/70.0 [00:00, model_arch='GraniteMoeHybridForCausalLM', exist_overrides={}, use_original_num_layers=False), 'model_impl': 'vllm', 'model': 'ibm-granite/granite-4.0-tiny-preview'} 2025-11-03T17:35:05.4562514Z 2025-11-03T17:35:05.4564579Z config.json: 0.00B [00:00, ?B/s] 2025-11-03T17:35:05.4564887Z config.json: 1.80kB [00:00, 9.90MB/s] 2025-11-03T17:35:11.8818471Z INFO 11-03 17:35:11 [model.py:653] Resolved architecture: GraniteMoeHybridForCausalLM 2025-11-03T17:35:11.8818997Z `torch_dtype` is deprecated! Use `dtype` instead! 2025-11-03T17:35:22.0232751Z ERROR 11-03 17:35:22 [config.py:298] Error retrieving safetensors: (ReadTimeoutError("HTTPSConnectionPool(host='huggingface.co', port=443): Read timed out. (read timeout=10)"), '(Request ID: aac3e1c7-1cef-4ae5-a43f-947353adb0e0)'), retrying 1 of 2 2025-11-03T17:35:24.1429819Z 2025-11-03T17:35:24.1433137Z model.safetensors.index.json: 0.00B [00:00, ?B/s] 2025-11-03T17:35:24.1433570Z model.safetensors.index.json: 48.9kB [00:00, 89.9MB/s] 2025-11-03T17:35:24.1463894Z 2025-11-03T17:35:24.1996699Z Parse safetensors files: 0% 0/3 [00:00= mamba page size. 2025-11-03T17:35:24.5039278Z INFO 11-03 17:35:24 [config.py:453] Padding mamba page size by 1.59% to ensure that mamba page size and attention page size are exactly equal. 2025-11-03T17:35:24.6480227Z 2025-11-03T17:35:24.6482712Z tokenizer_config.json: 0.00B [00:00, ?B/s] 2025-11-03T17:35:24.6483256Z tokenizer_config.json: 10.1kB [00:00, 46.0MB/s] 2025-11-03T17:35:24.9070454Z 2025-11-03T17:35:25.0144121Z tokenizer.json: 0.00B [00:00, ?B/s] 2025-11-03T17:35:25.0252989Z tokenizer.json: 2.76MB [00:00, 25.7MB/s] 2025-11-03T17:35:25.0253436Z tokenizer.json: 3.48MB [00:00, 29.4MB/s] 2025-11-03T17:35:25.1433606Z 2025-11-03T17:35:25.1434636Z special_tokens_map.json: 0% 0.00/801 [00:00, 'use_cudagraph': True, 'cudagraph_num_of_warmups': 1, 'cudagraph_capture_sizes': [512, 504, 496, 488, 480, 472, 464, 456, 448, 440, 432, 424, 416, 408, 400, 392, 384, 376, 368, 360, 352, 344, 336, 328, 320, 312, 304, 296, 288, 280, 272, 264, 256, 248, 240, 232, 224, 216, 208, 200, 192, 184, 176, 168, 160, 152, 144, 136, 128, 120, 112, 104, 96, 88, 80, 72, 64, 56, 48, 40, 32, 24, 16, 8, 4, 2, 1], 'cudagraph_copy_inputs': False, 'full_cuda_graph': True, 'use_inductor_graph_partition': False, 'pass_config': {}, 'max_capture_size': 512, 'local_cache_dir': None} 2025-11-03T17:35:26.0787306Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:35:26.0814822Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:35:26.0822921Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:35:26.0830540Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:35:26.0838418Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:35:26.0845843Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:35:26.0848750Z (EngineCore_DP0 pid=14368) INFO 11-03 17:35:26 [parallel_state.py:1231] rank 0 in world size 1 is assigned as DP rank 0, PP rank 0, TP rank 0, EP rank 0 2025-11-03T17:35:26.4321619Z (EngineCore_DP0 pid=14368) INFO 11-03 17:35:26 [topk_topp_sampler.py:59] Using FlashInfer for top-p & top-k sampling. 2025-11-03T17:35:26.4922679Z (EngineCore_DP0 pid=14368) INFO 11-03 17:35:26 [gpu_model_runner.py:2840] Starting to load model ibm-granite/granite-4.0-tiny-preview... 2025-11-03T17:35:26.7620839Z (EngineCore_DP0 pid=14368) INFO 11-03 17:35:26 [gpu_model_runner.py:2870] Loading model from scratch... 2025-11-03T17:35:27.3475545Z (EngineCore_DP0 pid=14368) INFO 11-03 17:35:27 [gpu_model_runner.py:2902] Model loading took 0.1857 GiB and 0.064259 seconds 2025-11-03T17:35:27.3481130Z (EngineCore_DP0 pid=14368) INFO 11-03 17:35:27 [kv_cache_utils.py:1199] GPU KV cache size: 1,717,960,704 tokens 2025-11-03T17:35:27.3482111Z (EngineCore_DP0 pid=14368) INFO 11-03 17:35:27 [kv_cache_utils.py:1204] Maximum concurrency for 131,072 tokens per request: 13107.00x 2025-11-03T17:35:27.8911225Z (EngineCore_DP0 pid=14368) INFO 11-03 17:35:27 [gc_utils.py:40] GC Debug Config. enabled:False,top_objects:-1 2025-11-03T17:35:27.8933900Z INFO 11-03 17:35:27 [llm.py:337] Supported tasks: ['generate'] 2025-11-03T17:35:28.3412110Z PASSED 2025-11-03T17:35:28.3552840Z models/test_initialization.py::test_can_initialize_large_subset[Grok1ModelForCausalLM] Fork a new process to run a test 14434 2025-11-03T17:35:28.3563634Z Fork a new process to run a test 0 2025-11-03T17:35:28.3838109Z INFO 11-03 17:35:28 [utils.py:239] non-default args: {'trust_remote_code': True, 'load_format': 'dummy', 'gpu_memory_utilization': 0.8, 'disable_log_stats': True, 'hf_overrides': functools.partial(, model_arch='Grok1ModelForCausalLM', exist_overrides={}, use_original_num_layers=False), 'model_impl': 'vllm', 'model': 'hpcai-tech/grok-1'} 2025-11-03T17:35:28.5594884Z 2025-11-03T17:35:28.5596807Z config.json: 0.00B [00:00, ?B/s] 2025-11-03T17:35:28.5597157Z config.json: 1.00kB [00:00, 5.65MB/s] 2025-11-03T17:35:28.6557624Z 2025-11-03T17:35:28.6558929Z configuration_grok1.py: 0.00B [00:00, ?B/s] 2025-11-03T17:35:28.6559313Z configuration_grok1.py: 2.13kB [00:00, 18.4MB/s] 2025-11-03T17:35:28.6649170Z A new version of the following files was downloaded from https://huggingface.co/hpcai-tech/grok-1: 2025-11-03T17:35:28.6649699Z - configuration_grok1.py 2025-11-03T17:35:28.6650331Z . Make sure to double-check they do not contain any added malicious code. To avoid downloading new versions of the code file, you can pin a revision. 2025-11-03T17:35:35.1595208Z INFO 11-03 17:35:35 [model.py:653] Resolved architecture: Grok1ModelForCausalLM 2025-11-03T17:35:35.1595712Z `torch_dtype` is deprecated! Use `dtype` instead! 2025-11-03T17:35:35.1848349Z INFO 11-03 17:35:35 [model.py:1714] Using max model len 8192 2025-11-03T17:35:35.1850553Z The argument `trust_remote_code` is to be used with Auto classes. It has no effect here and is ignored. 2025-11-03T17:35:35.2552587Z INFO 11-03 17:35:35 [scheduler.py:225] Chunked prefill is enabled with max_num_batched_tokens=8192. 2025-11-03T17:35:35.2951659Z 2025-11-03T17:35:35.2953323Z tokenizer_config.json: 0.00B [00:00, ?B/s] 2025-11-03T17:35:35.2953751Z tokenizer_config.json: 1.02kB [00:00, 8.31MB/s] 2025-11-03T17:35:35.4816295Z 2025-11-03T17:35:35.5174621Z tokenizer.json: 0.00B [00:00, ?B/s] 2025-11-03T17:35:35.5174980Z tokenizer.json: 9.14MB [00:00, 255MB/s] 2025-11-03T17:35:35.5912060Z 2025-11-03T17:35:35.5912904Z special_tokens_map.json: 0% 0.00/555 [00:00. This is expected, and simply means that the `legacy` (previous) behavior will be used so nothing changes for you. If you want to use the new behaviour, set `legacy=False`. This should only be set if you understand what it means, and thoroughly read the reason why this was added as explained in https://github.com/huggingface/transformers/pull/24565 - if you loaded a llama tokenizer from a GGUF file you can ignore this message. 2025-11-03T17:35:36.2201918Z (EngineCore_DP0 pid=14492) INFO 11-03 17:35:36 [core.py:727] Waiting for init message from front-end. 2025-11-03T17:35:36.2239205Z (EngineCore_DP0 pid=14492) INFO 11-03 17:35:36 [core.py:94] Initializing a V1 LLM engine (v0.11.0rc2.dev389+ge51928192.d20251103) with config: model='hpcai-tech/grok-1', speculative_config=None, tokenizer='hpcai-tech/grok-1', skip_tokenizer_init=False, tokenizer_mode=auto, revision=None, tokenizer_revision=None, trust_remote_code=True, dtype=torch.bfloat16, max_seq_len=8192, download_dir=None, load_format=dummy, tensor_parallel_size=1, pipeline_parallel_size=1, data_parallel_size=1, disable_custom_all_reduce=False, quantization=None, enforce_eager=False, kv_cache_dtype=auto, device_config=cuda, structured_outputs_config=StructuredOutputsConfig(backend='auto', disable_fallback=False, disable_any_whitespace=False, disable_additional_properties=False, reasoning_parser=''), observability_config=ObservabilityConfig(show_hidden_metrics_for_version=None, otlp_traces_endpoint=None, collect_detailed_traces=None), seed=0, served_model_name=hpcai-tech/grok-1, enable_prefix_caching=True, chunked_prefill_enabled=True, pooler_config=None, compilation_config={'level': 3, 'debug_dump_path': None, 'cache_dir': '', 'backend': '', 'custom_ops': [], 'splitting_ops': ['vllm.unified_attention', 'vllm.unified_attention_with_output', 'vllm.unified_mla_attention', 'vllm.unified_mla_attention_with_output', 'vllm.mamba_mixer2', 'vllm.mamba_mixer', 'vllm.short_conv', 'vllm.linear_attention', 'vllm.plamo2_mamba_mixer', 'vllm.gdn_attention', 'vllm.sparse_attn_indexer'], 'use_inductor': True, 'compile_sizes': [], 'inductor_compile_config': {'enable_auto_functionalized_v2': False}, 'inductor_passes': {}, 'cudagraph_mode': , 'use_cudagraph': True, 'cudagraph_num_of_warmups': 1, 'cudagraph_capture_sizes': [512, 504, 496, 488, 480, 472, 464, 456, 448, 440, 432, 424, 416, 408, 400, 392, 384, 376, 368, 360, 352, 344, 336, 328, 320, 312, 304, 296, 288, 280, 272, 264, 256, 248, 240, 232, 224, 216, 208, 200, 192, 184, 176, 168, 160, 152, 144, 136, 128, 120, 112, 104, 96, 88, 80, 72, 64, 56, 48, 40, 32, 24, 16, 8, 4, 2, 1], 'cudagraph_copy_inputs': False, 'full_cuda_graph': True, 'use_inductor_graph_partition': False, 'pass_config': {}, 'max_capture_size': 512, 'local_cache_dir': None} 2025-11-03T17:35:36.9263087Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:35:36.9290342Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:35:36.9298597Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:35:36.9306967Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:35:36.9315287Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:35:36.9323978Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:35:36.9326729Z (EngineCore_DP0 pid=14492) INFO 11-03 17:35:36 [parallel_state.py:1231] rank 0 in world size 1 is assigned as DP rank 0, PP rank 0, TP rank 0, EP rank 0 2025-11-03T17:35:37.2690743Z (EngineCore_DP0 pid=14492) INFO 11-03 17:35:37 [topk_topp_sampler.py:59] Using FlashInfer for top-p & top-k sampling. 2025-11-03T17:35:37.3684272Z (EngineCore_DP0 pid=14492) INFO 11-03 17:35:37 [gpu_model_runner.py:2840] Starting to load model hpcai-tech/grok-1... 2025-11-03T17:35:37.6285665Z (EngineCore_DP0 pid=14492) INFO 11-03 17:35:37 [gpu_model_runner.py:2870] Loading model from scratch... 2025-11-03T17:35:37.6951393Z (EngineCore_DP0 pid=14492) INFO 11-03 17:35:37 [cuda.py:404] Using Flash Attention backend on V1 engine. 2025-11-03T17:35:38.2683246Z (EngineCore_DP0 pid=14492) INFO 11-03 17:35:38 [gpu_model_runner.py:2902] Model loading took 3.9161 GiB and 0.125809 seconds 2025-11-03T17:35:38.2687996Z (EngineCore_DP0 pid=14492) INFO 11-03 17:35:38 [kv_cache_utils.py:1199] GPU KV cache size: 2,621,440 tokens 2025-11-03T17:35:38.2689323Z (EngineCore_DP0 pid=14492) INFO 11-03 17:35:38 [kv_cache_utils.py:1204] Maximum concurrency for 8,192 tokens per request: 320.00x 2025-11-03T17:35:39.0379222Z (EngineCore_DP0 pid=14492) INFO 11-03 17:35:39 [gc_utils.py:40] GC Debug Config. enabled:False,top_objects:-1 2025-11-03T17:35:39.0396021Z INFO 11-03 17:35:39 [llm.py:337] Supported tasks: ['generate'] 2025-11-03T17:35:39.5170834Z PASSED 2025-11-03T17:35:39.5312582Z models/test_initialization.py::test_can_initialize_large_subset[GteNewModel] Fork a new process to run a test 14558 2025-11-03T17:35:39.5323313Z Fork a new process to run a test 0 2025-11-03T17:35:39.5608379Z INFO 11-03 17:35:39 [utils.py:239] non-default args: {'trust_remote_code': True, 'load_format': 'dummy', 'gpu_memory_utilization': 0.8, 'disable_log_stats': True, 'hf_overrides': functools.partial(, model_arch='GteNewModel', exist_overrides={'architectures': ['GteNewModel']}, use_original_num_layers=False), 'model_impl': 'vllm', 'model': 'Alibaba-NLP/gte-base-en-v1.5'} 2025-11-03T17:35:39.6405997Z 2025-11-03T17:35:39.6408178Z config.json: 0.00B [00:00, ?B/s] 2025-11-03T17:35:39.6408488Z config.json: 1.35kB [00:00, 7.02MB/s] 2025-11-03T17:35:39.7260313Z INFO 11-03 17:35:39 [config.py:408] Replacing legacy 'type' key with 'rope_type' 2025-11-03T17:35:39.9007535Z 2025-11-03T17:35:39.9008585Z sentence_bert_config.json: 0% 0.00/54.0 [00:00, 'use_cudagraph': True, 'cudagraph_num_of_warmups': 1, 'cudagraph_capture_sizes': [512, 504, 496, 488, 480, 472, 464, 456, 448, 440, 432, 424, 416, 408, 400, 392, 384, 376, 368, 360, 352, 344, 336, 328, 320, 312, 304, 296, 288, 280, 272, 264, 256, 248, 240, 232, 224, 216, 208, 200, 192, 184, 176, 168, 160, 152, 144, 136, 128, 120, 112, 104, 96, 88, 80, 72, 64, 56, 48, 40, 32, 24, 16, 8, 4, 2, 1], 'cudagraph_copy_inputs': False, 'full_cuda_graph': False, 'use_inductor_graph_partition': False, 'pass_config': {}, 'max_capture_size': 512, 'local_cache_dir': None} 2025-11-03T17:35:47.6235963Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:35:47.6263417Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:35:47.6270270Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:35:47.6277918Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:35:47.6285597Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:35:47.6292840Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:35:47.6295500Z (EngineCore_DP0 pid=14617) INFO 11-03 17:35:47 [parallel_state.py:1231] rank 0 in world size 1 is assigned as DP rank 0, PP rank 0, TP rank 0, EP rank 0 2025-11-03T17:35:47.9674401Z (EngineCore_DP0 pid=14617) INFO 11-03 17:35:47 [topk_topp_sampler.py:59] Using FlashInfer for top-p & top-k sampling. 2025-11-03T17:35:47.9901019Z (EngineCore_DP0 pid=14617) INFO 11-03 17:35:47 [gpu_model_runner.py:2840] Starting to load model Alibaba-NLP/gte-base-en-v1.5... 2025-11-03T17:35:48.2475189Z (EngineCore_DP0 pid=14617) INFO 11-03 17:35:48 [gpu_model_runner.py:2870] Loading model from scratch... 2025-11-03T17:35:48.3147276Z (EngineCore_DP0 pid=14617) INFO 11-03 17:35:48 [cuda.py:404] Using Flash Attention backend on V1 engine. 2025-11-03T17:35:48.8863906Z (EngineCore_DP0 pid=14617) INFO 11-03 17:35:48 [gpu_model_runner.py:2902] Model loading took 0.0624 GiB and 0.127603 seconds 2025-11-03T17:35:48.9965882Z (EngineCore_DP0 pid=14617) INFO 11-03 17:35:48 [core.py:142] Disabling chunked prefill for model without KVCache 2025-11-03T17:35:48.9992294Z (EngineCore_DP0 pid=14617) INFO 11-03 17:35:48 [vllm.py:433] Only "last" pooling supports chunked prefill and prefix caching; disabling both. 2025-11-03T17:35:49.2605658Z (EngineCore_DP0 pid=14617) INFO 11-03 17:35:49 [gc_utils.py:40] GC Debug Config. enabled:False,top_objects:-1 2025-11-03T17:35:49.2620836Z INFO 11-03 17:35:49 [llm.py:337] Supported tasks: ['embed', 'encode'] 2025-11-03T17:35:49.6972718Z PASSED 2025-11-03T17:35:49.7113469Z models/test_initialization.py::test_can_initialize_large_subset[XverseForCausalLM] Fork a new process to run a test 14683 2025-11-03T17:35:49.7125706Z Fork a new process to run a test 0 2025-11-03T17:35:49.7406258Z INFO 11-03 17:35:49 [utils.py:239] non-default args: {'tokenizer': 'meta-llama/Llama-2-7b', 'trust_remote_code': True, 'load_format': 'dummy', 'gpu_memory_utilization': 0.8, 'disable_log_stats': True, 'hf_overrides': functools.partial(, model_arch='XverseForCausalLM', exist_overrides={}, use_original_num_layers=False), 'model_impl': 'vllm', 'model': 'xverse/XVERSE-7B-Chat'} 2025-11-03T17:35:49.9402800Z 2025-11-03T17:35:49.9404231Z config.json: 0% 0.00/687 [00:00. This is expected, and simply means that the `legacy` (previous) behavior will be used so nothing changes for you. If you want to use the new behaviour, set `legacy=False`. This should only be set if you understand what it means, and thoroughly read the reason why this was added as explained in https://github.com/huggingface/transformers/pull/24565 - if you loaded a llama tokenizer from a GGUF file you can ignore this message 2025-11-03T17:35:50.7890191Z You are using the default legacy behaviour of the . This is expected, and simply means that the `legacy` (previous) behavior will be used so nothing changes for you. If you want to use the new behaviour, set `legacy=False`. This should only be set if you understand what it means, and thoroughly read the reason why this was added as explained in https://github.com/huggingface/transformers/pull/24565 - if you loaded a llama tokenizer from a GGUF file you can ignore this message. 2025-11-03T17:35:51.0585868Z 2025-11-03T17:35:51.0586543Z generation_config.json: 0% 0.00/232 [00:00, 'use_cudagraph': True, 'cudagraph_num_of_warmups': 1, 'cudagraph_capture_sizes': [512, 504, 496, 488, 480, 472, 464, 456, 448, 440, 432, 424, 416, 408, 400, 392, 384, 376, 368, 360, 352, 344, 336, 328, 320, 312, 304, 296, 288, 280, 272, 264, 256, 248, 240, 232, 224, 216, 208, 200, 192, 184, 176, 168, 160, 152, 144, 136, 128, 120, 112, 104, 96, 88, 80, 72, 64, 56, 48, 40, 32, 24, 16, 8, 4, 2, 1], 'cudagraph_copy_inputs': False, 'full_cuda_graph': True, 'use_inductor_graph_partition': False, 'pass_config': {}, 'max_capture_size': 512, 'local_cache_dir': None} 2025-11-03T17:35:51.7510765Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:35:51.7539000Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:35:51.7546885Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:35:51.7554956Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:35:51.7562831Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:35:51.7571148Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:35:51.7573967Z (EngineCore_DP0 pid=14711) INFO 11-03 17:35:51 [parallel_state.py:1231] rank 0 in world size 1 is assigned as DP rank 0, PP rank 0, TP rank 0, EP rank 0 2025-11-03T17:35:52.0934824Z (EngineCore_DP0 pid=14711) INFO 11-03 17:35:52 [topk_topp_sampler.py:59] Using FlashInfer for top-p & top-k sampling. 2025-11-03T17:35:52.1626220Z (EngineCore_DP0 pid=14711) INFO 11-03 17:35:52 [gpu_model_runner.py:2840] Starting to load model xverse/XVERSE-7B-Chat... 2025-11-03T17:35:52.4172712Z (EngineCore_DP0 pid=14711) INFO 11-03 17:35:52 [gpu_model_runner.py:2870] Loading model from scratch... 2025-11-03T17:35:52.4843365Z (EngineCore_DP0 pid=14711) INFO 11-03 17:35:52 [cuda.py:404] Using Flash Attention backend on V1 engine. 2025-11-03T17:35:53.0497636Z (EngineCore_DP0 pid=14711) INFO 11-03 17:35:53 [gpu_model_runner.py:2902] Model loading took 1.9141 GiB and 0.127060 seconds 2025-11-03T17:35:53.0502485Z (EngineCore_DP0 pid=14711) INFO 11-03 17:35:53 [kv_cache_utils.py:1199] GPU KV cache size: 655,360 tokens 2025-11-03T17:35:53.0503244Z (EngineCore_DP0 pid=14711) INFO 11-03 17:35:53 [kv_cache_utils.py:1204] Maximum concurrency for 8,192 tokens per request: 80.00x 2025-11-03T17:35:53.8423270Z (EngineCore_DP0 pid=14711) INFO 11-03 17:35:53 [gc_utils.py:40] GC Debug Config. enabled:False,top_objects:-1 2025-11-03T17:35:53.8439369Z INFO 11-03 17:35:53 [llm.py:337] Supported tasks: ['generate'] 2025-11-03T17:35:54.2972966Z PASSED 2025-11-03T17:35:54.3114770Z models/test_initialization.py::test_can_initialize_large_subset[MixtralForCausalLM] Fork a new process to run a test 14777 2025-11-03T17:35:54.3125318Z Fork a new process to run a test 0 2025-11-03T17:35:54.3402491Z INFO 11-03 17:35:54 [utils.py:239] non-default args: {'load_format': 'dummy', 'gpu_memory_utilization': 0.8, 'disable_log_stats': True, 'hf_overrides': functools.partial(, model_arch='MixtralForCausalLM', exist_overrides={}, use_original_num_layers=False), 'model_impl': 'vllm', 'model': 'mistralai/Mixtral-8x7B-Instruct-v0.1'} 2025-11-03T17:35:54.4166678Z 2025-11-03T17:35:54.4168669Z config.json: 0% 0.00/720 [00:00, 'use_cudagraph': True, 'cudagraph_num_of_warmups': 1, 'cudagraph_capture_sizes': [512, 504, 496, 488, 480, 472, 464, 456, 448, 440, 432, 424, 416, 408, 400, 392, 384, 376, 368, 360, 352, 344, 336, 328, 320, 312, 304, 296, 288, 280, 272, 264, 256, 248, 240, 232, 224, 216, 208, 200, 192, 184, 176, 168, 160, 152, 144, 136, 128, 120, 112, 104, 96, 88, 80, 72, 64, 56, 48, 40, 32, 24, 16, 8, 4, 2, 1], 'cudagraph_copy_inputs': False, 'full_cuda_graph': True, 'use_inductor_graph_partition': False, 'pass_config': {}, 'max_capture_size': 512, 'local_cache_dir': None} 2025-11-03T17:36:02.2421704Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:36:02.2448924Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:36:02.2456256Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:36:02.2464368Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:36:02.2472217Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:36:02.2479921Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:36:02.2482931Z (EngineCore_DP0 pid=14855) INFO 11-03 17:36:02 [parallel_state.py:1231] rank 0 in world size 1 is assigned as DP rank 0, PP rank 0, TP rank 0, EP rank 0 2025-11-03T17:36:02.5836094Z (EngineCore_DP0 pid=14855) INFO 11-03 17:36:02 [topk_topp_sampler.py:59] Using FlashInfer for top-p & top-k sampling. 2025-11-03T17:36:02.6531347Z (EngineCore_DP0 pid=14855) INFO 11-03 17:36:02 [gpu_model_runner.py:2840] Starting to load model mistralai/Mixtral-8x7B-Instruct-v0.1... 2025-11-03T17:36:02.9086126Z (EngineCore_DP0 pid=14855) INFO 11-03 17:36:02 [gpu_model_runner.py:2870] Loading model from scratch... 2025-11-03T17:36:02.9759877Z (EngineCore_DP0 pid=14855) INFO 11-03 17:36:02 [cuda.py:404] Using Flash Attention backend on V1 engine. 2025-11-03T17:36:03.5468727Z (EngineCore_DP0 pid=14855) INFO 11-03 17:36:03 [gpu_model_runner.py:2902] Model loading took 1.2305 GiB and 0.127949 seconds 2025-11-03T17:36:03.5473079Z (EngineCore_DP0 pid=14855) INFO 11-03 17:36:03 [kv_cache_utils.py:1199] GPU KV cache size: 2,621,440 tokens 2025-11-03T17:36:03.5473954Z (EngineCore_DP0 pid=14855) INFO 11-03 17:36:03 [kv_cache_utils.py:1204] Maximum concurrency for 32,768 tokens per request: 80.00x 2025-11-03T17:36:04.0914262Z (EngineCore_DP0 pid=14855) INFO 11-03 17:36:04 [gc_utils.py:40] GC Debug Config. enabled:False,top_objects:-1 2025-11-03T17:36:04.0935581Z INFO 11-03 17:36:04 [llm.py:337] Supported tasks: ['generate'] 2025-11-03T17:36:04.5450282Z PASSED 2025-11-03T17:36:04.5591076Z models/test_initialization.py::test_can_initialize_large_subset[NVLM_D] Fork a new process to run a test 14921 2025-11-03T17:36:04.5601793Z Fork a new process to run a test 0 2025-11-03T17:36:04.5882709Z INFO 11-03 17:36:04 [utils.py:239] non-default args: {'trust_remote_code': True, 'load_format': 'dummy', 'gpu_memory_utilization': 0.8, 'disable_log_stats': True, 'hf_overrides': functools.partial(, model_arch='NVLM_D', exist_overrides={}, use_original_num_layers=False), 'model_impl': 'vllm', 'model': 'nvidia/NVLM-D-72B'} 2025-11-03T17:36:04.7053289Z 2025-11-03T17:36:04.7055476Z config.json: 0.00B [00:00, ?B/s] 2025-11-03T17:36:04.7055804Z config.json: 3.71kB [00:00, 17.6MB/s] 2025-11-03T17:36:04.7786309Z 2025-11-03T17:36:04.7787813Z configuration_nvlm_d.py: 0.00B [00:00, ?B/s] 2025-11-03T17:36:04.7788184Z configuration_nvlm_d.py: 3.80kB [00:00, 30.8MB/s] 2025-11-03T17:36:04.8226730Z 2025-11-03T17:36:04.8228544Z configuration_intern_vit.py: 0.00B [00:00, ?B/s] 2025-11-03T17:36:04.8229191Z configuration_intern_vit.py: 5.55kB [00:00, 37.6MB/s] 2025-11-03T17:36:04.8249810Z A new version of the following files was downloaded from https://huggingface.co/nvidia/NVLM-D-72B: 2025-11-03T17:36:04.8250327Z - configuration_intern_vit.py 2025-11-03T17:36:04.8251258Z . Make sure to double-check they do not contain any added malicious code. To avoid downloading new versions of the code file, you can pin a revision. 2025-11-03T17:36:04.8252099Z A new version of the following files was downloaded from https://huggingface.co/nvidia/NVLM-D-72B: 2025-11-03T17:36:04.8252601Z - configuration_nvlm_d.py 2025-11-03T17:36:04.8252863Z - configuration_intern_vit.py 2025-11-03T17:36:04.8253472Z . Make sure to double-check they do not contain any added malicious code. To avoid downloading new versions of the code file, you can pin a revision. 2025-11-03T17:36:05.0483798Z 2025-11-03T17:36:05.0484202Z preprocessor_config.json: 0% 0.00/287 [00:00, 'use_cudagraph': True, 'cudagraph_num_of_warmups': 1, 'cudagraph_capture_sizes': [512, 504, 496, 488, 480, 472, 464, 456, 448, 440, 432, 424, 416, 408, 400, 392, 384, 376, 368, 360, 352, 344, 336, 328, 320, 312, 304, 296, 288, 280, 272, 264, 256, 248, 240, 232, 224, 216, 208, 200, 192, 184, 176, 168, 160, 152, 144, 136, 128, 120, 112, 104, 96, 88, 80, 72, 64, 56, 48, 40, 32, 24, 16, 8, 4, 2, 1], 'cudagraph_copy_inputs': False, 'full_cuda_graph': True, 'use_inductor_graph_partition': False, 'pass_config': {}, 'max_capture_size': 512, 'local_cache_dir': None} 2025-11-03T17:36:13.6721468Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:36:13.6750914Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:36:13.6758595Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:36:13.6765881Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:36:13.6773735Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:36:13.6781855Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:36:13.6784936Z (EngineCore_DP0 pid=14979) INFO 11-03 17:36:13 [parallel_state.py:1231] rank 0 in world size 1 is assigned as DP rank 0, PP rank 0, TP rank 0, EP rank 0 2025-11-03T17:36:14.0215513Z (EngineCore_DP0 pid=14979) INFO 11-03 17:36:14 [topk_topp_sampler.py:59] Using FlashInfer for top-p & top-k sampling. 2025-11-03T17:36:14.1425267Z (EngineCore_DP0 pid=14979) WARNING 11-03 17:36:14 [__init__.py:2040] The following intended overrides are not keyword args and will be dropped: {'truncation'} 2025-11-03T17:36:14.1444459Z (EngineCore_DP0 pid=14979) WARNING 11-03 17:36:14 [__init__.py:2040] The following intended overrides are not keyword args and will be dropped: {'truncation'} 2025-11-03T17:36:14.1792236Z (EngineCore_DP0 pid=14979) INFO 11-03 17:36:14 [gpu_model_runner.py:2840] Starting to load model nvidia/NVLM-D-72B... 2025-11-03T17:36:14.4544455Z (EngineCore_DP0 pid=14979) INFO 11-03 17:36:14 [gpu_model_runner.py:2870] Loading model from scratch... 2025-11-03T17:36:14.4592823Z (EngineCore_DP0 pid=14979) INFO 11-03 17:36:14 [layer.py:497] MultiHeadAttention attn_backend: _Backend.FLASH_ATTN, use_upstream_fa: False 2025-11-03T17:36:14.5210655Z (EngineCore_DP0 pid=14979) INFO 11-03 17:36:14 [cuda.py:404] Using Flash Attention backend on V1 engine. 2025-11-03T17:36:15.1139718Z (EngineCore_DP0 pid=14979) INFO 11-03 17:36:15 [gpu_model_runner.py:2902] Model loading took 7.7171 GiB and 0.128413 seconds 2025-11-03T17:36:15.1144661Z (EngineCore_DP0 pid=14979) INFO 11-03 17:36:15 [kv_cache_utils.py:1199] GPU KV cache size: 2,621,440 tokens 2025-11-03T17:36:15.1145627Z (EngineCore_DP0 pid=14979) INFO 11-03 17:36:15 [kv_cache_utils.py:1204] Maximum concurrency for 98,304 tokens per request: 26.67x 2025-11-03T17:36:15.4922657Z (EngineCore_DP0 pid=14979) WARNING 11-03 17:36:15 [__init__.py:2040] The following intended overrides are not keyword args and will be dropped: {'truncation'} 2025-11-03T17:36:15.8777280Z (EngineCore_DP0 pid=14979) INFO 11-03 17:36:15 [gc_utils.py:40] GC Debug Config. enabled:False,top_objects:-1 2025-11-03T17:36:15.8799470Z INFO 11-03 17:36:15 [llm.py:337] Supported tasks: ['generate'] 2025-11-03T17:36:16.3789173Z PASSED 2025-11-03T17:36:16.3929683Z models/test_initialization.py::test_can_initialize_large_subset[GraniteMoeSharedForCausalLM] Fork a new process to run a test 15061 2025-11-03T17:36:16.3940326Z Fork a new process to run a test 0 2025-11-03T17:36:16.4217019Z INFO 11-03 17:36:16 [utils.py:239] non-default args: {'load_format': 'dummy', 'gpu_memory_utilization': 0.8, 'disable_log_stats': True, 'hf_overrides': functools.partial(, model_arch='GraniteMoeSharedForCausalLM', exist_overrides={}, use_original_num_layers=False), 'model_impl': 'vllm', 'model': 'ibm-research/moe-7b-1b-active-shared-experts'} 2025-11-03T17:36:16.5414031Z 2025-11-03T17:36:16.5414812Z config.json: 0% 0.00/901 [00:00, 'use_cudagraph': True, 'cudagraph_num_of_warmups': 1, 'cudagraph_capture_sizes': [512, 504, 496, 488, 480, 472, 464, 456, 448, 440, 432, 424, 416, 408, 400, 392, 384, 376, 368, 360, 352, 344, 336, 328, 320, 312, 304, 296, 288, 280, 272, 264, 256, 248, 240, 232, 224, 216, 208, 200, 192, 184, 176, 168, 160, 152, 144, 136, 128, 120, 112, 104, 96, 88, 80, 72, 64, 56, 48, 40, 32, 24, 16, 8, 4, 2, 1], 'cudagraph_copy_inputs': False, 'full_cuda_graph': True, 'use_inductor_graph_partition': False, 'pass_config': {}, 'max_capture_size': 512, 'local_cache_dir': None} 2025-11-03T17:36:24.5138193Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:36:24.5165556Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:36:24.5172967Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:36:24.5180818Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:36:24.5189237Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:36:24.5196629Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:36:24.5199490Z (EngineCore_DP0 pid=15125) INFO 11-03 17:36:24 [parallel_state.py:1231] rank 0 in world size 1 is assigned as DP rank 0, PP rank 0, TP rank 0, EP rank 0 2025-11-03T17:36:24.8525197Z (EngineCore_DP0 pid=15125) INFO 11-03 17:36:24 [topk_topp_sampler.py:59] Using FlashInfer for top-p & top-k sampling. 2025-11-03T17:36:24.9007966Z (EngineCore_DP0 pid=15125) INFO 11-03 17:36:24 [gpu_model_runner.py:2840] Starting to load model ibm-research/moe-7b-1b-active-shared-experts... 2025-11-03T17:36:25.1545357Z (EngineCore_DP0 pid=15125) INFO 11-03 17:36:25 [gpu_model_runner.py:2870] Loading model from scratch... 2025-11-03T17:36:25.2211745Z (EngineCore_DP0 pid=15125) INFO 11-03 17:36:25 [cuda.py:404] Using Flash Attention backend on V1 engine. 2025-11-03T17:36:25.7857496Z (EngineCore_DP0 pid=15125) INFO 11-03 17:36:25 [gpu_model_runner.py:2902] Model loading took 0.1748 GiB and 0.125664 seconds 2025-11-03T17:36:25.7861286Z (EngineCore_DP0 pid=15125) INFO 11-03 17:36:25 [kv_cache_utils.py:1199] GPU KV cache size: 5,242,880 tokens 2025-11-03T17:36:25.7862291Z (EngineCore_DP0 pid=15125) INFO 11-03 17:36:25 [kv_cache_utils.py:1204] Maximum concurrency for 4,096 tokens per request: 1280.00x 2025-11-03T17:36:26.6506848Z (EngineCore_DP0 pid=15125) INFO 11-03 17:36:26 [gc_utils.py:40] GC Debug Config. enabled:False,top_objects:-1 2025-11-03T17:36:26.6522182Z INFO 11-03 17:36:26 [llm.py:337] Supported tasks: ['generate'] 2025-11-03T17:36:27.0956506Z PASSED 2025-11-03T17:36:27.1097678Z models/test_initialization.py::test_can_initialize_large_subset[InternVLForConditionalGeneration] Fork a new process to run a test 15191 2025-11-03T17:36:27.1108541Z Fork a new process to run a test 0 2025-11-03T17:36:27.1385872Z INFO 11-03 17:36:27 [utils.py:239] non-default args: {'load_format': 'dummy', 'gpu_memory_utilization': 0.8, 'disable_log_stats': True, 'hf_overrides': functools.partial(, model_arch='InternVLForConditionalGeneration', exist_overrides={}, use_original_num_layers=False), 'model_impl': 'vllm', 'model': 'OpenGVLab/InternVL3-1B-hf'} 2025-11-03T17:36:27.2244425Z 2025-11-03T17:36:27.2247026Z config.json: 0.00B [00:00, ?B/s] 2025-11-03T17:36:27.2247375Z config.json: 1.96kB [00:00, 8.29MB/s] 2025-11-03T17:36:27.9733813Z 2025-11-03T17:36:27.9734560Z preprocessor_config.json: 0% 0.00/666 [00:00, 'use_cudagraph': True, 'cudagraph_num_of_warmups': 1, 'cudagraph_capture_sizes': [512, 504, 496, 488, 480, 472, 464, 456, 448, 440, 432, 424, 416, 408, 400, 392, 384, 376, 368, 360, 352, 344, 336, 328, 320, 312, 304, 296, 288, 280, 272, 264, 256, 248, 240, 232, 224, 216, 208, 200, 192, 184, 176, 168, 160, 152, 144, 136, 128, 120, 112, 104, 96, 88, 80, 72, 64, 56, 48, 40, 32, 24, 16, 8, 4, 2, 1], 'cudagraph_copy_inputs': False, 'full_cuda_graph': True, 'use_inductor_graph_partition': False, 'pass_config': {}, 'max_capture_size': 512, 'local_cache_dir': None} 2025-11-03T17:36:36.7608681Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:36:36.7638956Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:36:36.7646072Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:36:36.7654235Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:36:36.7661990Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:36:36.7670459Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:36:36.7673095Z (EngineCore_DP0 pid=15269) INFO 11-03 17:36:36 [parallel_state.py:1231] rank 0 in world size 1 is assigned as DP rank 0, PP rank 0, TP rank 0, EP rank 0 2025-11-03T17:36:37.1064611Z (EngineCore_DP0 pid=15269) INFO 11-03 17:36:37 [topk_topp_sampler.py:59] Using FlashInfer for top-p & top-k sampling. 2025-11-03T17:36:37.2243016Z (EngineCore_DP0 pid=15269) 2025-11-03T17:36:37.2243835Z processor_config.json: 0% 0.00/72.0 [00:00 8192). Running this sequence through the model will result in indexing errors 2025-11-03T17:36:41.8285332Z (EngineCore_DP0 pid=15269) INFO 11-03 17:36:41 [gpu_model_runner.py:2840] Starting to load model OpenGVLab/InternVL3-1B-hf... 2025-11-03T17:36:42.0844547Z (EngineCore_DP0 pid=15269) INFO 11-03 17:36:42 [gpu_model_runner.py:2870] Loading model from scratch... 2025-11-03T17:36:42.0882637Z (EngineCore_DP0 pid=15269) INFO 11-03 17:36:42 [layer.py:497] MultiHeadAttention attn_backend: _Backend.FLASH_ATTN, use_upstream_fa: False 2025-11-03T17:36:42.1154329Z (EngineCore_DP0 pid=15269) INFO 11-03 17:36:42 [cuda.py:404] Using Flash Attention backend on V1 engine. 2025-11-03T17:36:42.6844477Z (EngineCore_DP0 pid=15269) INFO 11-03 17:36:42 [gpu_model_runner.py:2902] Model loading took 0.5808 GiB and 0.091167 seconds 2025-11-03T17:36:42.6849207Z (EngineCore_DP0 pid=15269) INFO 11-03 17:36:42 [kv_cache_utils.py:1199] GPU KV cache size: 20,971,520 tokens 2025-11-03T17:36:42.6850166Z (EngineCore_DP0 pid=15269) INFO 11-03 17:36:42 [kv_cache_utils.py:1204] Maximum concurrency for 65,536 tokens per request: 320.00x 2025-11-03T17:36:46.9216848Z (EngineCore_DP0 pid=15269) INFO 11-03 17:36:46 [gc_utils.py:40] GC Debug Config. enabled:False,top_objects:-1 2025-11-03T17:36:46.9460949Z INFO 11-03 17:36:46 [llm.py:337] Supported tasks: ['generate'] 2025-11-03T17:36:47.4397987Z PASSED 2025-11-03T17:36:47.4539491Z models/test_initialization.py::test_can_initialize_large_subset[NemotronHForCausalLM] Fork a new process to run a test 15351 2025-11-03T17:36:47.4550945Z Fork a new process to run a test 0 2025-11-03T17:36:47.4828233Z INFO 11-03 17:36:47 [utils.py:239] non-default args: {'trust_remote_code': True, 'load_format': 'dummy', 'gpu_memory_utilization': 0.8, 'disable_log_stats': True, 'hf_overrides': functools.partial(, model_arch='NemotronHForCausalLM', exist_overrides={}, use_original_num_layers=False), 'model_impl': 'vllm', 'model': 'nvidia/Nemotron-H-8B-Base-8K'} 2025-11-03T17:36:47.5677872Z 2025-11-03T17:36:47.5680484Z config.json: 0.00B [00:00, ?B/s] 2025-11-03T17:36:47.5680892Z config.json: 1.50kB [00:00, 8.33MB/s] 2025-11-03T17:36:47.6306343Z 2025-11-03T17:36:47.6309466Z configuration_nemotron_h.py: 0.00B [00:00, ?B/s] 2025-11-03T17:36:47.6309912Z configuration_nemotron_h.py: 12.1kB [00:00, 29.7MB/s] 2025-11-03T17:36:47.6403693Z A new version of the following files was downloaded from https://huggingface.co/nvidia/Nemotron-H-8B-Base-8K: 2025-11-03T17:36:47.6404702Z - configuration_nemotron_h.py 2025-11-03T17:36:47.6405354Z . Make sure to double-check they do not contain any added malicious code. To avoid downloading new versions of the code file, you can pin a revision. 2025-11-03T17:36:54.1956543Z INFO 11-03 17:36:54 [model.py:653] Resolved architecture: NemotronHForCausalLM 2025-11-03T17:36:54.1957489Z `torch_dtype` is deprecated! Use `dtype` instead! 2025-11-03T17:36:54.2210970Z INFO 11-03 17:36:54 [model.py:1714] Using max model len 8192 2025-11-03T17:36:54.2212619Z The argument `trust_remote_code` is to be used with Auto classes. It has no effect here and is ignored. 2025-11-03T17:36:54.2868715Z INFO 11-03 17:36:54 [scheduler.py:225] Chunked prefill is enabled with max_num_batched_tokens=8192. 2025-11-03T17:36:54.2870475Z INFO 11-03 17:36:54 [config.py:313] Disabling cascade attention since it is not supported for hybrid models. 2025-11-03T17:36:54.3314378Z INFO 11-03 17:36:54 [config.py:429] Setting attention block size to 528 tokens to ensure that attention page size is >= mamba page size. 2025-11-03T17:36:54.3316841Z INFO 11-03 17:36:54 [config.py:453] Padding mamba page size by 0.19% to ensure that mamba page size and attention page size are exactly equal. 2025-11-03T17:36:54.3757263Z 2025-11-03T17:36:54.3762232Z tokenizer_config.json: 0.00B [00:00, ?B/s] 2025-11-03T17:36:54.3762655Z tokenizer_config.json: 178kB [00:00, 385MB/s] 2025-11-03T17:36:54.4726868Z 2025-11-03T17:36:54.9139532Z tokenizer.json: 0% 0.00/17.1M [00:00, 'use_cudagraph': True, 'cudagraph_num_of_warmups': 1, 'cudagraph_capture_sizes': [512, 504, 496, 488, 480, 472, 464, 456, 448, 440, 432, 424, 416, 408, 400, 392, 384, 376, 368, 360, 352, 344, 336, 328, 320, 312, 304, 296, 288, 280, 272, 264, 256, 248, 240, 232, 224, 216, 208, 200, 192, 184, 176, 168, 160, 152, 144, 136, 128, 120, 112, 104, 96, 88, 80, 72, 64, 56, 48, 40, 32, 24, 16, 8, 4, 2, 1], 'cudagraph_copy_inputs': False, 'full_cuda_graph': True, 'use_inductor_graph_partition': False, 'pass_config': {}, 'max_capture_size': 512, 'local_cache_dir': None} 2025-11-03T17:36:56.3289119Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:36:56.3316818Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:36:56.3325015Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:36:56.3332792Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:36:56.3341105Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:36:56.3350003Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:36:56.3352714Z (EngineCore_DP0 pid=15429) INFO 11-03 17:36:56 [parallel_state.py:1231] rank 0 in world size 1 is assigned as DP rank 0, PP rank 0, TP rank 0, EP rank 0 2025-11-03T17:36:56.6825571Z (EngineCore_DP0 pid=15429) INFO 11-03 17:36:56 [topk_topp_sampler.py:59] Using FlashInfer for top-p & top-k sampling. 2025-11-03T17:36:56.7592182Z (EngineCore_DP0 pid=15429) INFO 11-03 17:36:56 [gpu_model_runner.py:2840] Starting to load model nvidia/Nemotron-H-8B-Base-8K... 2025-11-03T17:36:57.0159625Z (EngineCore_DP0 pid=15429) INFO 11-03 17:36:57 [gpu_model_runner.py:2870] Loading model from scratch... 2025-11-03T17:36:57.0252241Z (EngineCore_DP0 pid=15429) INFO 11-03 17:36:57 [cuda.py:404] Using Flash Attention backend on V1 engine. 2025-11-03T17:36:57.6511913Z (EngineCore_DP0 pid=15429) INFO 11-03 17:36:57 [gpu_model_runner.py:2902] Model loading took 15.1125 GiB and 0.112880 seconds 2025-11-03T17:36:57.6524935Z (EngineCore_DP0 pid=15429) INFO 11-03 17:36:57 [kv_cache_utils.py:1199] GPU KV cache size: 93,456 tokens 2025-11-03T17:36:57.6526007Z (EngineCore_DP0 pid=15429) INFO 11-03 17:36:57 [kv_cache_utils.py:1204] Maximum concurrency for 8,192 tokens per request: 56.41x 2025-11-03T17:36:58.6290078Z (EngineCore_DP0 pid=15429) INFO 11-03 17:36:58 [gc_utils.py:40] GC Debug Config. enabled:False,top_objects:-1 2025-11-03T17:36:58.6304825Z INFO 11-03 17:36:58 [llm.py:337] Supported tasks: ['generate'] 2025-11-03T17:36:59.1114392Z PASSED 2025-11-03T17:36:59.1256007Z models/test_initialization.py::test_can_initialize_large_subset[NemotronForCausalLM] Fork a new process to run a test 15495 2025-11-03T17:36:59.1266750Z Fork a new process to run a test 0 2025-11-03T17:36:59.1543837Z INFO 11-03 17:36:59 [utils.py:239] non-default args: {'load_format': 'dummy', 'gpu_memory_utilization': 0.8, 'disable_log_stats': True, 'hf_overrides': functools.partial(, model_arch='NemotronForCausalLM', exist_overrides={}, use_original_num_layers=False), 'model_impl': 'vllm', 'model': 'nvidia/Minitron-8B-Base'} 2025-11-03T17:36:59.2733656Z 2025-11-03T17:36:59.2737235Z config.json: 0% 0.00/635 [00:00, 'use_cudagraph': True, 'cudagraph_num_of_warmups': 1, 'cudagraph_capture_sizes': [512, 504, 496, 488, 480, 472, 464, 456, 448, 440, 432, 424, 416, 408, 400, 392, 384, 376, 368, 360, 352, 344, 336, 328, 320, 312, 304, 296, 288, 280, 272, 264, 256, 248, 240, 232, 224, 216, 208, 200, 192, 184, 176, 168, 160, 152, 144, 136, 128, 120, 112, 104, 96, 88, 80, 72, 64, 56, 48, 40, 32, 24, 16, 8, 4, 2, 1], 'cudagraph_copy_inputs': False, 'full_cuda_graph': True, 'use_inductor_graph_partition': False, 'pass_config': {}, 'max_capture_size': 512, 'local_cache_dir': None} 2025-11-03T17:37:08.8644755Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:37:08.8672560Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:37:08.8679742Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:37:08.8687249Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:37:08.8694750Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:37:08.8702989Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:37:08.8705582Z (EngineCore_DP0 pid=15574) INFO 11-03 17:37:08 [parallel_state.py:1231] rank 0 in world size 1 is assigned as DP rank 0, PP rank 0, TP rank 0, EP rank 0 2025-11-03T17:37:09.2095047Z (EngineCore_DP0 pid=15574) INFO 11-03 17:37:09 [topk_topp_sampler.py:59] Using FlashInfer for top-p & top-k sampling. 2025-11-03T17:37:09.2841793Z (EngineCore_DP0 pid=15574) INFO 11-03 17:37:09 [gpu_model_runner.py:2840] Starting to load model nvidia/Minitron-8B-Base... 2025-11-03T17:37:09.5415244Z (EngineCore_DP0 pid=15574) INFO 11-03 17:37:09 [gpu_model_runner.py:2870] Loading model from scratch... 2025-11-03T17:37:09.6120536Z (EngineCore_DP0 pid=15574) INFO 11-03 17:37:09 [cuda.py:404] Using Flash Attention backend on V1 engine. 2025-11-03T17:37:10.1813107Z (EngineCore_DP0 pid=15574) INFO 11-03 17:37:10 [gpu_model_runner.py:2902] Model loading took 4.2662 GiB and 0.128946 seconds 2025-11-03T17:37:10.1818409Z (EngineCore_DP0 pid=15574) INFO 11-03 17:37:10 [kv_cache_utils.py:1199] GPU KV cache size: 2,621,440 tokens 2025-11-03T17:37:10.1819413Z (EngineCore_DP0 pid=15574) INFO 11-03 17:37:10 [kv_cache_utils.py:1204] Maximum concurrency for 4,096 tokens per request: 640.00x 2025-11-03T17:37:11.5544931Z (EngineCore_DP0 pid=15574) INFO 11-03 17:37:11 [gc_utils.py:40] GC Debug Config. enabled:False,top_objects:-1 2025-11-03T17:37:11.5558511Z INFO 11-03 17:37:11 [llm.py:337] Supported tasks: ['generate'] 2025-11-03T17:37:12.0394775Z PASSED 2025-11-03T17:37:12.0536866Z models/test_initialization.py::test_can_initialize_large_subset[ExaoneForCausalLM] Fork a new process to run a test 15640 2025-11-03T17:37:12.0547756Z Fork a new process to run a test 0 2025-11-03T17:37:12.0829588Z INFO 11-03 17:37:12 [utils.py:239] non-default args: {'trust_remote_code': True, 'load_format': 'dummy', 'gpu_memory_utilization': 0.8, 'disable_log_stats': True, 'hf_overrides': functools.partial(, model_arch='ExaoneForCausalLM', exist_overrides={}, use_original_num_layers=False), 'model_impl': 'vllm', 'model': 'LGAI-EXAONE/EXAONE-3.0-7.8B-Instruct'} 2025-11-03T17:37:12.1805191Z 2025-11-03T17:37:12.1805803Z config.json: 0% 0.00/878 [00:00, 'use_cudagraph': True, 'cudagraph_num_of_warmups': 1, 'cudagraph_capture_sizes': [512, 504, 496, 488, 480, 472, 464, 456, 448, 440, 432, 424, 416, 408, 400, 392, 384, 376, 368, 360, 352, 344, 336, 328, 320, 312, 304, 296, 288, 280, 272, 264, 256, 248, 240, 232, 224, 216, 208, 200, 192, 184, 176, 168, 160, 152, 144, 136, 128, 120, 112, 104, 96, 88, 80, 72, 64, 56, 48, 40, 32, 24, 16, 8, 4, 2, 1], 'cudagraph_copy_inputs': False, 'full_cuda_graph': True, 'use_inductor_graph_partition': False, 'pass_config': {}, 'max_capture_size': 512, 'local_cache_dir': None} 2025-11-03T17:37:20.6947061Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:37:20.6974197Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:37:20.6981572Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:37:20.6989079Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:37:20.6996524Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:37:20.7004367Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:37:20.7006860Z (EngineCore_DP0 pid=15698) INFO 11-03 17:37:20 [parallel_state.py:1231] rank 0 in world size 1 is assigned as DP rank 0, PP rank 0, TP rank 0, EP rank 0 2025-11-03T17:37:21.0365920Z (EngineCore_DP0 pid=15698) INFO 11-03 17:37:21 [topk_topp_sampler.py:59] Using FlashInfer for top-p & top-k sampling. 2025-11-03T17:37:21.1030154Z (EngineCore_DP0 pid=15698) INFO 11-03 17:37:21 [gpu_model_runner.py:2840] Starting to load model LGAI-EXAONE/EXAONE-3.0-7.8B-Instruct... 2025-11-03T17:37:21.3612643Z (EngineCore_DP0 pid=15698) INFO 11-03 17:37:21 [gpu_model_runner.py:2870] Loading model from scratch... 2025-11-03T17:37:21.4270995Z (EngineCore_DP0 pid=15698) INFO 11-03 17:37:21 [cuda.py:404] Using Flash Attention backend on V1 engine. 2025-11-03T17:37:21.9944585Z (EngineCore_DP0 pid=15698) INFO 11-03 17:37:21 [gpu_model_runner.py:2902] Model loading took 1.9698 GiB and 0.125178 seconds 2025-11-03T17:37:21.9948361Z (EngineCore_DP0 pid=15698) INFO 11-03 17:37:21 [kv_cache_utils.py:1199] GPU KV cache size: 2,621,440 tokens 2025-11-03T17:37:21.9949140Z (EngineCore_DP0 pid=15698) INFO 11-03 17:37:21 [kv_cache_utils.py:1204] Maximum concurrency for 4,096 tokens per request: 640.00x 2025-11-03T17:37:22.6056805Z (EngineCore_DP0 pid=15698) INFO 11-03 17:37:22 [gc_utils.py:40] GC Debug Config. enabled:False,top_objects:-1 2025-11-03T17:37:22.6078800Z INFO 11-03 17:37:22 [llm.py:337] Supported tasks: ['generate'] 2025-11-03T17:37:23.0605827Z PASSED 2025-11-03T17:37:23.0750031Z models/test_initialization.py::test_can_initialize_large_subset[TarsierForConditionalGeneration] Fork a new process to run a test 15764 2025-11-03T17:37:23.0762037Z Fork a new process to run a test 0 2025-11-03T17:37:23.1041312Z INFO 11-03 17:37:23 [utils.py:239] non-default args: {'load_format': 'dummy', 'gpu_memory_utilization': 0.8, 'disable_log_stats': True, 'hf_overrides': functools.partial(, model_arch='TarsierForConditionalGeneration', exist_overrides={}, use_original_num_layers=False), 'model_impl': 'vllm', 'model': 'omni-research/Tarsier-7b'} 2025-11-03T17:37:23.3105956Z 2025-11-03T17:37:23.3108298Z config.json: 0.00B [00:00, ?B/s] 2025-11-03T17:37:23.3108631Z config.json: 1.17kB [00:00, 5.84MB/s] 2025-11-03T17:37:23.4251358Z 2025-11-03T17:37:23.4251766Z preprocessor_config.json: 0% 0.00/505 [00:00, 'use_cudagraph': True, 'cudagraph_num_of_warmups': 1, 'cudagraph_capture_sizes': [512, 504, 496, 488, 480, 472, 464, 456, 448, 440, 432, 424, 416, 408, 400, 392, 384, 376, 368, 360, 352, 344, 336, 328, 320, 312, 304, 296, 288, 280, 272, 264, 256, 248, 240, 232, 224, 216, 208, 200, 192, 184, 176, 168, 160, 152, 144, 136, 128, 120, 112, 104, 96, 88, 80, 72, 64, 56, 48, 40, 32, 24, 16, 8, 4, 2, 1], 'cudagraph_copy_inputs': False, 'full_cuda_graph': True, 'use_inductor_graph_partition': False, 'pass_config': {}, 'max_capture_size': 512, 'local_cache_dir': None} 2025-11-03T17:37:38.8543045Z (EngineCore_DP0 pid=15844) /opt/conda/envs/py_3.12/lib/python3.12/site-packages/torch/library.py:356: UserWarning: Warning only once for all operators, other operators may also be overridden. 2025-11-03T17:37:38.8544335Z (EngineCore_DP0 pid=15844) Overriding a previously registered kernel for the same operator and the same dispatch key 2025-11-03T17:37:38.8546148Z (EngineCore_DP0 pid=15844) operator: xformers_python::_fused_allgather_and_linear_impl(Tensor scattered_input, Tensor[] weights, str process_group_name, Tensor(a3!)[] gathered_outputs, SymInt timeout_s, bool _wait, bool _memcpy, Tensor scale_scattered_input, Tensor?[] scales_weights) -> () 2025-11-03T17:37:38.8547702Z (EngineCore_DP0 pid=15844) registered at /opt/conda/envs/py_3.12/lib/python3.12/site-packages/torch/_library/custom_ops.py:922 2025-11-03T17:37:38.8548471Z (EngineCore_DP0 pid=15844) dispatch key: ADInplaceOrView 2025-11-03T17:37:38.8549183Z (EngineCore_DP0 pid=15844) previous kernel: no debug info 2025-11-03T17:37:38.8550209Z (EngineCore_DP0 pid=15844) new kernel: registered at /opt/conda/envs/py_3.12/lib/python3.12/site-packages/torch/_library/custom_ops.py:922 (Triggered internally at /var/lib/jenkins/workspace/aten/src/ATen/core/dispatch/OperatorEntry.cpp:208.) 2025-11-03T17:37:38.8551260Z (EngineCore_DP0 pid=15844) self.m.impl( 2025-11-03T17:37:39.3172931Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:37:39.3201987Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:37:39.3209003Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:37:39.3216437Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:37:39.3223408Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:37:39.3230515Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:37:39.3233114Z (EngineCore_DP0 pid=15844) INFO 11-03 17:37:39 [parallel_state.py:1231] rank 0 in world size 1 is assigned as DP rank 0, PP rank 0, TP rank 0, EP rank 0 2025-11-03T17:37:39.5568274Z (EngineCore_DP0 pid=15844) INFO 11-03 17:37:39 [topk_topp_sampler.py:59] Using FlashInfer for top-p & top-k sampling. 2025-11-03T17:37:39.6733706Z (EngineCore_DP0 pid=15844) Using a slow image processor as `use_fast` is unset and a slow processor was saved with this model. `use_fast=True` will be the default behavior in v4.52, even if the model was saved with a slow processor. This will result in minor differences in outputs. You'll still be able to use a slow processor with `use_fast=False`. 2025-11-03T17:37:40.3292716Z (EngineCore_DP0 pid=15844) INFO 11-03 17:37:40 [gpu_model_runner.py:2840] Starting to load model omni-research/Tarsier-7b... 2025-11-03T17:37:40.5610869Z (EngineCore_DP0 pid=15844) INFO 11-03 17:37:40 [gpu_model_runner.py:2870] Loading model from scratch... 2025-11-03T17:37:40.6122572Z (EngineCore_DP0 pid=15844) INFO 11-03 17:37:40 [cuda.py:404] Using Flash Attention backend on V1 engine. 2025-11-03T17:37:41.1220901Z (EngineCore_DP0 pid=15844) INFO 11-03 17:37:41 [gpu_model_runner.py:2902] Model loading took 0.9085 GiB and 0.103616 seconds 2025-11-03T17:37:41.3458367Z (EngineCore_DP0 pid=15844) INFO 11-03 17:37:41 [gpu_model_runner.py:3647] Encoder cache will be initialized with a budget of 8192 tokens, and profiled with 13 image items of the maximum feature size. 2025-11-03T17:37:42.8509372Z (EngineCore_DP0 pid=15844) INFO 11-03 17:37:42 [backends.py:594] Using cache directory: /var/lib/jenkins/.cache/vllm/torch_compile_cache/0315854aca/rank_0_0/backbone for vLLM's torch.compile 2025-11-03T17:37:42.8512864Z (EngineCore_DP0 pid=15844) INFO 11-03 17:37:42 [backends.py:608] Dynamo bytecode transform time: 1.24 s 2025-11-03T17:37:43.5007027Z (EngineCore_DP0 pid=15844) [rank0]:W1103 17:37:43.499000 15844 site-packages/torch/_inductor/utils.py:1665] [0/0] Not enough SMs to use max_autotune_gemm mode 2025-11-03T17:37:44.5632652Z (EngineCore_DP0 pid=15844) INFO 11-03 17:37:44 [backends.py:214] Cache the graph for dynamic shape for later use 2025-11-03T17:37:45.4667831Z (EngineCore_DP0 pid=15844) INFO 11-03 17:37:45 [backends.py:241] Compiling a graph for dynamic shape takes 2.41 s 2025-11-03T17:37:45.9184719Z (EngineCore_DP0 pid=15844) INFO 11-03 17:37:45 [monitor.py:33] torch.compile takes 3.65 s in total 2025-11-03T17:37:46.9122082Z (EngineCore_DP0 pid=15844) INFO 11-03 17:37:46 [gpu_worker.py:314] Available KV cache memory: 15.12 GiB 2025-11-03T17:37:47.1785276Z (EngineCore_DP0 pid=15844) INFO 11-03 17:37:47 [kv_cache_utils.py:1199] GPU KV cache size: 990,624 tokens 2025-11-03T17:37:47.1786615Z (EngineCore_DP0 pid=15844) INFO 11-03 17:37:47 [kv_cache_utils.py:1204] Maximum concurrency for 4,096 tokens per request: 241.85x 2025-11-03T17:37:47.4484537Z (EngineCore_DP0 pid=15844) 2025-11-03T17:37:47.5489361Z Capturing CUDA graphs (mixed prefill-decode, PIECEWISE): 0% 0/67 [00:00, model_arch='MistralForCausalLM', exist_overrides={}, use_original_num_layers=False), 'model_impl': 'vllm', 'model': 'mistralai/Mistral-7B-Instruct-v0.1'} 2025-11-03T17:37:49.5033509Z 2025-11-03T17:37:49.5034635Z config.json: 0% 0.00/571 [00:00, 'use_cudagraph': True, 'cudagraph_num_of_warmups': 1, 'cudagraph_capture_sizes': [512, 504, 496, 488, 480, 472, 464, 456, 448, 440, 432, 424, 416, 408, 400, 392, 384, 376, 368, 360, 352, 344, 336, 328, 320, 312, 304, 296, 288, 280, 272, 264, 256, 248, 240, 232, 224, 216, 208, 200, 192, 184, 176, 168, 160, 152, 144, 136, 128, 120, 112, 104, 96, 88, 80, 72, 64, 56, 48, 40, 32, 24, 16, 8, 4, 2, 1], 'cudagraph_copy_inputs': False, 'full_cuda_graph': True, 'use_inductor_graph_partition': False, 'pass_config': {}, 'max_capture_size': 512, 'local_cache_dir': None} 2025-11-03T17:37:50.9921619Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:37:50.9951141Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:37:50.9958480Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:37:50.9966456Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:37:50.9973743Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:37:50.9981764Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:37:50.9984579Z (EngineCore_DP0 pid=16132) INFO 11-03 17:37:50 [parallel_state.py:1231] rank 0 in world size 1 is assigned as DP rank 0, PP rank 0, TP rank 0, EP rank 0 2025-11-03T17:37:51.3360237Z (EngineCore_DP0 pid=16132) INFO 11-03 17:37:51 [topk_topp_sampler.py:59] Using FlashInfer for top-p & top-k sampling. 2025-11-03T17:37:51.4034393Z (EngineCore_DP0 pid=16132) INFO 11-03 17:37:51 [gpu_model_runner.py:2840] Starting to load model mistralai/Mistral-7B-Instruct-v0.1... 2025-11-03T17:37:51.6643000Z (EngineCore_DP0 pid=16132) INFO 11-03 17:37:51 [gpu_model_runner.py:2870] Loading model from scratch... 2025-11-03T17:37:51.7306523Z (EngineCore_DP0 pid=16132) INFO 11-03 17:37:51 [cuda.py:404] Using Flash Attention backend on V1 engine. 2025-11-03T17:37:52.3012218Z (EngineCore_DP0 pid=16132) INFO 11-03 17:37:52 [gpu_model_runner.py:2902] Model loading took 0.9024 GiB and 0.124514 seconds 2025-11-03T17:37:52.3016953Z (EngineCore_DP0 pid=16132) INFO 11-03 17:37:52 [kv_cache_utils.py:1199] GPU KV cache size: 2,621,440 tokens 2025-11-03T17:37:52.3018305Z (EngineCore_DP0 pid=16132) INFO 11-03 17:37:52 [kv_cache_utils.py:1204] Maximum concurrency for 32,768 tokens per request: 213.06x 2025-11-03T17:37:52.8671834Z (EngineCore_DP0 pid=16132) INFO 11-03 17:37:52 [gc_utils.py:40] GC Debug Config. enabled:False,top_objects:-1 2025-11-03T17:37:52.8688020Z INFO 11-03 17:37:52 [llm.py:337] Supported tasks: ['generate'] 2025-11-03T17:37:53.3159979Z PASSED 2025-11-03T17:37:53.3300211Z models/test_initialization.py::test_can_initialize_large_subset[ModernBertForSequenceClassification] Fork a new process to run a test 16198 2025-11-03T17:37:53.3310765Z Fork a new process to run a test 0 2025-11-03T17:37:53.3591777Z INFO 11-03 17:37:53 [utils.py:239] non-default args: {'load_format': 'dummy', 'gpu_memory_utilization': 0.8, 'disable_log_stats': True, 'hf_overrides': functools.partial(, model_arch='ModernBertForSequenceClassification', exist_overrides={}, use_original_num_layers=False), 'model_impl': 'vllm', 'model': 'Alibaba-NLP/gte-reranker-modernbert-base'} 2025-11-03T17:37:53.4633946Z 2025-11-03T17:37:53.4636116Z config.json: 0.00B [00:00, ?B/s] 2025-11-03T17:37:53.4636558Z config.json: 1.33kB [00:00, 7.33MB/s] 2025-11-03T17:37:59.9705818Z INFO 11-03 17:37:59 [model.py:915] Resolved `--runner auto` to `--runner pooling`. Pass the value explicitly to silence this message. 2025-11-03T17:37:59.9706633Z INFO 11-03 17:37:59 [model.py:653] Resolved architecture: ModernBertForSequenceClassification 2025-11-03T17:37:59.9707138Z `torch_dtype` is deprecated! Use `dtype` instead! 2025-11-03T17:37:59.9956672Z INFO 11-03 17:37:59 [model.py:1939] Downcasting torch.float32 to torch.float16. 2025-11-03T17:38:00.0348524Z 2025-11-03T17:38:00.0352998Z tokenizer_config.json: 0.00B [00:00, ?B/s] 2025-11-03T17:38:00.0353579Z tokenizer_config.json: 21.0kB [00:00, 43.7MB/s] 2025-11-03T17:38:00.0362086Z INFO 11-03 17:38:00 [model.py:1714] Using max model len 8192 2025-11-03T17:38:00.0666553Z INFO 11-03 17:38:00 [arg_utils.py:1725] (Disabling) chunked prefill by default 2025-11-03T17:38:00.0667365Z INFO 11-03 17:38:00 [arg_utils.py:1728] (Disabling) prefix caching by default 2025-11-03T17:38:00.2087938Z INFO 11-03 17:38:00 [vllm.py:433] Only "last" pooling supports chunked prefill and prefix caching; disabling both. 2025-11-03T17:38:00.3022788Z 2025-11-03T17:38:00.3161992Z tokenizer.json: 0.00B [00:00, ?B/s] 2025-11-03T17:38:00.3162553Z tokenizer.json: 3.58MB [00:00, 258MB/s] 2025-11-03T17:38:00.4111940Z 2025-11-03T17:38:00.4113228Z special_tokens_map.json: 0% 0.00/694 [00:00, 'use_cudagraph': True, 'cudagraph_num_of_warmups': 1, 'cudagraph_capture_sizes': [512, 504, 496, 488, 480, 472, 464, 456, 448, 440, 432, 424, 416, 408, 400, 392, 384, 376, 368, 360, 352, 344, 336, 328, 320, 312, 304, 296, 288, 280, 272, 264, 256, 248, 240, 232, 224, 216, 208, 200, 192, 184, 176, 168, 160, 152, 144, 136, 128, 120, 112, 104, 96, 88, 80, 72, 64, 56, 48, 40, 32, 24, 16, 8, 4, 2, 1], 'cudagraph_copy_inputs': False, 'full_cuda_graph': False, 'use_inductor_graph_partition': False, 'pass_config': {}, 'max_capture_size': 512, 'local_cache_dir': None} 2025-11-03T17:38:01.3642867Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:38:01.3673709Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:38:01.3681106Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:38:01.3688974Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:38:01.3696545Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:38:01.3704843Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:38:01.3708217Z (EngineCore_DP0 pid=16257) INFO 11-03 17:38:01 [parallel_state.py:1231] rank 0 in world size 1 is assigned as DP rank 0, PP rank 0, TP rank 0, EP rank 0 2025-11-03T17:38:01.7090026Z (EngineCore_DP0 pid=16257) INFO 11-03 17:38:01 [topk_topp_sampler.py:59] Using FlashInfer for top-p & top-k sampling. 2025-11-03T17:38:01.7293756Z (EngineCore_DP0 pid=16257) INFO 11-03 17:38:01 [gpu_model_runner.py:2840] Starting to load model Alibaba-NLP/gte-reranker-modernbert-base... 2025-11-03T17:38:01.9865999Z (EngineCore_DP0 pid=16257) INFO 11-03 17:38:01 [gpu_model_runner.py:2870] Loading model from scratch... 2025-11-03T17:38:02.0531943Z (EngineCore_DP0 pid=16257) INFO 11-03 17:38:02 [cuda.py:404] Using Flash Attention backend on V1 engine. 2025-11-03T17:38:02.6612321Z (EngineCore_DP0 pid=16257) INFO 11-03 17:38:02 [gpu_model_runner.py:2902] Model loading took 0.0837 GiB and 0.150153 seconds 2025-11-03T17:38:03.2262883Z (EngineCore_DP0 pid=16257) INFO 11-03 17:38:03 [core.py:142] Disabling chunked prefill for model without KVCache 2025-11-03T17:38:03.2288385Z (EngineCore_DP0 pid=16257) INFO 11-03 17:38:03 [vllm.py:433] Only "last" pooling supports chunked prefill and prefix caching; disabling both. 2025-11-03T17:38:03.4863153Z (EngineCore_DP0 pid=16257) INFO 11-03 17:38:03 [gc_utils.py:40] GC Debug Config. enabled:False,top_objects:-1 2025-11-03T17:38:03.4884178Z INFO 11-03 17:38:03 [llm.py:337] Supported tasks: ['score', 'classify', 'encode'] 2025-11-03T17:38:03.9227163Z PASSED 2025-11-03T17:38:03.9365696Z models/test_initialization.py::test_can_initialize_large_subset[Qwen2ForCausalLM] Fork a new process to run a test 16323 2025-11-03T17:38:03.9375931Z Fork a new process to run a test 0 2025-11-03T17:38:03.9654440Z INFO 11-03 17:38:03 [utils.py:239] non-default args: {'load_format': 'dummy', 'gpu_memory_utilization': 0.8, 'disable_log_stats': True, 'hf_overrides': functools.partial(, model_arch='Qwen2ForCausalLM', exist_overrides={}, use_original_num_layers=False), 'model_impl': 'vllm', 'model': 'Qwen/Qwen2-0.5B-Instruct'} 2025-11-03T17:38:04.0646305Z 2025-11-03T17:38:04.0647690Z config.json: 0% 0.00/659 [00:00, 'use_cudagraph': True, 'cudagraph_num_of_warmups': 1, 'cudagraph_capture_sizes': [512, 504, 496, 488, 480, 472, 464, 456, 448, 440, 432, 424, 416, 408, 400, 392, 384, 376, 368, 360, 352, 344, 336, 328, 320, 312, 304, 296, 288, 280, 272, 264, 256, 248, 240, 232, 224, 216, 208, 200, 192, 184, 176, 168, 160, 152, 144, 136, 128, 120, 112, 104, 96, 88, 80, 72, 64, 56, 48, 40, 32, 24, 16, 8, 4, 2, 1], 'cudagraph_copy_inputs': False, 'full_cuda_graph': True, 'use_inductor_graph_partition': False, 'pass_config': {}, 'max_capture_size': 512, 'local_cache_dir': None} 2025-11-03T17:38:05.9280654Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:38:05.9307908Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:38:05.9315228Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:38:05.9323682Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:38:05.9331867Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:38:05.9339494Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:38:05.9342861Z (EngineCore_DP0 pid=16331) INFO 11-03 17:38:05 [parallel_state.py:1231] rank 0 in world size 1 is assigned as DP rank 0, PP rank 0, TP rank 0, EP rank 0 2025-11-03T17:38:06.2717334Z (EngineCore_DP0 pid=16331) INFO 11-03 17:38:06 [topk_topp_sampler.py:59] Using FlashInfer for top-p & top-k sampling. 2025-11-03T17:38:06.3164689Z (EngineCore_DP0 pid=16331) INFO 11-03 17:38:06 [gpu_model_runner.py:2840] Starting to load model Qwen/Qwen2-0.5B-Instruct... 2025-11-03T17:38:06.5699824Z (EngineCore_DP0 pid=16331) INFO 11-03 17:38:06 [gpu_model_runner.py:2870] Loading model from scratch... 2025-11-03T17:38:06.6351397Z (EngineCore_DP0 pid=16331) INFO 11-03 17:38:06 [cuda.py:404] Using Flash Attention backend on V1 engine. 2025-11-03T17:38:07.2279899Z (EngineCore_DP0 pid=16331) INFO 11-03 17:38:07 [gpu_model_runner.py:2902] Model loading took 0.2861 GiB and 0.146916 seconds 2025-11-03T17:38:07.2283932Z (EngineCore_DP0 pid=16331) INFO 11-03 17:38:07 [kv_cache_utils.py:1199] GPU KV cache size: 20,971,520 tokens 2025-11-03T17:38:07.2285301Z (EngineCore_DP0 pid=16331) INFO 11-03 17:38:07 [kv_cache_utils.py:1204] Maximum concurrency for 32,768 tokens per request: 640.00x 2025-11-03T17:38:09.7555375Z (EngineCore_DP0 pid=16331) INFO 11-03 17:38:09 [gc_utils.py:40] GC Debug Config. enabled:False,top_objects:-1 2025-11-03T17:38:09.7569255Z INFO 11-03 17:38:09 [llm.py:337] Supported tasks: ['generate'] 2025-11-03T17:38:10.2206112Z PASSED 2025-11-03T17:38:10.2345683Z models/test_initialization.py::test_can_initialize_large_subset[MPTForCausalLM] Fork a new process to run a test 16397 2025-11-03T17:38:10.2356043Z Fork a new process to run a test 0 2025-11-03T17:38:10.2635025Z INFO 11-03 17:38:10 [utils.py:239] non-default args: {'load_format': 'dummy', 'gpu_memory_utilization': 0.8, 'disable_log_stats': True, 'hf_overrides': functools.partial(, model_arch='MPTForCausalLM', exist_overrides={}, use_original_num_layers=False), 'model_impl': 'vllm', 'model': 'mosaicml/mpt-7b'} 2025-11-03T17:38:10.3825799Z 2025-11-03T17:38:10.3828188Z config.json: 0.00B [00:00, ?B/s] 2025-11-03T17:38:10.3828667Z config.json: 1.23kB [00:00, 6.85MB/s] 2025-11-03T17:38:16.8072531Z INFO 11-03 17:38:16 [model.py:653] Resolved architecture: MPTForCausalLM 2025-11-03T17:38:16.8073219Z `torch_dtype` is deprecated! Use `dtype` instead! 2025-11-03T17:38:16.8323400Z INFO 11-03 17:38:16 [model.py:1714] Using max model len 2048 2025-11-03T17:38:17.0039552Z INFO 11-03 17:38:17 [scheduler.py:225] Chunked prefill is enabled with max_num_batched_tokens=8192. 2025-11-03T17:38:17.0377733Z 2025-11-03T17:38:17.0378982Z tokenizer_config.json: 0% 0.00/237 [00:00, 'use_cudagraph': True, 'cudagraph_num_of_warmups': 1, 'cudagraph_capture_sizes': [512, 504, 496, 488, 480, 472, 464, 456, 448, 440, 432, 424, 416, 408, 400, 392, 384, 376, 368, 360, 352, 344, 336, 328, 320, 312, 304, 296, 288, 280, 272, 264, 256, 248, 240, 232, 224, 216, 208, 200, 192, 184, 176, 168, 160, 152, 144, 136, 128, 120, 112, 104, 96, 88, 80, 72, 64, 56, 48, 40, 32, 24, 16, 8, 4, 2, 1], 'cudagraph_copy_inputs': False, 'full_cuda_graph': True, 'use_inductor_graph_partition': False, 'pass_config': {}, 'max_capture_size': 512, 'local_cache_dir': None} 2025-11-03T17:38:18.3119891Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:38:18.3147915Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:38:18.3155149Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:38:18.3163496Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:38:18.3170906Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:38:18.3178554Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:38:18.3181290Z (EngineCore_DP0 pid=16455) INFO 11-03 17:38:18 [parallel_state.py:1231] rank 0 in world size 1 is assigned as DP rank 0, PP rank 0, TP rank 0, EP rank 0 2025-11-03T17:38:18.6496245Z (EngineCore_DP0 pid=16455) INFO 11-03 17:38:18 [topk_topp_sampler.py:59] Using FlashInfer for top-p & top-k sampling. 2025-11-03T17:38:18.7138893Z (EngineCore_DP0 pid=16455) INFO 11-03 17:38:18 [gpu_model_runner.py:2840] Starting to load model mosaicml/mpt-7b... 2025-11-03T17:38:18.9713319Z (EngineCore_DP0 pid=16455) INFO 11-03 17:38:18 [gpu_model_runner.py:2870] Loading model from scratch... 2025-11-03T17:38:19.0149255Z (EngineCore_DP0 pid=16455) INFO 11-03 17:38:19 [cuda.py:404] Using Flash Attention backend on V1 engine. 2025-11-03T17:38:19.5672567Z (EngineCore_DP0 pid=16455) INFO 11-03 17:38:19 [gpu_model_runner.py:2902] Model loading took 0.7598 GiB and 0.101642 seconds 2025-11-03T17:38:19.5676467Z (EngineCore_DP0 pid=16455) INFO 11-03 17:38:19 [kv_cache_utils.py:1199] GPU KV cache size: 655,360 tokens 2025-11-03T17:38:19.5677375Z (EngineCore_DP0 pid=16455) INFO 11-03 17:38:19 [kv_cache_utils.py:1204] Maximum concurrency for 2,048 tokens per request: 320.00x 2025-11-03T17:38:20.0216661Z (EngineCore_DP0 pid=16455) INFO 11-03 17:38:20 [gc_utils.py:40] GC Debug Config. enabled:False,top_objects:-1 2025-11-03T17:38:20.0231750Z INFO 11-03 17:38:20 [llm.py:337] Supported tasks: ['generate'] 2025-11-03T17:38:20.4630240Z PASSED 2025-11-03T17:38:20.4770840Z models/test_initialization.py::test_can_initialize_large_subset[Starcoder2ForCausalLM] Fork a new process to run a test 16521 2025-11-03T17:38:20.4781355Z Fork a new process to run a test 0 2025-11-03T17:38:20.5059934Z INFO 11-03 17:38:20 [utils.py:239] non-default args: {'load_format': 'dummy', 'gpu_memory_utilization': 0.8, 'disable_log_stats': True, 'hf_overrides': functools.partial(, model_arch='Starcoder2ForCausalLM', exist_overrides={}, use_original_num_layers=False), 'model_impl': 'vllm', 'model': 'bigcode/starcoder2-3b'} 2025-11-03T17:38:20.6008138Z 2025-11-03T17:38:20.6008730Z config.json: 0% 0.00/700 [00:00, 'use_cudagraph': True, 'cudagraph_num_of_warmups': 1, 'cudagraph_capture_sizes': [512, 504, 496, 488, 480, 472, 464, 456, 448, 440, 432, 424, 416, 408, 400, 392, 384, 376, 368, 360, 352, 344, 336, 328, 320, 312, 304, 296, 288, 280, 272, 264, 256, 248, 240, 232, 224, 216, 208, 200, 192, 184, 176, 168, 160, 152, 144, 136, 128, 120, 112, 104, 96, 88, 80, 72, 64, 56, 48, 40, 32, 24, 16, 8, 4, 2, 1], 'cudagraph_copy_inputs': False, 'full_cuda_graph': True, 'use_inductor_graph_partition': False, 'pass_config': {}, 'max_capture_size': 512, 'local_cache_dir': None} 2025-11-03T17:38:28.5638352Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:38:28.5666029Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:38:28.5673403Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:38:28.5681300Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:38:28.5689110Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:38:28.5697112Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:38:28.5699873Z (EngineCore_DP0 pid=16579) INFO 11-03 17:38:28 [parallel_state.py:1231] rank 0 in world size 1 is assigned as DP rank 0, PP rank 0, TP rank 0, EP rank 0 2025-11-03T17:38:28.9204096Z (EngineCore_DP0 pid=16579) INFO 11-03 17:38:28 [topk_topp_sampler.py:59] Using FlashInfer for top-p & top-k sampling. 2025-11-03T17:38:28.9880418Z (EngineCore_DP0 pid=16579) INFO 11-03 17:38:28 [gpu_model_runner.py:2840] Starting to load model bigcode/starcoder2-3b... 2025-11-03T17:38:29.2531468Z (EngineCore_DP0 pid=16579) INFO 11-03 17:38:29 [gpu_model_runner.py:2870] Loading model from scratch... 2025-11-03T17:38:29.3203024Z (EngineCore_DP0 pid=16579) INFO 11-03 17:38:29 [cuda.py:404] Using Flash Attention backend on V1 engine. 2025-11-03T17:38:29.9030422Z (EngineCore_DP0 pid=16579) INFO 11-03 17:38:29 [gpu_model_runner.py:2902] Model loading took 0.4649 GiB and 0.126758 seconds 2025-11-03T17:38:29.9034812Z (EngineCore_DP0 pid=16579) INFO 11-03 17:38:29 [kv_cache_utils.py:1199] GPU KV cache size: 10,485,760 tokens 2025-11-03T17:38:29.9035774Z (EngineCore_DP0 pid=16579) INFO 11-03 17:38:29 [kv_cache_utils.py:1204] Maximum concurrency for 16,384 tokens per request: 852.22x 2025-11-03T17:38:31.2383167Z (EngineCore_DP0 pid=16579) INFO 11-03 17:38:31 [gc_utils.py:40] GC Debug Config. enabled:False,top_objects:-1 2025-11-03T17:38:31.2397322Z INFO 11-03 17:38:31 [llm.py:337] Supported tasks: ['generate'] 2025-11-03T17:38:31.6931252Z PASSED 2025-11-03T17:38:31.7070882Z models/test_initialization.py::test_can_initialize_large_subset[Qwen2_5_VLForConditionalGeneration] Fork a new process to run a test 16645 2025-11-03T17:38:31.7081691Z Fork a new process to run a test 0 2025-11-03T17:38:31.7359827Z INFO 11-03 17:38:31 [utils.py:239] non-default args: {'load_format': 'dummy', 'max_model_len': 4096, 'gpu_memory_utilization': 0.8, 'disable_log_stats': True, 'hf_overrides': functools.partial(, model_arch='Qwen2_5_VLForConditionalGeneration', exist_overrides={}, use_original_num_layers=False), 'model_impl': 'vllm', 'model': 'Qwen/Qwen2.5-VL-3B-Instruct'} 2025-11-03T17:38:31.8329550Z 2025-11-03T17:38:31.8334730Z config.json: 0.00B [00:00, ?B/s] 2025-11-03T17:38:31.8335714Z config.json: 1.37kB [00:00, 3.44MB/s] 2025-11-03T17:38:31.9240766Z 2025-11-03T17:38:31.9242356Z preprocessor_config.json: 0% 0.00/350 [00:00, 'use_cudagraph': True, 'cudagraph_num_of_warmups': 1, 'cudagraph_capture_sizes': [512, 504, 496, 488, 480, 472, 464, 456, 448, 440, 432, 424, 416, 408, 400, 392, 384, 376, 368, 360, 352, 344, 336, 328, 320, 312, 304, 296, 288, 280, 272, 264, 256, 248, 240, 232, 224, 216, 208, 200, 192, 184, 176, 168, 160, 152, 144, 136, 128, 120, 112, 104, 96, 88, 80, 72, 64, 56, 48, 40, 32, 24, 16, 8, 4, 2, 1], 'cudagraph_copy_inputs': False, 'full_cuda_graph': True, 'use_inductor_graph_partition': False, 'pass_config': {}, 'max_capture_size': 512, 'local_cache_dir': None} 2025-11-03T17:38:34.0657017Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:38:34.0686283Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:38:34.0693484Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:38:34.0701219Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:38:34.0709201Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:38:34.0716737Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:38:34.0719493Z (EngineCore_DP0 pid=16653) INFO 11-03 17:38:34 [parallel_state.py:1231] rank 0 in world size 1 is assigned as DP rank 0, PP rank 0, TP rank 0, EP rank 0 2025-11-03T17:38:34.4127770Z (EngineCore_DP0 pid=16653) INFO 11-03 17:38:34 [topk_topp_sampler.py:59] Using FlashInfer for top-p & top-k sampling. 2025-11-03T17:38:35.8776650Z (EngineCore_DP0 pid=16653) 2025-11-03T17:38:35.8778229Z chat_template.json: 0.00B [00:00, ?B/s] 2025-11-03T17:38:35.8778572Z chat_template.json: 1.05kB [00:00, 3.21MB/s] 2025-11-03T17:38:36.4197389Z (EngineCore_DP0 pid=16653) INFO 11-03 17:38:36 [gpu_model_runner.py:2840] Starting to load model Qwen/Qwen2.5-VL-3B-Instruct... 2025-11-03T17:38:36.7570035Z (EngineCore_DP0 pid=16653) INFO 11-03 17:38:36 [gpu_model_runner.py:2870] Loading model from scratch... 2025-11-03T17:38:36.8233535Z (EngineCore_DP0 pid=16653) INFO 11-03 17:38:36 [cuda.py:404] Using Flash Attention backend on V1 engine. 2025-11-03T17:38:37.4131641Z (EngineCore_DP0 pid=16653) INFO 11-03 17:38:37 [gpu_model_runner.py:2902] Model loading took 2.0974 GiB and 0.130729 seconds 2025-11-03T17:38:37.4135947Z (EngineCore_DP0 pid=16653) INFO 11-03 17:38:37 [kv_cache_utils.py:1199] GPU KV cache size: 10,485,760 tokens 2025-11-03T17:38:37.4136859Z (EngineCore_DP0 pid=16653) INFO 11-03 17:38:37 [kv_cache_utils.py:1204] Maximum concurrency for 4,096 tokens per request: 2560.00x 2025-11-03T17:38:38.9129168Z (EngineCore_DP0 pid=16653) INFO 11-03 17:38:38 [gc_utils.py:40] GC Debug Config. enabled:False,top_objects:-1 2025-11-03T17:38:38.9142707Z INFO 11-03 17:38:38 [llm.py:337] Supported tasks: ['generate'] 2025-11-03T17:38:39.3849417Z PASSED 2025-11-03T17:38:39.3988659Z models/test_initialization.py::test_can_initialize_large_subset[DeepseekV32ForCausalLM] Fork a new process to run a test 16719 2025-11-03T17:38:39.3999832Z Fork a new process to run a test 0 2025-11-03T17:38:39.4275580Z INFO 11-03 17:38:39 [utils.py:239] non-default args: {'load_format': 'dummy', 'gpu_memory_utilization': 0.8, 'disable_log_stats': True, 'hf_overrides': functools.partial(, model_arch='DeepseekV32ForCausalLM', exist_overrides={}, use_original_num_layers=False), 'model_impl': 'vllm', 'model': 'deepseek-ai/DeepSeek-V3.2-Exp'} 2025-11-03T17:38:39.5158291Z 2025-11-03T17:38:39.5160529Z config.json: 0.00B [00:00, ?B/s] 2025-11-03T17:38:39.5160831Z config.json: 1.55kB [00:00, 7.23MB/s] 2025-11-03T17:38:39.5588107Z You are using a model of type deepseek_v32 to instantiate a model of type deepseek_v3. This is not supported for all configurations of models and can yield errors. 2025-11-03T17:38:39.5597257Z INFO 11-03 17:38:39 [config.py:637] Detected quantization_config.scale_fmt=ue8m0; enabling UE8M0 for DeepGEMM. 2025-11-03T17:38:39.5599192Z INFO 11-03 17:38:39 [config.py:408] Replacing legacy 'type' key with 'rope_type' 2025-11-03T17:38:39.6398726Z INFO 11-03 17:38:39 [model.py:653] Resolved architecture: DeepseekV32ForCausalLM 2025-11-03T17:38:39.6399325Z `torch_dtype` is deprecated! Use `dtype` instead! 2025-11-03T17:38:39.6644275Z INFO 11-03 17:38:39 [model.py:1714] Using max model len 163840 2025-11-03T17:38:39.9396740Z INFO 11-03 17:38:39 [scheduler.py:225] Chunked prefill is enabled with max_num_batched_tokens=8192. 2025-11-03T17:38:39.9397468Z INFO 11-03 17:38:39 [config.py:480] Using custom fp8 kv-cache format for DeepSeekV3.2 2025-11-03T17:38:39.9428963Z INFO 11-03 17:38:39 [cuda.py:186] Forcing kv cache block size to 64 for FlashMLASparse backend. 2025-11-03T17:38:39.9793932Z 2025-11-03T17:38:39.9795790Z tokenizer_config.json: 0.00B [00:00, ?B/s] 2025-11-03T17:38:39.9796157Z tokenizer_config.json: 4.06kB [00:00, 28.3MB/s] 2025-11-03T17:38:40.0965525Z 2025-11-03T17:38:40.1284997Z tokenizer.json: 0.00B [00:00, ?B/s] 2025-11-03T17:38:40.1285450Z tokenizer.json: 7.85MB [00:00, 246MB/s] 2025-11-03T17:38:40.6077257Z 2025-11-03T17:38:40.6077505Z generation_config.json: 0% 0.00/171 [00:00, 'use_cudagraph': True, 'cudagraph_num_of_warmups': 1, 'cudagraph_capture_sizes': [512, 504, 496, 488, 480, 472, 464, 456, 448, 440, 432, 424, 416, 408, 400, 392, 384, 376, 368, 360, 352, 344, 336, 328, 320, 312, 304, 296, 288, 280, 272, 264, 256, 248, 240, 232, 224, 216, 208, 200, 192, 184, 176, 168, 160, 152, 144, 136, 128, 120, 112, 104, 96, 88, 80, 72, 64, 56, 48, 40, 32, 24, 16, 8, 4, 2, 1], 'cudagraph_copy_inputs': False, 'full_cuda_graph': True, 'use_inductor_graph_partition': False, 'pass_config': {}, 'max_capture_size': 512, 'local_cache_dir': None} 2025-11-03T17:38:41.3488567Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:38:41.3515838Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:38:41.3523532Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:38:41.3531147Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:38:41.3538963Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:38:41.3547592Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:38:41.3549948Z (EngineCore_DP0 pid=16727) INFO 11-03 17:38:41 [parallel_state.py:1231] rank 0 in world size 1 is assigned as DP rank 0, PP rank 0, TP rank 0, EP rank 0 2025-11-03T17:38:41.6968042Z (EngineCore_DP0 pid=16727) INFO 11-03 17:38:41 [topk_topp_sampler.py:59] Using FlashInfer for top-p & top-k sampling. 2025-11-03T17:38:41.8109402Z (EngineCore_DP0 pid=16727) INFO 11-03 17:38:41 [gpu_model_runner.py:2840] Starting to load model deepseek-ai/DeepSeek-V3.2-Exp... 2025-11-03T17:38:42.0702090Z (EngineCore_DP0 pid=16727) INFO 11-03 17:38:42 [gpu_model_runner.py:2870] Loading model from scratch... 2025-11-03T17:38:42.1620010Z (EngineCore_DP0 pid=16727) INFO 11-03 17:38:42 [cuda.py:275] Using Sparse MLA backend on V1 engine. 2025-11-03T17:38:42.1793429Z (EngineCore_DP0 pid=16727) WARNING 11-03 17:38:42 [fp8.py:150] DeepGEMM backend requested but not available. 2025-11-03T17:38:42.1794304Z (EngineCore_DP0 pid=16727) INFO 11-03 17:38:42 [fp8.py:165] Using Triton backend for FP8 MoE 2025-11-03T17:38:42.3005234Z (EngineCore_DP0 pid=16727) WARNING 11-03 17:38:42 [fp8_utils.py:785] Using default W8A8 Block FP8 kernel config. Performance might be sub-optimal! Config file not found at /opt/conda/envs/py_3.12/lib/python3.12/site-packages/vllm/model_executor/layers/quantization/utils/configs/N=32768,K=512,device_name=NVIDIA_L4,dtype=fp8_w8a8,block_shape=[128,128].json 2025-11-03T17:38:43.3261779Z (EngineCore_DP0 pid=16727) INFO 11-03 17:38:43 [gpu_model_runner.py:2902] Model loading took 4.4518 GiB and 0.731108 seconds 2025-11-03T17:38:43.3266957Z (EngineCore_DP0 pid=16727) INFO 11-03 17:38:43 [kv_cache_utils.py:1199] GPU KV cache size: 13,626,112 tokens 2025-11-03T17:38:43.3267870Z (EngineCore_DP0 pid=16727) INFO 11-03 17:38:43 [kv_cache_utils.py:1204] Maximum concurrency for 163,840 tokens per request: 83.17x 2025-11-03T17:38:44.4025077Z (EngineCore_DP0 pid=16727) INFO 11-03 17:38:44 [gc_utils.py:40] GC Debug Config. enabled:False,top_objects:-1 2025-11-03T17:38:44.4039263Z INFO 11-03 17:38:44 [llm.py:337] Supported tasks: ['generate'] 2025-11-03T17:38:44.9034285Z PASSED 2025-11-03T17:38:44.9173087Z models/test_initialization.py::test_can_initialize_large_subset[PhiMoEForCausalLM] Fork a new process to run a test 16797 2025-11-03T17:38:44.9183512Z Fork a new process to run a test 0 2025-11-03T17:38:44.9459203Z INFO 11-03 17:38:44 [utils.py:239] non-default args: {'trust_remote_code': True, 'load_format': 'dummy', 'gpu_memory_utilization': 0.8, 'disable_log_stats': True, 'hf_overrides': functools.partial(, model_arch='PhiMoEForCausalLM', exist_overrides={}, use_original_num_layers=False), 'model_impl': 'vllm', 'model': 'microsoft/Phi-3.5-MoE-instruct'} 2025-11-03T17:38:45.0238196Z 2025-11-03T17:38:45.0240676Z config.json: 0.00B [00:00, ?B/s] 2025-11-03T17:38:45.0241015Z config.json: 4.53kB [00:00, 18.5MB/s] 2025-11-03T17:38:45.1058653Z 2025-11-03T17:38:45.1060918Z configuration_phimoe.py: 0.00B [00:00, ?B/s] 2025-11-03T17:38:45.1061304Z configuration_phimoe.py: 12.3kB [00:00, 35.6MB/s] 2025-11-03T17:38:45.1164473Z A new version of the following files was downloaded from https://huggingface.co/microsoft/Phi-3.5-MoE-instruct: 2025-11-03T17:38:45.1165023Z - configuration_phimoe.py 2025-11-03T17:38:45.1165625Z . Make sure to double-check they do not contain any added malicious code. To avoid downloading new versions of the code file, you can pin a revision. 2025-11-03T17:38:45.1503133Z INFO 11-03 17:38:45 [config.py:408] Replacing legacy 'type' key with 'rope_type' 2025-11-03T17:38:51.5762630Z INFO 11-03 17:38:51 [model.py:653] Resolved architecture: PhiMoEForCausalLM 2025-11-03T17:38:51.5763120Z `torch_dtype` is deprecated! Use `dtype` instead! 2025-11-03T17:38:51.6015933Z INFO 11-03 17:38:51 [model.py:1714] Using max model len 131072 2025-11-03T17:38:51.6017695Z The argument `trust_remote_code` is to be used with Auto classes. It has no effect here and is ignored. 2025-11-03T17:38:51.6579608Z INFO 11-03 17:38:51 [scheduler.py:225] Chunked prefill is enabled with max_num_batched_tokens=8192. 2025-11-03T17:38:51.6990924Z 2025-11-03T17:38:51.6991996Z tokenizer_config.json: 0.00B [00:00, ?B/s] 2025-11-03T17:38:51.6992331Z tokenizer_config.json: 3.98kB [00:00, 38.9MB/s] 2025-11-03T17:38:51.8042494Z 2025-11-03T17:38:51.9336195Z tokenizer.model: 0% 0.00/500k [00:00, 'use_cudagraph': True, 'cudagraph_num_of_warmups': 1, 'cudagraph_capture_sizes': [512, 504, 496, 488, 480, 472, 464, 456, 448, 440, 432, 424, 416, 408, 400, 392, 384, 376, 368, 360, 352, 344, 336, 328, 320, 312, 304, 296, 288, 280, 272, 264, 256, 248, 240, 232, 224, 216, 208, 200, 192, 184, 176, 168, 160, 152, 144, 136, 128, 120, 112, 104, 96, 88, 80, 72, 64, 56, 48, 40, 32, 24, 16, 8, 4, 2, 1], 'cudagraph_copy_inputs': False, 'full_cuda_graph': True, 'use_inductor_graph_partition': False, 'pass_config': {}, 'max_capture_size': 512, 'local_cache_dir': None} 2025-11-03T17:38:52.9318968Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:38:52.9347742Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:38:52.9354811Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:38:52.9362749Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:38:52.9370548Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:38:52.9378316Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:38:52.9381525Z (EngineCore_DP0 pid=16875) INFO 11-03 17:38:52 [parallel_state.py:1231] rank 0 in world size 1 is assigned as DP rank 0, PP rank 0, TP rank 0, EP rank 0 2025-11-03T17:38:53.2786126Z (EngineCore_DP0 pid=16875) INFO 11-03 17:38:53 [topk_topp_sampler.py:59] Using FlashInfer for top-p & top-k sampling. 2025-11-03T17:38:53.3642073Z (EngineCore_DP0 pid=16875) INFO 11-03 17:38:53 [gpu_model_runner.py:2840] Starting to load model microsoft/Phi-3.5-MoE-instruct... 2025-11-03T17:38:53.6219524Z (EngineCore_DP0 pid=16875) INFO 11-03 17:38:53 [gpu_model_runner.py:2870] Loading model from scratch... 2025-11-03T17:38:53.6908046Z (EngineCore_DP0 pid=16875) INFO 11-03 17:38:53 [cuda.py:404] Using Flash Attention backend on V1 engine. 2025-11-03T17:38:54.2687696Z (EngineCore_DP0 pid=16875) INFO 11-03 17:38:54 [gpu_model_runner.py:2902] Model loading took 0.8927 GiB and 0.128726 seconds 2025-11-03T17:38:54.2691886Z (EngineCore_DP0 pid=16875) INFO 11-03 17:38:54 [kv_cache_utils.py:1199] GPU KV cache size: 2,621,440 tokens 2025-11-03T17:38:54.2693001Z (EngineCore_DP0 pid=16875) INFO 11-03 17:38:54 [kv_cache_utils.py:1204] Maximum concurrency for 131,072 tokens per request: 20.00x 2025-11-03T17:38:54.8057916Z (EngineCore_DP0 pid=16875) INFO 11-03 17:38:54 [gc_utils.py:40] GC Debug Config. enabled:False,top_objects:-1 2025-11-03T17:38:54.8080152Z INFO 11-03 17:38:54 [llm.py:337] Supported tasks: ['generate'] 2025-11-03T17:38:55.2699447Z PASSED 2025-11-03T17:38:55.2840564Z models/test_initialization.py::test_can_initialize_large_subset[Qwen2MoeForCausalLM] Fork a new process to run a test 16941 2025-11-03T17:38:55.2851317Z Fork a new process to run a test 0 2025-11-03T17:38:55.3129835Z INFO 11-03 17:38:55 [utils.py:239] non-default args: {'load_format': 'dummy', 'gpu_memory_utilization': 0.8, 'disable_log_stats': True, 'hf_overrides': functools.partial(, model_arch='Qwen2MoeForCausalLM', exist_overrides={}, use_original_num_layers=False), 'model_impl': 'vllm', 'model': 'Qwen/Qwen1.5-MoE-A2.7B-Chat'} 2025-11-03T17:38:55.3944024Z 2025-11-03T17:38:55.3945673Z config.json: 0% 0.00/920 [00:00, 'use_cudagraph': True, 'cudagraph_num_of_warmups': 1, 'cudagraph_capture_sizes': [512, 504, 496, 488, 480, 472, 464, 456, 448, 440, 432, 424, 416, 408, 400, 392, 384, 376, 368, 360, 352, 344, 336, 328, 320, 312, 304, 296, 288, 280, 272, 264, 256, 248, 240, 232, 224, 216, 208, 200, 192, 184, 176, 168, 160, 152, 144, 136, 128, 120, 112, 104, 96, 88, 80, 72, 64, 56, 48, 40, 32, 24, 16, 8, 4, 2, 1], 'cudagraph_copy_inputs': False, 'full_cuda_graph': True, 'use_inductor_graph_partition': False, 'pass_config': {}, 'max_capture_size': 512, 'local_cache_dir': None} 2025-11-03T17:39:03.4882545Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:39:03.4910557Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:39:03.4917631Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:39:03.4925443Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:39:03.4933330Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:39:03.4940508Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:39:03.4943295Z (EngineCore_DP0 pid=16999) INFO 11-03 17:39:03 [parallel_state.py:1231] rank 0 in world size 1 is assigned as DP rank 0, PP rank 0, TP rank 0, EP rank 0 2025-11-03T17:39:03.8289270Z (EngineCore_DP0 pid=16999) INFO 11-03 17:39:03 [topk_topp_sampler.py:59] Using FlashInfer for top-p & top-k sampling. 2025-11-03T17:39:03.8821815Z (EngineCore_DP0 pid=16999) INFO 11-03 17:39:03 [gpu_model_runner.py:2840] Starting to load model Qwen/Qwen1.5-MoE-A2.7B-Chat... 2025-11-03T17:39:04.1406304Z (EngineCore_DP0 pid=16999) INFO 11-03 17:39:04 [gpu_model_runner.py:2870] Loading model from scratch... 2025-11-03T17:39:04.2061847Z (EngineCore_DP0 pid=16999) INFO 11-03 17:39:04 [cuda.py:404] Using Flash Attention backend on V1 engine. 2025-11-03T17:39:04.7742702Z (EngineCore_DP0 pid=16999) INFO 11-03 17:39:04 [gpu_model_runner.py:2902] Model loading took 1.2959 GiB and 0.125911 seconds 2025-11-03T17:39:04.7747396Z (EngineCore_DP0 pid=16999) INFO 11-03 17:39:04 [kv_cache_utils.py:1199] GPU KV cache size: 1,310,720 tokens 2025-11-03T17:39:04.7748758Z (EngineCore_DP0 pid=16999) INFO 11-03 17:39:04 [kv_cache_utils.py:1204] Maximum concurrency for 32,768 tokens per request: 40.00x 2025-11-03T17:39:05.4540545Z (EngineCore_DP0 pid=16999) INFO 11-03 17:39:05 [gc_utils.py:40] GC Debug Config. enabled:False,top_objects:-1 2025-11-03T17:39:05.4556108Z INFO 11-03 17:39:05 [llm.py:337] Supported tasks: ['generate'] 2025-11-03T17:39:05.9104386Z PASSED 2025-11-03T17:39:05.9245116Z models/test_initialization.py::test_can_initialize_large_subset[DotsOCRForCausalLM] Fork a new process to run a test 17065 2025-11-03T17:39:05.9256345Z Fork a new process to run a test 0 2025-11-03T17:39:05.9533186Z INFO 11-03 17:39:05 [utils.py:239] non-default args: {'trust_remote_code': True, 'load_format': 'dummy', 'gpu_memory_utilization': 0.8, 'disable_log_stats': True, 'hf_overrides': functools.partial(, model_arch='DotsOCRForCausalLM', exist_overrides={}, use_original_num_layers=False), 'model_impl': 'vllm', 'model': 'rednote-hilab/dots.ocr'} 2025-11-03T17:39:06.0370542Z 2025-11-03T17:39:06.0372372Z config.json: 0.00B [00:00, ?B/s] 2025-11-03T17:39:06.0372679Z config.json: 1.47kB [00:00, 9.97MB/s] 2025-11-03T17:39:06.1235159Z 2025-11-03T17:39:06.1236553Z configuration_dots.py: 0.00B [00:00, ?B/s] 2025-11-03T17:39:06.1236926Z configuration_dots.py: 3.04kB [00:00, 12.5MB/s] 2025-11-03T17:39:06.1331430Z A new version of the following files was downloaded from https://huggingface.co/rednote-hilab/dots.ocr: 2025-11-03T17:39:06.1331943Z - configuration_dots.py 2025-11-03T17:39:06.1332564Z . Make sure to double-check they do not contain any added malicious code. To avoid downloading new versions of the code file, you can pin a revision. 2025-11-03T17:39:06.3455418Z 2025-11-03T17:39:06.3455892Z preprocessor_config.json: 0% 0.00/432 [00:00, 'use_cudagraph': True, 'cudagraph_num_of_warmups': 1, 'cudagraph_capture_sizes': [512, 504, 496, 488, 480, 472, 464, 456, 448, 440, 432, 424, 416, 408, 400, 392, 384, 376, 368, 360, 352, 344, 336, 328, 320, 312, 304, 296, 288, 280, 272, 264, 256, 248, 240, 232, 224, 216, 208, 200, 192, 184, 176, 168, 160, 152, 144, 136, 128, 120, 112, 104, 96, 88, 80, 72, 64, 56, 48, 40, 32, 24, 16, 8, 4, 2, 1], 'cudagraph_copy_inputs': False, 'full_cuda_graph': True, 'use_inductor_graph_partition': False, 'pass_config': {}, 'max_capture_size': 512, 'local_cache_dir': None} 2025-11-03T17:39:14.6042924Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:39:14.6072058Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:39:14.6079406Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:39:14.6086939Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:39:14.6094606Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:39:14.6102202Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:39:14.6104960Z (EngineCore_DP0 pid=17123) INFO 11-03 17:39:14 [parallel_state.py:1231] rank 0 in world size 1 is assigned as DP rank 0, PP rank 0, TP rank 0, EP rank 0 2025-11-03T17:39:14.9482753Z (EngineCore_DP0 pid=17123) INFO 11-03 17:39:14 [topk_topp_sampler.py:59] Using FlashInfer for top-p & top-k sampling. 2025-11-03T17:39:15.0909908Z (EngineCore_DP0 pid=17123) The image processor of type `Qwen2VLImageProcessor` is now loaded as a fast processor by default, even if the model checkpoint was saved with a slow processor. This is a breaking change and may produce slightly different outputs. To continue using the slow processor, instantiate this class with `use_fast=False`. Note that this behavior will be extended to all models in a future release. 2025-11-03T17:39:16.4490125Z (EngineCore_DP0 pid=17123) 2025-11-03T17:39:16.4491701Z chat_template.json: 0.00B [00:00, ?B/s] 2025-11-03T17:39:16.4492032Z chat_template.json: 1.11kB [00:00, 3.53MB/s] 2025-11-03T17:39:17.0676973Z (EngineCore_DP0 pid=17123) INFO 11-03 17:39:17 [gpu_model_runner.py:2840] Starting to load model rednote-hilab/dots.ocr... 2025-11-03T17:39:17.3299379Z (EngineCore_DP0 pid=17123) INFO 11-03 17:39:17 [gpu_model_runner.py:2870] Loading model from scratch... 2025-11-03T17:39:17.3598984Z (EngineCore_DP0 pid=17123) INFO 11-03 17:39:17 [cuda.py:404] Using Flash Attention backend on V1 engine. 2025-11-03T17:39:17.9420400Z (EngineCore_DP0 pid=17123) INFO 11-03 17:39:17 [gpu_model_runner.py:2902] Model loading took 1.1334 GiB and 0.090076 seconds 2025-11-03T17:39:17.9425299Z (EngineCore_DP0 pid=17123) INFO 11-03 17:39:17 [kv_cache_utils.py:1199] GPU KV cache size: 10,485,760 tokens 2025-11-03T17:39:17.9426228Z (EngineCore_DP0 pid=17123) INFO 11-03 17:39:17 [kv_cache_utils.py:1204] Maximum concurrency for 131,072 tokens per request: 80.00x 2025-11-03T17:39:19.4428747Z (EngineCore_DP0 pid=17123) INFO 11-03 17:39:19 [gc_utils.py:40] GC Debug Config. enabled:False,top_objects:-1 2025-11-03T17:39:19.4442418Z INFO 11-03 17:39:19 [llm.py:337] Supported tasks: ['generate'] 2025-11-03T17:39:19.9251776Z PASSED 2025-11-03T17:39:19.9393013Z models/test_initialization.py::test_can_initialize_large_subset[Qwen3NextMTP] Fork a new process to run a test 17189 2025-11-03T17:39:19.9403775Z Fork a new process to run a test 0 2025-11-03T17:39:19.9407712Z `transformers==4.56.2` installed, but `transformers>=4.56.3` is required to run this model. 2025-11-03T17:39:20.2339100Z PASSED 2025-11-03T17:39:20.2480997Z models/test_initialization.py::test_can_initialize_large_subset[BaichuanForCausalLM] Fork a new process to run a test 17190 2025-11-03T17:39:20.2491589Z Fork a new process to run a test 0 2025-11-03T17:39:20.2767303Z INFO 11-03 17:39:20 [utils.py:239] non-default args: {'trust_remote_code': True, 'load_format': 'dummy', 'gpu_memory_utilization': 0.8, 'disable_log_stats': True, 'hf_overrides': functools.partial(, model_arch='BaichuanForCausalLM', exist_overrides={}, use_original_num_layers=False), 'model_impl': 'vllm', 'model': 'baichuan-inc/Baichuan2-7B-chat'} 2025-11-03T17:39:20.4370897Z 2025-11-03T17:39:20.4372251Z config.json: 0% 0.00/758 [00:00, 'use_cudagraph': True, 'cudagraph_num_of_warmups': 1, 'cudagraph_capture_sizes': [512, 504, 496, 488, 480, 472, 464, 456, 448, 440, 432, 424, 416, 408, 400, 392, 384, 376, 368, 360, 352, 344, 336, 328, 320, 312, 304, 296, 288, 280, 272, 264, 256, 248, 240, 232, 224, 216, 208, 200, 192, 184, 176, 168, 160, 152, 144, 136, 128, 120, 112, 104, 96, 88, 80, 72, 64, 56, 48, 40, 32, 24, 16, 8, 4, 2, 1], 'cudagraph_copy_inputs': False, 'full_cuda_graph': True, 'use_inductor_graph_partition': False, 'pass_config': {}, 'max_capture_size': 512, 'local_cache_dir': None} 2025-11-03T17:39:29.2615074Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:39:29.2642959Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:39:29.2650498Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:39:29.2658885Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:39:29.2666690Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:39:29.2674709Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:39:29.2677288Z (EngineCore_DP0 pid=17268) INFO 11-03 17:39:29 [parallel_state.py:1231] rank 0 in world size 1 is assigned as DP rank 0, PP rank 0, TP rank 0, EP rank 0 2025-11-03T17:39:29.6123310Z (EngineCore_DP0 pid=17268) INFO 11-03 17:39:29 [topk_topp_sampler.py:59] Using FlashInfer for top-p & top-k sampling. 2025-11-03T17:39:29.6810146Z (EngineCore_DP0 pid=17268) INFO 11-03 17:39:29 [gpu_model_runner.py:2840] Starting to load model baichuan-inc/Baichuan2-7B-chat... 2025-11-03T17:39:29.9434454Z (EngineCore_DP0 pid=17268) INFO 11-03 17:39:29 [gpu_model_runner.py:2870] Loading model from scratch... 2025-11-03T17:39:30.0093143Z (EngineCore_DP0 pid=17268) INFO 11-03 17:39:30 [cuda.py:404] Using Flash Attention backend on V1 engine. 2025-11-03T17:39:30.5866740Z (EngineCore_DP0 pid=17268) INFO 11-03 17:39:30 [gpu_model_runner.py:2902] Model loading took 2.2959 GiB and 0.127177 seconds 2025-11-03T17:39:30.5872125Z (EngineCore_DP0 pid=17268) INFO 11-03 17:39:30 [kv_cache_utils.py:1199] GPU KV cache size: 655,360 tokens 2025-11-03T17:39:30.5872872Z (EngineCore_DP0 pid=17268) INFO 11-03 17:39:30 [kv_cache_utils.py:1204] Maximum concurrency for 4,096 tokens per request: 160.00x 2025-11-03T17:39:31.2026664Z (EngineCore_DP0 pid=17268) WARNING 11-03 17:39:31 [tokenizer.py:255] Using a slow tokenizer. This might cause a significant slowdown. Consider using a fast tokenizer instead. 2025-11-03T17:39:31.7077271Z (EngineCore_DP0 pid=17268) INFO 11-03 17:39:31 [gc_utils.py:40] GC Debug Config. enabled:False,top_objects:-1 2025-11-03T17:39:31.7091366Z INFO 11-03 17:39:31 [llm.py:337] Supported tasks: ['generate'] 2025-11-03T17:39:32.1768908Z PASSED 2025-11-03T17:39:32.1909297Z models/test_initialization.py::test_can_initialize_large_subset[Qwen2ForRewardModel] Fork a new process to run a test 17334 2025-11-03T17:39:32.1920224Z Fork a new process to run a test 0 2025-11-03T17:39:32.1924632Z `transformers==4.56.2` installed, but `transformers<=4.53` is required to run this model. Reason: HF model uses remote code that is not compatible with latest Transformers 2025-11-03T17:39:32.4804962Z PASSED 2025-11-03T17:39:32.4946213Z models/test_initialization.py::test_can_initialize_large_subset[MiniCPMO] Fork a new process to run a test 17335 2025-11-03T17:39:32.4957196Z Fork a new process to run a test 0 2025-11-03T17:39:32.5230965Z INFO 11-03 17:39:32 [utils.py:239] non-default args: {'trust_remote_code': True, 'load_format': 'dummy', 'gpu_memory_utilization': 0.8, 'disable_log_stats': True, 'hf_overrides': functools.partial(, model_arch='MiniCPMO', exist_overrides={}, use_original_num_layers=False), 'model_impl': 'vllm', 'model': 'openbmb/MiniCPM-o-2_6'} 2025-11-03T17:39:32.6192526Z 2025-11-03T17:39:32.6194408Z config.json: 0.00B [00:00, ?B/s] 2025-11-03T17:39:32.6194856Z config.json: 3.44kB [00:00, 18.2MB/s] 2025-11-03T17:39:32.6925402Z 2025-11-03T17:39:32.6927426Z configuration_minicpm.py: 0.00B [00:00, ?B/s] 2025-11-03T17:39:32.6927899Z configuration_minicpm.py: 7.55kB [00:00, 54.4MB/s] 2025-11-03T17:39:32.7543379Z 2025-11-03T17:39:32.7546287Z modeling_navit_siglip.py: 0.00B [00:00, ?B/s] 2025-11-03T17:39:32.7546753Z modeling_navit_siglip.py: 42.1kB [00:00, 126MB/s] 2025-11-03T17:39:32.7650123Z A new version of the following files was downloaded from https://huggingface.co/openbmb/MiniCPM-o-2_6: 2025-11-03T17:39:32.7650674Z - modeling_navit_siglip.py 2025-11-03T17:39:32.7651535Z . Make sure to double-check they do not contain any added malicious code. To avoid downloading new versions of the code file, you can pin a revision. 2025-11-03T17:39:32.7652376Z A new version of the following files was downloaded from https://huggingface.co/openbmb/MiniCPM-o-2_6: 2025-11-03T17:39:32.7652870Z - configuration_minicpm.py 2025-11-03T17:39:32.7653125Z - modeling_navit_siglip.py 2025-11-03T17:39:32.7653959Z . Make sure to double-check they do not contain any added malicious code. To avoid downloading new versions of the code file, you can pin a revision. 2025-11-03T17:39:33.0796318Z 2025-11-03T17:39:33.0796907Z preprocessor_config.json: 0% 0.00/714 [00:00, 'use_cudagraph': True, 'cudagraph_num_of_warmups': 1, 'cudagraph_capture_sizes': [512, 504, 496, 488, 480, 472, 464, 456, 448, 440, 432, 424, 416, 408, 400, 392, 384, 376, 368, 360, 352, 344, 336, 328, 320, 312, 304, 296, 288, 280, 272, 264, 256, 248, 240, 232, 224, 216, 208, 200, 192, 184, 176, 168, 160, 152, 144, 136, 128, 120, 112, 104, 96, 88, 80, 72, 64, 56, 48, 40, 32, 24, 16, 8, 4, 2, 1], 'cudagraph_copy_inputs': False, 'full_cuda_graph': True, 'use_inductor_graph_partition': False, 'pass_config': {}, 'max_capture_size': 512, 'local_cache_dir': None} 2025-11-03T17:39:41.5353244Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:39:41.5382334Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:39:41.5389804Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:39:41.5397638Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:39:41.5405360Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:39:41.5413073Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:39:41.5415875Z (EngineCore_DP0 pid=17393) INFO 11-03 17:39:41 [parallel_state.py:1231] rank 0 in world size 1 is assigned as DP rank 0, PP rank 0, TP rank 0, EP rank 0 2025-11-03T17:39:41.8803799Z (EngineCore_DP0 pid=17393) INFO 11-03 17:39:41 [topk_topp_sampler.py:59] Using FlashInfer for top-p & top-k sampling. 2025-11-03T17:39:42.1343999Z (EngineCore_DP0 pid=17393) 2025-11-03T17:39:42.1346226Z processing_minicpmo.py: 0.00B [00:00, ?B/s] 2025-11-03T17:39:42.1346679Z processing_minicpmo.py: 20.0kB [00:00, 66.7MB/s] 2025-11-03T17:39:42.2324429Z (EngineCore_DP0 pid=17393) 2025-11-03T17:39:42.2326470Z image_processing_minicpmv.py: 0.00B [00:00, ?B/s] 2025-11-03T17:39:42.2327272Z image_processing_minicpmv.py: 16.7kB [00:00, 72.6MB/s] 2025-11-03T17:39:42.3029808Z (EngineCore_DP0 pid=17393) Using a slow image processor as `use_fast` is unset and a slow processor was saved with this model. `use_fast=True` will be the default behavior in v4.52, even if the model was saved with a slow processor. This will result in minor differences in outputs. You'll still be able to use a slow processor with `use_fast=False`. 2025-11-03T17:39:51.7129390Z (EngineCore_DP0 pid=17393) INFO 11-03 17:39:51 [gpu_model_runner.py:2840] Starting to load model openbmb/MiniCPM-o-2_6... 2025-11-03T17:39:52.0487584Z (EngineCore_DP0 pid=17393) INFO 11-03 17:39:52 [gpu_model_runner.py:2870] Loading model from scratch... 2025-11-03T17:39:52.0727927Z (EngineCore_DP0 pid=17393) INFO 11-03 17:39:52 [cuda.py:404] Using Flash Attention backend on V1 engine. 2025-11-03T17:39:52.1488133Z (EngineCore_DP0 pid=17393) INFO 11-03 17:39:52 [layer.py:497] MultiHeadAttention attn_backend: _Backend.XFORMERS, use_upstream_fa: False 2025-11-03T17:39:52.7466618Z (EngineCore_DP0 pid=17393) INFO 11-03 17:39:52 [gpu_model_runner.py:2902] Model loading took 2.7360 GiB and 0.183769 seconds 2025-11-03T17:39:52.7471303Z (EngineCore_DP0 pid=17393) INFO 11-03 17:39:52 [kv_cache_utils.py:1199] GPU KV cache size: 5,242,880 tokens 2025-11-03T17:39:52.7472584Z (EngineCore_DP0 pid=17393) INFO 11-03 17:39:52 [kv_cache_utils.py:1204] Maximum concurrency for 32,768 tokens per request: 160.00x 2025-11-03T17:39:53.9816503Z (EngineCore_DP0 pid=17393) INFO 11-03 17:39:53 [gc_utils.py:40] GC Debug Config. enabled:False,top_objects:-1 2025-11-03T17:39:53.9832153Z INFO 11-03 17:39:53 [llm.py:337] Supported tasks: ['generate'] 2025-11-03T17:39:54.4723482Z PASSED 2025-11-03T17:39:54.4865394Z models/test_initialization.py::test_can_initialize_large_subset[DeepseekForCausalLM] Fork a new process to run a test 17475 2025-11-03T17:39:54.4876651Z Fork a new process to run a test 0 2025-11-03T17:39:54.5153891Z INFO 11-03 17:39:54 [utils.py:239] non-default args: {'load_format': 'dummy', 'gpu_memory_utilization': 0.8, 'disable_log_stats': True, 'hf_overrides': functools.partial(, model_arch='DeepseekForCausalLM', exist_overrides={}, use_original_num_layers=False), 'model_impl': 'vllm', 'model': 'deepseek-ai/deepseek-llm-7b-chat'} 2025-11-03T17:39:54.6089034Z 2025-11-03T17:39:54.6090186Z config.json: 0% 0.00/594 [00:00, 'use_cudagraph': True, 'cudagraph_num_of_warmups': 1, 'cudagraph_capture_sizes': [512, 504, 496, 488, 480, 472, 464, 456, 448, 440, 432, 424, 416, 408, 400, 392, 384, 376, 368, 360, 352, 344, 336, 328, 320, 312, 304, 296, 288, 280, 272, 264, 256, 248, 240, 232, 224, 216, 208, 200, 192, 184, 176, 168, 160, 152, 144, 136, 128, 120, 112, 104, 96, 88, 80, 72, 64, 56, 48, 40, 32, 24, 16, 8, 4, 2, 1], 'cudagraph_copy_inputs': False, 'full_cuda_graph': True, 'use_inductor_graph_partition': False, 'pass_config': {}, 'max_capture_size': 512, 'local_cache_dir': None} 2025-11-03T17:39:56.2547492Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:39:56.2575096Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:39:56.2582783Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:39:56.2590643Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:39:56.2598426Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:39:56.2606416Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:39:56.2609325Z (EngineCore_DP0 pid=17483) INFO 11-03 17:39:56 [parallel_state.py:1231] rank 0 in world size 1 is assigned as DP rank 0, PP rank 0, TP rank 0, EP rank 0 2025-11-03T17:39:56.6040293Z (EngineCore_DP0 pid=17483) INFO 11-03 17:39:56 [topk_topp_sampler.py:59] Using FlashInfer for top-p & top-k sampling. 2025-11-03T17:39:56.6690575Z (EngineCore_DP0 pid=17483) INFO 11-03 17:39:56 [gpu_model_runner.py:2840] Starting to load model deepseek-ai/deepseek-llm-7b-chat... 2025-11-03T17:39:56.9285525Z (EngineCore_DP0 pid=17483) INFO 11-03 17:39:56 [gpu_model_runner.py:2870] Loading model from scratch... 2025-11-03T17:39:56.9945037Z (EngineCore_DP0 pid=17483) INFO 11-03 17:39:56 [cuda.py:404] Using Flash Attention backend on V1 engine. 2025-11-03T17:39:57.5731903Z (EngineCore_DP0 pid=17483) INFO 11-03 17:39:57 [gpu_model_runner.py:2902] Model loading took 1.9405 GiB and 0.124794 seconds 2025-11-03T17:39:57.5736371Z (EngineCore_DP0 pid=17483) INFO 11-03 17:39:57 [kv_cache_utils.py:1199] GPU KV cache size: 655,360 tokens 2025-11-03T17:39:57.5737318Z (EngineCore_DP0 pid=17483) INFO 11-03 17:39:57 [kv_cache_utils.py:1204] Maximum concurrency for 4,096 tokens per request: 160.00x 2025-11-03T17:39:58.1871218Z (EngineCore_DP0 pid=17483) INFO 11-03 17:39:58 [gc_utils.py:40] GC Debug Config. enabled:False,top_objects:-1 2025-11-03T17:39:58.1893637Z INFO 11-03 17:39:58 [llm.py:337] Supported tasks: ['generate'] 2025-11-03T17:39:58.6440816Z PASSED 2025-11-03T17:39:58.6583309Z models/test_initialization.py::test_can_initialize_large_subset[LlamaForCausalLM] Fork a new process to run a test 17549 2025-11-03T17:39:58.6594515Z Fork a new process to run a test 0 2025-11-03T17:39:58.6874046Z INFO 11-03 17:39:58 [utils.py:239] non-default args: {'load_format': 'dummy', 'gpu_memory_utilization': 0.8, 'disable_log_stats': True, 'hf_overrides': functools.partial(, model_arch='LlamaForCausalLM', exist_overrides={}, use_original_num_layers=False), 'model_impl': 'vllm', 'model': 'meta-llama/Llama-3.2-1B-Instruct'} 2025-11-03T17:39:58.8600227Z INFO 11-03 17:39:58 [model.py:653] Resolved architecture: LlamaForCausalLM 2025-11-03T17:39:58.8600863Z `torch_dtype` is deprecated! Use `dtype` instead! 2025-11-03T17:39:58.8849516Z INFO 11-03 17:39:58 [model.py:1714] Using max model len 131072 2025-11-03T17:39:59.0493284Z INFO 11-03 17:39:59 [scheduler.py:225] Chunked prefill is enabled with max_num_batched_tokens=8192. 2025-11-03T17:39:59.5218655Z (EngineCore_DP0 pid=17556) INFO 11-03 17:39:59 [core.py:727] Waiting for init message from front-end. 2025-11-03T17:39:59.5257545Z (EngineCore_DP0 pid=17556) INFO 11-03 17:39:59 [core.py:94] Initializing a V1 LLM engine (v0.11.0rc2.dev389+ge51928192.d20251103) with config: model='meta-llama/Llama-3.2-1B-Instruct', speculative_config=None, tokenizer='meta-llama/Llama-3.2-1B-Instruct', skip_tokenizer_init=False, tokenizer_mode=auto, revision=None, tokenizer_revision=None, trust_remote_code=False, dtype=torch.bfloat16, max_seq_len=131072, download_dir=None, load_format=dummy, tensor_parallel_size=1, pipeline_parallel_size=1, data_parallel_size=1, disable_custom_all_reduce=False, quantization=None, enforce_eager=False, kv_cache_dtype=auto, device_config=cuda, structured_outputs_config=StructuredOutputsConfig(backend='auto', disable_fallback=False, disable_any_whitespace=False, disable_additional_properties=False, reasoning_parser=''), observability_config=ObservabilityConfig(show_hidden_metrics_for_version=None, otlp_traces_endpoint=None, collect_detailed_traces=None), seed=0, served_model_name=meta-llama/Llama-3.2-1B-Instruct, enable_prefix_caching=True, chunked_prefill_enabled=True, pooler_config=None, compilation_config={'level': 3, 'debug_dump_path': None, 'cache_dir': '', 'backend': '', 'custom_ops': [], 'splitting_ops': ['vllm.unified_attention', 'vllm.unified_attention_with_output', 'vllm.unified_mla_attention', 'vllm.unified_mla_attention_with_output', 'vllm.mamba_mixer2', 'vllm.mamba_mixer', 'vllm.short_conv', 'vllm.linear_attention', 'vllm.plamo2_mamba_mixer', 'vllm.gdn_attention', 'vllm.sparse_attn_indexer'], 'use_inductor': True, 'compile_sizes': [], 'inductor_compile_config': {'enable_auto_functionalized_v2': False}, 'inductor_passes': {}, 'cudagraph_mode': , 'use_cudagraph': True, 'cudagraph_num_of_warmups': 1, 'cudagraph_capture_sizes': [512, 504, 496, 488, 480, 472, 464, 456, 448, 440, 432, 424, 416, 408, 400, 392, 384, 376, 368, 360, 352, 344, 336, 328, 320, 312, 304, 296, 288, 280, 272, 264, 256, 248, 240, 232, 224, 216, 208, 200, 192, 184, 176, 168, 160, 152, 144, 136, 128, 120, 112, 104, 96, 88, 80, 72, 64, 56, 48, 40, 32, 24, 16, 8, 4, 2, 1], 'cudagraph_copy_inputs': False, 'full_cuda_graph': True, 'use_inductor_graph_partition': False, 'pass_config': {}, 'max_capture_size': 512, 'local_cache_dir': None} 2025-11-03T17:40:00.1886334Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:40:00.1913526Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:40:00.1920651Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:40:00.1928711Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:40:00.1936353Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:40:00.1944851Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:40:00.1947378Z (EngineCore_DP0 pid=17556) INFO 11-03 17:40:00 [parallel_state.py:1231] rank 0 in world size 1 is assigned as DP rank 0, PP rank 0, TP rank 0, EP rank 0 2025-11-03T17:40:00.5273386Z (EngineCore_DP0 pid=17556) INFO 11-03 17:40:00 [topk_topp_sampler.py:59] Using FlashInfer for top-p & top-k sampling. 2025-11-03T17:40:00.5908862Z (EngineCore_DP0 pid=17556) INFO 11-03 17:40:00 [gpu_model_runner.py:2840] Starting to load model meta-llama/Llama-3.2-1B-Instruct... 2025-11-03T17:40:00.8449243Z (EngineCore_DP0 pid=17556) INFO 11-03 17:40:00 [gpu_model_runner.py:2870] Loading model from scratch... 2025-11-03T17:40:00.9376145Z (EngineCore_DP0 pid=17556) INFO 11-03 17:40:00 [cuda.py:404] Using Flash Attention backend on V1 engine. 2025-11-03T17:40:01.4954232Z (EngineCore_DP0 pid=17556) INFO 11-03 17:40:01 [gpu_model_runner.py:2902] Model loading took 0.6192 GiB and 0.150960 seconds 2025-11-03T17:40:01.4958692Z (EngineCore_DP0 pid=17556) INFO 11-03 17:40:01 [kv_cache_utils.py:1199] GPU KV cache size: 5,242,880 tokens 2025-11-03T17:40:01.4959741Z (EngineCore_DP0 pid=17556) INFO 11-03 17:40:01 [kv_cache_utils.py:1204] Maximum concurrency for 131,072 tokens per request: 40.00x 2025-11-03T17:40:02.5898809Z (EngineCore_DP0 pid=17556) INFO 11-03 17:40:02 [gc_utils.py:40] GC Debug Config. enabled:False,top_objects:-1 2025-11-03T17:40:02.5916057Z INFO 11-03 17:40:02 [llm.py:337] Supported tasks: ['generate'] 2025-11-03T17:40:03.0644686Z PASSED 2025-11-03T17:40:03.0785861Z models/test_initialization.py::test_can_initialize_large_subset[Ernie4_5_MoeForCausalLM] Fork a new process to run a test 17622 2025-11-03T17:40:03.0796887Z Fork a new process to run a test 0 2025-11-03T17:40:03.1071661Z INFO 11-03 17:40:03 [utils.py:239] non-default args: {'load_format': 'dummy', 'gpu_memory_utilization': 0.8, 'disable_log_stats': True, 'hf_overrides': functools.partial(, model_arch='Ernie4_5_MoeForCausalLM', exist_overrides={}, use_original_num_layers=False), 'model_impl': 'vllm', 'model': 'baidu/ERNIE-4.5-21B-A3B-PT'} 2025-11-03T17:40:03.2068996Z 2025-11-03T17:40:03.2070407Z config.json: 0% 0.00/936 [00:00, 'use_cudagraph': True, 'cudagraph_num_of_warmups': 1, 'cudagraph_capture_sizes': [512, 504, 496, 488, 480, 472, 464, 456, 448, 440, 432, 424, 416, 408, 400, 392, 384, 376, 368, 360, 352, 344, 336, 328, 320, 312, 304, 296, 288, 280, 272, 264, 256, 248, 240, 232, 224, 216, 208, 200, 192, 184, 176, 168, 160, 152, 144, 136, 128, 120, 112, 104, 96, 88, 80, 72, 64, 56, 48, 40, 32, 24, 16, 8, 4, 2, 1], 'cudagraph_copy_inputs': False, 'full_cuda_graph': True, 'use_inductor_graph_partition': False, 'pass_config': {}, 'max_capture_size': 512, 'local_cache_dir': None} 2025-11-03T17:40:12.3764909Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:40:12.3792612Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:40:12.3799772Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:40:12.3807425Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:40:12.3814651Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:40:12.3822867Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:40:12.3825687Z (EngineCore_DP0 pid=17701) INFO 11-03 17:40:12 [parallel_state.py:1231] rank 0 in world size 1 is assigned as DP rank 0, PP rank 0, TP rank 0, EP rank 0 2025-11-03T17:40:12.7168967Z (EngineCore_DP0 pid=17701) INFO 11-03 17:40:12 [topk_topp_sampler.py:59] Using FlashInfer for top-p & top-k sampling. 2025-11-03T17:40:12.8002565Z (EngineCore_DP0 pid=17701) INFO 11-03 17:40:12 [gpu_model_runner.py:2840] Starting to load model baidu/ERNIE-4.5-21B-A3B-PT... 2025-11-03T17:40:13.0548057Z (EngineCore_DP0 pid=17701) INFO 11-03 17:40:13 [gpu_model_runner.py:2870] Loading model from scratch... 2025-11-03T17:40:13.1220453Z (EngineCore_DP0 pid=17701) INFO 11-03 17:40:13 [cuda.py:404] Using Flash Attention backend on V1 engine. 2025-11-03T17:40:13.6834373Z (EngineCore_DP0 pid=17701) INFO 11-03 17:40:13 [gpu_model_runner.py:2902] Model loading took 0.7310 GiB and 0.125065 seconds 2025-11-03T17:40:13.6838501Z (EngineCore_DP0 pid=17701) INFO 11-03 17:40:13 [kv_cache_utils.py:1199] GPU KV cache size: 5,242,880 tokens 2025-11-03T17:40:13.6839401Z (EngineCore_DP0 pid=17701) INFO 11-03 17:40:13 [kv_cache_utils.py:1204] Maximum concurrency for 131,072 tokens per request: 40.00x 2025-11-03T17:40:15.7262183Z (EngineCore_DP0 pid=17701) INFO 11-03 17:40:15 [gc_utils.py:40] GC Debug Config. enabled:False,top_objects:-1 2025-11-03T17:40:15.7275235Z INFO 11-03 17:40:15 [llm.py:337] Supported tasks: ['generate'] 2025-11-03T17:40:16.2045637Z PASSED 2025-11-03T17:40:16.2185448Z models/test_initialization.py::test_can_initialize_large_subset[MistralModel] Fork a new process to run a test 17767 2025-11-03T17:40:16.2195691Z Fork a new process to run a test 0 2025-11-03T17:40:16.2467177Z INFO 11-03 17:40:16 [utils.py:239] non-default args: {'load_format': 'dummy', 'gpu_memory_utilization': 0.8, 'disable_log_stats': True, 'hf_overrides': functools.partial(, model_arch='MistralModel', exist_overrides={}, use_original_num_layers=False), 'model_impl': 'vllm', 'model': 'intfloat/e5-mistral-7b-instruct'} 2025-11-03T17:40:16.3396897Z 2025-11-03T17:40:16.3398037Z config.json: 0% 0.00/629 [00:00, 'use_cudagraph': True, 'cudagraph_num_of_warmups': 1, 'cudagraph_capture_sizes': [512, 504, 496, 488, 480, 472, 464, 456, 448, 440, 432, 424, 416, 408, 400, 392, 384, 376, 368, 360, 352, 344, 336, 328, 320, 312, 304, 296, 288, 280, 272, 264, 256, 248, 240, 232, 224, 216, 208, 200, 192, 184, 176, 168, 160, 152, 144, 136, 128, 120, 112, 104, 96, 88, 80, 72, 64, 56, 48, 40, 32, 24, 16, 8, 4, 2, 1], 'cudagraph_copy_inputs': False, 'full_cuda_graph': False, 'use_inductor_graph_partition': False, 'pass_config': {}, 'max_capture_size': 512, 'local_cache_dir': None} 2025-11-03T17:40:18.2619046Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:40:18.2647731Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:40:18.2655949Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:40:18.2664656Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:40:18.2672591Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:40:18.2680616Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:40:18.2683251Z (EngineCore_DP0 pid=17795) INFO 11-03 17:40:18 [parallel_state.py:1231] rank 0 in world size 1 is assigned as DP rank 0, PP rank 0, TP rank 0, EP rank 0 2025-11-03T17:40:18.6095816Z (EngineCore_DP0 pid=17795) INFO 11-03 17:40:18 [topk_topp_sampler.py:59] Using FlashInfer for top-p & top-k sampling. 2025-11-03T17:40:18.6570693Z (EngineCore_DP0 pid=17795) INFO 11-03 17:40:18 [gpu_model_runner.py:2840] Starting to load model intfloat/e5-mistral-7b-instruct... 2025-11-03T17:40:18.9136515Z (EngineCore_DP0 pid=17795) INFO 11-03 17:40:18 [gpu_model_runner.py:2870] Loading model from scratch... 2025-11-03T17:40:18.9796635Z (EngineCore_DP0 pid=17795) INFO 11-03 17:40:18 [cuda.py:404] Using Flash Attention backend on V1 engine. 2025-11-03T17:40:19.5549282Z (EngineCore_DP0 pid=17795) INFO 11-03 17:40:19 [gpu_model_runner.py:2902] Model loading took 0.6582 GiB and 0.127195 seconds 2025-11-03T17:40:19.5553621Z (EngineCore_DP0 pid=17795) INFO 11-03 17:40:19 [kv_cache_utils.py:1199] GPU KV cache size: 2,621,440 tokens 2025-11-03T17:40:19.5554575Z (EngineCore_DP0 pid=17795) INFO 11-03 17:40:19 [kv_cache_utils.py:1204] Maximum concurrency for 32,768 tokens per request: 213.06x 2025-11-03T17:40:20.1231882Z (EngineCore_DP0 pid=17795) INFO 11-03 17:40:20 [gc_utils.py:40] GC Debug Config. enabled:False,top_objects:-1 2025-11-03T17:40:20.1247274Z INFO 11-03 17:40:20 [llm.py:337] Supported tasks: ['embed'] 2025-11-03T17:40:20.5736095Z PASSED 2025-11-03T17:40:20.5875402Z models/test_initialization.py::test_can_initialize_large_subset[GritLM] Fork a new process to run a test 17861 2025-11-03T17:40:20.5886220Z Fork a new process to run a test 0 2025-11-03T17:40:20.6165339Z INFO 11-03 17:40:20 [utils.py:239] non-default args: {'load_format': 'dummy', 'gpu_memory_utilization': 0.8, 'disable_log_stats': True, 'hf_overrides': functools.partial(, model_arch='GritLM', exist_overrides={}, use_original_num_layers=False), 'model_impl': 'vllm', 'model': 'parasail-ai/GritLM-7B-vllm'} 2025-11-03T17:40:20.7746267Z 2025-11-03T17:40:20.7747832Z config.json: 0% 0.00/934 [00:00, 'use_cudagraph': True, 'cudagraph_num_of_warmups': 1, 'cudagraph_capture_sizes': [512, 504, 496, 488, 480, 472, 464, 456, 448, 440, 432, 424, 416, 408, 400, 392, 384, 376, 368, 360, 352, 344, 336, 328, 320, 312, 304, 296, 288, 280, 272, 264, 256, 248, 240, 232, 224, 216, 208, 200, 192, 184, 176, 168, 160, 152, 144, 136, 128, 120, 112, 104, 96, 88, 80, 72, 64, 56, 48, 40, 32, 24, 16, 8, 4, 2, 1], 'cudagraph_copy_inputs': False, 'full_cuda_graph': False, 'use_inductor_graph_partition': False, 'pass_config': {}, 'max_capture_size': 512, 'local_cache_dir': None} 2025-11-03T17:40:28.8313921Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:40:28.8344665Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:40:28.8351637Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:40:28.8359331Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:40:28.8366977Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:40:28.8374620Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:40:28.8377278Z (EngineCore_DP0 pid=17940) INFO 11-03 17:40:28 [parallel_state.py:1231] rank 0 in world size 1 is assigned as DP rank 0, PP rank 0, TP rank 0, EP rank 0 2025-11-03T17:40:29.1789747Z (EngineCore_DP0 pid=17940) INFO 11-03 17:40:29 [topk_topp_sampler.py:59] Using FlashInfer for top-p & top-k sampling. 2025-11-03T17:40:29.3230426Z (EngineCore_DP0 pid=17940) INFO 11-03 17:40:29 [gpu_model_runner.py:2840] Starting to load model parasail-ai/GritLM-7B-vllm... 2025-11-03T17:40:29.5759295Z (EngineCore_DP0 pid=17940) INFO 11-03 17:40:29 [gpu_model_runner.py:2870] Loading model from scratch... 2025-11-03T17:40:29.6412030Z (EngineCore_DP0 pid=17940) INFO 11-03 17:40:29 [cuda.py:404] Using Flash Attention backend on V1 engine. 2025-11-03T17:40:30.3339069Z (EngineCore_DP0 pid=17940) INFO 11-03 17:40:30 [gpu_model_runner.py:2902] Model loading took 0.9024 GiB and 0.251635 seconds 2025-11-03T17:40:30.4508268Z (EngineCore_DP0 pid=17940) INFO 11-03 17:40:30 [core.py:142] Disabling chunked prefill for model without KVCache 2025-11-03T17:40:30.5532229Z (EngineCore_DP0 pid=17940) INFO 11-03 17:40:30 [vllm.py:433] Only "last" pooling supports chunked prefill and prefix caching; disabling both. 2025-11-03T17:40:30.5533127Z (EngineCore_DP0 pid=17940) INFO 11-03 17:40:30 [vllm.py:433] Only models using causal attention supports chunked prefill and prefix caching; disabling both. 2025-11-03T17:40:30.8216172Z (EngineCore_DP0 pid=17940) INFO 11-03 17:40:30 [gc_utils.py:40] GC Debug Config. enabled:False,top_objects:-1 2025-11-03T17:40:30.8240609Z INFO 11-03 17:40:30 [llm.py:337] Supported tasks: ['embed', 'encode'] 2025-11-03T17:40:31.3168334Z PASSED 2025-11-03T17:40:31.3309124Z models/test_initialization.py::test_can_initialize_large_subset[MiniCPM3ForCausalLM] Fork a new process to run a test 18006 2025-11-03T17:40:31.3320080Z Fork a new process to run a test 0 2025-11-03T17:40:31.3597860Z INFO 11-03 17:40:31 [utils.py:239] non-default args: {'trust_remote_code': True, 'load_format': 'dummy', 'gpu_memory_utilization': 0.8, 'disable_log_stats': True, 'hf_overrides': functools.partial(, model_arch='MiniCPM3ForCausalLM', exist_overrides={}, use_original_num_layers=False), 'model_impl': 'vllm', 'model': 'openbmb/MiniCPM3-4B'} 2025-11-03T17:40:31.4871690Z 2025-11-03T17:40:31.4874178Z config.json: 0.00B [00:00, ?B/s] 2025-11-03T17:40:31.4874500Z config.json: 1.93kB [00:00, 9.77MB/s] 2025-11-03T17:40:31.5584581Z 2025-11-03T17:40:31.5586636Z configuration_minicpm.py: 0.00B [00:00, ?B/s] 2025-11-03T17:40:31.5587003Z configuration_minicpm.py: 9.23kB [00:00, 53.3MB/s] 2025-11-03T17:40:31.5681644Z A new version of the following files was downloaded from https://huggingface.co/openbmb/MiniCPM3-4B: 2025-11-03T17:40:31.5682169Z - configuration_minicpm.py 2025-11-03T17:40:31.5682794Z . Make sure to double-check they do not contain any added malicious code. To avoid downloading new versions of the code file, you can pin a revision. 2025-11-03T17:40:31.6155800Z INFO 11-03 17:40:31 [config.py:408] Replacing legacy 'type' key with 'rope_type' 2025-11-03T17:40:37.9932035Z INFO 11-03 17:40:37 [model.py:653] Resolved architecture: MiniCPM3ForCausalLM 2025-11-03T17:40:37.9932531Z `torch_dtype` is deprecated! Use `dtype` instead! 2025-11-03T17:40:38.0190703Z INFO 11-03 17:40:38 [model.py:1714] Using max model len 32768 2025-11-03T17:40:38.0192647Z The argument `trust_remote_code` is to be used with Auto classes. It has no effect here and is ignored. 2025-11-03T17:40:38.0779190Z INFO 11-03 17:40:38 [scheduler.py:225] Chunked prefill is enabled with max_num_batched_tokens=8192. 2025-11-03T17:40:38.1278551Z 2025-11-03T17:40:38.1280683Z tokenizer_config.json: 0.00B [00:00, ?B/s] 2025-11-03T17:40:38.1281066Z tokenizer_config.json: 10.4kB [00:00, 50.1MB/s] 2025-11-03T17:40:38.2409652Z 2025-11-03T17:40:38.5517854Z tokenizer.model: 0% 0.00/1.18M [00:00, 'use_cudagraph': True, 'cudagraph_num_of_warmups': 1, 'cudagraph_capture_sizes': [512, 504, 496, 488, 480, 472, 464, 456, 448, 440, 432, 424, 416, 408, 400, 392, 384, 376, 368, 360, 352, 344, 336, 328, 320, 312, 304, 296, 288, 280, 272, 264, 256, 248, 240, 232, 224, 216, 208, 200, 192, 184, 176, 168, 160, 152, 144, 136, 128, 120, 112, 104, 96, 88, 80, 72, 64, 56, 48, 40, 32, 24, 16, 8, 4, 2, 1], 'cudagraph_copy_inputs': False, 'full_cuda_graph': True, 'use_inductor_graph_partition': False, 'pass_config': {}, 'max_capture_size': 512, 'local_cache_dir': None} 2025-11-03T17:40:39.6628189Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:40:39.6655446Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:40:39.6662648Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:40:39.6670253Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:40:39.6677632Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:40:39.6685722Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:40:39.6688552Z (EngineCore_DP0 pid=18084) INFO 11-03 17:40:39 [parallel_state.py:1231] rank 0 in world size 1 is assigned as DP rank 0, PP rank 0, TP rank 0, EP rank 0 2025-11-03T17:40:40.0151658Z (EngineCore_DP0 pid=18084) INFO 11-03 17:40:40 [topk_topp_sampler.py:59] Using FlashInfer for top-p & top-k sampling. 2025-11-03T17:40:40.0857158Z (EngineCore_DP0 pid=18084) INFO 11-03 17:40:40 [gpu_model_runner.py:2840] Starting to load model openbmb/MiniCPM3-4B... 2025-11-03T17:40:40.3464775Z (EngineCore_DP0 pid=18084) INFO 11-03 17:40:40 [gpu_model_runner.py:2870] Loading model from scratch... 2025-11-03T17:40:40.4134462Z (EngineCore_DP0 pid=18084) INFO 11-03 17:40:40 [cuda.py:404] Using Flash Attention backend on V1 engine. 2025-11-03T17:40:40.9894702Z (EngineCore_DP0 pid=18084) INFO 11-03 17:40:40 [gpu_model_runner.py:2902] Model loading took 0.4724 GiB and 0.128176 seconds 2025-11-03T17:40:40.9899108Z (EngineCore_DP0 pid=18084) INFO 11-03 17:40:40 [kv_cache_utils.py:1199] GPU KV cache size: 699,040 tokens 2025-11-03T17:40:40.9899890Z (EngineCore_DP0 pid=18084) INFO 11-03 17:40:40 [kv_cache_utils.py:1204] Maximum concurrency for 32,768 tokens per request: 21.33x 2025-11-03T17:40:41.5370869Z (EngineCore_DP0 pid=18084) INFO 11-03 17:40:41 [gc_utils.py:40] GC Debug Config. enabled:False,top_objects:-1 2025-11-03T17:40:41.5392809Z INFO 11-03 17:40:41 [llm.py:337] Supported tasks: ['generate'] 2025-11-03T17:40:41.9886942Z PASSED 2025-11-03T17:40:42.0028449Z models/test_initialization.py::test_can_initialize_large_subset[SmolLM3ForCausalLM] Fork a new process to run a test 18150 2025-11-03T17:40:42.0038943Z Fork a new process to run a test 0 2025-11-03T17:40:42.0317038Z INFO 11-03 17:40:42 [utils.py:239] non-default args: {'load_format': 'dummy', 'gpu_memory_utilization': 0.8, 'disable_log_stats': True, 'hf_overrides': functools.partial(, model_arch='SmolLM3ForCausalLM', exist_overrides={}, use_original_num_layers=False), 'model_impl': 'vllm', 'model': 'HuggingFaceTB/SmolLM3-3B'} 2025-11-03T17:40:42.1680989Z 2025-11-03T17:40:42.1682950Z config.json: 0.00B [00:00, ?B/s] 2025-11-03T17:40:42.1683257Z config.json: 1.92kB [00:00, 10.1MB/s] 2025-11-03T17:40:42.2959532Z INFO 11-03 17:40:42 [model.py:653] Resolved architecture: SmolLM3ForCausalLM 2025-11-03T17:40:42.2960291Z `torch_dtype` is deprecated! Use `dtype` instead! 2025-11-03T17:40:42.3200996Z INFO 11-03 17:40:42 [model.py:1714] Using max model len 65536 2025-11-03T17:40:42.4876083Z INFO 11-03 17:40:42 [scheduler.py:225] Chunked prefill is enabled with max_num_batched_tokens=8192. 2025-11-03T17:40:42.5192693Z 2025-11-03T17:40:42.5196427Z tokenizer_config.json: 0.00B [00:00, ?B/s] 2025-11-03T17:40:42.5196773Z tokenizer_config.json: 50.4kB [00:00, 154MB/s] 2025-11-03T17:40:42.6139912Z 2025-11-03T17:40:42.9151913Z tokenizer.json: 0% 0.00/17.2M [00:00, 'use_cudagraph': True, 'cudagraph_num_of_warmups': 1, 'cudagraph_capture_sizes': [512, 504, 496, 488, 480, 472, 464, 456, 448, 440, 432, 424, 416, 408, 400, 392, 384, 376, 368, 360, 352, 344, 336, 328, 320, 312, 304, 296, 288, 280, 272, 264, 256, 248, 240, 232, 224, 216, 208, 200, 192, 184, 176, 168, 160, 152, 144, 136, 128, 120, 112, 104, 96, 88, 80, 72, 64, 56, 48, 40, 32, 24, 16, 8, 4, 2, 1], 'cudagraph_copy_inputs': False, 'full_cuda_graph': True, 'use_inductor_graph_partition': False, 'pass_config': {}, 'max_capture_size': 512, 'local_cache_dir': None} 2025-11-03T17:40:44.4279503Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:40:44.4306907Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:40:44.4313652Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:40:44.4322048Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:40:44.4329570Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:40:44.4336983Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:40:44.4339726Z (EngineCore_DP0 pid=18179) INFO 11-03 17:40:44 [parallel_state.py:1231] rank 0 in world size 1 is assigned as DP rank 0, PP rank 0, TP rank 0, EP rank 0 2025-11-03T17:40:44.7675302Z (EngineCore_DP0 pid=18179) INFO 11-03 17:40:44 [topk_topp_sampler.py:59] Using FlashInfer for top-p & top-k sampling. 2025-11-03T17:40:44.8342970Z (EngineCore_DP0 pid=18179) INFO 11-03 17:40:44 [gpu_model_runner.py:2840] Starting to load model HuggingFaceTB/SmolLM3-3B... 2025-11-03T17:40:45.0888745Z (EngineCore_DP0 pid=18179) INFO 11-03 17:40:45 [gpu_model_runner.py:2870] Loading model from scratch... 2025-11-03T17:40:45.0890003Z (EngineCore_DP0 pid=18179) INFO 11-03 17:40:45 [transformers.py:493] Using Transformers backend. 2025-11-03T17:40:45.1253671Z (EngineCore_DP0 pid=18179) `torch_dtype` is deprecated! Use `dtype` instead! 2025-11-03T17:40:45.2072252Z (EngineCore_DP0 pid=18179) INFO 11-03 17:40:45 [cuda.py:404] Using Flash Attention backend on V1 engine. 2025-11-03T17:40:45.7727925Z (EngineCore_DP0 pid=18179) INFO 11-03 17:40:45 [gpu_model_runner.py:2902] Model loading took 0.6358 GiB and 0.177146 seconds 2025-11-03T17:40:45.7732687Z (EngineCore_DP0 pid=18179) INFO 11-03 17:40:45 [kv_cache_utils.py:1199] GPU KV cache size: 5,242,880 tokens 2025-11-03T17:40:45.7733481Z (EngineCore_DP0 pid=18179) INFO 11-03 17:40:45 [kv_cache_utils.py:1204] Maximum concurrency for 65,536 tokens per request: 80.00x 2025-11-03T17:40:47.1271969Z (EngineCore_DP0 pid=18179) INFO 11-03 17:40:47 [gc_utils.py:40] GC Debug Config. enabled:False,top_objects:-1 2025-11-03T17:40:47.1285651Z INFO 11-03 17:40:47 [llm.py:337] Supported tasks: ['generate'] 2025-11-03T17:40:47.6011552Z PASSED 2025-11-03T17:40:47.6154984Z models/test_initialization.py::test_can_initialize_large_subset[BambaForCausalLM] Fork a new process to run a test 18245 2025-11-03T17:40:47.6164857Z Fork a new process to run a test 0 2025-11-03T17:40:47.6446726Z INFO 11-03 17:40:47 [utils.py:239] non-default args: {'load_format': 'dummy', 'gpu_memory_utilization': 0.8, 'disable_log_stats': True, 'hf_overrides': functools.partial(, model_arch='BambaForCausalLM', exist_overrides={}, use_original_num_layers=False), 'model_impl': 'vllm', 'model': 'ibm-ai-platform/Bamba-9B-v1'} 2025-11-03T17:40:47.7509336Z 2025-11-03T17:40:47.7511083Z config.json: 0% 0.00/921 [00:00= mamba page size. 2025-11-03T17:41:14.3860406Z INFO 11-03 17:41:14 [config.py:453] Padding mamba page size by 0.69% to ensure that mamba page size and attention page size are exactly equal. 2025-11-03T17:41:14.4554143Z 2025-11-03T17:41:14.4558256Z tokenizer_config.json: 0.00B [00:00, ?B/s] 2025-11-03T17:41:14.4558655Z tokenizer_config.json: 50.6kB [00:00, 96.8MB/s] 2025-11-03T17:41:14.5546722Z 2025-11-03T17:41:14.5889800Z tokenizer.json: 0.00B [00:00, ?B/s] 2025-11-03T17:41:14.5890160Z tokenizer.json: 9.09MB [00:00, 264MB/s] 2025-11-03T17:41:14.6868209Z 2025-11-03T17:41:14.6869179Z special_tokens_map.json: 0% 0.00/301 [00:00, 'use_cudagraph': True, 'cudagraph_num_of_warmups': 1, 'cudagraph_capture_sizes': [512, 504, 496, 488, 480, 472, 464, 456, 448, 440, 432, 424, 416, 408, 400, 392, 384, 376, 368, 360, 352, 344, 336, 328, 320, 312, 304, 296, 288, 280, 272, 264, 256, 248, 240, 232, 224, 216, 208, 200, 192, 184, 176, 168, 160, 152, 144, 136, 128, 120, 112, 104, 96, 88, 80, 72, 64, 56, 48, 40, 32, 24, 16, 8, 4, 2, 1], 'cudagraph_copy_inputs': False, 'full_cuda_graph': True, 'use_inductor_graph_partition': False, 'pass_config': {}, 'max_capture_size': 512, 'local_cache_dir': None} 2025-11-03T17:41:15.9306683Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:41:15.9334348Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:41:15.9342020Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:41:15.9349196Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:41:15.9357158Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:41:15.9365217Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:41:15.9368038Z (EngineCore_DP0 pid=18303) INFO 11-03 17:41:15 [parallel_state.py:1231] rank 0 in world size 1 is assigned as DP rank 0, PP rank 0, TP rank 0, EP rank 0 2025-11-03T17:41:16.2721338Z (EngineCore_DP0 pid=18303) INFO 11-03 17:41:16 [topk_topp_sampler.py:59] Using FlashInfer for top-p & top-k sampling. 2025-11-03T17:41:16.3365125Z (EngineCore_DP0 pid=18303) INFO 11-03 17:41:16 [gpu_model_runner.py:2840] Starting to load model ibm-ai-platform/Bamba-9B-v1... 2025-11-03T17:41:16.5883021Z (EngineCore_DP0 pid=18303) INFO 11-03 17:41:16 [gpu_model_runner.py:2870] Loading model from scratch... 2025-11-03T17:41:17.1544027Z (EngineCore_DP0 pid=18303) INFO 11-03 17:41:17 [gpu_model_runner.py:2902] Model loading took 2.4767 GiB and 0.062165 seconds 2025-11-03T17:41:17.1549442Z (EngineCore_DP0 pid=18303) INFO 11-03 17:41:17 [kv_cache_utils.py:1199] GPU KV cache size: 20,332,544 tokens 2025-11-03T17:41:17.1550680Z (EngineCore_DP0 pid=18303) INFO 11-03 17:41:17 [kv_cache_utils.py:1204] Maximum concurrency for 4,096 tokens per request: 4964.00x 2025-11-03T17:41:17.8281473Z (EngineCore_DP0 pid=18303) INFO 11-03 17:41:17 [gc_utils.py:40] GC Debug Config. enabled:False,top_objects:-1 2025-11-03T17:41:17.8302607Z INFO 11-03 17:41:17 [llm.py:337] Supported tasks: ['generate'] 2025-11-03T17:41:18.2793645Z PASSED 2025-11-03T17:41:18.2933364Z models/test_initialization.py::test_can_initialize_large_subset[BertForTokenClassification] Fork a new process to run a test 18369 2025-11-03T17:41:18.2943904Z Fork a new process to run a test 0 2025-11-03T17:41:18.3214064Z INFO 11-03 17:41:18 [utils.py:239] non-default args: {'load_format': 'dummy', 'gpu_memory_utilization': 0.8, 'disable_log_stats': True, 'hf_overrides': functools.partial(, model_arch='BertForTokenClassification', exist_overrides={}, use_original_num_layers=False), 'model_impl': 'vllm', 'model': 'boltuix/NeuroBERT-NER'} 2025-11-03T17:41:18.5656977Z 2025-11-03T17:41:18.5658985Z config.json: 0.00B [00:00, ?B/s] 2025-11-03T17:41:18.5659303Z config.json: 1.84kB [00:00, 10.9MB/s] 2025-11-03T17:41:25.1630665Z INFO 11-03 17:41:25 [model.py:915] Resolved `--runner auto` to `--runner pooling`. Pass the value explicitly to silence this message. 2025-11-03T17:41:25.1631409Z INFO 11-03 17:41:25 [model.py:653] Resolved architecture: BertForTokenClassification 2025-11-03T17:41:25.1631884Z `torch_dtype` is deprecated! Use `dtype` instead! 2025-11-03T17:41:25.1881011Z INFO 11-03 17:41:25 [model.py:1939] Downcasting torch.float32 to torch.float16. 2025-11-03T17:41:25.2254722Z 2025-11-03T17:41:25.2255864Z tokenizer_config.json: 0.00B [00:00, ?B/s] 2025-11-03T17:41:25.2256266Z tokenizer_config.json: 1.30kB [00:00, 6.67MB/s] 2025-11-03T17:41:25.2264134Z INFO 11-03 17:41:25 [model.py:1714] Using max model len 512 2025-11-03T17:41:25.2650338Z INFO 11-03 17:41:25 [arg_utils.py:1725] (Disabling) chunked prefill by default 2025-11-03T17:41:25.2650843Z INFO 11-03 17:41:25 [arg_utils.py:1728] (Disabling) prefix caching by default 2025-11-03T17:41:25.4067445Z INFO 11-03 17:41:25 [vllm.py:433] Only "last" pooling supports chunked prefill and prefix caching; disabling both. 2025-11-03T17:41:25.5648201Z 2025-11-03T17:41:25.5675800Z vocab.txt: 0.00B [00:00, ?B/s] 2025-11-03T17:41:25.5676119Z vocab.txt: 232kB [00:00, 83.9MB/s] 2025-11-03T17:41:25.6108312Z 2025-11-03T17:41:25.6146454Z tokenizer.json: 0.00B [00:00, ?B/s] 2025-11-03T17:41:25.6147049Z tokenizer.json: 711kB [00:00, 189MB/s] 2025-11-03T17:41:25.6878862Z 2025-11-03T17:41:25.6879402Z special_tokens_map.json: 0% 0.00/695 [00:00, 'use_cudagraph': True, 'cudagraph_num_of_warmups': 1, 'cudagraph_capture_sizes': [512, 504, 496, 488, 480, 472, 464, 456, 448, 440, 432, 424, 416, 408, 400, 392, 384, 376, 368, 360, 352, 344, 336, 328, 320, 312, 304, 296, 288, 280, 272, 264, 256, 248, 240, 232, 224, 216, 208, 200, 192, 184, 176, 168, 160, 152, 144, 136, 128, 120, 112, 104, 96, 88, 80, 72, 64, 56, 48, 40, 32, 24, 16, 8, 4, 2, 1], 'cudagraph_copy_inputs': False, 'full_cuda_graph': False, 'use_inductor_graph_partition': False, 'pass_config': {}, 'max_capture_size': 512, 'local_cache_dir': None} 2025-11-03T17:41:26.5828822Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:41:26.5859616Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:41:26.5867003Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:41:26.5874666Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:41:26.5881888Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:41:26.5889806Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:41:26.5892492Z (EngineCore_DP0 pid=18428) INFO 11-03 17:41:26 [parallel_state.py:1231] rank 0 in world size 1 is assigned as DP rank 0, PP rank 0, TP rank 0, EP rank 0 2025-11-03T17:41:26.9210942Z (EngineCore_DP0 pid=18428) INFO 11-03 17:41:26 [topk_topp_sampler.py:59] Using FlashInfer for top-p & top-k sampling. 2025-11-03T17:41:26.9312676Z (EngineCore_DP0 pid=18428) INFO 11-03 17:41:26 [gpu_model_runner.py:2840] Starting to load model boltuix/NeuroBERT-NER... 2025-11-03T17:41:27.1810074Z (EngineCore_DP0 pid=18428) INFO 11-03 17:41:27 [gpu_model_runner.py:2870] Loading model from scratch... 2025-11-03T17:41:27.1858928Z (EngineCore_DP0 pid=18428) INFO 11-03 17:41:27 [cuda.py:404] Using Flash Attention backend on V1 engine. 2025-11-03T17:41:27.7528457Z (EngineCore_DP0 pid=18428) INFO 11-03 17:41:27 [gpu_model_runner.py:2902] Model loading took 0.0163 GiB and 0.064606 seconds 2025-11-03T17:41:27.8513880Z (EngineCore_DP0 pid=18428) INFO 11-03 17:41:27 [core.py:142] Disabling chunked prefill for model without KVCache 2025-11-03T17:41:27.8541654Z (EngineCore_DP0 pid=18428) INFO 11-03 17:41:27 [vllm.py:433] Only "last" pooling supports chunked prefill and prefix caching; disabling both. 2025-11-03T17:41:28.1155084Z (EngineCore_DP0 pid=18428) INFO 11-03 17:41:28 [gc_utils.py:40] GC Debug Config. enabled:False,top_objects:-1 2025-11-03T17:41:28.1169983Z INFO 11-03 17:41:28 [llm.py:337] Supported tasks: ['encode'] 2025-11-03T17:41:28.5330376Z PASSED 2025-11-03T17:41:28.5470379Z models/test_initialization.py::test_can_initialize_large_subset[Ovis2_5] Fork a new process to run a test 18493 2025-11-03T17:41:28.5481793Z Fork a new process to run a test 0 2025-11-03T17:41:28.5760031Z INFO 11-03 17:41:28 [utils.py:239] non-default args: {'trust_remote_code': True, 'load_format': 'dummy', 'gpu_memory_utilization': 0.8, 'disable_log_stats': True, 'hf_overrides': functools.partial(, model_arch='Ovis2_5', exist_overrides={}, use_original_num_layers=False), 'model_impl': 'vllm', 'model': 'AIDC-AI/Ovis2.5-2B'} 2025-11-03T17:41:28.6648536Z 2025-11-03T17:41:28.6650562Z config.json: 0.00B [00:00, ?B/s] 2025-11-03T17:41:28.6650885Z config.json: 2.00kB [00:00, 11.8MB/s] 2025-11-03T17:41:28.7447540Z 2025-11-03T17:41:28.7449407Z configuration_ovis2_5.py: 0.00B [00:00, ?B/s] 2025-11-03T17:41:28.7449805Z configuration_ovis2_5.py: 3.94kB [00:00, 15.5MB/s] 2025-11-03T17:41:28.7543720Z A new version of the following files was downloaded from https://huggingface.co/AIDC-AI/Ovis2.5-2B: 2025-11-03T17:41:28.7544320Z - configuration_ovis2_5.py 2025-11-03T17:41:28.7544950Z . Make sure to double-check they do not contain any added malicious code. To avoid downloading new versions of the code file, you can pin a revision. 2025-11-03T17:41:28.9817754Z 2025-11-03T17:41:28.9818782Z preprocessor_config.json: 0% 0.00/394 [00:00, 'use_cudagraph': True, 'cudagraph_num_of_warmups': 1, 'cudagraph_capture_sizes': [512, 504, 496, 488, 480, 472, 464, 456, 448, 440, 432, 424, 416, 408, 400, 392, 384, 376, 368, 360, 352, 344, 336, 328, 320, 312, 304, 296, 288, 280, 272, 264, 256, 248, 240, 232, 224, 216, 208, 200, 192, 184, 176, 168, 160, 152, 144, 136, 128, 120, 112, 104, 96, 88, 80, 72, 64, 56, 48, 40, 32, 24, 16, 8, 4, 2, 1], 'cudagraph_copy_inputs': False, 'full_cuda_graph': True, 'use_inductor_graph_partition': False, 'pass_config': {}, 'max_capture_size': 512, 'local_cache_dir': None} 2025-11-03T17:41:37.7211766Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:41:37.7239395Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:41:37.7246360Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:41:37.7255042Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:41:37.7262702Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:41:37.7270618Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:41:37.7273481Z (EngineCore_DP0 pid=18571) INFO 11-03 17:41:37 [parallel_state.py:1231] rank 0 in world size 1 is assigned as DP rank 0, PP rank 0, TP rank 0, EP rank 0 2025-11-03T17:41:38.0729957Z (EngineCore_DP0 pid=18571) INFO 11-03 17:41:38 [topk_topp_sampler.py:59] Using FlashInfer for top-p & top-k sampling. 2025-11-03T17:41:38.3185951Z (EngineCore_DP0 pid=18571) Using a slow image processor as `use_fast` is unset and a slow processor was saved with this model. `use_fast=True` will be the default behavior in v4.52, even if the model was saved with a slow processor. This will result in minor differences in outputs. You'll still be able to use a slow processor with `use_fast=False`. 2025-11-03T17:41:39.6779570Z (EngineCore_DP0 pid=18571) 2025-11-03T17:41:39.6779971Z chat_template.json: 0.00B [00:00, ?B/s] 2025-11-03T17:41:39.6780627Z chat_template.json: 1.69kB [00:00, 13.6MB/s] 2025-11-03T17:41:40.7348273Z (EngineCore_DP0 pid=18571) INFO 11-03 17:41:40 [gpu_model_runner.py:2840] Starting to load model AIDC-AI/Ovis2.5-2B... 2025-11-03T17:41:41.0843103Z (EngineCore_DP0 pid=18571) INFO 11-03 17:41:41 [gpu_model_runner.py:2870] Loading model from scratch... 2025-11-03T17:41:41.1112498Z (EngineCore_DP0 pid=18571) INFO 11-03 17:41:41 [cuda.py:404] Using Flash Attention backend on V1 engine. 2025-11-03T17:41:41.7360772Z (EngineCore_DP0 pid=18571) INFO 11-03 17:41:41 [gpu_model_runner.py:2902] Model loading took 2.2769 GiB and 0.127159 seconds 2025-11-03T17:41:41.7365323Z (EngineCore_DP0 pid=18571) INFO 11-03 17:41:41 [kv_cache_utils.py:1199] GPU KV cache size: 2,621,440 tokens 2025-11-03T17:41:41.7366252Z (EngineCore_DP0 pid=18571) INFO 11-03 17:41:41 [kv_cache_utils.py:1204] Maximum concurrency for 40,960 tokens per request: 64.00x 2025-11-03T17:41:43.1366824Z (EngineCore_DP0 pid=18571) INFO 11-03 17:41:43 [gc_utils.py:40] GC Debug Config. enabled:False,top_objects:-1 2025-11-03T17:41:43.1413051Z INFO 11-03 17:41:43 [llm.py:337] Supported tasks: ['generate'] 2025-11-03T17:41:43.6196235Z PASSED 2025-11-03T17:41:43.6335897Z models/test_initialization.py::test_can_initialize_large_subset[EagleMiniCPMForCausalLM] Fork a new process to run a test 18653 2025-11-03T17:41:43.6346968Z Fork a new process to run a test 0 2025-11-03T17:41:43.6348509Z Model is not available online 2025-11-03T17:41:43.9215636Z PASSED 2025-11-03T17:41:43.9354811Z models/test_initialization.py::test_can_initialize_large_subset[Tarsier2ForConditionalGeneration] Fork a new process to run a test 18654 2025-11-03T17:41:43.9365907Z Fork a new process to run a test 0 2025-11-03T17:41:43.9638421Z INFO 11-03 17:41:43 [utils.py:239] non-default args: {'load_format': 'dummy', 'gpu_memory_utilization': 0.8, 'disable_log_stats': True, 'hf_overrides': functools.partial(, model_arch='Tarsier2ForConditionalGeneration', exist_overrides={'architectures': ['Tarsier2ForConditionalGeneration']}, use_original_num_layers=False), 'model_impl': 'vllm', 'model': 'omni-research/Tarsier2-Recap-7b'} 2025-11-03T17:41:52.2561700Z 2025-11-03T17:41:52.5838556Z config.json: 0% 0.00/5.52k [00:00, 'use_cudagraph': True, 'cudagraph_num_of_warmups': 1, 'cudagraph_capture_sizes': [512, 504, 496, 488, 480, 472, 464, 456, 448, 440, 432, 424, 416, 408, 400, 392, 384, 376, 368, 360, 352, 344, 336, 328, 320, 312, 304, 296, 288, 280, 272, 264, 256, 248, 240, 232, 224, 216, 208, 200, 192, 184, 176, 168, 160, 152, 144, 136, 128, 120, 112, 104, 96, 88, 80, 72, 64, 56, 48, 40, 32, 24, 16, 8, 4, 2, 1], 'cudagraph_copy_inputs': False, 'full_cuda_graph': True, 'use_inductor_graph_partition': False, 'pass_config': {}, 'max_capture_size': 512, 'local_cache_dir': None} 2025-11-03T17:42:02.5173264Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:42:02.5202972Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:42:02.5210569Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:42:02.5218299Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:42:02.5227058Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:42:02.5235140Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:42:02.5237879Z (EngineCore_DP0 pid=18740) INFO 11-03 17:42:02 [parallel_state.py:1231] rank 0 in world size 1 is assigned as DP rank 0, PP rank 0, TP rank 0, EP rank 0 2025-11-03T17:42:02.8586708Z (EngineCore_DP0 pid=18740) INFO 11-03 17:42:02 [topk_topp_sampler.py:59] Using FlashInfer for top-p & top-k sampling. 2025-11-03T17:42:04.5058888Z (EngineCore_DP0 pid=18740) INFO 11-03 17:42:04 [gpu_model_runner.py:2840] Starting to load model omni-research/Tarsier2-Recap-7b... 2025-11-03T17:42:04.7615415Z (EngineCore_DP0 pid=18740) INFO 11-03 17:42:04 [gpu_model_runner.py:2870] Loading model from scratch... 2025-11-03T17:42:04.8863817Z (EngineCore_DP0 pid=18740) INFO 11-03 17:42:04 [cuda.py:404] Using Flash Attention backend on V1 engine. 2025-11-03T17:42:05.5135094Z (EngineCore_DP0 pid=18740) INFO 11-03 17:42:05 [gpu_model_runner.py:2902] Model loading took 15.5314 GiB and 0.241063 seconds 2025-11-03T17:42:05.5141896Z (EngineCore_DP0 pid=18740) INFO 11-03 17:42:05 [kv_cache_utils.py:1199] GPU KV cache size: 187,232 tokens 2025-11-03T17:42:05.5142817Z (EngineCore_DP0 pid=18740) INFO 11-03 17:42:05 [kv_cache_utils.py:1204] Maximum concurrency for 32,768 tokens per request: 5.71x 2025-11-03T17:42:07.5530703Z (EngineCore_DP0 pid=18740) INFO 11-03 17:42:07 [gc_utils.py:40] GC Debug Config. enabled:False,top_objects:-1 2025-11-03T17:42:07.5547076Z INFO 11-03 17:42:07 [llm.py:337] Supported tasks: ['generate'] 2025-11-03T17:42:08.0445795Z PASSED 2025-11-03T17:42:08.0584854Z models/test_initialization.py::test_can_initialize_large_subset[Glm4vMoeForConditionalGeneration] Fork a new process to run a test 18806 2025-11-03T17:42:08.0595403Z Fork a new process to run a test 0 2025-11-03T17:42:08.0873789Z INFO 11-03 17:42:08 [utils.py:239] non-default args: {'load_format': 'dummy', 'gpu_memory_utilization': 0.8, 'disable_log_stats': True, 'hf_overrides': functools.partial(, model_arch='Glm4vMoeForConditionalGeneration', exist_overrides={}, use_original_num_layers=False), 'model_impl': 'vllm', 'model': 'zai-org/GLM-4.5V'} 2025-11-03T17:42:08.1643311Z 2025-11-03T17:42:08.1645235Z config.json: 0.00B [00:00, ?B/s] 2025-11-03T17:42:08.1645563Z config.json: 1.85kB [00:00, 12.3MB/s] 2025-11-03T17:42:08.2793674Z 2025-11-03T17:42:08.2794124Z preprocessor_config.json: 0% 0.00/364 [00:00, 'use_cudagraph': True, 'cudagraph_num_of_warmups': 1, 'cudagraph_capture_sizes': [512, 504, 496, 488, 480, 472, 464, 456, 448, 440, 432, 424, 416, 408, 400, 392, 384, 376, 368, 360, 352, 344, 336, 328, 320, 312, 304, 296, 288, 280, 272, 264, 256, 248, 240, 232, 224, 216, 208, 200, 192, 184, 176, 168, 160, 152, 144, 136, 128, 120, 112, 104, 96, 88, 80, 72, 64, 56, 48, 40, 32, 24, 16, 8, 4, 2, 1], 'cudagraph_copy_inputs': False, 'full_cuda_graph': True, 'use_inductor_graph_partition': False, 'pass_config': {}, 'max_capture_size': 512, 'local_cache_dir': None} 2025-11-03T17:42:17.3265976Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:42:17.3295886Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:42:17.3304013Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:42:17.3311529Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:42:17.3319487Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:42:17.3328356Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:42:17.3331162Z (EngineCore_DP0 pid=18889) INFO 11-03 17:42:17 [parallel_state.py:1231] rank 0 in world size 1 is assigned as DP rank 0, PP rank 0, TP rank 0, EP rank 0 2025-11-03T17:42:17.6682829Z (EngineCore_DP0 pid=18889) INFO 11-03 17:42:17 [topk_topp_sampler.py:59] Using FlashInfer for top-p & top-k sampling. 2025-11-03T17:42:17.9425832Z (EngineCore_DP0 pid=18889) Using a slow image processor as `use_fast` is unset and a slow processor was saved with this model. `use_fast=True` will be the default behavior in v4.52, even if the model was saved with a slow processor. This will result in minor differences in outputs. You'll still be able to use a slow processor with `use_fast=False`. 2025-11-03T17:42:19.2249564Z (EngineCore_DP0 pid=18889) 2025-11-03T17:42:19.2250016Z video_preprocessor_config.json: 0% 0.00/365 [00:00, model_arch='MiniMaxM1ForCausalLM', exist_overrides={}, use_original_num_layers=False), 'model_impl': 'vllm', 'model': 'MiniMaxAI/MiniMax-M1-40k'} 2025-11-03T17:42:42.7688716Z 2025-11-03T17:42:42.7691214Z config.json: 0.00B [00:00, ?B/s] 2025-11-03T17:42:42.7691538Z config.json: 1.87kB [00:00, 10.1MB/s] 2025-11-03T17:42:42.8419655Z 2025-11-03T17:42:42.8422011Z configuration_minimax_m1.py: 0.00B [00:00, ?B/s] 2025-11-03T17:42:42.8422422Z configuration_minimax_m1.py: 7.30kB [00:00, 51.6MB/s] 2025-11-03T17:42:42.8509569Z A new version of the following files was downloaded from https://huggingface.co/MiniMaxAI/MiniMax-M1-40k: 2025-11-03T17:42:42.8510351Z - configuration_minimax_m1.py 2025-11-03T17:42:42.8511265Z . Make sure to double-check they do not contain any added malicious code. To avoid downloading new versions of the code file, you can pin a revision. 2025-11-03T17:42:42.8874897Z You are using a model of type minimax_m1 to instantiate a model of type MiniMaxM1. This is not supported for all configurations of models and can yield errors. 2025-11-03T17:42:43.0772663Z INFO 11-03 17:42:43 [model.py:653] Resolved architecture: MiniMaxM1ForCausalLM 2025-11-03T17:42:43.0773154Z `torch_dtype` is deprecated! Use `dtype` instead! 2025-11-03T17:42:43.1686565Z 2025-11-03T17:42:43.1707433Z model.safetensors.index.json: 0.00B [00:00, ?B/s] 2025-11-03T17:42:43.1716413Z model.safetensors.index.json: 823kB [00:00, 401MB/s] 2025-11-03T17:42:43.1977034Z 2025-11-03T17:42:43.3082206Z Parse safetensors files: 0% 0/413 [00:00= mamba page size. 2025-11-03T17:42:44.7750807Z 2025-11-03T17:42:44.7752851Z tokenizer_config.json: 0.00B [00:00, ?B/s] 2025-11-03T17:42:44.7753311Z tokenizer_config.json: 1.97kB [00:00, 15.9MB/s] 2025-11-03T17:42:44.8563806Z 2025-11-03T17:42:44.8767269Z vocab.json: 0.00B [00:00, ?B/s] 2025-11-03T17:42:44.8767597Z vocab.json: 4.71MB [00:00, 232MB/s] 2025-11-03T17:42:44.9091479Z 2025-11-03T17:42:44.9212391Z merges.txt: 0.00B [00:00, ?B/s] 2025-11-03T17:42:44.9212834Z merges.txt: 2.41MB [00:00, 200MB/s] 2025-11-03T17:42:44.9602561Z 2025-11-03T17:42:45.0035869Z tokenizer.json: 0.00B [00:00, ?B/s] 2025-11-03T17:42:45.0036287Z tokenizer.json: 9.73MB [00:00, 225MB/s] 2025-11-03T17:42:45.8078189Z You are using a model of type minimax_m1 to instantiate a model of type MiniMaxM1. This is not supported for all configurations of models and can yield errors. 2025-11-03T17:42:45.8314657Z (EngineCore_DP0 pid=18999) INFO 11-03 17:42:45 [core.py:727] Waiting for init message from front-end. 2025-11-03T17:42:45.8353992Z (EngineCore_DP0 pid=18999) INFO 11-03 17:42:45 [core.py:94] Initializing a V1 LLM engine (v0.11.0rc2.dev389+ge51928192.d20251103) with config: model='MiniMaxAI/MiniMax-M1-40k', speculative_config=None, tokenizer='MiniMaxAI/MiniMax-M1-40k', skip_tokenizer_init=False, tokenizer_mode=auto, revision=None, tokenizer_revision=None, trust_remote_code=True, dtype=torch.bfloat16, max_seq_len=10240000, download_dir=None, load_format=dummy, tensor_parallel_size=1, pipeline_parallel_size=1, data_parallel_size=1, disable_custom_all_reduce=False, quantization=None, enforce_eager=False, kv_cache_dtype=auto, device_config=cuda, structured_outputs_config=StructuredOutputsConfig(backend='auto', disable_fallback=False, disable_any_whitespace=False, disable_additional_properties=False, reasoning_parser=''), observability_config=ObservabilityConfig(show_hidden_metrics_for_version=None, otlp_traces_endpoint=None, collect_detailed_traces=None), seed=0, served_model_name=MiniMaxAI/MiniMax-M1-40k, enable_prefix_caching=False, chunked_prefill_enabled=True, pooler_config=None, compilation_config={'level': 3, 'debug_dump_path': None, 'cache_dir': '', 'backend': '', 'custom_ops': [], 'splitting_ops': ['vllm.unified_attention', 'vllm.unified_attention_with_output', 'vllm.unified_mla_attention', 'vllm.unified_mla_attention_with_output', 'vllm.mamba_mixer2', 'vllm.mamba_mixer', 'vllm.short_conv', 'vllm.linear_attention', 'vllm.plamo2_mamba_mixer', 'vllm.gdn_attention', 'vllm.sparse_attn_indexer'], 'use_inductor': True, 'compile_sizes': [], 'inductor_compile_config': {'enable_auto_functionalized_v2': False}, 'inductor_passes': {}, 'cudagraph_mode': , 'use_cudagraph': True, 'cudagraph_num_of_warmups': 1, 'cudagraph_capture_sizes': [512, 504, 496, 488, 480, 472, 464, 456, 448, 440, 432, 424, 416, 408, 400, 392, 384, 376, 368, 360, 352, 344, 336, 328, 320, 312, 304, 296, 288, 280, 272, 264, 256, 248, 240, 232, 224, 216, 208, 200, 192, 184, 176, 168, 160, 152, 144, 136, 128, 120, 112, 104, 96, 88, 80, 72, 64, 56, 48, 40, 32, 24, 16, 8, 4, 2, 1], 'cudagraph_copy_inputs': False, 'full_cuda_graph': True, 'use_inductor_graph_partition': False, 'pass_config': {}, 'max_capture_size': 512, 'local_cache_dir': None} 2025-11-03T17:42:46.5264726Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:42:46.5292987Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:42:46.5301218Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:42:46.5310660Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:42:46.5319728Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:42:46.5328255Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:42:46.5331095Z (EngineCore_DP0 pid=18999) INFO 11-03 17:42:46 [parallel_state.py:1231] rank 0 in world size 1 is assigned as DP rank 0, PP rank 0, TP rank 0, EP rank 0 2025-11-03T17:42:46.8738879Z (EngineCore_DP0 pid=18999) INFO 11-03 17:42:46 [topk_topp_sampler.py:59] Using FlashInfer for top-p & top-k sampling. 2025-11-03T17:42:47.8406801Z (EngineCore_DP0 pid=18999) INFO 11-03 17:42:47 [gpu_model_runner.py:2840] Starting to load model MiniMaxAI/MiniMax-M1-40k... 2025-11-03T17:42:48.1025640Z (EngineCore_DP0 pid=18999) INFO 11-03 17:42:48 [gpu_model_runner.py:2870] Loading model from scratch... 2025-11-03T17:42:48.7387623Z (EngineCore_DP0 pid=18999) INFO 11-03 17:42:48 [gpu_model_runner.py:2902] Model loading took 7.9713 GiB and 0.110557 seconds 2025-11-03T17:42:48.7392434Z (EngineCore_DP0 pid=18999) INFO 11-03 17:42:48 [kv_cache_utils.py:1199] GPU KV cache size: 52,428,800,000 tokens 2025-11-03T17:42:48.7393175Z (EngineCore_DP0 pid=18999) INFO 11-03 17:42:48 [kv_cache_utils.py:1204] Maximum concurrency for 10,240,000 tokens per request: 5120.00x 2025-11-03T17:42:49.6140425Z (EngineCore_DP0 pid=18999) INFO 11-03 17:42:49 [gc_utils.py:40] GC Debug Config. enabled:False,top_objects:-1 2025-11-03T17:42:49.6153429Z INFO 11-03 17:42:49 [llm.py:337] Supported tasks: ['generate'] 2025-11-03T17:42:51.0107123Z PASSED 2025-11-03T17:42:51.0246739Z models/test_initialization.py::test_can_initialize_large_subset[GptOssForCausalLM] Fork a new process to run a test 19065 2025-11-03T17:42:51.0257561Z Fork a new process to run a test 0 2025-11-03T17:42:51.0535900Z INFO 11-03 17:42:51 [utils.py:239] non-default args: {'load_format': 'dummy', 'gpu_memory_utilization': 0.8, 'disable_log_stats': True, 'hf_overrides': functools.partial(, model_arch='GptOssForCausalLM', exist_overrides={}, use_original_num_layers=False), 'model_impl': 'vllm', 'model': 'lmsys/gpt-oss-20b-bf16'} 2025-11-03T17:42:51.1974371Z 2025-11-03T17:42:51.1976359Z config.json: 0.00B [00:00, ?B/s] 2025-11-03T17:42:51.1976701Z config.json: 1.59kB [00:00, 8.79MB/s] 2025-11-03T17:42:57.6641607Z INFO 11-03 17:42:57 [model.py:653] Resolved architecture: GptOssForCausalLM 2025-11-03T17:42:57.6642103Z `torch_dtype` is deprecated! Use `dtype` instead! 2025-11-03T17:42:57.6889427Z INFO 11-03 17:42:57 [model.py:1714] Using max model len 131072 2025-11-03T17:42:57.8607225Z INFO 11-03 17:42:57 [scheduler.py:225] Chunked prefill is enabled with max_num_batched_tokens=8192. 2025-11-03T17:42:57.8608250Z INFO 11-03 17:42:57 [config.py:263] Overriding max cuda graph capture size to 992 for performance. 2025-11-03T17:42:57.8987621Z 2025-11-03T17:42:57.8989138Z tokenizer_config.json: 0.00B [00:00, ?B/s] 2025-11-03T17:42:57.8989539Z tokenizer_config.json: 4.20kB [00:00, 26.5MB/s] 2025-11-03T17:42:58.0046337Z 2025-11-03T17:42:58.2095519Z tokenizer.json: 0% 0.00/27.9M [00:00, 'use_cudagraph': True, 'cudagraph_num_of_warmups': 1, 'cudagraph_capture_sizes': [992, 976, 960, 944, 928, 912, 896, 880, 864, 848, 832, 816, 800, 784, 768, 752, 736, 720, 704, 688, 672, 656, 640, 624, 608, 592, 576, 560, 544, 528, 512, 496, 480, 464, 448, 432, 416, 400, 384, 368, 352, 336, 320, 304, 288, 272, 256, 248, 240, 232, 224, 216, 208, 200, 192, 184, 176, 168, 160, 152, 144, 136, 128, 120, 112, 104, 96, 88, 80, 72, 64, 56, 48, 40, 32, 24, 16, 8, 4, 2, 1], 'cudagraph_copy_inputs': False, 'full_cuda_graph': True, 'use_inductor_graph_partition': False, 'pass_config': {}, 'max_capture_size': 992, 'local_cache_dir': None} 2025-11-03T17:42:59.9407790Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:42:59.9435056Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:42:59.9441753Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:42:59.9449506Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:42:59.9457091Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:42:59.9465368Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:42:59.9467959Z (EngineCore_DP0 pid=19143) INFO 11-03 17:42:59 [parallel_state.py:1231] rank 0 in world size 1 is assigned as DP rank 0, PP rank 0, TP rank 0, EP rank 0 2025-11-03T17:43:00.2757138Z (EngineCore_DP0 pid=19143) INFO 11-03 17:43:00 [topk_topp_sampler.py:59] Using FlashInfer for top-p & top-k sampling. 2025-11-03T17:43:00.3724007Z (EngineCore_DP0 pid=19143) INFO 11-03 17:43:00 [gpu_model_runner.py:2840] Starting to load model lmsys/gpt-oss-20b-bf16... 2025-11-03T17:43:00.6260816Z (EngineCore_DP0 pid=19143) INFO 11-03 17:43:00 [gpu_model_runner.py:2870] Loading model from scratch... 2025-11-03T17:43:00.7343774Z (EngineCore_DP0 pid=19143) INFO 11-03 17:43:00 [cuda.py:359] Using Triton backend on V1 engine. 2025-11-03T17:43:01.2913257Z (EngineCore_DP0 pid=19143) INFO 11-03 17:43:01 [gpu_model_runner.py:2902] Model loading took 2.3320 GiB and 0.167044 seconds 2025-11-03T17:43:01.2918244Z (EngineCore_DP0 pid=19143) INFO 11-03 17:43:01 [kv_cache_utils.py:1199] GPU KV cache size: 5,242,880 tokens 2025-11-03T17:43:01.2919156Z (EngineCore_DP0 pid=19143) INFO 11-03 17:43:01 [kv_cache_utils.py:1204] Maximum concurrency for 131,072 tokens per request: 628.94x 2025-11-03T17:43:02.9213556Z (EngineCore_DP0 pid=19143) INFO 11-03 17:43:02 [gc_utils.py:40] GC Debug Config. enabled:False,top_objects:-1 2025-11-03T17:43:02.9228548Z INFO 11-03 17:43:02 [llm.py:337] Supported tasks: ['generate'] 2025-11-03T17:43:03.4373654Z PASSED 2025-11-03T17:43:03.4513680Z models/test_initialization.py::test_can_initialize_large_subset[Eagle3LlamaForCausalLM] Fork a new process to run a test 19225 2025-11-03T17:43:03.4523974Z Fork a new process to run a test 0 2025-11-03T17:43:03.4803340Z INFO 11-03 17:43:03 [utils.py:239] non-default args: {'tokenizer': 'meta-llama/Llama-3.1-8B-Instruct', 'trust_remote_code': True, 'load_format': 'dummy', 'max_model_len': 10240, 'gpu_memory_utilization': 0.8, 'disable_log_stats': True, 'hf_overrides': functools.partial(, model_arch='Eagle3LlamaForCausalLM', exist_overrides={}, use_original_num_layers=True), 'speculative_config': {'model': 'yuhuili/EAGLE3-LLaMA3.1-Instruct-8B', 'num_speculative_tokens': 1}, 'model_impl': 'vllm', 'model': 'meta-llama/Llama-3.1-8B-Instruct'} 2025-11-03T17:43:03.6379162Z 2025-11-03T17:43:03.6379997Z config.json: 0% 0.00/855 [00:00, 'use_cudagraph': True, 'cudagraph_num_of_warmups': 1, 'cudagraph_capture_sizes': [512, 504, 496, 488, 480, 472, 464, 456, 448, 440, 432, 424, 416, 408, 400, 392, 384, 376, 368, 360, 352, 344, 336, 328, 320, 312, 304, 296, 288, 280, 272, 264, 256, 248, 240, 232, 224, 216, 208, 200, 192, 184, 176, 168, 160, 152, 144, 136, 128, 120, 112, 104, 96, 88, 80, 72, 64, 56, 48, 40, 32, 24, 16, 8, 4, 2, 1], 'cudagraph_copy_inputs': False, 'full_cuda_graph': True, 'use_inductor_graph_partition': False, 'pass_config': {}, 'max_capture_size': 512, 'local_cache_dir': None} 2025-11-03T17:43:05.7487701Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:43:05.7515675Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:43:05.7524046Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:43:05.7532113Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:43:05.7539889Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:43:05.7548503Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:43:05.7551409Z (EngineCore_DP0 pid=19233) INFO 11-03 17:43:05 [parallel_state.py:1231] rank 0 in world size 1 is assigned as DP rank 0, PP rank 0, TP rank 0, EP rank 0 2025-11-03T17:43:06.0851080Z (EngineCore_DP0 pid=19233) INFO 11-03 17:43:06 [topk_topp_sampler.py:59] Using FlashInfer for top-p & top-k sampling. 2025-11-03T17:43:06.1002318Z (EngineCore_DP0 pid=19233) WARNING 11-03 17:43:06 [__init__.py:194] min_p, logit_bias, and min_tokens parameters won't currently work with speculative decoding enabled. 2025-11-03T17:43:06.1383088Z (EngineCore_DP0 pid=19233) INFO 11-03 17:43:06 [gpu_model_runner.py:2840] Starting to load model meta-llama/Llama-3.1-8B-Instruct... 2025-11-03T17:43:06.3898291Z (EngineCore_DP0 pid=19233) INFO 11-03 17:43:06 [gpu_model_runner.py:2870] Loading model from scratch... 2025-11-03T17:43:06.4737436Z (EngineCore_DP0 pid=19233) INFO 11-03 17:43:06 [cuda.py:404] Using Flash Attention backend on V1 engine. 2025-11-03T17:43:06.5867814Z (EngineCore_DP0 pid=19233) INFO 11-03 17:43:06 [gpu_model_runner.py:2879] Loading drafter model... 2025-11-03T17:43:06.6535216Z (EngineCore_DP0 pid=19233) INFO 11-03 17:43:06 [eagle.py:1005] Assuming the EAGLE head shares the same vocab embedding with the target model. 2025-11-03T17:43:06.6536787Z (EngineCore_DP0 pid=19233) INFO 11-03 17:43:06 [eagle.py:1043] The EAGLE head's lm_head will be loaded separately from the target model. 2025-11-03T17:43:07.1603118Z (EngineCore_DP0 pid=19233) INFO 11-03 17:43:07 [gpu_model_runner.py:2902] Model loading took 15.7807 GiB and 0.264026 seconds 2025-11-03T17:43:07.1611937Z (EngineCore_DP0 pid=19233) INFO 11-03 17:43:07 [kv_cache_utils.py:1199] GPU KV cache size: 79,424 tokens 2025-11-03T17:43:07.1612897Z (EngineCore_DP0 pid=19233) INFO 11-03 17:43:07 [kv_cache_utils.py:1204] Maximum concurrency for 10,240 tokens per request: 7.76x 2025-11-03T17:43:07.8252084Z (EngineCore_DP0 pid=19233) INFO 11-03 17:43:07 [gc_utils.py:40] GC Debug Config. enabled:False,top_objects:-1 2025-11-03T17:43:07.8268953Z INFO 11-03 17:43:07 [llm.py:337] Supported tasks: ['generate'] 2025-11-03T17:43:08.3023791Z PASSED 2025-11-03T17:43:08.3163105Z models/test_initialization.py::test_can_initialize_large_subset[Cohere2ForCausalLM] Fork a new process to run a test 19299 2025-11-03T17:43:08.3173983Z Fork a new process to run a test 0 2025-11-03T17:43:08.3456977Z INFO 11-03 17:43:08 [utils.py:239] non-default args: {'trust_remote_code': True, 'load_format': 'dummy', 'gpu_memory_utilization': 0.8, 'disable_log_stats': True, 'hf_overrides': functools.partial(, model_arch='Cohere2ForCausalLM', exist_overrides={}, use_original_num_layers=False), 'model_impl': 'vllm', 'model': 'CohereForAI/c4ai-command-r7b-12-2024'} 2025-11-03T17:43:08.5002854Z 2025-11-03T17:43:08.5004251Z config.json: 0% 0.00/1.02k [00:00, 'use_cudagraph': True, 'cudagraph_num_of_warmups': 1, 'cudagraph_capture_sizes': [512, 504, 496, 488, 480, 472, 464, 456, 448, 440, 432, 424, 416, 408, 400, 392, 384, 376, 368, 360, 352, 344, 336, 328, 320, 312, 304, 296, 288, 280, 272, 264, 256, 248, 240, 232, 224, 216, 208, 200, 192, 184, 176, 168, 160, 152, 144, 136, 128, 120, 112, 104, 96, 88, 80, 72, 64, 56, 48, 40, 32, 24, 16, 8, 4, 2, 1], 'cudagraph_copy_inputs': False, 'full_cuda_graph': True, 'use_inductor_graph_partition': False, 'pass_config': {}, 'max_capture_size': 512, 'local_cache_dir': None} 2025-11-03T17:43:11.5152623Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:43:11.5180022Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:43:11.5187479Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:43:11.5195648Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:43:11.5202890Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:43:11.5210595Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:43:11.5213210Z (EngineCore_DP0 pid=19327) INFO 11-03 17:43:11 [parallel_state.py:1231] rank 0 in world size 1 is assigned as DP rank 0, PP rank 0, TP rank 0, EP rank 0 2025-11-03T17:43:11.8578857Z (EngineCore_DP0 pid=19327) INFO 11-03 17:43:11 [topk_topp_sampler.py:59] Using FlashInfer for top-p & top-k sampling. 2025-11-03T17:43:11.9739154Z (EngineCore_DP0 pid=19327) INFO 11-03 17:43:11 [gpu_model_runner.py:2840] Starting to load model CohereForAI/c4ai-command-r7b-12-2024... 2025-11-03T17:43:12.2311152Z (EngineCore_DP0 pid=19327) INFO 11-03 17:43:12 [gpu_model_runner.py:2870] Loading model from scratch... 2025-11-03T17:43:12.3121783Z (EngineCore_DP0 pid=19327) INFO 11-03 17:43:12 [cuda.py:404] Using Flash Attention backend on V1 engine. 2025-11-03T17:43:12.8710855Z (EngineCore_DP0 pid=19327) INFO 11-03 17:43:12 [gpu_model_runner.py:2902] Model loading took 2.3909 GiB and 0.128765 seconds 2025-11-03T17:43:12.8715403Z (EngineCore_DP0 pid=19327) INFO 11-03 17:43:12 [kv_cache_utils.py:1199] GPU KV cache size: 2,621,440 tokens 2025-11-03T17:43:12.8716743Z (EngineCore_DP0 pid=19327) INFO 11-03 17:43:12 [kv_cache_utils.py:1204] Maximum concurrency for 132,096 tokens per request: 213.06x 2025-11-03T17:43:14.1534628Z (EngineCore_DP0 pid=19327) INFO 11-03 17:43:14 [gc_utils.py:40] GC Debug Config. enabled:False,top_objects:-1 2025-11-03T17:43:14.1549679Z INFO 11-03 17:43:14 [llm.py:337] Supported tasks: ['generate'] 2025-11-03T17:43:14.6549919Z PASSED 2025-11-03T17:43:14.6690630Z models/test_initialization.py::test_can_initialize_large_subset[GPT2ForSequenceClassification] Fork a new process to run a test 19393 2025-11-03T17:43:14.6701060Z Fork a new process to run a test 0 2025-11-03T17:43:14.6977582Z INFO 11-03 17:43:14 [utils.py:239] non-default args: {'load_format': 'dummy', 'gpu_memory_utilization': 0.8, 'disable_log_stats': True, 'hf_overrides': functools.partial(, model_arch='GPT2ForSequenceClassification', exist_overrides={}, use_original_num_layers=False), 'model_impl': 'vllm', 'model': 'nie3e/sentiment-polish-gpt2-small'} 2025-11-03T17:43:14.7809792Z 2025-11-03T17:43:14.7812118Z config.json: 0.00B [00:00, ?B/s] 2025-11-03T17:43:14.7812579Z config.json: 1.11kB [00:00, 5.72MB/s] 2025-11-03T17:43:21.3829967Z INFO 11-03 17:43:21 [model.py:915] Resolved `--runner auto` to `--runner pooling`. Pass the value explicitly to silence this message. 2025-11-03T17:43:21.3830861Z INFO 11-03 17:43:21 [model.py:967] Resolved `--convert auto` to `--convert classify`. Pass the value explicitly to silence this message. 2025-11-03T17:43:21.3831562Z INFO 11-03 17:43:21 [model.py:653] Resolved architecture: GPT2ForSequenceClassification 2025-11-03T17:43:21.3832029Z `torch_dtype` is deprecated! Use `dtype` instead! 2025-11-03T17:43:21.4077516Z INFO 11-03 17:43:21 [model.py:1939] Downcasting torch.float32 to torch.float16. 2025-11-03T17:43:21.4078122Z INFO 11-03 17:43:21 [model.py:1714] Using max model len 2048 2025-11-03T17:43:21.4357552Z INFO 11-03 17:43:21 [arg_utils.py:1725] (Enabling) chunked prefill by default 2025-11-03T17:43:21.4358040Z INFO 11-03 17:43:21 [arg_utils.py:1728] (Enabling) prefix caching by default 2025-11-03T17:43:21.5747455Z INFO 11-03 17:43:21 [scheduler.py:225] Chunked prefill is enabled with max_num_batched_tokens=8192. 2025-11-03T17:43:21.6232862Z 2025-11-03T17:43:21.6234770Z tokenizer_config.json: 0.00B [00:00, ?B/s] 2025-11-03T17:43:21.6235165Z tokenizer_config.json: 1.17kB [00:00, 8.68MB/s] 2025-11-03T17:43:21.6930942Z 2025-11-03T17:43:21.7013555Z vocab.json: 0.00B [00:00, ?B/s] 2025-11-03T17:43:21.7013970Z vocab.json: 907kB [00:00, 109MB/s] 2025-11-03T17:43:21.7532265Z 2025-11-03T17:43:21.7559564Z merges.txt: 0.00B [00:00, ?B/s] 2025-11-03T17:43:21.7559984Z merges.txt: 559kB [00:00, 208MB/s] 2025-11-03T17:43:21.8343069Z 2025-11-03T17:43:21.8428199Z tokenizer.json: 0.00B [00:00, ?B/s] 2025-11-03T17:43:21.8428640Z tokenizer.json: 2.34MB [00:00, 275MB/s] 2025-11-03T17:43:21.9101634Z 2025-11-03T17:43:21.9102328Z special_tokens_map.json: 0% 0.00/437 [00:00, 'use_cudagraph': True, 'cudagraph_num_of_warmups': 1, 'cudagraph_capture_sizes': [512, 504, 496, 488, 480, 472, 464, 456, 448, 440, 432, 424, 416, 408, 400, 392, 384, 376, 368, 360, 352, 344, 336, 328, 320, 312, 304, 296, 288, 280, 272, 264, 256, 248, 240, 232, 224, 216, 208, 200, 192, 184, 176, 168, 160, 152, 144, 136, 128, 120, 112, 104, 96, 88, 80, 72, 64, 56, 48, 40, 32, 24, 16, 8, 4, 2, 1], 'cudagraph_copy_inputs': False, 'full_cuda_graph': False, 'use_inductor_graph_partition': False, 'pass_config': {}, 'max_capture_size': 512, 'local_cache_dir': None} 2025-11-03T17:43:22.7917486Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:43:22.7945850Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:43:22.7953052Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:43:22.7961036Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:43:22.7968688Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:43:22.7976522Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:43:22.7979165Z (EngineCore_DP0 pid=19452) INFO 11-03 17:43:22 [parallel_state.py:1231] rank 0 in world size 1 is assigned as DP rank 0, PP rank 0, TP rank 0, EP rank 0 2025-11-03T17:43:23.1343374Z (EngineCore_DP0 pid=19452) INFO 11-03 17:43:23 [topk_topp_sampler.py:59] Using FlashInfer for top-p & top-k sampling. 2025-11-03T17:43:23.1532338Z (EngineCore_DP0 pid=19452) WARNING 11-03 17:43:23 [interfaces_base.py:74] The model () is missing the `get_input_embeddings` method. 2025-11-03T17:43:23.1543585Z (EngineCore_DP0 pid=19452) INFO 11-03 17:43:23 [gpu_model_runner.py:2840] Starting to load model nie3e/sentiment-polish-gpt2-small... 2025-11-03T17:43:23.4086311Z (EngineCore_DP0 pid=19452) INFO 11-03 17:43:23 [gpu_model_runner.py:2870] Loading model from scratch... 2025-11-03T17:43:23.4123996Z (EngineCore_DP0 pid=19452) INFO 11-03 17:43:23 [cuda.py:404] Using Flash Attention backend on V1 engine. 2025-11-03T17:43:23.9863521Z (EngineCore_DP0 pid=19452) INFO 11-03 17:43:23 [gpu_model_runner.py:2902] Model loading took 0.0904 GiB and 0.063119 seconds 2025-11-03T17:43:23.9867188Z (EngineCore_DP0 pid=19452) INFO 11-03 17:43:23 [kv_cache_utils.py:1199] GPU KV cache size: 3,495,248 tokens 2025-11-03T17:43:23.9868251Z (EngineCore_DP0 pid=19452) INFO 11-03 17:43:23 [kv_cache_utils.py:1204] Maximum concurrency for 2,048 tokens per request: 1706.66x 2025-11-03T17:43:24.7779442Z (EngineCore_DP0 pid=19452) INFO 11-03 17:43:24 [gc_utils.py:40] GC Debug Config. enabled:False,top_objects:-1 2025-11-03T17:43:24.7801760Z INFO 11-03 17:43:24 [llm.py:337] Supported tasks: ['classify'] 2025-11-03T17:43:25.2064326Z PASSED 2025-11-03T17:43:25.2208787Z models/test_initialization.py::test_can_initialize_large_subset[Qwen3VLForConditionalGeneration] Fork a new process to run a test 19518 2025-11-03T17:43:25.2218708Z Fork a new process to run a test 0 2025-11-03T17:43:25.2220321Z Model is not available online 2025-11-03T17:43:25.5113746Z PASSED 2025-11-03T17:43:25.5252717Z models/test_initialization.py::test_can_initialize_large_subset[DbrxForCausalLM] Fork a new process to run a test 19519 2025-11-03T17:43:25.5262463Z Fork a new process to run a test 0 2025-11-03T17:43:25.5541327Z INFO 11-03 17:43:25 [utils.py:239] non-default args: {'load_format': 'dummy', 'gpu_memory_utilization': 0.8, 'disable_log_stats': True, 'hf_overrides': functools.partial(, model_arch='DbrxForCausalLM', exist_overrides={}, use_original_num_layers=False), 'model_impl': 'vllm', 'model': 'databricks/dbrx-instruct'} 2025-11-03T17:43:25.6509703Z 2025-11-03T17:43:25.6512928Z config.json: 0% 0.00/733 [00:00, 'use_cudagraph': True, 'cudagraph_num_of_warmups': 1, 'cudagraph_capture_sizes': [512, 504, 496, 488, 480, 472, 464, 456, 448, 440, 432, 424, 416, 408, 400, 392, 384, 376, 368, 360, 352, 344, 336, 328, 320, 312, 304, 296, 288, 280, 272, 264, 256, 248, 240, 232, 224, 216, 208, 200, 192, 184, 176, 168, 160, 152, 144, 136, 128, 120, 112, 104, 96, 88, 80, 72, 64, 56, 48, 40, 32, 24, 16, 8, 4, 2, 1], 'cudagraph_copy_inputs': False, 'full_cuda_graph': True, 'use_inductor_graph_partition': False, 'pass_config': {}, 'max_capture_size': 512, 'local_cache_dir': None} 2025-11-03T17:43:34.0247263Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:43:34.0274866Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:43:34.0282359Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:43:34.0290522Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:43:34.0298256Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:43:34.0305963Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:43:34.0308980Z (EngineCore_DP0 pid=19577) INFO 11-03 17:43:34 [parallel_state.py:1231] rank 0 in world size 1 is assigned as DP rank 0, PP rank 0, TP rank 0, EP rank 0 2025-11-03T17:43:34.3590128Z (EngineCore_DP0 pid=19577) INFO 11-03 17:43:34 [topk_topp_sampler.py:59] Using FlashInfer for top-p & top-k sampling. 2025-11-03T17:43:34.4595441Z (EngineCore_DP0 pid=19577) INFO 11-03 17:43:34 [gpu_model_runner.py:2840] Starting to load model databricks/dbrx-instruct... 2025-11-03T17:43:34.7165765Z (EngineCore_DP0 pid=19577) INFO 11-03 17:43:34 [gpu_model_runner.py:2870] Loading model from scratch... 2025-11-03T17:43:34.7821867Z (EngineCore_DP0 pid=19577) INFO 11-03 17:43:34 [cuda.py:404] Using Flash Attention backend on V1 engine. 2025-11-03T17:43:34.7848769Z (EngineCore_DP0 pid=19577) WARNING 11-03 17:43:34 [vllm.py:821] `torch.compile` is turned on, but the model databricks/dbrx-instruct does not support it. Please open an issue on GitHub if you want it to be supported. 2025-11-03T17:43:35.2783900Z (EngineCore_DP0 pid=19577) INFO 11-03 17:43:35 [gpu_model_runner.py:2902] Model loading took 8.3752 GiB and 0.071043 seconds 2025-11-03T17:43:35.2788266Z (EngineCore_DP0 pid=19577) INFO 11-03 17:43:35 [kv_cache_utils.py:1199] GPU KV cache size: 2,621,440 tokens 2025-11-03T17:43:35.2789167Z (EngineCore_DP0 pid=19577) INFO 11-03 17:43:35 [kv_cache_utils.py:1204] Maximum concurrency for 32,768 tokens per request: 80.00x 2025-11-03T17:43:35.8644229Z (EngineCore_DP0 pid=19577) INFO 11-03 17:43:35 [gc_utils.py:40] GC Debug Config. enabled:False,top_objects:-1 2025-11-03T17:43:35.8666538Z INFO 11-03 17:43:35 [llm.py:337] Supported tasks: ['generate'] 2025-11-03T17:43:36.3373776Z PASSED 2025-11-03T17:43:36.3519534Z models/test_initialization.py::test_can_initialize_large_subset[AriaForConditionalGeneration] Fork a new process to run a test 19643 2025-11-03T17:43:36.3530623Z Fork a new process to run a test 0 2025-11-03T17:43:36.3812420Z INFO 11-03 17:43:36 [utils.py:239] non-default args: {'load_format': 'dummy', 'gpu_memory_utilization': 0.8, 'disable_log_stats': True, 'hf_overrides': functools.partial(, model_arch='AriaForConditionalGeneration', exist_overrides={}, use_original_num_layers=False), 'model_impl': 'vllm', 'model': 'rhymes-ai/Aria'} 2025-11-03T17:43:36.4604767Z 2025-11-03T17:43:36.4606860Z config.json: 0.00B [00:00, ?B/s] 2025-11-03T17:43:36.4607203Z config.json: 1.08kB [00:00, 5.98MB/s] 2025-11-03T17:43:36.5635522Z 2025-11-03T17:43:36.5636523Z preprocessor_config.json: 0.00B [00:00, ?B/s] 2025-11-03T17:43:36.5636925Z preprocessor_config.json: 1.21kB [00:00, 12.9MB/s] 2025-11-03T17:43:42.8926067Z INFO 11-03 17:43:42 [model.py:653] Resolved architecture: AriaForConditionalGeneration 2025-11-03T17:43:42.8926532Z `torch_dtype` is deprecated! Use `dtype` instead! 2025-11-03T17:43:42.9173824Z INFO 11-03 17:43:42 [model.py:1714] Using max model len 65536 2025-11-03T17:43:43.0907593Z INFO 11-03 17:43:43 [scheduler.py:225] Chunked prefill is enabled with max_num_batched_tokens=8192. 2025-11-03T17:43:43.1230821Z 2025-11-03T17:43:43.1235380Z tokenizer_config.json: 0.00B [00:00, ?B/s] 2025-11-03T17:43:43.1235734Z tokenizer_config.json: 110kB [00:00, 233MB/s] 2025-11-03T17:43:43.2488641Z 2025-11-03T17:43:43.5178765Z tokenizer.model: 0% 0.00/1.70M [00:00, 'use_cudagraph': True, 'cudagraph_num_of_warmups': 1, 'cudagraph_capture_sizes': [512, 504, 496, 488, 480, 472, 464, 456, 448, 440, 432, 424, 416, 408, 400, 392, 384, 376, 368, 360, 352, 344, 336, 328, 320, 312, 304, 296, 288, 280, 272, 264, 256, 248, 240, 232, 224, 216, 208, 200, 192, 184, 176, 168, 160, 152, 144, 136, 128, 120, 112, 104, 96, 88, 80, 72, 64, 56, 48, 40, 32, 24, 16, 8, 4, 2, 1], 'cudagraph_copy_inputs': False, 'full_cuda_graph': True, 'use_inductor_graph_partition': False, 'pass_config': {}, 'max_capture_size': 512, 'local_cache_dir': None} 2025-11-03T17:43:46.6865352Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:43:46.6892592Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:43:46.6899790Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:43:46.6907265Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:43:46.6914788Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:43:46.6922492Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:43:46.6925239Z (EngineCore_DP0 pid=19722) INFO 11-03 17:43:46 [parallel_state.py:1231] rank 0 in world size 1 is assigned as DP rank 0, PP rank 0, TP rank 0, EP rank 0 2025-11-03T17:43:47.0213318Z (EngineCore_DP0 pid=19722) INFO 11-03 17:43:47 [topk_topp_sampler.py:59] Using FlashInfer for top-p & top-k sampling. 2025-11-03T17:43:47.1827414Z (EngineCore_DP0 pid=19722) 2025-11-03T17:43:47.1829197Z processor_config.json: 0% 0.00/100 [00:00, model_arch='LlavaNextForConditionalGeneration', exist_overrides={}, use_original_num_layers=False), 'model_impl': 'vllm', 'model': 'llava-hf/llava-v1.6-mistral-7b-hf'} 2025-11-03T17:43:54.1883912Z 2025-11-03T17:43:54.1885831Z config.json: 0.00B [00:00, ?B/s] 2025-11-03T17:43:54.1886243Z config.json: 1.28kB [00:00, 7.82MB/s] 2025-11-03T17:44:25.2309970Z ERROR 11-03 17:44:25 [config.py:298] Error retrieving file list: 502 Server Error: Bad Gateway for url: https://huggingface.co/api/models/llava-hf/llava-v1.6-mistral-7b-hf/tree/main?recursive=True&expand=False, retrying 1 of 2 2025-11-03T17:44:27.3019433Z 2025-11-03T17:44:27.3019901Z preprocessor_config.json: 0% 0.00/772 [00:00, 'use_cudagraph': True, 'cudagraph_num_of_warmups': 1, 'cudagraph_capture_sizes': [512, 504, 496, 488, 480, 472, 464, 456, 448, 440, 432, 424, 416, 408, 400, 392, 384, 376, 368, 360, 352, 344, 336, 328, 320, 312, 304, 296, 288, 280, 272, 264, 256, 248, 240, 232, 224, 216, 208, 200, 192, 184, 176, 168, 160, 152, 144, 136, 128, 120, 112, 104, 96, 88, 80, 72, 64, 56, 48, 40, 32, 24, 16, 8, 4, 2, 1], 'cudagraph_copy_inputs': False, 'full_cuda_graph': True, 'use_inductor_graph_partition': False, 'pass_config': {}, 'max_capture_size': 512, 'local_cache_dir': None} 2025-11-03T17:44:42.3797845Z (EngineCore_DP0 pid=19884) /opt/conda/envs/py_3.12/lib/python3.12/site-packages/torch/library.py:356: UserWarning: Warning only once for all operators, other operators may also be overridden. 2025-11-03T17:44:42.3799171Z (EngineCore_DP0 pid=19884) Overriding a previously registered kernel for the same operator and the same dispatch key 2025-11-03T17:44:42.3800923Z (EngineCore_DP0 pid=19884) operator: xformers_python::_fused_allgather_and_linear_impl(Tensor scattered_input, Tensor[] weights, str process_group_name, Tensor(a3!)[] gathered_outputs, SymInt timeout_s, bool _wait, bool _memcpy, Tensor scale_scattered_input, Tensor?[] scales_weights) -> () 2025-11-03T17:44:42.3802771Z (EngineCore_DP0 pid=19884) registered at /opt/conda/envs/py_3.12/lib/python3.12/site-packages/torch/_library/custom_ops.py:922 2025-11-03T17:44:42.3803514Z (EngineCore_DP0 pid=19884) dispatch key: ADInplaceOrView 2025-11-03T17:44:42.3804042Z (EngineCore_DP0 pid=19884) previous kernel: no debug info 2025-11-03T17:44:42.3805117Z (EngineCore_DP0 pid=19884) new kernel: registered at /opt/conda/envs/py_3.12/lib/python3.12/site-packages/torch/_library/custom_ops.py:922 (Triggered internally at /var/lib/jenkins/workspace/aten/src/ATen/core/dispatch/OperatorEntry.cpp:208.) 2025-11-03T17:44:42.3806014Z (EngineCore_DP0 pid=19884) self.m.impl( 2025-11-03T17:44:42.7464805Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:44:42.7492994Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:44:42.7499963Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:44:42.7507366Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:44:42.7514293Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:44:42.7521624Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:44:42.7524007Z (EngineCore_DP0 pid=19884) INFO 11-03 17:44:42 [parallel_state.py:1231] rank 0 in world size 1 is assigned as DP rank 0, PP rank 0, TP rank 0, EP rank 0 2025-11-03T17:44:42.9889789Z (EngineCore_DP0 pid=19884) INFO 11-03 17:44:42 [topk_topp_sampler.py:59] Using FlashInfer for top-p & top-k sampling. 2025-11-03T17:44:43.1186543Z (EngineCore_DP0 pid=19884) 2025-11-03T17:44:43.1186954Z processor_config.json: 0% 0.00/176 [00:00, model_arch='ChatGLMForConditionalGeneration', exist_overrides={}, use_original_num_layers=False), 'model_impl': 'vllm', 'model': 'thu-coai/ShieldLM-6B-chatglm3'} 2025-11-03T17:44:53.7596464Z 2025-11-03T17:44:53.7598739Z config.json: 0.00B [00:00, ?B/s] 2025-11-03T17:44:53.7599064Z config.json: 1.51kB [00:00, 7.20MB/s] 2025-11-03T17:45:00.1946838Z INFO 11-03 17:45:00 [model.py:653] Resolved architecture: ChatGLMForConditionalGeneration 2025-11-03T17:45:00.1947667Z `torch_dtype` is deprecated! Use `dtype` instead! 2025-11-03T17:45:00.2196701Z INFO 11-03 17:45:00 [model.py:1714] Using max model len 8192 2025-11-03T17:45:00.2198939Z The argument `trust_remote_code` is to be used with Auto classes. It has no effect here and is ignored. 2025-11-03T17:45:00.3867975Z INFO 11-03 17:45:00 [scheduler.py:225] Chunked prefill is enabled with max_num_batched_tokens=8192. 2025-11-03T17:45:00.4254196Z 2025-11-03T17:45:00.4255275Z tokenizer_config.json: 0% 0.00/244 [00:00, 'use_cudagraph': True, 'cudagraph_num_of_warmups': 1, 'cudagraph_capture_sizes': [512, 504, 496, 488, 480, 472, 464, 456, 448, 440, 432, 424, 416, 408, 400, 392, 384, 376, 368, 360, 352, 344, 336, 328, 320, 312, 304, 296, 288, 280, 272, 264, 256, 248, 240, 232, 224, 216, 208, 200, 192, 184, 176, 168, 160, 152, 144, 136, 128, 120, 112, 104, 96, 88, 80, 72, 64, 56, 48, 40, 32, 24, 16, 8, 4, 2, 1], 'cudagraph_copy_inputs': False, 'full_cuda_graph': True, 'use_inductor_graph_partition': False, 'pass_config': {}, 'max_capture_size': 512, 'local_cache_dir': None} 2025-11-03T17:45:01.9605617Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:45:01.9633679Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:45:01.9640547Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:45:01.9648867Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:45:01.9659512Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:45:01.9670338Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:45:01.9673287Z (EngineCore_DP0 pid=20249) INFO 11-03 17:45:01 [parallel_state.py:1231] rank 0 in world size 1 is assigned as DP rank 0, PP rank 0, TP rank 0, EP rank 0 2025-11-03T17:45:02.3051444Z (EngineCore_DP0 pid=20249) INFO 11-03 17:45:02 [topk_topp_sampler.py:59] Using FlashInfer for top-p & top-k sampling. 2025-11-03T17:45:02.3758753Z (EngineCore_DP0 pid=20249) INFO 11-03 17:45:02 [gpu_model_runner.py:2840] Starting to load model thu-coai/ShieldLM-6B-chatglm3... 2025-11-03T17:45:02.6327671Z (EngineCore_DP0 pid=20249) INFO 11-03 17:45:02 [gpu_model_runner.py:2870] Loading model from scratch... 2025-11-03T17:45:02.6983877Z (EngineCore_DP0 pid=20249) INFO 11-03 17:45:02 [cuda.py:404] Using Flash Attention backend on V1 engine. 2025-11-03T17:45:03.2691433Z (EngineCore_DP0 pid=20249) INFO 11-03 17:45:03 [gpu_model_runner.py:2902] Model loading took 1.3741 GiB and 0.125880 seconds 2025-11-03T17:45:03.2695605Z (EngineCore_DP0 pid=20249) INFO 11-03 17:45:03 [kv_cache_utils.py:1199] GPU KV cache size: 10,485,760 tokens 2025-11-03T17:45:03.2696527Z (EngineCore_DP0 pid=20249) INFO 11-03 17:45:03 [kv_cache_utils.py:1204] Maximum concurrency for 8,192 tokens per request: 1280.00x 2025-11-03T17:45:03.4942629Z (EngineCore_DP0 pid=20249) WARNING 11-03 17:45:03 [tokenizer.py:255] Using a slow tokenizer. This might cause a significant slowdown. Consider using a fast tokenizer instead. 2025-11-03T17:45:04.7682202Z (EngineCore_DP0 pid=20249) INFO 11-03 17:45:04 [gc_utils.py:40] GC Debug Config. enabled:False,top_objects:-1 2025-11-03T17:45:04.7695148Z INFO 11-03 17:45:04 [llm.py:337] Supported tasks: ['generate'] 2025-11-03T17:45:05.2285696Z PASSED 2025-11-03T17:45:05.2429122Z models/test_initialization.py::test_can_initialize_large_subset[ArcticForCausalLM] Fork a new process to run a test 20315 2025-11-03T17:45:05.2439822Z Fork a new process to run a test 0 2025-11-03T17:45:05.2719777Z INFO 11-03 17:45:05 [utils.py:239] non-default args: {'trust_remote_code': True, 'load_format': 'dummy', 'gpu_memory_utilization': 0.8, 'disable_log_stats': True, 'hf_overrides': functools.partial(, model_arch='ArcticForCausalLM', exist_overrides={}, use_original_num_layers=False), 'model_impl': 'vllm', 'model': 'Snowflake/snowflake-arctic-instruct'} 2025-11-03T17:45:05.4009484Z 2025-11-03T17:45:05.4011489Z config.json: 0.00B [00:00, ?B/s] 2025-11-03T17:45:05.4011829Z config.json: 1.28kB [00:00, 7.30MB/s] 2025-11-03T17:45:05.4663058Z 2025-11-03T17:45:05.4665736Z configuration_arctic.py: 0.00B [00:00, ?B/s] 2025-11-03T17:45:05.4666153Z configuration_arctic.py: 9.46kB [00:00, 45.8MB/s] 2025-11-03T17:45:05.4759437Z A new version of the following files was downloaded from https://huggingface.co/Snowflake/snowflake-arctic-instruct: 2025-11-03T17:45:05.4760024Z - configuration_arctic.py 2025-11-03T17:45:05.4760696Z . Make sure to double-check they do not contain any added malicious code. To avoid downloading new versions of the code file, you can pin a revision. 2025-11-03T17:45:12.0030873Z INFO 11-03 17:45:12 [model.py:653] Resolved architecture: ArcticForCausalLM 2025-11-03T17:45:12.0031399Z `torch_dtype` is deprecated! Use `dtype` instead! 2025-11-03T17:45:12.0281637Z INFO 11-03 17:45:12 [model.py:1714] Using max model len 4096 2025-11-03T17:45:12.0283381Z The argument `trust_remote_code` is to be used with Auto classes. It has no effect here and is ignored. 2025-11-03T17:45:12.0913258Z INFO 11-03 17:45:12 [scheduler.py:225] Chunked prefill is enabled with max_num_batched_tokens=8192. 2025-11-03T17:45:12.1266595Z 2025-11-03T17:45:12.1268480Z tokenizer_config.json: 0.00B [00:00, ?B/s] 2025-11-03T17:45:12.1269466Z tokenizer_config.json: 1.58kB [00:00, 10.9MB/s] 2025-11-03T17:45:12.1631717Z 2025-11-03T17:45:12.1632821Z tokenization_arctic.py: 0.00B [00:00, ?B/s] 2025-11-03T17:45:12.1633317Z tokenization_arctic.py: 1.80kB [00:00, 15.6MB/s] 2025-11-03T17:45:12.1653057Z A new version of the following files was downloaded from https://huggingface.co/Snowflake/snowflake-arctic-instruct: 2025-11-03T17:45:12.1653644Z - tokenization_arctic.py 2025-11-03T17:45:12.1654247Z . Make sure to double-check they do not contain any added malicious code. To avoid downloading new versions of the code file, you can pin a revision. 2025-11-03T17:45:12.2551101Z 2025-11-03T17:45:12.4172616Z tokenizer.model: 0% 0.00/500k [00:00, 'use_cudagraph': True, 'cudagraph_num_of_warmups': 1, 'cudagraph_capture_sizes': [512, 504, 496, 488, 480, 472, 464, 456, 448, 440, 432, 424, 416, 408, 400, 392, 384, 376, 368, 360, 352, 344, 336, 328, 320, 312, 304, 296, 288, 280, 272, 264, 256, 248, 240, 232, 224, 216, 208, 200, 192, 184, 176, 168, 160, 152, 144, 136, 128, 120, 112, 104, 96, 88, 80, 72, 64, 56, 48, 40, 32, 24, 16, 8, 4, 2, 1], 'cudagraph_copy_inputs': False, 'full_cuda_graph': True, 'use_inductor_graph_partition': False, 'pass_config': {}, 'max_capture_size': 512, 'local_cache_dir': None} 2025-11-03T17:45:13.4312652Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:45:13.4341425Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:45:13.4349298Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:45:13.4356962Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:45:13.4364203Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:45:13.4372154Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:45:13.4375103Z (EngineCore_DP0 pid=20393) INFO 11-03 17:45:13 [parallel_state.py:1231] rank 0 in world size 1 is assigned as DP rank 0, PP rank 0, TP rank 0, EP rank 0 2025-11-03T17:45:13.7879668Z (EngineCore_DP0 pid=20393) INFO 11-03 17:45:13 [topk_topp_sampler.py:59] Using FlashInfer for top-p & top-k sampling. 2025-11-03T17:45:13.9022021Z (EngineCore_DP0 pid=20393) INFO 11-03 17:45:13 [gpu_model_runner.py:2840] Starting to load model Snowflake/snowflake-arctic-instruct... 2025-11-03T17:45:14.1662101Z (EngineCore_DP0 pid=20393) INFO 11-03 17:45:14 [gpu_model_runner.py:2870] Loading model from scratch... 2025-11-03T17:45:14.2319995Z (EngineCore_DP0 pid=20393) INFO 11-03 17:45:14 [cuda.py:404] Using Flash Attention backend on V1 engine. 2025-11-03T17:45:14.8185577Z (EngineCore_DP0 pid=20393) INFO 11-03 17:45:14 [gpu_model_runner.py:2902] Model loading took 1.7530 GiB and 0.126361 seconds 2025-11-03T17:45:14.8189914Z (EngineCore_DP0 pid=20393) INFO 11-03 17:45:14 [kv_cache_utils.py:1199] GPU KV cache size: 2,621,440 tokens 2025-11-03T17:45:14.8190857Z (EngineCore_DP0 pid=20393) INFO 11-03 17:45:14 [kv_cache_utils.py:1204] Maximum concurrency for 4,096 tokens per request: 640.00x 2025-11-03T17:45:14.9766792Z (EngineCore_DP0 pid=20393) WARNING 11-03 17:45:14 [tokenizer.py:255] Using a slow tokenizer. This might cause a significant slowdown. Consider using a fast tokenizer instead. 2025-11-03T17:45:15.4460249Z (EngineCore_DP0 pid=20393) INFO 11-03 17:45:15 [gc_utils.py:40] GC Debug Config. enabled:False,top_objects:-1 2025-11-03T17:45:15.4482326Z INFO 11-03 17:45:15 [llm.py:337] Supported tasks: ['generate'] 2025-11-03T17:45:15.9200409Z PASSED 2025-11-03T17:45:16.1557591Z models/test_initialization.py::test_can_initialize_large_subset[Lfm2ForCausalLM] SKIPPED 2025-11-03T17:45:16.1699639Z models/test_initialization.py::test_can_initialize_large_subset[H2OVLChatModel] Fork a new process to run a test 20459 2025-11-03T17:45:16.1710275Z Fork a new process to run a test 0 2025-11-03T17:45:16.1714240Z `transformers==4.56.2` installed, but `transformers<=4.48` is required to run this model. Reason: HF model is not compatible. 2025-11-03T17:45:16.4670187Z PASSED 2025-11-03T17:45:16.4813100Z models/test_initialization.py::test_can_initialize_large_subset[SeedOssForCausalLM] Fork a new process to run a test 20460 2025-11-03T17:45:16.4823699Z Fork a new process to run a test 0 2025-11-03T17:45:16.4825481Z Model is not available online 2025-11-03T17:45:16.7786375Z PASSED 2025-11-03T17:45:16.7929916Z models/test_initialization.py::test_can_initialize_large_subset[MiMoMTPModel] Fork a new process to run a test 20461 2025-11-03T17:45:16.7940360Z Fork a new process to run a test 0 2025-11-03T17:45:16.8220267Z INFO 11-03 17:45:16 [utils.py:239] non-default args: {'trust_remote_code': True, 'load_format': 'dummy', 'gpu_memory_utilization': 0.8, 'disable_log_stats': True, 'hf_overrides': functools.partial(, model_arch='MiMoMTPModel', exist_overrides={}, use_original_num_layers=False), 'speculative_config': {'model': 'XiaomiMiMo/MiMo-7B-RL', 'num_speculative_tokens': 1}, 'model_impl': 'vllm', 'model': 'XiaomiMiMo/MiMo-7B-RL'} 2025-11-03T17:45:17.6712027Z INFO 11-03 17:45:17 [model.py:653] Resolved architecture: MiMoForCausalLM 2025-11-03T17:45:17.6712557Z `torch_dtype` is deprecated! Use `dtype` instead! 2025-11-03T17:45:17.6991174Z INFO 11-03 17:45:17 [model.py:1714] Using max model len 32768 2025-11-03T17:45:17.6993242Z The argument `trust_remote_code` is to be used with Auto classes. It has no effect here and is ignored. 2025-11-03T17:45:24.1778889Z INFO 11-03 17:45:24 [model.py:653] Resolved architecture: MiMoMTPModel 2025-11-03T17:45:24.1779402Z INFO 11-03 17:45:24 [model.py:1714] Using max model len 32768 2025-11-03T17:45:24.1781709Z INFO 11-03 17:45:24 [scheduler.py:225] Chunked prefill is enabled with max_num_batched_tokens=8192. 2025-11-03T17:45:24.6094557Z (EngineCore_DP0 pid=20518) INFO 11-03 17:45:24 [core.py:727] Waiting for init message from front-end. 2025-11-03T17:45:24.6158970Z (EngineCore_DP0 pid=20518) INFO 11-03 17:45:24 [core.py:94] Initializing a V1 LLM engine (v0.11.0rc2.dev389+ge51928192.d20251103) with config: model='XiaomiMiMo/MiMo-7B-RL', speculative_config=SpeculativeConfig(method='mtp', model='XiaomiMiMo/MiMo-7B-RL', num_spec_tokens=1), tokenizer='XiaomiMiMo/MiMo-7B-RL', skip_tokenizer_init=False, tokenizer_mode=auto, revision=None, tokenizer_revision=None, trust_remote_code=True, dtype=torch.bfloat16, max_seq_len=32768, download_dir=None, load_format=dummy, tensor_parallel_size=1, pipeline_parallel_size=1, data_parallel_size=1, disable_custom_all_reduce=False, quantization=None, enforce_eager=False, kv_cache_dtype=auto, device_config=cuda, structured_outputs_config=StructuredOutputsConfig(backend='auto', disable_fallback=False, disable_any_whitespace=False, disable_additional_properties=False, reasoning_parser=''), observability_config=ObservabilityConfig(show_hidden_metrics_for_version=None, otlp_traces_endpoint=None, collect_detailed_traces=None), seed=0, served_model_name=XiaomiMiMo/MiMo-7B-RL, enable_prefix_caching=True, chunked_prefill_enabled=True, pooler_config=None, compilation_config={'level': 3, 'debug_dump_path': None, 'cache_dir': '', 'backend': '', 'custom_ops': [], 'splitting_ops': ['vllm.unified_attention', 'vllm.unified_attention_with_output', 'vllm.unified_mla_attention', 'vllm.unified_mla_attention_with_output', 'vllm.mamba_mixer2', 'vllm.mamba_mixer', 'vllm.short_conv', 'vllm.linear_attention', 'vllm.plamo2_mamba_mixer', 'vllm.gdn_attention', 'vllm.sparse_attn_indexer'], 'use_inductor': True, 'compile_sizes': [], 'inductor_compile_config': {'enable_auto_functionalized_v2': False}, 'inductor_passes': {}, 'cudagraph_mode': , 'use_cudagraph': True, 'cudagraph_num_of_warmups': 1, 'cudagraph_capture_sizes': [512, 504, 496, 488, 480, 472, 464, 456, 448, 440, 432, 424, 416, 408, 400, 392, 384, 376, 368, 360, 352, 344, 336, 328, 320, 312, 304, 296, 288, 280, 272, 264, 256, 248, 240, 232, 224, 216, 208, 200, 192, 184, 176, 168, 160, 152, 144, 136, 128, 120, 112, 104, 96, 88, 80, 72, 64, 56, 48, 40, 32, 24, 16, 8, 4, 2, 1], 'cudagraph_copy_inputs': False, 'full_cuda_graph': True, 'use_inductor_graph_partition': False, 'pass_config': {}, 'max_capture_size': 512, 'local_cache_dir': None} 2025-11-03T17:45:25.3472871Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:45:25.3500866Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:45:25.3508757Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:45:25.3516364Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:45:25.3524239Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:45:25.3532015Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:45:25.3534769Z (EngineCore_DP0 pid=20518) INFO 11-03 17:45:25 [parallel_state.py:1231] rank 0 in world size 1 is assigned as DP rank 0, PP rank 0, TP rank 0, EP rank 0 2025-11-03T17:45:25.6865688Z (EngineCore_DP0 pid=20518) INFO 11-03 17:45:25 [topk_topp_sampler.py:59] Using FlashInfer for top-p & top-k sampling. 2025-11-03T17:45:25.7012432Z (EngineCore_DP0 pid=20518) WARNING 11-03 17:45:25 [__init__.py:194] min_p, logit_bias, and min_tokens parameters won't currently work with speculative decoding enabled. 2025-11-03T17:45:25.7450147Z (EngineCore_DP0 pid=20518) INFO 11-03 17:45:25 [gpu_model_runner.py:2840] Starting to load model XiaomiMiMo/MiMo-7B-RL... 2025-11-03T17:45:25.9978866Z (EngineCore_DP0 pid=20518) INFO 11-03 17:45:25 [gpu_model_runner.py:2870] Loading model from scratch... 2025-11-03T17:45:26.0536388Z (EngineCore_DP0 pid=20518) INFO 11-03 17:45:26 [cuda.py:404] Using Flash Attention backend on V1 engine. 2025-11-03T17:45:26.1146763Z (EngineCore_DP0 pid=20518) INFO 11-03 17:45:26 [gpu_model_runner.py:2879] Loading drafter model... 2025-11-03T17:45:26.1277059Z (EngineCore_DP0 pid=20518) WARNING 11-03 17:45:26 [vllm.py:821] `torch.compile` is turned on, but the model XiaomiMiMo/MiMo-7B-RL does not support it. Please open an issue on GitHub if you want it to be supported. 2025-11-03T17:45:26.1282150Z (EngineCore_DP0 pid=20518) INFO 11-03 17:45:26 [eagle.py:1005] Assuming the EAGLE head shares the same vocab embedding with the target model. 2025-11-03T17:45:26.1282925Z (EngineCore_DP0 pid=20518) INFO 11-03 17:45:26 [eagle.py:1027] Loading EAGLE LM head weights from the target model. 2025-11-03T17:45:26.6402378Z (EngineCore_DP0 pid=20518) INFO 11-03 17:45:26 [gpu_model_runner.py:2902] Model loading took 3.0470 GiB and 0.130734 seconds 2025-11-03T17:45:26.6406774Z (EngineCore_DP0 pid=20518) INFO 11-03 17:45:26 [kv_cache_utils.py:1199] GPU KV cache size: 1,310,720 tokens 2025-11-03T17:45:26.6407678Z (EngineCore_DP0 pid=20518) INFO 11-03 17:45:26 [kv_cache_utils.py:1204] Maximum concurrency for 32,768 tokens per request: 40.00x 2025-11-03T17:45:27.3576883Z (EngineCore_DP0 pid=20518) INFO 11-03 17:45:27 [gc_utils.py:40] GC Debug Config. enabled:False,top_objects:-1 2025-11-03T17:45:27.3600219Z INFO 11-03 17:45:27 [llm.py:337] Supported tasks: ['generate'] 2025-11-03T17:45:27.8263000Z PASSED 2025-11-03T17:45:27.8405193Z models/test_initialization.py::test_can_initialize_large_subset[OlmoForCausalLM] Fork a new process to run a test 20584 2025-11-03T17:45:27.8416395Z Fork a new process to run a test 0 2025-11-03T17:45:27.8697850Z INFO 11-03 17:45:27 [utils.py:239] non-default args: {'load_format': 'dummy', 'gpu_memory_utilization': 0.8, 'disable_log_stats': True, 'hf_overrides': functools.partial(, model_arch='OlmoForCausalLM', exist_overrides={}, use_original_num_layers=False), 'model_impl': 'vllm', 'model': 'allenai/OLMo-1B-hf'} 2025-11-03T17:45:27.9475068Z 2025-11-03T17:45:27.9476427Z config.json: 0% 0.00/632 [00:00, 'use_cudagraph': True, 'cudagraph_num_of_warmups': 1, 'cudagraph_capture_sizes': [512, 504, 496, 488, 480, 472, 464, 456, 448, 440, 432, 424, 416, 408, 400, 392, 384, 376, 368, 360, 352, 344, 336, 328, 320, 312, 304, 296, 288, 280, 272, 264, 256, 248, 240, 232, 224, 216, 208, 200, 192, 184, 176, 168, 160, 152, 144, 136, 128, 120, 112, 104, 96, 88, 80, 72, 64, 56, 48, 40, 32, 24, 16, 8, 4, 2, 1], 'cudagraph_copy_inputs': False, 'full_cuda_graph': True, 'use_inductor_graph_partition': False, 'pass_config': {}, 'max_capture_size': 512, 'local_cache_dir': None} 2025-11-03T17:45:35.7617057Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:45:35.7645302Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:45:35.7652801Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:45:35.7661041Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:45:35.7669530Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:45:35.7677207Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:45:35.7680030Z (EngineCore_DP0 pid=20642) INFO 11-03 17:45:35 [parallel_state.py:1231] rank 0 in world size 1 is assigned as DP rank 0, PP rank 0, TP rank 0, EP rank 0 2025-11-03T17:45:36.1017047Z (EngineCore_DP0 pid=20642) INFO 11-03 17:45:36 [topk_topp_sampler.py:59] Using FlashInfer for top-p & top-k sampling. 2025-11-03T17:45:36.1503143Z (EngineCore_DP0 pid=20642) INFO 11-03 17:45:36 [gpu_model_runner.py:2840] Starting to load model allenai/OLMo-1B-hf... 2025-11-03T17:45:36.4031693Z (EngineCore_DP0 pid=20642) INFO 11-03 17:45:36 [gpu_model_runner.py:2870] Loading model from scratch... 2025-11-03T17:45:36.4682646Z (EngineCore_DP0 pid=20642) INFO 11-03 17:45:36 [cuda.py:404] Using Flash Attention backend on V1 engine. 2025-11-03T17:45:37.0299734Z (EngineCore_DP0 pid=20642) INFO 11-03 17:45:37 [gpu_model_runner.py:2902] Model loading took 0.3174 GiB and 0.123320 seconds 2025-11-03T17:45:37.0303439Z (EngineCore_DP0 pid=20642) INFO 11-03 17:45:37 [kv_cache_utils.py:1199] GPU KV cache size: 1,310,720 tokens 2025-11-03T17:45:37.0304212Z (EngineCore_DP0 pid=20642) INFO 11-03 17:45:37 [kv_cache_utils.py:1204] Maximum concurrency for 2,048 tokens per request: 640.00x 2025-11-03T17:45:37.4736816Z (EngineCore_DP0 pid=20642) INFO 11-03 17:45:37 [gc_utils.py:40] GC Debug Config. enabled:False,top_objects:-1 2025-11-03T17:45:37.4751676Z INFO 11-03 17:45:37 [llm.py:337] Supported tasks: ['generate'] 2025-11-03T17:45:37.9155665Z PASSED 2025-11-03T17:45:37.9297877Z models/test_initialization.py::test_can_initialize_large_subset[QwenVLForConditionalGeneration] Fork a new process to run a test 20708 2025-11-03T17:45:37.9309108Z Fork a new process to run a test 0 2025-11-03T17:45:37.9587045Z INFO 11-03 17:45:37 [utils.py:239] non-default args: {'trust_remote_code': True, 'load_format': 'dummy', 'gpu_memory_utilization': 0.8, 'disable_log_stats': True, 'hf_overrides': functools.partial(, model_arch='QwenVLForConditionalGeneration', exist_overrides={'architectures': ['QwenVLForConditionalGeneration']}, use_original_num_layers=False), 'model_impl': 'vllm', 'model': 'Qwen/Qwen-VL'} 2025-11-03T17:45:38.0345324Z 2025-11-03T17:45:38.0347653Z config.json: 0.00B [00:00, ?B/s] 2025-11-03T17:45:38.0348134Z config.json: 1.16kB [00:00, 5.57MB/s] 2025-11-03T17:45:38.1233728Z 2025-11-03T17:45:38.1235774Z configuration_qwen.py: 0.00B [00:00, ?B/s] 2025-11-03T17:45:38.1236149Z configuration_qwen.py: 2.09kB [00:00, 17.7MB/s] 2025-11-03T17:45:38.1325926Z A new version of the following files was downloaded from https://huggingface.co/Qwen/Qwen-VL: 2025-11-03T17:45:38.1326423Z - configuration_qwen.py 2025-11-03T17:45:38.1327064Z . Make sure to double-check they do not contain any added malicious code. To avoid downloading new versions of the code file, you can pin a revision. 2025-11-03T17:45:45.1503367Z INFO 11-03 17:45:45 [model.py:653] Resolved architecture: QwenVLForConditionalGeneration 2025-11-03T17:45:45.1503888Z `torch_dtype` is deprecated! Use `dtype` instead! 2025-11-03T17:45:45.1758425Z INFO 11-03 17:45:45 [model.py:1714] Using max model len 2048 2025-11-03T17:45:45.1761222Z The argument `trust_remote_code` is to be used with Auto classes. It has no effect here and is ignored. 2025-11-03T17:45:45.2333504Z INFO 11-03 17:45:45 [scheduler.py:225] Chunked prefill is enabled with max_num_batched_tokens=8192. 2025-11-03T17:45:45.2811541Z 2025-11-03T17:45:45.2812300Z tokenizer_config.json: 0% 0.00/173 [00:00, 'use_cudagraph': True, 'cudagraph_num_of_warmups': 1, 'cudagraph_capture_sizes': [512, 504, 496, 488, 480, 472, 464, 456, 448, 440, 432, 424, 416, 408, 400, 392, 384, 376, 368, 360, 352, 344, 336, 328, 320, 312, 304, 296, 288, 280, 272, 264, 256, 248, 240, 232, 224, 216, 208, 200, 192, 184, 176, 168, 160, 152, 144, 136, 128, 120, 112, 104, 96, 88, 80, 72, 64, 56, 48, 40, 32, 24, 16, 8, 4, 2, 1], 'cudagraph_copy_inputs': False, 'full_cuda_graph': True, 'use_inductor_graph_partition': False, 'pass_config': {}, 'max_capture_size': 512, 'local_cache_dir': None} 2025-11-03T17:45:50.8431242Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:45:50.8464581Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:45:50.8472738Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:45:50.8481024Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:45:50.8489457Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:45:50.8497798Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:45:50.8501004Z (EngineCore_DP0 pid=20766) INFO 11-03 17:45:50 [parallel_state.py:1231] rank 0 in world size 1 is assigned as DP rank 0, PP rank 0, TP rank 0, EP rank 0 2025-11-03T17:45:51.2049031Z (EngineCore_DP0 pid=20766) INFO 11-03 17:45:51 [topk_topp_sampler.py:59] Using FlashInfer for top-p & top-k sampling. 2025-11-03T17:45:51.6115075Z (EngineCore_DP0 pid=20766) WARNING 11-03 17:45:51 [__init__.py:2040] The following intended overrides are not keyword args and will be dropped: {'truncation'} 2025-11-03T17:45:51.6126235Z (EngineCore_DP0 pid=20766) WARNING 11-03 17:45:51 [__init__.py:2040] The following intended overrides are not keyword args and will be dropped: {'truncation'} 2025-11-03T17:45:51.6171064Z (EngineCore_DP0 pid=20766) INFO 11-03 17:45:51 [gpu_model_runner.py:2840] Starting to load model Qwen/Qwen-VL... 2025-11-03T17:45:51.8868040Z (EngineCore_DP0 pid=20766) INFO 11-03 17:45:51 [gpu_model_runner.py:2870] Loading model from scratch... 2025-11-03T17:45:51.9464243Z (EngineCore_DP0 pid=20766) INFO 11-03 17:45:51 [cuda.py:404] Using Flash Attention backend on V1 engine. 2025-11-03T17:45:52.6269255Z (EngineCore_DP0 pid=20766) INFO 11-03 17:45:52 [gpu_model_runner.py:2902] Model loading took 6.3133 GiB and 0.197016 seconds 2025-11-03T17:45:52.6274448Z (EngineCore_DP0 pid=20766) INFO 11-03 17:45:52 [kv_cache_utils.py:1199] GPU KV cache size: 655,360 tokens 2025-11-03T17:45:52.6275761Z (EngineCore_DP0 pid=20766) INFO 11-03 17:45:52 [kv_cache_utils.py:1204] Maximum concurrency for 2,048 tokens per request: 320.00x 2025-11-03T17:45:52.9382255Z (EngineCore_DP0 pid=20766) WARNING 11-03 17:45:52 [tokenizer.py:255] Using a slow tokenizer. This might cause a significant slowdown. Consider using a fast tokenizer instead. 2025-11-03T17:45:53.0226749Z (EngineCore_DP0 pid=20766) WARNING 11-03 17:45:53 [__init__.py:2040] The following intended overrides are not keyword args and will be dropped: {'truncation'} 2025-11-03T17:45:53.3297193Z (EngineCore_DP0 pid=20766) INFO 11-03 17:45:53 [gc_utils.py:40] GC Debug Config. enabled:False,top_objects:-1 2025-11-03T17:45:53.3312834Z INFO 11-03 17:45:53 [llm.py:337] Supported tasks: ['generate'] 2025-11-03T17:45:53.8468000Z PASSED 2025-11-03T17:45:53.8608959Z models/test_initialization.py::test_can_initialize_large_subset[Glm4MoeForCausalLM] Fork a new process to run a test 20832 2025-11-03T17:45:53.8620518Z Fork a new process to run a test 0 2025-11-03T17:45:53.8899863Z INFO 11-03 17:45:53 [utils.py:239] non-default args: {'load_format': 'dummy', 'gpu_memory_utilization': 0.8, 'disable_log_stats': True, 'hf_overrides': functools.partial(, model_arch='Glm4MoeForCausalLM', exist_overrides={}, use_original_num_layers=False), 'model_impl': 'vllm', 'model': 'zai-org/GLM-4.5'} 2025-11-03T17:45:53.9726739Z 2025-11-03T17:45:53.9729249Z config.json: 0.00B [00:00, ?B/s] 2025-11-03T17:45:53.9729582Z config.json: 1.01kB [00:00, 4.50MB/s] 2025-11-03T17:46:00.4025410Z INFO 11-03 17:46:00 [model.py:653] Resolved architecture: Glm4MoeForCausalLM 2025-11-03T17:46:00.4025913Z `torch_dtype` is deprecated! Use `dtype` instead! 2025-11-03T17:46:00.4272367Z INFO 11-03 17:46:00 [model.py:1714] Using max model len 131072 2025-11-03T17:46:00.5969008Z INFO 11-03 17:46:00 [scheduler.py:225] Chunked prefill is enabled with max_num_batched_tokens=8192. 2025-11-03T17:46:00.6434682Z 2025-11-03T17:46:00.6436719Z tokenizer_config.json: 0.00B [00:00, ?B/s] 2025-11-03T17:46:00.6437114Z tokenizer_config.json: 7.31kB [00:00, 41.4MB/s] 2025-11-03T17:46:00.8056818Z 2025-11-03T17:46:01.2209446Z tokenizer.json: 0% 0.00/20.0M [00:00, 'use_cudagraph': True, 'cudagraph_num_of_warmups': 1, 'cudagraph_capture_sizes': [512, 504, 496, 488, 480, 472, 464, 456, 448, 440, 432, 424, 416, 408, 400, 392, 384, 376, 368, 360, 352, 344, 336, 328, 320, 312, 304, 296, 288, 280, 272, 264, 256, 248, 240, 232, 224, 216, 208, 200, 192, 184, 176, 168, 160, 152, 144, 136, 128, 120, 112, 104, 96, 88, 80, 72, 64, 56, 48, 40, 32, 24, 16, 8, 4, 2, 1], 'cudagraph_copy_inputs': False, 'full_cuda_graph': True, 'use_inductor_graph_partition': False, 'pass_config': {}, 'max_capture_size': 512, 'local_cache_dir': None} 2025-11-03T17:46:02.8500106Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:46:02.8528847Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:46:02.8535990Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:46:02.8544180Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:46:02.8551935Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:46:02.8560215Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:46:02.8562918Z (EngineCore_DP0 pid=20912) INFO 11-03 17:46:02 [parallel_state.py:1231] rank 0 in world size 1 is assigned as DP rank 0, PP rank 0, TP rank 0, EP rank 0 2025-11-03T17:46:03.1993879Z (EngineCore_DP0 pid=20912) INFO 11-03 17:46:03 [topk_topp_sampler.py:59] Using FlashInfer for top-p & top-k sampling. 2025-11-03T17:46:03.3284837Z (EngineCore_DP0 pid=20912) INFO 11-03 17:46:03 [gpu_model_runner.py:2840] Starting to load model zai-org/GLM-4.5... 2025-11-03T17:46:03.5937084Z (EngineCore_DP0 pid=20912) INFO 11-03 17:46:03 [gpu_model_runner.py:2870] Loading model from scratch... 2025-11-03T17:46:03.6720698Z (EngineCore_DP0 pid=20912) INFO 11-03 17:46:03 [cuda.py:404] Using Flash Attention backend on V1 engine. 2025-11-03T17:46:04.2601723Z (EngineCore_DP0 pid=20912) INFO 11-03 17:46:04 [gpu_model_runner.py:2902] Model loading took 3.2931 GiB and 0.139203 seconds 2025-11-03T17:46:04.2607194Z (EngineCore_DP0 pid=20912) INFO 11-03 17:46:04 [kv_cache_utils.py:1199] GPU KV cache size: 2,621,440 tokens 2025-11-03T17:46:04.2608176Z (EngineCore_DP0 pid=20912) INFO 11-03 17:46:04 [kv_cache_utils.py:1204] Maximum concurrency for 131,072 tokens per request: 20.00x 2025-11-03T17:46:05.2795857Z (EngineCore_DP0 pid=20912) INFO 11-03 17:46:05 [gc_utils.py:40] GC Debug Config. enabled:False,top_objects:-1 2025-11-03T17:46:05.2810324Z INFO 11-03 17:46:05 [llm.py:337] Supported tasks: ['generate'] 2025-11-03T17:46:05.7936508Z PASSED 2025-11-03T17:46:05.8077454Z models/test_initialization.py::test_can_initialize_large_subset[StableLmForCausalLM] Fork a new process to run a test 20978 2025-11-03T17:46:05.8088545Z Fork a new process to run a test 0 2025-11-03T17:46:05.8366898Z INFO 11-03 17:46:05 [utils.py:239] non-default args: {'load_format': 'dummy', 'gpu_memory_utilization': 0.8, 'disable_log_stats': True, 'hf_overrides': functools.partial(, model_arch='StableLmForCausalLM', exist_overrides={}, use_original_num_layers=False), 'model_impl': 'vllm', 'model': 'stabilityai/stablelm-3b-4e1t'} 2025-11-03T17:46:06.0873756Z 2025-11-03T17:46:06.0874202Z config.json: 0% 0.00/599 [00:00, 'use_cudagraph': True, 'cudagraph_num_of_warmups': 1, 'cudagraph_capture_sizes': [512, 504, 496, 488, 480, 472, 464, 456, 448, 440, 432, 424, 416, 408, 400, 392, 384, 376, 368, 360, 352, 344, 336, 328, 320, 312, 304, 296, 288, 280, 272, 264, 256, 248, 240, 232, 224, 216, 208, 200, 192, 184, 176, 168, 160, 152, 144, 136, 128, 120, 112, 104, 96, 88, 80, 72, 64, 56, 48, 40, 32, 24, 16, 8, 4, 2, 1], 'cudagraph_copy_inputs': False, 'full_cuda_graph': True, 'use_inductor_graph_partition': False, 'pass_config': {}, 'max_capture_size': 512, 'local_cache_dir': None} 2025-11-03T17:46:13.9118647Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:46:13.9146326Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:46:13.9153243Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:46:13.9160965Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:46:13.9168867Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:46:13.9176938Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:46:13.9179106Z (EngineCore_DP0 pid=21036) INFO 11-03 17:46:13 [parallel_state.py:1231] rank 0 in world size 1 is assigned as DP rank 0, PP rank 0, TP rank 0, EP rank 0 2025-11-03T17:46:14.2477668Z (EngineCore_DP0 pid=21036) INFO 11-03 17:46:14 [topk_topp_sampler.py:59] Using FlashInfer for top-p & top-k sampling. 2025-11-03T17:46:14.3119787Z (EngineCore_DP0 pid=21036) INFO 11-03 17:46:14 [gpu_model_runner.py:2840] Starting to load model stabilityai/stablelm-3b-4e1t... 2025-11-03T17:46:14.5670542Z (EngineCore_DP0 pid=21036) INFO 11-03 17:46:14 [gpu_model_runner.py:2870] Loading model from scratch... 2025-11-03T17:46:14.6376636Z (EngineCore_DP0 pid=21036) INFO 11-03 17:46:14 [cuda.py:420] Using FlexAttention backend for head_size=80 on V1 engine. 2025-11-03T17:46:14.7548212Z (EngineCore_DP0 pid=21036) WARNING 11-03 17:46:14 [vllm.py:821] `torch.compile` is turned on, but the model stabilityai/stablelm-3b-4e1t does not support it. Please open an issue on GitHub if you want it to be supported. 2025-11-03T17:46:15.2563048Z (EngineCore_DP0 pid=21036) INFO 11-03 17:46:15 [gpu_model_runner.py:2902] Model loading took 0.6296 GiB and 0.190260 seconds 2025-11-03T17:46:15.2567526Z (EngineCore_DP0 pid=21036) INFO 11-03 17:46:15 [kv_cache_utils.py:1199] GPU KV cache size: 1,048,576 tokens 2025-11-03T17:46:15.2568521Z (EngineCore_DP0 pid=21036) INFO 11-03 17:46:15 [kv_cache_utils.py:1204] Maximum concurrency for 4,096 tokens per request: 256.00x 2025-11-03T17:46:15.7315891Z (EngineCore_DP0 pid=21036) INFO 11-03 17:46:15 [gc_utils.py:40] GC Debug Config. enabled:False,top_objects:-1 2025-11-03T17:46:15.7336849Z INFO 11-03 17:46:15 [llm.py:337] Supported tasks: ['generate'] 2025-11-03T17:46:16.1775882Z PASSED 2025-11-03T17:46:16.1917044Z models/test_initialization.py::test_can_initialize_large_subset[Phi3VForCausalLM] Fork a new process to run a test 21102 2025-11-03T17:46:16.1928223Z Fork a new process to run a test 0 2025-11-03T17:46:16.1932120Z `transformers==4.56.2` installed, but `transformers<=4.48` is required to run this model. Reason: Use of deprecated imports which have been removed. 2025-11-03T17:46:16.4819369Z PASSED 2025-11-03T17:46:16.4960567Z models/test_initialization.py::test_can_initialize_large_subset[TransformersMoEEmbeddingModel] Fork a new process to run a test 21103 2025-11-03T17:46:16.4971747Z Fork a new process to run a test 0 2025-11-03T17:46:16.4975718Z `transformers==4.56.2` installed, but `transformers>=4.57.0.dev0` is required to run this model. 2025-11-03T17:46:16.7910852Z PASSED 2025-11-03T17:46:16.8051836Z models/test_initialization.py::test_can_initialize_large_subset[Gemma3ForCausalLM] Fork a new process to run a test 21104 2025-11-03T17:46:16.8062470Z Fork a new process to run a test 0 2025-11-03T17:46:16.8334708Z INFO 11-03 17:46:16 [utils.py:239] non-default args: {'load_format': 'dummy', 'gpu_memory_utilization': 0.8, 'disable_log_stats': True, 'hf_overrides': functools.partial(, model_arch='Gemma3ForCausalLM', exist_overrides={}, use_original_num_layers=False), 'model_impl': 'vllm', 'model': 'google/gemma-3-1b-it'} 2025-11-03T17:46:16.9271817Z 2025-11-03T17:46:16.9273102Z config.json: 0% 0.00/899 [00:00, 'use_cudagraph': True, 'cudagraph_num_of_warmups': 1, 'cudagraph_capture_sizes': [512, 504, 496, 488, 480, 472, 464, 456, 448, 440, 432, 424, 416, 408, 400, 392, 384, 376, 368, 360, 352, 344, 336, 328, 320, 312, 304, 296, 288, 280, 272, 264, 256, 248, 240, 232, 224, 216, 208, 200, 192, 184, 176, 168, 160, 152, 144, 136, 128, 120, 112, 104, 96, 88, 80, 72, 64, 56, 48, 40, 32, 24, 16, 8, 4, 2, 1], 'cudagraph_copy_inputs': False, 'full_cuda_graph': True, 'use_inductor_graph_partition': False, 'pass_config': {}, 'max_capture_size': 512, 'local_cache_dir': None} 2025-11-03T17:46:26.7848035Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:46:26.7875392Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:46:26.7882712Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:46:26.7891219Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:46:26.7898434Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:46:26.7906428Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:46:26.7909444Z (EngineCore_DP0 pid=21183) INFO 11-03 17:46:26 [parallel_state.py:1231] rank 0 in world size 1 is assigned as DP rank 0, PP rank 0, TP rank 0, EP rank 0 2025-11-03T17:46:27.1283258Z (EngineCore_DP0 pid=21183) INFO 11-03 17:46:27 [topk_topp_sampler.py:59] Using FlashInfer for top-p & top-k sampling. 2025-11-03T17:46:27.2025672Z (EngineCore_DP0 pid=21183) INFO 11-03 17:46:27 [gpu_model_runner.py:2840] Starting to load model google/gemma-3-1b-it... 2025-11-03T17:46:27.4622056Z (EngineCore_DP0 pid=21183) INFO 11-03 17:46:27 [gpu_model_runner.py:2870] Loading model from scratch... 2025-11-03T17:46:27.5502990Z (EngineCore_DP0 pid=21183) INFO 11-03 17:46:27 [cuda.py:404] Using Flash Attention backend on V1 engine. 2025-11-03T17:46:28.1216775Z (EngineCore_DP0 pid=21183) INFO 11-03 17:46:28 [gpu_model_runner.py:2902] Model loading took 0.6289 GiB and 0.147303 seconds 2025-11-03T17:46:28.1221880Z (EngineCore_DP0 pid=21183) INFO 11-03 17:46:28 [kv_cache_utils.py:1199] GPU KV cache size: 10,485,760 tokens 2025-11-03T17:46:28.1222802Z (EngineCore_DP0 pid=21183) INFO 11-03 17:46:28 [kv_cache_utils.py:1204] Maximum concurrency for 32,768 tokens per request: 1202.50x 2025-11-03T17:46:30.9300417Z (EngineCore_DP0 pid=21183) INFO 11-03 17:46:30 [gc_utils.py:40] GC Debug Config. enabled:False,top_objects:-1 2025-11-03T17:46:30.9314475Z INFO 11-03 17:46:30 [llm.py:337] Supported tasks: ['generate'] 2025-11-03T17:46:31.4420022Z PASSED 2025-11-03T17:46:31.4560477Z models/test_initialization.py::test_can_initialize_large_subset[CwmForCausalLM] Fork a new process to run a test 21249 2025-11-03T17:46:31.4571888Z Fork a new process to run a test 0 2025-11-03T17:46:31.4573206Z Model is not available online 2025-11-03T17:46:31.7466735Z PASSED 2025-11-03T17:46:31.7607147Z models/test_initialization.py::test_can_initialize_large_subset[GPTNeoXForCausalLM] Fork a new process to run a test 21250 2025-11-03T17:46:31.7617336Z Fork a new process to run a test 0 2025-11-03T17:46:31.7895007Z INFO 11-03 17:46:31 [utils.py:239] non-default args: {'load_format': 'dummy', 'gpu_memory_utilization': 0.8, 'disable_log_stats': True, 'hf_overrides': functools.partial(, model_arch='GPTNeoXForCausalLM', exist_overrides={}, use_original_num_layers=False), 'model_impl': 'vllm', 'model': 'EleutherAI/pythia-70m'} 2025-11-03T17:46:31.8849041Z 2025-11-03T17:46:31.8850190Z config.json: 0% 0.00/567 [00:00, 'use_cudagraph': True, 'cudagraph_num_of_warmups': 1, 'cudagraph_capture_sizes': [512, 504, 496, 488, 480, 472, 464, 456, 448, 440, 432, 424, 416, 408, 400, 392, 384, 376, 368, 360, 352, 344, 336, 328, 320, 312, 304, 296, 288, 280, 272, 264, 256, 248, 240, 232, 224, 216, 208, 200, 192, 184, 176, 168, 160, 152, 144, 136, 128, 120, 112, 104, 96, 88, 80, 72, 64, 56, 48, 40, 32, 24, 16, 8, 4, 2, 1], 'cudagraph_copy_inputs': False, 'full_cuda_graph': True, 'use_inductor_graph_partition': False, 'pass_config': {}, 'max_capture_size': 512, 'local_cache_dir': None} 2025-11-03T17:46:39.7114500Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:46:39.7142170Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:46:39.7150274Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:46:39.7158730Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:46:39.7166799Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:46:39.7175303Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:46:39.7178442Z (EngineCore_DP0 pid=21308) INFO 11-03 17:46:39 [parallel_state.py:1231] rank 0 in world size 1 is assigned as DP rank 0, PP rank 0, TP rank 0, EP rank 0 2025-11-03T17:46:40.0493617Z (EngineCore_DP0 pid=21308) INFO 11-03 17:46:40 [topk_topp_sampler.py:59] Using FlashInfer for top-p & top-k sampling. 2025-11-03T17:46:40.0853317Z (EngineCore_DP0 pid=21308) INFO 11-03 17:46:40 [gpu_model_runner.py:2840] Starting to load model EleutherAI/pythia-70m... 2025-11-03T17:46:40.3413812Z (EngineCore_DP0 pid=21308) INFO 11-03 17:46:40 [gpu_model_runner.py:2870] Loading model from scratch... 2025-11-03T17:46:40.4072104Z (EngineCore_DP0 pid=21308) INFO 11-03 17:46:40 [cuda.py:404] Using Flash Attention backend on V1 engine. 2025-11-03T17:46:40.9692192Z (EngineCore_DP0 pid=21308) INFO 11-03 17:46:40 [gpu_model_runner.py:2902] Model loading took 0.1036 GiB and 0.124730 seconds 2025-11-03T17:46:40.9696588Z (EngineCore_DP0 pid=21308) INFO 11-03 17:46:40 [kv_cache_utils.py:1199] GPU KV cache size: 5,242,880 tokens 2025-11-03T17:46:40.9697518Z (EngineCore_DP0 pid=21308) INFO 11-03 17:46:40 [kv_cache_utils.py:1204] Maximum concurrency for 2,048 tokens per request: 2560.00x 2025-11-03T17:46:41.8488714Z (EngineCore_DP0 pid=21308) INFO 11-03 17:46:41 [gc_utils.py:40] GC Debug Config. enabled:False,top_objects:-1 2025-11-03T17:46:41.8503760Z INFO 11-03 17:46:41 [llm.py:337] Supported tasks: ['generate'] 2025-11-03T17:46:42.2827535Z PASSED 2025-11-03T17:46:42.2968437Z models/test_initialization.py::test_can_initialize_large_subset[ModernBertForTokenClassification] Fork a new process to run a test 21374 2025-11-03T17:46:42.2979405Z Fork a new process to run a test 0 2025-11-03T17:46:42.3259787Z INFO 11-03 17:46:42 [utils.py:239] non-default args: {'load_format': 'dummy', 'gpu_memory_utilization': 0.8, 'disable_log_stats': True, 'hf_overrides': functools.partial(, model_arch='ModernBertForTokenClassification', exist_overrides={}, use_original_num_layers=False), 'model_impl': 'vllm', 'model': 'disham993/electrical-ner-ModernBERT-base'} 2025-11-03T17:46:42.4584127Z 2025-11-03T17:46:42.4586239Z config.json: 0.00B [00:00, ?B/s] 2025-11-03T17:46:42.4586564Z config.json: 1.96kB [00:00, 10.6MB/s] 2025-11-03T17:46:48.8320875Z INFO 11-03 17:46:48 [model.py:915] Resolved `--runner auto` to `--runner pooling`. Pass the value explicitly to silence this message. 2025-11-03T17:46:48.8321952Z INFO 11-03 17:46:48 [model.py:653] Resolved architecture: ModernBertForTokenClassification 2025-11-03T17:46:48.8322420Z `torch_dtype` is deprecated! Use `dtype` instead! 2025-11-03T17:46:48.8570800Z INFO 11-03 17:46:48 [model.py:1939] Downcasting torch.float32 to torch.float16. 2025-11-03T17:46:48.8923759Z 2025-11-03T17:46:48.8925988Z tokenizer_config.json: 0.00B [00:00, ?B/s] 2025-11-03T17:46:48.8926330Z tokenizer_config.json: 20.9kB [00:00, 104MB/s] 2025-11-03T17:46:48.8934721Z INFO 11-03 17:46:48 [model.py:1714] Using max model len 8192 2025-11-03T17:46:48.9286354Z INFO 11-03 17:46:48 [arg_utils.py:1725] (Disabling) chunked prefill by default 2025-11-03T17:46:48.9286834Z INFO 11-03 17:46:48 [arg_utils.py:1728] (Disabling) prefix caching by default 2025-11-03T17:46:49.0699184Z INFO 11-03 17:46:49 [vllm.py:433] Only "last" pooling supports chunked prefill and prefix caching; disabling both. 2025-11-03T17:46:49.1988686Z 2025-11-03T17:46:49.2111426Z tokenizer.json: 0.00B [00:00, ?B/s] 2025-11-03T17:46:49.2111765Z tokenizer.json: 3.58MB [00:00, 289MB/s] 2025-11-03T17:46:49.2483099Z 2025-11-03T17:46:49.2483942Z added_tokens.json: 0.00B [00:00, ?B/s] 2025-11-03T17:46:49.2484303Z added_tokens.json: 1.99kB [00:00, 21.5MB/s] 2025-11-03T17:46:49.2862495Z 2025-11-03T17:46:49.2863758Z special_tokens_map.json: 0% 0.00/694 [00:00, 'use_cudagraph': True, 'cudagraph_num_of_warmups': 1, 'cudagraph_capture_sizes': [512, 504, 496, 488, 480, 472, 464, 456, 448, 440, 432, 424, 416, 408, 400, 392, 384, 376, 368, 360, 352, 344, 336, 328, 320, 312, 304, 296, 288, 280, 272, 264, 256, 248, 240, 232, 224, 216, 208, 200, 192, 184, 176, 168, 160, 152, 144, 136, 128, 120, 112, 104, 96, 88, 80, 72, 64, 56, 48, 40, 32, 24, 16, 8, 4, 2, 1], 'cudagraph_copy_inputs': False, 'full_cuda_graph': False, 'use_inductor_graph_partition': False, 'pass_config': {}, 'max_capture_size': 512, 'local_cache_dir': None} 2025-11-03T17:46:50.2538535Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:46:50.2569707Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:46:50.2577174Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:46:50.2584983Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:46:50.2593220Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:46:50.2601106Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:46:50.2603869Z (EngineCore_DP0 pid=21433) INFO 11-03 17:46:50 [parallel_state.py:1231] rank 0 in world size 1 is assigned as DP rank 0, PP rank 0, TP rank 0, EP rank 0 2025-11-03T17:46:50.5953427Z (EngineCore_DP0 pid=21433) INFO 11-03 17:46:50 [topk_topp_sampler.py:59] Using FlashInfer for top-p & top-k sampling. 2025-11-03T17:46:50.6156142Z (EngineCore_DP0 pid=21433) INFO 11-03 17:46:50 [gpu_model_runner.py:2840] Starting to load model disham993/electrical-ner-ModernBERT-base... 2025-11-03T17:46:50.8748997Z (EngineCore_DP0 pid=21433) INFO 11-03 17:46:50 [gpu_model_runner.py:2870] Loading model from scratch... 2025-11-03T17:46:50.9405500Z (EngineCore_DP0 pid=21433) INFO 11-03 17:46:50 [cuda.py:404] Using Flash Attention backend on V1 engine. 2025-11-03T17:46:51.5118729Z (EngineCore_DP0 pid=21433) INFO 11-03 17:46:51 [gpu_model_runner.py:2902] Model loading took 0.0837 GiB and 0.125765 seconds 2025-11-03T17:46:51.6905996Z (EngineCore_DP0 pid=21433) INFO 11-03 17:46:51 [core.py:142] Disabling chunked prefill for model without KVCache 2025-11-03T17:46:51.6931997Z (EngineCore_DP0 pid=21433) INFO 11-03 17:46:51 [vllm.py:433] Only "last" pooling supports chunked prefill and prefix caching; disabling both. 2025-11-03T17:46:51.9547797Z (EngineCore_DP0 pid=21433) INFO 11-03 17:46:51 [gc_utils.py:40] GC Debug Config. enabled:False,top_objects:-1 2025-11-03T17:46:51.9560773Z INFO 11-03 17:46:51 [llm.py:337] Supported tasks: ['encode'] 2025-11-03T17:46:52.3897872Z PASSED 2025-11-03T17:46:52.4038721Z models/test_initialization.py::test_can_initialize_large_subset[FalconForCausalLM] Fork a new process to run a test 21499 2025-11-03T17:46:52.4049560Z Fork a new process to run a test 0 2025-11-03T17:46:52.4330104Z INFO 11-03 17:46:52 [utils.py:239] non-default args: {'load_format': 'dummy', 'gpu_memory_utilization': 0.8, 'disable_log_stats': True, 'hf_overrides': functools.partial(, model_arch='FalconForCausalLM', exist_overrides={}, use_original_num_layers=False), 'model_impl': 'vllm', 'model': 'tiiuae/falcon-7b'} 2025-11-03T17:46:52.5437834Z 2025-11-03T17:46:52.5440280Z config.json: 0.00B [00:00, ?B/s] 2025-11-03T17:46:52.5440593Z config.json: 1.05kB [00:00, 5.30MB/s] 2025-11-03T17:46:52.6740727Z INFO 11-03 17:46:52 [model.py:653] Resolved architecture: FalconForCausalLM 2025-11-03T17:46:52.6741248Z `torch_dtype` is deprecated! Use `dtype` instead! 2025-11-03T17:46:52.6983715Z INFO 11-03 17:46:52 [model.py:1714] Using max model len 2048 2025-11-03T17:46:52.8681858Z INFO 11-03 17:46:52 [scheduler.py:225] Chunked prefill is enabled with max_num_batched_tokens=8192. 2025-11-03T17:46:52.9035397Z 2025-11-03T17:46:52.9036619Z tokenizer_config.json: 0% 0.00/287 [00:00, 'use_cudagraph': True, 'cudagraph_num_of_warmups': 1, 'cudagraph_capture_sizes': [512, 504, 496, 488, 480, 472, 464, 456, 448, 440, 432, 424, 416, 408, 400, 392, 384, 376, 368, 360, 352, 344, 336, 328, 320, 312, 304, 296, 288, 280, 272, 264, 256, 248, 240, 232, 224, 216, 208, 200, 192, 184, 176, 168, 160, 152, 144, 136, 128, 120, 112, 104, 96, 88, 80, 72, 64, 56, 48, 40, 32, 24, 16, 8, 4, 2, 1], 'cudagraph_copy_inputs': False, 'full_cuda_graph': True, 'use_inductor_graph_partition': False, 'pass_config': {}, 'max_capture_size': 512, 'local_cache_dir': None} 2025-11-03T17:46:53.9925715Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:46:53.9953701Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:46:53.9961404Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:46:53.9969461Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:46:53.9977688Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:46:53.9985997Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:46:53.9988975Z (EngineCore_DP0 pid=21507) INFO 11-03 17:46:53 [parallel_state.py:1231] rank 0 in world size 1 is assigned as DP rank 0, PP rank 0, TP rank 0, EP rank 0 2025-11-03T17:46:54.3371540Z (EngineCore_DP0 pid=21507) INFO 11-03 17:46:54 [topk_topp_sampler.py:59] Using FlashInfer for top-p & top-k sampling. 2025-11-03T17:46:54.4464018Z (EngineCore_DP0 pid=21507) INFO 11-03 17:46:54 [gpu_model_runner.py:2840] Starting to load model tiiuae/falcon-7b... 2025-11-03T17:46:54.7072552Z (EngineCore_DP0 pid=21507) INFO 11-03 17:46:54 [gpu_model_runner.py:2870] Loading model from scratch... 2025-11-03T17:46:54.7722810Z (EngineCore_DP0 pid=21507) INFO 11-03 17:46:54 [cuda.py:404] Using Flash Attention backend on V1 engine. 2025-11-03T17:46:55.3522997Z (EngineCore_DP0 pid=21507) INFO 11-03 17:46:55 [gpu_model_runner.py:2902] Model loading took 0.9383 GiB and 0.124364 seconds 2025-11-03T17:46:55.3528353Z (EngineCore_DP0 pid=21507) INFO 11-03 17:46:55 [kv_cache_utils.py:1199] GPU KV cache size: 41,943,040 tokens 2025-11-03T17:46:55.3529304Z (EngineCore_DP0 pid=21507) INFO 11-03 17:46:55 [kv_cache_utils.py:1204] Maximum concurrency for 2,048 tokens per request: 20480.00x 2025-11-03T17:46:59.1344171Z (EngineCore_DP0 pid=21507) INFO 11-03 17:46:59 [gc_utils.py:40] GC Debug Config. enabled:False,top_objects:-1 2025-11-03T17:46:59.1357436Z INFO 11-03 17:46:59 [llm.py:337] Supported tasks: ['generate'] 2025-11-03T17:46:59.6274798Z PASSED 2025-11-03T17:46:59.6415155Z models/test_initialization.py::test_can_initialize_large_subset[Phi4MMForCausalLM] Fork a new process to run a test 21573 2025-11-03T17:46:59.6426647Z Fork a new process to run a test 0 2025-11-03T17:46:59.6696695Z INFO 11-03 17:46:59 [utils.py:239] non-default args: {'trust_remote_code': True, 'load_format': 'dummy', 'gpu_memory_utilization': 0.8, 'disable_log_stats': True, 'hf_overrides': functools.partial(, model_arch='Phi4MMForCausalLM', exist_overrides={}, use_original_num_layers=False), 'model_impl': 'vllm', 'model': 'microsoft/Phi-4-multimodal-instruct'} 2025-11-03T17:46:59.7516681Z 2025-11-03T17:46:59.7519939Z config.json: 0.00B [00:00, ?B/s] 2025-11-03T17:46:59.7520293Z config.json: 4.63kB [00:00, 17.9MB/s] 2025-11-03T17:46:59.8231770Z 2025-11-03T17:46:59.8234103Z configuration_phi4mm.py: 0.00B [00:00, ?B/s] 2025-11-03T17:46:59.8234511Z configuration_phi4mm.py: 11.0kB [00:00, 53.3MB/s] 2025-11-03T17:46:59.8331839Z A new version of the following files was downloaded from https://huggingface.co/microsoft/Phi-4-multimodal-instruct: 2025-11-03T17:46:59.8332413Z - configuration_phi4mm.py 2025-11-03T17:46:59.8333027Z . Make sure to double-check they do not contain any added malicious code. To avoid downloading new versions of the code file, you can pin a revision. 2025-11-03T17:46:59.8629472Z INFO 11-03 17:46:59 [config.py:408] Replacing legacy 'type' key with 'rope_type' 2025-11-03T17:47:00.0485839Z 2025-11-03T17:47:00.0486722Z preprocessor_config.json: 0% 0.00/482 [00:00, 'use_cudagraph': True, 'cudagraph_num_of_warmups': 1, 'cudagraph_capture_sizes': [512, 504, 496, 488, 480, 472, 464, 456, 448, 440, 432, 424, 416, 408, 400, 392, 384, 376, 368, 360, 352, 344, 336, 328, 320, 312, 304, 296, 288, 280, 272, 264, 256, 248, 240, 232, 224, 216, 208, 200, 192, 184, 176, 168, 160, 152, 144, 136, 128, 120, 112, 104, 96, 88, 80, 72, 64, 56, 48, 40, 32, 24, 16, 8, 4, 2, 1], 'cudagraph_copy_inputs': False, 'full_cuda_graph': True, 'use_inductor_graph_partition': False, 'pass_config': {}, 'max_capture_size': 512, 'local_cache_dir': None} 2025-11-03T17:47:08.9788373Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:47:08.9817489Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:47:08.9825822Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:47:08.9833844Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:47:08.9841275Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:47:08.9849063Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:47:08.9851667Z (EngineCore_DP0 pid=21652) INFO 11-03 17:47:08 [parallel_state.py:1231] rank 0 in world size 1 is assigned as DP rank 0, PP rank 0, TP rank 0, EP rank 0 2025-11-03T17:47:09.3234251Z (EngineCore_DP0 pid=21652) INFO 11-03 17:47:09 [topk_topp_sampler.py:59] Using FlashInfer for top-p & top-k sampling. 2025-11-03T17:47:09.4541521Z (EngineCore_DP0 pid=21652) 2025-11-03T17:47:09.4541959Z processor_config.json: 0% 0.00/121 [00:00, model_arch='TeleFLMForCausalLM', exist_overrides={}, use_original_num_layers=False), 'model_impl': 'vllm', 'model': 'CofeAI/FLM-2-52B-Instruct-2407'} 2025-11-03T17:47:16.5663580Z 2025-11-03T17:47:16.5664695Z config.json: 0% 0.00/940 [00:00, 'use_cudagraph': True, 'cudagraph_num_of_warmups': 1, 'cudagraph_capture_sizes': [512, 504, 496, 488, 480, 472, 464, 456, 448, 440, 432, 424, 416, 408, 400, 392, 384, 376, 368, 360, 352, 344, 336, 328, 320, 312, 304, 296, 288, 280, 272, 264, 256, 248, 240, 232, 224, 216, 208, 200, 192, 184, 176, 168, 160, 152, 144, 136, 128, 120, 112, 104, 96, 88, 80, 72, 64, 56, 48, 40, 32, 24, 16, 8, 4, 2, 1], 'cudagraph_copy_inputs': False, 'full_cuda_graph': False, 'use_inductor_graph_partition': False, 'pass_config': {}, 'max_capture_size': 512, 'local_cache_dir': None} 2025-11-03T17:47:24.9409082Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:47:24.9437621Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:47:24.9444711Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:47:24.9452416Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:47:24.9460237Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:47:24.9468490Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:47:24.9471185Z (EngineCore_DP0 pid=21827) INFO 11-03 17:47:24 [parallel_state.py:1231] rank 0 in world size 1 is assigned as DP rank 0, PP rank 0, TP rank 0, EP rank 0 2025-11-03T17:47:25.2831273Z (EngineCore_DP0 pid=21827) INFO 11-03 17:47:25 [topk_topp_sampler.py:59] Using FlashInfer for top-p & top-k sampling. 2025-11-03T17:47:25.3597607Z (EngineCore_DP0 pid=21827) INFO 11-03 17:47:25 [gpu_model_runner.py:2840] Starting to load model CofeAI/FLM-2-52B-Instruct-2407... 2025-11-03T17:47:25.6180461Z (EngineCore_DP0 pid=21827) INFO 11-03 17:47:25 [gpu_model_runner.py:2870] Loading model from scratch... 2025-11-03T17:47:25.6840945Z (EngineCore_DP0 pid=21827) INFO 11-03 17:47:25 [cuda.py:404] Using Flash Attention backend on V1 engine. 2025-11-03T17:47:26.3024479Z (EngineCore_DP0 pid=21827) INFO 11-03 17:47:26 [gpu_model_runner.py:2902] Model loading took 2.7349 GiB and 0.170046 seconds 2025-11-03T17:47:26.3028629Z (EngineCore_DP0 pid=21827) INFO 11-03 17:47:26 [kv_cache_utils.py:1199] GPU KV cache size: 327,680 tokens 2025-11-03T17:47:26.3029933Z (EngineCore_DP0 pid=21827) INFO 11-03 17:47:26 [kv_cache_utils.py:1204] Maximum concurrency for 2,048 tokens per request: 160.00x 2025-11-03T17:47:26.6705444Z (EngineCore_DP0 pid=21827) WARNING 11-03 17:47:26 [tokenizer.py:255] Using a slow tokenizer. This might cause a significant slowdown. Consider using a fast tokenizer instead. 2025-11-03T17:47:27.0308855Z (EngineCore_DP0 pid=21827) INFO 11-03 17:47:27 [gc_utils.py:40] GC Debug Config. enabled:False,top_objects:-1 2025-11-03T17:47:27.0322094Z INFO 11-03 17:47:27 [llm.py:337] Supported tasks: ['embed'] 2025-11-03T17:47:27.5115203Z PASSED 2025-11-03T17:47:27.5256929Z models/test_initialization.py::test_can_initialize_large_subset[Cohere2VisionForConditionalGeneration] Fork a new process to run a test 21893 2025-11-03T17:47:27.5268125Z Fork a new process to run a test 0 2025-11-03T17:47:27.5542288Z INFO 11-03 17:47:27 [utils.py:239] non-default args: {'load_format': 'dummy', 'gpu_memory_utilization': 0.8, 'disable_log_stats': True, 'hf_overrides': functools.partial(, model_arch='Cohere2VisionForConditionalGeneration', exist_overrides={}, use_original_num_layers=False), 'model_impl': 'vllm', 'model': 'CohereLabs/command-a-vision-07-2025'} 2025-11-03T17:47:27.6728253Z 2025-11-03T17:47:27.6729294Z config.json: 0% 0.00/3.59k [00:00, 'use_cudagraph': True, 'cudagraph_num_of_warmups': 1, 'cudagraph_capture_sizes': [512, 504, 496, 488, 480, 472, 464, 456, 448, 440, 432, 424, 416, 408, 400, 392, 384, 376, 368, 360, 352, 344, 336, 328, 320, 312, 304, 296, 288, 280, 272, 264, 256, 248, 240, 232, 224, 216, 208, 200, 192, 184, 176, 168, 160, 152, 144, 136, 128, 120, 112, 104, 96, 88, 80, 72, 64, 56, 48, 40, 32, 24, 16, 8, 4, 2, 1], 'cudagraph_copy_inputs': False, 'full_cuda_graph': True, 'use_inductor_graph_partition': False, 'pass_config': {}, 'max_capture_size': 512, 'local_cache_dir': None} 2025-11-03T17:47:37.6354792Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:47:37.6383014Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:47:37.6390600Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:47:37.6399031Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:47:37.6406904Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:47:37.6415188Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:47:37.6418039Z (EngineCore_DP0 pid=21971) INFO 11-03 17:47:37 [parallel_state.py:1231] rank 0 in world size 1 is assigned as DP rank 0, PP rank 0, TP rank 0, EP rank 0 2025-11-03T17:47:37.9807772Z (EngineCore_DP0 pid=21971) INFO 11-03 17:47:37 [topk_topp_sampler.py:59] Using FlashInfer for top-p & top-k sampling. 2025-11-03T17:47:42.2564324Z (EngineCore_DP0 pid=21971) INFO 11-03 17:47:42 [gpu_model_runner.py:2840] Starting to load model CohereLabs/command-a-vision-07-2025... 2025-11-03T17:47:42.6712832Z (EngineCore_DP0 pid=21971) INFO 11-03 17:47:42 [gpu_model_runner.py:2870] Loading model from scratch... 2025-11-03T17:47:42.6834096Z (EngineCore_DP0 pid=21971) INFO 11-03 17:47:42 [layer.py:497] MultiHeadAttention attn_backend: _Backend.XFORMERS, use_upstream_fa: False 2025-11-03T17:47:42.7265031Z (EngineCore_DP0 pid=21971) INFO 11-03 17:47:42 [cuda.py:404] Using Flash Attention backend on V1 engine. 2025-11-03T17:47:43.3012158Z (EngineCore_DP0 pid=21971) INFO 11-03 17:47:43 [gpu_model_runner.py:2902] Model loading took 9.8899 GiB and 0.103605 seconds 2025-11-03T17:47:43.3016900Z (EngineCore_DP0 pid=21971) INFO 11-03 17:47:43 [kv_cache_utils.py:1199] GPU KV cache size: 2,621,440 tokens 2025-11-03T17:47:43.3017862Z (EngineCore_DP0 pid=21971) INFO 11-03 17:47:43 [kv_cache_utils.py:1204] Maximum concurrency for 500,000 tokens per request: 213.06x 2025-11-03T17:47:44.6433291Z (EngineCore_DP0 pid=21971) INFO 11-03 17:47:44 [gc_utils.py:40] GC Debug Config. enabled:False,top_objects:-1 2025-11-03T17:47:44.6462575Z INFO 11-03 17:47:44 [llm.py:337] Supported tasks: ['generate'] 2025-11-03T17:47:45.2600940Z PASSED 2025-11-03T17:47:45.2742655Z models/test_initialization.py::test_can_initialize_large_subset[Olmo2ForCausalLM] Fork a new process to run a test 22053 2025-11-03T17:47:45.2753411Z Fork a new process to run a test 0 2025-11-03T17:47:45.3030137Z INFO 11-03 17:47:45 [utils.py:239] non-default args: {'load_format': 'dummy', 'gpu_memory_utilization': 0.8, 'disable_log_stats': True, 'hf_overrides': functools.partial(, model_arch='Olmo2ForCausalLM', exist_overrides={}, use_original_num_layers=False), 'model_impl': 'vllm', 'model': 'allenai/OLMo-2-0425-1B'} 2025-11-03T17:47:45.3884700Z 2025-11-03T17:47:45.3886039Z config.json: 0% 0.00/623 [00:00, 'use_cudagraph': True, 'cudagraph_num_of_warmups': 1, 'cudagraph_capture_sizes': [512, 504, 496, 488, 480, 472, 464, 456, 448, 440, 432, 424, 416, 408, 400, 392, 384, 376, 368, 360, 352, 344, 336, 328, 320, 312, 304, 296, 288, 280, 272, 264, 256, 248, 240, 232, 224, 216, 208, 200, 192, 184, 176, 168, 160, 152, 144, 136, 128, 120, 112, 104, 96, 88, 80, 72, 64, 56, 48, 40, 32, 24, 16, 8, 4, 2, 1], 'cudagraph_copy_inputs': False, 'full_cuda_graph': True, 'use_inductor_graph_partition': False, 'pass_config': {}, 'max_capture_size': 512, 'local_cache_dir': None} 2025-11-03T17:47:53.4899738Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:47:53.4927645Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:47:53.4935172Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:47:53.4943181Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:47:53.4950894Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:47:53.4958643Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:47:53.4961816Z (EngineCore_DP0 pid=22111) INFO 11-03 17:47:53 [parallel_state.py:1231] rank 0 in world size 1 is assigned as DP rank 0, PP rank 0, TP rank 0, EP rank 0 2025-11-03T17:47:53.8335366Z (EngineCore_DP0 pid=22111) INFO 11-03 17:47:53 [topk_topp_sampler.py:59] Using FlashInfer for top-p & top-k sampling. 2025-11-03T17:47:53.8952319Z (EngineCore_DP0 pid=22111) INFO 11-03 17:47:53 [gpu_model_runner.py:2840] Starting to load model allenai/OLMo-2-0425-1B... 2025-11-03T17:47:54.1570047Z (EngineCore_DP0 pid=22111) INFO 11-03 17:47:54 [gpu_model_runner.py:2870] Loading model from scratch... 2025-11-03T17:47:54.1591553Z (EngineCore_DP0 pid=22111) INFO 11-03 17:47:54 [cuda.py:404] Using Flash Attention backend on V1 engine. 2025-11-03T17:47:54.7991864Z (EngineCore_DP0 pid=22111) INFO 11-03 17:47:54 [gpu_model_runner.py:2902] Model loading took 0.8916 GiB and 0.125839 seconds 2025-11-03T17:47:54.7996154Z (EngineCore_DP0 pid=22111) INFO 11-03 17:47:54 [kv_cache_utils.py:1199] GPU KV cache size: 1,310,720 tokens 2025-11-03T17:47:54.7997482Z (EngineCore_DP0 pid=22111) INFO 11-03 17:47:54 [kv_cache_utils.py:1204] Maximum concurrency for 4,096 tokens per request: 320.00x 2025-11-03T17:47:55.4073448Z (EngineCore_DP0 pid=22111) INFO 11-03 17:47:55 [gc_utils.py:40] GC Debug Config. enabled:False,top_objects:-1 2025-11-03T17:47:55.4087015Z INFO 11-03 17:47:55 [llm.py:337] Supported tasks: ['generate'] 2025-11-03T17:47:55.8586721Z PASSED 2025-11-03T17:47:55.8726471Z models/test_initialization.py::test_can_initialize_large_subset[OlmoeForCausalLM] Fork a new process to run a test 22177 2025-11-03T17:47:55.8737608Z Fork a new process to run a test 0 2025-11-03T17:47:55.9015484Z INFO 11-03 17:47:55 [utils.py:239] non-default args: {'load_format': 'dummy', 'gpu_memory_utilization': 0.8, 'disable_log_stats': True, 'hf_overrides': functools.partial(, model_arch='OlmoeForCausalLM', exist_overrides={}, use_original_num_layers=False), 'model_impl': 'vllm', 'model': 'allenai/OLMoE-1B-7B-0924-Instruct'} 2025-11-03T17:47:55.9788560Z 2025-11-03T17:47:55.9789877Z config.json: 0% 0.00/759 [00:00, 'use_cudagraph': True, 'cudagraph_num_of_warmups': 1, 'cudagraph_capture_sizes': [512, 504, 496, 488, 480, 472, 464, 456, 448, 440, 432, 424, 416, 408, 400, 392, 384, 376, 368, 360, 352, 344, 336, 328, 320, 312, 304, 296, 288, 280, 272, 264, 256, 248, 240, 232, 224, 216, 208, 200, 192, 184, 176, 168, 160, 152, 144, 136, 128, 120, 112, 104, 96, 88, 80, 72, 64, 56, 48, 40, 32, 24, 16, 8, 4, 2, 1], 'cudagraph_copy_inputs': False, 'full_cuda_graph': True, 'use_inductor_graph_partition': False, 'pass_config': {}, 'max_capture_size': 512, 'local_cache_dir': None} 2025-11-03T17:48:03.6120728Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:48:03.6148593Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:48:03.6155731Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:48:03.6163197Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:48:03.6170430Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:48:03.6178172Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:48:03.6180892Z (EngineCore_DP0 pid=22235) INFO 11-03 17:48:03 [parallel_state.py:1231] rank 0 in world size 1 is assigned as DP rank 0, PP rank 0, TP rank 0, EP rank 0 2025-11-03T17:48:03.9476138Z (EngineCore_DP0 pid=22235) INFO 11-03 17:48:03 [topk_topp_sampler.py:59] Using FlashInfer for top-p & top-k sampling. 2025-11-03T17:48:03.9961404Z (EngineCore_DP0 pid=22235) INFO 11-03 17:48:03 [gpu_model_runner.py:2840] Starting to load model allenai/OLMoE-1B-7B-0924-Instruct... 2025-11-03T17:48:04.2496904Z (EngineCore_DP0 pid=22235) INFO 11-03 17:48:04 [gpu_model_runner.py:2870] Loading model from scratch... 2025-11-03T17:48:04.3153206Z (EngineCore_DP0 pid=22235) INFO 11-03 17:48:04 [cuda.py:404] Using Flash Attention backend on V1 engine. 2025-11-03T17:48:04.8726128Z (EngineCore_DP0 pid=22235) INFO 11-03 17:48:04 [gpu_model_runner.py:2902] Model loading took 0.4395 GiB and 0.124353 seconds 2025-11-03T17:48:04.8731054Z (EngineCore_DP0 pid=22235) INFO 11-03 17:48:04 [kv_cache_utils.py:1199] GPU KV cache size: 1,310,720 tokens 2025-11-03T17:48:04.8731832Z (EngineCore_DP0 pid=22235) INFO 11-03 17:48:04 [kv_cache_utils.py:1204] Maximum concurrency for 4,096 tokens per request: 320.00x 2025-11-03T17:48:05.3254844Z (EngineCore_DP0 pid=22235) INFO 11-03 17:48:05 [gc_utils.py:40] GC Debug Config. enabled:False,top_objects:-1 2025-11-03T17:48:05.3269398Z INFO 11-03 17:48:05 [llm.py:337] Supported tasks: ['generate'] 2025-11-03T17:48:05.7753830Z PASSED 2025-11-03T17:48:05.7893651Z models/test_initialization.py::test_can_initialize_large_subset[MolmoForCausalLM] Fork a new process to run a test 22301 2025-11-03T17:48:05.7904525Z Fork a new process to run a test 0 2025-11-03T17:48:05.7908229Z `transformers==4.56.2` installed, but `transformers<=4.48` is required to run this model. Reason: Incorrectly-detected `tensorflow` import. 2025-11-03T17:48:06.0835868Z PASSED 2025-11-03T17:48:06.0975608Z models/test_initialization.py::test_can_initialize_large_subset[MedusaModel] Fork a new process to run a test 22302 2025-11-03T17:48:06.0987914Z Fork a new process to run a test 0 2025-11-03T17:48:06.1265308Z INFO 11-03 17:48:06 [utils.py:239] non-default args: {'load_format': 'dummy', 'gpu_memory_utilization': 0.8, 'disable_log_stats': True, 'hf_overrides': functools.partial(, model_arch='MedusaModel', exist_overrides={}, use_original_num_layers=False), 'speculative_config': {'model': 'abhigoyal/vllm-medusa-llama-68m-random', 'num_speculative_tokens': 1}, 'model_impl': 'vllm', 'model': 'JackFram/llama-68m'} 2025-11-03T17:48:06.3871995Z 2025-11-03T17:48:06.3873075Z config.json: 0% 0.00/545 [00:00. This is expected, and simply means that the `legacy` (previous) behavior will be used so nothing changes for you. If you want to use the new behaviour, set `legacy=False`. This should only be set if you understand what it means, and thoroughly read the reason why this was added as explained in https://github.com/huggingface/transformers/pull/24565 - if you loaded a llama tokenizer from a GGUF file you can ignore this message 2025-11-03T17:48:14.0780939Z You are using the default legacy behaviour of the . This is expected, and simply means that the `legacy` (previous) behavior will be used so nothing changes for you. If you want to use the new behaviour, set `legacy=False`. This should only be set if you understand what it means, and thoroughly read the reason why this was added as explained in https://github.com/huggingface/transformers/pull/24565 - if you loaded a llama tokenizer from a GGUF file you can ignore this message. 2025-11-03T17:48:14.2893811Z 2025-11-03T17:48:14.2894540Z generation_config.json: 0% 0.00/137 [00:00, 'use_cudagraph': True, 'cudagraph_num_of_warmups': 1, 'cudagraph_capture_sizes': [512, 504, 496, 488, 480, 472, 464, 456, 448, 440, 432, 424, 416, 408, 400, 392, 384, 376, 368, 360, 352, 344, 336, 328, 320, 312, 304, 296, 288, 280, 272, 264, 256, 248, 240, 232, 224, 216, 208, 200, 192, 184, 176, 168, 160, 152, 144, 136, 128, 120, 112, 104, 96, 88, 80, 72, 64, 56, 48, 40, 32, 24, 16, 8, 4, 2, 1], 'cudagraph_copy_inputs': False, 'full_cuda_graph': True, 'use_inductor_graph_partition': False, 'pass_config': {}, 'max_capture_size': 512, 'local_cache_dir': None} 2025-11-03T17:48:15.0056131Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:48:15.0083992Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:48:15.0090976Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:48:15.0098890Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:48:15.0106451Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:48:15.0114741Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:48:15.0117498Z (EngineCore_DP0 pid=22380) INFO 11-03 17:48:15 [parallel_state.py:1231] rank 0 in world size 1 is assigned as DP rank 0, PP rank 0, TP rank 0, EP rank 0 2025-11-03T17:48:15.3485641Z (EngineCore_DP0 pid=22380) INFO 11-03 17:48:15 [topk_topp_sampler.py:59] Using FlashInfer for top-p & top-k sampling. 2025-11-03T17:48:15.3505018Z (EngineCore_DP0 pid=22380) WARNING 11-03 17:48:15 [__init__.py:194] min_p, logit_bias, and min_tokens parameters won't currently work with speculative decoding enabled. 2025-11-03T17:48:15.3640691Z (EngineCore_DP0 pid=22380) INFO 11-03 17:48:15 [gpu_model_runner.py:2840] Starting to load model JackFram/llama-68m... 2025-11-03T17:48:15.6161372Z (EngineCore_DP0 pid=22380) INFO 11-03 17:48:15 [gpu_model_runner.py:2870] Loading model from scratch... 2025-11-03T17:48:15.6813802Z (EngineCore_DP0 pid=22380) INFO 11-03 17:48:15 [cuda.py:404] Using Flash Attention backend on V1 engine. 2025-11-03T17:48:15.7404613Z (EngineCore_DP0 pid=22380) INFO 11-03 17:48:15 [gpu_model_runner.py:2879] Loading drafter model... 2025-11-03T17:48:15.7416463Z (EngineCore_DP0 pid=22380) WARNING 11-03 17:48:15 [vllm.py:821] `torch.compile` is turned on, but the model JackFram/llama-68m does not support it. Please open an issue on GitHub if you want it to be supported. 2025-11-03T17:48:16.2445205Z (EngineCore_DP0 pid=22380) INFO 11-03 17:48:16 [gpu_model_runner.py:2902] Model loading took 0.1563 GiB and 0.126042 seconds 2025-11-03T17:48:16.2449770Z (EngineCore_DP0 pid=22380) INFO 11-03 17:48:16 [kv_cache_utils.py:1199] GPU KV cache size: 3,495,248 tokens 2025-11-03T17:48:16.2450865Z (EngineCore_DP0 pid=22380) INFO 11-03 17:48:16 [kv_cache_utils.py:1204] Maximum concurrency for 2,048 tokens per request: 1706.66x 2025-11-03T17:48:17.2557763Z (EngineCore_DP0 pid=22380) INFO 11-03 17:48:17 [gc_utils.py:40] GC Debug Config. enabled:False,top_objects:-1 2025-11-03T17:48:17.2571526Z INFO 11-03 17:48:17 [llm.py:337] Supported tasks: ['generate'] 2025-11-03T17:48:17.7055573Z PASSED 2025-11-03T17:48:17.7195967Z models/test_initialization.py::test_can_initialize_large_subset[GraniteForCausalLM] Fork a new process to run a test 22446 2025-11-03T17:48:17.7207577Z Fork a new process to run a test 0 2025-11-03T17:48:17.7485325Z INFO 11-03 17:48:17 [utils.py:239] non-default args: {'load_format': 'dummy', 'gpu_memory_utilization': 0.8, 'disable_log_stats': True, 'hf_overrides': functools.partial(, model_arch='GraniteForCausalLM', exist_overrides={}, use_original_num_layers=False), 'model_impl': 'vllm', 'model': 'ibm/PowerLM-3b'} 2025-11-03T17:48:17.8806846Z 2025-11-03T17:48:17.8807455Z config.json: 0% 0.00/755 [00:00, 'use_cudagraph': True, 'cudagraph_num_of_warmups': 1, 'cudagraph_capture_sizes': [512, 504, 496, 488, 480, 472, 464, 456, 448, 440, 432, 424, 416, 408, 400, 392, 384, 376, 368, 360, 352, 344, 336, 328, 320, 312, 304, 296, 288, 280, 272, 264, 256, 248, 240, 232, 224, 216, 208, 200, 192, 184, 176, 168, 160, 152, 144, 136, 128, 120, 112, 104, 96, 88, 80, 72, 64, 56, 48, 40, 32, 24, 16, 8, 4, 2, 1], 'cudagraph_copy_inputs': False, 'full_cuda_graph': True, 'use_inductor_graph_partition': False, 'pass_config': {}, 'max_capture_size': 512, 'local_cache_dir': None} 2025-11-03T17:48:28.6521828Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:48:28.6550339Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:48:28.6557886Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:48:28.6565681Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:48:28.6574104Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:48:28.6582148Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:48:28.6584967Z (EngineCore_DP0 pid=22507) INFO 11-03 17:48:28 [parallel_state.py:1231] rank 0 in world size 1 is assigned as DP rank 0, PP rank 0, TP rank 0, EP rank 0 2025-11-03T17:48:28.9991917Z (EngineCore_DP0 pid=22507) INFO 11-03 17:48:28 [topk_topp_sampler.py:59] Using FlashInfer for top-p & top-k sampling. 2025-11-03T17:48:29.0633532Z (EngineCore_DP0 pid=22507) INFO 11-03 17:48:29 [gpu_model_runner.py:2840] Starting to load model ibm/PowerLM-3b... 2025-11-03T17:48:29.3215629Z (EngineCore_DP0 pid=22507) INFO 11-03 17:48:29 [gpu_model_runner.py:2870] Loading model from scratch... 2025-11-03T17:48:29.3866181Z (EngineCore_DP0 pid=22507) INFO 11-03 17:48:29 [cuda.py:404] Using Flash Attention backend on V1 engine. 2025-11-03T17:48:29.9498497Z (EngineCore_DP0 pid=22507) INFO 11-03 17:48:29 [gpu_model_runner.py:2902] Model loading took 0.3707 GiB and 0.124002 seconds 2025-11-03T17:48:29.9503308Z (EngineCore_DP0 pid=22507) INFO 11-03 17:48:29 [kv_cache_utils.py:1199] GPU KV cache size: 1,165,072 tokens 2025-11-03T17:48:29.9504071Z (EngineCore_DP0 pid=22507) INFO 11-03 17:48:29 [kv_cache_utils.py:1204] Maximum concurrency for 4,096 tokens per request: 284.44x 2025-11-03T17:48:30.5145399Z (EngineCore_DP0 pid=22507) INFO 11-03 17:48:30 [gc_utils.py:40] GC Debug Config. enabled:False,top_objects:-1 2025-11-03T17:48:30.5167066Z INFO 11-03 17:48:30 [llm.py:337] Supported tasks: ['generate'] 2025-11-03T17:48:30.9677117Z PASSED 2025-11-03T17:48:30.9817096Z models/test_initialization.py::test_can_initialize_large_subset[Olmo3ForCausalLM] Fork a new process to run a test 22573 2025-11-03T17:48:30.9829101Z Fork a new process to run a test 0 2025-11-03T17:48:31.0108238Z INFO 11-03 17:48:31 [utils.py:239] non-default args: {'load_format': 'dummy', 'gpu_memory_utilization': 0.8, 'disable_log_stats': True, 'hf_overrides': functools.partial(, model_arch='Olmo3ForCausalLM', exist_overrides={}, use_original_num_layers=False), 'model_impl': 'vllm', 'model': 'shanearora/2025-sep-a-base-model'} 2025-11-03T17:48:31.1085186Z 2025-11-03T17:48:31.1087230Z config.json: 0.00B [00:00, ?B/s] 2025-11-03T17:48:31.1087518Z config.json: 1.46kB [00:00, 8.04MB/s] 2025-11-03T17:48:31.2110616Z INFO 11-03 17:48:31 [model.py:653] Resolved architecture: Olmo2ForCausalLM 2025-11-03T17:48:31.2111086Z `torch_dtype` is deprecated! Use `dtype` instead! 2025-11-03T17:48:31.2354481Z INFO 11-03 17:48:31 [model.py:1714] Using max model len 8192 2025-11-03T17:48:31.4573156Z INFO 11-03 17:48:31 [scheduler.py:225] Chunked prefill is enabled with max_num_batched_tokens=8192. 2025-11-03T17:48:31.4908597Z 2025-11-03T17:48:31.4910621Z tokenizer_config.json: 0.00B [00:00, ?B/s] 2025-11-03T17:48:31.4911001Z tokenizer_config.json: 4.34kB [00:00, 11.4MB/s] 2025-11-03T17:48:34.7307005Z 2025-11-03T17:48:34.7410676Z vocab.json: 0.00B [00:00, ?B/s] 2025-11-03T17:48:34.7410971Z vocab.json: 1.61MB [00:00, 154MB/s] 2025-11-03T17:48:34.7795721Z 2025-11-03T17:48:34.7846341Z merges.txt: 0.00B [00:00, ?B/s] 2025-11-03T17:48:34.7846614Z merges.txt: 917kB [00:00, 182MB/s] 2025-11-03T17:48:34.8195625Z 2025-11-03T17:48:34.8412307Z tokenizer.json: 0.00B [00:00, ?B/s] 2025-11-03T17:48:34.8412610Z tokenizer.json: 7.14MB [00:00, 330MB/s] 2025-11-03T17:48:34.9066952Z 2025-11-03T17:48:34.9067494Z special_tokens_map.json: 0% 0.00/125 [00:00, 'use_cudagraph': True, 'cudagraph_num_of_warmups': 1, 'cudagraph_capture_sizes': [512, 504, 496, 488, 480, 472, 464, 456, 448, 440, 432, 424, 416, 408, 400, 392, 384, 376, 368, 360, 352, 344, 336, 328, 320, 312, 304, 296, 288, 280, 272, 264, 256, 248, 240, 232, 224, 216, 208, 200, 192, 184, 176, 168, 160, 152, 144, 136, 128, 120, 112, 104, 96, 88, 80, 72, 64, 56, 48, 40, 32, 24, 16, 8, 4, 2, 1], 'cudagraph_copy_inputs': False, 'full_cuda_graph': True, 'use_inductor_graph_partition': False, 'pass_config': {}, 'max_capture_size': 512, 'local_cache_dir': None} 2025-11-03T17:48:35.9577517Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:48:35.9605535Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:48:35.9613396Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:48:35.9622099Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:48:35.9630319Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:48:35.9638221Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:48:35.9640995Z (EngineCore_DP0 pid=22581) INFO 11-03 17:48:35 [parallel_state.py:1231] rank 0 in world size 1 is assigned as DP rank 0, PP rank 0, TP rank 0, EP rank 0 2025-11-03T17:48:36.3035422Z (EngineCore_DP0 pid=22581) INFO 11-03 17:48:36 [topk_topp_sampler.py:59] Using FlashInfer for top-p & top-k sampling. 2025-11-03T17:48:36.3710276Z (EngineCore_DP0 pid=22581) INFO 11-03 17:48:36 [gpu_model_runner.py:2840] Starting to load model shanearora/2025-sep-a-base-model... 2025-11-03T17:48:36.6278557Z (EngineCore_DP0 pid=22581) INFO 11-03 17:48:36 [gpu_model_runner.py:2870] Loading model from scratch... 2025-11-03T17:48:36.6301413Z (EngineCore_DP0 pid=22581) INFO 11-03 17:48:36 [cuda.py:404] Using Flash Attention backend on V1 engine. 2025-11-03T17:48:37.2663801Z (EngineCore_DP0 pid=22581) INFO 11-03 17:48:37 [gpu_model_runner.py:2902] Model loading took 1.9102 GiB and 0.126653 seconds 2025-11-03T17:48:37.2668981Z (EngineCore_DP0 pid=22581) INFO 11-03 17:48:37 [kv_cache_utils.py:1199] GPU KV cache size: 655,360 tokens 2025-11-03T17:48:37.2669922Z (EngineCore_DP0 pid=22581) INFO 11-03 17:48:37 [kv_cache_utils.py:1204] Maximum concurrency for 8,192 tokens per request: 79.84x 2025-11-03T17:48:37.8497722Z (EngineCore_DP0 pid=22581) INFO 11-03 17:48:37 [gc_utils.py:40] GC Debug Config. enabled:False,top_objects:-1 2025-11-03T17:48:37.8519559Z INFO 11-03 17:48:37 [llm.py:337] Supported tasks: ['generate'] 2025-11-03T17:48:38.3235773Z PASSED 2025-11-03T17:48:38.3376602Z models/test_initialization.py::test_can_initialize_large_subset[StableLMEpochForCausalLM] Fork a new process to run a test 22647 2025-11-03T17:48:38.3387202Z Fork a new process to run a test 0 2025-11-03T17:48:38.3666755Z INFO 11-03 17:48:38 [utils.py:239] non-default args: {'load_format': 'dummy', 'gpu_memory_utilization': 0.8, 'disable_log_stats': True, 'hf_overrides': functools.partial(, model_arch='StableLMEpochForCausalLM', exist_overrides={}, use_original_num_layers=False), 'model_impl': 'vllm', 'model': 'stabilityai/stablelm-zephyr-3b'} 2025-11-03T17:48:38.4521306Z 2025-11-03T17:48:38.4522397Z config.json: 0% 0.00/599 [00:00, 'use_cudagraph': True, 'cudagraph_num_of_warmups': 1, 'cudagraph_capture_sizes': [512, 504, 496, 488, 480, 472, 464, 456, 448, 440, 432, 424, 416, 408, 400, 392, 384, 376, 368, 360, 352, 344, 336, 328, 320, 312, 304, 296, 288, 280, 272, 264, 256, 248, 240, 232, 224, 216, 208, 200, 192, 184, 176, 168, 160, 152, 144, 136, 128, 120, 112, 104, 96, 88, 80, 72, 64, 56, 48, 40, 32, 24, 16, 8, 4, 2, 1], 'cudagraph_copy_inputs': False, 'full_cuda_graph': True, 'use_inductor_graph_partition': False, 'pass_config': {}, 'max_capture_size': 512, 'local_cache_dir': None} 2025-11-03T17:48:39.8119763Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:48:39.8148149Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:48:39.8155889Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:48:39.8163781Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:48:39.8171580Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:48:39.8180118Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:48:39.8182913Z (EngineCore_DP0 pid=22655) INFO 11-03 17:48:39 [parallel_state.py:1231] rank 0 in world size 1 is assigned as DP rank 0, PP rank 0, TP rank 0, EP rank 0 2025-11-03T17:48:40.1590472Z (EngineCore_DP0 pid=22655) INFO 11-03 17:48:40 [topk_topp_sampler.py:59] Using FlashInfer for top-p & top-k sampling. 2025-11-03T17:48:40.2238873Z (EngineCore_DP0 pid=22655) INFO 11-03 17:48:40 [gpu_model_runner.py:2840] Starting to load model stabilityai/stablelm-zephyr-3b... 2025-11-03T17:48:40.4830288Z (EngineCore_DP0 pid=22655) INFO 11-03 17:48:40 [gpu_model_runner.py:2870] Loading model from scratch... 2025-11-03T17:48:40.5484903Z (EngineCore_DP0 pid=22655) INFO 11-03 17:48:40 [cuda.py:420] Using FlexAttention backend for head_size=80 on V1 engine. 2025-11-03T17:48:40.6515503Z (EngineCore_DP0 pid=22655) WARNING 11-03 17:48:40 [vllm.py:821] `torch.compile` is turned on, but the model stabilityai/stablelm-zephyr-3b does not support it. Please open an issue on GitHub if you want it to be supported. 2025-11-03T17:48:41.1637543Z (EngineCore_DP0 pid=22655) INFO 11-03 17:48:41 [gpu_model_runner.py:2902] Model loading took 0.6296 GiB and 0.171086 seconds 2025-11-03T17:48:41.1642934Z (EngineCore_DP0 pid=22655) INFO 11-03 17:48:41 [kv_cache_utils.py:1199] GPU KV cache size: 1,048,576 tokens 2025-11-03T17:48:41.1643714Z (EngineCore_DP0 pid=22655) INFO 11-03 17:48:41 [kv_cache_utils.py:1204] Maximum concurrency for 4,096 tokens per request: 256.00x 2025-11-03T17:48:41.6265634Z (EngineCore_DP0 pid=22655) INFO 11-03 17:48:41 [gc_utils.py:40] GC Debug Config. enabled:False,top_objects:-1 2025-11-03T17:48:41.6285814Z INFO 11-03 17:48:41 [llm.py:337] Supported tasks: ['generate'] 2025-11-03T17:48:42.0750042Z PASSED 2025-11-03T17:48:42.0890636Z models/test_initialization.py::test_can_initialize_large_subset[HunYuanMoEV1ForCausalLM] Fork a new process to run a test 22721 2025-11-03T17:48:42.0901321Z Fork a new process to run a test 0 2025-11-03T17:48:42.1179067Z INFO 11-03 17:48:42 [utils.py:239] non-default args: {'trust_remote_code': True, 'load_format': 'dummy', 'gpu_memory_utilization': 0.8, 'disable_log_stats': True, 'hf_overrides': functools.partial(, model_arch='HunYuanMoEV1ForCausalLM', exist_overrides={}, use_original_num_layers=False), 'model_impl': 'vllm', 'model': 'tencent/Hunyuan-A13B-Instruct'} 2025-11-03T17:48:42.2685027Z 2025-11-03T17:48:42.2687170Z config.json: 0.00B [00:00, ?B/s] 2025-11-03T17:48:42.2687493Z config.json: 3.38kB [00:00, 18.4MB/s] 2025-11-03T17:48:42.3387645Z 2025-11-03T17:48:42.3389715Z configuration_hunyuan.py: 0.00B [00:00, ?B/s] 2025-11-03T17:48:42.3390114Z configuration_hunyuan.py: 15.0kB [00:00, 85.1MB/s] 2025-11-03T17:48:42.3503880Z A new version of the following files was downloaded from https://huggingface.co/tencent/Hunyuan-A13B-Instruct: 2025-11-03T17:48:42.3504449Z - configuration_hunyuan.py 2025-11-03T17:48:42.3505099Z . Make sure to double-check they do not contain any added malicious code. To avoid downloading new versions of the code file, you can pin a revision. 2025-11-03T17:48:42.3881873Z You are using a model of type hunyuan_v1_moe to instantiate a model of type hunyuan. This is not supported for all configurations of models and can yield errors. 2025-11-03T17:48:42.3895383Z INFO 11-03 17:48:42 [config.py:408] Replacing legacy 'type' key with 'rope_type' 2025-11-03T17:48:48.8538202Z INFO 11-03 17:48:48 [model.py:653] Resolved architecture: HunYuanMoEV1ForCausalLM 2025-11-03T17:48:48.8538733Z `torch_dtype` is deprecated! Use `dtype` instead! 2025-11-03T17:48:48.8791704Z INFO 11-03 17:48:48 [model.py:1714] Using max model len 32768 2025-11-03T17:48:48.8793407Z The argument `trust_remote_code` is to be used with Auto classes. It has no effect here and is ignored. 2025-11-03T17:48:48.9363102Z INFO 11-03 17:48:48 [scheduler.py:225] Chunked prefill is enabled with max_num_batched_tokens=8192. 2025-11-03T17:48:48.9724598Z 2025-11-03T17:48:48.9727319Z tokenizer_config.json: 0.00B [00:00, ?B/s] 2025-11-03T17:48:48.9727978Z tokenizer_config.json: 1.02kB [00:00, 3.75MB/s] 2025-11-03T17:48:49.1651991Z 2025-11-03T17:48:49.4065802Z tokenizer.json: 0% 0.00/16.4M [00:00, 'use_cudagraph': True, 'cudagraph_num_of_warmups': 1, 'cudagraph_capture_sizes': [512, 504, 496, 488, 480, 472, 464, 456, 448, 440, 432, 424, 416, 408, 400, 392, 384, 376, 368, 360, 352, 344, 336, 328, 320, 312, 304, 296, 288, 280, 272, 264, 256, 248, 240, 232, 224, 216, 208, 200, 192, 184, 176, 168, 160, 152, 144, 136, 128, 120, 112, 104, 96, 88, 80, 72, 64, 56, 48, 40, 32, 24, 16, 8, 4, 2, 1], 'cudagraph_copy_inputs': False, 'full_cuda_graph': True, 'use_inductor_graph_partition': False, 'pass_config': {}, 'max_capture_size': 512, 'local_cache_dir': None} 2025-11-03T17:48:51.0498930Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:48:51.0526667Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:48:51.0533761Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:48:51.0542311Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:48:51.0550084Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:48:51.0557569Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:48:51.0560214Z (EngineCore_DP0 pid=22799) INFO 11-03 17:48:51 [parallel_state.py:1231] rank 0 in world size 1 is assigned as DP rank 0, PP rank 0, TP rank 0, EP rank 0 2025-11-03T17:48:51.3948037Z (EngineCore_DP0 pid=22799) INFO 11-03 17:48:51 [topk_topp_sampler.py:59] Using FlashInfer for top-p & top-k sampling. 2025-11-03T17:48:51.4900374Z (EngineCore_DP0 pid=22799) INFO 11-03 17:48:51 [gpu_model_runner.py:2840] Starting to load model tencent/Hunyuan-A13B-Instruct... 2025-11-03T17:48:51.7502146Z (EngineCore_DP0 pid=22799) INFO 11-03 17:48:51 [gpu_model_runner.py:2870] Loading model from scratch... 2025-11-03T17:48:51.8273305Z (EngineCore_DP0 pid=22799) INFO 11-03 17:48:51 [cuda.py:404] Using Flash Attention backend on V1 engine. 2025-11-03T17:48:52.4016154Z (EngineCore_DP0 pid=22799) INFO 11-03 17:48:52 [gpu_model_runner.py:2902] Model loading took 1.2754 GiB and 0.137872 seconds 2025-11-03T17:48:52.4020598Z (EngineCore_DP0 pid=22799) INFO 11-03 17:48:52 [kv_cache_utils.py:1199] GPU KV cache size: 2,621,440 tokens 2025-11-03T17:48:52.4022237Z (EngineCore_DP0 pid=22799) INFO 11-03 17:48:52 [kv_cache_utils.py:1204] Maximum concurrency for 32,768 tokens per request: 80.00x 2025-11-03T17:48:53.6390301Z (EngineCore_DP0 pid=22799) INFO 11-03 17:48:53 [gc_utils.py:40] GC Debug Config. enabled:False,top_objects:-1 2025-11-03T17:48:53.6404300Z INFO 11-03 17:48:53 [llm.py:337] Supported tasks: ['generate'] 2025-11-03T17:48:54.1190198Z PASSED 2025-11-03T17:48:54.1330808Z models/test_initialization.py::test_can_initialize_large_subset[AyaVisionForConditionalGeneration] Fork a new process to run a test 22865 2025-11-03T17:48:54.1340939Z Fork a new process to run a test 0 2025-11-03T17:48:54.1617082Z INFO 11-03 17:48:54 [utils.py:239] non-default args: {'load_format': 'dummy', 'gpu_memory_utilization': 0.8, 'disable_log_stats': True, 'hf_overrides': functools.partial(, model_arch='AyaVisionForConditionalGeneration', exist_overrides={}, use_original_num_layers=False), 'model_impl': 'vllm', 'model': 'CohereForAI/aya-vision-8b'} 2025-11-03T17:48:54.3583485Z 2025-11-03T17:48:54.3585236Z config.json: 0% 0.00/1.17k [00:00, 'use_cudagraph': True, 'cudagraph_num_of_warmups': 1, 'cudagraph_capture_sizes': [512, 504, 496, 488, 480, 472, 464, 456, 448, 440, 432, 424, 416, 408, 400, 392, 384, 376, 368, 360, 352, 344, 336, 328, 320, 312, 304, 296, 288, 280, 272, 264, 256, 248, 240, 232, 224, 216, 208, 200, 192, 184, 176, 168, 160, 152, 144, 136, 128, 120, 112, 104, 96, 88, 80, 72, 64, 56, 48, 40, 32, 24, 16, 8, 4, 2, 1], 'cudagraph_copy_inputs': False, 'full_cuda_graph': True, 'use_inductor_graph_partition': False, 'pass_config': {}, 'max_capture_size': 512, 'local_cache_dir': None} 2025-11-03T17:49:04.4254643Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:49:04.4282228Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:49:04.4289520Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:49:04.4297494Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:49:04.4305104Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:49:04.4312779Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:49:04.4315282Z (EngineCore_DP0 pid=22943) INFO 11-03 17:49:04 [parallel_state.py:1231] rank 0 in world size 1 is assigned as DP rank 0, PP rank 0, TP rank 0, EP rank 0 2025-11-03T17:49:04.7633638Z (EngineCore_DP0 pid=22943) INFO 11-03 17:49:04 [topk_topp_sampler.py:59] Using FlashInfer for top-p & top-k sampling. 2025-11-03T17:49:05.0183123Z (EngineCore_DP0 pid=22943) 2025-11-03T17:49:05.0183574Z processor_config.json: 0% 0.00/389 [00:00, model_arch='LongCatFlashMTPModel', exist_overrides={}, use_original_num_layers=False), 'speculative_config': {'model': 'meituan-longcat/LongCat-Flash-Chat', 'num_speculative_tokens': 1}, 'model_impl': 'vllm', 'model': 'meituan-longcat/LongCat-Flash-Chat'} 2025-11-03T17:49:12.6798241Z 2025-11-03T17:49:12.6800410Z config.json: 0.00B [00:00, ?B/s] 2025-11-03T17:49:12.6800887Z config.json: 1.00kB [00:00, 4.97MB/s] 2025-11-03T17:49:12.7679445Z 2025-11-03T17:49:12.7681484Z configuration_longcat_flash.py: 0.00B [00:00, ?B/s] 2025-11-03T17:49:12.7681982Z configuration_longcat_flash.py: 9.69kB [00:00, 47.9MB/s] 2025-11-03T17:49:12.7778088Z A new version of the following files was downloaded from https://huggingface.co/meituan-longcat/LongCat-Flash-Chat: 2025-11-03T17:49:12.7778678Z - configuration_longcat_flash.py 2025-11-03T17:49:12.7779336Z . Make sure to double-check they do not contain any added malicious code. To avoid downloading new versions of the code file, you can pin a revision. 2025-11-03T17:49:19.3837841Z INFO 11-03 17:49:19 [model.py:653] Resolved architecture: LongcatFlashForCausalLM 2025-11-03T17:49:19.3838364Z `torch_dtype` is deprecated! Use `dtype` instead! 2025-11-03T17:49:19.4677692Z 2025-11-03T17:49:19.4798558Z model.safetensors.index.json: 0.00B [00:00, ?B/s] 2025-11-03T17:49:19.4799092Z model.safetensors.index.json: 4.14MB [00:00, 345MB/s] 2025-11-03T17:49:19.5239430Z 2025-11-03T17:49:19.6613210Z Parse safetensors files: 0% 0/75 [00:00, 'use_cudagraph': True, 'cudagraph_num_of_warmups': 1, 'cudagraph_capture_sizes': [512, 504, 496, 488, 480, 472, 464, 456, 448, 440, 432, 424, 416, 408, 400, 392, 384, 376, 368, 360, 352, 344, 336, 328, 320, 312, 304, 296, 288, 280, 272, 264, 256, 248, 240, 232, 224, 216, 208, 200, 192, 184, 176, 168, 160, 152, 144, 136, 128, 120, 112, 104, 96, 88, 80, 72, 64, 56, 48, 40, 32, 24, 16, 8, 4, 2, 1], 'cudagraph_copy_inputs': False, 'full_cuda_graph': True, 'use_inductor_graph_partition': False, 'pass_config': {}, 'max_capture_size': 512, 'local_cache_dir': None} 2025-11-03T17:49:28.8419309Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:49:28.8449217Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:49:28.8458619Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:49:28.8467811Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:49:28.8476869Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:49:28.8485963Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:49:28.8488729Z (EngineCore_DP0 pid=23173) INFO 11-03 17:49:28 [parallel_state.py:1231] rank 0 in world size 1 is assigned as DP rank 0, PP rank 0, TP rank 0, EP rank 0 2025-11-03T17:49:29.1796861Z (EngineCore_DP0 pid=23173) INFO 11-03 17:49:29 [topk_topp_sampler.py:59] Using FlashInfer for top-p & top-k sampling. 2025-11-03T17:49:29.1952993Z (EngineCore_DP0 pid=23173) WARNING 11-03 17:49:29 [__init__.py:194] min_p, logit_bias, and min_tokens parameters won't currently work with speculative decoding enabled. 2025-11-03T17:49:29.2844268Z (EngineCore_DP0 pid=23173) INFO 11-03 17:49:29 [gpu_model_runner.py:2840] Starting to load model meituan-longcat/LongCat-Flash-Chat... 2025-11-03T17:49:29.5425994Z (EngineCore_DP0 pid=23173) INFO 11-03 17:49:29 [gpu_model_runner.py:2870] Loading model from scratch... 2025-11-03T17:49:29.5983136Z (EngineCore_DP0 pid=23173) INFO 11-03 17:49:29 [cuda.py:328] Using Triton MLA backend on V1 engine. 2025-11-03T17:49:29.6794922Z (EngineCore_DP0 pid=23173) INFO 11-03 17:49:29 [gpu_model_runner.py:2879] Loading drafter model... 2025-11-03T17:49:29.7006613Z (EngineCore_DP0 pid=23173) WARNING 11-03 17:49:29 [vllm.py:821] `torch.compile` is turned on, but the model meituan-longcat/LongCat-Flash-Chat does not support it. Please open an issue on GitHub if you want it to be supported. 2025-11-03T17:49:29.7014998Z (EngineCore_DP0 pid=23173) INFO 11-03 17:49:29 [eagle.py:1005] Assuming the EAGLE head shares the same vocab embedding with the target model. 2025-11-03T17:49:29.7016294Z (EngineCore_DP0 pid=23173) INFO 11-03 17:49:29 [eagle.py:1027] Loading EAGLE LM head weights from the target model. 2025-11-03T17:49:30.2225332Z (EngineCore_DP0 pid=23173) INFO 11-03 17:49:30 [gpu_model_runner.py:2902] Model loading took 5.0746 GiB and 0.159286 seconds 2025-11-03T17:49:30.2229437Z (EngineCore_DP0 pid=23173) INFO 11-03 17:49:30 [kv_cache_utils.py:1199] GPU KV cache size: 3,106,880 tokens 2025-11-03T17:49:30.2230417Z (EngineCore_DP0 pid=23173) INFO 11-03 17:49:30 [kv_cache_utils.py:1204] Maximum concurrency for 131,072 tokens per request: 23.70x 2025-11-03T17:49:31.2943087Z (EngineCore_DP0 pid=23173) INFO 11-03 17:49:31 [gc_utils.py:40] GC Debug Config. enabled:False,top_objects:-1 2025-11-03T17:49:31.2956549Z INFO 11-03 17:49:31 [llm.py:337] Supported tasks: ['generate'] 2025-11-03T17:49:31.8038577Z PASSED 2025-11-03T17:49:31.8178722Z models/test_initialization.py::test_can_initialize_large_subset[Qwen3NextForCausalLM] Fork a new process to run a test 23239 2025-11-03T17:49:31.8190173Z Fork a new process to run a test 0 2025-11-03T17:49:31.8193991Z `transformers==4.56.2` installed, but `transformers>=4.56.3` is required to run this model. 2025-11-03T17:49:32.1185331Z PASSED 2025-11-03T17:49:32.1324505Z models/test_initialization.py::test_can_initialize_large_subset[Llama_Nemotron_Nano_VL] Fork a new process to run a test 23240 2025-11-03T17:49:32.1334981Z Fork a new process to run a test 0 2025-11-03T17:49:32.1616985Z INFO 11-03 17:49:32 [utils.py:239] non-default args: {'trust_remote_code': True, 'load_format': 'dummy', 'gpu_memory_utilization': 0.8, 'disable_log_stats': True, 'hf_overrides': functools.partial(, model_arch='Llama_Nemotron_Nano_VL', exist_overrides={}, use_original_num_layers=False), 'model_impl': 'vllm', 'model': 'nvidia/Llama-3.1-Nemotron-Nano-VL-8B-V1'} 2025-11-03T17:49:32.2625729Z 2025-11-03T17:49:32.5040903Z config.json: 0% 0.00/7.65k [00:00, 'use_cudagraph': True, 'cudagraph_num_of_warmups': 1, 'cudagraph_capture_sizes': [512, 504, 496, 488, 480, 472, 464, 456, 448, 440, 432, 424, 416, 408, 400, 392, 384, 376, 368, 360, 352, 344, 336, 328, 320, 312, 304, 296, 288, 280, 272, 264, 256, 248, 240, 232, 224, 216, 208, 200, 192, 184, 176, 168, 160, 152, 144, 136, 128, 120, 112, 104, 96, 88, 80, 72, 64, 56, 48, 40, 32, 24, 16, 8, 4, 2, 1], 'cudagraph_copy_inputs': False, 'full_cuda_graph': True, 'use_inductor_graph_partition': False, 'pass_config': {}, 'max_capture_size': 512, 'local_cache_dir': None} 2025-11-03T17:49:50.6257680Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:49:50.6285631Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:49:50.6292643Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:49:50.6299665Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:49:50.6307025Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:49:50.6313998Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:49:50.6316599Z (EngineCore_DP0 pid=23323) INFO 11-03 17:49:50 [parallel_state.py:1231] rank 0 in world size 1 is assigned as DP rank 0, PP rank 0, TP rank 0, EP rank 0 2025-11-03T17:49:50.8697071Z (EngineCore_DP0 pid=23323) INFO 11-03 17:49:50 [topk_topp_sampler.py:59] Using FlashInfer for top-p & top-k sampling. 2025-11-03T17:49:51.0021850Z (EngineCore_DP0 pid=23323) Using a slow image processor as `use_fast` is unset and a slow processor was saved with this model. `use_fast=True` will be the default behavior in v4.52, even if the model was saved with a slow processor. This will result in minor differences in outputs. You'll still be able to use a slow processor with `use_fast=False`. 2025-11-03T17:49:51.0364107Z (EngineCore_DP0 pid=23323) 2025-11-03T17:49:51.0365418Z image_processing.py: 0.00B [00:00, ?B/s] 2025-11-03T17:49:51.0365748Z image_processing.py: 4.78kB [00:00, 29.7MB/s] 2025-11-03T17:49:51.0411681Z (EngineCore_DP0 pid=23323) A new version of the following files was downloaded from https://huggingface.co/nvidia/Llama-3.1-Nemotron-Nano-VL-8B-V1: 2025-11-03T17:49:51.0412493Z (EngineCore_DP0 pid=23323) - image_processing.py 2025-11-03T17:49:51.0413449Z (EngineCore_DP0 pid=23323) . Make sure to double-check they do not contain any added malicious code. To avoid downloading new versions of the code file, you can pin a revision. 2025-11-03T17:49:51.0507822Z (EngineCore_DP0 pid=23323) WARNING 11-03 17:49:51 [__init__.py:2040] The following intended overrides are not keyword args and will be dropped: {'truncation'} 2025-11-03T17:49:51.0512172Z (EngineCore_DP0 pid=23323) WARNING 11-03 17:49:51 [__init__.py:2040] The following intended overrides are not keyword args and will be dropped: {'truncation'} 2025-11-03T17:49:51.0996700Z (EngineCore_DP0 pid=23323) INFO 11-03 17:49:51 [gpu_model_runner.py:2840] Starting to load model nvidia/Llama-3.1-Nemotron-Nano-VL-8B-V1... 2025-11-03T17:49:51.3321408Z (EngineCore_DP0 pid=23323) INFO 11-03 17:49:51 [gpu_model_runner.py:2870] Loading model from scratch... 2025-11-03T17:49:51.4857646Z (EngineCore_DP0 pid=23323) /opt/conda/envs/py_3.12/lib/python3.12/site-packages/timm/models/registry.py:4: FutureWarning: Importing from timm.models.registry is deprecated, please import via timm.models 2025-11-03T17:49:51.4859034Z (EngineCore_DP0 pid=23323) warnings.warn(f"Importing from {__name__} is deprecated, please import via timm.models", FutureWarning) 2025-11-03T17:49:51.7650490Z (EngineCore_DP0 pid=23323) /opt/conda/envs/py_3.12/lib/python3.12/site-packages/torch/library.py:356: UserWarning: Warning only once for all operators, other operators may also be overridden. 2025-11-03T17:49:51.7651768Z (EngineCore_DP0 pid=23323) Overriding a previously registered kernel for the same operator and the same dispatch key 2025-11-03T17:49:51.7653768Z (EngineCore_DP0 pid=23323) operator: xformers_python::_fused_allgather_and_linear_impl(Tensor scattered_input, Tensor[] weights, str process_group_name, Tensor(a3!)[] gathered_outputs, SymInt timeout_s, bool _wait, bool _memcpy, Tensor scale_scattered_input, Tensor?[] scales_weights) -> () 2025-11-03T17:49:51.7655293Z (EngineCore_DP0 pid=23323) registered at /opt/conda/envs/py_3.12/lib/python3.12/site-packages/torch/_library/custom_ops.py:922 2025-11-03T17:49:51.7656082Z (EngineCore_DP0 pid=23323) dispatch key: ADInplaceOrView 2025-11-03T17:49:51.7656635Z (EngineCore_DP0 pid=23323) previous kernel: no debug info 2025-11-03T17:49:51.7657917Z (EngineCore_DP0 pid=23323) new kernel: registered at /opt/conda/envs/py_3.12/lib/python3.12/site-packages/torch/_library/custom_ops.py:922 (Triggered internally at /var/lib/jenkins/workspace/aten/src/ATen/core/dispatch/OperatorEntry.cpp:208.) 2025-11-03T17:49:51.7659032Z (EngineCore_DP0 pid=23323) self.m.impl( 2025-11-03T17:49:51.7837398Z (EngineCore_DP0 pid=23323) /opt/conda/envs/py_3.12/lib/python3.12/site-packages/timm/models/layers/__init__.py:48: FutureWarning: Importing from timm.models.layers is deprecated, please import via timm.layers 2025-11-03T17:49:51.7838726Z (EngineCore_DP0 pid=23323) warnings.warn(f"Importing from {__name__} is deprecated, please import via timm.layers", FutureWarning) 2025-11-03T17:49:51.7877174Z (EngineCore_DP0 pid=23323) No pretrained configuration specified for vit_huge_patch16_224 model. Using a default. Please add a config to the model pretrained_cfg registry or pass explicitly. 2025-11-03T17:49:51.9224111Z (EngineCore_DP0 pid=23323) INFO 11-03 17:49:51 [cuda.py:404] Using Flash Attention backend on V1 engine. 2025-11-03T17:49:52.4955958Z (EngineCore_DP0 pid=23323) INFO 11-03 17:49:52 [gpu_model_runner.py:2902] Model loading took 3.6833 GiB and 0.658304 seconds 2025-11-03T17:49:52.7454462Z (EngineCore_DP0 pid=23323) INFO 11-03 17:49:52 [gpu_model_runner.py:3647] Encoder cache will be initialized with a budget of 8192 tokens, and profiled with 2 image items of the maximum feature size. 2025-11-03T17:49:52.7660415Z (EngineCore_DP0 pid=23323) WARNING 11-03 17:49:52 [__init__.py:2040] The following intended overrides are not keyword args and will be dropped: {'truncation'} 2025-11-03T17:49:52.7664181Z (EngineCore_DP0 pid=23323) WARNING 11-03 17:49:52 [__init__.py:2040] The following intended overrides are not keyword args and will be dropped: {'truncation'} 2025-11-03T17:49:55.0313455Z (EngineCore_DP0 pid=23323) INFO 11-03 17:49:55 [backends.py:594] Using cache directory: /var/lib/jenkins/.cache/vllm/torch_compile_cache/b948255062/rank_0_0/backbone for vLLM's torch.compile 2025-11-03T17:49:55.0314600Z (EngineCore_DP0 pid=23323) INFO 11-03 17:49:55 [backends.py:608] Dynamo bytecode transform time: 1.27 s 2025-11-03T17:49:55.6843612Z (EngineCore_DP0 pid=23323) [rank0]:W1103 17:49:55.683000 23323 site-packages/torch/_inductor/utils.py:1665] [0/0] Not enough SMs to use max_autotune_gemm mode 2025-11-03T17:49:56.9817717Z (EngineCore_DP0 pid=23323) INFO 11-03 17:49:56 [backends.py:214] Cache the graph for dynamic shape for later use 2025-11-03T17:49:57.8775950Z (EngineCore_DP0 pid=23323) INFO 11-03 17:49:57 [backends.py:241] Compiling a graph for dynamic shape takes 2.64 s 2025-11-03T17:49:58.4390941Z (EngineCore_DP0 pid=23323) INFO 11-03 17:49:58 [monitor.py:33] torch.compile takes 3.91 s in total 2025-11-03T17:49:59.4635234Z (EngineCore_DP0 pid=23323) INFO 11-03 17:49:59 [gpu_worker.py:314] Available KV cache memory: 12.86 GiB 2025-11-03T17:49:59.7569546Z (EngineCore_DP0 pid=23323) INFO 11-03 17:49:59 [kv_cache_utils.py:1199] GPU KV cache size: 3,370,240 tokens 2025-11-03T17:49:59.7571030Z (EngineCore_DP0 pid=23323) INFO 11-03 17:49:59 [kv_cache_utils.py:1204] Maximum concurrency for 131,072 tokens per request: 25.71x 2025-11-03T17:50:00.0522864Z (EngineCore_DP0 pid=23323) 2025-11-03T17:50:00.1531690Z Capturing CUDA graphs (mixed prefill-decode, PIECEWISE): 0% 0/67 [00:00, model_arch='InternS1ForConditionalGeneration', exist_overrides={}, use_original_num_layers=False), 'model_impl': 'vllm', 'model': 'internlm/Intern-S1'} 2025-11-03T17:50:03.0998855Z 2025-11-03T17:50:03.1000910Z config.json: 0.00B [00:00, ?B/s] 2025-11-03T17:50:03.1001364Z config.json: 2.62kB [00:00, 14.7MB/s] 2025-11-03T17:50:03.1787737Z 2025-11-03T17:50:03.1790599Z configuration_interns1.py: 0.00B [00:00, ?B/s] 2025-11-03T17:50:03.1790999Z configuration_interns1.py: 10.2kB [00:00, 35.7MB/s] 2025-11-03T17:50:03.1884137Z A new version of the following files was downloaded from https://huggingface.co/internlm/Intern-S1: 2025-11-03T17:50:03.1884654Z - configuration_interns1.py 2025-11-03T17:50:03.1885301Z . Make sure to double-check they do not contain any added malicious code. To avoid downloading new versions of the code file, you can pin a revision. 2025-11-03T17:50:03.4472914Z 2025-11-03T17:50:03.4473396Z preprocessor_config.json: 0% 0.00/694 [00:00, 'use_cudagraph': True, 'cudagraph_num_of_warmups': 1, 'cudagraph_capture_sizes': [512, 504, 496, 488, 480, 472, 464, 456, 448, 440, 432, 424, 416, 408, 400, 392, 384, 376, 368, 360, 352, 344, 336, 328, 320, 312, 304, 296, 288, 280, 272, 264, 256, 248, 240, 232, 224, 216, 208, 200, 192, 184, 176, 168, 160, 152, 144, 136, 128, 120, 112, 104, 96, 88, 80, 72, 64, 56, 48, 40, 32, 24, 16, 8, 4, 2, 1], 'cudagraph_copy_inputs': False, 'full_cuda_graph': True, 'use_inductor_graph_partition': False, 'pass_config': {}, 'max_capture_size': 512, 'local_cache_dir': None} 2025-11-03T17:50:24.6802147Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:50:24.6831334Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:50:24.6838927Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:50:24.6846456Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:50:24.6854000Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:50:24.6861920Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:50:24.6864605Z (EngineCore_DP0 pid=23629) INFO 11-03 17:50:24 [parallel_state.py:1231] rank 0 in world size 1 is assigned as DP rank 0, PP rank 0, TP rank 0, EP rank 0 2025-11-03T17:50:25.0310550Z (EngineCore_DP0 pid=23629) INFO 11-03 17:50:25 [topk_topp_sampler.py:59] Using FlashInfer for top-p & top-k sampling. 2025-11-03T17:50:25.1807384Z (EngineCore_DP0 pid=23629) 2025-11-03T17:50:25.1808564Z processor_config.json: 0% 0.00/154 [00:00 32768). Running this sequence through the model will result in indexing errors 2025-11-03T17:50:34.4178288Z (EngineCore_DP0 pid=23629) INFO 11-03 17:50:34 [gpu_model_runner.py:2840] Starting to load model internlm/Intern-S1... 2025-11-03T17:50:34.6827617Z (EngineCore_DP0 pid=23629) INFO 11-03 17:50:34 [gpu_model_runner.py:2870] Loading model from scratch... 2025-11-03T17:50:34.6871141Z (EngineCore_DP0 pid=23629) INFO 11-03 17:50:34 [layer.py:497] MultiHeadAttention attn_backend: _Backend.FLASH_ATTN, use_upstream_fa: False 2025-11-03T17:50:34.7362446Z (EngineCore_DP0 pid=23629) INFO 11-03 17:50:34 [cuda.py:404] Using Flash Attention backend on V1 engine. 2025-11-03T17:50:35.3145092Z (EngineCore_DP0 pid=23629) INFO 11-03 17:50:35 [gpu_model_runner.py:2902] Model loading took 2.9280 GiB and 0.121223 seconds 2025-11-03T17:50:35.3149852Z (EngineCore_DP0 pid=23629) INFO 11-03 17:50:35 [kv_cache_utils.py:1199] GPU KV cache size: 5,242,880 tokens 2025-11-03T17:50:35.3150779Z (EngineCore_DP0 pid=23629) INFO 11-03 17:50:35 [kv_cache_utils.py:1204] Maximum concurrency for 65,536 tokens per request: 80.00x 2025-11-03T17:50:35.7381190Z (EngineCore_DP0 pid=23629) WARNING 11-03 17:50:35 [tokenizer.py:255] Using a slow tokenizer. This might cause a significant slowdown. Consider using a fast tokenizer instead. 2025-11-03T17:50:38.6886659Z (EngineCore_DP0 pid=23629) INFO 11-03 17:50:38 [gc_utils.py:40] GC Debug Config. enabled:False,top_objects:-1 2025-11-03T17:50:38.7119877Z INFO 11-03 17:50:38 [llm.py:337] Supported tasks: ['generate'] 2025-11-03T17:50:39.1843899Z PASSED 2025-11-03T17:50:39.1984526Z models/test_initialization.py::test_can_initialize_large_subset[Phi4MultimodalForCausalLM] Fork a new process to run a test 23695 2025-11-03T17:50:39.1994964Z Fork a new process to run a test 0 2025-11-03T17:50:39.2270210Z INFO 11-03 17:50:39 [utils.py:239] non-default args: {'load_format': 'dummy', 'gpu_memory_utilization': 0.8, 'disable_log_stats': True, 'revision': 'refs/pr/70', 'hf_overrides': functools.partial(, model_arch='Phi4MultimodalForCausalLM', exist_overrides={}, use_original_num_layers=False), 'model_impl': 'vllm', 'model': 'microsoft/Phi-4-multimodal-instruct'} 2025-11-03T17:50:39.3062617Z 2025-11-03T17:50:39.3065313Z config.json: 0.00B [00:00, ?B/s] 2025-11-03T17:50:39.3065749Z config.json: 3.71kB [00:00, 17.4MB/s] 2025-11-03T17:50:39.3430266Z INFO 11-03 17:50:39 [config.py:408] Replacing legacy 'type' key with 'rope_type' 2025-11-03T17:50:39.4122222Z 2025-11-03T17:50:39.4123795Z preprocessor_config.json: 0% 0.00/705 [00:00, 'use_cudagraph': True, 'cudagraph_num_of_warmups': 1, 'cudagraph_capture_sizes': [512, 504, 496, 488, 480, 472, 464, 456, 448, 440, 432, 424, 416, 408, 400, 392, 384, 376, 368, 360, 352, 344, 336, 328, 320, 312, 304, 296, 288, 280, 272, 264, 256, 248, 240, 232, 224, 216, 208, 200, 192, 184, 176, 168, 160, 152, 144, 136, 128, 120, 112, 104, 96, 88, 80, 72, 64, 56, 48, 40, 32, 24, 16, 8, 4, 2, 1], 'cudagraph_copy_inputs': False, 'full_cuda_graph': True, 'use_inductor_graph_partition': False, 'pass_config': {}, 'max_capture_size': 512, 'local_cache_dir': None} 2025-11-03T17:50:48.6343395Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:50:48.6372597Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:50:48.6379609Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:50:48.6387850Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:50:48.6395860Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:50:48.6403742Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:50:48.6406536Z (EngineCore_DP0 pid=23773) INFO 11-03 17:50:48 [parallel_state.py:1231] rank 0 in world size 1 is assigned as DP rank 0, PP rank 0, TP rank 0, EP rank 0 2025-11-03T17:50:48.9757153Z (EngineCore_DP0 pid=23773) INFO 11-03 17:50:48 [topk_topp_sampler.py:59] Using FlashInfer for top-p & top-k sampling. 2025-11-03T17:50:52.2183999Z (EngineCore_DP0 pid=23773) It is strongly recommended to pass the `sampling_rate` argument to `Phi4MultimodalFeatureExtractor()`. Failing to do so can result in silent errors that might be hard to debug. 2025-11-03T17:50:52.3870138Z (EngineCore_DP0 pid=23773) INFO 11-03 17:50:52 [gpu_model_runner.py:2840] Starting to load model microsoft/Phi-4-multimodal-instruct... 2025-11-03T17:50:52.6450462Z (EngineCore_DP0 pid=23773) INFO 11-03 17:50:52 [gpu_model_runner.py:2870] Loading model from scratch... 2025-11-03T17:50:52.7341828Z (EngineCore_DP0 pid=23773) INFO 11-03 17:50:52 [cuda.py:404] Using Flash Attention backend on V1 engine. 2025-11-03T17:50:53.3294475Z (EngineCore_DP0 pid=23773) INFO 11-03 17:50:53 [gpu_model_runner.py:2902] Model loading took 2.2617 GiB and 0.162406 seconds 2025-11-03T17:50:53.3298758Z (EngineCore_DP0 pid=23773) INFO 11-03 17:50:53 [kv_cache_utils.py:1199] GPU KV cache size: 2,621,440 tokens 2025-11-03T17:50:53.3299678Z (EngineCore_DP0 pid=23773) INFO 11-03 17:50:53 [kv_cache_utils.py:1204] Maximum concurrency for 131,072 tokens per request: 20.00x 2025-11-03T17:50:54.1387354Z (EngineCore_DP0 pid=23773) It is strongly recommended to pass the `sampling_rate` argument to `Phi4MultimodalFeatureExtractor()`. Failing to do so can result in silent errors that might be hard to debug. 2025-11-03T17:50:54.5984357Z (EngineCore_DP0 pid=23773) INFO 11-03 17:50:54 [gc_utils.py:40] GC Debug Config. enabled:False,top_objects:-1 2025-11-03T17:50:54.6031348Z INFO 11-03 17:50:54 [llm.py:337] Supported tasks: ['generate'] 2025-11-03T17:50:55.1070358Z PASSED 2025-11-03T17:50:55.1212476Z models/test_initialization.py::test_can_initialize_large_subset[Fairseq2LlamaForCausalLM] Fork a new process to run a test 23855 2025-11-03T17:50:55.1222465Z Fork a new process to run a test 0 2025-11-03T17:50:55.1499640Z INFO 11-03 17:50:55 [utils.py:239] non-default args: {'load_format': 'dummy', 'gpu_memory_utilization': 0.8, 'disable_log_stats': True, 'hf_overrides': functools.partial(, model_arch='Fairseq2LlamaForCausalLM', exist_overrides={}, use_original_num_layers=False), 'model_impl': 'vllm', 'model': 'mgleize/fairseq2-dummy-Llama-3.2-1B'} 2025-11-03T17:50:55.2256231Z 2025-11-03T17:50:55.2257693Z config.json: 0% 0.00/591 [00:00, 'use_cudagraph': True, 'cudagraph_num_of_warmups': 1, 'cudagraph_capture_sizes': [512, 504, 496, 488, 480, 472, 464, 456, 448, 440, 432, 424, 416, 408, 400, 392, 384, 376, 368, 360, 352, 344, 336, 328, 320, 312, 304, 296, 288, 280, 272, 264, 256, 248, 240, 232, 224, 216, 208, 200, 192, 184, 176, 168, 160, 152, 144, 136, 128, 120, 112, 104, 96, 88, 80, 72, 64, 56, 48, 40, 32, 24, 16, 8, 4, 2, 1], 'cudagraph_copy_inputs': False, 'full_cuda_graph': True, 'use_inductor_graph_partition': False, 'pass_config': {}, 'max_capture_size': 512, 'local_cache_dir': None} 2025-11-03T17:51:05.9576602Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:51:05.9604668Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:51:05.9612308Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:51:05.9620238Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:51:05.9628816Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:51:05.9636272Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:51:05.9639116Z (EngineCore_DP0 pid=23933) INFO 11-03 17:51:05 [parallel_state.py:1231] rank 0 in world size 1 is assigned as DP rank 0, PP rank 0, TP rank 0, EP rank 0 2025-11-03T17:51:06.3076895Z (EngineCore_DP0 pid=23933) INFO 11-03 17:51:06 [topk_topp_sampler.py:59] Using FlashInfer for top-p & top-k sampling. 2025-11-03T17:51:06.3882966Z (EngineCore_DP0 pid=23933) INFO 11-03 17:51:06 [gpu_model_runner.py:2840] Starting to load model mgleize/fairseq2-dummy-Llama-3.2-1B... 2025-11-03T17:51:06.6545636Z (EngineCore_DP0 pid=23933) INFO 11-03 17:51:06 [gpu_model_runner.py:2870] Loading model from scratch... 2025-11-03T17:51:06.7591060Z (EngineCore_DP0 pid=23933) INFO 11-03 17:51:06 [cuda.py:404] Using Flash Attention backend on V1 engine. 2025-11-03T17:51:07.3491360Z (EngineCore_DP0 pid=23933) INFO 11-03 17:51:07 [gpu_model_runner.py:2902] Model loading took 1.1094 GiB and 0.164524 seconds 2025-11-03T17:51:07.3496028Z (EngineCore_DP0 pid=23933) INFO 11-03 17:51:07 [kv_cache_utils.py:1199] GPU KV cache size: 5,242,880 tokens 2025-11-03T17:51:07.3496969Z (EngineCore_DP0 pid=23933) INFO 11-03 17:51:07 [kv_cache_utils.py:1204] Maximum concurrency for 131,072 tokens per request: 40.00x 2025-11-03T17:51:08.7133964Z (EngineCore_DP0 pid=23933) INFO 11-03 17:51:08 [gc_utils.py:40] GC Debug Config. enabled:False,top_objects:-1 2025-11-03T17:51:08.7154893Z INFO 11-03 17:51:08 [llm.py:337] Supported tasks: ['generate'] 2025-11-03T17:51:09.2012190Z PASSED 2025-11-03T17:51:09.2152352Z models/test_initialization.py::test_can_initialize_large_subset[HunYuanDenseV1ForCausalLM] Fork a new process to run a test 23999 2025-11-03T17:51:09.2162555Z Fork a new process to run a test 0 2025-11-03T17:51:09.2164353Z Model is not available online 2025-11-03T17:51:09.5055542Z PASSED 2025-11-03T17:51:09.5196048Z models/test_initialization.py::test_can_initialize_large_subset[DeepseekVLV2ForCausalLM] Fork a new process to run a test 24000 2025-11-03T17:51:09.5207598Z Fork a new process to run a test 0 2025-11-03T17:51:09.5211517Z `transformers==4.56.2` installed, but `transformers<=4.48` is required to run this model. Reason: HF model is not compatible. 2025-11-03T17:51:09.8029541Z PASSED 2025-11-03T17:51:09.8171834Z models/test_initialization.py::test_can_initialize_large_subset[FalconH1ForCausalLM] Fork a new process to run a test 24001 2025-11-03T17:51:09.8182847Z Fork a new process to run a test 0 2025-11-03T17:51:09.8460343Z INFO 11-03 17:51:09 [utils.py:239] non-default args: {'load_format': 'dummy', 'gpu_memory_utilization': 0.8, 'disable_log_stats': True, 'hf_overrides': functools.partial(, model_arch='FalconH1ForCausalLM', exist_overrides={}, use_original_num_layers=False), 'model_impl': 'vllm', 'model': 'tiiuae/Falcon-H1-0.5B-Base'} 2025-11-03T17:51:09.9625874Z 2025-11-03T17:51:09.9628078Z config.json: 0.00B [00:00, ?B/s] 2025-11-03T17:51:09.9628394Z config.json: 1.58kB [00:00, 7.89MB/s] 2025-11-03T17:51:16.4288873Z INFO 11-03 17:51:16 [model.py:653] Resolved architecture: FalconH1ForCausalLM 2025-11-03T17:51:16.4289370Z `torch_dtype` is deprecated! Use `dtype` instead! 2025-11-03T17:51:16.4542501Z INFO 11-03 17:51:16 [model.py:1714] Using max model len 16384 2025-11-03T17:51:16.6447139Z INFO 11-03 17:51:16 [scheduler.py:225] Chunked prefill is enabled with max_num_batched_tokens=8192. 2025-11-03T17:51:16.6447906Z INFO 11-03 17:51:16 [config.py:313] Disabling cascade attention since it is not supported for hybrid models. 2025-11-03T17:51:16.7004297Z INFO 11-03 17:51:16 [config.py:429] Setting attention block size to 800 tokens to ensure that attention page size is >= mamba page size. 2025-11-03T17:51:16.7005790Z INFO 11-03 17:51:16 [config.py:453] Padding mamba page size by 1.39% to ensure that mamba page size and attention page size are exactly equal. 2025-11-03T17:51:16.7413605Z 2025-11-03T17:51:16.7418517Z tokenizer_config.json: 0.00B [00:00, ?B/s] 2025-11-03T17:51:16.7418859Z tokenizer_config.json: 99.7kB [00:00, 220MB/s] 2025-11-03T17:51:16.8125578Z 2025-11-03T17:51:16.8215885Z tokenizer.json: 0.00B [00:00, ?B/s] 2025-11-03T17:51:16.8216226Z tokenizer.json: 2.35MB [00:00, 261MB/s] 2025-11-03T17:51:16.9212044Z 2025-11-03T17:51:16.9213681Z special_tokens_map.json: 0.00B [00:00, ?B/s] 2025-11-03T17:51:16.9214048Z special_tokens_map.json: 7.42kB [00:00, 58.0MB/s] 2025-11-03T17:51:17.0602795Z 2025-11-03T17:51:17.0603449Z generation_config.json: 0% 0.00/138 [00:00, 'use_cudagraph': True, 'cudagraph_num_of_warmups': 1, 'cudagraph_capture_sizes': [512, 504, 496, 488, 480, 472, 464, 456, 448, 440, 432, 424, 416, 408, 400, 392, 384, 376, 368, 360, 352, 344, 336, 328, 320, 312, 304, 296, 288, 280, 272, 264, 256, 248, 240, 232, 224, 216, 208, 200, 192, 184, 176, 168, 160, 152, 144, 136, 128, 120, 112, 104, 96, 88, 80, 72, 64, 56, 48, 40, 32, 24, 16, 8, 4, 2, 1], 'cudagraph_copy_inputs': False, 'full_cuda_graph': True, 'use_inductor_graph_partition': False, 'pass_config': {}, 'max_capture_size': 512, 'local_cache_dir': None} 2025-11-03T17:51:17.7536208Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:51:17.7564229Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:51:17.7572240Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:51:17.7579406Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:51:17.7586935Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:51:17.7595052Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:51:17.7597846Z (EngineCore_DP0 pid=24059) INFO 11-03 17:51:17 [parallel_state.py:1231] rank 0 in world size 1 is assigned as DP rank 0, PP rank 0, TP rank 0, EP rank 0 2025-11-03T17:51:18.1049685Z (EngineCore_DP0 pid=24059) INFO 11-03 17:51:18 [topk_topp_sampler.py:59] Using FlashInfer for top-p & top-k sampling. 2025-11-03T17:51:18.1460804Z (EngineCore_DP0 pid=24059) INFO 11-03 17:51:18 [gpu_model_runner.py:2840] Starting to load model tiiuae/Falcon-H1-0.5B-Base... 2025-11-03T17:51:18.4122446Z (EngineCore_DP0 pid=24059) INFO 11-03 17:51:18 [gpu_model_runner.py:2870] Loading model from scratch... 2025-11-03T17:51:18.4783370Z (EngineCore_DP0 pid=24059) INFO 11-03 17:51:18 [cuda.py:404] Using Flash Attention backend on V1 engine. 2025-11-03T17:51:19.0713385Z (EngineCore_DP0 pid=24059) INFO 11-03 17:51:19 [gpu_model_runner.py:2902] Model loading took 0.1515 GiB and 0.126629 seconds 2025-11-03T17:51:19.0721340Z (EngineCore_DP0 pid=24059) INFO 11-03 17:51:19 [kv_cache_utils.py:1199] GPU KV cache size: 10,485,600 tokens 2025-11-03T17:51:19.0722580Z (EngineCore_DP0 pid=24059) INFO 11-03 17:51:19 [kv_cache_utils.py:1204] Maximum concurrency for 16,384 tokens per request: 1191.55x 2025-11-03T17:51:19.5400517Z (EngineCore_DP0 pid=24059) INFO 11-03 17:51:19 [gc_utils.py:40] GC Debug Config. enabled:False,top_objects:-1 2025-11-03T17:51:19.5417502Z INFO 11-03 17:51:19 [llm.py:337] Supported tasks: ['generate'] 2025-11-03T17:51:19.9744456Z PASSED 2025-11-03T17:51:19.9885308Z models/test_initialization.py::test_can_initialize_large_subset[Emu3ForConditionalGeneration] Fork a new process to run a test 24125 2025-11-03T17:51:19.9895588Z Fork a new process to run a test 0 2025-11-03T17:51:20.0174024Z INFO 11-03 17:51:20 [utils.py:239] non-default args: {'load_format': 'dummy', 'gpu_memory_utilization': 0.8, 'disable_log_stats': True, 'hf_overrides': functools.partial(, model_arch='Emu3ForConditionalGeneration', exist_overrides={}, use_original_num_layers=False), 'model_impl': 'vllm', 'model': 'BAAI/Emu3-Chat-hf'} 2025-11-03T17:51:20.6443875Z INFO 11-03 17:51:20 [model.py:653] Resolved architecture: Emu3ForConditionalGeneration 2025-11-03T17:51:20.6444428Z `torch_dtype` is deprecated! Use `dtype` instead! 2025-11-03T17:51:20.6680640Z INFO 11-03 17:51:20 [model.py:1939] Downcasting torch.float32 to torch.bfloat16. 2025-11-03T17:51:20.6681218Z INFO 11-03 17:51:20 [model.py:1714] Using max model len 131072 2025-11-03T17:51:20.8931302Z INFO 11-03 17:51:20 [scheduler.py:225] Chunked prefill is enabled with max_num_batched_tokens=8192. 2025-11-03T17:51:41.8934289Z (EngineCore_DP0 pid=24132) INFO 11-03 17:51:41 [core.py:727] Waiting for init message from front-end. 2025-11-03T17:51:41.8969709Z (EngineCore_DP0 pid=24132) INFO 11-03 17:51:41 [core.py:94] Initializing a V1 LLM engine (v0.11.0rc2.dev389+ge51928192.d20251103) with config: model='BAAI/Emu3-Chat-hf', speculative_config=None, tokenizer='BAAI/Emu3-Chat-hf', skip_tokenizer_init=False, tokenizer_mode=auto, revision=None, tokenizer_revision=None, trust_remote_code=False, dtype=torch.bfloat16, max_seq_len=131072, download_dir=None, load_format=dummy, tensor_parallel_size=1, pipeline_parallel_size=1, data_parallel_size=1, disable_custom_all_reduce=False, quantization=None, enforce_eager=False, kv_cache_dtype=auto, device_config=cuda, structured_outputs_config=StructuredOutputsConfig(backend='auto', disable_fallback=False, disable_any_whitespace=False, disable_additional_properties=False, reasoning_parser=''), observability_config=ObservabilityConfig(show_hidden_metrics_for_version=None, otlp_traces_endpoint=None, collect_detailed_traces=None), seed=0, served_model_name=BAAI/Emu3-Chat-hf, enable_prefix_caching=True, chunked_prefill_enabled=True, pooler_config=None, compilation_config={'level': 3, 'debug_dump_path': None, 'cache_dir': '', 'backend': '', 'custom_ops': [], 'splitting_ops': ['vllm.unified_attention', 'vllm.unified_attention_with_output', 'vllm.unified_mla_attention', 'vllm.unified_mla_attention_with_output', 'vllm.mamba_mixer2', 'vllm.mamba_mixer', 'vllm.short_conv', 'vllm.linear_attention', 'vllm.plamo2_mamba_mixer', 'vllm.gdn_attention', 'vllm.sparse_attn_indexer'], 'use_inductor': True, 'compile_sizes': [], 'inductor_compile_config': {'enable_auto_functionalized_v2': False}, 'inductor_passes': {}, 'cudagraph_mode': , 'use_cudagraph': True, 'cudagraph_num_of_warmups': 1, 'cudagraph_capture_sizes': [512, 504, 496, 488, 480, 472, 464, 456, 448, 440, 432, 424, 416, 408, 400, 392, 384, 376, 368, 360, 352, 344, 336, 328, 320, 312, 304, 296, 288, 280, 272, 264, 256, 248, 240, 232, 224, 216, 208, 200, 192, 184, 176, 168, 160, 152, 144, 136, 128, 120, 112, 104, 96, 88, 80, 72, 64, 56, 48, 40, 32, 24, 16, 8, 4, 2, 1], 'cudagraph_copy_inputs': False, 'full_cuda_graph': True, 'use_inductor_graph_partition': False, 'pass_config': {}, 'max_capture_size': 512, 'local_cache_dir': None} 2025-11-03T17:51:42.6432950Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:51:42.6460859Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:51:42.6469701Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:51:42.6477993Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:51:42.6486100Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:51:42.6494309Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:51:42.6497310Z (EngineCore_DP0 pid=24132) INFO 11-03 17:51:42 [parallel_state.py:1231] rank 0 in world size 1 is assigned as DP rank 0, PP rank 0, TP rank 0, EP rank 0 2025-11-03T17:51:42.9953760Z (EngineCore_DP0 pid=24132) INFO 11-03 17:51:42 [topk_topp_sampler.py:59] Using FlashInfer for top-p & top-k sampling. 2025-11-03T17:52:16.6886551Z (EngineCore_DP0 pid=24132) INFO 11-03 17:52:16 [gpu_model_runner.py:2840] Starting to load model BAAI/Emu3-Chat-hf... 2025-11-03T17:52:17.1060842Z (EngineCore_DP0 pid=24132) INFO 11-03 17:52:17 [gpu_model_runner.py:2870] Loading model from scratch... 2025-11-03T17:52:17.1062345Z (EngineCore_DP0 pid=24132) INFO 11-03 17:52:17 [transformers.py:493] Using Transformers backend. 2025-11-03T17:52:17.1426424Z (EngineCore_DP0 pid=24132) `torch_dtype` is deprecated! Use `dtype` instead! 2025-11-03T17:52:17.3556954Z (EngineCore_DP0 pid=24132) INFO 11-03 17:52:17 [cuda.py:404] Using Flash Attention backend on V1 engine. 2025-11-03T17:52:17.9650062Z (EngineCore_DP0 pid=24132) INFO 11-03 17:52:17 [gpu_model_runner.py:2902] Model loading took 3.7283 GiB and 0.334326 seconds 2025-11-03T17:52:17.9654764Z (EngineCore_DP0 pid=24132) INFO 11-03 17:52:17 [kv_cache_utils.py:1199] GPU KV cache size: 2,621,440 tokens 2025-11-03T17:52:17.9657158Z (EngineCore_DP0 pid=24132) INFO 11-03 17:52:17 [kv_cache_utils.py:1204] Maximum concurrency for 131,072 tokens per request: 20.00x 2025-11-03T17:52:29.5924571Z (EngineCore_DP0 pid=24132) INFO 11-03 17:52:29 [gc_utils.py:40] GC Debug Config. enabled:False,top_objects:-1 2025-11-03T17:52:29.5937720Z INFO 11-03 17:52:29 [llm.py:337] Supported tasks: ['generate'] 2025-11-03T17:52:30.1036552Z PASSED 2025-11-03T17:52:30.1178903Z models/test_initialization.py::test_can_initialize_large_subset[BloomForCausalLM] Fork a new process to run a test 24198 2025-11-03T17:52:30.1189674Z Fork a new process to run a test 0 2025-11-03T17:52:30.1468857Z INFO 11-03 17:52:30 [utils.py:239] non-default args: {'load_format': 'dummy', 'gpu_memory_utilization': 0.8, 'disable_log_stats': True, 'hf_overrides': functools.partial(, model_arch='BloomForCausalLM', exist_overrides={}, use_original_num_layers=False), 'model_impl': 'vllm', 'model': 'bigscience/bloom-560m'} 2025-11-03T17:52:30.2328216Z 2025-11-03T17:52:30.2331151Z config.json: 0% 0.00/693 [00:00, 'use_cudagraph': True, 'cudagraph_num_of_warmups': 1, 'cudagraph_capture_sizes': [512, 504, 496, 488, 480, 472, 464, 456, 448, 440, 432, 424, 416, 408, 400, 392, 384, 376, 368, 360, 352, 344, 336, 328, 320, 312, 304, 296, 288, 280, 272, 264, 256, 248, 240, 232, 224, 216, 208, 200, 192, 184, 176, 168, 160, 152, 144, 136, 128, 120, 112, 104, 96, 88, 80, 72, 64, 56, 48, 40, 32, 24, 16, 8, 4, 2, 1], 'cudagraph_copy_inputs': False, 'full_cuda_graph': True, 'use_inductor_graph_partition': False, 'pass_config': {}, 'max_capture_size': 512, 'local_cache_dir': None} 2025-11-03T17:52:39.0704899Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:52:39.0732291Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:52:39.0739193Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:52:39.0747554Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:52:39.0755121Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:52:39.0762682Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:52:39.0765228Z (EngineCore_DP0 pid=24276) INFO 11-03 17:52:39 [parallel_state.py:1231] rank 0 in world size 1 is assigned as DP rank 0, PP rank 0, TP rank 0, EP rank 0 2025-11-03T17:52:39.4074505Z (EngineCore_DP0 pid=24276) INFO 11-03 17:52:39 [topk_topp_sampler.py:59] Using FlashInfer for top-p & top-k sampling. 2025-11-03T17:52:39.4544097Z (EngineCore_DP0 pid=24276) INFO 11-03 17:52:39 [gpu_model_runner.py:2840] Starting to load model bigscience/bloom-560m... 2025-11-03T17:52:39.7113767Z (EngineCore_DP0 pid=24276) INFO 11-03 17:52:39 [gpu_model_runner.py:2870] Loading model from scratch... 2025-11-03T17:52:39.7489394Z (EngineCore_DP0 pid=24276) INFO 11-03 17:52:39 [cuda.py:404] Using Flash Attention backend on V1 engine. 2025-11-03T17:52:40.3176653Z (EngineCore_DP0 pid=24276) INFO 11-03 17:52:40 [gpu_model_runner.py:2902] Model loading took 0.5020 GiB and 0.096373 seconds 2025-11-03T17:52:40.3180893Z (EngineCore_DP0 pid=24276) INFO 11-03 17:52:40 [kv_cache_utils.py:1199] GPU KV cache size: 2,621,440 tokens 2025-11-03T17:52:40.3181783Z (EngineCore_DP0 pid=24276) INFO 11-03 17:52:40 [kv_cache_utils.py:1204] Maximum concurrency for 2,048 tokens per request: 1280.00x 2025-11-03T17:52:41.6571740Z (EngineCore_DP0 pid=24276) INFO 11-03 17:52:41 [gc_utils.py:40] GC Debug Config. enabled:False,top_objects:-1 2025-11-03T17:52:41.6586011Z INFO 11-03 17:52:41 [llm.py:337] Supported tasks: ['generate'] 2025-11-03T17:52:42.1144482Z PASSED 2025-11-03T17:52:42.1284589Z models/test_initialization.py::test_can_initialize_large_subset[LLaMAForCausalLM] Fork a new process to run a test 24342 2025-11-03T17:52:42.1295497Z Fork a new process to run a test 0 2025-11-03T17:52:42.1296778Z Model is not available online 2025-11-03T17:52:42.4159035Z PASSED 2025-11-03T17:52:42.4299154Z models/test_initialization.py::test_can_initialize_large_subset[PersimmonForCausalLM] Fork a new process to run a test 24343 2025-11-03T17:52:42.4309766Z Fork a new process to run a test 0 2025-11-03T17:52:42.4590890Z INFO 11-03 17:52:42 [utils.py:239] non-default args: {'load_format': 'dummy', 'gpu_memory_utilization': 0.8, 'disable_log_stats': True, 'hf_overrides': functools.partial(, model_arch='PersimmonForCausalLM', exist_overrides={}, use_original_num_layers=False), 'model_impl': 'vllm', 'model': 'adept/persimmon-8b-chat'} 2025-11-03T17:52:42.5868330Z 2025-11-03T17:52:42.5869876Z config.json: 0% 0.00/635 [00:00. This is expected, and simply means that the `legacy` (previous) behavior will be used so nothing changes for you. If you want to use the new behaviour, set `legacy=False`. This should only be set if you understand what it means, and thoroughly read the reason why this was added as explained in https://github.com/huggingface/transformers/pull/24565 - if you loaded a llama tokenizer from a GGUF file you can ignore this message 2025-11-03T17:52:50.8681776Z You are using the default legacy behaviour of the . This is expected, and simply means that the `legacy` (previous) behavior will be used so nothing changes for you. If you want to use the new behaviour, set `legacy=False`. This should only be set if you understand what it means, and thoroughly read the reason why this was added as explained in https://github.com/huggingface/transformers/pull/24565 - if you loaded a llama tokenizer from a GGUF file you can ignore this message. 2025-11-03T17:53:13.2430247Z 2025-11-03T17:53:13.2430700Z generation_config.json: 0% 0.00/124 [00:00, 'use_cudagraph': True, 'cudagraph_num_of_warmups': 1, 'cudagraph_capture_sizes': [512, 504, 496, 488, 480, 472, 464, 456, 448, 440, 432, 424, 416, 408, 400, 392, 384, 376, 368, 360, 352, 344, 336, 328, 320, 312, 304, 296, 288, 280, 272, 264, 256, 248, 240, 232, 224, 216, 208, 200, 192, 184, 176, 168, 160, 152, 144, 136, 128, 120, 112, 104, 96, 88, 80, 72, 64, 56, 48, 40, 32, 24, 16, 8, 4, 2, 1], 'cudagraph_copy_inputs': False, 'full_cuda_graph': True, 'use_inductor_graph_partition': False, 'pass_config': {}, 'max_capture_size': 512, 'local_cache_dir': None} 2025-11-03T17:53:13.9668222Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:53:13.9696345Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:53:13.9704247Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:53:13.9711755Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:53:13.9720085Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:53:13.9728796Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:53:13.9731610Z (EngineCore_DP0 pid=24421) INFO 11-03 17:53:13 [parallel_state.py:1231] rank 0 in world size 1 is assigned as DP rank 0, PP rank 0, TP rank 0, EP rank 0 2025-11-03T17:53:14.3060757Z (EngineCore_DP0 pid=24421) INFO 11-03 17:53:14 [topk_topp_sampler.py:59] Using FlashInfer for top-p & top-k sampling. 2025-11-03T17:53:14.3870066Z (EngineCore_DP0 pid=24421) INFO 11-03 17:53:14 [gpu_model_runner.py:2840] Starting to load model adept/persimmon-8b-chat... 2025-11-03T17:53:14.6462487Z (EngineCore_DP0 pid=24421) INFO 11-03 17:53:14 [gpu_model_runner.py:2870] Loading model from scratch... 2025-11-03T17:53:14.7138322Z (EngineCore_DP0 pid=24421) INFO 11-03 17:53:14 [cuda.py:404] Using Flash Attention backend on V1 engine. 2025-11-03T17:53:15.2729655Z (EngineCore_DP0 pid=24421) INFO 11-03 17:53:15 [gpu_model_runner.py:2902] Model loading took 4.3761 GiB and 0.127308 seconds 2025-11-03T17:53:15.2733509Z (EngineCore_DP0 pid=24421) INFO 11-03 17:53:15 [kv_cache_utils.py:1199] GPU KV cache size: 655,360 tokens 2025-11-03T17:53:15.2734397Z (EngineCore_DP0 pid=24421) INFO 11-03 17:53:15 [kv_cache_utils.py:1204] Maximum concurrency for 16,384 tokens per request: 40.00x 2025-11-03T17:53:38.1735698Z (EngineCore_DP0 pid=24421) INFO 11-03 17:53:38 [gc_utils.py:40] GC Debug Config. enabled:False,top_objects:-1 2025-11-03T17:53:38.1749775Z INFO 11-03 17:53:38 [llm.py:337] Supported tasks: ['generate'] 2025-11-03T17:53:38.7152698Z PASSED 2025-11-03T17:53:38.7292137Z models/test_initialization.py::test_can_initialize_large_subset[Terratorch] Fork a new process to run a test 24487 2025-11-03T17:53:38.7303121Z Fork a new process to run a test 0 2025-11-03T17:53:38.7582368Z INFO 11-03 17:53:38 [utils.py:239] non-default args: {'skip_tokenizer_init': True, 'load_format': 'dummy', 'dtype': torch.float16, 'gpu_memory_utilization': 0.8, 'max_num_seqs': 32, 'disable_log_stats': True, 'hf_overrides': functools.partial(, model_arch='Terratorch', exist_overrides={}, use_original_num_layers=False), 'enforce_eager': True, 'model_impl': 'vllm', 'model': 'ibm-nasa-geospatial/Prithvi-EO-2.0-300M-TL-Sen1Floods11'} 2025-11-03T17:53:39.0653407Z INFO 11-03 17:53:39 [model.py:915] Resolved `--runner auto` to `--runner pooling`. Pass the value explicitly to silence this message. 2025-11-03T17:53:39.0654051Z INFO 11-03 17:53:39 [model.py:653] Resolved architecture: Terratorch 2025-11-03T17:53:39.0654561Z `torch_dtype` is deprecated! Use `dtype` instead! 2025-11-03T17:53:39.1181008Z ERROR 11-03 17:53:39 [config.py:298] Error retrieving safetensors: 'ibm-nasa-geospatial/Prithvi-EO-2.0-300M-TL-Sen1Floods11' is not a safetensors repo. Couldn't find 'model.safetensors.index.json' or 'model.safetensors' files., retrying 1 of 2 2025-11-03T17:53:41.1761638Z ERROR 11-03 17:53:41 [config.py:296] Error retrieving safetensors: 'ibm-nasa-geospatial/Prithvi-EO-2.0-300M-TL-Sen1Floods11' is not a safetensors repo. Couldn't find 'model.safetensors.index.json' or 'model.safetensors' files. 2025-11-03T17:53:41.1762840Z INFO 11-03 17:53:41 [model.py:1939] Downcasting torch.float32 to torch.float16. 2025-11-03T17:53:41.1764215Z WARNING 11-03 17:53:41 [model.py:2042] The model's config.json does not contain any of the following keys to determine the original maximum length of the model: ['max_position_embeddings', 'n_positions', 'max_seq_len', 'seq_length', 'model_max_length', 'max_target_positions', 'max_sequence_length', 'max_seq_length', 'seq_len']. Assuming the model's maximum length is 2048. 2025-11-03T17:53:41.1765507Z INFO 11-03 17:53:41 [model.py:1714] Using max model len 2048 2025-11-03T17:53:41.2085402Z INFO 11-03 17:53:41 [arg_utils.py:1725] (Disabling) chunked prefill by default 2025-11-03T17:53:41.2085883Z INFO 11-03 17:53:41 [arg_utils.py:1728] (Disabling) prefix caching by default 2025-11-03T17:53:41.3571893Z INFO 11-03 17:53:41 [vllm.py:354] Cudagraph is disabled under eager mode 2025-11-03T17:53:41.3572514Z INFO 11-03 17:53:41 [vllm.py:433] Only "last" pooling supports chunked prefill and prefix caching; disabling both. 2025-11-03T17:53:45.0132624Z WARNING 11-03 17:53:45 [__init__.py:2879] We must use the `spawn` multiprocessing start method. Overriding VLLM_WORKER_MULTIPROC_METHOD to 'spawn'. See https://docs.vllm.ai/en/latest/usage/troubleshooting.html#python-multiprocessing for more information. Reasons: CUDA is initialized 2025-11-03T17:53:45.5633002Z /opt/conda/envs/py_3.12/lib/python3.12/site-packages/torch/cuda/__init__.py:65: FutureWarning: The pynvml package is deprecated. Please install nvidia-ml-py instead. If you did not install pynvml directly, please report this to the maintainers of the package that installed pynvml for you. 2025-11-03T17:53:45.5634020Z import pynvml # type: ignore[import] 2025-11-03T17:53:47.8324307Z INFO 11-03 17:53:47 [__init__.py:224] Automatically detected platform cuda. 2025-11-03T17:53:50.3576252Z (EngineCore_DP0 pid=24496) INFO 11-03 17:53:50 [core.py:727] Waiting for init message from front-end. 2025-11-03T17:53:50.3837055Z (EngineCore_DP0 pid=24496) INFO 11-03 17:53:50 [core.py:94] Initializing a V1 LLM engine (v0.11.0rc2.dev389+ge51928192.d20251103) with config: model='ibm-nasa-geospatial/Prithvi-EO-2.0-300M-TL-Sen1Floods11', speculative_config=None, tokenizer='ibm-nasa-geospatial/Prithvi-EO-2.0-300M-TL-Sen1Floods11', skip_tokenizer_init=True, tokenizer_mode=auto, revision=None, tokenizer_revision=None, trust_remote_code=False, dtype=torch.float16, max_seq_len=2048, download_dir=None, load_format=dummy, tensor_parallel_size=1, pipeline_parallel_size=1, data_parallel_size=1, disable_custom_all_reduce=False, quantization=None, enforce_eager=True, kv_cache_dtype=auto, device_config=cuda, structured_outputs_config=StructuredOutputsConfig(backend='auto', disable_fallback=False, disable_any_whitespace=False, disable_additional_properties=False, reasoning_parser=''), observability_config=ObservabilityConfig(show_hidden_metrics_for_version=None, otlp_traces_endpoint=None, collect_detailed_traces=None), seed=0, served_model_name=ibm-nasa-geospatial/Prithvi-EO-2.0-300M-TL-Sen1Floods11, enable_prefix_caching=False, chunked_prefill_enabled=False, pooler_config=PoolerConfig(pooling_type='All', normalize=None, dimensions=None, enable_chunked_processing=None, max_embed_len=None, activation=None, logit_bias=None, softmax=None, step_tag_id=None, returned_token_ids=None), compilation_config={'level': 0, 'debug_dump_path': None, 'cache_dir': '', 'backend': '', 'custom_ops': [], 'splitting_ops': None, 'use_inductor': True, 'compile_sizes': [], 'inductor_compile_config': {'enable_auto_functionalized_v2': False}, 'inductor_passes': {}, 'cudagraph_mode': , 'use_cudagraph': False, 'cudagraph_num_of_warmups': 0, 'cudagraph_capture_sizes': [], 'cudagraph_copy_inputs': False, 'full_cuda_graph': False, 'use_inductor_graph_partition': False, 'pass_config': {}, 'max_capture_size': 0, 'local_cache_dir': None} 2025-11-03T17:53:54.2098508Z (EngineCore_DP0 pid=24496) /opt/conda/envs/py_3.12/lib/python3.12/site-packages/torch/library.py:356: UserWarning: Warning only once for all operators, other operators may also be overridden. 2025-11-03T17:53:54.2099582Z (EngineCore_DP0 pid=24496) Overriding a previously registered kernel for the same operator and the same dispatch key 2025-11-03T17:53:54.2101026Z (EngineCore_DP0 pid=24496) operator: xformers_python::_fused_allgather_and_linear_impl(Tensor scattered_input, Tensor[] weights, str process_group_name, Tensor(a3!)[] gathered_outputs, SymInt timeout_s, bool _wait, bool _memcpy, Tensor scale_scattered_input, Tensor?[] scales_weights) -> () 2025-11-03T17:53:54.2102362Z (EngineCore_DP0 pid=24496) registered at /opt/conda/envs/py_3.12/lib/python3.12/site-packages/torch/_library/custom_ops.py:922 2025-11-03T17:53:54.2103002Z (EngineCore_DP0 pid=24496) dispatch key: ADInplaceOrView 2025-11-03T17:53:54.2103425Z (EngineCore_DP0 pid=24496) previous kernel: no debug info 2025-11-03T17:53:54.2104452Z (EngineCore_DP0 pid=24496) new kernel: registered at /opt/conda/envs/py_3.12/lib/python3.12/site-packages/torch/_library/custom_ops.py:922 (Triggered internally at /var/lib/jenkins/workspace/aten/src/ATen/core/dispatch/OperatorEntry.cpp:208.) 2025-11-03T17:53:54.2105340Z (EngineCore_DP0 pid=24496) self.m.impl( 2025-11-03T17:53:54.2399244Z (EngineCore_DP0 pid=24496) /opt/conda/envs/py_3.12/lib/python3.12/site-packages/torch/__init__.py:1628: UserWarning: Please use the new API settings to control TF32 behavior, such as torch.backends.cudnn.conv.fp32_precision = 'tf32' or torch.backends.cuda.matmul.fp32_precision = 'ieee'. Old settings, e.g, torch.backends.cuda.matmul.allow_tf32 = True, torch.backends.cudnn.allow_tf32 = True, allowTF32CuDNN() and allowTF32CuBLAS() will be deprecated after Pytorch 2.9. Please see https://pytorch.org/docs/main/notes/cuda.html#tensorfloat-32-tf32-on-ampere-and-later-devices (Triggered internally at /var/lib/jenkins/workspace/aten/src/ATen/Context.cpp:45.) 2025-11-03T17:53:54.2401655Z (EngineCore_DP0 pid=24496) _C._set_float32_matmul_precision(precision) 2025-11-03T17:53:54.4759727Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:53:54.4786117Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:53:54.4792847Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:53:54.4800264Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:53:54.4807549Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:53:54.4814359Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:53:54.4817094Z (EngineCore_DP0 pid=24496) INFO 11-03 17:53:54 [parallel_state.py:1231] rank 0 in world size 1 is assigned as DP rank 0, PP rank 0, TP rank 0, EP rank 0 2025-11-03T17:53:54.7933434Z (EngineCore_DP0 pid=24496) INFO 11-03 17:53:54 [topk_topp_sampler.py:59] Using FlashInfer for top-p & top-k sampling. 2025-11-03T17:53:54.8022350Z (EngineCore_DP0 pid=24496) INFO 11-03 17:53:54 [gpu_model_runner.py:2840] Starting to load model ibm-nasa-geospatial/Prithvi-EO-2.0-300M-TL-Sen1Floods11... 2025-11-03T17:53:55.1092190Z (EngineCore_DP0 pid=24496) INFO 11-03 17:53:55 [gpu_model_runner.py:2870] Loading model from scratch... 2025-11-03T17:53:56.1284423Z (EngineCore_DP0 pid=24496) WARNING:root:Decoder UperNetDecoder does not have an `includes_head` attribute. Falling back to the value of the registry. 2025-11-03T17:53:56.1286350Z (EngineCore_DP0 pid=24496) /opt/conda/envs/py_3.12/lib/python3.12/site-packages/terratorch/models/decoders/upernet_decoder.py:37: UserWarning: DeprecationWarning: scale_modules is deprecated and will be removed in future versions. Use LearnedInterpolateToPyramidal neck instead. 2025-11-03T17:53:56.1287673Z (EngineCore_DP0 pid=24496) warnings.warn( 2025-11-03T17:53:56.7605996Z (EngineCore_DP0 pid=24496) INFO 11-03 17:53:56 [gpu_model_runner.py:2902] Model loading took 0.5962 GiB and 1.041108 seconds 2025-11-03T17:53:57.9561830Z (EngineCore_DP0 pid=24496) INFO 11-03 17:53:57 [core.py:240] init engine (profile, create kv cache, warmup model) took 1.20 seconds 2025-11-03T17:53:57.9563270Z (EngineCore_DP0 pid=24496) INFO 11-03 17:53:57 [core.py:142] Disabling chunked prefill for model without KVCache 2025-11-03T17:53:57.9608998Z (EngineCore_DP0 pid=24496) INFO 11-03 17:53:57 [vllm.py:354] Cudagraph is disabled under eager mode 2025-11-03T17:53:57.9609999Z (EngineCore_DP0 pid=24496) INFO 11-03 17:53:57 [vllm.py:433] Only "last" pooling supports chunked prefill and prefix caching; disabling both. 2025-11-03T17:53:58.2700359Z (EngineCore_DP0 pid=24496) INFO 11-03 17:53:58 [gc_utils.py:40] GC Debug Config. enabled:False,top_objects:-1 2025-11-03T17:53:58.2711053Z INFO 11-03 17:53:58 [llm.py:337] Supported tasks: ['encode'] 2025-11-03T17:53:58.4117268Z PASSED[rank0]:[W1103 17:53:58.339083754 ProcessGroupNCCL.cpp:1552] Warning: WARNING: destroy_process_group() was not called before program exit, which can leak resources. For more info, please see https://pytorch.org/docs/stable/distributed.html#shutdown (function operator()) 2025-11-03T17:53:58.6540377Z 2025-11-03T17:53:58.6682447Z models/test_initialization.py::test_can_initialize_large_subset[Qwen2AudioForConditionalGeneration] Fork a new process to run a test 24613 2025-11-03T17:53:58.6693080Z Fork a new process to run a test 0 2025-11-03T17:53:58.6970103Z INFO 11-03 17:53:58 [utils.py:239] non-default args: {'load_format': 'dummy', 'gpu_memory_utilization': 0.8, 'disable_log_stats': True, 'hf_overrides': functools.partial(, model_arch='Qwen2AudioForConditionalGeneration', exist_overrides={}, use_original_num_layers=False), 'model_impl': 'vllm', 'model': 'Qwen/Qwen2-Audio-7B-Instruct'} 2025-11-03T17:53:58.7960993Z 2025-11-03T17:53:58.7961457Z config.json: 0% 0.00/853 [00:00, 'use_cudagraph': True, 'cudagraph_num_of_warmups': 1, 'cudagraph_capture_sizes': [512, 504, 496, 488, 480, 472, 464, 456, 448, 440, 432, 424, 416, 408, 400, 392, 384, 376, 368, 360, 352, 344, 336, 328, 320, 312, 304, 296, 288, 280, 272, 264, 256, 248, 240, 232, 224, 216, 208, 200, 192, 184, 176, 168, 160, 152, 144, 136, 128, 120, 112, 104, 96, 88, 80, 72, 64, 56, 48, 40, 32, 24, 16, 8, 4, 2, 1], 'cudagraph_copy_inputs': False, 'full_cuda_graph': True, 'use_inductor_graph_partition': False, 'pass_config': {}, 'max_capture_size': 512, 'local_cache_dir': None} 2025-11-03T17:54:07.7899854Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:54:07.7927972Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:54:07.7936081Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:54:07.7943169Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:54:07.7950735Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:54:07.7958538Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:54:07.7961110Z (EngineCore_DP0 pid=24671) INFO 11-03 17:54:07 [parallel_state.py:1231] rank 0 in world size 1 is assigned as DP rank 0, PP rank 0, TP rank 0, EP rank 0 2025-11-03T17:54:08.1333305Z (EngineCore_DP0 pid=24671) INFO 11-03 17:54:08 [topk_topp_sampler.py:59] Using FlashInfer for top-p & top-k sampling. 2025-11-03T17:54:12.3835878Z (EngineCore_DP0 pid=24671) INFO 11-03 17:54:12 [gpu_model_runner.py:2840] Starting to load model Qwen/Qwen2-Audio-7B-Instruct... 2025-11-03T17:54:12.7267686Z (EngineCore_DP0 pid=24671) INFO 11-03 17:54:12 [gpu_model_runner.py:2870] Loading model from scratch... 2025-11-03T17:54:12.7573816Z (EngineCore_DP0 pid=24671) INFO 11-03 17:54:12 [cuda.py:404] Using Flash Attention backend on V1 engine. 2025-11-03T17:54:13.3276536Z (EngineCore_DP0 pid=24671) INFO 11-03 17:54:13 [gpu_model_runner.py:2902] Model loading took 2.8225 GiB and 0.089636 seconds 2025-11-03T17:54:13.3281133Z (EngineCore_DP0 pid=24671) INFO 11-03 17:54:13 [kv_cache_utils.py:1199] GPU KV cache size: 655,360 tokens 2025-11-03T17:54:13.3282151Z (EngineCore_DP0 pid=24671) INFO 11-03 17:54:13 [kv_cache_utils.py:1204] Maximum concurrency for 8,192 tokens per request: 80.00x 2025-11-03T17:54:14.0845460Z (EngineCore_DP0 pid=24671) INFO 11-03 17:54:14 [gc_utils.py:40] GC Debug Config. enabled:False,top_objects:-1 2025-11-03T17:54:14.0861835Z INFO 11-03 17:54:14 [llm.py:337] Supported tasks: ['generate'] 2025-11-03T17:54:14.5530708Z PASSED 2025-11-03T17:54:14.5673312Z models/test_initialization.py::test_can_initialize_large_subset[RobertaForSequenceClassification] Fork a new process to run a test 24753 2025-11-03T17:54:14.5684819Z Fork a new process to run a test 0 2025-11-03T17:54:14.5967807Z INFO 11-03 17:54:14 [utils.py:239] non-default args: {'load_format': 'dummy', 'gpu_memory_utilization': 0.8, 'disable_log_stats': True, 'hf_overrides': functools.partial(, model_arch='RobertaForSequenceClassification', exist_overrides={}, use_original_num_layers=False), 'model_impl': 'vllm', 'model': 'cross-encoder/quora-roberta-base'} 2025-11-03T17:54:14.7138644Z 2025-11-03T17:54:14.7139277Z config.json: 0% 0.00/608 [00:00, 'use_cudagraph': True, 'cudagraph_num_of_warmups': 1, 'cudagraph_capture_sizes': [512, 504, 496, 488, 480, 472, 464, 456, 448, 440, 432, 424, 416, 408, 400, 392, 384, 376, 368, 360, 352, 344, 336, 328, 320, 312, 304, 296, 288, 280, 272, 264, 256, 248, 240, 232, 224, 216, 208, 200, 192, 184, 176, 168, 160, 152, 144, 136, 128, 120, 112, 104, 96, 88, 80, 72, 64, 56, 48, 40, 32, 24, 16, 8, 4, 2, 1], 'cudagraph_copy_inputs': False, 'full_cuda_graph': False, 'use_inductor_graph_partition': False, 'pass_config': {}, 'max_capture_size': 512, 'local_cache_dir': None} 2025-11-03T17:54:16.4152936Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:54:16.4184121Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:54:16.4191466Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:54:16.4199661Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:54:16.4207262Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:54:16.4214760Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:54:16.4217687Z (EngineCore_DP0 pid=24761) INFO 11-03 17:54:16 [parallel_state.py:1231] rank 0 in world size 1 is assigned as DP rank 0, PP rank 0, TP rank 0, EP rank 0 2025-11-03T17:54:16.7654072Z (EngineCore_DP0 pid=24761) INFO 11-03 17:54:16 [topk_topp_sampler.py:59] Using FlashInfer for top-p & top-k sampling. 2025-11-03T17:54:16.7796174Z (EngineCore_DP0 pid=24761) INFO 11-03 17:54:16 [gpu_model_runner.py:2840] Starting to load model cross-encoder/quora-roberta-base... 2025-11-03T17:54:17.0367151Z (EngineCore_DP0 pid=24761) INFO 11-03 17:54:17 [gpu_model_runner.py:2870] Loading model from scratch... 2025-11-03T17:54:17.0423205Z (EngineCore_DP0 pid=24761) INFO 11-03 17:54:17 [cuda.py:404] Using Flash Attention backend on V1 engine. 2025-11-03T17:54:17.6190662Z (EngineCore_DP0 pid=24761) INFO 11-03 17:54:17 [gpu_model_runner.py:2902] Model loading took 0.0884 GiB and 0.066469 seconds 2025-11-03T17:54:17.8176108Z (EngineCore_DP0 pid=24761) INFO 11-03 17:54:17 [core.py:142] Disabling chunked prefill for model without KVCache 2025-11-03T17:54:17.9228224Z (EngineCore_DP0 pid=24761) INFO 11-03 17:54:17 [vllm.py:433] Only "last" pooling supports chunked prefill and prefix caching; disabling both. 2025-11-03T17:54:18.1908595Z (EngineCore_DP0 pid=24761) INFO 11-03 17:54:18 [gc_utils.py:40] GC Debug Config. enabled:False,top_objects:-1 2025-11-03T17:54:18.1930525Z INFO 11-03 17:54:18 [llm.py:337] Supported tasks: ['score', 'classify', 'encode'] 2025-11-03T17:54:18.6122420Z PASSED 2025-11-03T17:54:18.6263585Z models/test_initialization.py::test_can_initialize_large_subset[FuyuForCausalLM] Fork a new process to run a test 24827 2025-11-03T17:54:18.6274539Z Fork a new process to run a test 0 2025-11-03T17:54:18.6554723Z INFO 11-03 17:54:18 [utils.py:239] non-default args: {'load_format': 'dummy', 'gpu_memory_utilization': 0.8, 'disable_log_stats': True, 'hf_overrides': functools.partial(, model_arch='FuyuForCausalLM', exist_overrides={}, use_original_num_layers=False), 'model_impl': 'vllm', 'model': 'adept/fuyu-8b'} 2025-11-03T17:54:18.7327587Z 2025-11-03T17:54:18.7330923Z config.json: 0% 0.00/768 [00:00, 'use_cudagraph': True, 'cudagraph_num_of_warmups': 1, 'cudagraph_capture_sizes': [512, 504, 496, 488, 480, 472, 464, 456, 448, 440, 432, 424, 416, 408, 400, 392, 384, 376, 368, 360, 352, 344, 336, 328, 320, 312, 304, 296, 288, 280, 272, 264, 256, 248, 240, 232, 224, 216, 208, 200, 192, 184, 176, 168, 160, 152, 144, 136, 128, 120, 112, 104, 96, 88, 80, 72, 64, 56, 48, 40, 32, 24, 16, 8, 4, 2, 1], 'cudagraph_copy_inputs': False, 'full_cuda_graph': True, 'use_inductor_graph_partition': False, 'pass_config': {}, 'max_capture_size': 512, 'local_cache_dir': None} 2025-11-03T17:54:28.6892902Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:54:28.6921662Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:54:28.6929524Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:54:28.6937729Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:54:28.6946544Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:54:28.6954622Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:54:28.6957469Z (EngineCore_DP0 pid=24906) INFO 11-03 17:54:28 [parallel_state.py:1231] rank 0 in world size 1 is assigned as DP rank 0, PP rank 0, TP rank 0, EP rank 0 2025-11-03T17:54:29.0366970Z (EngineCore_DP0 pid=24906) INFO 11-03 17:54:29 [topk_topp_sampler.py:59] Using FlashInfer for top-p & top-k sampling. 2025-11-03T17:54:33.2696265Z (EngineCore_DP0 pid=24906) INFO 11-03 17:54:33 [gpu_model_runner.py:2840] Starting to load model adept/fuyu-8b... 2025-11-03T17:54:33.5304047Z (EngineCore_DP0 pid=24906) INFO 11-03 17:54:33 [gpu_model_runner.py:2870] Loading model from scratch... 2025-11-03T17:54:33.5572422Z (EngineCore_DP0 pid=24906) INFO 11-03 17:54:33 [cuda.py:404] Using Flash Attention backend on V1 engine. 2025-11-03T17:54:34.1365126Z (EngineCore_DP0 pid=24906) INFO 11-03 17:54:34 [gpu_model_runner.py:2902] Model loading took 4.3976 GiB and 0.086209 seconds 2025-11-03T17:54:34.1369606Z (EngineCore_DP0 pid=24906) INFO 11-03 17:54:34 [kv_cache_utils.py:1199] GPU KV cache size: 655,360 tokens 2025-11-03T17:54:34.1370509Z (EngineCore_DP0 pid=24906) INFO 11-03 17:54:34 [kv_cache_utils.py:1204] Maximum concurrency for 16,384 tokens per request: 40.00x 2025-11-03T17:54:35.2015849Z (EngineCore_DP0 pid=24906) INFO 11-03 17:54:35 [gc_utils.py:40] GC Debug Config. enabled:False,top_objects:-1 2025-11-03T17:54:35.2030902Z INFO 11-03 17:54:35 [llm.py:337] Supported tasks: ['generate'] 2025-11-03T17:54:35.8349893Z PASSED 2025-11-03T17:54:35.8490915Z models/test_initialization.py::test_can_initialize_large_subset[BertModel] Fork a new process to run a test 24988 2025-11-03T17:54:35.8501948Z Fork a new process to run a test 0 2025-11-03T17:54:35.8785358Z INFO 11-03 17:54:35 [utils.py:239] non-default args: {'load_format': 'dummy', 'gpu_memory_utilization': 0.8, 'disable_log_stats': True, 'hf_overrides': functools.partial(, model_arch='BertModel', exist_overrides={}, use_original_num_layers=False), 'model_impl': 'vllm', 'model': 'BAAI/bge-base-en-v1.5'} 2025-11-03T17:54:35.9545262Z 2025-11-03T17:54:35.9546270Z config.json: 0% 0.00/777 [00:00, 'use_cudagraph': True, 'cudagraph_num_of_warmups': 1, 'cudagraph_capture_sizes': [512, 504, 496, 488, 480, 472, 464, 456, 448, 440, 432, 424, 416, 408, 400, 392, 384, 376, 368, 360, 352, 344, 336, 328, 320, 312, 304, 296, 288, 280, 272, 264, 256, 248, 240, 232, 224, 216, 208, 200, 192, 184, 176, 168, 160, 152, 144, 136, 128, 120, 112, 104, 96, 88, 80, 72, 64, 56, 48, 40, 32, 24, 16, 8, 4, 2, 1], 'cudagraph_copy_inputs': False, 'full_cuda_graph': False, 'use_inductor_graph_partition': False, 'pass_config': {}, 'max_capture_size': 512, 'local_cache_dir': None} 2025-11-03T17:54:37.4714202Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:54:37.4746325Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:54:37.4754040Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:54:37.4762402Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:54:37.4770561Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:54:37.4778671Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:54:37.4781408Z (EngineCore_DP0 pid=24996) INFO 11-03 17:54:37 [parallel_state.py:1231] rank 0 in world size 1 is assigned as DP rank 0, PP rank 0, TP rank 0, EP rank 0 2025-11-03T17:54:37.8142735Z (EngineCore_DP0 pid=24996) INFO 11-03 17:54:37 [topk_topp_sampler.py:59] Using FlashInfer for top-p & top-k sampling. 2025-11-03T17:54:37.8247082Z (EngineCore_DP0 pid=24996) INFO 11-03 17:54:37 [gpu_model_runner.py:2840] Starting to load model BAAI/bge-base-en-v1.5... 2025-11-03T17:54:38.0828929Z (EngineCore_DP0 pid=24996) INFO 11-03 17:54:38 [gpu_model_runner.py:2870] Loading model from scratch... 2025-11-03T17:54:38.0873340Z (EngineCore_DP0 pid=24996) INFO 11-03 17:54:38 [cuda.py:404] Using Flash Attention backend on V1 engine. 2025-11-03T17:54:38.6635507Z (EngineCore_DP0 pid=24996) INFO 11-03 17:54:38 [gpu_model_runner.py:2902] Model loading took 0.0579 GiB and 0.065111 seconds 2025-11-03T17:54:38.7646215Z (EngineCore_DP0 pid=24996) INFO 11-03 17:54:38 [core.py:142] Disabling chunked prefill for model without KVCache 2025-11-03T17:54:38.8627666Z (EngineCore_DP0 pid=24996) INFO 11-03 17:54:38 [vllm.py:433] Only "last" pooling supports chunked prefill and prefix caching; disabling both. 2025-11-03T17:54:39.1265960Z (EngineCore_DP0 pid=24996) INFO 11-03 17:54:39 [gc_utils.py:40] GC Debug Config. enabled:False,top_objects:-1 2025-11-03T17:54:39.1280114Z INFO 11-03 17:54:39 [llm.py:337] Supported tasks: ['embed', 'encode'] 2025-11-03T17:54:39.5590335Z PASSED 2025-11-03T17:54:39.5730923Z models/test_initialization.py::test_can_initialize_large_subset[VoxtralForConditionalGeneration] Fork a new process to run a test 25062 2025-11-03T17:54:39.5742158Z Fork a new process to run a test 0 2025-11-03T17:54:39.5743897Z Model is not available online 2025-11-03T17:54:39.8637455Z PASSED 2025-11-03T17:54:39.8781038Z models/test_initialization.py::test_can_initialize_large_subset[TeleChat2ForCausalLM] Fork a new process to run a test 25063 2025-11-03T17:54:39.8792204Z Fork a new process to run a test 0 2025-11-03T17:54:39.9067195Z INFO 11-03 17:54:39 [utils.py:239] non-default args: {'trust_remote_code': True, 'load_format': 'dummy', 'gpu_memory_utilization': 0.8, 'disable_log_stats': True, 'hf_overrides': functools.partial(, model_arch='TeleChat2ForCausalLM', exist_overrides={}, use_original_num_layers=False), 'model_impl': 'vllm', 'model': 'Tele-AI/TeleChat2-3B'} 2025-11-03T17:54:39.9838500Z 2025-11-03T17:54:39.9840828Z config.json: 0.00B [00:00, ?B/s] 2025-11-03T17:54:39.9841312Z config.json: 1.04kB [00:00, 5.24MB/s] 2025-11-03T17:54:40.0579840Z 2025-11-03T17:54:40.0581951Z configuration_telechat2.py: 0.00B [00:00, ?B/s] 2025-11-03T17:54:40.0582574Z configuration_telechat2.py: 4.48kB [00:00, 25.6MB/s] 2025-11-03T17:54:40.0676376Z A new version of the following files was downloaded from https://huggingface.co/Tele-AI/TeleChat2-3B: 2025-11-03T17:54:40.0690754Z - configuration_telechat2.py 2025-11-03T17:54:40.0691397Z . Make sure to double-check they do not contain any added malicious code. To avoid downloading new versions of the code file, you can pin a revision. 2025-11-03T17:54:46.5924863Z INFO 11-03 17:54:46 [model.py:653] Resolved architecture: TeleChat2ForCausalLM 2025-11-03T17:54:46.5925349Z `torch_dtype` is deprecated! Use `dtype` instead! 2025-11-03T17:54:46.6585352Z ERROR 11-03 17:54:46 [config.py:298] Error retrieving safetensors: 'Tele-AI/TeleChat2-3B' is not a safetensors repo. Couldn't find 'model.safetensors.index.json' or 'model.safetensors' files., retrying 1 of 2 2025-11-03T17:54:48.7076799Z ERROR 11-03 17:54:48 [config.py:296] Error retrieving safetensors: 'Tele-AI/TeleChat2-3B' is not a safetensors repo. Couldn't find 'model.safetensors.index.json' or 'model.safetensors' files. 2025-11-03T17:54:48.7326380Z INFO 11-03 17:54:48 [model.py:1939] Downcasting torch.float32 to torch.bfloat16. 2025-11-03T17:54:48.7331693Z INFO 11-03 17:54:48 [model.py:1714] Using max model len 32768 2025-11-03T17:54:48.7334008Z The argument `trust_remote_code` is to be used with Auto classes. It has no effect here and is ignored. 2025-11-03T17:54:48.7895015Z INFO 11-03 17:54:48 [scheduler.py:225] Chunked prefill is enabled with max_num_batched_tokens=8192. 2025-11-03T17:54:48.8289198Z 2025-11-03T17:54:48.8291432Z tokenizer_config.json: 0.00B [00:00, ?B/s] 2025-11-03T17:54:48.8291906Z tokenizer_config.json: 5.07kB [00:00, 12.8MB/s] 2025-11-03T17:54:48.8649515Z 2025-11-03T17:54:48.8651607Z tokenization_telechat2.py: 0.00B [00:00, ?B/s] 2025-11-03T17:54:48.8652083Z tokenization_telechat2.py: 8.82kB [00:00, 29.5MB/s] 2025-11-03T17:54:48.8702003Z A new version of the following files was downloaded from https://huggingface.co/Tele-AI/TeleChat2-3B: 2025-11-03T17:54:48.8702522Z - tokenization_telechat2.py 2025-11-03T17:54:48.8703173Z . Make sure to double-check they do not contain any added malicious code. To avoid downloading new versions of the code file, you can pin a revision. 2025-11-03T17:54:48.9718221Z 2025-11-03T17:54:49.1481914Z tokenizer.model: 0% 0.00/2.20M [00:00, 'use_cudagraph': True, 'cudagraph_num_of_warmups': 1, 'cudagraph_capture_sizes': [512, 504, 496, 488, 480, 472, 464, 456, 448, 440, 432, 424, 416, 408, 400, 392, 384, 376, 368, 360, 352, 344, 336, 328, 320, 312, 304, 296, 288, 280, 272, 264, 256, 248, 240, 232, 224, 216, 208, 200, 192, 184, 176, 168, 160, 152, 144, 136, 128, 120, 112, 104, 96, 88, 80, 72, 64, 56, 48, 40, 32, 24, 16, 8, 4, 2, 1], 'cudagraph_copy_inputs': False, 'full_cuda_graph': True, 'use_inductor_graph_partition': False, 'pass_config': {}, 'max_capture_size': 512, 'local_cache_dir': None} 2025-11-03T17:54:50.4106458Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:54:50.4134212Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:54:50.4141738Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:54:50.4149197Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:54:50.4156515Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:54:50.4164323Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:54:50.4166760Z (EngineCore_DP0 pid=25141) INFO 11-03 17:54:50 [parallel_state.py:1231] rank 0 in world size 1 is assigned as DP rank 0, PP rank 0, TP rank 0, EP rank 0 2025-11-03T17:54:50.7492228Z (EngineCore_DP0 pid=25141) INFO 11-03 17:54:50 [topk_topp_sampler.py:59] Using FlashInfer for top-p & top-k sampling. 2025-11-03T17:54:50.8217720Z (EngineCore_DP0 pid=25141) INFO 11-03 17:54:50 [gpu_model_runner.py:2840] Starting to load model Tele-AI/TeleChat2-3B... 2025-11-03T17:54:51.0781846Z (EngineCore_DP0 pid=25141) INFO 11-03 17:54:51 [gpu_model_runner.py:2870] Loading model from scratch... 2025-11-03T17:54:51.1430024Z (EngineCore_DP0 pid=25141) INFO 11-03 17:54:51 [cuda.py:404] Using Flash Attention backend on V1 engine. 2025-11-03T17:54:51.7368644Z (EngineCore_DP0 pid=25141) INFO 11-03 17:54:51 [gpu_model_runner.py:2902] Model loading took 1.6778 GiB and 0.150800 seconds 2025-11-03T17:54:51.7373523Z (EngineCore_DP0 pid=25141) INFO 11-03 17:54:51 [kv_cache_utils.py:1199] GPU KV cache size: 873,808 tokens 2025-11-03T17:54:51.7374559Z (EngineCore_DP0 pid=25141) INFO 11-03 17:54:51 [kv_cache_utils.py:1204] Maximum concurrency for 32,768 tokens per request: 26.67x 2025-11-03T17:54:52.1123368Z (EngineCore_DP0 pid=25141) WARNING 11-03 17:54:52 [tokenizer.py:255] Using a slow tokenizer. This might cause a significant slowdown. Consider using a fast tokenizer instead. 2025-11-03T17:54:52.5897979Z (EngineCore_DP0 pid=25141) INFO 11-03 17:54:52 [gc_utils.py:40] GC Debug Config. enabled:False,top_objects:-1 2025-11-03T17:54:52.5914397Z INFO 11-03 17:54:52 [llm.py:337] Supported tasks: ['generate'] 2025-11-03T17:54:53.0610186Z PASSED 2025-11-03T17:54:53.0753160Z models/test_initialization.py::test_can_initialize_large_subset[Qwen3VLMoeForConditionalGeneration] Fork a new process to run a test 25207 2025-11-03T17:54:53.0765213Z Fork a new process to run a test 0 2025-11-03T17:54:53.0767252Z Model is not available online 2025-11-03T17:54:53.3636461Z PASSED 2025-11-03T17:54:53.3777787Z models/test_initialization.py::test_can_initialize_large_subset[Dots1ForCausalLM] Fork a new process to run a test 25208 2025-11-03T17:54:53.3788748Z Fork a new process to run a test 0 2025-11-03T17:54:53.4063647Z INFO 11-03 17:54:53 [utils.py:239] non-default args: {'load_format': 'dummy', 'gpu_memory_utilization': 0.8, 'disable_log_stats': True, 'hf_overrides': functools.partial(, model_arch='Dots1ForCausalLM', exist_overrides={}, use_original_num_layers=False), 'model_impl': 'vllm', 'model': 'rednote-hilab/dots.llm1.inst'} 2025-11-03T17:54:53.4999220Z 2025-11-03T17:54:53.5000049Z config.json: 0% 0.00/988 [00:00, 'use_cudagraph': True, 'cudagraph_num_of_warmups': 1, 'cudagraph_capture_sizes': [512, 504, 496, 488, 480, 472, 464, 456, 448, 440, 432, 424, 416, 408, 400, 392, 384, 376, 368, 360, 352, 344, 336, 328, 320, 312, 304, 296, 288, 280, 272, 264, 256, 248, 240, 232, 224, 216, 208, 200, 192, 184, 176, 168, 160, 152, 144, 136, 128, 120, 112, 104, 96, 88, 80, 72, 64, 56, 48, 40, 32, 24, 16, 8, 4, 2, 1], 'cudagraph_copy_inputs': False, 'full_cuda_graph': True, 'use_inductor_graph_partition': False, 'pass_config': {}, 'max_capture_size': 512, 'local_cache_dir': None} 2025-11-03T17:55:01.4977327Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:55:01.5004491Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:55:01.5012955Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:55:01.5021350Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:55:01.5028915Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:55:01.5036736Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:55:01.5039749Z (EngineCore_DP0 pid=25266) INFO 11-03 17:55:01 [parallel_state.py:1231] rank 0 in world size 1 is assigned as DP rank 0, PP rank 0, TP rank 0, EP rank 0 2025-11-03T17:55:01.8441290Z (EngineCore_DP0 pid=25266) INFO 11-03 17:55:01 [topk_topp_sampler.py:59] Using FlashInfer for top-p & top-k sampling. 2025-11-03T17:55:01.9135668Z (EngineCore_DP0 pid=25266) INFO 11-03 17:55:01 [gpu_model_runner.py:2840] Starting to load model rednote-hilab/dots.llm1.inst... 2025-11-03T17:55:02.1701873Z (EngineCore_DP0 pid=25266) INFO 11-03 17:55:02 [gpu_model_runner.py:2870] Loading model from scratch... 2025-11-03T17:55:02.2347534Z (EngineCore_DP0 pid=25266) INFO 11-03 17:55:02 [cuda.py:404] Using Flash Attention backend on V1 engine. 2025-11-03T17:55:02.7999742Z (EngineCore_DP0 pid=25266) INFO 11-03 17:55:02 [gpu_model_runner.py:2902] Model loading took 2.5821 GiB and 0.125022 seconds 2025-11-03T17:55:02.8005127Z (EngineCore_DP0 pid=25266) INFO 11-03 17:55:02 [kv_cache_utils.py:1199] GPU KV cache size: 655,360 tokens 2025-11-03T17:55:02.8005907Z (EngineCore_DP0 pid=25266) INFO 11-03 17:55:02 [kv_cache_utils.py:1204] Maximum concurrency for 32,768 tokens per request: 20.00x 2025-11-03T17:55:03.4686764Z (EngineCore_DP0 pid=25266) INFO 11-03 17:55:03 [gc_utils.py:40] GC Debug Config. enabled:False,top_objects:-1 2025-11-03T17:55:03.4709098Z INFO 11-03 17:55:03 [llm.py:337] Supported tasks: ['generate'] 2025-11-03T17:55:03.9279243Z PASSED 2025-11-03T17:55:03.9419621Z models/test_initialization.py::test_can_initialize_large_subset[EagleDeepSeekMTPModel] Fork a new process to run a test 25332 2025-11-03T17:55:03.9431585Z Fork a new process to run a test 0 2025-11-03T17:55:03.9705685Z INFO 11-03 17:55:03 [utils.py:239] non-default args: {'trust_remote_code': True, 'load_format': 'dummy', 'gpu_memory_utilization': 0.8, 'disable_log_stats': True, 'hf_overrides': functools.partial(, model_arch='EagleDeepSeekMTPModel', exist_overrides={}, use_original_num_layers=False), 'speculative_config': {'model': 'eagle618/eagle-deepseek-v3-random', 'num_speculative_tokens': 1}, 'model_impl': 'vllm', 'model': 'eagle618/deepseek-v3-random'} 2025-11-03T17:55:04.0568118Z 2025-11-03T17:55:04.0570253Z config.json: 0.00B [00:00, ?B/s] 2025-11-03T17:55:04.0570571Z config.json: 1.95kB [00:00, 9.96MB/s] 2025-11-03T17:55:04.1089171Z INFO 11-03 17:55:04 [config.py:408] Replacing legacy 'type' key with 'rope_type' 2025-11-03T17:55:04.1842555Z INFO 11-03 17:55:04 [model.py:653] Resolved architecture: DeepseekV3ForCausalLM 2025-11-03T17:55:04.1843017Z `torch_dtype` is deprecated! Use `dtype` instead! 2025-11-03T17:55:04.2087147Z INFO 11-03 17:55:04 [model.py:1714] Using max model len 163840 2025-11-03T17:55:04.3021012Z The argument `trust_remote_code` is to be used with Auto classes. It has no effect here and is ignored. 2025-11-03T17:55:04.6073050Z 2025-11-03T17:55:04.6074046Z config.json: 0.00B [00:00, ?B/s] 2025-11-03T17:55:04.6074436Z config.json: 1.79kB [00:00, 11.5MB/s] 2025-11-03T17:55:04.6365060Z INFO 11-03 17:55:04 [config.py:408] Replacing legacy 'type' key with 'rope_type' 2025-11-03T17:55:04.7712554Z INFO 11-03 17:55:04 [model.py:653] Resolved architecture: DeepSeekMTPModel 2025-11-03T17:55:04.7714432Z INFO 11-03 17:55:04 [model.py:1714] Using max model len 163840 2025-11-03T17:55:04.7723330Z INFO 11-03 17:55:04 [scheduler.py:225] Chunked prefill is enabled with max_num_batched_tokens=8192. 2025-11-03T17:55:04.8147771Z 2025-11-03T17:55:04.8148726Z tokenizer_config.json: 0.00B [00:00, ?B/s] 2025-11-03T17:55:04.8149074Z tokenizer_config.json: 3.13kB [00:00, 19.8MB/s] 2025-11-03T17:55:04.9282524Z 2025-11-03T17:55:04.9578822Z tokenizer.json: 0.00B [00:00, ?B/s] 2025-11-03T17:55:04.9579168Z tokenizer.json: 7.85MB [00:00, 266MB/s] 2025-11-03T17:55:05.6161338Z INFO 11-03 17:55:05 [config.py:408] Replacing legacy 'type' key with 'rope_type' 2025-11-03T17:55:05.6387206Z (EngineCore_DP0 pid=25340) INFO 11-03 17:55:05 [core.py:727] Waiting for init message from front-end. 2025-11-03T17:55:05.6424481Z (EngineCore_DP0 pid=25340) INFO 11-03 17:55:05 [core.py:94] Initializing a V1 LLM engine (v0.11.0rc2.dev389+ge51928192.d20251103) with config: model='eagle618/deepseek-v3-random', speculative_config=SpeculativeConfig(method='eagle', model='eagle618/eagle-deepseek-v3-random', num_spec_tokens=1), tokenizer='eagle618/deepseek-v3-random', skip_tokenizer_init=False, tokenizer_mode=auto, revision=None, tokenizer_revision=None, trust_remote_code=True, dtype=torch.bfloat16, max_seq_len=163840, download_dir=None, load_format=dummy, tensor_parallel_size=1, pipeline_parallel_size=1, data_parallel_size=1, disable_custom_all_reduce=False, quantization=fp8, enforce_eager=False, kv_cache_dtype=auto, device_config=cuda, structured_outputs_config=StructuredOutputsConfig(backend='auto', disable_fallback=False, disable_any_whitespace=False, disable_additional_properties=False, reasoning_parser=''), observability_config=ObservabilityConfig(show_hidden_metrics_for_version=None, otlp_traces_endpoint=None, collect_detailed_traces=None), seed=0, served_model_name=eagle618/deepseek-v3-random, enable_prefix_caching=True, chunked_prefill_enabled=True, pooler_config=None, compilation_config={'level': 3, 'debug_dump_path': None, 'cache_dir': '', 'backend': '', 'custom_ops': ['+quant_fp8'], 'splitting_ops': ['vllm.unified_attention', 'vllm.unified_attention_with_output', 'vllm.unified_mla_attention', 'vllm.unified_mla_attention_with_output', 'vllm.mamba_mixer2', 'vllm.mamba_mixer', 'vllm.short_conv', 'vllm.linear_attention', 'vllm.plamo2_mamba_mixer', 'vllm.gdn_attention', 'vllm.sparse_attn_indexer'], 'use_inductor': True, 'compile_sizes': [], 'inductor_compile_config': {'enable_auto_functionalized_v2': False}, 'inductor_passes': {}, 'cudagraph_mode': , 'use_cudagraph': True, 'cudagraph_num_of_warmups': 1, 'cudagraph_capture_sizes': [512, 504, 496, 488, 480, 472, 464, 456, 448, 440, 432, 424, 416, 408, 400, 392, 384, 376, 368, 360, 352, 344, 336, 328, 320, 312, 304, 296, 288, 280, 272, 264, 256, 248, 240, 232, 224, 216, 208, 200, 192, 184, 176, 168, 160, 152, 144, 136, 128, 120, 112, 104, 96, 88, 80, 72, 64, 56, 48, 40, 32, 24, 16, 8, 4, 2, 1], 'cudagraph_copy_inputs': False, 'full_cuda_graph': True, 'use_inductor_graph_partition': False, 'pass_config': {}, 'max_capture_size': 512, 'local_cache_dir': None} 2025-11-03T17:55:06.3680170Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:55:06.3707810Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:55:06.3715597Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:55:06.3723461Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:55:06.3731480Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:55:06.3739298Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:55:06.3742245Z (EngineCore_DP0 pid=25340) INFO 11-03 17:55:06 [parallel_state.py:1231] rank 0 in world size 1 is assigned as DP rank 0, PP rank 0, TP rank 0, EP rank 0 2025-11-03T17:55:06.7147572Z (EngineCore_DP0 pid=25340) INFO 11-03 17:55:06 [topk_topp_sampler.py:59] Using FlashInfer for top-p & top-k sampling. 2025-11-03T17:55:06.7295474Z (EngineCore_DP0 pid=25340) WARNING 11-03 17:55:06 [__init__.py:194] min_p, logit_bias, and min_tokens parameters won't currently work with speculative decoding enabled. 2025-11-03T17:55:06.7923336Z (EngineCore_DP0 pid=25340) INFO 11-03 17:55:06 [gpu_model_runner.py:2840] Starting to load model eagle618/deepseek-v3-random... 2025-11-03T17:55:07.0517923Z (EngineCore_DP0 pid=25340) INFO 11-03 17:55:07 [gpu_model_runner.py:2870] Loading model from scratch... 2025-11-03T17:55:07.1311901Z (EngineCore_DP0 pid=25340) INFO 11-03 17:55:07 [cuda.py:328] Using Triton MLA backend on V1 engine. 2025-11-03T17:55:07.1465995Z (EngineCore_DP0 pid=25340) WARNING 11-03 17:55:07 [fp8.py:150] DeepGEMM backend requested but not available. 2025-11-03T17:55:07.1466820Z (EngineCore_DP0 pid=25340) INFO 11-03 17:55:07 [fp8.py:165] Using Triton backend for FP8 MoE 2025-11-03T17:55:07.2141028Z (EngineCore_DP0 pid=25340) WARNING 11-03 17:55:07 [fp8_utils.py:785] Using default W8A8 Block FP8 kernel config. Performance might be sub-optimal! Config file not found at /opt/conda/envs/py_3.12/lib/python3.12/site-packages/vllm/model_executor/layers/quantization/utils/configs/N=8192,K=512,device_name=NVIDIA_L4,dtype=fp8_w8a8,block_shape=[128,128].json 2025-11-03T17:55:07.7121394Z (EngineCore_DP0 pid=25340) INFO 11-03 17:55:07 [gpu_model_runner.py:2879] Loading drafter model... 2025-11-03T17:55:07.7214720Z (EngineCore_DP0 pid=25340) INFO 11-03 17:55:07 [eagle.py:1005] Assuming the EAGLE head shares the same vocab embedding with the target model. 2025-11-03T17:55:07.7215681Z (EngineCore_DP0 pid=25340) INFO 11-03 17:55:07 [eagle.py:1027] Loading EAGLE LM head weights from the target model. 2025-11-03T17:55:08.2592147Z (EngineCore_DP0 pid=25340) INFO 11-03 17:55:08 [gpu_model_runner.py:2902] Model loading took 2.2133 GiB and 0.670020 seconds 2025-11-03T17:55:08.2596924Z (EngineCore_DP0 pid=25340) INFO 11-03 17:55:08 [kv_cache_utils.py:1199] GPU KV cache size: 4,660,336 tokens 2025-11-03T17:55:08.2598073Z (EngineCore_DP0 pid=25340) INFO 11-03 17:55:08 [kv_cache_utils.py:1204] Maximum concurrency for 163,840 tokens per request: 28.44x 2025-11-03T17:55:09.3969670Z (EngineCore_DP0 pid=25340) INFO 11-03 17:55:09 [gc_utils.py:40] GC Debug Config. enabled:False,top_objects:-1 2025-11-03T17:55:09.3983486Z INFO 11-03 17:55:09 [llm.py:337] Supported tasks: ['generate'] 2025-11-03T17:55:09.8812199Z PASSED 2025-11-03T17:55:09.8952476Z models/test_initialization.py::test_can_initialize_large_subset[Zamba2ForCausalLM] Fork a new process to run a test 25410 2025-11-03T17:55:09.8962853Z Fork a new process to run a test 0 2025-11-03T17:55:09.9244164Z INFO 11-03 17:55:09 [utils.py:239] non-default args: {'load_format': 'dummy', 'gpu_memory_utilization': 0.8, 'disable_log_stats': True, 'hf_overrides': functools.partial(, model_arch='Zamba2ForCausalLM', exist_overrides={}, use_original_num_layers=False), 'model_impl': 'vllm', 'model': 'Zyphra/Zamba2-7B-instruct'} 2025-11-03T17:55:10.0467125Z 2025-11-03T17:55:10.0469483Z config.json: 0.00B [00:00, ?B/s] 2025-11-03T17:55:10.0469938Z config.json: 2.41kB [00:00, 12.6MB/s] 2025-11-03T17:55:16.5480748Z INFO 11-03 17:55:16 [model.py:653] Resolved architecture: Zamba2ForCausalLM 2025-11-03T17:55:16.5481241Z `torch_dtype` is deprecated! Use `dtype` instead! 2025-11-03T17:55:16.7163597Z 2025-11-03T17:55:16.7166812Z model.safetensors.index.json: 0.00B [00:00, ?B/s] 2025-11-03T17:55:16.7167378Z model.safetensors.index.json: 63.3kB [00:00, 170MB/s] 2025-11-03T17:55:16.7203268Z 2025-11-03T17:55:16.8011589Z Parse safetensors files: 0% 0/3 [00:00= mamba page size. 2025-11-03T17:55:17.0866246Z INFO 11-03 17:55:17 [config.py:453] Padding mamba page size by 43.05% to ensure that mamba page size and attention page size are exactly equal. 2025-11-03T17:55:17.1452443Z 2025-11-03T17:55:17.1454078Z tokenizer_config.json: 0.00B [00:00, ?B/s] 2025-11-03T17:55:17.1454510Z tokenizer_config.json: 1.37kB [00:00, 10.1MB/s] 2025-11-03T17:55:18.2512263Z 2025-11-03T17:55:18.2596213Z tokenizer.json: 0.00B [00:00, ?B/s] 2025-11-03T17:55:18.2596679Z tokenizer.json: 1.80MB [00:00, 214MB/s] 2025-11-03T17:55:18.4684253Z 2025-11-03T17:55:18.4685811Z special_tokens_map.json: 0% 0.00/504 [00:00, 'use_cudagraph': True, 'cudagraph_num_of_warmups': 1, 'cudagraph_capture_sizes': [512, 504, 496, 488, 480, 472, 464, 456, 448, 440, 432, 424, 416, 408, 400, 392, 384, 376, 368, 360, 352, 344, 336, 328, 320, 312, 304, 296, 288, 280, 272, 264, 256, 248, 240, 232, 224, 216, 208, 200, 192, 184, 176, 168, 160, 152, 144, 136, 128, 120, 112, 104, 96, 88, 80, 72, 64, 56, 48, 40, 32, 24, 16, 8, 4, 2, 1], 'cudagraph_copy_inputs': False, 'full_cuda_graph': True, 'use_inductor_graph_partition': False, 'pass_config': {}, 'max_capture_size': 512, 'local_cache_dir': None} 2025-11-03T17:55:19.4679244Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:55:19.4707109Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:55:19.4714765Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:55:19.4723616Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:55:19.4731741Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:55:19.4739808Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:55:19.4742870Z (EngineCore_DP0 pid=25471) INFO 11-03 17:55:19 [parallel_state.py:1231] rank 0 in world size 1 is assigned as DP rank 0, PP rank 0, TP rank 0, EP rank 0 2025-11-03T17:55:19.8142204Z (EngineCore_DP0 pid=25471) INFO 11-03 17:55:19 [topk_topp_sampler.py:59] Using FlashInfer for top-p & top-k sampling. 2025-11-03T17:55:19.8787101Z (EngineCore_DP0 pid=25471) INFO 11-03 17:55:19 [gpu_model_runner.py:2840] Starting to load model Zyphra/Zamba2-7B-instruct... 2025-11-03T17:55:20.1397424Z (EngineCore_DP0 pid=25471) INFO 11-03 17:55:20 [gpu_model_runner.py:2870] Loading model from scratch... 2025-11-03T17:55:20.1422794Z (EngineCore_DP0 pid=25471) INFO 11-03 17:55:20 [cuda.py:404] Using Flash Attention backend on V1 engine. 2025-11-03T17:55:20.9017823Z (EngineCore_DP0 pid=25471) INFO 11-03 17:55:20 [gpu_model_runner.py:2902] Model loading took 13.7859 GiB and 0.235309 seconds 2025-11-03T17:55:20.9036951Z (EngineCore_DP0 pid=25471) WARNING 11-03 17:55:20 [kv_cache_utils.py:949] Add 10 padding layers, may waste at most 12.35% KV cache memory 2025-11-03T17:55:20.9044631Z (EngineCore_DP0 pid=25471) INFO 11-03 17:55:20 [kv_cache_utils.py:1199] GPU KV cache size: 3,600 tokens 2025-11-03T17:55:20.9045947Z (EngineCore_DP0 pid=25471) INFO 11-03 17:55:20 [kv_cache_utils.py:1204] Maximum concurrency for 4,096 tokens per request: 6.45x 2025-11-03T17:55:21.4234488Z (EngineCore_DP0 pid=25471) INFO 11-03 17:55:21 [gc_utils.py:40] GC Debug Config. enabled:False,top_objects:-1 2025-11-03T17:55:21.4256632Z INFO 11-03 17:55:21 [llm.py:337] Supported tasks: ['generate'] 2025-11-03T17:55:21.8878098Z PASSED 2025-11-03T17:55:21.9018334Z models/test_initialization.py::test_can_initialize_large_subset[CLIPModel] Fork a new process to run a test 25537 2025-11-03T17:55:21.9029662Z Fork a new process to run a test 0 2025-11-03T17:55:21.9311401Z INFO 11-03 17:55:21 [utils.py:239] non-default args: {'load_format': 'dummy', 'gpu_memory_utilization': 0.8, 'disable_log_stats': True, 'hf_overrides': functools.partial(, model_arch='CLIPModel', exist_overrides={}, use_original_num_layers=False), 'model_impl': 'vllm', 'model': 'openai/clip-vit-base-patch32'} 2025-11-03T17:55:22.0106462Z 2025-11-03T17:55:22.0108410Z config.json: 0.00B [00:00, ?B/s] 2025-11-03T17:55:22.0108730Z config.json: 4.19kB [00:00, 24.3MB/s] 2025-11-03T17:55:22.1101863Z 2025-11-03T17:55:22.1103302Z preprocessor_config.json: 0% 0.00/316 [00:00, 'use_cudagraph': True, 'cudagraph_num_of_warmups': 1, 'cudagraph_capture_sizes': [512, 504, 496, 488, 480, 472, 464, 456, 448, 440, 432, 424, 416, 408, 400, 392, 384, 376, 368, 360, 352, 344, 336, 328, 320, 312, 304, 296, 288, 280, 272, 264, 256, 248, 240, 232, 224, 216, 208, 200, 192, 184, 176, 168, 160, 152, 144, 136, 128, 120, 112, 104, 96, 88, 80, 72, 64, 56, 48, 40, 32, 24, 16, 8, 4, 2, 1], 'cudagraph_copy_inputs': False, 'full_cuda_graph': False, 'use_inductor_graph_partition': False, 'pass_config': {}, 'max_capture_size': 512, 'local_cache_dir': None} 2025-11-03T17:55:32.0724074Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:55:32.0751676Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:55:32.0759759Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:55:32.0767546Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:55:32.0775175Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:55:32.0783452Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:55:32.0786505Z (EngineCore_DP0 pid=25596) INFO 11-03 17:55:32 [parallel_state.py:1231] rank 0 in world size 1 is assigned as DP rank 0, PP rank 0, TP rank 0, EP rank 0 2025-11-03T17:55:32.4124878Z (EngineCore_DP0 pid=25596) INFO 11-03 17:55:32 [topk_topp_sampler.py:59] Using FlashInfer for top-p & top-k sampling. 2025-11-03T17:55:32.4268295Z (EngineCore_DP0 pid=25596) Using a slow image processor as `use_fast` is unset and a slow processor was saved with this model. `use_fast=True` will be the default behavior in v4.52, even if the model was saved with a slow processor. This will result in minor differences in outputs. You'll still be able to use a slow processor with `use_fast=False`. 2025-11-03T17:55:33.1880126Z (EngineCore_DP0 pid=25596) WARNING 11-03 17:55:33 [processing.py:1091] CLIPProcessor did not return `BatchFeature`. Make sure to match the behaviour of `ProcessorMixin` when implementing custom processors. 2025-11-03T17:55:33.2118686Z (EngineCore_DP0 pid=25596) INFO 11-03 17:55:33 [gpu_model_runner.py:2840] Starting to load model openai/clip-vit-base-patch32... 2025-11-03T17:55:33.4643254Z (EngineCore_DP0 pid=25596) INFO 11-03 17:55:33 [gpu_model_runner.py:2870] Loading model from scratch... 2025-11-03T17:55:33.4665748Z (EngineCore_DP0 pid=25596) INFO 11-03 17:55:33 [cuda.py:404] Using Flash Attention backend on V1 engine. 2025-11-03T17:55:33.4721256Z (EngineCore_DP0 pid=25596) INFO 11-03 17:55:33 [layer.py:497] MultiHeadAttention attn_backend: _Backend.FLASH_ATTN, use_upstream_fa: False 2025-11-03T17:55:33.5122536Z (EngineCore_DP0 pid=25596) WARNING 11-03 17:55:33 [vllm.py:821] `torch.compile` is turned on, but the model openai/clip-vit-base-patch32 does not support it. Please open an issue on GitHub if you want it to be supported. 2025-11-03T17:55:34.0149117Z (EngineCore_DP0 pid=25596) INFO 11-03 17:55:34 [gpu_model_runner.py:2902] Model loading took 0.0724 GiB and 0.049921 seconds 2025-11-03T17:55:34.0152972Z (EngineCore_DP0 pid=25596) INFO 11-03 17:55:34 [kv_cache_utils.py:1199] GPU KV cache size: 5,242,880 tokens 2025-11-03T17:55:34.0153751Z (EngineCore_DP0 pid=25596) INFO 11-03 17:55:34 [kv_cache_utils.py:1204] Maximum concurrency for 77 tokens per request: 65536.00x 2025-11-03T17:55:34.8769405Z (EngineCore_DP0 pid=25596) INFO 11-03 17:55:34 [gc_utils.py:40] GC Debug Config. enabled:False,top_objects:-1 2025-11-03T17:55:34.8785694Z INFO 11-03 17:55:34 [llm.py:337] Supported tasks: ['embed'] 2025-11-03T17:55:35.3117951Z PASSED 2025-11-03T17:55:35.3259846Z models/test_initialization.py::test_can_initialize_large_subset[ErnieMTPModel] Fork a new process to run a test 25678 2025-11-03T17:55:35.3270955Z Fork a new process to run a test 0 2025-11-03T17:55:35.3550141Z INFO 11-03 17:55:35 [utils.py:239] non-default args: {'trust_remote_code': True, 'load_format': 'dummy', 'gpu_memory_utilization': 0.8, 'disable_log_stats': True, 'hf_overrides': functools.partial(, model_arch='ErnieMTPModel', exist_overrides={}, use_original_num_layers=False), 'speculative_config': {'model': 'baidu/ERNIE-4.5-21B-A3B-PT', 'num_speculative_tokens': 1}, 'model_impl': 'vllm', 'model': 'baidu/ERNIE-4.5-21B-A3B-PT'} 2025-11-03T17:55:35.5499052Z INFO 11-03 17:55:35 [model.py:653] Resolved architecture: Ernie4_5_MoeForCausalLM 2025-11-03T17:55:35.5499539Z `torch_dtype` is deprecated! Use `dtype` instead! 2025-11-03T17:55:35.5742652Z INFO 11-03 17:55:35 [model.py:1714] Using max model len 131072 2025-11-03T17:55:35.5745089Z The argument `trust_remote_code` is to be used with Auto classes. It has no effect here and is ignored. 2025-11-03T17:55:42.1015283Z INFO 11-03 17:55:42 [model.py:653] Resolved architecture: ErnieMTPModel 2025-11-03T17:55:42.1016125Z INFO 11-03 17:55:42 [model.py:1714] Using max model len 131072 2025-11-03T17:55:42.1019421Z INFO 11-03 17:55:42 [scheduler.py:225] Chunked prefill is enabled with max_num_batched_tokens=8192. 2025-11-03T17:55:43.2542896Z (EngineCore_DP0 pid=25735) INFO 11-03 17:55:43 [core.py:727] Waiting for init message from front-end. 2025-11-03T17:55:43.2581041Z (EngineCore_DP0 pid=25735) INFO 11-03 17:55:43 [core.py:94] Initializing a V1 LLM engine (v0.11.0rc2.dev389+ge51928192.d20251103) with config: model='baidu/ERNIE-4.5-21B-A3B-PT', speculative_config=SpeculativeConfig(method='mtp', model='baidu/ERNIE-4.5-21B-A3B-PT', num_spec_tokens=1), tokenizer='baidu/ERNIE-4.5-21B-A3B-PT', skip_tokenizer_init=False, tokenizer_mode=auto, revision=None, tokenizer_revision=None, trust_remote_code=True, dtype=torch.bfloat16, max_seq_len=131072, download_dir=None, load_format=dummy, tensor_parallel_size=1, pipeline_parallel_size=1, data_parallel_size=1, disable_custom_all_reduce=False, quantization=None, enforce_eager=False, kv_cache_dtype=auto, device_config=cuda, structured_outputs_config=StructuredOutputsConfig(backend='auto', disable_fallback=False, disable_any_whitespace=False, disable_additional_properties=False, reasoning_parser=''), observability_config=ObservabilityConfig(show_hidden_metrics_for_version=None, otlp_traces_endpoint=None, collect_detailed_traces=None), seed=0, served_model_name=baidu/ERNIE-4.5-21B-A3B-PT, enable_prefix_caching=True, chunked_prefill_enabled=True, pooler_config=None, compilation_config={'level': 3, 'debug_dump_path': None, 'cache_dir': '', 'backend': '', 'custom_ops': [], 'splitting_ops': ['vllm.unified_attention', 'vllm.unified_attention_with_output', 'vllm.unified_mla_attention', 'vllm.unified_mla_attention_with_output', 'vllm.mamba_mixer2', 'vllm.mamba_mixer', 'vllm.short_conv', 'vllm.linear_attention', 'vllm.plamo2_mamba_mixer', 'vllm.gdn_attention', 'vllm.sparse_attn_indexer'], 'use_inductor': True, 'compile_sizes': [], 'inductor_compile_config': {'enable_auto_functionalized_v2': False}, 'inductor_passes': {}, 'cudagraph_mode': , 'use_cudagraph': True, 'cudagraph_num_of_warmups': 1, 'cudagraph_capture_sizes': [512, 504, 496, 488, 480, 472, 464, 456, 448, 440, 432, 424, 416, 408, 400, 392, 384, 376, 368, 360, 352, 344, 336, 328, 320, 312, 304, 296, 288, 280, 272, 264, 256, 248, 240, 232, 224, 216, 208, 200, 192, 184, 176, 168, 160, 152, 144, 136, 128, 120, 112, 104, 96, 88, 80, 72, 64, 56, 48, 40, 32, 24, 16, 8, 4, 2, 1], 'cudagraph_copy_inputs': False, 'full_cuda_graph': True, 'use_inductor_graph_partition': False, 'pass_config': {}, 'max_capture_size': 512, 'local_cache_dir': None} 2025-11-03T17:55:43.9260269Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:55:43.9287910Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:55:43.9295851Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:55:43.9303782Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:55:43.9311260Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:55:43.9319425Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:55:43.9323327Z (EngineCore_DP0 pid=25735) INFO 11-03 17:55:43 [parallel_state.py:1231] rank 0 in world size 1 is assigned as DP rank 0, PP rank 0, TP rank 0, EP rank 0 2025-11-03T17:55:44.2667900Z (EngineCore_DP0 pid=25735) INFO 11-03 17:55:44 [topk_topp_sampler.py:59] Using FlashInfer for top-p & top-k sampling. 2025-11-03T17:55:44.2814150Z (EngineCore_DP0 pid=25735) WARNING 11-03 17:55:44 [__init__.py:194] min_p, logit_bias, and min_tokens parameters won't currently work with speculative decoding enabled. 2025-11-03T17:55:44.3343788Z (EngineCore_DP0 pid=25735) INFO 11-03 17:55:44 [gpu_model_runner.py:2840] Starting to load model baidu/ERNIE-4.5-21B-A3B-PT... 2025-11-03T17:55:44.5923125Z (EngineCore_DP0 pid=25735) INFO 11-03 17:55:44 [gpu_model_runner.py:2870] Loading model from scratch... 2025-11-03T17:55:44.6487668Z (EngineCore_DP0 pid=25735) INFO 11-03 17:55:44 [cuda.py:404] Using Flash Attention backend on V1 engine. 2025-11-03T17:55:44.7082913Z (EngineCore_DP0 pid=25735) INFO 11-03 17:55:44 [gpu_model_runner.py:2879] Loading drafter model... 2025-11-03T17:55:44.7160368Z (EngineCore_DP0 pid=25735) WARNING 11-03 17:55:44 [vllm.py:821] `torch.compile` is turned on, but the model baidu/ERNIE-4.5-21B-A3B-PT does not support it. Please open an issue on GitHub if you want it to be supported. 2025-11-03T17:55:44.7167056Z (EngineCore_DP0 pid=25735) INFO 11-03 17:55:44 [eagle.py:1005] Assuming the EAGLE head shares the same vocab embedding with the target model. 2025-11-03T17:55:44.7168085Z (EngineCore_DP0 pid=25735) INFO 11-03 17:55:44 [eagle.py:1027] Loading EAGLE LM head weights from the target model. 2025-11-03T17:55:45.2222402Z (EngineCore_DP0 pid=25735) INFO 11-03 17:55:45 [gpu_model_runner.py:2902] Model loading took 0.9917 GiB and 0.124745 seconds 2025-11-03T17:55:45.2227330Z (EngineCore_DP0 pid=25735) INFO 11-03 17:55:45 [kv_cache_utils.py:1199] GPU KV cache size: 2,621,440 tokens 2025-11-03T17:55:45.2228276Z (EngineCore_DP0 pid=25735) INFO 11-03 17:55:45 [kv_cache_utils.py:1204] Maximum concurrency for 131,072 tokens per request: 20.00x 2025-11-03T17:55:46.9077595Z (EngineCore_DP0 pid=25735) INFO 11-03 17:55:46 [gc_utils.py:40] GC Debug Config. enabled:False,top_objects:-1 2025-11-03T17:55:46.9090950Z INFO 11-03 17:55:46 [llm.py:337] Supported tasks: ['generate'] 2025-11-03T17:55:47.3807599Z PASSED 2025-11-03T17:55:47.3948655Z models/test_initialization.py::test_can_initialize_large_subset[TransformersMoEForCausalLM] Fork a new process to run a test 25801 2025-11-03T17:55:47.3959890Z Fork a new process to run a test 0 2025-11-03T17:55:47.3963775Z `transformers==4.56.2` installed, but `transformers>=4.57.0.dev0` is required to run this model. 2025-11-03T17:55:47.6893432Z PASSED 2025-11-03T17:55:47.7036219Z models/test_initialization.py::test_can_initialize_large_subset[GraniteSpeechForConditionalGeneration] Fork a new process to run a test 25802 2025-11-03T17:55:47.7047706Z Fork a new process to run a test 0 2025-11-03T17:55:47.7318398Z INFO 11-03 17:55:47 [utils.py:239] non-default args: {'load_format': 'dummy', 'gpu_memory_utilization': 0.8, 'disable_log_stats': True, 'hf_overrides': functools.partial(, model_arch='GraniteSpeechForConditionalGeneration', exist_overrides={}, use_original_num_layers=False), 'model_impl': 'vllm', 'model': 'ibm-granite/granite-speech-3.3-2b'} 2025-11-03T17:55:47.8094846Z 2025-11-03T17:55:47.8096872Z config.json: 0.00B [00:00, ?B/s] 2025-11-03T17:55:47.8097172Z config.json: 2.41kB [00:00, 12.9MB/s] 2025-11-03T17:55:47.9125030Z 2025-11-03T17:55:47.9125764Z preprocessor_config.json: 0% 0.00/2.00 [00:00, 'use_cudagraph': True, 'cudagraph_num_of_warmups': 1, 'cudagraph_capture_sizes': [512, 504, 496, 488, 480, 472, 464, 456, 448, 440, 432, 424, 416, 408, 400, 392, 384, 376, 368, 360, 352, 344, 336, 328, 320, 312, 304, 296, 288, 280, 272, 264, 256, 248, 240, 232, 224, 216, 208, 200, 192, 184, 176, 168, 160, 152, 144, 136, 128, 120, 112, 104, 96, 88, 80, 72, 64, 56, 48, 40, 32, 24, 16, 8, 4, 2, 1], 'cudagraph_copy_inputs': False, 'full_cuda_graph': True, 'use_inductor_graph_partition': False, 'pass_config': {}, 'max_capture_size': 512, 'local_cache_dir': None} 2025-11-03T17:55:55.7157433Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:55:55.7186171Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:55:55.7193231Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:55:55.7200760Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:55:55.7207988Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:55:55.7215393Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:55:55.7218123Z (EngineCore_DP0 pid=25860) INFO 11-03 17:55:55 [parallel_state.py:1231] rank 0 in world size 1 is assigned as DP rank 0, PP rank 0, TP rank 0, EP rank 0 2025-11-03T17:55:56.0564408Z (EngineCore_DP0 pid=25860) INFO 11-03 17:55:56 [topk_topp_sampler.py:59] Using FlashInfer for top-p & top-k sampling. 2025-11-03T17:55:57.3642160Z (EngineCore_DP0 pid=25860) INFO 11-03 17:55:57 [gpu_model_runner.py:2840] Starting to load model ibm-granite/granite-speech-3.3-2b... 2025-11-03T17:55:57.6223795Z (EngineCore_DP0 pid=25860) INFO 11-03 17:55:57 [gpu_model_runner.py:2870] Loading model from scratch... 2025-11-03T17:55:57.6702978Z (EngineCore_DP0 pid=25860) INFO 11-03 17:55:57 [cuda.py:404] Using Flash Attention backend on V1 engine. 2025-11-03T17:55:58.2858590Z (EngineCore_DP0 pid=25860) INFO 11-03 17:55:58 [gpu_model_runner.py:2902] Model loading took 0.4359 GiB and 0.141164 seconds 2025-11-03T17:55:58.2862993Z (EngineCore_DP0 pid=25860) INFO 11-03 17:55:58 [kv_cache_utils.py:1199] GPU KV cache size: 5,242,880 tokens 2025-11-03T17:55:58.2863764Z (EngineCore_DP0 pid=25860) INFO 11-03 17:55:58 [kv_cache_utils.py:1204] Maximum concurrency for 131,072 tokens per request: 40.00x 2025-11-03T17:55:59.2953943Z (EngineCore_DP0 pid=25860) INFO 11-03 17:55:59 [gc_utils.py:40] GC Debug Config. enabled:False,top_objects:-1 2025-11-03T17:55:59.2969649Z INFO 11-03 17:55:59 [llm.py:337] Supported tasks: ['generate'] 2025-11-03T17:55:59.7607052Z PASSED 2025-11-03T17:55:59.7750134Z models/test_initialization.py::test_can_initialize_large_subset[Qwen2_5OmniForConditionalGeneration] Fork a new process to run a test 25942 2025-11-03T17:55:59.7760616Z Fork a new process to run a test 0 2025-11-03T17:55:59.8038964Z INFO 11-03 17:55:59 [utils.py:239] non-default args: {'load_format': 'dummy', 'gpu_memory_utilization': 0.8, 'disable_log_stats': True, 'hf_overrides': functools.partial(, model_arch='Qwen2_5OmniForConditionalGeneration', exist_overrides={}, use_original_num_layers=False), 'model_impl': 'vllm', 'model': 'Qwen/Qwen2.5-Omni-7B-AWQ'} 2025-11-03T17:55:59.8744321Z 2025-11-03T17:55:59.8747502Z config.json: 0.00B [00:00, ?B/s] 2025-11-03T17:55:59.8747820Z config.json: 15.4kB [00:00, 51.4MB/s] 2025-11-03T17:55:59.9233634Z Unrecognized keys in `rope_scaling` for 'rope_type'='default': {'mrope_section'} 2025-11-03T17:56:00.0052240Z 2025-11-03T17:56:00.0053630Z preprocessor_config.json: 0% 0.00/667 [00:00, 'use_cudagraph': True, 'cudagraph_num_of_warmups': 1, 'cudagraph_capture_sizes': [512, 504, 496, 488, 480, 472, 464, 456, 448, 440, 432, 424, 416, 408, 400, 392, 384, 376, 368, 360, 352, 344, 336, 328, 320, 312, 304, 296, 288, 280, 272, 264, 256, 248, 240, 232, 224, 216, 208, 200, 192, 184, 176, 168, 160, 152, 144, 136, 128, 120, 112, 104, 96, 88, 80, 72, 64, 56, 48, 40, 32, 24, 16, 8, 4, 2, 1], 'cudagraph_copy_inputs': False, 'full_cuda_graph': True, 'use_inductor_graph_partition': False, 'pass_config': {}, 'max_capture_size': 512, 'local_cache_dir': None} 2025-11-03T17:56:02.8780071Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:56:02.8808759Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:56:02.8815977Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:56:02.8824497Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:56:02.8832199Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:56:02.8839691Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:56:02.8842289Z (EngineCore_DP0 pid=25971) INFO 11-03 17:56:02 [parallel_state.py:1231] rank 0 in world size 1 is assigned as DP rank 0, PP rank 0, TP rank 0, EP rank 0 2025-11-03T17:56:03.2266683Z (EngineCore_DP0 pid=25971) INFO 11-03 17:56:03 [topk_topp_sampler.py:59] Using FlashInfer for top-p & top-k sampling. 2025-11-03T17:56:05.6999078Z (EngineCore_DP0 pid=25971) INFO 11-03 17:56:05 [gpu_model_runner.py:2840] Starting to load model Qwen/Qwen2.5-Omni-7B-AWQ... 2025-11-03T17:56:06.0408701Z (EngineCore_DP0 pid=25971) INFO 11-03 17:56:06 [gpu_model_runner.py:2870] Loading model from scratch... 2025-11-03T17:56:06.0410278Z (EngineCore_DP0 pid=25971) WARNING 11-03 17:56:06 [qwen2_5_omni_thinker.py:865] flash_attn is not available, the model may not yield the exactly same result as the transformers implementation in the audio tower part. 2025-11-03T17:56:06.1624072Z (EngineCore_DP0 pid=25971) INFO 11-03 17:56:06 [cuda.py:404] Using Flash Attention backend on V1 engine. 2025-11-03T17:56:06.7871922Z (EngineCore_DP0 pid=25971) INFO 11-03 17:56:06 [gpu_model_runner.py:2902] Model loading took 4.6412 GiB and 0.220666 seconds 2025-11-03T17:56:06.7875911Z (EngineCore_DP0 pid=25971) INFO 11-03 17:56:06 [kv_cache_utils.py:1199] GPU KV cache size: 5,242,880 tokens 2025-11-03T17:56:06.7876810Z (EngineCore_DP0 pid=25971) INFO 11-03 17:56:06 [kv_cache_utils.py:1204] Maximum concurrency for 32,768 tokens per request: 160.00x 2025-11-03T17:56:07.9468627Z (EngineCore_DP0 pid=25971) INFO 11-03 17:56:07 [gc_utils.py:40] GC Debug Config. enabled:False,top_objects:-1 2025-11-03T17:56:07.9484970Z INFO 11-03 17:56:07 [llm.py:337] Supported tasks: ['generate'] 2025-11-03T17:56:08.4413202Z PASSED 2025-11-03T17:56:08.4553852Z models/test_initialization.py::test_can_initialize_large_subset[RForConditionalGeneration] Fork a new process to run a test 26037 2025-11-03T17:56:08.4564294Z Fork a new process to run a test 0 2025-11-03T17:56:08.4847694Z INFO 11-03 17:56:08 [utils.py:239] non-default args: {'trust_remote_code': True, 'load_format': 'dummy', 'gpu_memory_utilization': 0.8, 'disable_log_stats': True, 'hf_overrides': functools.partial(, model_arch='RForConditionalGeneration', exist_overrides={}, use_original_num_layers=False), 'model_impl': 'vllm', 'model': 'YannQi/R-4B'} 2025-11-03T17:56:08.5587366Z 2025-11-03T17:56:08.5589132Z config.json: 0.00B [00:00, ?B/s] 2025-11-03T17:56:08.5589448Z config.json: 2.19kB [00:00, 13.9MB/s] 2025-11-03T17:56:08.6213537Z 2025-11-03T17:56:08.6215108Z configuration_r.py: 0.00B [00:00, ?B/s] 2025-11-03T17:56:08.6215453Z configuration_r.py: 3.59kB [00:00, 14.4MB/s] 2025-11-03T17:56:08.6311300Z A new version of the following files was downloaded from https://huggingface.co/YannQi/R-4B: 2025-11-03T17:56:08.6311768Z - configuration_r.py 2025-11-03T17:56:08.6312370Z . Make sure to double-check they do not contain any added malicious code. To avoid downloading new versions of the code file, you can pin a revision. 2025-11-03T17:56:08.8814730Z 2025-11-03T17:56:08.8815007Z preprocessor_config.json: 0% 0.00/745 [00:00, 'use_cudagraph': True, 'cudagraph_num_of_warmups': 1, 'cudagraph_capture_sizes': [512, 504, 496, 488, 480, 472, 464, 456, 448, 440, 432, 424, 416, 408, 400, 392, 384, 376, 368, 360, 352, 344, 336, 328, 320, 312, 304, 296, 288, 280, 272, 264, 256, 248, 240, 232, 224, 216, 208, 200, 192, 184, 176, 168, 160, 152, 144, 136, 128, 120, 112, 104, 96, 88, 80, 72, 64, 56, 48, 40, 32, 24, 16, 8, 4, 2, 1], 'cudagraph_copy_inputs': False, 'full_cuda_graph': True, 'use_inductor_graph_partition': False, 'pass_config': {}, 'max_capture_size': 512, 'local_cache_dir': None} 2025-11-03T17:56:25.0643518Z (EngineCore_DP0 pid=26118) /opt/conda/envs/py_3.12/lib/python3.12/site-packages/torch/library.py:356: UserWarning: Warning only once for all operators, other operators may also be overridden. 2025-11-03T17:56:25.0645139Z (EngineCore_DP0 pid=26118) Overriding a previously registered kernel for the same operator and the same dispatch key 2025-11-03T17:56:25.0647304Z (EngineCore_DP0 pid=26118) operator: xformers_python::_fused_allgather_and_linear_impl(Tensor scattered_input, Tensor[] weights, str process_group_name, Tensor(a3!)[] gathered_outputs, SymInt timeout_s, bool _wait, bool _memcpy, Tensor scale_scattered_input, Tensor?[] scales_weights) -> () 2025-11-03T17:56:25.0649322Z (EngineCore_DP0 pid=26118) registered at /opt/conda/envs/py_3.12/lib/python3.12/site-packages/torch/_library/custom_ops.py:922 2025-11-03T17:56:25.0650228Z (EngineCore_DP0 pid=26118) dispatch key: ADInplaceOrView 2025-11-03T17:56:25.0650830Z (EngineCore_DP0 pid=26118) previous kernel: no debug info 2025-11-03T17:56:25.0652756Z (EngineCore_DP0 pid=26118) new kernel: registered at /opt/conda/envs/py_3.12/lib/python3.12/site-packages/torch/_library/custom_ops.py:922 (Triggered internally at /var/lib/jenkins/workspace/aten/src/ATen/core/dispatch/OperatorEntry.cpp:208.) 2025-11-03T17:56:25.0653794Z (EngineCore_DP0 pid=26118) self.m.impl( 2025-11-03T17:56:25.7761831Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:56:25.7802641Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:56:25.7809707Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:56:25.7817257Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:56:25.7824739Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:56:25.7831599Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:56:25.7834251Z (EngineCore_DP0 pid=26118) INFO 11-03 17:56:25 [parallel_state.py:1231] rank 0 in world size 1 is assigned as DP rank 0, PP rank 0, TP rank 0, EP rank 0 2025-11-03T17:56:26.0233884Z (EngineCore_DP0 pid=26118) INFO 11-03 17:56:26 [topk_topp_sampler.py:59] Using FlashInfer for top-p & top-k sampling. 2025-11-03T17:56:26.1342463Z (EngineCore_DP0 pid=26118) 2025-11-03T17:56:26.1342997Z processor_config.json: 0% 0.00/330 [00:00, model_arch='Qwen3ForCausalLM', exist_overrides={}, use_original_num_layers=False), 'model_impl': 'vllm', 'model': 'Qwen/Qwen3-8B'} 2025-11-03T17:56:41.2915060Z INFO 11-03 17:56:41 [model.py:653] Resolved architecture: Qwen3ForCausalLM 2025-11-03T17:56:41.2915543Z `torch_dtype` is deprecated! Use `dtype` instead! 2025-11-03T17:56:41.3159508Z INFO 11-03 17:56:41 [model.py:1714] Using max model len 40960 2025-11-03T17:56:41.4894436Z INFO 11-03 17:56:41 [scheduler.py:225] Chunked prefill is enabled with max_num_batched_tokens=8192. 2025-11-03T17:56:42.0085333Z (EngineCore_DP0 pid=26493) INFO 11-03 17:56:42 [core.py:727] Waiting for init message from front-end. 2025-11-03T17:56:42.0115421Z (EngineCore_DP0 pid=26493) INFO 11-03 17:56:42 [core.py:94] Initializing a V1 LLM engine (v0.11.0rc2.dev389+ge51928192.d20251103) with config: model='Qwen/Qwen3-8B', speculative_config=None, tokenizer='Qwen/Qwen3-8B', skip_tokenizer_init=False, tokenizer_mode=auto, revision=None, tokenizer_revision=None, trust_remote_code=False, dtype=torch.bfloat16, max_seq_len=40960, download_dir=None, load_format=dummy, tensor_parallel_size=1, pipeline_parallel_size=1, data_parallel_size=1, disable_custom_all_reduce=False, quantization=None, enforce_eager=False, kv_cache_dtype=auto, device_config=cuda, structured_outputs_config=StructuredOutputsConfig(backend='auto', disable_fallback=False, disable_any_whitespace=False, disable_additional_properties=False, reasoning_parser=''), observability_config=ObservabilityConfig(show_hidden_metrics_for_version=None, otlp_traces_endpoint=None, collect_detailed_traces=None), seed=0, served_model_name=Qwen/Qwen3-8B, enable_prefix_caching=True, chunked_prefill_enabled=True, pooler_config=None, compilation_config={'level': 3, 'debug_dump_path': None, 'cache_dir': '', 'backend': '', 'custom_ops': [], 'splitting_ops': ['vllm.unified_attention', 'vllm.unified_attention_with_output', 'vllm.unified_mla_attention', 'vllm.unified_mla_attention_with_output', 'vllm.mamba_mixer2', 'vllm.mamba_mixer', 'vllm.short_conv', 'vllm.linear_attention', 'vllm.plamo2_mamba_mixer', 'vllm.gdn_attention', 'vllm.sparse_attn_indexer'], 'use_inductor': True, 'compile_sizes': [], 'inductor_compile_config': {'enable_auto_functionalized_v2': False}, 'inductor_passes': {}, 'cudagraph_mode': , 'use_cudagraph': True, 'cudagraph_num_of_warmups': 1, 'cudagraph_capture_sizes': [512, 504, 496, 488, 480, 472, 464, 456, 448, 440, 432, 424, 416, 408, 400, 392, 384, 376, 368, 360, 352, 344, 336, 328, 320, 312, 304, 296, 288, 280, 272, 264, 256, 248, 240, 232, 224, 216, 208, 200, 192, 184, 176, 168, 160, 152, 144, 136, 128, 120, 112, 104, 96, 88, 80, 72, 64, 56, 48, 40, 32, 24, 16, 8, 4, 2, 1], 'cudagraph_copy_inputs': False, 'full_cuda_graph': True, 'use_inductor_graph_partition': False, 'pass_config': {}, 'max_capture_size': 512, 'local_cache_dir': None} 2025-11-03T17:56:42.6906016Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:56:42.6933645Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:56:42.6941202Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:56:42.6949330Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:56:42.6956470Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:56:42.6964154Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:56:42.6966709Z (EngineCore_DP0 pid=26493) INFO 11-03 17:56:42 [parallel_state.py:1231] rank 0 in world size 1 is assigned as DP rank 0, PP rank 0, TP rank 0, EP rank 0 2025-11-03T17:56:43.0312207Z (EngineCore_DP0 pid=26493) INFO 11-03 17:56:43 [topk_topp_sampler.py:59] Using FlashInfer for top-p & top-k sampling. 2025-11-03T17:56:43.1114811Z (EngineCore_DP0 pid=26493) INFO 11-03 17:56:43 [gpu_model_runner.py:2840] Starting to load model Qwen/Qwen3-8B... 2025-11-03T17:56:43.3670993Z (EngineCore_DP0 pid=26493) INFO 11-03 17:56:43 [gpu_model_runner.py:2870] Loading model from scratch... 2025-11-03T17:56:43.4355967Z (EngineCore_DP0 pid=26493) INFO 11-03 17:56:43 [cuda.py:404] Using Flash Attention backend on V1 engine. 2025-11-03T17:56:44.0075649Z (EngineCore_DP0 pid=26493) INFO 11-03 17:56:44 [gpu_model_runner.py:2902] Model loading took 2.6895 GiB and 0.128970 seconds 2025-11-03T17:56:44.0080751Z (EngineCore_DP0 pid=26493) INFO 11-03 17:56:44 [kv_cache_utils.py:1199] GPU KV cache size: 2,621,440 tokens 2025-11-03T17:56:44.0081702Z (EngineCore_DP0 pid=26493) INFO 11-03 17:56:44 [kv_cache_utils.py:1204] Maximum concurrency for 40,960 tokens per request: 64.00x 2025-11-03T17:56:44.8035188Z (EngineCore_DP0 pid=26493) INFO 11-03 17:56:44 [gc_utils.py:40] GC Debug Config. enabled:False,top_objects:-1 2025-11-03T17:56:44.8052285Z INFO 11-03 17:56:44 [llm.py:337] Supported tasks: ['generate'] 2025-11-03T17:56:45.2712463Z PASSED 2025-11-03T17:56:45.2855570Z models/test_initialization.py::test_can_initialize_large_subset[NomicBertModel] Fork a new process to run a test 26559 2025-11-03T17:56:45.2867538Z Fork a new process to run a test 0 2025-11-03T17:56:45.3148434Z INFO 11-03 17:56:45 [utils.py:239] non-default args: {'trust_remote_code': True, 'load_format': 'dummy', 'gpu_memory_utilization': 0.8, 'disable_log_stats': True, 'hf_overrides': functools.partial(, model_arch='NomicBertModel', exist_overrides={}, use_original_num_layers=False), 'model_impl': 'vllm', 'model': 'nomic-ai/nomic-embed-text-v2-moe'} 2025-11-03T17:56:45.3972864Z 2025-11-03T17:56:45.3975221Z config.json: 0.00B [00:00, ?B/s] 2025-11-03T17:56:45.3975657Z config.json: 2.48kB [00:00, 11.3MB/s] 2025-11-03T17:56:45.4656473Z 2025-11-03T17:56:45.4658350Z configuration_hf_nomic_bert.py: 0.00B [00:00, ?B/s] 2025-11-03T17:56:45.4658768Z configuration_hf_nomic_bert.py: 1.96kB [00:00, 12.6MB/s] 2025-11-03T17:56:45.4797996Z A new version of the following files was downloaded from https://huggingface.co/nomic-ai/nomic-bert-2048: 2025-11-03T17:56:45.4798554Z - configuration_hf_nomic_bert.py 2025-11-03T17:56:45.7096571Z . Make sure to double-check they do not contain any added malicious code. To avoid downloading new versions of the code file, you can pin a revision. 2025-11-03T17:56:45.7097168Z 2025-11-03T17:56:45.7097317Z sentence_bert_config.json: 0% 0.00/53.0 [00:00, 'use_cudagraph': True, 'cudagraph_num_of_warmups': 1, 'cudagraph_capture_sizes': [512, 504, 496, 488, 480, 472, 464, 456, 448, 440, 432, 424, 416, 408, 400, 392, 384, 376, 368, 360, 352, 344, 336, 328, 320, 312, 304, 296, 288, 280, 272, 264, 256, 248, 240, 232, 224, 216, 208, 200, 192, 184, 176, 168, 160, 152, 144, 136, 128, 120, 112, 104, 96, 88, 80, 72, 64, 56, 48, 40, 32, 24, 16, 8, 4, 2, 1], 'cudagraph_copy_inputs': False, 'full_cuda_graph': False, 'use_inductor_graph_partition': False, 'pass_config': {}, 'max_capture_size': 512, 'local_cache_dir': None} 2025-11-03T17:56:54.5178918Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:56:54.5206896Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:56:54.5214100Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:56:54.5222468Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:56:54.5230862Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:56:54.5238495Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:56:54.5241566Z (EngineCore_DP0 pid=26640) INFO 11-03 17:56:54 [parallel_state.py:1231] rank 0 in world size 1 is assigned as DP rank 0, PP rank 0, TP rank 0, EP rank 0 2025-11-03T17:56:54.8696129Z (EngineCore_DP0 pid=26640) INFO 11-03 17:56:54 [topk_topp_sampler.py:59] Using FlashInfer for top-p & top-k sampling. 2025-11-03T17:56:54.8851985Z (EngineCore_DP0 pid=26640) INFO 11-03 17:56:54 [gpu_model_runner.py:2840] Starting to load model nomic-ai/nomic-embed-text-v2-moe... 2025-11-03T17:56:55.1429650Z (EngineCore_DP0 pid=26640) INFO 11-03 17:56:55 [gpu_model_runner.py:2870] Loading model from scratch... 2025-11-03T17:56:55.2080648Z (EngineCore_DP0 pid=26640) INFO 11-03 17:56:55 [cuda.py:404] Using Flash Attention backend on V1 engine. 2025-11-03T17:56:55.7909256Z (EngineCore_DP0 pid=26640) INFO 11-03 17:56:55 [gpu_model_runner.py:2902] Model loading took 0.3718 GiB and 0.128326 seconds 2025-11-03T17:56:56.5669034Z (EngineCore_DP0 pid=26640) INFO 11-03 17:56:56 [core.py:142] Disabling chunked prefill for model without KVCache 2025-11-03T17:56:56.5725233Z (EngineCore_DP0 pid=26640) INFO 11-03 17:56:56 [vllm.py:433] Only "last" pooling supports chunked prefill and prefix caching; disabling both. 2025-11-03T17:56:56.8345151Z (EngineCore_DP0 pid=26640) INFO 11-03 17:56:56 [gc_utils.py:40] GC Debug Config. enabled:False,top_objects:-1 2025-11-03T17:56:56.8361395Z INFO 11-03 17:56:56 [llm.py:337] Supported tasks: ['embed', 'encode'] 2025-11-03T17:56:57.3251818Z PASSED 2025-11-03T17:56:57.3394600Z models/test_initialization.py::test_can_initialize_large_subset[Qwen2VLForConditionalGeneration] Fork a new process to run a test 26706 2025-11-03T17:56:57.3405817Z Fork a new process to run a test 0 2025-11-03T17:56:57.3687678Z INFO 11-03 17:56:57 [utils.py:239] non-default args: {'load_format': 'dummy', 'gpu_memory_utilization': 0.8, 'disable_log_stats': True, 'hf_overrides': functools.partial(, model_arch='Qwen2VLForConditionalGeneration', exist_overrides={}, use_original_num_layers=False), 'model_impl': 'vllm', 'model': 'Qwen/Qwen2-VL-2B-Instruct'} 2025-11-03T17:56:57.4944580Z 2025-11-03T17:56:57.4946783Z config.json: 0.00B [00:00, ?B/s] 2025-11-03T17:56:57.4947107Z config.json: 1.20kB [00:00, 6.02MB/s] 2025-11-03T17:56:57.5934119Z 2025-11-03T17:56:57.5934762Z preprocessor_config.json: 0% 0.00/347 [00:00, 'use_cudagraph': True, 'cudagraph_num_of_warmups': 1, 'cudagraph_capture_sizes': [512, 504, 496, 488, 480, 472, 464, 456, 448, 440, 432, 424, 416, 408, 400, 392, 384, 376, 368, 360, 352, 344, 336, 328, 320, 312, 304, 296, 288, 280, 272, 264, 256, 248, 240, 232, 224, 216, 208, 200, 192, 184, 176, 168, 160, 152, 144, 136, 128, 120, 112, 104, 96, 88, 80, 72, 64, 56, 48, 40, 32, 24, 16, 8, 4, 2, 1], 'cudagraph_copy_inputs': False, 'full_cuda_graph': True, 'use_inductor_graph_partition': False, 'pass_config': {}, 'max_capture_size': 512, 'local_cache_dir': None} 2025-11-03T17:56:59.5845775Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:56:59.5875061Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:56:59.5882438Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:56:59.5890629Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:56:59.5898595Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:56:59.5906733Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:56:59.5909690Z (EngineCore_DP0 pid=26714) INFO 11-03 17:56:59 [parallel_state.py:1231] rank 0 in world size 1 is assigned as DP rank 0, PP rank 0, TP rank 0, EP rank 0 2025-11-03T17:56:59.9291773Z (EngineCore_DP0 pid=26714) INFO 11-03 17:56:59 [topk_topp_sampler.py:59] Using FlashInfer for top-p & top-k sampling. 2025-11-03T17:57:01.6071836Z (EngineCore_DP0 pid=26714) 2025-11-03T17:57:01.6073735Z chat_template.json: 0.00B [00:00, ?B/s] 2025-11-03T17:57:01.6074247Z chat_template.json: 1.05kB [00:00, 3.25MB/s] 2025-11-03T17:57:02.1490767Z (EngineCore_DP0 pid=26714) INFO 11-03 17:57:02 [gpu_model_runner.py:2840] Starting to load model Qwen/Qwen2-VL-2B-Instruct... 2025-11-03T17:57:02.4843178Z (EngineCore_DP0 pid=26714) INFO 11-03 17:57:02 [gpu_model_runner.py:2870] Loading model from scratch... 2025-11-03T17:57:02.5536450Z (EngineCore_DP0 pid=26714) INFO 11-03 17:57:02 [cuda.py:404] Using Flash Attention backend on V1 engine. 2025-11-03T17:57:03.1310688Z (EngineCore_DP0 pid=26714) INFO 11-03 17:57:03 [gpu_model_runner.py:2902] Model loading took 1.7941 GiB and 0.135002 seconds 2025-11-03T17:57:03.1315419Z (EngineCore_DP0 pid=26714) INFO 11-03 17:57:03 [kv_cache_utils.py:1199] GPU KV cache size: 10,485,760 tokens 2025-11-03T17:57:03.1316546Z (EngineCore_DP0 pid=26714) INFO 11-03 17:57:03 [kv_cache_utils.py:1204] Maximum concurrency for 32,768 tokens per request: 320.00x 2025-11-03T17:57:04.6227716Z (EngineCore_DP0 pid=26714) INFO 11-03 17:57:04 [gc_utils.py:40] GC Debug Config. enabled:False,top_objects:-1 2025-11-03T17:57:04.6241596Z INFO 11-03 17:57:04 [llm.py:337] Supported tasks: ['generate'] 2025-11-03T17:57:05.1030846Z PASSED 2025-11-03T17:57:05.1173055Z models/test_initialization.py::test_can_initialize_large_subset[JambaForSequenceClassification] Fork a new process to run a test 26780 2025-11-03T17:57:05.1184351Z Fork a new process to run a test 0 2025-11-03T17:57:05.1453237Z INFO 11-03 17:57:05 [utils.py:239] non-default args: {'load_format': 'dummy', 'gpu_memory_utilization': 0.8, 'disable_log_stats': True, 'hf_overrides': functools.partial(, model_arch='JambaForSequenceClassification', exist_overrides={}, use_original_num_layers=False), 'model_impl': 'vllm', 'model': 'ai21labs/Jamba-tiny-reward-dev'} 2025-11-03T17:57:05.2308861Z 2025-11-03T17:57:05.2311014Z config.json: 0.00B [00:00, ?B/s] 2025-11-03T17:57:05.2311381Z config.json: 1.14kB [00:00, 6.32MB/s] 2025-11-03T17:57:11.6524628Z INFO 11-03 17:57:11 [model.py:915] Resolved `--runner auto` to `--runner pooling`. Pass the value explicitly to silence this message. 2025-11-03T17:57:11.6525582Z INFO 11-03 17:57:11 [model.py:653] Resolved architecture: JambaForSequenceClassification 2025-11-03T17:57:11.6526210Z `torch_dtype` is deprecated! Use `dtype` instead! 2025-11-03T17:57:11.6774510Z INFO 11-03 17:57:11 [model.py:1714] Using max model len 262144 2025-11-03T17:57:11.7073493Z INFO 11-03 17:57:11 [arg_utils.py:1725] (Enabling) chunked prefill by default 2025-11-03T17:57:11.7074185Z INFO 11-03 17:57:11 [arg_utils.py:1728] (Enabling) prefix caching by default 2025-11-03T17:57:11.8480347Z INFO 11-03 17:57:11 [scheduler.py:225] Chunked prefill is enabled with max_num_batched_tokens=8192. 2025-11-03T17:57:11.8481344Z INFO 11-03 17:57:11 [config.py:306] Hybrid or mamba-based model detected without support for prefix caching: disabling. 2025-11-03T17:57:11.8482291Z INFO 11-03 17:57:11 [config.py:313] Disabling cascade attention since it is not supported for hybrid models. 2025-11-03T17:57:11.9019866Z INFO 11-03 17:57:11 [config.py:429] Setting attention block size to 80 tokens to ensure that attention page size is >= mamba page size. 2025-11-03T17:57:11.9023981Z INFO 11-03 17:57:11 [config.py:453] Padding mamba page size by 5.26% to ensure that mamba page size and attention page size are exactly equal. 2025-11-03T17:57:11.9386131Z 2025-11-03T17:57:11.9388585Z tokenizer_config.json: 0.00B [00:00, ?B/s] 2025-11-03T17:57:11.9388966Z tokenizer_config.json: 14.3kB [00:00, 61.4MB/s] 2025-11-03T17:57:12.0622579Z 2025-11-03T17:57:12.3188366Z tokenizer.model: 0% 0.00/1.12M [00:00, 'use_cudagraph': True, 'cudagraph_num_of_warmups': 1, 'cudagraph_capture_sizes': [512, 504, 496, 488, 480, 472, 464, 456, 448, 440, 432, 424, 416, 408, 400, 392, 384, 376, 368, 360, 352, 344, 336, 328, 320, 312, 304, 296, 288, 280, 272, 264, 256, 248, 240, 232, 224, 216, 208, 200, 192, 184, 176, 168, 160, 152, 144, 136, 128, 120, 112, 104, 96, 88, 80, 72, 64, 56, 48, 40, 32, 24, 16, 8, 4, 2, 1], 'cudagraph_copy_inputs': False, 'full_cuda_graph': False, 'use_inductor_graph_partition': False, 'pass_config': {}, 'max_capture_size': 512, 'local_cache_dir': None} 2025-11-03T17:57:13.3752885Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:57:13.3780743Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:57:13.3788600Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:57:13.3796592Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:57:13.3804717Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:57:13.3812429Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:57:13.3815092Z (EngineCore_DP0 pid=26859) INFO 11-03 17:57:13 [parallel_state.py:1231] rank 0 in world size 1 is assigned as DP rank 0, PP rank 0, TP rank 0, EP rank 0 2025-11-03T17:57:13.7260755Z (EngineCore_DP0 pid=26859) INFO 11-03 17:57:13 [topk_topp_sampler.py:59] Using FlashInfer for top-p & top-k sampling. 2025-11-03T17:57:13.7629458Z (EngineCore_DP0 pid=26859) INFO 11-03 17:57:13 [gpu_model_runner.py:2840] Starting to load model ai21labs/Jamba-tiny-reward-dev... 2025-11-03T17:57:14.0317535Z (EngineCore_DP0 pid=26859) INFO 11-03 17:57:14 [gpu_model_runner.py:2870] Loading model from scratch... 2025-11-03T17:57:14.6345652Z (EngineCore_DP0 pid=26859) INFO 11-03 17:57:14 [gpu_model_runner.py:2902] Model loading took 0.1348 GiB and 0.063817 seconds 2025-11-03T17:57:14.6350596Z (EngineCore_DP0 pid=26859) INFO 11-03 17:57:14 [kv_cache_utils.py:1199] GPU KV cache size: 68,719,476,736 tokens 2025-11-03T17:57:14.6351890Z (EngineCore_DP0 pid=26859) INFO 11-03 17:57:14 [kv_cache_utils.py:1204] Maximum concurrency for 262,144 tokens per request: 262144.00x 2025-11-03T17:57:15.6501485Z (EngineCore_DP0 pid=26859) INFO 11-03 17:57:15 [gc_utils.py:40] GC Debug Config. enabled:False,top_objects:-1 2025-11-03T17:57:15.6519258Z INFO 11-03 17:57:15 [llm.py:337] Supported tasks: ['classify'] 2025-11-03T17:57:16.1145439Z PASSED 2025-11-03T17:57:16.1287556Z models/test_initialization.py::test_can_initialize_large_subset[Glm4MoeMTPModel] Fork a new process to run a test 26925 2025-11-03T17:57:16.1298725Z Fork a new process to run a test 0 2025-11-03T17:57:16.1300471Z Model is not available online 2025-11-03T17:57:16.4170672Z PASSED 2025-11-03T17:57:16.4315095Z models/test_initialization.py::test_can_initialize_large_subset[MambaForCausalLM] Fork a new process to run a test 26926 2025-11-03T17:57:16.4325207Z Fork a new process to run a test 0 2025-11-03T17:57:16.4602334Z INFO 11-03 17:57:16 [utils.py:239] non-default args: {'load_format': 'dummy', 'gpu_memory_utilization': 0.8, 'disable_log_stats': True, 'hf_overrides': functools.partial(, model_arch='MambaForCausalLM', exist_overrides={}, use_original_num_layers=False), 'model_impl': 'vllm', 'model': 'state-spaces/mamba-130m-hf'} 2025-11-03T17:57:16.8177178Z 2025-11-03T17:57:16.8178042Z config.json: 0% 0.00/895 [00:00, 'use_cudagraph': True, 'cudagraph_num_of_warmups': 1, 'cudagraph_capture_sizes': [512, 504, 496, 488, 480, 472, 464, 456, 448, 440, 432, 424, 416, 408, 400, 392, 384, 376, 368, 360, 352, 344, 336, 328, 320, 312, 304, 296, 288, 280, 272, 264, 256, 248, 240, 232, 224, 216, 208, 200, 192, 184, 176, 168, 160, 152, 144, 136, 128, 120, 112, 104, 96, 88, 80, 72, 64, 56, 48, 40, 32, 24, 16, 8, 4, 2, 1], 'cudagraph_copy_inputs': False, 'full_cuda_graph': True, 'use_inductor_graph_partition': False, 'pass_config': {}, 'max_capture_size': 512, 'local_cache_dir': None} 2025-11-03T17:57:18.2320062Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:57:18.2348261Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:57:18.2355657Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:57:18.2363394Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:57:18.2371109Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:57:18.2379254Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:57:18.2382211Z (EngineCore_DP0 pid=26934) INFO 11-03 17:57:18 [parallel_state.py:1231] rank 0 in world size 1 is assigned as DP rank 0, PP rank 0, TP rank 0, EP rank 0 2025-11-03T17:57:18.5813146Z (EngineCore_DP0 pid=26934) INFO 11-03 17:57:18 [topk_topp_sampler.py:59] Using FlashInfer for top-p & top-k sampling. 2025-11-03T17:57:18.7506367Z (EngineCore_DP0 pid=26934) INFO 11-03 17:57:18 [gpu_model_runner.py:2840] Starting to load model state-spaces/mamba-130m-hf... 2025-11-03T17:57:19.0098781Z (EngineCore_DP0 pid=26934) INFO 11-03 17:57:19 [gpu_model_runner.py:2870] Loading model from scratch... 2025-11-03T17:57:19.5988999Z (EngineCore_DP0 pid=26934) INFO 11-03 17:57:19 [gpu_model_runner.py:2902] Model loading took 0.0793 GiB and 0.060799 seconds 2025-11-03T17:57:19.5994214Z (EngineCore_DP0 pid=26934) INFO 11-03 17:57:19 [kv_cache_utils.py:1199] GPU KV cache size: 376,750,080 tokens 2025-11-03T17:57:19.5995338Z (EngineCore_DP0 pid=26934) INFO 11-03 17:57:19 [kv_cache_utils.py:1204] Maximum concurrency for 2,048 tokens per request: 183960.00x 2025-11-03T17:57:20.1050193Z (EngineCore_DP0 pid=26934) INFO 11-03 17:57:20 [gc_utils.py:40] GC Debug Config. enabled:False,top_objects:-1 2025-11-03T17:57:20.1063844Z INFO 11-03 17:57:20 [llm.py:337] Supported tasks: ['generate'] 2025-11-03T17:57:20.5315890Z PASSED 2025-11-03T17:57:20.5461473Z models/test_initialization.py::test_can_initialize_large_subset[InternLMForCausalLM] Fork a new process to run a test 27000 2025-11-03T17:57:20.5472822Z Fork a new process to run a test 0 2025-11-03T17:57:20.5744503Z INFO 11-03 17:57:20 [utils.py:239] non-default args: {'trust_remote_code': True, 'load_format': 'dummy', 'gpu_memory_utilization': 0.8, 'disable_log_stats': True, 'hf_overrides': functools.partial(, model_arch='InternLMForCausalLM', exist_overrides={}, use_original_num_layers=False), 'model_impl': 'vllm', 'model': 'internlm/internlm-chat-7b'} 2025-11-03T17:57:20.6637782Z 2025-11-03T17:57:20.6639463Z config.json: 0% 0.00/731 [00:00, 'use_cudagraph': True, 'cudagraph_num_of_warmups': 1, 'cudagraph_capture_sizes': [512, 504, 496, 488, 480, 472, 464, 456, 448, 440, 432, 424, 416, 408, 400, 392, 384, 376, 368, 360, 352, 344, 336, 328, 320, 312, 304, 296, 288, 280, 272, 264, 256, 248, 240, 232, 224, 216, 208, 200, 192, 184, 176, 168, 160, 152, 144, 136, 128, 120, 112, 104, 96, 88, 80, 72, 64, 56, 48, 40, 32, 24, 16, 8, 4, 2, 1], 'cudagraph_copy_inputs': False, 'full_cuda_graph': True, 'use_inductor_graph_partition': False, 'pass_config': {}, 'max_capture_size': 512, 'local_cache_dir': None} 2025-11-03T17:57:22.8109100Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:57:22.8136977Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:57:22.8144640Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:57:22.8152106Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:57:22.8159908Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:57:22.8167535Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:57:22.8170335Z (EngineCore_DP0 pid=27029) INFO 11-03 17:57:22 [parallel_state.py:1231] rank 0 in world size 1 is assigned as DP rank 0, PP rank 0, TP rank 0, EP rank 0 2025-11-03T17:57:23.1606920Z (EngineCore_DP0 pid=27029) INFO 11-03 17:57:23 [topk_topp_sampler.py:59] Using FlashInfer for top-p & top-k sampling. 2025-11-03T17:57:23.2278467Z (EngineCore_DP0 pid=27029) INFO 11-03 17:57:23 [gpu_model_runner.py:2840] Starting to load model internlm/internlm-chat-7b... 2025-11-03T17:57:23.4914546Z (EngineCore_DP0 pid=27029) INFO 11-03 17:57:23 [gpu_model_runner.py:2870] Loading model from scratch... 2025-11-03T17:57:23.5566411Z (EngineCore_DP0 pid=27029) INFO 11-03 17:57:23 [cuda.py:404] Using Flash Attention backend on V1 engine. 2025-11-03T17:57:24.1531947Z (EngineCore_DP0 pid=27029) INFO 11-03 17:57:24 [gpu_model_runner.py:2902] Model loading took 1.9517 GiB and 0.125273 seconds 2025-11-03T17:57:24.1535970Z (EngineCore_DP0 pid=27029) INFO 11-03 17:57:24 [kv_cache_utils.py:1199] GPU KV cache size: 655,360 tokens 2025-11-03T17:57:24.1536688Z (EngineCore_DP0 pid=27029) INFO 11-03 17:57:24 [kv_cache_utils.py:1204] Maximum concurrency for 2,048 tokens per request: 320.00x 2025-11-03T17:57:24.4548935Z (EngineCore_DP0 pid=27029) WARNING 11-03 17:57:24 [tokenizer.py:255] Using a slow tokenizer. This might cause a significant slowdown. Consider using a fast tokenizer instead. 2025-11-03T17:57:24.9655582Z (EngineCore_DP0 pid=27029) INFO 11-03 17:57:24 [gc_utils.py:40] GC Debug Config. enabled:False,top_objects:-1 2025-11-03T17:57:24.9671087Z INFO 11-03 17:57:24 [llm.py:337] Supported tasks: ['generate'] 2025-11-03T17:57:25.4197240Z PASSED 2025-11-03T17:57:25.4339634Z models/test_initialization.py::test_can_initialize_large_subset[XLMRobertaForSequenceClassification] Fork a new process to run a test 27095 2025-11-03T17:57:25.4350535Z Fork a new process to run a test 0 2025-11-03T17:57:25.4625969Z INFO 11-03 17:57:25 [utils.py:239] non-default args: {'load_format': 'dummy', 'gpu_memory_utilization': 0.8, 'disable_log_stats': True, 'hf_overrides': functools.partial(, model_arch='XLMRobertaForSequenceClassification', exist_overrides={}, use_original_num_layers=False), 'model_impl': 'vllm', 'model': 'BAAI/bge-reranker-v2-m3'} 2025-11-03T17:57:25.5356876Z 2025-11-03T17:57:25.5358392Z config.json: 0% 0.00/795 [00:00, 'use_cudagraph': True, 'cudagraph_num_of_warmups': 1, 'cudagraph_capture_sizes': [512, 504, 496, 488, 480, 472, 464, 456, 448, 440, 432, 424, 416, 408, 400, 392, 384, 376, 368, 360, 352, 344, 336, 328, 320, 312, 304, 296, 288, 280, 272, 264, 256, 248, 240, 232, 224, 216, 208, 200, 192, 184, 176, 168, 160, 152, 144, 136, 128, 120, 112, 104, 96, 88, 80, 72, 64, 56, 48, 40, 32, 24, 16, 8, 4, 2, 1], 'cudagraph_copy_inputs': False, 'full_cuda_graph': False, 'use_inductor_graph_partition': False, 'pass_config': {}, 'max_capture_size': 512, 'local_cache_dir': None} 2025-11-03T17:57:27.9690095Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:57:27.9721028Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:57:27.9728933Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:57:27.9737209Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:57:27.9745768Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:57:27.9753409Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:57:27.9756458Z (EngineCore_DP0 pid=27124) INFO 11-03 17:57:27 [parallel_state.py:1231] rank 0 in world size 1 is assigned as DP rank 0, PP rank 0, TP rank 0, EP rank 0 2025-11-03T17:57:28.3108398Z (EngineCore_DP0 pid=27124) INFO 11-03 17:57:28 [topk_topp_sampler.py:59] Using FlashInfer for top-p & top-k sampling. 2025-11-03T17:57:28.3371850Z (EngineCore_DP0 pid=27124) INFO 11-03 17:57:28 [gpu_model_runner.py:2840] Starting to load model BAAI/bge-reranker-v2-m3... 2025-11-03T17:57:28.5946556Z (EngineCore_DP0 pid=27124) INFO 11-03 17:57:28 [gpu_model_runner.py:2870] Loading model from scratch... 2025-11-03T17:57:28.6003550Z (EngineCore_DP0 pid=27124) INFO 11-03 17:57:28 [cuda.py:404] Using Flash Attention backend on V1 engine. 2025-11-03T17:57:29.1861151Z (EngineCore_DP0 pid=27124) INFO 11-03 17:57:29 [gpu_model_runner.py:2902] Model loading took 0.5200 GiB and 0.066685 seconds 2025-11-03T17:57:29.9764516Z (EngineCore_DP0 pid=27124) INFO 11-03 17:57:29 [core.py:142] Disabling chunked prefill for model without KVCache 2025-11-03T17:57:30.0631543Z (EngineCore_DP0 pid=27124) INFO 11-03 17:57:30 [vllm.py:433] Only "last" pooling supports chunked prefill and prefix caching; disabling both. 2025-11-03T17:57:30.3245196Z (EngineCore_DP0 pid=27124) INFO 11-03 17:57:30 [gc_utils.py:40] GC Debug Config. enabled:False,top_objects:-1 2025-11-03T17:57:30.3259834Z INFO 11-03 17:57:30 [llm.py:337] Supported tasks: ['score', 'classify', 'encode'] 2025-11-03T17:57:30.8035998Z PASSED 2025-11-03T17:57:30.8176422Z models/test_initialization.py::test_can_initialize_large_subset[Phi3ForCausalLM] Fork a new process to run a test 27190 2025-11-03T17:57:30.8186922Z Fork a new process to run a test 0 2025-11-03T17:57:30.8464666Z INFO 11-03 17:57:30 [utils.py:239] non-default args: {'load_format': 'dummy', 'gpu_memory_utilization': 0.8, 'disable_log_stats': True, 'hf_overrides': functools.partial(, model_arch='Phi3ForCausalLM', exist_overrides={}, use_original_num_layers=False), 'model_impl': 'vllm', 'model': 'microsoft/Phi-3-mini-4k-instruct'} 2025-11-03T17:57:30.9173149Z 2025-11-03T17:57:30.9174410Z config.json: 0% 0.00/967 [00:00, 'use_cudagraph': True, 'cudagraph_num_of_warmups': 1, 'cudagraph_capture_sizes': [512, 504, 496, 488, 480, 472, 464, 456, 448, 440, 432, 424, 416, 408, 400, 392, 384, 376, 368, 360, 352, 344, 336, 328, 320, 312, 304, 296, 288, 280, 272, 264, 256, 248, 240, 232, 224, 216, 208, 200, 192, 184, 176, 168, 160, 152, 144, 136, 128, 120, 112, 104, 96, 88, 80, 72, 64, 56, 48, 40, 32, 24, 16, 8, 4, 2, 1], 'cudagraph_copy_inputs': False, 'full_cuda_graph': True, 'use_inductor_graph_partition': False, 'pass_config': {}, 'max_capture_size': 512, 'local_cache_dir': None} 2025-11-03T17:57:39.0115715Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:57:39.0142826Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:57:39.0149951Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:57:39.0157581Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:57:39.0165330Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:57:39.0173216Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:57:39.0176237Z (EngineCore_DP0 pid=27268) INFO 11-03 17:57:39 [parallel_state.py:1231] rank 0 in world size 1 is assigned as DP rank 0, PP rank 0, TP rank 0, EP rank 0 2025-11-03T17:57:39.3584920Z (EngineCore_DP0 pid=27268) INFO 11-03 17:57:39 [topk_topp_sampler.py:59] Using FlashInfer for top-p & top-k sampling. 2025-11-03T17:57:39.4228402Z (EngineCore_DP0 pid=27268) INFO 11-03 17:57:39 [gpu_model_runner.py:2840] Starting to load model microsoft/Phi-3-mini-4k-instruct... 2025-11-03T17:57:39.6803331Z (EngineCore_DP0 pid=27268) INFO 11-03 17:57:39 [gpu_model_runner.py:2870] Loading model from scratch... 2025-11-03T17:57:39.7455641Z (EngineCore_DP0 pid=27268) INFO 11-03 17:57:39 [cuda.py:404] Using Flash Attention backend on V1 engine. 2025-11-03T17:57:40.3211618Z (EngineCore_DP0 pid=27268) INFO 11-03 17:57:40 [gpu_model_runner.py:2902] Model loading took 0.5789 GiB and 0.123813 seconds 2025-11-03T17:57:40.3215982Z (EngineCore_DP0 pid=27268) INFO 11-03 17:57:40 [kv_cache_utils.py:1199] GPU KV cache size: 873,808 tokens 2025-11-03T17:57:40.3216760Z (EngineCore_DP0 pid=27268) INFO 11-03 17:57:40 [kv_cache_utils.py:1204] Maximum concurrency for 4,096 tokens per request: 212.50x 2025-11-03T17:57:40.7854577Z (EngineCore_DP0 pid=27268) INFO 11-03 17:57:40 [gc_utils.py:40] GC Debug Config. enabled:False,top_objects:-1 2025-11-03T17:57:40.7867853Z INFO 11-03 17:57:40 [llm.py:337] Supported tasks: ['generate'] 2025-11-03T17:57:41.2344479Z PASSED 2025-11-03T17:57:41.2487209Z models/test_initialization.py::test_can_initialize_large_subset[SolarForCausalLM] Fork a new process to run a test 27334 2025-11-03T17:57:41.2497579Z Fork a new process to run a test 0 2025-11-03T17:57:41.2775090Z INFO 11-03 17:57:41 [utils.py:239] non-default args: {'trust_remote_code': True, 'load_format': 'dummy', 'gpu_memory_utilization': 0.8, 'disable_log_stats': True, 'hf_overrides': functools.partial(, model_arch='SolarForCausalLM', exist_overrides={}, use_original_num_layers=False), 'model_impl': 'vllm', 'model': 'upstage/solar-pro-preview-instruct'} 2025-11-03T17:57:41.4314179Z 2025-11-03T17:57:41.4316217Z config.json: 0.00B [00:00, ?B/s] 2025-11-03T17:57:41.4316535Z config.json: 1.03kB [00:00, 5.34MB/s] 2025-11-03T17:57:41.5087036Z 2025-11-03T17:57:41.5089637Z configuration_solar.py: 0.00B [00:00, ?B/s] 2025-11-03T17:57:41.5089986Z configuration_solar.py: 10.1kB [00:00, 30.3MB/s] 2025-11-03T17:57:41.5187791Z A new version of the following files was downloaded from https://huggingface.co/upstage/solar-pro-preview-instruct: 2025-11-03T17:57:41.5188366Z - configuration_solar.py 2025-11-03T17:57:41.5188977Z . Make sure to double-check they do not contain any added malicious code. To avoid downloading new versions of the code file, you can pin a revision. 2025-11-03T17:57:48.0135857Z INFO 11-03 17:57:48 [model.py:653] Resolved architecture: SolarForCausalLM 2025-11-03T17:57:48.0136488Z `torch_dtype` is deprecated! Use `dtype` instead! 2025-11-03T17:57:48.0387956Z INFO 11-03 17:57:48 [model.py:1714] Using max model len 4096 2025-11-03T17:57:48.0389749Z The argument `trust_remote_code` is to be used with Auto classes. It has no effect here and is ignored. 2025-11-03T17:57:48.1043756Z INFO 11-03 17:57:48 [scheduler.py:225] Chunked prefill is enabled with max_num_batched_tokens=8192. 2025-11-03T17:57:48.1464799Z 2025-11-03T17:57:48.1467239Z tokenizer_config.json: 0.00B [00:00, ?B/s] 2025-11-03T17:57:48.2752419Z tokenizer_config.json: 24.6kB [00:00, 121MB/s] 2025-11-03T17:57:48.2752672Z 2025-11-03T17:57:48.4692489Z tokenizer.model: 0% 0.00/500k [00:00, 'use_cudagraph': True, 'cudagraph_num_of_warmups': 1, 'cudagraph_capture_sizes': [512, 504, 496, 488, 480, 472, 464, 456, 448, 440, 432, 424, 416, 408, 400, 392, 384, 376, 368, 360, 352, 344, 336, 328, 320, 312, 304, 296, 288, 280, 272, 264, 256, 248, 240, 232, 224, 216, 208, 200, 192, 184, 176, 168, 160, 152, 144, 136, 128, 120, 112, 104, 96, 88, 80, 72, 64, 56, 48, 40, 32, 24, 16, 8, 4, 2, 1], 'cudagraph_copy_inputs': False, 'full_cuda_graph': True, 'use_inductor_graph_partition': False, 'pass_config': {}, 'max_capture_size': 512, 'local_cache_dir': None} 2025-11-03T17:57:49.4715703Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:57:49.4743400Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:57:49.4750949Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:57:49.4757976Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:57:49.4765705Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:57:49.4773236Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:57:49.4776038Z (EngineCore_DP0 pid=27412) INFO 11-03 17:57:49 [parallel_state.py:1231] rank 0 in world size 1 is assigned as DP rank 0, PP rank 0, TP rank 0, EP rank 0 2025-11-03T17:57:49.8259202Z (EngineCore_DP0 pid=27412) INFO 11-03 17:57:49 [topk_topp_sampler.py:59] Using FlashInfer for top-p & top-k sampling. 2025-11-03T17:57:49.9250156Z (EngineCore_DP0 pid=27412) INFO 11-03 17:57:49 [gpu_model_runner.py:2840] Starting to load model upstage/solar-pro-preview-instruct... 2025-11-03T17:57:50.1817050Z (EngineCore_DP0 pid=27412) INFO 11-03 17:57:50 [gpu_model_runner.py:2870] Loading model from scratch... 2025-11-03T17:57:50.2467531Z (EngineCore_DP0 pid=27412) INFO 11-03 17:57:50 [cuda.py:404] Using Flash Attention backend on V1 engine. 2025-11-03T17:57:50.8394512Z (EngineCore_DP0 pid=27412) INFO 11-03 17:57:50 [gpu_model_runner.py:2902] Model loading took 1.2510 GiB and 0.148184 seconds 2025-11-03T17:57:50.8399771Z (EngineCore_DP0 pid=27412) INFO 11-03 17:57:50 [kv_cache_utils.py:1199] GPU KV cache size: 2,097,152 tokens 2025-11-03T17:57:50.8400790Z (EngineCore_DP0 pid=27412) INFO 11-03 17:57:50 [kv_cache_utils.py:1204] Maximum concurrency for 4,096 tokens per request: 510.01x 2025-11-03T17:57:51.4072903Z (EngineCore_DP0 pid=27412) INFO 11-03 17:57:51 [gc_utils.py:40] GC Debug Config. enabled:False,top_objects:-1 2025-11-03T17:57:51.4094340Z INFO 11-03 17:57:51 [llm.py:337] Supported tasks: ['generate'] 2025-11-03T17:57:51.8736395Z PASSED 2025-11-03T17:57:51.8877059Z models/test_initialization.py::test_can_initialize_large_subset[Gemma2Model] Fork a new process to run a test 27478 2025-11-03T17:57:51.8888196Z Fork a new process to run a test 0 2025-11-03T17:57:51.9167040Z INFO 11-03 17:57:51 [utils.py:239] non-default args: {'load_format': 'dummy', 'gpu_memory_utilization': 0.8, 'disable_log_stats': True, 'hf_overrides': functools.partial(, model_arch='Gemma2Model', exist_overrides={}, use_original_num_layers=False), 'model_impl': 'vllm', 'model': 'BAAI/bge-multilingual-gemma2'} 2025-11-03T17:57:52.0122513Z 2025-11-03T17:57:52.0123985Z config.json: 0% 0.00/897 [00:00, 'use_cudagraph': True, 'cudagraph_num_of_warmups': 1, 'cudagraph_capture_sizes': [512, 504, 496, 488, 480, 472, 464, 456, 448, 440, 432, 424, 416, 408, 400, 392, 384, 376, 368, 360, 352, 344, 336, 328, 320, 312, 304, 296, 288, 280, 272, 264, 256, 248, 240, 232, 224, 216, 208, 200, 192, 184, 176, 168, 160, 152, 144, 136, 128, 120, 112, 104, 96, 88, 80, 72, 64, 56, 48, 40, 32, 24, 16, 8, 4, 2, 1], 'cudagraph_copy_inputs': False, 'full_cuda_graph': False, 'use_inductor_graph_partition': False, 'pass_config': {}, 'max_capture_size': 512, 'local_cache_dir': None} 2025-11-03T17:57:55.0643463Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:57:55.0670892Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:57:55.0678342Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:57:55.0686391Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:57:55.0694078Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:57:55.0702060Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:57:55.0704710Z (EngineCore_DP0 pid=27507) INFO 11-03 17:57:55 [parallel_state.py:1231] rank 0 in world size 1 is assigned as DP rank 0, PP rank 0, TP rank 0, EP rank 0 2025-11-03T17:57:55.4008054Z (EngineCore_DP0 pid=27507) INFO 11-03 17:57:55 [topk_topp_sampler.py:59] Using FlashInfer for top-p & top-k sampling. 2025-11-03T17:57:55.4579658Z (EngineCore_DP0 pid=27507) INFO 11-03 17:57:55 [gpu_model_runner.py:2840] Starting to load model BAAI/bge-multilingual-gemma2... 2025-11-03T17:57:55.7122127Z (EngineCore_DP0 pid=27507) INFO 11-03 17:57:55 [gpu_model_runner.py:2870] Loading model from scratch... 2025-11-03T17:57:55.7813363Z (EngineCore_DP0 pid=27507) INFO 11-03 17:57:55 [cuda.py:404] Using Flash Attention backend on V1 engine. 2025-11-03T17:57:56.3471970Z (EngineCore_DP0 pid=27507) INFO 11-03 17:57:56 [gpu_model_runner.py:2902] Model loading took 2.0825 GiB and 0.132153 seconds 2025-11-03T17:57:56.3475956Z (EngineCore_DP0 pid=27507) INFO 11-03 17:57:56 [kv_cache_utils.py:1199] GPU KV cache size: 1,310,720 tokens 2025-11-03T17:57:56.3477206Z (EngineCore_DP0 pid=27507) INFO 11-03 17:57:56 [kv_cache_utils.py:1204] Maximum concurrency for 8,192 tokens per request: 159.69x 2025-11-03T17:57:57.5771409Z (EngineCore_DP0 pid=27507) INFO 11-03 17:57:57 [gc_utils.py:40] GC Debug Config. enabled:False,top_objects:-1 2025-11-03T17:57:57.5786439Z INFO 11-03 17:57:57 [llm.py:337] Supported tasks: ['embed'] 2025-11-03T17:57:58.0543285Z PASSED 2025-11-03T17:57:58.0685040Z models/test_initialization.py::test_can_initialize_large_subset[Step3VLForConditionalGeneration] Fork a new process to run a test 27573 2025-11-03T17:57:58.0695692Z Fork a new process to run a test 0 2025-11-03T17:57:58.0974776Z INFO 11-03 17:57:58 [utils.py:239] non-default args: {'trust_remote_code': True, 'load_format': 'dummy', 'gpu_memory_utilization': 0.8, 'disable_log_stats': True, 'hf_overrides': functools.partial(, model_arch='Step3VLForConditionalGeneration', exist_overrides={}, use_original_num_layers=False), 'model_impl': 'vllm', 'model': 'stepfun-ai/step3'} 2025-11-03T17:57:58.2943247Z INFO 11-03 17:57:58 [model.py:653] Resolved architecture: Step3VLForConditionalGeneration 2025-11-03T17:57:58.2944069Z `torch_dtype` is deprecated! Use `dtype` instead! 2025-11-03T17:57:58.3186049Z INFO 11-03 17:57:58 [model.py:1714] Using max model len 65536 2025-11-03T17:57:58.3188580Z The argument `trust_remote_code` is to be used with Auto classes. It has no effect here and is ignored. 2025-11-03T17:57:58.4874583Z INFO 11-03 17:57:58 [scheduler.py:225] Chunked prefill is enabled with max_num_batched_tokens=8192. 2025-11-03T17:57:59.4451582Z (EngineCore_DP0 pid=27580) INFO 11-03 17:57:59 [core.py:727] Waiting for init message from front-end. 2025-11-03T17:57:59.4486114Z (EngineCore_DP0 pid=27580) INFO 11-03 17:57:59 [core.py:94] Initializing a V1 LLM engine (v0.11.0rc2.dev389+ge51928192.d20251103) with config: model='stepfun-ai/step3', speculative_config=None, tokenizer='stepfun-ai/step3', skip_tokenizer_init=False, tokenizer_mode=auto, revision=None, tokenizer_revision=None, trust_remote_code=True, dtype=torch.bfloat16, max_seq_len=65536, download_dir=None, load_format=dummy, tensor_parallel_size=1, pipeline_parallel_size=1, data_parallel_size=1, disable_custom_all_reduce=False, quantization=None, enforce_eager=False, kv_cache_dtype=auto, device_config=cuda, structured_outputs_config=StructuredOutputsConfig(backend='auto', disable_fallback=False, disable_any_whitespace=False, disable_additional_properties=False, reasoning_parser=''), observability_config=ObservabilityConfig(show_hidden_metrics_for_version=None, otlp_traces_endpoint=None, collect_detailed_traces=None), seed=0, served_model_name=stepfun-ai/step3, enable_prefix_caching=True, chunked_prefill_enabled=True, pooler_config=None, compilation_config={'level': 3, 'debug_dump_path': None, 'cache_dir': '', 'backend': '', 'custom_ops': [], 'splitting_ops': ['vllm.unified_attention', 'vllm.unified_attention_with_output', 'vllm.unified_mla_attention', 'vllm.unified_mla_attention_with_output', 'vllm.mamba_mixer2', 'vllm.mamba_mixer', 'vllm.short_conv', 'vllm.linear_attention', 'vllm.plamo2_mamba_mixer', 'vllm.gdn_attention', 'vllm.sparse_attn_indexer'], 'use_inductor': True, 'compile_sizes': [], 'inductor_compile_config': {'enable_auto_functionalized_v2': False}, 'inductor_passes': {}, 'cudagraph_mode': , 'use_cudagraph': True, 'cudagraph_num_of_warmups': 1, 'cudagraph_capture_sizes': [512, 504, 496, 488, 480, 472, 464, 456, 448, 440, 432, 424, 416, 408, 400, 392, 384, 376, 368, 360, 352, 344, 336, 328, 320, 312, 304, 296, 288, 280, 272, 264, 256, 248, 240, 232, 224, 216, 208, 200, 192, 184, 176, 168, 160, 152, 144, 136, 128, 120, 112, 104, 96, 88, 80, 72, 64, 56, 48, 40, 32, 24, 16, 8, 4, 2, 1], 'cudagraph_copy_inputs': False, 'full_cuda_graph': True, 'use_inductor_graph_partition': False, 'pass_config': {}, 'max_capture_size': 512, 'local_cache_dir': None} 2025-11-03T17:58:00.1507490Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:58:00.1535108Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:58:00.1542988Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:58:00.1551248Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:58:00.1559403Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:58:00.1567779Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:58:00.1570965Z (EngineCore_DP0 pid=27580) INFO 11-03 17:58:00 [parallel_state.py:1231] rank 0 in world size 1 is assigned as DP rank 0, PP rank 0, TP rank 0, EP rank 0 2025-11-03T17:58:00.4957723Z (EngineCore_DP0 pid=27580) INFO 11-03 17:58:00 [topk_topp_sampler.py:59] Using FlashInfer for top-p & top-k sampling. 2025-11-03T17:58:00.6031551Z (EngineCore_DP0 pid=27580) INFO 11-03 17:58:00 [gpu_model_runner.py:2840] Starting to load model stepfun-ai/step3... 2025-11-03T17:58:00.8665203Z (EngineCore_DP0 pid=27580) INFO 11-03 17:58:00 [gpu_model_runner.py:2870] Loading model from scratch... 2025-11-03T17:58:00.8839117Z (EngineCore_DP0 pid=27580) INFO 11-03 17:58:00 [layer.py:497] MultiHeadAttention attn_backend: _Backend.XFORMERS, use_upstream_fa: False 2025-11-03T17:58:00.9512364Z (EngineCore_DP0 pid=27580) INFO 11-03 17:58:00 [cuda.py:404] Using Flash Attention backend on V1 engine. 2025-11-03T17:58:01.5366653Z (EngineCore_DP0 pid=27580) INFO 11-03 17:58:01 [gpu_model_runner.py:2902] Model loading took 5.3915 GiB and 0.145262 seconds 2025-11-03T17:58:01.5371115Z (EngineCore_DP0 pid=27580) INFO 11-03 17:58:01 [kv_cache_utils.py:1199] GPU KV cache size: 10,485,760 tokens 2025-11-03T17:58:01.5372640Z (EngineCore_DP0 pid=27580) INFO 11-03 17:58:01 [kv_cache_utils.py:1204] Maximum concurrency for 65,536 tokens per request: 160.00x 2025-11-03T17:58:03.4256564Z (EngineCore_DP0 pid=27580) INFO 11-03 17:58:03 [gc_utils.py:40] GC Debug Config. enabled:False,top_objects:-1 2025-11-03T17:58:03.4270650Z INFO 11-03 17:58:03 [llm.py:337] Supported tasks: ['generate'] 2025-11-03T17:58:03.9236406Z PASSED 2025-11-03T17:58:03.9377751Z models/test_initialization.py::test_can_initialize_large_subset[Gemma3TextModel] Fork a new process to run a test 27646 2025-11-03T17:58:03.9389284Z Fork a new process to run a test 0 2025-11-03T17:58:03.9674156Z INFO 11-03 17:58:03 [utils.py:239] non-default args: {'load_format': 'dummy', 'gpu_memory_utilization': 0.8, 'disable_log_stats': True, 'hf_overrides': functools.partial(, model_arch='Gemma3TextModel', exist_overrides={}, use_original_num_layers=False), 'model_impl': 'vllm', 'model': 'google/embeddinggemma-300m'} 2025-11-03T17:58:04.1029823Z 2025-11-03T17:58:04.1030900Z config.json: 0% 0.00/1.49k [00:00, 'use_cudagraph': True, 'cudagraph_num_of_warmups': 1, 'cudagraph_capture_sizes': [512, 504, 496, 488, 480, 472, 464, 456, 448, 440, 432, 424, 416, 408, 400, 392, 384, 376, 368, 360, 352, 344, 336, 328, 320, 312, 304, 296, 288, 280, 272, 264, 256, 248, 240, 232, 224, 216, 208, 200, 192, 184, 176, 168, 160, 152, 144, 136, 128, 120, 112, 104, 96, 88, 80, 72, 64, 56, 48, 40, 32, 24, 16, 8, 4, 2, 1], 'cudagraph_copy_inputs': False, 'full_cuda_graph': False, 'use_inductor_graph_partition': False, 'pass_config': {}, 'max_capture_size': 512, 'local_cache_dir': None} 2025-11-03T17:58:14.4061883Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:58:14.4093195Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:58:14.4100269Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:58:14.4108770Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:58:14.4116416Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:58:14.4124141Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:58:14.4126486Z (EngineCore_DP0 pid=27725) INFO 11-03 17:58:14 [parallel_state.py:1231] rank 0 in world size 1 is assigned as DP rank 0, PP rank 0, TP rank 0, EP rank 0 2025-11-03T17:58:14.7516013Z (EngineCore_DP0 pid=27725) INFO 11-03 17:58:14 [topk_topp_sampler.py:59] Using FlashInfer for top-p & top-k sampling. 2025-11-03T17:58:14.7724741Z (EngineCore_DP0 pid=27725) INFO 11-03 17:58:14 [gpu_model_runner.py:2840] Starting to load model google/embeddinggemma-300m... 2025-11-03T17:58:15.0265366Z (EngineCore_DP0 pid=27725) INFO 11-03 17:58:15 [gpu_model_runner.py:2870] Loading model from scratch... 2025-11-03T17:58:15.1135487Z (EngineCore_DP0 pid=27725) INFO 11-03 17:58:15 [cuda.py:404] Using Flash Attention backend on V1 engine. 2025-11-03T17:58:15.2393957Z (EngineCore_DP0 pid=27725) 2025-11-03T17:58:15.2394353Z config.json: 0% 0.00/134 [00:00=4.57.0.dev0` is required to run this model. 2025-11-03T17:58:19.1904758Z PASSED 2025-11-03T17:58:19.2047170Z models/test_initialization.py::test_can_initialize_large_subset[AquilaModel] Fork a new process to run a test 27813 2025-11-03T17:58:19.2057339Z Fork a new process to run a test 0 2025-11-03T17:58:19.2337750Z INFO 11-03 17:58:19 [utils.py:239] non-default args: {'trust_remote_code': True, 'load_format': 'dummy', 'gpu_memory_utilization': 0.8, 'disable_log_stats': True, 'hf_overrides': functools.partial(, model_arch='AquilaModel', exist_overrides={}, use_original_num_layers=False), 'model_impl': 'vllm', 'model': 'BAAI/AquilaChat-7B'} 2025-11-03T17:58:19.3481999Z 2025-11-03T17:58:19.3483332Z config.json: 0% 0.00/707 [00:00, 'use_cudagraph': True, 'cudagraph_num_of_warmups': 1, 'cudagraph_capture_sizes': [512, 504, 496, 488, 480, 472, 464, 456, 448, 440, 432, 424, 416, 408, 400, 392, 384, 376, 368, 360, 352, 344, 336, 328, 320, 312, 304, 296, 288, 280, 272, 264, 256, 248, 240, 232, 224, 216, 208, 200, 192, 184, 176, 168, 160, 152, 144, 136, 128, 120, 112, 104, 96, 88, 80, 72, 64, 56, 48, 40, 32, 24, 16, 8, 4, 2, 1], 'cudagraph_copy_inputs': False, 'full_cuda_graph': True, 'use_inductor_graph_partition': False, 'pass_config': {}, 'max_capture_size': 512, 'local_cache_dir': None} 2025-11-03T17:58:21.1140199Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:58:21.1168165Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:58:21.1175425Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:58:21.1182805Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:58:21.1190579Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:58:21.1198254Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:58:21.1200920Z (EngineCore_DP0 pid=27821) INFO 11-03 17:58:21 [parallel_state.py:1231] rank 0 in world size 1 is assigned as DP rank 0, PP rank 0, TP rank 0, EP rank 0 2025-11-03T17:58:21.4497163Z (EngineCore_DP0 pid=27821) INFO 11-03 17:58:21 [topk_topp_sampler.py:59] Using FlashInfer for top-p & top-k sampling. 2025-11-03T17:58:21.5166494Z (EngineCore_DP0 pid=27821) INFO 11-03 17:58:21 [gpu_model_runner.py:2840] Starting to load model BAAI/AquilaChat-7B... 2025-11-03T17:58:21.7694456Z (EngineCore_DP0 pid=27821) INFO 11-03 17:58:21 [gpu_model_runner.py:2870] Loading model from scratch... 2025-11-03T17:58:21.8343632Z (EngineCore_DP0 pid=27821) INFO 11-03 17:58:21 [cuda.py:404] Using Flash Attention backend on V1 engine. 2025-11-03T17:58:22.3951975Z (EngineCore_DP0 pid=27821) INFO 11-03 17:58:22 [gpu_model_runner.py:2902] Model loading took 1.9048 GiB and 0.124629 seconds 2025-11-03T17:58:22.3956798Z (EngineCore_DP0 pid=27821) INFO 11-03 17:58:22 [kv_cache_utils.py:1199] GPU KV cache size: 655,360 tokens 2025-11-03T17:58:23.0066564Z (EngineCore_DP0 pid=27821) INFO 11-03 17:58:22 [kv_cache_utils.py:1204] Maximum concurrency for 2,048 tokens per request: 320.00x 2025-11-03T17:58:23.0068300Z (EngineCore_DP0 pid=27821) INFO 11-03 17:58:23 [gc_utils.py:40] GC Debug Config. enabled:False,top_objects:-1 2025-11-03T17:58:23.0090590Z INFO 11-03 17:58:23 [llm.py:337] Supported tasks: ['generate'] 2025-11-03T17:58:23.4596038Z PASSED 2025-11-03T17:58:23.4737665Z models/test_initialization.py::test_can_initialize_large_subset[LongcatFlashForCausalLM] Fork a new process to run a test 27887 2025-11-03T17:58:23.4748363Z Fork a new process to run a test 0 2025-11-03T17:58:23.5026320Z INFO 11-03 17:58:23 [utils.py:239] non-default args: {'trust_remote_code': True, 'load_format': 'dummy', 'gpu_memory_utilization': 0.8, 'disable_log_stats': True, 'hf_overrides': functools.partial(, model_arch='LongcatFlashForCausalLM', exist_overrides={}, use_original_num_layers=False), 'model_impl': 'vllm', 'model': 'meituan-longcat/LongCat-Flash-Chat'} 2025-11-03T17:58:23.8867333Z INFO 11-03 17:58:23 [model.py:653] Resolved architecture: LongcatFlashForCausalLM 2025-11-03T17:58:23.8867839Z `torch_dtype` is deprecated! Use `dtype` instead! 2025-11-03T17:58:24.0196835Z 2025-11-03T17:58:24.1206693Z Parse safetensors files: 0% 0/75 [00:00, 'use_cudagraph': True, 'cudagraph_num_of_warmups': 1, 'cudagraph_capture_sizes': [512, 504, 496, 488, 480, 472, 464, 456, 448, 440, 432, 424, 416, 408, 400, 392, 384, 376, 368, 360, 352, 344, 336, 328, 320, 312, 304, 296, 288, 280, 272, 264, 256, 248, 240, 232, 224, 216, 208, 200, 192, 184, 176, 168, 160, 152, 144, 136, 128, 120, 112, 104, 96, 88, 80, 72, 64, 56, 48, 40, 32, 24, 16, 8, 4, 2, 1], 'cudagraph_copy_inputs': False, 'full_cuda_graph': True, 'use_inductor_graph_partition': False, 'pass_config': {}, 'max_capture_size': 512, 'local_cache_dir': None} 2025-11-03T17:58:25.7765446Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:58:25.7794993Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:58:25.7803199Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:58:25.7812220Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:58:25.7820908Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:58:25.7830613Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:58:25.7833328Z (EngineCore_DP0 pid=27915) INFO 11-03 17:58:25 [parallel_state.py:1231] rank 0 in world size 1 is assigned as DP rank 0, PP rank 0, TP rank 0, EP rank 0 2025-11-03T17:58:26.1158902Z (EngineCore_DP0 pid=27915) INFO 11-03 17:58:26 [topk_topp_sampler.py:59] Using FlashInfer for top-p & top-k sampling. 2025-11-03T17:58:26.2283536Z (EngineCore_DP0 pid=27915) INFO 11-03 17:58:26 [gpu_model_runner.py:2840] Starting to load model meituan-longcat/LongCat-Flash-Chat... 2025-11-03T17:58:26.4853745Z (EngineCore_DP0 pid=27915) INFO 11-03 17:58:26 [gpu_model_runner.py:2870] Loading model from scratch... 2025-11-03T17:58:26.5515531Z (EngineCore_DP0 pid=27915) INFO 11-03 17:58:26 [cuda.py:328] Using Triton MLA backend on V1 engine. 2025-11-03T17:58:27.1347642Z (EngineCore_DP0 pid=27915) INFO 11-03 17:58:27 [gpu_model_runner.py:2902] Model loading took 4.3434 GiB and 0.143235 seconds 2025-11-03T17:58:27.1352727Z (EngineCore_DP0 pid=27915) INFO 11-03 17:58:27 [kv_cache_utils.py:1199] GPU KV cache size: 4,660,336 tokens 2025-11-03T17:58:27.1353711Z (EngineCore_DP0 pid=27915) INFO 11-03 17:58:27 [kv_cache_utils.py:1204] Maximum concurrency for 131,072 tokens per request: 35.56x 2025-11-03T17:58:28.2423454Z (EngineCore_DP0 pid=27915) INFO 11-03 17:58:28 [gc_utils.py:40] GC Debug Config. enabled:False,top_objects:-1 2025-11-03T17:58:28.2438920Z INFO 11-03 17:58:28 [llm.py:337] Supported tasks: ['generate'] 2025-11-03T17:58:28.7405578Z PASSED 2025-11-03T17:58:28.7547231Z models/test_initialization.py::test_can_initialize_large_subset[InternLM3ForCausalLM] Fork a new process to run a test 27981 2025-11-03T17:58:28.7558873Z Fork a new process to run a test 0 2025-11-03T17:58:28.7831583Z INFO 11-03 17:58:28 [utils.py:239] non-default args: {'trust_remote_code': True, 'load_format': 'dummy', 'gpu_memory_utilization': 0.8, 'disable_log_stats': True, 'hf_overrides': functools.partial(, model_arch='InternLM3ForCausalLM', exist_overrides={}, use_original_num_layers=False), 'model_impl': 'vllm', 'model': 'internlm/internlm3-8b-instruct'} 2025-11-03T17:58:28.8583467Z 2025-11-03T17:58:28.8585657Z config.json: 0.00B [00:00, ?B/s] 2025-11-03T17:58:28.8585960Z config.json: 1.01kB [00:00, 5.12MB/s] 2025-11-03T17:58:28.9237342Z 2025-11-03T17:58:28.9241026Z configuration_internlm3.py: 0.00B [00:00, ?B/s] 2025-11-03T17:58:28.9241425Z configuration_internlm3.py: 10.5kB [00:00, 18.9MB/s] 2025-11-03T17:58:28.9331728Z A new version of the following files was downloaded from https://huggingface.co/internlm/internlm3-8b-instruct: 2025-11-03T17:58:28.9332507Z - configuration_internlm3.py 2025-11-03T17:58:28.9333129Z . Make sure to double-check they do not contain any added malicious code. To avoid downloading new versions of the code file, you can pin a revision. 2025-11-03T17:58:29.1675347Z INFO 11-03 17:58:29 [model.py:653] Resolved architecture: InternLM3ForCausalLM 2025-11-03T17:58:29.1675823Z `torch_dtype` is deprecated! Use `dtype` instead! 2025-11-03T17:58:29.1919120Z INFO 11-03 17:58:29 [model.py:1714] Using max model len 196608 2025-11-03T17:58:29.1920593Z The argument `trust_remote_code` is to be used with Auto classes. It has no effect here and is ignored. 2025-11-03T17:58:29.2577805Z INFO 11-03 17:58:29 [scheduler.py:225] Chunked prefill is enabled with max_num_batched_tokens=8192. 2025-11-03T17:58:29.2920674Z 2025-11-03T17:58:29.2922954Z tokenizer_config.json: 0.00B [00:00, ?B/s] 2025-11-03T17:58:29.2923320Z tokenizer_config.json: 5.69kB [00:00, 37.5MB/s] 2025-11-03T17:58:29.3414291Z 2025-11-03T17:58:29.3417290Z tokenization_internlm3.py: 0.00B [00:00, ?B/s] 2025-11-03T17:58:29.3417695Z tokenization_internlm3.py: 13.2kB [00:00, 34.6MB/s] 2025-11-03T17:58:29.3472054Z A new version of the following files was downloaded from https://huggingface.co/internlm/internlm3-8b-instruct: 2025-11-03T17:58:29.3472613Z - tokenization_internlm3.py 2025-11-03T17:58:29.3473242Z . Make sure to double-check they do not contain any added malicious code. To avoid downloading new versions of the code file, you can pin a revision. 2025-11-03T17:58:29.6148931Z 2025-11-03T17:58:29.7999910Z tokenizer.model: 0% 0.00/2.48M [00:00, 'use_cudagraph': True, 'cudagraph_num_of_warmups': 1, 'cudagraph_capture_sizes': [512, 504, 496, 488, 480, 472, 464, 456, 448, 440, 432, 424, 416, 408, 400, 392, 384, 376, 368, 360, 352, 344, 336, 328, 320, 312, 304, 296, 288, 280, 272, 264, 256, 248, 240, 232, 224, 216, 208, 200, 192, 184, 176, 168, 160, 152, 144, 136, 128, 120, 112, 104, 96, 88, 80, 72, 64, 56, 48, 40, 32, 24, 16, 8, 4, 2, 1], 'cudagraph_copy_inputs': False, 'full_cuda_graph': True, 'use_inductor_graph_partition': False, 'pass_config': {}, 'max_capture_size': 512, 'local_cache_dir': None} 2025-11-03T17:58:31.3505007Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:58:31.3532508Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:58:31.3539404Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:58:31.3547676Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:58:31.3555305Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:58:31.3563199Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:58:31.3565854Z (EngineCore_DP0 pid=28009) INFO 11-03 17:58:31 [parallel_state.py:1231] rank 0 in world size 1 is assigned as DP rank 0, PP rank 0, TP rank 0, EP rank 0 2025-11-03T17:58:31.6935081Z (EngineCore_DP0 pid=28009) INFO 11-03 17:58:31 [topk_topp_sampler.py:59] Using FlashInfer for top-p & top-k sampling. 2025-11-03T17:58:31.7867520Z (EngineCore_DP0 pid=28009) INFO 11-03 17:58:31 [gpu_model_runner.py:2840] Starting to load model internlm/internlm3-8b-instruct... 2025-11-03T17:58:32.0449764Z (EngineCore_DP0 pid=28009) INFO 11-03 17:58:32 [gpu_model_runner.py:2870] Loading model from scratch... 2025-11-03T17:58:32.1117878Z (EngineCore_DP0 pid=28009) INFO 11-03 17:58:32 [cuda.py:404] Using Flash Attention backend on V1 engine. 2025-11-03T17:58:32.6822280Z (EngineCore_DP0 pid=28009) INFO 11-03 17:58:32 [gpu_model_runner.py:2902] Model loading took 2.3086 GiB and 0.126222 seconds 2025-11-03T17:58:32.6826916Z (EngineCore_DP0 pid=28009) INFO 11-03 17:58:32 [kv_cache_utils.py:1199] GPU KV cache size: 10,485,760 tokens 2025-11-03T17:58:32.6827787Z (EngineCore_DP0 pid=28009) INFO 11-03 17:58:32 [kv_cache_utils.py:1204] Maximum concurrency for 196,608 tokens per request: 53.33x 2025-11-03T17:58:33.0837401Z (EngineCore_DP0 pid=28009) WARNING 11-03 17:58:33 [tokenizer.py:255] Using a slow tokenizer. This might cause a significant slowdown. Consider using a fast tokenizer instead. 2025-11-03T17:58:34.4716049Z (EngineCore_DP0 pid=28009) INFO 11-03 17:58:34 [gc_utils.py:40] GC Debug Config. enabled:False,top_objects:-1 2025-11-03T17:58:34.4729911Z INFO 11-03 17:58:34 [llm.py:337] Supported tasks: ['generate'] 2025-11-03T17:58:35.0021650Z PASSED 2025-11-03T17:58:35.0162476Z models/test_initialization.py::test_can_initialize_large_subset[JambaForCausalLM] Fork a new process to run a test 28075 2025-11-03T17:58:35.0174290Z Fork a new process to run a test 0 2025-11-03T17:58:35.0453312Z INFO 11-03 17:58:35 [utils.py:239] non-default args: {'load_format': 'dummy', 'gpu_memory_utilization': 0.8, 'disable_log_stats': True, 'hf_overrides': functools.partial(, model_arch='JambaForCausalLM', exist_overrides={}, use_original_num_layers=False), 'model_impl': 'vllm', 'model': 'ai21labs/AI21-Jamba-1.5-Mini'} 2025-11-03T17:58:35.2729551Z 2025-11-03T17:58:35.2730807Z config.json: 0% 0.00/932 [00:00= mamba page size. 2025-11-03T17:58:42.0438903Z INFO 11-03 17:58:42 [config.py:453] Padding mamba page size by 5.26% to ensure that mamba page size and attention page size are exactly equal. 2025-11-03T17:58:42.2052630Z 2025-11-03T17:58:42.2054727Z tokenizer_config.json: 0% 0.00/14.2k [00:00. This is expected, and simply means that the `legacy` (previous) behavior will be used so nothing changes for you. If you want to use the new behaviour, set `legacy=False`. This should only be set if you understand what it means, and thoroughly read the reason why this was added as explained in https://github.com/huggingface/transformers/pull/24565 - if you loaded a llama tokenizer from a GGUF file you can ignore this message. 2025-11-03T17:58:43.3047973Z 2025-11-03T17:58:43.3049003Z generation_config.json: 0% 0.00/126 [00:00, 'use_cudagraph': True, 'cudagraph_num_of_warmups': 1, 'cudagraph_capture_sizes': [512, 504, 496, 488, 480, 472, 464, 456, 448, 440, 432, 424, 416, 408, 400, 392, 384, 376, 368, 360, 352, 344, 336, 328, 320, 312, 304, 296, 288, 280, 272, 264, 256, 248, 240, 232, 224, 216, 208, 200, 192, 184, 176, 168, 160, 152, 144, 136, 128, 120, 112, 104, 96, 88, 80, 72, 64, 56, 48, 40, 32, 24, 16, 8, 4, 2, 1], 'cudagraph_copy_inputs': False, 'full_cuda_graph': True, 'use_inductor_graph_partition': False, 'pass_config': {}, 'max_capture_size': 512, 'local_cache_dir': None} 2025-11-03T17:58:43.9747674Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:58:43.9775014Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:58:43.9782440Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:58:43.9790104Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:58:43.9798079Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:58:43.9806210Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:58:43.9809434Z (EngineCore_DP0 pid=28154) INFO 11-03 17:58:43 [parallel_state.py:1231] rank 0 in world size 1 is assigned as DP rank 0, PP rank 0, TP rank 0, EP rank 0 2025-11-03T17:58:44.3186005Z (EngineCore_DP0 pid=28154) INFO 11-03 17:58:44 [topk_topp_sampler.py:59] Using FlashInfer for top-p & top-k sampling. 2025-11-03T17:58:44.4077625Z (EngineCore_DP0 pid=28154) INFO 11-03 17:58:44 [gpu_model_runner.py:2840] Starting to load model ai21labs/AI21-Jamba-1.5-Mini... 2025-11-03T17:58:44.6640580Z (EngineCore_DP0 pid=28154) INFO 11-03 17:58:44 [gpu_model_runner.py:2870] Loading model from scratch... 2025-11-03T17:58:45.2403525Z (EngineCore_DP0 pid=28154) INFO 11-03 17:58:45 [gpu_model_runner.py:2902] Model loading took 1.5245 GiB and 0.062653 seconds 2025-11-03T17:58:45.2408963Z (EngineCore_DP0 pid=28154) INFO 11-03 17:58:45 [kv_cache_utils.py:1199] GPU KV cache size: 8,589,934,592 tokens 2025-11-03T17:58:45.2410326Z (EngineCore_DP0 pid=28154) INFO 11-03 17:58:45 [kv_cache_utils.py:1204] Maximum concurrency for 262,144 tokens per request: 32768.00x 2025-11-03T17:58:45.8510994Z (EngineCore_DP0 pid=28154) INFO 11-03 17:58:45 [gc_utils.py:40] GC Debug Config. enabled:False,top_objects:-1 2025-11-03T17:58:45.8532239Z INFO 11-03 17:58:45 [llm.py:337] Supported tasks: ['generate'] 2025-11-03T17:58:46.3247205Z PASSED 2025-11-03T17:58:46.3389816Z models/test_initialization.py::test_can_initialize_large_subset[PhiForCausalLM] Fork a new process to run a test 28220 2025-11-03T17:58:46.3399582Z Fork a new process to run a test 0 2025-11-03T17:58:46.3672321Z INFO 11-03 17:58:46 [utils.py:239] non-default args: {'load_format': 'dummy', 'gpu_memory_utilization': 0.8, 'disable_log_stats': True, 'hf_overrides': functools.partial(, model_arch='PhiForCausalLM', exist_overrides={}, use_original_num_layers=False), 'model_impl': 'vllm', 'model': 'microsoft/phi-2'} 2025-11-03T17:58:46.4483533Z 2025-11-03T17:58:46.4485501Z config.json: 0% 0.00/735 [00:00, 'use_cudagraph': True, 'cudagraph_num_of_warmups': 1, 'cudagraph_capture_sizes': [512, 504, 496, 488, 480, 472, 464, 456, 448, 440, 432, 424, 416, 408, 400, 392, 384, 376, 368, 360, 352, 344, 336, 328, 320, 312, 304, 296, 288, 280, 272, 264, 256, 248, 240, 232, 224, 216, 208, 200, 192, 184, 176, 168, 160, 152, 144, 136, 128, 120, 112, 104, 96, 88, 80, 72, 64, 56, 48, 40, 32, 24, 16, 8, 4, 2, 1], 'cudagraph_copy_inputs': False, 'full_cuda_graph': True, 'use_inductor_graph_partition': False, 'pass_config': {}, 'max_capture_size': 512, 'local_cache_dir': None} 2025-11-03T17:58:54.2818621Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:58:54.2846662Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:58:54.2854476Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:58:54.2862402Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:58:54.2870179Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:58:54.2878055Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:58:54.2881030Z (EngineCore_DP0 pid=28278) INFO 11-03 17:58:54 [parallel_state.py:1231] rank 0 in world size 1 is assigned as DP rank 0, PP rank 0, TP rank 0, EP rank 0 2025-11-03T17:58:54.6223481Z (EngineCore_DP0 pid=28278) INFO 11-03 17:58:54 [topk_topp_sampler.py:59] Using FlashInfer for top-p & top-k sampling. 2025-11-03T17:58:54.6869996Z (EngineCore_DP0 pid=28278) INFO 11-03 17:58:54 [gpu_model_runner.py:2840] Starting to load model microsoft/phi-2... 2025-11-03T17:58:54.9455393Z (EngineCore_DP0 pid=28278) INFO 11-03 17:58:54 [gpu_model_runner.py:2870] Loading model from scratch... 2025-11-03T17:58:55.0106599Z (EngineCore_DP0 pid=28278) INFO 11-03 17:58:55 [cuda.py:420] Using FlexAttention backend for head_size=80 on V1 engine. 2025-11-03T17:58:55.6430138Z (EngineCore_DP0 pid=28278) INFO 11-03 17:58:55 [gpu_model_runner.py:2902] Model loading took 0.6355 GiB and 0.186806 seconds 2025-11-03T17:58:55.6434705Z (EngineCore_DP0 pid=28278) INFO 11-03 17:58:55 [kv_cache_utils.py:1199] GPU KV cache size: 1,048,576 tokens 2025-11-03T17:58:55.6435623Z (EngineCore_DP0 pid=28278) INFO 11-03 17:58:55 [kv_cache_utils.py:1204] Maximum concurrency for 2,048 tokens per request: 512.00x 2025-11-03T17:58:56.1318537Z (EngineCore_DP0 pid=28278) INFO 11-03 17:58:56 [gc_utils.py:40] GC Debug Config. enabled:False,top_objects:-1 2025-11-03T17:58:56.1335458Z INFO 11-03 17:58:56 [llm.py:337] Supported tasks: ['generate'] 2025-11-03T17:58:56.5755182Z PASSED 2025-11-03T17:58:56.5898479Z models/test_initialization.py::test_can_initialize_large_subset[GemmaForCausalLM] Fork a new process to run a test 28344 2025-11-03T17:58:56.5910523Z Fork a new process to run a test 0 2025-11-03T17:58:56.6199662Z INFO 11-03 17:58:56 [utils.py:239] non-default args: {'load_format': 'dummy', 'gpu_memory_utilization': 0.8, 'disable_log_stats': True, 'hf_overrides': functools.partial(, model_arch='GemmaForCausalLM', exist_overrides={}, use_original_num_layers=False), 'model_impl': 'vllm', 'model': 'google/gemma-1.1-2b-it'} 2025-11-03T17:58:56.7789380Z 2025-11-03T17:58:56.7790499Z config.json: 0% 0.00/618 [00:00, 'use_cudagraph': True, 'cudagraph_num_of_warmups': 1, 'cudagraph_capture_sizes': [512, 504, 496, 488, 480, 472, 464, 456, 448, 440, 432, 424, 416, 408, 400, 392, 384, 376, 368, 360, 352, 344, 336, 328, 320, 312, 304, 296, 288, 280, 272, 264, 256, 248, 240, 232, 224, 216, 208, 200, 192, 184, 176, 168, 160, 152, 144, 136, 128, 120, 112, 104, 96, 88, 80, 72, 64, 56, 48, 40, 32, 24, 16, 8, 4, 2, 1], 'cudagraph_copy_inputs': False, 'full_cuda_graph': True, 'use_inductor_graph_partition': False, 'pass_config': {}, 'max_capture_size': 512, 'local_cache_dir': None} 2025-11-03T17:59:06.3753703Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:59:06.3780822Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:59:06.3788359Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:59:06.3795736Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:59:06.3803904Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:59:06.3811670Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:59:06.3814560Z (EngineCore_DP0 pid=28424) INFO 11-03 17:59:06 [parallel_state.py:1231] rank 0 in world size 1 is assigned as DP rank 0, PP rank 0, TP rank 0, EP rank 0 2025-11-03T17:59:06.7096186Z (EngineCore_DP0 pid=28424) INFO 11-03 17:59:06 [topk_topp_sampler.py:59] Using FlashInfer for top-p & top-k sampling. 2025-11-03T17:59:06.7720107Z (EngineCore_DP0 pid=28424) INFO 11-03 17:59:06 [gpu_model_runner.py:2840] Starting to load model google/gemma-1.1-2b-it... 2025-11-03T17:59:07.0303125Z (EngineCore_DP0 pid=28424) INFO 11-03 17:59:07 [gpu_model_runner.py:2870] Loading model from scratch... 2025-11-03T17:59:07.0988563Z (EngineCore_DP0 pid=28424) INFO 11-03 17:59:07 [cuda.py:404] Using Flash Attention backend on V1 engine. 2025-11-03T17:59:07.6582397Z (EngineCore_DP0 pid=28424) INFO 11-03 17:59:07 [gpu_model_runner.py:2902] Model loading took 1.1856 GiB and 0.126894 seconds 2025-11-03T17:59:07.6587159Z (EngineCore_DP0 pid=28424) INFO 11-03 17:59:07 [kv_cache_utils.py:1199] GPU KV cache size: 10,485,760 tokens 2025-11-03T17:59:07.6588038Z (EngineCore_DP0 pid=28424) INFO 11-03 17:59:07 [kv_cache_utils.py:1204] Maximum concurrency for 8,192 tokens per request: 1280.00x 2025-11-03T17:59:09.6858791Z (EngineCore_DP0 pid=28424) INFO 11-03 17:59:09 [gc_utils.py:40] GC Debug Config. enabled:False,top_objects:-1 2025-11-03T17:59:09.6873574Z INFO 11-03 17:59:09 [llm.py:337] Supported tasks: ['generate'] 2025-11-03T17:59:10.1668680Z PASSED 2025-11-03T17:59:10.1809582Z models/test_initialization.py::test_can_initialize_large_subset[TransformersForCausalLM] Fork a new process to run a test 28490 2025-11-03T17:59:10.1819834Z Fork a new process to run a test 0 2025-11-03T17:59:10.2102435Z INFO 11-03 17:59:10 [utils.py:239] non-default args: {'trust_remote_code': True, 'load_format': 'dummy', 'gpu_memory_utilization': 0.8, 'disable_log_stats': True, 'hf_overrides': functools.partial(, model_arch='TransformersForCausalLM', exist_overrides={}, use_original_num_layers=False), 'model_impl': 'transformers', 'model': 'hmellor/Ilama-3.2-1B'} 2025-11-03T17:59:11.0484896Z INFO 11-03 17:59:11 [model.py:653] Resolved architecture: TransformersForCausalLM 2025-11-03T17:59:11.0485695Z `torch_dtype` is deprecated! Use `dtype` instead! 2025-11-03T17:59:11.0727444Z INFO 11-03 17:59:11 [model.py:1939] Downcasting torch.float32 to torch.bfloat16. 2025-11-03T17:59:11.0730375Z INFO 11-03 17:59:11 [model.py:1714] Using max model len 131072 2025-11-03T17:59:11.0733085Z The argument `trust_remote_code` is to be used with Auto classes. It has no effect here and is ignored. 2025-11-03T17:59:11.1285635Z INFO 11-03 17:59:11 [scheduler.py:225] Chunked prefill is enabled with max_num_batched_tokens=8192. 2025-11-03T17:59:11.7468773Z (EngineCore_DP0 pid=28497) INFO 11-03 17:59:11 [core.py:727] Waiting for init message from front-end. 2025-11-03T17:59:11.7502688Z (EngineCore_DP0 pid=28497) INFO 11-03 17:59:11 [core.py:94] Initializing a V1 LLM engine (v0.11.0rc2.dev389+ge51928192.d20251103) with config: model='hmellor/Ilama-3.2-1B', speculative_config=None, tokenizer='hmellor/Ilama-3.2-1B', skip_tokenizer_init=False, tokenizer_mode=auto, revision=None, tokenizer_revision=None, trust_remote_code=True, dtype=torch.bfloat16, max_seq_len=131072, download_dir=None, load_format=dummy, tensor_parallel_size=1, pipeline_parallel_size=1, data_parallel_size=1, disable_custom_all_reduce=False, quantization=None, enforce_eager=False, kv_cache_dtype=auto, device_config=cuda, structured_outputs_config=StructuredOutputsConfig(backend='auto', disable_fallback=False, disable_any_whitespace=False, disable_additional_properties=False, reasoning_parser=''), observability_config=ObservabilityConfig(show_hidden_metrics_for_version=None, otlp_traces_endpoint=None, collect_detailed_traces=None), seed=0, served_model_name=hmellor/Ilama-3.2-1B, enable_prefix_caching=True, chunked_prefill_enabled=True, pooler_config=None, compilation_config={'level': 3, 'debug_dump_path': None, 'cache_dir': '', 'backend': '', 'custom_ops': [], 'splitting_ops': ['vllm.unified_attention', 'vllm.unified_attention_with_output', 'vllm.unified_mla_attention', 'vllm.unified_mla_attention_with_output', 'vllm.mamba_mixer2', 'vllm.mamba_mixer', 'vllm.short_conv', 'vllm.linear_attention', 'vllm.plamo2_mamba_mixer', 'vllm.gdn_attention', 'vllm.sparse_attn_indexer'], 'use_inductor': True, 'compile_sizes': [], 'inductor_compile_config': {'enable_auto_functionalized_v2': False}, 'inductor_passes': {}, 'cudagraph_mode': , 'use_cudagraph': True, 'cudagraph_num_of_warmups': 1, 'cudagraph_capture_sizes': [512, 504, 496, 488, 480, 472, 464, 456, 448, 440, 432, 424, 416, 408, 400, 392, 384, 376, 368, 360, 352, 344, 336, 328, 320, 312, 304, 296, 288, 280, 272, 264, 256, 248, 240, 232, 224, 216, 208, 200, 192, 184, 176, 168, 160, 152, 144, 136, 128, 120, 112, 104, 96, 88, 80, 72, 64, 56, 48, 40, 32, 24, 16, 8, 4, 2, 1], 'cudagraph_copy_inputs': False, 'full_cuda_graph': True, 'use_inductor_graph_partition': False, 'pass_config': {}, 'max_capture_size': 512, 'local_cache_dir': None} 2025-11-03T17:59:12.4889052Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:59:12.4916560Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:59:12.4924688Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:59:12.4932772Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:59:12.4940516Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:59:12.4948383Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:59:12.4951037Z (EngineCore_DP0 pid=28497) INFO 11-03 17:59:12 [parallel_state.py:1231] rank 0 in world size 1 is assigned as DP rank 0, PP rank 0, TP rank 0, EP rank 0 2025-11-03T17:59:12.8317951Z (EngineCore_DP0 pid=28497) INFO 11-03 17:59:12 [topk_topp_sampler.py:59] Using FlashInfer for top-p & top-k sampling. 2025-11-03T17:59:13.0407216Z (EngineCore_DP0 pid=28497) INFO 11-03 17:59:13 [gpu_model_runner.py:2840] Starting to load model hmellor/Ilama-3.2-1B... 2025-11-03T17:59:13.3020814Z (EngineCore_DP0 pid=28497) INFO 11-03 17:59:13 [gpu_model_runner.py:2870] Loading model from scratch... 2025-11-03T17:59:13.3022069Z (EngineCore_DP0 pid=28497) INFO 11-03 17:59:13 [transformers.py:493] Using Transformers backend. 2025-11-03T17:59:13.3351147Z (EngineCore_DP0 pid=28497) `torch_dtype` is deprecated! Use `dtype` instead! 2025-11-03T17:59:13.4513228Z (EngineCore_DP0 pid=28497) INFO 11-03 17:59:13 [cuda.py:404] Using Flash Attention backend on V1 engine. 2025-11-03T17:59:14.0232320Z (EngineCore_DP0 pid=28497) INFO 11-03 17:59:14 [gpu_model_runner.py:2902] Model loading took 0.6035 GiB and 0.207320 seconds 2025-11-03T17:59:14.0237661Z (EngineCore_DP0 pid=28497) INFO 11-03 17:59:14 [kv_cache_utils.py:1199] GPU KV cache size: 5,242,880 tokens 2025-11-03T17:59:14.0238795Z (EngineCore_DP0 pid=28497) INFO 11-03 17:59:14 [kv_cache_utils.py:1204] Maximum concurrency for 131,072 tokens per request: 40.00x 2025-11-03T17:59:15.2938151Z (EngineCore_DP0 pid=28497) INFO 11-03 17:59:15 [gc_utils.py:40] GC Debug Config. enabled:False,top_objects:-1 2025-11-03T17:59:15.2952617Z INFO 11-03 17:59:15 [llm.py:337] Supported tasks: ['generate'] 2025-11-03T17:59:15.7785299Z PASSED 2025-11-03T17:59:15.7929133Z models/test_initialization.py::test_can_initialize_large_subset[Glm4ForCausalLM] Fork a new process to run a test 28563 2025-11-03T17:59:15.7940282Z Fork a new process to run a test 0 2025-11-03T17:59:15.8221952Z INFO 11-03 17:59:15 [utils.py:239] non-default args: {'load_format': 'dummy', 'gpu_memory_utilization': 0.8, 'disable_log_stats': True, 'hf_overrides': functools.partial(, model_arch='Glm4ForCausalLM', exist_overrides={}, use_original_num_layers=False), 'model_impl': 'vllm', 'model': 'zai-org/GLM-4-9B-0414'} 2025-11-03T17:59:16.0919740Z 2025-11-03T17:59:16.0921938Z config.json: 0% 0.00/689 [00:00, 'use_cudagraph': True, 'cudagraph_num_of_warmups': 1, 'cudagraph_capture_sizes': [512, 504, 496, 488, 480, 472, 464, 456, 448, 440, 432, 424, 416, 408, 400, 392, 384, 376, 368, 360, 352, 344, 336, 328, 320, 312, 304, 296, 288, 280, 272, 264, 256, 248, 240, 232, 224, 216, 208, 200, 192, 184, 176, 168, 160, 152, 144, 136, 128, 120, 112, 104, 96, 88, 80, 72, 64, 56, 48, 40, 32, 24, 16, 8, 4, 2, 1], 'cudagraph_copy_inputs': False, 'full_cuda_graph': True, 'use_inductor_graph_partition': False, 'pass_config': {}, 'max_capture_size': 512, 'local_cache_dir': None} 2025-11-03T17:59:24.8499090Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:59:24.8529947Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:59:24.8537383Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:59:24.8546563Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:59:24.8554317Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:59:24.8562096Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:59:24.8564859Z (EngineCore_DP0 pid=28641) INFO 11-03 17:59:24 [parallel_state.py:1231] rank 0 in world size 1 is assigned as DP rank 0, PP rank 0, TP rank 0, EP rank 0 2025-11-03T17:59:25.1929188Z (EngineCore_DP0 pid=28641) INFO 11-03 17:59:25 [topk_topp_sampler.py:59] Using FlashInfer for top-p & top-k sampling. 2025-11-03T17:59:25.2827341Z (EngineCore_DP0 pid=28641) INFO 11-03 17:59:25 [gpu_model_runner.py:2840] Starting to load model zai-org/GLM-4-9B-0414... 2025-11-03T17:59:25.5400947Z (EngineCore_DP0 pid=28641) INFO 11-03 17:59:25 [gpu_model_runner.py:2870] Loading model from scratch... 2025-11-03T17:59:25.6229951Z (EngineCore_DP0 pid=28641) INFO 11-03 17:59:25 [cuda.py:404] Using Flash Attention backend on V1 engine. 2025-11-03T17:59:26.1934250Z (EngineCore_DP0 pid=28641) INFO 11-03 17:59:26 [gpu_model_runner.py:2902] Model loading took 2.6973 GiB and 0.143267 seconds 2025-11-03T17:59:26.1938986Z (EngineCore_DP0 pid=28641) INFO 11-03 17:59:26 [kv_cache_utils.py:1199] GPU KV cache size: 10,485,760 tokens 2025-11-03T17:59:26.1940235Z (EngineCore_DP0 pid=28641) INFO 11-03 17:59:26 [kv_cache_utils.py:1204] Maximum concurrency for 32,768 tokens per request: 320.00x 2025-11-03T17:59:28.0708748Z (EngineCore_DP0 pid=28641) INFO 11-03 17:59:28 [gc_utils.py:40] GC Debug Config. enabled:False,top_objects:-1 2025-11-03T17:59:28.0723471Z INFO 11-03 17:59:28 [llm.py:337] Supported tasks: ['generate'] 2025-11-03T17:59:28.5668026Z PASSED 2025-11-03T17:59:28.5810934Z models/test_initialization.py::test_can_initialize_large_subset[TransformersMoEForMultimodalLM] Fork a new process to run a test 28707 2025-11-03T17:59:28.5821905Z Fork a new process to run a test 0 2025-11-03T17:59:28.5825615Z `transformers==4.56.2` installed, but `transformers>=4.57.0.dev0` is required to run this model. 2025-11-03T17:59:28.8707981Z PASSED 2025-11-03T17:59:28.8851689Z models/test_initialization.py::test_can_initialize_large_subset[GraniteMoeForCausalLM] Fork a new process to run a test 28708 2025-11-03T17:59:28.8862274Z Fork a new process to run a test 0 2025-11-03T17:59:28.9142822Z INFO 11-03 17:59:28 [utils.py:239] non-default args: {'load_format': 'dummy', 'gpu_memory_utilization': 0.8, 'disable_log_stats': True, 'hf_overrides': functools.partial(, model_arch='GraniteMoeForCausalLM', exist_overrides={}, use_original_num_layers=False), 'model_impl': 'vllm', 'model': 'ibm/PowerMoE-3b'} 2025-11-03T17:59:29.1404428Z 2025-11-03T17:59:29.1405540Z config.json: 0% 0.00/928 [00:00, 'use_cudagraph': True, 'cudagraph_num_of_warmups': 1, 'cudagraph_capture_sizes': [512, 504, 496, 488, 480, 472, 464, 456, 448, 440, 432, 424, 416, 408, 400, 392, 384, 376, 368, 360, 352, 344, 336, 328, 320, 312, 304, 296, 288, 280, 272, 264, 256, 248, 240, 232, 224, 216, 208, 200, 192, 184, 176, 168, 160, 152, 144, 136, 128, 120, 112, 104, 96, 88, 80, 72, 64, 56, 48, 40, 32, 24, 16, 8, 4, 2, 1], 'cudagraph_copy_inputs': False, 'full_cuda_graph': True, 'use_inductor_graph_partition': False, 'pass_config': {}, 'max_capture_size': 512, 'local_cache_dir': None} 2025-11-03T17:59:37.3315540Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:59:37.3343576Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:59:37.3351374Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:59:37.3358900Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:59:37.3367029Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:59:37.3374694Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:59:37.3377435Z (EngineCore_DP0 pid=28766) INFO 11-03 17:59:37 [parallel_state.py:1231] rank 0 in world size 1 is assigned as DP rank 0, PP rank 0, TP rank 0, EP rank 0 2025-11-03T17:59:37.6770073Z (EngineCore_DP0 pid=28766) INFO 11-03 17:59:37 [topk_topp_sampler.py:59] Using FlashInfer for top-p & top-k sampling. 2025-11-03T17:59:37.7254185Z (EngineCore_DP0 pid=28766) INFO 11-03 17:59:37 [gpu_model_runner.py:2840] Starting to load model ibm/PowerMoE-3b... 2025-11-03T17:59:37.9854468Z (EngineCore_DP0 pid=28766) INFO 11-03 17:59:37 [gpu_model_runner.py:2870] Loading model from scratch... 2025-11-03T17:59:38.0522573Z (EngineCore_DP0 pid=28766) INFO 11-03 17:59:38 [cuda.py:404] Using Flash Attention backend on V1 engine. 2025-11-03T17:59:38.6218549Z (EngineCore_DP0 pid=28766) INFO 11-03 17:59:38 [gpu_model_runner.py:2902] Model loading took 0.1616 GiB and 0.126503 seconds 2025-11-03T17:59:38.6222920Z (EngineCore_DP0 pid=28766) INFO 11-03 17:59:38 [kv_cache_utils.py:1199] GPU KV cache size: 5,242,880 tokens 2025-11-03T17:59:38.6223916Z (EngineCore_DP0 pid=28766) INFO 11-03 17:59:38 [kv_cache_utils.py:1204] Maximum concurrency for 4,096 tokens per request: 1280.00x 2025-11-03T17:59:39.5189815Z (EngineCore_DP0 pid=28766) INFO 11-03 17:59:39 [gc_utils.py:40] GC Debug Config. enabled:False,top_objects:-1 2025-11-03T17:59:39.5203269Z INFO 11-03 17:59:39 [llm.py:337] Supported tasks: ['generate'] 2025-11-03T17:59:39.9593478Z PASSED 2025-11-03T17:59:39.9735162Z models/test_initialization.py::test_can_initialize_large_subset[Ernie4_5_VLMoeForConditionalGeneration] Fork a new process to run a test 28832 2025-11-03T17:59:39.9745144Z Fork a new process to run a test 0 2025-11-03T17:59:40.0030201Z INFO 11-03 17:59:40 [utils.py:239] non-default args: {'trust_remote_code': True, 'load_format': 'dummy', 'gpu_memory_utilization': 0.8, 'disable_log_stats': True, 'hf_overrides': functools.partial(, model_arch='Ernie4_5_VLMoeForConditionalGeneration', exist_overrides={}, use_original_num_layers=False), 'model_impl': 'vllm', 'model': 'baidu/ERNIE-4.5-VL-28B-A3B-PT'} 2025-11-03T17:59:40.0865012Z 2025-11-03T17:59:40.0866939Z config.json: 0.00B [00:00, ?B/s] 2025-11-03T17:59:40.0867296Z config.json: 2.01kB [00:00, 13.2MB/s] 2025-11-03T17:59:40.1511204Z 2025-11-03T17:59:40.1515864Z configuration_ernie4_5_vl.py: 0.00B [00:00, ?B/s] 2025-11-03T17:59:40.1516296Z configuration_ernie4_5_vl.py: 28.3kB [00:00, 54.8MB/s] 2025-11-03T17:59:40.1644145Z A new version of the following files was downloaded from https://huggingface.co/baidu/ERNIE-4.5-VL-28B-A3B-PT: 2025-11-03T17:59:40.1644687Z - configuration_ernie4_5_vl.py 2025-11-03T17:59:40.1645338Z . Make sure to double-check they do not contain any added malicious code. To avoid downloading new versions of the code file, you can pin a revision. 2025-11-03T17:59:40.1994302Z INFO 11-03 17:59:40 [config.py:408] Replacing legacy 'type' key with 'rope_type' 2025-11-03T17:59:40.3919044Z 2025-11-03T17:59:40.3919384Z preprocessor_config.json: 0% 0.00/477 [00:00, 'use_cudagraph': True, 'cudagraph_num_of_warmups': 1, 'cudagraph_capture_sizes': [512, 504, 496, 488, 480, 472, 464, 456, 448, 440, 432, 424, 416, 408, 400, 392, 384, 376, 368, 360, 352, 344, 336, 328, 320, 312, 304, 296, 288, 280, 272, 264, 256, 248, 240, 232, 224, 216, 208, 200, 192, 184, 176, 168, 160, 152, 144, 136, 128, 120, 112, 104, 96, 88, 80, 72, 64, 56, 48, 40, 32, 24, 16, 8, 4, 2, 1], 'cudagraph_copy_inputs': False, 'full_cuda_graph': True, 'use_inductor_graph_partition': False, 'pass_config': {}, 'max_capture_size': 512, 'local_cache_dir': None} 2025-11-03T17:59:51.5358855Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:59:51.5386083Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:59:51.5393606Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:59:51.5401765Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:59:51.5409256Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:59:51.5416722Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T17:59:51.5419461Z (EngineCore_DP0 pid=28911) INFO 11-03 17:59:51 [parallel_state.py:1231] rank 0 in world size 1 is assigned as DP rank 0, PP rank 0, TP rank 0, EP rank 0 2025-11-03T17:59:51.8857124Z (EngineCore_DP0 pid=28911) INFO 11-03 17:59:51 [topk_topp_sampler.py:59] Using FlashInfer for top-p & top-k sampling. 2025-11-03T17:59:57.1393038Z (EngineCore_DP0 pid=28911) 2025-11-03T17:59:57.1393430Z chat_template.json: 0.00B [00:00, ?B/s] 2025-11-03T17:59:57.1393730Z chat_template.json: 3.19kB [00:00, 23.5MB/s] 2025-11-03T17:59:59.2024221Z (EngineCore_DP0 pid=28911) INFO 11-03 17:59:59 [gpu_model_runner.py:2840] Starting to load model baidu/ERNIE-4.5-VL-28B-A3B-PT... 2025-11-03T17:59:59.4893824Z (EngineCore_DP0 pid=28911) INFO 11-03 17:59:59 [gpu_model_runner.py:2870] Loading model from scratch... 2025-11-03T17:59:59.5851270Z (EngineCore_DP0 pid=28911) INFO 11-03 17:59:59 [cuda.py:404] Using Flash Attention backend on V1 engine. 2025-11-03T17:59:59.5878045Z (EngineCore_DP0 pid=28911) WARNING 11-03 17:59:59 [vllm.py:821] `torch.compile` is turned on, but the model baidu/ERNIE-4.5-VL-28B-A3B-PT does not support it. Please open an issue on GitHub if you want it to be supported. 2025-11-03T18:00:00.1228017Z (EngineCore_DP0 pid=28911) INFO 11-03 18:00:00 [gpu_model_runner.py:2902] Model loading took 2.2676 GiB and 0.109294 seconds 2025-11-03T18:00:00.1232178Z (EngineCore_DP0 pid=28911) INFO 11-03 18:00:00 [kv_cache_utils.py:1199] GPU KV cache size: 5,242,880 tokens 2025-11-03T18:00:00.1233348Z (EngineCore_DP0 pid=28911) INFO 11-03 18:00:00 [kv_cache_utils.py:1204] Maximum concurrency for 131,072 tokens per request: 40.00x 2025-11-03T18:00:00.5042677Z (EngineCore_DP0 pid=28911) WARNING 11-03 18:00:00 [tokenizer.py:255] Using a slow tokenizer. This might cause a significant slowdown. Consider using a fast tokenizer instead. 2025-11-03T18:00:01.3939622Z (EngineCore_DP0 pid=28911) INFO 11-03 18:00:01 [gc_utils.py:40] GC Debug Config. enabled:False,top_objects:-1 2025-11-03T18:00:01.3953890Z INFO 11-03 18:00:01 [llm.py:337] Supported tasks: ['generate'] 2025-11-03T18:00:01.8765688Z PASSED 2025-11-03T18:00:01.8907065Z models/test_initialization.py::test_can_initialize_large_subset[Plamo2ForCausalLM] Fork a new process to run a test 28977 2025-11-03T18:00:01.8918361Z Fork a new process to run a test 0 2025-11-03T18:00:01.8922653Z `transformers==4.56.2` installed, but `transformers<=4.55.4` is required to run this model. Reason: HF model uses remote code that is not compatible with latest Transformers 2025-11-03T18:00:02.1832307Z PASSED 2025-11-03T18:00:02.1974211Z models/test_initialization.py::test_can_initialize_large_subset[Qwen2ForProcessRewardModel] Fork a new process to run a test 28978 2025-11-03T18:00:02.1985509Z Fork a new process to run a test 0 2025-11-03T18:00:02.1989538Z `transformers==4.56.2` installed, but `transformers<=4.53` is required to run this model. Reason: HF model uses remote code that is not compatible with latest Transformers 2025-11-03T18:00:02.4901112Z PASSED 2025-11-03T18:00:02.5045698Z models/test_initialization.py::test_can_initialize_large_subset[RobertaForMaskedLM] Fork a new process to run a test 28979 2025-11-03T18:00:02.5056614Z Fork a new process to run a test 0 2025-11-03T18:00:02.5336954Z INFO 11-03 18:00:02 [utils.py:239] non-default args: {'load_format': 'dummy', 'gpu_memory_utilization': 0.8, 'disable_log_stats': True, 'hf_overrides': functools.partial(, model_arch='RobertaForMaskedLM', exist_overrides={}, use_original_num_layers=False), 'model_impl': 'vllm', 'model': 'sentence-transformers/all-roberta-large-v1'} 2025-11-03T18:00:02.6341582Z 2025-11-03T18:00:02.6342978Z config.json: 0% 0.00/650 [00:00, 'use_cudagraph': True, 'cudagraph_num_of_warmups': 1, 'cudagraph_capture_sizes': [256, 248, 240, 232, 224, 216, 208, 200, 192, 184, 176, 168, 160, 152, 144, 136, 128, 120, 112, 104, 96, 88, 80, 72, 64, 56, 48, 40, 32, 24, 16, 8, 4, 2, 1], 'cudagraph_copy_inputs': False, 'full_cuda_graph': False, 'use_inductor_graph_partition': False, 'pass_config': {}, 'max_capture_size': 256, 'local_cache_dir': None} 2025-11-03T18:00:04.8724485Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T18:00:04.8755629Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T18:00:04.8762788Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T18:00:04.8770677Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T18:00:04.8778146Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T18:00:04.8786123Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T18:00:04.8788676Z (EngineCore_DP0 pid=28987) INFO 11-03 18:00:04 [parallel_state.py:1231] rank 0 in world size 1 is assigned as DP rank 0, PP rank 0, TP rank 0, EP rank 0 2025-11-03T18:00:05.2171294Z (EngineCore_DP0 pid=28987) INFO 11-03 18:00:05 [topk_topp_sampler.py:59] Using FlashInfer for top-p & top-k sampling. 2025-11-03T18:00:05.2310972Z (EngineCore_DP0 pid=28987) INFO 11-03 18:00:05 [gpu_model_runner.py:2840] Starting to load model sentence-transformers/all-roberta-large-v1... 2025-11-03T18:00:05.4835623Z (EngineCore_DP0 pid=28987) INFO 11-03 18:00:05 [gpu_model_runner.py:2870] Loading model from scratch... 2025-11-03T18:00:05.4889125Z (EngineCore_DP0 pid=28987) INFO 11-03 18:00:05 [cuda.py:404] Using Flash Attention backend on V1 engine. 2025-11-03T18:00:06.0613297Z (EngineCore_DP0 pid=28987) INFO 11-03 18:00:06 [gpu_model_runner.py:2902] Model loading took 0.1211 GiB and 0.065674 seconds 2025-11-03T18:00:06.2301055Z (EngineCore_DP0 pid=28987) INFO 11-03 18:00:06 [core.py:142] Disabling chunked prefill for model without KVCache 2025-11-03T18:00:06.3027696Z (EngineCore_DP0 pid=28987) INFO 11-03 18:00:06 [vllm.py:433] Only "last" pooling supports chunked prefill and prefix caching; disabling both. 2025-11-03T18:00:06.5645446Z (EngineCore_DP0 pid=28987) INFO 11-03 18:00:06 [gc_utils.py:40] GC Debug Config. enabled:False,top_objects:-1 2025-11-03T18:00:06.5659628Z INFO 11-03 18:00:06 [llm.py:337] Supported tasks: ['embed', 'encode'] 2025-11-03T18:00:06.9876139Z PASSED 2025-11-03T18:00:07.0017396Z models/test_initialization.py::test_can_initialize_large_subset[GPTBigCodeForCausalLM] Fork a new process to run a test 29053 2025-11-03T18:00:07.0028022Z Fork a new process to run a test 0 2025-11-03T18:00:07.0313240Z INFO 11-03 18:00:07 [utils.py:239] non-default args: {'load_format': 'dummy', 'gpu_memory_utilization': 0.8, 'disable_log_stats': True, 'hf_overrides': functools.partial(, model_arch='GPTBigCodeForCausalLM', exist_overrides={}, use_original_num_layers=False), 'model_impl': 'vllm', 'model': 'bigcode/starcoder'} 2025-11-03T18:00:07.1887704Z 2025-11-03T18:00:07.1889060Z config.json: 0% 0.00/1.05k [00:00, 'use_cudagraph': True, 'cudagraph_num_of_warmups': 1, 'cudagraph_capture_sizes': [512, 504, 496, 488, 480, 472, 464, 456, 448, 440, 432, 424, 416, 408, 400, 392, 384, 376, 368, 360, 352, 344, 336, 328, 320, 312, 304, 296, 288, 280, 272, 264, 256, 248, 240, 232, 224, 216, 208, 200, 192, 184, 176, 168, 160, 152, 144, 136, 128, 120, 112, 104, 96, 88, 80, 72, 64, 56, 48, 40, 32, 24, 16, 8, 4, 2, 1], 'cudagraph_copy_inputs': False, 'full_cuda_graph': True, 'use_inductor_graph_partition': False, 'pass_config': {}, 'max_capture_size': 512, 'local_cache_dir': None} 2025-11-03T18:00:15.4520742Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T18:00:15.4543090Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T18:00:15.4550612Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T18:00:15.4557954Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T18:00:15.4566278Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T18:00:15.4574301Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T18:00:15.4577053Z (EngineCore_DP0 pid=29111) INFO 11-03 18:00:15 [parallel_state.py:1231] rank 0 in world size 1 is assigned as DP rank 0, PP rank 0, TP rank 0, EP rank 0 2025-11-03T18:00:15.7899518Z (EngineCore_DP0 pid=29111) INFO 11-03 18:00:15 [topk_topp_sampler.py:59] Using FlashInfer for top-p & top-k sampling. 2025-11-03T18:00:15.8877722Z (EngineCore_DP0 pid=29111) INFO 11-03 18:00:15 [gpu_model_runner.py:2840] Starting to load model bigcode/starcoder... 2025-11-03T18:00:16.1480005Z (EngineCore_DP0 pid=29111) INFO 11-03 18:00:16 [gpu_model_runner.py:2870] Loading model from scratch... 2025-11-03T18:00:16.1521879Z (EngineCore_DP0 pid=29111) INFO 11-03 18:00:16 [cuda.py:404] Using Flash Attention backend on V1 engine. 2025-11-03T18:00:16.7211174Z (EngineCore_DP0 pid=29111) INFO 11-03 18:00:16 [gpu_model_runner.py:2902] Model loading took 1.3634 GiB and 0.063309 seconds 2025-11-03T18:00:16.7214998Z (EngineCore_DP0 pid=29111) INFO 11-03 18:00:16 [kv_cache_utils.py:1199] GPU KV cache size: 20,971,520 tokens 2025-11-03T18:00:16.7216340Z (EngineCore_DP0 pid=29111) INFO 11-03 18:00:16 [kv_cache_utils.py:1204] Maximum concurrency for 8,192 tokens per request: 2560.00x 2025-11-03T18:00:19.0247721Z (EngineCore_DP0 pid=29111) INFO 11-03 18:00:19 [gc_utils.py:40] GC Debug Config. enabled:False,top_objects:-1 2025-11-03T18:00:19.0262280Z INFO 11-03 18:00:19 [llm.py:337] Supported tasks: ['generate'] 2025-11-03T18:00:19.4961706Z PASSED 2025-11-03T18:00:19.5103217Z models/test_initialization.py::test_can_initialize_large_subset[PaliGemmaForConditionalGeneration] Fork a new process to run a test 29177 2025-11-03T18:00:19.5114491Z Fork a new process to run a test 0 2025-11-03T18:00:19.5402396Z INFO 11-03 18:00:19 [utils.py:239] non-default args: {'load_format': 'dummy', 'gpu_memory_utilization': 0.8, 'disable_log_stats': True, 'hf_overrides': functools.partial(, model_arch='PaliGemmaForConditionalGeneration', exist_overrides={}, use_original_num_layers=False), 'model_impl': 'vllm', 'model': 'google/paligemma-3b-mix-224'} 2025-11-03T18:00:19.6625152Z 2025-11-03T18:00:19.6626415Z config.json: 0% 0.00/1.03k [00:00, 'use_cudagraph': True, 'cudagraph_num_of_warmups': 1, 'cudagraph_capture_sizes': [512, 504, 496, 488, 480, 472, 464, 456, 448, 440, 432, 424, 416, 408, 400, 392, 384, 376, 368, 360, 352, 344, 336, 328, 320, 312, 304, 296, 288, 280, 272, 264, 256, 248, 240, 232, 224, 216, 208, 200, 192, 184, 176, 168, 160, 152, 144, 136, 128, 120, 112, 104, 96, 88, 80, 72, 64, 56, 48, 40, 32, 24, 16, 8, 4, 2, 1], 'cudagraph_copy_inputs': False, 'full_cuda_graph': True, 'use_inductor_graph_partition': False, 'pass_config': {}, 'max_capture_size': 512, 'local_cache_dir': None} 2025-11-03T18:00:29.8999161Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T18:00:29.9030042Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T18:00:29.9036936Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T18:00:29.9044325Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T18:00:29.9051979Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T18:00:29.9060359Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T18:00:29.9062862Z (EngineCore_DP0 pid=29257) INFO 11-03 18:00:29 [parallel_state.py:1231] rank 0 in world size 1 is assigned as DP rank 0, PP rank 0, TP rank 0, EP rank 0 2025-11-03T18:00:30.2455489Z (EngineCore_DP0 pid=29257) INFO 11-03 18:00:30 [topk_topp_sampler.py:59] Using FlashInfer for top-p & top-k sampling. 2025-11-03T18:00:30.8088938Z (EngineCore_DP0 pid=29257) Using a slow image processor as `use_fast` is unset and a slow processor was saved with this model. `use_fast=True` will be the default behavior in v4.52, even if the model was saved with a slow processor. This will result in minor differences in outputs. You'll still be able to use a slow processor with `use_fast=False`. 2025-11-03T18:00:35.8611684Z (EngineCore_DP0 pid=29257) You are passing both `text` and `images` to `PaliGemmaProcessor`. The processor expects special image tokens in the text, as many tokens as there are images per each text. It is recommended to add `` tokens in the very beginning of your text. For this call, we will infer how many images each text has and add special tokens. 2025-11-03T18:00:36.2750266Z (EngineCore_DP0 pid=29257) INFO 11-03 18:00:36 [gpu_model_runner.py:2840] Starting to load model google/paligemma-3b-mix-224... 2025-11-03T18:00:36.5385291Z (EngineCore_DP0 pid=29257) INFO 11-03 18:00:36 [gpu_model_runner.py:2870] Loading model from scratch... 2025-11-03T18:00:36.5433961Z (EngineCore_DP0 pid=29257) INFO 11-03 18:00:36 [layer.py:497] MultiHeadAttention attn_backend: _Backend.XFORMERS, use_upstream_fa: False 2025-11-03T18:00:36.5689072Z (EngineCore_DP0 pid=29257) INFO 11-03 18:00:36 [cuda.py:404] Using Flash Attention backend on V1 engine. 2025-11-03T18:00:36.5699583Z (EngineCore_DP0 pid=29257) WARNING 11-03 18:00:36 [gemma.py:67] Gemma's activation function was incorrectly set to exact GeLU in the config JSON file when it was initially released. Changing the activation function to approximate GeLU (`gelu_pytorch_tanh`). If you want to use the legacy `gelu_pytorch_tanh`, edit the config JSON to set `hidden_activation=gelu_pytorch_tanh` instead of `hidden_act`. See https://github.com/huggingface/transformers/pull/29402 for more details. 2025-11-03T18:00:37.1501127Z (EngineCore_DP0 pid=29257) INFO 11-03 18:00:37 [gpu_model_runner.py:2902] Model loading took 1.2249 GiB and 0.087306 seconds 2025-11-03T18:00:37.1505401Z (EngineCore_DP0 pid=29257) INFO 11-03 18:00:37 [kv_cache_utils.py:1199] GPU KV cache size: 10,485,760 tokens 2025-11-03T18:00:37.1506290Z (EngineCore_DP0 pid=29257) INFO 11-03 18:00:37 [kv_cache_utils.py:1204] Maximum concurrency for 8,192 tokens per request: 1280.00x 2025-11-03T18:00:37.9421758Z (EngineCore_DP0 pid=29257) You are passing both `text` and `images` to `PaliGemmaProcessor`. The processor expects special image tokens in the text, as many tokens as there are images per each text. It is recommended to add `` tokens in the very beginning of your text. For this call, we will infer how many images each text has and add special tokens. 2025-11-03T18:00:39.1180551Z (EngineCore_DP0 pid=29257) INFO 11-03 18:00:39 [gc_utils.py:40] GC Debug Config. enabled:False,top_objects:-1 2025-11-03T18:00:39.1194991Z INFO 11-03 18:00:39 [llm.py:337] Supported tasks: ['generate'] 2025-11-03T18:00:39.6215783Z PASSED 2025-11-03T18:00:39.6358041Z models/test_initialization.py::test_can_initialize_large_subset[GPTJForCausalLM] Fork a new process to run a test 29339 2025-11-03T18:00:39.6369767Z Fork a new process to run a test 0 2025-11-03T18:00:39.6654805Z INFO 11-03 18:00:39 [utils.py:239] non-default args: {'load_format': 'dummy', 'gpu_memory_utilization': 0.8, 'disable_log_stats': True, 'hf_overrides': functools.partial(, model_arch='GPTJForCausalLM', exist_overrides={}, use_original_num_layers=False), 'model_impl': 'vllm', 'model': 'Milos/slovak-gpt-j-405M'} 2025-11-03T18:00:39.7576762Z 2025-11-03T18:00:39.7578132Z config.json: 0% 0.00/836 [00:00, 'use_cudagraph': True, 'cudagraph_num_of_warmups': 1, 'cudagraph_capture_sizes': [512, 504, 496, 488, 480, 472, 464, 456, 448, 440, 432, 424, 416, 408, 400, 392, 384, 376, 368, 360, 352, 344, 336, 328, 320, 312, 304, 296, 288, 280, 272, 264, 256, 248, 240, 232, 224, 216, 208, 200, 192, 184, 176, 168, 160, 152, 144, 136, 128, 120, 112, 104, 96, 88, 80, 72, 64, 56, 48, 40, 32, 24, 16, 8, 4, 2, 1], 'cudagraph_copy_inputs': False, 'full_cuda_graph': True, 'use_inductor_graph_partition': False, 'pass_config': {}, 'max_capture_size': 512, 'local_cache_dir': None} 2025-11-03T18:00:47.6504293Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T18:00:47.6532010Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T18:00:47.6539455Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T18:00:47.6547466Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T18:00:47.6555253Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T18:00:47.6563458Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T18:00:47.6566090Z (EngineCore_DP0 pid=29397) INFO 11-03 18:00:47 [parallel_state.py:1231] rank 0 in world size 1 is assigned as DP rank 0, PP rank 0, TP rank 0, EP rank 0 2025-11-03T18:00:47.9885909Z (EngineCore_DP0 pid=29397) INFO 11-03 18:00:47 [topk_topp_sampler.py:59] Using FlashInfer for top-p & top-k sampling. 2025-11-03T18:00:48.0288968Z (EngineCore_DP0 pid=29397) INFO 11-03 18:00:48 [gpu_model_runner.py:2840] Starting to load model Milos/slovak-gpt-j-405M... 2025-11-03T18:00:48.2852310Z (EngineCore_DP0 pid=29397) INFO 11-03 18:00:48 [gpu_model_runner.py:2870] Loading model from scratch... 2025-11-03T18:00:48.3505530Z (EngineCore_DP0 pid=29397) INFO 11-03 18:00:48 [cuda.py:404] Using Flash Attention backend on V1 engine. 2025-11-03T18:00:48.9096147Z (EngineCore_DP0 pid=29397) INFO 11-03 18:00:48 [gpu_model_runner.py:2902] Model loading took 0.2157 GiB and 0.124343 seconds 2025-11-03T18:00:48.9101719Z (EngineCore_DP0 pid=29397) INFO 11-03 18:00:48 [kv_cache_utils.py:1199] GPU KV cache size: 2,621,440 tokens 2025-11-03T18:00:48.9102663Z (EngineCore_DP0 pid=29397) INFO 11-03 18:00:48 [kv_cache_utils.py:1204] Maximum concurrency for 2,048 tokens per request: 1280.00x 2025-11-03T18:00:49.7202745Z (EngineCore_DP0 pid=29397) INFO 11-03 18:00:49 [gc_utils.py:40] GC Debug Config. enabled:False,top_objects:-1 2025-11-03T18:00:49.7217340Z INFO 11-03 18:00:49 [llm.py:337] Supported tasks: ['generate'] 2025-11-03T18:00:50.1587789Z PASSED 2025-11-03T18:00:50.1729408Z models/test_initialization.py::test_can_initialize_large_subset[BaiChuanForCausalLM] Fork a new process to run a test 29463 2025-11-03T18:00:50.1740398Z Fork a new process to run a test 0 2025-11-03T18:00:50.2017922Z INFO 11-03 18:00:50 [utils.py:239] non-default args: {'trust_remote_code': True, 'load_format': 'dummy', 'gpu_memory_utilization': 0.8, 'disable_log_stats': True, 'hf_overrides': functools.partial(, model_arch='BaiChuanForCausalLM', exist_overrides={}, use_original_num_layers=False), 'model_impl': 'vllm', 'model': 'baichuan-inc/Baichuan-7B'} 2025-11-03T18:00:50.3212870Z 2025-11-03T18:00:50.3213789Z config.json: 0% 0.00/656 [00:00, 'use_cudagraph': True, 'cudagraph_num_of_warmups': 1, 'cudagraph_capture_sizes': [512, 504, 496, 488, 480, 472, 464, 456, 448, 440, 432, 424, 416, 408, 400, 392, 384, 376, 368, 360, 352, 344, 336, 328, 320, 312, 304, 296, 288, 280, 272, 264, 256, 248, 240, 232, 224, 216, 208, 200, 192, 184, 176, 168, 160, 152, 144, 136, 128, 120, 112, 104, 96, 88, 80, 72, 64, 56, 48, 40, 32, 24, 16, 8, 4, 2, 1], 'cudagraph_copy_inputs': False, 'full_cuda_graph': True, 'use_inductor_graph_partition': False, 'pass_config': {}, 'max_capture_size': 512, 'local_cache_dir': None} 2025-11-03T18:00:58.6513902Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T18:00:58.6542333Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T18:00:58.6550121Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T18:00:58.6558426Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T18:00:58.6566580Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T18:00:58.6575590Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T18:00:58.6578509Z (EngineCore_DP0 pid=29541) INFO 11-03 18:00:58 [parallel_state.py:1231] rank 0 in world size 1 is assigned as DP rank 0, PP rank 0, TP rank 0, EP rank 0 2025-11-03T18:00:58.9959352Z (EngineCore_DP0 pid=29541) INFO 11-03 18:00:58 [topk_topp_sampler.py:59] Using FlashInfer for top-p & top-k sampling. 2025-11-03T18:00:59.0635539Z (EngineCore_DP0 pid=29541) INFO 11-03 18:00:59 [gpu_model_runner.py:2840] Starting to load model baichuan-inc/Baichuan-7B... 2025-11-03T18:00:59.3201310Z (EngineCore_DP0 pid=29541) INFO 11-03 18:00:59 [gpu_model_runner.py:2870] Loading model from scratch... 2025-11-03T18:00:59.3852392Z (EngineCore_DP0 pid=29541) INFO 11-03 18:00:59 [cuda.py:404] Using Flash Attention backend on V1 engine. 2025-11-03T18:00:59.9658857Z (EngineCore_DP0 pid=29541) INFO 11-03 18:00:59 [gpu_model_runner.py:2902] Model loading took 1.3545 GiB and 0.125776 seconds 2025-11-03T18:00:59.9663811Z (EngineCore_DP0 pid=29541) INFO 11-03 18:00:59 [kv_cache_utils.py:1199] GPU KV cache size: 655,360 tokens 2025-11-03T18:00:59.9665056Z (EngineCore_DP0 pid=29541) INFO 11-03 18:00:59 [kv_cache_utils.py:1204] Maximum concurrency for 4,096 tokens per request: 160.00x 2025-11-03T18:01:00.1872682Z (EngineCore_DP0 pid=29541) WARNING 11-03 18:01:00 [tokenizer.py:255] Using a slow tokenizer. This might cause a significant slowdown. Consider using a fast tokenizer instead. 2025-11-03T18:01:00.5999477Z (EngineCore_DP0 pid=29541) INFO 11-03 18:01:00 [gc_utils.py:40] GC Debug Config. enabled:False,top_objects:-1 2025-11-03T18:01:00.6015543Z INFO 11-03 18:01:00 [llm.py:337] Supported tasks: ['generate'] 2025-11-03T18:01:01.0605823Z PASSED 2025-11-03T18:01:01.0747499Z models/test_initialization.py::test_implicit_converted_models[GemmaForSequenceClassification] Fork a new process to run a test 29607 2025-11-03T18:01:01.0758156Z Fork a new process to run a test 0 2025-11-03T18:01:01.1042239Z INFO 11-03 18:01:01 [utils.py:239] non-default args: {'load_format': 'dummy', 'gpu_memory_utilization': 0.8, 'disable_log_stats': True, 'hf_overrides': functools.partial(, model_arch='GemmaForSequenceClassification', exist_overrides={'architectures': ['GemmaForSequenceClassification'], 'classifier_from_token': ['Yes'], 'method': 'no_post_processing'}, use_original_num_layers=False), 'model_impl': 'vllm', 'model': 'BAAI/bge-reranker-v2-gemma'} 2025-11-03T18:01:01.2358943Z 2025-11-03T18:01:01.2359935Z config.json: 0% 0.00/659 [00:00, 'use_cudagraph': True, 'cudagraph_num_of_warmups': 1, 'cudagraph_capture_sizes': [512, 504, 496, 488, 480, 472, 464, 456, 448, 440, 432, 424, 416, 408, 400, 392, 384, 376, 368, 360, 352, 344, 336, 328, 320, 312, 304, 296, 288, 280, 272, 264, 256, 248, 240, 232, 224, 216, 208, 200, 192, 184, 176, 168, 160, 152, 144, 136, 128, 120, 112, 104, 96, 88, 80, 72, 64, 56, 48, 40, 32, 24, 16, 8, 4, 2, 1], 'cudagraph_copy_inputs': False, 'full_cuda_graph': False, 'use_inductor_graph_partition': False, 'pass_config': {}, 'max_capture_size': 512, 'local_cache_dir': None} 2025-11-03T18:01:04.1050622Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T18:01:04.1077993Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T18:01:04.1085526Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T18:01:04.1093545Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T18:01:04.1101036Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T18:01:04.1108587Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T18:01:04.1111265Z (EngineCore_DP0 pid=29636) INFO 11-03 18:01:04 [parallel_state.py:1231] rank 0 in world size 1 is assigned as DP rank 0, PP rank 0, TP rank 0, EP rank 0 2025-11-03T18:01:04.4483522Z (EngineCore_DP0 pid=29636) INFO 11-03 18:01:04 [topk_topp_sampler.py:59] Using FlashInfer for top-p & top-k sampling. 2025-11-03T18:01:04.4890196Z (EngineCore_DP0 pid=29636) INFO 11-03 18:01:04 [gpu_model_runner.py:2840] Starting to load model BAAI/bge-reranker-v2-gemma... 2025-11-03T18:01:04.7447722Z (EngineCore_DP0 pid=29636) INFO 11-03 18:01:04 [gpu_model_runner.py:2870] Loading model from scratch... 2025-11-03T18:01:04.8134121Z (EngineCore_DP0 pid=29636) INFO 11-03 18:01:04 [cuda.py:404] Using Flash Attention backend on V1 engine. 2025-11-03T18:01:04.8144743Z (EngineCore_DP0 pid=29636) WARNING 11-03 18:01:04 [gemma.py:67] Gemma's activation function was incorrectly set to exact GeLU in the config JSON file when it was initially released. Changing the activation function to approximate GeLU (`gelu_pytorch_tanh`). If you want to use the legacy `gelu`, edit the config JSON to set `hidden_activation=gelu` instead of `hidden_act`. See https://github.com/huggingface/transformers/pull/29402 for more details. 2025-11-03T18:01:05.3757569Z (EngineCore_DP0 pid=29636) INFO 11-03 18:01:05 [gpu_model_runner.py:2902] Model loading took 1.1856 GiB and 0.128830 seconds 2025-11-03T18:01:05.3761236Z (EngineCore_DP0 pid=29636) INFO 11-03 18:01:05 [kv_cache_utils.py:1199] GPU KV cache size: 10,485,760 tokens 2025-11-03T18:01:05.3762131Z (EngineCore_DP0 pid=29636) INFO 11-03 18:01:05 [kv_cache_utils.py:1204] Maximum concurrency for 8,192 tokens per request: 1280.00x 2025-11-03T18:01:07.4392945Z (EngineCore_DP0 pid=29636) INFO 11-03 18:01:07 [gc_utils.py:40] GC Debug Config. enabled:False,top_objects:-1 2025-11-03T18:01:07.4406767Z INFO 11-03 18:01:07 [llm.py:337] Supported tasks: ['score', 'classify'] 2025-11-03T18:01:07.9224637Z PASSED 2025-11-03T18:01:07.9367712Z models/test_initialization.py::test_implicit_converted_models[LlamaForSequenceClassification] Fork a new process to run a test 29702 2025-11-03T18:01:07.9377428Z Fork a new process to run a test 0 2025-11-03T18:01:07.9654366Z INFO 11-03 18:01:07 [utils.py:239] non-default args: {'load_format': 'dummy', 'gpu_memory_utilization': 0.8, 'disable_log_stats': True, 'hf_overrides': functools.partial(, model_arch='LlamaForSequenceClassification', exist_overrides={}, use_original_num_layers=False), 'model_impl': 'vllm', 'model': 'Skywork/Skywork-Reward-V2-Llama-3.2-1B'} 2025-11-03T18:01:08.0776732Z 2025-11-03T18:01:08.0777468Z config.json: 0% 0.00/991 [00:00, 'use_cudagraph': True, 'cudagraph_num_of_warmups': 1, 'cudagraph_capture_sizes': [512, 504, 496, 488, 480, 472, 464, 456, 448, 440, 432, 424, 416, 408, 400, 392, 384, 376, 368, 360, 352, 344, 336, 328, 320, 312, 304, 296, 288, 280, 272, 264, 256, 248, 240, 232, 224, 216, 208, 200, 192, 184, 176, 168, 160, 152, 144, 136, 128, 120, 112, 104, 96, 88, 80, 72, 64, 56, 48, 40, 32, 24, 16, 8, 4, 2, 1], 'cudagraph_copy_inputs': False, 'full_cuda_graph': False, 'use_inductor_graph_partition': False, 'pass_config': {}, 'max_capture_size': 512, 'local_cache_dir': None} 2025-11-03T18:01:11.2375702Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T18:01:11.2403133Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T18:01:11.2410899Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T18:01:11.2419416Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T18:01:11.2427313Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T18:01:11.2435090Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T18:01:11.2437862Z (EngineCore_DP0 pid=29730) INFO 11-03 18:01:11 [parallel_state.py:1231] rank 0 in world size 1 is assigned as DP rank 0, PP rank 0, TP rank 0, EP rank 0 2025-11-03T18:01:11.5783824Z (EngineCore_DP0 pid=29730) INFO 11-03 18:01:11 [topk_topp_sampler.py:59] Using FlashInfer for top-p & top-k sampling. 2025-11-03T18:01:11.6385543Z (EngineCore_DP0 pid=29730) INFO 11-03 18:01:11 [gpu_model_runner.py:2840] Starting to load model Skywork/Skywork-Reward-V2-Llama-3.2-1B... 2025-11-03T18:01:11.8969929Z (EngineCore_DP0 pid=29730) INFO 11-03 18:01:11 [gpu_model_runner.py:2870] Loading model from scratch... 2025-11-03T18:01:12.0025139Z (EngineCore_DP0 pid=29730) INFO 11-03 18:01:12 [cuda.py:404] Using Flash Attention backend on V1 engine. 2025-11-03T18:01:12.5731977Z (EngineCore_DP0 pid=29730) INFO 11-03 18:01:12 [gpu_model_runner.py:2902] Model loading took 0.6192 GiB and 0.165214 seconds 2025-11-03T18:01:12.5735725Z (EngineCore_DP0 pid=29730) INFO 11-03 18:01:12 [kv_cache_utils.py:1199] GPU KV cache size: 5,242,880 tokens 2025-11-03T18:01:12.5737605Z (EngineCore_DP0 pid=29730) INFO 11-03 18:01:12 [kv_cache_utils.py:1204] Maximum concurrency for 131,072 tokens per request: 40.00x 2025-11-03T18:01:13.9156427Z (EngineCore_DP0 pid=29730) INFO 11-03 18:01:13 [gc_utils.py:40] GC Debug Config. enabled:False,top_objects:-1 2025-11-03T18:01:13.9170061Z INFO 11-03 18:01:13 [llm.py:337] Supported tasks: ['score', 'classify'] 2025-11-03T18:01:14.4055249Z PASSED 2025-11-03T18:01:14.4193973Z models/test_initialization.py::test_implicit_converted_models[Qwen2ForSequenceClassification] Fork a new process to run a test 29796 2025-11-03T18:01:14.4205159Z Fork a new process to run a test 0 2025-11-03T18:01:14.4472896Z INFO 11-03 18:01:14 [utils.py:239] non-default args: {'load_format': 'dummy', 'gpu_memory_utilization': 0.8, 'disable_log_stats': True, 'hf_overrides': functools.partial(, model_arch='Qwen2ForSequenceClassification', exist_overrides={}, use_original_num_layers=False), 'model_impl': 'vllm', 'model': 'jason9693/Qwen2.5-1.5B-apeach'} 2025-11-03T18:01:14.5249055Z 2025-11-03T18:01:14.5250006Z config.json: 0% 0.00/950 [00:00, 'use_cudagraph': True, 'cudagraph_num_of_warmups': 1, 'cudagraph_capture_sizes': [512, 504, 496, 488, 480, 472, 464, 456, 448, 440, 432, 424, 416, 408, 400, 392, 384, 376, 368, 360, 352, 344, 336, 328, 320, 312, 304, 296, 288, 280, 272, 264, 256, 248, 240, 232, 224, 216, 208, 200, 192, 184, 176, 168, 160, 152, 144, 136, 128, 120, 112, 104, 96, 88, 80, 72, 64, 56, 48, 40, 32, 24, 16, 8, 4, 2, 1], 'cudagraph_copy_inputs': False, 'full_cuda_graph': False, 'use_inductor_graph_partition': False, 'pass_config': {}, 'max_capture_size': 512, 'local_cache_dir': None} 2025-11-03T18:01:17.0905764Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T18:01:17.0934284Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T18:01:17.0941006Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T18:01:17.0948914Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T18:01:17.0957049Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T18:01:17.0964879Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T18:01:17.0967589Z (EngineCore_DP0 pid=29826) INFO 11-03 18:01:17 [parallel_state.py:1231] rank 0 in world size 1 is assigned as DP rank 0, PP rank 0, TP rank 0, EP rank 0 2025-11-03T18:01:17.4312019Z (EngineCore_DP0 pid=29826) INFO 11-03 18:01:17 [topk_topp_sampler.py:59] Using FlashInfer for top-p & top-k sampling. 2025-11-03T18:01:17.4913844Z (EngineCore_DP0 pid=29826) INFO 11-03 18:01:17 [gpu_model_runner.py:2840] Starting to load model jason9693/Qwen2.5-1.5B-apeach... 2025-11-03T18:01:17.7503933Z (EngineCore_DP0 pid=29826) INFO 11-03 18:01:17 [gpu_model_runner.py:2870] Loading model from scratch... 2025-11-03T18:01:17.8203775Z (EngineCore_DP0 pid=29826) INFO 11-03 18:01:17 [cuda.py:404] Using Flash Attention backend on V1 engine. 2025-11-03T18:01:18.3951015Z (EngineCore_DP0 pid=29826) INFO 11-03 18:01:18 [gpu_model_runner.py:2902] Model loading took 0.5540 GiB and 0.129675 seconds 2025-11-03T18:01:18.3955223Z (EngineCore_DP0 pid=29826) INFO 11-03 18:01:18 [kv_cache_utils.py:1199] GPU KV cache size: 10,485,760 tokens 2025-11-03T18:01:18.3956194Z (EngineCore_DP0 pid=29826) INFO 11-03 18:01:18 [kv_cache_utils.py:1204] Maximum concurrency for 131,072 tokens per request: 80.00x 2025-11-03T18:01:20.0271456Z (EngineCore_DP0 pid=29826) INFO 11-03 18:01:20 [gc_utils.py:40] GC Debug Config. enabled:False,top_objects:-1 2025-11-03T18:01:20.0284666Z INFO 11-03 18:01:20 [llm.py:337] Supported tasks: ['classify'] 2025-11-03T18:01:20.5094568Z PASSED 2025-11-03T18:01:20.5233255Z models/test_initialization.py::test_implicit_converted_models[Qwen3ForSequenceClassification] Fork a new process to run a test 29892 2025-11-03T18:01:20.5244847Z Fork a new process to run a test 0 2025-11-03T18:01:20.5525572Z INFO 11-03 18:01:20 [utils.py:239] non-default args: {'load_format': 'dummy', 'gpu_memory_utilization': 0.8, 'disable_log_stats': True, 'hf_overrides': functools.partial(, model_arch='Qwen3ForSequenceClassification', exist_overrides={}, use_original_num_layers=False), 'model_impl': 'vllm', 'model': 'tomaarsen/Qwen3-Reranker-0.6B-seq-cls'} 2025-11-03T18:01:20.8037497Z 2025-11-03T18:01:20.8038641Z config.json: 0% 0.00/844 [00:00, 'use_cudagraph': True, 'cudagraph_num_of_warmups': 1, 'cudagraph_capture_sizes': [512, 504, 496, 488, 480, 472, 464, 456, 448, 440, 432, 424, 416, 408, 400, 392, 384, 376, 368, 360, 352, 344, 336, 328, 320, 312, 304, 296, 288, 280, 272, 264, 256, 248, 240, 232, 224, 216, 208, 200, 192, 184, 176, 168, 160, 152, 144, 136, 128, 120, 112, 104, 96, 88, 80, 72, 64, 56, 48, 40, 32, 24, 16, 8, 4, 2, 1], 'cudagraph_copy_inputs': False, 'full_cuda_graph': False, 'use_inductor_graph_partition': False, 'pass_config': {}, 'max_capture_size': 512, 'local_cache_dir': None} 2025-11-03T18:01:23.0606673Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T18:01:23.0634176Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T18:01:23.0641414Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T18:01:23.0649285Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T18:01:23.0656758Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T18:01:23.0664988Z [Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0 2025-11-03T18:01:23.0667222Z (EngineCore_DP0 pid=29920) INFO 11-03 18:01:23 [parallel_state.py:1231] rank 0 in world size 1 is assigned as DP rank 0, PP rank 0, TP rank 0, EP rank 0 2025-11-03T18:01:23.4067167Z (EngineCore_DP0 pid=29920) INFO 11-03 18:01:23 [topk_topp_sampler.py:59] Using FlashInfer for top-p & top-k sampling. 2025-11-03T18:01:23.4429091Z (EngineCore_DP0 pid=29920) INFO 11-03 18:01:23 [gpu_model_runner.py:2840] Starting to load model tomaarsen/Qwen3-Reranker-0.6B-seq-cls... 2025-11-03T18:01:23.7011749Z (EngineCore_DP0 pid=29920) INFO 11-03 18:01:23 [gpu_model_runner.py:2870] Loading model from scratch... 2025-11-03T18:01:23.7748170Z (EngineCore_DP0 pid=29920) INFO 11-03 18:01:23 [cuda.py:404] Using Flash Attention backend on V1 engine. 2025-11-03T18:01:24.3569202Z (EngineCore_DP0 pid=29920) INFO 11-03 18:01:24 [gpu_model_runner.py:2902] Model loading took 0.3284 GiB and 0.135755 seconds 2025-11-03T18:01:24.3572899Z (EngineCore_DP0 pid=29920) INFO 11-03 18:01:24 [kv_cache_utils.py:1199] GPU KV cache size: 2,621,440 tokens 2025-11-03T18:01:24.3573892Z (EngineCore_DP0 pid=29920) INFO 11-03 18:01:24 [kv_cache_utils.py:1204] Maximum concurrency for 40,960 tokens per request: 64.00x 2025-11-03T18:01:25.2772178Z (EngineCore_DP0 pid=29920) INFO 11-03 18:01:25 [gc_utils.py:40] GC Debug Config. enabled:False,top_objects:-1 2025-11-03T18:01:25.2796243Z INFO 11-03 18:01:25 [llm.py:337] Supported tasks: ['score', 'classify'] 2025-11-03T18:01:25.7399603Z PASSED 2025-11-03T18:01:25.7399907Z 2025-11-03T18:01:25.7400467Z =============================== warnings summary =============================== 2025-11-03T18:01:25.7401291Z ../../../../../../opt/conda/envs/py_3.12/lib/python3.12/site-packages/torch/cuda/__init__.py:65 2025-11-03T18:01:25.7402615Z /opt/conda/envs/py_3.12/lib/python3.12/site-packages/torch/cuda/__init__.py:65: FutureWarning: The pynvml package is deprecated. Please install nvidia-ml-py instead. If you did not install pynvml directly, please report this to the maintainers of the package that installed pynvml for you. 2025-11-03T18:01:25.7403894Z import pynvml # type: ignore[import] 2025-11-03T18:01:25.7404074Z 2025-11-03T18:01:25.7404171Z :488 2025-11-03T18:01:25.7404580Z :488: DeprecationWarning: builtin type SwigPyPacked has no __module__ attribute 2025-11-03T18:01:25.7404921Z 2025-11-03T18:01:25.7405045Z :488 2025-11-03T18:01:25.7405621Z :488: DeprecationWarning: builtin type SwigPyObject has no __module__ attribute 2025-11-03T18:01:25.7406095Z 2025-11-03T18:01:25.7406552Z ../../../../../../opt/conda/envs/py_3.12/lib/python3.12/site-packages/schemathesis/generation/coverage.py:305 2025-11-03T18:01:25.7407753Z /opt/conda/envs/py_3.12/lib/python3.12/site-packages/schemathesis/generation/coverage.py:305: DeprecationWarning: jsonschema.exceptions.RefResolutionError is deprecated as of version 4.18.0. If you wish to catch potential reference resolution errors, directly catch referencing.exceptions.Unresolvable. 2025-11-03T18:01:25.7408815Z ref_error: type[Exception] = jsonschema.RefResolutionError, 2025-11-03T18:01:25.7409027Z 2025-11-03T18:01:25.7410669Z tests/models/test_initialization.py: 224 warnings 2025-11-03T18:01:25.7411445Z /var/lib/jenkins/workspace/vllm/tests/utils.py:872: DeprecationWarning: This process (pid=3677) is multi-threaded, use of fork() may lead to deadlocks in the child. 2025-11-03T18:01:25.7412025Z pid = os.fork() 2025-11-03T18:01:25.7412131Z 2025-11-03T18:01:25.7412327Z -- Docs: https://docs.pytest.org/en/stable/how-to/capture-warnings.html 2025-11-03T18:01:25.7415062Z ========== 224 passed, 1 skipped, 228 warnings in 2680.85s (0:44:40) =========== 2025-11-03T18:01:26.0004928Z sys:1: DeprecationWarning: builtin type swigvarlink has no __module__ attribute 2025-11-03T18:01:27.1418875Z 2025-11-03 18:01:27,141 [INFO] cli.lib.core.vllm.lib: Finish running step: pytest -v -s models/test_initialization.py 2025-11-03T18:01:27.1419850Z 2025-11-03 18:01:27,141 [INFO] cli.lib.core.vllm.lib: Done. All tests passed 2025-11-03T18:01:27.1573637Z 2025-11-03 18:01:27,156 [INFO] cli.lib.core.vllm.vllm_test: Double check installed packages 2025-11-03T18:01:27.3218301Z 2025-11-03 18:01:27,321 [INFO] cli.lib.common.pip_helper: torch already exist with version: 2.10.0a0+git3f6538f 2025-11-03T18:01:27.3233608Z 2025-11-03 18:01:27,323 [INFO] cli.lib.common.pip_helper: xformers already exist with version: 0.0.33+5d4b92a5.d20251103 2025-11-03T18:01:27.3275486Z 2025-11-03 18:01:27,327 [INFO] cli.lib.common.pip_helper: torchvision already exist with version: 0.25.0a0+cfbc5c2 2025-11-03T18:01:27.3300293Z 2025-11-03 18:01:27,329 [INFO] cli.lib.common.pip_helper: torchaudio already exist with version: 2.10.0a0+3b0e7a6 2025-11-03T18:01:27.3408538Z 2025-11-03 18:01:27,340 [INFO] cli.lib.common.pip_helper: vllm already exist with version: 0.11.0rc2.dev389+ge51928192.d20251103 2025-11-03T18:01:27.3416332Z 2025-11-03 18:01:27,341 [INFO] cli.lib.core.vllm.vllm_test: Done. checked installed packages 2025-11-03T18:01:28.4318331Z + sccache_epilogue 2025-11-03T18:01:28.4368808Z + echo '::group::Sccache Compilation Log' 2025-11-03T18:01:28.4369444Z ##[group]Sccache Compilation Log 2025-11-03T18:01:28.4369766Z + echo '=================== sccache compilation log ===================' 2025-11-03T18:01:28.4370131Z =================== sccache compilation log =================== 2025-11-03T18:01:28.4376120Z + python /var/lib/jenkins/workspace/.ci/pytorch/print_sccache_log.py /var/lib/jenkins/sccache_error.log 2025-11-03T18:01:28.4855533Z + echo '=========== If your build fails, please take a look at the log above for possible reasons ===========' 2025-11-03T18:01:28.4856211Z =========== If your build fails, please take a look at the log above for possible reasons =========== 2025-11-03T18:01:28.4856647Z + sccache --show-stats 2025-11-03T18:01:28.4954727Z Compile requests 13 2025-11-03T18:01:28.4955173Z Compile requests executed 0 2025-11-03T18:01:28.4955463Z Cache hits 0 2025-11-03T18:01:28.4955729Z Cache misses 0 2025-11-03T18:01:28.4955990Z Cache hits rate - 2025-11-03T18:01:28.4956255Z Cache timeouts 0 2025-11-03T18:01:28.4956514Z Cache read errors 0 2025-11-03T18:01:28.4956774Z Forced recaches 0 2025-11-03T18:01:28.4957021Z Cache write errors 0 2025-11-03T18:01:28.4957278Z Cache errors 0 2025-11-03T18:01:28.4957551Z Compilations 0 2025-11-03T18:01:28.4957815Z Compilation failures 0 2025-11-03T18:01:28.4958085Z Non-cacheable compilations 0 2025-11-03T18:01:28.4958352Z Non-cacheable calls 0 2025-11-03T18:01:28.4958672Z Non-compilation calls 13 2025-11-03T18:01:28.4959173Z Unsupported compiler calls 0 2025-11-03T18:01:28.4959484Z Average cache write 0.000 s 2025-11-03T18:01:28.4959776Z Average compiler 0.000 s 2025-11-03T18:01:28.4960066Z Average cache read hit 0.000 s 2025-11-03T18:01:28.4960350Z Failed distributed compilations 0 2025-11-03T18:01:28.4960747Z Cache location s3, name: ossci-compiler-cache-circleci-v2, prefix: / 2025-11-03T18:01:28.4961149Z Version (client) 0.10.0 2025-11-03T18:01:28.4961428Z + sccache --stop-server 2025-11-03T18:01:28.4981628Z Stopping sccache server... 2025-11-03T18:01:28.4990122Z Compile requests 13 2025-11-03T18:01:28.4990540Z Compile requests executed 0 2025-11-03T18:01:28.4991029Z Cache hits 0 2025-11-03T18:01:28.4991284Z Cache misses 0 2025-11-03T18:01:28.4991552Z Cache hits rate - 2025-11-03T18:01:28.4991805Z Cache timeouts 0 2025-11-03T18:01:28.4992058Z Cache read errors 0 2025-11-03T18:01:28.4992315Z Forced recaches 0 2025-11-03T18:01:28.4992687Z Cache write errors 0 2025-11-03T18:01:28.4992934Z Cache errors 0 2025-11-03T18:01:28.4993363Z Compilations 0 2025-11-03T18:01:28.4993784Z Compilation failures 0 2025-11-03T18:01:28.4994069Z Non-cacheable compilations 0 2025-11-03T18:01:28.4994447Z Non-cacheable calls 0 2025-11-03T18:01:28.4994728Z Non-compilation calls 13 2025-11-03T18:01:28.4994994Z Unsupported compiler calls 0 2025-11-03T18:01:28.4995264Z Average cache write 0.000 s 2025-11-03T18:01:28.4995553Z Average compiler 0.000 s 2025-11-03T18:01:28.4995822Z Average cache read hit 0.000 s 2025-11-03T18:01:28.4996097Z Failed distributed compilations 0 2025-11-03T18:01:28.4996502Z Cache location s3, name: ossci-compiler-cache-circleci-v2, prefix: / 2025-11-03T18:01:28.4996925Z Version (client) 0.10.0 2025-11-03T18:01:28.4997348Z + echo ::endgroup:: 2025-11-03T18:01:28.4997920Z ##[endgroup] 2025-11-03T18:01:28.5000996Z + cleanup_workspace 2025-11-03T18:01:28.5007671Z + echo 'sudo may print the following warning message that can be ignored. The chown command will still run.' 2025-11-03T18:01:28.5008240Z sudo may print the following warning message that can be ignored. The chown command will still run. 2025-11-03T18:01:28.5008868Z + echo ' sudo: setrlimit(RLIMIT_STACK): Operation not permitted' 2025-11-03T18:01:28.5009219Z sudo: setrlimit(RLIMIT_STACK): Operation not permitted 2025-11-03T18:01:28.5014223Z + echo 'For more details refer to https://github.com/sudo-project/sudo/issues/42' 2025-11-03T18:01:28.5014661Z For more details refer to https://github.com/sudo-project/sudo/issues/42 2025-11-03T18:01:28.5015001Z + sudo chown -R 1000 /var/lib/jenkins/workspace 2025-11-03T18:01:29.5622119Z ##[group]Run pytorch/test-infra/.github/actions/upload-benchmark-results@main 2025-11-03T18:01:29.5622752Z with: 2025-11-03T18:01:29.5623087Z benchmark-results-dir: test/test-reports 2025-11-03T18:01:29.5623505Z dry-run: false 2025-11-03T18:01:29.5623806Z schema-version: v3 2025-11-03T18:01:29.5624376Z github-token: *** 2025-11-03T18:01:29.5624669Z env: 2025-11-03T18:01:29.5624943Z GIT_DEFAULT_BRANCH: main 2025-11-03T18:01:29.5625377Z GPU_FLAG: --gpus all -e NVIDIA_DRIVER_CAPABILITIES=all 2025-11-03T18:01:29.5626106Z DOCKER_CONTAINER_ID: c5e792ba8f7ad58ea91feb6e358bcdc693306479a78a475d1e0ca6fca41e9d73 2025-11-03T18:01:29.5626744Z ##[endgroup] 2025-11-03T18:01:29.5656624Z ##[group]Run set -eux 2025-11-03T18:01:29.5656852Z set -eux 2025-11-03T18:01:29.5657005Z  2025-11-03T18:01:29.5657160Z if [[ -n "" ]]; then 2025-11-03T18:01:29.5657354Z  source "" 2025-11-03T18:01:29.5657523Z fi 2025-11-03T18:01:29.5657800Z python3 -mpip install boto3==1.35.33 psutil==7.0.0 pynvml==12.0.0 2025-11-03T18:01:29.5658116Z  2025-11-03T18:01:29.5658269Z DEVICE_NAME="" 2025-11-03T18:01:29.5658465Z DEVICE_TYPE="" 2025-11-03T18:01:29.5658639Z  2025-11-03T18:01:29.5658808Z if command -v nvidia-smi; then 2025-11-03T18:01:29.5659127Z  # NB: I'm using PyTorch here to get the device name, however, it needs to 2025-11-03T18:01:29.5659534Z  # install the correct version of PyTorch manually for now. Any PyTorch 2025-11-03T18:01:29.5659916Z  # version is fine, I just use 2.7.1 to satify PYPIDEP linter 2025-11-03T18:01:29.5660221Z  python3 -mpip install torch==2.7.1 2025-11-03T18:01:29.5660595Z elif command -v rocminfo; then 2025-11-03T18:01:29.5660898Z  # NB: Installing torch on ROCm runner with pip here causes CI to fail 2025-11-03T18:01:29.5661299Z  # with a memoryview is too large error only on MI300 runners. Is pip 2025-11-03T18:01:29.5661803Z  # version on ROCm runner there too old? As a workaround, let's use the 2025-11-03T18:01:29.5662144Z  # GPU device name coming from rocminfo instead 2025-11-03T18:01:29.5662401Z  DEVICE_NAME=rocm 2025-11-03T18:01:29.5662733Z  DEVICE_TYPE=$(rocminfo | grep "Marketing Name" | tail -n1 | awk -F':' '{print $2}' | xargs) 2025-11-03T18:01:29.5663095Z fi 2025-11-03T18:01:29.5663246Z  2025-11-03T18:01:29.5663431Z echo "DEVICE_NAME=$DEVICE_NAME" >> $GITHUB_ENV 2025-11-03T18:01:29.5663703Z echo "DEVICE_TYPE=$DEVICE_TYPE" >> $GITHUB_ENV 2025-11-03T18:01:29.5675908Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2025-11-03T18:01:29.5676185Z env: 2025-11-03T18:01:29.5676342Z GIT_DEFAULT_BRANCH: main 2025-11-03T18:01:29.5676582Z GPU_FLAG: --gpus all -e NVIDIA_DRIVER_CAPABILITIES=all 2025-11-03T18:01:29.5676979Z DOCKER_CONTAINER_ID: c5e792ba8f7ad58ea91feb6e358bcdc693306479a78a475d1e0ca6fca41e9d73 2025-11-03T18:01:29.5677341Z ##[endgroup] 2025-11-03T18:01:29.5712910Z + [[ -n '' ]] 2025-11-03T18:01:29.5713431Z + python3 -mpip install boto3==1.35.33 psutil==7.0.0 pynvml==12.0.0 2025-11-03T18:01:30.0782109Z Defaulting to user installation because normal site-packages is not writeable 2025-11-03T18:01:31.1142969Z Collecting boto3==1.35.33 2025-11-03T18:01:31.1319749Z Downloading boto3-1.35.33-py3-none-any.whl (139 kB) 2025-11-03T18:01:31.4172535Z Collecting psutil==7.0.0 2025-11-03T18:01:31.4213925Z Downloading psutil-7.0.0-cp36-abi3-manylinux_2_12_x86_64.manylinux2010_x86_64.manylinux_2_17_x86_64.manylinux2014_x86_64.whl (277 kB) 2025-11-03T18:01:31.4492784Z Collecting pynvml==12.0.0 2025-11-03T18:01:31.4539482Z Downloading pynvml-12.0.0-py3-none-any.whl (26 kB) 2025-11-03T18:01:31.4994771Z Collecting s3transfer<0.11.0,>=0.10.0 2025-11-03T18:01:31.5034688Z Downloading s3transfer-0.10.4-py3-none-any.whl (83 kB) 2025-11-03T18:01:31.5090312Z Requirement already satisfied: jmespath<2.0.0,>=0.7.1 in /usr/lib/python3.9/site-packages (from boto3==1.35.33) (0.10.0) 2025-11-03T18:01:32.5768998Z Collecting botocore<1.36.0,>=1.35.33 2025-11-03T18:01:32.5808488Z Downloading botocore-1.35.99-py3-none-any.whl (13.3 MB) 2025-11-03T18:01:32.8088003Z Collecting nvidia-ml-py<13.0.0a0,>=12.0.0 2025-11-03T18:01:32.8149661Z Downloading nvidia_ml_py-12.575.51-py3-none-any.whl (47 kB) 2025-11-03T18:01:32.8232629Z Requirement already satisfied: python-dateutil<3.0.0,>=2.1 in /usr/lib/python3.9/site-packages (from botocore<1.36.0,>=1.35.33->boto3==1.35.33) (2.8.1) 2025-11-03T18:01:32.8241019Z Requirement already satisfied: urllib3<1.27,>=1.25.4 in /usr/lib/python3.9/site-packages (from botocore<1.36.0,>=1.35.33->boto3==1.35.33) (1.25.10) 2025-11-03T18:01:32.9496320Z Requirement already satisfied: six>=1.5 in /usr/lib/python3.9/site-packages (from python-dateutil<3.0.0,>=2.1->botocore<1.36.0,>=1.35.33->boto3==1.35.33) (1.15.0) 2025-11-03T18:01:33.0664619Z Installing collected packages: botocore, s3transfer, nvidia-ml-py, pynvml, psutil, boto3 2025-11-03T18:01:33.5876130Z Attempting uninstall: nvidia-ml-py 2025-11-03T18:01:33.5877245Z Found existing installation: nvidia-ml-py 11.525.84 2025-11-03T18:01:33.5890151Z Uninstalling nvidia-ml-py-11.525.84: 2025-11-03T18:01:33.6110944Z Successfully uninstalled nvidia-ml-py-11.525.84 2025-11-03T18:01:33.6580560Z Attempting uninstall: psutil 2025-11-03T18:01:33.6581271Z Found existing installation: psutil 5.9.8 2025-11-03T18:01:33.6657368Z Uninstalling psutil-5.9.8: 2025-11-03T18:01:33.6663806Z Successfully uninstalled psutil-5.9.8 2025-11-03T18:01:33.8144880Z Successfully installed boto3-1.35.33 botocore-1.35.99 nvidia-ml-py-12.575.51 psutil-7.0.0 pynvml-12.0.0 s3transfer-0.10.4 2025-11-03T18:01:33.9013455Z + DEVICE_NAME= 2025-11-03T18:01:33.9013872Z + DEVICE_TYPE= 2025-11-03T18:01:33.9014244Z + command -v nvidia-smi 2025-11-03T18:01:33.9014683Z + python3 -mpip install torch==2.7.1 2025-11-03T18:01:33.9015009Z /usr/bin/nvidia-smi 2025-11-03T18:01:34.1215234Z Defaulting to user installation because normal site-packages is not writeable 2025-11-03T18:01:34.3606129Z Collecting torch==2.7.1 2025-11-03T18:01:34.3810287Z Downloading torch-2.7.1-cp39-cp39-manylinux_2_28_x86_64.whl (821.1 MB) 2025-11-03T18:01:46.9422284Z Collecting nvidia-cufft-cu12==11.3.0.4 2025-11-03T18:01:46.9455872Z Downloading nvidia_cufft_cu12-11.3.0.4-py3-none-manylinux2014_x86_64.manylinux_2_17_x86_64.whl (200.2 MB) 2025-11-03T18:01:49.3256326Z Collecting triton==3.3.1 2025-11-03T18:01:49.3340866Z Downloading triton-3.3.1-cp39-cp39-manylinux_2_27_x86_64.manylinux_2_28_x86_64.whl (155.6 MB) 2025-11-03T18:01:50.8890064Z Collecting nvidia-cusparselt-cu12==0.6.3 2025-11-03T18:01:50.8930336Z Downloading nvidia_cusparselt_cu12-0.6.3-py3-none-manylinux2014_x86_64.whl (156.8 MB) 2025-11-03T18:01:52.4604243Z Requirement already satisfied: typing-extensions>=4.10.0 in /home/ec2-user/.local/lib/python3.9/site-packages (from torch==2.7.1) (4.15.0) 2025-11-03T18:01:52.5181224Z Collecting fsspec 2025-11-03T18:01:52.5223724Z Downloading fsspec-2025.10.0-py3-none-any.whl (200 kB) 2025-11-03T18:01:52.5517390Z Collecting nvidia-curand-cu12==10.3.7.77 2025-11-03T18:01:52.5556830Z Downloading nvidia_curand_cu12-10.3.7.77-py3-none-manylinux2014_x86_64.manylinux_2_17_x86_64.whl (56.3 MB) 2025-11-03T18:01:53.1448218Z Collecting nvidia-cuda-nvrtc-cu12==12.6.77 2025-11-03T18:01:53.1487355Z Downloading nvidia_cuda_nvrtc_cu12-12.6.77-py3-none-manylinux2014_x86_64.whl (23.7 MB) 2025-11-03T18:01:53.4084013Z Collecting sympy>=1.13.3 2025-11-03T18:01:53.4123572Z Downloading sympy-1.14.0-py3-none-any.whl (6.3 MB) 2025-11-03T18:01:53.5110581Z Collecting nvidia-cudnn-cu12==9.5.1.17 2025-11-03T18:01:53.5145607Z Downloading nvidia_cudnn_cu12-9.5.1.17-py3-none-manylinux_2_28_x86_64.whl (571.0 MB) 2025-11-03T18:02:01.7288826Z Collecting networkx 2025-11-03T18:02:01.7324472Z Downloading networkx-3.2.1-py3-none-any.whl (1.6 MB) 2025-11-03T18:02:01.7548145Z Requirement already satisfied: jinja2 in /usr/lib/python3.9/site-packages (from torch==2.7.1) (2.11.3) 2025-11-03T18:02:01.7821596Z Collecting nvidia-cusparse-cu12==12.5.4.2 2025-11-03T18:02:01.7861113Z Downloading nvidia_cusparse_cu12-12.5.4.2-py3-none-manylinux2014_x86_64.manylinux_2_17_x86_64.whl (216.6 MB) 2025-11-03T18:02:04.4174270Z Collecting nvidia-cublas-cu12==12.6.4.1 2025-11-03T18:02:04.4212595Z Downloading nvidia_cublas_cu12-12.6.4.1-py3-none-manylinux2014_x86_64.manylinux_2_17_x86_64.whl (393.1 MB) 2025-11-03T18:02:09.7566588Z Collecting nvidia-cuda-cupti-cu12==12.6.80 2025-11-03T18:02:09.7608624Z Downloading nvidia_cuda_cupti_cu12-12.6.80-py3-none-manylinux2014_x86_64.manylinux_2_17_x86_64.whl (8.9 MB) 2025-11-03T18:02:09.8738732Z Collecting nvidia-cusolver-cu12==11.7.1.2 2025-11-03T18:02:09.8775519Z Downloading nvidia_cusolver_cu12-11.7.1.2-py3-none-manylinux2014_x86_64.manylinux_2_17_x86_64.whl (158.2 MB) 2025-11-03T18:02:11.6139085Z Collecting nvidia-nccl-cu12==2.26.2 2025-11-03T18:02:11.6173298Z Downloading nvidia_nccl_cu12-2.26.2-py3-none-manylinux2014_x86_64.manylinux_2_17_x86_64.whl (201.3 MB) 2025-11-03T18:02:13.9122186Z Collecting nvidia-cufile-cu12==1.11.1.6 2025-11-03T18:02:13.9163606Z Downloading nvidia_cufile_cu12-1.11.1.6-py3-none-manylinux2014_x86_64.manylinux_2_17_x86_64.whl (1.1 MB) 2025-11-03T18:02:13.9525907Z Collecting nvidia-nvtx-cu12==12.6.77 2025-11-03T18:02:13.9562528Z Downloading nvidia_nvtx_cu12-12.6.77-py3-none-manylinux2014_x86_64.manylinux_2_17_x86_64.whl (89 kB) 2025-11-03T18:02:13.9842163Z Collecting nvidia-cuda-runtime-cu12==12.6.77 2025-11-03T18:02:13.9880218Z Downloading nvidia_cuda_runtime_cu12-12.6.77-py3-none-manylinux2014_x86_64.manylinux_2_17_x86_64.whl (897 kB) 2025-11-03T18:02:14.0585505Z Collecting filelock 2025-11-03T18:02:14.0644587Z Downloading filelock-3.19.1-py3-none-any.whl (15 kB) 2025-11-03T18:02:14.0930178Z Collecting nvidia-nvjitlink-cu12==12.6.85 2025-11-03T18:02:14.0964392Z Downloading nvidia_nvjitlink_cu12-12.6.85-py3-none-manylinux2010_x86_64.manylinux_2_12_x86_64.whl (19.7 MB) 2025-11-03T18:02:14.3209893Z Requirement already satisfied: setuptools>=40.8.0 in /usr/lib/python3.9/site-packages (from triton==3.3.1->torch==2.7.1) (59.6.0) 2025-11-03T18:02:14.3486336Z Collecting mpmath<1.4,>=1.1.0 2025-11-03T18:02:14.3522423Z Downloading mpmath-1.3.0-py3-none-any.whl (536 kB) 2025-11-03T18:02:14.4312738Z Requirement already satisfied: MarkupSafe>=0.23 in /usr/lib64/python3.9/site-packages (from jinja2->torch==2.7.1) (1.1.1) 2025-11-03T18:02:14.7389768Z Installing collected packages: nvidia-nvjitlink-cu12, nvidia-cusparse-cu12, nvidia-cublas-cu12, mpmath, triton, sympy, nvidia-nvtx-cu12, nvidia-nccl-cu12, nvidia-cusparselt-cu12, nvidia-cusolver-cu12, nvidia-curand-cu12, nvidia-cufile-cu12, nvidia-cufft-cu12, nvidia-cudnn-cu12, nvidia-cuda-runtime-cu12, nvidia-cuda-nvrtc-cu12, nvidia-cuda-cupti-cu12, networkx, fsspec, filelock, torch 2025-11-03T18:02:22.4230035Z WARNING: The scripts proton and proton-viewer are installed in '/home/ec2-user/.local/bin' which is not on PATH. 2025-11-03T18:02:22.4230915Z Consider adding this directory to PATH or, if you prefer to suppress this warning, use --no-warn-script-location. 2025-11-03T18:02:25.8983784Z WARNING: The script isympy is installed in '/home/ec2-user/.local/bin' which is not on PATH. 2025-11-03T18:02:25.8984576Z Consider adding this directory to PATH or, if you prefer to suppress this warning, use --no-warn-script-location. 2025-11-03T18:02:51.9400076Z WARNING: The scripts torchfrtrace and torchrun are installed in '/home/ec2-user/.local/bin' which is not on PATH. 2025-11-03T18:02:51.9400942Z Consider adding this directory to PATH or, if you prefer to suppress this warning, use --no-warn-script-location. 2025-11-03T18:02:52.1770876Z Successfully installed filelock-3.19.1 fsspec-2025.10.0 mpmath-1.3.0 networkx-3.2.1 nvidia-cublas-cu12-12.6.4.1 nvidia-cuda-cupti-cu12-12.6.80 nvidia-cuda-nvrtc-cu12-12.6.77 nvidia-cuda-runtime-cu12-12.6.77 nvidia-cudnn-cu12-9.5.1.17 nvidia-cufft-cu12-11.3.0.4 nvidia-cufile-cu12-1.11.1.6 nvidia-curand-cu12-10.3.7.77 nvidia-cusolver-cu12-11.7.1.2 nvidia-cusparse-cu12-12.5.4.2 nvidia-cusparselt-cu12-0.6.3 nvidia-nccl-cu12-2.26.2 nvidia-nvjitlink-cu12-12.6.85 nvidia-nvtx-cu12-12.6.77 sympy-1.14.0 torch-2.7.1 triton-3.3.1 2025-11-03T18:02:52.6798427Z + echo DEVICE_NAME= 2025-11-03T18:02:52.6798749Z + echo DEVICE_TYPE= 2025-11-03T18:02:52.6824690Z ##[group]Run set -eux 2025-11-03T18:02:52.6824936Z set -eux 2025-11-03T18:02:52.6825101Z  2025-11-03T18:02:52.6825277Z if [[ -z "${GITHUB_TOKEN}" ]]; then 2025-11-03T18:02:52.6825772Z  echo "Missing github-token input" 2025-11-03T18:02:52.6826093Z  exit 1 2025-11-03T18:02:52.6826252Z fi 2025-11-03T18:02:52.6836381Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2025-11-03T18:02:52.6836690Z env: 2025-11-03T18:02:52.6836844Z GIT_DEFAULT_BRANCH: main 2025-11-03T18:02:52.6837091Z GPU_FLAG: --gpus all -e NVIDIA_DRIVER_CAPABILITIES=all 2025-11-03T18:02:52.6837502Z DOCKER_CONTAINER_ID: c5e792ba8f7ad58ea91feb6e358bcdc693306479a78a475d1e0ca6fca41e9d73 2025-11-03T18:02:52.6837876Z DEVICE_NAME: 2025-11-03T18:02:52.6838042Z DEVICE_TYPE: 2025-11-03T18:02:52.6838402Z GITHUB_TOKEN: *** 2025-11-03T18:02:52.6838584Z ##[endgroup] 2025-11-03T18:02:52.6869887Z + [[ -z *** ]] 2025-11-03T18:02:52.7003462Z ##[group]Run pytorch/test-infra/.github/actions/get-workflow-job-id@main 2025-11-03T18:02:52.7003767Z with: 2025-11-03T18:02:52.7004039Z github-token: *** 2025-11-03T18:02:52.7004211Z env: 2025-11-03T18:02:52.7004368Z GIT_DEFAULT_BRANCH: main 2025-11-03T18:02:52.7004616Z GPU_FLAG: --gpus all -e NVIDIA_DRIVER_CAPABILITIES=all 2025-11-03T18:02:52.7005014Z DOCKER_CONTAINER_ID: c5e792ba8f7ad58ea91feb6e358bcdc693306479a78a475d1e0ca6fca41e9d73 2025-11-03T18:02:52.7005472Z DEVICE_NAME: 2025-11-03T18:02:52.7005640Z DEVICE_TYPE: 2025-11-03T18:02:52.7005802Z ##[endgroup] 2025-11-03T18:02:52.7031734Z ##[group]Run set -eux 2025-11-03T18:02:52.7031928Z set -eux 2025-11-03T18:02:52.7032101Z  2025-11-03T18:02:52.7032443Z python3 "${GITHUB_ACTION_PATH}/../../scripts/get_workflow_job_id.py" "${GITHUB_RUN_ID}" "${RUNNER_NAME}" 2025-11-03T18:02:52.7039750Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2025-11-03T18:02:52.7040028Z env: 2025-11-03T18:02:52.7040202Z GIT_DEFAULT_BRANCH: main 2025-11-03T18:02:52.7040449Z GPU_FLAG: --gpus all -e NVIDIA_DRIVER_CAPABILITIES=all 2025-11-03T18:02:52.7040844Z DOCKER_CONTAINER_ID: c5e792ba8f7ad58ea91feb6e358bcdc693306479a78a475d1e0ca6fca41e9d73 2025-11-03T18:02:52.7041199Z DEVICE_NAME: 2025-11-03T18:02:52.7041363Z DEVICE_TYPE: 2025-11-03T18:02:52.7041640Z GITHUB_TOKEN: *** 2025-11-03T18:02:52.7041814Z ##[endgroup] 2025-11-03T18:02:52.7071146Z + python3 /home/ec2-user/actions-runner/_work/_actions/pytorch/test-infra/main/.github/actions/get-workflow-job-id/../../scripts/get_workflow_job_id.py 19040285420 i-0aa7ffe8f5edfd321 2025-11-03T18:02:53.4665132Z setting job-id=54383051168 2025-11-03T18:02:53.4665746Z setting job-name=ci-vllm-test / test (vllm_basic_models_test, 1, 1, linux.g6.4xlarge.experimental.nvidia.gpu) 2025-11-03T18:02:53.4847820Z ##[group]Run set -eux 2025-11-03T18:02:53.4848176Z set -eux 2025-11-03T18:02:53.4848342Z  2025-11-03T18:02:53.4848500Z if [[ -n "" ]]; then 2025-11-03T18:02:53.4848699Z  source "" 2025-11-03T18:02:53.4848874Z fi 2025-11-03T18:02:53.4849028Z  2025-11-03T18:02:53.4849305Z python3 "${GITHUB_ACTION_PATH}/../../scripts/benchmarks/gather_metadata.py" \ 2025-11-03T18:02:53.4849674Z  --schema-version "${SCHEMA_VERSION}" \ 2025-11-03T18:02:53.4849919Z  --repo "${REPO}" \ 2025-11-03T18:02:53.4850123Z  --head-branch "${HEAD_BRANCH}" \ 2025-11-03T18:02:53.4850368Z  --head-sha "${HEAD_SHA}" \ 2025-11-03T18:02:53.4850605Z  --workflow-id "${WORKFLOW_RUN_ID}" \ 2025-11-03T18:02:53.4850851Z  --run-attempt "${RUN_ATTEMPT}" \ 2025-11-03T18:02:53.4851235Z  --job-id "${JOB_ID}" \ 2025-11-03T18:02:53.4851460Z  --job-name "${JOB_NAME}" 2025-11-03T18:02:53.4859236Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2025-11-03T18:02:53.4859535Z env: 2025-11-03T18:02:53.4859695Z GIT_DEFAULT_BRANCH: main 2025-11-03T18:02:53.4859941Z GPU_FLAG: --gpus all -e NVIDIA_DRIVER_CAPABILITIES=all 2025-11-03T18:02:53.4860340Z DOCKER_CONTAINER_ID: c5e792ba8f7ad58ea91feb6e358bcdc693306479a78a475d1e0ca6fca41e9d73 2025-11-03T18:02:53.4860703Z DEVICE_NAME: 2025-11-03T18:02:53.4860874Z DEVICE_TYPE: 2025-11-03T18:02:53.4861033Z SCHEMA_VERSION: v3 2025-11-03T18:02:53.4861219Z REPO: pytorch/pytorch 2025-11-03T18:02:53.4861411Z HEAD_BRANCH: refs/heads/main 2025-11-03T18:02:53.4861729Z HEAD_SHA: 3f6538febd727b782e6e13cfd026a309fb14351d 2025-11-03T18:02:53.4861988Z WORKFLOW_RUN_ID: 19040285420 2025-11-03T18:02:53.4862178Z RUN_ATTEMPT: 1 2025-11-03T18:02:53.4862337Z JOB_ID: 54383051168 2025-11-03T18:02:53.4862680Z JOB_NAME: ci-vllm-test / test (vllm_basic_models_test, 1, 1, linux.g6.4xlarge.experimental.nvidia.gpu) 2025-11-03T18:02:53.4863049Z ##[endgroup] 2025-11-03T18:02:53.4891066Z + [[ -n '' ]] 2025-11-03T18:02:53.4893082Z + python3 /home/ec2-user/actions-runner/_work/_actions/pytorch/test-infra/main/.github/actions/upload-benchmark-results/../../scripts/benchmarks/gather_metadata.py --schema-version v3 --repo pytorch/pytorch --head-branch refs/heads/main --head-sha 3f6538febd727b782e6e13cfd026a309fb14351d --workflow-id 19040285420 --run-attempt 1 --job-id 54383051168 --job-name 'ci-vllm-test / test (vllm_basic_models_test, 1, 1, linux.g6.4xlarge.experimental.nvidia.gpu)' 2025-11-03T18:02:53.5297991Z ##[group]Run set -eux 2025-11-03T18:02:53.5298190Z set -eux 2025-11-03T18:02:53.5298462Z  2025-11-03T18:02:53.5298617Z if [[ -n "" ]]; then 2025-11-03T18:02:53.5298807Z  source "" 2025-11-03T18:02:53.5298973Z fi 2025-11-03T18:02:53.5299118Z  2025-11-03T18:02:53.5299417Z python3 "${GITHUB_ACTION_PATH}/../../scripts/benchmarks/gather_runners_info.py" 2025-11-03T18:02:53.5307118Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2025-11-03T18:02:53.5307412Z env: 2025-11-03T18:02:53.5307576Z GIT_DEFAULT_BRANCH: main 2025-11-03T18:02:53.5307816Z GPU_FLAG: --gpus all -e NVIDIA_DRIVER_CAPABILITIES=all 2025-11-03T18:02:53.5308208Z DOCKER_CONTAINER_ID: c5e792ba8f7ad58ea91feb6e358bcdc693306479a78a475d1e0ca6fca41e9d73 2025-11-03T18:02:53.5308567Z DEVICE_NAME: 2025-11-03T18:02:53.5308734Z DEVICE_TYPE: 2025-11-03T18:02:53.5308886Z ##[endgroup] 2025-11-03T18:02:53.5336678Z + [[ -n '' ]] 2025-11-03T18:02:53.5337466Z + python3 /home/ec2-user/actions-runner/_work/_actions/pytorch/test-infra/main/.github/actions/upload-benchmark-results/../../scripts/benchmarks/gather_runners_info.py 2025-11-03T18:02:54.3907972Z /home/ec2-user/.local/lib/python3.9/site-packages/torch/_subclasses/functional_tensor.py:276: UserWarning: Failed to initialize NumPy: No module named 'numpy' (Triggered internally at /pytorch/torch/csrc/utils/tensor_numpy.cpp:81.) 2025-11-03T18:02:54.3909114Z cpu = _conversion_method_template(device=torch.device("cpu")) 2025-11-03T18:02:55.2179046Z ##[group]Run set -eux 2025-11-03T18:02:55.2179263Z set -eux 2025-11-03T18:02:55.2179423Z  2025-11-03T18:02:55.2179598Z # TODO (huydhn): Implement this part 2025-11-03T18:02:55.2179875Z echo "dependencies={}" >> "${GITHUB_OUTPUT}" 2025-11-03T18:02:55.2187834Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2025-11-03T18:02:55.2188114Z env: 2025-11-03T18:02:55.2188267Z GIT_DEFAULT_BRANCH: main 2025-11-03T18:02:55.2188525Z GPU_FLAG: --gpus all -e NVIDIA_DRIVER_CAPABILITIES=all 2025-11-03T18:02:55.2188942Z DOCKER_CONTAINER_ID: c5e792ba8f7ad58ea91feb6e358bcdc693306479a78a475d1e0ca6fca41e9d73 2025-11-03T18:02:55.2189297Z DEVICE_NAME: 2025-11-03T18:02:55.2189460Z DEVICE_TYPE: 2025-11-03T18:02:55.2189615Z ##[endgroup] 2025-11-03T18:02:55.2218503Z + echo 'dependencies={}' 2025-11-03T18:02:55.2310882Z ##[group]Run set -eux 2025-11-03T18:02:55.2311089Z set -eux 2025-11-03T18:02:55.2311271Z  2025-11-03T18:02:55.2311423Z if [[ -n "" ]]; then 2025-11-03T18:02:55.2311625Z  source "" 2025-11-03T18:02:55.2311801Z fi 2025-11-03T18:02:55.2311949Z  2025-11-03T18:02:55.2312132Z if [[ ! -d "${BENCHMARK_RESULTS_DIR}" ]]; then 2025-11-03T18:02:55.2312445Z  echo "${BENCHMARK_RESULTS_DIR} does not exist, skipping" 2025-11-03T18:02:55.2312794Z  # We don't want the job to fail if the directory doesn't exist 2025-11-03T18:02:55.2313079Z  exit 0 2025-11-03T18:02:55.2313239Z fi 2025-11-03T18:02:55.2313394Z  2025-11-03T18:02:55.2313561Z if [[ "${DRY_RUN}" == "true" ]]; then 2025-11-03T18:02:55.2313902Z  python3 "${GITHUB_ACTION_PATH}/../../scripts/upload_benchmark_results.py" \ 2025-11-03T18:02:55.2314324Z  --benchmark-results-dir "${BENCHMARK_RESULTS_DIR}" \ 2025-11-03T18:02:55.2314625Z  --metadata "${BENCHMARK_METADATA}" \ 2025-11-03T18:02:55.2314875Z  --runners "${RUNNER_INFO}" \ 2025-11-03T18:02:55.2315132Z  --dependencies "${DEPENDENCIES}" \ 2025-11-03T18:02:55.2315365Z  --dry-run 2025-11-03T18:02:55.2315547Z else 2025-11-03T18:02:55.2315825Z  python3 "${GITHUB_ACTION_PATH}/../../scripts/upload_benchmark_results.py" \ 2025-11-03T18:02:55.2316214Z  --benchmark-results-dir "${BENCHMARK_RESULTS_DIR}" \ 2025-11-03T18:02:55.2316507Z  --metadata "${BENCHMARK_METADATA}" \ 2025-11-03T18:02:55.2316761Z  --runners "${RUNNER_INFO}" \ 2025-11-03T18:02:55.2317257Z  --dependencies "${DEPENDENCIES}" 2025-11-03T18:02:55.2317492Z fi 2025-11-03T18:02:55.2325018Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2025-11-03T18:02:55.2325304Z env: 2025-11-03T18:02:55.2325477Z GIT_DEFAULT_BRANCH: main 2025-11-03T18:02:55.2325734Z GPU_FLAG: --gpus all -e NVIDIA_DRIVER_CAPABILITIES=all 2025-11-03T18:02:55.2326141Z DOCKER_CONTAINER_ID: c5e792ba8f7ad58ea91feb6e358bcdc693306479a78a475d1e0ca6fca41e9d73 2025-11-03T18:02:55.2326500Z DEVICE_NAME: 2025-11-03T18:02:55.2326659Z DEVICE_TYPE: 2025-11-03T18:02:55.2326845Z BENCHMARK_RESULTS_DIR: test/test-reports 2025-11-03T18:02:55.2327067Z DRY_RUN: false 2025-11-03T18:02:55.2328029Z BENCHMARK_METADATA: {"timestamp": 1762192973, "schema_version": "v3", "name": "ci-vllm-test / test (vllm_basic_models_test, 1, 1, linux.g6.4xlarge.experimental.nvidia.gpu)", "repo": "pytorch/pytorch", "head_branch": "refs/heads/main", "head_sha": "3f6538febd727b782e6e13cfd026a309fb14351d", "workflow_id": 19040285420, "run_attempt": 1, "job_id": 54383051168} 2025-11-03T18:02:55.2329355Z RUNNER_INFO: [{"cpu_info": "x86_64", "cpu_count": 16, "avail_mem_in_gb": 60, "extra_info": {"hostname": "ip-10-0-2-76.ec2.internal"}, "name": "cuda", "type": "NVIDIA L4", "gpu_count": 1, "avail_gpu_mem_in_gb": 22}] 2025-11-03T18:02:55.2329882Z DEPENDENCIES: {} 2025-11-03T18:02:55.2330047Z ##[endgroup] 2025-11-03T18:02:55.2356755Z + [[ -n '' ]] 2025-11-03T18:02:55.2357148Z + [[ ! -d test/test-reports ]] 2025-11-03T18:02:55.2357554Z + echo 'test/test-reports does not exist, skipping' 2025-11-03T18:02:55.2357864Z + exit 0 2025-11-03T18:02:55.2358096Z test/test-reports does not exist, skipping 2025-11-03T18:02:55.2489799Z ##[group]Run cat test/**/*_toprint.log || true 2025-11-03T18:02:55.2490129Z cat test/**/*_toprint.log || true 2025-11-03T18:02:55.2497070Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2025-11-03T18:02:55.2497356Z env: 2025-11-03T18:02:55.2497524Z GIT_DEFAULT_BRANCH: main 2025-11-03T18:02:55.2497776Z GPU_FLAG: --gpus all -e NVIDIA_DRIVER_CAPABILITIES=all 2025-11-03T18:02:55.2498165Z DOCKER_CONTAINER_ID: c5e792ba8f7ad58ea91feb6e358bcdc693306479a78a475d1e0ca6fca41e9d73 2025-11-03T18:02:55.2498526Z DEVICE_NAME: 2025-11-03T18:02:55.2498692Z DEVICE_TYPE: 2025-11-03T18:02:55.2498852Z ##[endgroup] 2025-11-03T18:02:55.2596827Z cat: 'test/**/*_toprint.log': No such file or directory 2025-11-03T18:02:55.2752403Z ##[group]Run kill "$MONITOR_SCRIPT_PID" 2025-11-03T18:02:55.2752779Z kill "$MONITOR_SCRIPT_PID" 2025-11-03T18:02:55.2759985Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2025-11-03T18:02:55.2760264Z env: 2025-11-03T18:02:55.2760429Z GIT_DEFAULT_BRANCH: main 2025-11-03T18:02:55.2760668Z GPU_FLAG: --gpus all -e NVIDIA_DRIVER_CAPABILITIES=all 2025-11-03T18:02:55.2761061Z DOCKER_CONTAINER_ID: c5e792ba8f7ad58ea91feb6e358bcdc693306479a78a475d1e0ca6fca41e9d73 2025-11-03T18:02:55.2761423Z DEVICE_NAME: 2025-11-03T18:02:55.2761597Z DEVICE_TYPE: 2025-11-03T18:02:55.2761766Z MONITOR_SCRIPT_PID: 61182 2025-11-03T18:02:55.2761949Z ##[endgroup] 2025-11-03T18:02:55.2951986Z Prepare all required actions 2025-11-03T18:02:55.2952360Z Getting action download info 2025-11-03T18:02:55.4444688Z Download action repository 'seemethere/upload-artifact-s3@v5' (SHA:baba72d0712b404f646cebe0730933554ebce96a) 2025-11-03T18:02:55.9891627Z Download action repository 'actions/upload-artifact@v4' (SHA:ea165f8d65b6e75b540449e92b4886f43607fa02) 2025-11-03T18:02:57.7052255Z ##[group]Run ./.github/actions/upload-test-artifacts 2025-11-03T18:02:57.7052516Z with: 2025-11-03T18:02:57.7052852Z file-suffix: test-vllm_basic_models_test-1-1-linux.g6.4xlarge.experimental.nvidia.gpu_54383051168 2025-11-03T18:02:57.7053250Z s3-bucket: gha-artifacts 2025-11-03T18:02:57.7053432Z env: 2025-11-03T18:02:57.7053587Z GIT_DEFAULT_BRANCH: main 2025-11-03T18:02:57.7053822Z GPU_FLAG: --gpus all -e NVIDIA_DRIVER_CAPABILITIES=all 2025-11-03T18:02:57.7054231Z DOCKER_CONTAINER_ID: c5e792ba8f7ad58ea91feb6e358bcdc693306479a78a475d1e0ca6fca41e9d73 2025-11-03T18:02:57.7054735Z DEVICE_NAME: 2025-11-03T18:02:57.7054895Z DEVICE_TYPE: 2025-11-03T18:02:57.7055053Z ##[endgroup] 2025-11-03T18:02:57.7161637Z ##[group]Run # Remove any previous test jsons if they exist 2025-11-03T18:02:57.7161972Z # Remove any previous test jsons if they exist 2025-11-03T18:02:57.7162245Z rm -f test-jsons-*.zip 2025-11-03T18:02:57.7162560Z zip -r "test-jsons-${FILE_SUFFIX}.zip" test/test-reports -i '*.json' 2025-11-03T18:02:57.7170337Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2025-11-03T18:02:57.7170612Z env: 2025-11-03T18:02:57.7170768Z GIT_DEFAULT_BRANCH: main 2025-11-03T18:02:57.7171009Z GPU_FLAG: --gpus all -e NVIDIA_DRIVER_CAPABILITIES=all 2025-11-03T18:02:57.7171397Z DOCKER_CONTAINER_ID: c5e792ba8f7ad58ea91feb6e358bcdc693306479a78a475d1e0ca6fca41e9d73 2025-11-03T18:02:57.7171767Z DEVICE_NAME: 2025-11-03T18:02:57.7171935Z DEVICE_TYPE: 2025-11-03T18:02:57.7172279Z FILE_SUFFIX: test-vllm_basic_models_test-1-1-linux.g6.4xlarge.experimental.nvidia.gpu_54383051168 2025-11-03T18:02:57.7172657Z ##[endgroup] 2025-11-03T18:02:57.8279776Z zip warning: name not matched: test/test-reports 2025-11-03T18:02:57.8281660Z zip warning: zip file empty 2025-11-03T18:02:57.8358423Z ##[group]Run # Remove any previous test reports if they exist 2025-11-03T18:02:57.8358875Z # Remove any previous test reports if they exist 2025-11-03T18:02:57.8359156Z rm -f test-reports-*.zip 2025-11-03T18:02:57.8359501Z zip -r "test-reports-${FILE_SUFFIX}.zip" test/test-reports -i '*.xml' -i '*.csv' 2025-11-03T18:02:57.8366653Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2025-11-03T18:02:57.8366929Z env: 2025-11-03T18:02:57.8367092Z GIT_DEFAULT_BRANCH: main 2025-11-03T18:02:57.8367339Z GPU_FLAG: --gpus all -e NVIDIA_DRIVER_CAPABILITIES=all 2025-11-03T18:02:57.8367735Z DOCKER_CONTAINER_ID: c5e792ba8f7ad58ea91feb6e358bcdc693306479a78a475d1e0ca6fca41e9d73 2025-11-03T18:02:57.8368119Z DEVICE_NAME: 2025-11-03T18:02:57.8368298Z DEVICE_TYPE: 2025-11-03T18:02:57.8368637Z FILE_SUFFIX: test-vllm_basic_models_test-1-1-linux.g6.4xlarge.experimental.nvidia.gpu_54383051168 2025-11-03T18:02:57.8369021Z ##[endgroup] 2025-11-03T18:02:57.8467785Z zip warning: name not matched: test/test-reports 2025-11-03T18:02:57.8468684Z zip warning: zip file empty 2025-11-03T18:02:57.8533402Z ##[group]Run # Remove any previous usage logs if they exist 2025-11-03T18:02:57.8533815Z # Remove any previous usage logs if they exist 2025-11-03T18:02:57.8534140Z rm -f logs-*.zip 2025-11-03T18:02:57.8534446Z zip "logs-${FILE_SUFFIX}.zip" 'usage_log.txt' || true 2025-11-03T18:02:57.8534883Z zip -r "logs-${FILE_SUFFIX}.zip" test/test-reports -i '*.log' || true 2025-11-03T18:02:57.8542340Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2025-11-03T18:02:57.8542623Z env: 2025-11-03T18:02:57.8542779Z GIT_DEFAULT_BRANCH: main 2025-11-03T18:02:57.8543045Z GPU_FLAG: --gpus all -e NVIDIA_DRIVER_CAPABILITIES=all 2025-11-03T18:02:57.8563205Z DOCKER_CONTAINER_ID: c5e792ba8f7ad58ea91feb6e358bcdc693306479a78a475d1e0ca6fca41e9d73 2025-11-03T18:02:57.8563834Z DEVICE_NAME: 2025-11-03T18:02:57.8564037Z DEVICE_TYPE: 2025-11-03T18:02:57.8564408Z FILE_SUFFIX: test-vllm_basic_models_test-1-1-linux.g6.4xlarge.experimental.nvidia.gpu_54383051168 2025-11-03T18:02:57.8565068Z ##[endgroup] 2025-11-03T18:02:57.8700807Z adding: usage_log.txt (deflated 94%) 2025-11-03T18:02:57.8715857Z zip warning: name not matched: test/test-reports 2025-11-03T18:02:57.8716254Z 2025-11-03T18:02:57.8716901Z zip error: Nothing to do! (logs-test-vllm_basic_models_test-1-1-linux.g6.4xlarge.experimental.nvidia.gpu_54383051168.zip) 2025-11-03T18:02:57.8799341Z ##[group]Run # Remove any previous debugging artifacts if they exist 2025-11-03T18:02:57.8799729Z # Remove any previous debugging artifacts if they exist 2025-11-03T18:02:57.8800140Z rm -f debug-*.zip 2025-11-03T18:02:57.8800362Z if [ -d 'test/debug' ]; then 2025-11-03T18:02:57.8800629Z  zip -r "debug-${FILE_SUFFIX}.zip" test/debug 2025-11-03T18:02:57.8800873Z fi 2025-11-03T18:02:57.8807883Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2025-11-03T18:02:57.8808176Z env: 2025-11-03T18:02:57.8808340Z GIT_DEFAULT_BRANCH: main 2025-11-03T18:02:57.8808584Z GPU_FLAG: --gpus all -e NVIDIA_DRIVER_CAPABILITIES=all 2025-11-03T18:02:57.8808986Z DOCKER_CONTAINER_ID: c5e792ba8f7ad58ea91feb6e358bcdc693306479a78a475d1e0ca6fca41e9d73 2025-11-03T18:02:57.8809358Z DEVICE_NAME: 2025-11-03T18:02:57.8809534Z DEVICE_TYPE: 2025-11-03T18:02:57.8809864Z FILE_SUFFIX: test-vllm_basic_models_test-1-1-linux.g6.4xlarge.experimental.nvidia.gpu_54383051168 2025-11-03T18:02:57.8810261Z ##[endgroup] 2025-11-03T18:02:57.8900921Z ##[group]Run seemethere/upload-artifact-s3@v5 2025-11-03T18:02:57.8901160Z with: 2025-11-03T18:02:57.8901332Z s3-bucket: gha-artifacts 2025-11-03T18:02:57.8901687Z s3-prefix: pytorch/pytorch/19040285420/1/artifact 2025-11-03T18:02:57.8901935Z retention-days: 14 2025-11-03T18:02:57.8902123Z if-no-files-found: warn 2025-11-03T18:02:57.8902328Z path: test-jsons-*.zip 2025-11-03T18:02:57.8902509Z name: artifact 2025-11-03T18:02:57.8902662Z region: us-east-1 2025-11-03T18:02:57.8902822Z env: 2025-11-03T18:02:57.8903091Z GIT_DEFAULT_BRANCH: main 2025-11-03T18:02:57.8903339Z GPU_FLAG: --gpus all -e NVIDIA_DRIVER_CAPABILITIES=all 2025-11-03T18:02:57.8903733Z DOCKER_CONTAINER_ID: c5e792ba8f7ad58ea91feb6e358bcdc693306479a78a475d1e0ca6fca41e9d73 2025-11-03T18:02:57.8904086Z DEVICE_NAME: 2025-11-03T18:02:57.8904248Z DEVICE_TYPE: 2025-11-03T18:02:57.8904412Z ##[endgroup] 2025-11-03T18:02:58.3574620Z NOTE: s3-prefix specified, ignoring name parameter 2025-11-03T18:02:58.3575064Z With the provided path, there will be 1 file uploaded 2025-11-03T18:02:58.3575465Z Uploading to s3 prefix: pytorch/pytorch/19040285420/1/artifact 2025-11-03T18:02:58.3643481Z Starting upload of test-jsons-test-vllm_basic_models_test-1-1-linux.g6.4xlarge.experimental.nvidia.gpu_54383051168.zip 2025-11-03T18:02:58.4703402Z Finished upload of test-jsons-test-vllm_basic_models_test-1-1-linux.g6.4xlarge.experimental.nvidia.gpu_54383051168.zip 2025-11-03T18:02:58.4948870Z ##[group]Run seemethere/upload-artifact-s3@v5 2025-11-03T18:02:58.4949108Z with: 2025-11-03T18:02:58.4949283Z s3-bucket: gha-artifacts 2025-11-03T18:02:58.4949524Z s3-prefix: pytorch/pytorch/19040285420/1/artifact 2025-11-03T18:02:58.4949780Z retention-days: 14 2025-11-03T18:02:58.4949961Z if-no-files-found: error 2025-11-03T18:02:58.4950169Z path: test-reports-*.zip 2025-11-03T18:02:58.4950360Z name: artifact 2025-11-03T18:02:58.4950524Z region: us-east-1 2025-11-03T18:02:58.4950681Z env: 2025-11-03T18:02:58.4950833Z GIT_DEFAULT_BRANCH: main 2025-11-03T18:02:58.4951071Z GPU_FLAG: --gpus all -e NVIDIA_DRIVER_CAPABILITIES=all 2025-11-03T18:02:58.4951491Z DOCKER_CONTAINER_ID: c5e792ba8f7ad58ea91feb6e358bcdc693306479a78a475d1e0ca6fca41e9d73 2025-11-03T18:02:58.4952040Z DEVICE_NAME: 2025-11-03T18:02:58.4952217Z DEVICE_TYPE: 2025-11-03T18:02:58.4952387Z ##[endgroup] 2025-11-03T18:02:59.0006281Z NOTE: s3-prefix specified, ignoring name parameter 2025-11-03T18:02:59.0006814Z With the provided path, there will be 1 file uploaded 2025-11-03T18:02:59.0007294Z Uploading to s3 prefix: pytorch/pytorch/19040285420/1/artifact 2025-11-03T18:02:59.0074225Z Starting upload of test-reports-test-vllm_basic_models_test-1-1-linux.g6.4xlarge.experimental.nvidia.gpu_54383051168.zip 2025-11-03T18:02:59.1292129Z Finished upload of test-reports-test-vllm_basic_models_test-1-1-linux.g6.4xlarge.experimental.nvidia.gpu_54383051168.zip 2025-11-03T18:02:59.1610609Z ##[group]Run seemethere/upload-artifact-s3@v5 2025-11-03T18:02:59.1610857Z with: 2025-11-03T18:02:59.1611024Z s3-bucket: gha-artifacts 2025-11-03T18:02:59.1611274Z s3-prefix: pytorch/pytorch/19040285420/1/artifact 2025-11-03T18:02:59.1611666Z retention-days: 14 2025-11-03T18:02:59.1611841Z if-no-files-found: ignore 2025-11-03T18:02:59.1612043Z path: logs-*.zip 2025-11-03T18:02:59.1612208Z name: artifact 2025-11-03T18:02:59.1612381Z region: us-east-1 2025-11-03T18:02:59.1612539Z env: 2025-11-03T18:02:59.1612694Z GIT_DEFAULT_BRANCH: main 2025-11-03T18:02:59.1612939Z GPU_FLAG: --gpus all -e NVIDIA_DRIVER_CAPABILITIES=all 2025-11-03T18:02:59.1613345Z DOCKER_CONTAINER_ID: c5e792ba8f7ad58ea91feb6e358bcdc693306479a78a475d1e0ca6fca41e9d73 2025-11-03T18:02:59.1613695Z DEVICE_NAME: 2025-11-03T18:02:59.1613855Z DEVICE_TYPE: 2025-11-03T18:02:59.1614024Z ##[endgroup] 2025-11-03T18:02:59.4548726Z NOTE: s3-prefix specified, ignoring name parameter 2025-11-03T18:02:59.4549437Z With the provided path, there will be 1 file uploaded 2025-11-03T18:02:59.4550085Z Uploading to s3 prefix: pytorch/pytorch/19040285420/1/artifact 2025-11-03T18:02:59.4615446Z Starting upload of logs-test-vllm_basic_models_test-1-1-linux.g6.4xlarge.experimental.nvidia.gpu_54383051168.zip 2025-11-03T18:02:59.5864708Z Finished upload of logs-test-vllm_basic_models_test-1-1-linux.g6.4xlarge.experimental.nvidia.gpu_54383051168.zip 2025-11-03T18:02:59.6225759Z ##[group]Run seemethere/upload-artifact-s3@v5 2025-11-03T18:02:59.6225999Z with: 2025-11-03T18:02:59.6226164Z s3-bucket: gha-artifacts 2025-11-03T18:02:59.6226402Z s3-prefix: pytorch/pytorch/19040285420/1/artifact 2025-11-03T18:02:59.6226759Z retention-days: 14 2025-11-03T18:02:59.6226936Z if-no-files-found: ignore 2025-11-03T18:02:59.6227141Z path: debug-*.zip 2025-11-03T18:02:59.6227312Z name: artifact 2025-11-03T18:02:59.6227476Z region: us-east-1 2025-11-03T18:02:59.6227632Z env: 2025-11-03T18:02:59.6227790Z GIT_DEFAULT_BRANCH: main 2025-11-03T18:02:59.6228039Z GPU_FLAG: --gpus all -e NVIDIA_DRIVER_CAPABILITIES=all 2025-11-03T18:02:59.6228444Z DOCKER_CONTAINER_ID: c5e792ba8f7ad58ea91feb6e358bcdc693306479a78a475d1e0ca6fca41e9d73 2025-11-03T18:02:59.6228795Z DEVICE_NAME: 2025-11-03T18:02:59.6228966Z DEVICE_TYPE: 2025-11-03T18:02:59.6229122Z ##[endgroup] 2025-11-03T18:02:59.9113199Z No files were found with the provided path: debug-*.zip. No artifacts will be uploaded. 2025-11-03T18:02:59.9430109Z ##[group]Run # shellcheck disable=SC2156 2025-11-03T18:02:59.9430425Z # shellcheck disable=SC2156 2025-11-03T18:02:59.9430848Z find . -iname "core.[1-9]*" -exec docker exec "${DOCKER_CONTAINER_ID}" sh -c "gdb python {} -ex 'bt' -ex 'q'" \; 2025-11-03T18:02:59.9438883Z shell: /usr/bin/bash -e {0} 2025-11-03T18:02:59.9439089Z env: 2025-11-03T18:02:59.9439247Z GIT_DEFAULT_BRANCH: main 2025-11-03T18:02:59.9439482Z GPU_FLAG: --gpus all -e NVIDIA_DRIVER_CAPABILITIES=all 2025-11-03T18:02:59.9439900Z DOCKER_CONTAINER_ID: c5e792ba8f7ad58ea91feb6e358bcdc693306479a78a475d1e0ca6fca41e9d73 2025-11-03T18:02:59.9440256Z DEVICE_NAME: 2025-11-03T18:02:59.9440421Z DEVICE_TYPE: 2025-11-03T18:02:59.9440578Z ##[endgroup] 2025-11-03T18:03:00.3814788Z Prepare all required actions 2025-11-03T18:03:00.3815144Z Getting action download info 2025-11-03T18:03:00.5333998Z ##[group]Run ./.github/actions/upload-utilization-stats 2025-11-03T18:03:00.5334248Z with: 2025-11-03T18:03:00.5334400Z job_id: 54383051168 2025-11-03T18:03:00.5334746Z job_name: ci-vllm-test / test (vllm_basic_models_test, 1, 1, linux.g6.4xlarge.experimental.nvidia.gpu) 2025-11-03T18:03:00.5335125Z workflow_name: vllm-test 2025-11-03T18:03:00.5335323Z workflow_run_id: 19040285420 2025-11-03T18:03:00.5335513Z workflow_attempt: 1 2025-11-03T18:03:00.5335678Z env: 2025-11-03T18:03:00.5335830Z GIT_DEFAULT_BRANCH: main 2025-11-03T18:03:00.5336064Z GPU_FLAG: --gpus all -e NVIDIA_DRIVER_CAPABILITIES=all 2025-11-03T18:03:00.5336479Z DOCKER_CONTAINER_ID: c5e792ba8f7ad58ea91feb6e358bcdc693306479a78a475d1e0ca6fca41e9d73 2025-11-03T18:03:00.5336839Z DEVICE_NAME: 2025-11-03T18:03:00.5337001Z DEVICE_TYPE: 2025-11-03T18:03:00.5337154Z ##[endgroup] 2025-11-03T18:03:00.5433386Z ##[group]Run echo "workflow_id: 19040285420" 2025-11-03T18:03:00.5433752Z echo "workflow_id: 19040285420" 2025-11-03T18:03:00.5433989Z echo "workflow_attempt: 1" 2025-11-03T18:03:00.5434211Z echo "workflow_Name: vllm-test" 2025-11-03T18:03:00.5434441Z echo "job_id: 54383051168" 2025-11-03T18:03:00.5434858Z echo "job_name: ci-vllm-test / test (vllm_basic_models_test, 1, 1, linux.g6.4xlarge.experimental.nvidia.gpu)" 2025-11-03T18:03:00.5435291Z echo "artifact_prefix: " 2025-11-03T18:03:00.5435504Z python3 --version 2025-11-03T18:03:00.5443472Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2025-11-03T18:03:00.5443750Z env: 2025-11-03T18:03:00.5443908Z GIT_DEFAULT_BRANCH: main 2025-11-03T18:03:00.5444162Z GPU_FLAG: --gpus all -e NVIDIA_DRIVER_CAPABILITIES=all 2025-11-03T18:03:00.5444562Z DOCKER_CONTAINER_ID: c5e792ba8f7ad58ea91feb6e358bcdc693306479a78a475d1e0ca6fca41e9d73 2025-11-03T18:03:00.5444940Z DEVICE_NAME: 2025-11-03T18:03:00.5445119Z DEVICE_TYPE: 2025-11-03T18:03:00.5445288Z ##[endgroup] 2025-11-03T18:03:00.5473690Z workflow_id: 19040285420 2025-11-03T18:03:00.5473974Z workflow_attempt: 1 2025-11-03T18:03:00.5474207Z workflow_Name: vllm-test 2025-11-03T18:03:00.5474439Z job_id: 54383051168 2025-11-03T18:03:00.5474867Z job_name: ci-vllm-test / test (vllm_basic_models_test, 1, 1, linux.g6.4xlarge.experimental.nvidia.gpu) 2025-11-03T18:03:00.5475528Z artifact_prefix: 2025-11-03T18:03:00.5489758Z Python 3.9.23 2025-11-03T18:03:00.5640035Z ##[group]Run nick-fields/retry@v3.0.0 2025-11-03T18:03:00.5640253Z with: 2025-11-03T18:03:00.5640393Z shell: bash 2025-11-03T18:03:00.5640551Z timeout_minutes: 5 2025-11-03T18:03:00.5640732Z max_attempts: 5 2025-11-03T18:03:00.5640903Z retry_wait_seconds: 30 2025-11-03T18:03:00.5641283Z command: set -eu python3 -m pip install python-dateutil==2.8.2 boto3==1.35.42 pandas==2.1.3 dataclasses_json==0.6.7 2025-11-03T18:03:00.5641704Z polling_interval_seconds: 1 2025-11-03T18:03:00.5641907Z warning_on_retry: true 2025-11-03T18:03:00.5642104Z continue_on_error: false 2025-11-03T18:03:00.5642286Z env: 2025-11-03T18:03:00.5642433Z GIT_DEFAULT_BRANCH: main 2025-11-03T18:03:00.5642669Z GPU_FLAG: --gpus all -e NVIDIA_DRIVER_CAPABILITIES=all 2025-11-03T18:03:00.5643078Z DOCKER_CONTAINER_ID: c5e792ba8f7ad58ea91feb6e358bcdc693306479a78a475d1e0ca6fca41e9d73 2025-11-03T18:03:00.5643431Z DEVICE_NAME: 2025-11-03T18:03:00.5643591Z DEVICE_TYPE: 2025-11-03T18:03:00.5643753Z ##[endgroup] 2025-11-03T18:03:00.9056565Z Defaulting to user installation because normal site-packages is not writeable 2025-11-03T18:03:01.1082403Z Collecting python-dateutil==2.8.2 2025-11-03T18:03:01.1243906Z Downloading python_dateutil-2.8.2-py2.py3-none-any.whl (247 kB) 2025-11-03T18:03:02.0749918Z Collecting boto3==1.35.42 2025-11-03T18:03:02.0801901Z Downloading boto3-1.35.42-py3-none-any.whl (139 kB) 2025-11-03T18:03:02.6169889Z Collecting pandas==2.1.3 2025-11-03T18:03:02.6206684Z Downloading pandas-2.1.3-cp39-cp39-manylinux_2_17_x86_64.manylinux2014_x86_64.whl (12.3 MB) 2025-11-03T18:03:02.9213522Z Requirement already satisfied: dataclasses_json==0.6.7 in /home/ec2-user/.local/lib/python3.9/site-packages (0.6.7) 2025-11-03T18:03:02.9228386Z Requirement already satisfied: six>=1.5 in /usr/lib/python3.9/site-packages (from python-dateutil==2.8.2) (1.15.0) 2025-11-03T18:03:02.9269449Z Requirement already satisfied: s3transfer<0.11.0,>=0.10.0 in /home/ec2-user/.local/lib/python3.9/site-packages (from boto3==1.35.42) (0.10.4) 2025-11-03T18:03:02.9274144Z Requirement already satisfied: jmespath<2.0.0,>=0.7.1 in /usr/lib/python3.9/site-packages (from boto3==1.35.42) (0.10.0) 2025-11-03T18:03:02.9277661Z Requirement already satisfied: botocore<1.36.0,>=1.35.42 in /home/ec2-user/.local/lib/python3.9/site-packages (from boto3==1.35.42) (1.35.99) 2025-11-03T18:03:02.9827798Z Requirement already satisfied: pytz>=2020.1 in /usr/lib/python3.9/site-packages (from pandas==2.1.3) (2022.7.1) 2025-11-03T18:03:03.0386350Z Collecting tzdata>=2022.1 2025-11-03T18:03:03.0420799Z Downloading tzdata-2025.2-py2.py3-none-any.whl (347 kB) 2025-11-03T18:03:03.8887228Z Collecting numpy<2,>=1.22.4 2025-11-03T18:03:03.8953048Z Downloading numpy-1.26.4-cp39-cp39-manylinux_2_17_x86_64.manylinux2014_x86_64.whl (18.2 MB) 2025-11-03T18:03:04.3382542Z Requirement already satisfied: typing-inspect<1,>=0.4.0 in /home/ec2-user/.local/lib/python3.9/site-packages (from dataclasses_json==0.6.7) (0.9.0) 2025-11-03T18:03:04.3384363Z Requirement already satisfied: marshmallow<4.0.0,>=3.18.0 in /home/ec2-user/.local/lib/python3.9/site-packages (from dataclasses_json==0.6.7) (3.26.1) 2025-11-03T18:03:04.3451711Z Requirement already satisfied: urllib3<1.27,>=1.25.4 in /usr/lib/python3.9/site-packages (from botocore<1.36.0,>=1.35.42->boto3==1.35.42) (1.25.10) 2025-11-03T18:03:04.3534963Z Requirement already satisfied: packaging>=17.0 in /home/ec2-user/.local/lib/python3.9/site-packages (from marshmallow<4.0.0,>=3.18.0->dataclasses_json==0.6.7) (25.0) 2025-11-03T18:03:04.3626754Z Requirement already satisfied: typing-extensions>=3.7.4 in /home/ec2-user/.local/lib/python3.9/site-packages (from typing-inspect<1,>=0.4.0->dataclasses_json==0.6.7) (4.15.0) 2025-11-03T18:03:04.3630138Z Requirement already satisfied: mypy-extensions>=0.3.0 in /home/ec2-user/.local/lib/python3.9/site-packages (from typing-inspect<1,>=0.4.0->dataclasses_json==0.6.7) (1.1.0) 2025-11-03T18:03:04.6397697Z Installing collected packages: python-dateutil, tzdata, numpy, pandas, boto3 2025-11-03T18:03:09.2675315Z Attempting uninstall: boto3 2025-11-03T18:03:09.2676513Z Found existing installation: boto3 1.35.33 2025-11-03T18:03:09.2785779Z Uninstalling boto3-1.35.33: 2025-11-03T18:03:09.2800810Z Successfully uninstalled boto3-1.35.33 2025-11-03T18:03:09.3815505Z Successfully installed boto3-1.35.42 numpy-1.26.4 pandas-2.1.3 python-dateutil-2.8.2 tzdata-2025.2 2025-11-03T18:03:09.6402704Z Command completed after 1 attempt(s). 2025-11-03T18:03:09.6537311Z ##[group]Run python3 -m tools.stats.upload_utilization_stats.upload_utilization_stats \ 2025-11-03T18:03:09.6539663Z python3 -m tools.stats.upload_utilization_stats.upload_utilization_stats \ 2025-11-03T18:03:09.6540077Z  --workflow-run-id "19040285420" \ 2025-11-03T18:03:09.6540355Z  --workflow-name "vllm-test" \ 2025-11-03T18:03:09.6540604Z  --workflow-run-attempt "1" \ 2025-11-03T18:03:09.6540835Z  --job-id "54383051168" \ 2025-11-03T18:03:09.6541260Z  --job-name "ci-vllm-test / test (vllm_basic_models_test, 1, 1, linux.g6.4xlarge.experimental.nvidia.gpu)" \ 2025-11-03T18:03:09.6541788Z  --local-path "" \ 2025-11-03T18:03:09.6542002Z  --artifact-prefix "" 2025-11-03T18:03:09.6549922Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2025-11-03T18:03:09.6550194Z env: 2025-11-03T18:03:09.6550359Z GIT_DEFAULT_BRANCH: main 2025-11-03T18:03:09.6550794Z GPU_FLAG: --gpus all -e NVIDIA_DRIVER_CAPABILITIES=all 2025-11-03T18:03:09.6551210Z DOCKER_CONTAINER_ID: c5e792ba8f7ad58ea91feb6e358bcdc693306479a78a475d1e0ca6fca41e9d73 2025-11-03T18:03:09.6551571Z DEVICE_NAME: 2025-11-03T18:03:09.6551731Z DEVICE_TYPE: 2025-11-03T18:03:09.6551887Z ##[endgroup] 2025-11-03T18:03:11.6630024Z repo: pytorch/pytorch 2025-11-03T18:03:11.6630417Z Search for test log in s3 bucket: ossci-utilization 2025-11-03T18:03:11.6631039Z Downloading logs-test-vllm_basic_models_test-1-1-linux.g6.4xlarge.experimental.nvidia.gpu_54383051168.zip 2025-11-03T18:03:11.6631902Z extracting usage_log.txt from zip file logs-test-vllm_basic_models_test-1-1-linux.g6.4xlarge.experimental.nvidia.gpu_54383051168.zip 2025-11-03T18:03:11.6632532Z Converted Log Model: UtilizationMetadata: 2025-11-03T18:03:11.6633835Z UtilizationMetadata(level='metadata', workflow_id='19040285420', job_id='54383051168', workflow_name='vllm-test', job_name='ci-vllm-test / test (vllm_basic_models_test, 1, 1, linux.g6.4xlarge.experimental.nvidia.gpu)', usage_collect_interval=1.0, data_model_version=1.5, start_at=1762189548, gpu_count=1, cpu_count=16, gpu_type='pynvml', error=None) 2025-11-03T18:03:11.6635443Z [Db Segments] detected pytest cmd: 12, generated segments: 12 2025-11-03T18:03:11.6635817Z [db model] Peek db timeseries 2025-11-03T18:03:11.6636053Z :{ 2025-11-03T18:03:11.6636227Z "created_at": 1762192991, 2025-11-03T18:03:11.6636482Z "type": "utilization", 2025-11-03T18:03:11.6636704Z "tags": [ 2025-11-03T18:03:11.6636864Z "record" 2025-11-03T18:03:11.6637034Z ], 2025-11-03T18:03:11.6637184Z "time_stamp": 1762189548, 2025-11-03T18:03:11.6637380Z "repo": "pytorch/pytorch", 2025-11-03T18:03:11.6637580Z "workflow_id": 19040285420, 2025-11-03T18:03:11.6637773Z "run_attempt": 1, 2025-11-03T18:03:11.6637946Z "job_id": 54383051168, 2025-11-03T18:03:11.6638134Z "workflow_name": "vllm-test", 2025-11-03T18:03:11.6638521Z "job_name": "ci-vllm-test / test (vllm_basic_models_test, 1, 1, linux.g6.4xlarge.experimental.nvidia.gpu)", 2025-11-03T18:03:11.6638898Z "json_data": "{}" 2025-11-03T18:03:11.6639062Z } 2025-11-03T18:03:11.6639413Z Writing 1 documents to S3 ossci-utilization/util_metadata/v_1.5/pytorch/pytorch/19040285420/1/54383051168/metadata 2025-11-03T18:03:11.6640070Z Done! Finish writing document to S3 ossci-utilization/util_metadata/v_1.5/pytorch/pytorch/19040285420/1/54383051168/metadata 2025-11-03T18:03:11.6640731Z Writing 685 documents to S3 ossci-utilization/util_timeseries/v_1.5/pytorch/pytorch/19040285420/1/54383051168/time_series 2025-11-03T18:03:11.6641543Z Done! Finish writing document to S3 ossci-utilization/util_timeseries/v_1.5/pytorch/pytorch/19040285420/1/54383051168/time_series 2025-11-03T18:03:11.7578616Z ##[group]Run pytorch/test-infra/.github/actions/teardown-linux@main 2025-11-03T18:03:11.7578977Z with: 2025-11-03T18:03:11.7579127Z env: 2025-11-03T18:03:11.7579284Z GIT_DEFAULT_BRANCH: main 2025-11-03T18:03:11.7579531Z GPU_FLAG: --gpus all -e NVIDIA_DRIVER_CAPABILITIES=all 2025-11-03T18:03:11.7579942Z DOCKER_CONTAINER_ID: c5e792ba8f7ad58ea91feb6e358bcdc693306479a78a475d1e0ca6fca41e9d73 2025-11-03T18:03:11.7580312Z DEVICE_NAME: 2025-11-03T18:03:11.7580467Z DEVICE_TYPE: 2025-11-03T18:03:11.7580622Z ##[endgroup] 2025-11-03T18:03:11.7672840Z ##[group]Run set -eou pipefail 2025-11-03T18:03:11.7673085Z set -eou pipefail 2025-11-03T18:03:11.7673279Z  2025-11-03T18:03:11.7673559Z echo "Holding runner for 2 hours until all ssh sessions have logged out" 2025-11-03T18:03:11.7673897Z for _ in $(seq 1440); do 2025-11-03T18:03:11.7674131Z  # Break if no ssh session exists anymore 2025-11-03T18:03:11.7674405Z  if [ "$(who)" = "" ]; then 2025-11-03T18:03:11.7674610Z  break 2025-11-03T18:03:11.7674766Z  fi 2025-11-03T18:03:11.7674924Z  echo "." 2025-11-03T18:03:11.7675105Z  sleep 5 2025-11-03T18:03:11.7675267Z done 2025-11-03T18:03:11.7683070Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2025-11-03T18:03:11.7683355Z env: 2025-11-03T18:03:11.7683512Z GIT_DEFAULT_BRANCH: main 2025-11-03T18:03:11.7683749Z GPU_FLAG: --gpus all -e NVIDIA_DRIVER_CAPABILITIES=all 2025-11-03T18:03:11.7684145Z DOCKER_CONTAINER_ID: c5e792ba8f7ad58ea91feb6e358bcdc693306479a78a475d1e0ca6fca41e9d73 2025-11-03T18:03:11.7684508Z DEVICE_NAME: 2025-11-03T18:03:11.7684674Z DEVICE_TYPE: 2025-11-03T18:03:11.7684831Z ##[endgroup] 2025-11-03T18:03:11.7713502Z Holding runner for 2 hours until all ssh sessions have logged out 2025-11-03T18:03:11.8234349Z ##[group]Run # ignore expansion of "docker ps -q" since it could be empty 2025-11-03T18:03:11.8234752Z # ignore expansion of "docker ps -q" since it could be empty 2025-11-03T18:03:11.8235063Z # shellcheck disable=SC2046 2025-11-03T18:03:11.8235313Z docker stop $(docker ps -q) || true 2025-11-03T18:03:11.8235561Z # Prune all of the docker images 2025-11-03T18:03:11.8235806Z docker system prune -af 2025-11-03T18:03:11.8243152Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2025-11-03T18:03:11.8243433Z env: 2025-11-03T18:03:11.8243586Z GIT_DEFAULT_BRANCH: main 2025-11-03T18:03:11.8243832Z GPU_FLAG: --gpus all -e NVIDIA_DRIVER_CAPABILITIES=all 2025-11-03T18:03:11.8244227Z DOCKER_CONTAINER_ID: c5e792ba8f7ad58ea91feb6e358bcdc693306479a78a475d1e0ca6fca41e9d73 2025-11-03T18:03:11.8244583Z DEVICE_NAME: 2025-11-03T18:03:11.8244755Z DEVICE_TYPE: 2025-11-03T18:03:11.8244909Z ##[endgroup] 2025-11-03T18:03:31.7023973Z c5e792ba8f7a 2025-11-03T18:03:41.3398852Z Deleted Containers: 2025-11-03T18:03:41.3399350Z c5e792ba8f7ad58ea91feb6e358bcdc693306479a78a475d1e0ca6fca41e9d73 2025-11-03T18:03:41.3399690Z 2025-11-03T18:03:50.6275216Z Deleted Images: 2025-11-03T18:03:50.6276058Z untagged: 308535385114.dkr.ecr.us-east-1.amazonaws.com/pytorch/ci-image:pytorch-linux-jammy-cuda12.8-cudnn9-py3.12-gcc11-vllm-30147fb242f88ab8f13e14b76ab186d189d8605c 2025-11-03T18:03:50.6277219Z untagged: 308535385114.dkr.ecr.us-east-1.amazonaws.com/pytorch/ci-image@sha256:655b37e47c7ab0fab00339d6162ef718600c6b4e76f1601297e6a07026860230 2025-11-03T18:03:50.6278049Z deleted: sha256:123a0a8d2d8124dd2cca825f1276a192d04ee365eb7de45a4829163bae42bcc5 2025-11-03T18:03:50.6278614Z deleted: sha256:31820faffa36fa10d707f81682d96bdb5de2207e3612a625bc2add174a16f33b 2025-11-03T18:03:50.6279167Z deleted: sha256:1bbaafd08d25d515d9317d42cef8b0d6d389624b4d01c65b2374e72e58a98012 2025-11-03T18:03:50.6279910Z deleted: sha256:3a98336219906852cadb9ffa8b260359b85b0ef2ce63f194b2ad17340bf2c5d6 2025-11-03T18:03:50.6280340Z deleted: sha256:a73360da0fd9b9cc724c9a24f957b8b4bd0936a4f97ee5c2fc41a13d6994b0bb 2025-11-03T18:03:50.6281004Z deleted: sha256:38cc6c78e3939cce4347b386ded8480f5b3ba61185cda08e95c53c55b71d1698 2025-11-03T18:03:50.6281474Z deleted: sha256:77104ac67f1c4a71d44f95e55721e88d0d338ccb804df56f916ea3edd8c57177 2025-11-03T18:03:50.6281931Z deleted: sha256:e81aea95b6a623c6eea401a37492286ab3ee28e5042b808f2c9328ba1ba29d53 2025-11-03T18:03:50.6282391Z deleted: sha256:b0de82ab4f930d742562dc8beffaeb4f9e2e37888069111da820e249172d0c04 2025-11-03T18:03:50.6282837Z deleted: sha256:47d87661555964c91d4fc524b43b15505dcbecdf57b6167adcf9774f1c597d53 2025-11-03T18:03:50.6283268Z deleted: sha256:ce3967b7450d7c2a5833ce475b87987bb427bb87bb87d421aaa98cfe8fe093d3 2025-11-03T18:03:50.6283697Z deleted: sha256:65cde699f23961ef3b15b672e8715d50af40ec180991a3fe50c708d5d955da37 2025-11-03T18:03:50.6284130Z deleted: sha256:bda9c0867d0e7ac8556b4cfc02c04d75bd9e845a6b579e50ddbeac76f48824ae 2025-11-03T18:03:50.6284566Z deleted: sha256:b185a3d739322209684243ec5a50898a86f3c18f0058921dcaff683e1a759f53 2025-11-03T18:03:50.6284998Z deleted: sha256:f3b6f40043f4f16c2f36594521bcd3561ae8d9e84c3a1baac1fdd6be2b1d2503 2025-11-03T18:03:50.6285438Z deleted: sha256:c5d92fcf72e25e113b2514a4e18a1faf280ff97600c44e089c8add4ba4bfed67 2025-11-03T18:03:50.6285873Z deleted: sha256:e1b08cb82b0287ee00554e71dc6e69dd5747e05995a19d5be2427a0a0e2ea298 2025-11-03T18:03:50.6286295Z deleted: sha256:f514e296acfd635f5184645775136c340b4bdab12aa80c62965f8fd66a9adc2d 2025-11-03T18:03:50.6286744Z deleted: sha256:a7b8036e533224113c12edff1c24cbd4e6d1b2bfc15ea49643f262a00309092b 2025-11-03T18:03:50.6287175Z deleted: sha256:f23436ff58fe28485a737310c5367061602a90ab5856989457daa060a372a0ae 2025-11-03T18:03:50.6287611Z deleted: sha256:6e1a14c6f6f6089f5b4be23f93de15f5c6898accb4fdc2c577df25bef047c66c 2025-11-03T18:03:50.6288051Z deleted: sha256:cb7a3bd4bcc1a4223e27af60de87c05873f833c649e227b9afb63f5bb208cbb5 2025-11-03T18:03:50.6288490Z deleted: sha256:45210d5bac2c05cfd830e1097010555216604bafbfb71ea5495a0e4b224b7f9a 2025-11-03T18:03:50.6288941Z deleted: sha256:afdad6a8611fbabce05e6714dfffad4c1675e10d3b1e039749ce204b0f8cd41c 2025-11-03T18:03:50.6289375Z deleted: sha256:a44823411001fe4089d1c72c3d5de14d0b2687232a5d8ced0818589e3325316d 2025-11-03T18:03:50.6289801Z deleted: sha256:f263db78292410b172d222dbe7c398e9d8c87663eb725ca5f3ea998b4b2ed080 2025-11-03T18:03:50.6290233Z deleted: sha256:13ecce5bfe277268a0de6e0d01ae45821caad15c1a608da44f82bbc57cbdc020 2025-11-03T18:03:50.6290783Z deleted: sha256:9ab9deacc335d4c360367e7f061aaf66216e015d9d058a73260f35efe0b73b9d 2025-11-03T18:03:50.6291226Z deleted: sha256:bfed55b6ddec7b775d5fa270eb2be23a41a6069ca17ea5b3dd991f8f67712dd8 2025-11-03T18:03:50.6291673Z deleted: sha256:982e6e5dfbb4728f4a51c83a1896cc16587d550ae6f515f89fa807dc87046bfb 2025-11-03T18:03:50.6292107Z deleted: sha256:df42847c61a0101cc7348e49cd2a500cdbbde2f218e4d6fe58ab752b51a43fac 2025-11-03T18:03:50.6292546Z deleted: sha256:de33546992ae1fce63d51acabeec23ebeeaba447c7155cfee96ac2d115b3640f 2025-11-03T18:03:50.6293023Z deleted: sha256:45ba9418147153ef58cb0e6509c4a9139163afeccd32dc80e8a335b32c050b07 2025-11-03T18:03:50.6293468Z deleted: sha256:4e36f594c176514603ccb1a2c4d5e4900ed546cb9499dcc3f6061aab9519c57f 2025-11-03T18:03:50.6293906Z deleted: sha256:a83e194f5cc8b355dda2be872a19f4dc25ed2314ee7952374b67baf90977c9da 2025-11-03T18:03:50.6294335Z deleted: sha256:445972ee71b4614ff3f1d009fba94556bf17e09b2c99292a908ebf5e9529e1c7 2025-11-03T18:03:50.6294758Z deleted: sha256:d2e38abb0bff4a6968be64628e6a661db235e0530d5925a552991f3fe1ce1a30 2025-11-03T18:03:50.6295202Z deleted: sha256:7275b58f0e992bd8031ef2ea708bd2c3b080ecaaecbbebbb2d438ef4413913c1 2025-11-03T18:03:50.6295635Z deleted: sha256:030aae20394d8c1c0d34388b4856c6a2966f30f4cd8393cd33d1defccd696a55 2025-11-03T18:03:50.6296066Z deleted: sha256:f4d5ea7a039db19152ce6bc9827e57a4de69ed4fb053f8e50de52eac2410daca 2025-11-03T18:03:50.6296565Z deleted: sha256:545e10edde5c7ee4d66227f76f9b63db14d5f1b03e8588f429c3961260a2bf09 2025-11-03T18:03:50.6296990Z deleted: sha256:8eced9c38b8d1316270f6ab88b14de97a711d07c7b668f568bd501dc518e7ced 2025-11-03T18:03:50.6297523Z deleted: sha256:709ec82360eef380a961d01fea671225c186880c6cf3375daca37712840f8549 2025-11-03T18:03:50.6297950Z deleted: sha256:7bb05462f24d343c9c9e7e2d24c62d8ba06495e5e07f040ab95a8eaf29724acf 2025-11-03T18:03:50.6298381Z deleted: sha256:0dfdf4b3e5ed9f1d5f79747ed2d29e2b87752d07331442c42efa1a613c1011f9 2025-11-03T18:03:50.6298815Z deleted: sha256:efb8a361d5f170f02f902aaad54acb9dcd69e2b4da2ae2470934d8dad2d6cb25 2025-11-03T18:03:50.6299264Z deleted: sha256:9e8da81757d04e51ce5c02529d343ff92c351b8435f829866a45df37cd4a65eb 2025-11-03T18:03:50.6299689Z deleted: sha256:6a408c52616752ec8052b86a8896a66ea203cfedebd61abc3e660ed783e6778d 2025-11-03T18:03:50.6300120Z deleted: sha256:7cf69cce58132deb3c49815b6e82debd5687c108df52ebc97770b29c97d0c1b6 2025-11-03T18:03:50.6300548Z deleted: sha256:64540da67f68a44e7e37dd5c80e72066f8394abdd1d890073f301097f6234e04 2025-11-03T18:03:50.6300990Z deleted: sha256:3c3d38edfc6f883c7e76990a02f0e977d3d284835cabe63f596c134a4f0408c1 2025-11-03T18:03:50.6301560Z deleted: sha256:dbfcbb7dcb1e0a93c3b284a39a54ac20951ca01bd43d222ef6c21a9a958ef19d 2025-11-03T18:03:50.6302007Z deleted: sha256:89230198d83d52abcd62d73aa0c0919ae60e4c3b0552805ad87fb0e14c78f0ff 2025-11-03T18:03:50.6302440Z deleted: sha256:529d43aa0a1d6c2a11e0ae029bb4a68c2a3cde415a77d094886d6db98794e336 2025-11-03T18:03:50.6302866Z deleted: sha256:a8066515cbb348517543b217f1dcc01167d7202a7dbea9745f573c02834d65a3 2025-11-03T18:03:50.6313459Z deleted: sha256:a13e9abfebb1053045eead7615d90161552798de57c43eaf337b3e64a1140915 2025-11-03T18:03:50.6314029Z deleted: sha256:19ee032c28086715c5dfc7960eaa38dfb68db347946624410c344101d7dee4fb 2025-11-03T18:03:50.6314499Z deleted: sha256:c3a1419c3b6e2c7a3aa39b66ce90f210b6de4804a6a494babc21bda3b8ddb728 2025-11-03T18:03:50.6314949Z deleted: sha256:f6dd774d7904ac47308018e060dd1a6460f1e6b0b495e4637d6dec0e766bbef9 2025-11-03T18:03:50.6315412Z deleted: sha256:b14fbc43504939ea343b2c5ba35bba56b52e208d1262a383ef185124c12f5738 2025-11-03T18:03:50.6315880Z deleted: sha256:58dac273a49f6d6ffd5ecaf85dcf3874a239a37db02320098059eb5686d33ae8 2025-11-03T18:03:50.6316318Z deleted: sha256:7c4c937e97b6d494fb2e144665b1594212d1b854a880e74265e7f2d70f3c3944 2025-11-03T18:03:50.6316748Z deleted: sha256:bf342185a1f51131a608e59d62ad65da006e6322d6aa191876417eef3c624118 2025-11-03T18:03:50.6317174Z deleted: sha256:121408b55e8d21cee32f17ecb5d380729586f6b7d8198be62423161a316f5909 2025-11-03T18:03:50.6317608Z deleted: sha256:fd8ef0586f9366378d192cf475e6e783c5815185a528b077d533f9ecb4f6881d 2025-11-03T18:03:50.6318145Z deleted: sha256:4b4bbdfa85b1154871d9d703b6578679d6ad436b0d157acc03589b20f7e51174 2025-11-03T18:03:50.6318576Z deleted: sha256:a7ec280e3d580ac545fdddb47716864c39c8d5cbd27c48487fb39f6d52ec5813 2025-11-03T18:03:50.6319015Z deleted: sha256:af6a46762136c8cc3cca20b4f523662bda54d2a17a70900b436f5829c2694ab0 2025-11-03T18:03:50.6319459Z deleted: sha256:ed75bd3dae248978cc1d7f962d6313404c2f2cf4c5f5defa282a77834011b95a 2025-11-03T18:03:50.6319908Z deleted: sha256:baab5a5b13fbe97055c0eee16f014775c75da39ee1bfd50ce919e42fd0740eb3 2025-11-03T18:03:50.6320356Z deleted: sha256:abc859f44381c472c6110fe3ea1d5db4f58416a2e4ac461bd7a636be635d19f1 2025-11-03T18:03:50.6320790Z deleted: sha256:120e4c672fc1d0e7e22a7dcfb14bc009fdc45208d24cc41e4fed2c865d4c61b4 2025-11-03T18:03:50.6321544Z deleted: sha256:3a901973b8d19183073f3925d3581e2124a7e019528f651c18ed984dc8920633 2025-11-03T18:03:50.6321966Z deleted: sha256:13d9cf2669287254c4fc4922600894d75d97fb756c308b92ae4067f685b3b9c4 2025-11-03T18:03:50.6322398Z deleted: sha256:d7cc64ff2a7a5cdf4d2e027c93b11862cb741b68bd813300f0e850d03b5894de 2025-11-03T18:03:50.6322829Z deleted: sha256:ddb9c9f2a33a6ff22294d8f0450ec0da44874318928f90f1f53c9e51bd736c90 2025-11-03T18:03:50.6323282Z deleted: sha256:b615ed2509100fbf1d31da5895e95df9b6d890ef5b8cb725aea547c2310d370d 2025-11-03T18:03:50.6323719Z deleted: sha256:cf4cba224210966dd0f57c902103bbe1142ffc9ae140509ba29c98b14df150d2 2025-11-03T18:03:50.6324251Z deleted: sha256:63cc9363306d393f09857fa7d93605c66c884fdcbbb254c37474b6cf4f293a35 2025-11-03T18:03:50.6324819Z deleted: sha256:767e56ba346ae714b6e6b816baa839051145ed78cfa0e4524a86cc287b0c4b00 2025-11-03T18:03:50.6325196Z untagged: public.ecr.aws/docker/library/python:3.13 2025-11-03T18:03:50.6325683Z untagged: public.ecr.aws/docker/library/python@sha256:12513c633252a28bcfee85839aa384e1af322f11275779c6645076c6cd0cfe52 2025-11-03T18:03:50.6326241Z deleted: sha256:438f5c84513b2fc181c3e3cf09b621e66b769789fab066c9c436a35b8588d29e 2025-11-03T18:03:50.6326678Z deleted: sha256:007b97ee297d1b90751d5850e62f9ebc070d655634bd08aa0a4b83fea3a53a14 2025-11-03T18:03:50.6327101Z deleted: sha256:ee90a75564a713c0ee50c40d6581495f346f765072ae44828bf219e3da9fb55c 2025-11-03T18:03:50.6327530Z deleted: sha256:788fe66fc657a4e83d7248083c07adcd949d1e8a803c2835d66e07bff863f4d8 2025-11-03T18:03:50.6327970Z deleted: sha256:a60dcbc8722b7ecb7798ea22a4dad988c0ce2ccb1efb598aa8deee75026f3e1d 2025-11-03T18:03:50.6328413Z deleted: sha256:f8b7333a4ac07b3cf4c5976461e665c3445b16dacf992d1d3130e2d39ebe2570 2025-11-03T18:03:50.6328856Z deleted: sha256:77c2e291dd8342baaa9f49b9265f1925e1357f6a98a61f736d060eaaf3cd7b3b 2025-11-03T18:03:50.6329274Z deleted: sha256:f2522c6ed78b338a9e272dd5038005d008f74729e036073e837f701f221b99ba 2025-11-03T18:03:50.6329527Z 2025-11-03T18:03:50.6329620Z Total reclaimed space: 66.66GB 2025-11-03T18:03:50.6406105Z Post job cleanup. 2025-11-03T18:03:50.6451678Z Post job cleanup. 2025-11-03T18:03:50.7384235Z [command]/usr/bin/git version 2025-11-03T18:03:50.7439159Z git version 2.50.1 2025-11-03T18:03:50.7472730Z Copying '/home/ec2-user/.gitconfig' to '/home/ec2-user/actions-runner/_work/_temp/456946b8-9a03-4fe9-ad58-9dd0de1add22/.gitconfig' 2025-11-03T18:03:50.7482225Z Temporarily overriding HOME='/home/ec2-user/actions-runner/_work/_temp/456946b8-9a03-4fe9-ad58-9dd0de1add22' before making global git config changes 2025-11-03T18:03:50.7482928Z Adding repository directory to the temporary git global config as a safe directory 2025-11-03T18:03:50.7487048Z [command]/usr/bin/git config --global --add safe.directory /home/ec2-user/actions-runner/_work/pytorch/pytorch 2025-11-03T18:03:50.7528769Z [command]/usr/bin/git config --local --name-only --get-regexp core\.sshCommand 2025-11-03T18:03:50.7568457Z [command]/usr/bin/git submodule foreach --recursive sh -c "git config --local --name-only --get-regexp 'core\.sshCommand' && git config --local --unset-all 'core.sshCommand' || :" 2025-11-03T18:03:50.7947083Z Entering 'android/libs/fbjni' 2025-11-03T18:03:50.8022000Z Entering 'third_party/FP16' 2025-11-03T18:03:50.8093193Z Entering 'third_party/FXdiv' 2025-11-03T18:03:50.8164197Z Entering 'third_party/NNPACK' 2025-11-03T18:03:50.8234909Z Entering 'third_party/NVTX' 2025-11-03T18:03:50.8309472Z Entering 'third_party/VulkanMemoryAllocator' 2025-11-03T18:03:50.8382927Z Entering 'third_party/XNNPACK' 2025-11-03T18:03:50.8467281Z Entering 'third_party/aiter' 2025-11-03T18:03:50.8541658Z Entering 'third_party/aiter/3rdparty/composable_kernel' 2025-11-03T18:03:50.8623900Z Entering 'third_party/benchmark' 2025-11-03T18:03:50.8694669Z Entering 'third_party/composable_kernel' 2025-11-03T18:03:50.8777334Z Entering 'third_party/cpp-httplib' 2025-11-03T18:03:50.8851666Z Entering 'third_party/cpuinfo' 2025-11-03T18:03:50.8923977Z Entering 'third_party/cudnn_frontend' 2025-11-03T18:03:50.8995225Z Entering 'third_party/cutlass' 2025-11-03T18:03:50.9074768Z Entering 'third_party/fbgemm' 2025-11-03T18:03:50.9146413Z Entering 'third_party/fbgemm/external/asmjit' 2025-11-03T18:03:50.9220742Z Entering 'third_party/fbgemm/external/composable_kernel' 2025-11-03T18:03:50.9299526Z Entering 'third_party/fbgemm/external/cpuinfo' 2025-11-03T18:03:50.9374645Z Entering 'third_party/fbgemm/external/cutlass' 2025-11-03T18:03:50.9450982Z Entering 'third_party/fbgemm/external/googletest' 2025-11-03T18:03:50.9524171Z Entering 'third_party/fbgemm/external/hipify_torch' 2025-11-03T18:03:50.9592599Z Entering 'third_party/fbgemm/external/json' 2025-11-03T18:03:50.9666916Z Entering 'third_party/flash-attention' 2025-11-03T18:03:50.9739991Z Entering 'third_party/flash-attention/csrc/composable_kernel' 2025-11-03T18:03:50.9816632Z Entering 'third_party/flash-attention/csrc/cutlass' 2025-11-03T18:03:50.9899560Z Entering 'third_party/flatbuffers' 2025-11-03T18:03:50.9975307Z Entering 'third_party/fmt' 2025-11-03T18:03:51.0045428Z Entering 'third_party/gemmlowp/gemmlowp' 2025-11-03T18:03:51.0117220Z Entering 'third_party/gloo' 2025-11-03T18:03:51.0190859Z Entering 'third_party/googletest' 2025-11-03T18:03:51.0261178Z Entering 'third_party/ideep' 2025-11-03T18:03:51.0332388Z Entering 'third_party/ideep/mkl-dnn' 2025-11-03T18:03:51.0411931Z Entering 'third_party/ittapi' 2025-11-03T18:03:51.0485031Z Entering 'third_party/kineto' 2025-11-03T18:03:51.0554644Z Entering 'third_party/kineto/libkineto/third_party/dynolog' 2025-11-03T18:03:51.0625522Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/DCGM' 2025-11-03T18:03:51.0695912Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/cpr' 2025-11-03T18:03:51.0765514Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/fmt' 2025-11-03T18:03:51.0839795Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/gflags' 2025-11-03T18:03:51.0910097Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/gflags/doc' 2025-11-03T18:03:51.0984682Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/glog' 2025-11-03T18:03:51.1055345Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/googletest' 2025-11-03T18:03:51.1129634Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/json' 2025-11-03T18:03:51.1202097Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/pfs' 2025-11-03T18:03:51.1273366Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/prometheus-cpp' 2025-11-03T18:03:51.1345055Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/prometheus-cpp/3rdparty/civetweb' 2025-11-03T18:03:51.1415564Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/prometheus-cpp/3rdparty/googletest' 2025-11-03T18:03:51.1492530Z Entering 'third_party/kineto/libkineto/third_party/fmt' 2025-11-03T18:03:51.1565315Z Entering 'third_party/kineto/libkineto/third_party/googletest' 2025-11-03T18:03:51.1637508Z Entering 'third_party/kleidiai' 2025-11-03T18:03:51.1710899Z Entering 'third_party/mimalloc' 2025-11-03T18:03:51.1783859Z Entering 'third_party/nlohmann' 2025-11-03T18:03:51.1855581Z Entering 'third_party/onnx' 2025-11-03T18:03:51.1939387Z Entering 'third_party/onnx/third_party/pybind11' 2025-11-03T18:03:51.2013935Z Entering 'third_party/opentelemetry-cpp' 2025-11-03T18:03:51.2084006Z Entering 'third_party/opentelemetry-cpp/third_party/benchmark' 2025-11-03T18:03:51.2153081Z Entering 'third_party/opentelemetry-cpp/third_party/googletest' 2025-11-03T18:03:51.2223008Z Entering 'third_party/opentelemetry-cpp/third_party/ms-gsl' 2025-11-03T18:03:51.2291637Z Entering 'third_party/opentelemetry-cpp/third_party/nlohmann-json' 2025-11-03T18:03:51.2364405Z Entering 'third_party/opentelemetry-cpp/third_party/opentelemetry-proto' 2025-11-03T18:03:51.2432424Z Entering 'third_party/opentelemetry-cpp/third_party/opentracing-cpp' 2025-11-03T18:03:51.2501987Z Entering 'third_party/opentelemetry-cpp/third_party/prometheus-cpp' 2025-11-03T18:03:51.2569889Z Entering 'third_party/opentelemetry-cpp/third_party/prometheus-cpp/3rdparty/civetweb' 2025-11-03T18:03:51.2642360Z Entering 'third_party/opentelemetry-cpp/third_party/prometheus-cpp/3rdparty/googletest' 2025-11-03T18:03:51.2715878Z Entering 'third_party/opentelemetry-cpp/tools/vcpkg' 2025-11-03T18:03:51.2806982Z Entering 'third_party/pocketfft' 2025-11-03T18:03:51.2881069Z Entering 'third_party/protobuf' 2025-11-03T18:03:51.2953443Z Entering 'third_party/protobuf/third_party/benchmark' 2025-11-03T18:03:51.3022699Z Entering 'third_party/protobuf/third_party/googletest' 2025-11-03T18:03:51.3095595Z Entering 'third_party/psimd' 2025-11-03T18:03:51.3166050Z Entering 'third_party/pthreadpool' 2025-11-03T18:03:51.3235328Z Entering 'third_party/pybind11' 2025-11-03T18:03:51.3312096Z Entering 'third_party/python-peachpy' 2025-11-03T18:03:51.3381955Z Entering 'third_party/sleef' 2025-11-03T18:03:51.3455063Z Entering 'third_party/tensorpipe' 2025-11-03T18:03:51.3524408Z Entering 'third_party/tensorpipe/third_party/googletest' 2025-11-03T18:03:51.3595060Z Entering 'third_party/tensorpipe/third_party/libnop' 2025-11-03T18:03:51.3663352Z Entering 'third_party/tensorpipe/third_party/libuv' 2025-11-03T18:03:51.3733268Z Entering 'third_party/tensorpipe/third_party/pybind11' 2025-11-03T18:03:51.3800474Z Entering 'third_party/tensorpipe/third_party/pybind11/tools/clang' 2025-11-03T18:03:51.3899114Z [command]/usr/bin/git config --local --name-only --get-regexp http\.https\:\/\/github\.com\/\.extraheader 2025-11-03T18:03:51.3923523Z http.https://github.com/.extraheader 2025-11-03T18:03:51.3933103Z [command]/usr/bin/git config --local --unset-all http.https://github.com/.extraheader 2025-11-03T18:03:51.3967100Z [command]/usr/bin/git submodule foreach --recursive sh -c "git config --local --name-only --get-regexp 'http\.https\:\/\/github\.com\/\.extraheader' && git config --local --unset-all 'http.https://github.com/.extraheader' || :" 2025-11-03T18:03:51.4330937Z Entering 'android/libs/fbjni' 2025-11-03T18:03:51.4378526Z http.https://github.com/.extraheader 2025-11-03T18:03:51.4423257Z Entering 'third_party/FP16' 2025-11-03T18:03:51.4470016Z http.https://github.com/.extraheader 2025-11-03T18:03:51.4513568Z Entering 'third_party/FXdiv' 2025-11-03T18:03:51.4559804Z http.https://github.com/.extraheader 2025-11-03T18:03:51.4605057Z Entering 'third_party/NNPACK' 2025-11-03T18:03:51.4653654Z http.https://github.com/.extraheader 2025-11-03T18:03:51.4697720Z Entering 'third_party/NVTX' 2025-11-03T18:03:51.4742522Z http.https://github.com/.extraheader 2025-11-03T18:03:51.4785917Z Entering 'third_party/VulkanMemoryAllocator' 2025-11-03T18:03:51.4835227Z http.https://github.com/.extraheader 2025-11-03T18:03:51.4883068Z Entering 'third_party/XNNPACK' 2025-11-03T18:03:51.4930005Z http.https://github.com/.extraheader 2025-11-03T18:03:51.4987599Z Entering 'third_party/aiter' 2025-11-03T18:03:51.5035455Z http.https://github.com/.extraheader 2025-11-03T18:03:51.5082039Z Entering 'third_party/aiter/3rdparty/composable_kernel' 2025-11-03T18:03:51.5128570Z http.https://github.com/.extraheader 2025-11-03T18:03:51.5183053Z Entering 'third_party/benchmark' 2025-11-03T18:03:51.5231026Z http.https://github.com/.extraheader 2025-11-03T18:03:51.5273426Z Entering 'third_party/composable_kernel' 2025-11-03T18:03:51.5320940Z http.https://github.com/.extraheader 2025-11-03T18:03:51.5372707Z Entering 'third_party/cpp-httplib' 2025-11-03T18:03:51.5418801Z http.https://github.com/.extraheader 2025-11-03T18:03:51.5467299Z Entering 'third_party/cpuinfo' 2025-11-03T18:03:51.5515741Z http.https://github.com/.extraheader 2025-11-03T18:03:51.5563063Z Entering 'third_party/cudnn_frontend' 2025-11-03T18:03:51.5608634Z http.https://github.com/.extraheader 2025-11-03T18:03:51.5654446Z Entering 'third_party/cutlass' 2025-11-03T18:03:51.5700347Z http.https://github.com/.extraheader 2025-11-03T18:03:51.5753616Z Entering 'third_party/fbgemm' 2025-11-03T18:03:51.5800115Z http.https://github.com/.extraheader 2025-11-03T18:03:51.5845920Z Entering 'third_party/fbgemm/external/asmjit' 2025-11-03T18:03:51.5890880Z http.https://github.com/.extraheader 2025-11-03T18:03:51.5934763Z Entering 'third_party/fbgemm/external/composable_kernel' 2025-11-03T18:03:51.5979945Z http.https://github.com/.extraheader 2025-11-03T18:03:51.6031986Z Entering 'third_party/fbgemm/external/cpuinfo' 2025-11-03T18:03:51.6083378Z http.https://github.com/.extraheader 2025-11-03T18:03:51.6129605Z Entering 'third_party/fbgemm/external/cutlass' 2025-11-03T18:03:51.6176521Z http.https://github.com/.extraheader 2025-11-03T18:03:51.6234064Z Entering 'third_party/fbgemm/external/googletest' 2025-11-03T18:03:51.6279761Z http.https://github.com/.extraheader 2025-11-03T18:03:51.6324585Z Entering 'third_party/fbgemm/external/hipify_torch' 2025-11-03T18:03:51.6371022Z http.https://github.com/.extraheader 2025-11-03T18:03:51.6414133Z Entering 'third_party/fbgemm/external/json' 2025-11-03T18:03:51.6460359Z http.https://github.com/.extraheader 2025-11-03T18:03:51.6508059Z Entering 'third_party/flash-attention' 2025-11-03T18:03:51.6554575Z http.https://github.com/.extraheader 2025-11-03T18:03:51.6603647Z Entering 'third_party/flash-attention/csrc/composable_kernel' 2025-11-03T18:03:51.6648772Z http.https://github.com/.extraheader 2025-11-03T18:03:51.6699258Z Entering 'third_party/flash-attention/csrc/cutlass' 2025-11-03T18:03:51.6750110Z http.https://github.com/.extraheader 2025-11-03T18:03:51.6803546Z Entering 'third_party/flatbuffers' 2025-11-03T18:03:51.6849593Z http.https://github.com/.extraheader 2025-11-03T18:03:51.6896882Z Entering 'third_party/fmt' 2025-11-03T18:03:51.6942697Z http.https://github.com/.extraheader 2025-11-03T18:03:51.6984642Z Entering 'third_party/gemmlowp/gemmlowp' 2025-11-03T18:03:51.7030827Z http.https://github.com/.extraheader 2025-11-03T18:03:51.7074078Z Entering 'third_party/gloo' 2025-11-03T18:03:51.7120525Z http.https://github.com/.extraheader 2025-11-03T18:03:51.7166065Z Entering 'third_party/googletest' 2025-11-03T18:03:51.7211350Z http.https://github.com/.extraheader 2025-11-03T18:03:51.7255452Z Entering 'third_party/ideep' 2025-11-03T18:03:51.7300629Z http.https://github.com/.extraheader 2025-11-03T18:03:51.7342764Z Entering 'third_party/ideep/mkl-dnn' 2025-11-03T18:03:51.7389082Z http.https://github.com/.extraheader 2025-11-03T18:03:51.7443705Z Entering 'third_party/ittapi' 2025-11-03T18:03:51.7489943Z http.https://github.com/.extraheader 2025-11-03T18:03:51.7534023Z Entering 'third_party/kineto' 2025-11-03T18:03:51.7579893Z http.https://github.com/.extraheader 2025-11-03T18:03:51.7623891Z Entering 'third_party/kineto/libkineto/third_party/dynolog' 2025-11-03T18:03:51.7669956Z http.https://github.com/.extraheader 2025-11-03T18:03:51.7712972Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/DCGM' 2025-11-03T18:03:51.7759607Z http.https://github.com/.extraheader 2025-11-03T18:03:51.7805366Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/cpr' 2025-11-03T18:03:51.7851989Z http.https://github.com/.extraheader 2025-11-03T18:03:51.7898444Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/fmt' 2025-11-03T18:03:51.7947157Z http.https://github.com/.extraheader 2025-11-03T18:03:51.7992632Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/gflags' 2025-11-03T18:03:51.8039943Z http.https://github.com/.extraheader 2025-11-03T18:03:51.8082798Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/gflags/doc' 2025-11-03T18:03:51.8130494Z http.https://github.com/.extraheader 2025-11-03T18:03:51.8177609Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/glog' 2025-11-03T18:03:51.8227623Z http.https://github.com/.extraheader 2025-11-03T18:03:51.8273041Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/googletest' 2025-11-03T18:03:51.8319470Z http.https://github.com/.extraheader 2025-11-03T18:03:51.8363417Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/json' 2025-11-03T18:03:51.8409827Z http.https://github.com/.extraheader 2025-11-03T18:03:51.8456269Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/pfs' 2025-11-03T18:03:51.8501577Z http.https://github.com/.extraheader 2025-11-03T18:03:51.8546378Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/prometheus-cpp' 2025-11-03T18:03:51.8591930Z http.https://github.com/.extraheader 2025-11-03T18:03:51.8636897Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/prometheus-cpp/3rdparty/civetweb' 2025-11-03T18:03:51.8682113Z http.https://github.com/.extraheader 2025-11-03T18:03:51.8730133Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/prometheus-cpp/3rdparty/googletest' 2025-11-03T18:03:51.8777621Z http.https://github.com/.extraheader 2025-11-03T18:03:51.8830335Z Entering 'third_party/kineto/libkineto/third_party/fmt' 2025-11-03T18:03:51.8877553Z http.https://github.com/.extraheader 2025-11-03T18:03:51.8921373Z Entering 'third_party/kineto/libkineto/third_party/googletest' 2025-11-03T18:03:51.8968343Z http.https://github.com/.extraheader 2025-11-03T18:03:51.9016058Z Entering 'third_party/kleidiai' 2025-11-03T18:03:51.9061944Z http.https://github.com/.extraheader 2025-11-03T18:03:51.9105728Z Entering 'third_party/mimalloc' 2025-11-03T18:03:51.9154667Z http.https://github.com/.extraheader 2025-11-03T18:03:51.9200140Z Entering 'third_party/nlohmann' 2025-11-03T18:03:51.9248146Z http.https://github.com/.extraheader 2025-11-03T18:03:51.9293824Z Entering 'third_party/onnx' 2025-11-03T18:03:51.9339564Z http.https://github.com/.extraheader 2025-11-03T18:03:51.9398774Z Entering 'third_party/onnx/third_party/pybind11' 2025-11-03T18:03:51.9446165Z http.https://github.com/.extraheader 2025-11-03T18:03:51.9495657Z Entering 'third_party/opentelemetry-cpp' 2025-11-03T18:03:51.9541316Z http.https://github.com/.extraheader 2025-11-03T18:03:51.9585308Z Entering 'third_party/opentelemetry-cpp/third_party/benchmark' 2025-11-03T18:03:51.9630673Z http.https://github.com/.extraheader 2025-11-03T18:03:51.9674550Z Entering 'third_party/opentelemetry-cpp/third_party/googletest' 2025-11-03T18:03:51.9719901Z http.https://github.com/.extraheader 2025-11-03T18:03:51.9762908Z Entering 'third_party/opentelemetry-cpp/third_party/ms-gsl' 2025-11-03T18:03:51.9808999Z http.https://github.com/.extraheader 2025-11-03T18:03:51.9852714Z Entering 'third_party/opentelemetry-cpp/third_party/nlohmann-json' 2025-11-03T18:03:51.9901335Z http.https://github.com/.extraheader 2025-11-03T18:03:51.9947235Z Entering 'third_party/opentelemetry-cpp/third_party/opentelemetry-proto' 2025-11-03T18:03:51.9991998Z http.https://github.com/.extraheader 2025-11-03T18:03:52.0038465Z Entering 'third_party/opentelemetry-cpp/third_party/opentracing-cpp' 2025-11-03T18:03:52.0086102Z http.https://github.com/.extraheader 2025-11-03T18:03:52.0132207Z Entering 'third_party/opentelemetry-cpp/third_party/prometheus-cpp' 2025-11-03T18:03:52.0177788Z http.https://github.com/.extraheader 2025-11-03T18:03:52.0220905Z Entering 'third_party/opentelemetry-cpp/third_party/prometheus-cpp/3rdparty/civetweb' 2025-11-03T18:03:52.0268687Z http.https://github.com/.extraheader 2025-11-03T18:03:52.0316040Z Entering 'third_party/opentelemetry-cpp/third_party/prometheus-cpp/3rdparty/googletest' 2025-11-03T18:03:52.0361793Z http.https://github.com/.extraheader 2025-11-03T18:03:52.0408133Z Entering 'third_party/opentelemetry-cpp/tools/vcpkg' 2025-11-03T18:03:52.0456344Z http.https://github.com/.extraheader 2025-11-03T18:03:52.0520665Z Entering 'third_party/pocketfft' 2025-11-03T18:03:52.0569123Z http.https://github.com/.extraheader 2025-11-03T18:03:52.0613072Z Entering 'third_party/protobuf' 2025-11-03T18:03:52.0660801Z http.https://github.com/.extraheader 2025-11-03T18:03:52.0706236Z Entering 'third_party/protobuf/third_party/benchmark' 2025-11-03T18:03:52.0755005Z http.https://github.com/.extraheader 2025-11-03T18:03:52.0800672Z Entering 'third_party/protobuf/third_party/googletest' 2025-11-03T18:03:52.0848027Z http.https://github.com/.extraheader 2025-11-03T18:03:52.0895110Z Entering 'third_party/psimd' 2025-11-03T18:03:52.0941072Z http.https://github.com/.extraheader 2025-11-03T18:03:52.0985221Z Entering 'third_party/pthreadpool' 2025-11-03T18:03:52.1031337Z http.https://github.com/.extraheader 2025-11-03T18:03:52.1077123Z Entering 'third_party/pybind11' 2025-11-03T18:03:52.1127774Z http.https://github.com/.extraheader 2025-11-03T18:03:52.1172346Z Entering 'third_party/python-peachpy' 2025-11-03T18:03:52.1220543Z http.https://github.com/.extraheader 2025-11-03T18:03:52.1264873Z Entering 'third_party/sleef' 2025-11-03T18:03:52.1310390Z http.https://github.com/.extraheader 2025-11-03T18:03:52.1354008Z Entering 'third_party/tensorpipe' 2025-11-03T18:03:52.1400660Z http.https://github.com/.extraheader 2025-11-03T18:03:52.1443404Z Entering 'third_party/tensorpipe/third_party/googletest' 2025-11-03T18:03:52.1491031Z http.https://github.com/.extraheader 2025-11-03T18:03:52.1535237Z Entering 'third_party/tensorpipe/third_party/libnop' 2025-11-03T18:03:52.1580737Z http.https://github.com/.extraheader 2025-11-03T18:03:52.1625550Z Entering 'third_party/tensorpipe/third_party/libuv' 2025-11-03T18:03:52.1670996Z http.https://github.com/.extraheader 2025-11-03T18:03:52.1714925Z Entering 'third_party/tensorpipe/third_party/pybind11' 2025-11-03T18:03:52.1761686Z http.https://github.com/.extraheader 2025-11-03T18:03:52.1802964Z Entering 'third_party/tensorpipe/third_party/pybind11/tools/clang' 2025-11-03T18:03:52.1850959Z http.https://github.com/.extraheader 2025-11-03T18:03:52.2001399Z A job completed hook has been configured by the self-hosted runner administrator 2025-11-03T18:03:52.2031838Z ##[group]Run '/home/ec2-user/runner-scripts/after_job.sh' 2025-11-03T18:03:52.2038865Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2025-11-03T18:03:52.2039151Z ##[endgroup] 2025-11-03T18:03:59.5006514Z Cleaning up orphan processes